When Autism Research Conflicts: A Conversation with Dr. Gina Green

April 9, 2026

Key Takeaways

  • The most important distinction in autism treatment research is between prospective controlled studies and retrospective chart reviews. Dr. Gina Green, a former president of the Association for Professional Behavior Analysts and one of the field’s most cited methodologists, says this distinction determines “the confidence we can have about conclusions.”
  • Retrospective studies look backward at existing records not collected to answer a research question. 
  • Retrospective studies are highly susceptible to bias, particularly when researchers have a financial or organizational stake in the outcome. 
  • Outcome measurement methodology matters as much as study design. 
  • No single study, however well-designed, can establish a general truth about autism treatment. 
  • A 2025 meta-analysis by Sigmund Eldevik and colleagues, drawing individual-level data from fifteen prospective controlled studies of early intensive behavioral intervention, found effect sizes of 0.66 for adaptive behavior, 0.87 for intellectual functioning, and 1.36 for autism severity reduction. 
  • Families, insurers, and policymakers can apply a consistent evaluative framework.

When a new study on autism treatment lands, it often arrives with a headline and a conclusion. Fifteen hours of therapy is enough. Thirty hours is better. Intensive intervention works, or it does not.

For parents trying to make decisions about their child’s care, for insurers setting ABA coverage policies, for legislators writing mandates, the task of sorting through these claims can feel impossible. The studies seem to contradict one another. The language is technical, while the stakes are enormous.

Not all research is created equal, and the distinction between study types matters more than most people realize. Understanding a few basic principles, what separates a well-designed study from a poorly designed one, and why some conclusions deserve more confidence than others, can transform a confusing landscape into something navigable.

Dr. Gina Green has spent four decades in the field. She holds a Ph.D. in psychology from Utah State University and an honorary doctorate from Queen’s University Belfast, awarded for her work in autism. She is a former president of the Association for Behavior Analysis International, a Fellow of the American Psychological Association, and co-editor of Behavioral Intervention for Young Children with Autism, a foundational text in the field. In 2000, Psychology Today named her Mental Health Professional of the Year.

When it comes to evaluating autism research, few people have thought longer or harder about what constitutes credible evidence.

“The really important and significant differences between a well-designed study and a poorly designed one,” Green said in an interview with Acuity Media Network, “are between prospective controlled studies and uncontrolled retrospective chart reviews.” That distinction, she argues, determines “the confidence we can have about conclusions.”

Prospective vs. Retrospective ABA Studies: Why the Distinction Determines What You Can Conclude

A retrospective study looks backward. Researchers gather existing records, billing data, clinical notes, assessment scores, and search for patterns. A prospective study looks forward. Researchers define their questions in advance, carefully specify their methods, and collect data as the study unfolds.

The distinction is not merely procedural. A retrospective chart review uses data “not collected to address a research question,” Green explains. The records might have been gathered for insurance billing or regulatory compliance. They were never designed to answer the question the researchers are now asking of them.

Consider a physician trying to determine whether a medication is effective. She could look back at patient records and observe that people who took the medication saw their blood pressure improve. But those records would not tell her whether the improvement came from the medication, from dietary changes the patients made, from other life events, or from some combination of factors. “You can’t be sure,” Green says, “whether the outcome was due to the medication, was it due to lifestyle changes, to some other aspect of their health, or some other life events.”

A prospective study, by contrast, can isolate these variables. The researcher specifies in advance which patients will receive the medication and which will serve as a comparison group. She defines exactly how outcomes will be measured. She controls for confounding factors, age, weight, pre-existing conditions, that might otherwise muddy the results.

That said, retrospective studies do have their uses. They can do a great job with identifying associations between variables and patterns worth investigating further. But they cannot establish that one thing caused another. “At most,” Green says, “that kind of look back at data might identify some variables that are associated or correlated.” The leap from correlation to causation requires a different kind of study altogether.

How Bias Undermines Retrospective Autism Research

Retrospective studies face another challenge: they are, in Green’s words, “highly susceptible to bias.” The researchers examining existing records bring their own expectations, conscious or not, to the data. Without the guardrails of a prospective design, predetermined hypotheses, blinded evaluators, pre-registered analysis plans, the potential for selective interpretation grows.

The issue becomes particularly fraught when the researchers have a stake in the outcome. Most academic journals require authors to disclose financial conflicts of interest. But disclosure does not eliminate the problem. When a study uses data from providers within a network that the researchers manage, the question of objectivity becomes harder to set aside.

Some researchers attempt to address this by having independent professionals analyze the same data and compare their conclusions. “That doesn’t guarantee either of them are accurate,” Green notes, “but at least puts a separate set of eyes on the data.” The absence of such safeguards is worth noticing. 

Why Outcome Measurement Methodology Matters in Autism Treatment Research

Even well-designed studies can be undermined by weak measurement. In autism research, the Vineland Adaptive Behavior Scale is a common tool for assessing everyday skills, self-care, communication, and daily living. It is widely used and has acceptable evidence of validity. But it also has limitations that are often overlooked.

The Vineland is what researchers call an indirect measure. It relies on a parent, caregiver, or therapist reporting their impressions of a child’s abilities. No one directly observes or tests the child’s skills in a standardized way. The assessment captures what someone remembers and perceives, not necessarily what the child can do.

Studies have found that parent and therapist Vineland scores for the same child often differ. “Not surprising,” Green says. “Parents have different priorities. They interact with their children in different environments, in different situations than does the teacher at school or the interventionist who’s working with them in a clinic.” These discrepancies do not invalidate the Vineland, but they should temper the confidence placed in any single score.

That’s not to mention that the third edition of the Vineland has prompted additional concerns. At least one research team has raised questions about its applicability to individuals with autism specifically. “You can’t assume,” Green says, “that everything that applied to the second edition of an instrument like the Vineland is also true of the third edition, because the third edition has been changed.”

Why Single ABA Studies Cannot Establish General Truths: The Case for Meta-Analyses

Even a well-designed prospective study with rigorous measurement has limits. It tells us what happened with a particular sample of participants under particular conditions. It does not, by itself, tell us what will happen for other children in other settings.

“One study tells us this is what was reported to have happened with this particular sample of participants under these particular conditions,” Green explains. “We need metaanalyses and reviews that look at the entire body of research, the many replications done by independent investigators in different locations, different countries even, and what that body of literature as a whole tells us.”

Meta-analyses aggregate data from multiple studies to identify patterns that hold across different samples and settings. The strongest meta-analyses go further still, using individual participant data rather than just group averages. This approach is particularly valuable for autism research, given the well-documented heterogeneity of the population. People with autism are affected in vastly different ways; group averages can obscure important variations in how individuals respond to treatment.

A meta-analysis published this year in Autism Research by Sigmund Eldevik and colleagues at Oslo Metropolitan University exemplifies this approach. The researchers gathered individual-level data from fifteen controlled prospective studies of early intensive behavioral intervention, encompassing 341 children who received treatment and 280 in comparison groups. Rather than relying solely on group averages, they examined how each child responded, allowing for a more granular analysis of what predicts outcomes.

Treatment intensity, the number of weekly therapy hours, significantly predicted outcomes across all measures.

The results were substantial: effect sizes of 0.66 for improvement in adaptive behavior, 0.87 for intellectual functioning, and 1.36 for reductions in autism severity. But perhaps more important than any single finding is what the study represents: a synthesis of the best available evidence, drawn from multiple countries and research teams, analyzed at the level of individual children.

This is the kind of evidence that informs what researchers call “generally accepted standards of care,” the benchmarks that professional organizations develop based on the best available science. At any given moment, the evidence will be imperfect and incomplete. But some evidence is stronger than others, and the hierarchy matters.

A Practical Framework for Evaluating ABA Research Quality

Readers encountering a new study can ask a few basic questions. Was this study prospective or retrospective? If retrospective, does it claim to establish causation, or does it appropriately limit itself to identifying associations? Were the researchers independent, or did they have financial ties to the outcomes? What safeguards were in place to reduce bias?

How were outcomes measured? Were they direct assessments or indirect reports? If indirect, whose perspective do they capture, and what might be missing? Does the study acknowledge the limitations of its measures?

And perhaps most importantly: is this a single study, or does it fit within a larger body of research? A single study, however well-designed, cannot establish a general truth. It takes replication across different contexts, independent investigators, and rigorous synthesis to build confidence in a conclusion.

“When conducting a research study, we’re careful not to extrapolate beyond the findings of the particular conditions of our particular study,” Green says. The same caution should apply to anyone reading research. The difference between what one study found and what the evidence shows is the difference between anecdote and knowledge.

Autism, as Green notes, “has always been kind of mysterious.” The uncertainties surrounding its causes and variations have created space for all manner of claims. In such an environment, the tools of evaluation matter more, not less. Knowing how to read research is not a guarantee against confusion, but it is certainly a start.

Ethan Webb is a staff writer at Acuity Media Network, where he covers the business of autism and behavioral health care. His reporting examines how financial pressures, policy changes, and market consolidation shape the ABA industry — and what that means for providers and families. Ethan holds a BFA in Creative Writing from Emerson College and brings more than seven years of professional writing and editing experience spanning healthcare, finance, and business journalism. He has served as Managing Editor of Dental Lifestyles Magazine and has ghostwritten multiple titles that reached the USA Today and Wall Street Journal bestseller lists.