The 30-second answer
Data science is medium risk on average — but the within-role spread is the widest of any knowledge profession we track. If you spend most of your day training standard models, running EDA notebooks, generating reports, and engineering features from defined schemas, your AEI score is likely between 65 and 78. If you spend most of it on problem framing, research design, and translating business questions into measurable hypotheses, you're between 18 and 30. Both profiles carry the same "data scientist" title in the hiring market.
The automation question — will AI replace data scientists? — is the wrong frame. The right question is: which layer of the data science stack does AI now own, and which layer requires irreplaceable human judgment? That's what the AEI task-level decomposition measures, and it's what determines where you sit in the risk distribution.
AutoML is real — and already running in your organization
AutoML platforms — Google Vertex AI AutoML, AWS SageMaker Autopilot, DataRobot, H2O.ai — have crossed the capability threshold where they can match or exceed hand-tuned models on standard tabular tasks with minimal configuration. Enterprise adoption of AutoML grew 94% year-over-year in 2025, driven by platform teams embedding it into data infrastructure. Claude, GPT-4o, and Gemini can now write production-quality EDA pipelines, feature engineering code, and statistical summaries from schema descriptions alone.
For data scientists whose primary output is model files and standardized reports, this is not a future concern — it is the current competitive baseline. The Eloundou et al. study published in Science (2024) rated data science and analytics occupations at approximately 94% theoretical AI task coverage for execution-layer tasks across 19,265 occupational tasks. That ceiling has not been reached in practice — but the gap is closing.
What the numbers actually mean for data scientists in 2026
Theoretical coverage and observed automation diverge because of organizational friction: data governance constraints, lack of labeled ground-truth for novel problems, trust gaps in AI-generated model explanations, and the fundamental difficulty of specifying the right problem to solve. The Anthropic Economic Index (March 2026) shows 36% observed automation for data and analytics roles — meaningful, but less than half of the theoretical ceiling.
The 36% number is rising steadily. Data scientists who primarily execute known problems against defined data sets face a narrowing window. Those who own the upstream question of what to measure and why are in the most durable position in the field.
Execution vs research: where the 48-point gap lives
The AEI framework identifies Human Alpha Calibration (HAC) — tasks where human judgment produces outcomes AI cannot replicate at equivalent quality. For data scientists, HAC tasks cluster at the top of the problem-solving stack:
- Problem framing — deciding what question is worth answering
- Research design — choosing methodologies under uncertainty and causal ambiguity
- Business translation — converting stakeholder intuitions into measurable hypotheses
- Experiment strategy — designing experiments that isolate causal signal from noise
These tasks score 18–25% on the TLD automation scale. AI is good at answering defined questions; it is poor at recognizing which questions are worth asking. That asymmetry is where the research-lead data scientist's durable advantage lives.
Task-level breakdown for data scientists
Below is the per-task AEI scoring for the nine most-cited data science tasks. Weight each by the share of your working week it consumes to estimate your personal AEI.
| Task | AI Score | Verdict |
|---|---|---|
| Report generation & dashboarding | 75% | High Risk |
| AutoML & standard model training | 72% | High Risk |
| Exploratory data analysis (EDA) | 68% | High Risk |
| Feature engineering (defined schema) | 65% | Medium |
| Model monitoring & drift detection | 58% | Medium |
| Experiment strategy design | 25% | Low Risk |
| Research design & methodology | 22% | Low Risk |
| Business question translation | 20% | Low Risk |
| Problem framing | 18% | Low Risk |