Role Definition
| Field | Value |
|---|---|
| Job Title | Historian |
| Seniority Level | Mid-Level |
| Primary Function | Researches, analyzes, interprets, and presents the past by systematically collecting information from primary and secondary sources. Conducts archival research, synthesizes historical evidence into publications, reports, and presentations. Works in government agencies, museums, cultural heritage organizations, consulting firms, or academia. Splits time between archival research (25%), interpretation and synthesis (25%), writing and publication (20%), data management (10%), teaching/public education (10%), and advisory/methodology work (10%). |
| What This Role Is NOT | NOT an archivist (25-4011 — manages record collections, not historical interpretation). NOT an anthropologist/archeologist (19-3091 — field-based excavation and ethnographic research). NOT a museum curator (25-4012 — collections management focus). NOT a postsecondary history teacher (25-1125 — primarily teaching, scored separately). This is SOC 19-3093 — the research and analysis historian. |
| Typical Experience | 5-10 years. Master's degree required for most positions; PhD required for academic and senior government roles. Specialization in a historical period, region, or methodology (e.g., public history, digital humanities, military history). |
Seniority note: Entry-level historians (0-2 years) performing routine archival cataloging and research assistance would score Red — more data processing, less interpretation. Senior/principal historians (10+ years) directing research programs, advising policy, and serving as expert witnesses would score upper Yellow or borderline Green — more goal-setting, accountability, and advisory work.
Protective Principles + AI Growth Correlation
| Principle | Score (0-3) | Rationale |
|---|---|---|
| Embodied Physicality | 0 | Fully desk-based and digital. Archival visits are structured, predictable settings — reading rooms, digital repositories. No unstructured physical environments. |
| Deep Interpersonal Connection | 1 | Some interpersonal component — oral history interviews, stakeholder consultation with descendant communities, public presentations. But most work is solitary research and writing. Trust matters for oral history but is not the core value delivery. |
| Goal-Setting & Moral Judgment | 2 | Formulates research questions, selects interpretive frameworks, makes ethical decisions about historical representation, and exercises professional judgment about source reliability and historical significance. Significant interpretation within scholarly frameworks, though constrained by evidentiary standards rather than setting organizational direction. |
| Protective Total | 3/9 | |
| AI Growth Correlation | 0 | Demand driven by government historic preservation mandates (NHPA Section 106/110), museum programming, academic positions, and public history projects — independent of AI adoption rates. AI is a tool within the role, not a demand driver. |
Quick screen result: Protective 3 + Correlation 0 — likely Yellow. Modest judgment protection but no physical barriers and limited interpersonal protection. Proceed to quantify.
Task Decomposition (Agentic AI Scoring)
| Task | Time % | Score (1-5) | Weighted | Aug/Disp | Rationale |
|---|---|---|---|---|---|
| Archival research and primary source analysis | 25% | 3 | 0.75 | AUGMENTATION | AI-powered search engines index vast digital archives, HTR (Handwritten Text Recognition) deciphers manuscripts, NLP extracts named entities and dates from historical texts at scale. Human still selects sources, evaluates provenance, and interprets context — but AI handles the discovery and extraction layer. |
| Historical interpretation and synthesis | 25% | 2 | 0.50 | AUGMENTATION | Developing historical arguments, contextualizing events within broader narratives, and constructing interpretive frameworks. AI can summarize and identify patterns but cannot originate novel historical arguments or evaluate competing historiographic perspectives. Human judgment core. |
| Report writing and publication | 20% | 4 | 0.80 | DISPLACEMENT | Drafting historical reports, compliance documentation (Section 106/110 reviews), exhibit text, and publication manuscripts. AI agents generate first drafts, structure arguments, and format citations with minimal oversight. Academic peer-reviewed writing still human-led but heavily AI-accelerated. Government/consulting reports increasingly AI-generated with human review. |
| Data collection and database management | 10% | 4 | 0.40 | DISPLACEMENT | Building and maintaining historical databases, cataloging sources, managing digital collections, entering metadata. AI handles data entry, OCR processing, metadata tagging, and cross-referencing at scale. Routine data management is near-fully automatable. |
| Teaching and public education | 10% | 1 | 0.10 | NOT INVOLVED | Leading tours, delivering lectures, conducting educational programs, oral history interviews. Requires human presence, pedagogical judgment, and interpersonal engagement. Irreducible human task — trust and connection IS the value. |
| Stakeholder consultation and advisory work | 5% | 2 | 0.10 | NOT INVOLVED | Advising government agencies on historic preservation, consulting with descendant communities, serving as expert witness, policy recommendations. Requires professional judgment, cultural sensitivity, and institutional credibility. |
| Research design and methodology | 5% | 2 | 0.10 | AUGMENTATION | Designing research projects, selecting methodological approaches, formulating hypotheses about historical events and processes. AI assists with literature review but cannot originate research questions grounded in historiographic tradition. |
| Total | 100% | 2.75 |
Task Resistance Score: 6.00 - 2.75 = 3.25/5.0
Displacement/Augmentation split: 30% displacement, 55% augmentation, 15% not involved.
Reinstatement check (Acemoglu): Yes — AI creates new tasks: validating AI-generated historical summaries for accuracy (hallucination detection is critical in history where fabricated citations are unacceptable), training and curating domain-specific NLP models for historical text analysis, managing digital humanities projects, and interpreting AI-discovered patterns in large archival datasets. The role is transforming from primary source discoverer to AI-output validator and interpretive specialist.
Evidence Score
| Dimension | Score (-2 to 2) | Evidence |
|---|---|---|
| Job Posting Trends | -1 | BLS projects -2% decline 2022-2032 for historians (SOC 19-3093) — slower than average. Only 3,400 employed with ~300 annual openings, mostly replacements. Academic history positions declining alongside broader humanities contraction. Public history and CRM consulting postings stable but small market. |
| Company Actions | -1 | No major AI-specific layoffs but structural decline. University humanities departments shrinking — history PhD programs reporting fewer tenure-track openings. Government agencies (NPS, USACE, state historic preservation offices) maintaining positions but not expanding. Cultural heritage consulting firms adopting AI tools to reduce labor hours per project. No named mass layoffs but the profession is contracting organically. |
| Wage Trends | 0 | Median $74,050 (BLS 2024). Government historians earn above median; academic positions vary widely. Wages tracking inflation — no real-terms decline or growth. No AI-driven premium emerging yet. Digital humanities specialists may command modest premium but data is sparse. |
| AI Tool Maturity | -1 | Production tools performing core research and writing tasks: NLP/NER tools extract entities from historical texts at scale (spaCy, NLTK, Transkribus for HTR), GPT-4/Claude generate first-draft reports and summaries, topic modeling identifies themes across document collections, AI-powered archival search engines (ArchivesSpace AI, digital repository tools) accelerate discovery. Tools augment 55% and displace 30% of task time. Not yet eliminating positions but compressing person-hours per project. |
| Expert Consensus | 0 | Mixed. American Historical Association acknowledges AI as transformative tool requiring adaptation. Digital humanities community optimistic about AI augmentation. No broad consensus on displacement — most see AI as productivity enhancer for existing historians rather than role eliminator. However, the tiny size of the profession (3,400) means even modest productivity gains reduce headcount pressure. |
| Total | -3 |
Barrier Assessment
Reframed question: What prevents AI execution even when programmatically possible?
| Barrier | Score (0-2) | Rationale |
|---|---|---|
| Regulatory/Licensing | 0 | No professional license required. Section 106/110 reviews require "qualified professionals" meeting Secretary of Interior standards (36 CFR 61) but this is a qualification standard, not a statutory license. No PE-style bar to entry. |
| Physical Presence | 0 | Fully remote/digital possible for most work. Archival visits are structured (reading rooms, digital collections). No unstructured physical environments. Oral history interviews require presence but are a small share of work. |
| Union/Collective Bargaining | 0 | Federal historians (NPS, USACE) have some civil service protections via AFGE but union representation is weak across the profession. Academic positions sometimes unionized (AAUP) but most historians are in non-unionized settings. Minimal friction against headcount reduction. |
| Liability/Accountability | 1 | Moderate stakes. Section 106 compliance errors can delay or halt federal construction projects. Expert witness testimony carries professional credibility consequences. Misrepresentation of historical evidence in government reports has legal implications. But no one goes to prison for a bad historical analysis — consequences are reputational and procedural, not criminal. |
| Cultural/Ethical | 1 | Some cultural resistance to AI-generated historical narratives — concerns about hallucination, bias in historical AI models (training data reflects existing historiographic biases), and the importance of human judgment in representing sensitive historical events (slavery, genocide, indigenous history). Professional norms emphasize evidence-based interpretation by trained historians, but cultural barrier is moderate, not strong. Society does not have the same visceral resistance to AI-generated history as to AI-generated medical diagnoses or legal judgments. |
| Total | 2/10 |
AI Growth Correlation Check
Confirmed at 0 (neutral). Demand for historians is driven by historic preservation mandates (Section 106/110 triggered by federal construction projects), museum programming, academic positions, and public history initiatives — none of which correlate with AI adoption rates. AI is a tool within the role (NLP for text analysis, GPT for drafting), not a driver of demand for it. The profession's small size (3,400) means AI-driven productivity gains could reduce headcount without requiring formal displacement — fewer historians doing more work per person.
JobZone Composite Score (AIJRI)
| Input | Value |
|---|---|
| Task Resistance Score | 3.25/5.0 |
| Evidence Modifier | 1.0 + (-3 x 0.04) = 0.88 |
| Barrier Modifier | 1.0 + (2 x 0.02) = 1.04 |
| Growth Modifier | 1.0 + (0 x 0.05) = 1.00 |
Raw: 3.25 x 0.88 x 1.04 x 1.00 = 2.9744
JobZone Score: (2.9744 - 0.54) / 7.93 x 100 = 30.7/100
Zone: YELLOW (Green >=48, Yellow 25-47, Red <25)
Sub-Label Determination
| Metric | Value |
|---|---|
| % of task time scoring 3+ | 55% |
| AI Growth Correlation | 0 |
| Sub-label | Yellow (Urgent) — AIJRI 25-47 AND >=40% of task time scores 3+ |
Assessor override: None — formula score accepted. The 30.7 sits mid-Yellow, 17.3 points below the Green boundary. Comparable to Economist (31.6), which shares the same social science research profile with heavy writing and data analysis exposure. The weak barrier score (2/10) is the key differentiator from Anthropologist/Archeologist (39.4, barriers 7/10) — historians lack the physical fieldwork protection, NAGPRA/tribal sovereignty cultural barriers, and regulatory licensing that protect archeologists. Without even the modest 2/10 barriers, the score would drop to 29.5.
Assessor Commentary
Score vs Reality Check
The Yellow (Urgent) label is honest. Historians face a compounding challenge: the profession is already tiny (3,400 workers), the academic job market for humanities PhDs has been contracting for over a decade, and AI now automates the research-discovery and report-writing layers that historically required significant person-hours. The 30.7 score sits comfortably in Yellow — not Red because interpretive and advisory work provides meaningful task resistance (3.25), but not Green because evidence is negative, barriers are nearly absent, and the 55% of task time at score 3+ represents substantial automation exposure. The score is well-calibrated against Economist (31.6) and Librarian (33.2) — similar knowledge-worker profiles with heavy data/writing components and weak structural barriers.
What the Numbers Don't Capture
- Micro-profession vulnerability — With only 3,400 workers, even small AI-driven productivity gains (10-20% fewer person-hours per project) could eliminate hundreds of positions without formal layoffs. The profession is too small for meaningful statistical tracking of AI displacement.
- Academic humanities contraction — History PhD programs are shrinking for structural reasons (declining enrollment, budget cuts, adjunctification) independent of AI. AI accelerates an existing decline rather than causing a new one.
- Digital humanities bifurcation — Historians who adopt computational methods (NLP, GIS, data visualization) are creating a distinct subspecialty with different demand dynamics. Traditional archival historians and digital humanities historians may diverge in AI resistance.
- Function-spending vs people-spending — Government agencies and museums investing in digital collections and AI-powered search tools may maintain or increase their history-related spending while reducing historian headcount. More history output, fewer historians.
Who Should Worry (and Who Shouldn't)
If you are a public historian or government historian working in historic preservation (Section 106/110 compliance), museum interpretation, or cultural heritage advisory roles — where you spend most of your time consulting with stakeholders, presenting to the public, and exercising professional judgment about historic significance — you are more secure than the 30.7 suggests. Your interpersonal, advisory, and regulatory compliance work resists automation.
If you are a research historian whose primary output is written reports, literature syntheses, and archival data compilation — particularly in CRM consulting or academic research assistance — you are more at risk. AI agents can search archives, extract entities from historical texts, synthesize secondary sources, and draft standardized reports with minimal human oversight. The historian who primarily discovers and compiles information is on a converging trajectory with AI capabilities.
The single biggest factor separating the safe version from the at-risk version is interpretive originality. Historians who generate novel arguments, challenge existing narratives, and exercise judgment about contested historical questions are doing work AI cannot originate. Historians who primarily compile, summarize, and report known historical facts are doing work AI already does competently.
What This Means
The role in 2028: The surviving historian uses AI to search archives in minutes instead of months, processes thousands of historical documents through NLP pipelines, and generates first-draft reports and compliance documentation with AI agents. But the core of the role — developing original historical arguments, interpreting contested evidence, advising on historic preservation policy, and communicating history to the public — remains human. The profession will be smaller, more productive per capita, and more concentrated in interpretive, advisory, and public-facing roles. Pure research compilation positions will contract.
Survival strategy:
- Shift toward interpretive and advisory work — Build expertise in historic preservation consulting, expert witness testimony, policy advisory roles, and public history programming where human judgment and professional credibility are non-negotiable. Move away from pure research compilation.
- Master digital humanities and AI tools — Become proficient with NLP for historical text analysis (spaCy, Transkribus), GIS for spatial history, topic modeling, and AI-powered archival search. The historian who directs and validates AI outputs commands a premium over the historian who does manually what AI does faster.
- Develop a public-facing specialization — Oral history, documentary consulting, museum interpretation, educational programming, or media commentary. These interpersonal, trust-based activities are irreducible human tasks that AI cannot perform.
Where to look next. If you are considering a career shift, these Green Zone roles share transferable skills with historians:
- Museum Technician and Conservator (Mid-Level) (AIJRI 49.8) — your archival, collections, and cultural heritage knowledge transfers directly; physical artifact handling and conservation judgment resist automation
- Education Administrator, K-12 (Mid-to-Senior) (AIJRI 59.9) — research design, curriculum knowledge, and institutional leadership skills transfer; strong interpersonal and goal-setting protection
- Social and Community Service Manager (Mid-to-Senior) (AIJRI 48.9) — stakeholder engagement, program management, and community advisory skills transfer; interpersonal and judgment protection
Browse all scored roles at jobzonerisk.com to find the right fit for your skills and interests.
Timeline: 3-5 years for significant transformation. NLP, HTR, and generative AI tools are already production-grade for historical text analysis and report generation. The small size of the profession (3,400) means productivity gains compress headcount quickly. Interpretive and advisory work provides the longer runway.