JAMA's AI Policy: What You Can Use, What You Must Disclose, and Where the Lines Are
JAMA requires detailed AI disclosure in Methods including tool name, version, and manufacturer, prohibits AI authorship, and applies the same policy across all 14 JAMA Network journals.
Senior Researcher, Oncology & Cell Biology
Author context
Specializes in manuscript preparation and peer review strategy for oncology and cell biology, with deep experience evaluating submissions to Nature Medicine, JCO, Cancer Cell, and Cell-family journals.
Next step
Choose the next useful decision step first.
Use the guide or checklist that matches this page's intent before you ask for a manuscript-level diagnostic.
JAMA was among the fastest major medical journals to formalize its AI policy, and the rules it set have influenced how the entire JAMA Network, 13 specialty journals plus JAMA Network Open, handles AI-assisted manuscripts. If you're targeting any JAMA title, the policy is stricter than what you'll find at most Springer Nature journals and more prescriptive about disclosure format than almost any other publisher.
The core rules
JAMA's AI policy boils down to four requirements:
- AI can't be an author. Large language models and generative AI tools don't meet ICMJE authorship criteria. They can't design studies, interpret data, take public accountability, or approve final manuscripts. JAMA won't accept submissions that list AI tools as co-authors.
- All AI use must be disclosed in Methods. If you used ChatGPT, Claude, Copilot, or any other AI tool during manuscript preparation, including drafting, editing, data analysis, or literature review, you must describe this in the Methods section.
- The disclosure must be specific. JAMA doesn't accept vague statements like "AI tools assisted with writing." You need to name the tool, specify the version, identify the manufacturer, and describe what the tool was used for.
- Authors are fully responsible. Using an AI tool doesn't shift any responsibility. Every author listed on the paper must be able to vouch for the accuracy and integrity of all content, including sections where AI assisted.
How JAMA's policy compares to the AMA publisher-wide stance
JAMA is published by the American Medical Association, and the AMA's position on AI in scholarly publishing aligns closely with JAMA's editorial policy. But there are important distinctions:
AMA's broader position: The AMA has issued organizational statements on AI in healthcare that go beyond publishing, covering clinical AI, algorithmic bias, and AI regulation. These don't directly affect your manuscript submission.
JAMA's editorial specifics: JAMA's policy is more granular than the AMA's general principles. While the AMA endorses transparency in AI use, JAMA specifies exactly where disclosure goes (Methods section), what format it takes (tool name, version, manufacturer, use case), and what consequences follow non-compliance.
JAMA Network consistency: The editorial office applies the same AI policy across all JAMA Network journals. You won't find different rules at JAMA Oncology versus JAMA Cardiology. This is unusual, some publishers let individual journal editors set their own AI standards.
The practical takeaway: if you've read JAMA's policy, you've read the policy for all 14 JAMA Network titles. You don't need to check each journal separately.
Writing the disclosure statement
JAMA expects more detail in the AI disclosure than most journals. Here's what a properly formatted statement looks like:
"During preparation of this manuscript, the authors used ChatGPT (GPT-4, OpenAI, San Francisco, CA) to assist with editing the Discussion section for clarity and conciseness. The authors reviewed and revised all AI-generated suggestions and take full responsibility for the content of the published article."
Key elements JAMA wants:
- Tool name (ChatGPT, Claude, Copilot, etc.)
- Version (GPT-4, Claude 3.5, etc.)
- Manufacturer and location (OpenAI, San Francisco, CA)
- Specific use case (editing the Discussion, improving language, generating code for statistical analysis)
- Author responsibility statement (confirming human oversight and accountability)
A statement that wouldn't pass JAMA's standards:
"AI tools were used during manuscript preparation."
This is too vague. It doesn't tell the reader what tool was used, what it did, or which parts of the manuscript it touched. JAMA's editors will ask you to revise.
Multiple AI tools
If you used more than one tool, disclose each separately:
"The authors used ChatGPT (GPT-4, OpenAI) to improve the readability of the Introduction and Claude (Claude 3.5, Anthropic, San Francisco, CA) to assist with drafting the statistical analysis code in R. All AI outputs were reviewed and verified by the authors, who take full responsibility for the published content."
What counts as AI use requiring disclosure
JAMA's threshold for disclosure is lower than many researchers expect. Here's the breakdown:
Use case | Disclosure required? | Notes |
|---|---|---|
Grammar/spell check (Grammarly, Word) | No | Standard writing tools aren't covered |
Language polishing with ChatGPT | Yes | Even minor edits require disclosure |
Rewriting paragraphs for clarity | Yes | Methods section disclosure |
Generating first drafts | Yes | Risky but permitted with disclosure |
Literature search with AI tools | Yes | Describe scope and limitations |
Statistical code generation | Yes | Specify what code was generated |
AI-assisted figure creation | Yes | Describe what the AI contributed |
AI-generated images (DALL-E, etc.) | Prohibited | Same as most medical journals |
Translation from another language | Yes | Specify source and target language |
Summarizing references | Yes | State which references were AI-summarized |
Data analysis with AI assistance | Yes | Detail what analyses the AI performed |
The dividing line is roughly this: if the tool uses generative AI or large language models, disclose it. If it's a traditional software tool (spell checkers, reference managers, statistical packages), you don't need to.
What happens if you don't disclose
JAMA treats undisclosed AI use as a publication ethics violation, similar in severity to undisclosed conflicts of interest or undisclosed duplicate submission. The consequences escalate based on when the violation is discovered:
During peer review: The manuscript is returned to the authors with a request to add proper AI disclosure. If the omission appears deliberate, the editor may reject the paper outright. JAMA editors have noted in editorials that transparency failures undermine trust in the entire manuscript.
After acceptance but before publication: The production team adds the disclosure to the Methods section with the authors' consent. If authors refuse to disclose, the acceptance may be rescinded.
After publication: This is where consequences get serious. JAMA can issue:
- A correction (if the AI use was minor and didn't affect scientific content)
- An expression of concern (if the scope of undisclosed AI use is unclear)
- A retraction (if AI-generated content is found to contain fabricated data or unverifiable claims)
JAMA also follows ICMJE guidelines on publication ethics investigations. A finding of deliberate non-disclosure could result in the authors being flagged to their institution and to other journals where they serve as authors or reviewers.
The career risk is real. JAMA publishes roughly 11% of submitted manuscripts. Getting your paper through that filter and then having it corrected or retracted for an avoidable disclosure issue would be a significant setback. Over-disclosure costs you nothing. Under-disclosure can cost you the paper.
The JAMA Network scope
JAMA's AI policy applies uniformly across these titles:
Journal | Focus area |
|---|---|
JAMA | General medicine |
JAMA Internal Medicine | Internal medicine |
JAMA Oncology | Cancer |
JAMA Cardiology | Cardiovascular |
JAMA Neurology | Neuroscience/neurology |
JAMA Pediatrics | Pediatrics |
JAMA Surgery | Surgery |
JAMA Psychiatry | Psychiatry |
JAMA Dermatology | Dermatology |
JAMA Ophthalmology | Ophthalmology |
JAMA Otolaryngology | Head and neck surgery |
JAMA Health Forum | Health policy |
JAMA Network Open | All medical disciplines (open access) |
JAMA Network Open deserves special mention. It's the network's only fully open-access journal, with a much higher acceptance rate than JAMA itself (~20% vs ~11%). The AI policy is identical, but the volume of submissions is significantly higher, meaning more AI-assisted manuscripts flow through its editorial system.
Timeline of JAMA's AI policy
Date | Development |
|---|---|
January 2023 | JAMA publishes editorial addressing ChatGPT and authorship |
February 2023 | Formal AI disclosure requirement added to Instructions for Authors |
Mid 2023 | Policy refined to require tool name, version, and manufacturer |
2024 | ICMJE updates authorship guidelines; JAMA aligns immediately |
2025–2026 | Policy stable; enforcement integrated into submission workflow |
JAMA moved fast. The editorial acknowledging AI's impact on medical publishing appeared within weeks of ChatGPT's widespread adoption. The journal didn't wait for ICMJE consensus, it set its own rules first and aligned with ICMJE updates as they came.
How JAMA compares to other top medical journals
Feature | JAMA | NEJM | The Lancet | The BMJ |
|---|---|---|---|---|
AI authorship | Prohibited | Prohibited | Prohibited | Prohibited |
Disclosure location | Methods | Methods + cover letter | Methods | Methods |
Detail required | High (tool, version, manufacturer) | Moderate | Moderate | Moderate |
AI-generated images | Prohibited | Prohibited | Prohibited | Prohibited |
Copy editing exemption | No (disclose all AI use) | No | Limited | Yes (basic tools) |
Scope | All JAMA Network titles | NEJM + NEJM Group | All Lancet titles | BMJ + BMJ specialty journals |
JAMA's disclosure requirements are the most prescriptive of the big four medical journals. Requiring the manufacturer's name and location goes beyond what NEJM or The Lancet ask for. This reflects JAMA's broader editorial philosophy of maximum transparency in methods reporting.
The copy editing distinction matters. Nature exempts basic grammar tools from disclosure. JAMA doesn't draw that line as clearly, if you used ChatGPT to fix grammar, JAMA still wants to know. When in doubt, disclose.
Practical advice for JAMA submissions
Before you write:
- Decide upfront which AI tools you'll use and for what purpose. This makes the disclosure statement easier to write later.
- Keep a log of AI interactions. If you use ChatGPT to edit three paragraphs, note which paragraphs and what prompts you used. You probably won't need this level of detail for the disclosure, but having it protects you if questions arise.
During manuscript preparation:
- Don't use AI to generate clinical data, patient descriptions, or case details. JAMA's editors are experienced enough to spot fabricated clinical scenarios, and AI-generated medical content carries serious accuracy risks.
- If you use AI for statistical code, verify every output independently. JAMA's statistical reviewers will check your methods regardless.
- Draft your disclosure statement as you write, not after. It's easier to remember what you used AI for while you're doing it.
At submission:
- Include the AI disclosure in the Methods section, formatted as described above.
- If you didn't use any AI tools, you don't need a negative disclosure statement. JAMA doesn't require authors to state "no AI tools were used."
- Double-check that all co-authors are aware of the AI disclosure. JAMA requires all authors to approve the final manuscript, and that includes the AI disclosure.
Common mistakes:
- Listing ChatGPT in the Acknowledgments section instead of Methods. JAMA specifically requires Methods placement.
- Disclosing AI use for the abstract but not specifying which sections of the main text were AI-assisted.
- Forgetting to include the version number. "ChatGPT" isn't specific enough, JAMA wants "ChatGPT (GPT-4)" or "ChatGPT (GPT-3.5)."
- Assuming your institution's AI policy overrides JAMA's. If your university allows unrestricted AI use, that doesn't mean JAMA will accept an undisclosed AI-written manuscript.
AI in JAMA's own editorial process
It's worth noting that JAMA has been transparent about exploring AI tools in its own operations. The journal has published research on AI's ability to answer medical exam questions, generate clinical text, and detect errors in manuscripts. JAMA's editors aren't anti-AI, they're pro-transparency.
This matters because some researchers worry that disclosing AI use will bias editors against their manuscript. JAMA's published position suggests the opposite: editors are more concerned about hidden AI use than disclosed use. A paper with a clear, honest AI disclosure is better positioned than one where undisclosed AI use is later discovered.
Bottom line
JAMA's AI policy is clear and applies network-wide: use AI if it helps your work, but disclose everything in Methods with the tool name, version, manufacturer, and specific use case. AI can't be an author. AI-generated images aren't allowed. The disclosure requirements are more detailed than most journals, but that's by design, JAMA wants enough information for readers to assess how AI influenced the final paper. The biggest risk isn't using AI; it's using AI without telling JAMA about it.
Before Submitting to JAMA
A pre-submission manuscript review can catch statistical reporting gaps and scope mismatches before JAMA's 3-day desk triage.
Sources
- JAMA Instructions for Authors
- JAMA editorial: Nonhuman "Authors" and Implications for the Integrity of Scientific Publication and Medical Knowledge
- ICMJE Recommendations: Defining the Role of Authors and Contributors
- JAMA Network Open author guidelines
- AMA policy on augmented intelligence in health care
Reference library
Use the core publishing datasets alongside this guide
This article answers one part of the publishing decision. The reference library covers the recurring questions that usually come next: how selective journals are, how long review takes, and what the submission requirements look like across journals.
Dataset / reference guide
Peer Review Timelines by Journal
Reference-grade journal timeline data that authors, labs, and writing centers can cite when discussing realistic review timing.
Dataset / benchmark
Biomedical Journal Acceptance Rates
A field-organized acceptance-rate guide that works as a neutral benchmark when authors are deciding how selective to target.
Reference table
Journal Submission Specs
A high-utility submission table covering word limits, figure caps, reference limits, and formatting expectations.
Before you upload
Choose the next useful decision step first.
Move from this article into the next decision-support step. The scan works best once the journal and submission plan are clearer.
Use the scan once the manuscript and target journal are concrete enough to evaluate.
Anthropic Privacy Partner. Zero-retention manuscript processing.
Where to go next
Conversion step
Choose the next useful decision step first.
Use the scan once the manuscript and target journal are concrete enough to evaluate.