JCI's AI Policy: What Authors Need to Know About AI Disclosure at the Journal of Clinical Investigation
The JCI requires AI disclosure in Methods, prohibits AI authorship, and sets its own editorial policy through the ASCI rather than inheriting rules from a commercial publisher.
Senior Researcher, Oncology & Cell Biology
Author context
Specializes in manuscript preparation and peer review strategy for oncology and cell biology, with deep experience evaluating submissions to Nature Medicine, JCO, Cancer Cell, and Cell-family journals.
Next step
Choose the next useful decision step first.
Use the guide or checklist that matches this page's intent before you ask for a manuscript-level diagnostic.
The Journal of Clinical Investigation holds a unique position in biomedical publishing. It's not owned by a commercial publisher, it's published by the American Society for Clinical Investigation, a membership society of physician-scientists. This means the JCI doesn't inherit its AI policy from a Springer Nature, Elsevier, or Wiley corporate directive. The editorial board sets its own rules, and those rules reflect how physician-scientists think about transparency in translational research.
The JCI's AI policy
The JCI's AI policy aligns with ICMJE guidelines while adding journal-specific expectations:
- AI can't be an author. Consistent with ICMJE criteria, AI tools don't qualify for authorship because they can't take accountability, approve manuscripts, or be responsible for the published work.
- AI use must be disclosed in Methods. If you used any generative AI tool during manuscript preparation, including language editing, code generation, literature review, or data visualization, describe it in the Methods section.
- AI-generated images are prohibited. No figures or visual content produced by generative AI tools.
- Authors are responsible for all content. Every listed author must vouch for the accuracy and integrity of the manuscript, including any AI-assisted sections.
- The disclosure should be specific. Name the tool, the version, and what it was used for. Vague statements aren't sufficient.
How the JCI's policy differs from publisher-driven journals
Most journals covered in AI policy guides inherit their rules from a parent publisher. Nature Medicine follows Springer Nature. Cancer Cell follows Cell Press. Gut follows BMJ Publishing Group. The JCI is different, it's an ASCI journal with editorial independence.
What this means in practice:
No publisher-wide cascade. When Springer Nature updates its AI policy, all 3,000+ journals follow automatically. When the JCI updates its policy, it's a decision by the JCI's editorial board. This makes changes slower but more deliberate.
ICMJE alignment is the foundation. The JCI has long followed ICMJE recommendations closely. Its AI policy builds directly on the ICMJE's position on AI and authorship, rather than being filtered through a corporate publishing policy.
The JCI Insight connection. JCI Insight is the JCI's open-access companion journal. Both journals follow the same AI policy set by the ASCI editorial board. If you've read the JCI's rules, you've read JCI Insight's rules.
Aspect | JCI / JCI Insight | Nature Medicine | JAMA |
|---|---|---|---|
Policy source | ASCI editorial board | Springer Nature corporate | AMA editorial office |
ICMJE alignment | Direct | Through publisher | Through publisher |
Number of journals covered | 2 (JCI + JCI Insight) | 3,000+ | 14 (JAMA Network) |
Policy update process | Editorial board decision | Corporate policy update | Editorial office decision |
Companion journal | JCI Insight | Nature Communications (different scope) | JAMA Network Open |
The translational research context
The JCI publishes research that bridges basic science and clinical medicine, the "bench to bedside" work that physician-scientists specialize in. This creates specific AI considerations:
Patient data in translational studies
JCI papers frequently include human subject data: patient cohorts, biomarker analyses, clinical correlations, genetic studies. The same data privacy rules that apply at clinical journals apply here: don't input patient data into cloud-based AI tools.
For translational papers that include both mouse model data and human patient data, the AI disclosure should address both:
"No human participant data or animal experimental records were processed through cloud-based AI tools. ChatGPT (GPT-4, OpenAI) was used to improve the language of the Discussion section. All AI-suggested text was reviewed by the clinical investigators (J.R. and S.M.)."
Mechanistic claims with clinical implications
The JCI's editorial identity centers on mechanistic insight with disease relevance. Papers don't just describe phenomena, they connect molecular mechanisms to clinical conditions. If AI tools helped write sections that make these connections, the disclosure should reflect this, and the claims themselves should be entirely human-generated.
AI-polished prose about a signaling pathway is fine if the scientific content is accurate. AI-generated claims about how that pathway connects to a disease mechanism aren't acceptable without careful human verification.
Multi-omics and systems biology
Many JCI papers integrate genomics, transcriptomics, proteomics, and metabolomics data. If AI helped with analysis code for any of these, disclose it:
"The authors used GitHub Copilot (Microsoft) to assist with writing R scripts for the RNA-seq differential expression analysis and the GSEA pathway enrichment analysis. All scripts were validated against published tutorial datasets and manually verified by the bioinformatics team (L.K.). The multi-omics integration pipeline was developed without AI assistance."
Writing the disclosure for the JCI
For a translational research paper:
"During preparation of this manuscript, the authors used ChatGPT (GPT-4, OpenAI) to improve the readability of the Introduction and Discussion sections. No AI tools were used for data analysis, clinical interpretation, or mechanistic conclusions. All AI-suggested text was reviewed by the corresponding author (A.B.) and verified against the experimental data. The authors take full responsibility for the published content."
For a clinical investigation paper:
"The authors used Claude (Claude 3.5, Anthropic) to edit the Methods section for language clarity. No patient data was processed through cloud-based AI tools. The statistical analysis was performed using SAS 9.4 by the study biostatistician (C.D.). Clinical outcome assessments and their interpretations are entirely the work of the investigator team. The authors take full responsibility for the content."
For a basic science paper with disease relevance:
"During manuscript preparation, ChatGPT (GPT-4, OpenAI) was used to assist with editing the Results section. GitHub Copilot (Microsoft) helped write Python scripts for the ChIP-seq peak calling analysis. All code was validated against ENCODE reference datasets. The disease mechanism described in the Discussion is based on the experimental data and was formulated by the authors without AI assistance."
What requires disclosure at the JCI
Use case | Disclosure required? | Notes |
|---|---|---|
Grammar/spell check | No | Standard tools exempt |
ChatGPT for language editing | Yes | Methods section |
AI for analysis code | Yes | Specify which analyses; confirm validation |
AI for figure creation | Prohibited if generative | Data-derived plots are fine |
Translation of manuscript | Yes | Name tool and languages |
AI for literature search | Yes | Describe scope |
AI for statistical code | Yes | Confirm biostatistician validation |
AI for clinical data interpretation | Shouldn't be used | Human investigators only |
AI for cover letter | Not required | Cover letter isn't published |
AI for grant-to-paper text adaptation | Yes | Generating new manuscript text |
Consequences of non-disclosure
The JCI follows COPE and ICMJE guidelines for handling publication ethics issues:
During review:
- Editor contacts corresponding author
- Disclosure must be added to Methods
- Deliberate concealment may result in rejection
- If AI use affected clinical or mechanistic claims, additional review may be required
After publication:
- Correction for minor undisclosed language editing
- Expression of concern if scope is unclear or if it affects scientific content
- Retraction for fabricated data or false mechanistic/clinical claims
- Institutional notification for serious cases
The physician-scientist community: The ASCI has approximately 3,400 active members, and many serve as JCI reviewers and editorial board members. The community is small enough that publication ethics issues become known quickly. A correction at the JCI affects not just the paper but the author's standing within the ASCI membership, a community that elects its members based on scientific achievement and integrity.
How the JCI's policy compares to its publisher-wide context
The JCI doesn't have a "publisher-wide" policy in the way Nature Medicine has Springer Nature's policy or Cancer Cell has Cell Press's. The ASCI publishes two journals, the JCI and JCI Insight, and the editorial board sets the AI policy directly.
Aspect | JCI approach | Publisher-driven journals |
|---|---|---|
Policy source | ASCI editorial board | Corporate publisher |
Number of journals covered | 2 | Hundreds to thousands |
Update mechanism | Board decision | Corporate policy update |
ICMJE alignment | Direct, foundational | Filtered through publisher |
Journal-specific flexibility | Maximum | Limited by publisher rules |
This independence has advantages: the JCI can update its AI policy faster than a 3,000-journal publisher can cascade changes. It also means the policy reflects physician-scientists' priorities rather than a publishing company's legal team's preferences.
Comparison with other translational medicine journals
Feature | JCI | Nature Medicine | Science Translational Medicine | Journal of Experimental Medicine | PNAS |
|---|---|---|---|---|---|
Publisher | ASCI | Springer Nature | AAAS | Rockefeller UP | NAS |
AI authorship | Prohibited | Prohibited | Prohibited | Prohibited | Prohibited |
Disclosure location | Methods | Methods | Methods/Acknowledgments | Methods | Methods + Author Contributions |
AI image ban | Yes | Yes | Yes | Yes | Yes |
Publisher-wide policy | No (journal-specific) | Yes (3,000+ journals) | Yes (AAAS journals) | No (journal-specific) | No (NAS-specific) |
Translational focus | Primary | Significant | Primary | Significant | Mixed |
Science Translational Medicine (AAAS) follows the same AI policy as Science. PNAS has its own policy set by the National Academy of Sciences, which requires AI disclosure in both Methods and Author Contributions. The JCI's approach, Methods-only disclosure with ICMJE alignment, is the most straightforward of the group.
Practical advice for JCI submissions
For translational papers:
- Clearly separate AI writing tools from research computational tools
- If your paper connects a molecular mechanism to a disease, the mechanistic reasoning must be entirely human-generated
- Disclose AI use in the cover letter as a courtesy, even though the JCI primarily requires Methods disclosure
For papers with patient cohorts:
- Don't process patient data through cloud AI
- If AI helped with analysis code for clinical data, confirm that the code was validated by the study biostatistician
- Clinical phenotyping and outcome assessments should be described by the clinical team, not AI
For JCI Insight submissions:
- Same policy as the JCI. Don't look for different rules.
- JCI Insight is open access, so your disclosure is publicly visible to all readers
Before submission checklist:
- [ ] AI disclosure in Methods with tool name, version, use case
- [ ] No patient data processed through cloud AI tools
- [ ] Clinical and mechanistic interpretations are human-generated
- [ ] Analysis code validated independently
- [ ] All co-authors aware of AI disclosure
- [ ] ICMJE authorship criteria confirmed for all listed authors
A free manuscript assessment can help verify your JCI submission meets the journal's editorial and ethical requirements before you submit.
Sources
Reference library
Use the core publishing datasets alongside this guide
This article answers one part of the publishing decision. The reference library covers the recurring questions that usually come next: how selective journals are, how long review takes, and what the submission requirements look like across journals.
Dataset / reference guide
Peer Review Timelines by Journal
Reference-grade journal timeline data that authors, labs, and writing centers can cite when discussing realistic review timing.
Dataset / benchmark
Biomedical Journal Acceptance Rates
A field-organized acceptance-rate guide that works as a neutral benchmark when authors are deciding how selective to target.
Reference table
Journal Submission Specs
A high-utility submission table covering word limits, figure caps, reference limits, and formatting expectations.
Before you upload
Choose the next useful decision step first.
Move from this article into the next decision-support step. The scan works best once the journal and submission plan are clearer.
Use the scan once the manuscript and target journal are concrete enough to evaluate.
Anthropic Privacy Partner. Zero-retention manuscript processing.
Where to go next
Conversion step
Choose the next useful decision step first.
Use the scan once the manuscript and target journal are concrete enough to evaluate.