Pre-Submission Review for Ecology and Evolution Papers: What Nature Ecology & Evolution Reviewers Expect
Ecology and evolution manuscripts need robust field data, proper sampling design, and conclusions that scale appropriately from the study system to broader principles.
Senior Researcher, Oncology & Cell Biology
Author context
Specializes in manuscript preparation and peer review strategy for oncology and cell biology, with deep experience evaluating submissions to Nature Medicine, JCO, Cancer Cell, and Cell-family journals.
Readiness scan
Find out if this manuscript is ready to submit.
Run the Free Readiness Scan before you submit. Catch the issues editors reject on first read.
How to use this page well
These pages work best when they behave like tools, not essays. Use the quick structure first, then apply it to the exact journal and manuscript situation.
Question | What to do |
|---|---|
Use this page for | Building a point-by-point response that is easy for reviewers and editors to trust. |
Start with | State the reviewer concern clearly, then pair each response with the exact evidence or revision. |
Common mistake | Sounding defensive or abstract instead of specific about what changed. |
Best next step | Turn the response into a visible checklist or matrix before you finalize the letter. |
Decision cue: Ecology and evolution papers face a unique challenge: the gap between what a single study system can tell us and the broader principles authors want to draw. Reviewers at Nature Ecology & Evolution, Ecology Letters, and the Proceedings of the Royal Society expect manuscripts to bridge this gap honestly. Overgeneralizing from one population, one season, or one site is the most common framing failure in ecological manuscripts.
Check your ecology manuscript readiness in 60 seconds with the free scan.
What ecology and evolution reviewers screen for
Sampling design and statistical power
Ecological data are inherently variable. Reviewers check whether the sampling design is adequate for the questions being asked:
- spatial and temporal replication sufficient for the conclusions
- pseudoreplication identified and addressed
- hierarchical or nested sampling structures modeled appropriately (mixed models)
- sample sizes justified for the effect sizes expected
- seasonal and annual variation accounted for (not just one snapshot)
- environmental covariates measured and included in analyses
Generalizability of conclusions
A study of one population in one habitat during one season has limited generalizability. Reviewers evaluate whether the conclusions appropriately scale:
- are claims specific to the study system, or do they claim generality?
- is the study system representative of the broader question?
- are alternative explanations for observed patterns addressed?
- are phylogenetic relationships accounted for in comparative analyses?
Data and code sharing
Ecology and evolution have strong norms around data sharing:
- ecological data deposited in Dryad, Figshare, or field-specific repositories
- phylogenetic trees deposited in TreeBASE or Dryad
- R scripts or Python code deposited with documentation
- raw data (not just processed summaries) available
Ethical and permit considerations
- collection permits documented for field studies
- IACUC approval for animal experiments
- CITES compliance for protected species
- indigenous knowledge and benefit-sharing agreements documented where relevant
The ecology and evolution pre-submission checklist
For field-based ecology
- spatial and temporal replication described and justified
- environmental covariates measured and included
- sampling design avoids pseudoreplication
- statistical methods match the data structure (mixed models for nested designs)
- effect sizes reported alongside p-values
- raw data deposited in public repository
For phylogenetic comparative studies
- phylogeny source specified and justified
- phylogenetic uncertainty accounted for (multiple trees, Bayesian posterior)
- comparative methods appropriate for the data type (PGLS, MCMCglmm)
- trait data sources documented
For evolutionary ecology
- fitness measures appropriate and justified
- selection gradients estimated correctly
- genetic versus environmental effects distinguished where possible
- population structure accounted for
For all ecology/evolution manuscripts
- conclusions scaled to what the data support
- alternative explanations discussed
- data and code publicly available
- permits and approvals documented
- ARRIVE guidelines followed for animal experiments
Where pre-submission review helps in ecology
The Manusights free readiness scan evaluates methodology and journal fit in about 60 seconds. For ecology manuscripts, journal-specific calibration is important because Nature Ecology & Evolution has very different expectations from Ecology or Oecologia.
The $29 AI Diagnostic provides verified citations, catching missing references to recent competing studies, and figure-level feedback for data visualization. For top-tier submissions, Manusights Expert Review connects you with ecology reviewers who know what those journals prioritize.
How top ecology journals compare
Feature | Nature Ecology & Evolution | Ecology Letters | Proc R Soc B | Ecology |
|---|---|---|---|---|
Scope | Broadest, highest impact | Conceptual advances | Broad biology/ecology | Classic ecology |
Desk rejection | ~70% | ~50% | ~50% | ~40% |
Key requirement | Broad significance beyond ecology | Conceptual novelty | Biological insight | Methodological rigor |
Best for | Major ecological/evolutionary advances | New conceptual frameworks | Solid biology with evolutionary context | Rigorous ecological studies |
On this page
Reference library
Use the core publishing datasets alongside this guide
This article answers one part of the publishing decision. The reference library covers the recurring questions that usually come next: how selective journals are, how long review takes, and what the submission requirements look like across journals.
Dataset / reference guide
Peer Review Timelines by Journal
Reference-grade journal timeline data that authors, labs, and writing centers can cite when discussing realistic review timing.
Dataset / benchmark
Biomedical Journal Acceptance Rates
A field-organized acceptance-rate guide that works as a neutral benchmark when authors are deciding how selective to target.
Reference table
Journal Submission Specs
A high-utility submission table covering word limits, figure caps, reference limits, and formatting expectations.
Final step
Find out if this manuscript is ready to submit.
Run the Free Readiness Scan. See score, top issues, and journal-fit signals before you submit.
Anthropic Privacy Partner. Zero-retention manuscript processing.
Need deeper scientific feedback? See Expert Review Options
Where to go next
Start here
Same journal, next question
Supporting reads
Conversion step
Find out if this manuscript is ready to submit.
Anthropic Privacy Partner. Zero-retention manuscript processing.