International Journal of Computer Vision Submission Guide
A practical International Journal of Computer Vision (IJCV) submission guide for vision researchers evaluating their work against the journal's technical bar.
Senior Researcher, Oncology & Cell Biology
Author context
Specializes in manuscript preparation and peer review strategy for oncology and cell biology, with deep experience evaluating submissions to Nature Medicine, JCO, Cancer Cell, and Cell-family journals.
Readiness scan
Find out if this manuscript is ready to submit.
Run the Free Readiness Scan before you submit. Catch the issues editors reject on first read.
Quick answer: This International Journal of Computer Vision submission guide is for vision researchers evaluating their work against IJCV's technical bar. The journal is selective (~15-20% acceptance, 30-40% desk rejection). The editorial standard requires substantial technical contribution beyond conference versions and comprehensive experimental validation.
If you're targeting IJCV, the main risk is insufficient extension beyond conference version, missing baseline comparisons, or weak theoretical contribution.
From our manuscript review practice
Of submissions we've reviewed for IJCV, the most consistent desk-rejection trigger is insufficient technical contribution beyond a prior conference version.
How this page was created
This page was researched from IJCV's author guidelines, Springer editorial-policy materials, Clarivate JCR data, and Manusights internal analysis of submissions to IJCV and adjacent venues.
IJCV Journal Metrics
Metric | Value |
|---|---|
Impact Factor (2024 JCR) | 11.6 |
5-Year Impact Factor | ~14+ |
CiteScore | 24.0 |
Acceptance Rate | ~15-20% |
Desk Rejection Rate | ~30-40% |
First Decision | 4-6 months |
Publisher | Springer |
Source: Clarivate JCR 2024, Springer editorial disclosures (accessed April 2026).
IJCV Submission Requirements and Timeline
Requirement | Details |
|---|---|
Submission portal | Springer Editorial Manager |
Article types | Regular Paper, Short Paper, Survey |
Article length | 20-30 pages |
Cover letter | Required |
First decision | 4-6 months |
Peer review duration | 6-12 months |
Source: IJCV author guidelines.
Submission snapshot
What to pressure-test | What should already be true before upload |
|---|---|
Technical contribution | Substantial advance beyond any prior conference version |
Experimental validation | Comprehensive baselines on standard benchmarks |
Theoretical contribution | Mathematical or algorithmic novelty |
Conference-extension distinction | Cover letter quantifies new contributions over prior CVPR/ICCV/ECCV |
Reproducibility | Code and data documentation |
What this page is for
Use this page when deciding:
- whether the technical contribution is substantial enough for IJCV
- whether experimental validation meets IJCV's bar
- whether the conference-to-journal extension is sufficient
What should already be in the package
- a clear technical contribution beyond conference version
- comprehensive experimental validation against state-of-the-art
- mathematical or algorithmic novelty
- reproducibility materials
- a cover letter quantifying new contributions
Package mistakes that trigger early rejection
- Insufficient extension beyond conference version.
- Missing comprehensive baseline comparisons.
- Engineering applications without theoretical contribution.
- Thin reproducibility materials.
What makes IJCV a distinct target
IJCV is a flagship computer-vision journal.
Theory + experiment requirement: the journal differentiates from CVPR/ICCV/ECCV conference papers by demanding deeper analysis and comprehensive experiments.
Conference-extension expectation: IJCV expects journal versions to add at least 30% new content beyond conference versions.
The 30-40% desk rejection rate: decisive editorial screen.
What a strong cover letter sounds like
The strongest IJCV cover letters establish:
- the technical contribution
- the substantial extension beyond conference version
- the experimental validation scope
- the theoretical novelty
Diagnosing pre-submission problems
Problem | Fix |
|---|---|
Conference extension is thin | Add deeper theoretical analysis and additional experiments |
Baseline comparisons are incomplete | Add state-of-the-art baselines |
Theoretical contribution is weak | Strengthen mathematical analysis |
How IJCV compares against nearby alternatives
Method note: the comparison reflects published author guidelines and Manusights internal analysis. We have not personally been IJCV authors; the boundary is publicly documented editorial behavior. Pros and cons are based on documented editorial scope.
Factor | IJCV | IEEE TPAMI | IEEE Transactions on Image Processing | Computer Vision and Image Understanding |
|---|---|---|---|---|
Best fit (pros) | Computer-vision research with substantial extension | Broader pattern analysis | Image processing focus | Broader vision and image research |
Think twice if (cons) | Topic is broader pattern analysis | Topic is vision-only | Topic is general computer vision | Topic is high-impact vision |
Submit If
- the technical contribution is substantial beyond conference version
- experimental validation is comprehensive
- theoretical contribution is clearly stated
- reproducibility materials are complete
Think Twice If
- the manuscript is a thin extension of a conference paper
- baseline comparisons are incomplete
- the work fits IEEE TPAMI or specialty venue better
What to read next
Before upload, run your manuscript through an IJCV technical contribution readiness check.
In our pre-submission review work with manuscripts targeting IJCV
In our pre-submission review work with computer-vision manuscripts targeting IJCV, three patterns generate the most consistent desk rejections.
In our experience, roughly 35% of IJCV desk rejections trace to insufficient extension beyond conference version. In our experience, roughly 25% involve missing comprehensive baseline comparisons. In our experience, roughly 20% arise from weak theoretical contribution.
- Insufficient extension beyond conference version. IJCV expects journal versions to add substantial new content. We observe submissions that are minor extensions of CVPR/ICCV/ECCV papers routinely desk-rejected.
- Missing comprehensive baseline comparisons. IJCV editors expect comparison to state-of-the-art baselines. We see manuscripts comparing only to outdated baselines routinely returned.
- Weak theoretical contribution. IJCV expects mathematical or algorithmic novelty. We find papers framed as engineering improvements without theoretical analysis routinely redirected. An IJCV technical contribution readiness check can identify whether the package supports a submission.
Clarivate JCR 2024 bibliometric data places IJCV among top computer-vision journals.
What we look for during pre-submission diagnostics
In pre-submission diagnostic work for top computer-vision journals, we consistently see four signals that distinguish strong submissions from weak ones. First, the journal version must add substantial new content beyond any prior conference paper. Second, experimental validation should cover state-of-the-art baselines on standard benchmarks. Third, theoretical contribution should be clearly stated. Fourth, reproducibility materials should be available.
How conference-to-journal extension framing matters
The single most consistent feedback class we deliver in pre-submission diagnostics for IJCV is the conference-extension distinction. IJCV expects journal versions to add at least 30% new content beyond conference versions. Submissions that primarily reformat conference papers routinely receive "insufficient extension" feedback during desk screening. We coach authors to articulate the new contributions explicitly. If the new contributions reduce to "we provide more details," the extension is structurally weak. If they read like "we add a new theoretical analysis showing X, prove convergence under Y assumptions, and demonstrate generalization to Z domain," the extension is structurally substantial.
Common pre-submission diagnostic patterns we encounter
Beyond the rubric checks, three pre-submission diagnostic patterns recur most often in the manuscripts we review for IJCV. First, manuscripts where the contribution section uses generic language without specifying baseline comparisons are flagged at desk for insufficient detail. Second, manuscripts that lack engagement with the journal's recent issues are at risk of being told the contribution doesn't fit the publication conversation. Third, manuscripts with reproducibility materials marked as "available upon request" are increasingly flagged.
What separates strong from weak submissions at this tier
The strongest manuscripts we coach distinguish themselves on three operational behaviors. First, they confine the cover letter to one page. Second, they include a one-sentence elevator pitch articulating the technical contribution. Third, they identify the specific recent IJCV articles that this manuscript builds on and the specific competing work.
How editorial triage shapes IJCV submission strategy
Beyond the rubric checks, IJCV editorial triage operates on limited time per manuscript. Editors typically scan the abstract, introduction, contributions section, and experimental tables before deciding whether to invite reviewer engagement. Manuscripts that bury the technical contribution or require multiple readings to identify the central advance fare worse than manuscripts that lead with their strongest signal. We coach researchers to design the abstract, contributions section, and experimental tables for fast assessment: each independently conveys the contribution, the methodological rigor, and the empirical performance, rather than relying on linear reading of the full manuscript.
Readiness check
Run the scan against the requirements while they're in front of you.
See score, top issues, and journal-fit signals before you submit.
Final pre-submission checklist
Manuscripts checking these five items consistently clear the editorial screen at higher rates: (1) clear technical contribution statement in the cover letter's first paragraph, (2) explicit conference-extension quantification with line-level differentiation from the prior version, (3) state-of-the-art baseline comparisons on standard benchmarks with statistical significance testing where applicable, (4) reproducibility materials provided as supplementary code repositories with documented dependencies, (5) discussion of limitations, computational complexity, and future research directions integrated into the conclusions section rather than treated as an afterthought.
Frequently asked questions
Submit through Springer Editorial Manager. The journal accepts unsolicited Regular Papers, Short Papers, and Survey articles on computer vision. The cover letter should establish the technical contribution and distinguish from prior conference work.
IJCV's 2024 impact factor is around 11.6. Acceptance rate runs ~15-20% with desk-rejection around 30-40%. Median first decisions in 4-6 months.
Original research on computer vision: image and video understanding, recognition, 3D vision, generative models, vision-language models, and applications. The journal expects substantial technical contributions beyond conference-paper extensions.
Most reasons: insufficient extension beyond conference version, missing comprehensive experimental validation, scope mismatch, or thin theoretical contribution.
Sources
Before you upload
Choose the next useful decision step first.
Move from this article into the next decision-support step. The scan works best once the journal and submission plan are clearer.
Use the scan once the manuscript and target journal are concrete enough to evaluate.
Anthropic Privacy Partner. Zero-retention manuscript processing.
Where to go next
Supporting reads
Conversion step
Choose the next useful decision step first.
Use the scan once the manuscript and target journal are concrete enough to evaluate.