Is Your Paper Ready for Applied Sciences (MDPI)? An Honest Look at the Broadest Open-Access Journal
Applied Sciences (MDPI) accepts 45-50% of submissions and evaluates technical soundness over novelty. Understand the APC, review speed, and when this broad open-access venue is the right fit.
Senior Researcher, Oncology & Cell Biology
Author context
Specializes in manuscript preparation and peer review strategy for oncology and cell biology, with deep experience evaluating submissions to Nature Medicine, JCO, Cancer Cell, and Cell-family journals.
Next step
Choose the next useful decision step first.
Use the guide or checklist that matches this page's intent before you ask for a manuscript-level diagnostic.
Applied Sciences is MDPI's broadest multidisciplinary journal, and that's saying something for a publisher that already runs over 400 titles. It covers everything from mechanical engineering to food science to quantum computing to clinical biomechanics. If you can attach the word "applied" to your discipline, there's probably a section editor waiting to handle your manuscript. That breadth is intentional, and it's the main reason the journal publishes over 15,000 papers per year, making it one of the largest academic journals in existence.
But volume isn't quality, and you shouldn't confuse one for the other. Here's what you need to know before submitting.
Applied Sciences at a glance
Applied Sciences accepts roughly 45-50% of submissions, charges an APC of about $2,400 USD, and returns first decisions in 2-4 weeks. It's fully open access, indexed in Web of Science and Scopus, and carries an impact factor around 2.5. That makes it a mid-tier OA journal with fast turnaround and low barriers to entry.
Metric | Applied Sciences (MDPI) |
|---|---|
Impact Factor (2024 JCR) | ~2.5 |
CiteScore (2024) | ~5.3 |
Acceptance Rate | ~45-50% |
Annual Publications | 15,000+ |
APC | ~$2,400 USD |
Review Time (First Decision) | 2-4 weeks |
Submission to Publication | 6-10 weeks |
Peer Review Type | Single-blind |
Open Access | Fully OA (CC BY 4.0) |
Indexed In | Web of Science (SCIE), Scopus |
Publisher | MDPI (Basel, Switzerland) |
Quartile | Q2 in most applied categories |
Those numbers tell you exactly what Applied Sciences is: a fast, accessible, technically screened venue. It isn't trying to compete with Nature or Science. It's competing with Scientific Reports, PLOS ONE, IEEE Access, and Heliyon for the massive middle band of papers that are sound but wouldn't survive the selectivity filter at a top-tier field journal.
The MDPI publishing model and what it means for you
You can't understand Applied Sciences without understanding MDPI. The publisher operates a model that's fundamentally different from Elsevier, Springer Nature, or Wiley, and it shapes everything about the editorial experience.
Speed is the product. MDPI's entire workflow is built around fast turnaround. Two to four weeks for first decision isn't a marketing claim; it's what most authors actually experience. The publisher uses an in-house editorial team that manages the review process aggressively, sending multiple reminders to reviewers and replacing non-responsive ones quickly. If you've ever waited six months for a decision at an Elsevier journal, MDPI's pace will feel almost disorienting.
Guest editors run special issues. A large fraction of Applied Sciences papers come through special issues, which are organized by external guest editors who recruit submissions around a specific theme. This isn't inherently bad, but it changes the review dynamics. Guest editors vary widely in how strict they are. Some run a tight ship with genuine peer review. Others are more lenient, and the acceptance rates for certain special issues can run noticeably higher than the journal average.
Volume is a feature, not a bug. MDPI's business model depends on publishing a lot of papers. Each one generates an APC. This doesn't mean they'll publish anything, but it does mean the editorial incentive structure is different from a subscription journal. At a subscription journal, the editor's job is to fill a fixed number of pages with the best possible content. At an APC-funded megajournal, the editor's job is to filter out technically flawed work while keeping the pipeline moving. That distinction matters.
I'm not saying this makes Applied Sciences a bad journal. It doesn't. But you should go in with your eyes open about the model you're participating in.
What Applied Sciences editors actually screen for
Despite the broad scope and relatively high acceptance rate, Applied Sciences does reject papers. Here's what trips people up.
Technical soundness is the floor. The journal doesn't ask whether your results are exciting or novel. It asks whether your methods are appropriate, your data supports your conclusions, and your statistical analysis is correct. This is the same "soundness not significance" model that PLOS ONE popularized. If your experimental design is solid, your paper has a real chance here even if the results aren't earth-shattering.
Scope fit still matters. "Applied" is the operative word. Purely theoretical work without any connection to practical application doesn't fit. A mathematical proof without engineering context, a philosophical analysis without empirical grounding, or a basic science study with no applied angle will get desk-rejected. You don't need to solve an industrial problem, but there should be a clear line between your findings and some real-world application.
Completeness over ambition. Applied Sciences wants a self-contained paper. That means complete methods, full datasets (or data availability statements), proper controls, and conclusions that don't overreach. A modest study done well will fare better than an ambitious study with obvious gaps.
English language quality. MDPI journals have a reputation for being more forgiving on language than some competitors, but Applied Sciences editors do flag papers with poor English during initial screening. If English isn't your first language, get a colleague to proofread or use a language editing service. The journal offers one through MDPI, though you're free to use any service you prefer.
The special issue question
Let's talk honestly about special issues, because they're a defining feature of Applied Sciences and a common source of confusion.
MDPI publishes hundreds of special issues per year across its journals, and Applied Sciences alone hosts a staggering number of them. You'll often receive email invitations to submit to a special issue. These invitations aren't scams, but they aren't exactly curated either. MDPI sends them broadly, often based on keyword matching with your previous publications.
Here's what you should know:
Special issue papers go through peer review. They aren't pay-to-publish bypasses. Reviewers evaluate them, and papers do get rejected.
But the review intensity can vary. Guest editors pick their own reviewers and set their own standards (within MDPI's framework). A special issue organized by a well-known research group in your field will likely have rigorous review. A special issue from an editor you've never heard of, on a topic so broad it could include almost anything, may not be held to the same bar.
Your paper won't be labeled differently. Whether you submit to a special issue or the regular track, your published paper looks the same in databases and citations. Readers won't know unless they look at the special issue collection page.
My advice: if you're invited to a special issue that genuinely matches your work and is run by someone credible in your field, it's a fine route. If the invitation feels random and the topic is suspiciously broad, submit through the regular track instead. You'll get the same journal and potentially tighter review.
How Applied Sciences compares to similar journals
If you're considering Applied Sciences, you're probably also weighing these alternatives.
Feature | Applied Sciences (MDPI) | Scientific Reports | PLOS ONE | IEEE Access | Heliyon |
|---|---|---|---|---|---|
Publisher | MDPI | Springer Nature | PLOS | IEEE | Elsevier |
Impact Factor | ~2.5 | 3.8 | 2.6 | 3.4 | 2.8 |
Acceptance Rate | ~45-50% | ~42% | ~31% | ~48% | ~35-40% |
APC | ~$2,400 | ~$2,490 | ~$2,290 | ~$1,750 | ~$2,310 |
Review Speed | 2-4 weeks | 4-8 weeks | 4-8 weeks | 4-6 weeks | 4-8 weeks |
Scope | All applied science/engineering | All natural sciences | All sciences | Electrical/computer engineering | All sciences |
Open Access | Fully OA | Fully OA | Fully OA | Fully OA | Fully OA |
Applied Sciences vs. Scientific Reports. Scientific Reports has a higher IF and slightly lower acceptance rate, which gives it a stronger selectivity signal. Review takes longer, though. If you need speed and don't mind the lower IF, Applied Sciences wins on turnaround. If you want the Springer Nature brand and a bit more prestige, Scientific Reports is worth the wait.
Applied Sciences vs. PLOS ONE. PLOS ONE is harder to get into (31% acceptance) and carries the cachet of being the journal that invented the mega-journal model. It's also run by a nonprofit, which some authors prefer on principle. Review takes about twice as long as at Applied Sciences.
Applied Sciences vs. IEEE Access. For electrical engineering, computer science, and adjacent fields, IEEE Access is the more natural home. It's got a higher IF and the IEEE brand matters in those communities. But IEEE Access doesn't cover the life sciences, materials, or environmental engineering work that Applied Sciences handles.
Applied Sciences vs. Heliyon. Heliyon (Elsevier) covers a similarly broad scope and has a lower acceptance rate. If you want the Elsevier ecosystem and slightly more selectivity, Heliyon is the alternative. Applied Sciences is faster.
Five reasons papers get rejected at Applied Sciences
Even with a 45-50% acceptance rate, half of all submissions don't make it. Here's why.
1. The paper isn't actually applied. I've seen purely theoretical manuscripts submitted here because the scope looked broad enough. It isn't. If your paper doesn't connect to a practical problem, application, or technology, it won't pass triage.
2. Incomplete methods. If someone can't reproduce your work from what you've written, reviewers will flag it. This is the most common reviewer complaint across all MDPI journals. Don't assume readers know your standard protocols. Spell them out.
3. Statistics that don't hold up. Underpowered studies, inappropriate tests, missing error bars, and cherry-picked results. Applied Sciences reviewers may not always be the top experts in your narrow subfield, but they can spot basic statistical problems.
4. Recycled content. MDPI runs plagiarism checks using iThenticate. Papers with high text overlap with the authors' previous work get flagged. If you're building on a conference paper, you need to add at least 50% new content and cite the conference version.
5. Overreaching conclusions. Stating that your lab-scale demonstration "could revolutionize the industry" without any evidence for scalability or economic viability. Keep your claims proportional to your evidence. Reviewers appreciate honesty about limitations.
Self-archiving and green open access
Since Applied Sciences is fully OA under CC BY 4.0, you can deposit the published version anywhere: your institutional repository, ResearchGate, your personal website. There are no embargo periods. This is one of the genuine advantages of MDPI's OA model. You own the copyright to your work under the Creative Commons license.
The pre-submission checklist
Before you submit, run through these questions:
- Is your work genuinely applied, with a clear connection to a real-world problem or technology?
- Are your methods detailed enough for independent reproduction?
- Do your statistical analyses match your study design and sample size?
- Have you run a plagiarism check against your own previous publications?
- Are your figures publication-quality with proper labeling and resolution?
- Have you included a data availability statement?
- If submitting to a special issue, have you verified the guest editor's credibility?
- Does your abstract contain specific results (numbers, percentages, effect sizes), not just vague statements?
If you answered yes to all of these, your paper is probably ready. If you're unsure about technical soundness or methods completeness, run your manuscript through a free Manusights AI review before submitting. It'll catch structural gaps, missing methods details, and statistical red flags faster than self-review after you've been staring at the same draft for weeks.
The honest bottom line
Applied Sciences isn't a prestige journal. It doesn't pretend to be. What it offers is fast, affordable, indexed open-access publication for technically sound applied research. That's a legitimate need, and the journal fills it well. Over 15,000 papers a year can't all be wrong.
But you should be strategic about when to use it. If your paper is competitive for a Q1 field journal, send it there first. You won't get the same career signal from an Applied Sciences publication as you would from a top-tier venue in your discipline. If your paper is solid but not strong enough for the top tier, or if you need fast publication for a grant deadline, or if you're building a publication record early in your career and need accessible venues, Applied Sciences is a reasonable choice.
Don't treat it as a dumping ground for rejected manuscripts. Treat it as what it is: a large, fast, open-access journal that values technical quality over flashy results. Submit your best version of the paper, not the version that got rejected elsewhere with "we'll fix it in revisions" still lingering in the methods section. Even with a 45-50% acceptance rate, sloppy work gets caught.
- Scopus Source Details for Applied Sciences
- MDPI APC information: https://www.mdpi.com/apc
- MDPI editorial process documentation: https://www.mdpi.com/editorial_process
Sources
- MDPI Applied Sciences journal homepage: https://www.mdpi.com/journal/applsci
- Clarivate Journal Citation Reports (2024 edition)
Reference library
Use the core publishing datasets alongside this guide
This article answers one part of the publishing decision. The reference library covers the recurring questions that usually come next: how selective journals are, how long review takes, and what the submission requirements look like across journals.
Dataset / reference guide
Peer Review Timelines by Journal
Reference-grade journal timeline data that authors, labs, and writing centers can cite when discussing realistic review timing.
Dataset / benchmark
Biomedical Journal Acceptance Rates
A field-organized acceptance-rate guide that works as a neutral benchmark when authors are deciding how selective to target.
Reference table
Journal Submission Specs
A high-utility submission table covering word limits, figure caps, reference limits, and formatting expectations.
Before you upload
Choose the next useful decision step first.
Move from this article into the next decision-support step. The scan works best once the journal and submission plan are clearer.
Use the scan once the manuscript and target journal are concrete enough to evaluate.
Anthropic Privacy Partner. Zero-retention manuscript processing.
Where to go next
Supporting reads
Conversion step
Choose the next useful decision step first.
Use the scan once the manuscript and target journal are concrete enough to evaluate.