Quick Start
If you encounter any installation issues, please check the FAQ first.
Online Experience
Official online web application
The official online version has the same functionality as the client, with a beautiful interface and rich features, requires login to use
Gradio-based online demo
A WebUI developed based on Gradio, with a simple interface and only core parsing functionality, no login required
Local Deployment
Warning
Prerequisites - Hardware and Software Environment Support
To ensure the stability and reliability of the project, we have optimized and tested only specific hardware and software environments during development. This ensures that users can achieve optimal performance and encounter the fewest compatibility issues when deploying and running the project on recommended system configurations.
By concentrating our resources and efforts on mainstream environments, our team can more efficiently resolve potential bugs and timely develop new features.
In non-mainstream environments, due to the diversity of hardware and software configurations, as well as compatibility issues with third-party dependencies, we cannot guarantee 100% usability of the project. Therefore, for users who wish to use this project in non-recommended environments, we suggest carefully reading the documentation and FAQ first, as most issues have corresponding solutions in the FAQ. Additionally, we encourage community feedback on issues so that we can gradually expand our support range.
| Parsing Backend | pipeline | *-auto-engine | *-http-client | ||
|---|---|---|---|---|---|
| hybrid | vlm | hybrid | vlm | ||
| Backend Features | Good Compatibility | High Hardware Requirements | For OpenAI Compatible Servers2 | ||
| Accuracy1 | 82+ | 90+ | |||
| Operating System | Linux3 / Windows4 / macOS5 | ||||
| Pure CPU Support | ✅ | ❌ | ✅ | ||
| GPU Acceleration | Volta and later architecture GPUs or Apple Silicon | Not Required | |||
| Min VRAM | 6GB | 10GB | 8GB | 3GB | |
| RAM | Min 16GB+, Recommended 32GB+ | 8GB | |||
| Disk Space | 20GB+, SSD Recommended | 2GB | |||
| Python Version | 3.10-3.13 | ||||
1 Accuracy metrics are the End-to-End Evaluation Overall scores from OmniDocBench (v1.5), based on the latest version of MinerU.
2 Servers compatible with OpenAI API, such as local model servers or remote model services deployed via inference frameworks like vLLM/SGLang/LMDeploy.
3 Linux only supports distributions from 2019 and later.
4 Since the key dependency ray does not support Python 3.13 on Windows, only versions 3.10~3.12 are supported.
5 macOS requires version 14.0 or later.
Install MinerU
Install MinerU using pip or uv
pip install --upgrade pip
pip install uv
uv pip install -U "mineru[all]"
Install MinerU from source code
git clone https://github.com/opendatalab/MinerU.git
cd MinerU
uv pip install -e .[all]
Tip
mineru[all] includes all core features, compatible with Windows / Linux / macOS systems, suitable for most users.
If you need to specify the inference framework for the VLM model, or only intend to install a lightweight client on an edge device, please refer to the documentation Extension Modules Installation Guide.
Deploy MinerU using Docker
MinerU provides a convenient Docker deployment method, which helps quickly set up the environment and solve some tricky environment compatibility issues. You can get the Docker Deployment Instructions in the documentation.
Using MinerU
If your device meets the GPU acceleration requirements in the table above, you can use a simple command line for document parsing:
mineru -p <input_path> -o <output_path>
pipeline to run in a pure CPU environment:
mineru -p <input_path> -o <output_path> -b pipeline
You can use MinerU for PDF parsing through various methods such as command line, API, and WebUI. For detailed instructions, please refer to the Usage Guide.