Welcome to the official repository of UniMERNet, a solution that converts images of mathematical expressions into LaTeX, suitable for a wide range of real-world scenarios.
2024.06.06 🎉🎉 Open-sourced evaluation code for UniMER dataset.
2024.05.06 🎉🎉 Open-sourced UniMER dataset, including UniMER-1M for model training and UniMER-Test for MER evaluation.
2024.05.06 🎉🎉 Added Streamlit formula recognition demo and provided local deployment App.
2024.04.24 🎉🎉 Paper now available on ArXiv.
2024.04.24 🎉🎉 Inference code and checkpoints have been released.
DirectRecognition.mp4
MunualSelection.mp4
git clone https://github.com/opendatalab/UniMERNet.git
cd UniMERNet/models
# Download the model and tokenizer individually or use git-lfs
git lfs install
git clone https://huggingface.co/wanderkid/unimernet
conda create -n unimernet python=3.10
conda activate unimernet
pip install --upgrade unimernet
-
Streamlit Application: For an interactive and user-friendly experience, use our Streamlit-based GUI. This application allows real-time formula recognition and rendering.
unimernet_gui
Ensure you have the latest version of UniMERNet installed (
pip install --upgrade unimernet
) for the streamlit GUI application. -
Command-line Demo: Predict LaTeX code from an image.
python demo.py
-
Jupyter Notebook Demo: Recognize and render formula from an image.
jupyter-lab ./demo.ipynb
Download the UniMER-Test dataset and extract it to the following directory:
./data/UniMER-Test
python test.py --cfg configs/demo.yaml
UniMERNet significantly outperforms mainstream models in recognizing real-world mathematical expressions, demonstrating superior performance across Simple Printed Expressions (SPE), Complex Printed Expressions (CPE), Screen-Captured Expressions (SCE), and Handwritten Expressions (HWE), as evidenced by the comparative BLEU Score evaluation.
UniMERNet excels in visual recognition of challenging samples, outperforming other methods.
The UniMER dataset is a specialized collection curated to advance the field of Mathematical Expression Recognition (MER). It encompasses the comprehensive UniMER-1M training set, featuring over one million instances that represent a diverse and intricate range of mathematical expressions, coupled with the UniMER Test Set, meticulously designed to benchmark MER models against real-world scenarios. The dataset details are as follows:
UniMER-1M Training Set:
- Total Samples: 1,061,791 Latex-Image pairs
- Composition: A balanced mix of concise and complex, extended formula expressions
- Aim: To train robust, high-accuracy MER models, enhancing recognition precision and generalization
UniMER Test Set:
- Total Samples: 23,757, categorized into four types of expressions:
- Simple Printed Expressions (SPE): 6,762 samples
- Complex Printed Expressions (CPE): 5,921 samples
- Screen Capture Expressions (SCE): 4,742 samples
- Handwritten Expressions (HWE): 6,332 samples
- Purpose: To provide a thorough evaluation of MER models across a spectrum of real-world conditions
You can download the dataset from OpenDataLab (recommended for users in China) or HuggingFace.
- Release inference code and checkpoints of UniMERNet.
- Release UniMER-1M and UniMER-Test.
- Open-source the Streamlit formula recognition GUI application.
- Release the training code for UniMERNet.
If you find our models / code / papers useful in your research, please consider giving us a star ⭐ and citing our work 📝, thank you :)
@misc{wang2024unimernet,
title={UniMERNet: A Universal Network for Real-World Mathematical Expression Recognition},
author={Bin Wang and Zhuangcheng Gu and Chao Xu and Bo Zhang and Botian Shi and Conghui He},
year={2024},
eprint={2404.15254},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
- VIGC. The model framework is dependent on VIGC.
- Texify. A mainstream MER algorithm, UniMERNet data processing refers to Texify.
- Latex-OCR. Another mainstream MER algorithm.
- Donut. The UniMERNet's Transformer Encoder-Decoder are referenced from Donut.
- Nougat. The tokenizer uses Nougat.
If you have any questions, comments, or suggestions, please do not hesitate to contact us at [email protected].