lucasvanmol12k downloadsGenerate LaTeX equations from images in your vault or clipboard.
⚠️ Inference API issues ⚠️
The HuggingFace Inference API is not working as Hugging Face is currently not supporting image-to-text models. Check the issue here for updates. Running locally is still supported & working.
Generate Latex equations from images and screenshots inside your vault.
Ctrl+Alt+V if you use it often!).By default, this plugin uses the HuggingFace inference API. Here's how you get your API key:
read access token in your Hugging Face profile settings. If you already have other access tokens I recommend creating one specifically for this plugin.Alternatively, you can run the model locally. This requires installing an accompanying python package. Install it using pip (or, preferably pipx):
pip install latex-ocr-server
You can check if it is installed by running
python -m latex_ocr_server --version
Open Obsidian and navigate to the Community Plugins section and enable the plugin. Then head to the LatexOCR settings tab, enable "Use local model" and configure it.

You will first need to set the python path that the plugin will use to run the model in the LatexOCR settings. You can then check if it's working using the button below it. Once this is done, press "(Re)start Server".
Note that the first time you do this, the model needs to be downloaded from huggingface, and is around ~1.4 GB. You can check the status of this download in the LatexOCR settings tab by pressing "Check Status".
The status bar at the bottom will indicate the status of the server.
| Status | Meaning |
|---|---|
| LatexOCR ✅ | server online |
| LatexOCR ⚙️ | server loading |
| LatexOCR 🌐 | downloading model |
| LatexOCR ❌ | server unreachable |
If the server is online but you encounter an error getting a response from the server, ensure that the Cache dir filepath from the plugin settings is pointing to a valid folder and that the model has successfully been saved there.
You can check if GPU support is working by running:
python -m latex_ocr_server info --gpu-available
If you want GPU support, follow the instructions at https://pytorch.org/get-started/locally/ to install pytorch with CUDA. Note you may need to uninstall torch first. torchvision and torchaudio is not required.
Massive thanks to NormXU for training and releasing the model.