EasyOCR: Optical character recognition for 80+ languages
PyTorch OCR library using CRAFT and CRNN models.
Learn more about EasyOCR
EasyOCR is a Python-based OCR library built on PyTorch that uses deep learning models for text detection and recognition. The system employs CRAFT for text detection and CRNN with LSTM layers for character recognition, supporting 80+ languages across various writing systems including Latin, Chinese, Arabic, Devanagari, and Cyrillic. Model weights are automatically downloaded on first use and cached locally, with support for both GPU and CPU inference modes. The library is commonly used for document processing, scene text recognition, and information extraction from images in production and research applications.
80+ Language Support
Single unified API handles multiple writing systems including Latin, Chinese, Arabic, Devanagari, and Cyrillic. Language models can be combined for simultaneous multi-language recognition with automatic downloading and local caching.
Flexible Input Handling
Accepts file paths, OpenCV objects, numpy arrays, and image URLs without conversion code. Returns configurable output formats—detailed results with bounding boxes and confidence scores, or simplified text-only extraction.
CPU and GPU Inference
Runs in GPU-accelerated or CPU-only modes for deployment flexibility across hardware constraints. Models load once per session and remain cached in memory for fast subsequent inference calls.
import easyocr
reader = easyocr.Reader(['en'])
result = reader.readtext('invoice.jpg')
for detection in result:
bbox, text, confidence = detection
print(f'{text} (confidence: {confidence:.2f})')Compatibility fixes and improvements for better system integration.
- –Fix several compatibilities
Compatibility fixes and improvements for better system integration.
- –Fix several compatibilities
Major release adding Apple Silicon support and resolving multiple compatibility issues.
- –Add Apple Silicon support
- –Fix several compatibilities
Related Repositories
Discover similar tools and frameworks used by developers
vLLM
Fast, memory-efficient LLM inference engine with PagedAttention for production deployments at scale.
Codex CLI
OpenAI's command-line coding assistant that runs locally with ChatGPT integration for terminal use.
Llama
PyTorch inference for Meta's Llama language models.
Summarize
CLI and browser extension that generates summaries from URLs, files, videos, podcasts, and other media sources.
Civitai
Community platform for sharing Stable Diffusion models, embeddings, and AI generation assets.