layoutlm-base-uncased is a multimodal transformer model developed by Microsoft for document image understanding tasks. It incorporates both text and layout (position) features to effectively process structured documents like forms, invoices, and receipts. This base version has 113 million parameters and is pre-trained on 11 million documents from the IIT-CDIP dataset. LayoutLM enables better performance in tasks where the spatial arrangement of text plays a crucial role. The model uses a standard BERT-like architecture but enriches input with 2D positional embeddings. It achieves state-of-the-art results in form understanding and information extraction benchmarks. This model is particularly useful for document AI applications like document classification, question answering, and named entity recognition.
Features
- Combines text and layout (bounding box) embeddings
- Pre-trained on 11 million scanned document images
- Supports document image understanding and information extraction
- Uses 12 transformer layers with 768 hidden units and 12 attention heads
- Trained on the IIT-CDIP 1.0 dataset for 2 epochs
- Compatible with Hugging Face Transformers, PyTorch, and TensorFlow
- Licensed under the permissive MIT license
- Achieves SOTA on datasets like FUNSD and SROIE