BERT base uncased

A 110M-parameter encoder-only Transformer (BERT) pretrained on BookCorpus and English Wikipedia using masked language modeling and next sentence prediction. Provides bidirectional text representations for downstream tasks like classification, token classification, and question answering. Available on Hugging Face with Apache-2.0 license and ready-to-use across Transformers (PyTorch, TensorFlow, JAX) with export options (ONNX, Core ML).

Key Information

  • Category: Language Models
  • Source: Huggingface
  • Tags: fill-mask
  • Last updated: February 24, 2026

Structured Metrics

No structured metrics captured yet.

Links

Canonical source: https://huggingface.co/google-bert/bert-base-uncased