bert-base-chinese
by google-bert
--- language: zh license: apache-2.0 --- - Model Details - Uses - Risks, Limitations and Biases - Training - Evaluation - How to Get Started With the Model This model has been pre-trained for Chinese, training and random input masking has been applied independently to word pieces (as in the original...
π§ Architecture Explorer
Neural network architecture
About
- Model Details - Uses - Risks, Limitations and Biases - Training - Evaluation - How to Get Started With the Model This model has been pre-trained for Chinese, training and random input masking has been applied independently to word pieces (as in the original BERT paper). - **Developed by:** Google - **Model Type:** Fill-Mask - **Language(s):** Chinese - **License:** Apache 2.0 - **Parent Model:** See the BERT base uncased model for more information ab...
π Limitations & Considerations
- β’ Benchmark scores may vary based on evaluation methodology and hardware configuration.
- β’ VRAM requirements are estimates; actual usage depends on quantization and batch size.
- β’ FNI scores are relative rankings and may change as new models are added.
- β’ Data source: [{"source_platform":"huggingface","source_url":"https://huggingface.co/google-bert/bert-base-chinese","fetched_at":"2025-12-18T04:21:59.001Z","adapter_version":"3.2.0"}]
π Related Resources
π Related Papers
No related papers linked yet. Check the model's official documentation for research papers.
π Training Datasets
Training data information not available. Refer to the original model card for details.
π Related Models
Data unavailable