Pre-trained models for BERT (Bidirectional Encoder Representations from Transformers) are publicly available.
Download and unzip an archive with the BERT-Base, Multilingual Uncased Model.
After the archive is unzipped, the directory uncased_L-12_H-768_A-12 is created and contains the following files:
bert_config.jsonbert_model.ckpt.data-00000-of-00001bert_model.ckpt.indexbert_model.ckpt.metavocab.txtPre-trained model meta-graph files are bert_model.ckpt.*.
To generate the BERT Intermediate Representation (IR) of the model, run the Model Optimizer with the following parameters: