VinBERT is a combination of two powerful Vietnamese language models: Vintern-1b-v2 and PhoBERT. With VinBERT, we create a language model optimized to better serve applications in the Vietnamese language, including tasks such as text classification, entity extraction, and more.
- VinBERT leverages the strengths of Vintern-1b-v2 and PhoBERT, providing high efficiency and accuracy for Vietnamese NLP applications.
- It supports distributed training on multiple GPUs and AWS Sagemaker infrastructure, optimizing time and resources.
- cuda:
Data parallelismandModel parallelismare supported with backendnccl - xla :
Data parallelismare supported with backendxla
- An AWS account with access to Sagemaker.
- An environment set up to interact with AWS CLI and Sagemaker.
- You have quota to use
ml.p4d.24xlargeandml.trn1.32xlargeinstances.
pip install -r requirements.txt- Prepare the environment: pull docker image flash attn base from dockerhub:
vantufit/flash-attn-cuda
docker pull vantufit/flash-attn-cuda- Run the job:
- Configure parameters such as instance type, number of GPUs, and batch size.
- Run the following command to initiate the job:
export INSTANCE=ml.p4d.24xlarge python training.py
- Run the job:
export INSTANCE=ml.trn1.32xlarge python training.py
-
Implement Tensor parallelism with
neuronx_distributed -
Monitoring training process
