Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
- The model conversion script is the same as for the Llama2 family:
Here is a sample usage:
python src/transformers/models/llama/convert_llama_weights_to_hf.py \
--input_dir /path/to/downloaded/llama/weights --model_size 7B --output_dir /output/path
Note that executing the script requires enough CPU RAM to host the whole model in float16 precision (even if the biggest versions
come in several checkpoints they each contain a part of each weight of the model, so we need to load them all in RAM).