SageMaker deployment script doesn't work
The script of deploying on sagemaker doesn't work
Did you found any way to make it run.?
While trying to host in sagemaker, I get the following error message:
"""
The tokenizer class you load from this checkpoint is 'CodeLlamaTokenizer'.
The class this function is called from is 'LlamaTokenizer'.
"""
I believe it probably has to do with this disclaimer in the "Model usage":
To use this model, please make sure to install transformers from main until the next version is released:
pip install git+https://github.com/huggingface/transformers.git@main accelerate
Since the scripts gets aHugging Face Deep Learning Container image from here:
get_huggingface_llm_image_uri("huggingface",version="0.9.3")
Hugging face team probably has to update this version to include new transformers class.
I could be totally wrong anyway.
@lucasbonservizzi is right. Code Llama just landed in Transformers main branch. There is not yet even a release so there is no sagemaker container available.
@philschmid Do you have any instructions on how would I be able to create my own container with TGI and the new transformers branch in order to try Code llama on SageMaker?
I got this error : You are using a model of type llama to instantiate a model of type . This is not supported for all configurations of models and can yield errors.
Is it related to the same issue, and is there any ETA on the release please?
Try changing the huggingface version to 0.8.2 in the deployment script
Try changing the huggingface version to 0.8.2 in the deployment script
I haven't tried your suggestion yet. However, it seems awkward that version 0.8.2 works and 0.9.3 doesn't.
As I previously commented, the error seems to be that CodeLlama models seems to be using a brand new class of tokenizer (CodeLlamaTokenizer), which has not been implemented on the newest TGI version (0.9.3) let alone an older version like 0.8.2. Anyway, I could be mistaken and downgrading might work.
In addition, if anyone has any tutorials on how to build your own TGI container it would be nice to learn. I haven't found out any tutorials on how to do it.
@lucasbonservizzi
I tested the new TGI release (1.0.3) and it seems to work pretty well.
https://github.com/aws/deep-learning-containers/pull/3286
The code snippet you get from "deploy" is updated as well.