In this example we will go through the steps required for fine-tuning foundation models on Amazon SageMaker by using @remote decorator for executing SageMaker Training jobs.
You can run this repository from Amazon SageMaker Studio or from your local IDE.
For additional information, take a look at the AWS Blog Fine-tune Falcon 7B and other LLMs on Amazon SageMaker with @remote decorator
This notebook is inspired by Philipp Schmid Blogs
The notebooks are currently using the latest PyTorch Training Container available for the region us-east-1
. If you are running the notebooks in a different region, make sure to update the ImageUri in the file config.yaml.
- Navigate [Available Deep Learning Containers Images](Available Deep Learning Containers Images)
- Select the right Hugging Face TGI container for model training based on your selected region
- Update ImageUri in the file config.yaml
- [Supervised - QLoRA] Falcon-7B
- [Supervised - QLoRA, FSDP] Llama-13B
- [Self-supervised - QLoRA, FSDP] Llama-13B
- [Self-supervised - QLoRA] Mistral-7B
- [Supervised - QLoRA, FSDP] Mixtral-8x7B
- [Supervised - QLoRA, DDP] Code-Llama 13B
- [Supervised - QLORA, DDP] Llama-3 8B
- [Supervised - QLoRA, DDP] Llama-3.1 8B
- [Supervised - QLoRA, DDP] Arcee AI Llama-3.1 Supernova Lite
- [Supervised - QLoRA] Llama-3.2 1B
- [Supervised - QLoRA] Llama-3.2 3B
- [Supervised - QLoRA, FSDP] Codestral-22B
- [Supervised - LoRA] TinyLlama 1.1B
- [Supervised - LoRA] Arcee Lite 1.5B
- [Supervised - LoRA] SmolLM2-1.7B-Instruct
- [Supervised - QLORA, FSDP] Qwen 2.5 7B
- [Supervised - QLORA] Falcon3 3B
- [Supervised - QLORA, FSDP] Falcon3 7B
- [Supervised - QLORA, FSDP] Llama-3.1 70B
- [Self-supervised - DoRA, FSDP] Mistral-7B v0.3
- [Supervised - QLORA, FSDP] Llama-3.3 70B