- We will leverage PEFT library from Hugging Face ecosystem, as well as QLoRA for more memory efficient finetuning.
- Using Llama-2-7B-bf16-sharded Model
- And Articles_Constitution_3300_Instruction_Set Dataset from HuggingFace.
- For our experiment we will need accelerate, peft, transformers, datasets and TRL to leverage the recent SFTTrainer. We will use bitsandbytes to quantize the base model into 4bit. We will also install einops as it is a requirement to load Falcon models.