During this tutorial session we will get to know how LLMs are trained. We will also explore creating our own LLM. You'll get a chance to prompt LLM, finetune an LLM, perform retrieval augmented generation, and implement LLM agents!
Slides available at: https://www.canva.com/design/DAGBJ2Rzgv8/ob5MPDkmOMoziqH22qtKoA/edit?utm_content=DAGBJ2Rzgv8&utm_campaign=designshare&utm_medium=link2&utm_source=sharebutton
The overview of the tutorial can be found below:
Mistral 7B: https://arxiv.org/abs/2310.06825
Chain of Thought Prompting: https://arxiv.org/abs/2310.04959
- Load Mistral-7B model.
- Prompt engineering for better results.
- Metatags for better model performance.
- Chain-of-thought prompting.
RAG: https://arxiv.org/abs/2005.11401
- Load Mistral-7B model.
- Use custom data.
- Use prompts that interact with your data.
QLoRA: https://arxiv.org/abs/2305.14314
- Load Mistral-7B model.
- Prepare dataset.
- Use QLoRA to align the model.
OLMO: https://arxiv.org/abs/2402.00838
- Get started with OLMO framework.
- Load training data.
- Create a training job.