5, 10 or 20 seats+ for your team - learn more
In this liveProject series, you’ll take on the role of a data scientist at X Education, tasked with building a self-hosted LLM to cut costs and boost productivity. You’ll code the Llama 3.2 architecture from scratch to master transformers, then work to fine-tune it for lead classification for the marketing team, and adapt it for HTML code generation to support your overworked developer colleagues. Using Unsloth to speed up training on free Colab GPUs, you’ll finish with a fully working LLM that’s flexible, efficient, and free from vendor lock-in.
In this liveProject, you’ll build a whole large language model from the ground up! You’ll step into the role of a team member at X Education, tasked by your managers with understanding how AI works by creating one of your own. Following the open source Llama 3.2 architecture, you’ll start by coding modular architecture components and assembling them into a GPU-ready model. You’ll implement a tokenizer that seamlessly integrates with the system, fine-tune the model on a sample dataset, and load pre-trained weights to unlock full functionality. Finally, you’ll generate, structure, and refine text outputs in a chat format, until you have a polished, user-friendly model running locally on your hardware.
X Education gets thousands of potential leads every day—so why do the sales team keep chasing cold leads instead of promising prospects? In this liveProject, your boss has tasked you to solve this problem by creating a powerful lead-scoring model using LLM technology. You’ll take a base purpose model and train it on X Education’s data to become a focused machine lead evaluating machine! Begin by setting up a resource-efficient environment and load the pre-trained model, then transform X Education’s 9,000-lead dataset into tokenized prompts. You’ll fine-tune Llama 3.2 to classify leads for conversion potential, and finally evaluate its performance against a few-shot baseline.
AI will be based on a pretrained Llama 3.2 model and be entirely based on-site with no reliance on external services. You’ll set up your Colab environment with Unsloth and preprocess a commit-style dataset into instruction–input–output triples. Next, benchmark the base model’s code generation skills, then fine-tune it with LoRA. Finally, you’ll evaluate the fine-tuned model against the baseline, comparing consistency, correctness, and output quality.
This liveProject is for intermediate Python programmers who understand both data processing and machine learning basics, and are keen to adapt these skills to creating and tuning Large Language Models.
Build Your LLM and Fine-Tune it for Real Tasks project for free