Skip to content

This project fine-tunes the LLaMA-3.1-8B model using LoRA adapters for parameter-efficient training. It leverages chat templates for conversation structuring, utilizes 4-bit quantization for memory efficiency, and saves the fine-tuned model for deployment on the Hugging Face Hub.

Notifications You must be signed in to change notification settings

SURESHBEEKHANI/Finetune-LLAMA-2-On-Your-DataSet-AutoTrain-From-Hugging-Face

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 

Repository files navigation

Fine-Tuning and Inference with LLaMA-3.1-8B Model

This project demonstrates how to fine-tune the LLaMA-3.1-8B model using LoRA adapters, apply chat templates, and save the model for inference. The model is trained on local data, optimized for parameter-efficient training, and deployed to the Hugging Face Hub.

Overview

  • Model: LLaMA-3.1-8B with 4-bit quantization for efficient memory usage.
  • Techniques: Fine-tuning using LoRA (Low-Rank Adaptation) adapters, gradient checkpointing, and chat templates.
  • Data: Custom local CSV file used for training.
  • Goal: Train and deploy a chatbot model capable of handling user input in conversation-based formats.

Features

  • LoRA Adapters: Parameter-efficient fine-tuning.
  • Quantization: Efficient memory usage with 4-bit precision.
  • Chat Templates: Structured conversation flow with tokenization.
  • Model Deployment: Saving and pushing models to Hugging Face Hub in different quantization formats.

Installation

To run this project, you'll need to install the required packages. You can set this up in Google Colab or your local environment:

pip install torch transformers datasets pandas unsloth trl

About

This project fine-tunes the LLaMA-3.1-8B model using LoRA adapters for parameter-efficient training. It leverages chat templates for conversation structuring, utilizes 4-bit quantization for memory efficiency, and saves the fine-tuned model for deployment on the Hugging Face Hub.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published