Skip to content

This repository contains my experiments with the LLMs related to fine-tuning, prompt-tuning, PEFT, mechanic interpretability.

Notifications You must be signed in to change notification settings

trilokpadhi/Explorations-with-LLMs

Repository files navigation

Project Structure

  • PEFT - Contains the experiments related to the PEFT technique.
    • Prompt-Tuning - Contains the experiments related to the Prompt-Tuning technique.
    • Adapter-Tuning - Contains the experiments related to the Adapter-Tuning technique.
  • transformer-circuits - Contains the experiments related to the Mechanical Interpretation technique.
  • embeddings - How to extract embeddings from the LLMs

Techniques Used

This repository utilizes several techniques for fine-tuning and prompt-tuning of Language Models (LLMs). Here are some of the key techniques used:

  1. PEFT (Parameter Efficient Fine Tuning): This technique involves creating effective prompts that guide the model to generate the desired output.

  2. Mechanistic-Interpretability:

Datasets

Models

  1. Llama-2
  2. Gemma
  3. Mistral

Libraries

  1. Huggingface Transformers
  2. PyTorch Lightning
  3. AdapterHub
  4. TransformerLens

Experiments

Results

  • TBD

Resources

  • Repositories: -litgpt by Lighting AI : LitGPT is a command-line tool designed to easily finetune, pretrain, evaluate, and deploy 20+ LLMs on your own data. It features highly-optimized training recipes for the world's most powerful open-source large-language-models (LLMs). It is built on top of the Hugging Face Transformers library and PyTorch Lightning.

  • Libraries: -AdapterHub : AdapterHub is a library for using and sharing adapters for fine-tuning LLMs. It provides a simple API for adding adapters to LLMs and fine-tuning them on downstream tasks. It also offers a repository of pre-trained adapters for various tasks and languages.

  • Blogs:

    • LESSWRONG - A community blog focused on rationality, AI alignment, and other topics.
    • Chris Olah's Blog - A blog by Chris Olah, a researcher at OpenAI, that covers a wide range of topics related to AI and machine learning.
    • Neel Nanda's Blog - A blog by Neel Nanda, a researcher at DeepMind, that covers topics related to AI alignment, interpretability, and other areas.
    • Transformer Circuits Thread
  • Podcasts:

About

This repository contains my experiments with the LLMs related to fine-tuning, prompt-tuning, PEFT, mechanic interpretability.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published