Transformer Explainer is an interactive visualization tool designed to help anyone learn how Transformer-based models like GPT work. It runs a live GPT-2 model right in your browser, allowing you to experiment with your own text and observe in real time how internal components and operations of the Transformer work together to predict the next tokens.
transformer-explainer.mp4 |
|
🚀 Live Demo | 📺 Demo Video |
- Node.js 20 or higher
- NPM
git clone https://github.com/poloclub/transformer-explainer.git
cd transformer-explainer
npm install
npm run dev
Then, on your web browser, access http://localhost:5173.
Transformer Explainer was created by Aeree Cho, Grace C. Kim, Alexander Karpekov, Alec Helbling, Jay Wang, Seongmin Lee, Benjamin Hoover, and Polo Chau at the Georgia Institute of Technology.
The software is available under the MIT License.
If you have any questions, feel free to open an issue or contact Aeree Cho or any of the contributors listed above.