Implementation of Switch Transformers from the paper: "Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity"
-
Updated
Apr 4, 2025 - Python
Implementation of Switch Transformers from the paper: "Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity"
The implementation of mixtures for different tasks.
🤪🧠💥 Mixture of Idiots (MoI): A Python project exploring 'Mixture of Models' (MOM) to solve complex problems by combining outputs from multiple LLMs (OpenAI, MistralAI, Gemini) using King, Duopoly, and Democracy architectures. Sometimes, a team of 'idiots' is surprisingly brilliant!
Add a description, image, and links to the mixture-of-models topic page so that developers can more easily learn about it.
To associate your repository with the mixture-of-models topic, visit your repo's landing page and select "manage topics."