Pinned Loading
Repositories
Showing 10 of 13 repositories
- hybrid-distillation Public
fla-org/hybrid-distillation’s past year of commit activity - flash-linear-attention Public
🚀 Efficient implementations of state-of-the-art linear attention models
fla-org/flash-linear-attention’s past year of commit activity - distillation-fla Public Forked from OpenSparseLLMs/Linearization
Distillation pipeline from pretrained Transformers to customized FLA models
fla-org/distillation-fla’s past year of commit activity - vllm Public Forked from vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
fla-org/vllm’s past year of commit activity - fla-rl Public
A minimal RL frame work for scaling FLA models on long-horizon reasoning and agentic scenarios.
fla-org/fla-rl’s past year of commit activity
Top languages
Loading…
Most used topics
Loading…