Skip to content

Adapting Starcoderplus for Multimodal Experts

License

Notifications You must be signed in to change notification settings

huu4ontocord/aurora-m

 
 

Repository files navigation

Aurora-M

This is code to finetune and run Aurora-M, an open source Starcoderplus based model trained on 400B additional tokens of multilingual and multidomain data, and adapted for multimodal understanding using the BakLLaVA/LLaVA 1.5 code base. The 400B additional tokens were trained with BigCode's Megatron fork. This model is intended for mixture of experts (MoE) adapation using the M*DEL MoE adapatation. See our M*DEL project page for more details.

Compute provided by the LUMI Supercomputer center and JUWELS Supercomptuer center. Thank you!

Also check out our BakLLaVA project, which is a cooperation between the AI Open source organizations: LAION, Ontocord, Skunkworks OSS AI group and AI Alignment Lab.

Aurora Over BakLLaVA

About

Adapting Starcoderplus for Multimodal Experts

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 91.1%
  • Shell 4.4%
  • JavaScript 2.3%
  • HTML 1.8%
  • CSS 0.4%