Skip to content

Finetuning an LLM (7B Billion parameters or less) with Mac M4 Mini #1647

Answered by awni
Hujaifa-Git asked this question in Q&A
Discussion options

You must be logged in to vote

I would look around on the internet for an estimate of fp32 / fp16 FLOPs for the machines you are interested in. The speed difference for LoRA / QLoRA training tends to follow the difference in peak flops since it's a very compute bound workflow.

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@Hujaifa-Git
Comment options

@awni
Comment options

awni Dec 5, 2024
Maintainer

Answer selected by Hujaifa-Git
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants