Add support for the Training Method for finetuning, and for Direct-Preference Optimization (DPO) #648
Annotations
2 errors
|
pre-commit
Canceling since a higher priority waiting request for 'pre-commit-Vprov/dpo_python' exists
|
|
pre-commit
The operation was canceled.
|