@@ -85,7 +85,7 @@ def download_model(
85
85
private : bool = False ,
86
86
ollama_upload : bool = False ,
87
87
hf_upload : bool = False ,
88
- clean_run : bool = False , # New parameter
88
+ clean_run : bool = False ,
89
89
):
90
90
logger .info (f"Downloading model { model_id } ..." )
91
91
import subprocess
@@ -159,8 +159,8 @@ def convert_to_gguf(
159
159
branch : str = "" ,
160
160
filter_path : str = "" ,
161
161
ollama_upload : bool = False ,
162
- hf_upload : bool = False , # New parameter
163
- clean_run : bool = False , # New parameter
162
+ hf_upload : bool = False ,
163
+ clean_run : bool = False ,
164
164
):
165
165
"""Convert model to GGUF format with multiple quantization types and push to Ollama"""
166
166
logger .info (f"Converting model with quantization types: { quanttypes } " )
@@ -246,7 +246,7 @@ def push_to_ollama(
246
246
modelname : str ,
247
247
source_model_id : str ,
248
248
username : str ,
249
- clean_run : bool = False , # New parameter
249
+ clean_run : bool = False ,
250
250
):
251
251
"""Push converted models to Ollama using tags for different quantizations"""
252
252
logger .info ("Pushing models to Ollama..." )
@@ -375,7 +375,7 @@ def main(
375
375
private : bool = False ,
376
376
ollama_upload : bool = False ,
377
377
hf_upload : bool = False ,
378
- clean_run : bool = False , # New parameter
378
+ clean_run : bool = False ,
379
379
):
380
380
logger .info (f"Starting conversion process for { modelowner } /{ modelname } " )
381
381
converter = ModelConverter ()
0 commit comments