@@ -132,21 +132,21 @@ def get_openai_api_key():
132
132
failed = 0 ,
133
133
total = 0 ,
134
134
),
135
- LLMModels .LLAMA2_7B_CHAT_GPTQ : LLMModelDetails (
136
- name = LLMModels .LLAMA2_7B_CHAT_GPTQ ,
137
- input_cost_per_1k_tokens = 0 ,
138
- output_cost_per_1k_tokens = 0 ,
139
- max_length = 4096 ,
140
- llm = get_llama_chat_model (
141
- LLMModels .LLAMA2_7B_CHAT_GPTQ .value ,
142
- model_kwargs = {"temperature" : 0 }
143
- ),
144
- input_tokens = 0 ,
145
- output_tokens = 0 ,
146
- succeeded = 0 ,
147
- failed = 0 ,
148
- total = 0 ,
149
- ),
135
+ # LLMModels.LLAMA2_7B_CHAT_GPTQ: LLMModelDetails(
136
+ # name=LLMModels.LLAMA2_7B_CHAT_GPTQ,
137
+ # input_cost_per_1k_tokens=0,
138
+ # output_cost_per_1k_tokens=0,
139
+ # max_length=4096,
140
+ # llm=get_llama_chat_model(
141
+ # LLMModels.LLAMA2_7B_CHAT_GPTQ.value,
142
+ # model_kwargs={"temperature": 0}
143
+ # ),
144
+ # input_tokens=0,
145
+ # output_tokens=0,
146
+ # succeeded=0,
147
+ # failed=0,
148
+ # total=0,
149
+ # ),
150
150
# LLMModels.LLAMA2_13B_CHAT_GPTQ: LLMModelDetails(
151
151
# name=LLMModels.LLAMA2_13B_CHAT_GPTQ,
152
152
# input_cost_per_1k_tokens=0,
@@ -162,13 +162,43 @@ def get_openai_api_key():
162
162
# failed=0,
163
163
# total=0,
164
164
# ),
165
- LLMModels .GOOGLE_GEMMA_2B_INSTRUCT : LLMModelDetails (
166
- name = LLMModels .GOOGLE_GEMMA_2B_INSTRUCT ,
165
+ # LLMModels.GOOGLE_GEMMA_2B_INSTRUCT: LLMModelDetails(
166
+ # name=LLMModels.GOOGLE_GEMMA_2B_INSTRUCT,
167
+ # input_cost_per_1k_tokens=0,
168
+ # output_cost_per_1k_tokens=0,
169
+ # max_length=8192,
170
+ # llm=get_gemma_chat_model(
171
+ # LLMModels.GOOGLE_GEMMA_2B_INSTRUCT.value,
172
+ # model_kwargs={"temperature": 0}
173
+ # ),
174
+ # input_tokens=0,
175
+ # output_tokens=0,
176
+ # succeeded=0,
177
+ # failed=0,
178
+ # total=0,
179
+ # ),
180
+ # LLMModels.GOOGLE_GEMMA_7B_INSTRUCT: LLMModelDetails(
181
+ # name=LLMModels.GOOGLE_GEMMA_7B_INSTRUCT,
182
+ # input_cost_per_1k_tokens=0,
183
+ # output_cost_per_1k_tokens=0,
184
+ # max_length=8192,
185
+ # llm=get_gemma_chat_model(
186
+ # LLMModels.GOOGLE_GEMMA_7B_INSTRUCT.value,
187
+ # model_kwargs={"temperature": 0}
188
+ # ),
189
+ # input_tokens=0,
190
+ # output_tokens=0,
191
+ # succeeded=0,
192
+ # failed=0,
193
+ # total=0,
194
+ # ),
195
+ LLMModels .CODELLAMA_7B_INSTRUCT_HF : LLMModelDetails (
196
+ name = LLMModels .CODELLAMA_7B_INSTRUCT_HF ,
167
197
input_cost_per_1k_tokens = 0 ,
168
198
output_cost_per_1k_tokens = 0 ,
169
199
max_length = 8192 ,
170
- llm = get_gemma_chat_model (
171
- LLMModels .GOOGLE_GEMMA_2B_INSTRUCT .value ,
200
+ llm = get_llama_chat_model (
201
+ LLMModels .CODELLAMA_7B_INSTRUCT_HF .value ,
172
202
model_kwargs = {"temperature" : 0 }
173
203
),
174
204
input_tokens = 0 ,
@@ -177,13 +207,13 @@ def get_openai_api_key():
177
207
failed = 0 ,
178
208
total = 0 ,
179
209
),
180
- # LLMModels.GOOGLE_GEMMA_7B_INSTRUCT : LLMModelDetails(
181
- # name=LLMModels.GOOGLE_GEMMA_7B_INSTRUCT ,
210
+ # LLMModels.CODELLAMA_13B_INSTRUCT_HF : LLMModelDetails(
211
+ # name=LLMModels.CODELLAMA_13B_INSTRUCT_HF ,
182
212
# input_cost_per_1k_tokens=0,
183
213
# output_cost_per_1k_tokens=0,
184
214
# max_length=8192,
185
- # llm=get_gemma_chat_model (
186
- # LLMModels.GOOGLE_GEMMA_7B_INSTRUCT .value,
215
+ # llm=get_llama_chat_model (
216
+ # LLMModels.CODELLAMA_13B_INSTRUCT_HF .value,
187
217
# model_kwargs={"temperature": 0}
188
218
# ),
189
219
# input_tokens=0,
0 commit comments