qwerty45-uiop commited on
Commit
f219e66
·
verified ·
1 Parent(s): 04a5269

Update src/streamlit_app.py

Browse files
Files changed (1) hide show
  1. src/streamlit_app.py +278 -49
src/streamlit_app.py CHANGED
@@ -203,86 +203,315 @@ LLM_DATABASE = {
203
 
204
  ],
205
  "code": [
206
- {"name": "CodeGen-2B", "size": "1.8GB", "description": "Salesforce code model", "parameters": "2B", "context": "2K"},
207
- {"name": "StarCoder-1B", "size": "1.1GB", "description": "BigCode project", "parameters": "1B", "context": "8K"}
208
- ],
 
 
 
 
 
 
 
 
 
 
209
  "chat": [
210
- {"name": "Alpaca-3B", "size": "2.0GB", "description": "Stanford's instruction model", "parameters": "3B", "context": "2K"},
211
- {"name": "Vicuna-3B", "size": "2.1GB", "description": "ChatGPT-style training", "parameters": "3B", "context": "2K"}
 
 
 
 
 
 
 
 
 
 
 
212
  ]
213
  },
 
 
 
 
 
 
 
 
214
  "moderate_low": { # 5-6GB
215
  "general": [
216
- {"name": "Phi-2", "size": "5.2GB", "description": "Microsoft's 2.7B model", "parameters": "2.7B", "context": "2K"},
217
- {"name": "Gemma-7B-it", "size": "4.2GB", "description": "Google instruction tuned", "parameters": "7B", "context": "8K"},
218
- {"name": "Mistral-7B-v0.1", "size": "4.1GB", "description": "Mistral AI base model", "parameters": "7B", "context": "8K"},
219
- {"name": "Llama-2-7B", "size": "4.0GB", "description": "Meta's foundation model", "parameters": "7B", "context": "4K"}
220
- ],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
221
  "code": [
222
- {"name": "CodeLlama-7B", "size": "3.8GB", "description": "Meta's code specialist", "parameters": "7B", "context": "16K"},
223
- {"name": "StarCoder-7B", "size": "4.0GB", "description": "Code generation expert", "parameters": "7B", "context": "8K"}
 
 
 
 
 
 
 
 
 
 
 
224
  ],
225
  "chat": [
226
- {"name": "Zephyr-7B-beta", "size": "4.2GB", "description": "HuggingFace chat model", "parameters": "7B", "context": "32K"},
227
- {"name": "Neural-Chat-7B", "size": "4.1GB", "description": "Intel optimized", "parameters": "7B", "context": "32K"}
 
 
 
 
 
 
 
 
 
 
 
228
  ]
229
  },
230
  "moderate": { # 7-8GB
231
  "general": [
232
- {"name": "Llama-2-7B-Chat", "size": "3.5GB", "description": "Meta's popular chat model", "parameters": "7B", "context": "4K"},
233
- {"name": "Mistral-7B-Instruct-v0.2", "size": "4.1GB", "description": "Latest Mistral instruct", "parameters": "7B", "context": "32K"},
234
- {"name": "Qwen-7B-Chat", "size": "4.0GB", "description": "Alibaba's multilingual", "parameters": "7B", "context": "32K"},
235
- {"name": "Solar-10.7B-Instruct", "size": "5.8GB", "description": "Upstage's efficient model", "parameters": "10.7B", "context": "4K"}
236
- ],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
237
  "code": [
238
- {"name": "CodeLlama-7B-Instruct", "size": "3.8GB", "description": "Instruction-tuned CodeLlama", "parameters": "7B", "context": "16K"},
239
- {"name": "WizardCoder-7B", "size": "4.0GB", "description": "Enhanced coding abilities", "parameters": "7B", "context": "16K"},
240
- {"name": "Phind-CodeLlama-34B-v2", "size": "4.2GB", "description": "4-bit quantized version", "parameters": "34B", "context": "16K"}
241
- ],
242
- "reasoning": [
243
- {"name": "WizardMath-7B", "size": "4.0GB", "description": "Mathematical reasoning", "parameters": "7B", "context": "2K"},
244
- {"name": "MetaMath-7B", "size": "3.9GB", "description": "Math problem solving", "parameters": "7B", "context": "2K"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
245
  ]
246
  },
247
  "good": { # 9-16GB
248
  "general": [
249
- {"name": "Llama-2-13B-Chat", "size": "7.3GB", "description": "Larger Llama variant", "parameters": "13B", "context": "4K"},
250
- {"name": "Vicuna-13B-v1.5", "size": "7.2GB", "description": "Enhanced Vicuna", "parameters": "13B", "context": "16K"},
251
- {"name": "OpenChat-3.5", "size": "7.1GB", "description": "High-quality chat model", "parameters": "7B", "context": "8K"},
252
- {"name": "Nous-Hermes-2-Mixtral-8x7B-DPO", "size": "12.9GB", "description": "4-bit quantized MoE", "parameters": "47B", "context": "32K"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
253
  ],
254
  "code": [
255
- {"name": "CodeLlama-13B-Instruct", "size": "7.3GB", "description": "Larger code model", "parameters": "13B", "context": "16K"},
256
- {"name": "WizardCoder-15B", "size": "8.2GB", "description": "Advanced coding", "parameters": "15B", "context": "16K"},
257
- {"name": "StarCoder-15B", "size": "8.5GB", "description": "Large code model", "parameters": "15B", "context": "8K"}
 
 
 
 
 
 
 
 
 
 
258
  ],
259
  "multimodal": [
260
- {"name": "LLaVA-7B", "size": "7.0GB", "description": "Vision + language", "parameters": "7B", "context": "2K"},
261
- {"name": "MiniGPT-4-7B", "size": "6.8GB", "description": "Multimodal chat", "parameters": "7B", "context": "2K"},
262
- {"name": "Instructblip-7B", "size": "7.2GB", "description": "Instruction-tuned VLM", "parameters": "7B", "context": "2K"}
 
 
 
 
 
 
 
 
 
 
263
  ],
264
  "reasoning": [
265
- {"name": "WizardMath-13B", "size": "7.3GB", "description": "Advanced math", "parameters": "13B", "context": "2K"},
266
- {"name": "Orca-2-13B", "size": "7.4GB", "description": "Microsoft reasoning", "parameters": "13B", "context": "4K"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
267
  ]
268
  },
269
  "high": { # 17-32GB
270
  "general": [
271
- {"name": "Mixtral-8x7B-Instruct-v0.1", "size": "26.9GB", "description": "Mixture of experts", "parameters": "47B", "context": "32K"},
272
- {"name": "Llama-2-70B-Chat", "size": "38.0GB", "description": "8-bit quantized", "parameters": "70B", "context": "4K"},
273
- {"name": "Yi-34B-Chat", "size": "19.5GB", "description": "01.AI's large model", "parameters": "34B", "context": "200K"},
274
- {"name": "Nous-Hermes-2-Yi-34B", "size": "19.2GB", "description": "Enhanced Yi variant", "parameters": "34B", "context": "200K"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
275
  ],
276
  "code": [
277
- {"name": "CodeLlama-34B-Instruct", "size": "19.0GB", "description": "Large code specialist", "parameters": "34B", "context": "16K"},
278
- {"name": "DeepSeek-Coder-33B", "size": "18.5GB", "description": "DeepSeek's coder", "parameters": "33B", "context": "16K"},
279
- {"name": "WizardCoder-34B", "size": "19.2GB", "description": "Enterprise coding", "parameters": "34B", "context": "16K"}
 
 
 
 
 
 
 
 
 
 
280
  ],
 
 
 
 
 
 
 
 
 
 
 
 
281
  "reasoning": [
282
- {"name": "WizardMath-70B", "size": "38.5GB", "description": "8-bit quantized math", "parameters": "70B", "context": "2K"},
283
- {"name": "MetaMath-70B", "size": "38.0GB", "description": "8-bit math reasoning", "parameters": "70B", "context": "2K"}
284
- ]
285
- },
 
 
 
 
 
 
286
  "ultra_high": { # >32GB
287
  "general": [
288
  {"name": "Llama-2-70B", "size": "130GB", "description": "Full precision", "parameters": "70B", "context": "4K"},
 
203
 
204
  ],
205
  "code": [
206
+
207
+ { "name": "CodeGen-2B", "size": "1.8GB", "description": "Salesforce's code generation model" },
208
+ { "name": "StarCoder-1B", "size": "1.1GB", "description": "BigCode's programming assistant" },
209
+ { "name": "InCoder-1B", "size": "1.0GB", "description": "Facebook's code infilling model" },
210
+ { "name": "PolyCoder-2.7B", "size": "2.7GB", "description": "Carnegie Mellon's code model" },
211
+ { "name": "CodeParrot-small", "size": "1.5GB", "description": "HuggingFace's Python code model" },
212
+ { "name": "SantaCoder-1.1B", "size": "1.1GB", "description": "BigCode's multilingual code model" },
213
+ { "name": "GPT-Code-2B", "size": "2.0GB", "description": "Code-specialized GPT variant" },
214
+ { "name": "AlphaCode-2B", "size": "2.2GB", "description": "DeepMind's programming model" },
215
+ { "name": "Codex-2B", "size": "2.0GB", "description": "OpenAI's code generation model" },
216
+ { "name": "TabNine-2B", "size": "2.1GB", "description": "AI code completion assistant" }
217
+
218
+ ],
219
  "chat": [
220
+
221
+ { "name": "Alpaca-3B", "size": "2.0GB", "description": "Stanford's instruction-following model" },
222
+ { "name": "Vicuna-3B", "size": "2.1GB", "description": "UC Berkeley's chat model" },
223
+ { "name": "Dolly-3B", "size": "2.2GB", "description": "Databricks' instruction-tuned model" },
224
+ { "name": "OpenAssistant-3B", "size": "2.3GB", "description": "LAION's assistant model" },
225
+ { "name": "StableVicuna-3B", "size": "2.1GB", "description": "Stable version of Vicuna" },
226
+ { "name": "MPT-3B-Chat", "size": "2.0GB", "description": "MosaicML's chat variant" },
227
+ { "name": "RedPajama-Chat-3B", "size": "2.1GB", "description": "Together AI's chat model" },
228
+ { "name": "OpenChatKit-3B", "size": "2.2GB", "description": "Together AI's open chat model" },
229
+ { "name": "Koala-3B", "size": "2.0GB", "description": "UC Berkeley's dialogue model" },
230
+ { "name": "Guanaco-3B", "size": "2.1GB", "description": "QLoRA fine-tuned model" }
231
+
232
+
233
  ]
234
  },
235
+ "reasoning": [
236
+ { "name": "WizardMath-7B", "size": "4.0GB", "description": "Mathematical reasoning specialist" },
237
+ { "name": "MAmmoTH-7B", "size": "4.1GB", "description": "Mathematical reasoning model" },
238
+ { "name": "MetaMath-7B", "size": "3.9GB", "description": "Mathematical problem solver" },
239
+ { "name": "Abel-7B", "size": "4.0GB", "description": "Advanced reasoning capabilities" },
240
+ { "name": "Orca-2-7B", "size": "4.1GB", "description": "Microsoft's reasoning specialist" }
241
+ ]
242
+ },
243
  "moderate_low": { # 5-6GB
244
  "general": [
245
+
246
+ { "name": "Phi-2", "size": "5.2GB", "description": "Microsoft's advanced 2.7B parameter model" },
247
+ { "name": "Gemma-7B", "size": "4.2GB", "description": "Google's efficient 7B model" },
248
+ { "name": "Mistral-7B-v0.1", "size": "4.1GB", "description": "Mistral AI's foundation model" },
249
+ { "name": "OpenLLaMA-7B", "size": "4.0GB", "description": "Open source 7B language model" },
250
+ { "name": "MPT-7B", "size": "4.3GB", "description": "MosaicML's transformer model" },
251
+ { "name": "Falcon-7B", "size": "4.1GB", "description": "TII's instruction model" },
252
+ { "name": "Pythia-6.9B", "size": "6.9GB", "description": "EleutherAI's large training model" },
253
+ { "name": "BLOOM-7B", "size": "7.0GB", "description": "BigScience's multilingual foundation model" },
254
+ { "name": "OLMo-7B", "size": "4.2GB", "description": "Allen AI's open language model" },
255
+ { "name": "Llama-7B", "size": "4.0GB", "description": "Meta's foundation model" },
256
+ { "name": "StableLM-7B", "size": "4.1GB", "description": "Stability AI's larger model" },
257
+ { "name": "RedPajama-7B", "size": "4.0GB", "description": "Together AI's 7B model" },
258
+ { "name": "OpenLLaMA-7B-v2", "size": "4.1GB", "description": "Improved OpenLLaMA version" },
259
+ { "name": "Vicuna-7B", "size": "3.9GB", "description": "UC Berkeley's 7B chat model" },
260
+ { "name": "Alpaca-7B", "size": "3.8GB", "description": "Stanford's instruction model" },
261
+ { "name": "GPT-NeoX-6B", "size": "6.0GB", "description": "EleutherAI's improved model" },
262
+ { "name": "OPT-6.7B", "size": "6.7GB", "description": "Meta's 6.7B parameter model" },
263
+ { "name": "T5-large", "size": "3.0GB", "description": "Large Text-to-Text Transfer" },
264
+ { "name": "FLAN-T5-large", "size": "3.2GB", "description": "Instruction-tuned T5 large" },
265
+ { "name": "UL2-base", "size": "4.0GB", "description": "Unified Language Learner base" }
266
+ ]
267
+
268
  "code": [
269
+
270
+ { "name": "CodeLlama-7B", "size": "3.8GB", "description": "Meta's specialized code model" },
271
+ { "name": "StarCoder-7B", "size": "4.0GB", "description": "Advanced code generation model" },
272
+ { "name": "SantaCoder-1.1B", "size": "1.2GB", "description": "Multilingual code model" },
273
+ { "name": "CodeGen-6B", "size": "6.0GB", "description": "Salesforce's larger code model" },
274
+ { "name": "CodeT5p-6B", "size": "6.2GB", "description": "Salesforce's code understanding model" },
275
+ { "name": "InCoder-6B", "size": "6.0GB", "description": "Facebook's large infilling model" },
276
+ { "name": "PolyCoder-6B", "size": "6.1GB", "description": "Carnegie Mellon's large code model" },
277
+ { "name": "AlphaCode-7B", "size": "4.0GB", "description": "DeepMind's competitive programming" },
278
+ { "name": "Codex-7B", "size": "4.1GB", "description": "OpenAI's advanced code model" },
279
+ { "name": "WizardCoder-7B", "size": "4.0GB", "description": "Microsoft's coding wizard" }
280
+
281
+
282
  ],
283
  "chat": [
284
+
285
+ { "name": "Zephyr-7B-beta", "size": "4.2GB", "description": "HuggingFace's chat specialist" },
286
+ { "name": "Neural-Chat-7B", "size": "4.1GB", "description": "Intel's optimized chat model" },
287
+ { "name": "OpenChat-7B", "size": "4.0GB", "description": "High-quality conversation model" },
288
+ { "name": "Nous-Hermes-7B", "size": "4.1GB", "description": "NousResearch's assistant model" },
289
+ { "name": "StableBeluga-7B", "size": "4.2GB", "description": "Stability AI's chat model" },
290
+ { "name": "Llama-2-7B-Chat", "size": "3.9GB", "description": "Meta's chat-optimized model" },
291
+ { "name": "Vicuna-7B-v1.3", "size": "3.9GB", "description": "Improved Vicuna chat model" },
292
+ { "name": "WizardLM-7B", "size": "4.0GB", "description": "Microsoft's instruction model" },
293
+ { "name": "Orca-Mini-7B", "size": "4.1GB", "description": "Microsoft's reasoning model" },
294
+ { "name": "Samantha-7B", "size": "4.0GB", "description": "Eric Hartford's assistant model" }
295
+
296
+
297
  ]
298
  },
299
  "moderate": { # 7-8GB
300
  "general": [
301
+
302
+ { "name": "Llama-2-7B-Chat", "size": "3.5GB", "description": "Meta's popular chat model (4-bit)" },
303
+ { "name": "Mistral-7B-Instruct-v0.2", "size": "4.1GB", "description": "Latest Mistral instruction model" },
304
+ { "name": "Qwen-7B-Chat", "size": "4.0GB", "description": "Alibaba's multilingual model" },
305
+ { "name": "Baichuan2-7B-Chat", "size": "4.1GB", "description": "Chinese LLM with strong capabilities" },
306
+ { "name": "Yi-6B-Chat", "size": "3.8GB", "description": "01.AI's bilingual chat model" },
307
+ { "name": "InternLM-7B-Chat", "size": "4.0GB", "description": "Shanghai AI Lab's model" },
308
+ { "name": "ChatGLM3-6B", "size": "3.7GB", "description": "Tsinghua's latest chat model" },
309
+ { "name": "Aquila-7B", "size": "4.1GB", "description": "BAAI's Chinese-English model" },
310
+ { "name": "Skywork-13B", "size": "7.2GB", "description": "Kunlun's bilingual model" },
311
+ { "name": "Llama-2-7B", "size": "3.8GB", "description": "Meta's base foundation model" },
312
+ { "name": "Mistral-7B-v0.1", "size": "4.0GB", "description": "Original Mistral foundation" },
313
+ { "name": "Solar-10.7B", "size": "5.4GB", "description": "Upstage's efficient model" },
314
+ { "name": "Nous-Hermes-2-7B", "size": "4.0GB", "description": "NousResearch's improved model" },
315
+ { "name": "OpenHermes-2.5-7B", "size": "4.1GB", "description": "Teknium's assistant model" },
316
+ { "name": "Starling-LM-7B", "size": "4.0GB", "description": "Berkeley's RLAIF model" },
317
+ { "name": "Openchat-3.5-7B", "size": "4.0GB", "description": "OpenChat's latest version" },
318
+ { "name": "Dolphin-2.2.1-7B", "size": "4.1GB", "description": "Eric Hartford's uncensored model" },
319
+ { "name": "PlatYi-7B", "size": "4.0GB", "description": "01.AI's chat-optimized model" },
320
+ { "name": "TinyLlama-1.1B-Chat", "size": "1.1GB", "description": "Compact conversational model" },
321
+ { "name": "DeepSeek-LLM-7B", "size": "4.2GB", "description": "DeepSeek's language model" }
322
+
323
+
324
  "code": [
325
+
326
+ { "name": "CodeLlama-7B-Instruct", "size": "3.8GB", "description": "Instruction-tuned code specialist" },
327
+ { "name": "WizardCoder-7B", "size": "4.0GB", "description": "Enhanced coding capabilities" },
328
+ { "name": "Phind-CodeLlama-7B-v2", "size": "3.9GB", "description": "Code search optimized model" },
329
+ { "name": "Magicoder-7B", "size": "4.0GB", "description": "OSS-Instruct trained code model" },
330
+ { "name": "DeepSeek-Coder-7B", "size": "3.9GB", "description": "DeepSeek's coding specialist" },
331
+ { "name": "WizardCoder-Python-7B", "size": "4.0GB", "description": "Python-specialized coding model" },
332
+ { "name": "StarCoder-7B", "size": "4.0GB", "description": "BigCode's 7B programming model" },
333
+ { "name": "CodeT5p-7B", "size": "4.1GB", "description": "Salesforce's code understanding" },
334
+ { "name": "InstructCodeT5p-7B", "size": "4.2GB", "description": "Instruction-tuned CodeT5p" },
335
+ { "name": "CodeGen2-7B", "size": "4.0GB", "description": "Salesforce's improved code model" },
336
+ { "name": "SantaCoder-7B", "size": "4.1GB", "description": "BigCode's multilingual coder" },
337
+ { "name": "Replit-Code-7B", "size": "4.0GB", "description": "Replit's code completion model" },
338
+ { "name": "Code-Alpaca-7B", "size": "3.9GB", "description": "Stanford's code instruction model" },
339
+ { "name": "UnixCoder-7B", "size": "4.0GB", "description": "Microsoft's large code model" }
340
+ ]
341
+ "chat": [
342
+
343
+ { "name": "Vicuna-7B-v1.5", "size": "3.9GB", "description": "Enhanced conversational model" },
344
+ { "name": "ChatGLM2-6B", "size": "3.7GB", "description": "Tsinghua's bilingual chat model" },
345
+ { "name": "Baize-7B", "size": "4.0GB", "description": "Self-chat trained model" },
346
+ { "name": "OpenBuddy-7B", "size": "4.0GB", "description": "Cross-lingual AI assistant" },
347
+ { "name": "Koala-7B", "size": "3.9GB", "description": "UC Berkeley's dialogue model" },
348
+ { "name": "GPT4All-7B", "size": "4.0GB", "description": "Nomic AI's local chat model" },
349
+ { "name": "Wizard-Vicuna-7B", "size": "4.1GB", "description": "Combined instruction model" },
350
+ { "name": "Manticore-7B", "size": "4.0GB", "description": "Multi-domain chat model" },
351
+ { "name": "Airoboros-7B", "size": "4.1GB", "description": "Context-aware chat model" },
352
+ { "name": "Samantha-1.2-7B", "size": "4.0GB", "description": "Empathetic AI assistant" }
353
+
354
+
355
+
356
+ ]
357
+ "reasoning": [
358
+ { "name": "MetaMath-7B", "size": "3.9GB", "description": "Mathematical problem solving" },
359
+ { "name": "Abel-7B", "size": "4.0GB", "description": "Advanced reasoning capabilities" },
360
+ { "name": "WizardMath-7B-V1.1", "size": "4.0GB", "description": "Enhanced math reasoning" },
361
+ { "name": "MAmmoTH-7B", "size": "4.1GB", "description": "Mathematical reasoning model" },
362
+ { "name": "Orca-2-7B", "size": "4.2GB", "description": "Microsoft's reasoning model" },
363
+ { "name": "OpenOrca-7B", "size": "4.0GB", "description": "Open-source Orca variant" }
364
+ ]
365
+ "multilingual": [
366
+
367
+ { "name": "Qwen-7B", "size": "4.0GB", "description": "Alibaba's multilingual foundation" },
368
+ { "name": "Baichuan2-7B", "size": "4.1GB", "description": "Chinese-English bilingual" },
369
+ { "name": "InternLM-7B", "size": "4.0GB", "description": "Shanghai AI Lab multilingual" },
370
+ { "name": "Chinese-LLaMA-2-7B", "size": "4.0GB", "description": "Chinese-optimized Llama" },
371
+ { "name": "Vigogne-7B", "size": "4.1GB", "description": "French instruction model" }
372
+
373
+
374
  ]
375
  },
376
  "good": { # 9-16GB
377
  "general": [
378
+
379
+ { "name": "Llama-2-13B-Chat", "size": "7.3GB", "description": "Larger Llama variant (4-bit)" },
380
+ { "name": "Vicuna-13B-v1.5", "size": "7.2GB", "description": "Enhanced large chat model" },
381
+ { "name": "OpenChat-3.5-13B", "size": "7.1GB", "description": "High-quality large chat model" },
382
+ { "name": "Qwen-14B-Chat", "size": "7.8GB", "description": "Alibaba's advanced model" },
383
+ { "name": "Baichuan2-13B-Chat", "size": "7.5GB", "description": "Large Chinese language model" },
384
+ { "name": "Yi-34B-Chat (8-bit)", "size": "19.5GB", "description": "01.AI's flagship model" },
385
+ { "name": "Nous-Hermes-13B", "size": "7.3GB", "description": "NousResearch's large assistant" },
386
+ { "name": "WizardLM-13B", "size": "7.2GB", "description": "Microsoft's instruction model" },
387
+ { "name": "Alpaca-13B", "size": "7.0GB", "description": "Stanford's large instruction model" },
388
+ { "name": "Llama-2-13B", "size": "7.0GB", "description": "Meta's 13B foundation model" },
389
+ { "name": "MPT-30B", "size": "15.0GB", "description": "MosaicML's large transformer" },
390
+ { "name": "Falcon-40B (8-bit)", "size": "20.0GB", "description": "TII's large instruction model" },
391
+ { "name": "Guanaco-13B", "size": "7.1GB", "description": "QLoRA fine-tuned model" },
392
+ { "name": "Orca-13B", "size": "7.4GB", "description": "Microsoft's reasoning model" },
393
+ { "name": "Platypus-13B", "size": "7.2GB", "description": "Fine-tuned Llama variant" },
394
+ { "name": "WizardLM-13B-V1.2", "size": "7.3GB", "description": "Improved WizardLM" },
395
+ { "name": "Nous-Hermes-2-13B", "size": "7.4GB", "description": "Enhanced Hermes model" },
396
+ { "name": "OpenOrca-13B", "size": "7.2GB", "description": "Open-source Orca recreation" },
397
+ { "name": "Airoboros-13B", "size": "7.3GB", "description": "Context-aware large model" },
398
+ { "name": "MythoMax-13B", "size": "7.2GB", "description": "Roleplay-optimized model" }
399
+
400
+
401
  ],
402
  "code": [
403
+
404
+ { "name": "CodeLlama-13B-Instruct", "size": "7.3GB", "description": "Large code generation model" },
405
+ { "name": "WizardCoder-15B", "size": "8.2GB", "description": "Advanced coding assistant" },
406
+ { "name": "StarCoder-15B", "size": "8.5GB", "description": "Large programming model" },
407
+ { "name": "CodeT5p-16B", "size": "8.8GB", "description": "Salesforce's large code model" },
408
+ { "name": "Phind-CodeLlama-34B (8-bit)", "size": "19.0GB", "description": "Large code search model" },
409
+ { "name": "DeepSeek-Coder-33B (8-bit)", "size": "18.5GB", "description": "Large coding specialist" },
410
+ { "name": "CodeLlama-13B-Python", "size": "7.4GB", "description": "Python-specialized CodeLlama" },
411
+ { "name": "WizardCoder-Python-13B", "size": "7.3GB", "description": "Python coding wizard" },
412
+ { "name": "InstructCodeT5p-16B", "size": "8.9GB", "description": "Large instruction code model" },
413
+ { "name": "CodeGen2-16B", "size": "8.7GB", "description": "Salesforce's large code model" }
414
+
415
+
416
  ],
417
  "multimodal": [
418
+
419
+ { "name": "LLaVA-13B", "size": "7.5GB", "description": "Large vision-language model" },
420
+ { "name": "MiniGPT-4-13B", "size": "7.2GB", "description": "Multimodal conversational AI" },
421
+ { "name": "InstructBLIP-13B", "size": "7.8GB", "description": "Vision-language instruction model" },
422
+ { "name": "BLIP-2-FlanT5-XL", "size": "4.8GB", "description": "Salesforce's vision-language model" },
423
+ { "name": "Flamingo-9B", "size": "9.0GB", "description": "DeepMind's few-shot learning model" },
424
+ { "name": "LLaVA-1.5-13B", "size": "7.6GB", "description": "Improved LLaVA model" },
425
+ { "name": "Otter-13B", "size": "7.4GB", "description": "Multi-modal instruction tuned" },
426
+ { "name": "mPLUG-Owl-14B", "size": "8.0GB", "description": "Alibaba's multimodal model" },
427
+ { "name": "InternLM-XComposer-7B", "size": "7.0GB", "description": "Vision-language composition" },
428
+ { "name": "Qwen-VL-7B", "size": "7.2GB", "description": "Qwen vision-language model" }
429
+
430
+
431
  ],
432
  "reasoning": [
433
+
434
+ { "name": "WizardMath-13B", "size": "7.3GB", "description": "Advanced mathematical reasoning" },
435
+ { "name": "Orca-2-13B", "size": "7.4GB", "description": "Microsoft's reasoning specialist" },
436
+ { "name": "MetaMath-13B", "size": "7.2GB", "description": "Mathematical problem solver" },
437
+ { "name": "MAmmoTH-13B", "size": "7.3GB", "description": "Large mathematical reasoning model" },
438
+ { "name": "Abel-13B", "size": "7.4GB", "description": "Advanced reasoning capabilities" },
439
+ { "name": "Goat-13B", "size": "7.2GB", "description": "Arithmetic reasoning specialist" },
440
+ { "name": "OpenOrca-Platypus-13B", "size": "7.3GB", "description": "Combined reasoning model" }
441
+
442
+ ]
443
+ "multilingual": [
444
+ { name: "Qwen-14B", size: "7.8GB", description: "Alibaba's large multilingual" },
445
+ { name: "Baichuan2-13B", size: "7.5GB", description: "Large Chinese-English model" },
446
+ { name: "InternLM-20B", size: "11.0GB", description: "Shanghai AI Lab's large model" },
447
+ { name: "Chinese-Alpaca-Plus-13B", size: "7.4GB", description: "Enhanced Chinese model" },
448
+ { name: "Polyglot-Ko-13B", size: "7.3GB", description: "Large Korean model" }
449
+
450
+
451
  ]
452
  },
453
  "high": { # 17-32GB
454
  "general": [
455
+
456
+ { "name": "Mixtral-8x7B-Instruct-v0.1", "size": "26.9GB", "description": "Mixture of experts model (4-bit)" },
457
+ { "name": "Llama-2-70B-Chat (8-bit)", "size": "38.0GB", "description": "Large language model" },
458
+ { "name": "Yi-34B-Chat", "size": "19.5GB", "description": "01.AI's flagship model" },
459
+ { "name": "Qwen-72B (4-bit)", "size": "36.0GB", "description": "Alibaba's largest model" },
460
+ { "name": "DeepSeek-67B", "size": "35.0GB", "description": "Advanced reasoning model" },
461
+ { "name": "Nous-Hermes-2-Mixtral-8x7B", "size": "26.9GB", "description": "NousResearch's MoE model" },
462
+ { "name": "Solar-10.7B", "size": "10.7GB", "description": "Upstage's efficient model" },
463
+ { "name": "Dolphin-2.5-Mixtral-8x7B", "size": "26.9GB", "description": "Uncensored Mixtral variant" },
464
+ { "name": "Llama-2-70B", "size": "35.0GB", "description": "Meta's flagship model (8-bit)" },
465
+ { "name": "Falcon-40B", "size": "20.0GB", "description": "TII's large model" },
466
+ { "name": "MPT-30B", "size": "15.0GB", "description": "MosaicML's 30B model" },
467
+ { "name": "Nous-Hermes-2-Yi-34B", "size": "19.6GB", "description": "Enhanced Yi model" },
468
+ { "name": "OpenHermes-2.5-Mistral-7B", "size": "4.1GB", "description": "Teknium's Mistral variant" },
469
+ { "name": "Starling-LM-7B-alpha", "size": "4.2GB", "description": "Berkeley's RLAIF model" },
470
+ { "name": "NeuralBeagle-14B", "size": "8.0GB", "description": "MLP KAT merged model" },
471
+ { "name": "Goliath-120B (4-bit)", "size": "60.0GB", "description": "Large merged model" },
472
+ { "name": "Xwin-LM-70B (8-bit)", "size": "38.5GB", "description": "Xwin team's large model" },
473
+ { "name": "Airoboros-L2-70B (8-bit)", "size": "38.0GB", "description": "Large context model" }
474
+
475
+
476
  ],
477
  "code": [
478
+
479
+ { "name": "CodeLlama-34B-Instruct", "size": "19.0GB", "description": "Large specialized coder" },
480
+ { "name": "DeepSeek-Coder-33B", "size": "18.5GB", "description": "Advanced code generation" },
481
+ { "name": "WizardCoder-34B", "size": "19.2GB", "description": "Enterprise-grade coding" },
482
+ { "name": "StarCoder2-15B", "size": "8.5GB", "description": "Next-gen programming model" },
483
+ { "name": "Phind-CodeLlama-34B", "size": "19.0GB", "description": "Code search specialized model" },
484
+ { "name": "Magicoder-34B", "size": "19.1GB", "description": "Large OSS-Instruct model" },
485
+ { "name": "CodeLlama-34B-Python", "size": "19.1GB", "description": "Python-specialized large model" },
486
+ { "name": "WizardCoder-Python-34B", "size": "19.2GB", "description": "Large Python specialist" },
487
+ { "name": "StarCoder-15.5B", "size": "8.8GB", "description": "Enhanced StarCoder" },
488
+ { "name": "Code-Alpaca-34B", "size": "18.9GB", "description": "Large code instruction model" }
489
+
490
+
491
  ],
492
+ "chat": [
493
+
494
+
495
+ { "name": "Vicuna-33B", "size": "18.5GB", "description": "Large conversational model" },
496
+ { "name": "Guanaco-65B (4-bit)", "size": "33.0GB", "description": "Large instruction-tuned model" },
497
+ { "name": "Alpaca-30B", "size": "18.0GB", "description": "Large Stanford model" },
498
+ { "name": "OpenBuddy-34B", "size": "19.0GB", "description": "Large cross-lingual assistant" },
499
+ { "name": "WizardLM-30B", "size": "17.0GB", "description": "Large instruction model" },
500
+ { "name": "Nous-Hermes-Llama2-70B (8-bit)", "size": "38.2GB", "description": "Large Hermes variant" },
501
+ { "name": "Airoboros-65B (4-bit)", "size": "33.5GB", "description": "Large context chat model" },
502
+ { "name": "MythoMax-L2-13B", "size": "7.4GB", "description": "Roleplay optimized" }
503
+ ],
504
  "reasoning": [
505
+
506
+ { "name": "WizardMath-70B (8-bit)", "size": "38.5GB", "description": "Premier math reasoning" },
507
+ { "name": "MetaMath-70B (8-bit)", "size": "38.0GB", "description": "Advanced mathematical AI" },
508
+ { "name": "Goat-70B (8-bit)", "size": "35.0GB", "description": "Arithmetic reasoning specialist" },
509
+ { "name": "MAmmoTH-70B (8-bit)", "size": "38.2GB", "description": "Large mathematical model" },
510
+ { "name": "Orca-2-13B", "size": "7.4GB", "description": "Microsoft's reasoning model" },
511
+ { "name": "Abel-70B (8-bit)", "size": "38.1GB", "description": "Large reasoning model" }
512
+ ]
513
+ }
514
+ },
515
  "ultra_high": { # >32GB
516
  "general": [
517
  {"name": "Llama-2-70B", "size": "130GB", "description": "Full precision", "parameters": "70B", "context": "4K"},