Update src/streamlit_app.py
Browse files- src/streamlit_app.py +278 -49
src/streamlit_app.py
CHANGED
@@ -203,86 +203,315 @@ LLM_DATABASE = {
|
|
203 |
|
204 |
],
|
205 |
"code": [
|
206 |
-
|
207 |
-
|
208 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
209 |
"chat": [
|
210 |
-
|
211 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
212 |
]
|
213 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
214 |
"moderate_low": { # 5-6GB
|
215 |
"general": [
|
216 |
-
|
217 |
-
|
218 |
-
|
219 |
-
|
220 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
221 |
"code": [
|
222 |
-
|
223 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
224 |
],
|
225 |
"chat": [
|
226 |
-
|
227 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
228 |
]
|
229 |
},
|
230 |
"moderate": { # 7-8GB
|
231 |
"general": [
|
232 |
-
|
233 |
-
|
234 |
-
|
235 |
-
|
236 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
237 |
"code": [
|
238 |
-
|
239 |
-
|
240 |
-
|
241 |
-
|
242 |
-
|
243 |
-
|
244 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
245 |
]
|
246 |
},
|
247 |
"good": { # 9-16GB
|
248 |
"general": [
|
249 |
-
|
250 |
-
|
251 |
-
|
252 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
253 |
],
|
254 |
"code": [
|
255 |
-
|
256 |
-
|
257 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
258 |
],
|
259 |
"multimodal": [
|
260 |
-
|
261 |
-
|
262 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
263 |
],
|
264 |
"reasoning": [
|
265 |
-
|
266 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
267 |
]
|
268 |
},
|
269 |
"high": { # 17-32GB
|
270 |
"general": [
|
271 |
-
|
272 |
-
|
273 |
-
|
274 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
275 |
],
|
276 |
"code": [
|
277 |
-
|
278 |
-
|
279 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
280 |
],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
281 |
"reasoning": [
|
282 |
-
|
283 |
-
|
284 |
-
|
285 |
-
},
|
|
|
|
|
|
|
|
|
|
|
|
|
286 |
"ultra_high": { # >32GB
|
287 |
"general": [
|
288 |
{"name": "Llama-2-70B", "size": "130GB", "description": "Full precision", "parameters": "70B", "context": "4K"},
|
|
|
203 |
|
204 |
],
|
205 |
"code": [
|
206 |
+
|
207 |
+
{ "name": "CodeGen-2B", "size": "1.8GB", "description": "Salesforce's code generation model" },
|
208 |
+
{ "name": "StarCoder-1B", "size": "1.1GB", "description": "BigCode's programming assistant" },
|
209 |
+
{ "name": "InCoder-1B", "size": "1.0GB", "description": "Facebook's code infilling model" },
|
210 |
+
{ "name": "PolyCoder-2.7B", "size": "2.7GB", "description": "Carnegie Mellon's code model" },
|
211 |
+
{ "name": "CodeParrot-small", "size": "1.5GB", "description": "HuggingFace's Python code model" },
|
212 |
+
{ "name": "SantaCoder-1.1B", "size": "1.1GB", "description": "BigCode's multilingual code model" },
|
213 |
+
{ "name": "GPT-Code-2B", "size": "2.0GB", "description": "Code-specialized GPT variant" },
|
214 |
+
{ "name": "AlphaCode-2B", "size": "2.2GB", "description": "DeepMind's programming model" },
|
215 |
+
{ "name": "Codex-2B", "size": "2.0GB", "description": "OpenAI's code generation model" },
|
216 |
+
{ "name": "TabNine-2B", "size": "2.1GB", "description": "AI code completion assistant" }
|
217 |
+
|
218 |
+
],
|
219 |
"chat": [
|
220 |
+
|
221 |
+
{ "name": "Alpaca-3B", "size": "2.0GB", "description": "Stanford's instruction-following model" },
|
222 |
+
{ "name": "Vicuna-3B", "size": "2.1GB", "description": "UC Berkeley's chat model" },
|
223 |
+
{ "name": "Dolly-3B", "size": "2.2GB", "description": "Databricks' instruction-tuned model" },
|
224 |
+
{ "name": "OpenAssistant-3B", "size": "2.3GB", "description": "LAION's assistant model" },
|
225 |
+
{ "name": "StableVicuna-3B", "size": "2.1GB", "description": "Stable version of Vicuna" },
|
226 |
+
{ "name": "MPT-3B-Chat", "size": "2.0GB", "description": "MosaicML's chat variant" },
|
227 |
+
{ "name": "RedPajama-Chat-3B", "size": "2.1GB", "description": "Together AI's chat model" },
|
228 |
+
{ "name": "OpenChatKit-3B", "size": "2.2GB", "description": "Together AI's open chat model" },
|
229 |
+
{ "name": "Koala-3B", "size": "2.0GB", "description": "UC Berkeley's dialogue model" },
|
230 |
+
{ "name": "Guanaco-3B", "size": "2.1GB", "description": "QLoRA fine-tuned model" }
|
231 |
+
|
232 |
+
|
233 |
]
|
234 |
},
|
235 |
+
"reasoning": [
|
236 |
+
{ "name": "WizardMath-7B", "size": "4.0GB", "description": "Mathematical reasoning specialist" },
|
237 |
+
{ "name": "MAmmoTH-7B", "size": "4.1GB", "description": "Mathematical reasoning model" },
|
238 |
+
{ "name": "MetaMath-7B", "size": "3.9GB", "description": "Mathematical problem solver" },
|
239 |
+
{ "name": "Abel-7B", "size": "4.0GB", "description": "Advanced reasoning capabilities" },
|
240 |
+
{ "name": "Orca-2-7B", "size": "4.1GB", "description": "Microsoft's reasoning specialist" }
|
241 |
+
]
|
242 |
+
},
|
243 |
"moderate_low": { # 5-6GB
|
244 |
"general": [
|
245 |
+
|
246 |
+
{ "name": "Phi-2", "size": "5.2GB", "description": "Microsoft's advanced 2.7B parameter model" },
|
247 |
+
{ "name": "Gemma-7B", "size": "4.2GB", "description": "Google's efficient 7B model" },
|
248 |
+
{ "name": "Mistral-7B-v0.1", "size": "4.1GB", "description": "Mistral AI's foundation model" },
|
249 |
+
{ "name": "OpenLLaMA-7B", "size": "4.0GB", "description": "Open source 7B language model" },
|
250 |
+
{ "name": "MPT-7B", "size": "4.3GB", "description": "MosaicML's transformer model" },
|
251 |
+
{ "name": "Falcon-7B", "size": "4.1GB", "description": "TII's instruction model" },
|
252 |
+
{ "name": "Pythia-6.9B", "size": "6.9GB", "description": "EleutherAI's large training model" },
|
253 |
+
{ "name": "BLOOM-7B", "size": "7.0GB", "description": "BigScience's multilingual foundation model" },
|
254 |
+
{ "name": "OLMo-7B", "size": "4.2GB", "description": "Allen AI's open language model" },
|
255 |
+
{ "name": "Llama-7B", "size": "4.0GB", "description": "Meta's foundation model" },
|
256 |
+
{ "name": "StableLM-7B", "size": "4.1GB", "description": "Stability AI's larger model" },
|
257 |
+
{ "name": "RedPajama-7B", "size": "4.0GB", "description": "Together AI's 7B model" },
|
258 |
+
{ "name": "OpenLLaMA-7B-v2", "size": "4.1GB", "description": "Improved OpenLLaMA version" },
|
259 |
+
{ "name": "Vicuna-7B", "size": "3.9GB", "description": "UC Berkeley's 7B chat model" },
|
260 |
+
{ "name": "Alpaca-7B", "size": "3.8GB", "description": "Stanford's instruction model" },
|
261 |
+
{ "name": "GPT-NeoX-6B", "size": "6.0GB", "description": "EleutherAI's improved model" },
|
262 |
+
{ "name": "OPT-6.7B", "size": "6.7GB", "description": "Meta's 6.7B parameter model" },
|
263 |
+
{ "name": "T5-large", "size": "3.0GB", "description": "Large Text-to-Text Transfer" },
|
264 |
+
{ "name": "FLAN-T5-large", "size": "3.2GB", "description": "Instruction-tuned T5 large" },
|
265 |
+
{ "name": "UL2-base", "size": "4.0GB", "description": "Unified Language Learner base" }
|
266 |
+
]
|
267 |
+
|
268 |
"code": [
|
269 |
+
|
270 |
+
{ "name": "CodeLlama-7B", "size": "3.8GB", "description": "Meta's specialized code model" },
|
271 |
+
{ "name": "StarCoder-7B", "size": "4.0GB", "description": "Advanced code generation model" },
|
272 |
+
{ "name": "SantaCoder-1.1B", "size": "1.2GB", "description": "Multilingual code model" },
|
273 |
+
{ "name": "CodeGen-6B", "size": "6.0GB", "description": "Salesforce's larger code model" },
|
274 |
+
{ "name": "CodeT5p-6B", "size": "6.2GB", "description": "Salesforce's code understanding model" },
|
275 |
+
{ "name": "InCoder-6B", "size": "6.0GB", "description": "Facebook's large infilling model" },
|
276 |
+
{ "name": "PolyCoder-6B", "size": "6.1GB", "description": "Carnegie Mellon's large code model" },
|
277 |
+
{ "name": "AlphaCode-7B", "size": "4.0GB", "description": "DeepMind's competitive programming" },
|
278 |
+
{ "name": "Codex-7B", "size": "4.1GB", "description": "OpenAI's advanced code model" },
|
279 |
+
{ "name": "WizardCoder-7B", "size": "4.0GB", "description": "Microsoft's coding wizard" }
|
280 |
+
|
281 |
+
|
282 |
],
|
283 |
"chat": [
|
284 |
+
|
285 |
+
{ "name": "Zephyr-7B-beta", "size": "4.2GB", "description": "HuggingFace's chat specialist" },
|
286 |
+
{ "name": "Neural-Chat-7B", "size": "4.1GB", "description": "Intel's optimized chat model" },
|
287 |
+
{ "name": "OpenChat-7B", "size": "4.0GB", "description": "High-quality conversation model" },
|
288 |
+
{ "name": "Nous-Hermes-7B", "size": "4.1GB", "description": "NousResearch's assistant model" },
|
289 |
+
{ "name": "StableBeluga-7B", "size": "4.2GB", "description": "Stability AI's chat model" },
|
290 |
+
{ "name": "Llama-2-7B-Chat", "size": "3.9GB", "description": "Meta's chat-optimized model" },
|
291 |
+
{ "name": "Vicuna-7B-v1.3", "size": "3.9GB", "description": "Improved Vicuna chat model" },
|
292 |
+
{ "name": "WizardLM-7B", "size": "4.0GB", "description": "Microsoft's instruction model" },
|
293 |
+
{ "name": "Orca-Mini-7B", "size": "4.1GB", "description": "Microsoft's reasoning model" },
|
294 |
+
{ "name": "Samantha-7B", "size": "4.0GB", "description": "Eric Hartford's assistant model" }
|
295 |
+
|
296 |
+
|
297 |
]
|
298 |
},
|
299 |
"moderate": { # 7-8GB
|
300 |
"general": [
|
301 |
+
|
302 |
+
{ "name": "Llama-2-7B-Chat", "size": "3.5GB", "description": "Meta's popular chat model (4-bit)" },
|
303 |
+
{ "name": "Mistral-7B-Instruct-v0.2", "size": "4.1GB", "description": "Latest Mistral instruction model" },
|
304 |
+
{ "name": "Qwen-7B-Chat", "size": "4.0GB", "description": "Alibaba's multilingual model" },
|
305 |
+
{ "name": "Baichuan2-7B-Chat", "size": "4.1GB", "description": "Chinese LLM with strong capabilities" },
|
306 |
+
{ "name": "Yi-6B-Chat", "size": "3.8GB", "description": "01.AI's bilingual chat model" },
|
307 |
+
{ "name": "InternLM-7B-Chat", "size": "4.0GB", "description": "Shanghai AI Lab's model" },
|
308 |
+
{ "name": "ChatGLM3-6B", "size": "3.7GB", "description": "Tsinghua's latest chat model" },
|
309 |
+
{ "name": "Aquila-7B", "size": "4.1GB", "description": "BAAI's Chinese-English model" },
|
310 |
+
{ "name": "Skywork-13B", "size": "7.2GB", "description": "Kunlun's bilingual model" },
|
311 |
+
{ "name": "Llama-2-7B", "size": "3.8GB", "description": "Meta's base foundation model" },
|
312 |
+
{ "name": "Mistral-7B-v0.1", "size": "4.0GB", "description": "Original Mistral foundation" },
|
313 |
+
{ "name": "Solar-10.7B", "size": "5.4GB", "description": "Upstage's efficient model" },
|
314 |
+
{ "name": "Nous-Hermes-2-7B", "size": "4.0GB", "description": "NousResearch's improved model" },
|
315 |
+
{ "name": "OpenHermes-2.5-7B", "size": "4.1GB", "description": "Teknium's assistant model" },
|
316 |
+
{ "name": "Starling-LM-7B", "size": "4.0GB", "description": "Berkeley's RLAIF model" },
|
317 |
+
{ "name": "Openchat-3.5-7B", "size": "4.0GB", "description": "OpenChat's latest version" },
|
318 |
+
{ "name": "Dolphin-2.2.1-7B", "size": "4.1GB", "description": "Eric Hartford's uncensored model" },
|
319 |
+
{ "name": "PlatYi-7B", "size": "4.0GB", "description": "01.AI's chat-optimized model" },
|
320 |
+
{ "name": "TinyLlama-1.1B-Chat", "size": "1.1GB", "description": "Compact conversational model" },
|
321 |
+
{ "name": "DeepSeek-LLM-7B", "size": "4.2GB", "description": "DeepSeek's language model" }
|
322 |
+
|
323 |
+
|
324 |
"code": [
|
325 |
+
|
326 |
+
{ "name": "CodeLlama-7B-Instruct", "size": "3.8GB", "description": "Instruction-tuned code specialist" },
|
327 |
+
{ "name": "WizardCoder-7B", "size": "4.0GB", "description": "Enhanced coding capabilities" },
|
328 |
+
{ "name": "Phind-CodeLlama-7B-v2", "size": "3.9GB", "description": "Code search optimized model" },
|
329 |
+
{ "name": "Magicoder-7B", "size": "4.0GB", "description": "OSS-Instruct trained code model" },
|
330 |
+
{ "name": "DeepSeek-Coder-7B", "size": "3.9GB", "description": "DeepSeek's coding specialist" },
|
331 |
+
{ "name": "WizardCoder-Python-7B", "size": "4.0GB", "description": "Python-specialized coding model" },
|
332 |
+
{ "name": "StarCoder-7B", "size": "4.0GB", "description": "BigCode's 7B programming model" },
|
333 |
+
{ "name": "CodeT5p-7B", "size": "4.1GB", "description": "Salesforce's code understanding" },
|
334 |
+
{ "name": "InstructCodeT5p-7B", "size": "4.2GB", "description": "Instruction-tuned CodeT5p" },
|
335 |
+
{ "name": "CodeGen2-7B", "size": "4.0GB", "description": "Salesforce's improved code model" },
|
336 |
+
{ "name": "SantaCoder-7B", "size": "4.1GB", "description": "BigCode's multilingual coder" },
|
337 |
+
{ "name": "Replit-Code-7B", "size": "4.0GB", "description": "Replit's code completion model" },
|
338 |
+
{ "name": "Code-Alpaca-7B", "size": "3.9GB", "description": "Stanford's code instruction model" },
|
339 |
+
{ "name": "UnixCoder-7B", "size": "4.0GB", "description": "Microsoft's large code model" }
|
340 |
+
]
|
341 |
+
"chat": [
|
342 |
+
|
343 |
+
{ "name": "Vicuna-7B-v1.5", "size": "3.9GB", "description": "Enhanced conversational model" },
|
344 |
+
{ "name": "ChatGLM2-6B", "size": "3.7GB", "description": "Tsinghua's bilingual chat model" },
|
345 |
+
{ "name": "Baize-7B", "size": "4.0GB", "description": "Self-chat trained model" },
|
346 |
+
{ "name": "OpenBuddy-7B", "size": "4.0GB", "description": "Cross-lingual AI assistant" },
|
347 |
+
{ "name": "Koala-7B", "size": "3.9GB", "description": "UC Berkeley's dialogue model" },
|
348 |
+
{ "name": "GPT4All-7B", "size": "4.0GB", "description": "Nomic AI's local chat model" },
|
349 |
+
{ "name": "Wizard-Vicuna-7B", "size": "4.1GB", "description": "Combined instruction model" },
|
350 |
+
{ "name": "Manticore-7B", "size": "4.0GB", "description": "Multi-domain chat model" },
|
351 |
+
{ "name": "Airoboros-7B", "size": "4.1GB", "description": "Context-aware chat model" },
|
352 |
+
{ "name": "Samantha-1.2-7B", "size": "4.0GB", "description": "Empathetic AI assistant" }
|
353 |
+
|
354 |
+
|
355 |
+
|
356 |
+
]
|
357 |
+
"reasoning": [
|
358 |
+
{ "name": "MetaMath-7B", "size": "3.9GB", "description": "Mathematical problem solving" },
|
359 |
+
{ "name": "Abel-7B", "size": "4.0GB", "description": "Advanced reasoning capabilities" },
|
360 |
+
{ "name": "WizardMath-7B-V1.1", "size": "4.0GB", "description": "Enhanced math reasoning" },
|
361 |
+
{ "name": "MAmmoTH-7B", "size": "4.1GB", "description": "Mathematical reasoning model" },
|
362 |
+
{ "name": "Orca-2-7B", "size": "4.2GB", "description": "Microsoft's reasoning model" },
|
363 |
+
{ "name": "OpenOrca-7B", "size": "4.0GB", "description": "Open-source Orca variant" }
|
364 |
+
]
|
365 |
+
"multilingual": [
|
366 |
+
|
367 |
+
{ "name": "Qwen-7B", "size": "4.0GB", "description": "Alibaba's multilingual foundation" },
|
368 |
+
{ "name": "Baichuan2-7B", "size": "4.1GB", "description": "Chinese-English bilingual" },
|
369 |
+
{ "name": "InternLM-7B", "size": "4.0GB", "description": "Shanghai AI Lab multilingual" },
|
370 |
+
{ "name": "Chinese-LLaMA-2-7B", "size": "4.0GB", "description": "Chinese-optimized Llama" },
|
371 |
+
{ "name": "Vigogne-7B", "size": "4.1GB", "description": "French instruction model" }
|
372 |
+
|
373 |
+
|
374 |
]
|
375 |
},
|
376 |
"good": { # 9-16GB
|
377 |
"general": [
|
378 |
+
|
379 |
+
{ "name": "Llama-2-13B-Chat", "size": "7.3GB", "description": "Larger Llama variant (4-bit)" },
|
380 |
+
{ "name": "Vicuna-13B-v1.5", "size": "7.2GB", "description": "Enhanced large chat model" },
|
381 |
+
{ "name": "OpenChat-3.5-13B", "size": "7.1GB", "description": "High-quality large chat model" },
|
382 |
+
{ "name": "Qwen-14B-Chat", "size": "7.8GB", "description": "Alibaba's advanced model" },
|
383 |
+
{ "name": "Baichuan2-13B-Chat", "size": "7.5GB", "description": "Large Chinese language model" },
|
384 |
+
{ "name": "Yi-34B-Chat (8-bit)", "size": "19.5GB", "description": "01.AI's flagship model" },
|
385 |
+
{ "name": "Nous-Hermes-13B", "size": "7.3GB", "description": "NousResearch's large assistant" },
|
386 |
+
{ "name": "WizardLM-13B", "size": "7.2GB", "description": "Microsoft's instruction model" },
|
387 |
+
{ "name": "Alpaca-13B", "size": "7.0GB", "description": "Stanford's large instruction model" },
|
388 |
+
{ "name": "Llama-2-13B", "size": "7.0GB", "description": "Meta's 13B foundation model" },
|
389 |
+
{ "name": "MPT-30B", "size": "15.0GB", "description": "MosaicML's large transformer" },
|
390 |
+
{ "name": "Falcon-40B (8-bit)", "size": "20.0GB", "description": "TII's large instruction model" },
|
391 |
+
{ "name": "Guanaco-13B", "size": "7.1GB", "description": "QLoRA fine-tuned model" },
|
392 |
+
{ "name": "Orca-13B", "size": "7.4GB", "description": "Microsoft's reasoning model" },
|
393 |
+
{ "name": "Platypus-13B", "size": "7.2GB", "description": "Fine-tuned Llama variant" },
|
394 |
+
{ "name": "WizardLM-13B-V1.2", "size": "7.3GB", "description": "Improved WizardLM" },
|
395 |
+
{ "name": "Nous-Hermes-2-13B", "size": "7.4GB", "description": "Enhanced Hermes model" },
|
396 |
+
{ "name": "OpenOrca-13B", "size": "7.2GB", "description": "Open-source Orca recreation" },
|
397 |
+
{ "name": "Airoboros-13B", "size": "7.3GB", "description": "Context-aware large model" },
|
398 |
+
{ "name": "MythoMax-13B", "size": "7.2GB", "description": "Roleplay-optimized model" }
|
399 |
+
|
400 |
+
|
401 |
],
|
402 |
"code": [
|
403 |
+
|
404 |
+
{ "name": "CodeLlama-13B-Instruct", "size": "7.3GB", "description": "Large code generation model" },
|
405 |
+
{ "name": "WizardCoder-15B", "size": "8.2GB", "description": "Advanced coding assistant" },
|
406 |
+
{ "name": "StarCoder-15B", "size": "8.5GB", "description": "Large programming model" },
|
407 |
+
{ "name": "CodeT5p-16B", "size": "8.8GB", "description": "Salesforce's large code model" },
|
408 |
+
{ "name": "Phind-CodeLlama-34B (8-bit)", "size": "19.0GB", "description": "Large code search model" },
|
409 |
+
{ "name": "DeepSeek-Coder-33B (8-bit)", "size": "18.5GB", "description": "Large coding specialist" },
|
410 |
+
{ "name": "CodeLlama-13B-Python", "size": "7.4GB", "description": "Python-specialized CodeLlama" },
|
411 |
+
{ "name": "WizardCoder-Python-13B", "size": "7.3GB", "description": "Python coding wizard" },
|
412 |
+
{ "name": "InstructCodeT5p-16B", "size": "8.9GB", "description": "Large instruction code model" },
|
413 |
+
{ "name": "CodeGen2-16B", "size": "8.7GB", "description": "Salesforce's large code model" }
|
414 |
+
|
415 |
+
|
416 |
],
|
417 |
"multimodal": [
|
418 |
+
|
419 |
+
{ "name": "LLaVA-13B", "size": "7.5GB", "description": "Large vision-language model" },
|
420 |
+
{ "name": "MiniGPT-4-13B", "size": "7.2GB", "description": "Multimodal conversational AI" },
|
421 |
+
{ "name": "InstructBLIP-13B", "size": "7.8GB", "description": "Vision-language instruction model" },
|
422 |
+
{ "name": "BLIP-2-FlanT5-XL", "size": "4.8GB", "description": "Salesforce's vision-language model" },
|
423 |
+
{ "name": "Flamingo-9B", "size": "9.0GB", "description": "DeepMind's few-shot learning model" },
|
424 |
+
{ "name": "LLaVA-1.5-13B", "size": "7.6GB", "description": "Improved LLaVA model" },
|
425 |
+
{ "name": "Otter-13B", "size": "7.4GB", "description": "Multi-modal instruction tuned" },
|
426 |
+
{ "name": "mPLUG-Owl-14B", "size": "8.0GB", "description": "Alibaba's multimodal model" },
|
427 |
+
{ "name": "InternLM-XComposer-7B", "size": "7.0GB", "description": "Vision-language composition" },
|
428 |
+
{ "name": "Qwen-VL-7B", "size": "7.2GB", "description": "Qwen vision-language model" }
|
429 |
+
|
430 |
+
|
431 |
],
|
432 |
"reasoning": [
|
433 |
+
|
434 |
+
{ "name": "WizardMath-13B", "size": "7.3GB", "description": "Advanced mathematical reasoning" },
|
435 |
+
{ "name": "Orca-2-13B", "size": "7.4GB", "description": "Microsoft's reasoning specialist" },
|
436 |
+
{ "name": "MetaMath-13B", "size": "7.2GB", "description": "Mathematical problem solver" },
|
437 |
+
{ "name": "MAmmoTH-13B", "size": "7.3GB", "description": "Large mathematical reasoning model" },
|
438 |
+
{ "name": "Abel-13B", "size": "7.4GB", "description": "Advanced reasoning capabilities" },
|
439 |
+
{ "name": "Goat-13B", "size": "7.2GB", "description": "Arithmetic reasoning specialist" },
|
440 |
+
{ "name": "OpenOrca-Platypus-13B", "size": "7.3GB", "description": "Combined reasoning model" }
|
441 |
+
|
442 |
+
]
|
443 |
+
"multilingual": [
|
444 |
+
{ name: "Qwen-14B", size: "7.8GB", description: "Alibaba's large multilingual" },
|
445 |
+
{ name: "Baichuan2-13B", size: "7.5GB", description: "Large Chinese-English model" },
|
446 |
+
{ name: "InternLM-20B", size: "11.0GB", description: "Shanghai AI Lab's large model" },
|
447 |
+
{ name: "Chinese-Alpaca-Plus-13B", size: "7.4GB", description: "Enhanced Chinese model" },
|
448 |
+
{ name: "Polyglot-Ko-13B", size: "7.3GB", description: "Large Korean model" }
|
449 |
+
|
450 |
+
|
451 |
]
|
452 |
},
|
453 |
"high": { # 17-32GB
|
454 |
"general": [
|
455 |
+
|
456 |
+
{ "name": "Mixtral-8x7B-Instruct-v0.1", "size": "26.9GB", "description": "Mixture of experts model (4-bit)" },
|
457 |
+
{ "name": "Llama-2-70B-Chat (8-bit)", "size": "38.0GB", "description": "Large language model" },
|
458 |
+
{ "name": "Yi-34B-Chat", "size": "19.5GB", "description": "01.AI's flagship model" },
|
459 |
+
{ "name": "Qwen-72B (4-bit)", "size": "36.0GB", "description": "Alibaba's largest model" },
|
460 |
+
{ "name": "DeepSeek-67B", "size": "35.0GB", "description": "Advanced reasoning model" },
|
461 |
+
{ "name": "Nous-Hermes-2-Mixtral-8x7B", "size": "26.9GB", "description": "NousResearch's MoE model" },
|
462 |
+
{ "name": "Solar-10.7B", "size": "10.7GB", "description": "Upstage's efficient model" },
|
463 |
+
{ "name": "Dolphin-2.5-Mixtral-8x7B", "size": "26.9GB", "description": "Uncensored Mixtral variant" },
|
464 |
+
{ "name": "Llama-2-70B", "size": "35.0GB", "description": "Meta's flagship model (8-bit)" },
|
465 |
+
{ "name": "Falcon-40B", "size": "20.0GB", "description": "TII's large model" },
|
466 |
+
{ "name": "MPT-30B", "size": "15.0GB", "description": "MosaicML's 30B model" },
|
467 |
+
{ "name": "Nous-Hermes-2-Yi-34B", "size": "19.6GB", "description": "Enhanced Yi model" },
|
468 |
+
{ "name": "OpenHermes-2.5-Mistral-7B", "size": "4.1GB", "description": "Teknium's Mistral variant" },
|
469 |
+
{ "name": "Starling-LM-7B-alpha", "size": "4.2GB", "description": "Berkeley's RLAIF model" },
|
470 |
+
{ "name": "NeuralBeagle-14B", "size": "8.0GB", "description": "MLP KAT merged model" },
|
471 |
+
{ "name": "Goliath-120B (4-bit)", "size": "60.0GB", "description": "Large merged model" },
|
472 |
+
{ "name": "Xwin-LM-70B (8-bit)", "size": "38.5GB", "description": "Xwin team's large model" },
|
473 |
+
{ "name": "Airoboros-L2-70B (8-bit)", "size": "38.0GB", "description": "Large context model" }
|
474 |
+
|
475 |
+
|
476 |
],
|
477 |
"code": [
|
478 |
+
|
479 |
+
{ "name": "CodeLlama-34B-Instruct", "size": "19.0GB", "description": "Large specialized coder" },
|
480 |
+
{ "name": "DeepSeek-Coder-33B", "size": "18.5GB", "description": "Advanced code generation" },
|
481 |
+
{ "name": "WizardCoder-34B", "size": "19.2GB", "description": "Enterprise-grade coding" },
|
482 |
+
{ "name": "StarCoder2-15B", "size": "8.5GB", "description": "Next-gen programming model" },
|
483 |
+
{ "name": "Phind-CodeLlama-34B", "size": "19.0GB", "description": "Code search specialized model" },
|
484 |
+
{ "name": "Magicoder-34B", "size": "19.1GB", "description": "Large OSS-Instruct model" },
|
485 |
+
{ "name": "CodeLlama-34B-Python", "size": "19.1GB", "description": "Python-specialized large model" },
|
486 |
+
{ "name": "WizardCoder-Python-34B", "size": "19.2GB", "description": "Large Python specialist" },
|
487 |
+
{ "name": "StarCoder-15.5B", "size": "8.8GB", "description": "Enhanced StarCoder" },
|
488 |
+
{ "name": "Code-Alpaca-34B", "size": "18.9GB", "description": "Large code instruction model" }
|
489 |
+
|
490 |
+
|
491 |
],
|
492 |
+
"chat": [
|
493 |
+
|
494 |
+
|
495 |
+
{ "name": "Vicuna-33B", "size": "18.5GB", "description": "Large conversational model" },
|
496 |
+
{ "name": "Guanaco-65B (4-bit)", "size": "33.0GB", "description": "Large instruction-tuned model" },
|
497 |
+
{ "name": "Alpaca-30B", "size": "18.0GB", "description": "Large Stanford model" },
|
498 |
+
{ "name": "OpenBuddy-34B", "size": "19.0GB", "description": "Large cross-lingual assistant" },
|
499 |
+
{ "name": "WizardLM-30B", "size": "17.0GB", "description": "Large instruction model" },
|
500 |
+
{ "name": "Nous-Hermes-Llama2-70B (8-bit)", "size": "38.2GB", "description": "Large Hermes variant" },
|
501 |
+
{ "name": "Airoboros-65B (4-bit)", "size": "33.5GB", "description": "Large context chat model" },
|
502 |
+
{ "name": "MythoMax-L2-13B", "size": "7.4GB", "description": "Roleplay optimized" }
|
503 |
+
],
|
504 |
"reasoning": [
|
505 |
+
|
506 |
+
{ "name": "WizardMath-70B (8-bit)", "size": "38.5GB", "description": "Premier math reasoning" },
|
507 |
+
{ "name": "MetaMath-70B (8-bit)", "size": "38.0GB", "description": "Advanced mathematical AI" },
|
508 |
+
{ "name": "Goat-70B (8-bit)", "size": "35.0GB", "description": "Arithmetic reasoning specialist" },
|
509 |
+
{ "name": "MAmmoTH-70B (8-bit)", "size": "38.2GB", "description": "Large mathematical model" },
|
510 |
+
{ "name": "Orca-2-13B", "size": "7.4GB", "description": "Microsoft's reasoning model" },
|
511 |
+
{ "name": "Abel-70B (8-bit)", "size": "38.1GB", "description": "Large reasoning model" }
|
512 |
+
]
|
513 |
+
}
|
514 |
+
},
|
515 |
"ultra_high": { # >32GB
|
516 |
"general": [
|
517 |
{"name": "Llama-2-70B", "size": "130GB", "description": "Full precision", "parameters": "70B", "context": "4K"},
|