NitinBot001 commited on
Commit
fcb7620
·
verified ·
1 Parent(s): d0d9e2f

Create config.yaml

Browse files
Files changed (1) hide show
  1. config.yaml +52 -0
config.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # LocalAI Configuration for fast lightweight models
2
+ name: tinyllama
3
+ backend: llama
4
+ parameters:
5
+ # Model file - will be downloaded automatically
6
+ model: tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf
7
+
8
+ # Performance settings
9
+ threads: 4
10
+ context_size: 2048
11
+ batch: 512
12
+
13
+ # Speed optimizations
14
+ f16: true
15
+ low_vram: true
16
+
17
+ # Generation settings for faster responses
18
+ temperature: 0.1
19
+ top_p: 0.9
20
+ top_k: 10
21
+ repeat_penalty: 1.1
22
+
23
+ # Template for chat format
24
+ template:
25
+ chat: |
26
+ <|system|>
27
+ You are a helpful assistant.</s>
28
+ <|user|>
29
+ {{.Input}}</s>
30
+ <|assistant|>
31
+ completion: |
32
+ {{.Input}}
33
+
34
+ ---
35
+ name: qwen2-0.5b
36
+ backend: llama
37
+ parameters:
38
+ model: qwen2-0_5b-instruct-q4_k_m.gguf
39
+ threads: 4
40
+ context_size: 1024
41
+ batch: 256
42
+ f16: true
43
+ low_vram: true
44
+ temperature: 0.1
45
+ top_p: 0.9
46
+ template:
47
+ chat: |
48
+ <|im_start|>system
49
+ You are a helpful assistant.<|im_end|>
50
+ <|im_start|>user
51
+ {{.Input}}<|im_end|>
52
+ <|im_start|>assistant