nightmedia commited on
Commit
bc4a068
Β·
verified Β·
1 Parent(s): 1e6f5f8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -69,7 +69,7 @@ Base Instruct VLTO Variants
69
  - Formal logic: "Which object has greater inertia, a truck or car?"
70
  ```bash
71
  Base Instruct VLTO Variants
72
- 0.442–0.445 0.435–0.441
73
  ```
74
  βš–οΈ Base instruct wins by tiny margins β€” VLTO models prioritize real-world intuition over textbook logic.
75
 
@@ -82,7 +82,7 @@ This is intentional: multimodal training focuses on "how things work" in practic
82
  Term Meaning
83
  qx85x 5-bit storage for most weights + 8-bit embeddings/attention
84
  qx86x 6-bit storage for most weights + 8-bit embeddings/attention
85
- hi Group size 32 for quantization (finer precision control)
86
  ```
87
  πŸ’‘ The "8-bit" components (embeddings, attention heads) are critical for language tasks β€” protecting them from aggressive compression preserves nuance.
88
 
 
69
  - Formal logic: "Which object has greater inertia, a truck or car?"
70
  ```bash
71
  Base Instruct VLTO Variants
72
+ 0.442–0.445 0.435–0.441
73
  ```
74
  βš–οΈ Base instruct wins by tiny margins β€” VLTO models prioritize real-world intuition over textbook logic.
75
 
 
82
  Term Meaning
83
  qx85x 5-bit storage for most weights + 8-bit embeddings/attention
84
  qx86x 6-bit storage for most weights + 8-bit embeddings/attention
85
+ hi Group size 32 for quantization (finer precision control)
86
  ```
87
  πŸ’‘ The "8-bit" components (embeddings, attention heads) are critical for language tasks β€” protecting them from aggressive compression preserves nuance.
88