BaoKhuong commited on
Commit
b92ef69
·
verified ·
1 Parent(s): 6a696b5

Update model.py

Browse files
Files changed (1) hide show
  1. model.py +1 -1
model.py CHANGED
@@ -104,7 +104,7 @@ def get_finance_llama_pipeline():
104
 
105
  # Configure llama.cpp loader
106
  threads = int(os.getenv("LLAMA_THREADS", str(os.cpu_count() or 2)))
107
- ctx = int(os.getenv("LLAMA_CTX", "4096")) #Mặc định 2048 -> 1024
108
  batch = int(os.getenv("LLAMA_BATCH", "64")) #Mặc định 128 -> 64
109
  use_mmap = os.getenv("LLAMA_MMAP", "1") in {"1","true","True","yes"}
110
  use_mlock = os.getenv("LLAMA_MLOCK", "0") in {"1","true","True","yes"}
 
104
 
105
  # Configure llama.cpp loader
106
  threads = int(os.getenv("LLAMA_THREADS", str(os.cpu_count() or 2)))
107
+ ctx = int(os.getenv("LLAMA_CTX", "131072")) #Mặc định 2048 -> 1024
108
  batch = int(os.getenv("LLAMA_BATCH", "64")) #Mặc định 128 -> 64
109
  use_mmap = os.getenv("LLAMA_MMAP", "1") in {"1","true","True","yes"}
110
  use_mlock = os.getenv("LLAMA_MLOCK", "0") in {"1","true","True","yes"}