nikravan commited on
Commit
f4c4266
·
verified ·
1 Parent(s): 0cd51ce

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +12 -6
app.py CHANGED
@@ -29,14 +29,19 @@ def respond(
29
 
30
  from transformers import AutoModelForCausalLM, AutoTokenizer
31
 
32
- MODEL_PATH = "THUDM/GLM-4-Z1-32B-0414"
33
 
34
- tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
35
- model = AutoModelForCausalLM.from_pretrained(MODEL_PATH, device_map="auto")
36
 
37
- message = [{"role": "user", "content": "Let a, b be positive real numbers such that ab = a + b + 3. Determine the range of possible values for a + b."}]
38
 
39
- .device)
 
 
 
 
 
40
 
41
  generate_kwargs = {
42
  "input_ids": inputs["input_ids"],
@@ -46,7 +51,8 @@ generate_kwargs = {
46
  }
47
  out = model.generate(**generate_kwargs)
48
  print(tokenizer.decode(out[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True))
49
-
 
50
  for message in client.chat_completion(
51
  messages,
52
  max_tokens=max_tokens,
 
29
 
30
  from transformers import AutoModelForCausalLM, AutoTokenizer
31
 
32
+ MODEL_PATH = "THUDM/GLM-4-Z1-32B-0414"
33
 
34
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
35
+ model = AutoModelForCausalLM.from_pretrained(MODEL_PATH, device_map="auto")
36
 
37
+ message = [{"role": "user", "content": "Let a, b be positive real numbers such that ab = a + b + 3. Determine the range of possible values for a + b."}]
38
 
39
+ inputs = tokenizer.apply_chat_template(
40
+ message,
41
+ return_tensors="pt",
42
+ add_generation_prompt=True,
43
+ return_dict=True,
44
+ ).to(model.device)
45
 
46
  generate_kwargs = {
47
  "input_ids": inputs["input_ids"],
 
51
  }
52
  out = model.generate(**generate_kwargs)
53
  print(tokenizer.decode(out[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True))
54
+
55
+
56
  for message in client.chat_completion(
57
  messages,
58
  max_tokens=max_tokens,