Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +1 -0
- README.md +63 -0
- amaas.yaml +85 -0
- chat_template.jinja +118 -0
- config.json +429 -0
- configuration.json +1 -0
- generation_config.json +14 -0
- model-00001-of-00046.safetensors +3 -0
- model-00002-of-00046.safetensors +3 -0
- model-00003-of-00046.safetensors +3 -0
- model-00004-of-00046.safetensors +3 -0
- model-00005-of-00046.safetensors +3 -0
- model-00006-of-00046.safetensors +3 -0
- model-00007-of-00046.safetensors +3 -0
- model-00008-of-00046.safetensors +3 -0
- model-00009-of-00046.safetensors +3 -0
- model-00010-of-00046.safetensors +3 -0
- model-00011-of-00046.safetensors +3 -0
- model-00012-of-00046.safetensors +3 -0
- model-00013-of-00046.safetensors +3 -0
- model-00014-of-00046.safetensors +3 -0
- model-00015-of-00046.safetensors +3 -0
- model-00016-of-00046.safetensors +3 -0
- model-00017-of-00046.safetensors +3 -0
- model-00018-of-00046.safetensors +3 -0
- model-00019-of-00046.safetensors +3 -0
- model-00020-of-00046.safetensors +3 -0
- model-00021-of-00046.safetensors +3 -0
- model-00022-of-00046.safetensors +3 -0
- model-00023-of-00046.safetensors +3 -0
- model-00024-of-00046.safetensors +3 -0
- model-00025-of-00046.safetensors +3 -0
- model-00026-of-00046.safetensors +3 -0
- model-00027-of-00046.safetensors +3 -0
- model-00028-of-00046.safetensors +3 -0
- model-00029-of-00046.safetensors +3 -0
- model-00030-of-00046.safetensors +3 -0
- model-00031-of-00046.safetensors +3 -0
- model-00032-of-00046.safetensors +3 -0
- model-00033-of-00046.safetensors +3 -0
- model-00034-of-00046.safetensors +3 -0
- model-00035-of-00046.safetensors +3 -0
- model-00036-of-00046.safetensors +3 -0
- model-00037-of-00046.safetensors +3 -0
- model-00038-of-00046.safetensors +3 -0
- model-00039-of-00046.safetensors +3 -0
- model-00040-of-00046.safetensors +3 -0
- model-00041-of-00046.safetensors +3 -0
- model-00042-of-00046.safetensors +3 -0
- model-00043-of-00046.safetensors +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
|
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
language:
|
| 4 |
+
- zh
|
| 5 |
+
- en
|
| 6 |
+
base_model:
|
| 7 |
+
- zai-org/GLM-4.5-Air-Base
|
| 8 |
+
pipeline_tag: image-text-to-text
|
| 9 |
+
library_name: transformers
|
| 10 |
+
---
|
| 11 |
+
|
| 12 |
+
# GLM-4.5V
|
| 13 |
+
|
| 14 |
+
<div align="center">
|
| 15 |
+
<img src=https://raw.githubusercontent.com/zai-org/GLM-V/refs/heads/main/resources/logo.svg width="40%"/>
|
| 16 |
+
</div>
|
| 17 |
+
<p align="center">
|
| 18 |
+
👋 加入我们的<a href="https://github.com/zai-org/GLM-V/blob/main/resources/WECHAT.md" target="_blank"> 微信群 </a>。
|
| 19 |
+
<br>
|
| 20 |
+
📖 查看 <a href="https://arxiv.org/abs/2507.01006" target="_blank">论文</a> 。
|
| 21 |
+
<br>
|
| 22 |
+
📍 在 <a href="https://www.bigmodel.cn">智谱大模型开放平台</a> 使用 GLM-V 系列模型的API服务。
|
| 23 |
+
</p>
|
| 24 |
+
|
| 25 |
+
## 模型介绍
|
| 26 |
+
|
| 27 |
+
视觉语言大模型(VLM)已经成为智能系统的关键基石。随着真实世界的智能任务越来越复杂,VLM模型也亟需在基本的多模态感知之外,逐渐增强复杂任务中的推理能力,提升自身的准确性、全面性和智能化程度,使得复杂问题解决、长上下文理解、多模态智能体等智能任务成为可能。
|
| 28 |
+
|
| 29 |
+
我们希望通过我们的开源工作,与社区一起探索技术前沿,同时赋能更多开发者发挥创意做出更多好玩的应用。
|
| 30 |
+
|
| 31 |
+
GLM-4.5V 基于智谱新一代旗舰文本基座模型 GLM-4.5-Air(106B参数,12B激活),延续 GLM-4.1V-Thinking 技术路线,在 42 个公开视觉多模态榜单中综合效果达到同级别开源模型 SOTA 性能,涵盖图像、视频、文档理解以及 GUI Agent 等常见任务。
|
| 32 |
+
|
| 33 |
+

|
| 34 |
+
|
| 35 |
+
在多模态榜单之外,我们更重视模型在真实场景下的表现与可用性。GLM-4.5V 通过高效混合训练,具备覆盖不同种视觉内容的处理能力,实现全场景视觉推理,包括:
|
| 36 |
+
|
| 37 |
+
- 图像推理(场景理解、复杂多图分析、位置识别)
|
| 38 |
+
- 视频理解(长视频分镜分析、事件识别)
|
| 39 |
+
- GUI 任务(屏幕读取、图标识别、桌面操作辅助)
|
| 40 |
+
- 复杂图表与长文档解析(研报分析、信息提取)
|
| 41 |
+
- Grounding 能力(精准定位视觉元素)
|
| 42 |
+
|
| 43 |
+
同时,模型新增 “思考模式” 开关,用户可灵活选择快速响应或深度推理,平衡效率与效果,该开关的使用方式与GLM-4.5 语言模型相同。
|
| 44 |
+
|
| 45 |
+
## 快速开始
|
| 46 |
+
|
| 47 |
+
请通过我们的 [Github](https://github.com/zai-org/GLM-V) 了解更多代码信息。
|
| 48 |
+
|
| 49 |
+
## 引用论文
|
| 50 |
+
|
| 51 |
+
如果您使用了本模型,请引用以下论文:
|
| 52 |
+
|
| 53 |
+
```bibtex
|
| 54 |
+
@misc{glmvteam2025glm41vthinkingversatilemultimodalreasoning,
|
| 55 |
+
title={GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning},
|
| 56 |
+
author={GLM-V Team and Wenyi Hong and Wenmeng Yu and Xiaotao Gu and Guo Wang and Guobing Gan and Haomiao Tang and Jiale Cheng and Ji Qi and Junhui Ji and Lihang Pan and Shuaiqi Duan and Weihan Wang and Yan Wang and Yean Cheng and Zehai He and Zhe Su and Zhen Yang and Ziyang Pan and Aohan Zeng and Baoxu Wang and Boyan Shi and Changyu Pang and Chenhui Zhang and Da Yin and Fan Yang and Guoqing Chen and Jiazheng Xu and Jiali Chen and Jing Chen and Jinhao Chen and Jinghao Lin and Jinjiang Wang and Junjie Chen and Leqi Lei and Letian Gong and Leyi Pan and Mingzhi Zhang and Qinkai Zheng and Sheng Yang and Shi Zhong and Shiyu Huang and Shuyuan Zhao and Siyan Xue and Shangqin Tu and Shengbiao Meng and Tianshu Zhang and Tianwei Luo and Tianxiang Hao and Wenkai Li and Wei Jia and Xin Lyu and Xuancheng Huang and Yanling Wang and Yadong Xue and Yanfeng Wang and Yifan An and Yifan Du and Yiming Shi and Yiheng Huang and Yilin Niu and Yuan Wang and Yuanchang Yue and Yuchen Li and Yutao Zhang and Yuxuan Zhang and Zhanxiao Du and Zhenyu Hou and Zhao Xue and Zhengxiao Du and Zihan Wang and Peng Zhang and Debing Liu and Bin Xu and Juanzi Li and Minlie Huang and Yuxiao Dong and Jie Tang},
|
| 57 |
+
year={2025},
|
| 58 |
+
eprint={2507.01006},
|
| 59 |
+
archivePrefix={arXiv},
|
| 60 |
+
primaryClass={cs.CV},
|
| 61 |
+
url={https://arxiv.org/abs/2507.01006},
|
| 62 |
+
}
|
| 63 |
+
```
|
amaas.yaml
ADDED
|
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
meta:
|
| 2 |
+
backend_type: "ftransformers"
|
| 3 |
+
categories: ""
|
| 4 |
+
dir_path: "GLM-4.5V-GPU-weight"
|
| 5 |
+
display_model_name: "GLM-4.5V-GPU-weight"
|
| 6 |
+
display_model_quanted_type: "BF16"
|
| 7 |
+
family: "ZhipuAI"
|
| 8 |
+
local_path: "GLM-4.5V-GPU-weight"
|
| 9 |
+
model_id: "ApproachingAI2024/GLM-4.5V-GPU-weight"
|
| 10 |
+
name: "GLM-4.5V-GPU-weight"
|
| 11 |
+
parameters:
|
| 12 |
+
cpuinfer: "80"
|
| 13 |
+
amx-weight-path: "ApproachingAI2024/GLM-4.5V-CPU-weight"
|
| 14 |
+
mem-fraction-static: "0.98"
|
| 15 |
+
amx-method: "AMXINT8"
|
| 16 |
+
attention-backend: "flashinfer"
|
| 17 |
+
subpool-count: ""
|
| 18 |
+
tool-call-parser: ""
|
| 19 |
+
quanted_type: "BF16"
|
| 20 |
+
required_disk: "216895848448"
|
| 21 |
+
required_dram: "107946704896"
|
| 22 |
+
required_vram: "121869697024"
|
| 23 |
+
subpool_count: "2"
|
| 24 |
+
type: "vlm"
|
| 25 |
+
weight_size: "106"
|
| 26 |
+
|
| 27 |
+
integrity:
|
| 28 |
+
files:
|
| 29 |
+
"chat_template.jinja": "3858"
|
| 30 |
+
"config.json": "23130"
|
| 31 |
+
"configuration.json": "51"
|
| 32 |
+
"generation_config.json": "234"
|
| 33 |
+
"model-00001-of-00046.safetensors": "2957438824"
|
| 34 |
+
"model-00002-of-00046.safetensors": "2470328656"
|
| 35 |
+
"model-00003-of-00046.safetensors": "2470328656"
|
| 36 |
+
"model-00004-of-00046.safetensors": "2470328656"
|
| 37 |
+
"model-00005-of-00046.safetensors": "2470328656"
|
| 38 |
+
"model-00006-of-00046.safetensors": "2470328656"
|
| 39 |
+
"model-00007-of-00046.safetensors": "2470328656"
|
| 40 |
+
"model-00008-of-00046.safetensors": "2470328656"
|
| 41 |
+
"model-00009-of-00046.safetensors": "2470328656"
|
| 42 |
+
"model-00010-of-00046.safetensors": "2470329824"
|
| 43 |
+
"model-00011-of-00046.safetensors": "2470329824"
|
| 44 |
+
"model-00012-of-00046.safetensors": "2470329824"
|
| 45 |
+
"model-00013-of-00046.safetensors": "2470329824"
|
| 46 |
+
"model-00014-of-00046.safetensors": "2470329824"
|
| 47 |
+
"model-00015-of-00046.safetensors": "2470329824"
|
| 48 |
+
"model-00016-of-00046.safetensors": "2470329824"
|
| 49 |
+
"model-00017-of-00046.safetensors": "2470329824"
|
| 50 |
+
"model-00018-of-00046.safetensors": "2470329824"
|
| 51 |
+
"model-00019-of-00046.safetensors": "2470329824"
|
| 52 |
+
"model-00020-of-00046.safetensors": "2470329824"
|
| 53 |
+
"model-00021-of-00046.safetensors": "2470329824"
|
| 54 |
+
"model-00022-of-00046.safetensors": "2470329824"
|
| 55 |
+
"model-00023-of-00046.safetensors": "2470329824"
|
| 56 |
+
"model-00024-of-00046.safetensors": "2470329824"
|
| 57 |
+
"model-00025-of-00046.safetensors": "2470329824"
|
| 58 |
+
"model-00026-of-00046.safetensors": "2470329824"
|
| 59 |
+
"model-00027-of-00046.safetensors": "2470329824"
|
| 60 |
+
"model-00028-of-00046.safetensors": "2470329824"
|
| 61 |
+
"model-00029-of-00046.safetensors": "2470329824"
|
| 62 |
+
"model-00030-of-00046.safetensors": "2470329824"
|
| 63 |
+
"model-00031-of-00046.safetensors": "2470329824"
|
| 64 |
+
"model-00032-of-00046.safetensors": "2470329824"
|
| 65 |
+
"model-00033-of-00046.safetensors": "2470329824"
|
| 66 |
+
"model-00034-of-00046.safetensors": "2470329824"
|
| 67 |
+
"model-00035-of-00046.safetensors": "2470329824"
|
| 68 |
+
"model-00036-of-00046.safetensors": "2470329824"
|
| 69 |
+
"model-00037-of-00046.safetensors": "2470329824"
|
| 70 |
+
"model-00038-of-00046.safetensors": "2470329824"
|
| 71 |
+
"model-00039-of-00046.safetensors": "2470329824"
|
| 72 |
+
"model-00040-of-00046.safetensors": "2470329824"
|
| 73 |
+
"model-00041-of-00046.safetensors": "2470329824"
|
| 74 |
+
"model-00042-of-00046.safetensors": "2470329824"
|
| 75 |
+
"model-00043-of-00046.safetensors": "2470329824"
|
| 76 |
+
"model-00044-of-00046.safetensors": "2470329824"
|
| 77 |
+
"model-00045-of-00046.safetensors": "3149890536"
|
| 78 |
+
"model-00046-of-00046.safetensors": "3520860512"
|
| 79 |
+
"model.safetensors.index.json": "5865342"
|
| 80 |
+
"preprocessor_config.json": "364"
|
| 81 |
+
"README.md": "4043"
|
| 82 |
+
"tokenizer_config.json": "7307"
|
| 83 |
+
"tokenizer.json": "19970699"
|
| 84 |
+
"video_preprocessor_config.json": "365"
|
| 85 |
+
sha256: "e903ac50faf8b68c7d469c20b309a8e2670066adc09cff02f65ca4ea2d17e444"
|
chat_template.jinja
ADDED
|
@@ -0,0 +1,118 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[gMASK]<sop>
|
| 2 |
+
{%- if tools -%}
|
| 3 |
+
<|system|>
|
| 4 |
+
# Tools
|
| 5 |
+
|
| 6 |
+
You may call one or more functions to assist with the user query.
|
| 7 |
+
|
| 8 |
+
You are provided with function signatures within <tools></tools> XML tags:
|
| 9 |
+
<tools>
|
| 10 |
+
{% for tool in tools %}
|
| 11 |
+
{{ tool | tojson(ensure_ascii=False) }}
|
| 12 |
+
{% endfor %}
|
| 13 |
+
</tools>
|
| 14 |
+
|
| 15 |
+
For each function call, output the function name and arguments within the following XML format:
|
| 16 |
+
<tool_call>{function-name}
|
| 17 |
+
<arg_key>{arg-key-1}</arg_key>
|
| 18 |
+
<arg_value>{arg-value-1}</arg_value>
|
| 19 |
+
<arg_key>{arg-key-2}</arg_key>
|
| 20 |
+
<arg_value>{arg-value-2}</arg_value>
|
| 21 |
+
...
|
| 22 |
+
</tool_call>{%- endif -%}
|
| 23 |
+
{%- macro visible_text(content) -%}
|
| 24 |
+
{%- if content is string -%}
|
| 25 |
+
{{- content }}
|
| 26 |
+
{%- elif content is iterable and content is not mapping -%}
|
| 27 |
+
{%- for item in content -%}
|
| 28 |
+
{%- if item is mapping and item.type == 'text' -%}
|
| 29 |
+
{{- item.text }}
|
| 30 |
+
{%- elif item is mapping and (item.type == 'image' or 'image' in item) -%}
|
| 31 |
+
<|begin_of_image|><|image|><|end_of_image|>
|
| 32 |
+
{%- elif item is mapping and (item.type == 'video' or 'video' in item) -%}
|
| 33 |
+
<|begin_of_video|><|video|><|end_of_video|>
|
| 34 |
+
{%- elif item is string -%}
|
| 35 |
+
{{- item }}
|
| 36 |
+
{%- endif -%}
|
| 37 |
+
{%- endfor -%}
|
| 38 |
+
{%- else -%}
|
| 39 |
+
{{- content }}
|
| 40 |
+
{%- endif -%}
|
| 41 |
+
{%- endmacro -%}
|
| 42 |
+
{%- set ns = namespace(last_user_index=-1) %}
|
| 43 |
+
{%- for m in messages %}
|
| 44 |
+
{%- if m.role == 'user' %}
|
| 45 |
+
{% set ns.last_user_index = loop.index0 -%}
|
| 46 |
+
{%- endif %}
|
| 47 |
+
{%- endfor %}
|
| 48 |
+
{% for m in messages %}
|
| 49 |
+
{%- if m.role == 'user' -%}<|user|>
|
| 50 |
+
{% if m.content is string %}
|
| 51 |
+
{{ m.content }}
|
| 52 |
+
{%- else %}
|
| 53 |
+
{%- for item in m.content %}
|
| 54 |
+
{% if item.type == 'video' or 'video' in item %}
|
| 55 |
+
<|begin_of_video|><|video|><|end_of_video|>{% elif item.type == 'image' or 'image' in item %}
|
| 56 |
+
<|begin_of_image|><|image|><|end_of_image|>{% elif item.type == 'text' %}
|
| 57 |
+
{{ item.text }}
|
| 58 |
+
{%- endif %}
|
| 59 |
+
{%- endfor %}
|
| 60 |
+
{%- endif %}
|
| 61 |
+
{{- '/nothink' if (enable_thinking is defined and not enable_thinking and not visible_text(m.content).endswith("/nothink")) else '' -}}
|
| 62 |
+
{%- elif m.role == 'assistant' -%}
|
| 63 |
+
<|assistant|>
|
| 64 |
+
{%- set reasoning_content = '' %}
|
| 65 |
+
{%- set content = visible_text(m.content) %}
|
| 66 |
+
{%- if m.reasoning_content is string %}
|
| 67 |
+
{%- set reasoning_content = m.reasoning_content %}
|
| 68 |
+
{%- else %}
|
| 69 |
+
{%- if '</think>' in content %}
|
| 70 |
+
{%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
|
| 71 |
+
{%- set content = content.split('</think>')[-1].lstrip('\n') %}
|
| 72 |
+
{%- endif %}
|
| 73 |
+
{%- endif %}
|
| 74 |
+
{%- if loop.index0 > ns.last_user_index and reasoning_content -%}
|
| 75 |
+
{{ '\n<think>' + reasoning_content.strip() + '</think>'}}
|
| 76 |
+
{%- else -%}
|
| 77 |
+
{{ '\n<think></think>' }}
|
| 78 |
+
{%- endif -%}
|
| 79 |
+
{%- if content.strip() -%}
|
| 80 |
+
{{ '\n' + content.strip() }}
|
| 81 |
+
{%- endif -%}
|
| 82 |
+
{% if m.tool_calls %}
|
| 83 |
+
{% for tc in m.tool_calls %}
|
| 84 |
+
{%- if tc.function %}
|
| 85 |
+
{%- set tc = tc.function %}
|
| 86 |
+
{%- endif %}
|
| 87 |
+
{{ '\n<tool_call>' + tc.name }}
|
| 88 |
+
{% set _args = tc.arguments %}
|
| 89 |
+
{% for k, v in _args.items() %}
|
| 90 |
+
<arg_key>{{ k }}</arg_key>
|
| 91 |
+
<arg_value>{{ v | tojson(ensure_ascii=False) if v is not string else v }}</arg_value>
|
| 92 |
+
{% endfor %}
|
| 93 |
+
</tool_call>{% endfor %}
|
| 94 |
+
{% endif %}
|
| 95 |
+
{%- elif m.role == 'tool' -%}
|
| 96 |
+
{%- if m.content is string -%}
|
| 97 |
+
{%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
|
| 98 |
+
{{- '<|observation|>' }}
|
| 99 |
+
{%- endif %}
|
| 100 |
+
{{- '\n<tool_response>\n' }}
|
| 101 |
+
{{- m.content }}
|
| 102 |
+
{{- '\n</tool_response>' }}
|
| 103 |
+
{%- else -%}
|
| 104 |
+
<|observation|>{% for tr in m.content %}
|
| 105 |
+
|
| 106 |
+
<tool_response>
|
| 107 |
+
{{ tr.output if tr.output is defined else tr }}
|
| 108 |
+
</tool_response>{% endfor -%}
|
| 109 |
+
{% endif -%}
|
| 110 |
+
{%- elif m.role == 'system' -%}
|
| 111 |
+
<|system|>
|
| 112 |
+
{{ visible_text(m.content) }}
|
| 113 |
+
{%- endif -%}
|
| 114 |
+
{%- endfor -%}
|
| 115 |
+
{%- if add_generation_prompt -%}
|
| 116 |
+
<|assistant|>
|
| 117 |
+
{{'<think></think>\n' if (enable_thinking is defined and not enable_thinking) else ''}}
|
| 118 |
+
{%- endif -%}
|
config.json
ADDED
|
@@ -0,0 +1,429 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"Glm4vMoeForConditionalGeneration"
|
| 4 |
+
],
|
| 5 |
+
"model_type": "glm4v_moe",
|
| 6 |
+
"text_config": {
|
| 7 |
+
"pad_token_id": 151329,
|
| 8 |
+
"vocab_size": 151552,
|
| 9 |
+
"eos_token_id": [
|
| 10 |
+
151329,
|
| 11 |
+
151336,
|
| 12 |
+
151338
|
| 13 |
+
],
|
| 14 |
+
"image_end_token_id": 151340,
|
| 15 |
+
"image_start_token_id": 151339,
|
| 16 |
+
"image_token_id": 151363,
|
| 17 |
+
"head_dim": 128,
|
| 18 |
+
"attention_bias": true,
|
| 19 |
+
"attention_dropout": 0.0,
|
| 20 |
+
"first_k_dense_replace": 1,
|
| 21 |
+
"hidden_act": "silu",
|
| 22 |
+
"hidden_size": 4096,
|
| 23 |
+
"initializer_range": 0.02,
|
| 24 |
+
"intermediate_size": 10944,
|
| 25 |
+
"max_position_embeddings": 65536,
|
| 26 |
+
"model_type": "glm4v_moe_text",
|
| 27 |
+
"moe_intermediate_size": 1408,
|
| 28 |
+
"n_group": 1,
|
| 29 |
+
"n_routed_experts": 128,
|
| 30 |
+
"n_shared_experts": 1,
|
| 31 |
+
"norm_topk_prob": true,
|
| 32 |
+
"num_attention_heads": 96,
|
| 33 |
+
"num_experts_per_tok": 8,
|
| 34 |
+
"num_hidden_layers": 46,
|
| 35 |
+
"num_key_value_heads": 8,
|
| 36 |
+
"partial_rotary_factor": 0.5,
|
| 37 |
+
"quantization_config": {
|
| 38 |
+
"config_groups": {
|
| 39 |
+
"group_0": {
|
| 40 |
+
"input_activations": null,
|
| 41 |
+
"output_activations": null,
|
| 42 |
+
"targets": [
|
| 43 |
+
"Linear"
|
| 44 |
+
],
|
| 45 |
+
"weights": {
|
| 46 |
+
"actorder": null,
|
| 47 |
+
"block_structure": null,
|
| 48 |
+
"dynamic": false,
|
| 49 |
+
"group_size": null,
|
| 50 |
+
"num_bits": 8,
|
| 51 |
+
"observer": "minmax",
|
| 52 |
+
"observer_kwargs": {},
|
| 53 |
+
"strategy": "channel",
|
| 54 |
+
"symmetric": true,
|
| 55 |
+
"type": "int"
|
| 56 |
+
}
|
| 57 |
+
}
|
| 58 |
+
},
|
| 59 |
+
"format": "pack-quantized",
|
| 60 |
+
"global_compression_ratio": null,
|
| 61 |
+
"ignore": [
|
| 62 |
+
"model.language_model.layers.0.self_attn.q_proj",
|
| 63 |
+
"model.language_model.layers.0.self_attn.k_proj",
|
| 64 |
+
"model.language_model.layers.0.self_attn.v_proj",
|
| 65 |
+
"model.language_model.layers.0.self_attn.o_proj",
|
| 66 |
+
"model.language_model.layers.0.mlp.gate_proj",
|
| 67 |
+
"model.language_model.layers.0.mlp.up_proj",
|
| 68 |
+
"model.language_model.layers.0.mlp.down_proj",
|
| 69 |
+
"model.language_model.layers.1.self_attn.q_proj",
|
| 70 |
+
"model.language_model.layers.1.self_attn.k_proj",
|
| 71 |
+
"model.language_model.layers.1.self_attn.v_proj",
|
| 72 |
+
"model.language_model.layers.1.self_attn.o_proj",
|
| 73 |
+
"model.language_model.layers.1.mlp.shared_experts.gate_proj",
|
| 74 |
+
"model.language_model.layers.1.mlp.shared_experts.up_proj",
|
| 75 |
+
"model.language_model.layers.1.mlp.shared_experts.down_proj",
|
| 76 |
+
"model.language_model.layers.2.self_attn.q_proj",
|
| 77 |
+
"model.language_model.layers.2.self_attn.k_proj",
|
| 78 |
+
"model.language_model.layers.2.self_attn.v_proj",
|
| 79 |
+
"model.language_model.layers.2.self_attn.o_proj",
|
| 80 |
+
"model.language_model.layers.2.mlp.shared_experts.gate_proj",
|
| 81 |
+
"model.language_model.layers.2.mlp.shared_experts.up_proj",
|
| 82 |
+
"model.language_model.layers.2.mlp.shared_experts.down_proj",
|
| 83 |
+
"model.language_model.layers.3.self_attn.q_proj",
|
| 84 |
+
"model.language_model.layers.3.self_attn.k_proj",
|
| 85 |
+
"model.language_model.layers.3.self_attn.v_proj",
|
| 86 |
+
"model.language_model.layers.3.self_attn.o_proj",
|
| 87 |
+
"model.language_model.layers.3.mlp.shared_experts.gate_proj",
|
| 88 |
+
"model.language_model.layers.3.mlp.shared_experts.up_proj",
|
| 89 |
+
"model.language_model.layers.3.mlp.shared_experts.down_proj",
|
| 90 |
+
"model.language_model.layers.4.self_attn.q_proj",
|
| 91 |
+
"model.language_model.layers.4.self_attn.k_proj",
|
| 92 |
+
"model.language_model.layers.4.self_attn.v_proj",
|
| 93 |
+
"model.language_model.layers.4.self_attn.o_proj",
|
| 94 |
+
"model.language_model.layers.4.mlp.shared_experts.gate_proj",
|
| 95 |
+
"model.language_model.layers.4.mlp.shared_experts.up_proj",
|
| 96 |
+
"model.language_model.layers.4.mlp.shared_experts.down_proj",
|
| 97 |
+
"model.language_model.layers.5.self_attn.q_proj",
|
| 98 |
+
"model.language_model.layers.5.self_attn.k_proj",
|
| 99 |
+
"model.language_model.layers.5.self_attn.v_proj",
|
| 100 |
+
"model.language_model.layers.5.self_attn.o_proj",
|
| 101 |
+
"model.language_model.layers.5.mlp.shared_experts.gate_proj",
|
| 102 |
+
"model.language_model.layers.5.mlp.shared_experts.up_proj",
|
| 103 |
+
"model.language_model.layers.5.mlp.shared_experts.down_proj",
|
| 104 |
+
"model.language_model.layers.6.self_attn.q_proj",
|
| 105 |
+
"model.language_model.layers.6.self_attn.k_proj",
|
| 106 |
+
"model.language_model.layers.6.self_attn.v_proj",
|
| 107 |
+
"model.language_model.layers.6.self_attn.o_proj",
|
| 108 |
+
"model.language_model.layers.6.mlp.shared_experts.gate_proj",
|
| 109 |
+
"model.language_model.layers.6.mlp.shared_experts.up_proj",
|
| 110 |
+
"model.language_model.layers.6.mlp.shared_experts.down_proj",
|
| 111 |
+
"model.language_model.layers.7.self_attn.q_proj",
|
| 112 |
+
"model.language_model.layers.7.self_attn.k_proj",
|
| 113 |
+
"model.language_model.layers.7.self_attn.v_proj",
|
| 114 |
+
"model.language_model.layers.7.self_attn.o_proj",
|
| 115 |
+
"model.language_model.layers.7.mlp.shared_experts.gate_proj",
|
| 116 |
+
"model.language_model.layers.7.mlp.shared_experts.up_proj",
|
| 117 |
+
"model.language_model.layers.7.mlp.shared_experts.down_proj",
|
| 118 |
+
"model.language_model.layers.8.self_attn.q_proj",
|
| 119 |
+
"model.language_model.layers.8.self_attn.k_proj",
|
| 120 |
+
"model.language_model.layers.8.self_attn.v_proj",
|
| 121 |
+
"model.language_model.layers.8.self_attn.o_proj",
|
| 122 |
+
"model.language_model.layers.8.mlp.shared_experts.gate_proj",
|
| 123 |
+
"model.language_model.layers.8.mlp.shared_experts.up_proj",
|
| 124 |
+
"model.language_model.layers.8.mlp.shared_experts.down_proj",
|
| 125 |
+
"model.language_model.layers.9.self_attn.q_proj",
|
| 126 |
+
"model.language_model.layers.9.self_attn.k_proj",
|
| 127 |
+
"model.language_model.layers.9.self_attn.v_proj",
|
| 128 |
+
"model.language_model.layers.9.self_attn.o_proj",
|
| 129 |
+
"model.language_model.layers.9.mlp.shared_experts.gate_proj",
|
| 130 |
+
"model.language_model.layers.9.mlp.shared_experts.up_proj",
|
| 131 |
+
"model.language_model.layers.9.mlp.shared_experts.down_proj",
|
| 132 |
+
"model.language_model.layers.10.self_attn.q_proj",
|
| 133 |
+
"model.language_model.layers.10.self_attn.k_proj",
|
| 134 |
+
"model.language_model.layers.10.self_attn.v_proj",
|
| 135 |
+
"model.language_model.layers.10.self_attn.o_proj",
|
| 136 |
+
"model.language_model.layers.10.mlp.shared_experts.gate_proj",
|
| 137 |
+
"model.language_model.layers.10.mlp.shared_experts.up_proj",
|
| 138 |
+
"model.language_model.layers.10.mlp.shared_experts.down_proj",
|
| 139 |
+
"model.language_model.layers.11.self_attn.q_proj",
|
| 140 |
+
"model.language_model.layers.11.self_attn.k_proj",
|
| 141 |
+
"model.language_model.layers.11.self_attn.v_proj",
|
| 142 |
+
"model.language_model.layers.11.self_attn.o_proj",
|
| 143 |
+
"model.language_model.layers.11.mlp.shared_experts.gate_proj",
|
| 144 |
+
"model.language_model.layers.11.mlp.shared_experts.up_proj",
|
| 145 |
+
"model.language_model.layers.11.mlp.shared_experts.down_proj",
|
| 146 |
+
"model.language_model.layers.12.self_attn.q_proj",
|
| 147 |
+
"model.language_model.layers.12.self_attn.k_proj",
|
| 148 |
+
"model.language_model.layers.12.self_attn.v_proj",
|
| 149 |
+
"model.language_model.layers.12.self_attn.o_proj",
|
| 150 |
+
"model.language_model.layers.12.mlp.shared_experts.gate_proj",
|
| 151 |
+
"model.language_model.layers.12.mlp.shared_experts.up_proj",
|
| 152 |
+
"model.language_model.layers.12.mlp.shared_experts.down_proj",
|
| 153 |
+
"model.language_model.layers.13.self_attn.q_proj",
|
| 154 |
+
"model.language_model.layers.13.self_attn.k_proj",
|
| 155 |
+
"model.language_model.layers.13.self_attn.v_proj",
|
| 156 |
+
"model.language_model.layers.13.self_attn.o_proj",
|
| 157 |
+
"model.language_model.layers.13.mlp.shared_experts.gate_proj",
|
| 158 |
+
"model.language_model.layers.13.mlp.shared_experts.up_proj",
|
| 159 |
+
"model.language_model.layers.13.mlp.shared_experts.down_proj",
|
| 160 |
+
"model.language_model.layers.14.self_attn.q_proj",
|
| 161 |
+
"model.language_model.layers.14.self_attn.k_proj",
|
| 162 |
+
"model.language_model.layers.14.self_attn.v_proj",
|
| 163 |
+
"model.language_model.layers.14.self_attn.o_proj",
|
| 164 |
+
"model.language_model.layers.14.mlp.shared_experts.gate_proj",
|
| 165 |
+
"model.language_model.layers.14.mlp.shared_experts.up_proj",
|
| 166 |
+
"model.language_model.layers.14.mlp.shared_experts.down_proj",
|
| 167 |
+
"model.language_model.layers.15.self_attn.q_proj",
|
| 168 |
+
"model.language_model.layers.15.self_attn.k_proj",
|
| 169 |
+
"model.language_model.layers.15.self_attn.v_proj",
|
| 170 |
+
"model.language_model.layers.15.self_attn.o_proj",
|
| 171 |
+
"model.language_model.layers.15.mlp.shared_experts.gate_proj",
|
| 172 |
+
"model.language_model.layers.15.mlp.shared_experts.up_proj",
|
| 173 |
+
"model.language_model.layers.15.mlp.shared_experts.down_proj",
|
| 174 |
+
"model.language_model.layers.16.self_attn.q_proj",
|
| 175 |
+
"model.language_model.layers.16.self_attn.k_proj",
|
| 176 |
+
"model.language_model.layers.16.self_attn.v_proj",
|
| 177 |
+
"model.language_model.layers.16.self_attn.o_proj",
|
| 178 |
+
"model.language_model.layers.16.mlp.shared_experts.gate_proj",
|
| 179 |
+
"model.language_model.layers.16.mlp.shared_experts.up_proj",
|
| 180 |
+
"model.language_model.layers.16.mlp.shared_experts.down_proj",
|
| 181 |
+
"model.language_model.layers.17.self_attn.q_proj",
|
| 182 |
+
"model.language_model.layers.17.self_attn.k_proj",
|
| 183 |
+
"model.language_model.layers.17.self_attn.v_proj",
|
| 184 |
+
"model.language_model.layers.17.self_attn.o_proj",
|
| 185 |
+
"model.language_model.layers.17.mlp.shared_experts.gate_proj",
|
| 186 |
+
"model.language_model.layers.17.mlp.shared_experts.up_proj",
|
| 187 |
+
"model.language_model.layers.17.mlp.shared_experts.down_proj",
|
| 188 |
+
"model.language_model.layers.18.self_attn.q_proj",
|
| 189 |
+
"model.language_model.layers.18.self_attn.k_proj",
|
| 190 |
+
"model.language_model.layers.18.self_attn.v_proj",
|
| 191 |
+
"model.language_model.layers.18.self_attn.o_proj",
|
| 192 |
+
"model.language_model.layers.18.mlp.shared_experts.gate_proj",
|
| 193 |
+
"model.language_model.layers.18.mlp.shared_experts.up_proj",
|
| 194 |
+
"model.language_model.layers.18.mlp.shared_experts.down_proj",
|
| 195 |
+
"model.language_model.layers.19.self_attn.q_proj",
|
| 196 |
+
"model.language_model.layers.19.self_attn.k_proj",
|
| 197 |
+
"model.language_model.layers.19.self_attn.v_proj",
|
| 198 |
+
"model.language_model.layers.19.self_attn.o_proj",
|
| 199 |
+
"model.language_model.layers.19.mlp.shared_experts.gate_proj",
|
| 200 |
+
"model.language_model.layers.19.mlp.shared_experts.up_proj",
|
| 201 |
+
"model.language_model.layers.19.mlp.shared_experts.down_proj",
|
| 202 |
+
"model.language_model.layers.20.self_attn.q_proj",
|
| 203 |
+
"model.language_model.layers.20.self_attn.k_proj",
|
| 204 |
+
"model.language_model.layers.20.self_attn.v_proj",
|
| 205 |
+
"model.language_model.layers.20.self_attn.o_proj",
|
| 206 |
+
"model.language_model.layers.20.mlp.shared_experts.gate_proj",
|
| 207 |
+
"model.language_model.layers.20.mlp.shared_experts.up_proj",
|
| 208 |
+
"model.language_model.layers.20.mlp.shared_experts.down_proj",
|
| 209 |
+
"model.language_model.layers.21.self_attn.q_proj",
|
| 210 |
+
"model.language_model.layers.21.self_attn.k_proj",
|
| 211 |
+
"model.language_model.layers.21.self_attn.v_proj",
|
| 212 |
+
"model.language_model.layers.21.self_attn.o_proj",
|
| 213 |
+
"model.language_model.layers.21.mlp.shared_experts.gate_proj",
|
| 214 |
+
"model.language_model.layers.21.mlp.shared_experts.up_proj",
|
| 215 |
+
"model.language_model.layers.21.mlp.shared_experts.down_proj",
|
| 216 |
+
"model.language_model.layers.22.self_attn.q_proj",
|
| 217 |
+
"model.language_model.layers.22.self_attn.k_proj",
|
| 218 |
+
"model.language_model.layers.22.self_attn.v_proj",
|
| 219 |
+
"model.language_model.layers.22.self_attn.o_proj",
|
| 220 |
+
"model.language_model.layers.22.mlp.shared_experts.gate_proj",
|
| 221 |
+
"model.language_model.layers.22.mlp.shared_experts.up_proj",
|
| 222 |
+
"model.language_model.layers.22.mlp.shared_experts.down_proj",
|
| 223 |
+
"model.language_model.layers.23.self_attn.q_proj",
|
| 224 |
+
"model.language_model.layers.23.self_attn.k_proj",
|
| 225 |
+
"model.language_model.layers.23.self_attn.v_proj",
|
| 226 |
+
"model.language_model.layers.23.self_attn.o_proj",
|
| 227 |
+
"model.language_model.layers.23.mlp.shared_experts.gate_proj",
|
| 228 |
+
"model.language_model.layers.23.mlp.shared_experts.up_proj",
|
| 229 |
+
"model.language_model.layers.23.mlp.shared_experts.down_proj",
|
| 230 |
+
"model.language_model.layers.24.self_attn.q_proj",
|
| 231 |
+
"model.language_model.layers.24.self_attn.k_proj",
|
| 232 |
+
"model.language_model.layers.24.self_attn.v_proj",
|
| 233 |
+
"model.language_model.layers.24.self_attn.o_proj",
|
| 234 |
+
"model.language_model.layers.24.mlp.shared_experts.gate_proj",
|
| 235 |
+
"model.language_model.layers.24.mlp.shared_experts.up_proj",
|
| 236 |
+
"model.language_model.layers.24.mlp.shared_experts.down_proj",
|
| 237 |
+
"model.language_model.layers.25.self_attn.q_proj",
|
| 238 |
+
"model.language_model.layers.25.self_attn.k_proj",
|
| 239 |
+
"model.language_model.layers.25.self_attn.v_proj",
|
| 240 |
+
"model.language_model.layers.25.self_attn.o_proj",
|
| 241 |
+
"model.language_model.layers.25.mlp.shared_experts.gate_proj",
|
| 242 |
+
"model.language_model.layers.25.mlp.shared_experts.up_proj",
|
| 243 |
+
"model.language_model.layers.25.mlp.shared_experts.down_proj",
|
| 244 |
+
"model.language_model.layers.26.self_attn.q_proj",
|
| 245 |
+
"model.language_model.layers.26.self_attn.k_proj",
|
| 246 |
+
"model.language_model.layers.26.self_attn.v_proj",
|
| 247 |
+
"model.language_model.layers.26.self_attn.o_proj",
|
| 248 |
+
"model.language_model.layers.26.mlp.shared_experts.gate_proj",
|
| 249 |
+
"model.language_model.layers.26.mlp.shared_experts.up_proj",
|
| 250 |
+
"model.language_model.layers.26.mlp.shared_experts.down_proj",
|
| 251 |
+
"model.language_model.layers.27.self_attn.q_proj",
|
| 252 |
+
"model.language_model.layers.27.self_attn.k_proj",
|
| 253 |
+
"model.language_model.layers.27.self_attn.v_proj",
|
| 254 |
+
"model.language_model.layers.27.self_attn.o_proj",
|
| 255 |
+
"model.language_model.layers.27.mlp.shared_experts.gate_proj",
|
| 256 |
+
"model.language_model.layers.27.mlp.shared_experts.up_proj",
|
| 257 |
+
"model.language_model.layers.27.mlp.shared_experts.down_proj",
|
| 258 |
+
"model.language_model.layers.28.self_attn.q_proj",
|
| 259 |
+
"model.language_model.layers.28.self_attn.k_proj",
|
| 260 |
+
"model.language_model.layers.28.self_attn.v_proj",
|
| 261 |
+
"model.language_model.layers.28.self_attn.o_proj",
|
| 262 |
+
"model.language_model.layers.28.mlp.shared_experts.gate_proj",
|
| 263 |
+
"model.language_model.layers.28.mlp.shared_experts.up_proj",
|
| 264 |
+
"model.language_model.layers.28.mlp.shared_experts.down_proj",
|
| 265 |
+
"model.language_model.layers.29.self_attn.q_proj",
|
| 266 |
+
"model.language_model.layers.29.self_attn.k_proj",
|
| 267 |
+
"model.language_model.layers.29.self_attn.v_proj",
|
| 268 |
+
"model.language_model.layers.29.self_attn.o_proj",
|
| 269 |
+
"model.language_model.layers.29.mlp.shared_experts.gate_proj",
|
| 270 |
+
"model.language_model.layers.29.mlp.shared_experts.up_proj",
|
| 271 |
+
"model.language_model.layers.29.mlp.shared_experts.down_proj",
|
| 272 |
+
"model.language_model.layers.30.self_attn.q_proj",
|
| 273 |
+
"model.language_model.layers.30.self_attn.k_proj",
|
| 274 |
+
"model.language_model.layers.30.self_attn.v_proj",
|
| 275 |
+
"model.language_model.layers.30.self_attn.o_proj",
|
| 276 |
+
"model.language_model.layers.30.mlp.shared_experts.gate_proj",
|
| 277 |
+
"model.language_model.layers.30.mlp.shared_experts.up_proj",
|
| 278 |
+
"model.language_model.layers.30.mlp.shared_experts.down_proj",
|
| 279 |
+
"model.language_model.layers.31.self_attn.q_proj",
|
| 280 |
+
"model.language_model.layers.31.self_attn.k_proj",
|
| 281 |
+
"model.language_model.layers.31.self_attn.v_proj",
|
| 282 |
+
"model.language_model.layers.31.self_attn.o_proj",
|
| 283 |
+
"model.language_model.layers.31.mlp.shared_experts.gate_proj",
|
| 284 |
+
"model.language_model.layers.31.mlp.shared_experts.up_proj",
|
| 285 |
+
"model.language_model.layers.31.mlp.shared_experts.down_proj",
|
| 286 |
+
"model.language_model.layers.32.self_attn.q_proj",
|
| 287 |
+
"model.language_model.layers.32.self_attn.k_proj",
|
| 288 |
+
"model.language_model.layers.32.self_attn.v_proj",
|
| 289 |
+
"model.language_model.layers.32.self_attn.o_proj",
|
| 290 |
+
"model.language_model.layers.32.mlp.shared_experts.gate_proj",
|
| 291 |
+
"model.language_model.layers.32.mlp.shared_experts.up_proj",
|
| 292 |
+
"model.language_model.layers.32.mlp.shared_experts.down_proj",
|
| 293 |
+
"model.language_model.layers.33.self_attn.q_proj",
|
| 294 |
+
"model.language_model.layers.33.self_attn.k_proj",
|
| 295 |
+
"model.language_model.layers.33.self_attn.v_proj",
|
| 296 |
+
"model.language_model.layers.33.self_attn.o_proj",
|
| 297 |
+
"model.language_model.layers.33.mlp.shared_experts.gate_proj",
|
| 298 |
+
"model.language_model.layers.33.mlp.shared_experts.up_proj",
|
| 299 |
+
"model.language_model.layers.33.mlp.shared_experts.down_proj",
|
| 300 |
+
"model.language_model.layers.34.self_attn.q_proj",
|
| 301 |
+
"model.language_model.layers.34.self_attn.k_proj",
|
| 302 |
+
"model.language_model.layers.34.self_attn.v_proj",
|
| 303 |
+
"model.language_model.layers.34.self_attn.o_proj",
|
| 304 |
+
"model.language_model.layers.34.mlp.shared_experts.gate_proj",
|
| 305 |
+
"model.language_model.layers.34.mlp.shared_experts.up_proj",
|
| 306 |
+
"model.language_model.layers.34.mlp.shared_experts.down_proj",
|
| 307 |
+
"model.language_model.layers.35.self_attn.q_proj",
|
| 308 |
+
"model.language_model.layers.35.self_attn.k_proj",
|
| 309 |
+
"model.language_model.layers.35.self_attn.v_proj",
|
| 310 |
+
"model.language_model.layers.35.self_attn.o_proj",
|
| 311 |
+
"model.language_model.layers.35.mlp.shared_experts.gate_proj",
|
| 312 |
+
"model.language_model.layers.35.mlp.shared_experts.up_proj",
|
| 313 |
+
"model.language_model.layers.35.mlp.shared_experts.down_proj",
|
| 314 |
+
"model.language_model.layers.36.self_attn.q_proj",
|
| 315 |
+
"model.language_model.layers.36.self_attn.k_proj",
|
| 316 |
+
"model.language_model.layers.36.self_attn.v_proj",
|
| 317 |
+
"model.language_model.layers.36.self_attn.o_proj",
|
| 318 |
+
"model.language_model.layers.36.mlp.shared_experts.gate_proj",
|
| 319 |
+
"model.language_model.layers.36.mlp.shared_experts.up_proj",
|
| 320 |
+
"model.language_model.layers.36.mlp.shared_experts.down_proj",
|
| 321 |
+
"model.language_model.layers.37.self_attn.q_proj",
|
| 322 |
+
"model.language_model.layers.37.self_attn.k_proj",
|
| 323 |
+
"model.language_model.layers.37.self_attn.v_proj",
|
| 324 |
+
"model.language_model.layers.37.self_attn.o_proj",
|
| 325 |
+
"model.language_model.layers.37.mlp.shared_experts.gate_proj",
|
| 326 |
+
"model.language_model.layers.37.mlp.shared_experts.up_proj",
|
| 327 |
+
"model.language_model.layers.37.mlp.shared_experts.down_proj",
|
| 328 |
+
"model.language_model.layers.38.self_attn.q_proj",
|
| 329 |
+
"model.language_model.layers.38.self_attn.k_proj",
|
| 330 |
+
"model.language_model.layers.38.self_attn.v_proj",
|
| 331 |
+
"model.language_model.layers.38.self_attn.o_proj",
|
| 332 |
+
"model.language_model.layers.38.mlp.shared_experts.gate_proj",
|
| 333 |
+
"model.language_model.layers.38.mlp.shared_experts.up_proj",
|
| 334 |
+
"model.language_model.layers.38.mlp.shared_experts.down_proj",
|
| 335 |
+
"model.language_model.layers.39.self_attn.q_proj",
|
| 336 |
+
"model.language_model.layers.39.self_attn.k_proj",
|
| 337 |
+
"model.language_model.layers.39.self_attn.v_proj",
|
| 338 |
+
"model.language_model.layers.39.self_attn.o_proj",
|
| 339 |
+
"model.language_model.layers.39.mlp.shared_experts.gate_proj",
|
| 340 |
+
"model.language_model.layers.39.mlp.shared_experts.up_proj",
|
| 341 |
+
"model.language_model.layers.39.mlp.shared_experts.down_proj",
|
| 342 |
+
"model.language_model.layers.40.self_attn.q_proj",
|
| 343 |
+
"model.language_model.layers.40.self_attn.k_proj",
|
| 344 |
+
"model.language_model.layers.40.self_attn.v_proj",
|
| 345 |
+
"model.language_model.layers.40.self_attn.o_proj",
|
| 346 |
+
"model.language_model.layers.40.mlp.shared_experts.gate_proj",
|
| 347 |
+
"model.language_model.layers.40.mlp.shared_experts.up_proj",
|
| 348 |
+
"model.language_model.layers.40.mlp.shared_experts.down_proj",
|
| 349 |
+
"model.language_model.layers.41.self_attn.q_proj",
|
| 350 |
+
"model.language_model.layers.41.self_attn.k_proj",
|
| 351 |
+
"model.language_model.layers.41.self_attn.v_proj",
|
| 352 |
+
"model.language_model.layers.41.self_attn.o_proj",
|
| 353 |
+
"model.language_model.layers.41.mlp.shared_experts.gate_proj",
|
| 354 |
+
"model.language_model.layers.41.mlp.shared_experts.up_proj",
|
| 355 |
+
"model.language_model.layers.41.mlp.shared_experts.down_proj",
|
| 356 |
+
"model.language_model.layers.42.self_attn.q_proj",
|
| 357 |
+
"model.language_model.layers.42.self_attn.k_proj",
|
| 358 |
+
"model.language_model.layers.42.self_attn.v_proj",
|
| 359 |
+
"model.language_model.layers.42.self_attn.o_proj",
|
| 360 |
+
"model.language_model.layers.42.mlp.shared_experts.gate_proj",
|
| 361 |
+
"model.language_model.layers.42.mlp.shared_experts.up_proj",
|
| 362 |
+
"model.language_model.layers.42.mlp.shared_experts.down_proj",
|
| 363 |
+
"model.language_model.layers.43.self_attn.q_proj",
|
| 364 |
+
"model.language_model.layers.43.self_attn.k_proj",
|
| 365 |
+
"model.language_model.layers.43.self_attn.v_proj",
|
| 366 |
+
"model.language_model.layers.43.self_attn.o_proj",
|
| 367 |
+
"model.language_model.layers.43.mlp.shared_experts.gate_proj",
|
| 368 |
+
"model.language_model.layers.43.mlp.shared_experts.up_proj",
|
| 369 |
+
"model.language_model.layers.43.mlp.shared_experts.down_proj",
|
| 370 |
+
"model.language_model.layers.44.self_attn.q_proj",
|
| 371 |
+
"model.language_model.layers.44.self_attn.k_proj",
|
| 372 |
+
"model.language_model.layers.44.self_attn.v_proj",
|
| 373 |
+
"model.language_model.layers.44.self_attn.o_proj",
|
| 374 |
+
"model.language_model.layers.44.mlp.shared_experts.gate_proj",
|
| 375 |
+
"model.language_model.layers.44.mlp.shared_experts.up_proj",
|
| 376 |
+
"model.language_model.layers.44.mlp.shared_experts.down_proj",
|
| 377 |
+
"model.language_model.layers.45.self_attn.q_proj",
|
| 378 |
+
"model.language_model.layers.45.self_attn.k_proj",
|
| 379 |
+
"model.language_model.layers.45.self_attn.v_proj",
|
| 380 |
+
"model.language_model.layers.45.self_attn.o_proj",
|
| 381 |
+
"model.language_model.layers.45.mlp.shared_experts.gate_proj",
|
| 382 |
+
"model.language_model.layers.45.mlp.shared_experts.up_proj",
|
| 383 |
+
"model.language_model.layers.45.mlp.shared_experts.down_proj",
|
| 384 |
+
"lm_head"
|
| 385 |
+
],
|
| 386 |
+
"kv_cache_scheme": null,
|
| 387 |
+
"quant_method": "compressed-tensors",
|
| 388 |
+
"quantization_status": "compressed"
|
| 389 |
+
},
|
| 390 |
+
"rms_norm_eps": 1e-05,
|
| 391 |
+
"torch_dtype": "bfloat16",
|
| 392 |
+
"rope_scaling": {
|
| 393 |
+
"rope_type": "default",
|
| 394 |
+
"mrope_section": [
|
| 395 |
+
8,
|
| 396 |
+
12,
|
| 397 |
+
12
|
| 398 |
+
]
|
| 399 |
+
},
|
| 400 |
+
"rope_theta": 10000.0,
|
| 401 |
+
"routed_scaling_factor": 1.0,
|
| 402 |
+
"topk_group": 1,
|
| 403 |
+
"use_cache": true,
|
| 404 |
+
"use_qk_norm": false
|
| 405 |
+
},
|
| 406 |
+
"torch_dtype": "bfloat16",
|
| 407 |
+
"transformers_version": "4.55.0.dev0",
|
| 408 |
+
"video_end_token_id": 151342,
|
| 409 |
+
"video_start_token_id": 151341,
|
| 410 |
+
"video_token_id": 151364,
|
| 411 |
+
"vision_config": {
|
| 412 |
+
"attention_bias": false,
|
| 413 |
+
"attention_dropout": 0.0,
|
| 414 |
+
"depth": 24,
|
| 415 |
+
"hidden_act": "silu",
|
| 416 |
+
"hidden_size": 1536,
|
| 417 |
+
"image_size": 336,
|
| 418 |
+
"in_channels": 3,
|
| 419 |
+
"initializer_range": 0.02,
|
| 420 |
+
"intermediate_size": 10944,
|
| 421 |
+
"model_type": "glm4v_moe",
|
| 422 |
+
"num_heads": 12,
|
| 423 |
+
"out_hidden_size": 4096,
|
| 424 |
+
"patch_size": 14,
|
| 425 |
+
"rms_norm_eps": 1e-05,
|
| 426 |
+
"spatial_merge_size": 2,
|
| 427 |
+
"temporal_patch_size": 2
|
| 428 |
+
}
|
| 429 |
+
}
|
configuration.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"framework":"Pytorch","task":"image-text-to-text"}
|
generation_config.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"do_sample": true,
|
| 4 |
+
"eos_token_id": [
|
| 5 |
+
151329,
|
| 6 |
+
151336,
|
| 7 |
+
151338
|
| 8 |
+
],
|
| 9 |
+
"pad_token_id": 151329,
|
| 10 |
+
"temperature": 1.0,
|
| 11 |
+
"top_k": 1,
|
| 12 |
+
"top_p": 0.0001,
|
| 13 |
+
"transformers_version": "4.55.0.dev"
|
| 14 |
+
}
|
model-00001-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6da6166fec18f3cd248b70df065edd2d7b6727b3953417117f925a082d0ed0b
|
| 3 |
+
size 2957438824
|
model-00002-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:edc7ee3f7775f171797cf4406a909af2a42bf10754223f1c4a5a38be49c5e81f
|
| 3 |
+
size 2470328656
|
model-00003-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fac990d1a026db27fd8215be8991c0f4485516b49cb5a5dfc2286922a988ae83
|
| 3 |
+
size 2470328656
|
model-00004-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:173ee1fbce09c0ba235dbb40e349ab2c33d38d28a8bb3efb6408b9ab17e70713
|
| 3 |
+
size 2470328656
|
model-00005-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ab9c26eafa6eb1459deadfb84c3c38ac560fd0c203be03f864afd3f8789cd6bb
|
| 3 |
+
size 2470328656
|
model-00006-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e72f6c9753625ff12cf0e000a56cd9fed301a9a1698309b262211b876424a443
|
| 3 |
+
size 2470328656
|
model-00007-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1770966db00e74ec7f92a6c4041ca15cad26a63aec8f34438a47ee33d05f92da
|
| 3 |
+
size 2470328656
|
model-00008-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:151567c13dbd8bab79ab1040077ce8005b13e4459c616af7bef5674b81b5724a
|
| 3 |
+
size 2470328656
|
model-00009-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:15093df20ad7e00bf15750aad2615f745e4c2d1b6ffcd61d66a71dd0391109a7
|
| 3 |
+
size 2470328656
|
model-00010-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fad89472a5c2a8bc1ae5cff24cabbbaeeb3083f1b1e7d5b33befd052818c6a8c
|
| 3 |
+
size 2470329824
|
model-00011-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7ffd7ab6dcc6ff105dda87562bc5d947e2bff2d61bbc9f7c1fbc0a6f9b82d3ab
|
| 3 |
+
size 2470329824
|
model-00012-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:915e0a4dee0cbbd5425b2699238dd31f1220711b6e08c0b3e2aa936c58dd573b
|
| 3 |
+
size 2470329824
|
model-00013-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0da21b0ae635c613fbb77ab994e3a669238c8ee52c4a17bc95119a312fc1b20
|
| 3 |
+
size 2470329824
|
model-00014-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c92b579b5fa798f40bf7a62c033abeb934c2fa840c167dfe3d0efec5c9e7dde
|
| 3 |
+
size 2470329824
|
model-00015-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:47ff7dec9ad39349da07653b80355d68f0e7066b8c8d4edef03eebcd53b892b2
|
| 3 |
+
size 2470329824
|
model-00016-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aa639eec3b5fa22796ec13299499b6deea89320e9763885cf74b05513cfa80e8
|
| 3 |
+
size 2470329824
|
model-00017-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fde1cccdcec907f618d96e08c3e97ffcb39c92b8b11a5967162a0969fc733d71
|
| 3 |
+
size 2470329824
|
model-00018-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7ce384b24a5ec8d38945891b3f46d11a79d291f338b133bf22022fd62d800a42
|
| 3 |
+
size 2470329824
|
model-00019-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7890d557b26f8694b52fa25129e5600cb164977d79b991e94669643361a903af
|
| 3 |
+
size 2470329824
|
model-00020-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2025beb796d51c67f86a784144c211801566ededaa9677c56dc7f53cc9252e55
|
| 3 |
+
size 2470329824
|
model-00021-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a615f16ecb17d4fbb9870d3e665fb678df3c29e323dd4ac00dcb72aaef469f5c
|
| 3 |
+
size 2470329824
|
model-00022-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:39e098cd237f2c67ac210e8927a522173195de44d71018a34a28a1bfe838e072
|
| 3 |
+
size 2470329824
|
model-00023-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0ad010a7b62129d6011a1c3a5de995328af9182a2103e937ae283ad9061c066e
|
| 3 |
+
size 2470329824
|
model-00024-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ca04b8fcb28f9e1373f17cd6f564ee0670fbacf425f926b6b027a5f80ae02ee7
|
| 3 |
+
size 2470329824
|
model-00025-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6155196817dc9ceecbae528f48ee84466840b2fdffee5b5faee23d1a0a1892db
|
| 3 |
+
size 2470329824
|
model-00026-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3ea9d9158ce865521725b672396d4146895c477fa778d54b274a8eec9c1cc9c9
|
| 3 |
+
size 2470329824
|
model-00027-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b36a7c5fb11e96016e8957c3d32cc23e1e03f34748b7453deb271ddf8db21b56
|
| 3 |
+
size 2470329824
|
model-00028-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b7b95c0feebf15ecb9c7f1cf49a3bb516f80da8da2299475bf75263ce0865b0e
|
| 3 |
+
size 2470329824
|
model-00029-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3ae7393ba1b208da259ed9afa5fa98d798506daef1064f081a17a6b7ceb61982
|
| 3 |
+
size 2470329824
|
model-00030-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c0b047ea2eb206ba4dc9f493b1a54ffa4b167f33c65576339e48e7862aec1bbb
|
| 3 |
+
size 2470329824
|
model-00031-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9f5a19faf5b18a8f76112798d64fc8de527e4bb42bb952475479ab5c180985d
|
| 3 |
+
size 2470329824
|
model-00032-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e9016195362cae34e1965b5d4c519cd12f08fd1adeedfa1310a9db32b8ebd171
|
| 3 |
+
size 2470329824
|
model-00033-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b2f079a1691e05ba9bf72b921269389261a59a296a51f2e0d297960f4e043c83
|
| 3 |
+
size 2470329824
|
model-00034-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:33e8a15a1300af48423b9f70183c4271dbd3de7a356155f25499d76c7a8dca34
|
| 3 |
+
size 2470329824
|
model-00035-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41cba281c5cf0bca9e35aad8b1e892b904accf0b176cbe76a6c7eeed7d97bb71
|
| 3 |
+
size 2470329824
|
model-00036-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6079ab450f05a901b0f003c9a3f50a7f089e789ddd1a9a206af7db614f696fc
|
| 3 |
+
size 2470329824
|
model-00037-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4e839e34b8ae4f5a214e3f0a346066d6296fad9f17b7afc50d1c7ba662d30aa4
|
| 3 |
+
size 2470329824
|
model-00038-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f5a7d2af48237f720fba95b05c9e3fe41255c3dfece5ee970088d471668243e4
|
| 3 |
+
size 2470329824
|
model-00039-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b7c14f85704f134f9e1dff38525e7a117727dff8026f889960ee9228708aaa95
|
| 3 |
+
size 2470329824
|
model-00040-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:989c36dbb81681603051580eafdb53198c1ef9526cff72c6a89f6d728dd0bb68
|
| 3 |
+
size 2470329824
|
model-00041-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d2eaf29bccabdb1bb51f6ae2d06676e170d6127b0e6bf86bbc90b33078a445c
|
| 3 |
+
size 2470329824
|
model-00042-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8fa1be2876aa10116d6f62e903262b404e73c393aa27e9ff9e752677fa723f4
|
| 3 |
+
size 2470329824
|
model-00043-of-00046.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:052c09a7b69fab6dc41d8b5e8d53b2a27b50d3bc74e1db4aff209d911383622b
|
| 3 |
+
size 2470329824
|