ngxson HF Staff commited on
Commit
94982a1
·
verified ·
1 Parent(s): 1da7ad3

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ GLM-4.6V-Flash-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
37
+ mmproj-GLM-4.6V-Flash-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
GLM-4.6V-Flash-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10910372ab8a31a11b46d34cc54b1e8ff275e893c3d8787008497d000ac0e4e6
3
+ size 6166577888
README.md ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: zai-org/GLM-4.6V-Flash
3
+ ---
4
+
5
+ # GLM-4.6V-Flash-GGUF
6
+
7
+ This model is converted from [zai-org/GLM-4.6V-Flash](https://huggingface.co/zai-org/GLM-4.6V-Flash) to GGUF using `convert_hf_to_gguf.py`
8
+
9
+ To use it:
10
+
11
+ ```
12
+ llama-server -hf ggml-org/GLM-4.6V-Flash-GGUF
13
+ ```
14
+
mmproj-GLM-4.6V-Flash-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d3af471d3a52020278528480f44b608592b42584e2e65828b40bc8851b653ac
3
+ size 980055680