Upload nvidia.Minitron-4B-Base.Q4_K_M.gguf with huggingface_hub
Browse files- .gitattributes +1 -0
- nvidia.Minitron-4B-Base.Q4_K_M.gguf +3 -0
.gitattributes
CHANGED
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
nvidia.Minitron-4B-Base.f16.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
nvidia.Minitron-4B-Base.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
nvidia.Minitron-4B-Base.f16.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
nvidia.Minitron-4B-Base.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
nvidia.Minitron-4B-Base.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
nvidia.Minitron-4B-Base.Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26c238ff51ec79c42903f8d11338ab7e997a9bf431056ef67c594542a475b119
|
3 |
+
size 2697385824
|