You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We support parsing [GGUF](https://github.com/ggerganov/ggml/blob/master/docs/gguf.md) files with the following tensor types:
3
3
- F16
4
4
- F32
5
5
- Q4_0
6
6
- Q6_K
7
7
8
+
If an unsupported type is encountered while parsing a GGUF file, an exception is raised.
8
9
9
-
### Download
10
-
First download a GGUF model and tokenizer. In this example, we use GGUF Q4_0 format.
10
+
We now go over an example of using GGUF files in the torchchat flow.
11
+
12
+
### Download resources
13
+
First download a GGUF model and tokenizer. In this example, we use a Q4_0 GGUF file. (Note that Q4_0 is only the dominant tensor type in the file, but the file also contains GGUF tensors of types Q6_K, F16, and F32.)
0 commit comments