Hugging Face GGUF 模型可视化
Hugging Face GGUF 模型可视化
- 1. Finding GGUF files (检索 GGUF 模型)
- 2. Viewer for metadata & tensors info (可视化 GGUF 模型)
- References
无知小儿,仙家雄霸天下,依附强者才是唯一的出路。否则天地虽大,也让你们无路可走!
GGUF
https://huggingface.co/docs/hub/gguf
Hugging Face Hub supports all file formats, but has built-in features for GGUF format, a binary format that is optimized for quick loading and saving of models, making it highly efficient for inference purposes.
GGUF
https://github.com/ggerganov/ggml/blob/master/docs/gguf.md
GGUF is designed for use with GGML and other executors. GGUF was developed by Georgi Gerganov who is also the developer of llama.cpp
, a popular C/C++ LLM inference framework. Models initially developed in frameworks like PyTorch can be converted to GGUF format for use with those engines.
Georgi Gerganov
https://huggingface.co/ggerganov
As we can see in this graph, unlike tensor-only file formats like safetensors - which is also a recommended model format for the Hub - GGUF encodes both the tensors and a standardized set of metadata.
Safetensors
https://huggingface.co/docs/safetensors
1. Finding GGUF files (检索 GGUF 模型)
You can browse all models with GGUF files filtering by the GGUF tag: https://huggingface.co/models?library=gguf.
Moreover, you can use https://huggingface.co/spaces/ggml-org/gguf-my-repo tool to convert/quantize your model weights into GGUF weights.
For example, you can check out TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF for seeing GGUF files in action.
TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF
https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF
deepseek-ai/DeepSeek-V3
https://huggingface.co/deepseek-ai/DeepSeek-V3
2. Viewer for metadata & tensors info (可视化 GGUF 模型)
The Hub has a viewer for GGUF files that lets a user check out metadata & tensors info (name, shape, precison). The viewer is available on model page & files page.
点击下图绿色矩形框区域图标即可进入 Hugging Face GGUF 模型可视化界面。
TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF
https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF/tree/main
MaziyarPanahi/Llama-3.2-1B-Instruct-GGUF
https://huggingface.co/MaziyarPanahi/Llama-3.2-1B-Instruct-GGUF/tree/main
https://huggingface.co/MaziyarPanahi/Llama-3.2-1B-Instruct-GGUF/tree/main?show_file_info=Llama-3.2-1B-Instruct.Q4_K_M.gguf
References
[1] Yongqiang Cheng, https://yongqiang.blog.csdn.net/
[2] huggingface/gguf, https://github.com/huggingface/huggingface.js/tree/main/packages/gguf