Hugging Face GGUF 模型可视化

Hugging Face GGUF 模型可视化

  • [1. Finding GGUF files (检索 GGUF 模型)](#1. Finding GGUF files (检索 GGUF 模型))
  • [2. Viewer for metadata & tensors info (可视化 GGUF 模型)](#2. Viewer for metadata & tensors info (可视化 GGUF 模型))
  • References

无知小儿,仙家雄霸天下,依附强者才是唯一的出路。否则天地虽大,也让你们无路可走!

GGUF
https://huggingface.co/docs/hub/gguf

Hugging Face Hub supports all file formats, but has built-in features for GGUF format, a binary format that is optimized for quick loading and saving of models, making it highly efficient for inference purposes.

GGUF
https://github.com/ggerganov/ggml/blob/master/docs/gguf.md

GGUF is designed for use with GGML and other executors. GGUF was developed by Georgi Gerganov who is also the developer of llama.cpp, a popular C/C++ LLM inference framework. Models initially developed in frameworks like PyTorch can be converted to GGUF format for use with those engines.

Georgi Gerganov
https://huggingface.co/ggerganov

As we can see in this graph, unlike tensor-only file formats like safetensors - which is also a recommended model format for the Hub - GGUF encodes both the tensors and a standardized set of metadata.

Safetensors
https://huggingface.co/docs/safetensors

1. Finding GGUF files (检索 GGUF 模型)

You can browse all models with GGUF files filtering by the GGUF tag: https://huggingface.co/models?library=gguf.

Moreover, you can use https://huggingface.co/spaces/ggml-org/gguf-my-repo tool to convert/quantize your model weights into GGUF weights.

For example, you can check out TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF for seeing GGUF files in action.

TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF
https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF

deepseek-ai/DeepSeek-V3
https://huggingface.co/deepseek-ai/DeepSeek-V3

2. Viewer for metadata & tensors info (可视化 GGUF 模型)

The Hub has a viewer for GGUF files that lets a user check out metadata & tensors info (name, shape, precison). The viewer is available on model page & files page.

点击下图绿色矩形框区域图标即可进入 Hugging Face GGUF 模型可视化界面。

TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF
https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF/tree/main

MaziyarPanahi/Llama-3.2-1B-Instruct-GGUF
https://huggingface.co/MaziyarPanahi/Llama-3.2-1B-Instruct-GGUF/tree/main

https://huggingface.co/MaziyarPanahi/Llama-3.2-1B-Instruct-GGUF/tree/main?show_file_info=Llama-3.2-1B-Instruct.Q4_K_M.gguf

References

[1] Yongqiang Cheng, https://yongqiang.blog.csdn.net/

[2] huggingface/gguf, https://github.com/huggingface/huggingface.js/tree/main/packages/gguf

相关推荐
Yongqiang Cheng11 天前
Hugging Face Transformers and Meta Llama
transformers·hugging face·meta llama
HuggingFace2 个月前
开源开发者指南:欧盟《人工智能法案》解读
hugging face
HuggingFace3 个月前
为数据集而生的 SQL 控制台
hugging face·sql sonsole
HuggingFace3 个月前
Hugging Face 与 TruffleHog 合作,实现风险预警
hugging face·trufflehog
算家云3 个月前
stable-zero123模型构建指南
人工智能·aigc·gpu·图像生成·hugging face·3d模型生成·comfyui、
Hoper.J5 个月前
使用 HFD 加快 Hugging Face 模型和数据集的下载
llm·aigc·hugging face·hfd
HuggingFace5 个月前
ggml 简介
hugging face·ggml
HuggingFace6 个月前
Llama 3.1 - 405B、70B 和 8B 的多语言与长上下文能力解析
hugging face·llama 3.1