From 84cfb993472046ad7a1d76d1ee69e636e64198cd Mon Sep 17 00:00:00 2001 From: Mishig Davaadorj Date: Fri, 15 Mar 2024 11:06:37 +0000 Subject: [PATCH] Fix gguf doc page style --- docs/hub/gguf.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/hub/gguf.md b/docs/hub/gguf.md index 95cc1b555..6c394e54c 100644 --- a/docs/hub/gguf.md +++ b/docs/hub/gguf.md @@ -2,8 +2,8 @@ Hugging Face Hub supports all file formats, but has built-in features for [GGUF format](https://github.com/ggerganov/ggml/blob/master/docs/gguf.md), a binary format that is optimized for quick loading and saving of models, making it highly efficient for inference purposes. GGUF is designed for use with GGML and other executors. GGUF was developed by [@ggerganov](https://huggingface.co/ggerganov) who is also the developer of [llama.cpp](https://github.com/ggerganov/llama.cpp), a popular C/C++ LLM inference framework. Models initially developed in frameworks like PyTorch can be converted to GGUF format for use with those engines. -
- +
+
As we can see in this graph, unlike tensor-only file formats like [safetensors](https://huggingface.co/docs/safetensors) – which is also a recommended model format for the Hub – GGUF encodes both the tensors and a standardized set of metadata.