Georgi Gerganov - ggml - llama.cpp - whisper.cpp

Georgi Gerganov - ggml - llama.cpp - whisper.cpp

  • [1. Georgi Gerganov](#1. Georgi Gerganov)
    • [1.1. Projects](#1.1. Projects)
  • [2. `ggml`](#2. ggml)
  • [3. `llama.cpp`](#3. llama.cpp)
  • [4. `whisper.cpp`](#4. whisper.cpp)
  • References

1. Georgi Gerganov

https://github.com/ggerganov
https://ggerganov.com/

ggml-org
https://github.com/ggml-org

GGML - AI at the edge
https://ggml.ai/

ggml.ai is a company founded by Georgi Gerganov to support the development of ggml.

ggml is a tensor library for machine learning to enable large models and high performance on commodity hardware. It is used by llama.cpp and whisper.cpp.

1.1. Projects

  • whisper.cpp

https://github.com/ggerganov/whisper.cpp

High-performance inference of OpenAI's Whisper automatic speech recognition model

The project provides a high-quality speech-to-text solution that runs on Mac, Windows, Linux, iOS, Android, Raspberry Pi, and Web

  • llama.cpp

https://github.com/ggerganov/llama.cpp

Inference of Meta's LLaMA model (and others) in pure C/C++

The project provides efficient inference across a wide range of hardware and serves as the foundation for numerous LLM-based applications

2. ggml

https://github.com/ggerganov/ggml

Tensor library for machine learning

Some of the development is currently happening in the llama.cpp and whisper.cpp repos.

复制代码
sync : llama.cpp
sync : whisper.cpp

3. llama.cpp

https://github.com/ggerganov/llama.cpp

Inference of Meta's LLaMA model (and others) in pure C/C++

The llama.cpp project is the main playground for developing new features for the ggml library.

复制代码
sync : ggml

4. whisper.cpp

https://github.com/ggerganov/whisper.cpp

High-performance inference of OpenAI's Whisper automatic speech recognition (ASR) model.

The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library.

复制代码
sync : ggml
sync : ggml + llama.cpp
sync : ggml and llama.cpp

References

1\] Yongqiang Cheng, \[2\] Introduction to ggml,

相关推荐
d1z8882 天前
(二十)32天GPU测试从入门到精通-llama.cpp CPU/GPU 混合推理day18
人工智能·llama·显卡·llama.cpp
gergul3 天前
在llama-cpp-python中使用自己编译的llama.cpp,解决pip install llama-cpp-python报错
python·llama·llama.cpp·llamacpppython
晨欣6 天前
单卡 48GB 实测:Gemma 4 26B A4B、Gemma 4 31B、gpt-oss-20b 三模型部署与并发对比
google·openai·nvidia·vllm·llama.cpp·gpt-oss-20b·gemma4
belldeep9 天前
AI: ggml llama.cpp 与 BitNet 模型介绍
人工智能·llama.cpp·bitnet·gguf·ggml
未来之窗软件服务11 天前
SenseVoicecpp ggml-rpc.cpp大模型[AI人工智能(七十七)]—东方仙盟
人工智能·rpc·ggml·仙盟创梦ide·东方仙盟
love530love17 天前
【独家资源】Windows 本地部署微软 BitNet b1.58: Flash Attention + CUDA GPU 加速 (sm_86) + AVX2 优化 + 1.58bit 量化
人工智能·windows·microsoft·llama.cpp·bitnet·flash attention·bitlinear_cpp
love530love1 个月前
OpenClaw搭配LM Studio VS Ollama:Windows CUDA实战深度对比与完全配置指南
人工智能·windows·vllm·ollama·llama.cpp·lm studio·openclaw
seaside20031 个月前
ggml交叉编译安卓部署大模型
安卓·交叉编译·ggml
晨欣1 个月前
llama.cpp 设计巧思:多模态模型拆分加载,按需使用视觉能力(配图由谷歌的Nano Banana模型倾情生成)
llm·谷歌·cursor·llama.cpp·gguf模型·gpt5.4
JohnCHsu1 个月前
性能干翻235B,单卡私有化部署OpenClaw
ai·agent·llama.cpp·openclaw