Georgi Gerganov - ggml - llama.cpp - whisper.cpp

Georgi Gerganov - ggml - llama.cpp - whisper.cpp

  • [1. Georgi Gerganov](#1. Georgi Gerganov)
    • [1.1. Projects](#1.1. Projects)
  • [2. `ggml`](#2. ggml)
  • [3. `llama.cpp`](#3. llama.cpp)
  • [4. `whisper.cpp`](#4. whisper.cpp)
  • References

1. Georgi Gerganov

https://github.com/ggerganov
https://ggerganov.com/

ggml-org
https://github.com/ggml-org

GGML - AI at the edge
https://ggml.ai/

ggml.ai is a company founded by Georgi Gerganov to support the development of ggml.

ggml is a tensor library for machine learning to enable large models and high performance on commodity hardware. It is used by llama.cpp and whisper.cpp.

1.1. Projects

  • whisper.cpp

https://github.com/ggerganov/whisper.cpp

High-performance inference of OpenAI's Whisper automatic speech recognition model

The project provides a high-quality speech-to-text solution that runs on Mac, Windows, Linux, iOS, Android, Raspberry Pi, and Web

  • llama.cpp

https://github.com/ggerganov/llama.cpp

Inference of Meta's LLaMA model (and others) in pure C/C++

The project provides efficient inference across a wide range of hardware and serves as the foundation for numerous LLM-based applications

2. ggml

https://github.com/ggerganov/ggml

Tensor library for machine learning

Some of the development is currently happening in the llama.cpp and whisper.cpp repos.

sync : llama.cpp
sync : whisper.cpp

3. llama.cpp

https://github.com/ggerganov/llama.cpp

Inference of Meta's LLaMA model (and others) in pure C/C++

The llama.cpp project is the main playground for developing new features for the ggml library.

sync : ggml

4. whisper.cpp

https://github.com/ggerganov/whisper.cpp

High-performance inference of OpenAI's Whisper automatic speech recognition (ASR) model.

The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library.

sync : ggml
sync : ggml + llama.cpp
sync : ggml and llama.cpp

References

[1] Yongqiang Cheng, https://yongqiang.blog.csdn.net/

[2] Introduction to ggml, https://huggingface.co/blog/introduction-to-ggml

相关推荐
带电的小王18 天前
whisper.cpp: Android端测试 -- Android端手机部署音频大模型
android·智能手机·llm·whisper·音频大模型·whisper.cpp
带电的小王19 天前
llama.cpp:PC端测试 MobileVLM -- 电脑端部署图生文大模型
llm·llama.cpp·vlm·mobilevlm·图生文
AI.愚人自愈1 个月前
【2024 Dec 超实时】编辑安装llama.cpp并运行llama
llm·llama.cpp
沉下心来学鲁班2 个月前
欺诈文本分类检测(十八):基于llama.cpp+CPU推理
人工智能·语言模型·分类·cpu·llama.cpp
幽影相随3 个月前
构建llama.cpp并在linux上使用gpu
llm·llama.cpp
Hoper.J3 个月前
GPTQ vs AWQ vs GGUF(GGML) 速览和 GGUF 文件命名规范
模型量化·gptq·awq·gguf·ggml
telllong4 个月前
使用llama.cpp 在推理MiniCPM-1.2B模型
c++·llama·llama.cpp
shizidushu4 个月前
Docker下使用llama.cpp部署带Function calling和Json Mode功能的Mistral 7B模型
docker·语言模型·llama·llama.cpp·mistral-7b
穷人小水滴4 个月前
(章节 3.1) 本地运行 AI 有多慢 ? 大模型推理测速 (llama.cpp, Intel GPU A770)
人工智能·llama·llama.cpp·a770·vulkan·gnu/linux·sycl