ggml
Here are 107 public repositories matching this topic...
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
-
Updated
Apr 9, 2025 - Python
Stable Diffusion and Flux in pure C/C++
-
Updated
Mar 9, 2025 - C++
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
-
Updated
Mar 23, 2025 - C++
Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
-
Updated
Dec 3, 2024 - JavaScript
Suno AI's Bark model in C/C++ for fast text-to-speech generation
-
Updated
Nov 16, 2024 - C++
Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)
-
Updated
Aug 8, 2023 - C++
CLIP inference in plain C/C++ with no extra dependencies
-
Updated
Aug 18, 2024 - C++
Inference Vision Transformer (ViT) in plain C/C++ with ggml
-
Updated
Apr 11, 2024 - C++
This custom_node for ComfyUI adds one-click "Virtual VRAM" for any GGUF UNet and CLIP loader, managing the offload of layers to DRAM or VRAM to maximize the latent space of your card. Also includes nodes for directly loading entire components (UNet, CLIP, VAE) onto the device you choose. Includes 16 examples covering common use cases.
-
Updated
Apr 4, 2025 - Python
Improve this page
Add a description, image, and links to the ggml topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the ggml topic, visit your repo's landing page and select "manage topics."