--- ## metas - &llamacpp name: "llama-cpp" alias: "llama-cpp" license: mit icon: https://user-images.githubusercontent.com/1991296/230134379-7181e485-c521-4d23-a0d6-f7b3b61ba524.png description: | LLM inference in C/C++ urls: - https://github.com/ggerganov/llama.cpp tags: - text-to-text - LLM - CPU - GPU - Metal - CUDA - HIP capabilities: default: "cpu-llama-cpp" nvidia: "cuda12-llama-cpp" intel: "intel-sycl-f16-llama-cpp" amd: "rocm-llama-cpp" metal: "metal-llama-cpp" vulkan: "vulkan-llama-cpp" nvidia-l4t: "nvidia-l4t-arm64-llama-cpp" darwin-x86: "darwin-x86-llama-cpp" - &whispercpp name: "whisper" alias: "whisper" license: mit icon: https://user-images.githubusercontent.com/1991296/235238348-05d0f6a4-da44-4900-a1de-d0707e75b763.jpeg description: | Port of OpenAI's Whisper model in C/C++ urls: - https://github.com/ggml-org/whisper.cpp tags: - audio-transcription - CPU - GPU - CUDA - HIP capabilities: default: "cpu-whisper" nvidia: "cuda12-whisper" intel: "intel-sycl-f16-whisper" metal: "metal-whisper" amd: "rocm-whisper" vulkan: "vulkan-whisper" nvidia-l4t: "nvidia-l4t-arm64-whisper" - &stablediffusionggml name: "stablediffusion-ggml" alias: "stablediffusion-ggml" license: mit icon: https://github.com/leejet/stable-diffusion.cpp/raw/master/assets/cat_with_sd_cpp_42.png description: | Stable Diffusion and Flux in pure C/C++ urls: - https://github.com/leejet/stable-diffusion.cpp tags: - image-generation - CPU - GPU - Metal - CUDA - HIP capabilities: default: "cpu-stablediffusion-ggml" nvidia: "cuda12-stablediffusion-ggml" intel: "intel-sycl-f16-stablediffusion-ggml" # amd: "rocm-stablediffusion-ggml" vulkan: "vulkan-stablediffusion-ggml" nvidia-l4t: "nvidia-l4t-arm64-stablediffusion-ggml" metal: "metal-stablediffusion-ggml" # darwin-x86: "darwin-x86-stablediffusion-ggml" - &rfdetr name: "rfdetr" alias: "rfdetr" license: apache-2.0 icon: https://avatars.githubusercontent.com/u/53104118?s=200&v=4 description: | RF-DETR is a real-time, transformer-based object detection model architecture developed by Roboflow and released under the Apache 2.0 license. RF-DETR is the first real-time model to exceed 60 AP on the Microsoft COCO benchmark alongside competitive performance at base sizes. It also achieves state-of-the-art performance on RF100-VL, an object detection benchmark that measures model domain adaptability to real world problems. RF-DETR is fastest and most accurate for its size when compared current real-time objection models. RF-DETR is small enough to run on the edge using Inference, making it an ideal model for deployments that need both strong accuracy and real-time performance. urls: - https://github.com/roboflow/rf-detr tags: - object-detection - rfdetr - gpu - cpu capabilities: nvidia: "cuda12-rfdetr" intel: "intel-rfdetr" #amd: "rocm-rfdetr" nvidia-l4t: "nvidia-l4t-arm64-rfdetr" default: "cpu-rfdetr" - &vllm name: "vllm" license: apache-2.0 urls: - https://github.com/vllm-project/vllm tags: - text-to-text - multimodal - GPTQ - AWQ - AutoRound - INT4 - INT8 - FP8 icon: https://raw.githubusercontent.com/vllm-project/vllm/main/docs/assets/logos/vllm-logo-text-dark.png description: | vLLM is a fast and easy-to-use library for LLM inference and serving. Originally developed in the Sky Computing Lab at UC Berkeley, vLLM has evolved into a community-driven project with contributions from both academia and industry. vLLM is fast with: State-of-the-art serving throughput Efficient management of attention key and value memory with PagedAttention Continuous batching of incoming requests Fast model execution with CUDA/HIP graph Quantizations: GPTQ, AWQ, AutoRound, INT4, INT8, and FP8 Optimized CUDA kernels, including integration with FlashAttention and FlashInfer Speculative decoding Chunked prefill alias: "vllm" capabilities: nvidia: "cuda12-vllm" amd: "rocm-vllm" intel: "intel-vllm" - &mlx name: "mlx" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-mlx" icon: https://avatars.githubusercontent.com/u/102832242?s=200&v=4 urls: - https://github.com/ml-explore/mlx-lm mirrors: - localai/localai-backends:latest-metal-darwin-arm64-mlx license: MIT description: | Run LLMs with MLX tags: - text-to-text - LLM - MLX - &mlx-vlm name: "mlx-vlm" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-mlx-vlm" icon: https://avatars.githubusercontent.com/u/102832242?s=200&v=4 urls: - https://github.com/Blaizzy/mlx-vlm mirrors: - localai/localai-backends:latest-metal-darwin-arm64-mlx-vlm license: MIT description: | Run Vision-Language Models with MLX tags: - text-to-text - multimodal - vision-language - LLM - MLX - &mlx-audio name: "mlx-audio" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-mlx-audio" icon: https://avatars.githubusercontent.com/u/102832242?s=200&v=4 urls: - https://github.com/Blaizzy/mlx-audio mirrors: - localai/localai-backends:latest-metal-darwin-arm64-mlx-audio license: MIT description: | Run Audio Models with MLX tags: - audio-to-text - audio-generation - text-to-audio - LLM - MLX - &rerankers name: "rerankers" alias: "rerankers" capabilities: nvidia: "cuda12-rerankers" intel: "intel-rerankers" amd: "rocm-rerankers" - &transformers name: "transformers" icon: https://camo.githubusercontent.com/26569a27b8a30a488dd345024b71dbc05da7ff1b2ba97bb6080c9f1ee0f26cc7/68747470733a2f2f68756767696e67666163652e636f2f64617461736574732f68756767696e67666163652f646f63756d656e746174696f6e2d696d616765732f7265736f6c76652f6d61696e2f7472616e73666f726d6572732f7472616e73666f726d6572735f61735f615f6d6f64656c5f646566696e6974696f6e2e706e67 alias: "transformers" license: apache-2.0 description: | Transformers acts as the model-definition framework for state-of-the-art machine learning models in text, computer vision, audio, video, and multimodal model, for both inference and training. It centralizes the model definition so that this definition is agreed upon across the ecosystem. transformers is the pivot across frameworks: if a model definition is supported, it will be compatible with the majority of training frameworks (Axolotl, Unsloth, DeepSpeed, FSDP, PyTorch-Lightning, ...), inference engines (vLLM, SGLang, TGI, ...), and adjacent modeling libraries (llama.cpp, mlx, ...) which leverage the model definition from transformers. urls: - https://github.com/huggingface/transformers tags: - text-to-text - multimodal capabilities: nvidia: "cuda12-transformers" intel: "intel-transformers" amd: "rocm-transformers" - &diffusers name: "diffusers" icon: https://raw.githubusercontent.com/huggingface/diffusers/main/docs/source/en/imgs/diffusers_library.jpg description: | πŸ€— Diffusers is the go-to library for state-of-the-art pretrained diffusion models for generating images, audio, and even 3D structures of molecules. Whether you're looking for a simple inference solution or training your own diffusion models, πŸ€— Diffusers is a modular toolbox that supports both. urls: - https://github.com/huggingface/diffusers tags: - image-generation - video-generation - diffusion-models license: apache-2.0 alias: "diffusers" capabilities: nvidia: "cuda12-diffusers" intel: "intel-diffusers" amd: "rocm-diffusers" nvidia-l4t: "nvidia-l4t-diffusers" metal: "metal-diffusers" default: "cpu-diffusers" - &exllama2 name: "exllama2" urls: - https://github.com/turboderp-org/exllamav2 tags: - text-to-text - LLM - EXL2 license: MIT description: | ExLlamaV2 is an inference library for running local LLMs on modern consumer GPUs. alias: "exllama2" capabilities: nvidia: "cuda12-exllama2" intel: "intel-exllama2" - &faster-whisper icon: https://avatars.githubusercontent.com/u/1520500?s=200&v=4 description: | faster-whisper is a reimplementation of OpenAI's Whisper model using CTranslate2, which is a fast inference engine for Transformer models. This implementation is up to 4 times faster than openai/whisper for the same accuracy while using less memory. The efficiency can be further improved with 8-bit quantization on both CPU and GPU. urls: - https://github.com/SYSTRAN/faster-whisper tags: - speech-to-text - Whisper license: MIT name: "faster-whisper" capabilities: nvidia: "cuda12-faster-whisper" intel: "intel-faster-whisper" amd: "rocm-faster-whisper" - &kokoro icon: https://avatars.githubusercontent.com/u/166769057?v=4 description: | Kokoro is an open-weight TTS model with 82 million parameters. Despite its lightweight architecture, it delivers comparable quality to larger models while being significantly faster and more cost-efficient. With Apache-licensed weights, Kokoro can be deployed anywhere from production environments to personal projects. urls: - https://huggingface.co/hexgrad/Kokoro-82M - https://github.com/hexgrad/kokoro tags: - text-to-speech - TTS - LLM license: apache-2.0 alias: "kokoro" name: "kokoro" capabilities: nvidia: "cuda12-kokoro" intel: "intel-kokoro" amd: "rocm-kokoro" nvidia-l4t: "nvidia-l4t-kokoro" - &coqui urls: - https://github.com/idiap/coqui-ai-TTS description: | 🐸 Coqui TTS is a library for advanced Text-to-Speech generation. πŸš€ Pretrained models in +1100 languages. πŸ› οΈ Tools for training new models and fine-tuning existing models in any language. πŸ“š Utilities for dataset analysis and curation. tags: - text-to-speech - TTS license: mpl-2.0 name: "coqui" alias: "coqui" capabilities: nvidia: "cuda12-coqui" intel: "intel-coqui" amd: "rocm-coqui" icon: https://avatars.githubusercontent.com/u/1338804?s=200&v=4 - &bark urls: - https://github.com/suno-ai/bark description: | Bark is a transformer-based text-to-audio model created by Suno. Bark can generate highly realistic, multilingual speech as well as other audio - including music, background noise and simple sound effects. The model can also produce nonverbal communications like laughing, sighing and crying. To support the research community, we are providing access to pretrained model checkpoints, which are ready for inference and available for commercial use. tags: - text-to-speech - TTS license: MIT name: "bark" alias: "bark" capabilities: cuda: "cuda12-bark" intel: "intel-bark" rocm: "rocm-bark" icon: https://avatars.githubusercontent.com/u/99442120?s=200&v=4 - &barkcpp urls: - https://github.com/PABannier/bark.cpp description: | With bark.cpp, our goal is to bring real-time realistic multilingual text-to-speech generation to the community. Plain C/C++ implementation without dependencies AVX, AVX2 and AVX512 for x86 architectures CPU and GPU compatible backends Mixed F16 / F32 precision 4-bit, 5-bit and 8-bit integer quantization Metal and CUDA backends Models supported Bark Small Bark Large tags: - text-to-speech - TTS license: MIT icon: https://github.com/PABannier/bark.cpp/raw/main/assets/banner.png name: "bark-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-bark-cpp" mirrors: - localai/localai-backends:latest-bark-cpp alias: "bark-cpp" - &chatterbox urls: - https://github.com/resemble-ai/chatterbox description: | Resemble AI's first production-grade open source TTS model. Licensed under MIT, Chatterbox has been benchmarked against leading closed-source systems like ElevenLabs, and is consistently preferred in side-by-side evaluations. Whether you're working on memes, videos, games, or AI agents, Chatterbox brings your content to life. It's also the first open source TTS model to support emotion exaggeration control, a powerful feature that makes your voices stand out. tags: - text-to-speech - TTS license: MIT icon: https://private-user-images.githubusercontent.com/660224/448166653-bd8c5f03-e91d-4ee5-b680-57355da204d1.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3NTAxOTE0MDAsIm5iZiI6MTc1MDE5MTEwMCwicGF0aCI6Ii82NjAyMjQvNDQ4MTY2NjUzLWJkOGM1ZjAzLWU5MWQtNGVlNS1iNjgwLTU3MzU1ZGEyMDRkMS5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjUwNjE3JTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI1MDYxN1QyMDExNDBaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT1hMmI1NGY3OGFiZTlhNGFkNTVlYTY4NTIwMWEzODRiZGE4YzdhNGQ5MGNhNzE3MDYyYTA2NDIxYTkyYzhiODkwJlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCJ9.mR9kM9xX0TdzPuSpuspCllHYQiq79dFQ2rtuNvjrl6w name: "chatterbox" alias: "chatterbox" capabilities: nvidia: "cuda12-chatterbox" metal: "metal-chatterbox" default: "cpu-chatterbox" nvidia-l4t: "nvidia-l4t-arm64-chatterbox" - &piper name: "piper" uri: "quay.io/go-skynet/local-ai-backends:latest-piper" icon: https://github.com/OHF-Voice/piper1-gpl/raw/main/etc/logo.png urls: - https://github.com/rhasspy/piper - https://github.com/mudler/go-piper mirrors: - localai/localai-backends:latest-piper license: MIT description: | A fast, local neural text to speech system tags: - text-to-speech - TTS - &silero-vad name: "silero-vad" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-silero-vad" icon: https://user-images.githubusercontent.com/12515440/89997349-b3523080-dc94-11ea-9906-ca2e8bc50535.png urls: - https://github.com/snakers4/silero-vad mirrors: - localai/localai-backends:latest-cpu-silero-vad description: | Silero VAD: pre-trained enterprise-grade Voice Activity Detector. Silero VAD is a voice activity detection model that can be used to detect whether a given audio contains speech or not. tags: - voice-activity-detection - VAD - silero-vad - CPU - &local-store name: "local-store" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-local-store" mirrors: - localai/localai-backends:latest-cpu-local-store urls: - https://github.com/mudler/LocalAI description: | Local Store is a local-first, self-hosted, and open-source vector database. tags: - vector-database - local-first - open-source - CPU license: MIT - &huggingface name: "huggingface" uri: "quay.io/go-skynet/local-ai-backends:latest-huggingface" mirrors: - localai/localai-backends:latest-huggingface icon: https://huggingface.co/front/assets/huggingface_logo-noborder.svg urls: - https://huggingface.co/docs/hub/en/api description: | HuggingFace is a backend which uses the huggingface API to run models. tags: - LLM - huggingface license: MIT - &kitten-tts name: "kitten-tts" uri: "quay.io/go-skynet/local-ai-backends:latest-kitten-tts" mirrors: - localai/localai-backends:latest-kitten-tts urls: - https://github.com/KittenML/KittenTTS description: | Kitten TTS is a text-to-speech model that can generate speech from text. tags: - text-to-speech - TTS license: apache-2.0 - &neutts name: "neutts" urls: - https://github.com/neuphonic/neutts-air description: | NeuTTS Air is the world’s first super-realistic, on-device, TTS speech language model with instant voice cloning. Built off a 0.5B LLM backbone, NeuTTS Air brings natural-sounding speech, real-time performance, built-in security and speaker cloning to your local device - unlocking a new category of embedded voice agents, assistants, toys, and compliance-safe apps. tags: - text-to-speech - TTS license: apache-2.0 capabilities: default: "cpu-neutts" nvidia: "cuda12-neutts" amd: "rocm-neutts" nvidia-l4t: "nvidia-l4t-neutts" - !!merge <<: *neutts name: "neutts-development" capabilities: default: "cpu-neutts-development" nvidia: "cuda12-neutts-development" amd: "rocm-neutts-development" nvidia-l4t: "nvidia-l4t-neutts-development" - !!merge <<: *neutts name: "cpu-neutts" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-neutts" mirrors: - localai/localai-backends:latest-cpu-neutts - !!merge <<: *neutts name: "cuda12-neutts" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-neutts" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-neutts - !!merge <<: *neutts name: "rocm-neutts" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-neutts" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-neutts - !!merge <<: *neutts name: "nvidia-l4t-neutts" uri: "quay.io/go-skynet/local-ai-backends:latest-nvidia-l4t-arm64-neutts" mirrors: - localai/localai-backends:latest-nvidia-l4t-arm64-neutts - !!merge <<: *neutts name: "cpu-neutts-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-neutts" mirrors: - localai/localai-backends:master-cpu-neutts - !!merge <<: *neutts name: "cuda12-neutts-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-neutts" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-neutts - !!merge <<: *neutts name: "rocm-neutts-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-neutts" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-neutts - !!merge <<: *neutts name: "nvidia-l4t-neutts-development" uri: "quay.io/go-skynet/local-ai-backends:master-nvidia-l4t-arm64-neutts" mirrors: - localai/localai-backends:master-nvidia-l4t-arm64-neutts - !!merge <<: *mlx name: "mlx-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-mlx" mirrors: - localai/localai-backends:master-metal-darwin-arm64-mlx - !!merge <<: *mlx-vlm name: "mlx-vlm-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-mlx-vlm" mirrors: - localai/localai-backends:master-metal-darwin-arm64-mlx-vlm - !!merge <<: *mlx-audio name: "mlx-audio-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-mlx-audio" mirrors: - localai/localai-backends:master-metal-darwin-arm64-mlx-audio - !!merge <<: *kitten-tts name: "kitten-tts-development" uri: "quay.io/go-skynet/local-ai-backends:master-kitten-tts" mirrors: - localai/localai-backends:master-kitten-tts - !!merge <<: *huggingface name: "huggingface-development" uri: "quay.io/go-skynet/local-ai-backends:master-huggingface" mirrors: - localai/localai-backends:master-huggingface - !!merge <<: *local-store name: "local-store-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-local-store" mirrors: - localai/localai-backends:master-cpu-local-store - !!merge <<: *silero-vad name: "silero-vad-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-silero-vad" mirrors: - localai/localai-backends:master-cpu-silero-vad - !!merge <<: *piper name: "piper-development" uri: "quay.io/go-skynet/local-ai-backends:master-piper" mirrors: - localai/localai-backends:master-piper ## llama-cpp - !!merge <<: *llamacpp name: "darwin-x86-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-darwin-x86-llama-cpp" mirrors: - localai/localai-backends:latest-darwin-x86-llama-cpp - !!merge <<: *llamacpp name: "darwin-x86-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-darwin-x86-llama-cpp" mirrors: - localai/localai-backends:master-darwin-x86-llama-cpp - !!merge <<: *llamacpp name: "nvidia-l4t-arm64-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-nvidia-l4t-arm64-llama-cpp" mirrors: - localai/localai-backends:latest-nvidia-l4t-arm64-llama-cpp - !!merge <<: *llamacpp name: "nvidia-l4t-arm64-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-nvidia-l4t-arm64-llama-cpp" mirrors: - localai/localai-backends:master-nvidia-l4t-arm64-llama-cpp - !!merge <<: *llamacpp name: "cpu-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-llama-cpp" mirrors: - localai/localai-backends:latest-cpu-llama-cpp - !!merge <<: *llamacpp name: "cpu-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-llama-cpp" mirrors: - localai/localai-backends:master-cpu-llama-cpp - !!merge <<: *llamacpp name: "cuda11-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-llama-cpp" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-llama-cpp - !!merge <<: *llamacpp name: "cuda12-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-llama-cpp" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-llama-cpp - !!merge <<: *llamacpp name: "rocm-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-llama-cpp" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-llama-cpp - !!merge <<: *llamacpp name: "intel-sycl-f32-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f32-llama-cpp" mirrors: - localai/localai-backends:latest-gpu-intel-sycl-f32-llama-cpp - !!merge <<: *llamacpp name: "intel-sycl-f16-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f16-llama-cpp" mirrors: - localai/localai-backends:latest-gpu-intel-sycl-f16-llama-cpp - !!merge <<: *llamacpp name: "vulkan-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-vulkan-llama-cpp" mirrors: - localai/localai-backends:latest-gpu-vulkan-llama-cpp - !!merge <<: *llamacpp name: "vulkan-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-vulkan-llama-cpp" mirrors: - localai/localai-backends:master-gpu-vulkan-llama-cpp - !!merge <<: *llamacpp name: "metal-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-llama-cpp" mirrors: - localai/localai-backends:latest-metal-darwin-arm64-llama-cpp - !!merge <<: *llamacpp name: "metal-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-llama-cpp" mirrors: - localai/localai-backends:master-metal-darwin-arm64-llama-cpp - !!merge <<: *llamacpp name: "cuda11-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-llama-cpp" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-llama-cpp - !!merge <<: *llamacpp name: "cuda12-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-llama-cpp" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-llama-cpp - !!merge <<: *llamacpp name: "rocm-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-llama-cpp" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-llama-cpp - !!merge <<: *llamacpp name: "intel-sycl-f32-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f32-llama-cpp" mirrors: - localai/localai-backends:master-gpu-intel-sycl-f32-llama-cpp - !!merge <<: *llamacpp name: "intel-sycl-f16-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f16-llama-cpp" mirrors: - localai/localai-backends:master-gpu-intel-sycl-f16-llama-cpp ## whisper - !!merge <<: *whispercpp name: "nvidia-l4t-arm64-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-nvidia-l4t-arm64-whisper" mirrors: - localai/localai-backends:latest-nvidia-l4t-arm64-whisper - !!merge <<: *whispercpp name: "nvidia-l4t-arm64-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-nvidia-l4t-arm64-whisper" mirrors: - localai/localai-backends:master-nvidia-l4t-arm64-whisper - !!merge <<: *whispercpp name: "cpu-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-whisper" mirrors: - localai/localai-backends:latest-cpu-whisper - !!merge <<: *whispercpp name: "metal-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-whisper" mirrors: - localai/localai-backends:latest-metal-darwin-arm64-whisper - !!merge <<: *whispercpp name: "metal-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-whisper" mirrors: - localai/localai-backends:master-metal-darwin-arm64-whisper - !!merge <<: *whispercpp name: "cpu-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-whisper" mirrors: - localai/localai-backends:master-cpu-whisper - !!merge <<: *whispercpp name: "cuda11-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-whisper" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-whisper - !!merge <<: *whispercpp name: "cuda12-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-whisper" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-whisper - !!merge <<: *whispercpp name: "rocm-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-whisper" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-whisper - !!merge <<: *whispercpp name: "intel-sycl-f32-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f32-whisper" mirrors: - localai/localai-backends:latest-gpu-intel-sycl-f32-whisper - !!merge <<: *whispercpp name: "intel-sycl-f16-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f16-whisper" mirrors: - localai/localai-backends:latest-gpu-intel-sycl-f16-whisper - !!merge <<: *whispercpp name: "vulkan-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-vulkan-whisper" mirrors: - localai/localai-backends:latest-gpu-vulkan-whisper - !!merge <<: *whispercpp name: "vulkan-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-vulkan-whisper" mirrors: - localai/localai-backends:master-gpu-vulkan-whisper - !!merge <<: *whispercpp name: "metal-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-whisper" mirrors: - localai/localai-backends:latest-metal-darwin-arm64-whisper - !!merge <<: *whispercpp name: "metal-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-whisper" mirrors: - localai/localai-backends:master-metal-darwin-arm64-whisper - !!merge <<: *whispercpp name: "cuda11-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-whisper" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-whisper - !!merge <<: *whispercpp name: "cuda12-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-whisper" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-whisper - !!merge <<: *whispercpp name: "rocm-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-whisper" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-whisper - !!merge <<: *whispercpp name: "intel-sycl-f32-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f32-whisper" mirrors: - localai/localai-backends:master-gpu-intel-sycl-f32-whisper - !!merge <<: *whispercpp name: "intel-sycl-f16-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f16-whisper" mirrors: - localai/localai-backends:master-gpu-intel-sycl-f16-whisper ## stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "cpu-stablediffusion-ggml" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-stablediffusion-ggml" mirrors: - localai/localai-backends:latest-cpu-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "cpu-stablediffusion-ggml-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-stablediffusion-ggml" mirrors: - localai/localai-backends:master-cpu-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "metal-stablediffusion-ggml" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-stablediffusion-ggml" mirrors: - localai/localai-backends:latest-metal-darwin-arm64-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "metal-stablediffusion-ggml-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-stablediffusion-ggml" mirrors: - localai/localai-backends:master-metal-darwin-arm64-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "vulkan-stablediffusion-ggml" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-vulkan-stablediffusion-ggml" mirrors: - localai/localai-backends:latest-gpu-vulkan-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "vulkan-stablediffusion-ggml-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-vulkan-stablediffusion-ggml" mirrors: - localai/localai-backends:master-gpu-vulkan-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "cuda12-stablediffusion-ggml" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-stablediffusion-ggml" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "intel-sycl-f32-stablediffusion-ggml" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f32-stablediffusion-ggml" - !!merge <<: *stablediffusionggml name: "intel-sycl-f16-stablediffusion-ggml" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f16-stablediffusion-ggml" mirrors: - localai/localai-backends:latest-gpu-intel-sycl-f16-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "cuda11-stablediffusion-ggml" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-stablediffusion-ggml" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "cuda12-stablediffusion-ggml-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-stablediffusion-ggml" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "intel-sycl-f32-stablediffusion-ggml-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f32-stablediffusion-ggml" mirrors: - localai/localai-backends:master-gpu-intel-sycl-f32-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "intel-sycl-f16-stablediffusion-ggml-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f16-stablediffusion-ggml" mirrors: - localai/localai-backends:master-gpu-intel-sycl-f16-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "cuda11-stablediffusion-ggml-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-stablediffusion-ggml" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "nvidia-l4t-arm64-stablediffusion-ggml-development" uri: "quay.io/go-skynet/local-ai-backends:master-nvidia-l4t-arm64-stablediffusion-ggml" mirrors: - localai/localai-backends:master-nvidia-l4t-arm64-stablediffusion-ggml - !!merge <<: *stablediffusionggml name: "nvidia-l4t-arm64-stablediffusion-ggml" uri: "quay.io/go-skynet/local-ai-backends:latest-nvidia-l4t-arm64-stablediffusion-ggml" mirrors: - localai/localai-backends:latest-nvidia-l4t-arm64-stablediffusion-ggml # vllm - !!merge <<: *vllm name: "vllm-development" capabilities: nvidia: "cuda12-vllm-development" amd: "rocm-vllm-development" intel: "intel-vllm-development" - !!merge <<: *vllm name: "cuda12-vllm" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-vllm" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-vllm - !!merge <<: *vllm name: "rocm-vllm" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-vllm" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-vllm - !!merge <<: *vllm name: "intel-vllm" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-vllm" mirrors: - localai/localai-backends:latest-gpu-intel-vllm - !!merge <<: *vllm name: "cuda12-vllm-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-vllm" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-vllm - !!merge <<: *vllm name: "rocm-vllm-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-vllm" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-vllm - !!merge <<: *vllm name: "intel-vllm-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-vllm" mirrors: - localai/localai-backends:master-gpu-intel-vllm # rfdetr - !!merge <<: *rfdetr name: "rfdetr-development" capabilities: nvidia: "cuda12-rfdetr-development" intel: "intel-rfdetr-development" #amd: "rocm-rfdetr-development" nvidia-l4t: "nvidia-l4t-arm64-rfdetr-development" default: "cpu-rfdetr-development" - !!merge <<: *rfdetr name: "cuda12-rfdetr" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-rfdetr" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-rfdetr - !!merge <<: *rfdetr name: "intel-rfdetr" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-rfdetr" mirrors: - localai/localai-backends:latest-gpu-intel-rfdetr # - !!merge <<: *rfdetr # name: "rocm-rfdetr" # uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-hipblas-rfdetr" # mirrors: # - localai/localai-backends:latest-gpu-hipblas-rfdetr - !!merge <<: *rfdetr name: "nvidia-l4t-arm64-rfdetr" uri: "quay.io/go-skynet/local-ai-backends:latest-nvidia-l4t-arm64-rfdetr" mirrors: - localai/localai-backends:latest-nvidia-l4t-arm64-rfdetr - !!merge <<: *rfdetr name: "cpu-rfdetr" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-rfdetr" mirrors: - localai/localai-backends:latest-cpu-rfdetr - !!merge <<: *rfdetr name: "cuda12-rfdetr-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-rfdetr" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-rfdetr - !!merge <<: *rfdetr name: "intel-rfdetr-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-rfdetr" mirrors: - localai/localai-backends:master-gpu-intel-rfdetr # - !!merge <<: *rfdetr # name: "rocm-rfdetr-development" # uri: "quay.io/go-skynet/local-ai-backends:master-gpu-hipblas-rfdetr" # mirrors: # - localai/localai-backends:master-gpu-hipblas-rfdetr - !!merge <<: *rfdetr name: "cpu-rfdetr-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-rfdetr" mirrors: - localai/localai-backends:master-cpu-rfdetr - !!merge <<: *rfdetr name: "intel-rfdetr" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-rfdetr" mirrors: - localai/localai-backends:latest-gpu-intel-rfdetr ## Rerankers - !!merge <<: *rerankers name: "rerankers-development" capabilities: nvidia: "cuda12-rerankers-development" intel: "intel-rerankers-development" amd: "rocm-rerankers-development" - !!merge <<: *rerankers name: "cuda11-rerankers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-rerankers" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-rerankers - !!merge <<: *rerankers name: "cuda12-rerankers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-rerankers" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-rerankers - !!merge <<: *rerankers name: "intel-rerankers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-rerankers" mirrors: - localai/localai-backends:latest-gpu-intel-rerankers - !!merge <<: *rerankers name: "rocm-rerankers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-rerankers" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-rerankers - !!merge <<: *rerankers name: "cuda11-rerankers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-rerankers" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-rerankers - !!merge <<: *rerankers name: "cuda12-rerankers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-rerankers" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-rerankers - !!merge <<: *rerankers name: "rocm-rerankers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-rerankers" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-rerankers - !!merge <<: *rerankers name: "intel-rerankers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-rerankers" mirrors: - localai/localai-backends:master-gpu-intel-rerankers ## Transformers - !!merge <<: *transformers name: "transformers-development" capabilities: nvidia: "cuda12-transformers-development" intel: "intel-transformers-development" amd: "rocm-transformers-development" - !!merge <<: *transformers name: "cuda12-transformers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-transformers" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-transformers - !!merge <<: *transformers name: "rocm-transformers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-transformers" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-transformers - !!merge <<: *transformers name: "intel-transformers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-transformers" mirrors: - localai/localai-backends:latest-gpu-intel-transformers - !!merge <<: *transformers name: "cuda11-transformers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-transformers" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-transformers - !!merge <<: *transformers name: "cuda11-transformers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-transformers" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-transformers - !!merge <<: *transformers name: "cuda12-transformers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-transformers" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-transformers - !!merge <<: *transformers name: "rocm-transformers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-transformers" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-transformers - !!merge <<: *transformers name: "intel-transformers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-transformers" mirrors: - localai/localai-backends:master-gpu-intel-transformers ## Diffusers - !!merge <<: *diffusers name: "diffusers-development" capabilities: nvidia: "cuda12-diffusers-development" intel: "intel-diffusers-development" amd: "rocm-diffusers-development" nvidia-l4t: "nvidia-l4t-diffusers-development" metal: "metal-diffusers-development" default: "cpu-diffusers-development" - !!merge <<: *diffusers name: "cpu-diffusers" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-diffusers" mirrors: - localai/localai-backends:latest-cpu-diffusers - !!merge <<: *diffusers name: "cpu-diffusers-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-diffusers" mirrors: - localai/localai-backends:master-cpu-diffusers - !!merge <<: *diffusers name: "nvidia-l4t-diffusers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-l4t-diffusers" mirrors: - localai/localai-backends:latest-gpu-nvidia-l4t-diffusers - !!merge <<: *diffusers name: "nvidia-l4t-diffusers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-l4t-diffusers" mirrors: - localai/localai-backends:master-gpu-nvidia-l4t-diffusers - !!merge <<: *diffusers name: "cuda12-diffusers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-diffusers" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-diffusers - !!merge <<: *diffusers name: "rocm-diffusers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-diffusers" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-diffusers - !!merge <<: *diffusers name: "cuda11-diffusers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-diffusers" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-diffusers - !!merge <<: *diffusers name: "intel-diffusers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-diffusers" mirrors: - localai/localai-backends:latest-gpu-intel-diffusers - !!merge <<: *diffusers name: "cuda11-diffusers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-diffusers" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-diffusers - !!merge <<: *diffusers name: "cuda12-diffusers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-diffusers" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-diffusers - !!merge <<: *diffusers name: "rocm-diffusers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-diffusers" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-diffusers - !!merge <<: *diffusers name: "intel-diffusers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-diffusers" mirrors: - localai/localai-backends:master-gpu-intel-diffusers - !!merge <<: *diffusers name: "metal-diffusers" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-diffusers" mirrors: - localai/localai-backends:latest-metal-darwin-arm64-diffusers - !!merge <<: *diffusers name: "metal-diffusers-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-diffusers" mirrors: - localai/localai-backends:master-metal-darwin-arm64-diffusers ## exllama2 - !!merge <<: *exllama2 name: "exllama2-development" capabilities: nvidia: "cuda12-exllama2-development" intel: "intel-exllama2-development" - !!merge <<: *exllama2 name: "cuda11-exllama2" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-exllama2" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-exllama2 - !!merge <<: *exllama2 name: "cuda12-exllama2" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-exllama2" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-exllama2 - !!merge <<: *exllama2 name: "cuda11-exllama2-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-exllama2" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-exllama2 - !!merge <<: *exllama2 name: "cuda12-exllama2-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-exllama2" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-exllama2 ## kokoro - !!merge <<: *kokoro name: "kokoro-development" capabilities: nvidia: "cuda12-kokoro-development" intel: "intel-kokoro-development" amd: "rocm-kokoro-development" nvidia-l4t: "nvidia-l4t-kokoro-development" - !!merge <<: *kokoro name: "cuda11-kokoro-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-kokoro" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-kokoro - !!merge <<: *kokoro name: "cuda12-kokoro-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-kokoro" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-kokoro - !!merge <<: *kokoro name: "rocm-kokoro-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-kokoro" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-kokoro - !!merge <<: *kokoro name: "intel-kokoro" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-kokoro" mirrors: - localai/localai-backends:latest-gpu-intel-kokoro - !!merge <<: *kokoro name: "intel-kokoro-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-kokoro" mirrors: - localai/localai-backends:master-gpu-intel-kokoro - !!merge <<: *kokoro name: "nvidia-l4t-kokoro" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-l4t-kokoro" mirrors: - localai/localai-backends:latest-gpu-nvidia-l4t-kokoro - !!merge <<: *kokoro name: "nvidia-l4t-kokoro-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-l4t-kokoro" mirrors: - localai/localai-backends:master-gpu-nvidia-l4t-kokoro - !!merge <<: *kokoro name: "cuda11-kokoro" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-kokoro" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-kokoro - !!merge <<: *kokoro name: "cuda12-kokoro" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-kokoro" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-kokoro - !!merge <<: *kokoro name: "rocm-kokoro" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-kokoro" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-kokoro ## faster-whisper - !!merge <<: *faster-whisper name: "faster-whisper-development" capabilities: nvidia: "cuda12-faster-whisper-development" intel: "intel-faster-whisper-development" amd: "rocm-faster-whisper-development" - !!merge <<: *faster-whisper name: "cuda11-faster-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-faster-whisper" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-faster-whisper - !!merge <<: *faster-whisper name: "cuda12-faster-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-faster-whisper" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-faster-whisper - !!merge <<: *faster-whisper name: "rocm-faster-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-faster-whisper" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-faster-whisper - !!merge <<: *faster-whisper name: "intel-faster-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-faster-whisper" mirrors: - localai/localai-backends:latest-gpu-intel-faster-whisper - !!merge <<: *faster-whisper name: "intel-faster-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-faster-whisper" mirrors: - localai/localai-backends:master-gpu-intel-faster-whisper ## coqui - !!merge <<: *coqui name: "coqui-development" capabilities: nvidia: "cuda12-coqui-development" intel: "intel-coqui-development" amd: "rocm-coqui-development" - !!merge <<: *coqui name: "cuda11-coqui" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-coqui" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-coqui - !!merge <<: *coqui name: "cuda12-coqui" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-coqui" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-coqui - !!merge <<: *coqui name: "cuda11-coqui-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-coqui" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-coqui - !!merge <<: *coqui name: "cuda12-coqui-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-coqui" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-coqui - !!merge <<: *coqui name: "rocm-coqui-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-coqui" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-coqui - !!merge <<: *coqui name: "intel-coqui" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-coqui" mirrors: - localai/localai-backends:latest-gpu-intel-coqui - !!merge <<: *coqui name: "intel-coqui-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-coqui" mirrors: - localai/localai-backends:master-gpu-intel-coqui - !!merge <<: *coqui name: "rocm-coqui" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-coqui" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-coqui ## bark - !!merge <<: *bark name: "bark-development" capabilities: nvidia: "cuda12-bark-development" intel: "intel-bark-development" amd: "rocm-bark-development" - !!merge <<: *bark name: "cuda11-bark-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-bark" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-bark - !!merge <<: *bark name: "cuda11-bark" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-bark" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-bark - !!merge <<: *bark name: "rocm-bark-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-bark" mirrors: - localai/localai-backends:master-gpu-rocm-hipblas-bark - !!merge <<: *bark name: "intel-bark" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-bark" mirrors: - localai/localai-backends:latest-gpu-intel-bark - !!merge <<: *bark name: "intel-bark-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-bark" mirrors: - localai/localai-backends:master-gpu-intel-bark - !!merge <<: *bark name: "cuda12-bark" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-bark" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-bark - !!merge <<: *bark name: "rocm-bark" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-bark" mirrors: - localai/localai-backends:latest-gpu-rocm-hipblas-bark - !!merge <<: *bark name: "cuda12-bark-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-bark" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-bark - !!merge <<: *barkcpp name: "bark-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-bark-cpp" alias: "bark-cpp" ## chatterbox - !!merge <<: *chatterbox name: "chatterbox-development" capabilities: nvidia: "cuda12-chatterbox-development" metal: "metal-chatterbox-development" default: "cpu-chatterbox-development" nvidia-l4t: "nvidia-l4t-arm64-chatterbox" - !!merge <<: *chatterbox name: "cpu-chatterbox" uri: "quay.io/go-skynet/local-ai-backends:latest-cpu-chatterbox" mirrors: - localai/localai-backends:latest-cpu-chatterbox - !!merge <<: *chatterbox name: "cpu-chatterbox-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-chatterbox" mirrors: - localai/localai-backends:master-cpu-chatterbox - !!merge <<: *chatterbox name: "nvidia-l4t-arm64-chatterbox" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-l4t-arm64-chatterbox" mirrors: - localai/localai-backends:latest-gpu-nvidia-l4t-arm64-chatterbox - !!merge <<: *chatterbox name: "nvidia-l4t-arm64-chatterbox-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-l4t-arm64-chatterbox" mirrors: - localai/localai-backends:master-gpu-nvidia-l4t-arm64-chatterbox - !!merge <<: *chatterbox name: "metal-chatterbox" uri: "quay.io/go-skynet/local-ai-backends:latest-metal-darwin-arm64-chatterbox" mirrors: - localai/localai-backends:latest-metal-darwin-arm64-chatterbox - !!merge <<: *chatterbox name: "metal-chatterbox-development" uri: "quay.io/go-skynet/local-ai-backends:master-metal-darwin-arm64-chatterbox" mirrors: - localai/localai-backends:master-metal-darwin-arm64-chatterbox - !!merge <<: *chatterbox name: "cuda12-chatterbox-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-chatterbox" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-chatterbox - !!merge <<: *chatterbox name: "cuda11-chatterbox" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-chatterbox" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-chatterbox - !!merge <<: *chatterbox name: "cuda11-chatterbox-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-chatterbox" mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-11-chatterbox - !!merge <<: *chatterbox name: "cuda12-chatterbox" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-12-chatterbox" mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-chatterbox