AI Providers
ServerAssistantAI supports a wide range of AI providers for both language models (LLMs) and embeddings.
What Are Providers?
Providers are configurable components that enable ServerAssistantAI's customizability and flexibility. They serve as the backbone of the plugin's ability to integrate with various AI services, giving server owners the power to tailor the AI assistant's capabilities to their specific needs.
ServerAssistantAI offers several ways to configure and extend its functionality through a flexible provider system for embedding models, chat models (LLMs), and question detection, including Built-in Providers (ready-to-use), Addon Providers (additional providers installed via addons), Pre-configured OpenAI-compatible Providers (built-in providers with predefined endpoint URLs), and Custom Providers with Custom Base URLs (integration with any OpenAI API-compatible service using a custom endpoint URL).
To configure a provider, simply specify its name and any required options in the config.yml file using the following format:
section_name:
provider: 'openai' # Example provider
option1: valueEach provider has its own set of options. Some options are shared across providers, while others may have the same name but behave differently depending on the provider. The specific functionality and configuration are determined by the selected provider.
Supported AI Providers
Built-in
LLM & Embedding
Free & Paid
Provides access to Cohere's language and embedding models, with RAG capabilities to improve performance.
Addon
LLM & Embedding
Paid
Allows integration with Azure OpenAI Service for both LLM and embedding capabilities.
Pre-configured
LLM & Embedding
Paid
Run the latest ML models with ease using DeepInfra's simple REST API.
Pre-configured
LLM & Embedding
Paid
Fireworks.ai is a fast inference platform for serving generative AI models efficiently.
Addon
LLM & Embedding
Free & Paid
Provides access to Google's most advanced Gemini generative AI models.
Pre-configured
LLM
Temporarily Free
Utilizes Groq's LPU (Language Processing Unit) Inference Engine for fast LLM inference.
Addon
LLM & Embedding
Free
Provides access to thousands of open-source models for free through the HuggingFace Inference API.
Pre-configured
LLM
Paid
Kolank is an AI routing platform that connects to various models, ensuring high-quality responses.
Custom baseURL
LLM & Embedding
Self-hosted
A desktop app for running local models on your computer, supporting models from HuggingFace.
Custom baseURL
LLM & Embedding
Self-hosted
Open-source, OpenAI drop-in alternative REST API for local inferencing without a GPU.
Addon
LLM & Embedding
Paid
Integrates Mistral AI models for both LLM and embedding capabilities.
Pre-configured
LLM & Embedding
Free & Paid
Offers reliable API access to GPT-4, Gemini 1.5, Llama 3B, and various other language and embedding models.
Pre-configured
LLM
Free & Paid
Integrates NVIDIA's optimized AI models for efficient LLM functionality.
Pre-configured
LLM
Paid
Harness the latest AI innovations with OctoAI's efficient, reliable, and customizable AI systems for your apps.
Custom baseURL
LLM
Self-hosted
Allows self-hosting of Ollama, a lightweight framework for running language models locally.
Custom baseURL
LLM
Self-hosted
Allows developers to run any open-source LLMs (Llama 3.1, Qwen2, Phi3 and more) or custom models.
Pre-configured
LLM
Free & Paid
Standardized API for switching between models and providers, prioritizing price or performance.
Pre-configured
LLM
Paid
Perplexity AI's API enables users to use Perplexity Models and Open-Source LLMs.
Pre-configured
LLM & Embedding
Paid
Fast, cost-efficient, and scalable inference for open-source models like Llama-3.
Pre-configured
LLM
Paid
Offers language models like Yi-1.5, delivering strong performance in instruction-following.
Pre-configured
LLM & Embedding
Free & Paid
Brings xAI's powerful Grok models, enabling advanced text capabilities.
OpenAI-Variant
Custom baseURL
LLM &/or Embedding
Free or Paid
Allows integration with any OpenAI API-compatible service. Users can set up custom endpoints by specifying the base URL in the config.yml file.
All providers that are not built-in or OpenAI variants require the installation of their respective addons, which are available for free.
Last updated
Was this helpful?