Configuring Providers
ServerAssistantAI uses a flexible provider system for chat models (LLMs), embedding models, and question detection. Users can choose from built-in providers, addon providers, pre-configured OpenAI-compatible providers, or custom providers with custom base URLs.
Built-in Providers
Built-in providers, such as Cohere and OpenAI, are included with ServerAssistantAI and can be configured directly in the config.yml
file.
To use a built-in provider:
Set the
provider
field to the provider name (e.g.,cohere
oropenai
) in theconfig.yml
file for either LLM and/or Embedding.Add the appropriate
model
name for the chosen provider.Add the API key to the corresponding section in the
credentials.yml
file, if not already present.Reload the plugin to apply the changes.
Addon Providers
Addon providers, such as Anthropic and Google-AIStudio, require the installation of their respective addons, which are available for free in the #saai-addons
Discord channel.
To use an addon provider:
Download the addon
.jar
file from the#saai-addons
Discord channel.Place the addon file in the
plugins/
directory.Restart the server to load the addon.
Set the
provider
field to the addon provider name (e.g.,anthropic
,google-aistudio
) in theconfig.yml
file for either LLM and/or Embedding.Add the appropriate
model
name for the chosen provider.Add the API key in the newly generated field in the
credentials.yml
file.Reload the plugin to apply the changes.
Pre-configured OpenAI-compatible Providers
Pre-configured OpenAI-compatible providers, such as Groq and Perplexity, are already built into ServerAssistantAI and can be easily configured using their respective endpoint URLs.
To use a pre-configured OpenAI-compatible provider:
Set the
provider
field to the pre-configured provider name (e.g.,groq
,perplexity
) in theconfig.yml
file for either LLM and/or Embedding.Add the appropriate
model
name for the chosen provider.Reload the plugin to generate the new field in the
credentials.yml
file.Fill out the credentials in the newly generated field in the
credentials.yml
file.Reload the plugin again to apply the changes.
Custom Providers with Custom Base URLs
Custom providers with custom base URLs, such as LM-Studio and Ollama, allow users to integrate with any OpenAI API-compatible service by specifying the custom endpoint URL in the config.yml
file.
To use a custom provider with a custom base URL:
Set the
provider
field toopenai-variant
in theconfig.yml
file for either LLM and/or Embedding.Add the appropriate
model
name for the chosen provider.Add the
base_url
field under the provider field and set it to your custom URL, without including/chat/completions
if it is an LLM and/embeddings
if it is an embedding model within the URL. For example, if an LLM's endpoint URL is http://localhost:1234/v1/chat/completions, then thebase_url
should be http://localhost:1234/v1:Add the API key to the OpenAI section in the
credentials.yml
file. If there is no API key, add random information to the API key field.Reload the plugin to apply the changes.
Remember to reload ServerAssistantAI after making changes to the config.yml
file for the changes to take effect.