Configuring Providers
Last updated
Was this helpful?
Last updated
Was this helpful?
ServerAssistantAI uses a flexible provider system for , , and . Users can choose from built-in providers, addon providers, pre-configured OpenAI-compatible providers, or custom providers with custom base URLs.
Built-in providers, such as Cohere and OpenAI, are included with ServerAssistantAI and can be configured directly in the config.yml
file.
To use a built-in provider:
Set the provider
field to the provider name (e.g., cohere
or openai
) in the config.yml
file for either LLM and/or Embedding.
Add the appropriate model
name for the chosen provider.
Add the API key to the corresponding section in the credentials.yml
file, if not already present.
Reload the plugin to apply the changes.
Addon providers, such as Anthropic and Google-AIStudio, require the installation of their respective addons, which are available for free in the #saai-addons
Discord channel.
To use an addon provider:
Download the addon .jar
file from the #saai-addons
Discord channel.
Place the addon file in the plugins/
directory.
Restart the server to load the addon.
Set the provider
field to the addon provider name (e.g., anthropic
, google-aistudio
) in the config.yml
file for either LLM and/or Embedding.
Add the appropriate model
name for the chosen provider.
Add the API key in the newly generated field in the credentials.yml
file.
Reload the plugin to apply the changes.
Pre-configured OpenAI-compatible providers, such as Groq and Perplexity, are already built into ServerAssistantAI and can be easily configured using their respective endpoint URLs.
To use a pre-configured OpenAI-compatible provider:
Set the provider
field to the pre-configured provider name (e.g., groq
, perplexity
) in the config.yml
file for either LLM and/or Embedding.
Add the appropriate model
name for the chosen provider.
Reload the plugin to generate the new field in the credentials.yml
file.
Fill out the credentials in the newly generated field in the credentials.yml
file.
Reload the plugin again to apply the changes.
Custom providers with custom base URLs, such as LM-Studio and Ollama, allow users to integrate with any OpenAI API-compatible service by specifying the custom endpoint URL in the config.yml
file.
To use a custom provider with a custom base URL:
Set the provider
field to openai-variant
in the config.yml
file for either LLM and/or Embedding.
Add the appropriate model
name for the chosen provider.
Add the base_url
field under the provider field and set it to your custom URL, without including /chat/completions
if it is an LLM and /embeddings
if it is an embedding model within the URL. For example, if an LLM's endpoint URL is http://localhost:1234/v1/chat/completions, then the base_url
should be http://localhost:1234/v1:
Add the API key to the OpenAI section in the credentials.yml
file. If there is no API key, add random information to the API key field.
Reload the plugin to apply the changes.
Remember to reload ServerAssistantAI after making changes to the config.yml
file for the changes to take effect.