Groq
Groq offers probably the fastest LLM inference service as of writing. Unfortunately, the billing settings are not yet avail and the free tier is limited.
Check Groq docs to obtain an api key
Configure the Toolkit API
Locate the file ./config/api/.env
and add the following configurations
LLM_SERVICE=groq
# GROQ_API_KEY Provide Groq api key for chat
GROQ_API_KEY='api key'
## the following are the defaults, edit if needed
# GROQ_MODEL Default Groq model used as fallback
GROQ_MODEL='mixtral-8x7b-32768'
# GROQ_CHAT_MODELS Supported chat models from Groq. Leave empty to allow all available.
GROQ_CHAT_MODELS='gemma-7b-it,gemma2-9b-it,llama2-70b-4096,llama3-70b-8192,llama3-8b-8192,mixtral-8x7b-32768,llama3-groq-8b-8192-tool-use-preview,llama3-groq-70b-8192-tool-use-preview'