Models
RecurseChat offers versatile support for AI models, including a default suite of models as well as the capability to create extensively customized models.
Default Models
We support Meta Llama 3 (opens in a new tab), Mistral (opens in a new tab), Llava (opens in a new tab), WizardCoder Python (opens in a new tab), as well as OpenAI GPT-3 and GPT-4 models (If you bring your own API key) out of the box. Note that to use GPT-4, you would need to have made $5 payment to your OpenAI account (usage tier 1 (opens in a new tab)).
Local GGUF Models
We support custom GGUF models with your local GGUF files. You can create any number of local GGUF model and delete GGUF models.
RecurseChat is powered by llama.cpp (opens in a new tab) under the hood. The app should work with most GGUF models that llama.cpp supports.
OpenAI Chat Completion Models
In addition, we support chat completion endpoints are compatible with OpenAI API endpoints. One example below is using Ollama with RecurseChat:
While RecurseChat doesn't need Ollama to operate (It has a built-in llama.cpp executable), You can use local Ollama with RecurseChat through Ollama's support of OpenAI compatible API (opens in a new tab).
The steps are:
- Serve an Ollama model with
ollama serve - Create a new OpenAI API model: go to top right of the model tab, and click the "New Model" button, and click "New OpenAI Chat Completion model".
- To use Ollama model, set base url to http://127.0.0.1:11434/v1 and set model id to an Ollama model id like mistral.