Quick start
Prerequisites
Before you start using Twinny you need to have access to an inference provider. An inference provider is a local or cloud hosted server that runs the AI models.
The recommended way to do this is to use Ollama. Ollama makes it easy to run your models locally and exposes them as an OpenAI compatible API. Performance will depend on your hardware and chosen model, see Ollama’s documentation for more information.
Installing the extension
Installing Ollama as an inference provider
- Visit Install Ollama and follow the instructions to install Ollama on your machine.
- Choose a model from the list of models available on Ollama. The recommended models are codellama:7b-instruct for chat and codellama:7b-code for fill-in-middle.
Once both the extension and Ollama are installed you can start using Twinny.
- Open VS code (if already open a restart might be needed) and press
CTRL+SHIFT+Z CTRL+SHIFT+T
to open the side panel.
You should see the 🤖 icon indicating that Twinny is ready to use. The icon will change to a spinner when Twinny is making a call to the inference provider.