Files
neovim_config/docs/OLLAMA_QUICK_SETUP.md
2026-02-13 08:49:20 -06:00

1.2 KiB

Quick Setup for Other Machines

Step 1: Find Your Ollama Server's Tailscale IP

On your Ollama server machine, run:

tailscale ip -4

Example output: 100.123.45.67

Step 2: Set Environment Variable on Other Machines

Add this to your shell config file (~/.zshrc, ~/.bashrc, or ~/.config/fish/config.fish):

export OLLAMA_ENDPOINT="http://100.123.45.67:11434"

Replace 100.123.45.67 with your actual Tailscale IP.

Step 3: Verify Connection

Test the connection:

curl http://100.123.45.67:11434/api/tags

You should see a JSON response with available models.

Step 4: Use in Neovim

Start Neovim and press <leader>cll to chat with Ollama!

Troubleshooting

Issue Solution
Connection refused Verify Ollama is running: curl http://100.123.45.67:11434/api/tags
Tailscale not working Run tailscale status on both machines
Model not found Pull the model on the server: ollama pull mistral
Slow responses Try a smaller model or check server resources

Available Keymaps

  • <leader>cll - Chat with Ollama
  • <leader>cc - Chat with Claude Haiku
  • <leader>cs - Chat with Claude Sonnet
  • <leader>co - Chat with Claude Opus