Replace Ollama with llama-server (OpenAI-compatible API) #14

Merged
ryan merged 1 commits from feature/llama-cpp-integration into main 2026-01-31 21:41:19 -05:00
Owner
  • Update llm.py to use OpenAI client with custom base_url for llama-server
  • Update agents.py to use ChatOpenAI instead of ChatOllama
  • Remove unused ollama imports from main.py, chunker.py, query.py
  • Add LLAMA_SERVER_URL and LLAMA_MODEL_NAME env vars
  • Remove ollama and langchain-ollama dependencies

Co-Authored-By: Claude Opus 4.5 noreply@anthropic.com

- Update llm.py to use OpenAI client with custom base_url for llama-server - Update agents.py to use ChatOpenAI instead of ChatOllama - Remove unused ollama imports from main.py, chunker.py, query.py - Add LLAMA_SERVER_URL and LLAMA_MODEL_NAME env vars - Remove ollama and langchain-ollama dependencies Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
ryan added 1 commit 2026-01-31 21:40:33 -05:00
- Update llm.py to use OpenAI client with custom base_url for llama-server
- Update agents.py to use ChatOpenAI instead of ChatOllama
- Remove unused ollama imports from main.py, chunker.py, query.py
- Add LLAMA_SERVER_URL and LLAMA_MODEL_NAME env vars
- Remove ollama and langchain-ollama dependencies

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
ryan merged commit f0f72cce36 into main 2026-01-31 21:41:19 -05:00
Sign in to join this conversation.
No Reviewers
No Label
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: ryan/simbarag#14