Skip to content

feat: support togetherAI via /completions #1724

feat: support togetherAI via /completions

feat: support togetherAI via /completions #1724

Workflow file for this run

name: Endpoint (Ollama)
env:
OLLAMA_BASE_URL: "http://localhost:11434"
on:
push:
branches: [ main ]
pull_request:
branches: [ main ]
jobs:
test:
runs-on: ubuntu-latest
timeout-minutes: 15
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Start Ollama Server
run: |
curl -fsSL https://ollama.com/install.sh | sh
ollama serve &
sleep 10 # wait for server
ollama pull thewindmom/hermes-3-llama-3.1-8b
ollama pull mxbai-embed-large
- name: "Setup Python, Poetry and Dependencies"
uses: packetcoders/action-setup-cache-python-poetry@main
with:
python-version: "3.12"
poetry-version: "1.8.2"
install-args: "-E dev -E ollama"
- name: Test LLM endpoint
run: |
poetry run pytest -s -vv tests/test_endpoints.py::test_llm_endpoint_ollama
- name: Test embedding endpoint
run: |
poetry run pytest -s -vv tests/test_endpoints.py::test_embedding_endpoint_ollama
- name: Test provider
run: |
poetry run pytest -s -vv tests/test_providers.py::test_ollama