Skip to content

Test / ollama

Test / ollama #178

Workflow file for this run

# This workflow comes from https://github.com/ofek/hatch-mypyc
# https://github.com/ofek/hatch-mypyc/blob/5a198c0ba8660494d02716cfc9d79ce4adfb1442/.github/workflows/test.yml
name: Test / ollama
on:
schedule:
- cron: "0 0 * * *"
pull_request:
paths:
- "integrations/ollama/**"
- ".github/workflows/ollama.yml"
defaults:
run:
working-directory: integrations/ollama
concurrency:
group: ollama-${{ github.head_ref }}
cancel-in-progress: true
env:
PYTHONUNBUFFERED: "1"
FORCE_COLOR: "1"
LLM_FOR_TESTS: "orca-mini"
EMBEDDER_FOR_TESTS: "nomic-embed-text"
jobs:
run:
name: Python ${{ matrix.python-version }} on ${{ startsWith(matrix.os, 'macos-') && 'macOS' || startsWith(matrix.os, 'windows-') && 'Windows' || 'Linux' }}
runs-on: ${{ matrix.os }}
strategy:
fail-fast: false
matrix:
os: [ubuntu-latest]
python-version: ["3.9","3.10","3.11"]
services:
ollama:
image: ollama/ollama:latest
options: --name ollama
ports:
- 11434:11434
steps:
- uses: actions/checkout@v4
- name: Set up Python ${{ matrix.python-version }}
uses: actions/setup-python@v5
with:
python-version: ${{ matrix.python-version }}
- name: Install Hatch
run: pip install --upgrade hatch
- name: Lint
if: matrix.python-version == '3.9'
run: hatch run lint:all
- name: Pull the LLM in the Ollama service
run: docker exec ollama ollama pull ${{ env.LLM_FOR_TESTS }}
- name: Pull the Embedding Model in the Ollama service
run: docker exec ollama ollama pull ${{ env.EMBEDDER_FOR_TESTS }}
- name: Generate docs
if: matrix.python-version == '3.9' && runner.os == 'Linux'
run: hatch run docs
- name: Run tests
run: hatch run cov