diff --git a/.devcontainer/Containerfile b/.devcontainer/Containerfile new file mode 100644 index 000000000..4ef78ce5a --- /dev/null +++ b/.devcontainer/Containerfile @@ -0,0 +1,9 @@ +FROM quay.io/containers/podman:v4.9.3 + +USER root + +COPY playground/tests/requirements.txt . + +RUN dnf install -y python3.11 python3-pip buildah git && \ + dnf clean all && \ + pip3 install -r requirements.txt diff --git a/.devcontainer/devcontainer.json b/.devcontainer/devcontainer.json new file mode 100644 index 000000000..d9945aa9f --- /dev/null +++ b/.devcontainer/devcontainer.json @@ -0,0 +1,12 @@ +{ + "name": "recepies", + "build": { + "dockerfile": "Containerfile", + "context": ".." + }, + "privileged": true, + "containerEnv": { + "REGISTRY": "ghcr.io", + "IMAGE_NAME": "ai-lab-recipes/playground" + } +} diff --git a/.github/workflows/playground.yaml b/.github/workflows/playground.yaml new file mode 100644 index 000000000..c377ea64f --- /dev/null +++ b/.github/workflows/playground.yaml @@ -0,0 +1,48 @@ +name: playground + +on: + pull_request: + branches: + - main + push: + branches: + - main + +env: + REGISTRY: ghcr.io + IMAGE_NAME: ${{ github.repository_owner }}/playground + +jobs: + build-and-push-image: + runs-on: ubuntu-latest + permissions: + contents: read + packages: write + services: + registry: + image: registry:2.8.3 + ports: + - 5000:5000 + steps: + - uses: actions/checkout@v4.1.1 + + - name: Login to ghcr + uses: docker/login-action@v3.1.0 + with: + registry: ${{ env.REGISTRY }} + username: ${{ github.actor }} + password: ${{ secrets.GITHUB_TOKEN }} + + - name: Buildah Action + uses: redhat-actions/buildah-build@v2.13 + with: + image: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }} + tags: latest + containerfiles: ./playground/Containerfile + context: playground + + - name: Set up Python + uses: actions/setup-python@v5.0.0 + + - name: Run tests + run: make -f playground/Makefile test diff --git a/.gitignore b/.gitignore index 3c2129ef7..50cf3e3cd 100644 --- a/.gitignore +++ b/.gitignore @@ -1,3 +1,4 @@ *.gguf *.bin *_pycache_* +port_check.lock diff --git a/playground/Containerfile b/playground/Containerfile index e47b81dd9..89536f5b3 100644 --- a/playground/Containerfile +++ b/playground/Containerfile @@ -1,8 +1,7 @@ -FROM registry.access.redhat.com/ubi9/python-39:latest +FROM registry.access.redhat.com/ubi9/python-311:1-52 WORKDIR /locallm COPY requirements.txt /locallm/requirements.txt -RUN pip install --upgrade pip -RUN pip install --no-cache-dir --upgrade -r /locallm/requirements.txt +RUN pip install --no-cache-dir --verbose -r /locallm/requirements.txt COPY run.sh run.sh EXPOSE 8001 ENTRYPOINT [ "sh", "run.sh" ] diff --git a/playground/Makefile b/playground/Makefile new file mode 100644 index 000000000..997861ec7 --- /dev/null +++ b/playground/Makefile @@ -0,0 +1,18 @@ +.PHONY: build +build: + podman build -f playground/Containerfile -t ghcr.io/ai-lab-recipes/playground --format docker playground + +models/llama-2-7b-chat.Q5_K_S.gguf: + curl -s -S -L -f https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGUF/resolve/main/llama-2-7b-chat.Q5_K_S.gguf -z $@ -o $@.tmp && mv -f $@.tmp $@ 2>/dev/null || rm -f $@.tmp $@ + +.PHONY: download +download: + pip install -r playground/tests/requirements.txt + +.PHONY: run +run: install models/llama-2-7b-chat.Q5_K_S.gguf + podman run -it -d -p 8001:8001 -v ./models:/locallm/models:ro,Z -e MODEL_PATH=models/llama-2-7b-chat.Q5_K_S.gguf -e HOST=0.0.0.0 -e PORT=8001 --net=host ghcr.io/redhat-et/playground + +.PHONY: test +test: models/llama-2-7b-chat.Q5_K_S.gguf download + pytest --log-cli-level NOTSET diff --git a/playground/README.md b/playground/README.md index cdc8caaed..3a5ca62db 100644 --- a/playground/README.md +++ b/playground/README.md @@ -3,7 +3,11 @@ From this directory, ```bash -podman build -t playground:image . +podman build -t playground . +``` +or +```bash +make -f Containerfile build ``` ### Download Model @@ -20,6 +24,10 @@ cd ../models wget cd ../ ``` +or +```bash +make -f Containerfile download +``` ### Deploy Model Service @@ -34,7 +42,11 @@ podman run --rm -it -d \ -e MODEL_PATH=models/ \ -e HOST=0.0.0.0 \ -e PORT=8001 \ - playground:image` + playground` +``` +or +```bash +make -f Containerfile run ``` #### Multiple Model Service: @@ -68,5 +80,14 @@ podman run --rm -it -d \ -p 8001:8001 \ -v Local/path/to/locallm/models:/locallm/models:ro,Z \ -e CONFIG_PATH=models/ \ - playground:image + playground +``` + +### DEV environment + +The environment is implemented with devcontainer technology. + +Running tests +```bash +make -f Containerfile test ``` diff --git a/playground/requirements.txt b/playground/requirements.txt index bbea3dd85..a3ffc4c08 100644 --- a/playground/requirements.txt +++ b/playground/requirements.txt @@ -1 +1,2 @@ -llama-cpp-python[server] \ No newline at end of file +llama-cpp-python[server]==0.2.57 +pip==24.0 diff --git a/playground/tests/__init__.py b/playground/tests/__init__.py new file mode 100644 index 000000000..e69de29bb diff --git a/playground/tests/conftest.py b/playground/tests/conftest.py new file mode 100644 index 000000000..96deb7dda --- /dev/null +++ b/playground/tests/conftest.py @@ -0,0 +1,32 @@ +import pytest_container +import os + + +MS = pytest_container.Container( + url=f"containers-storage:{os.environ['REGISTRY']}/{os.environ['IMAGE_NAME']}", + volume_mounts=[ + pytest_container.container.BindMount( + container_path="/locallm/models", + host_path="./models", + flags=["ro"] + ) + ], + extra_environment_variables={ + "MODEL_PATH": "models/llama-2-7b-chat.Q5_K_S.gguf", + "HOST": "0.0.0.0", + "PORT": "8001" + }, + forwarded_ports=[ + pytest_container.PortForwarding( + container_port=8001, + host_port=8001 + ) + ], + extra_launch_args=["--net=host"] + ) + +def pytest_generate_tests(metafunc): + pytest_container.auto_container_parametrize(metafunc) + +def pytest_addoption(parser): + pytest_container.add_logging_level_options(parser) diff --git a/playground/tests/requirements.txt b/playground/tests/requirements.txt new file mode 100644 index 000000000..751d336dc --- /dev/null +++ b/playground/tests/requirements.txt @@ -0,0 +1,6 @@ +pip==24.0 +pytest-container==0.4.0 +pytest-testinfra==10.1.0 +pytest==8.1.1 +requests==2.31.0 +tenacity==8.2.3 diff --git a/playground/tests/test_alive.py b/playground/tests/test_alive.py new file mode 100644 index 000000000..fcad510a0 --- /dev/null +++ b/playground/tests/test_alive.py @@ -0,0 +1,13 @@ +import pytest_container +from .conftest import MS +import tenacity + +CONTAINER_IMAGES = [MS] + + +def test_etc_os_release_present(auto_container: pytest_container.container.ContainerData): + assert auto_container.connection.file("/etc/os-release").exists + +@tenacity.retry(stop=tenacity.stop_after_attempt(5), wait=tenacity.wait_exponential()) +def test_alive(auto_container: pytest_container.container.ContainerData, host): + host.run_expect([0],f"curl http://localhost:{auto_container.forwarded_ports[0].host_port}",).stdout.strip()