diff --git a/chatbot/ai-lab.yaml b/chatbot/ai-lab.yaml index 760155f91..6f0bbcc84 100644 --- a/chatbot/ai-lab.yaml +++ b/chatbot/ai-lab.yaml @@ -5,8 +5,8 @@ application: description: This is a Streamlit chat demo application. containers: - name: llamacpp-server - contextdir: ../playground - containerfile: Containerfile + contextdir: ../model_servers/llamacpp_python + containerfile: ./base/Containerfile model-service: true backend: - llama diff --git a/code-generation/ai-lab.yaml b/code-generation/ai-lab.yaml index b91db12c3..e18b075e8 100644 --- a/code-generation/ai-lab.yaml +++ b/code-generation/ai-lab.yaml @@ -5,8 +5,8 @@ application: description: This is a code-generation demo application. containers: - name: llamacpp-server - contextdir: ../playground - containerfile: Containerfile + contextdir: ../model_servers/llamacpp_python + containerfile: ./base/Containerfile model-service: true backend: - llama diff --git a/image_understanding/ai-lab.yaml b/image_understanding/ai-lab.yaml index f60403876..76e7ac96e 100644 --- a/image_understanding/ai-lab.yaml +++ b/image_understanding/ai-lab.yaml @@ -4,8 +4,8 @@ application: description: This demo uses a Multimodal model to analyze the content of images. containers: - name: llamacpp-server - contextdir: ../playground - containerfile: Containerfile + contextdir: ../model_servers/llamacpp_python + containerfile: ./base/Containerfile model-service: true backend: - llama diff --git a/model_servers/llamacpp/src/run.sh b/model_servers/llamacpp/src/run.sh deleted file mode 100644 index 5d3131334..000000000 --- a/model_servers/llamacpp/src/run.sh +++ /dev/null @@ -1,2 +0,0 @@ -#!/bin/bash -python -m llama_cpp.server --model ${MODEL_PATH} --host ${HOST:=0.0.0.0} --port ${PORT:=8001} --n_gpu_layers -1 diff --git a/playground/README.md b/model_servers/llamacpp_python/README.md similarity index 100% rename from playground/README.md rename to model_servers/llamacpp_python/README.md diff --git a/model_servers/llamacpp/base/Containerfile b/model_servers/llamacpp_python/base/Containerfile similarity index 86% rename from model_servers/llamacpp/base/Containerfile rename to model_servers/llamacpp_python/base/Containerfile index 362874f48..de459fb91 100644 --- a/model_servers/llamacpp/base/Containerfile +++ b/model_servers/llamacpp_python/base/Containerfile @@ -4,4 +4,4 @@ COPY src . RUN pip install --upgrade pip RUN pip install --no-cache-dir --upgrade -r /locallm/requirements.txt EXPOSE 8001 -ENTRYPOINT [ "sh", "run.sh" ] +ENTRYPOINT [ "sh", "run.sh" ] \ No newline at end of file diff --git a/model_servers/llamacpp/cuda/Containerfile b/model_servers/llamacpp_python/cuda/Containerfile similarity index 100% rename from model_servers/llamacpp/cuda/Containerfile rename to model_servers/llamacpp_python/cuda/Containerfile diff --git a/model_servers/llamacpp/cuda/README.md b/model_servers/llamacpp_python/cuda/README.md similarity index 100% rename from model_servers/llamacpp/cuda/README.md rename to model_servers/llamacpp_python/cuda/README.md diff --git a/model_servers/llamacpp/src/requirements.txt b/model_servers/llamacpp_python/src/requirements.txt similarity index 100% rename from model_servers/llamacpp/src/requirements.txt rename to model_servers/llamacpp_python/src/requirements.txt diff --git a/playground/run.sh b/model_servers/llamacpp_python/src/run.sh similarity index 100% rename from playground/run.sh rename to model_servers/llamacpp_python/src/run.sh diff --git a/playground/tooling_options.ipynb b/model_servers/llamacpp_python/tooling_options.ipynb similarity index 100% rename from playground/tooling_options.ipynb rename to model_servers/llamacpp_python/tooling_options.ipynb diff --git a/model_servers/llamacpp/vulkan/Containerfile b/model_servers/llamacpp_python/vulkan/Containerfile similarity index 100% rename from model_servers/llamacpp/vulkan/Containerfile rename to model_servers/llamacpp_python/vulkan/Containerfile diff --git a/playground/Containerfile b/playground/Containerfile deleted file mode 100644 index e47b81dd9..000000000 --- a/playground/Containerfile +++ /dev/null @@ -1,8 +0,0 @@ -FROM registry.access.redhat.com/ubi9/python-39:latest -WORKDIR /locallm -COPY requirements.txt /locallm/requirements.txt -RUN pip install --upgrade pip -RUN pip install --no-cache-dir --upgrade -r /locallm/requirements.txt -COPY run.sh run.sh -EXPOSE 8001 -ENTRYPOINT [ "sh", "run.sh" ] diff --git a/playground/requirements.txt b/playground/requirements.txt deleted file mode 100644 index bbea3dd85..000000000 --- a/playground/requirements.txt +++ /dev/null @@ -1 +0,0 @@ -llama-cpp-python[server] \ No newline at end of file diff --git a/rag/ai-lab.yaml b/rag/ai-lab.yaml index 3919bdf0c..a88ddbeb2 100644 --- a/rag/ai-lab.yaml +++ b/rag/ai-lab.yaml @@ -5,8 +5,8 @@ application: description: This is a RAG demo application. containers: - name: llamacpp-server - contextdir: ../playground - containerfile: Containerfile + contextdir: ../model_servers/llamacpp_python + containerfile: ./base/Containerfile model-service: true backend: - llama diff --git a/summarizer-langchain/ai-lab.yaml b/summarizer-langchain/ai-lab.yaml index 65c9e3b10..3062a51ff 100644 --- a/summarizer-langchain/ai-lab.yaml +++ b/summarizer-langchain/ai-lab.yaml @@ -5,8 +5,8 @@ application: description: This is a Streamlit demo application for summarizing text. containers: - name: llamacpp-server - contextdir: ../playground - containerfile: Containerfile + contextdir: ../model_servers/llamacpp_python + containerfile: ./base/Containerfile model-service: true backend: - llama