Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update README #92

Merged
merged 1 commit into from
Mar 22, 2024
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
21 changes: 5 additions & 16 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,11 +1,10 @@
# Locallm

This repo contains artifacts that can be used to build and run LLM (Large Language Model) services locally on your Mac using podman.
These containerized LLM services can be used to help developers quickly prototype new LLM based applications, without the need for relying
on any other externally hosted services. Since they are already containerized, it also helps developers move from their prototype
to production quicker.
This repo contains recipes for building and running containerized AI and LLM Applications locally with podman.

## Current Locallm Services:
These containerized AI recipes can be used to help developers quickly prototype new AI and LLM based applications, without the need for relying on any other externally hosted services. Since they are already containerized, it also helps developers move quickly from prototype to production.

## Current Recipes:

* [Model Service](#model-service)
* [Chatbot](#chatbot)
Expand All @@ -25,7 +24,7 @@ A simple chatbot using the [Streamlit UI](https://docs.streamlit.io/). Learn how

### Text Summarization

An LLM app that can summarize arbitrarily long text inputs with the [streamlit UI](https://docs.streamlit.io/). Learn how to build and run thisapplication here:
An LLM app that can summarize arbitrarily long text inputs with the [Streamlit UI](https://docs.streamlit.io/). Learn how to build and run thisapplication here:
[Text Summarization](/summarizer-langchain/).

### Code generation
Expand All @@ -46,13 +45,3 @@ Learn how to build and run this model training job here: [Fine-tuning](/finetune

Images for all sample applications and models are tracked in [locallm-images.md](./locallm-images.md)

## Architecture
![](/assets/arch.jpg)

The diagram above indicates the general architecture for each of the individual applications contained in this repo.
The core code available here is the "LLM Task Service" and the "API Server", bundled together under `./playground`.
With an appropriately chosen model, [./playground/Containerfile] can build an image to run the model-service.
Model services are intended to be light-weight and run with smaller hardware footprints (hence the `locallm` name),
but they can be run on any hardware that supports containers and can be scaled up if needed.

Within each sample application folders, there is an inference implementation in the `./builds` folder with a Containerfile for building the image. These examples show how a developer might interact with the model service based on their requirements.