Skip to content

Commit

Permalink
Merge pull request #154 from janhq/new-api-ref-update
Browse files Browse the repository at this point in the history
fix navigating + change api path to api-ref
  • Loading branch information
tikikun authored Nov 21, 2023
2 parents 29dff69 + 4de8456 commit 831d9ba
Show file tree
Hide file tree
Showing 12 changed files with 40 additions and 18 deletions.
File renamed without changes.
11 changes: 11 additions & 0 deletions docs/docs/examples/chatbox.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
---
title: Nitro with Chatbox
---

:::info COMING SOON
:::

<!--
## What is Chatbox?
## How to use Nitro as backend -->
4 changes: 2 additions & 2 deletions docs/docs/features/chat.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ To send a single query to your chosen LLM, follow these steps:
<div style={{ width: '50%', float: 'left', clear: 'left' }}>

```bash title="Nitro"
curl http://localhost:3928/inferences/llamacpp/chat_completion \
curl http://localhost:3928/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "",
Expand Down Expand Up @@ -53,7 +53,7 @@ For ongoing conversations or multiple queries, the dialog request feature is ide
<div style={{ width: '50%', float: 'left', clear: 'left' }}>

```bash title="Nitro"
curl http://localhost:3928/inferences/llamacpp/chat_completion \
curl http://localhost:3928/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"messages": [
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/features/embed.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ Here’s an example showing how to get the embedding result from the model:
<div style={{ width: '50%', float: 'left', clear: 'left' }}>

```bash title="Nitro" {1}
curl http://localhost:3928/inferences/llamacpp/embedding \
curl http://localhost:3928/v1/embeddings \
-H 'Content-Type: application/json' \
-d '{
"input": "Hello",
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/features/prompt.md
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ curl http://localhost:3928/inferences/llamacpp/loadmodel \
### Testing the Assistant

```bash title="Pirate Assistant"
curl http://localhost:3928/inferences/llamacpp/chat_completion \
curl http://localhost:3928/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"messages": [
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/new/about.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ For instance, compare the Nitro inference call:
<div style={{ width: '50%', float: 'left', clear: 'left' }}>

```bash title="Nitro chat completion"
curl http://localhost:3928/inferences/llamacpp/chat_completion \
curl http://localhost:3928/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-3.5-turbo",
Expand Down
4 changes: 2 additions & 2 deletions docs/docs/new/install.md
Original file line number Diff line number Diff line change
Expand Up @@ -141,7 +141,7 @@ Simple testcase with nitro, after starting the server, you can run the following
"embedding": false
}'
# Send a prompt request to nitro
curl -s --location 'http://localhost:3928/inferences/llamacpp/chat_completion' \
curl -s --location 'http://localhost:3928/v1/chat/completions' \
--header 'Content-Type: application/json' \
--data '{
"messages": [
Expand Down Expand Up @@ -172,7 +172,7 @@ Simple testcase with nitro, after starting the server, you can run the following
# Send a prompt request to nitro
set "curl_data2={\"messages\":[{\"content\":\"Hello there\",\"role\":\"assistant\"},{\"content\":\"Write a long and sad story for me\",\"role\":\"user\"}],\"stream\":true,\"model\":\"gpt-3.5-turbo\",\"max_tokens\":100,\"stop\":[\"hello\"],\"frequency_penalty\":0,\"presence_penalty\":0,\"temperature\":0.7}"
curl.exe -s -w "%%{http_code}" --location "http://localhost:3928/inferences/llamacpp/chat_completion" ^
curl.exe -s -w "%%{http_code}" --location "http://localhost:3928/v1/chat/completions" ^
--header "Content-Type: application/json" ^
--data "%curl_data2%"
```
Expand Down
20 changes: 17 additions & 3 deletions docs/docs/new/quickstart.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ Next, we need to download a model. For this example, we'll use the [Llama2 7B ch
- Create a `/model` and navigate into it:
```bash
mkdir model && cd model
wget https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGUF/resolve/main/llama-2-7b-chat.Q5_K_M.gguf?download=true
wget -O llama-2-7b-model.gguf https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGUF/resolve/main/llama-2-7b-chat.Q5_K_M.gguf?download=true
```

## Step 3: Run Nitro server
Expand All @@ -43,14 +43,28 @@ To check if the Nitro server is running:
curl http://localhost:3928/healthz
```

## Step 4: Making an Inference
## Step 4: Load model

To load the model to Nitro server, you need to run:

```bash title="Load model"
curl http://localhost:3928/inferences/llamacpp/loadmodel \
-H 'Content-Type: application/json' \
-d '{
"llama_model_path": "/model/llama-2-7b-model.gguf",
"ctx_len": 512,
"ngl": 100,
}'
```

## Step 5: Making an Inference

Finally, let's make an actual inference call using Nitro.

- In your terminal, execute:

```bash title="Nitro Inference"
curl http://localhost:3928/inferences/llamacpp/chat_completion \
curl http://localhost:3928/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"messages": [
Expand Down
2 changes: 1 addition & 1 deletion docs/docusaurus.config.js
Original file line number Diff line number Diff line change
Expand Up @@ -103,7 +103,7 @@ const config = {
{
spec: "openapi/NitroAPI.yaml", // can be local file, url, or parsed json object
// spec: "openapi/OpenAIAPI.yaml",
route: "/api/",
route: "/api-reference/",
},
],
theme: {
Expand Down
4 changes: 2 additions & 2 deletions docs/openapi/NitroAPI.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -134,7 +134,7 @@ paths:
schema:
$ref: "#/components/schemas/StatusResponse"

/inferences/llamacpp/embedding:
/v1/embeddings:
post:
operationId: createEmbedding
tags:
Expand Down Expand Up @@ -162,7 +162,7 @@ paths:
schema:
$ref: "#/components/schemas/CreateEmbeddingResponse"

/inferences/llamacpp/chat_completion:
/v1/chat/completions:
post:
operationId: createChatCompletion
tags:
Expand Down
4 changes: 2 additions & 2 deletions docs/sidebars.js
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ const sidebars = {
label: "Guides",
collapsible: false,
collapsed: false,
items: ["examples/llm"],
items: ["examples/chatbox"],
},
// {
// type: "category",
Expand All @@ -61,7 +61,7 @@ const sidebars = {
],

apiSidebar: [
"api"
"api-reference"
],

// communitySidebar: [
Expand Down
3 changes: 0 additions & 3 deletions docs/src/styles/base.scss
Original file line number Diff line number Diff line change
@@ -1,7 +1,4 @@
@layer base {
html {
@apply scroll-smooth;
}
html[data-theme="light"] {
--ifm-background-color: white;
--ifm-color-primary: #2563eb; /* New Primary Blue */
Expand Down

0 comments on commit 831d9ba

Please sign in to comment.