-
Notifications
You must be signed in to change notification settings - Fork 135
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Restructure the docs + simplify features
- Loading branch information
1 parent
44462ae
commit 279de3b
Showing
11 changed files
with
242 additions
and
118 deletions.
There are no files selected for viewing
This file was deleted.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file was deleted.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,67 @@ | ||
--- | ||
title: About Nitro | ||
slug: /docs | ||
--- | ||
|
||
Nitro is a fast, lightweight (3mb) inference server that can be embedded in apps to run local AI. Nitro can be used to run a variety of popular open source AI models, and provides an OpenAI-compatible API. | ||
|
||
Nitro is used to power [Jan](https://jan.ai), a open source alternative to OpenAI's platform that can be run on your own computer or server. | ||
|
||
|
||
Nitro is a fast, lightweight, and embeddable inference engine, powering [Jan](https://jan.ai/). Developed in C++, it's specially optimized for use in edge computing and is ready for deployment in products. | ||
|
||
⚡ Discover more about Nitro on [GitHub](https://github.com/janhq/nitro) | ||
|
||
## Why Nitro? | ||
|
||
### Lightweight & Fast | ||
|
||
- Old materials | ||
- At a mere 3MB, Nitro is a testament to efficiency. This stark difference in size makes Nitro an ideal choice for applications. | ||
- Nitro is designed to blend seamlessly into your application without restricting the use of other tools. This flexibility is a crucial advantage. | ||
- **Quick Setup:** | ||
Nitro can be up and running in about 10 seconds. This rapid deployment means you can focus more on development and less on installation processes. | ||
|
||
- Old material | ||
- Nitro uses the `drogon` C++17/20 HTTP application framework, which makes a significant difference. This framework is known for its speed, ensuring that Nitro processes data swiftly. This means your applications can make quick decisions based on complex data, a crucial factor in today's fast-paced digital environment. | ||
- Nitro elevates its game with drogon cpp, a C++ production-ready web framework. Its non-blocking socket IO ensures that your web services are efficient, robust, and reliable. | ||
- [Batching Inference](features/batch) | ||
- Non-blocking Socket IO | ||
|
||
### OpenAI-compatible API | ||
|
||
- [ ] OpenAI-compatible | ||
- [ ] Given examples | ||
- [ ] What is not covered? (e.g. Assistants, Tools -> See Jan) | ||
|
||
- Extends OpenAI's API with helpful model methods | ||
- e.g. Load/Unload model | ||
- e.g. Checking model status | ||
- [Unload model](features/load-unload) | ||
- With Nitro, you gain more control over `llama.cpp` features. You can now stop background slot processing and unload models as needed. This level of control optimizes resource usage and enhances application performance. | ||
|
||
### Cross-Platform | ||
|
||
- [ ] Cross-platform | ||
|
||
### Multi-modal | ||
|
||
- [ ] Hint at what's coming | ||
|
||
## Architecture | ||
|
||
- [ ] Link to Specifications | ||
|
||
## Support | ||
|
||
- [ ] File a Github Issue | ||
- [ ] Go to Discord | ||
|
||
## Contributing | ||
|
||
- [ ] Link to Github | ||
|
||
## Acknowledgements | ||
|
||
- [drogon](https://github.com/drogonframework/drogon): The fast C++ web framework supporting either C++17 or C++14 | ||
- [llama.cpp](https://github.com/ggerganov/llama.cpp): Inference of LLaMA model in pure C/C++ |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,7 @@ | ||
--- | ||
title: Architecture | ||
--- | ||
|
||
We should only have 1 document | ||
- [ ] Refactor system/architecture | ||
- [ ] Refactor system/key-concepts |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,4 @@ | ||
--- | ||
title: Install from Source | ||
slug: /install | ||
--- |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,34 @@ | ||
--- | ||
title: Quickstart | ||
--- | ||
|
||
- Objective | ||
- Quickstart shows the "power" of the system very quickly | ||
- Combine | ||
- [ ] nitro/using-nitro | ||
- [ ] nitro/installation | ||
- [ ] nitro/first-call | ||
|
||
## Getting Nitro | ||
|
||
- [ ] Overview of the different ways to install nitro | ||
- [ ] via npm | ||
- [ ] via pip | ||
- [ ] via shell script | ||
- [ ] Link to other page for "Build from Source" (tedious, not happy path) | ||
- [ ] What does installing Nitro do? (what changes in your system?) | ||
|
||
## Downloading a Model | ||
|
||
- Recommend an actual model to download | ||
|
||
## Check Nitro server | ||
|
||
```zsh title="Nitro Health Status" | ||
curl -X GET http://localhost:3928/healthz | ||
``` | ||
|
||
## Making an Inference | ||
|
||
- Make an actual inference call using Nitro | ||
- Talk about OpenAI compatibility |
Oops, something went wrong.