-
-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
docs(swagger): core more localai/openai endpoints #2904
Merged
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
✅ Deploy Preview for localai ready!
To edit notification comments on pull requests, go to your Netlify site configuration. |
This comment was marked as off-topic.
This comment was marked as off-topic.
Signed-off-by: Ettore Di Giacinto <[email protected]>
Signed-off-by: Dave <[email protected]>
dave-gray101
approved these changes
Jul 18, 2024
truecharts-admin
referenced
this pull request
in truecharts/public
Jul 24, 2024
…9.1 by renovate (#24152) This PR contains the following updates: | Package | Update | Change | |---|---|---| | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1-aio-cpu` -> `v2.19.1-aio-cpu` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1-aio-gpu-nvidia-cuda-11` -> `v2.19.1-aio-gpu-nvidia-cuda-11` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1-aio-gpu-nvidia-cuda-12` -> `v2.19.1-aio-gpu-nvidia-cuda-12` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1-cublas-cuda11-ffmpeg-core` -> `v2.19.1-cublas-cuda11-ffmpeg-core` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1-cublas-cuda11-core` -> `v2.19.1-cublas-cuda11-core` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1-cublas-cuda12-ffmpeg-core` -> `v2.19.1-cublas-cuda12-ffmpeg-core` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1-cublas-cuda12-core` -> `v2.19.1-cublas-cuda12-core` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1-ffmpeg-core` -> `v2.19.1-ffmpeg-core` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.17.1` -> `v2.19.1` | --- > [!WARNING] > Some dependencies could not be looked up. Check the Dependency Dashboard for more information. --- ### Release Notes <details> <summary>mudler/LocalAI (docker.io/localai/localai)</summary> ### [`v2.19.1`](https://togithub.com/mudler/LocalAI/releases/tag/v2.19.1) [Compare Source](https://togithub.com/mudler/LocalAI/compare/v2.19.0...v2.19.1) ![local-ai-release-219-shadow](https://togithub.com/user-attachments/assets/c5d7c930-656f-410d-aab9-455a466925fe) ##### LocalAI 2.19.1 is out! :mega: ##### TLDR; Summary spotlight - 🖧 Federated Instances via P2P: LocalAI now supports federated instances with P2P, offering both load-balanced and non-load-balanced options. - 🎛️ P2P Dashboard: A new dashboard to guide and assist in setting up P2P instances with auto-discovery using shared tokens. - 🔊 TTS Integration: Text-to-Speech (TTS) is now included in the binary releases. - 🛠️ Enhanced Installer: The installer script now supports setting up federated instances. - 📥 Model Pulling: Models can now be pulled directly via URL. - 🖼️ WebUI Enhancements: Visual improvements and cleanups to the WebUI and model lists. - 🧠 llama-cpp Backend: The llama-cpp (grpc) backend now supports embedding ( https://localai.io/features/embeddings/#llamacpp-embeddings ) - ⚙️ Tool Support: Small enhancements to tools with disabled grammars. ##### 🖧 LocalAI Federation and AI swarms <p align="center"> <img src="https://github.com/user-attachments/assets/17b39f8a-fc41-47d9-b846-b3a88307813b"/> </p> LocalAI is revolutionizing the future of distributed AI workloads by making it simpler and more accessible. No more complex setups, Docker or Kubernetes configurations – LocalAI allows you to create your own AI cluster with minimal friction. By auto-discovering and sharing work or weights of the LLM model across your existing devices, LocalAI aims to scale both horizontally and vertically with ease. ##### How it works? Starting LocalAI with `--p2p` generates a shared token for connecting multiple instances: and that's all you need to create AI clusters, eliminating the need for intricate network setups. Simply navigate to the "Swarm" section in the WebUI and follow the on-screen instructions. For fully shared instances, initiate LocalAI with `--p2p --federated` and adhere to the Swarm section's guidance. This feature, while still experimental, offers a tech preview quality experience. ##### Federated LocalAI Launch multiple LocalAI instances and cluster them together to share requests across the cluster. The "Swarm" tab in the WebUI provides one-liner instructions on connecting various LocalAI instances using a shared token. Instances will auto-discover each other, even across different networks. ![346663124-1d2324fd-8b55-4fa2-9856-721a467969c2](https://togithub.com/user-attachments/assets/19ebd44a-20ff-412c-b92f-cfb8efbe4b21) Check out a demonstration video: [Watch now](https://www.youtube.com/watch?v=pH8Bv\_\_9cnA) ##### LocalAI P2P Workers Distribute weights across nodes by starting multiple LocalAI workers, currently available only on the llama.cpp backend, with plans to expand to other backends soon. ![346663124-1d2324fd-8b55-4fa2-9856-721a467969c2](https://togithub.com/user-attachments/assets/b8cadddf-a467-49cf-a1ed-8850de95366d) Check out a demonstration video: [Watch now](https://www.youtube.com/watch?v=ePH8PGqMSpo) ##### What's Changed ##### Bug fixes :bug: - fix: make sure the GNUMake jobserver is passed to cmake for the llama.cpp build by [@​cryptk](https://togithub.com/cryptk) in [https://github.com/mudler/LocalAI/pull/2697](https://togithub.com/mudler/LocalAI/pull/2697) - Using exec when starting a backend instead of spawning a new process by [@​a17t](https://togithub.com/a17t) in [https://github.com/mudler/LocalAI/pull/2720](https://togithub.com/mudler/LocalAI/pull/2720) - fix(cuda): downgrade default version from 12.5 to 12.4 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2707](https://togithub.com/mudler/LocalAI/pull/2707) - fix: Lora loading by [@​vaaale](https://togithub.com/vaaale) in [https://github.com/mudler/LocalAI/pull/2893](https://togithub.com/mudler/LocalAI/pull/2893) - fix: short-circuit when nodes aren't detected by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2909](https://togithub.com/mudler/LocalAI/pull/2909) - fix: do not list txt files as potential models by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2910](https://togithub.com/mudler/LocalAI/pull/2910) ##### 🖧 P2P area - feat(p2p): Federation and AI swarms by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2723](https://togithub.com/mudler/LocalAI/pull/2723) - feat(p2p): allow to disable DHT and use only LAN by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2751](https://togithub.com/mudler/LocalAI/pull/2751) ##### Exciting New Features 🎉 - Allows to remove a backend from the list by [@​mauromorales](https://togithub.com/mauromorales) in [https://github.com/mudler/LocalAI/pull/2721](https://togithub.com/mudler/LocalAI/pull/2721) - ci(Makefile): adds tts in binary releases by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2695](https://togithub.com/mudler/LocalAI/pull/2695) - feat: HF `/scan` endpoint by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2566](https://togithub.com/mudler/LocalAI/pull/2566) - feat(model-list): be consistent, skip known files from listing by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2760](https://togithub.com/mudler/LocalAI/pull/2760) - feat(models): pull models from urls by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2750](https://togithub.com/mudler/LocalAI/pull/2750) - feat(webui): show also models without a config in the welcome page by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2772](https://togithub.com/mudler/LocalAI/pull/2772) - feat(install.sh): support federated install by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2752](https://togithub.com/mudler/LocalAI/pull/2752) - feat(llama.cpp): support embeddings endpoints by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2871](https://togithub.com/mudler/LocalAI/pull/2871) - feat(functions): parse broken JSON when we parse the raw results, use dynamic rules for grammar keys by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2912](https://togithub.com/mudler/LocalAI/pull/2912) - feat(federation): add load balanced option by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2915](https://togithub.com/mudler/LocalAI/pull/2915) ##### 🧠 Models - models(gallery): :arrow_up: update checksum by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2701](https://togithub.com/mudler/LocalAI/pull/2701) - models(gallery): add l3-8b-everything-cot by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2705](https://togithub.com/mudler/LocalAI/pull/2705) - models(gallery): add hercules-5.0-qwen2-7b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2708](https://togithub.com/mudler/LocalAI/pull/2708) - models(gallery): add llama3-8b-darkidol-2.2-uncensored-1048k-iq-imatrix by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2710](https://togithub.com/mudler/LocalAI/pull/2710) - models(gallery): add llama-3-llamilitary by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2711](https://togithub.com/mudler/LocalAI/pull/2711) - models(gallery): add tess-v2.5-gemma-2-27b-alpha by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2712](https://togithub.com/mudler/LocalAI/pull/2712) - models(gallery): add arcee-agent by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2713](https://togithub.com/mudler/LocalAI/pull/2713) - models(gallery): add gemma2-daybreak by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2714](https://togithub.com/mudler/LocalAI/pull/2714) - models(gallery): add L3-Stheno-Maid-Blackroot-Grand-HORROR-16B-GGUF by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2715](https://togithub.com/mudler/LocalAI/pull/2715) - models(gallery): add qwen2-7b-instruct-v0.8 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2717](https://togithub.com/mudler/LocalAI/pull/2717) - models(gallery): add internlm2\_5-7b-chat-1m by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2719](https://togithub.com/mudler/LocalAI/pull/2719) - models(gallery): add gemma-2-9b-it-sppo-iter3 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2722](https://togithub.com/mudler/LocalAI/pull/2722) - models(gallery): add llama-3\_8b_unaligned_alpha by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2727](https://togithub.com/mudler/LocalAI/pull/2727) - models(gallery): add l3-8b-lunaris-v1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2729](https://togithub.com/mudler/LocalAI/pull/2729) - models(gallery): add llama-3\_8b_unaligned_alpha_rp_soup-i1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2734](https://togithub.com/mudler/LocalAI/pull/2734) - models(gallery): add hathor_respawn-l3-8b-v0.8 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2738](https://togithub.com/mudler/LocalAI/pull/2738) - models(gallery): add llama3-8b-instruct-replete-adapted by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2739](https://togithub.com/mudler/LocalAI/pull/2739) - models(gallery): add llama-3-perky-pat-instruct-8b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2740](https://togithub.com/mudler/LocalAI/pull/2740) - models(gallery): add l3-uncen-merger-omelette-rp-v0.2-8b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2741](https://togithub.com/mudler/LocalAI/pull/2741) - models(gallery): add nymph\_8b-i1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2742](https://togithub.com/mudler/LocalAI/pull/2742) - models(gallery): add smegmma-9b-v1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2743](https://togithub.com/mudler/LocalAI/pull/2743) - models(gallery): add hathor_tahsin-l3-8b-v0.85 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2762](https://togithub.com/mudler/LocalAI/pull/2762) - models(gallery): add replete-coder-instruct-8b-merged by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2782](https://togithub.com/mudler/LocalAI/pull/2782) - models(gallery): add arliai-llama-3-8b-formax-v1.0 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2783](https://togithub.com/mudler/LocalAI/pull/2783) - models(gallery): add smegmma-deluxe-9b-v1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2784](https://togithub.com/mudler/LocalAI/pull/2784) - models(gallery): add l3-ms-astoria-8b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2785](https://togithub.com/mudler/LocalAI/pull/2785) - models(gallery): add halomaidrp-v1.33-15b-l3-i1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2786](https://togithub.com/mudler/LocalAI/pull/2786) - models(gallery): add llama-3-patronus-lynx-70b-instruct by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2788](https://togithub.com/mudler/LocalAI/pull/2788) - models(gallery): add llamax3 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2849](https://togithub.com/mudler/LocalAI/pull/2849) - models(gallery): add arliai-llama-3-8b-dolfin-v0.5 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2852](https://togithub.com/mudler/LocalAI/pull/2852) - models(gallery): add tiger-gemma-9b-v1-i1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2853](https://togithub.com/mudler/LocalAI/pull/2853) - feat: models(gallery): add deepseek-v2-lite by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2658](https://togithub.com/mudler/LocalAI/pull/2658) - models(gallery): :arrow_up: update checksum by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2860](https://togithub.com/mudler/LocalAI/pull/2860) - models(gallery): add phi-3.1-mini-4k-instruct by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2863](https://togithub.com/mudler/LocalAI/pull/2863) - models(gallery): :arrow_up: update checksum by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2887](https://togithub.com/mudler/LocalAI/pull/2887) - models(gallery): add ezo model series (llama3, gemma) by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2891](https://togithub.com/mudler/LocalAI/pull/2891) - models(gallery): add l3-8b-niitama-v1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2895](https://togithub.com/mudler/LocalAI/pull/2895) - models(gallery): add mathstral-7b-v0.1-imat by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2901](https://togithub.com/mudler/LocalAI/pull/2901) - models(gallery): add MythicalMaid/EtherealMaid 15b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2902](https://togithub.com/mudler/LocalAI/pull/2902) - models(gallery): add flammenai/Mahou-1.3d-mistral-7B by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2903](https://togithub.com/mudler/LocalAI/pull/2903) - models(gallery): add big-tiger-gemma-27b-v1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2918](https://togithub.com/mudler/LocalAI/pull/2918) - models(gallery): add phillama-3.8b-v0.1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2920](https://togithub.com/mudler/LocalAI/pull/2920) - models(gallery): add qwen2-wukong-7b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2921](https://togithub.com/mudler/LocalAI/pull/2921) - models(gallery): add einstein-v4-7b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2922](https://togithub.com/mudler/LocalAI/pull/2922) - models(gallery): add gemma-2b-translation-v0.150 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2923](https://togithub.com/mudler/LocalAI/pull/2923) - models(gallery): add emo-2b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2924](https://togithub.com/mudler/LocalAI/pull/2924) - models(gallery): add celestev1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2925](https://togithub.com/mudler/LocalAI/pull/2925) ##### 📖 Documentation and examples - :arrow_up: Update docs version mudler/LocalAI by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2699](https://togithub.com/mudler/LocalAI/pull/2699) - examples(gha): add example on how to run LocalAI in Github actions by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2716](https://togithub.com/mudler/LocalAI/pull/2716) - docs(swagger): enhance coverage of APIs by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2753](https://togithub.com/mudler/LocalAI/pull/2753) - docs(swagger): comment LocalAI gallery endpoints and rerankers by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2854](https://togithub.com/mudler/LocalAI/pull/2854) - docs: add a note on benchmarks by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2857](https://togithub.com/mudler/LocalAI/pull/2857) - docs(swagger): cover p2p endpoints by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2862](https://togithub.com/mudler/LocalAI/pull/2862) - ci: use github action by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2899](https://togithub.com/mudler/LocalAI/pull/2899) - docs: update try-it-out.md by [@​eltociear](https://togithub.com/eltociear) in [https://github.com/mudler/LocalAI/pull/2906](https://togithub.com/mudler/LocalAI/pull/2906) - docs(swagger): core more localai/openai endpoints by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2904](https://togithub.com/mudler/LocalAI/pull/2904) - docs: more swagger, update docs by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2907](https://togithub.com/mudler/LocalAI/pull/2907) - feat(swagger): update swagger by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2916](https://togithub.com/mudler/LocalAI/pull/2916) ##### 👒 Dependencies - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2700](https://togithub.com/mudler/LocalAI/pull/2700) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2704](https://togithub.com/mudler/LocalAI/pull/2704) - deps(whisper.cpp): update to latest commit by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2709](https://togithub.com/mudler/LocalAI/pull/2709) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2718](https://togithub.com/mudler/LocalAI/pull/2718) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2725](https://togithub.com/mudler/LocalAI/pull/2725) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2736](https://togithub.com/mudler/LocalAI/pull/2736) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2744](https://togithub.com/mudler/LocalAI/pull/2744) - :arrow_up: Update ggerganov/whisper.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2746](https://togithub.com/mudler/LocalAI/pull/2746) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2747](https://togithub.com/mudler/LocalAI/pull/2747) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2755](https://togithub.com/mudler/LocalAI/pull/2755) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2767](https://togithub.com/mudler/LocalAI/pull/2767) - :arrow_up: Update ggerganov/whisper.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2756](https://togithub.com/mudler/LocalAI/pull/2756) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2774](https://togithub.com/mudler/LocalAI/pull/2774) - chore(deps): Update Dependencies by [@​reneleonhardt](https://togithub.com/reneleonhardt) in [https://github.com/mudler/LocalAI/pull/2538](https://togithub.com/mudler/LocalAI/pull/2538) - chore(deps): Bump dependabot/fetch-metadata from 2.1.0 to 2.2.0 by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2791](https://togithub.com/mudler/LocalAI/pull/2791) - chore(deps): Bump llama-index from 0.9.48 to 0.10.55 in /examples/chainlit by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2795](https://togithub.com/mudler/LocalAI/pull/2795) - chore(deps): Bump openai from 1.33.0 to 1.35.13 in /examples/functions by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2793](https://togithub.com/mudler/LocalAI/pull/2793) - chore(deps): Bump nginx from 1.a.b.c to 1.27.0 in /examples/k8sgpt by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2790](https://togithub.com/mudler/LocalAI/pull/2790) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/coqui by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2798](https://togithub.com/mudler/LocalAI/pull/2798) - chore(deps): Bump inflect from 7.0.0 to 7.3.1 in /backend/python/openvoice by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2796](https://togithub.com/mudler/LocalAI/pull/2796) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/parler-tts by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2797](https://togithub.com/mudler/LocalAI/pull/2797) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/petals by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2799](https://togithub.com/mudler/LocalAI/pull/2799) - chore(deps): Bump causal-conv1d from 1.2.0.post2 to 1.4.0 in /backend/python/mamba by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2792](https://togithub.com/mudler/LocalAI/pull/2792) - chore(deps): Bump docs/themes/hugo-theme-relearn from `c25bc2a` to `1b2e139` by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2801](https://togithub.com/mudler/LocalAI/pull/2801) - chore(deps): Bump tenacity from 8.3.0 to 8.5.0 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2803](https://togithub.com/mudler/LocalAI/pull/2803) - chore(deps): Bump openai from 1.33.0 to 1.35.13 in /examples/langchain-chroma by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2794](https://togithub.com/mudler/LocalAI/pull/2794) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/bark by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2805](https://togithub.com/mudler/LocalAI/pull/2805) - chore(deps): Bump streamlit from 1.30.0 to 1.36.0 in /examples/streamlit-bot by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2804](https://togithub.com/mudler/LocalAI/pull/2804) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/diffusers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2807](https://togithub.com/mudler/LocalAI/pull/2807) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/exllama2 by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2809](https://togithub.com/mudler/LocalAI/pull/2809) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/common/template by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2802](https://togithub.com/mudler/LocalAI/pull/2802) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/autogptq by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2800](https://togithub.com/mudler/LocalAI/pull/2800) - chore(deps): Bump weaviate-client from 4.6.4 to 4.6.5 in /examples/chainlit by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2811](https://togithub.com/mudler/LocalAI/pull/2811) - chore(deps): Bump gradio from 4.36.1 to 4.37.1 in /backend/python/openvoice in the pip group by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2815](https://togithub.com/mudler/LocalAI/pull/2815) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/vall-e-x by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2812](https://togithub.com/mudler/LocalAI/pull/2812) - chore(deps): Bump certifi from 2024.6.2 to 2024.7.4 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2814](https://togithub.com/mudler/LocalAI/pull/2814) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/transformers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2817](https://togithub.com/mudler/LocalAI/pull/2817) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/sentencetransformers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2813](https://togithub.com/mudler/LocalAI/pull/2813) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/rerankers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2819](https://togithub.com/mudler/LocalAI/pull/2819) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/parler-tts by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2818](https://togithub.com/mudler/LocalAI/pull/2818) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/vllm by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2820](https://togithub.com/mudler/LocalAI/pull/2820) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/coqui by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2825](https://togithub.com/mudler/LocalAI/pull/2825) - chore(deps): Bump faster-whisper from 0.9.0 to 1.0.3 in /backend/python/openvoice by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2829](https://togithub.com/mudler/LocalAI/pull/2829) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/exllama by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2841](https://togithub.com/mudler/LocalAI/pull/2841) - chore(deps): Bump scipy from 1.13.0 to 1.14.0 in /backend/python/transformers-musicgen by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2842](https://togithub.com/mudler/LocalAI/pull/2842) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2846](https://togithub.com/mudler/LocalAI/pull/2846) - chore(deps): Bump langchain from 0.2.3 to 0.2.7 in /examples/functions by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2806](https://togithub.com/mudler/LocalAI/pull/2806) - chore(deps): Bump mamba-ssm from 1.2.0.post1 to 2.2.2 in /backend/python/mamba by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2821](https://togithub.com/mudler/LocalAI/pull/2821) - chore(deps): Bump pydantic from 2.7.3 to 2.8.2 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2832](https://togithub.com/mudler/LocalAI/pull/2832) - chore(deps): Bump langchain from 0.2.3 to 0.2.7 in /examples/langchain-chroma by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2822](https://togithub.com/mudler/LocalAI/pull/2822) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/bark by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2831](https://togithub.com/mudler/LocalAI/pull/2831) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/diffusers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2833](https://togithub.com/mudler/LocalAI/pull/2833) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/autogptq by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2816](https://togithub.com/mudler/LocalAI/pull/2816) - chore(deps): Bump gradio from 4.36.1 to 4.38.1 in /backend/python/openvoice by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2840](https://togithub.com/mudler/LocalAI/pull/2840) - chore(deps): Bump the pip group across 1 directory with 2 updates by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2848](https://togithub.com/mudler/LocalAI/pull/2848) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/transformers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2837](https://togithub.com/mudler/LocalAI/pull/2837) - chore(deps): Bump sentence-transformers from 2.5.1 to 3.0.1 in /backend/python/sentencetransformers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2826](https://togithub.com/mudler/LocalAI/pull/2826) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/vall-e-x by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2830](https://togithub.com/mudler/LocalAI/pull/2830) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/rerankers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2834](https://togithub.com/mudler/LocalAI/pull/2834) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/vllm by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2839](https://togithub.com/mudler/LocalAI/pull/2839) - chore(deps): Bump librosa from 0.9.1 to 0.10.2.post1 in /backend/python/openvoice by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2836](https://togithub.com/mudler/LocalAI/pull/2836) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/transformers-musicgen by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2843](https://togithub.com/mudler/LocalAI/pull/2843) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/mamba by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2808](https://togithub.com/mudler/LocalAI/pull/2808) - chore(deps): Bump llama-index from 0.10.43 to 0.10.55 in /examples/langchain-chroma by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2810](https://togithub.com/mudler/LocalAI/pull/2810) - chore(deps): Bump langchain from 0.2.3 to 0.2.7 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2824](https://togithub.com/mudler/LocalAI/pull/2824) - chore(deps): Bump numpy from 1.26.4 to 2.0.0 in /backend/python/openvoice by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2823](https://togithub.com/mudler/LocalAI/pull/2823) - chore(deps): Bump grpcio from 1.64.0 to 1.64.1 in /backend/python/transformers-musicgen by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2844](https://togithub.com/mudler/LocalAI/pull/2844) - build(deps): bump docker/build-push-action from 5 to 6 by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2592](https://togithub.com/mudler/LocalAI/pull/2592) - chore(deps): Bump chromadb from 0.5.0 to 0.5.4 in /examples/langchain-chroma by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2828](https://togithub.com/mudler/LocalAI/pull/2828) - chore(deps): Bump torch from 2.2.0 to 2.3.1 in /backend/python/mamba by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2835](https://togithub.com/mudler/LocalAI/pull/2835) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2851](https://togithub.com/mudler/LocalAI/pull/2851) - chore(deps): Bump setuptools from 69.5.1 to 70.3.0 in /backend/python/sentencetransformers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2838](https://togithub.com/mudler/LocalAI/pull/2838) - chore: update edgevpn dependency by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2855](https://togithub.com/mudler/LocalAI/pull/2855) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2859](https://togithub.com/mudler/LocalAI/pull/2859) - chore(deps): Bump langchain from 0.2.7 to 0.2.8 in /examples/functions by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2873](https://togithub.com/mudler/LocalAI/pull/2873) - chore(deps): Bump langchain from 0.2.7 to 0.2.8 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2874](https://togithub.com/mudler/LocalAI/pull/2874) - chore(deps): Bump numexpr from 2.10.0 to 2.10.1 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2877](https://togithub.com/mudler/LocalAI/pull/2877) - chore: :arrow_up: Update ggerganov/whisper.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2885](https://togithub.com/mudler/LocalAI/pull/2885) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2886](https://togithub.com/mudler/LocalAI/pull/2886) - chore(deps): Bump debugpy from 1.8.1 to 1.8.2 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2878](https://togithub.com/mudler/LocalAI/pull/2878) - chore(deps): Bump langchain-community from 0.2.5 to 0.2.7 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2875](https://togithub.com/mudler/LocalAI/pull/2875) - chore(deps): Bump langchain from 0.2.7 to 0.2.8 in /examples/langchain-chroma by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2872](https://togithub.com/mudler/LocalAI/pull/2872) - chore(deps): Bump openai from 1.33.0 to 1.35.13 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2876](https://togithub.com/mudler/LocalAI/pull/2876) - chore: :arrow_up: Update ggerganov/whisper.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2898](https://togithub.com/mudler/LocalAI/pull/2898) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2897](https://togithub.com/mudler/LocalAI/pull/2897) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2905](https://togithub.com/mudler/LocalAI/pull/2905) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2917](https://togithub.com/mudler/LocalAI/pull/2917) ##### Other Changes - ci: add pipelines for discord notifications by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2703](https://togithub.com/mudler/LocalAI/pull/2703) - ci(arm64): fix gRPC build by adding googletest to CMakefile by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2754](https://togithub.com/mudler/LocalAI/pull/2754) - fix: arm builds via disabling abseil tests by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2758](https://togithub.com/mudler/LocalAI/pull/2758) - ci(grpc): disable ABSEIL tests by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2759](https://togithub.com/mudler/LocalAI/pull/2759) - ci(deps): add libgmock-dev by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2761](https://togithub.com/mudler/LocalAI/pull/2761) - fix abseil test issue \[attempt 3] by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2769](https://togithub.com/mudler/LocalAI/pull/2769) - feat(swagger): update swagger by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2766](https://togithub.com/mudler/LocalAI/pull/2766) - ci: Do not test the full matrix on PRs by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2771](https://togithub.com/mudler/LocalAI/pull/2771) - Git fetch specific branch instead of full tree during build by [@​LoricOSC](https://togithub.com/LoricOSC) in [https://github.com/mudler/LocalAI/pull/2748](https://togithub.com/mudler/LocalAI/pull/2748) - fix(ci): small fixups to checksum_checker.sh by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2776](https://togithub.com/mudler/LocalAI/pull/2776) - fix(ci): fixup correct path for check_and_update.py by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2777](https://togithub.com/mudler/LocalAI/pull/2777) - fixes to `check_and_update.py` script by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2778](https://togithub.com/mudler/LocalAI/pull/2778) - Update remaining git clones to git fetch by [@​LoricOSC](https://togithub.com/LoricOSC) in [https://github.com/mudler/LocalAI/pull/2779](https://togithub.com/mudler/LocalAI/pull/2779) - feat(scripts): add scripts to help adding new models to the gallery by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2789](https://togithub.com/mudler/LocalAI/pull/2789) - build: speedup `git submodule update` with `--single-branch` by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2847](https://togithub.com/mudler/LocalAI/pull/2847) - Revert "chore(deps): Bump inflect from 7.0.0 to 7.3.1 in /backend/python/openvoice" by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2856](https://togithub.com/mudler/LocalAI/pull/2856) - Revert "chore(deps): Bump librosa from 0.9.1 to 0.10.2.post1 in /backend/python/openvoice" by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2861](https://togithub.com/mudler/LocalAI/pull/2861) - feat(swagger): update swagger by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2858](https://togithub.com/mudler/LocalAI/pull/2858) - Revert "chore(deps): Bump numpy from 1.26.4 to 2.0.0 in /backend/python/openvoice" by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2868](https://togithub.com/mudler/LocalAI/pull/2868) - feat(swagger): update swagger by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2884](https://togithub.com/mudler/LocalAI/pull/2884) - fix: update grpcio version to match version used in builds by [@​cryptk](https://togithub.com/cryptk) in [https://github.com/mudler/LocalAI/pull/2888](https://togithub.com/mudler/LocalAI/pull/2888) - fix: cleanup indentation and remove duplicate dockerfile stanza by [@​cryptk](https://togithub.com/cryptk) in [https://github.com/mudler/LocalAI/pull/2889](https://togithub.com/mudler/LocalAI/pull/2889) - ci: add workflow to comment new Opened PRs by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2892](https://togithub.com/mudler/LocalAI/pull/2892) - build: fix go.mod - don't import ourself by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2896](https://togithub.com/mudler/LocalAI/pull/2896) - feat(swagger): update swagger by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2908](https://togithub.com/mudler/LocalAI/pull/2908) - refactor: move federated server logic to its own service by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2914](https://togithub.com/mudler/LocalAI/pull/2914) - refactor: groundwork - add pkg/concurrency and the associated test file by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2745](https://togithub.com/mudler/LocalAI/pull/2745) ##### New Contributors - [@​a17t](https://togithub.com/a17t) made their first contribution in [https://github.com/mudler/LocalAI/pull/2720](https://togithub.com/mudler/LocalAI/pull/2720) - [@​LoricOSC](https://togithub.com/LoricOSC) made their first contribution in [https://github.com/mudler/LocalAI/pull/2748](https://togithub.com/mudler/LocalAI/pull/2748) - [@​vaaale](https://togithub.com/vaaale) made their first contribution in [https://github.com/mudler/LocalAI/pull/2893](https://togithub.com/mudler/LocalAI/pull/2893) **Full Changelog**: https://github.com/mudler/LocalAI/compare/v2.18.1...v2.19.0 ### [`v2.19.0`](https://togithub.com/mudler/LocalAI/releases/tag/v2.19.0) [Compare Source](https://togithub.com/mudler/LocalAI/compare/v2.18.1...v2.19.0) ![local-ai-release-219-shadow](https://togithub.com/user-attachments/assets/c5d7c930-656f-410d-aab9-455a466925fe) ##### LocalAI 2.19.0 is out! :mega: ##### TLDR; Summary spotlight - 🖧 Federated Instances via P2P: LocalAI now supports federated instances with P2P, offering both load-balanced and non-load-balanced options. - 🎛️ P2P Dashboard: A new dashboard to guide and assist in setting up P2P instances with auto-discovery using shared tokens. - 🔊 TTS Integration: Text-to-Speech (TTS) is now included in the binary releases. - 🛠️ Enhanced Installer: The installer script now supports setting up federated instances. - 📥 Model Pulling: Models can now be pulled directly via URL. - 🖼️ WebUI Enhancements: Visual improvements and cleanups to the WebUI and model lists. - 🧠 llama-cpp Backend: The llama-cpp (grpc) backend now supports embedding ( https://localai.io/features/embeddings/#llamacpp-embeddings ) - ⚙️ Tool Support: Small enhancements to tools with disabled grammars. ##### 🖧 LocalAI Federation and AI swarms <p align="center"> <img src="https://github.com/user-attachments/assets/17b39f8a-fc41-47d9-b846-b3a88307813b"/> </p> LocalAI is revolutionizing the future of distributed AI workloads by making it simpler and more accessible. No more complex setups, Docker or Kubernetes configurations – LocalAI allows you to create your own AI cluster with minimal friction. By auto-discovering and sharing work or weights of the LLM model across your existing devices, LocalAI aims to scale both horizontally and vertically with ease. ##### How it works? Starting LocalAI with `--p2p` generates a shared token for connecting multiple instances: and that's all you need to create AI clusters, eliminating the need for intricate network setups. Simply navigate to the "Swarm" section in the WebUI and follow the on-screen instructions. For fully shared instances, initiate LocalAI with `--p2p --federated` and adhere to the Swarm section's guidance. This feature, while still experimental, offers a tech preview quality experience. ##### Federated LocalAI Launch multiple LocalAI instances and cluster them together to share requests across the cluster. The "Swarm" tab in the WebUI provides one-liner instructions on connecting various LocalAI instances using a shared token. Instances will auto-discover each other, even across different networks. ![346663124-1d2324fd-8b55-4fa2-9856-721a467969c2](https://togithub.com/user-attachments/assets/19ebd44a-20ff-412c-b92f-cfb8efbe4b21) Check out a demonstration video: [Watch now](https://www.youtube.com/watch?v=pH8Bv\_\_9cnA) ##### LocalAI P2P Workers Distribute weights across nodes by starting multiple LocalAI workers, currently available only on the llama.cpp backend, with plans to expand to other backends soon. ![346663124-1d2324fd-8b55-4fa2-9856-721a467969c2](https://togithub.com/user-attachments/assets/b8cadddf-a467-49cf-a1ed-8850de95366d) Check out a demonstration video: [Watch now](https://www.youtube.com/watch?v=ePH8PGqMSpo) ##### What's Changed ##### Bug fixes :bug: - fix: make sure the GNUMake jobserver is passed to cmake for the llama.cpp build by [@​cryptk](https://togithub.com/cryptk) in [https://github.com/mudler/LocalAI/pull/2697](https://togithub.com/mudler/LocalAI/pull/2697) - Using exec when starting a backend instead of spawning a new process by [@​a17t](https://togithub.com/a17t) in [https://github.com/mudler/LocalAI/pull/2720](https://togithub.com/mudler/LocalAI/pull/2720) - fix(cuda): downgrade default version from 12.5 to 12.4 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2707](https://togithub.com/mudler/LocalAI/pull/2707) - fix: Lora loading by [@​vaaale](https://togithub.com/vaaale) in [https://github.com/mudler/LocalAI/pull/2893](https://togithub.com/mudler/LocalAI/pull/2893) - fix: short-circuit when nodes aren't detected by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2909](https://togithub.com/mudler/LocalAI/pull/2909) - fix: do not list txt files as potential models by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2910](https://togithub.com/mudler/LocalAI/pull/2910) ##### 🖧 P2P area - feat(p2p): Federation and AI swarms by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2723](https://togithub.com/mudler/LocalAI/pull/2723) - feat(p2p): allow to disable DHT and use only LAN by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2751](https://togithub.com/mudler/LocalAI/pull/2751) ##### Exciting New Features 🎉 - Allows to remove a backend from the list by [@​mauromorales](https://togithub.com/mauromorales) in [https://github.com/mudler/LocalAI/pull/2721](https://togithub.com/mudler/LocalAI/pull/2721) - ci(Makefile): adds tts in binary releases by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2695](https://togithub.com/mudler/LocalAI/pull/2695) - feat: HF `/scan` endpoint by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2566](https://togithub.com/mudler/LocalAI/pull/2566) - feat(model-list): be consistent, skip known files from listing by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2760](https://togithub.com/mudler/LocalAI/pull/2760) - feat(models): pull models from urls by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2750](https://togithub.com/mudler/LocalAI/pull/2750) - feat(webui): show also models without a config in the welcome page by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2772](https://togithub.com/mudler/LocalAI/pull/2772) - feat(install.sh): support federated install by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2752](https://togithub.com/mudler/LocalAI/pull/2752) - feat(llama.cpp): support embeddings endpoints by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2871](https://togithub.com/mudler/LocalAI/pull/2871) - feat(functions): parse broken JSON when we parse the raw results, use dynamic rules for grammar keys by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2912](https://togithub.com/mudler/LocalAI/pull/2912) - feat(federation): add load balanced option by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2915](https://togithub.com/mudler/LocalAI/pull/2915) ##### 🧠 Models - models(gallery): :arrow_up: update checksum by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2701](https://togithub.com/mudler/LocalAI/pull/2701) - models(gallery): add l3-8b-everything-cot by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2705](https://togithub.com/mudler/LocalAI/pull/2705) - models(gallery): add hercules-5.0-qwen2-7b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2708](https://togithub.com/mudler/LocalAI/pull/2708) - models(gallery): add llama3-8b-darkidol-2.2-uncensored-1048k-iq-imatrix by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2710](https://togithub.com/mudler/LocalAI/pull/2710) - models(gallery): add llama-3-llamilitary by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2711](https://togithub.com/mudler/LocalAI/pull/2711) - models(gallery): add tess-v2.5-gemma-2-27b-alpha by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2712](https://togithub.com/mudler/LocalAI/pull/2712) - models(gallery): add arcee-agent by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2713](https://togithub.com/mudler/LocalAI/pull/2713) - models(gallery): add gemma2-daybreak by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2714](https://togithub.com/mudler/LocalAI/pull/2714) - models(gallery): add L3-Stheno-Maid-Blackroot-Grand-HORROR-16B-GGUF by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2715](https://togithub.com/mudler/LocalAI/pull/2715) - models(gallery): add qwen2-7b-instruct-v0.8 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2717](https://togithub.com/mudler/LocalAI/pull/2717) - models(gallery): add internlm2\_5-7b-chat-1m by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2719](https://togithub.com/mudler/LocalAI/pull/2719) - models(gallery): add gemma-2-9b-it-sppo-iter3 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2722](https://togithub.com/mudler/LocalAI/pull/2722) - models(gallery): add llama-3\_8b_unaligned_alpha by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2727](https://togithub.com/mudler/LocalAI/pull/2727) - models(gallery): add l3-8b-lunaris-v1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2729](https://togithub.com/mudler/LocalAI/pull/2729) - models(gallery): add llama-3\_8b_unaligned_alpha_rp_soup-i1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2734](https://togithub.com/mudler/LocalAI/pull/2734) - models(gallery): add hathor_respawn-l3-8b-v0.8 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2738](https://togithub.com/mudler/LocalAI/pull/2738) - models(gallery): add llama3-8b-instruct-replete-adapted by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2739](https://togithub.com/mudler/LocalAI/pull/2739) - models(gallery): add llama-3-perky-pat-instruct-8b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2740](https://togithub.com/mudler/LocalAI/pull/2740) - models(gallery): add l3-uncen-merger-omelette-rp-v0.2-8b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2741](https://togithub.com/mudler/LocalAI/pull/2741) - models(gallery): add nymph\_8b-i1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2742](https://togithub.com/mudler/LocalAI/pull/2742) - models(gallery): add smegmma-9b-v1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2743](https://togithub.com/mudler/LocalAI/pull/2743) - models(gallery): add hathor_tahsin-l3-8b-v0.85 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2762](https://togithub.com/mudler/LocalAI/pull/2762) - models(gallery): add replete-coder-instruct-8b-merged by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2782](https://togithub.com/mudler/LocalAI/pull/2782) - models(gallery): add arliai-llama-3-8b-formax-v1.0 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2783](https://togithub.com/mudler/LocalAI/pull/2783) - models(gallery): add smegmma-deluxe-9b-v1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2784](https://togithub.com/mudler/LocalAI/pull/2784) - models(gallery): add l3-ms-astoria-8b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2785](https://togithub.com/mudler/LocalAI/pull/2785) - models(gallery): add halomaidrp-v1.33-15b-l3-i1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2786](https://togithub.com/mudler/LocalAI/pull/2786) - models(gallery): add llama-3-patronus-lynx-70b-instruct by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2788](https://togithub.com/mudler/LocalAI/pull/2788) - models(gallery): add llamax3 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2849](https://togithub.com/mudler/LocalAI/pull/2849) - models(gallery): add arliai-llama-3-8b-dolfin-v0.5 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2852](https://togithub.com/mudler/LocalAI/pull/2852) - models(gallery): add tiger-gemma-9b-v1-i1 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2853](https://togithub.com/mudler/LocalAI/pull/2853) - feat: models(gallery): add deepseek-v2-lite by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2658](https://togithub.com/mudler/LocalAI/pull/2658) - models(gallery): :arrow_up: update checksum by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2860](https://togithub.com/mudler/LocalAI/pull/2860) - models(gallery): add phi-3.1-mini-4k-instruct by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2863](https://togithub.com/mudl </details> --- ### Configuration 📅 **Schedule**: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 **Automerge**: Enabled. ♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 **Ignore**: Close this PR and you won't be reminded about these updates again. --- - [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check this box --- This PR has been generated by [Renovate Bot](https://togithub.com/renovatebot/renovate). <!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiIzNy40NDAuNiIsInVwZGF0ZWRJblZlciI6IjM3LjQ0MC42IiwidGFyZ2V0QnJhbmNoIjoibWFzdGVyIiwibGFiZWxzIjpbImF1dG9tZXJnZSIsInVwZGF0ZS9kb2NrZXIvZ2VuZXJhbC9ub24tbWFqb3IiXX0=-->
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
This PR adds swagger endpoints for part of the assistant api, and for part of the localai endpoints