-
-
Notifications
You must be signed in to change notification settings - Fork 341
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ollama starcoder2:15b inline completions model does not work #896
Comments
My first guess would be that the model just takes so long on your machine. Until you can confirm that it works in the chat but not with the completion, then this would be a reasonable assumption. |
Also, are you using the latest JupyterLab version? |
From your pictures it looks like you are expecting the completion to appear in a new line (based on the position of your cursor). Currently it works by completing the text you star typing in an existing line, you need type 2 or 3 characters and wait. What exact version of JupyterLab are you using? |
My question is if chat model with the same exact model works, not with a different model. Conversly, if you try using Gemma for inline completion does it work for you? |
however, starcoder for code suggestions is much better than gemma2. |
How do i unsubscribe from this Github.com repository thread or channel? I
want to monitor activity more passively, because I get a message anytime
anyone posts anything, and it is maxing out my Gmail account everyday.
I keep trying but apparently I have only unsubscribing to a single thread?
Thank you.
Best Regards,
Rich Lysakowski, Ph.D.
Data Scientist, AI & Analytics Engineer,
and Senior Business Systems Analyst
781-640-2048 mobile
…On Sun, Jul 14, 2024 at 6:45 PM Pedro Asevedo ***@***.***> wrote:
image.png (view on web)
<https://github.com/user-attachments/assets/99a296cd-6ae4-4395-8e4b-8c0b90507d11>
testing starcoder2 in both models (completion model) and inline completion
model, nothing happen....
:(
—
Reply to this email directly, view it on GitHub
<#896 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ACHJVL2SGNYTYFZXFNIJDGDZML5RPAVCNFSM6AAAAABK3OHIE6VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDEMRXGUYDMOJUHE>
.
You are receiving this because you are subscribed to this thread.Message
ID: ***@***.***>
|
If other models work but starcoder does not work, it is likely a problem with the model, or in this case also possibly your GPU having less memory than required to run it sufficiently fast to be useful (but I see you managed to run deepseek-coder v2:16b, so unless deepseek-coder is quantized and stardocder is not it would not align with the theory of hardware issue). @richlysakowski on the main page of the repository (https://github.com/jupyterlab/jupyter-ai) you will see "Unwatch" button, with options to only watch releases. |
When I put a part of the code, and the model returns several explanations, and the complete code, with explanations and etc... and I wanted it to complete only the missing part. For example, I insert "import pandas as", I expect the model of complete with "pd", but it repeats the import pandas as pd and puts some random explanations, you know? I just want you to complete what I'm writing, don't rewrite everything. I don't know if my doubt was very clear and if this can be configurable. |
If you use streaming mode, this should have been fixed in #879 (which will be included in 2.19 release) |
Thanks very much! This release is launch today? |
Thanks for testing, the Ollama provider is experimental so there may be issues to iron out. Things I would suspect:
You already know the answer, as it was provided in #646 (comment). Otherwise, there are no systematic tests for individual models. |
I can reproduce the prefix trimming issue with all providers in 2.19.0, whether streaming or not. |
Ah no, this was still |
I liked some results that the gemma2:9b model gave me in response, we could go more in-depth about this model, I test some models daily, the most famous ones and analyze their responses, as I test the models I will report here or elsewhere specific topic. |
@pedrogutobjj did you have a chance to test the latest release, v2.19.1, which includes #900? Is it any better? |
Hey @krassowski , morning! i tested some lines of codes this morning, here are the results. |
Is this what you would expect or not? To me it looks like syntactically valid response. Of course a bit useless, but this is down to the ability of the model you use. |
The logic is correct, I mean the overlaps, I don't know if it was clear about this, for example: I inserted "def sum_matrizes(matrix1, matrix2): ...... then comes the autocomplete part, it repeats the def sum_matrizes again as a suggestion, since I already inserted it at the beginning of the code, I don't know if I was able to be clear in my suggestion. |
I see that, but it looks like the model is at fault here. It first inserted "import numpy as" and only then started the "def sum_matrizes(matrix1, matrix2):" part again. Previously you were testing with |
Thanks! Just to help me reproduce, where was your cursor when you invoked the online completer? |
on top of the cell. |
Do you mean that your cursor was before in the first line here: def soma_matrices(matriz1, matriz2):| or in the new line: def soma_matrices(matriz1, matriz2):
| or in the new line after tab: def soma_matrices(matriz1, matriz2):
| |
Hi everyone, I had previously posted about this "error", could anyone help me? As you can see, when trying to make the inline completion model work nothing happens, only the GPU usage goes to 100% practically and nothing happens in the code line. Follow the inline completion configuration screens and my Jupyter screens.
The text was updated successfully, but these errors were encountered: