-
Notifications
You must be signed in to change notification settings - Fork 161
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to run it in VS Code? #3
Comments
I think you need to add a comment.
|
In current version, i have disabled autocomplete for empty lines, but this was a mistake. Also keep in mind that sometimes neural network would recommend nothing. |
Could it somehow pull the base models from https://huggingface.co/deepseek-ai be useful? |
I also am not clear how this plugin is supposed to work. For me ollama is running on a remote host with 48G of ram on the video card. The ollama systemd unit is running and I can see the port is open:
The above is just to demonstrate that a closed port refuses the connection. Am I supposed to do
When I try this the model downloads and I have an interactive prompt. I set both the user and workspace settings in the plugin to the following: However, its not quite clear how I should know if Llama coder is working |
You don't need to download it manually, plugin would download it and shows the download indicator if it is not. Deep Seek model is discussed in #2 |
thats for responding. It's still not clear what to expect. It didn't seem to do anything when i traced the network calls on the port. Is there any way to debug the vscode plugin? |
You can try to open Output window of a plugin, it is named "llama code" |
So I am not exactly sure how to line up my expectations with what this plugin provides. For example, I tried typing comments, doc strings and a simple for-loop in python. I tried this with and without ollama running on the remote host and all I see in the output window for Llama Coder is
I would have expected some messages about ollama being off line. Is it possible that someone might provide a known working example of what I should expect from the plugin? Thanks for your efforts! Really appreciated |
I recommend to try a fresh version with a smallish models and you probably would see a difference. |
I wanted to just touch base again about this issue. So I couldn't get the Llama coder plugin to work remotely. So what I did was I used the VSCode ssh extension to connect to and code on the box that I am running ollama on. When I did this I do see the activity in the output tab
Ollama isn't using my GPUs for some reason but that isn't llama-coder's issue |
I installed |
I'm having the same problem with remote installs, i have a remote server with 64GB and want to use that, ollama is installed and active via systemd, i have set the bindings properly, and the firewall and ports are all correct my ollama install is on the machine "server01.local" on the default port using the api to pull the list of installed models from the machine with VSC on it gives me $ curl http://server01.local:11434/api/tags | jq I have set that host URL in the settings but nothing seems to appear the only thing I saw in the llama coder output window is: 2023-12-02 17:19:15.879 [info] Llama Coder is activated. |
@thawkins what does the llama coder output return when you write to an existing file? |
I also have problem on how to run it in VS Code. Ollama is installed and works locally, but when I click the "Llama Coder" at the right bottom corner (status bar) of VS Code it does nothing |
I also am having problems. Can't figure out how to get it to complete anything (on macOS) M2. I ran:
Restarted VSCodium (after installing the extension manually using the The Output of the plugin looks like this:
Halp? |
it looks good, just codellama returns a line break |
The question is, how do I get it to do anything else? 😅 |
I would recommend to try deepseek models, they are superior! |
Deepseek have just released a 67b parameter version, it will be a little while before it's available on ollama. |
I have changed default model to deepseek-1b one with q4 quantization that would take only ~700mb of VRAM/RAM. It works faster and provide auto-complete more often. |
I tried with deepseek but it still doesn't show any suggestions in the UI. I tried pressing the tab key, ctrl+space, neither worked. On macOS M2 using VSCodium. |
It shows automatically as a grey text, it is not tab or ctrl+space, you just type. Can you show a snipped where you are trying to do so? |
OK, it seems to be sometimes working... maybe I'm just not understanding how and when it works. For example, here it won't complete (in a mounted () {
this.pushNotificationGranted =
}, I feel like the tool could use a "manual autocomplete" mode where the user forces it to generate a completion. Maybe I'm just using it wrong. |
Same here. Been staring at VS Code for the past half hr wondering how to use this. What good is this if you don't know/can't use it? |
I assume the language doesn't matter. I have been having some problems with python (specifically gradio). Let me refine this statement. I will try changing to the deepseek its possible that its an issue with the LLM and not the plugin |
So here is a code snippet that didn't return anything for me
The output window shows this
The ollama service on the remote host has the following log generated as i typed
I have verified that the model downloaded. I am not including the output window of this because I think that's largely pointless. NVTOP shows the model is loaded into video ram
I have installed this on the remote host using the vscode remote ssh ability. The remote host is running RHEL 9. These are the plugin settings for the RHEL 9 remote host: If I attempt the same process (i.e. installing the plugin on the remote host) but the remote host is a Raspberry Pi 4 running Raspberry Pi OS I do see things with just simply typing Output window:
Ollama logs:
These are the plugin settings for the Raspberry PI remote host: What other information can I provide that would be useful? |
I do not get how I am supposed to be invoking this :( I have llama with codellama:13b-code-q4_K_M running on my local. I can use ollama locally and ask questions. The Vscode extension does not seem to be doing or invoking anything. Is there a shortcut to invoke it? It creates activity there it is connected but no results
|
VSCode downloaded the model but the extension does nothing. |
Similar issue here. Just installed. Configured remote host. Nothing changed, nothing happens. Upd: after enabling |
Trying to use it on VSCode connected to WSL, the ollama itself is running on a remote machine and I port-forward 11434 to local, but nothing works. In the extension output I see
Note that if I open a browser at |
2024-11-07 11:45:47.035 [info] Running AI completion... cant do this work |
Hi.
I installed it locally on my M1 and it works in CLI.
When i click on Llama Coder in top right corner (status bar) of VS Code it does nothing.
Sorry for question, maybe its too obvious for me.
The text was updated successfully, but these errors were encountered: