-
Notifications
You must be signed in to change notification settings - Fork 8.3k
Description
Bug Report
I tried to use GPT4All as a local LLM server with an OpenAI-like API for serving as a code copilot via the continue plugin for VSCode.
Unfortunately, whatever I tried, it did not work.
The server is correctly detected and all models are correctly loaded, using continue prerelease. However, when trying to send any message to gpt4all from continue, the response seems to be empty.
However, when I do my own curl query it works, so I don't know how to debug this further.
I have tried with Ollama and Koboldcpp (via the OpenAI-like API, same settings as for GPT4All - of course I changed the ports), and it worked for both flawlessly.
This seems to me to be an incompatibility in the API. Continue is expecting something that GPT4All is not providing or not in the expected format.
Steps to Reproduce
- Install GPT4All and enable the OpenAI-like API, change port to 8000, then restart.
- Install the continue extension in VSCode, switch to prerelease.
- In the Continue tab, click on the "+" at the bottom left of the panel to add a new server, then select "Other OpenAI-compatible API". Then select "Autodetect".
- Go back where there was the "+", now click on the button with text just on the left, and select a model from GPT4All (it should appear as "OpenAI - name_of_model".
- Try to chat with the model from continue (just input some text in the textbox above in the same panel". After some time to load, there will be no response, but no error either, it will just go to the next textbox as if the model correctly answered but it was just an empty response.
Expected Behavior
Continue should get non-empty responses from GPT4All.
Your Environment
- GPT4All version (if applicable): v2.7.3 (latest at the time of writing)
- Operating System: Windows 11 Pro
- Chat model used (if applicable): deepseek-coder-1.3b (but I also tried 6.7b, stablecode-3b, openhermes-2.5, etc).