This chapter assumes familiarity with REST APIs.
Coding Agents
NeuralDrive is designed to provide high-performance local inference for coding agents and IDE integrations. By following a standard OpenAI-compatible pattern, you can connect your favorite development tools to NeuralDrive with minimal configuration.
Generic Configuration Pattern
Most coding agents use a common set of parameters for OpenAI-compatible providers:
- Provider: OpenAI Compatible (or Custom)
- Base URL:
https://neuraldrive.local:8443/v1 - API Key:
nd-xxxxxxxxxxxxxxxxxxxx - Model Name: e.g.,
llama3.1:8b,codestral, orstarcoder2
Important Note on TLS Trust
Because NeuralDrive uses a self-signed certificate, most agents will fail to connect unless you explicitly trust the NeuralDrive CA. Before configuring your agent, ensure you have followed the steps in the TLS Trust chapter to install the certificate on your local machine.
Cursor
To use NeuralDrive with Cursor:
- Open Cursor Settings > Models.
- In the OpenAI API section, toggle the switch to Enabled.
- Click Override OpenAI Base URL and enter:
https://neuraldrive.local:8443/v1 - Enter your NeuralDrive API key in the API Key field.
- Under Models, add your desired model names (e.g.,
llama3.1:8b). - Ensure other OpenAI models are disabled if you wish to force local inference.
Continue
For the Continue VS Code or JetBrains extension, edit your config.json:
{
"models": [
{
"title": "NeuralDrive",
"provider": "openai",
"baseUrl": "https://neuraldrive.local:8443/v1",
"apiKey": "nd-xxxxxxxxxxxxxxxxxxxx",
"model": "llama3.1:8b"
}
]
}
If you experience TLS verification errors, you may need to set the NODE_EXTRA_CA_CERTS environment variable to point to your neuraldrive-ca.crt file before launching your IDE.
Aider
Aider supports OpenAI-compatible endpoints through environment variables. Use the following command to start Aider with NeuralDrive:
export OPENAI_API_BASE=https://neuraldrive.local:8443/v1
export OPENAI_API_KEY=nd-xxxxxxxxxxxxxxxxxxxx
export REQUESTS_CA_BUNDLE=/path/to/neuraldrive-ca.crt
aider --model openai/llama3.1:8b
By setting REQUESTS_CA_BUNDLE, you ensure that Aider's underlying Python libraries trust the self-signed certificate.
Open Interpreter
Open Interpreter can be configured to use NeuralDrive by specifying the base URL and model:
export SSL_CERT_FILE=/path/to/neuraldrive-ca.crt
interpreter --model openai/llama3.1:8b \
--api_base https://neuraldrive.local:8443/v1 \
--api_key nd-xxxxxxxxxxxxxxxxxxxx
Using --api_base directs traffic to NeuralDrive, while SSL_CERT_FILE handles the TLS verification requirements for the Python environment.