title | description | author | ms.author | ms.date |
---|---|---|---|---|
Use local language model with Dev Proxy |
How to configure Dev Proxy to use a local language model |
waldekmastykarz |
wmastyka |
04/15/2025 |
By connecting Dev Proxy to a local language model, you can improve Dev Proxy's functionality. Select Dev Proxy plugins use the language model to improve their output, where natural language is involved. By using a local language model, you can benefit from the improved functionality without incurring extra costs.
Dev Proxy supports Ollama and LM Studio as the client for local language models. To set them up on your machine, follow the instructions in their documentation.
To configure Dev Proxy to use a local language model, use the languageModel
setting in the devproxyrc.json
file.
{
"languageModel": {
"enabled": true
}
}
You can use the following options as part of the languageModel
setting:
Option | Description | Default value |
---|---|---|
cacheResponses |
Specifies whether to cache responses from the language model. | true |
client |
The name of the client to use. Allowed values: LMStudio , Ollama |
Ollama |
enabled |
Specifies whether to use a local language model. | false |
model |
The name of the language model to use. | llama3.2 |
url |
The URL of the local language model client. | http://localhost:11434 |
By default, Dev Proxy uses the standard Ollama configuration with the llama3.2 model. It also caches responses from the language model, which means, that for the same prompt, you get an immediate response without waiting for the language model to process it.
Important
When using a local language model, be sure to start your local language model client before starting Dev Proxy.
The following Dev Proxy plugins use a local language model if available: