diff --git a/docs/tutorial/continue-dev.md b/docs/tutorial/continue-dev.md index b5d360f..08362e8 100644 --- a/docs/tutorial/continue-dev.md +++ b/docs/tutorial/continue-dev.md @@ -68,6 +68,7 @@ Make sure you pull the model into your ollama instance/s beforehand. "title": "Granite Code", "provider": "openai", "model": "granite-code:latest", + "useLegacyCompletionsEndpoint": false, "apiBase": "http://YOUROPENWEBUI/ollama/v1", "apiKey": "sk-YOUR-API-KEY" } @@ -83,6 +84,7 @@ Make sure you pull the model into your ollama instance/s beforehand. "title": "Granite Code", "provider": "openai", "model": "granite-code:latest", + "useLegacyCompletionsEndpoint": false, "apiBase": "http://localhost:3000/ollama/v1", "apiKey": "sk-YOUR-API-KEY" } @@ -98,4 +100,4 @@ Select it and you should now be chatting via Open WebUI (and or any [pipelines]( You can do this for as many models you would like to use, altough any model should work, you should use a model that is designed for code. -See the continue documentation for additional continue configuration, [Continue Documentation](https://docs.continue.dev/reference/Model%20Providers/openai) \ No newline at end of file +See the continue documentation for additional continue configuration, [Continue Documentation](https://docs.continue.dev/reference/Model%20Providers/openai)