I've setup ollama (llama3) on my pc.
.env file has the following fields:
LLM_URL=http://127.0.0.1:11434
LLM_MODEL=ollama
The error:
For more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/404
00:17:02 - INFO - [362/365] score=0 Software Developer
00:17:02 - INFO - HTTP Request: POST http://127.0.0.1:11434/chat/completions "HTTP/1.1 404 Not Found"
00:17:02 - ERROR - LLM error scoring job 'Senior Cloud Security Engineer': Client error '404 Not Found' for url 'http://127.0.0.1:11434/chat/completions'
It seems to be defaulting to the /chat/completions endpoint for local LLMs which seems to be causing the issue. On looking at the code, it seems that the assumed response will not work either.
I've setup ollama (llama3) on my pc.
.env file has the following fields:
The error:
It seems to be defaulting to the /chat/completions endpoint for local LLMs which seems to be causing the issue. On looking at the code, it seems that the assumed response will not work either.