Trouble setting up using RAG with llama: error raised by inference endpoint: HTTPConnectionPool(host=’localhost’, port=11434):
There are multiple, but similar demos on how to set up using RAG with llama on the internet. With following line (similar in each example),