Skip to content

Latest commit

 

History

History
45 lines (39 loc) · 851 Bytes

README.md

File metadata and controls

45 lines (39 loc) · 851 Bytes

LLAMA 3.1 Self deploy

trying to run Ollama locally for RAG purposes. Interact with the model using rest API (WIP)

llama3.1-local-demo.mp4

Installation

Prerequesite

Docker installed Ollama installed

Download model locally

ollama run llama3.1:8b

Run redis

docker run --name redis-container -p 6379:6379 -d redis

Run LLM-Service

cd llm-service
pip install -r requirements.txt
uvicorn llm:app --reload

Go to http://127.0.0.1:8000/docs

Run UI frontend

  1. Cd in frontend directory
    cd frontend
    
  2. Install dependencies
    npm install
    
  3. Run the server
    npm run dev
    

Go to http://localhost:3000