Llama4U is a privacy-focused AI assistant developed using Ollama, LangChain and Llama3. A completely free AI solution that can be hosted locally, while providing online capabilities in a responsible and user-controllable way.
APIs that have usage limitations or require keys to be registered with an online account won't be added to this project.
- Host
llama3
model from Ollama on your computer. - Clone this repository.
There are 2 usage modes:
pip install -U langchain-cli && langchain serve
- The default server is hosted on
127.0.0.1
orlocalhost
and port8000
. - Playground can be accessed at
<host_ip>:<port>/llama4u/playground
.
cd app/ && pip install -e .
llama4u
llama4u --help
for full CLI.
/search
: Perform online search using DuckDuckGo
- Perplexity AI
- ChatGPT/GPT4o
- Powerful CPU or Nvidia GPU (>=8G VRAM)
- Ubuntu 22.04
- Works on WSL2 with nvidia CUDA
Use these steps to setup nvidia cuda drivers, if GPU is not being used:
# nvidia GPU setup for Ubuntu 22.04
curl -fSsL https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/3bf863cc.pub | sudo gpg --dearmor | sudo tee /usr/share/keyrings/nvidia-drivers.gpg > /dev/null 2>&1
echo 'deb [signed-by=/usr/share/keyrings/nvidia-drivers.gpg] https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/ /' | sudo tee /etc/apt/sources.list.d/nvidia-drivers.list
sudo apt update
sudo apt install cuda-toolkit-12-4
export PATH=/usr/local/cuda-12/bin:~/.local/bin:${PATH}
export CUDACXX=$(which nvcc)
if -z $CUDACXX; then
echo "nvcc not found in PATH."
exit /b 1
fi
echo $CUDACXX && $CUDACXX --version
- Meta, for the open source Llama models
- Ollama
- LangChain community