-
Notifications
You must be signed in to change notification settings - Fork 620
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Step by step instructions to install and run the Llama Stack on Linux and Mac #40
Comments
Your error message says "Conda environment 'ollama' exists". Did you run Step 4 more than once? What does "conda env list|grep ollama" show? Can you try "llama distribution install --spec local-ollama --name ollama2" assuming "ollama2" doesn't exist then use "ollama2" instead of "ollama" in Steps 5 and 6. |
I see |
Noted - thank you. I can create other conda envs successfully. Also no need for root; I just tried that route for debugging this issue. Running as non-root has the same issue |
@amkoupaei Dont have hands on an unbuntu machine to try this right now but some early debugging seems like if we update line 111 in
This might fix the issue for you. Can you give this a try and see if this fixes it for you ? |
Unfortunately, it did not work either. |
I just tried on a fresh EC2 too and it worked for me - the complete log of "llama distribution" is here. What's your log or diff look like? @amkoupaei |
here is the logs: |
Really odd. Can you run |
I simplified a bit: meta-llama/llama-stack@0d933ac Can you see if this helps? |
yes, that succeeds - giving the location of the python installation. Thank you all for your help/support. |
@hardikjshah @dltn we need to host these instructions (these are great!) somewhere in our READMEs or instructions for Ollama. What would be the right place? |
Running the command : |
hi @HabebNawatha, please try the quick start guide here to run llama stack with mac. |
I managed to make the Llama Stack server and client work with Ollama on both EC2 (with 24GB GPU) and Mac (tested on 2021 M1 and 2019 2.4GHz i9 MBP, both with 32GB memory). Steps are below:
Otherwise, download the Ollama zip for Mac here, unzip it and double click the Ollama.app to move it to the Applications folder.
to download the Llama 3.1 8B model and then run:
to confirm it works by entering some question and expecting Llama's answer.
You should see (and hit enter to accept default settings for Configuring..., except n & n for the two questions related to llama_guard_shield & prompt_guard_shield):
llama-agentic-system
folder, then:and either (on Mac)
or (on Linux)
You should see output starting with (Note: If you start the script right after Step 5, especially on a slower machine such as 2019 Mac with 2.4GHz i9, you may see "httpcore.ReadTimeout" because the Llama model is still being loaded; wait a moment and retry (a few times) should work):
and on the first Terminal that runs
llama distribution start --name ollama --port 5000
, you should see:Bonus: To see the tool calling (see here and here for more info) in action, try the
hello.py
example, which asks Llama "Which players played in the winning team of the NBA western conference semifinals of 2024, please use tools" whose answer needs a web search tool, followed by a prompt "Hello". On Mac, run (replacelocalhost
with[::]
on Linux):And you should see the output returning "BuiltinTool.brave_search" below (if you see "httpcore.ReadTimeout", retry should work):
If you delete "please use tools" in the prompt of
hello.py
, not wanting to beg, you'll likely see the output:By setting an appropriate system prompt, or switching to a bigger sized Llama 3.1 model - details coming soon - you'd see you don't have to be too polite to make Llama comfortable but yourself not.
The text was updated successfully, but these errors were encountered: