Using Ollama to run Llama3.1
July 29, 2024
- download
ollama
- https://ollama.com/
App:
- run
ollama run llama3.1
, this will spin up inference
Python:
pip install ollama
import ollama
response = ollama.chat(model='llama3.1', messages=[
{
'role': 'user',
'content': 'Why is the sky blue?',
},
])
print(response['message']['content'])
Related: