In this tutorial, we will learn how to use Llama-3 locally.
- Using Llama 3 With Ollama
- Accessing the Ollama API using CURL
- Accessing the Ollama API using Python Package
- Integrating the Llama 3 in VSCode
- Developing the AI Application Locally using Langchain, Ollama, Chroma, and Langchain Hub.
Install Ollama:
curl -fsSL https://ollama.com/install.sh | sh
Downloading Llama3 and using it:
ollama run llama3
Serving the Llama3 locally:
ollama serve
Installing the required Python packages:
pip install -r requirements.txt
Using CURL:
bash curl_ollama.sh
Using Python package:
python python_api_ollama.py
Using Async and Stream:
python async_stream_ollama.py
Read the tutorial
Run the following langchain_ollama.ipynb
for building the vector store and testing the code.
Then, run:
python AI_app.py
Output:
Question: Who is Ali?
Answer: Based on the provided context, Ali is a character in the story whose life and tragic death are being pieced together through flashbacks and conversations with those who knew him.
Question: