Voice activated Assistant
This project implements a voice assistant using OpenAI's Whisper for speech recognition and Ollama's Gemma model for natural language processing.
- Wake word detection ("Hey Assistant")
- Speech-to-text conversion using Whisper
- Natural language processing using Gemma:2b
- Text-to-speech output
- Non-blocking response output and simultaneous threaded speaking
- Python 3.9+
- Ollama (with Gemma:2b model)
- OpenAI Whisper
- SpeechRecognition
- PyAudio
- pyttsx3 (for non-macOS systems)
-
Clone this repository:
git clone https://github.com/MikeyBeez/JoeJoe.git cd JoeJoe
-
Create and activate a Conda environment:
conda create -n voice_assistant python=3.9 conda activate voice_assistant
-
Install required packages:
conda install -c conda-forge speechrecognition conda install -c conda-forge pyttsx3 pip install ollama openai-whisper pyaudio
-
Install Ollama (See Ollama.com for instructions) and pull the Gemma:2b model:
ollama pull gemma:2b
-
Activate the Conda environment:
conda activate voice_assistant
-
Run the script:
python main.py
-
Say "jarvis" to wake up the assistant, then speak your query.
- Ensure your microphone is properly set up and recognized by your system.
- The first run may take some time as it downloads the Whisper models.
- This project runs Whisper on CPU. For better performance, consider using a GPU if available.
MIT