Skip to content

This Python script enables hands-free interaction with a local Llama2 language model. It captures voice commands from the microphone, sends them to Llama2 for natural language processing, and converts the model's textual responses into speech.

License

Notifications You must be signed in to change notification settings

Marvin-VW/python-ollama-local

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 

Repository files navigation

python-ollama-local

Overview

This project enables hands-free voice interactions through a local installation of the Llama2 language model. Users can provide voice commands, which are then processed by Llama2 for natural language understanding. The textual responses are converted into speech using the "edge-tts" tool, and the generated audio is played back sequentially.

Requirements

Ensure you have the following Python libraries and tools installed:

Usage

  1. Clone the Repository:

    git clone https://github.com/Marvin-VW/python-ollama-local/blob/main/main.py
    cd voice-interaction-project
    
  2. Install Dependencies:

    pip install -r requirements.txt
    
  3. Run the Script:

    python voice_interaction.py
    

Features

  • Captures voice commands from the microphone.
  • Utilizes local Llama2 for natural language processing.
  • Converts textual responses to speech using "edge-tts."
  • Sequential playback of generated audio files.

Contribution

Feel free to contribute to the project by opening issues or submitting pull requests.

License

This project is licensed under the MIT License.

About

This Python script enables hands-free interaction with a local Llama2 language model. It captures voice commands from the microphone, sends them to Llama2 for natural language processing, and converts the model's textual responses into speech.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages