Skip to content

RagaviDevX/Jarvis-Model

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

1 Commit
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

Jarvis-Model ๐Ÿค–

A premium, modular AI assistant with a futuristic HUD interface, advanced automation skills, and living responsiveness.

Python PyQt6 Groq License: MIT


โœจ Overview

Jarvis-Model is a voice-controlled AI assistant that blends a stunning "Stararkโ€‘style" HUD with powerful automation capabilities. Built in Python using PyQt6 for the GUI and Groq's LLM engine for natural language understanding, Jarvis offers both voice and text modes, making it suitable for any environment.

The assistant is designed around a modular skill systemโ€”new capabilities can be dropped into the skills/ directory and are instantly available.


๐Ÿš€ Core Features

  • Futuristic HUD
    • Animated Arc Reactor core that pulses with voice activity
    • Dynamic Hexagon Panel background visualizations
    • Realโ€‘time Telemetry bars for system feedback
  • Dual Interaction Modes
    • Voice Mode: Handsโ€‘free interaction via speech recognition & TTS
    • Text Mode: Silent CLI for debugging or quiet environments
  • Modular Skill Architecture
    • Dropโ€‘in skill modules in skills/
    • Easy extension without touching core code
  • Rich Skill Set
    • Web & communication (Google search, WhatsApp, Email)
    • Vision & sensing (YOLO object detection, camera, screenshot, Gemini Live)
    • System control (volume, brightness, app management, file ops)
    • Context awareness (date/time, longโ€‘term memory)

๐Ÿ“ฆ Installation

Prerequisites

  • Python 3.10+
  • A Groq API Key (free tier available)
  • Git (to clone the repository)

Steps

  1. Clone the repository

    git clone https://github.com/RagaviDevX/Jarvis-Model.git
    cd Jarvis-Model
  2. Install dependencies

    pip install -r requirements.txt

    Note: The project relies on PyQt6 for the GUI and ultralytics for vision.

  3. Configure environment Create a .env file in the project root:

    GROQ_API_KEY=your_groq_key_here
    # Add additional keys as required by specific skills (e.g., WhatsApp, Email)
  4. Launch the assistant

    • Voice mode (with GUI)
      python main.py
    • Textโ€‘only mode
      python main.py --text

๐Ÿ“‚ Project Structure

Jarvis-Model/
โ”œโ”€โ”€ Project_JARVIS-main/   # Core source code (kept for legacy compatibility)
โ”‚   โ”œโ”€โ”€ core/              # Engine, voice processing, skill registry
โ”‚   โ”œโ”€โ”€ gui/               # PyQt6 application logic & rendering
โ”‚   โ”œโ”€โ”€ skills/            # Individual capability modules
โ”‚   โ”œโ”€โ”€ assets/            # Images, icons, arc reactor graphics
โ”‚   โ”œโ”€โ”€ requirements.txt   # Python dependencies
โ”‚   โ””โ”€โ”€ main.py            # Entry point
โ”œโ”€โ”€ .env                   # Environment variables (create from .env.template)
โ”œโ”€โ”€ README.md              # This file
โ””โ”€โ”€ .gitignore

Tip: The actual runnable code lives inside Project_JARVIS-main/. The topโ€‘level folder is kept clean for GitHub presentation.


๐Ÿ› ๏ธ Available Skills

Category Skill Description
Web & Communication web_ops Google search & website opening
whatsapp_skill Automated WhatsApp messaging via Selenium
email_ops Send, read, and manage emails
Vision & Sensing detection_skill Realโ€‘time object detection (YOLOv8)
camera_skill Capture photos & process visual input
screenshot_ops Take & analyze screenshots
gemini_live_skill Advanced multimodal interaction
System Control system_ops Volume, brightness, app management
file_ops Create, read, organize files
memory_ops Longโ€‘term memory & context awareness
datetime_ops Date & time utilities
Text text_ops Text manipulation & processing
Weather weather_ops Fetch weather information

Skills are autoโ€‘loaded from the skills/ directory. To add a new skill, create a Python file following the existing template and register it in core/registry.py.


โ–ถ๏ธ Usage Examples

Voice Interaction

  1. Run python main.py.
  2. The HUD appears with a pulsating Arc Reactor.
  3. Speak naturallyโ€”e.g., โ€œWhatโ€™s the weather today?โ€ or โ€œOpen YouTube.โ€
  4. Click the reactor to pause/resume listening.

Text Interaction

python main.py --text
> What is the capital of Japan?
> Tokyo is the capital of Japan.
>

Debugging

  • Use --text mode to see logs directly in the terminal.
  • Individual skills can be tested by importing them in a Python REPL.

๐Ÿค Contributing

We welcome contributions! Please follow these steps:

  1. Fork the repository.
  2. Create a feature branch (git checkout -b feature/amazing-skill).
  3. Commit your changes (git commit -m 'Add amazing skill').
  4. Push to the branch (git push origin feature/amazing-skill).
  5. Open a Pull Request.

Please ensure:

  • Code follows PEP 8 (we use ruff for linting).
  • New skills include a short docstring and are placed in skills/.
  • Update README.md if you add new major features.

๐Ÿ“œ License

This project is licensed under the MIT License โ€“ see the LICENSE file for details.


๐Ÿ™‹โ€โ™‚๏ธ Support & Contact

  • Author: Ragavi DevX
  • GitHub: @RagaviDevX
  • Questions? Open an issue or drop a message in the Discussions tab.

Jarvis-Model โ€“ Your intelligent companion, ready to assist with voice, vision, and automation.
Built with โค๏ธ using Python, PyQt6, and Groq.

About

Jarvis-Model ๐Ÿค– A premium, modular AI assistant with a futuristic HUD interface, advanced automation skills, and living responsiveness.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages