A comprehensive AGI pipeline integrating NLP, Computer Vision, and Speech Processing using pre-trained models.
- Text generation with T5
- Object detection with YOLO
- Speech-to-text with Whisper
- Text-to-speech with Pyttsx3
-
Clone the repository:
git clone https://github.com/yourusername/agi-pipeline.git
-
Navigate to the project directory:
cd agi-pipeline
-
Create and activate a virtual environment:
python -m venv venv source venv/bin/activate # On Windows use `venv\Scripts\activate`
-
Install dependencies:
pip install -r requirements.txt
-
Run the FastAPI application:
uvicorn main:app --reload
-
Access the API at
http://127.0.0.1:8000
.
-
Build the Docker image:
docker build -t agi-pipeline:1.0.1 .
-
Run the Docker container:
docker run -p 8000:8000 agi-pipeline:1.0.1
Feel free to open issues or submit pull requests!
This project is licensed under the MIT License - see the LICENSE file for details.