AudioInsights is a powerful tool designed to transform audio input from various events such as meetings, lectures, speeches, etc., into formal and structured reports. The primary objective is to provide users with automated minutes for meetings, formal lecture notes, and more, utilizing cutting-edge technologies for audio transcription and document generation.
-
Audio Transcription with OpenAI's Whisper: Leverage the advanced audio transcription capabilities of OpenAI's Whisper API to convert spoken words into written text accurately.
-
Django Backend: The backend of AudioInsights is built using Django, a high-level Python web framework, ensuring robustness, security, and scalability.
-
React Frontend: The frontend is developed using React, a popular JavaScript library for building user interfaces. This provides a seamless and interactive user experience.
-
GPT API for Document Generation: Generate formal reports, meeting minutes, and lecture notes effortlessly with the GPT API, incorporating state-of-the-art natural language processing.
To run AudioInsights locally, follow these steps:
-
Clone the repository:
git clone https://github.com/Sandesh-Pyakurel/AudioInsight.git
-
Navigate to the project's backend directory:
cd AudioInsight/backend
-
Install dependencies and start server:
pip install -r requirements.txt
python3 manage.py migrate
python3 manage.py runserver
-
Navigate to the project's frontend directory:
cd ../fend/fendui
-
Install dependencies:
npm install
npm start
-
Access the application at http://localhost:3000 in your browser.
-
Upload the audio file of the event you want to transcribe and analyze.
-
The Whisper API will transcribe the audio content, converting it into written text.
-
Utilize the React-based frontend to navigate through the transcribed content.
-
Leverage the GPT API to generate formal reports, minutes, or lecture notes based on the transcribed audio.