๐๏ธ TransMeet โ AI-Powered Meeting Summarizer
Turn your meeting recordings into clear, structured minutes using LLMs like Groq Whisper and Google Speech Recognition.
๐ Features
- โ
Audio Transcription โ Automatically convert
.wavor.mp3files into text - ๐ง LLM-Powered Summarization โ Generate concise and structured meeting minutes
- ๐ Groq & Google Support โ Choose between Groq Whisper models or Google Speech API
- ๐ช Automatic Chunking โ Splits large files intelligently for smoother transcription
- โ๏ธ Fully Customizable โ Pick your preferred transcription and summarization models
- ๐งพ CLI & Python API โ Use it from the terminal or integrate in your Python workflows
- ๐ Clean Output โ Saves transcripts and summaries neatly in your desired folder
๐ฆ Installation
Dependencies
sudo apt-get update && sudo apt-get install -y ffmpeg gcc && sudo apt-get clean && sudo rm -rf /var/lib/apt/lists/*
๐ Setup
Set your GROQ API Key/OPENAI API Key in your environment variables.
export GROQ_API_KEY=your_groq_api_keyTo make this permanent:
echo 'export GROQ_API_KEY=your_groq_api_key' >> ~/.bashrc
If using OPENAI, set the
OPENAI_API_KEYsimilarly. For Google Speech, no API key is needed; it uses the default model.
๐งโ๐ป How to Use
โ Option 1: Import as a Python Module
from transmeet import generate_meeting_transcript_and_minutes generate_meeting_transcript_and_minutes( meeting_audio_file="/path/to/audio.wav", output_dir="complete_path_to_output_dir/", transcription_client="groq", # or "openai" transcription_model="whisper-large-v3-turbo", # change as per your need llm_client="groq", # or "openai" llm_model="llama-3.3-70b-versatile", # change as per your need )
This will save two files in your output directory:
transcription_<timestamp>.txtmeeting_minutes_<timestamp>.md
๐ง Option 2: Use the CLI
๐น Basic Usage (Default: GROQ)
transmeet -i /path/to/audio.wav -o output/
๐ธ Advanced Usage
transmeet \ -i /path/to/audio.wav \ -o output/ \ --transcription-client groq \ --transcription-model whisper-large-v3-turbo \ --llm-client groq \ --llm-model llama-3.3-70b-versatile \
๐๏ธ Output Structure
output/
โโโ transcriptions/
โ โโโ transcription_20250510_213038.txt
โโโ meeting_minutes/
โ โโโ meeting_minutes_20250510_213041.md
๐งช Supported Formats
.wav.mp3
โ๏ธ CLI Options
| Argument | Description |
|---|---|
-i, --audio-path |
Path to the input audio file |
-o, --output-dir |
Output directory (default: output/) |
--transcription-client |
groq or google (default: groq) |
--transcription-model |
e.g., whisper-large-v3-turbo |
--llm-client |
groq or openai (default: groq) |
--llm-model |
e.g., llama-3.3-70b-versatile |
๐ค LLM Models
- Groq Whisper:
whisper-large,whisper-large-v3-turbo, etc. - Google Speech: Model defaults to their API standard
- LLMs for minutes:
llama-3,mixtral,gpt-4, etc. (Groq/OpenAI)
๐ Roadmap
- Add support for multi-language meetings
- Speaker diarization support
- Upload directly to Notion or Google Docs
- Slack/Discord bots
๐งโ๐ Author
Deepak Raj ๐จโ๐ป GitHub โข ๐ LinkedIN
๐ค Contributing
Pull requests are welcome! Found a bug or need a feature? Open an issue or submit a PR.