Metadata-Version: 2.4
Name: meetcap
Version: 1.2.7
Summary: Offline meeting recorder & summarizer for macOS
Author-email: Juan <juanqui@villafam.com>
License-Expression: MIT
Keywords: macos,meeting,recorder,summarization,transcription
Classifier: Development Status :: 4 - Beta
Classifier: Environment :: Console
Classifier: Intended Audience :: End Users/Desktop
Classifier: License :: OSI Approved :: MIT License
Classifier: Operating System :: MacOS
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.10
Classifier: Programming Language :: Python :: 3.11
Classifier: Programming Language :: Python :: 3.12
Classifier: Topic :: Multimedia :: Sound/Audio :: Capture/Recording
Requires-Python: >=3.10
Requires-Dist: llama-cpp-python>=0.3.15
Requires-Dist: psutil>=6.1.0
Requires-Dist: pynput>=1.8.1
Requires-Dist: rich>=14.1.0
Requires-Dist: toml>=0.10.2
Requires-Dist: tomli>=2.2.1; python_version < '3.11'
Requires-Dist: typer>=0.16.0
Requires-Dist: urllib3>=2.5.0
Provides-Extra: dev
Requires-Dist: black>=25.1.0; extra == 'dev'
Requires-Dist: bump2version>=1.0.1; extra == 'dev'
Requires-Dist: coverage[toml]>=7.6.9; extra == 'dev'
Requires-Dist: mypy>=1.13.0; extra == 'dev'
Requires-Dist: pytest-asyncio>=0.25.0; extra == 'dev'
Requires-Dist: pytest-cov>=6.2.1; extra == 'dev'
Requires-Dist: pytest-mock>=3.14.0; extra == 'dev'
Requires-Dist: pytest-timeout>=2.4.0; extra == 'dev'
Requires-Dist: pytest>=8.4.1; extra == 'dev'
Requires-Dist: ruff>=0.12.8; extra == 'dev'
Provides-Extra: mlx-stt
Requires-Dist: mlx-whisper>=0.4.2; extra == 'mlx-stt'
Provides-Extra: stt
Requires-Dist: faster-whisper>=1.2.0; extra == 'stt'
Provides-Extra: vosk-stt
Requires-Dist: scikit-learn>=1.3.0; extra == 'vosk-stt'
Requires-Dist: soundfile>=0.13.1; extra == 'vosk-stt'
Requires-Dist: vosk>=0.3.44; extra == 'vosk-stt'
Description-Content-Type: text/markdown

# meetcap

Offline meeting recorder & summarizer for macOS

## Features

- Records both system audio and microphone simultaneously
- 100% offline operation - no network connections
- Local transcription using Whisper
- Local summarization using Qwen3-4B via llama.cpp
- Simple CLI workflow: start recording → stop with hotkey → get transcript & summary

## Installation

### Prerequisites

1. **Install ffmpeg**:
   ```bash
   brew install ffmpeg
   ```

2. **Install BlackHole** (for system audio capture):
   - Download from: https://github.com/ExistentialAudio/BlackHole
   - Install the 2ch version

   > **Why BlackHole?** macOS doesn't provide a native way to capture system audio (sounds from other apps, video calls, etc.) as an input source. BlackHole creates a virtual audio device that routes system output into an input that meetcap can record.

3. **Configure Audio Routing**:

   Open **Audio MIDI Setup** (Applications → Utilities) and create two devices:

   **Step 1: Multi-Output Device** (so you can hear AND record system audio):
   - Click "+" → Create Multi-Output Device
   - Check: **Built-in Output** (or your preferred speakers/headphones)
   - Check: **BlackHole 2ch**
   - Right-click this device → "Use This Device For Sound Output"
   - Name it "Multi-Output Device" (or similar)

   **Step 2: Aggregate Device** (combines system audio + microphone):
   - Click "+" → Create Aggregate Device
   - Check: **BlackHole 2ch** (captures system audio routed from step 1)
   - Check: **Your Microphone** (Built-in Microphone, USB mic, etc.)
   - Set your microphone as **clock source** (right-click → Use This Device as Clock Source)
   - Check **Drift Correction** for your microphone
   - Name it "Aggregate Device" (or similar)

   > **How it works:** System audio flows to Multi-Output → BlackHole → Aggregate Device, where it combines with your microphone for recording.

4. **Models** (selected and downloaded during setup):
   - **Whisper models**: large-v3 (default), large-v3-turbo, or small
   - **Vosk models** (with speaker identification): small (507MB), standard (1.8GB), or gigaspeech (3.3GB)
   - **LLM models**: Qwen3-4B-Thinking (default), Qwen3-4B-Instruct, or GPT-OSS-20B

### Install meetcap

**macOS (Apple Silicon recommended):**
```bash
pip install "meetcap[mlx-stt]"
```

**Other platforms or Intel Macs:**
```bash
pip install "meetcap[stt]"
```

**With Vosk for speaker identification:**
```bash
pip install "meetcap[vosk-stt]"
```

**First-time setup:**
```bash
# Run interactive setup wizard (downloads models, tests permissions)
meetcap setup

# Quick system verification
meetcap verify

# See all available commands
meetcap --help
```

> **Note:** The `mlx-stt` extra includes MLX Whisper for faster transcription on Apple Silicon. The `stt` extra uses faster-whisper which works on all platforms.

## Usage

### Basic Recording

```bash
# Start recording (uses default/best audio device)
meetcap record

# Press ⌘+⇧+S or Ctrl-C to stop recording
```

### Process Existing Audio Files

```bash
# Transcribe and summarize an existing audio file
meetcap summarize path/to/meeting.m4a

# Specify output directory
meetcap summarize recording.m4a --out ./results
```

### Reprocess Recordings

```bash
# Reprocess both transcript and summary with different models
meetcap reprocess 2025_Jan_15_TeamStandup

# Reprocess only the summary (keep existing transcript)
meetcap reprocess 2025_Jan_15_TeamStandup --mode summary

# Use a different STT engine for transcription
meetcap reprocess /path/to/recording --stt mlx

# Use a custom LLM model for summarization
meetcap reprocess recording_dir --llm ~/.meetcap/models/custom_model.gguf

# Skip confirmation prompt
meetcap reprocess recording_dir --yes
```

The reprocess command allows you to regenerate transcripts and summaries with different models, useful for:
- Testing different STT engines (faster-whisper vs mlx-whisper)
- Trying different LLM models for better summaries
- Fixing issues with previous processing
- Experimenting with model settings

### Commands

```bash
# First-time setup (interactive wizard)
meetcap setup

# List available audio devices
meetcap devices

# Quick system verification
meetcap verify

# Start recording a meeting
meetcap record --device "Aggregate Device" --out ~/MyRecordings

# Process existing audio file (m4a, wav, mp3, etc.)
meetcap summarize samples/meeting.m4a --out ./processed

# Reprocess a recording with different models
meetcap reprocess 2025_Jan_15_TeamStandup --mode stt
```

### Configuration

Edit `~/.meetcap/config.toml` to customize:
- Default audio device
- STT engine selection (faster-whisper, mlx-whisper, vosk)
- Model settings (defaults to auto-downloaded models)
- LLM context size (16k, 32k, 64k, or 128k tokens)
- Hotkey combinations
- Output directories

Models are automatically downloaded to `~/.meetcap/models/` on first use.

#### STT Engine Options

meetcap supports multiple speech-to-text engines:

1. **Faster-Whisper** (default on Intel Macs):
   - Best accuracy for general transcription
   - Models: large-v3, large-v3-turbo, small
   - Use: `--stt fwhisper`

2. **MLX-Whisper** (default on Apple Silicon):
   - Optimized for Apple Silicon performance
   - Models: large-v3-turbo, large-v3-mlx, small-mlx
   - Use: `--stt mlx`

3. **Vosk** (speaker identification):
   - Offline speech recognition with speaker diarization
   - Identifies different speakers in the meeting
   - Models: small (507MB), standard (1.8GB), gigaspeech (3.3GB)
   - Use: `--stt vosk`
   - Enable diarization in config:
   ```toml
   [models]
   stt_engine = "vosk"
   enable_speaker_diarization = true
   ```

Speaker identification improves summaries by attributing statements to specific speakers.

#### Context Size Settings

The LLM context size determines how much transcript text can be processed at once:
- **16k tokens**: ~30 minute meetings
- **32k tokens**: ~1 hour meetings (default)
- **64k tokens**: ~2 hour meetings
- **128k tokens**: 3+ hour meetings

You can configure this during `meetcap setup` or manually edit `~/.meetcap/config.toml`:
```toml
[llm]
n_ctx = 32768  # Context size in tokens
```

**Automatic Batching**: When a transcript exceeds the configured context size, meetcap automatically:
1. Splits the transcript into overlapping chunks that fit within the context
2. Generates summaries for each chunk independently
3. Merges all chunk summaries into a final comprehensive summary

This ensures even very long meetings (4+ hours) can be processed, though larger context sizes provide better results by maintaining more continuity.

## Permissions

Grant these permissions to your terminal app:
1. **Microphone**: System Preferences → Privacy & Security → Microphone
2. **Input Monitoring**: System Preferences → Privacy & Security → Input Monitoring

## Troubleshooting

### Audio Setup Issues

**No audio devices found:**
- Ensure BlackHole is installed and Audio MIDI Setup devices are created
- Run `meetcap devices` to list available devices
- Try restarting your terminal or system after setup

**Can't hear system audio during recording:**
- Check that Multi-Output Device is set as system output (right-click in Audio MIDI Setup)
- Ensure Built-in Output is checked in your Multi-Output Device
- Test by playing music - you should hear it through your speakers

**Recording only captures microphone or only system audio:**
- Verify Aggregate Device includes both BlackHole 2ch AND your microphone
- Check that microphone is set as clock source in Aggregate Device
- Ensure Drift Correction is enabled for your microphone
- Use `meetcap record --device "Aggregate Device"` to specify the device explicitly

**Volume control keys causing errors:**
- This is a known compatibility issue with some pynput versions - the app should continue working normally despite the error message

## Output Files

Each recording session creates a dedicated folder with all meeting artifacts organized together:

```
~/Recordings/meetcap/
├── 2025_Jan_15_TeamStandup/
│   ├── recording.wav                 # Audio recording
│   ├── recording.transcript.txt      # Plain text transcript
│   ├── recording.transcript.json     # Transcript with timestamps
│   └── recording.summary.md          # Meeting summary with AI-generated insights
├── 2025_Jan_16_ProductReview/
│   └── ... (same structure)
└── ...
```

### Folder Naming Convention

Meeting folders are automatically named using:
- **Date**: `YYYY_MMM_DD` format (e.g., `2025_Jan_15`)
- **Title**: AI-generated from meeting content in PascalCase (e.g., `TeamStandup`)

The AI analyzes your meeting transcript to generate a concise, descriptive title that captures the main topic discussed.

### Configuring Output Directory

The default output directory is `~/Recordings/meetcap/`. You can change this:

1. **During setup**: Run `meetcap setup` and specify your preferred directory
2. **In config file**: Edit `~/.meetcap/config.toml`:
   ```toml
   [paths]
   out_dir = "~/Documents/MyMeetings"  # Your custom path
   ```
3. **Via environment variable**:
   ```bash
   export MEETCAP_OUT_DIR="~/Documents/MyMeetings"
   ```

## License

MIT
