The Prompting & Context OS

Dictate. Capture. Compress. Inject.
Send context-rich prompts to any AI — without leaving your workflow.

macOS 14+ · Apple Silicon & Intel · MIT License

Vapor — floating prompt window with context tray and screenshot shelf

Everything between you and the AI

Vapor sits between your brain and any LLM — capturing context, compressing prompts, and injecting them where you need them.

Voice Dictation

Hold Fn to speak — on-device transcription via Apple Speech. No cloud, no latency, no privacy concerns.

Voice dictation with live transcript in editor

Prompt Compression

40–60% token reduction. Strip filler, fuse concepts, preserve meaning. Choose Local LLM (free, on-device) or OpenRouter (cloud).

Prompt compression with token counts

Browser Injection

Send compressed prompts directly into ChatGPT, Claude, Gemini, Grok, Perplexity — no copy-paste needed.

Compressed prompt injected into AI chat

Screenshot Shelf

Auto-detects screenshots on your Desktop. Add them to context with one keypress. Vapor sees what you see.

Screenshot shelf with detected screenshots

Research Interrogation

Scan live browser tabs for structured data — tables, JSON, XHR feeds, articles. Capture it all into context.

Research interrogation showing discovered browser sources

Context Tray

Captured pages, articles, and research — all in one sidebar. Search, filter, and insert context directly into your prompts.

Context tray sidebar with captured pages

Three steps. Zero friction.

1

Dictate or type

Hold the Fn key to speak, or type your prompt. Vapor captures your intent — voice or keyboard, your choice.

Voice dictation in editor
2

Compress

Hit ⌘↩. Vapor strips filler words and fuses concepts into dense, token-efficient form. 40–60% reduction, meaning preserved.

Compression with stats
3

Send

Hit ⌘⇧P. Vapor injects the compressed prompt directly into your AI chat tab. Auto-submit optional.

Prompt injected into AI chat

Along the way, context flows in automatically:

Screenshots Browser captures Entity extraction Summaries & citations Context Tray

Two ways to compress. Both best-in-class.

Local LLM

Free. Private. On-device.

  • Cost: Free
  • Privacy: 100% on-device
  • Latency: <1 second
  • Models: Phi-4 Mini (2.3 GB), Qwen 3 4B (2.4 GB), Qwen 2.5 7B (4.7 GB)
  • Setup: One-click download in Settings
Local LLM settings

OpenRouter

Cloud. Powerful. Configurable.

  • Cost: ~$0.01/1M tokens
  • Privacy: Cloud-based
  • Latency: ~1–2 seconds
  • Models: GLM-5 default, any OpenRouter model
  • Setup: Paste API key in Settings
OpenRouter settings

Your AI tabs. Connected.

The Chrome extension links Vapor directly to your AI chat interfaces. Three steps to set up.

1

Load the extension — included in the DMG. Open chrome://extensions, enable Developer mode, click "Load unpacked".

2

Copy the auth token — open Vapor Settings → Browser → Authentication → Copy.

3

Paste into the extension — click the Vapor icon → Settings → Paste → Save. Connected!

Chrome extension popup

Supported AI sites:

ChatGPT Claude Gemini Grok Perplexity + any site via DOM picker

Built in the open

Vapor is MIT-licensed. No telemetry, no tracking, no lock-in. Fork it, modify it, ship it.

MIT License No Telemetry On-Device First

Get Vapor

Download DMG

Includes the app and Chrome extension. Drag to Applications, load the extension, and go.

Download Latest Release

macOS 14+ (Sonoma) · Apple Silicon & Intel

Build from Source

Clone the repo, open in Xcode, build and run.

git clone https://github.com/memetic-research-labs/vapor.git
cd vapor/Vapor
open Vapor.xcodeproj
Build Instructions