what is MCP?
MCP (Model Context Protocol) is an open standard that allows AI assistants to connect to external data sources and tools. screenpipe implements MCP to expose your screen and audio history to compatible AI applications.quick setup
the fastest way to get started:Add to your AI assistant
see the specific setup guides below for your application:
transport modes
screenpipe MCP supports two transport modes:stdio (standard input/output)
for local AI assistants running on the same machine:- Claude Desktop
- Cursor
- VS Code extensions (Cline, Continue)
- local MCP clients
HTTP (network transport)
for remote AI assistants or network access:- AI assistants on different machines
- OpenClaw running on VPS
- team deployments
- network testing
available tools
when connected via MCP, AI assistants get access to these tools:search-content
search through recorded content with filtering:| parameter | type | description |
|---|---|---|
q | string | search query (optional - omit to get recent content) |
content_type | string | all, ocr, audio, input, or accessibility (default: all) |
limit | integer | max results (default: 10) |
offset | integer | pagination offset (default: 0) |
start_time | string | ISO 8601 UTC start time (e.g., 2024-01-15T10:00:00Z) |
end_time | string | ISO 8601 UTC end time |
app_name | string | filter by app (e.g., Google Chrome, Slack) |
window_name | string | filter by window title |
min_length | integer | minimum content length |
max_length | integer | maximum content length |
include_frames | boolean | include base64 screenshots (OCR only) |
speaker_ids | string | comma-separated speaker IDs for audio filtering (e.g., 1,2,3) |
speaker_name | string | filter audio by speaker name (case-insensitive partial match) |
all- everything (OCR + audio + accessibility + input)ocr- screen text from screenshots (legacy)audio- audio transcriptions from meetings/callsinput- user actions (clicks, keystrokes, clipboard, app switches)accessibility- accessibility tree text (preferred for screen content)
export-video
create video exports from screen recordings:| parameter | type | description |
|---|---|---|
start_time | string | ISO 8601 UTC start time (required) |
end_time | string | ISO 8601 UTC end time (required) |
fps | number | frames per second (default: 1.0) |
activity-summary
get lightweight activity overview for a time range:- app usage with active minutes and frame counts
- recent accessibility texts
- audio speaker summary
list-meetings
list detected meetings with duration, app, and attendees (requires smart transcription mode).search-elements
search structured UI elements (accessibility tree nodes and OCR text blocks) - lighter than search-content for targeted lookups.frame-context
get accessibility text, parsed tree nodes, and extracted URLs for a specific frame.example queries
try these in any MCP-compatible AI assistant:testing your setup
use MCP Inspector to test your configuration:http://localhost:5173 where you can:
- test all available tools
- view tool schemas
- send test queries
- debug responses
manual configuration
building from source
for development or custom builds:troubleshooting
MCP not connecting?- ensure screenpipe is running:
curl http://localhost:3030/health - restart your AI assistant after config changes
- check logs in the AI assistant’s developer console
- verify screenpipe has data:
curl "http://localhost:3030/search?limit=1" - ensure screen recording permissions are granted
- check the time range you’re querying
- grant accessibility permissions in System Settings > Privacy & Security > Accessibility
- ensure screenpipe app is listed and enabled
- reduce limit parameter
- narrow the time range
- avoid using
include_framesunless needed (large responses)