AutoReview — Opportunity Analysis

Desktop/web tool that detects user computer usage patterns to unlock automations. Research completed 2026-03-10.

$2B
Task mining market (2025)
$10B
Projected by 2033 (25% CAGR)
$500M+
Aggregate competitor funding
$1.3B
Scribe valuation (category leader)
The honest take

This is a real market with real demand, but the competitive landscape is intense. The window for a new entrant exists because (a) most funded players are enterprise-focused, leaving SMB/prosumer underserved, and (b) the "discover + execute" loop isn't solved well by anyone yet. But you'd be entering against well-capitalized teams with strong pedigrees.

Key Findings

Recommended Strategy

  1. Safe Bet first — Screenpipe plugin to validate demand (~$2-5K, 4 weeks)
  2. If validated — Chrome extension as real MVP
  3. If traction — Full desktop agent

Competitive Landscape

10+ competitors mapped across discovery, execution, and adjacent categories.

Direct Competitors

Autostep.ai YC X25 $1.1M

WhatOperational intelligence platform — desktop app observes employees, surfaces repetitive tasks ranked by $ impact
ApproachDesktop app, passive observation (exact capture method undisclosed)
TeamAidan Pratt (Georgia Tech, Kleiner Perkins Fellow, 8VC Fellow)
Positioning"The operational intelligence layer every company needs before deploying agents"
PricingNot public (sales-led)

Cofia.ai YC W26

WhatAI automations that write themselves — observes work, detects patterns, proactively generates automations
ApproachSystem events + anonymized network traffic (NOT screen capture)
TeamPaola Martinez (Stanford, ex-Brilliant PM) + Moses Wayne (ex-Duolingo Eng Director, $1B+ revenue)
Positioning"You never wrote a prompt. You never described your workflow. Cofia never asked."
StatusPre-launch. Demo Day: March 24, 2026

Edra.ai $30.3M

WhatAI automation platform — reverse-engineers operations, generates editable SOPs/playbooks, deploys agents
ApproachSystem-level data analysis — connects to existing tools, no screen recording
TeamEugen Alpeza (8yr Palantir, US Commercial Lead), Yannis Karamanlakis (Palantir ML). ~15 employees
Positioning"The automation platform that writes itself" — Palantir playbook applied to automation
Investors8VC (seed), likely Sequoia (Series A). Mentioned in Sequoia's "Services: The New Software"

More Funded Competitors

CompanyFundingWhatInvestors
Scribe$75M / $1.3B valAuto-documents workflows. 5M+ users, 78K paid orgs. Moving into automation discovery.StepStone, Redpoint, Tiger Global
SKAN.ai$54MComputer vision on desktops → automation-ready process maps. Legacy systems, VDI.Dell Technologies Capital, Cathay, Citi
Mimica AI$34.8MDesktop recording → process maps → trains AI agents on real workflows.Paladin Capital, Khosla Ventures
Sola AIa16z-backedScreen recording → automated workflow bots. 5x revenue growth. YC S23.a16z
SorocoUndisclosed"Work graph" platform. Fortune 500 customers, 30 countries.
FortressIQAcquiredThe OG: CV + NLP to learn tasks from desktop footage. Now part of Automation Anywhere.

Adjacent: Desktop AI Agents

CompanyFundingWhat
Simular AI$27MFull Mac/Windows PC control agent. Ex-DeepMind. Felicis + Nvidia.
Adept AI$415MFoundation model for desktop actions. Partially acquired by Amazon.
CyberdeskYC S25Self-learning computer use agent for developers.
AutoComputerYC F24Zero-setup desktop RPA via text prompts.
BasepilotYC W24AI coworkers that learn from browser demonstrations.

Big Tech + Incumbents

PlayerMove
AnthropicClaude Computer Use — 80%+ success on standard office tasks
OpenAIOperator — browser-only automation, ChatGPT Pro ($200/mo)
GoogleProject Mariner — Chrome agent + Gemini Computer Use API
UiPathAdded Task Mining product for desktop discovery
Automation AnywhereAcquired FortressIQ for CV-based process discovery
MicrosoftPower Automate + Process Mining + Copilot
Celonis$3B+ valuation, process mining leader, adding task mining

Landscape Map

DISCOVER <————————————> EXECUTE | | Enterprise Scribe ($1.3B) Mimica UiPath Automation Anywhere (funded) SKAN ($54M) Soroco Simular Adept Edra ($30.3M) | | Early-stage Autostep (YC) < AutoReview? > Sola (a16z) (scrappy) Cofia (YC) Cyberdesk (YC) Basepilot (YC) | | Big Tech Celonis Microsoft Recall Claude/GPT/Gemini Power Automate Computer Use APIs
The gap

Most funded players are enterprise-first ($100K+ ACVs, long sales cycles). The prosumer/SMB segment — "individual knowledge worker who wants to automate their own repetitive tasks" — is underserved.

Technical Approaches

Three approaches analyzed with tradeoffs. Industry converging on OS-level events + screenshot fallback.

Comparison Matrix

DimensionScreen Recording + AIBrowser ExtensionOS-Level Monitoring
ComplexityMediumLowestHighest
PlatformPer-OS captureCross-platformPer-OS (2 implementations)
Privacy riskHighestMediumLowest
Inference cost/day$0.10-$0.50$0.01-$0.10$0.01-$0.05
MVP timeline6-8 weeks4-6 weeks4-6w (Mac) / 8-12w (cross)
Data richnessHighest (visual)Medium (web only)High (structured)
CoverageEverything on screenBrowser only (40-60%)All apps, no visual
CPU overhead5-15%<1%2-5%
Storage5-10 GB/mo<100 MB/mo200-500 MB/mo
GDPRRequires DPIAMediumLower risk

Approach A: Screen Recording + AI

Capture screenshots periodically → OCR/Vision extraction → LLM analysis for workflow patterns.

Approach B: Browser Extension

Chrome Manifest V3 extension captures URL patterns, DOM interactions, click targets.

Approach C: OS-Level Monitoring

Track active window, accessibility tree, file events, input patterns via OS APIs.

Recommended: Hybrid Architecture

Layer 1 — OS-Level Events (always on)
  → App switches, window titles, file operations
  → Cheapest, lowest privacy risk, broadest coverage

Layer 2 — Accessibility Tree (on demand)
  → When repetitive pattern detected, snapshot UI elements
  → Rich structured data without screenshots

Layer 3 — Screenshot (rare fallback)
  → Only when accessibility tree unavailable
  → Encrypted, auto-purged after 24h

Layer 4 — Browser Extension (optional enrichment)
  → For web-heavy workflows, URL patterns + DOM interactions

Inference Cost Optimization

StrategyImpact
Event-driven capture (not time-based)90% data reduction vs fixed-interval
Hierarchical analysis — cheap local OCR, cloud LLM on batches10-20x cost reduction
Local models for screening (Moondream 1.6B via Ollama)$0/call for initial classification
Text-first, vision-secondSend text to LLM, not images
Batch 5-15 min of activity into single LLM call12 calls/hour vs 720
Cache known patternsNear-zero cost for repeat patterns
Optimized target

$0.03-$0.20/user/day ($0.60-$4/user/month)

Stack Recommendation

ComponentTechnology
Core daemonPython 3.11+
macOS APIspyobjc (accessibility tree, window tracking)
Accessibility treemacapptree (MacPaw)
Screen capture (fallback)mss (cross-platform)
OCRApple Vision framework (free, on-device)
Local LLMOllama + Moondream/LLaVA
Cloud LLMClaude Haiku 4.5
StorageSQLite
DashboardFlask/FastAPI + HTML or Electron

Verdict: Python MVP is the right call. Fast iteration, rich ML ecosystem, CTO can ship quickly. Rewrite performance-critical paths to Rust later if needed.

MVP vs. Future Vision

Progressive build path from 4-week prototype to the automation OS.

MVP — "Pattern Detective" (4-6 weeks, macOS only)

  1. Background daemon tracks active window (app name + title) every 2 seconds
  2. Logs app switches, time-per-window, file opens/saves
  3. Every 10 minutes, batches events → Claude Haiku: "What's repetitive?"
  4. Stores detected patterns in SQLite
  5. Menubar app: "You've done [pattern X] 12 times this week. Want to automate it?"
  6. Links to suggested automation (AppleScript, Shortcuts, Zapier) — does NOT execute

Cost per user: ~$0.03/day ($0.90/month)
Success criteria: 3+ real repetitive patterns detected per user per week

V1 — "Pattern Detective + Automation Suggest" (3-4 months)

V2 — "AutoPilot" (6-12 months)

Future Vision — "The Automation OS"

Every knowledge worker has a personal AI that learns how they work, progressively automates the boring parts, and compounds institutional knowledge. The tool becomes indispensable because it gets smarter with every day of use — unlike traditional automation that's static once built.

Moat

The longer a user runs AutoReview, the more patterns it knows, the harder it is to switch. Workflow data is the new lock-in.

Risk Assessment

Three biggest risks and three strategic options.

Top 3 Risks

1. Competitive density — HIGH

$500M+ in aggregate funding across competitors. Scribe at $1.3B with 5M users. Mimica, SKAN, Edra with deep enterprise relationships. Autostep and Cofia attacking same angle with YC backing. Breaking through requires a sharp wedge.

2. Privacy backlash — MEDIUM

Microsoft Recall faced massive backlash despite being local-only. "Watches what you do on your computer" triggers visceral reactions. Enterprise needs SOC2/HIPAA. EU needs GDPR. Getting this wrong kills trust permanently.

3. Pattern-to-automation gap — MEDIUM

Detecting patterns is easy. Generating useful automations that work reliably is hard. The jump from "you copy-paste 20x/day" to "here's a working automation" is where most tools stall. If suggestions aren't actionable, it's just a dashboard.

Menu of Options

OptionCostTimelineRisk
Quick Win: Chrome extension ~$5K 4 weeks Limited to browser, may not show full vision
Moonshot: Full desktop agent $30-50K 3-4 months Expensive, entering well-funded arena
Safe Bet: Screenpipe plugin ~$2-5K 4 weeks Dependent on Screenpipe, limited differentiation
Decision

Safe Bet first (Screenpipe plugin) → validate demand cheaply → Chrome extension if validated → Full desktop agent only if traction proves it.

Screenpipe Ecosystem

Deep dive on the platform we'd build on top of. 17K+ stars, pipe API, data model, constraints.

17.2K
GitHub stars
10K+
Discord members
6,630
Total commits
$400
Lifetime license

Architecture

Event-driven capture (not continuous recording). Triggers on: app switches, clicks, scrolling, clipboard, typing pauses, idle fallback (~5s).

LayerHow
Text extractionAccessibility tree first → OCR fallback (Apple Vision / Windows OCR / Tesseract)
Audio30s chunks → local Whisper + speaker diarization
StorageSQLite (~/.screenpipe/db.sqlite) + JPEG frames + audio chunks
Footprint~600MB RAM, 5-10% CPU, 5-10GB/month

Data Model (API at localhost:3030)

TableKey Fields
framesframe_id, file_path, timestamp
ocr_texttext, app_name, window_name, browser_url, timestamp, focused
elementsaccessibility tree data
audio_transcriptionstranscription, speaker_id, device_name, device_type
ui_eventstype (text/click/app_switch/window_focus/clipboard/scroll), app/window context
Key capability

/raw_sql endpoint — arbitrary SQL against the full database. No LLM overhead for data queries. This is our secret weapon for pattern detection.

Pipe System (Plugin Architecture)

A pipe = single pipe.md file in ~/.screenpipe/pipes/{name}/

---
schedule: every 30m    # cron, "every Xm/h", or "manual"
enabled: true
provider: anthropic
model: claude-sonnet
---

[Your prompt here — AI agent executes it, querying localhost:3030]

Constraints

Existing Pipes

Time tracking (Toggl), daily journal (Obsidian), standup reports. No automation detection pipes exist. This is our first-mover opportunity.

WebSocket API

/ws/events — real-time event stream, outside the pipe system. Enables a companion service for real-time pattern detection.

Build Opportunity

First mover

Nobody in the Screenpipe ecosystem is doing automation detection. The /raw_sql endpoint lets us run analytics without LLM overhead. Hybrid architecture: lightweight pipe for scheduled reports + standalone Python service for real-time detection via WebSocket.

Cofia vs Autostep vs AutoReview

Head-to-head comparison of the three earliest-stage players. Clear positioning differentiation.

Head-to-Head

DimensionCofiaAutostepAutoReview
StagePre-launch (Demo Day Mar 24)~10 months post-YCDiscovery
Team2 (strong duo)1 (solo)TBD
FundingYC standard ($500K)$1.1M (YC + Neo)Bootstrapped
Capture methodSystem events + network trafficDesktop appScreenpipe (existing infra)
Core pitch"Automations that write themselves""Uncover repetitive tasks""Your personal automation consultant"
Target buyerIndividual / team leadCOO / VP OpsPower user (bottom-up)
ComplianceNoneSOC 2 + HIPAAN/A (local-first)
Surveillance feelLowerHigher ("shadow them 8h/day")Zero (your own data)
MoatPattern detection qualityCompounding operational dataOpen source community + Screenpipe ecosystem

Market Segmentation

SegmentPlayerBuyer
Top-down enterprise surveillanceAutostepCOO / VP Ops
Magic zero-config agent factoryCofiaTeam lead
Bottom-up privacy-first individualAutoReviewPower user

Exploitable Gaps

  1. Open source + privacy-first — only GDPR-viable option. Local-first, auditable, no surveillance.
  2. Individual-first — user opts in to help themselves, not get monitored by their boss.
  3. Existing distribution — Screenpipe's 17K+ stars, 10K+ Discord = built-in audience.
  4. No surveillance optics — "your data, your machine, your automations."
  5. Cost structure — can undercut enterprise SaaS dramatically.
  6. No technical moat from competitors yet — both <1 year old, tiny teams.

Where They Beat Us

  1. Enterprise sales — Autostep has SOC 2/HIPAA, sells to C-suite
  2. Team-level aggregation — both do multi-user; AutoReview is individual-only
  3. Founder pedigree — Duolingo + Stanford (Cofia) and Cognition backing (Autostep)
  4. Zero-config magic — Cofia's pitch is cleaner than "install Screenpipe + install pipe"
Bottom line

Three different products for three different buyers. The biggest risk isn't competition from these two — it's that one of them raises a big round and expands into your lane. Speed matters. Cofia hasn't even had Demo Day yet. Autostep is one person. The window is open.

Screenpipe Plugin MVP

Product definition for the "Safe Bet" — validate demand on Screenpipe's infrastructure before building our own.

What It Is

A Screenpipe pipe that analyzes captured screen activity, detects repetitive workflow patterns, and suggests automations. "Your personal automation consultant that watches how you work."

The Problem

Knowledge workers repeat the same workflows dozens of times per week without realizing it — copying data between apps, switching between tools, manually compiling reports. They know automation exists but don't know what to automate or how.

How It Works

Screenpipe captures your activity (already running)
        ↓
AutoReview analyzes patterns (our pipe)
        ↓
"You did [X] 14 times this week → here's how to automate it"
        ↓
One-click setup: AppleScript / Shortcut / Zapier / n8n recipe

What the User Sees

🔄 AutoReview found 3 new patterns this week:

1. Salesforce → Google Sheets copy-paste (14x/week, ~45 min)
   → Suggested: Zapier integration [Set up →]

2. Jira ticket → Slack message (23x/week, ~30 min)
   → Suggested: Jira-Slack automation [Set up →]

3. Daily report compilation from 3 tabs (5x/week, ~1 hour)
   → Suggested: n8n workflow [Set up →]

Estimated time savings: 2.25 hours/week

Target User

Screenpipe power user — already privacy-conscious, already bought into local-first, already tech-savvy enough to install a pipe. Not mass-market consumer. A power tool for the 17K+ Screenpipe community.

Persona: "Alex, senior PM at a SaaS company. Uses Screenpipe to search past meetings. Doesn't realize they spend 6 hours/week on copy-paste workflows between Jira, Slack, and Google Sheets."

Success Criteria

Ship → measure within 30 days:

  1. Detection accuracy: 3+ real repetitive patterns per user per week
  2. Suggestion quality: Actionable (user can actually set them up)
  3. Retention: Users keep the pipe enabled after 2 weeks
  4. Signal: Screenpipe Discord users talk about it / request features

Pricing

Start free/open source to validate demand and build distribution. Monetize in v2 when automation execution is added. Options later: freemium ($5/mo for suggestions), or one-time $29-49 (matches Screenpipe's model).

MVP Architecture

Two-component design optimized for Screenpipe's constraints.

Why Two Components

Component 1: Pattern Detector (Python service)

Connects toScreenpipe's /raw_sql endpoint
FrequencyRuns SQL analytics every 15 minutes
Tables usedui_events, ocr_text, elements
DetectsApp-switch loops, copy-paste cycles, repetitive navigation, form fills, time-of-day rituals
StoresOwn SQLite DB with frequency counts, time estimates, confidence scores
Cost$0 — pure SQL analytics, no LLM calls

Component 2: Automation Advisor (Screenpipe pipe)

ScheduleEvery 4 hours (or manual trigger)
InputReads from Pattern Detector's DB (pre-aggregated)
LLMClaude Haiku — "Given these patterns, suggest specific automations"
OutputRanked automatable workflows with setup instructions
Cost~$0.01-$0.03 per run = $0.03-$0.09/day

Pattern Detection Logic (v1 — rule-based, no ML)

Pattern TypeDetection MethodSQL Approach
App-switch loopsSame 2-3 app sequence in 5 minGroup ui_events by session, find recurring sequences
Copy-paste bridgesClipboard between different appsJoin clipboard events with app context, count cross-app pairs
Repetitive navigationSame URL/window dailyGroup by window_name + browser_url, find daily recurrence
Form fillsSame fields populated repeatedlyDetect repeated text input in same app/window
Time-of-day ritualsSame app sequence at similar timesCluster by hour-of-day, find stable patterns

Automation Output Types

ComplexityOutputExample
EasyLink to existing Zapier/Make template"Connect Salesforce → Google Sheets"
MediumGenerated AppleScript / macOS Shortcut"When you open Jira, auto-post to Slack"
Hardn8n/Zapier workflow JSON + guideMulti-step with API connections

v1 does NOT execute automations. It suggests and links. Execution is v2.

Stack

ComponentTechnology
Pattern DetectorPython 3.11+, SQLite, requests
Automation AdvisorScreenpipe pipe (pipe.md)
LLMClaude Haiku 4.5 via Anthropic API
Notificationspync (macOS) or osascript
DistributionGitHub repo + Screenpipe pipe install URL

Risks & Mitigations

RiskMitigation
Screenpipe API changes break usPin to specific API version, monitor releases
Pattern detection too noisyConservative: min 5 repetitions/week before surfacing
Suggestions not actionableCurated template library, not generated from scratch
Screenpipe builds this themselvesShip fast, build community, become the de facto pipe
Low install base for our nicheAlso distribute standalone (reads DB directly)

Scope & Timeline

4 weeks to beta. ~$2-5K total cost.

In Scope (MVP)

Out of Scope (v2+)

Timeline

WeekMilestone
1Pattern Detector: SQL queries for 5 pattern types, running against Screenpipe DB
2Automation Advisor pipe: Claude Haiku integration, weekly report generation
3Notifications, polish, README, install instructions
4Beta release to Screenpipe Discord, collect feedback
Total investment

4 weeks to beta. ~$2-5K cost (developer time). $0.03-$0.09/day inference per user.

Build Path

PhaseWhatCostTimeline
NowScreenpipe plugin (validate demand)$2-5K4 weeks
If validatedChrome extension (real MVP)~$5K4-6 weeks
If tractionFull desktop agent$30-50K3-4 months

Compliance & Legal Risk

9 jurisdictions analyzed. Three architectural decisions neutralize most risk across all markets.

Bottom line

Continuous screen recording for employee monitoring is effectively illegal in the EU and high-risk in California, Illinois, and Brazil. Metadata-only + local-first is the legally defensible architecture.

Risk Summary by Jurisdiction

JurisdictionRiskBiggest ThreatKey Mitigation
US FederalMediumECPA interception claimsNotice + consent + exclude personal devices
CaliforniaHighCCPA employee data + PC 632 audio consentFull privacy notice, no audio, data minimization
IllinoisHighBIPA private lawsuits ($5K/violation)Face detection/blurring, no biometric processing
New YorkMedium§ 52-c notice requirementTemplate compliance notice
MexicoMediumNew 2025 data protection law + right to disconnectLocal storage, telework agreement, working hours limits
ChileMediumNew GDPR-style law effective Dec 2026Proportionality, written consent, DPIA
BrazilHighLGPD fines (BRL 50M) + labor court moral harassmentAggregated metrics over screen capture, DPIA
ColombiaMediumConstitutional privacy challenge + SIC finesConsent, proportionality, purpose limitation
EU (GDPR)HighEUR 20M / 4% turnover fines, works council requirementsMetadata-only, local-first, DPIA, no continuous recording

Three Decisions That Neutralize Most Risk

Local-first
Eliminates data transfer obligations in every jurisdiction
No audio
California PC 632 makes audio capture impossible at scale
Face blur
Kills Illinois BIPA exposure entirely

Competitor Compliance Comparison

FeatureScribeSKANScreenpipeAutostep
SOC 2Type IIYesEnterprise onlyYes
HIPAAUnverifiedNoEnterprise onlyYes
Local storageNo (cloud)HybridYes (default)Unknown
Continuous recordingNo (user-initiated)Yes (process flows)YesUnknown
PII redactionEnterprise tierReal-time MLNone built-inUnknown
Key insight

Scribe avoids the compliance problem entirely (user-initiated, no continuous recording). SKAN uses real-time PII redaction. Screenpipe goes local-only. SOC 2 Type II is table stakes for enterprise sales in any market.

United States Compliance

Federal framework + California, Illinois, and New York deep dives.

Federal Framework

ECPA (18 U.S.C. §§ 2510-2522)

CFAA (18 U.S.C. § 1030)

FTC Act § 5


California High Risk

CCPA/CPRA

California PC § 632 (Two-Party Consent)

Critical risk

California is an all-party consent state for recording confidential communications. If the tool records Zoom/Teams audio, every participant must consent. Practically impossible at scale. Penalties: $2,500-$10,000 + up to 1 year jail. Civil: $5,000/violation.


Illinois High Risk

BIPA (740 ILCS 14)

Mitigation

Face detection + blurring kills BIPA exposure entirely. Metadata-only avoids it completely.


New York Medium Risk

NY Civil Rights Law § 52-c

SHIELD Act

Latin America Compliance

Mexico, Chile, Brazil, Colombia — all modernizing data protection toward GDPR-level standards.

Mexico Medium Risk

Key LawLFPDPPP (new law enacted March 2025 replacing 2010 version)
ConsentMust be freely given, specific, informed, documented. Aviso de Privacidad required before collection.
Employer monitoringLegal under telework amendments but must be in written telework agreement. Right to disconnect — monitoring outside working hours prohibited.
ARCO RightsAccess, Rectification, Cancellation, Opposition (similar to GDPR)
PenaltiesUp to ~$320K USD (may increase under new law)
NOM-037Telework safety standard requires employers to establish supervision mechanisms + right to disconnect

Chile Medium (Rising)

Current LawLey 19.628 (1999, outdated, weak enforcement)
New LawLey 21.719 — approved Aug 2024, effective December 1, 2026. GDPR-style overhaul with new National Data Protection Agency.
ConsentExpress written consent required. Cannot access private email. Automated tracking challengeable via Fundamental Rights Tutelage Procedure.
ConstitutionArt. 19(4) — guarantees privacy. Disproportionate monitoring = unconstitutional.
Penalties (new law)Minor: ~$32K USD. Serious: ~$320K. Very serious: ~$640K or 2-4% revenue. [Estimated]

Brazil High Risk

Key LawLGPD (Lei 13.709/2018) — GDPR-aligned
Legal basisLegitimate interest or consent. Consent problematic in employment (power imbalance).
Screen recordingContinuous capture "increasingly viewed as excessive." Many employers switching to aggregated metrics.
DPIARequired for high-risk processing (screen recording likely qualifies)
Penalties2% of revenue in Brazil, up to BRL 50M (~$10M USD) per violation + labor court indemnification for privacy damages
CLT riskCourts have ruled excessive monitoring = moral harassment (assédio moral)

Colombia Medium Risk

Key LawLey 1581 de 2012 (Habeas Data) + Constitutional Art. 15
ConsentExplicit consent required. Must inform of specific purposes.
Employer monitoringConstitutional Court allows monitoring of company devices under subordination power, subject to proportionality.
Cannot monitorPersonal communications, personal devices, activity outside work scope
PenaltiesUp to 2,000x minimum monthly wage (~$520K USD). Enforced by SIC.

European Union (GDPR)

The strictest regime. Continuous screen recording effectively illegal for employee monitoring.

Screen Recording Legality

TypeStatusEvidence
Continuous recordingEffectively illegalNo DPA has endorsed it. CNIL, EDPB, German BAG all against.
Periodic screenshotsRiskyNorwegian DPA fined for this (2021). Spanish DPA: only with narrow purpose.
Metadata-onlyMost defensibleSatisfies Art. 5(1)(c) data minimization. Multiple DPAs endorse.
Keystroke loggingUniversally prohibitedCNIL, German DPAs, Italian Garante all prohibit.

Key GDPR Articles

ArticleRequirementImpact on Us
Art. 6(1)(f)Legitimate interest + balancing testContinuous recording fails proportionality in most DPA interpretations
Art. 9Special category data protectionsScreen content inevitably captures health, political, religious data
Art. 25Privacy by design & defaultDefault must be metadata-only, local, shortest retention, filtering ON
Art. 35DPIA mandatoryThis tool triggers 5 of EDPB's criteria. DPIA required.
Art. 22Automated decision-makingIf pattern detection affects employment decisions, human review required
ePrivacy Art. 5Confidentiality of communicationsCapturing emails/messages on screen = interception. Exclude comms apps by default.

Country-Specific Strictness

CountryKey Extra RequirementStrictness
GermanyWorks council consent mandatory (BetrVG §87). 16 state DPAs. Covert monitoring only with criminal suspicion.Strictest
FranceCSE consultation required. CNIL: permanent screen capture = disproportionate.Very strict
ItalyTrade union agreement or Labour Inspectorate authorization. Garante prohibited screenshot software.Very strict
SpainEmployee digital privacy rights (LOPDGDD 87-91). More permissive with notice + proportionality.Moderate
NetherlandsWorks council consent (WOR Art. 27). Dutch DPA active on remote monitoring.Strict

Real GDPR Fines for Employee Monitoring

CaseFineWhat Happened
H&M (Hamburg DPA, 2020)EUR 35.3MExtensive recording of employee personal info (health, family, religion)
notebooksbilliger.de (2021)EUR 10.4MDisproportionate video surveillance of employees for 2+ years
Greek company (HDPA, 2022)EUR 150KEmail monitoring without legal basis or transparency
Italian companies (Garante, 2021-23)InjunctionsOrdered to cease screenshot capture and GPS tracking
Penalties

Art. 83(5): Up to EUR 20M or 4% global turnover for violating basic processing principles. H&M's EUR 35.3M fine should calibrate risk appetite.

Recommended Privacy Architecture

Designed for compliance across all 9 jurisdictions. Local-first, metadata-default, consent-aware.

Data Capture Tiers

TierWhatRiskDefault
Tier 1: MetadataApp names, window titles, timestamps, durationsLowON
Tier 2: Accessibility treeUI element data without visual contentMediumOFF (opt-in)
Tier 3: Screen recordingVisual capture of screen contentHighOFF (opt-in per session)

Mandatory Defaults

Consent Flows by Market

MarketRequirement
US FederalWritten notice + acknowledgment
CaliforniaCCPA privacy notice + right to opt-out of sensitive PI
IllinoisWritten BIPA consent if any biometric processing
New York§ 52-c written notice + acknowledgment + posted notice
MexicoAviso de Privacidad + telework agreement + right to disconnect
ChileWritten informed consent + proportionality
BrazilLGPD notice + legitimate interest assessment
ColombiaHabeas data notice + explicit consent + purpose specification
EU/GDPRDPIA + employee notification + works council (DE/FR/IT/NL) + Art. 25 by design

Retention Policies

Data TypeDefaultMaximum
Raw screenshots24 hours7 days
Metadata events7 days30 days
Extracted patterns30 days90 days
Anonymized aggregates1 yearIndefinite

Cloud LLM Integration (Cross-Border Transfer)

Enterprise vs Personal Use

FeaturePersonalEnterprise
Legal basisHousehold exemption / consentLegitimate interest + DPIA
Consent flowSimple onboardingAdmin deployment + employee notification
Works councilN/ARequired (DE, FR, IT, NL, ES)
Stealth modeOptional (own device)Prohibited
Data accessUser onlyUser + admin (audit logged)

Certification Roadmap

PhaseCertificationTimeline
Pre-launchGDPR compliance assessmentBefore launch
0-6 monthsSOC 2 Type I6 months
6-12 monthsSOC 2 Type II + ISO 2700112 months
12-18 monthsISO 27701 (privacy management)18 months

Design Gate: MVP Compliance

Compliance requirements specific to the Screenpipe plugin MVP. What we must build before shipping.

Key insight

Screenpipe has no consent flow per pipe. Pipes get full access to all data. We own our own consent UI, privacy notice, and data processing disclosures.

Risk by Jurisdiction

JurisdictionRiskKey Requirement
GDPRHighClaude API calls = international data transfer. Need Anthropic DPA, DPIA, privacy notice.
ChileLow (rising)New GDPR-style law Dec 2026. Plan for it, don't gate launch.
US/CCPAMediumPrivacy policy disclosing Claude API sharing. Right-to-know/delete.
BIPALowWe don't process biometrics. Triggers only if we add image analysis.

Data Boundaries

DataCan Collect?Notes
App names, timestampsYesMetadata — lowest risk across all jurisdictions
UI event patternsYesClicks, switches, scroll counts
Aggregated patternsYes"User switched A→B 15 times"
Raw screenshot imagesNoNot needed, highest compliance burden
Audio transcriptionsNoTwo-party consent issues (CA PC 632)
Clipboard contentNoMay contain passwords, sensitive data
Email/chat body textNoECPA interception risk

What Goes to Claude API

Only: App names + frequencies + durations. No raw text, no timestamps, no window titles, no clipboard, no audio.

{
  "patterns": [
    {
      "type": "app_switch_loop",
      "apps": ["Salesforce", "Google Sheets"],
      "frequency_per_week": 14,
      "avg_duration_seconds": 180,
      "confidence": 0.85
    }
  ]
}

Required Consent Flow

ElementRequired?Notes
First-run disclosureYes"AutoReview sends anonymized pattern data to Claude API"
Opt-in for API callsYesDon't auto-send on install
Data previewRecommendedShow what will be sent before first API call
Pause/disable toggleYesStop AutoReview without stopping Screenpipe
Data deletionYesDelete all AutoReview-generated data
App exclusion listRecommendedExclude 1Password, email, chat by default

Build Checklist

  1. Build our own consent/onboarding flow (Screenpipe provides none)
  2. Text sanitization layer between Screenpipe data and Claude API
  3. Anthropic DPA signed; verify data retention/training policies
  4. Store patterns in our own DB with retention/deletion controls
  5. Default to metadata-only — escalate to OCR text only when needed
  6. No audio, no screenshots, no clipboard in MVP
  7. Ship with privacy policy disclosing Claude API data flow

Design Gate: Architecture Plan

Two-process design: Pattern Detector (Python, $0) + Automation Advisor (pipe.md, Haiku).

System Diagram

USER'S MACHINE (macOS) ┌──────────────────────────────────────────────────────────┘ │ │ │ SCREENPIPE (existing) │ │ ┌─────────────┘ ┌──────────────────────────┘ │ │ │ Capture │───>│ SQLite DB │ │ │ │ Engine │ │ (ocr_text, ui_events, │ │ │ └─────────────┘ │ elements, frames) │ │ │ └────────────┬─────────────┘ │ │ │ │ │ /raw_sql│:3030 │ │ │ │ │ AUTOREVIEW COMPONENTS │ │ │ ┌────────────────────────────────┬──────────────┘ │ │ │ ▼ │ │ │ │ Pattern Detector ──> patterns.db │ │ │ │ (Python daemon) (our SQLite) │ │ │ │ polls every 15m │ │ │ │ │ $0 inference ▼ │ │ │ │ Automation Advisor ──> Weekly Report │ │ │ │ (pipe.md, 4h) (Markdown) │ │ │ │ │ │ │ │ │ │ │ macOS Notification │ │ │ └───────┬───────────────────────────────────┘ │ │ │ summarized patterns only │ └─────────┬────────────────────────────────────────────────┘ ▼ Claude Haiku 4.5 (Anthropic API) ~$0.013/run = $2.28/month

Why Two Processes

Component A: Pattern Detector

LanguagePython 3.11+
RoleStandalone daemon. Queries Screenpipe's SQLite via /raw_sql every 15 min. Runs 5 detection algorithms.
Dependenciesrequests, sqlite3 (stdlib), schedule, pync
Runs asBackground process via launchd plist (macOS)
LLM cost$0 — pure SQL + Python frequency analysis

Component B: Automation Advisor

LanguageMarkdown (pipe.md) — executed by Screenpipe's built-in AI agent
ScheduleEvery 4 hours
InputReads from Pattern Detector's SQLite (pre-aggregated data)
LLMClaude Haiku 4.5 — "Given these patterns, suggest automations"
Cost~$0.01-$0.03 per run

Inference Cost Model

ModelCost/RunCost/Day (6 runs)Cost/MonthRecommendation
Haiku 4.5$0.013$0.076$2.28Use for MVP
Sonnet 4$0.047$0.279$8.37v2 paid tier
Opus 4$0.237$1.42$42.53Never
With optimizations

$0.50–$1.00/month per user with caching, skip-when-no-new-patterns, and reduced frequency.

Data Flow: Three-Stage Privacy Funnel

RAW SCREENPIPE DATA PATTERN DETECTOR SENT TO CLAUDE API (stays in Screenpipe DB) (stays in AutoReview DB) (leaves machine) Window titles ──────────┐ App names ────────────┤──► "App A → App B loop, ──► "app_switch_loop: UI events ────────────┤ 14x/week, ~3 min each" Salesforce → Sheets, Timestamps ───────────┤ 14x/week, ~3 min" OCR text ─────────────┤ (aggregated, no raw Clipboard content ──────┘ content, no timestamps) (app names only)

API Endpoints Used

EndpointUsed ByPurpose
POST /raw_sqlPattern DetectorAll analytics. The workhorse.
GET /searchNot in MVPReserve for v2.
WS /ws/eventsNot in MVPReserve for v2 real-time.

AutoReview SQLite Schema

CREATE TABLE patterns (
    id              INTEGER PRIMARY KEY,
    pattern_type    TEXT NOT NULL,
    signature_hash  TEXT NOT NULL UNIQUE,
    apps            TEXT NOT NULL,        -- JSON array
    sequence        TEXT,
    frequency_week  REAL,
    avg_duration_s  REAL,
    confidence      REAL,                 -- 0.0-1.0
    first_seen      TEXT NOT NULL,
    last_seen       TEXT NOT NULL,
    status          TEXT DEFAULT 'active'
);

CREATE TABLE suggestions (
    id              INTEGER PRIMARY KEY,
    pattern_id      INTEGER REFERENCES patterns(id),
    suggestion_type TEXT NOT NULL,        -- zapier | applescript | shortcut | n8n
    difficulty      TEXT NOT NULL,        -- easy | medium | hard
    title           TEXT NOT NULL,
    description     TEXT NOT NULL,
    setup_url       TEXT,
    llm_model       TEXT NOT NULL,
    status          TEXT DEFAULT 'pending'
);

File Structure

~/.screenpipe/pipes/autoreview-advisor/
└── pipe.md                    # Automation Advisor

~/.autoreview/
├── patterns.db                # Our SQLite database
├── config.yaml                # User preferences
├── detector.py                # Pattern Detector daemon
├── reports/
│   ├── latest.md              # Most recent report
│   └── weekly_2026w10.md      # Archived reports
└── logs/
    └── detector.log

Key Decisions

  1. Two-process vs single pipe: Global semaphore makes single-pipe unworkable for 15-min polling.
  2. /raw_sql over /search: We need GROUP BY, window functions, time-range filtering.
  3. Haiku over Sonnet: 3.7x cheaper. Structured task, quality sufficient.
  4. No WebSocket in MVP: 15-min polling catches same patterns with slight delay.
  5. Markdown over dashboard: Zero frontend deps. Validates core value first.

Design Gate: Scope & Wireframe

What the user sees, what we detect, and what's explicitly out of scope.

User Journey

PhaseWhat Happens
Install (2 min)screenpipe pipe install + pip install autoreview-detector + autoreview-detector start
Day 1Nothing visible. Detector silently builds pattern data. One notification: "Learning your workflows."
DailySystem notification when threshold met: "Slack ↔ Jira copy-paste — 8 times today (~20 min)"
WeeklyAutomation Advisor fires. Generates Markdown report. Notification: "3 patterns, ~4.5h/week recoverable"
ActionUser reads report, follows suggestion links, edits config to dismiss/snooze

Weekly Report (The "Dashboard")

┌─────────────────────────────────────────────────────────────────┘ │ AUTOREVIEW — WEEKLY REPORT │ │ Week of March 3–9, 2026 │ │ │ │ Patterns: 5 | New: 2 | Time wasted: 4h 35m/week │ │ Automatable: 3 of 5 (3h 10m recoverable) │ │ │ │ ┌─────────────────────────────────────────────────────────┘ │ │ │ #1 — Salesforce → Sheets copy-paste NEW │ │ │ │ 14x/week | ~45 min/week | Confidence: HIGH │ │ │ │ Suggestion: Zapier sync (EASY, ~10 min setup) │ │ │ │ Saved: 45 min/week = 39 hours/year │ │ │ └─────────────────────────────────────────────────────────┘ │ │ │ │ ┌─────────────────────────────────────────────────────────┘ │ │ │ #2 — Jira → Slack ticket sharing │ │ │ │ 23x/week | ~30 min/week | Confidence: HIGH │ │ │ │ Suggestion: Jira Automation rule (EASY, ~15 min) │ │ │ └─────────────────────────────────────────────────────────┘ │ │ │ │ ┌─────────────────────────────────────────────────────────┘ │ │ │ #3 — Morning report ritual (9:00–9:25 AM daily) │ │ │ │ 5/5 weekdays | ~2h 5m/week | Confidence: VERY HIGH │ │ │ │ Suggestion: n8n workflow (HARD, ~45 min setup) │ │ │ └─────────────────────────────────────────────────────────┘ │ │ │ │ Previously automated: 0 | Time saved: 0h | Weeks tracked: 1 │ └─────────────────────────────────────────────────────────────────┘

V1 Pattern Detection

#PatternDetectionThresholdPriority
1App-switch loopsSQL + sliding window3+ repeats of 2-app sequence in 10 minHighest
2Copy-paste bridgesSQL join on clipboard events3+ cross-app cycles in 1 hourHigh
3Time-of-day ritualsSQL group + sequence matchingSame 3+ app sequence, 3+ daysHigh
4Repetitive navigationSQL on URLs/window titlesSame URL on 3+ days, 5+ visitsMedium
5Form fillsSQL + fuzzy text match10+ inputs in same app+windowLow
Build order

Ship with 1–3 working. Patterns 4–5 can be partial or v1.1.

Suggestion Output Format

### Suggestion: Salesforce → Google Sheets sync

What you're doing manually:
You copied data from Salesforce and pasted it into Google Sheets
14 times this week. This took approximately 45 minutes.

How to automate it:
Zapier has a pre-built integration that syncs Salesforce records
to Google Sheets automatically.

Difficulty: EASY (no code, ~10 min setup)
Estimated time saved: 45 min/week = 39 hours/year

User Actions (per pattern)

  1. Follow suggestion — click link or run generated script
  2. Dismiss — add to config.yaml ignored list
  3. Snooze — resurfaces after 2 weeks
  4. Mark automated — enables progress tracking

All via config file edits or report links. No GUI in v1. Power-user-appropriate for Screenpipe audience.

What's NOT in V1

ExcludedWhyWhen
Dashboard UIMarkdown reports validate value firstv2
Automation executionTrust/security cliff. Prove detection firstv2
Real-time detection15-min polling catches same patternsv2
Windows/LinuxmacOS-primary audiencev2
Custom patternsDiscovery is the value, not configv2
Team featuresIndividual-first. Teams = surveillancev3
ML detectionRule-based SQL catches obvious winsv3
Audio/meetingsDifferent productNever
Technical risk

Only unknown: Screenpipe's exact DB schema needs validation against a live instance. First engineering task: connect to real Screenpipe, run PRAGMA table_info(ui_events), verify column names.

Mexico Compliance (LFPDPPP 2025)

Mexico's data protection law was fully replaced March 2025. Significantly easier than GDPR.

Key finding

Dissociated data (datos disociados) is explicitly exempt from the LFPDPPP. Our sanitization layer (app names + frequencies only, no user identity) has a strong argument for exemption from the law entirely for the API transmission.

LFPDPPP 2025 vs. GDPR

DimensionMexicoGDPRFor Us
Legal basisConsent-dominatedSix bases incl. legitimate interestMexico easier
Employee consentAccepted as validPresumed not freely givenMexico easier
DPOMandatory for ALLConditionalMexico stricter
DPIANot requiredMandatory for high-riskMexico easier
International transfersConsent onlyComplex (SCCs, adequacy, BCRs)Mexico easier
Penalties~$1.7M USD maxEUR 20M or 4% turnoverMexico easier
EnforcementSABG (political, transitional)Independent DPAsMexico easier
Screen recordingNot specifically regulatedEffectively illegalMexico easier
Right to disconnectExplicit in labor lawSome member states onlyMexico stricter

Institutional Change: INAI → SABG

Consent Requirements

Use CaseConsent TypeMechanism
Consumer/personal useTacitShow Aviso at first run; user proceeds = consent
Enterprise/employeeExpressWritten telework agreement
International transfer (Claude API)ExpressClick-through at first run (valid for non-sensitive data)

Aviso de Privacidad Requirements (Art. 15-16)

Cost: ~$500-1,000 USD for Mexican lawyer review.

Cross-Border Transfer to Claude API

Right to Disconnect (NOM-037)

Enterprise only

Teleworkers (40%+ remote): monitoring must stop outside working hours. AutoReview enterprise mode needs automatic pause. Consumer/personal use is NOT affected. Penalties: 50-5,000 UMA (~$285-$28,285 USD). No case law yet.

What AutoReview Must Do for Mexico

  1. Spanish-language Aviso de Privacidad (lawyer-reviewed)
  2. Click-through consent at first run (non-sensitive data = sufficient)
  3. Express consent for Claude API transfer
  4. Designate DPO (mandatory for all controllers)
  5. ARCO rights for AutoReview-generated data
  6. Enterprise mode: auto-pause outside working hours

Mexico: Architecture Changes

Zero architecture changes. All Mexico work is localization + prompt engineering.

Bottom line

The two-process design (Pattern Detector + Automation Advisor) holds completely for Mexico. No code architecture changes needed.

What Actually Changes

AreaChangeScopeImpact
Report languageSpanish prompt in pipe.mdPrompt engineeringNONE
Notification stringsTranslate ~10 stringsLocalizationNONE
Mexican Spanishtú vs usted, Mexican vocabPrompt engineeringNONE
App awarenessWhatsApp, CONTPAQi, TrelloPrompt engineeringNONE
Config languageSpanish comments in config.yamlLocalizationNONE
System app namesSpanish macOS localizationsConfigNONE
API retryExponential backoffMinor codeNONE
Windows supportCONTPAQi/Aspel are Windows-onlySignificantv1.1

Mexican SaaS Ecosystem

CategoryUS DefaultMexico Priority
CommunicationSlackWhatsApp Business (92% of population uses WhatsApp)
AccountingQuickBooksCONTPAQi, Aspel (both Windows-only)
TaxSAT portal (CFDI), Facturama, Xepelin
CRMSalesforceHubSpot (huge in MX SMEs), Clientify
Project mgmtJiraTrello, Monday.com, ClickUp
ERPSAPCONTPAQi Comercial, Bind ERP, Odoo MX
HR/PayrollRhinus, Runa, Worky
#1 pattern in Mexico

WhatsApp → spreadsheet copy-paste. 87% of LATAM companies use WhatsApp for business communication. The Slack→Jira example becomes WhatsApp→Google Sheets in Mexico.

Spanish Language Support

Platform Risk: macOS in Mexico

~13%
macOS share in Mexico
~28%
macOS share in US
~77%
Windows share in Mexico

CONTPAQi and Aspel (Mexico's dominant accounting tools) are Windows-only. Windows support moves from v2 to v1.1 for Mexico market.

Infrastructure

Mexico: Build Plan

4-week plan with day-by-day Week 1. CTO needs ~5-6 hours/week.

What Changes from Base Plan

AreaBase PlanMexico Change
LanguageEnglishAll output in Mexican Spanish
ConsentPrivacy policyAviso de Privacidad (specific legal instrument)
App detectionSalesforce, Jira, SlackWhatsApp, CONTPAQi, Trello, HubSpot
DistributionScreenpipe DiscordCoders Mexico Slack, #DevsMX, WhatsApp groups
TestingEnglish speakersMexican macOS users with Spanish QA

CTO Role vs. Claude Code

ResponsibilityCTOClaude Code
Install Screenpipe, validate schemaX
All Python code (detectors, daemon, tests)X
SQL queriesReviewWrite
pipe.md promptReviewWrite
Aviso de PrivacidadLawyer reviewDraft
Integration tests against live ScreenpipeX
Spanish language QAX
Beta tester recruitmentX
Community outreachPost & engageDraft posts
Anthropic API setupX

CTO Time Commitment

6-8h
Week 1: schema validation, integration tests
4-6h
Week 2: Haiku QA, Spanish review
3-4h
Week 3: installer test, recruit testers
6-8h
Week 4: beta coordination, feedback

Week 1: Day by Day

DayTaskWho
1-2Project scaffolding, screenpipe_client.py, db.py (SQLite schema), config.py with locale: es-MXClaude Code
1Schema validation: PRAGMA table_info(ui_events) via /raw_sql. Document exact column names.CTO
3-4App-switch detector, copy-paste detector, daily ritual detector. Unit tests with Mexican app names.Claude Code
4Integration test: run detectors against live Screenpipe DB. Report results.CTO (2h)
5Navigation detector, form fill detector (stub OK), daemon orchestrator, Mexican app name mapping.Claude Code
5End-to-end: start daemon, let run 2 hours, check patterns.db.CTO (30m)

Decisions Needed Before Coding

  1. Mexico-first only, or bilingual? (Rec: Spanish only for v1)
  2. Free for beta? (Rec: yes)
  3. GitHub private or public? (Rec: private during beta)
  4. Does CTO have 3-5 macOS users in Mexico for beta?
  5. Who pays Anthropic API? (~$2/month)
  6. Mexican lawyer contact for Aviso? (~$500-1K)

2-Week Stripped-Down Version

If you want to validate faster, cut to 10 days. You lose patterns 4-5, polished installer, lawyer-reviewed Aviso. You keep the core loop: detect → suggest in Spanish.

Day 1:  Screenpipe install + schema validation + scaffolding
Day 2:  App-switch + copy-paste detectors (highest value)
Day 3:  Daily ritual detector + daemon orchestrator
Day 4:  Integration test against live Screenpipe
Day 5:  pipe.md with Haiku (Mexican Spanish)
Day 6-7: Report template + notifications (Spanish)
Day 8:  Aviso de Privacidad draft + consent flow
Day 9:  Installer, launchd, README (Spanish)
Day 10: Clean install test. Ship to 2-3 testers.
Critical path

Schema validation (CTO, Day 1) → SQL queries work → detectors complete → Haiku integration → E2E test → ship to testers. The blocker is always schema validation.

Mexico: Risks & Distribution

What's easier, what's harder, and where to find users.

What's Easier (vs. US/EU)

  1. Privacy law lighter than GDPR. No DPIA. No authority registration. Aviso is a document, not a system.
  2. No state-level patchwork. One federal law covers all of Mexico (unlike US: CCPA + BIPA + NYC).
  3. Smaller market = faster feedback. Know your first 10 users personally.
  4. Zero competition in Spanish. Cofia, Autostep, Edra are all English-first, US-focused.
  5. CTO network in Mexico = beta testers are a phone call away.

What's Harder

  1. Screenpipe adoption in Mexico unknown. 17K stars are English-speaking. Could be 5-50 users in Mexico.
  2. macOS market share lower (~12-15% vs ~28% in US).
  3. Mexican SaaS tools less automation-friendly. CONTPAQi/Aspel are desktop Windows apps. No Zapier integrations.
  4. Spanish quality is a trust signal. Bad machine Spanish = "this wasn't built for us."
  5. Distribution is fragmented. No single Mexican HackerNews.

What Could Go Wrong

RiskMitigation
<10 Screenpipe users in MexicoPosition AutoReview as the reason TO install Screenpipe
Pattern detection misses Mexican appsFocus on universal patterns (browser, Google Workspace, WhatsApp Web)
Haiku generates Castilian SpanishExplicit system prompt + native speaker QA
Nobody cares ("just hire an assistant")Target remote workers/freelancers competing globally
Screenpipe API breaksPin version, error handling

Minimum Viable Test

The test

3 machines (CTO + 2 testers) in Mexico. 1 week. Success: 3+ real patterns per user, Spanish reads naturally, suggestions are actionable, no crashes. Cost: ~$0 Claude Code + 20h CTO + ~$2 API.

Distribution Channels

ChannelTypeWhy
Coders MexicoSlack communityActive Mexican dev community
Twitter/X #DevsMXSocialMexican tech Twitter very active
WhatsApp groupsDirect messagingHow things actually spread in Mexico
Telegram groupsMessagingPopular among Mexican devs
LinkedInProfessionalMexican tech pros very active
GDG chaptersMeetupCDMX, Guadalajara, Monterrey
Product Hunt (ES)LaunchGets LATAM tech press pickup
NOT Screenpipe Discord

English-speaking, primarily US/EU. Wrong audience for a Mexico-first launch. Use Mexican tech communities instead.

Cross-Platform Strategy

macOS is ~12-15% market share in Mexico. CONTPAQi/Aspel are Windows-only. What does Windows support take?

Screenpipe Windows Status (Current Issues)

IssueImpact on AutoReview
Shallow Chromium/Electron accessibility tree (#2362)HIGH HubSpot, Salesforce, SAT portal run in browsers. 2-3 of 5 detection algorithms produce worse results.
Windows OCR fallback broken (#2324)HIGH When accessibility tree fails, no text fallback.
Windows Defender false positivesMEDIUM Deletes Screenpipe binary. Users must whitelist.
Onboarding stuck on Windows (#2396)MEDIUM First-run experience broken for some users.

Our Code Changes for Windows

ComponentmacOS (current)Windows (needed)Effort
Notificationsosascriptdesktop-notifier (cross-platform)~2 hours
Background servicelaunchd plistTask Scheduler or NSSM~4 hours
File paths~/.autoreview/pathlib.Path everywhere~2 hours
Installerinstall.shPowerShell or PyInstaller + Inno Setup~1 day
Detection algorithmsNo change (pure Python + SQL)0
Screenpipe API callsNo change (HTTP to localhost:3030)0
Total Windows port effort: ~2-3 days

The real risk isn't our code — it's Screenpipe's capture quality on Windows.

Three Options

A: Mac-First, Windows Later

Timeline: 4 weeks Mac + 1 week Windows later

Pro: Ship faster. Validate on stable platform. Screenpipe bugs may get fixed.

Con: Misses 75-80% of Mexico market.

B: Cross-Platform Day 1

Timeline: 5-6 weeks

Pro: Full Mexico market from launch.

Con: Slower. Debugging Screenpipe's Windows bugs becomes our problem.

C: Mac MVP + Windows Spike RECOMMENDED

Timeline: 4 weeks + 1 day spike in week 2

Pro: Best of both. Ship Mac fast. Derisk Windows in week 2. Write cross-platform Python from day 1 (free).

Con: If spike fails, we know early but wait for Screenpipe fixes.

Screenpipe Dependency

How much risk are we taking by building on someone else's platform?

Pricing & Business Model

TierPriceWhat You Get
Open-source CLIFree forever (MIT)Full recording, local search, API, pipes
Desktop App$400 one-timeGUI, auto-updates, settings UI
Desktop + Pro$600 one-timeAbove + 1yr cloud sync, cloud AI
Pro renewal~$39/monthCloud features only

The core engine and API we depend on are MIT-licensed. Users don't need the $400 app — CLI runs for free.

Platform Risk Matrix

ScenarioLikelihoodImpactMitigation
API changes significantlyHIGHMediumAdapter pattern. Pin versions. Abstract all calls.
Goes paid / changes pricingLOWLowFork the MIT-licensed core.
Gets acquiredMEDIUMMediumMIT license protects us. Fork if needed.
Shuts downLOWHighFork MIT core. Maintain our own builds.
Removes pipe systemMEDIUMLowWe should bypass pipes anyway (Decision 4).

Community Health

Traction

17,177
GitHub Stars

1,451 forks. 157 contributors. 85 open issues.

Bus Factor

1
RED FLAG

Founder (Louis Beaumont) has 82% of commits. Essentially a one-person project.

Release Cadence

~3/day
Releases

Active but chaotic. No semver discipline. Jumped from v0.3.x to v2.2.x.

Three Non-Negotiable Guardrails

If we proceed with Screenpipe
  1. Adapter layer — Zero direct Screenpipe API calls from business logic. Everything through our abstraction.
  2. Version pinning — Lock to specific Screenpipe version per release. Test upgrades explicitly.
  3. Fork-ready architecture — Modular enough to swap in a maintained fork within a week.

Alternatives to Screenpipe

What else could we build on top of — or replace it with?

Comparison Matrix

ApproachPlatformData RichnessTimelineMaintenanceDistribution
Screenpipe pluginMac + Windows (degraded)Full4 weeksLowHard (install Screenpipe first)
Build own captureMac + WindowsFull14-16 weeksVery highMedium
ActivityWatchMac + Win + LinuxLow (app names only)3 weeksLowMedium
Browser extensionAll (Chrome/Firefox/Edge)Medium (60-70%)5-6 weeksMediumEasy (Chrome Web Store)
Hybrid: Extension + agentAllHigh7-8 weeksMediumMedium

Detailed Assessment

Build Own Capture Layer

10-12 weeks just for capture. Python libs: mss (screenshots), pyobjc (Apple Vision OCR), pywinauto (Windows a11y).

Verdict: Turns us into a screen capture company before we've validated the product. Not recommended.

ActivityWatch (16.9k stars)

Captures app/window names and browser URLs. Does NOT capture in-app content, form fields, or accessibility tree.

Verdict: Not granular enough. Only covers 2 of 5 detection algorithms (app-switch, daily ritual).

Browser Extension STRONG OPTION

Captures 60-70% of knowledge worker patterns: tab switching, form fills (DOM), copy-paste, navigation, interaction time.

Misses: Desktop app switching, non-browser apps (CONTPAQi, Aspel).

Distribution: Chrome Web Store vs "install Screenpipe + permissions + our tool." Dramatically easier.

Mexico reach: Chrome ~65% market share.

OpenAdapt (1.5k stars)

Captures screenshots + accessibility + input on Mac/Windows. Built for demo recording, not passive monitoring.

Verdict: Worth stealing ideas from their capture modules, not using as a platform.

The Big Strategic Question

The Mexico #1 pattern (WhatsApp → spreadsheet) works with both approaches. The question is: do we need desktop app detection, or does browser detection cover enough? This changes the entire MVP.

Python/Django Stack

Our CTO knows Python/Django. Does the architecture fit?

Architecture Fit

ComponentLanguage TodayPython Works?Notes
Pattern DetectorPythonYESAlready Python. requests, sqlite3, schedule.
Automation Advisorpipe.md (Screenpipe agent)YESReplace with anthropic SDK — more control.
ReportsMarkdown filesYESjinja2 + markdown
NotificationsosascriptYESsubprocess.run() or desktop-notifier
Dashboard (v1.1)N/AYESFlask or static HTML generation

The pipe.md Decision

pipe.md is NOT Python — it's a Markdown file executed by Screenpipe's AI agent. Should we keep it?

A: Keep pipe.md

Pro: Zero code to maintain.

Con: Locked into Screenpipe's agent. Limited prompt control. If pipe spec changes, we break. Outsources our core value.

B: Python Script RECOMMENDED

Pro: Full control over prompts, model, output, errors, retries, logging.

Con: We manage scheduling + API key.

The advisor is our core value — don't outsource it to someone else's runtime.

C: Django Management Cmd

Option B inside Django's command framework.

Verdict: Django overhead for a CLI script. No benefit.

Django vs Flask vs Nothing

CriterionDjangoFlaskNo Framework
Right-sized for "render 1 report"NoBarelyYES
Admin panelBuilt-inNoNo
Auth systemBuilt-inNoNo
CTO learning curveZeroLowN/A
Recommendation by version
  • MVP: No framework. Generate static HTML from Python (Jinja2 → file). User opens ~/.autoreview/reports/index.html.
  • v1.1: Flask. 1 file, 3 routes, minimal overhead.
  • v2+: Django. When you need admin panel, user accounts, multi-device sync, REST API.

Recommended File Structure (Pure Python, No pipe.md)

autoreview/
├── pyproject.toml
├── src/autoreview/
│   ├── main.py                 # Entry point — detector + scheduler
│   ├── config.py               # Paths, intervals, API key
│   ├── detector/
│   │   ├── runner.py           # Polls Screenpipe, orchestrates
│   │   ├── screenpipe.py       # HTTP adapter for localhost:3030
│   │   ├── algorithms/
│   │   │   ├── app_switch.py
│   │   │   ├── copy_paste.py
│   │   │   ├── daily_ritual.py
│   │   │   ├── navigation.py
│   │   │   └── form_fill.py
│   │   └── storage.py          # Write to patterns.db
│   ├── advisor/
│   │   ├── analyzer.py         # Read patterns, call Claude
│   │   └── prompts.py          # Prompt templates
│   ├── reporter/
│   │   ├── generator.py        # Jinja2 → Markdown + HTML
│   │   └── templates/
│   └── notifier.py             # Cross-platform notifications
├── tests/
└── scripts/
    └── install_launchd.py

Decision Matrix

Summary for CTO discussion. Four decisions, one big strategic fork.

Four Decisions

#DecisionRecommendedAlternativeKey Risk
1Cross-platformMac MVP + Windows spike in week 2Cross-platform day 1 (+1.5 weeks)Screenpipe Windows capture quality
2Screenpipe dependencyYes, with adapter + pinning + fork-readyBuild own capture (+10 weeks)Bus factor 1. No API stability.
3Platform approachBrowser extension OR Screenpipe pluginHybrid extension + agentExtension misses desktop apps. Screenpipe limits distribution.
4StackPure Python. No pipe.md. Flask for v1.1.Keep pipe.md + Djangopipe.md outsources core value. Django overkill.

The Big Fork: Screenpipe Plugin vs Browser Extension

This is the real strategic question. Everything else is tactical.

Screenpipe PluginBrowser Extension
Data richnessFull (screen, OCR, a11y, audio)60-70% (browser only)
Timeline4 weeks5-6 weeks
DistributionInstall Screenpipe + our toolChrome Web Store
Mexico reachUnknown (maybe 50 users?)~65% of all users
WindowsDegraded (capture bugs)Works everywhere
Dependency riskHigh (Screenpipe)Low (Chrome APIs stable)
CONTPAQi/AspelCan detect (if Windows + OCR works)Cannot detect (desktop apps)
WhatsApp Web → SheetsCan detectCan detect
Key Insight

The Mexico #1 pattern (WhatsApp → spreadsheet) works with both approaches. The question is whether we need desktop app detection or browser-only covers enough to validate demand.

Recommended Next Steps

  1. CTO decides: Screenpipe plugin or browser extension?
  2. If Screenpipe: Validate DB schema on live instance (blocking).
  3. If extension: Prototype Chrome extension tab/form tracking (1-2 days).
  4. Either way: Pure Python stack, no pipe.md, Flask for v1.1.

Sources

All claims sourced. Key references below.

Competitor Research

Compliance Research

Technical Research