Agent activity monitoring
See how Claude Code, Cursor, Windsurf, and other AI coding workflows use files, shell commands, network calls, repos, secrets, and local tools across enrolled developer devices.

Private beta for AI coding assistant security
Hosted security control plane for teams adopting AI coding assistants, MCP servers, and AI-built applications. Lighthouse helps AppSec see risky agent behavior, investigate prompt-injection chains, and prepare policy controls around the tools developers already use.
What it protects
Coding assistants now read repos, call tools, inspect terminals, connect to MCP servers, and write production code. Lighthouse gives security teams a managed place to observe, govern, and investigate that activity during a controlled beta rollout.
See how Claude Code, Cursor, Windsurf, and other AI coding workflows use files, shell commands, network calls, repos, secrets, and local tools across enrolled developer devices.
Discover registered and shadow MCP servers, inspect exposed tools, and review risky access paths before they become unmanaged dependencies.
Connect hostile tool output, model behavior, and follow-on actions into an evidence chain your security team can investigate.
Scan applications built with agents for common high-impact flaws such as IDOR, SSRF, unsafe execution, weak signing, and injection patterns.
How it works
Lighthouse is built for the workflow AppSec needs when coding agents become part of day-to-day engineering.
Connect a small developer cohort and capture agent activity with signed ingestion.
See tool calls, MCP exposure, repos, devices, sessions, and risky behavior in one dashboard.
Rebuild prompt-injection, insider-misuse, and AI-built app findings with evidence attached.
Apply findings-as-code policy, route evidence to SIEM or tickets, and export reports.

Analyst workflow
Lighthouse separates insider misuse, prompt injection, and normal development activity into lanes that match early response workflows. Analysts can move from fleet view to one session timeline without rebuilding context by hand.
Product proof
Lighthouse is being shaped around the evidence security leaders ask for before onboarding: what it finds, how it is operated, and how the hosted beta is controlled.
In a blind rediscovery benchmark against published AI-app advisories, Lighthouse found 13 of 28 in-scope issues: 10/17 in langflow and 3/11 in open-webui.
HMAC-signed agent ingestion, offline queue health, findings-as-code policy, audit logs, SIEM and ticket routing, PDF reports, and SARIF export are implemented.
GitHub Actions OIDC image delivery, signed agent artifacts, migration checks, and smoke gates are already part of the beta process.

Private beta