██████╗ ██████╗ ███████╗███╗   ██╗██╗    ██╗███████╗██████╗
██╔═══██╗██╔══██╗██╔════╝████╗  ██║██║    ██║██╔════╝██╔══██╗
██║   ██║██████╔╝█████╗  ██╔██╗ ██║██║ █╗ ██║█████╗  ██████╔╝
██║   ██║██╔═══╝ ██╔══╝  ██║╚██╗██║██║███╗██║██╔══╝  ██╔══██╗
╚██████╔╝██║     ███████╗██║ ╚████║╚███╔███╔╝███████╗██████╔╝
 ╚═════╝ ╚═╝     ╚══════╝╚═╝  ╚═══╝ ╚══╝╚══╝ ╚══════╝╚═════╝

SSH into any website

Search, grep, watch, diff and explore web pages like a filesystem. Built for AI agents. Works from any agent harness.

# Browse any website
$ ssh openweb.sh "ls /web/index/docs.stripe.com"
index.md    docs/    api/    payments/    webhooks/

# Search across all pages
$ ssh openweb.sh "grep -rl 'webhook' /web/index/docs.stripe.com"
/web/index/docs.stripe.com/docs/webhooks/overview.md
/web/index/docs.stripe.com/docs/payments/checkout.md
/web/index/docs.stripe.com/docs/api/events.md

# Read a specific page
$ ssh openweb.sh "cat /web/index/docs.stripe.com/docs/webhooks/overview.md"

# One-liner: list pages on a site
$ ssh openweb.sh "find /web/index/nyt.com -name '*.md' | head -20"

# See what changed since the last crawl
$ ssh openweb.sh "cd /web/index/jobs.netflix.com && git diff HEAD~1"

What is OpenWeb?

OpenWeb crawls websites and converts them to plain markdown files you can search with standard Unix commands. No SDK, no API client — just bash.

  • [1]
    Explicit. Every crawled page is a readable markdown file on a navigable filesystem. You can see exactly what's been indexed, inspect it, search it, and manage it — nothing is hidden in a black-box embedding store.
  • [2]
    File over app. Web content is stored as plain markdown files. That means it's interoperable — pipe it through grep, jq, awk, or any Unix tool. Feed it to any AI. The entire Unix toolkit just works.
  • [3]
    No SDK, simple bash. No client library, no API wrapper, no dependency. Just ssh openweb.sh — cat, grep, find, diff. If your agent can write bash, it can use OpenWeb.
  • [4]
    BYOAI. Plug any AI into this data — Claude, GPT, Gemini, open-source models, your own fine-tune. OpenWeb gives you structured web content and a personal filesystem; you choose what reasons over it.
  • [5]
    Yours. Your crawled data lives in your session, accessible via your SSH key. It's not locked inside a proprietary system — you can extract, copy, and move it however you want. You control your information.
  • [6]
    Collective index. No need to burn compute crawling or spend tokens converting HTML to markdown. The web is already mounted as a shared drive — if someone indexed a page before you, it's instantly available.

How it works

Three steps from zero to searching any website's content.

01
Connect

SSH into openweb.sh. Access any site at /web/index/{domain}/.

02
Index

Pages are converted to markdown and mounted as a virtual filesystem.

03
Explore

Use standard bash commands to search, read, and navigate the content.

Use cases

  • [*]
    Knowledge base Build a personal or company knowledge base from any set of websites — docs, wikis, blogs — all as searchable markdown files your AI agents can reason over
  • [*]
    Personalized onboarding Crawl a user's website to tailor their first-run experience with relevant context
  • [*]
    AI CMO Let marketing agents research competitor sites, track messaging changes, and draft positioning
  • [*]
    Competitive intelligence Agents compare feature pages across competitors to surface gaps and opportunities
  • [*]
    Scheduled signal tracking Monitor websites on a cron for pricing changes, new features, or content updates

Pricing

1 credit = 1 page crawled. Reading cached content is always free.

Free
$0
100 credits, one-time
Hobby
€8 /mo
3,000 credits/mo
Standard
€42 /mo
100,000 credits/mo
Growth
€167 /mo
500,000 credits/mo

Get started free. Upgrade anytime via openweb auth

$ ssh openweb.sh

Coming soon