Show HN: tomcp.org – Turn any URL into an MCP server

(github.com)

31 points | by ami3466 3 hours ago

7 comments

  • dennisy 56 minutes ago
    I am not quite clear why this adds value over a simple web fetch tool which does not require configuration per site.
  • mbreese 2 hours ago
    I think this is a good idea in general, but perhaps a bit too simple. It looks like this only works for static sites, right? It then performs a JS fetch to pull in the html code and then converts it (in a quick and dirty manner) to markdown.

    I know this is pointing to the GH repo, but I’d love to know more about why the author chose to build it this way. I suspect it keeps costs low/free. But why CF workers? How much processing can you get done for free here?

    I’m not sure how you could do much more in a CF worker, but this might be too simple to be useful on many sites.

    Example: I had to pull in a docs site that was built for a project I’m working on. We wanted an LLM to be able to use the docs in their responses. However, the site was based on VitePress. I didn’t have access to the source markdown files, so I wrote an MCP fetcher that uses a dockerized headless chrome instance to load the page. I then pull the innerHTML directly from the processed DOM. It’s probably overkill, but an example of when this tool might not work.

    But — if you have a static site, this tool could be a very simple way to configure MCP access. It’s a nice idea!

  • bsima 2 hours ago
    Who is tom and why is he copying?
  • bakies 2 hours ago
    I thought this is what the web_fetch tools already did? Tools are configured through MCP also, right? So why am I prepending a URL, and not just using the web_fetch tool that already works?

    Does this skirt the robots.txt by chance? Not being to fetch any web page is really bugging me and I'm hoping to use a better web_fetch that isn't censored. I'm just going to copy/paste the content anyway.

    • mbreese 2 hours ago
      I think the idea here is that the web_fetch is restricted to the target site. I might want to include my documentation in an MCP server (from docs.example.com), but that doesn’t mean I want the full web available.
  • aritex 1 hour ago
    This is a clever solution to a real problem. I could use this for quick turn around from webpage kb to the mcp. Thanks for sharing.
  • _pdp_ 1 hour ago
    Fun idea although I thought the industry is leaning towards using llms.txt.
    • mbreese 1 hour ago
      Isn’t that for scraping? I think this is for injecting (or making that possible) to add an MCP front end to a site.

      Different use cases, I think.

  • SilentM68 2 hours ago
    Cool :)