supermemory icon indicating copy to clipboard operation
supermemory copied to clipboard

[Feature Request] enable crawling when indexing a URL

Open Mikethebot44 opened this issue 2 months ago • 3 comments

Summary

Add a toggle to enable crawling and indexing subpages when scraping a URL, with tabbed display for each subpage's content. Useful for comprehensive indexing of documentation sites like vercel.com/docs.

Problem

Currently, indexing a URL scrapes only the specific page's content, excluding subpages. This limits utility for sites with distributed content, such as documentation or blogs.

Proposed Solution

Add a boolean toggle in the URL input dialog to enable subpage crawling. Update scraping logic to crawl and index subpages if toggled. Modify the content display UI to use tabs for each subpage, replacing the single content container.

Alternatives Considered

Implement a separate "crawl site" feature instead of integrating into URL indexing. Use a third-party crawling service for subpage discovery. Limit to manual subpage selection rather than automatic crawling.

Additional Context

UI could feature tabs labeled by subpage path (e.g., /docs/api, /docs/guides) for easy navigation of indexed content.

Happy to start getting to work on this myself but looking to hear thoughts

Mikethebot44 avatar Oct 09 '25 08:10 Mikethebot44

ENG-310

linear[bot] avatar Oct 09 '25 08:10 linear[bot]

@MaheshtheDev @Dhravya can i work this ?

ViSiOnOp19cr avatar Oct 09 '25 09:10 ViSiOnOp19cr

from dhravya:

Hey, this would be a backend change that's required, and this is in the roadmap as well!

linear[bot] avatar Oct 17 '25 16:10 linear[bot]