Technical Audit
This utility is a high-performance node optimized for modern browser environments. All data processing is executed client-side, ensuring zero knowledge transfer to external servers.
Extract all URLs from any XML sitemap. Find duplicate entries, analyze folder depth for crawl budget optimization, and export all links to CSV.
Loading your experience...
Please wait a moment
This utility is a high-performance node optimized for modern browser environments. All data processing is executed client-side, ensuring zero knowledge transfer to external servers.
Paste your sitemap URL or upload the XML file. The tool instantly parses all URLs, showing you the full list with their path depth and any duplicate entries. Export everything to CSV with one click.
Duplicate URLs in a sitemap waste your crawl budget — Google has a limit on how many pages it crawls per day. Duplicate entries also signal poor site structure and can split your ranking signals between the same content.
Path depth is the number of folder levels in a URL (counted by slashes after the domain). /blog/post = depth 2. Shallower URLs (depth 1-2) are generally prioritized by search engines and rank more easily.
A single sitemap file can contain up to 50,000 URLs with a maximum file size of 50MB. For larger sites, use a sitemap index file pointing to multiple sitemap files. Only include indexable, canonical URLs.
No. Only include pages you want Google to index: main content pages, blog posts, product pages, category pages. Exclude: admin pages, login pages, thank-you pages, duplicate content, and pages with noindex tags.