Input
Extracted URLs
0 URLs Found

Private & Secure

File uploads are parsed locally in your browser. URL fetches use a secure server-side proxy.

Universal Parsing

Handles standard XML sitemaps, sitemap index files, and malformed formats via regex fallback.

Flexible Export

Copy to clipboard or download as .txt for Excel, Google Sheets, or crawler tools.

How to Extract URLs from a Sitemap

1

Enter or Upload Your Sitemap

Paste any sitemap URL (sitemap.xml, sitemap_index.xml) into the field above, or upload an XML/TXT file directly from your computer.

2

Fetch & Parse Automatically

Click Fetch & Extract. The tool retrieves the XML and extracts every <loc> URL β€” including nested sitemap indexes.

3

Filter, Clean & Export

Remove duplicates, strip query parameters, then copy your clean URL list or download it as a .txt file.

Common Use Cases

SEO Audits

Get a full inventory of indexed pages to compare against Google Search Console or your crawler.

Content Inventory

Catalog all published URLs for site migrations, redirects mapping, or content gap analysis.

Broken Link Detection

Feed extracted URLs into link checkers to find 404 errors and broken redirect chains.

Competitor Analysis

Extract URLs from competitor sitemaps to map their content structure and publishing frequency.

FAQ

FAQ

What is an XML sitemap?
An XML sitemap is a file that lists all important URLs of a website, helping search engines like Google discover and index your pages efficiently. It typically lives at /sitemap.xml or /sitemap_index.xml.
Can I extract URLs from a sitemap index file?
Yes. Our tool handles both standard sitemaps and sitemap index files that link to multiple sub-sitemaps. All <loc> entries are extracted in one pass.
Is there a limit on the number of URLs?
No hard limit. The tool can handle large sitemaps with tens of thousands of URLs. Very large files may take a few extra seconds to process.
Is my data private?
Absolutely. File uploads are parsed entirely in your browser β€” nothing is sent to our servers. When fetching via URL, we use a secure proxy that does not log or store any data.
What file formats are supported?
We support standard XML sitemaps (.xml), sitemap index files, and plain text URL lists (.txt). Malformed XML is handled via a regex fallback.
What does 'Remove Params' do?
It strips query string parameters from each URL (everything after the ?), leaving only the base path. Duplicate paths are automatically removed after this step.