Bulk Index Checker
Paste up to 50 URLs and check whether each one is indexed by Google — with live progress, results table, and CSV export.
- Go to searchapi.io and click Sign up
- Register with your email — no credit card needed for the free plan
- Go to Dashboard → API Key and copy your key
- Paste below — saved in your browser so you only do this once
| # | URL | Status | Page title |
|---|
Set up Google Search Console
Go to search.google.com/search-console and add your site as a property. Verify ownership by adding a DNS TXT record, uploading an HTML file, or using the Google Analytics method. Search Console is your command center for everything indexing-related — you cannot properly manage indexing without it.
One-time setupCreate and submit your XML sitemap
A sitemap is a file that lists every URL on your site you want Google to index. Most CMS platforms (WordPress, Shopify, Wix) generate one automatically — usually at yoursite.com/sitemap.xml. In Search Console, go to Sitemaps → Add a new sitemap and submit the URL. Google will crawl it within days and discover all your pages.
Submit once, Google re-reads it automaticallyCheck robots.txt isn't blocking Google
Visit yoursite.com/robots.txt and make sure there's no Disallow: / rule blocking Googlebot from crawling your pages. This is a common mistake after site migrations or when a staging site's settings are accidentally copied to production. One wrong line in robots.txt can deindex your entire site.
Check immediately if pages are missingScan for noindex meta tags
Check that your pages don't have <meta name="robots" content="noindex"> in the <head>. This tag instructs Google not to include the page in its index. It's often left on accidentally after development, or set incorrectly by SEO plugins. In Search Console, the URL Inspection tool shows you exactly whether a noindex tag was detected on any page.
Inspect any non-indexed URL in Search ConsoleBuild internal links to new pages
Google discovers new pages by following links. If a new page has no internal links pointing to it from other indexed pages, Googlebot may never find it — even if it's in your sitemap. Add links from your homepage, relevant category pages, or a related posts section on similar articles. The more pathways to a page, the faster it gets found and indexed.
Critical for new contentRequest indexing for specific URLs
In Search Console, go to URL Inspection → enter your URL → Request Indexing. This signals Google to prioritize crawling that specific page. It doesn't guarantee immediate indexing but typically gets new content indexed within 24–72 hours on established sites. Use this after publishing important new pages or after making major updates to existing ones.
Use after publishing or major updatesGet backlinks from other indexed sites
When an external site with Google-indexed pages links to yours, Google follows that link and discovers your content. For brand-new domains with no history, this is often the fastest way to get indexed. Submit your site to relevant directories, write guest posts, or get listed in industry resources. Even one quality backlink from a trusted site can trigger Google to crawl and index your content quickly.
Essential for new domainsEnsure pages have enough unique content
Google may crawl a page but choose not to index it if the content is thin, duplicated from another page, or near-identical to other URLs on your site. Each page should have at least 300–500 words of unique, useful content. Avoid publishing placeholder pages, duplicate content across multiple URLs, or pages that are just lists of links with no real text.
Ongoing content quality checkCheck canonical tags aren't redirecting Google
A canonical tag (<link rel="canonical" href="...">) tells Google which version of a page is the "official" one. If your page has a canonical pointing to a different URL, Google will index that other URL instead of yours. This often happens with e-commerce sites that have multiple filter/sort variations of the same page. Use Search Console's URL Inspection to check the canonical Google is using.
Check for e-commerce and CMS sitesMonitor and re-check after 72 hours
After completing the steps above, wait 48–72 hours and run this bulk checker again. For new sites or new domains, allow 2–4 weeks before worrying — Google takes longer to trust brand new sites. If pages are still not indexed after several weeks, go back to Search Console's Pages report and check the specific reason Google gives for excluding each URL (e.g. "Crawled – currently not indexed", "Duplicate without canonical", "Blocked by noindex").
Ongoing monitoring🚫 Noindex tag
A meta robots or X-Robots-Tag header is explicitly telling Google not to index the page. Check page source and HTTP headers.
🤖 Blocked by robots.txt
A Disallow rule prevents Googlebot from even crawling the page. Google can't index what it can't read.
🔗 No internal links
Google discovers pages through links. An "orphan page" with no links from other pages is hard for Googlebot to find.
📄 Thin or duplicate content
Google skips low-value pages. If content is too short, near-identical to another page, or scraped, it may not be indexed.
🔄 Wrong canonical
A canonical tag pointing to another URL tells Google to index that other URL instead of this one.
⏳ New site or page
Brand new domains can take weeks to get crawled. New pages on established sites usually index in 24–72 hours.