Robots.txt Tester
Inspect how homeup.vn controls crawler access, blocked paths, sitemap references and AI crawler rules.
Robots.txt Status
Robots.txt Status
Present
Score
84
/100
ยท Strong
View Full Robots.txt
Robots.txt Content Preview
# As a condition of accessing this website, you agree to abide by the following # content signals: # (a) If a Content-Signal = yes, you may collect content for the corresponding # use. # (b) If a Content-Signal = no, you may not collect content for the # corresponding use. # (c) If the website operator does not include a Content-Signal for a # corresponding use, the website operator neither grants nor restricts # permission via Content-Signal with respect to the corresponding use. # The content signals and their meanings are: # search: building a search index and providing search results (e.g., returning # hyperlinks and short excerpts from your website's contents). Search does not # include providing AI-generated search summaries. # ai-input: inputting content into one or more AI models (e.g., retrieval # augmented generation, grounding, or other real-time taking of content for # generative AI search answers). # ai-train: training or fine-tuning AI models. # ANY RESTRICTIONS EXPRESSED VIA CONTENT SIGNALS ARE EXPRESS RESERVATIONS OF # RIGHTS UNDER ARTICLE 4 OF THE EUROPEAN UNION DIRECTIVE 2019/790 ON COPYRIGHT # AND RELATED RIGHTS IN THE DIGITAL SINGLE MARKET. # BEGIN Cloudflare Managed content User-agent: * Content-Signal: search=yes,ai-train=no Allow: / User-agent: Amazonbot Disallow: / User-agent: Applebot-Extended Disallow: / User-agent: Bytespider Disallow: / User-agent: CCBot Disallow: / User-agent: ClaudeBot Disallow: / User-agent: CloudflareBrowserRenderingCrawler Disallow: / User-agent: Google-Extended Disallow: / User-agent: GPTBot Disallow: / User-agent: meta-externalagent Disallow: / # END Cloudflare Managed Content User-agent: * Allow: / Disallow: /xoso Disallow: /xoso/ Disallow: /xoso/api Disallow: /xoso/refresh
User-agent Rules
| User-agent(s) | Allowed paths | Disallowed paths |
|---|---|---|
| * |
|
No explicit Disallow rules. |
| amazonbot | No explicit Allow rules. |
|
| applebot-extended | No explicit Allow rules. |
|
| bytespider | No explicit Allow rules. |
|
| ccbot | No explicit Allow rules. |
|
| claudebot | No explicit Allow rules. |
|
| cloudflarebrowserrenderingcrawler | No explicit Allow rules. |
|
| google-extended | No explicit Allow rules. |
|
| gptbot | No explicit Allow rules. |
|
| meta-externalagent | No explicit Allow rules. |
|
| * |
|
|
Blocked and Allowed Paths
| Blocked paths |
|
|---|---|
| Allowed paths |
|
| Crawl-delay | No Crawl-delay directive detected. |
Sitemaps Detected
No Sitemap directives found in robots.txt.
AI Crawler Policy
At least one AI crawler (such as GPTBot, ChatGPT-User, ClaudeBot, PerplexityBot or Google-Extended) appears to be blocked by robots.txt.
Issues Found
- At least one user-agent has Disallow: / which blocks the entire site.
- robots.txt does not reference any sitemap URLs.
Recommendations
- Add a Sitemap directive in robots.txt pointing to your primary XML sitemap.
- Avoid blocking the entire site (Disallow: /); restrict only sensitive or low-value paths instead.
- Review your AI crawler policy for GPTBot, ChatGPT-User, ClaudeBot, PerplexityBot and Google-Extended to ensure it matches your content strategy.
- Ensure important pages, CSS and JavaScript assets are crawlable so search engines can fully render your site.