Configure robots.txt (Built-in Solution)
From version 1.14.0, the Tyk Developer Portal includes built-in support for customizing therobots.txt
file, which is the standard way to instruct search engines and other well-behaved crawlers about which parts of your site they should not access.
To configure this:
- Log in to the Admin Portal
- Navigate to Settings > General
- Scroll down to the robots.txt Settings section
- Edit the content to control crawler access
robots.txt
configuration would look like:
robots.txt
configuration.
Implement Additional HTTP Headers
You can add custom response headers to further discourage crawling:X-Robots-Tag: noindex, nofollow
- Similar to robots.txt but as an HTTP headerCache-Control: no-store, no-cache, must-revalidate
- Prevents caching
Best Practices
- Regularly check your server logs for unusual crawling patterns
- Consider using a CAPTCHA for registration forms to prevent automated sign-ups (not supported natively by Tyk Developer Portal at this time)
- Use JavaScript-based content rendering for sensitive information, as basic crawlers may not execute JavaScript
robots.txt
rules or use sophisticated techniques to mimic human behavior.