Troubleshooting Setting Up a Tracked Site
What’s Covered?
This guide covers troubleshooting tips for resolving errors encountered while setting up a site to track in a Campaign.
Troubleshooting Errors Encountered When Setting up a Moz Pro Campaign
Our Crawlers Were Unable to Access That URL
Some folks receive this response when they try to set up a Campaign. Let's go through some steps to ensure you've entered the correct details and checking your site is crawlable.
First of all, please check the following:
- Check you've entered the correct site
- Check there aren't any extra blank spaces at the beginning or end of your URL
- Enter your URL into your browser and check it's publicly accessible
Still stuck? Let's go Through Some Other Troubleshooting Steps
To create a Campaign, we need to receive a valid HTTP response from the web server hosting your site.
That may sound rather complicated, but it basically means that our crawler, Rogerbot, must receive a response from your website's server that enables us to crawl your site. So even if your site is accessible in the browser there may be something blocking our crawler behind the scenes.
1. Test your site by sending an HTTP response using any online HTTP tool
There are many ways to run a test. One third-party tool that we use is httpstatus. Enter your site, for example moz.proxybugs.com, and click Submit. A valid URL will return a 200 OK or 301 response. If you receive any other error, then you should report this to your website administrator or hosting provider to investigate and resolve.
2. Check you're not blocking AWS servers
Our Campaign crawler, rogerbot, is hosted with AWS. In some cases sites may have firewalled AWS servers, which would block rogerbot from accessing your site. The best way to fix this is the check with your website administrator or hosting provider to investigate and resolve.
3. Check you’re not blocking our crawler
In some instances the http tool will return a valid http status and we will still not be able to crawl a site. In this case you'll need to speak to your website administrator or hosting provider to see if they can check their server logs for responses to User-agent: rogerbot. Please ensure your site and server are not blocking rogerbot. Blocking our crawler will result in a forbidden response from the server and you will not be able to move forward with Campaign creation.
Related Articles
Was this article helpful?
Yes! Amazing! Yes! It was what I needed. Meh. It wasn’t really what I was looking for. No, it wasn’t helpful at all.
Thanks for the feedback.