When I try to run the web crawler against a site we host, it fails with this error:
Failed HTTP request: Unable to request "< domain >" because it resolved to only private/invalid addresses
The site in question would resolve to a 10.n.n.n ip address. Is the crawler configured to reject that? Is there a way to override that behavior?