Crawlbase’s Frequently Asked Questions
Top frequently asked questions
Do I need to know how to code to use the Crawling API?
At the moment the only way to use Crawlbase is by integrating it to your product or service, therefore you will need to know development or know/pay someone which can help you.
Please note that we have some very easy to integrate libraries.
Where can I get the API keys?
You can get the API keys or request tokens from the Crawlbase Account Documentation page.
Can the 30 URLs-per-second limit be increased for large-scale crawls?
The 30 URLs-per-second limit applies to LinkedIn crawls. For other websites, we can evaluate and potentially increase the limit on a case-by-case basis. Please contact us to discuss your specific needs.
Can the Smart Proxy be used with HTTPS?
Yes, the Smart Proxy can be used for HTTPS websites, but the connection to the proxy must be done with http. So when connecting to the proxy use http://smartproxy.crawlbase.com
and not https://smartproxy.crawlbase.com
.
All HTTPS traffic that you send to any website using the Smart Proxy, will continue to be securely transfered and we won't be able to access it in any way.
Do you sell a list of proxies?
We want to offer more than just a list or lists of proxies, therefore our API and Crawler products are built on top of thousands of residential and datacenter worldwide proxies combined with artificial intelligence plus our engineering team which helps delivering the best data results possible. If you are looking just for proxies and you don't want to get any of the other benefits, we recommend that you use the Smart Proxy
Is it possible to change my account email address?
Yes, you can change your account email address. To do so, visit the Account page and look for the "Change email address" option. After updating your email address, you will receive a confirmation email at the new address. Please confirm the change by following the instructions in the email.
Live monitor wordings
"Waiting" means that your requests are in your crawler queue waiting to be processed. "Concurrent crawlers" are the requests that are being crawled at the same time. Concurrent crawlers gets increased by our system if you have many pages to crawl, we also monitor crawlers and increase or decrease the concurrency depending on the pool. "Sets to be retried" are your requests that failed for any reason, they land in your crawler retry queue and are processed with a retry rate up until maximum 110 retries.
What if my webhook endpoint is down?
If your Crawler callback is down, you are notified by email, your crawlers get paused and your last failed request due to downtime at your endpoint, is set to be retried. Your crawlers get resumed when your endpoint becomes available automatically. Our monitoring system checks your endpoint every minute.
Need help? Contact us
Please contact us for any type of query regarding products
Start crawling and scraping the web today
Try it free. No credit card required. Instant set-up.
Crawl product data at scale