Automated crawlers such as WebCopy can scan websites substantially faster than a human browser. This can result in small servers being overloaded, or server administrators to put in place automated blocks for clients that pull too much data at once. WebCopy includes some basic limit settings that the discerning user can enable to comply with the rules of remote hosts.
Disabling limits
To disable all limits and let the crawler run at maximum speed
- From the Project Properties dialogue, select the Speed Limits category
- Select the Do not use limits option
Limiting to specific requests per second
To allow only a maximum number of URLs to be processed per second
- From the Project Properties dialogue, select the Speed Limits category
- Select the Limit to requests per second option
- Enter the maximum number of requests WebCopy is allowed to perform in the Maximum requests per second field
Limiting to specific requests per minute
To allow only a maximum number of URLs to be processed per minute
- From the Project Properties dialogue, select the Speed Limits category
- Select the Limit to requests per minute option
- Enter the maximum number of requests WebCopy is allowed to perform in the Maximum requests per minute field