As well as using rules to control copying a website, you can also specify limits such as the maximum number of files to copy, or to ignore files that are too large or too small.
Maximum File Size
To instruct WebCopy to skip downloading files above a specific size:
- From the Project Properties dialogue, select the Limits category
- Check the Maximum file size option
- Enter a file size, in KiB (1024 bytes)
Note
This setting does not apply to HTML or CSS files.
Minimum File Size
To instruct WebCopy to skip downloading files below a specific size:
- From the Project Properties dialogue, select the Limits category
- Check the Minimum file size option
- Enter a file size, in KiB (1024 bytes)
Note
This setting does not apply to HTML or CSS files.
See Also
Configuring the Crawler
Working with local files
- Extracting inline data
- Remapping extensions
- Remapping local files
- Updating local time stamps
- Using query string parameters in local filenames
Controlling the crawl
- Content types
- Crawling multiple URLs
- Crawling outside the base URL
- Downloading all resources
- Including additional domains
- Including sub and sibling domains
- Limiting downloads by file count
- Limiting scans by depth
- Limiting scans by distance
- Scanning data attributes
- Setting speed limits
- Working with Rules
JavaScript
Security
- Crawling private areas
- Manually logging into a website
- TLS/SSL certificate options
- Working with Forms
- Working with Passwords
Modifying URLs
Creating a site map
Advanced
- Aborting the crawl using HTTP status codes
- Cookies
- Defining custom headers
- HEAD vs GET for preliminary requests
- HTTP Compression
- Origin reports
- Redirects
- Saving link data in a Crawler Project
- Setting the web page language
- Specifying a User Agent
- Specifying accepted content types
- Using Keep-Alive