Crawling can be left to try and scan as much of a website as it can access, or you can be limit it to only crawl to a certain depth.


Scan depth checks only apply to the main domain being crawled

How does WebCopy determine depth?

WebCopy determines the depth of a URL by looking at the number of path components it is made up of, excluding the document name if possible.


Configuring a scan depth

  1. From the Project Properties dialogue, select the General category
  2. Check the Limit crawl depth option
  3. Enter the maximum level that WebCopy will scan


Scan depth is taken from base domain, not the starting address

See Also

Configuring the Crawler

Working with local files

Controlling the crawl



Modifying URLs

Creating a site map


Deprecated features

© 2010-2024 Cyotek Ltd. All Rights Reserved.
Documentation version 1.10 (buildref #185.15779), last modified 2024-03-31. Generated 2024-03-31 14:04 using Cyotek HelpWrite Professional version 6.19.1