Crawling can be left to try and scan as much of a website as it can access, or you can be limit it to only crawl to a certain depth.
Scan depth checks only apply to the main domain being crawled
WebCopy determines the depth of a URL by looking at the number of path components it is made up of, excluding the document name if possible.
URL | Depth |
---|---|
http://www.example.com/ | 0 |
http://www.example.com/index.html | 0 |
http://www.example.com/products/ | 1 |
http://www.example.com/products/index.html | 1 |
http://www.example.com/products/webcopy | 2 |
Scan depth is taken from base domain, not the starting address