After you have scanned a website, you can use WebCopy's built in tools for viewing link data, or export them to CSV for use in other applications.
Tip
As well as the export process below, many lists also support CSV export - right click inside a list, or when the list has focus, press either the Menu key or Shift + F10 and choose Export from the context menu.
Exporting to CSV
- From the File menu, click Export...
- Enter or select the name of the file to export to
- Click Export to export the links list with default options
Customising the export
You can optionally customise the export
- From the Export CSV dialogue, click Advanced to show optional settings
- To prevent the header row with column names from being included, uncheck the Include header row option
- To use a separator other than a comma, enter the character to use in the Separator field
- To select which columns are present in the export, check or uncheck entries in the Columns list
Column Reference
Column | Description |
---|---|
URL | The remote URL |
Type | The content type |
Size | The content size |
Title | The document title |
Description | The document description |
HTTP Status | The HTTP status reported by the server when WebCopy probed the URL |
HTTP Description | The description of the HTTP Status |
Skip Reason | The reason why WebCopy skipped this URL, if applicable |
Status | The WebCopy status for this URL |
Source | The source of the URL |
Last Modified | The date and time the remote resource was last modified, if provided by the remote host |
Last Downloaded | The date and time WebCopy last downloaded this file |
Distance | The distance of the URL from the root URL of the crawler project |
Scan Index | The scan index that WebCopy last processed this URL |
Excluded | Specifies if WebCopy is excluding this URL |
Redirects To | The destination address if this URL redirects to another |
Error | Any error information WebCopy encountered trying to process this URL |
ID | An internal identifier for this URL. |
Relationship | The relationship between this URL and the root URL of the crawler project |
Depth | The depth of the URL from the remote host |