Start crawling

Modified on Sun, 19 Nov 2023 at 11:24 PM

1. Default settings by tabs:

1.1. General settings

1.2. Advanced settings

1.3. User Agent settings

1.4. Restrictions settings

2. Crawling:

2.1. Crawling process

2.2. Changes in settings during crawling

2.3. Pause and resume crawling

2.4. Data analysis

3. Saving the project and exporting the results


Netpeak Spider provides a wide range of custom settings and templates. In this article, you’ll learn how to start crawling quickly and how to configure Netpeak Spider.

 

1. Default settings

The default settings template is suitable for general analysis of the site status and standard SEO tasks. It affects such settings tabs:

  • General
  • Advanced
  • Virtual robots.txt
  • Scraping
  • User Agent
  • Restrictions 
  • Rules

You can always return to default template by choosing it from a template dropdown menu in the upper settings window and it will be applied to the program settings tabs listed above.


Default Settings

1.1. General settings

By default the ‘General‘ settings tab contains: 


General settings


  • Crawling in 10 threads → the optimal number of threads that most sites can handle. You can always change this number. Increase in the number of threads will speed up the crawling process, and also increase the server load, when the decrease in the number of threads will reduce the load.

Note that when you crawl a site with a large number of threads, the server on which the site is located, may not be able to withstand the load and begin to return an incorrect server response code or temporarily stop responding to requests. 

  • No delay between requests → the time interval in milliseconds which the crawler accesses the server. For load-sensitive servers or protected sites, we recommend setting a certain value so as not to overload the server, and imitate user behavior.
  • Maximum waiting time for server response – 30 000 milliseconds. If the server did not respond during this time, the ‘Timeout‘ error will be shown in the ‘Status code‘ field, and the page will be considered broken.
  • Enable JavaScript rendering → by default this feature is disabled. The feature allows you to crawl a URL with JavaScript code execution. This is necessary when part or all of the content is generated using JS frameworks and CSR (Client Side Rendering) technology. As the statistics show, the number of sites using CSR is much smaller than the number of sites that do not use this technology.
  • Crawl all subdomains → allows considering subdomains as a part of the current website. All subdomains will be crawled and links to them will be considered internal.
  • Check images → this feature allows you to collect common SEO parameters of images and detect ‘Broken images‘ and ‘Max image size‘ issues.
  • Check PDF → allows collecting common SEO parameters.
  • Check other MIME types → allows collecting common SEO parameters for RSS feeds, archives, documents, video and audio files.


1.2. Advanced settings

By default the ‘Advanced‘ settings tab contains: 


Advanced settings

1. All indexing instructions, except Canonical, are disabled. So you don't risk missing important pages in the general analysis. 


2. Crawl URLs from the tag:

  • Hreflang → allows you to follow the links from hreflang attribute in the tag of the section of HTTP response header ‘Link: rel=”alternate”‘. This is necessary to ensure the correct settings of the hreflang attribute. Note that Netpeak Spider will crawl URLs in hreflang (both internal and external) despite ‘Crawl all external links‘ settings.  
  • AMP HTML → allows you to follow and check links where AMP technology is introduced.
  • Other → allows you to check other URLs from the tags in the section. Note that this setting ignores the rel=”stylesheet” (CSS), hreflang, rel=”next/prev”, and rel=”amphtml” directives because they are covered by other settings. 


3. Cookie file is allowed. Check in case the analyzed website is closed for all requests without a cookie file. Also, all requests will be tracked within one session. Otherwise, every new request will generate a new session.


1.3. User Agent

User Agent

1.4 Restriction settings

Restriction settings

Max number of redirects – 5. In this case, Netpeak Spider will follow four redirects and the fifth one would be considered as an error. In the ‘Issue restrictions’ field you can adjust values for some parameters on your own to flexibly configure the program specifically for your current tasks. It was also based on the differences in requirements for each search engine. 


2. Crawling

2.1. Crawling process

To start crawling, add the site address to the ‘Initial URL’ bar and press the ‘Start’ button.

Start Crawling

On the ‘Dashboard‘ tab, the program shows the current information about the crawling process and applied settings. 


Dashboard

2.2. Changes in settings during crawling

Only the number of threads can be changed directly during the crawling process. The changes will take effect immediately after you click ‘OK‘. 


To change other settings (in the ‘Settings‘ of the main menu or on the side panel of the ‘Settings‘ tab), you need to suspend crawling by pressing the ‘Pause‘ button. Such changes will take effect when you resume crawling (‘Start‘ button) for uncrawled URLs. The results obtained before will remain unchanged. 


The ‘Restart‘ button on the control panel will restore the results and resume crawling to get data for all URLs according to the new settings. 


2.3. Pause and resume crawling

Crawling can be stopped at any stage by using the ‘Pause‘ button. To continue crawling (e.g. after changing settings), hit ‘Start‘, and it will start from where it had stopped. 


2.4. Data analysis

When the crawling is finished (or paused), the data analysis will automatically begin: the program will detect duplicates, incoming links, Canonical chains and calculate internal PageRank, counting the issues according to configurations of parameters in the ‘Restrictions‘ settings tab. Also the program gets data from Google Search Console, Google Analytics, Yandex.Metrika, and queries from Google Search Console which rank in this search engine, if before crawling the corresponding parameters were enabled on the sidebar. 


You can interrupt the analysis anytime, and resume in the ‘Analysis‘ menu on a control panel.


Data analysis

3. Saving the project and exporting the results 

When the crawling is finished or stopped, you can save the project. In order to save the project, select ‘Save... ‘ (Ctrl+S), or ‘Quick Save‘ (Ctrl+Shift+S) on the control panel in the ‘Project‘ dropdown menu.  

Saving a project via ‘Save... ‘ allows you to specify a directory and create a custom file name. When saving via ‘Quick Save‘, the project would be saved in the default folder (usually the Desktop one), and under the default name.  


Saving the project and exporting the results

If necessary, continue crawling later or on another device:

1. Save the project.

2. Open a saved project in the ‘Project‘ → ‘Recent projects‘ or ‘Open‘ (Ctrl+O) on the control panel.


Open the project

To export all or part of the results for further work you can:

  • Choose the necessary report template in the ‘Export‘ menu of the control panel or in the main menu. Download current results from ‘Dashboard’ via the ‘Export... ‘ button in the upper left corner. This method is useful when you apply filters, or work within database tables.

Export



Was this article helpful?

That’s Great!

Thank you for your feedback

Sorry! We couldn't be helpful

Thank you for your feedback

Let us know how can we improve this article!

Select atleast one of the reasons
CAPTCHA verification is required.

Feedback sent

We appreciate your effort and will try to fix the article