Skip to content

Loaded URLs from site lists too slow, how can I optimize it?

Hello all.

I am having a small issue at the moment, I scrapped around.. 2.6million urls that took a while.. and optimized it such as removing duplicates, etc etc..

Now, I only scrapped contextual platforms, articles, web 2's and that's it.. then I used the identify and sort option and it all loaded into the identified site list files, 2.6m worth..

I setup a tiered link campaign of three tiers..

It's been running for the past.. 3 days, and it's extremely slow.. on my first three tiers, the secondary are blog comments and whatnot that isn't using global site list because it doesn't have any links in there for those platforms..

What can I do to optimize this, for my first tiers that is actively searching those global site lists they're really really slow, showing loaded #/### urls from site list.. what can I do to make this goes faster.. it has nothing to do with the VPS for sure..


I turned off all of the other tiers, and right now it's only running about 1-9 threads at the most for the first tier..
I am using private proxies and have had no problems with them.


  • SvenSven
    why not importing your urls directly to the project?
  • How should I do that now that I have already loaded them into the identified site list? Should I do an export of all the identified then then load them directly into the project?
  • s4nt0ss4nt0s Houston, Texas
    edited December 2015
    @ibrazilian2 - You can import your site list into the project by right clicking on the project > import target urls > from site lists > identified. 
    Thanked by 1giggioman
  • Thanks, things are running much smoother now, all URLs are being parsed.. although can't say much are being verified but submitted.. the numbers are much higher now.
Sign In or Register to comment.