Dirty verified lists kill your LPM...
Satans_Apprentice
SERLists.com
in Need Help
Over time, your verified site lists will have links that go bad. If you are building links from your site lists and you have bad links, it will kill your LPM in a big way. A dirty site list is the difference between 20 LPM AND 100 LPM. Cleaning out your verified list will pay big dividends. The big problem occurs when you delete projects or directly import site lists. The cleanup functions in SER are tied to actual links in projects. If those actual links are deleted(deleted project), or never existed in the first place (imported site list) SER can't delete them. It appears that failed downloads hammer your LPM. If you are getting a lot of download failed, you probably have a dirty site list.
Tagged:
Comments
Here's an example: you buy a verified list (which I would recommend) and import it to your verified site list. A good portion of those URLs will be bad, and you won't be able to build links to them. They will now sit in your list, generating "download failed" or "no engine match" messages which kill your LPM. Since links were never built in your projects with those URLs, they can't be cleaned using the built in tools.
Check out my other thread on the right way to import site lists for how to solve the problem.
SER has to check the link is present on the page to confirm it, how can it check that if the project that created it no longer exists? SER doesn't even know what link to look for in that scenario.
All you need to do is main options > advanced > tools > Clean-Up (check and remove non working)
Then do all your global lists
It checks links are live and a recognised engine. If the engine has changed, it sorts that as well
It does not matter if you have killed a project that put the links into the lists. it still checks the links
Over time, web sites will die or change what type of platform is on there
Im looking at about 1 in 3 sites being dead. No idea what percentage has changed engine
Other options where list are compared or links are compared against past lists are apparently not a good idea because of the memory usage that would be required to load the whole list, according to Sven.
It does not run through the projects
It runs through the global lists
Im checking over 21 million links from my global lists at present. That's just submitted and verified global lists
Just doing everything. Engines that are no longer used anywhere etc inc in those lists
Once LPM goes low you will need to start from the beginning, brand new list - that's why it's ideal if you have another installation to get a new verified list ready for when the current one has died (all projects have used all available links in that list).
Does that makes sense?
- If you let SER scrape, then it scrapes random url's for random projects - Not very efficient.
- If you post from verified list that is better because all projects will use url's that you know SER can post to. Much more efficient, but run SER for 6 months and you will find your verified list becomes full of dead links (Clean up only removes url's that SER can not identify - Identify does not mean postable - Like if you run "Sort and Identify" function, SER can not post to most of those url's even though it identified them)
- So maximum speed is gained from using a freshly verified list for all projects. That's all it is really.
But it's only really viable if you are doing big numbers with multiple installations.
Here's a real-life example:
- List scraped with SB and tested on a VPS
- After one month collect all verified's from VPS and use for real projects on dedicated server. LPM will fly because all links are freshly verified (i.e. working links)
- Start new verified list on VPS and after one month remove current list from dedicated server and put this new verified list from VPS in it's place.
If you don't believe it is true go check your verified list, open the contextual engines and randomly check some links. You will see so many that redirect you to another page (article removed etc). SER will identify those url's as good during clean up because the engine is still the same.
But in reality you won't get a link there or it will be deleted soon after verification.
EDIT:
Here's an example: I had 36,000 unique domains from verified contextual links, i started a new project and imported all of those links. First run i got 1200 verified, 1200 from 36000! I'm running them again to hopefully get a few more, but the rest are dead so no point keep trying to post to them.
If this is the case, then you dont have to worry because that will clean out your verified list as it will remove after you run it and it doesnt work (so other projects wont keep trying to submit to).