Any reason to test both an inner page and homepage in current version?
googlealchemist
Anywhere I want
I dont know the original post I read it on, but it was something about how gsa processes urls when you import a new harvest of domains/urls to test to see if the software can get a link from it.
if i import domain.com/page123 it will try to look for an opportunity to place a link on that specific page, if it cant, it will navigate automatically to the root and look for an opportunity there before moving on to the next site?
if i just import the root directly, it will look for a link opportunity on the homepage...a registration/login/submit article type option or comment or guestbook form etc...but if there is no link opportunity directly there... then it will navigate to an inner page, based on the cms, to test for a link opportunity there?
and that someone had also mentioned i might get some additional submitted/verified urls if i took any inner pages i had uploaded but failed to get a link from, and stripped them down to the root and ran them thru again that way
am i remembering this correctly?
either way...with the current version of gsa which might make that a moot point regardless...whats the best practice for this?
i have a ton of urls that are inner pages and i have a ton of urls that are just the root domains from various scrapes.
is it worth the resources for me to just strip the whole list down to the roots, add that back to my master list, and then get rid of the duplicates just on a url vs domain level?
or is it better now overall to just strip the whole list do to the root domains, and only upload that list...and not bother with any of the specific inner urls?
or maybe it doesnt matter in the slightest and i should just upload my mix of roots/inner urls that have been de dupplicated on a domain level
just trying to be as efficient as possible here when were talking about processing tens of millions of domains
thanks
if i import domain.com/page123 it will try to look for an opportunity to place a link on that specific page, if it cant, it will navigate automatically to the root and look for an opportunity there before moving on to the next site?
if i just import the root directly, it will look for a link opportunity on the homepage...a registration/login/submit article type option or comment or guestbook form etc...but if there is no link opportunity directly there... then it will navigate to an inner page, based on the cms, to test for a link opportunity there?
and that someone had also mentioned i might get some additional submitted/verified urls if i took any inner pages i had uploaded but failed to get a link from, and stripped them down to the root and ran them thru again that way
am i remembering this correctly?
either way...with the current version of gsa which might make that a moot point regardless...whats the best practice for this?
i have a ton of urls that are inner pages and i have a ton of urls that are just the root domains from various scrapes.
is it worth the resources for me to just strip the whole list down to the roots, add that back to my master list, and then get rid of the duplicates just on a url vs domain level?
or is it better now overall to just strip the whole list do to the root domains, and only upload that list...and not bother with any of the specific inner urls?
or maybe it doesnt matter in the slightest and i should just upload my mix of roots/inner urls that have been de dupplicated on a domain level
just trying to be as efficient as possible here when were talking about processing tens of millions of domains
thanks
Comments