Questions About Filtering Regarding Imported Scraped Target URLs Into a Project
I have a scraped list of URLs which are all relevant to my keywords and those are the only URLs I want to use to get links. I right clicked on my project and imported target urls from the file I had all the scraped URLs on. I've unchecked every single "search engines to use" engine, unchecked the "use URLs from global site lists if enabled", and unchecked the "analyze and post to competitor's backlinks" box. I think I did that correctly to ensure I'm only using my imported URLs to create links. Let me know if I forgot anything.
ANYWAY, I have 3 questions about how GSA SER interprets those URLs after I've imported them:
1 - I'm not sure what kinds of platforms/engines the scraped URLs are/extend to, so should I just go ahead and check every single platform under the "where to submit" main interface in the project and subsequently fill out every single box, basically so that just in case I have any URLs which are that type, GSA SER will have that info? GSA SER figures out what platforms these URLs are as it goes, correct? I've never imported a URL list so I have no idea.
2 - If I only want to post to PR 1 and above for my Tier 1 campaign (which this project is), can I go into the "Filter URLs" section of Options and set that accordingly and it will go ahead and skip the PR 0 and N/A URLs? If so, what does it do with the PR 0 and N/A URLs? Will it leave them in the "remaining target URL's" so that I can go ahead and use them for the second tier?
3 - Lastly, this is a bit less related but when I scrape URLs via Scrapebox to give to GSA SER, should I trim them to root or leave them as the full URL before I import them to GSA SER? I don't know if I'm ruining GSA SER's ability to create a link if I trim it to root... I really don't know how GSA SER functions in creating that link. Any insight here or on any of these questions would be huge.
Thanks so much fellow GSA SERers!
ANYWAY, I have 3 questions about how GSA SER interprets those URLs after I've imported them:
1 - I'm not sure what kinds of platforms/engines the scraped URLs are/extend to, so should I just go ahead and check every single platform under the "where to submit" main interface in the project and subsequently fill out every single box, basically so that just in case I have any URLs which are that type, GSA SER will have that info? GSA SER figures out what platforms these URLs are as it goes, correct? I've never imported a URL list so I have no idea.
2 - If I only want to post to PR 1 and above for my Tier 1 campaign (which this project is), can I go into the "Filter URLs" section of Options and set that accordingly and it will go ahead and skip the PR 0 and N/A URLs? If so, what does it do with the PR 0 and N/A URLs? Will it leave them in the "remaining target URL's" so that I can go ahead and use them for the second tier?
3 - Lastly, this is a bit less related but when I scrape URLs via Scrapebox to give to GSA SER, should I trim them to root or leave them as the full URL before I import them to GSA SER? I don't know if I'm ruining GSA SER's ability to create a link if I trim it to root... I really don't know how GSA SER functions in creating that link. Any insight here or on any of these questions would be huge.
Thanks so much fellow GSA SERers!
Comments
1A - What would be the reasoning behind unchecking platforms like url shortner and web 2.0's? Are they generally spammy and best kept for a lower tier?
1B - My reasoning for wanting to import directly into the project vs. "identify and sort in" is that if I import directly into the project then I'll have full control that I'm just building links from that list, whereas if I "identify and sort in", then I'll have to use the global "identify" list in my project to get those links, but at the same time I'll be SIMULTANEOUSLY getting other identified targets which might be unrelated/from other keywords which already existed in the global identify list. Am I correct with this thought process? So basically if I import directly into the project, there won't be any mixing with other identified global targets for that project.
I just saw that you said "you can also save them a custom file if needed" which probably means you get what I'm saying. So does that include sorting them first then letting you save to a custom file? How do I save them as a custom file out of curiosity?
2 - Got it, basically just import the exact same scraped list file into my second tier but with a filter to ONLY build links from URLs with PR 0 or N/A.
3 - Thanks for the info, I just had a thought though. I guess if I trim to root, GSA SER will get the PR of the domain as a whole whereas if I leave the URLs as is, it'll collect the PR of that specific page on the domain, correct? That would give drastically different results if I was using the PR filter in my project, wouldn't it? That's something to think about if that's the case.
Thanks again!
Otherwise:
1 - Ah I got you, yeah I was blanking but yes it's simply an issue of quality. Good call.
2 - And there it is; I never tried doing the "identify platform and sort in" function. I guess that's probably the way to do it if I want to identify the links ahead of time to know what they are.
3 - True, that's another way of doing it. Good call.
4 - Totally forgot about that feature; that's exactly what I need. I can leave the URLs as is then decide in SER which PR I want to look at, great.
Thanks so much for the clarity on these questions!
I was about to say it'd be great if there was a tool to split them all up by category but then I saw the GSA Pi. Good tool, I'm thinking about getting it if I keep up with scraping my own links.
I've spent the last couple weeks harvesting lots of URLs using footprints from both GSA SER and elsewhere online and want to create links using GSA SER to my sites using these harvested URLs, but they're obviously all mixed together and not sorted by platform. I guess I don't know if GSA PI is necessary or if I can just tick off the platforms I want to create links with by using GSA SER, load in my thousands of unsorted URLs, and GSA SER will just figure out which URLs will work and which won't.
Any insight on this would be great, thanks!