NON contextual link list (Spam links)?
Hi guys,
I joined the GSA family a couple of days ago and everything worked just great since then, but now I hit sort of a bump in the road and I'm not quite sure how to approach this problem and was hoping you guys could give me some input to solve it.
The campaign I'm talking about is a german one which already limits the amount of scrapable niche relevant links by a lot. However, I was able to scrape enough links (at least I guess) to start a tiered quality campaign. This means that my tiers will all consist of contextual quality posts and the only difference between the tiers is going to be the quality (OBLs, PR, etc.).
So what I need now is to scrape for URLs for my secondary links (the links coming in 'from the side' of the link pyramid). They will mostly consist of blog comments, guestbooks, etc. so really spammy stuff.
The problem I encountered is that I literally have no idea how to scrape for these links. I'd appreciate it if you guys could help me out on this one.
Regards.
I joined the GSA family a couple of days ago and everything worked just great since then, but now I hit sort of a bump in the road and I'm not quite sure how to approach this problem and was hoping you guys could give me some input to solve it.
The campaign I'm talking about is a german one which already limits the amount of scrapable niche relevant links by a lot. However, I was able to scrape enough links (at least I guess) to start a tiered quality campaign. This means that my tiers will all consist of contextual quality posts and the only difference between the tiers is going to be the quality (OBLs, PR, etc.).
So what I need now is to scrape for URLs for my secondary links (the links coming in 'from the side' of the link pyramid). They will mostly consist of blog comments, guestbooks, etc. so really spammy stuff.
The problem I encountered is that I literally have no idea how to scrape for these links. I'd appreciate it if you guys could help me out on this one.
Regards.
Comments
Thanks anyways.
One question though - if you were to create spam links to your contextual tiered links, would you even bother combining the footprints with your (broad) KW? I'm just wondering, since they don't neccessarily have to be niche related.
He told me that I shouldn't bother scraping for URLs using SER's built-in scraper. Instead I should either invest in good lists (30-40$) or buy Scrapebox/Hrefer.
Now, I was under the impression that SER's built in scraper isn't that bad - at least if you're scraping for niche relevant sites.
However, I am nowhere near 5k+ links and I think this is at least what I need to be able to spam my lower tiers with secondary links.
Do I absolutely need a scraping tool like Scrapebox? As of right now I could afford it, but I don't feel like wasting money on something that I don't absolutely need right now.
What's your oppinion?
@fakenickahl These are exactly the two things I've heard of. 1. Their proxy service sucks and 2. You'll need a huge amount of proxies.
@ImLazy @fakenickahl Let's say I'll try free GScraper and do some basic scraping for low/medium comp projects, will 20-30 proxies suffice? Because that's pretty much how many I have access to.
Also, comparing free GScraper to Scrapebox (quality wise) any difference? I'm only talking about the scraping here, not the additional features, like KW suggester, etc.
Thanks guys, you're very helpful.
@fakenickahl Wow.. didn't expect an answer like that. In all the videos/guides/threads/.. I've found that people mostly use something between 10-50 proxies. How'd you define 'shitload'?
Please keep in mind that I don't intend to scrape 24/7 for the next month and create the biggest list in the history of scraping. I'm rather looking for a decent list that can be used for secondary spam links to my tiered link structure.
I feel like there's not enough info about scraping. I've watched pretty much every tutorial video posted on this forum and a couple of others, but I still don't really know how to properly scrape for KWs.
I read about people scraping lists with 100-500k verified links in it, but when I scrape for related KWs using SER Scraper I end up with 2-4k.
Right now I'm just punchin in a couple of my KWs + variations and LSIs, don't select foot prints, hit 'Scrape' and then I hope for the best. I assume that's not the proper way, is it?
- Take a set of 5k or 10k keywords (save these and use for all scrapes).
- Scrape only 1 footprint with your keywords. Record the total URLs scraped.
- Set up 1 project in SER for only that 1 scrape. Let it run until it has completely finished. Record how many verified URLs.
Now you have total URLs scraped and total URLs verified for one footprint. With some simple maths you can get a % verified.
Repeat for all footprints. It's a long process but if you want to find the best footprints it works.
Thanks.
I have a huge list of keywords, 10 million or more. Lots of languages. So i split them into chunks of 10k or whatever and scrape like that. I have footprints organised into batches too.
Sent you a PM.
One general question about foot prints and KWs:
If I use the SER scraper tool and ONLY use my KW list (no foot prints) will this make the the tool scrape every website possible? And is there a difference between doing it like that and manually adding every footprint from the list + merge them with the KWs from a text file?
Because right now I don't quite understand if the foot prints can be used as filters to scrape for particular platforms, or if they're neccessary for you to get results.
The problem is SER won't be able to post to most of them so you need to refine your scrape with footprints.