Diji1 has perfectly answered your question Unfortunately I believe it's against the forum rules to talk about other forums here, so I cannot do that. It's not too hard to find, though.
Thank you for the compliments Will keep it updated as often as I can. Happy scraping! I'm sure you'll get tons of untapped sites, especially with the other languages.
Yes, that is on my to do list. However, not many platforms have a "foreign" version so it may or may not be quite limited. I'll make sure I'll get some out after I add a few more new languages (which are coming very soon).
@Yurium: Just using articles SB crashed loading all keywords. (memory error)
What I did is split the articles footprints and (EN) keywords into 10 files, each the same number of entries.
Then I opened 10 instances of Scrapebox and loaded the keywords-1 list, added the footprints-1 list and scraped. Rinse and repeat for the other 9.
Whilst Scrapebox is a "Swiss Army Tool" for many things it's crap with large lists and URL scraping. I'd recommend paying the money for GSraper - does the same thing as above only at speeds that have to be seen to be believed and using much larger lists. It leaves Scrapebox in the dust as far as scraping URLS.
@OnkelMicha Some of our lists have millions of keywords, so importing and scraping ALL of them at one go would cause certain programs to crash. Try splitting them up into lists of 10 and scraping them one by one as what Diji1 recommended. That would be a lot more efficient than scraping an entire list at one go.
@gtsurfs Yes, you can, but that would be a lot less efficient than using a scraper as SER's main purpose is to create links, not scrape. I highly recommend you to get a scraper to use along with SER. If you know what you're doing, you can earn the money back easily. If you really are tight on budget, then I guess you could do that. Just load a list up when creating a project.
@AlexR My footprint list consists of the ones I've managed to collect over the years (my personal footprint list) as well as the ones in SER. There's a lot more footprints than the ones included in SER such as those for edu\gov sites.
@Cherub .... I imported some of the foreign language KW lists into Gscraper and even after converting them to UTF-8 I am getting alot of garbled characters. Any suggestions? Thx
@sweeppicker Why did you convert them to UTF-8? I've already encoded non ISO languages for scraping on Google so you can just load those up directly, they are equivalent versions The reason why you are seeing garbled characters is because your text editor does not have the proper language pack to utilize.
@joland I process all orders manually to prevent any leaking of the lists. Please be patient, as promised it'll be sent out WITHIN 24 hours
@FuryKyle - I thought I'd post this question here so that others who are using GSA SER as their scraper could benefit. I purchased the lifetime package and I'm looking at the keyword list folder that has 10 text files containing keywords in different languages. How many keywords are in each language file and how do you recommend splitting these text files into groups of 100K keywords?
Sorry for the noob questions. Thanks for your help!
@RayBan According to my records, the list was sent to the email address you provided in the form. Could you check your junk mail and let me know? I'll have it resent if you still can't find it.
@gtsurfs You can load them up to the KW section of GSA to find out how many keywords are there for any list you choose. I personally use SB to split up lists, but you can use free programs out there such as the one already mentioned in one of the pages before this. Here's the URL - http://sourceforge.net/projects/textwedge/
OnkelMicha Most text editors can handle them by default, so there's no need to install or add anything extra. It won't freeze if you don't have the pack installed, you simply get a bunch of jarbled text with squares everywhere. Use the encoded versions to scrape on Google for those non ISO encoded languages. That's why they're there
Ordered, just to see if there are some nice footprints to add to our collection. Also would be cool, if you can shoot us an invoice for the purchase to our mail (for accounting purposes)?
@andy1024 There's absolutely no need to scrape for more footprints. There's almost a billion in my list, trust me, it's way more than enough. Just merge the platform footprints with the language ones and you're good to go.
@DonAntonio I'm not sure why that happened, but I've dropped you a PM.
Comments
I would like to see Korean KW list if possible.
Don't make me wait so long..
Bought your list this morning, still waiting to receive it?
What I did is split the articles footprints and (EN) keywords into 10 files, each the same number of entries.
Then I opened 10 instances of Scrapebox and loaded the keywords-1 list, added the footprints-1 list and scraped. Rinse and repeat for the other 9.
Whilst Scrapebox is a "Swiss Army Tool" for many things it's crap with large lists and URL scraping. I'd recommend paying the money for GSraper - does the same thing as above only at speeds that have to be seen to be believed and using much larger lists. It leaves Scrapebox in the dust as far as scraping URLS.
@OnkelMicha
Some of our lists have millions of keywords, so importing and scraping ALL of them at one go would cause certain programs to crash. Try splitting them up into lists of 10 and scraping them one by one as what Diji1 recommended. That would be a lot more efficient than scraping an entire list at one go.
@gtsurfs
Yes, you can, but that would be a lot less efficient than using a scraper as SER's main purpose is to create links, not scrape. I highly recommend you to get a scraper to use along with SER. If you know what you're doing, you can earn the money back easily. If you really are tight on budget, then I guess you could do that. Just load a list up when creating a project.
@AlexR
My footprint list consists of the ones I've managed to collect over the years (my personal footprint list) as well as the ones in SER. There's a lot more footprints than the ones included in SER such as those for edu\gov sites.
Why did you convert them to UTF-8? I've already encoded non ISO languages for scraping on Google so you can just load those up directly, they are equivalent versions
@joland
I process all orders manually to prevent any leaking of the lists. Please be patient, as promised it'll be sent out WITHIN 24 hours
thank you.
An invoice? For what?
Just sent out the keyword lists to all buyers. Will add a Korean keywords list very soon.
Sorry for the noob questions. Thanks for your help!
One thing I'm not sure you answered, if I don't jave the language packs installed on my computer so can it not handle them?
I tried to load some of the asian stuff and it freezes as well.
@Eagleflux
Just sent it.
@RayBan
According to my records, the list was sent to the email address you provided in the form. Could you check your junk mail and let me know? I'll have it resent if you still can't find it.
@gtsurfs
You can load them up to the KW section of GSA to find out how many keywords are there for any list you choose. I personally use SB to split up lists, but you can use free programs out there such as the one already mentioned in one of the pages before this. Here's the URL - http://sourceforge.net/projects/textwedge/
OnkelMicha
Most text editors can handle them by default, so there's no need to install or add anything extra. It won't freeze if you don't have the pack installed, you simply get a bunch of jarbled text with squares everywhere. Use the encoded versions to scrape on Google for those non ISO encoded languages. That's why they're there
Please send ******olt@gmail.com
It's pretty much plug and play. Merge a keyword language list with footprints and you're good to go.
There's absolutely no need to scrape for more footprints. There's almost a billion in my list, trust me, it's way more than enough. Just merge the platform footprints with the language ones and you're good to go.
@DonAntonio
I'm not sure why that happened, but I've dropped you a PM.