there is no ftp server in SER of course. You have to scrape proxies on one PC (it's a one PC license) and provide the found proxies to all SER instances if you do not plan to buy Proxy Scraper for all your running SER licenses. This proxy can be provided by automatic export to a webserver, by email, ftp server or whatsoever.
hi guys, i just bought proxy scraper and this are my settings:
*Activate internal proxy server is checked "on"
Use proxies with tag: I only select "google" because i want to get proxies which can solve recaptchas.
From my understand, for internal proxy server, this is meant for use if i have gsa ser (or any other software) running on the same physical machine as proxy scraper?
Once I have this set, i went to GSA ser, proxies settings and i set these:
Check "automatically search for new proxies every X mins". And i entered the ip 127.0.0.1 which is where gsa ser accesses the new proxies every X mins?
I also checked "Disable proxies when detected to be down" for both public and private
With the above settings I am assuming this gsa ser copy will auto get the proxies from proxy scraper every X mins?
Once the above is done, I went to take a look at captcha breaker and to be honest im getting lotsa hard to solve recaptchas from the proxies. I have already set "google" proxies as the filter. Do I need such a filter if i am not scraping?
if you use the Proxy Scraper's internal proxy inside of SER as an only proxy solution, you should not use the options to disable proxies. This would cause problems as 127.0.0.1 actually are all proxies in the list of Proxy Scraper. It picks one randomly for each request sent to it.
@sven it seems like the public proxies i find die really fast, they are reported as ok in scraper but when i import into gsa ser, almost 90% are dead using the proxy tester. just curious how frequently show gsa ser pull the latest proxy from the ftp (supplied by proxy scraper)?
are we talking about 5-10 mins interval? or is 1-2 hours ok?
@hardcorenuker i experienced this problem too in all GSA product when i want to use this scrapped ip adresses... When i import to ScrapeBox this ip-s 99,99% are always dead. I use only fresh scrapped proxyes, not old ones, anonyms, and only fast ones.
You must know, the public proxies are not always reliable, and this scraper are a very new gsa product, but sven update the software very often, and we must believe him, that this scrapper will be a reliable tool for scrapping in the future...
Just bought GSA Proxy scraper, I'm new to scraping proxies and would appreciate the manual when is it going to be available? Meanwhile can anyone offer advice on setting it up to run automatically with GSA SER
@sven when i tested again google search in gsa ser, most of them came up red. but just mins ago they were reported as ok in proxy scraper. was wondering if im doing anything wrong?
@hardcorenuker the tests are almost identical in Search Engine Ranker and Proxy Scraper. What happens if you test them again in the Scraper once they are red in SER?
@GreyGable Im working hard on writing something up for the scraper. But it's not really a complicated program. All default options should be fine for most users.
Well you can already use it in it like in any other tools that offer proxy input.
a) Enable the internal proxy server in Proxy Scraper -> use it in the software 127.0.0.1:8080 (unless you changed it). This will use a random proxy from the list added in Proxy Scraper
b) Enable auto-export of proxies to a text file. In GSA tools you can set things up to import them in intervals.
Oh I checked out the auto export is looking good too. just can I send proxies to the PS to check it out and get back the result? and when auto upload an html file every 60 min thats always get refreshed and rechecked the ips? and there is any discount if I already own more of your products ? :P
I'd like to know if there is a way to exclude tags when exporting? For example I want to export all kinds of proxies except those with google tag. I want to use those with google tag exclusively for other projects. Is this currently possible?
@spiritfly - If you go to settings > export options, then setup an export job, you will be prompted with this window on the last step which lets you choose the specific proxies you want: http://i.imgur.com/WGV4R0t.png
You can setup multiple export jobs using different options so for one project you can export all proxies, for another you can export just google, etc.
Sorry to be so thick but can't find out how to import proxies from Proxy Scraper to SER. I've tried Configure proxies > Add Proxy > Import from file but I get a "No Proxies found" when opening my file from Proxy Scraper, is this because SER asks for a file with host:port:login:password? How do i get SER to import the proxy.txt file from PS every X minutes?
@spiritfly you are right, the option to export will export proxies with at least one of the tags. If you want to exclude proxies with a certain tag then I have to add another option. This will be in next update.
>ability to exclude proxies with a domain for exports
thats the second checkbox on the filter dialog.
>can i filter / delete the text included ip-s? where can i find that function?
I don't understand what you mean by that. On the same export filter dialog oyu can set what to export (e.g. google only) but now you can also define to export everything BUT e.g. google.
@s4nt0s I know about that feature and it's extremely helpful and well thought. But I'm afraid there isn't a way to exclude tags. For example, export all proxies EXCEPT google tagged. Since there is no way to tag all proxies, but google passed ones, there is no way to export them without google passed.
And this is very important because I want to use google proxies only for google scraping and the rest of them either for bing or whatever. But I definitely don't want to burn google passed proxies for bing and other things.
@sven it would be cool to include a "loader" like freecap or widecap, where you can assign rules for specific programs (proxychaining etc... no seo purpose). this would open a new salesroad for non-seo-markets and it would open the gsa proxy-server function for programs that doesn't have proxy support. with a loader you can tunnel all programs to a local port or a set of proxies even if they doesn't support proxies.
When I hit the Add button and add from clipboard a set of proxies, it goes through and scans them and then when finished, a popup comes up and says: Unable to save proxies to file!
But when the system is running normal ( meaning, automatcially pulling proxies itself without me hitting the Add button ), it saves them to the file just fine.
unable to save proxies is for the global list of proxies. They are saved in appdata folder. If thats not working, then something on your OS is restricting access to it.
same for that file-provider. Are you sure it has access to the file? I see it is working for me. Maybe simply start the program as admin to see if all works then.
Hello Sven, can i import Proxy Sources, in Test Version i dont find something, only one URL. Or is there no need to import sources, will GSA update the Sources? Thanks Peter
Bought the Tool, and enter Register Informations, when i enter them, the Program say Please Restart. After Restart the Tool is not registered! Please help!
Hello Sven, im a little bit disappointed about this Tool. It dosnt find not really more Proxys then the integrated Proxy Scraper in GSA Ranker. Or im doing something wrong? Peter
Comments
in options you have to add a custom test. Then you check it and all further tests are done against it.
An option to allow certain speed is not really required as you can just lower the timeout values and you skip all proxies being slower.
@Dikkill yes ther eis a interval you can set to test proxies each xyz minutes
@hardcorenuker
there is no ftp server in SER of course. You have to scrape proxies on one PC (it's a one PC license) and provide the found proxies to all SER instances if you do not plan to buy Proxy Scraper for all your running SER licenses. This proxy can be provided by automatic export to a webserver, by email, ftp server or whatsoever.
Export to ftp in proxy scrapper and then import in gsa automatically each X minutes from your source
Meanwhile can anyone offer advice on setting it up to run automatically with GSA SER
Well you can already use it in it like in any other tools that offer proxy input.
a) Enable the internal proxy server in Proxy Scraper -> use it in the software 127.0.0.1:8080 (unless you changed it). This will use a random proxy from the list added in Proxy Scraper
b) Enable auto-export of proxies to a text file. In GSA tools you can set things up to import them in intervals.
I've tried Configure proxies > Add Proxy > Import from file but I get a "No Proxies found" when opening my file from Proxy Scraper, is this because SER asks for a file with host:port:login:password?
How do i get SER to import the proxy.txt file from PS every X minutes?
>ability to exclude proxies with a domain for exports
thats the second checkbox on the filter dialog.
>can i filter / delete the text included ip-s? where can i find that function?
I don't understand what you mean by that. On the same export filter dialog oyu can set what to export (e.g. google only) but now you can also define to export everything BUT e.g. google.
And this is very important because I want to use google proxies only for google scraping and the rest of them either for bing or whatever. But I definitely don't want to burn google passed proxies for bing and other things.
However, when I manually export the list from the button in the main interface, it works fine.
Oddly, it hasnt given me the warning again though.
But when the system is running normal ( meaning, automatcially pulling proxies itself without me hitting the Add button ), it saves them to the file just fine.
Settings -> Provider
I enter in C:\Users\Administrator\Desktop\New Text Document.txt and it accepts it, but never pulls it during a run.
unable to save proxies is for the global list of proxies. They are saved in appdata folder. If thats not working, then something on your OS is restricting access to it.
same for that file-provider. Are you sure it has access to the file? I see it is working for me. Maybe simply start the program as admin to see if all works then.
can i import Proxy Sources, in Test Version i dont find something, only one URL.
Or is there no need to import sources, will GSA update the Sources?
Thanks
Peter
After Restart the Tool is not registered!
Please help!
@pbsolution for license issues email GSA directly please. Im sure it's some copy/paste error.
updating sources is however not really required as I am doing this all the time and adding many many of them.
Thanks
@wcwong8
option 1) enable internal proxy in Proxy Scraper and use it in SER (127.0.0.1:8080)
option 2) create an export in Proxy Scraper and add an proxy provider in SER to read proxies from a file
im a little bit disappointed about this Tool. It dosnt find not really more Proxys then the integrated Proxy Scraper in GSA Ranker. Or im doing something wrong?
Peter