Scrapebox - The Proxy Question
Hi guys,
first of all, I do realise there's a forum dedicated to Scrapebox, but the answers I got there were hardly satisfying, so I decided to rely on you guys
I've recently started scraping by myself and bought Scrapebox and went with a public proxy service. The service is 20$/month and I get 2x ~15k proxies every day. And this is exactly where the problem is.
The public proxies I get burn out pretty fast and therefore I'll have to exchange them at least once every 24h. This is quite annoying and costs a lot of time, because I have to abort the scrape, export unfinished KWs, reload new proxies, load the KWs, etc. And this issue makes it pretty much impossible for me to do big scrapes.
So my question is - Would you guys suggest getting ~40 semi dedicated proxies for scraping and maybe reduce the simultaneous connections per proxy to avoid proxy bans, or is there like a minimum amount for private proxies if I wanted to use them for scraping and should I therefore stay with my public proxy service?
I've read a couple of guides and some even considered amounts as low as 25 proxies to be enough for beginners.
Any suggestions?
first of all, I do realise there's a forum dedicated to Scrapebox, but the answers I got there were hardly satisfying, so I decided to rely on you guys
I've recently started scraping by myself and bought Scrapebox and went with a public proxy service. The service is 20$/month and I get 2x ~15k proxies every day. And this is exactly where the problem is.
The public proxies I get burn out pretty fast and therefore I'll have to exchange them at least once every 24h. This is quite annoying and costs a lot of time, because I have to abort the scrape, export unfinished KWs, reload new proxies, load the KWs, etc. And this issue makes it pretty much impossible for me to do big scrapes.
So my question is - Would you guys suggest getting ~40 semi dedicated proxies for scraping and maybe reduce the simultaneous connections per proxy to avoid proxy bans, or is there like a minimum amount for private proxies if I wanted to use them for scraping and should I therefore stay with my public proxy service?
I've read a couple of guides and some even considered amounts as low as 25 proxies to be enough for beginners.
Any suggestions?
Comments
I have 100 dedicated and i use them in SB 24/7 with 300 threads and never have any problems (They are being used on 3 servers for SER too).
The only problem is yours are semi-dedicated, so other people might be getting them banned already.
Well right now I've reduced my proxy package to 10 proxies, because my projects are rather small and I wanted to save some costs.
But I do see your point though. I guess semi private proxies work just fine for SER submissions, but if I only have like ~30-50 semi dedicated proxies and a couple of these get bans, even if it's just a temp. ban then this will severely affect my scraping progress and I might have to stop it, just like I'd have to do it with my public proxy lists right now.
So I guess what you're saying is if I wanted to use private proxies for scraping I should go for fully private proxies and not semi private, right?
@fakenickahl Yes, I've heard about GScraper's proxy handling. Much better than SB. I actually can't believe that a tool as great as SB handles its proxies so incredibly bad. Well, I guess I'll have to live with it until I can afford GScraper or fully private proxies.
Thanks for the input guys.
I forgot to mention that, @tixxpff ignore what i said before. Under normal circumstances those proxies would be burnt out in SB.
@fakenickahl is absolutely right there.
I tried 2 proxy services (which I won't name here, but I can give them out via PM if requested).
One was giving out daily proxy updates, which required me to stop my projects, export my KWs, import the new proxies, reimport my KWs and continue scraping. I guess this method does work, but it wasn't my preferred choice.
Service 2 was giving out a (small) list of proxies through which you would connect to a huge pool of proxies which are constantly updated. So pretty much the same as service 1, but without the annoying proxy changing.
I guess it doesn't really matter for GScraper, since you can import proxies on the fly, but for Scrapebox reimporting proxies and KWs was a huge pain in the ass.