Ok so quick update: as soon as I changed all sleep() parameters in the main python file to 15, my success rates went WAY up! In fact the only errors I get are on the recipient Sites' ends.
Thanks @sven as well as @CL67F for making this possible! It's at the very least Twice if not Thrice as accurate as XEvil! And it's at least 4 times faster
Ok so just a quick thing: it does try to repeat tasks that are unable to be solved.
Example: one of the websites I'm trying to post to keeps giving me an SSL_CERT error, which, by the way, I have tried telling Python to stop trying if that's the case:
if e == 'SSL_ERROR_BAD_CERT_DOMAIN' or 'SEC_ERROR_UNKNOWN_ISSUER' or 'SEC_ERROR_UNKNOWN' or 'SEC_ERROR_EXPIRED_CERTIFICATE' or 'Timeout 30000ms exceeded.':
try_connection = 0
force_new_proxy = False
However, it's still trying new proxies when in fact it's just going to fail every time. What should I do in order to tell the code to just skip the site entirely if those errors occur? I can press ctrl+C but that's just not practical if I'm running this overnight....
On Windows 10, please pay attention to Windows Defender CPU ! I didn"t manage to find a solution, I changed Antivirus !!! Ahhh, these fucking Windows Defender Antimalware & SmartScreen !!!! Got a headache !!!!
backlinkaddict Forgive me as I sure wish I had statistical proof! I just notice how XEvil takes at least 60-90 seconds per query to solve, and it really is only 50/50 solve success-fail.
With this solver, on the other hand, it's a solve every 10-12 seconds! I can see it in my web contact dashboard as "sent" as well (if it was an error like an 'unsolvable' XEvil and web contact would tell me).
Speaking of which, quick update and Good News simultaneously: it turns out I just needed to modify my aforementioned code as such:
if e == 'SSL_ERROR_BAD_CERT_DOMAIN' or 'SEC_ERROR_UNKNOWN_ISSUER' or 'SEC_ERROR_UNKNOWN' or 'SEC_ERROR_EXPIRED_CERTIFICATE' or 'Timeout 30000ms exceeded.':
By the way @CL67F and @backlinkaddict my script is still hanging after some time (sometimes after 5 minutes running or 10 minutes running the script just stops in perpetuity) - have you had the same problem? If so, how did y'all fix it?
@daviddig1 what's your PC config for xevil ? I've recently started using it and ranging 15-20s per solve with 200 threads. using reproxy and their proxies are working great. I've tried this script initially but gave up pretty soon cos of constant errors and issues. I really hope this script becomes a part of GSA CB as recaptcha solving is much needed these days.
Ok @rastarr or @CL67F I'm afraid I'm still having to press Ctrl+c or outright restart the bat file - any ideas as to how to make sure the script doesn't stop all of a sudden?
Ok @rastarr or @CL67F I'm afraid I'm still having to press Ctrl+c or outright restart the bat file - any ideas as to how to make sure the script doesn't stop all of a sudden?
When I was using my original version, this was a daily process, unfortunately. The framework doesn't always clean up unused previous version of chrome/firefox, resulting in ever increasing memory usage. That was one of the reasons I eventually stopped using it. Frameworks like Selenium, Playwright and the like are for low volume web scraping duties. They were never meant to be used in such high-volume activities like captcha solving within GSA like products. They only have rudimentary memory management, at best.
Ok @rastarr or @CL67F I'm afraid I'm still having to press Ctrl+c or outright restart the bat file - any ideas as to how to make sure the script doesn't stop all of a sudden?
When I was using my original version, this was a daily process, unfortunately. The framework doesn't always clean up unused previous version of chrome/firefox, resulting in ever increasing memory usage. That was one of the reasons I eventually stopped using it. Frameworks like Selenium, Playwright and the like are for low volume web scraping duties. They were never meant to be used in such high-volume activities like captcha solving within GSA like products. They only have rudimentary memory management, at best.
Oh ok no problem thanks for letting me know - would you recommend I turn down the number of threads in the.py file?
Ok @rastarr or @CL67F I'm afraid I'm still having to press Ctrl+c or outright restart the bat file - any ideas as to how to make sure the script doesn't stop all of a sudden?
When I was using my original version, this was a daily process, unfortunately. The framework doesn't always clean up unused previous version of chrome/firefox, resulting in ever increasing memory usage. That was one of the reasons I eventually stopped using it. Frameworks like Selenium, Playwright and the like are for low volume web scraping duties. They were never meant to be used in such high-volume activities like captcha solving within GSA like products. They only have rudimentary memory management, at best.
Oh ok no problem thanks for letting me know - would you recommend I turn down the number of threads in the.py file?
from testing point of view, I'd start with a very low number of threads, run it for a day and see how it plays out. It's irrelevant on the thread number though. Those ophaned chrome/firefox instances will grow so you need to find a happy medium that you can bear.
@daviddig1 what's your PC config for xevil ? I've recently started using it and ranging 15-20s per solve with 200 threads. using reproxy and their proxies are working great. I've tried this script initially but gave up pretty soon cos of constant errors and issues. I really hope this script becomes a part of GSA CB as recaptcha solving is much needed these days.
Hi @dp001 I had xevil set to 350 threads with 500 proxies - I will say it is much faster at solving for v3 BUT it gave a lot of captcha_unsolvable status codes in 'website contact'
Ok @rastarr do you know how I can turn off logging to capv2v3.log and capv2v3_err.log? Not only are the resulting log lines superfluous , They get huge very fast, and I have a strong feeling turning them off means the code will be more stable since it doesn't have to write to the 2 files...
Ok @rastarr do you know how I can turn off logging to capv2v3.log and capv2v3_err.log? Not only are the resulting log lines superfluous , They get huge very fast, and I have a strong feeling turning them off means the code will be more stable since it doesn't have to write to the 2 files...
@rastarr ok so I restored the original py file with a number of loggings turned off, and it runs perfectly smoothly, but I'm now running into my original problem that, it's trying to solve sites (switching proxies over and over) that are constantly returning an exception like a bad cert or a constant timeout error.
So: I just need to know how to stop looping the same sites with exception errors - otherwise it's just a waste of time as well as local resources - I've tried but my lack of python knowledge has bit me in the butt.
I just had a thought: why not just turn off the dialogue messages in cmd in their entirety? I really don't need to see them in cmd, and it could be the culprit of how it has to be Ctrl+C'd or restarted periodically
2.b A list of proxies (optional) If file is absent, this functionnality is disabled. The proxy is selected in random order. The proxy is tried 3 times then considered as « failed », therefore replaced by an other random proxy in list. The previous proxy is deleted from the list. • _proxies_v2.txt : list of proxies for recaptchav2 • _proxies_v3.txt : list of proxies for recaptchav3 Example : 182.25.68.36 :9985 (each line) Example : 182.25.68.36:9985;user;password (each line) 2.c Rotating proxies (optional) If file is present, this option has priority. If file is absent, this functionnality is disabled. Each access to the address gives a new proxy. Proxies are in rotation
The proxy is tried 3 times then considered as « failed », therefore replaced by an other random proxy from address. • _rotating_proxies_v2.txt : one rotating address for proxies for recaptchav2 • _rotating_proxies_v3.txt : one rotating address for proxies for recaptchav3 Example : h ttp://p.webshare.io:9999/ (one unique line)
daviddig1 Your local host is down? It's local loopback address. Maybe there is just nothing listening at the time, something in your host file, or another program trying to listen?
Comments
In fact the only errors I get are on the recipient Sites' ends.
Thanks @sven as well as @CL67F for making this possible! It's at the very least Twice if not Thrice as accurate as XEvil! And it's at least 4 times faster
Example: one of the websites I'm trying to post to keeps giving me an SSL_CERT error, which, by the way, I have tried telling Python to stop trying if that's the case:
if e == 'SSL_ERROR_BAD_CERT_DOMAIN' or 'SEC_ERROR_UNKNOWN_ISSUER' or 'SEC_ERROR_UNKNOWN' or 'SEC_ERROR_EXPIRED_CERTIFICATE' or 'Timeout 30000ms exceeded.':
However, it's still trying new proxies when in fact it's just going to fail every time.
What should I do in order to tell the code to just skip the site entirely if those errors occur? I can press ctrl+C but that's just not practical if I'm running this overnight....
I didn"t manage to find a solution, I changed Antivirus !!!
Ahhh, these fucking Windows Defender Antimalware & SmartScreen !!!!
Got a headache !!!!
All my GSA Softwares are working great !
Using API !
With this solver, on the other hand, it's a solve every 10-12 seconds! I can see it in my web contact dashboard as "sent" as well (if it was an error like an 'unsolvable' XEvil and web contact would tell me).
Speaking of which, quick update and Good News simultaneously: it turns out I just needed to modify my aforementioned code as such:
if e == 'SSL_ERROR_BAD_CERT_DOMAIN' or 'SEC_ERROR_UNKNOWN_ISSUER' or 'SEC_ERROR_UNKNOWN' or 'SEC_ERROR_EXPIRED_CERTIFICATE' or 'Timeout 30000ms exceeded.':
resetTempo(page_url)
By the way indeed I'm using 500 proxies from WebShare
I've tried this script initially but gave up pretty soon cos of constant errors and issues. I really hope this script becomes a part of GSA CB as recaptcha solving is much needed these days.
The framework doesn't always clean up unused previous version of chrome/firefox, resulting in ever increasing memory usage. That was one of the reasons I eventually stopped using it.
Frameworks like Selenium, Playwright and the like are for low volume web scraping duties. They were never meant to be used in such high-volume activities like captcha solving within GSA like products. They only have rudimentary memory management, at best.
It's irrelevant on the thread number though. Those ophaned chrome/firefox instances will grow so you need to find a happy medium that you can bear.
So: I just need to know how to stop looping the same sites with exception errors - otherwise it's just a waste of time as well as local resources - I've tried but my lack of python knowledge has bit me in the butt.
"username": "test2024-rotate",
"password": "passwordsr4l0zers"
})
If file is absent, this functionnality is disabled. The proxy is selected in random
order. The proxy is tried 3 times then considered as « failed », therefore
replaced by an other random proxy in list. The previous proxy is deleted from
the list.
• _proxies_v2.txt : list of proxies for recaptchav2
• _proxies_v3.txt : list of proxies for recaptchav3
Example : 182.25.68.36 :9985 (each line)
Example : 182.25.68.36:9985;user;password (each line)
2.c Rotating proxies (optional)
If file is present, this option has priority. If file is absent, this functionnality is
disabled. Each access to the address gives a new proxy. Proxies are in rotation
an other random proxy from address.
• _rotating_proxies_v2.txt : one rotating address for proxies for
recaptchav2
• _rotating_proxies_v3.txt : one rotating address for proxies for
recaptchav3
Example : h ttp://p.webshare.io:9999/ (one unique line)
help
yes gsa ser license breaker license
ffmeg install Link ?
21:11:59: [-] 191/441 ReCaptcha solving failed - Capv2v3: RECAPTCHA_UNSOLVABLE (SiteKey: 6LdvXBYcAAAAAJ_wjXuVdRUpc9H1GBGD_c0keuSD, Type: v2)
21:13:07: [-] 229/436 captcha service Capv2v3 is not responding/can`t solve