Skip to content

GSA SER Posting working wrong - already parsed

rysioslaw2rysioslaw2 hitman.agency
Hi,
I use GSA SER.
I made 5mln links with comment form on a pages. When I posted I saw many 'already parsed' info in log window.
After investigating I almost understand how it's work. I think, when posting fail on one link, never more try to post on same domain.
I made log to .txt file and made new project. The new project with imported "already parsed" links gave me a lot of good valuable and verified links, but also new already parsed.
In effect I get about 80% verified links from links which was signed as already parsed in first project.
In my opinion, this is a logical error in the operation of GSA SER. The fact that the first or second link in the same domain failed does not mean that the third and fourth are incorrect!

My question: Can I disable this feature and never see the "already parsed" message again?

Tagged:

Comments

  • SvenSven www.GSA-Online.de
    SER will set a whole domain as "already parsed" if it failed to submit or this engine is not URL based (e.g. wordpess article). For comments it will use every URL and only add it as "already parsed" if it submitted there before (and no duplicate posting is enabled) or it does not detect ot as a comment plattfrom or fails to submit there.
  • rysioslaw2rysioslaw2 hitman.agency
    I have no duplicate urls in my list.
    I think this logic is wrong because engine.ini (in this case General Blogs.ini) can't recognize engine. It's recognized only is it possible to post on exact page. In this case SER not recognize urls to pages like: contact-me about-me etc, as a wordpress and flag domain as "already parsed".

    For explain:
    I made a test: I made 4 links to my page: 3 links to blog posts where I have comment form and 4th to contact form (it's bad link but still wordpress engine).
    I try to post to Blog Comment / General Blog. When my list have links order: 1,2,3,4 - I have made 3 comments, but when I change order to: 4,1,2,3 I get: "no engine matches" and 3 links "already parsed".

    That problem appear in another situation. When I have 20 proxy and one of them is blacklisted, websites with antibot protection (cloudflare sg-captcha or siteground) not opening when SER try to use blacklisted proxy. In that case good proxy posts on that website, blacklisted have "no engine matches" and after this attempt, the entire domain is marked as "already parsed".

  • rysioslaw2rysioslaw2 hitman.agency
    @backlinkaddict Thank you for your help, I appreciate it.
    In my case, I have huge list, many many millions of links to wordpress page where was comment form on a page. Because of the I was very surprised when I saw more than 200 logs "already parsed". I think it's appearing because proxy blacklisting.
    Of course the best idea is to make script or another way to solve this situation. Another program when find something like that try to use another proxy.

    GSA SER know if there is antibot script. I think @Sven know that very good. 
    When response is similar to ex:

    response:
    <html><head><link rel="icon" href="data:;"><meta http-equiv="refresh" content="0;/.well-known/sgcaptcha/?r=%2F%D9%8A%D8%A7-%D8%A3%D9%85%D8%A7%D9%87-%D8%A3
    %D8%BA%D9%84%D9%82%D9%8A-%D8%A7%D9%84%D8%A8%D8%A7%D8%A8-%D9%88%D9%86%D8%A7%D9%85%D9%8A%2F&y=ipr:xxx.xxx.xxx.xxx:1710612134.417"></meta></head></html>

    it is certain that antibot is installed. Of course, this is just an example.

    If in that case SER will resolve that problem - will be very good, if SER will try to use another proxy - will be good, if SER will do nothing - will be enough, but when SER will flag whole domain as "already parsed" and in the future won't to try posting - it will be VERY BAD.

    GSA SER has a proxy scraper and should attempt to post to URLs even if it failed to post before.

Sign In or Register to comment.