Why Getting Already Parsed when I have NOT run these URL's?
I ran a list of scraped blogs (contextual blog posts) through my 3 tears.
On T1 I had PR set at a min of PR4
T2 and T2 I allowed any PR
As time has gone by targets are getting harder to find, I've decided to allow t1 to use domains
with a PR of 3.
I set min PR on T1 from PR4 to PR3 and re loaded JUST the PR3 domains from the initial run.
I was surprised when all of them produced an "already parsed" message.
None of these domains had been posted to on T1 before because the PR was too low on them before I reset it, so why the already parsed message?
Is the already parse message a global thing?
How can I get around it, must I delete history (would rather not)
Any ideas?
Comments
My educated guess is:
Yes, the URL's were 'parsed' in the sense that it looked at those targets already and it didn't meet the PR criteria.
So now you decide to change the PR benchmark. But the problem is that SER already has processed/parsed this URL - and now it's in history.
I don't think there is any way around deleting the history. In fact, I think this is exactly what @sven envisioned as an example of why you want to clear out your history.
I remember him writing on here about a slightly different example but very similar (it was a long time ago and hope I am getting it correct). I think it was where a user imported a bunch of targets, but didn't have that engine turned on at the time. So those targets were processed when there was 'no engine' and were basically parsed but no dice. So to allow them to be processed with that engine turned on, you need to clear the history. I think that is very similar to your situation.