A Few Questions?
1. Will it be ok if I run multiple projects at the same time that write the identified urls into the same folder?
2. Will it be ok to run a Remove Duplicates Project on a folder where another project is identifying and saving the identified urls into at the same time?
3. Will it be ok if I run Remove Duplicates on a folder in which Scrapebox saves the harvested urls using the automator plugin (NOT harvested_sessions folder from scrapebox) while both running at the same time?
4. I'm running a monitoring project on a folder where scrapebox saves the urls. The file in that folder has around 74k urls and the monitoring project shows processed urls as: 115k and it's still counting. Can you elaborate what's all this about? (3 more projects are doing the same thing)
5. A best practice question: Is it better to use the harvested_sessions folder from scrapebox and process all urls through one project (all engines selected) OR harvest with scrapebox engine by engine(currently only articles and wiki) and then create a project in PI for each engine separately?
I have a few more, but I'd like to get these sorted out first
2. Will it be ok to run a Remove Duplicates Project on a folder where another project is identifying and saving the identified urls into at the same time?
3. Will it be ok if I run Remove Duplicates on a folder in which Scrapebox saves the harvested urls using the automator plugin (NOT harvested_sessions folder from scrapebox) while both running at the same time?
4. I'm running a monitoring project on a folder where scrapebox saves the urls. The file in that folder has around 74k urls and the monitoring project shows processed urls as: 115k and it's still counting. Can you elaborate what's all this about? (3 more projects are doing the same thing)
5. A best practice question: Is it better to use the harvested_sessions folder from scrapebox and process all urls through one project (all engines selected) OR harvest with scrapebox engine by engine(currently only articles and wiki) and then create a project in PI for each engine separately?
I have a few more, but I'd like to get these sorted out first
Comments
Maybe I should process unidentified separately again, I dunno.. What do you think?
It's amazing how well things combine with Platform Identifier! Scraping seemed too complicated and not worth at first, but Platform Identifier made me want to love scraping!
First I share the opinion with a few pros here that MOZ metrics can easily be spammed because I've seen many such sites. Despite that I still think it's the most accurate metric currently and I definitely plan to try and build a list using DA/PA filter to see how it goes.
Second thing is that I don't have a moz pro account yet. Actually I do have one, but it's expiring soon and I won't be renewing it. Can a free moz account be used to check that many links? I thought free moz accounts had some checking limits?