Not sure if it's a bug, or GSA has a different way of doing the math. The text file has exactly 500k websites. For some reason, GSA counted almost 600k and it still has some to go
It's more like a funny thing than a problem.
EDIT1: Also, the remove duplicate DOMAINS leaves duplicate URLS, because every time I do that, I still have duplicate URLs left.