I imagine it would be very difficult for someone to start a website which hosts a lot of copyrighted material and claim they are genuinely archiving copyrighted material. This would be the defense of every piracy site ever, if it were feasible, I imagine.
I am not a copyright lawyer, and I welcome correction on this.
Developing an archival API for those who want their site archived is perfectly fine, though this is probably what robots.txt is for.