Why should we archive Websites?
Websites are informational resources documenting various events, subjects, and changes in society, and it is often imperative that this information is preserved beyond its fleeting online lifespan. In the past, important information recorded in print and other physical forms would be archived decades after its original creation. In the digital present, however, this information is must be archived as it appears online and at regular intervals.
How are Websites archived?
The UML captures and preserves copies of websites using Archive-It and the Internet Archive.
Is personal information captured in Web archiving?
The UML captures and preserves copies of websites that are already publicly accessible.
Do you archive Websites that are password-protected?
No, the UML cannot and will not password-protected Websites.
What about sites that block Web Crawlers?
The UML respects robot exclusion protocols which restrict the crawling of certain online content.
What about Copyright?
Copyright remains with the creator of the Website. Copyright is not transferred to the UML.
What types of files are captured through Web archiving?
In most cases, the crawl of a Website will capture everything from html code, css files, and javascript files to PDF, image, and video files.