InternetArchiveBot

What is InternetArchiveBot?

About

InternetArchiveBot is a developer helper operated by Internet Archive. If you think this is incorrect or can provide additional detail about its purpose, please contact us. You can see how often InternetArchiveBot visits your website by setting up Dark Visitors Agent Analytics.

Expected Behavior

Developer helpers generally visit websites on direct behalf of a user who is evaluating some feature or attribute. Visits can be one-off or part of a larger crawl.

Type

Developer Helper
Used by developers to test website functionality

Detail

Operated By Internet Archive
Last Updated 17 hours ago

Insights

Top Website Robots.txts

0%
0% of top websites are blocking InternetArchiveBot
Learn How →

Country of Origin

United States
InternetArchiveBot normally visits from the United States

Global Traffic

The percentage of all internet traffic coming from Developer Helpers

Top Visited Website Categories

News
Sports
Games
People and Society
Health
How Do I Get These Insights for My Website?
Use the WordPress plugin, Node.js package, or API to get started in seconds.

Robots.txt

Should I Block InternetArchiveBot?

Probably not. Developer helpers are normally used to optimize or find problems with your website.

How Do I Block InternetArchiveBot?

You can block InternetArchiveBot or limit its access by setting user agent token rules in your website's robots.txt. Set up Dark Visitors Agent Analytics to check whether it's actually following them.

How Do I Block All Developer Helpers?
Serve a continuously updating robots.txt that blocks new developer helpers automatically.
User Agent String IABot/2.0 (+https://meta.wikimedia.org/wiki/InternetArchiveBot/FAQ_for_sysadmins) (Checking if link from Wikipedia is broken and needs removal)
# In your robots.txt ...

User-agent: InternetArchiveBot # https://darkvisitors.com/agents/internetarchivebot
Disallow: /

⚠️ Manual Robots.txt Editing Is Not Scalable

New agents are created every day. We recommend setting up Dark Visitors Automatic Robots.txt if you want to block all agents of this type.