crawlerids
Crawlerids, commonly written as crawler IDs, are unique identifiers assigned to web crawlers to identify themselves to websites and services. They may be used to differentiate traffic from different crawlers and to facilitate access control, analytics, and policy enforcement. They can be embedded in user-agent strings, HTTP headers, or in request parameters.
In practice, some crawlers register with a directory or service and receive a persistent ID, sometimes combined
Purposes include tracking crawler activity, enabling site operators to apply crawl quotas, allowlists/blocklists, and gather stats
Limitations include susceptibility to spoofing, incomplete adoption, and privacy concerns. Not all crawlers expose IDs, and
Related concepts include the user agent, robots.txt, robots meta tags, crawl delays, and access tokens. There