Notes on using GNU Emacs' Tramp system in an unusual shell environment
Summary
The article argues that overly generic HTTP User-Agent headers hinder access to sites and advocates for identifiable user agents that include real URLs describing the software or activity. It highlights a surge of high-volume crawlers aimed at data collection for LLM training and suggests blocking or throttling such traffic to reduce server load.