The reason this works is unironically because of Googlebot. These paywall sites want to be at the top of search engine results, so they have to exempt Googlebot and other spiders from the paywall block, otherwise the data can't be indexed. All archive does is spoof the common spider user-agent headers on its requests, and poof, they get in.
The reason this works is unironically because of Googlebot. These paywall sites want to be at the top of search engine results, so they have to exempt Googlebot and other spiders from the paywall block, otherwise the data can't be indexed. All archive does is spoof the common spider user-agent headers on its requests, and poof, they get in.