Skip to content

Conversation

@giulianoriccio
Copy link

Crawling bot that doesn't respect the robots.txt rules and continues to make requests even on overloaded servers.

Makes requests through Amazon AWS or Bytedance owned network.

Crawling bot that doesn't respect the robots.txt rules and continues to make requests even on overloaded servers.

Makes requests through Amazon AWS or Bytedance owned network.
Crawling bot that doesn't respect the robots.txt rules and continues to make requests even on overloaded servers.

Makes requests through Amazon AWS or Bytedance owned network.
@Cyrille37
Copy link
Contributor

Any other reviewer ?

@LaurenceJJones
Copy link
Member

again as per #82 we cannot make presumptions on what user classes as "bad", we should pass out contextual lists then user can easily disable or enable blocking those rather than "these are bad"

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants