Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

spider fix: use internal download utilities for robots.txt #590

Merged
merged 2 commits into from
May 8, 2024

Conversation

adbar
Copy link
Owner

@adbar adbar commented May 8, 2024

This change makes the crawler more robust by using Trafilatura's download function instead of the default urllib.robotparser one.

@adbar adbar linked an issue May 8, 2024 that may be closed by this pull request
Copy link

codecov bot commented May 8, 2024

Codecov Report

All modified and coverable lines are covered by tests ✅

Project coverage is 97.90%. Comparing base (efe38bb) to head (af0822c).

Additional details and impacted files
@@            Coverage Diff             @@
##           master     #590      +/-   ##
==========================================
+ Coverage   97.81%   97.90%   +0.09%     
==========================================
  Files          21       21              
  Lines        3437     3443       +6     
==========================================
+ Hits         3362     3371       +9     
+ Misses         75       72       -3     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@adbar adbar merged commit 92bdd6e into master May 8, 2024
15 checks passed
@adbar adbar deleted the fix_robots_download branch May 8, 2024 09:58
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

No timeout in urllib.robotparser with focused_crawler
1 participant