Releases: scrapy/scrapy
2.3.0
Hihglights:
-
Feed exports now support Google Cloud Storage as a storage backend
-
The new
FEED_EXPORT_BATCH_ITEM_COUNT
setting allows to deliver output items in batches of up to the specified number of items.It also serves as a workaround for delayed file delivery, which causes Scrapy to only start item delivery after the crawl has finished when using certain storage backends (S3, FTP, and now GCS).
-
The base implementation of item loaders has been moved into a separate library, itemloaders, allowing usage from outside Scrapy and a separate release schedule
2.2.1
The startproject
command no longer makes unintended changes to the permissions of files in the destination folder, such as removing execution permissions.
2.2.0
Highlights:
- Python 3.5.2+ is required now
- dataclass objects and attrs objects are now valid item types
- New
TextResponse.json
method - New
bytes_received
signal that allows canceling response download CookiesMiddleware
fixes
2.1.0
Highlights:
- New
FEEDS
setting to export to multiple feeds - New
Response.ip_address
attribute
2.0.1
2.0.0
Highlights:
- Python 2 support has been removed
- Partial coroutine syntax support and experimental asyncio support
- New Response.follow_all method
- FTP support for media pipelines
- New Response.certificate attribute
- IPv6 support through DNS_RESOLVER
1.7.4
1.7.3
1.7.2
1.7.0
Highlights:
- Improvements for crawls targeting multiple domains
- A cleaner way to pass arguments to callbacks
- A new class for JSON requests
- Improvements for rule-based spiders
- New features for feed exports