-
Notifications
You must be signed in to change notification settings - Fork 319
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add always_enqueue
option to Request
for bypassing deduplication
#547
Comments
Hi @vdusek, I'm interested in working on this enhancement to add an
I'm familiar with Python and HTTP clients, and I'd be happy to work on this enhancement. Let me know if you need any clarification or have additional requirements before I start implementing the solution. Thank you for considering my contribution! |
Hi @belloibrahv, thanks for your interest in Crawlee.
yes
You can probably generate a standard unique key and then append some random string using crypto_random_object_id.
In that case, raise an exception.
I am not aware of any. |
/assign |
@paradoxxx09 We don't assign issues for hacktoberfest. If you want to work on this, open a PR. First mergeable one gets merged. |
always_enqueue
option (or use a better name for it, but avoid negative terms) as an input parameter to theRequest.from_url
constructor.unique_key
, ensuring that each request is always enqueued and processed.unique_key
andalways_enqueue=True
are provided.dont_filter
option in Scrapy (docs).The text was updated successfully, but these errors were encountered: