Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Partial index promotion due to scrappy spider signals not being handled #540

Open
ProTip opened this issue Jan 15, 2021 · 0 comments
Open

Comments

@ProTip
Copy link

ProTip commented Jan 15, 2021

I have personally experienced Ctrl-C resulting in an incomplete index.

The scrappy documentation for spider_closed signal, https://docs.scrapy.org/en/latest/topics/signals.html#scrapy.signals.spider_closed , mentions that the reason for the closing should be finished under normal circumstances. However the reason could be shutdown in the case of a Ctrl-C signal stopping the spider.

It doesn't appear the that doc spider is patching into any of the spider signals. In the case that a spider does not finish naturally the scraper is unaware and will proceed to promote the incomplete temp index.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant