Skip to content

the crawler is collecting data slowly and memory usage is increasing #1190

Answered by janbuchar
ycq0125 asked this question in Q&A
Discussion options

You must be logged in to vote

Hi @ycq0125! I noticed that you're making a new BeautifulSoupCrawler instance for each crawled host. Is there any chance you could reuse the same instance? Also, if you want to ingest requests from Redis, perhaps you could implement that as a RequestLoader or RequestManager (former if you just want to read from Redis, latter if you also want Redis to handle retries and additional links).

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by ycq0125
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants