Skip to content

[Bug]: deep crawl crawls same url multiple times #843

Closed
@eliaweiss

Description

@eliaweiss

crawl4ai version

0.5.0.post4

Expected Behavior

should not crawl the same url

Current Behavior

i see in the log the urls are crawled multiple times:

NFO:crawlApp:Processed page 1: https://out-door.co.il
[FETCH]... ↓ https://out-door.co.il/... | Status: True | Time: 3.48s
[SCRAPE].. ◆ https://out-door.co.il/... | Time: 0.237s
[COMPLETE] ● https://out-door.co.il/... | Status: True | Total: 3.72s
....
[FETCH]... ↓ https://out-door.co.il/... | Status: True | Time: 25.25s
[SCRAPE].. ◆ https://out-door.co.il/... | Time: 0.34s
[COMPLETE] ● https://out-door.co.il/... | Status: True | Total: 25.60s
INFO:app.process_crawl:✅ Result already saved: https://out-door.co.il/

Is this reproducible?

Yes

Inputs Causing the Bug

Steps to Reproduce

Code snippets

OS

ubuntu

Python version

3.12.3

Browser

No response

Browser version

No response

Error logs & Screenshots (if applicable)

No response

Metadata

Metadata

Assignees

Labels

✅ ReleasedBug fix, enhancement, FR that's released🐞 BugSomething isn't working📌 Root causedidentified the root cause of bug

Projects

No projects

Relationships

None yet

Development

No branches or pull requests

Issue actions