Skip to content

Commit

Permalink
chore(bench): add large benches
Browse files Browse the repository at this point in the history
  • Loading branch information
j-mendez committed Dec 27, 2023
1 parent 50ad1c6 commit e50dd74
Show file tree
Hide file tree
Showing 2 changed files with 38 additions and 2 deletions.
20 changes: 19 additions & 1 deletion bench/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -32,4 +32,22 @@ pages found 200
elasped duration 5.860108852386475
```

Linux performance for Spider-RS increases around 10x especially on Arm.
Test url: `https://a11ywatch.com` (medium)
648 pages

| `libraries` | `speed` |
| :-------------------------------- | :------ |
| **`spider-rs: crawl 10 samples`** | `2s` |
| **`scrapy: crawl 10 samples`** | `7.7s` |

Test url: `https://espn.com` (large)
150,387 pages

| `libraries` | `speed` |
| :-------------------------------- | :------ |
| **`spider-rs: crawl 10 samples`** | `186s` |
| **`scrapy: crawl 10 samples`** | `1h+` |

Scrapy used too much memory, crawl cancelled after an hour.

Note: The performance scales the larger the website and if throttling is needed. Linux benchmarks are about 10x faster than macOS for spider-rs.
20 changes: 19 additions & 1 deletion book/src/benchmarks.md
Original file line number Diff line number Diff line change
Expand Up @@ -50,4 +50,22 @@ Test url: `https://rsseau.fr` (medium)
| **`spider-rs: crawl 10 samples`** | `2.5s` |
| **`scrapy: crawl 10 samples`** | `10s` |

The performance scales the larger the website and if throttling is needed. Linux benchmarks are about 10x faster than macOS for spider-rs.
Test url: `https://a11ywatch.com` (medium)
648 pages

| `libraries` | `speed` |
| :-------------------------------- | :------ |
| **`spider-rs: crawl 10 samples`** | `2s` |
| **`scrapy: crawl 10 samples`** | `7.7s` |

Test url: `https://espn.com` (large)
150,387 pages

| `libraries` | `speed` |
| :-------------------------------- | :------ |
| **`spider-rs: crawl 10 samples`** | `186s` |
| **`scrapy: crawl 10 samples`** | `1h+` |

Scrapy used too much memory, crawl cancelled after an hour.

Note: The performance scales the larger the website and if throttling is needed. Linux benchmarks are about 10x faster than macOS for spider-rs.

0 comments on commit e50dd74

Please sign in to comment.