Skip to content

GH-hawolf/exploit_db_spider

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 

Repository files navigation

Features

  • 用于爬取Exploit_Database上的数据,包括exploit文件、exploit_id、Author等主要信息。
  • 可基于关键字来过滤exploit_title,以爬取特定exploit。
  • 可基于exploit_id,爬去一定范围内的exploit。

Usage

  • Option1 - router_exploits_spider.py
    • 默认模式,过滤'router',start_id: 42740,end_id: 19943
  • Option2 - router_exploits_spider.py target_str start_id end_id
    • 参数
    • target_str: string,用于过滤exploit_title
    • start_id: 开始爬取的exploit_id
    • end_id: 结束爬去的exploit_id

Save

  • dir exploits
    • 保存爬取到的exploits,文件名为exploit_id
  • dir info
    • 保存爬取到每个exploit的主要信息,文件名为exploit_id
  • exploit_id_done.txt
    • 保存以本次规则为基础,已经爬取过的exploit_id
  • save_DecodeError_url.txt
    • 保存以本次规则为基础,出现解码错误的url
  • save_timeout_url.txt
    • 保存以本次规则为基础,出现访问超时的url

About

A script for crawling exploits database

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages