A powerful web scraping and security scanning tool designed for penetration testing and security analysis. This tool combines website scraping capabilities with vulnerability detection to help identify common security issues.
ScorpionV3 is a dual-purpose tool developed for security professionals, penetration testers, and web developers who need to analyze websites for both content and security vulnerabilities. It can crawl websites recursively, download assets, and scan for common security issues such as XSS vulnerabilities, SQL injection points, information disclosure, and outdated software.
This project was created to simplify the initial reconnaissance phase of security assessments and provide actionable insight into potential vulnerabilities.
- Recursive website crawling
- Asset extraction (CSS, JS, images, videos, etc.)
- Proxy support
- Random User-Agent rotation
- Cross-Site Scripting (XSS) vulnerability detection
- SQL Injection vulnerability detection
- Information disclosure identification
- Software version detection
- HTTPS/SSL configuration checking
- Security header analysis
- Batch scanning of multiple URLs
- Detailed vulnerability reporting
- Customizable scan depth
- Authentication support
- Custom vulnerability rule definitions
- Export to PDF/HTML reports
- Python 3.7+
- Chrome browser (for Selenium-based scraping)
git clone https://github.com/greenbugx/ScorpionV3.git
cd ScorpionV3python -m venv venv
source venv/bin/activate # On Windows: venv\Scripts\activatepip install -r requirements.txtCreate a file named proxies.txt in the root directory and add your proxies (one per line) in the following format:
ip:port
ip:port:username:password
Create a file named UserAgent.txt in the root directory and add your User-Agents (one per line).
python main.pyWhen you run the tool, you'll be presented with the following options:
- Start Scraping - Scrape a website and save its assets
- Start Security Scan (Single URL) - Perform a security scan on a single URL
- Batch Security Scan (from file) - Scan multiple URLs from a list
- Quit - Exit the program
- Select option 1 from the main menu
- Enter the URL to scrape
- Provide a folder name to save the scraped content
- Specify the crawling depth (0 for single page, 1+ for recursive crawling)
- Select option 2 from the main menu
- Enter the URL to scan
- The tool will scan the URL and save results to the
security_scans/single_scansdirectory
- Create a text file with one URL per line
- Select option 3 from the main menu
- Enter the path to your URL list file
- Specify the number of parallel scans to run
- Results will be saved to the
security_scans/batch_scan_[timestamp]directory
The tool generates the following outputs:
- Scraped websites - Saved in the
scraped_sites/[folder_name]directory - Security scan results - Saved as JSON files in the
security_scans/directory - Log file - All activities are logged in
scraper.log
This tool is provided for educational and professional security assessment purposes only. Unauthorized scanning of websites may violate computer crime laws and other regulations.
- Always obtain explicit permission before scanning any website you don't own.
- Respect robots.txt and website terms of service.
- Do not use this tool to cause harm, disruption, or unauthorized access to systems.
- Use responsible scanning practices - avoid excessive requests that could cause denial of service.
- You are solely responsible for how you use this tool and any consequences thereof.
- Test on your own websites or dedicated security testing environments
- Use public test sites specifically designed for security testing
- Consider setting up local vulnerable applications like OWASP WebGoat or DVWA for practice
This project is licensed under the MIT License - see the LICENSE file for details.
Contributions are welcome! Please feel free to submit a Pull Request.
- Fork the project
- Create your feature branch (
git checkout -b feature/amazing-feature) - Commit your changes (
git commit -m 'Add some amazing feature') - Push to the branch (
git push origin feature/amazing-feature) - Open a Pull Request
- This tool uses BeautifulSoup for HTML parsing
- Selenium is used for browser automation