Skip to content
forked from tb0hdan/domains

Domains Project: Sorted list of internet domains

License

Notifications You must be signed in to change notification settings

KTZgraph/domains

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Welcome to Domains Project!

Sorted list of internet domains

This dataset was created as an experiment with hopes of being useful.

Dataset statistics

  • After reaching 9 million domains I've switched to archived storage. Please use freely available XZ to unpack files.

  • After reaching 30 million domains I've moved domain directories to /data so repository doesn't have it's README at the very bottom.

Milestones:

  • 10 Million
  • 20 Million
  • 30 Million
  • 50 Million
  • 100 Million
  • 150 Million
  • 300 Million

Random facts:

  • More than 1TB of Internet traffic is just 3 Mbytes of compressed data
  • 1 million domains is just 5 Mbytes compressed
  • More than 86TB of Internet traffic is necessary to get 50 million domains (1.73TB / 1 million).
  • Only 230 Mbytes of disk space is required to store 50 million domains in a compressed form
  • 1Gbit fully saturated link is good for about 5 million new domains every day
  • 8c/16t and 64 Gbytes of RAM machine is good for about 2 million new domains every day
  • 2 ISC Bind9 instances (>400 Mbytes RSS each) are required to get 2 million new domains every day

Using dataset

This repository empoys Git LFS technology, therefore user has to use both git lfs and xz to retrieve data. Cloning procedure is as follows:

git clone https://github.com/tb0hdan/domains.git
cd domains
./unpack.sh

Search engines and crawlers

Crawlers

All data in this dataset is gathered using Scrapy and Colly frameworks.

Crawler code for this project is available at: Domains Crawler

Yacy

Yacy is a great opensource search engine. Here's my post on Yacy forum: https://searchlab.eu/t/domain-list-for-easier-search-bootstrapping/231

Research

This dataset can be used for research. There are papers that cover different topics. I'm just going to leave links to them here for reference.

Re-registration and general statistics

Analysis of the Internet Domain Names Re-registration Market

Lexical analysis of malicious domains.

Detection of malicious domains through lexical analysis

Malicious Domain Names Detection Algorithm Based on Lexical Analysis and Feature Quantification

Detecting Malicious URLs Using Lexical Analysis

About

Domains Project: Sorted list of internet domains

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Shell 100.0%