Skip to content
forked from dmlc/xgboost

Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow

License

Notifications You must be signed in to change notification settings

chenzihao008/xgboost

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

xgboost: eXtreme Gradient Boosting

A General purpose gradient boosting (tree) library.

Authors:

  • Tianqi Chen, project creater
  • Kailong Chen, contributes regression module

Turorial and Documentation: https://github.com/tqchen/xgboost/wiki

Features

  • Sparse feature format:
    • Sparse feature format allows easy handling of missing values, and improve computation efficiency.
  • Push the limit on single machine:
    • Efficient implementation that optimizes memory and computation.
  • Layout of gradient boosting algorithm to support generic tasks, see project wiki.

Supported key components

  • Gradient boosting models:
    • regression tree (GBRT)
    • linear model/lasso
  • Objectives to support tasks:
    • regression
    • classification
  • OpenMP implementation

Planned components

  • More objective to support tasks:
    • ranking
    • matrix factorization
    • structured prediction

File extension convention

  • .h are interface, utils and data structures, with detailed comment;
  • .cpp are implementations that will be compiled, with less comment;
  • .hpp are implementations that will be included by .cpp, with less comment

About

Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • C++ 42.3%
  • Python 18.6%
  • Cuda 16.7%
  • Scala 8.5%
  • R 7.2%
  • Java 4.1%
  • Other 2.6%