Skip to content
/ xgboost Public
forked from dmlc/xgboost

Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Flink and DataFlow

License

Notifications You must be signed in to change notification settings

shaocf/xgboost

Repository files navigation

DMLC/XGBoost

Build Status Gitter chat for developers at https://gitter.im/dmlc/xgboost

An optimized general purpose gradient boosting library. The library is parallelized, and also provides an optimized distributed version. It implements machine learning algorithms under the Gradient Boosting framework, including Generalized Linear Model (GLM) and Gradient Boosted Decision Trees (GBDT). XGBoost can also be distributed and scale to Terascale data

Check out our Committers and Contributors who help make xgboost better.

Documentation: Documentation of dmlc/xgboost

Issue Tracker: https://github.com/dmlc/xgboost/issues

Please join XGBoost User Group to ask questions and share your experience on xgboost.

  • Use issue tracker for bug reports, feature requests etc.
  • Use the user group to post your experience, ask questions about general usages.

Distributed Version: Distributed XGBoost

Highlights of Usecases: Highlight Links

XGBoost is part of Distributed Machine Learning Common projects

What's New

Contributing to XGBoost

XGBoost has been developed and used by a group of active community members. Everyone is more than welcome to contribute. It is a way to make the project better and more accessible to more users.

Features

  • Easily accessible in python, R, Julia, CLI
  • Fast speed and memory efficient
    • Can be more than 10 times faster than GBM in sklearn and R
    • Handles sparse matrices, support external memory
  • Accurate prediction, and used extensively by data scientists and kagglers
  • Distributed and Portable
    • The distributed version runs on Hadoop (YARN), MPI, SGE etc.
    • Scales to billions of examples and beyond

Build

  • Run bash build.sh (you can also type make)

Version

  • Current version xgboost-0.4, a lot improvment has been made since 0.3
    • Change log in CHANGES.md
    • This version is compatible with 0.3x versions

XGBoost in Graphlab Create

About

Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Flink and DataFlow

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • C++ 41.0%
  • R 14.9%
  • Scala 14.6%
  • Python 13.5%
  • Java 7.1%
  • Cuda 4.9%
  • Other 4.0%