• Stars
    star
    25,402
  • Rank 757 (Top 0.02 %)
  • Language
    C++
  • License
    Apache License 2.0
  • Created about 10 years ago
  • Updated 15 days ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow

eXtreme Gradient Boosting

Build Status XGBoost-CI Documentation Status GitHub license CRAN Status Badge PyPI version Conda version Optuna Twitter OpenSSF Scorecard

Community | Documentation | Resources | Contributors | Release Notes

XGBoost is an optimized distributed gradient boosting library designed to be highly efficient, flexible and portable. It implements machine learning algorithms under the Gradient Boosting framework. XGBoost provides a parallel tree boosting (also known as GBDT, GBM) that solve many data science problems in a fast and accurate way. The same code runs on major distributed environment (Kubernetes, Hadoop, SGE, Dask, Spark, PySpark) and can solve problems beyond billions of examples.

License

ยฉ Contributors, 2021. Licensed under an Apache-2 license.

Contribute to XGBoost

XGBoost has been developed and used by a group of active community members. Your help is very valuable to make the package better for everyone. Checkout the Community Page.

Reference

  • Tianqi Chen and Carlos Guestrin. XGBoost: A Scalable Tree Boosting System. In 22nd SIGKDD Conference on Knowledge Discovery and Data Mining, 2016
  • XGBoost originates from research project at University of Washington.

Sponsors

Become a sponsor and get a logo here. See details at Sponsoring the XGBoost Project. The funds are used to defray the cost of continuous integration and testing infrastructure (https://xgboost-ci.net).

Open Source Collective sponsors

Backers on Open Collective Sponsors on Open Collective

Sponsors

[Become a sponsor]

NVIDIA

Backers

[Become a backer]

More Repositories

1

dgl

Python package built to ease deep learning on graph, on top of existing DL frameworks.
Python
12,855
star
2

gluon-cv

Gluon CV Toolkit
Python
5,724
star
3

gluon-nlp

NLP made easy
Python
2,548
star
4

nnvm

C++
1,655
star
5

decord

An efficient video loader for deep learning with smart shuffling that's super easy to digest
C++
1,567
star
6

ps-lite

A lightweight parameter server interface
C++
1,502
star
7

minpy

NumPy interface with mixed backend execution
Python
1,112
star
8

mshadow

Matrix Shadow:Lightweight CPU/GPU Matrix and Tensor Template Library in C++/CUDA for (Deep) Machine Learning
C++
1,098
star
9

cxxnet

move forward to https://github.com/dmlc/mxnet
C++
1,025
star
10

dmlc-core

A common bricks library for building scalable and portable distributed machine learning.
C++
861
star
11

dlpack

common in-memory tensor structure
Python
829
star
12

treelite

Universal model exchange and serialization format for decision tree forests
C++
703
star
13

minerva

Minerva: a fast and flexible tool for deep learning on multi-GPU. It provides ndarray programming interface, just like Numpy. Python bindings and C++ bindings are both available. The resulting code can be run on CPU or GPU. Multi-GPU support is very easy.
C++
689
star
14

parameter_server

moved to https://github.com/dmlc/ps-lite
C++
645
star
15

mxnet-notebooks

Notebooks for MXNet
Jupyter Notebook
613
star
16

rabit

Reliable Allreduce and Broadcast Interface for distributed machine learning
C++
507
star
17

mxnet.js

MXNetJS: Javascript Package for Deep Learning in Browser (without server)
JavaScript
435
star
18

MXNet.jl

MXNet Julia Package - flexible and efficient deep learning in Julia
372
star
19

tensorboard

Standalone TensorBoard for visualizing in deep learning
Python
370
star
20

wormhole

Deprecated
C++
341
star
21

mxnet-memonger

Sublinear memory optimization for deep learning, reduce GPU memory cost to train deeper nets
Python
308
star
22

difacto

Distributed Factorization Machines
C++
296
star
23

XGBoost.jl

XGBoost Julia Package
Julia
280
star
24

mxnet-model-gallery

Pre-trained Models of DMLC Project
266
star
25

HalideIR

Symbolic Expression and Statement Module for new DSLs
C++
202
star
26

GNNLens2

Visualization tool for Graph Neural Networks
TypeScript
202
star
27

mxnet-gtc-tutorial

MXNet Tutorial for NVidia GTC 2016.
Jupyter Notebook
131
star
28

experimental-lda

C++
127
star
29

MXNet.cpp

C++ interface for mxnet
C++
114
star
30

experimental-mf

cache-friendly multithread matrix factorization
C++
86
star
31

web-data

The repo to host all the web data including images for documents in dmlc projects.
Jupyter Notebook
80
star
32

nnvm-fusion

Kernel Fusion and Runtime Compilation Based on NNVM
C++
64
star
33

dmlc.github.io

HTML
27
star
34

cub

Cuda
18
star
35

tl2cgen

TL2cgen (TreeLite 2 C GENerator) is a model compiler for decision tree models
C++
13
star
36

mxnet-deepmark

Benchmark speed and other issues internally, before push to deep-mark
Python
7
star
37

mxnet-examples

MXNet Example
6
star
38

xgboost-bench

Python
4
star
39

drat

Drat Repository for DMLC R packages
4
star
40

nn-examples

1
star
41

gluon-nlp-notebooks

1
star
42

docs-redirect-for-mxnet

redirect mxnet.readthedocs.io to mxnet.io
Python
1
star