koptimizer/my_PaperLog

XGBoost: A Scalable Tree Boosting System

koptimizer opened this issue ยท 0 comments

๐Ÿ“‹ ๋…ผ๋ฌธ์˜ ์ •๋ณด๋ฅผ ์•Œ๋ ค์ฃผ์„ธ์š”.

  • XGBoost A Scalable Tree Boosting System
  • Tianqi Chen & Carlos Guestrin
  • KDD '16: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
  • 2016-08

๐Ÿ“ƒ Abstract

Tree boosting is a highly effective and widely used machine learning method. In this paper, we describe a scalable end-to-end tree boosting system called XGBoost, which is used widely by data scientists to achieve state-of-the-art results on many machine learning challenges. We propose a novel sparsity-aware algorithm for sparse data and weighted quantile sketch for approximate tree learning. More importantly, we provide insights on cache access patterns, data compression and sharding to build a scalable tree boosting system. By combining these insights, XGBoost scales beyond billions of examples using far fewer resources than existing systems.

๐Ÿ”Ž ์–ด๋–ค ๋…ผ๋ฌธ์ธ์ง€ ์†Œ๊ฐœํ•ด์ฃผ์„ธ์š”.

  • Scalableํ•œ ๋จธ์‹ ๋Ÿฌ๋‹ ํ”„๋ ˆ์ž„์›์ธ XGboost์— ๋Œ€ํ•œ ๋…ผ๋ฌธ์ž…๋‹ˆ๋‹ค.
  • ๊ฒฝ์ง„๋Œ€ํšŒ๋‚˜ ๊ฐ์ข… ๋ถ„์„์—์„œ ๋‹น์—ฐํ•˜๊ฒŒ ์ป๋˜ ํ”„๋ ˆ์ž„์›์ด๋ผ์„œ ๋„๋ฐ์ฒด ์–ด๋–ป๊ฒŒ ๋งŒ๋“ค์—ˆ๊ธธ๋ž˜ ๊ทธ๋ ‡๊ฒŒ ์„ฑ๋Šฅ์ด ์ข‹์€ ๊ฒƒ์ธ์ง€ ํ™•์ธํ•ด๋ณด๋ ค๊ณ  ํ•ฉ๋‹ˆ๋‹ค.

๐Ÿ”‘ ํ•ต์‹ฌ ํ‚ค์›Œ๋“œ๋ฅผ ์ ์–ด์ฃผ์„ธ์š”.

  • Scalable, Gradient boosting, Sparsity-aware, parallel

๐Ÿ“Ž URL