Libros importados con hasta 40% OFF + Envío gratis a todo USA  Ver más

menú

0
  • argentina
  • chile
  • colombia
  • españa
  • méxico
  • perú
  • estados unidos
  • internacional
portada XGBoost. The Extreme Gradient Boosting for Mining Applications (en Inglés)
Formato
Libro Físico
Editorial
Idioma
Inglés
N° páginas
60
Encuadernación
Tapa Blanda
Dimensiones
21.0 x 14.8 x 0.4 cm
Peso
0.09 kg.
ISBN13
9783668660618

XGBoost. The Extreme Gradient Boosting for Mining Applications (en Inglés)

Nonita Sharma (Autor) · Grin Verlag · Tapa Blanda

XGBoost. The Extreme Gradient Boosting for Mining Applications (en Inglés) - Sharma, Nonita

Libro Físico

$ 39.87

$ 50.50

Ahorras: $ 10.63

21% descuento
  • Estado: Nuevo
Se enviará desde nuestra bodega entre el Lunes 13 de Mayo y el Martes 14 de Mayo.
Lo recibirás en cualquier lugar de Estados Unidos entre 1 y 3 días hábiles luego del envío.

Reseña del libro "XGBoost. The Extreme Gradient Boosting for Mining Applications (en Inglés)"

Technical Report from the year 2017 in the subject Computer Science - Internet, New Technologies, grade: 8, language: English, abstract: Tree boosting has empirically proven to be a highly effective and versatile approach for data-driven modelling. The core argument is that tree boosting can adaptively determine the local neighbourhoods of the model thereby taking the bias-variance trade-off into consideration during model fitting. Recently, a tree boosting method known as XGBoost has gained popularity by providing higher accuracy. XGBoost further introduces some improvements which allow it to deal with the bias-variance trade-off even more carefully. In this research work, we propose to demonstrate the use of an adaptive procedure i.e. Learned Loss (LL) to update the loss function as the boosting proceeds. Accuracy of the proposed algorithm i.e. XGBoost with Learned Loss boosting function is evaluated using test/train method, K-fold cross validation, and Stratified cross validation method and compared with the state of the art algorithms viz. XGBoost, AdaBoost, AdaBoost-NN, Linear Regression(LR), Neural Network(NN), Decision Tree(DT), Support Vector Machine(SVM), bagging-DT, bagging-NN and Random Forest algorithms. The parameters evaluated are accuracy, Type 1 error and Type 2 error (in Percentages). This study uses total ten years of historical data from Jan 2007 to Aug 2017 of two stock market indices CNX Nifty and S&P BSE Sensex which are highly voluminous. Further, in this research work, we will investigate how XGBoost differs from the more traditional ensemble techniques. Moreover, we will discuss the regularization techniques that these methods offer and the effect these have on the models. In addition to this, we will attempt to answer the question of why XGBoost seems to win so many competitions. To do this, we will provide some arguments for why tree boosting, and in particular XGBoost, seems to be such a highly effective and versatile approa

Opiniones del libro

Ver más opiniones de clientes
  • 0% (0)
  • 0% (0)
  • 0% (0)
  • 0% (0)
  • 0% (0)

Preguntas frecuentes sobre el libro

Todos los libros de nuestro catálogo son Originales.
El libro está escrito en Inglés.
La encuadernación de esta edición es Tapa Blanda.

Preguntas y respuestas sobre el libro

¿Tienes una pregunta sobre el libro? Inicia sesión para poder agregar tu propia pregunta.

Opiniones sobre Buscalibre

Ver más opiniones de clientes