XGBoost Machine Learning for Data Science and Kaggle

Udemy XGBoost Machine Learning for Data Science and Kaggle

Register & Get access to index


The future world is the AI era of machine learning, so mastering the application of machine learning is equivalent to getting a key to the future career. If you can only learn one tool or algorithm for machine learning or building predictive models now, what is this tool? Without a doubt, that is Xgboost! If you are going to participate in a Kaggle contest, what is your preferred modeling tool? Again, the answer is Xgboost! This is proven by countless experienced data scientists and new comers. Therefore, you must register for this course!

The Xgboost is so famous in Kaggle contests because of its excellent accuracy, speed and stability. For example, according to the survey, more than 70% the top kaggle winners said they have used XGBoost.

The Xgboost is really useful and performs manifold functionalities in the data science world; this powerful algorithm is so frequently utilized to predict various types of targets – continuous, binary, categorical data, it is also found Xgboost very effective to solve different multiclass or multilabel classification problems. In addition, the contests on Kaggle platform covered almost all the applications and industries in the world, such as retail business, banking, insurance, pharmaceutical research, traffic control and credit risk management.

The Xgboost is powerful, but it is not that easy to exercise it full capabilities without expert’s guidance. For example, to successfully implement the Xgboost algorithm, you also need to understand and adjust many parameter settings. For doing so, I will teach you the underlying algorithm so you are able to configure the Xgboost that tailor to different data and application scenarios. In addition, I will provide intensive lectures on feature engineering, feature selection and parameters tuning aiming at Xgboost. So, after training you should also be able to prepare the suitable data or features that can well feed the XGBoost model.

This course is really practical but not lacking in theory; we start from decision trees and its related concepts and components, transferring to constructing the gradient boot methods, then leading to the Xgboost modeling. The math and statistics are mildly applied to explain the mechanisms in all machine learning methods. We use the Python pandas data frames to deal with data exploration and cleaning. One significant feature of this course is that we have used many Python program examples to demonstrate every single knowledge point and skill you have learned in the lecture.
First release
Last update
0.00 star(s) 0 ratings