Gradient boosted decision trees can be used for both regression and classification problems in big data. For regression, the goal is to minimize the error between the predicted and actual values. The algorithm builds the prediction function iteratively by fitting a decision tree to the residuals of the previous iteration. For classification, it models the probability of class membership using a sigmoid function of the ensemble of decision trees. The principle of maximum likelihood is used to find the function that maximizes the logarithm of the likelihood of the training data.