Linear regression is a supervised learning technique used to predict continuous valued outputs. It fits a linear equation to the training data to find the relationship between independent variables (x) and the dependent variable (y).
Gradient descent is an algorithm used to minimize the cost function in linear regression. It works by iteratively updating the parameters (θ values) in the direction of the steepest descent as calculated by the partial derivatives of the cost function with respect to each parameter. The learning rate (α) controls the step size in each iteration.
Multivariate linear regression extends the technique to problems with multiple independent variables by representing the hypothesis and parameters as vectors and matrices, allowing gradient descent to optimize the parameters to fit the linear model