Multi-linear regression extends simple linear regression to handle multiple input features simultaneously. Instead of predicting from just one variable, we can use several features to make more accurate predictions.

In civil engineering, this is particularly valuable for projects like predicting housing costs from multiple property characteristics, estimating structural loads from various design parameters, or forecasting traffic flow from multiple environmental factors.
• Click feature boxes to enable/disable them (colored = enabled, gray = disabled). The model will automatically re-train.
• Using the plot below, compare the actual vs predicted prices. The closer the points are to the diagonal line, the better the predictions
• Try different feature combinations to see how they affect performance, and observe how metrics change with different feature sets

Which combination of features maximises R²? Why do you think that is?
Understanding the Results:

R² values: 0.0-0.3 (poor), 0.3-0.7 (moderate), 0.7+ (good fit)
MSE (Mean Squared Error): Lower values indicate better fit - measures average squared differences
MAE (Mean Absolute Error): More intuitive than MSE - average absolute prediction error
Feature importance: Notice which features contribute most to prediction accuracy
Overfitting vs Underfitting: Too few features may underfit; all features don't always improve performance
MAE=0.58
Mean Absolute Error
MSE=0.615
Mean Squared Error
R2=0.465
Coefficient of Determination

Mathematical Foundations

Multiple linear regression extends the single-variable framework to handle multiple features simultaneously. With p features, the linear model becomes y^=β0+β1x1+β2x2++βpxp, where β0 is the intercept (previously denoted α in single-variable regression) and β1,,βp are the weights for each feature.

For our housing example: Price=β0+β1Area+β2Bedrooms+β3Bathrooms+β4Stories. Each feature contributes independently to the prediction, weighted by its corresponding parameter.

Developed by Kevin Yu & Panagiotis Angeloudis