Spaces:
Sleeping
Sleeping
AjithKSenthil
commited on
Commit
·
00a7e28
1
Parent(s):
76aaad4
added comments for interpretation of validation
Browse files
ChatAttachmentAnalysisWithValidation.py
CHANGED
@@ -37,3 +37,18 @@ test_preds = rfr.predict(X_test)
|
|
37 |
test_mse = mean_squared_error(y_test, test_preds)
|
38 |
test_mae = mean_absolute_error(y_test, test_preds)
|
39 |
print(f"Test MSE: {test_mse:.2f}, Test MAE: {test_mae:.2f}")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
test_mse = mean_squared_error(y_test, test_preds)
|
38 |
test_mae = mean_absolute_error(y_test, test_preds)
|
39 |
print(f"Test MSE: {test_mse:.2f}, Test MAE: {test_mae:.2f}")
|
40 |
+
|
41 |
+
# The validation set is used during the model building process to assess how well the model is performing.
|
42 |
+
# It helps tune the model's hyperparameters, prevent overfitting and select the best performing model.
|
43 |
+
# A lower Mean Squared Error (MSE) and Mean Absolute Error (MAE) on the validation set indicate a better fit of the model.
|
44 |
+
# These metrics measure the difference between the predicted and actual values.
|
45 |
+
# Validation MSE: The average of the squares of the differences between the predicted and actual values in the validation set.
|
46 |
+
# Validation MAE: The average of the absolute differences between the predicted and actual values in the validation set.
|
47 |
+
|
48 |
+
# Once we are confident about our model's parameters and performance, we test it on unseen data - the test set.
|
49 |
+
# The test set provides the final measure of the model's performance.
|
50 |
+
# It helps us understand how the model will generalize to new, unseen data.
|
51 |
+
# A lower Mean Squared Error (MSE) and Mean Absolute Error (MAE) on the test set also indicate a better fit of the model.
|
52 |
+
# Test MSE: The average of the squares of the differences between the predicted and actual values in the test set.
|
53 |
+
# Test MAE: The average of the absolute differences between the predicted and actual values in the test set.
|
54 |
+
# Note that if the model's performance on the test set is significantly worse than on the training set, it may be an indication of overfitting.
|