Login
Congrats in choosing to up-skill for your bright career! Please share correct details.
Home / Blog / Data Science Digital Book / Model Evaluation Techniques
Bharani Kumar Depuru is a well known IT personality from Hyderabad. He is the Founder and Director of AiSPRY and 360DigiTMG. Bharani Kumar is an IIT and ISB alumni with more than 18+ years of experience, he held prominent positions in the IT elites like HSBC, ITC Infotech, Infosys, and Deloitte. He is a prevalent IT consultant specializing in Industrial Revolution 4.0 implementation, Data Analytics practice setup, Artificial Intelligence, Big Data Analytics, Industrial IoT, Business Intelligence and Business Management. Bharani Kumar is also the chief trainer at 360DigiTMG with more than Ten years of experience and has been making the IT transition journey easy for his students. 360DigiTMG is at the forefront of delivering quality education, thereby bridging the gap between academia and industry.
The set of error functions below can be used to assess the model if the output variable 'Y' is continuous.
Error = Predicted Value - Actual Value (Actual Value is also called as Ground Truth Value)
Click here to learn Data Science in Hyderabad
MAEin-sample naive is the mean absolute error produced by a naive forecast
If the ‘Y’ is Discrete variable (Classification Models) then we can use the following list:
Click here to learn Data Analytics in Bangalore
can be used for multi-class classification models as well as binary classifications.
To compare anticipated values with actual values, confusion matrix is employed.
Click here to learn Data Analytics in Hyderabad
Click here to learn Artificial Intelligence in Bangalore
Click here to learn Artificial Intelligence in Hyderabad
Click here to learn Machine Learning in Hyderabad
F1 score is the harmonic mean of precision and recall. Closer the ‘F1’ value to 1, better the accuracy.
A cross table or contingency table are other names for a confusion matrix. Here is a multi-class categorization issue example.
Values off the diagonal are incorrectly predicted, whereas values along the diagonal are correctly predicted.
Since World War II, the Receiver Operating Characteristic Curve has been employed to discern between genuine signals and false alarms.
The 'True Positive Rate (TPR)' and 'False Positive Rate (FPR)' are plotted on the Y-axis and X-axis, respectively, of the ROC curve.
Accuracy is represented graphically using the ROC curve.
The cut-off value is also determined using the ROC curve.
Examples include: Risk Neutral: The probability should be > 0.5 as the cut-off value to classify a customer under the "will default" category; Risk Taking: The probability should be > 0.8 as the cut-off value to classify a customer under the "will default" category; or Risk Averse: The probability should be > 0.3 as the cut-off value to classify a customer under the "will default" category.
AUC (Area Under the Curve) may be determined numerically if one needs to assess the accuracy.
0.9 - 1.0 = A (outstanding)
0.8 - 0.9 = B (excellent/good)
0.7 - 0.8 = C (acceptable/fair)
0.6 - 0.7 = D (poor)
0.5 - 0.6 = F (no discrimination)
Click here to learn Data Science Course, Data Science Course in Hyderabad, Data Science Course in Bangalore
360DigiTMG - Data Science, IR 4.0, AI, Machine Learning Training in Malaysia
Level 16, 1 Sentral, Jalan Stesen Sentral 5, Kuala Lumpur Sentral, 50470 Kuala Lumpur, Wilayah Persekutuan Kuala Lumpur, Malaysia
+60 19-383 1378
Didn’t receive OTP? Resend
Let's Connect! Please share your details here