TY - JOUR
T1 - An adaptive cost-sensitive learning approach in neural networks to minimize local training–test class distributions mismatch
AU - Volk, Ohad
AU - Singer, Gonen
N1 - Publisher Copyright:
© 2023 The Author(s)
PY - 2024/3
Y1 - 2024/3
N2 - We design an adaptive learning algorithm for binary classification problems whose objective is to reduce the cost of misclassified instances derived from the consequences of errors. Our algorithm (Adaptive Cost-Sensitive Learning — AdaCSL) adaptively adjusts the loss function to bridge the difference between the class distributions between subgroups of samples in the training and validation data sets. This adjustment is made for samples with similar predicted probabilities, in such a way that the local cost decreases. This process usually leads to a reduction in cost when applied to the test data set (i.e., local training–test class distributions mismatch). We present empirical evidence that neural networks used with the proposed algorithm yield better cost results on several data sets compared to other approaches. In addition, the proposed AdaCSL algorithm can optimize evaluation metrics other than cost. We present an experiment that demonstrates how utilizing the AdaCSL algorithm generates superior accuracy results. The AdaCSL algorithm can be used for applications in which the training set is noisy or when large variability may occur between the training and validation data sets, such as the classification of disease severity for a given subject based on other subjects. Our code is available at https://github.com/OhadVolk/AdaCSL.
AB - We design an adaptive learning algorithm for binary classification problems whose objective is to reduce the cost of misclassified instances derived from the consequences of errors. Our algorithm (Adaptive Cost-Sensitive Learning — AdaCSL) adaptively adjusts the loss function to bridge the difference between the class distributions between subgroups of samples in the training and validation data sets. This adjustment is made for samples with similar predicted probabilities, in such a way that the local cost decreases. This process usually leads to a reduction in cost when applied to the test data set (i.e., local training–test class distributions mismatch). We present empirical evidence that neural networks used with the proposed algorithm yield better cost results on several data sets compared to other approaches. In addition, the proposed AdaCSL algorithm can optimize evaluation metrics other than cost. We present an experiment that demonstrates how utilizing the AdaCSL algorithm generates superior accuracy results. The AdaCSL algorithm can be used for applications in which the training set is noisy or when large variability may occur between the training and validation data sets, such as the classification of disease severity for a given subject based on other subjects. Our code is available at https://github.com/OhadVolk/AdaCSL.
KW - Adaptive loss function
KW - Cost-sensitive
KW - Misclassification costs
KW - Training-test mismatch
UR - http://www.scopus.com/inward/record.url?scp=85180533975&partnerID=8YFLogxK
U2 - 10.1016/j.iswa.2023.200316
DO - 10.1016/j.iswa.2023.200316
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:85180533975
SN - 2667-3053
VL - 21
JO - Intelligent Systems with Applications
JF - Intelligent Systems with Applications
M1 - 200316
ER -