Show simple item record

Files in this item

Thumbnail

Item metadata

dc.contributor.authorConnor, Richard
dc.contributor.authorDearle, Al
dc.contributor.authorClaydon, Ben
dc.contributor.authorVadicamo, Lucia
dc.date.accessioned2024-06-03T13:30:01Z
dc.date.available2024-06-03T13:30:01Z
dc.date.issued2024-05-30
dc.identifier302500976
dc.identifier705b9e5e-94e6-49bb-acbb-5c5c7bdf1583
dc.identifier.citationConnor , R , Dearle , A , Claydon , B & Vadicamo , L 2024 , ' Correlations of cross-entropy loss in machine learning ' , Entropy , vol. 26 , no. 6 , 491 . https://doi.org/10.3390/e26060491en
dc.identifier.issn1099-4300
dc.identifier.urihttps://hdl.handle.net/10023/29980
dc.description.abstractCross-entropy loss is crucial in training many deep neural networks. In this context, we show a number of novel and strong correlations among various related divergence functions. In particular, we demonstrate that, in some circumstances, (a) cross-entropy is almost perfectly correlated with the little-known triangular divergence, and (b) cross-entropy is strongly correlated with the Euclidean distance over the logits from which the softmax is derived. The consequences of these observations are as follows. First, triangular divergence may be used as a cheaper alternative to cross-entropy. Second, logits can be used as features in a Euclidean space which is strongly synergistic with the classification process. This justifies the use of Euclidean distance over logits as a measure of similarity, in cases where the network is trained using softmax and cross-entropy. We establish these correlations via empirical observation, supported by a mathematical explanation encompassing a number of strongly related divergence functions.
dc.format.extent16
dc.format.extent2747131
dc.language.isoeng
dc.relation.ispartofEntropyen
dc.subjectSoftmaxen
dc.subjectCross-entropyen
dc.subjectf-divergenceen
dc.subjectKullback-Liebler divergenceen
dc.subjectJensen-Shannon 12 divergenceen
dc.subjectTriangular divergenceen
dc.subjectQA75 Electronic computers. Computer scienceen
dc.subjectT-NDASen
dc.subject.lccQA75en
dc.titleCorrelations of cross-entropy loss in machine learningen
dc.typeJournal articleen
dc.contributor.institutionUniversity of St Andrews. School of Computer Scienceen
dc.identifier.doi10.3390/e26060491
dc.description.statusPeer revieweden
dc.identifier.urlhttps://www.mdpi.com/1099-4300/26/6/491en


This item appears in the following Collection(s)

Show simple item record