| Veröffentlichte Version Download ( PDF | 690kB) | Lizenz: Creative Commons Namensnennung 4.0 International |
Taming numerical imprecision by adapting the KL divergence to negative probabilities
Pfahler, Simon, Georg, Peter, Schill, Rudolf
, Klever, Maren, Grasedyck, Lars, Spang, Rainer
und Wettig, Tilo
(2024)
Taming numerical imprecision by adapting the KL divergence to negative probabilities.
Statistics and Computing 34 (168).
Veröffentlichungsdatum dieses Volltextes: 04 Apr 2024 04:21
Artikel
DOI zum Zitieren dieses Dokuments: 10.5283/epub.58034
Zusammenfassung
The Kullback-Leibler (KL) divergence is frequently used in data science. For discrete distributions on large state spaces, approximations of probability vectors may result in a few small negative entries, rendering the KL divergence undefined. We address this problem by introducing a parameterized family of substitute divergence measures, the shifted KL (sKL) divergence measures. Our approach is ...
The Kullback-Leibler (KL) divergence is frequently used in data science. For discrete distributions on large state spaces, approximations of probability vectors may result in a few small negative entries, rendering the KL divergence undefined. We address this problem by introducing a parameterized family of substitute divergence measures, the shifted KL (sKL) divergence measures. Our approach is generic and does not increase the computational overhead. We show that the sKL divergence shares important theoretical properties with the KL divergence and discuss how its shift parameters should be chosen. If Gaussian noise is added to a probability vector, we prove that the average sKL divergence converges to the KL divergence for small enough noise. We also show that our method solves the problem of negative entries in an application from computational oncology, the optimization of Mutual Hazard Networks for cancer progression using tensor-train approximations.
Alternative Links zum Volltext
Beteiligte Einrichtungen
Details
| Dokumentenart | Artikel | ||||
| Titel eines Journals oder einer Zeitschrift | Statistics and Computing | ||||
| Verlag: | Springer Nature | ||||
|---|---|---|---|---|---|
| Band: | 34 | ||||
| Nummer des Zeitschriftenheftes oder des Kapitels: | 168 | ||||
| Datum | 13 August 2024 | ||||
| Institutionen | Medizin > Institut für Funktionelle Genomik > Lehrstuhl für Statistische Bioinformatik (Prof. Spang) Informatik und Data Science > Fachbereich Bioinformatik > Lehrstuhl für Statistische Bioinformatik (Prof. Spang) Physik > Institut für Theoretische Physik > Lehrstuhl Professor Braun > Arbeitsgruppe Tilo Wettig | ||||
| Identifikationsnummer |
| ||||
| Verwandte URLs |
| ||||
| Stichwörter / Keywords | Kullback-Leibler divergence, Approximate Bayesian computation, Statistical optimization, Mutual Hazard Networks, Tensor trains | ||||
| Dewey-Dezimal-Klassifikation | 000 Informatik, Informationswissenschaft, allgemeine Werke > 004 Informatik 500 Naturwissenschaften und Mathematik > 530 Physik | ||||
| Status | Veröffentlicht | ||||
| Begutachtet | Ja, diese Version wurde begutachtet | ||||
| An der Universität Regensburg entstanden | Zum Teil | ||||
| URN der UB Regensburg | urn:nbn:de:bvb:355-epub-580343 | ||||
| Dokumenten-ID | 58034 |
Downloadstatistik
Downloadstatistik