The Analysis of Deep Neural Networks by Information Theory: From Explainability to Generalization

Shujian Yu*

*Corresponding author for this work

Research output: Chapter in Book / Report / Conference proceedingConference contributionAcademicpeer-review

Abstract

Despite their great success in many artificial intelligence tasks, deep neural networks (DNNs) still suffer from a few limitations, such as poor generalization behavior for out-of-distribution (OOD) data and the “black-box” nature. Information theory offers fresh insights to solve these challenges. In this short paper, we briefly review the recent developments in this area, and highlight our contributions.

Original languageEnglish
Title of host publicationAAAI-23 Special Programs, IAAI-23, EAAI-23, Student Papers and Demonstrations
Subtitle of host publicationThirty-Seventh AAAI Conference on Artificial Intelligence Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence Thirteenth Symposium on Educational Advances in Artificial Intelligence
EditorsBrian Williams, Yiling Chen, Jennifer Neville
PublisherAAAI Press
Pages15462-15462
Number of pages1
Volume13
ISBN (Electronic)9781577358800
DOIs
Publication statusPublished - 2023
Event37th AAAI Conference on Artificial Intelligence, AAAI 2023 - Washington, United States
Duration: 7 Feb 202314 Feb 2023

Publication series

NameProceedings of the xxth AAAI Conference on Artificial Intelligence
Number13
Volume37

Conference

Conference37th AAAI Conference on Artificial Intelligence, AAAI 2023
Country/TerritoryUnited States
CityWashington
Period7/02/2314/02/23

Bibliographical note

Publisher Copyright:
Copyright © 2023, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.

Fingerprint

Dive into the research topics of 'The Analysis of Deep Neural Networks by Information Theory: From Explainability to Generalization'. Together they form a unique fingerprint.

Cite this