Results 1 to 10 of about 1,753,158 (289)
Interpretable deep learning: interpretation, interpretability, trustworthiness, and beyond [PDF]
Deep neural networks have been well-known for their superb handling of various machine learning and artificial intelligence tasks. However, due to their over-parameterized black-box nature, it is often difficult to understand the prediction results of deep models.
Li, Xuhong +7 more
openaire +3 more sources
Analyzing Employee Attrition Using Explainable AI for Strategic HR Decision-Making
Employee attrition and high turnover have become critical challenges faced by various sectors in today’s competitive job market. In response to these pressing issues, organizations are increasingly turning to artificial intelligence (AI) to predict ...
Gabriel Marín Díaz +2 more
doaj +1 more source
The adaptation of deep learning models within safety-critical systems cannot rely only on good prediction performance but needs to provide interpretable and robust explanations for their decisions.
Domjan Barić +3 more
doaj +1 more source
Prediction or interpretability?
The journal published a review of the literature on recursive partition in epidemiological research comparing two decision tree methods: classification and regression trees (CARTs) and conditional inference trees (CITs).
Stefano Nembrini
doaj +1 more source
Re-interpreting rules interpretability
Abstract Trustworthy machine learning requires a high level of interpretability of machine learning models, yet many models are inherently black-boxes. Training interpretable models instead—or using them to mimic the black-box model—seems like a viable solution. In practice, however, these interpretable models are still unintelligible
Adilova, L. +3 more
openaire +2 more sources
Summary: Free-text clinical notes in electronic health records are more difficult for data mining while the structured diagnostic codes can be missing or erroneous. To improve the quality of diagnostic codes, this work extracts diagnostic codes from free-
Xianghao Zhan +3 more
doaj +1 more source
A New Interpretable Unsupervised Anomaly Detection Method Based on Residual Explanation
Despite the superior performance in modeling complex patterns to address challenging problems, the black-box nature of Deep Learning (DL) methods impose limitations to their application in real-world critical domains.
David F. N. Oliveira +8 more
doaj +1 more source
Double Prior Network for Multidegradation Remote Sensing Image Super-Resolution
Image super-resolution (SR) is widely used in remote sensing because it can effectively increase image details. Neural networks have shown remarkable performance in recent years, benefitting from their end-to-end training.
Mengyang Shi +3 more
doaj +1 more source
The influence of Artificial Intelligence is growing, as is the need to make it as explainable as possible. Explainability is one of the main obstacles that AI faces today on the way to more practical implementation.
Jurgita Černevičienė +1 more
doaj +1 more source
Emulating quantum dynamics with neural networks via knowledge distillation
We introduce an efficient training framework for constructing machine learning-based emulators and demonstrate its capability by training an artificial neural network to predict the time evolution of quantum wave packets propagating through a potential ...
Yu Yao +6 more
doaj +1 more source

