Recognition: unknown
Deep Kalman Filters
read the original abstract
Kalman Filters are one of the most influential models of time-varying phenomena. They admit an intuitive probabilistic interpretation, have a simple functional form, and enjoy widespread adoption in a variety of disciplines. Motivated by recent variational methods for learning deep generative models, we introduce a unified algorithm to efficiently learn a broad spectrum of Kalman filters. Of particular interest is the use of temporal generative models for counterfactual inference. We investigate the efficacy of such models for counterfactual inference, and to that end we introduce the "Healing MNIST" dataset where long-term structure, noise and actions are applied to sequences of digits. We show the efficacy of our method for modeling this dataset. We further show how our model can be used for counterfactual inference for patients, based on electronic health record data of 8,000 patients over 4.5 years.
This paper has not been read by Pith yet.
Forward citations
Cited by 7 Pith papers
-
Mastering Atari with Discrete World Models
DreamerV2 reaches human-level performance on 55 Atari games by learning behaviors inside a separately trained discrete-latent world model.
-
Dream to Control: Learning Behaviors by Latent Imagination
Dreamer learns to control from images by imagining and optimizing behaviors in a learned latent world model, outperforming prior methods on 20 visual tasks in data efficiency and final performance.
-
RT-Transformer: The Transformer Block as a Spherical State Estimator
Transformer components arise as the natural solution to precision-weighted directional state estimation on the hypersphere.
-
Coupled-NeuralHP: Directional Temporal Coupling Between AI Innovation Exposure and Public Response
Coupled-NeuralHP finds that AI patent streams forecast public response trends better than baselines in one direction while the reverse link is unsupported, with no robust 2022 regime shift detected.
-
Cognitive Flexibility as a Latent Structural Operator for Bayesian State Estimation
Cognitive Flexibility is a new representation-level operator for Bayesian filters that dynamically selects latent structures via predictive scores to reduce inconsistency under mismatch while preserving the recursion ...
-
Adaptive Learned State Estimation based on KalmanNet
AM-KNet adds sensor-specific modules, hypernetwork conditioning on target type and pose, and Joseph-form covariance estimation to KalmanNet, yielding better accuracy and stability than base KalmanNet on nuScenes and V...
-
CognitiveTwin: Robust Multi-Modal Digital Twins for Predicting Cognitive Decline in Alzheimer's Disease
CognitiveTwin combines Transformer multi-modal fusion and Deep Markov Models on longitudinal AD data to deliver personalized cognitive decline predictions that are fair across demographics and robust to missing data.
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.