A Data-Transparent Probabilistic Model of Temporal Propositional Abstraction

Published: 17 Sept 2025, Last Modified: 06 Nov 2025ACS 2025 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Probability theory, Propositional logic, Abstraction, Generative models, Transparency, Data scarcity, Markov chains, Hidden Markov models
Abstract: Standard probabilistic models face fundamental challenges such as data scarcity, a large hypothesis space, and poor data transparency. To address these challenges, we propose a novel probabilistic model of data-driven temporal propositional reasoning. Unlike conventional probabilistic models where data is a product of domain knowledge encoded in the probabilistic model, we explore the reverse direction where domain knowledge is a product of data encoded in the probabilistic model. This more data-driven perspective suggests no distinction between maximum likelihood parameter learning and temporal propositional reasoning. We show that our probabilistic model is equivalent to a highest-order, i.e., full-memory, Markov chain, and it can also be viewed as a hidden Markov model requiring no distinction between hidden and observable variables. We discuss that limits provide a natural and mathematically rigorous way to handle data scarcity, including the zero-frequency problem. We also discuss that a probability distribution over data generated by our probabilistic model helps data transparency by revealing influential data used in predictions. The reproducibility of this theoretical work is fully demonstrated by the included proofs.
Paper Track: Technical paper
Submission Number: 30
Loading