Abstract
Uncertain knowledge can be modeled by using graded probabilities rather than binary truth-values, but so far a completely satisfactory integration of logic and probability has been lacking. In particular the inability of confirming universal hypotheses has plagued most if not all systems so far. We address this problem head on. The main technical problem to be discussed is the following: Given a set of sentences, each having some probability of being true, what probability should be ascribed to other (query) sentences? A natural wish-list, among others, is that the probability distribution (i) is consistent with the knowledge base, (ii) allows for a consistent inference procedure and in particular (iii) reduces to deductive logic in the limit of probabilities being 0 and 1, (iv) allows (Bayesian) inductive reasoning and (v) learning in the limit and in particular (vi) allows confirmation of universally quanti- fied hypotheses/sentences. We show that probabilities satisfying (i)-(vi) exist, and present necessary and sufficient conditions (Gaifman and Cournot). The theory is a step towards a globally consistent and empirically satisfactory unification of probability and logic.
Original language | English |
---|---|
Title of host publication | IJCAI International Joint Conference on Artificial Intelligence |
Place of Publication | USA |
Publisher | AAAI Press |
Pages | 65-72 |
Edition | Peer Reviewed |
ISBN (Print) | 9781577356332 |
Publication status | Published - 2013 |
Event | 23rd International Joint Conference on Artificial Intelligence, IJCAI 2013 - Beijing China Duration: 1 Jan 2013 → … http://ijcai.org/papers13/contents.php |
Conference
Conference | 23rd International Joint Conference on Artificial Intelligence, IJCAI 2013 |
---|---|
Period | 1/01/13 → … |
Other | August 3-9 2013 |
Internet address |