TY - GEN
T1 - Learning abductive reasoning using random examples
AU - Juba, Brendan
N1 - Publisher Copyright:
© 2016, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
PY - 2016
Y1 - 2016
N2 - We consider a new formulation of abduction in which degrees of "plausibility" of explanations, along with the rules of the domain, are learned from concrete examples (settings of attributes). Our version of abduction thus falls in the "learning to reason" framework of Khardon and Roth. Such approaches enable us to capture a natural notion of "plausibility" in a domain while avoiding the extremely difficult problem of specifying an explicit representation of what is "plausible." We specifically consider the question of which syntactic classes of formulas have efficient algorithms for abduction. We find that the class of k-DNF explanations can be found in polynomial time for any fixed k; but, we also find evidence that even weak versions of our abduction task are intractable for the usual class of conjunctions. This evidence is provided by a connection to the usual, inductive PAC-learning model proposed by Valiant. We also consider an exceptiontolerant variant of abduction. We observe that it is possible for polynomial-Time algorithms to tolerate a few adversarially chosen exceptions, again for the class of k-DNF explanations. All of the algorithms we study are particularly simple, and indeed are variants of a rule proposed by Mill.
AB - We consider a new formulation of abduction in which degrees of "plausibility" of explanations, along with the rules of the domain, are learned from concrete examples (settings of attributes). Our version of abduction thus falls in the "learning to reason" framework of Khardon and Roth. Such approaches enable us to capture a natural notion of "plausibility" in a domain while avoiding the extremely difficult problem of specifying an explicit representation of what is "plausible." We specifically consider the question of which syntactic classes of formulas have efficient algorithms for abduction. We find that the class of k-DNF explanations can be found in polynomial time for any fixed k; but, we also find evidence that even weak versions of our abduction task are intractable for the usual class of conjunctions. This evidence is provided by a connection to the usual, inductive PAC-learning model proposed by Valiant. We also consider an exceptiontolerant variant of abduction. We observe that it is possible for polynomial-Time algorithms to tolerate a few adversarially chosen exceptions, again for the class of k-DNF explanations. All of the algorithms we study are particularly simple, and indeed are variants of a rule proposed by Mill.
UR - https://www.scopus.com/pages/publications/85007164461
M3 - Conference contribution
AN - SCOPUS:85007164461
T3 - 30th AAAI Conference on Artificial Intelligence, AAAI 2016
SP - 999
EP - 1007
BT - 30th AAAI Conference on Artificial Intelligence, AAAI 2016
PB - AAAI press
T2 - 30th AAAI Conference on Artificial Intelligence, AAAI 2016
Y2 - 12 February 2016 through 17 February 2016
ER -