neural-symbolic learning and reasoning

24
Neural-Symbolic Learning and Reasoning DAGSTUHL SEMINAR 14381, SEPTEMBER 2014 ARTUR D’AVILA GARCEZ AAAI SPRING SYMPOSIUM, STANFORD UNIVERSITY, 23 MARCH 2015 1

Upload: others

Post on 17-Mar-2022

7 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Neural-Symbolic Learning and Reasoning

Neural-Symbolic Learning and ReasoningDAGSTUHL SEMINAR 14381, SEPTEMBER 2014

ARTUR D’AVILA GARCEZ

AAAI SPRING SYMPOSIUM, STANFORD UNIVERSITY, 23 MARCH 2015 1

Page 2: Neural-Symbolic Learning and Reasoning

Neural-Symbolic Learning and Reasoning: Contributions and ChallengesArtur d’Avila Garcez1, Tarek R. Besold2, Luc de Raedt3, Peter Földiak4, Pascal Hitzler5, Thomas Icard6, Kai-Uwe Kühnberger2, Luis C. Lamb7, Risto Miikkulainen8, Daniel L. Silver9

Knowledge representation: computer science logic

Consolidation: knowledge extraction and transfer learning

Killer app: big data + descriptions (e.g. reasoning about video and audio)

Neural-Symbolic Learning and Reasoning Association: www.neural-symbolic.org

Page 3: Neural-Symbolic Learning and Reasoning

Representation precedes Learning

We need a language for describing the alternative algorithms that a network of neurons may be implementing…

Computer Science Logic + Neural Computation

GOAL of NSI: Learning from experience and reasoning about what has been learned from an uncertain environment in a computationally efficient way

Les Valiant

Page 4: Neural-Symbolic Learning and Reasoning

History of neural-symbolic integration (1)1988: P Smolensky, On the proper treatment of connectionism, BBS:11(1); J McCarthy (commentary), Epistemological challenges for connectionism

1990: G Hinton, Preface to the special issue on connectionist symbol processing, Artificial Intelligence 46,1-4

1993: L Shastri and V Ajjanagadde, From simple associations to systematic reasoning: A connectionist representation of rules, variables and dynamic bindings using temporal synchrony, BBS:16 (SHRUTI)

1994: G Towell, J Shavlik, Knowledge-Based Artificial Neural Networks. Artif. Intell. 70(1-2): 119-165 (KBANN)

1997: M Craven, J Shavlik, Understanding Time-Series Networks: A Case Study in Rule Extraction. Int. J. Neural Syst. 8(4): 373-384 (TREPAN)

2001: A Browne, R Sun, Connectionist inference models. Neural Networks 14(10): 1331-1355

2002: A d'Avila Garcez, K. Broda and D Gabbay, Neural-Symbolic Learning Systems: Foundations and Applications, Perspectives in Neural Computing, Springer-Verlag. (CILP)

Page 5: Neural-Symbolic Learning and Reasoning

History of neural-symbolic integration (2)2006: A d'Avila Garcez, L Lamb, A Connectionist Computational Model for Epistemic and Temporal Reasoning. Neural Computation 18(7): 1711-1738

2007: S Bader, P Hitzler, S Hölldobler, A. Witzel, A Fully Connectionist Model Generator for Covered First-Order Logic Programs. IJCAI 2007: 666-671

2007: Y Bengio, Y LeCun, Scaling learning algorithms towards AI. Large-scale kernel machines 34(5) (Representation Learning)

2009: A d'Avila Garcez, L Lamb, D. Gabbay, Neural-Symbolic Cognitive Reasoning. Cognitive Technologies, Springer

2010: D. Endres, P. Foldiak, U. Priss, An application of formal concept analysis to semantic neural decoding. Annals of Mathematics and Artificial Intelligence 57(3-4): 233-248

2012: G Pinkas, P Lima, S Cohen, A Dynamic Binding Mechanism for Retrieving and Unifying Complex Predicate-Logic Knowledge. ICANN (1) 2012: 482-490

2014: M. Franca, G. Zaverucha, A d'Avila Garcez, Fast relational learning using bottom clause propositionalizationwith artificial neural networks. Machine Learning 94(1): 81-104 (CILP++)

Page 6: Neural-Symbolic Learning and Reasoning

Neural-Symbolic Cycle

CILP++ system (Connectionist Inductive Logic Programming): download from http://sourceforge.net/projects/cilppp/

Background knowledge

Revised knowledge

Neural Network

Trained Network

Data

Consolidation

Translation

Extraction

Page 7: Neural-Symbolic Learning and Reasoning

Goal-directed Reasoning and Learningfactorial(0,1).

factorial(N,F) :- N>0, N1 is N-1, factorial(N1,F1), F is N * F1.

?- factorial(3,W).

W=6

Inductive Logic Programming in a Neural Network (CILP++)

Teleo-reactive programs (N Nilsson) for cognitive robotics(The order of the rules is important; higher rules have priority)

Tightly coupled and loosely coupled neural-symbolic systems

Page 8: Neural-Symbolic Learning and Reasoning

Neural-Symbolic SystemsNeural-symbolic methodology: translation algorithms to and from symbolic and connectionist models

In search of robustness and explanations; “combining the logical nature of reasoning and the statistical nature of learning”, L. Valiant

Logic ProgrammingNonmonotonic logic

Modal, temporal, epistemic, intuitionistic logicFirst-order logic

Higher-order logicAbductive reasoning

Argumentation

Feedforward networksHopfield networks

Elman networksBoltzmann machines

Recurrent Temporal RBMsARTMAPS

Deep Belief Networks

Page 9: Neural-Symbolic Learning and Reasoning

CILP Translation Algorithm

THEOREM 1: For any logic program P there exists a neural network N such that N computes P

Page 10: Neural-Symbolic Learning and Reasoning

Early application

Background Knowledge (14 rules):

Promoter IF Contact AND Conformation

Contact IF Minus10 AND Minus35

Minus35 IF @-36'ttgac'

Minus35 IF @-37'cttgac'

Conformation IF @-47'caa*tt*ac' AND @-22'g***t*c' AND @-8'gcgcc*cc'

10-fold cross-validation on set of 106 examples

CILP networks learn faster than backprop and KBANN, and perform slightly better than backpropand KBANN. We attribute this to the soundness of the CILP translation (i.e. the above theorem).

Page 11: Neural-Symbolic Learning and Reasoning

CILP Extraction AlgorithmKnowledge is extracted by querying the trained network;

A partial ordering helps guide the search, reducing complexity on the average case;

A proof of soundness guarantees that the rules approximate the network;

Rule simplification and visualization as state transition diagram can help experts validate results

Page 12: Neural-Symbolic Learning and Reasoning

Extensions of CILPModal, Temporal, Epistemic, Intuitionistic, Preference Logic, Abductive Reasoning, Value-based Argumentation.

New applications include: normative reasoning (robocup), temporal logic learning (model checking), software model adaptation (business process evolution), training and assessment in simulators (driving test).

CILP network ensembles;Represent and compute possible worlds; Modularity for learning;Modelling uncertainty through disjunctive information

Page 13: Neural-Symbolic Learning and Reasoning

Connectionist Modal/Temporal Logic1,2

1 Garcez, Lamb, Gabbay. Connectionist Modal Logic. Theoretical Computer Science, 371: 34-53, 2007.2 Garcez, Lamb. Connectionist Model for Epistemic and Temporal Reasoning. Neural Computation, 18:1711-1738, July 2006.

Use CILP network ensembles;

Propositional Modal Logic = decidable fragment of FOL with two variables

Full solution of Muddy Children puzzle

THEOREM 2: For any modal/temporal logic program P there exists an ensemble of neural networks N such that N computes P.

Page 14: Neural-Symbolic Learning and Reasoning

Fibring Neural Networks3

(neuromodulation?)A neuron that is a network!

Fibred networks approximate any polynomial function in an unbounded domain, as opposed to each of A, B,… which are universal approximators in compact domains.

3 Garcez, Gabbay. Fibring Neural Networks. AAAI 2004, July 2004.

Page 15: Neural-Symbolic Learning and Reasoning

Fibred Network Ensembles4

4 Garcez, Lamb, Gabbay. Neural-Symbolic Cognitive Reasoning. Springer, 2009.

Levels of abstraction

Modularity

Expressiveness

Relations and Specializations

Page 16: Neural-Symbolic Learning and Reasoning

Applications (1)

Training and Assessment in Simulators:

Learning new information from observation of experts and trainees at task execution and reasoning about this information online to provide feedback to the user

Uses knowledge insertion and extraction from RTRBMs

L. de Penning, A. S. d'Avila Garcez, L. C. Lamb and J. J. Meyer. A Neural-Symbolic Cognitive

Agent for Online Learning and Reasoning. IJCAI'11, July 2011

Page 17: Neural-Symbolic Learning and Reasoning

Applications (2)Learning Normative Rules of the RoboCup Competition

Software Model Verification and Adaptation

G. Boella, S. Colombo-Tosatto, A. S. d'Avila Garcez, V. Genovese, A. Perotti and L. van der Torre. Learning and Reasoning about Norms using Neural-Symbolic Systems. In Proc. AAMAS'12, Valencia, Spain, June 2012.

R. V. Borges, A. S. d'Avila Garcez and L. C. Lamb. Learning and Representing Temporal Knowledge in Recurrent Networks. IEEE Transactions on Neural Networks, 22(12):2409 - 2421, December 2011.

A. Perotti, A. S. d'Avila Garcez and Guido Boella, Neural Networks for Runtime Verification. The 2014 International Joint Conference on Neural Networks (IJCNN 2014), IEEE WCCI, Beijing, China, July 2014.

Page 18: Neural-Symbolic Learning and Reasoning

Relational Learning – CILP++A neural-symbolic system that can solve ILP problems (by translating BK, mode declarations and ILP examples into vectors for training with backpropagation using CILP networks)

M. França, G. Zaverucha, A. S. d'Avila Garcez, Fast relational learning using bottom clause propositionalization with artificial neural networks. Machine Learning 94(1): 81-104 (2014)

Page 19: Neural-Symbolic Learning and Reasoning

Work in progressEfficient First Order Logic Learning: encoding vs. propositionalisation?

◦ A number of attempts at first-order (and higher-order) logic learning via encoding: semantic approach (Hitzler et al), fibring, Topos (Osnabruck), association rules (SHRUTI), unification (Pinkas, etc.)

◦ Alternatively, relational learning via CILP++ (relevant for the analysis of complex networks: drug design in bioinformatics, link analysis in social networks, etc.)

◦ CILP++ (download it from sourceforge) competition: (probabilistic) ILP, MLNs, BLOG, StarAI (lifted inference)

Effective Knowledge Extraction from Very Large Networks: understanding multiple layers of representation, including semi-supervised learning and deep networks ◦ Visual intelligence applications: describing the actions in videos

Page 20: Neural-Symbolic Learning and Reasoning

Knowledge extraction from RBMs[Pinkas 1995]: rule extraction from symmetric networks using penalty logic; proved equivalence between conjunctive normal form and energy functions

[Penning et al 2011]: extraction of temporal logic rules from RTRBMs using sampling; extracts rules of the form:

hypothesis(t) iff belief(1) Ʌ ... Ʌ belief(n) Ʌ hypothesis(t-1)

[Tran and Garcez, 2012, 2014]: rule extraction from DBNs using confidence-value similar to penalty logic but maintaining implicational form; extraction without sampling

Page 21: Neural-Symbolic Learning and Reasoning

Low-cost representation for RBMs…Decompose the network into rules (pruning weights; averaging the absolute values of the remaining weights)

Page 22: Neural-Symbolic Learning and Reasoning

…guaranteed to minimise information loss

P Norvig, On Chomsky and the Two Cultures of Statistical Learning, Ninth Neural-Symbolic Learning and Reasoning Workshop NeSy13, IJCAI-13, Beijing, China, August 2013

Page 23: Neural-Symbolic Learning and Reasoning

ConclusionTo integrate the statistical nature of learning and the logical nature of reasoning

To develop effective computational systems for integrated reasoning and learning

To apply and evaluate such systems in complex new tasks…

Page 24: Neural-Symbolic Learning and Reasoning

Advertisements

Neural Relational Learning through Semi-Propositionalization of Bottom Clauses, M. Franca, A. d’Avila Garcez and G. Zaverucha (Poster at AAAI Spring KRR 2015)

Journal of Logic and Computation: learning and reasoning corner, A. d’Avila Garcezand L. Valiant (editors)

NeSy15 at IJCAI 2015, Buenos Aires, http://www.neural-symbolic.org/NeSy15/

NeSy 16 in New York (60 years of the Dartmouth conference)