Advanced Topics in Artificial Intelligence

Robots

Program


  1. Graphical Systems (Probabilistic)

  2. Logic Based Aproaches

  3. Other representations

  4. Integrated Systems

Instructor

Slides

Probabilistic Systems

Crowd

Main slides, also check quick reading slides:

  1. Adrian Weller, MLSALT4 graphical model

  2. Marc Toussaint University of Stuttgart Summer 2015, Machine Learning, Graphical Models

  3. For detailed information, try Daphne Koller Open Class Slides

Not-so-quick Reading:

a. Daphne Koller and Nir Friedman, Probabilistic Graphical Models: everything you wanted and everything you didn't

b. Kevin P. Murphy, Machine Learning A Probabilistic Learning: A probabilistic view of the world.

c. David Barber, Bayesian Reasoning and Machine Learning: pdf from author available/

Propositional Inference

Following slides discuss the connection between SAT solvers, BDDs and trees:

  1. Binary Decision Diagrams are one of the most widely used tools in CS. Their application to BN was proposed by Minato et al but they are not a very popular approach to compile BNs. Model Counting is also not widely used.

  2. The Problog language was initially implemented on BDDs. The ProbLog2 system can use BDDs. model-counting or trees.

  3. Arithmetic Circuits are a very efficient approach for bayesian inference; the Darwiche group recently proposed an extension, SDDs.

  4. The connection to constraint solving is discussed by Dechter.

Optimisation

Optimisation is a fundamental tool for modern machine learning and other areas of AI. Techniques are often based on work from the Operations Research and Constraint communities.

Classes

self-driving

Aula I - Graphical Models [14/02]

Aula II - Graphical Models [21/02]

Aula III - Graphical Models [28/02]

Aula IV - Graphical Models [28/02]

Aula V - Graphical Models vs Logic[28/02]

Aula VI - Graphical Models vs Logic [28/02]

Aula VII - Learning as Optimisation [28/02]

Aula VIII -Improving on Logistic Regression [28/02]

Aula IX - Neural Networks [28/02]

Aula X - Tensorflow

Aula XI - Deep Networks [28/02]

Grading

watson

  1. Mini Project: 4x3 Valores, or 2x3

  2. Project: 6 Valores, or none

  3. Exam: 8 Valores


Mini-Projects

Submission:

To submit you either/or must - brief presentation in class; - source + small report + run log (eg, jupyter notebook) to be sent to vscosta AT fc.up.pt, subj TAIA.

Deadlines:

Presentation

Short presentation on your favorite AI topic.

Support Material

Tutorials

Why attend a tutorial? Introduce, explain and comment on the material on n ai tutorial. Look for good tutorials at:

- IJCAI
- AAAI
- ECAI
- ICML
- NIPS 
- ECML
- KDD
- ICDM

Graphical Models

Evaluation:

Programs will be evaluated based on a short report and interaction with authors. Parameters are:

  1. Running Time: compare selection functions and/or vs other packages;

  2. Memory Size

  3. Try different network sizes

Minimum: you should at least run on the Asia and Cancer Networks

Machine Learning

Given two of the following datasets, plus a dataset of your choosing, compare machine learning models in terms of accuracy and running times:

Machine Learning

Given two of the following datasets, plus a dataset of your choosing, compare Tensorflow Neural Network models. One of the models learning models in terms of accuracy and running times. The report should include:

Datasets:

Evaluation:

Typical Exam Questions

I include a number of questions that would be typical of the exam. The first question should be worth 2 points (or more). The other questions will be circa 1 point:

  1. Given the Bayesian network not depicted here, and assuming that you have evidence on variables X1 and X2, give the sequence of steps that would be taken by the VE algorithm to compute the posterior probability on Y.

  2. For each step, please describe the operation it would perform, and report both the input and output sizes.

  3. Explain the difference between generative and discriminate models, by comparing the Naive Bayes classifier with another model.

  4. Of the two trees above, which one is a proper BDD. Why?

  5. Fig ... shows the structure of a GAN. What is a GAN for, and how does it work?

  6. Stochastic gradient descent is widely combined with mini-batches. What is a mini-batch and why is this done?

  7. TensorFlow is often used through the KERAS API. Give an example of setting up KERAS for a neural network with 2 internal layers.