Université Lyon 1
Arqus
Accueil  >>  Modèles Graphiques Probabilistes
  • Unité d'enseignement : Modèles Graphiques Probabilistes
Nombre de crédits de l'UE : 3
Code APOGEE : INF2354M
    Responsabilité de l'UE :
AUSSEM ALEXANDRE
 alexandre.aussemuniv-lyon1.fr
04.72.43.44.66
    Type d'enseignement
Nb heures *
Cours Magistraux (CM)
18 h
Travaux Dirigés (TD)
6 h
Travaux Pratiques (TP)
6 h

* Ces horaires sont donnés à titre indicatif.

    Compétences attestées (transversales, spécifiques) :
Non rédigé
    Programme de l'UE / Thématiques abordées :

In this class, you will learn the basics of the Probabilistic Graphical Models (PGM) representation and how to construct them, using both human knowledge and machine learning techniques.

What are Probabilistic Graphical Models?

Uncertainty is unavoidable in real-world applications: we can almost never predict with certainty what will happen in the future, and even in the present and the past, many important aspects of the world are not observed with certainty. Probability theory gives us the basic foundation to model our beliefs about the different possible states of the world, and to update these beliefs as new evidence is obtained. These beliefs can be combined with individual preferences to help guide our actions, and even in selecting which observations to make. While probability theory has existed since the 17th century, our ability to use it effectively on large problems involving many inter-related variables is fairly recent, and is due largely to the development of a framework known as Probabilistic Graphical Models (PGMs). This framework, which spans methods such as Bayesian networks and Markov random fields, uses ideas from discrete data structures in computer science to efficiently encode and manipulate probability distributions over high-dimensional spaces, often involving hundreds or even many thousands of variables. These methods have been used in an enormous range of application domains, which include: web search, medical and fault diagnosis, image understanding, reconstruction of biological networks, speech recognition, natural language processing, decoding of messages sent over a noisy communication channel, robot navigation, and many more. The PGM framework provides an essential tool for anyone who wants to learn how to reason coherently from limited and noisy observations.

In this class, you will learn the basics of the PGM representation and how to construct them, using both human knowledge and machine learning techniques; you will also learn algorithms for using a PGM to reach conclusions about the world from limited and noisy evidence, and for making good decisions under uncertainty. The class covers both the theoretical underpinnings of the PGM framework and practical skills needed to apply these techniques to new problems.

Topics covered include:

1.The Bayesian network and Markov network representation, including extensions for reasoning over domains that change over time and over domains with a variable number of entities
2.Reasoning and inference methods, including exact inference (variable elimination, clique trees) and approximate inference (belief propagation message passing, Markov chain Monte Carlo methods) 3.Learning parameters and structure in PGMs

4.Using a PGM for decision making under uncertainty.

There will be short weekly review quizzes and programming assignments (R) focusing on case studies and applications of PGMs to real-world problems:

    Parcours / Spécialité / Filière / Option utilisant cette UE :
Date de la dernière mise-à-jour : 08/09/2022
SELECT MEN_ID, `MEN_DIP_ABREVIATION`, `MEN_TITLE`, `PAR_TITLE`, `PAR_ID` FROM parcours INNER JOIN ue_parcours ON PAR_ID_FK=PAR_ID INNER JOIN mention ON MEN_ID = PAR_MENTION_FK WHERE PAR_ACTIVATE = 0 AND UE_ID_FK='16820' ORDER BY `MEN_DIP_ABREVIATION`, `MEN_TITLE`, `PAR_TITLE`