Probabilistic graphical models (PGMs) are a rich framework for encoding probability distributions over complex domains: joint (multivariate) distributions over large numbers of random variables that interact with each other. These representations sit at the intersection of statistics and computer science, relying on concepts from probability theory, graph algorithms, machine learning, and more. They are the basis for the state-of-the-art methods in a wide variety of applications, such as medical diagnosis, image understanding, speech recognition, natural language processing, and many, many more. They are also a foundational tool in formulating many machine learning problems.
The Leland Stanford Junior University, commonly referred to as Stanford University or Stanford, is an American private research university located in Stanford, California on an 8,180-acre (3,310 ha) campus near Palo Alto, California, United States.
- 5 stars
- 4 stars
- 3 stars
- 2 stars
- 1 star
來自PROBABILISTIC GRAPHICAL MODELS 2: INFERENCE的熱門評論
Just like the first course of the specialization, this course is really good. It is well organized and taught in the best way which really helped me to implement similar ideas for my projects.
I have clearly learnt a lot during this course. Even though some things should be updated and maybe completed, I would definitely recommend it to anyone whose interest lies in PGMs.
Very good course. Subject is quiet complex: lack of concrete examples to make sure concepts well understood. Had to review each the Course twice to understand concepts well
Thanks a lot for professor D.K.'s great course for PGM inference part. Really a very good starting point for PGM model and preparation for learning part.
I learned pretty much from this course. It answered my quandaries from the representation course, and as well deepened my understanding of PGM.
great course, though really advanced. would like a bit more examples especially regarding the coding. worth it overally
I would have like to complete the honors assignments, unfortunately, I'm not fluent in Matlab. Otherwise, great course!
Very interesting course. However, even after completing it with honors, I feel like I don't understand a lot.
Great introduction.\n\nIt would be great to have more examples included in the lectures and slides.
Great course, except that the programming assignments are in Matlab rather than Python
Great introduction to inference. Requires some extra reading from the textbook.
Had a wonderful and enriching fun filled experience, Thank you Daphne Ma'am
Great course! Expect to spend significant time reviewing the material.
This course induces lateral thinking and deep reasoning.
關於 概率图模型 專項課程
Learning Outcomes: By the end of this course, you will be able to take a given PGM and
Execute the basic steps of a variable elimination or message passing algorithm
Understand how properties of the graph structure influence the complexity of exact inference, and thereby estimate whether exact inference is likely to be feasible
Go through the basic steps of an MCMC algorithm, both Gibbs sampling and Metropolis Hastings
Understand how properties of the PGM influence the efficacy of sampling methods, and thereby estimate whether MCMC algorithms are likely to be effective
Design Metropolis Hastings proposal distributions that are more likely to give good results
Compute a MAP assignment by exact inference
Honors track learners will be able to implement message passing algorithms and MCMC algorithms, and apply them to a real world problem