https://www.youtube.com/c/AIPursuit/join
Subscribe to the channel:
https://www.youtube.com/c/AIPursuit?sub_confirmation=1
Support and Donation:
Paypal ⇢ https://paypal.me/tayhengee
Patreon ⇢ https://www.patreon.com/hengee
BTC ⇢ bc1q2r7eymlf20576alvcmryn28tgrvxqw5r30cmpu
ETH ⇢ 0x58c4bD4244686F3b4e636EfeBD159258A5513744
Doge ⇢ DSGNbzuS1s6x81ZSbSHHV5uGDxJXePeyKy
Wanted to own BTC, ETH, or even Dogecoin? Kickstart your crypto portfolio with the largest crypto market Binance with my affiliate link:
https://accounts.binance.com/en/register?ref=27700065
—————————————————————————————–
The video was published under the license of the Creative Commons Attribution license (reuse allowed) and is reposted for educational purposes.
Source: https://youtu.be/tVwV14YkbYs
Course website: http://bit.ly/pDL-home
0:00:00 – Week 7 – Lecture
We introduced the concept of the energy-based models and the intention for different approaches other than feed-forward networks. To solve the difficulty of the inference in EBM, latent variables are used to provide auxiliary information and enable multiple possible predictions. Finally, the EBM can generalize to probabilistic model with more flexible scoring functions.
0:01:04 – Energy-based model concept
0:15:04 – Latent-variable EBM: inference
0:28:23 – EBM vs. probabilistic models
We discussed self-supervised learning, introduced how to train an Energy-based models, discussed Latent Variable EBM, specifically with an explained K-means example. We also introduced Contrastive Methods, explained a denoising autoencoder with a topographic map, the training process, and how it can be used, followed by an introduction to BERT. Finally, we talked about Contrastive Divergence, also explained using a topographic map.
0:44:43 – Self-supervised learning
1:05:57 – Training an Energy-Based Model
1:19:27 – Latent Variable EBM, K-means example, Contrastive Methods
Add comment