- Видео 70
- Просмотров 171 975
Deep Shallownet
Грузия
Добавлен 16 июн 2019
Artificial Intelligence has great potential for positively impacting the world. By making a freely accessible learning resource, the channel aims to accelerate the integration of AI technologies that would improve the quality of life for everyone.
--------------------------------------------------
Channel art: pixabay.com/illustrations/processor-cpu-computer-chip-board-2217771/#_=_
--------------------------------------------------
Channel art: pixabay.com/illustrations/processor-cpu-computer-chip-board-2217771/#_=_
K-means++ & Lloyd's algorithm
Hands-on example of K-means++ initialization algorithm and Lloyd's algorithm of iterative refinement for K-means clustering.
References:
K means++ - Arthur, D.; Vassilvitskii, S. (2007). "k-means++: the advantages of careful seeding" (PDF). Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms. Society for Industrial and Applied Mathematics Philadelphia, PA, USA. pp. 1027-1035.
Lloyd - Lloyd, Stuart P. (1982), "Least squares quantization in PCM" (PDF), IEEE Transactions on Information Theory, 28 (2): 129-137,
Plot - www.desmos.com/calculator/o3zfw1mxjy
-----------------------------
Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua ...
References:
K means++ - Arthur, D.; Vassilvitskii, S. (2007). "k-means++: the advantages of careful seeding" (PDF). Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms. Society for Industrial and Applied Mathematics Philadelphia, PA, USA. pp. 1027-1035.
Lloyd - Lloyd, Stuart P. (1982), "Least squares quantization in PCM" (PDF), IEEE Transactions on Information Theory, 28 (2): 129-137,
Plot - www.desmos.com/calculator/o3zfw1mxjy
-----------------------------
Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua ...
Просмотров: 6 268
Видео
Principal Component Analysis (PCA) - Step by Step
Просмотров 3984 года назад
A step-by-step by hand principal component analysis on a toy dataset with visualizations. Covariance Matrix: ruclips.net/video/Ggtsxmyx5rM/видео.html Eigen Decomposition: ruclips.net/video/inBYfXAOZDA/видео.html Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint Plot: www....
Decision Tree - Gini Impurity
Просмотров 8304 года назад
Building a decision tree with Gini impurity, by hand. Decision tree learning, top-down greedy approach/algorithm. 0:00 - Interpreting a tree 1:18 - Attribute and split types 2:22 - Building a decision tree 2:45 - Gini Impurity 5:26 - Building a tree intuitively 6:24 - Notes on overfitting Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courv...
Radial Basis Function Kernel - Gaussian Kernel
Просмотров 7 тыс.4 года назад
Radial Basis Function Kernel considered as a measure of similarity and showing how it corresponds to a dot product. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint Derivation of space: pages.cs.wisc.edu/~matthewb/pages/notes/pdf/svms/RBFKernel.pdf
Kernel Trick Visualization, Derivation, and Explanation.
Просмотров 1,4 тыс.4 года назад
What Kernel Trick does visually, deriving the polynomial kernel of degree two, writing down the kernel perceptron algorithm. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Animation made with www.geogebra.org Slides made in: Microsoft PowerPoint
Perceptron
Просмотров 1584 года назад
By hand numerical example of finding a decision boundary using a perceptron learning algorithm and using it for classification. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint Graph: www.desmos.com/calculator/vnoky7b2fn
Soft Margin SVM
Просмотров 5394 года назад
Comparing Hard Margin with Soft Margin SVM, and finding the soft margin classifier in Python. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint Graph: www.desmos.com/calculator/dd0nuurs3e CVX solver: cvxopt.org/
Maximum Margin Classifier, SVM - Support Vector Machine
Просмотров 1,2 тыс.4 года назад
Explanation and example of finding a separating hyperplane using linear programming. 0:00 - Maximum Margin Classifier 2:20 - Linear Programming with CVX Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint Graphs made with Desmos: www.desmos.com/calculator/rp0lniv4oj CVX sol...
Logistic Regression | Binary Logistic Regression
Просмотров 914 года назад
Binary Logistic Regression with an example of fitting to the data. Erratum: At 3:02 and 3:15, log-likelihoods need a negative sign, because we're finding the derivative of the negative log-likelihood. View correct slide at: ibb.co/YTXK51M or bit.ly/3esE0cD Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Sl...
Bayesian Inference & Maximum a Posteriori Estimation | Bayesian Statistics
Просмотров 1 тыс.4 года назад
Bayesian Statistics with examples. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint Plots: Desmos
Linear Regression as Maximum Likelihood
Просмотров 2974 года назад
Justifying Linear Regression as a Maximum Likelihood Estimation. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint
Maximum Likelihood Estimation
Просмотров 2934 года назад
Likelihood and Maximum Likelihood Estimation Explained. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint Graphs: www.desmos.com/calculator
Consistency in Estimators, Bias of Consistent Estimators
Просмотров 6 тыс.4 года назад
Consistent Estimators and their Bias. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint
Bias & Variance Tradeoff with MSE
Просмотров 2574 года назад
Bias and Variance Tradeoff, representing MSE as bias and variance. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint
Variance and Standard Error of an Estimator/Statistic
Просмотров 1,5 тыс.4 года назад
Showing the importance of variance in estimators and finding the variance of the sample mean. Recommended to read along: Deep Learning An MIT Press book Ian Goodfellow and Yoshua Bengio and Aaron Courville www.deeplearningbook.org/ Slides made in: Microsoft PowerPoint
Point Estimators & Function Estimators
Просмотров 1624 года назад
Point Estimators & Function Estimators
Weight Decay - L2 Regularization Example
Просмотров 4,5 тыс.4 года назад
Weight Decay - L2 Regularization Example
Overfitting And Underfitting In Machine Learning
Просмотров 1894 года назад
Overfitting And Underfitting In Machine Learning
Backtracking Line Search in Gradient Descent
Просмотров 18 тыс.5 лет назад
Backtracking Line Search in Gradient Descent
Lipschitz Continuity | Lipschitz Condition
Просмотров 12 тыс.5 лет назад
Lipschitz Continuity | Lipschitz Condition
Thanks, very informative!
The most underrated channel for deep learning and probability.
pretty sure you forgot to include the 1/2 at the end.
You explained this in a minute better than every single other video I've seen
thanks! simple and easy to understand
rank of a matrix?
Isn't these data separable already in 2D space?
There are any application of Hadamard product on a matrix?
Yes, for example in the calculation of LSTM gates and gated recurrent units (GRUS) in deep learning
once you obtain the xnew=0.18 you start again the method with x=0.18 and beta=1? or you restart the method with x=0.18 and your last epsilon? or the exercise is over?
nice explanation
Brief and precise
So difficult to understand
it helps thank you
God bless you boss. I have really struggled today for close to 24 hours in search for g2bbage online till I got your video.
how / why is the the first term constant?
It seems that the y in the 1st row is the x_2 in the 3rd row ?
at 1:15
What is meant by fair ,biased and unbiased coin?
Enna koop ada e paryunnathu uttaram kittunila
This video really helped me clear it up and your example was very simple and useful! I am confused as to why you made beta = 0.707... what made you choose this? is it just a standard?
It is given. alpha lies between 0 and 0.5 and beta lies between 0 and 1.
good for quick recap
Super clear explanation!
Thank you for the video, can you help me how to prove that is unbiased in this question? Question: Compare the average height of employees in Google with the average height in the United States, do you think it is an unbiased estimate? If not, how to prove it is not mathced?
nice. would have been even cooler if you visualized the iterations of the converging centroids
Thanks for the great video! One question: 1:03 where do you get or derived that while condition? Can you please provide some materials? Thanks!
The while condition is just there to let you know to continue the inequality (updating epsilon or t in some textbooks) until it no longer holds. Then your Xc will be the last value where the inequality did not longer hold. Evaluate your objective function at this Xc to mimimize it.
@@leotorres300 thanks for the reply! I understand the logic. But I just want to know math of this while condition. why exactly is in that form? is it from gradient or what? it cannot be any arbitrary condition
Thank you
better than any other videos
"...ℝ , superscript 10..." . Whoopsie, TTS.
Perfect, go ahead 👏
Thank you!
Your annotation is blocking the screen. Do not use annotations pls.
Examples of unbiased but inconsistent estimator may pls be given
very lucid but would be more helpful with numerical example, thanks
*Robot* this video was *effective*
It was not made by a robot, but thanks 😂
Can you please share the link of the slides?
easyupload.io/ujy6cq link will work for 7 days.
lol, that first slide answered all of my questions.
I'm glad it helped 😄
Much appreciated
Thank you for watching! :)
thank you for easy explanation
Thank you for watching! :)
How is it useful ? Any real time examples? Or examples in Data Science preprocessing?
Both distributions can be used as a prior for Bayesian Inference. The probability of the failure of a battery over time can be modeled as exponential distribution. With time, the probability of a failure grows exponentially. Laplace distributions are encountered in finances, you can check the paper "Modelling and predicting market risk with Laplace-Gaussian mixture distributions".
Thanks for this perfect video Please could you provide me any resource that solves examples about Gaussian Maximum likelihood?
Applied statistics and probablility for engineers 6th edition by Douglas Montgomery and George Runger chapter 7.4.2, example 7-13 may be what you're looking for, but I'm not sure.
Thank you sir i will check it God place you 🌸😍💐
well this example have a problem how can one coins affect the probability to the other coins
Thank you for the question. In the example, the outcome of one coin does not affect the outcome of the other. As mentioned at 1:00 , the two events are independent, that is why when we calculate P(A=H|B=T) it equals P(A=H). If we had dependent events, it could have been the case that P(A=H|B=T) != P(A=H). Regardless of the dependency, it is true that P(A=H|B=T) = P(A=H, B=T)/P(B=T). The example showed how to use the formula.
@@deepshallownet5206 OK u should state that the experiment is independent
do i need to read the book? i'm asking coz i have it and there are lot of elaboration in the book. so i'm asking is watching videos sufficient enough?
It really depends on what you goal is. This book gives less elaboration than the standalone books dedicated to the linear algebra. For example there are no exercises, which you will always find in the linear algebra books. It covers only the minimum, which is required to understand rest of the deep learning, but there still are some details in the book that may be hard to grasp at first. However after watching these videos, the book should be easier to understand. Also, this playlist omits the principal component analysis derived with just linear algebra. If you want to understand that part, you will have to read the chapter. Anyway, reading the book would not hurt. The book often skips the examples of concepts making them hard to understand. But these videos explain them in an easier way and provide examples.
keep it up man.doing great job!!!!!!