@@stronggun2014 longer one has machine learning, deep learning, and reinforcement learning video parts.. however this one is only deep learning with videos on image classification and text classification
⭐ course contents ⭐ (0:00:00) what is a Neural Network? (0:26:34) Loading & Looking at Data (0:39:38) creating a model (0:56:48) using the model to make predictions (1:07:11) Text Classification p1 (1:28:37) what is an Embedding Layer? Text Classification P2 (1:42:30) Training the model - Text classification P3 (1:52:35) saving & Loading Models - Text Classification 4 (2:07:09) How to install Tensorflow GPU on Linux -- Learn to code for free and get a developer job:
(0:00:00) What is a Neural Network? (0:26:34) How to load & look at data (0:39:38) How to create a model (0:56:48) How to use the model to make predictions (1:07:11) Text Classification (part 1) (1:28:37) What is an Embedding Layer? Text Classification (part 2) (1:42:30) How to train the model - Text Classification (part 3) (1:52:35) How to saving & loading models - Text Classification (part 4) (2:07:09) How to install TensorFlow GPU on Linux
i am 13 years old and don't speak and understand very much englisch, but i am now watching the video at 25 min and i have understand how neural networks work and every thing else expect the activation function, thanks
Great tutorial so far, just a quick correction that the sigmoid activation function ranges between 0 and 1. What you had drawn was actually the tanh activation function, that ranges between -1 and 1. Cheers!
Newbie in almost every aspects in what is said in this video. Took me a whole day to get to 35:10. Had issues (1). downloading python. The correct version is python3.6.1 for tensorflow 2.0.0, to avoid "from google.protobuf.pyext import _message ImportError: DLL load failed: The specified procedure could not be found." errors, (2). Tensorflow 2.0.0, not 2.0.0alpha0, to avoid many many many "future warnings" (3). "Cache entry deserialization failed, entry ignored", solved by opening command prompt as administrator! (4). many typo error from my own fault. Almost given up. Using windows10 pro, CPU, intel64. I know the problem relates to my special settings, but might happen to other new users. This is a great video for beginners, thou.
A lot of people are asking what versions of python and windows can be used to run TensorFlow 2.0. I've dug into this for you all. A lot of info is from the official site, some is from github issues and published articles regarding TF 2.0, so at the time of writing this should be accurate information. First, operating systems. TF 2.0 was tested and is officially supported on the following *64-bit* systems: * Windows 7 or later. * Ubuntu 16.04 or later. * macOS 10.12.6 (Sierra) or later - note that these versions do not offer GPU support. * Raspbian 9.0 or later. Python versions that are currently supported are: * Python 3.6 (but NOT Python 3.7, despite its recent release.) * Python 2.7.
Somebody has probably already answered this but "verbose" means descriptive. If your were to enable the verbose property on an object it would normally give a lot more details about something, whether that is debug information or just printing output. Also Great Video Tim! I am a big fan of the tutorials on cutting edge technology as they are difficult to find elsewhere :) Keep up the great work
1:47:14 - The verbose parameter is a simple debugging tool which prints the status of epochs while the model is being trained. In the case, verbose=1 displays the epoch number with a little decoration. Please feel free to correct if any and add more info.
your're right. You can use verbose to track model's progress, if you dont want to see it's learning process you can switch it off. verbose = 0 - you see absolute no progress, just when training is finished program will go on, verbose = 2 - you'll see which epoch is already trained, verbose = 1 you'll see number of epoch (for example Epoch 11/100), progress bar in %, and some parameters (loss function, and accuracy).
By far (!) the best video i have found for beginners in neural networks. And i viewed a lot. Love it! What killed me is the writing of numbers from the bottom to the top. Never seen writing numbers that way.
Hey Tim, great content so far. I would recommend in future videos to reduce the point size on your pen so that your handwriting is a little clearer. Thanks for putting this together. Very well constructed explanations.
Lists and Arrays are very different data structures in the way they work during runtime. Lists are mutable objects, that is, one can add, replace, remove an element in a list whereas Arrays are immutable objects, that is, one cannot change data elements once an array has been created (you can delete the array entirely however). Lists can contain multiple types of data elements in multiple combinations, such as, a list inside a list inside a list, an array inside a nested list, a tuple, a dictionary, a string, an integer, a float, a timestamp, etc., whereas an Array must have all data elements of the same type. There are of course many other differences.
Took me hours to figure this out: If training the model takes really long, this might be the fix: Don't use IDLE to run the Code. Just use CMD/Terminal.
When you divide your data by the max value, you are essentially "normalizing" it. "Shrinking it" or more appropriately quantizing the data is a different process involving reducing the number of significant bits and the quantization step.
For everyone who as problems installing the pip package, its only for python versions up to 3.8. If you are running python 3.9 or higher install the alpha version of the pip package
FOR ANYONE WHO CANT DOWNLOAD TENSOR FLOW please do these steps it took me hours to figure out 1.if it doesn’t work uninstall ALL versions of python and and andacanda, including ALL files related to it 2. Reinstall and anaconda and set to be in path when the installer shows up 3. then just open your command prompt not the anaconda one then your done
Fwiw: Sigmoid is not one specific function, but a class of functions that plot roughly as a flattened S. Not all map to a range of (-1,1). The most commonly used sigmoid is the 'logistic function' which maps to (0,1) - good for feeding into probability distributions. The activation function used in this video could be a hyperbolic tangent, an error function, etc. A visit to the Sigmoid function page on Wikipedia might be helpful.
if you want the longest length for the text classifier tut it's 2697 or use the code bellow longest_length = 0; joinedlist = train_data + test_data for i in joinedlist: if len(joinedlist[i]) > longest_length: longest_length = len(joinedlist[i])
19:25 This is not the sigmoid function as you say in the video. This is the hyperbolic tangent. The sigmoid function maps any value between 0 and 1, not between -1 and 1.
These functions are known as non-linear activation functions common ones known are the sigmoid function and hyperbolic tangent but modern ones used are Rectified Linear Unit (ReLU), Leaky LU and Exponential LU
thank you very much man, I'm a philosophy student who's trying to find a way in understing AI, that's hard but videos like this are a HUGE help, thank you very much, keep it up
Thanks for taking time and explaining this concepts with example. it helps many beginners. I teach advanced statistics and this details will help many people to understand fundamentals. nice job.
This is incredible. I'm an hour in, and I feel like I've learned more practical application in this video than I have in all of the other ML research I've done combined.
If you don't feel like going through an installation process rn, just use google colab, it has numpy, pandas, matplotlib, and tensorflow already available to use.
If training the model takes longer than expected, just run the cmd.exe or python.exe or Spyder or whatever you use using graphics processor. Just right click on the file in the location and click "Run with graphics processor" in Windows 10 and use the non-integrated one.
When you are working with a single neuron, like at the beginning of your video, you have only one bias term. It makes no sense to have one per connection. Your excitation function should have been \sum_{i=1}^4 { w_i \cdot v_i } + b. This neuron requires five parameters instead of eight.
You don’t need machine learning, just basic regression and statistical models. Oh yeah, and tens of thousands of dollars a month in obscure and alpha generating data feeds. Or co locate a server at a stock exchange and front run the guys running the models.
At 1:05:25 you mentioned if we need to make predictions on a single image from the dataset we just need to put [test_images[7]] instead of test_images. correction: prediction = model.predict(np.array([test_images[7]])) we need to pass a NumPy array to the model else one would run into this error: ValueError: Input 0 of layer dense is incompatible with the layer: expected axis -1 of input shape to have value 784 but received input with shape [None, 28]
Thank you so much Tim! As a beginner into tensorflow and machine learning, you were so excellent and helped me grasp the basic tips! Thank you again!! 😀
37:53 its not the RGB values. They are the values obtained after using back propagation and adding biases. Its like conversion of images of T-shirt into a matrix that a computer can understand and then compare that matrix.
Flattening the input leads to a loss in information in the sense that geometrical dependencies are not retrievable. Usally this is solved by a convolutional layer (which is commonly used in image recognition). There is nothing wrong with choosing a simpler example in a tutorial but imo you should not give the impression that this is normal. In addition lists and arrays are not interchangeable since they are different data structures.
When shifting the word index by 3 and add the special token into the dictionary for "" and "" and so on, the whole system will have a mismatch of words, because such shift is not made to the data accordingly.
Oooo awesome man! I literally just saw a video from your main channel I'm assuming! Awesome content my friend. Im trying to familiarize myself with the AI community on RUclips. Have a good one!😃👍
For those facing problems installing TensorFlow and other libraries I recommend using Google Colab as it already has all the required libraries per-installed.👍
for the length of the reviews i used for i in range(len(test_data)): sum = sum+len(test_data[i]) print(sum/len(test_data)) to simply calculate the average length, the average review is 230 words long i think thats a good limit
From the top comments, I couldn't see but verbose means how much detail the program goes into, verbose is used a ton in linux commands so check them out!
Nice video! I'm trying to get into machine learning but I hate how it requires and NVIDIA graphics card. Even a Macbook can't get enough performance out of the box for it. I feel like there must be some other way.
My biggest problem with the TensorFlow guidelines is that there is no direct instruction on how to use your own data. The official TensorFlow guidelines do not clearly address all the requirements. I have also wondered why, for example, it is made so difficult to convert text into a TensorFlow-compatible data format. The official TensorFlow guidelines say "load CSV data" is a good example. The first section is really clear and simple when measuring the size of a clam or whatever it is. In the second section explaining the survivors of the Titanic the data conversion to a format suitable for TensorFlow is a pain in the ass. Why hasn't TensorFlow done anything to alleviate this.
255 is white and 0 is black is the correct mapping of color if it is greyscale image and not vice versa. If color image, the 3 numbers will be in each pixel (height x weight); each number of those 3 represents intensity of one color each(rgb is the default order used in Matplotlib. OpenCV converts in a different order ie., bgr) If doubt still exists run this code in python ide after importing numpy as np and matplotlib.pyplot as plt and try changing values of red, green and blue variables red, green, blue = 255,255,0 ; rgb = (red, green, blue); pixel = 1*[1*rgb]; breadth, height = 25, 25; image = np.array(height*[breadth*pixel]);plt.imshow(image);plt.show();
56:27 Training acc marginally larger than the val acc indicates that a high variance problem, and extends the training epochs doesn't sound like a good idea. Tricks as regularizations, dropout or just reduce the number of parameters should be work.
You can avoid downgrading numpy by changing you pickle value to True in the IMDB file located in your keras packages. line 85 in the file change to this with np.load(path, allow_pickle=True) as f:
+1 for the ‘plz’, but when you’re requesting an additional feature for your convenience, it doesn’t hurt to show a little gratitude for what has already been provided or a preemptive ‘thank you’ for what you want. Thank you. - the ‘Please & Thank You Police’
Well the Andrew NG's course is great too, but I feel you can learn the math behind it from 3blue1brown's playlist. And also this course is specifically intro to tensorflow and Deep learning and isn't as detailed as Coursera's MOOC.
Thanks for that Review as å tutorial and the preps and effort in Ur workspace. It took me just å little to catch the nearby and parental prints in tensor, which came from cran-R as spss clone. Now as New Tool for the unforseeanable..... Thx and Rock on 🤘
amazing tutorial but may help to use jupyter notebooks so that you don't have to run the whole code in the command line every time and can just run block by block :)
It was a great first half session (I watched till that point). How do we tell our neural network output neurons (output1 to output9) to predict the specific labels. For ex: output 1 neuron should predict the output of Trouser. Where do we really mention that?
Hey Tim, I have only one question. Since you called the dataset "imdb", I really want to know where it is actually stored. So far, I knew I had to download it. But where does it download from and where is it stored in my storage? Doing great work !!!
Great tutorial! One question about the globalaveragepooling layer. After embedding we are actually taking the average of the embedding features over all the word vectors and not the average of every individual vector? Say we have 2 words in a sentence that we want to predict the sentiment of: "Very nice" -> [1,1,1,1], [2,2,2,2] -> 2 words, 2 word vectors with 4 embedding features (contexts). The correct way is to take the average over these vectors so the lower dimensional output is [1.5, 1.5, 1.5, 1.5], that we then pass to the dense layer. And the incorrect way is to output a 2 dimensional vector averaging the 2 vectors individually -> output: [1, 2]? Just averaging every word vector individually and passing every single one in a new vector doesn't make sense to me and would just throw away the context.
If you want to learn even more about TensorFlow, check out this 7-hour course: ruclips.net/video/tPYj3fFJGjk/видео.html
Which one is better?
Should I watch this first or the other one?
@@stronggun2014 longer one has machine learning, deep learning, and reinforcement learning video parts.. however this one is only deep learning with videos on image classification and text classification
am i the only one getting an error?
is anyone getting this error?
Failed to load the native TensorFlow runtime.
does the course demonstrate how to use gpus?
Hope you guys liked it! If you want more machine learning and AI tutorials check out my channel 🔥
Thank you for this presentation Tim. God bless you.
Already subscribed. Could you please reply to my question. I am beginner in this.
Big fan of your videos and I was so excited to see you doing this course
Thank you Tim!!!
Thank you for this Tim!
⭐ course contents ⭐
(0:00:00) what is a Neural Network?
(0:26:34) Loading & Looking at Data
(0:39:38) creating a model
(0:56:48) using the model to make predictions
(1:07:11) Text Classification p1
(1:28:37) what is an Embedding Layer? Text Classification P2
(1:42:30) Training the model - Text classification P3
(1:52:35) saving & Loading Models - Text Classification 4
(2:07:09) How to install Tensorflow GPU on Linux
--
Learn to code for free and get a developer job:
Thank you!!
That's helpful tnx
@@feridakifzade9070mmmmm. :
Hmmmm
as if it wasn't already in the description
Well, ppl can click on the timestamps here, so I think it is kind of useful.
(0:00:00) What is a Neural Network?
(0:26:34) How to load & look at data
(0:39:38) How to create a model
(0:56:48) How to use the model to make predictions
(1:07:11) Text Classification (part 1)
(1:28:37) What is an Embedding Layer? Text Classification (part 2)
(1:42:30) How to train the model - Text Classification (part 3)
(1:52:35) How to saving & loading models - Text Classification (part 4)
(2:07:09) How to install TensorFlow GPU on Linux
i am 13 years old and don't speak and understand very much englisch, but i am now watching the video at 25 min and i have understand how neural networks work and every thing else expect the activation function, thanks
Great tutorial so far, just a quick correction that the sigmoid activation function ranges between 0 and 1. What you had drawn was actually the tanh activation function, that ranges between -1 and 1. Cheers!
Correct, I just came to write that
Newbie in almost every aspects in what is said in this video. Took me a whole day to get to 35:10. Had issues (1). downloading python. The correct version is python3.6.1 for tensorflow 2.0.0, to avoid "from google.protobuf.pyext import _message
ImportError: DLL load failed: The specified procedure could not be found." errors, (2). Tensorflow 2.0.0, not 2.0.0alpha0, to avoid many many many "future warnings" (3). "Cache entry deserialization failed, entry ignored", solved by opening command prompt as administrator! (4). many typo error from my own fault. Almost given up. Using windows10 pro, CPU, intel64. I know the problem relates to my special settings, but might happen to other new users. This is a great video for beginners, thou.
you explain this better then most college professors do and it doesn't cost me my future in student loans
I spent 3days to watch and figure out the basic theory behind this tutorial, thank you very much~!
A lot of people are asking what versions of python and windows can be used to run TensorFlow 2.0. I've dug into this for you all. A lot of info is from the official site, some is from github issues and published articles regarding TF 2.0, so at the time of writing this should be accurate information.
First, operating systems. TF 2.0 was tested and is officially supported on the following *64-bit* systems:
* Windows 7 or later.
* Ubuntu 16.04 or later.
* macOS 10.12.6 (Sierra) or later - note that these versions do not offer GPU support.
* Raspbian 9.0 or later.
Python versions that are currently supported are:
* Python 3.6 (but NOT Python 3.7, despite its recent release.)
* Python 2.7.
Somebody has probably already answered this but "verbose" means descriptive. If your were to enable the verbose property on an object it would normally give a lot more details about something, whether that is debug information or just printing output. Also Great Video Tim! I am a big fan of the tutorials on cutting edge technology as they are difficult to find elsewhere :) Keep up the great work
for those who are having problem with the predict method, replace it with predict_classes:
model.predict_classes([test_review])
Bro you saved me....thanks
1:47:14 - The verbose parameter is a simple debugging tool which prints the status of epochs while the model is being trained. In the case, verbose=1 displays the epoch number with a little decoration. Please feel free to correct if any and add more info.
your're right. You can use verbose to track model's progress, if you dont want to see it's learning process you can switch it off. verbose = 0 - you see absolute no progress, just when training is finished program will go on, verbose = 2 - you'll see which epoch is already trained, verbose = 1 you'll see number of epoch (for example Epoch 11/100), progress bar in %, and some parameters (loss function, and accuracy).
By far (!) the best video i have found for beginners in neural networks. And i viewed a lot. Love it!
What killed me is the writing of numbers from the bottom to the top. Never seen writing numbers that way.
He became my favorite by saying he doesn't know what Verbose is.
that comment gave me less confidence in him
Hey Tim, great content so far. I would recommend in future videos to reduce the point size on your pen so that your handwriting is a little clearer. Thanks for putting this together. Very well constructed explanations.
Lists and Arrays are very different data structures in the way they work during runtime. Lists are mutable objects, that is, one can add, replace, remove an element in a list whereas Arrays are immutable objects, that is, one cannot change data elements once an array has been created (you can delete the array entirely however). Lists can contain multiple types of data elements in multiple combinations, such as, a list inside a list inside a list, an array inside a nested list, a tuple, a dictionary, a string, an integer, a float, a timestamp, etc., whereas an Array must have all data elements of the same type. There are of course many other differences.
i have high school math and i understook what an activation function is so is very well explained!
Took me hours to figure this out:
If training the model takes really long, this might be the fix:
Don't use IDLE to run the Code. Just use CMD/Terminal.
When you divide your data by the max value, you are essentially "normalizing" it. "Shrinking it" or more appropriately quantizing the data is a different process involving reducing the number of significant bits and the quantization step.
For everyone who as problems installing the pip package, its only for python versions up to 3.8. If you are running python 3.9 or higher install the alpha version of the pip package
I'm excited to learn this with Tim! He taught me Java thoroughly!
1:19:30 He prints the codes from the train_data[0], but presents the string results from test_data[0]. So there's a mismatch. Hopefully, it's helpful.
Was using google collab to implement this tutorial. Thank you for the great content.
FOR ANYONE WHO CANT DOWNLOAD TENSOR FLOW please do these steps it took me hours to figure out
1.if it doesn’t work uninstall ALL versions of python and and andacanda, including ALL files related to it
2. Reinstall and anaconda and set to be in path when the installer shows up
3. then just open your command prompt not the anaconda one then your done
This tutorial is well put together. I was looking to learn more about neural nets and TensorFlow. this is perfect for a beginner in the field.
Fwiw: Sigmoid is not one specific function, but a class of functions that plot roughly as a flattened S. Not all map to a range of (-1,1). The most commonly used sigmoid is the 'logistic function' which maps to (0,1) - good for feeding into probability distributions. The activation function used in this video could be a hyperbolic tangent, an error function, etc. A visit to the Sigmoid function page on Wikipedia might be helpful.
if you want the longest length for the text classifier tut it's 2697 or use the code bellow
longest_length = 0;
joinedlist = train_data + test_data
for i in joinedlist:
if len(joinedlist[i]) > longest_length:
longest_length = len(joinedlist[i])
Best beginner tutorial on entire RUclips!! Thank you so much! Very good explained.
at 00:37:00 255 represents white and 0 represents black.
Great video! keep it up.
19:25
This is not the sigmoid function as you say in the video. This is the hyperbolic tangent. The sigmoid function maps any value between 0 and 1, not between -1 and 1.
Yup you're correct, the sigmoid function's y coordinate is bounded by 0 and 1.
But if you do 1-g(z)^2 it is a value between 0-1
These functions are known as non-linear activation functions common ones known are the sigmoid function and hyperbolic tangent but modern ones used are Rectified Linear Unit (ReLU), Leaky LU and Exponential LU
We can also use a value between -1 and 1 to represent strongly positive and negative values during testing
Just for the people who may have been confused
thank you very much man, I'm a philosophy student who's trying to find a way in understing AI, that's hard but videos like this are a HUGE help, thank you very much, keep it up
Thanks for taking time and explaining this concepts with example. it helps many beginners. I teach advanced statistics and this details will help many people to understand fundamentals. nice job.
This is incredible. I'm an hour in, and I feel like I've learned more practical application in this video than I have in all of the other ML research I've done combined.
Thanks for your video. I really appreciate the simplicity of your explanations and your humility is refreshing! Will watch more.
Was using google collab to implement this tutorial. Thank you for the great content.
I'm excited to learn this with Tim! He taught me Java thoroughly!
Who is botting?
If you don't feel like going through an installation process rn, just use google colab, it has numpy, pandas, matplotlib, and tensorflow already available to use.
If training the model takes longer than expected, just run the cmd.exe or python.exe or Spyder or whatever you use using graphics processor. Just right click on the file in the location and click "Run with graphics processor" in Windows 10 and use the non-integrated one.
When you are working with a single neuron, like at the beginning of your video, you have only one bias term. It makes no sense to have one per connection. Your excitation function should have been \sum_{i=1}^4 { w_i \cdot v_i } + b. This neuron requires five parameters instead of eight.
Thank you for pointing that out. I was really confused by that math for a bit. So I'm assuming you'd have 1 bias per node in the output layer?
@@LimitedWard You have a bias term per node in the hidden and output layers.
I was looking for this comment
Full Machine Learning For Finance - Quantitative Trading for Beginners (2019) PLEASE!!!
You don’t need machine learning, just basic regression and statistical models. Oh yeah, and tens of thousands of dollars a month in obscure and alpha generating data feeds. Or co locate a server at a stock exchange and front run the guys running the models.
@@AlexCell33 What are the steps to learn machine learning for finance?
Best tensorflow tutorial ive ever seen, thanks for this one!
have yopu completed the course ?? How was it?
At 1:05:25 you mentioned if we need to make predictions on a single image from the dataset we just need to put [test_images[7]] instead of test_images.
correction:
prediction = model.predict(np.array([test_images[7]]))
we need to pass a NumPy array to the model else one would run into this error:
ValueError: Input 0 of layer dense is incompatible with the layer: expected axis -1 of input shape to have value 784 but received input with shape [None, 28]
Very nice explanation of neural netwiorks and using the same to predict the fashion mnist. This makes a lot of sense to me, thanks for this.
Thank you so much Tim!
As a beginner into tensorflow and machine learning, you were so excellent and helped me grasp the basic tips! Thank you again!! 😀
37:53 its not the RGB values. They are the values obtained after using back propagation and adding biases. Its like conversion of images of T-shirt into a matrix that a computer can understand and then compare that matrix.
Very well done tutorial. Nice intro to TF and Neural Networks with some quick and easy to follow examples. So much to learn!
So cool to find a video on this subject where the teacher does not have a heavy accent
Flattening the input leads to a loss in information in the sense that geometrical dependencies are not retrievable. Usally this is solved by a convolutional layer (which is commonly used in image recognition). There is nothing wrong with choosing a simpler example in a tutorial but imo you should not give the impression that this is normal.
In addition lists and arrays are not interchangeable since they are different data structures.
When shifting the word index by 3 and add the special token into the dictionary for "" and "" and so on, the whole system will have a mismatch of words, because such shift is not made to the data accordingly.
Thanks for this series we need more series of videos on tensorflow 2.0
You can save your model like this and reuse it to save your time.
model.save('abc.h5')
model = keras.models.load_model('abc.h5')
Oooo awesome man! I literally just saw a video from your main channel I'm assuming! Awesome content my friend. Im trying to familiarize myself with the AI community on RUclips. Have a good one!😃👍
The tanh function belongs to [-1:1]. Sigmoid activation function is [0:1].
hot
For those facing problems installing TensorFlow and other libraries I recommend using Google Colab as it already has all the required libraries per-installed.👍
for the length of the reviews i used
for i in range(len(test_data)):
sum = sum+len(test_data[i])
print(sum/len(test_data))
to simply calculate the average length, the average review is 230 words long i think thats a good limit
Shorter and more pythonic:
s = sum(len(i) for i in test_data)
print(s/len(test_data))
Whenever you write something like...
for index in range(len(something)):
print(something[i])
...just use...
for element in something:
print(element)
Surprise power-move at 1:20:59 .... typing 'cmd' inside a Windows Explorer search bar to launch CMD in that directory. I never knew that!
verbose is just an interface showing the appearance of the model training progress. setting verbose=1 sets a different visual style and so on
Thank You FreeCodeCamp you people know what I want . I was just about to give ₹15000 for learning this this .
Thanks I love you❤😄
From the top comments, I couldn't see but verbose means how much detail the program goes into, verbose is used a ton in linux commands so check them out!
The universal greeting has become "Hey guys".
Nice video! I'm trying to get into machine learning but I hate how it requires and NVIDIA graphics card. Even a Macbook can't get enough performance out of the box for it. I feel like there must be some other way.
Chris Hawkes can use google colab
@@maheshthapa920 yeah i agree with this, u can use g colab
This line keeps throwing errors
fitModel = model.fit(x_train, y_train, epochs=40, batch_size=512, validation_data=(x_val, y_val), verbose=1)
Hats off to you dude. Crystal clear everything .
Best tutorial on neural networks
Yeah! I handle to run this tutorial with >>PyCharm +Anaconda{Py3.7, TF2.0.0}+offline data
My biggest problem with the TensorFlow guidelines is that there is no direct instruction on how to use your own data. The official TensorFlow guidelines do not clearly address all the requirements. I have also wondered why, for example, it is made so difficult to convert text into a TensorFlow-compatible data format.
The official TensorFlow guidelines say "load CSV data" is a good example. The first section is really clear and simple when measuring the size of a clam or whatever it is. In the second section explaining the survivors of the Titanic the data conversion to a format suitable for TensorFlow is a pain in the ass. Why hasn't TensorFlow done anything to alleviate this.
Got caught up for two hours and finally found out that python 3.8 is incompatible, get the 3.7 64 bit version
thank you
had the same problem, thanks
@@danielwohlgemuth7874 bist du deutsch und hast du evtl Interesse daran "zusammen zu lernen" ?
RoadSurfer Official 👏🏻🐣
Awesome video. Thank you. Now I feel ready to dive in with a book I picked up
this video is waaayyyy better than a course that i actually paid for. Stupid me
Best tutorial I've ever seen, learned so much about neural networks.
Must watch for basics of TensorFlow. Good Tutorial
To answer your question at 1:47:25 , verbose is an optional argument which can be used to report more information about an operation in your program
255 is white and 0 is black is the correct mapping of color if it is greyscale image and not vice versa. If color image, the 3 numbers will be in each pixel (height x weight); each number of those 3 represents intensity of one color each(rgb is the default order used in Matplotlib. OpenCV converts in a different order ie., bgr)
If doubt still exists run this code in python ide after importing numpy as np and matplotlib.pyplot as plt and try changing values of red, green and blue variables
red, green, blue = 255,255,0 ; rgb = (red, green, blue); pixel = 1*[1*rgb]; breadth, height = 25, 25;
image = np.array(height*[breadth*pixel]);plt.imshow(image);plt.show();
Love every minute of this video!great tutorial!thank you so much!
Pretty neat, tensor flow grow so much
Oh man, Thank you so much, Tim! I also follow your channel!
thank you so much ! i was just tired of the indian tutorials. you just made my next few days
56:27 Training acc marginally larger than the val acc indicates that a high variance problem, and extends the training epochs doesn't sound like a good idea. Tricks as regularizations, dropout or just reduce the number of parameters should be work.
Wanna see the step by step video tech how to train a machine learning model!
Very clear explanation and engaging.. Also people do not like 'daataa' it's data
thanks so much dawg, helped me out with my AI assignment, you da man!
🥺
I'm excited to learn this with Tim! He taught me Java thoroughly!
You can avoid downgrading numpy by changing you pickle value to True in the IMDB file located in your keras packages.
line 85 in the file change to this with np.load(path, allow_pickle=True) as f:
Hey a small suggestion. When you provide the timeline of topics in the description, plz make them clickable for easier and faster access
+1 for the ‘plz’, but when you’re requesting an additional feature for your convenience, it doesn’t hurt to show a little gratitude for what has already been provided or a preemptive ‘thank you’ for what you want. Thank you. - the ‘Please & Thank You Police’
I am a vfx artist I know basic python damn this is beautiful and also damn hard 😳😳😳
Great tutorial. Thanks. You might want to use Jupyter with Visual Studio Code, so you don't need to run from the beginning every time.
Python legends (continue the list):
- Tech With Tim
- Sentdex
- Siraj Raval
Dr chuck(python for everybody)
sentdex tops the list
Mike Dane
Siraj is fake. He is clueless about pretty much everything he talks about. Somebody else prepares the work and he just presents it.
@@hhhgggds I am totally agree with you
I took a 11 week course from coursera in ML to learn Neural Network
Tim: LOL
How does it compare to this 2 hour course? I would be so interested to know! :) Thank you
Well the Andrew NG's course is great too, but I feel you can learn the math behind it from 3blue1brown's playlist. And also this course is specifically intro to tensorflow and Deep learning and isn't as detailed as Coursera's MOOC.
love you man. really a good teacher.
There is Radeon-compatible version of Tensorflow, called tensorflow-rocm :)
It really only works on Vegas, but yes, it exists and works okay.
@@Bengt.Lueers I've made it to run on RX 570. Nothing special.
@@BlueCombPL Yes, TF-ROCm runs on gfx 803, too. It is however buggy to a point, where I say it does not work.
2:07 maybe chatbot? Yes bro pls.. My pathetic try to make it +1 for NN chatbot
Great explanation of the activation function.
Thanks for that Review as å tutorial and the preps and effort in Ur workspace. It took me just å little to catch the nearby and parental prints in tensor, which came from cran-R as spss clone. Now as New Tool for the unforseeanable..... Thx and Rock on 🤘
Super clear, easy to follow explanations, THANKS!
the best video for start, its better if you can make another video for more advanced nn works with ts2...
amazing tutorial but may help to use jupyter notebooks so that you don't have to run the whole code in the command line every time and can just run block by block :)
this is the best tentorflow tutorial hand down
It was a great first half session (I watched till that point). How do we tell our neural network output neurons (output1 to output9) to predict the specific labels. For ex: output 1 neuron should predict the output of Trouser. Where do we really mention that?
y labels represent the actual output.
Hey Tim, I have only one question. Since you called the dataset "imdb", I really want to know where it is actually stored. So far, I knew I had to download it. But where does it download from and where is it stored in my storage? Doing great work !!!
Great tutorial! One question about the globalaveragepooling layer. After embedding we are actually taking the average of the embedding features over all the word vectors and not the average of every individual vector? Say we have 2 words in a sentence that we want to predict the sentiment of: "Very nice" -> [1,1,1,1], [2,2,2,2] -> 2 words, 2 word vectors with 4 embedding features (contexts). The correct way is to take the average over these vectors so the lower dimensional output is [1.5, 1.5, 1.5, 1.5], that we then pass to the dense layer. And the incorrect way is to output a 2 dimensional vector averaging the 2 vectors individually -> output: [1, 2]? Just averaging every word vector individually and passing every single one in a new vector doesn't make sense to me and would just throw away the context.