Sign language detection with Python and Scikit Learn | Landmark detection | Computer vision tutorial

Поделиться
HTML-код
  • Опубликовано: 5 окт 2024

Комментарии • 398

  • @ComputerVisionEngineer
    @ComputerVisionEngineer  Год назад +6

    Did you enjoy this video? Try my premium courses! 😃🙌😊
    ● Hands-On Computer Vision in the Cloud: Building an AWS-based Real Time Number Plate Recognition System bit.ly/3RXrE1Y
    ● End-To-End Computer Vision: Build and Deploy a Video Summarization API bit.ly/3tyQX0M
    ● Computer Vision on Edge: Real Time Number Plate Recognition on an Edge Device bit.ly/4dYodA7
    ● Machine Learning Entrepreneur: How to start your entrepreneurial journey as a freelancer and content creator bit.ly/4bFLeaC
    Learn to create AI-based prototypes in the Computer Vision School! www.computervision.school 😃🚀🎓

  • @jesussachez5468
    @jesussachez5468 Год назад +18

    Hello from Mexico!
    I love your job, I did each step in the same way as you, and I had no difficulties, I really feel very grateful for the time you spent teaching us.
    Congratulations teacher!
    👨‍🏫

  • @joque4
    @joque4 6 месяцев назад +10

    For all who are getting errors like "inhomogeneous shapes" while training on big datasets take into account that the MP Hands processing not always return 42 features (sometimes it just doesn't predict the coordinates well enough).
    To avoid this situations always check the length of every array. You must have the same amount of images and labels, and the labels (landmark coordinates) should have the same shapes.
    Just remove the samples that doesn't return all the landmarks or doesn't work well with the Mediapipe hands solution, to ensure all the data has the same shape and to avoid these numpy errors (and bad models).

    • @RAHUL-dt5xm
      @RAHUL-dt5xm 5 месяцев назад +1

      can you help me. when I trained only one gesture nothing else, but the system detects untrained gestures as the trained gesture why? any idea

    • @aryanrana-o6n
      @aryanrana-o6n 5 месяцев назад +1

      can you please share the changed code

    • @mohamedlhachimi2933
      @mohamedlhachimi2933 4 месяца назад +2

      i think guys to solve this problem we had to tell the collect data script to save just frames where he could detect our hands else we will store bad models that will ends with this getting errors like "inhomogeneous shapes" , i actually try to solved this problem by not moving my hand when collecting data and making my model else you can try this code to check your images that are stored
      This script will only print the paths of the images that are deleted due to no hands being detected. It won't display any image windows.
      ##########################################"
      import os
      import cv2
      import mediapipe as mp
      def process_and_show(image_path, mp_drawing):
      mp_hands = mp.solutions.hands
      hands = mp_hands.Hands()

      # Read the image
      image = cv2.imread(image_path)
      image_rgb = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)

      # Detect hands and landmarks
      results = hands.process(image_rgb)

      if not results.multi_hand_landmarks:
      print(f"Deleted image: {image_path}")
      # Delete the image with no hands detected
      os.remove(image_path)
      # Path to your data folder containing subfolders
      data_folder = "data"
      mp_drawing = mp.solutions.drawing_utils
      mp_drawing_styles = mp.solutions.drawing_styles
      # Iterate through subfolders
      for folder_name in os.listdir(data_folder):
      folder_path = os.path.join(data_folder, folder_name)
      if os.path.isdir(folder_path):
      print(f"Checking images in folder: {folder_name}")
      # Iterate through images in the folder
      for filename in os.listdir(folder_path):
      if filename.endswith(".jpg") or filename.endswith(".png"):
      image_path = os.path.join(folder_path, filename)
      process_and_show(image_path, mp_drawing)

    • @pawnidixit1084
      @pawnidixit1084 2 месяца назад

      I understood the problem but can't really put it in the program. could you explain it please?

    • @clementdethoor5533
      @clementdethoor5533 22 дня назад +1

      Just add in create_dataset :
      if (len(data_aux) == 42):
      data.append(data_aux)
      labels.append(dir_)

  • @shwetaevangeline
    @shwetaevangeline 5 месяцев назад +3

    Thank you so much, sir for this wonderful project. I've completed my term project easily with the help of your video. Loved how we can create our own data instead of getting it from somewhere else.

  • @vignesh.v4247
    @vignesh.v4247 Месяц назад +1

    The best tutorial ever!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!

  • @moMo-zu4ds
    @moMo-zu4ds 3 месяца назад

    Having trouble with my ML project now, but so happy to find your video. Thanks for all the work!!

  • @e2mnaturals442
    @e2mnaturals442 8 месяцев назад +2

    hello from Nigeria
    i must say thanks for this video
    it was short, precise and educative
    yes, i had some errors which i was able to handle due to my past knowledge on Deep Learning. And for those that had issues with the disparity in the length of the data, you can always pad to its maximum length
    currently, i have a model that can identify 26 classes correctly and i will definitely increase the classes. i made each classes to have 700 images under different lighting condition
    thanks for all you do.

    • @ijaspr5486
      @ijaspr5486 8 месяцев назад

      bro can you send me the file for your project

    • @e2mnaturals442
      @e2mnaturals442 8 месяцев назад

      @@ijaspr5486 like the whole file?

    • @rarir0012
      @rarir0012 5 месяцев назад

      Could you share your GitHub link of your project?

    • @aryanrana-o6n
      @aryanrana-o6n 5 месяцев назад

      @@e2mnaturals442 yes like github code or i give you my social media id

    • @TheDreamsandTears
      @TheDreamsandTears 3 месяца назад

      can you share your code? I'm having somre errors, while I try do identify the letters. Also, in your code, could you do with signs with both hands and with movements? @e2mnaturals442

  • @WelcomeToMyLife888
    @WelcomeToMyLife888 Год назад +5

    great tutorial on how to organize the project into separate steps!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +2

      Good organization is the key to a successful project I am happy you enjoyed the video! 😄🙌

  • @thesoftwareguy2183
    @thesoftwareguy2183 6 месяцев назад +1

    Sir!! You have my respect I have really learned lots of things in your whole video . Just keep making this ML/DL Project videos , that you have done like implementing from scratch any exciting ML/DL project.
    Just Keep Going Sir!!!
    Thankyou So much!!✨✨✨✨✨✨❤❤❤❤❤❤

  • @aryanrana-o6n
    @aryanrana-o6n 5 месяцев назад +1

    Really Thank you sir. Great Project you helped me a lot to learn many things. After multiple errors solving finally i succeeded in making full project.

  • @artiste9357
    @artiste9357 9 месяцев назад +1

    Thanks a lot! I really appreciate keeping this under an hour as well :))

  • @ajisumiardi6736
    @ajisumiardi6736 2 месяца назад +1

    You're great, Man,, thank you for teaching us and put lots of research first to ensure Windows user can replicate the project too,,
    .
    let me leave a logs here for other Windows users:
    1. dont forget using packages with exactly same version as mentioned in requirements_windows.txt.
    2. Use numpy 1.23.3 version,, I take a sneak peek to your terminal output that give me information if you use numpy with that version,, at first my terminal installed numpy 2.0 version, but no luck, and then dowgrade it,,
    3. If you succesfully Instal Cmake via terminal, but still got error when compiling, I suggest you to install it by install Visual Studio first
    I've spent my first 4 hours dealing with those error before finally made it,,

  • @duleamihai2202
    @duleamihai2202 11 месяцев назад +21

    For those who faces the error where it can't convert the 'data' values from dictionary data_dict, just make sure that in photo samples you are giving the full hand because if not, there will be inconsistent data and the lists will not have the same lenght inside the data_dict['data']. Do again the photos retrieve part and all should be fine

  • @sudarsonbharathwaaj1412
    @sudarsonbharathwaaj1412 8 месяцев назад

    Thanks a lot bro, I watched many videos and i wasted a lot of time and finally found your video and done my project.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  8 месяцев назад +1

      You are welcome! Glad it was helpful! 😃

    • @RohanVector
      @RohanVector 8 месяцев назад +1

      Please send your github link please

    • @RohanVector
      @RohanVector 8 месяцев назад +1

      I got lot of error bro please please please please

  • @1hpxalphaop741
    @1hpxalphaop741 6 месяцев назад

    srsly like the best video, now i can train my custom hand gestures etc. even, thank youu❤❤

  • @John-xi2im
    @John-xi2im 6 месяцев назад

    very awesome tutorial with brilliant idea and conceptualization. Thanks a lost Felipe!

  • @LEDAT-AI
    @LEDAT-AI Год назад +6

    Hello, I have watched your video and found it very informative. However, I was wondering if you could make a video for recognizing different characters for a sequence of movements, for example, the letter "J" or "Z." Thank you for your video.

  • @ivanvijandi2052
    @ivanvijandi2052 15 дней назад

    Mas argentino imposible jsjs, Gran video!

  • @sourabhchandra1740
    @sourabhchandra1740 Год назад +6

    Hlo Sir, very nice video.... I also want to make a similar project ... But there will a bit difference.. I want to generate the entire subtitle for people who can't speak using their hand gestures during video conferencing in real time.
    Can you please guide me with the same ... Bcoz I completely a beginner. Your help will be appreciated. Thanks in advance. 😀

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +4

      Hey Sourabh, it sounds like a complex and very cool project! I would start by saving all the symbols you detect, its confidence score, and the duration of time you detect them so you can analyze this info later on. This is going to help you to understand the problem a little better and also it is going to help you to define rules in order to achieve your goal. 😃💪

    • @Abhaykumar-bu7ei
      @Abhaykumar-bu7ei Год назад

      Hi Sourabh were you able to make it if yes could you please share some update or code for the same

  • @kane_jester
    @kane_jester 11 месяцев назад +4

    sir , the projects get closed if more hands are placed in the real-time video , i know that randomforest classifier uses only certain features , is there a way so that the program doesnt close if more hands are in the video

  • @mariamartinez4860
    @mariamartinez4860 10 месяцев назад +2

    why does it close when you put another hand?

  • @yaranassar1208
    @yaranassar1208 5 месяцев назад +1

    Hii!! I loved your video. I learned a lot. I just have one question, if at the end I want to form a sentence and print it, how can I save each character on the screen to have a full sentence at the end?

  • @hayatlr3000
    @hayatlr3000 Год назад +4

    great tutorial so helpful for my pfe project i actually have to do hand recognition identification biometric only but the hand contour you explained so well the part "this is the most important thing" and I really need help when it comes to the approach of how i can solve this if it? is possible for you to help me by doing a video of it ?cause its the first time for me working with python i usually work with Matlab. thank you again for this video

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +2

      Hey Hayat, I am glad you found it helpful! 😄 Do you mean making a video about how to be strategic when starting a project and choose the most promising approach? Sure, I can do a video about problem solving strategies! 😃🙌

    • @luongtranle2979
      @luongtranle2979 Год назад

      Do you have file word report ?

  • @paradisofernando97
    @paradisofernando97 Год назад

    i love all that, you are very clearly and simply 😍😍

  • @susanlaime1318
    @susanlaime1318 12 дней назад

    Hello! Thank you so much for the tutorial!! :)
    Although I have trouble when trying to find the script's code at the very beginning, how can I get the code and connect my camera to get the 100 frames? Is it on GitHub? With what name? It seems to be there only the code that we built in the video...

  • @emnahamdi-wq4mz
    @emnahamdi-wq4mz 11 месяцев назад +2

    Hi! Great tutorial thank you. I have a question: does this program have data augmentation? and did u calculate the sensibility and accuracy of the program?

  • @radriyansyah05
    @radriyansyah05 4 дня назад

    Sir, if i would create 10 samples different of course what should i do at the stage of using the source code? please give me an example🙏🏻

  • @Yousef_Osman2000
    @Yousef_Osman2000 17 дней назад +1

    how do i get that function 18:10 ?

  • @prithvisingh2851
    @prithvisingh2851 11 месяцев назад +1

    I have trained my model using only numbers' data. It is working but the problem is it is only showing the numbers 9 or 1 in the frame. Do you think it's because of unclear data or problem in the training model.
    BTW great tutorial 👍

  • @bdtamilgamers8083
    @bdtamilgamers8083 Год назад +1

    Sir only 9 character can be trained plz help me to train 26 character

  • @nilayguler8397
    @nilayguler8397 6 месяцев назад

    Thanks a lot! I really appreciate keeping this under an hour as well :)) We are trying to implement this model in Flutter to develop a mobile app. How can we create Flutter integration ?

  • @georgevalentin9483
    @georgevalentin9483 Год назад +2

    I checked the github repo and there are some changes compared to the video. Why are you substracting the min of x_ from x (data_aux.append(x - min(x_))), also for y ? Why is it necessary to do that instead of just append x the way it is to the array. I saw u did that in the data processing and also in the model testing. Thanks a lot!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Hey George! Yeah, I sent that change in a new commit. It makes the solution more robust, you could think about it as a way of 'normalization'. This makes the classifier learn better than the (x, y) position of each landmark is not that important, the distance of each landmark to each other landmark is what matters most! 😃💪

    • @georgevalentin9483
      @georgevalentin9483 Год назад

      @@ComputerVisionEngineer Thanks a lot for the answer! I thought it has something to do with the mediapipe library and is a must, but it actually makes sense to be some kind of normalization. Thanks for you time!

  • @UtsavKuntalwad
    @UtsavKuntalwad 9 месяцев назад +2

    Hello, i was adding new alphabets to the dataset and got this error , unable to solve : " File "D:\Major project\.Major Project\code\train_classifier.py", line 11, in
    data = np.asarray(data_dict['data'])
    ValueError: setting an array element with a sequence. The requested array has an inhomogeneous shape after 1 dimensions. The detected shape was (400,) + inhomogeneous part."

  • @abdulbarisoylemez2817
    @abdulbarisoylemez2817 Год назад

    thank you my teacher, great a video , i tried it myself, I did it :)

  • @RohanVector
    @RohanVector 7 месяцев назад +1

    Some hand sign have two hand ,than what we can do that situation ?

  • @dinithnisal643
    @dinithnisal643 Год назад +2

    hello Sir, I follow your video for learning about computer vision .
    So I have a trouble with "DATA_DIR = './data'" , Is this file need to import from somewhere or should we need to prepare them? Can you help me to solve this?

    • @peterbarasa9190
      @peterbarasa9190 Год назад +1

      am also thinking the same. The images seem no to be there

  • @raziehahmadi4185
    @raziehahmadi4185 4 месяца назад

    Thanks for your good tutorial
    How to act for the rest of the letters?

  • @febriandewanto2447
    @febriandewanto2447 6 месяцев назад

    Thank you, very clear what was taught. I want to ask what if the dataset from a public video had the initial and final movements? whether the start and end frames go into training . and using deep learning?

  • @000HASNAIN_AHMED
    @000HASNAIN_AHMED 11 месяцев назад

    Thank you sir for this video

  • @iinfinixvilla389
    @iinfinixvilla389 3 месяца назад

    Hola from India sir, Sir i enjoyed your video very much. sir, I have a small doubt can you tell me how to check and the accuracy of the model being trained.

  • @saurabhmishra7487
    @saurabhmishra7487 4 месяца назад +1

    The app crashes when using both hands. How can I fix this?

  • @MrFurious0007
    @MrFurious0007 Год назад +2

    Hello , great tutorial 😀can this same approach be applied for british sign language because that uses both hands to make gestures , also can this be deployed in the real world and used at production level ?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      You would need to make some edits in order to use it with both hands but I guess it would work, yes. Regarding the performance, yeah you could train it and improve it so it can be used at a production level. 🙌

    • @MrFurious0007
      @MrFurious0007 11 месяцев назад

      thanks @@ComputerVisionEngineer 😁i'll try and see if it works out

    • @MrFurious0007
      @MrFurious0007 11 месяцев назад +1

      Hey @@ComputerVisionEngineer , its not working efficiently for the british sign lang , maybe because it uses both hands , do you have any suggestions on how i can build up my project , it'll be a huge help , thanks

  • @fragileaf1778
    @fragileaf1778 8 месяцев назад +1

    The camera crashes when I show more than one hand. Can you tell me how it can be fixed?

  • @vamsianurag3415
    @vamsianurag3415 Год назад +2

    Hi, while going through this code i'm getting model_dict = pickle.load(open('./model.p', 'rb'))
    FileNotFoundError: [Errno 2] No such file or directory: './model.p' and I didn't find any model.p file in your repository

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Hey, you can create the model yourself following the steps I describe in the video. 😃🙌

  • @philipokposo6265
    @philipokposo6265 Год назад

    Great content, thank you so much.

  • @swagatbaruah522
    @swagatbaruah522 Год назад +1

    EVERYTHING IS WORKING FINE, EXCEPT FOR THE FACT THAT THE MY FINAL PROGRAM IS UNABLE TO RECOGNIZE ANY SIGN. IT JUST GIVE EVERY SIGN THE SAME LABEL WHATEVER THERE IS IN THE INDEX 0 OF THE LABEL LIST. I don't understand why its not working???

  • @ShivamKumar-oj4zp
    @ShivamKumar-oj4zp Год назад

    Just wanted to tell you that your project is very famous in SMIT 😊

  • @VnZR_
    @VnZR_ 11 месяцев назад +1

    Hi... Since many signs involve some type of movement, I wonder if videos could be used in place of pictures. I hope you can reply to me because your video is very helpful for us. Thanks in advance.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  11 месяцев назад +2

      Yes, you could try with video classification. 🙌

    • @VnZR_
      @VnZR_ 3 месяца назад

      ​@@ComputerVisionEngineer how to insert video type in pycharm?

    • @VnZR_
      @VnZR_ 3 месяца назад

      I hope you can help us..thank you

    • @VnZR_
      @VnZR_ 3 месяца назад

      Is there a front - end that can connect in pycharm?

  • @zeroboom4
    @zeroboom4 6 месяцев назад +1

    I have tried it with arabic Sign language,and it did not working correctly, I get one letter almost every time and it's wrong letter, any ideas that can help me train the model. I got the dataset from kaggle.

  • @livelife846
    @livelife846 11 месяцев назад

    Thank you so much it's helpful for me 😊

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  11 месяцев назад

      Glad to hear it is helpful! 😃🙌

    • @RohanVector
      @RohanVector 8 месяцев назад

      size.width>0 && size.height>0 in function 'cv::imshow' error sir

  • @arif-emre-yaman
    @arif-emre-yaman Год назад

    another great project

  • @szmasclips1774
    @szmasclips1774 3 месяца назад

    Great video but How do you do the collecting images part of the code?

  • @frankleonardoluquelineros9389
    @frankleonardoluquelineros9389 Год назад

    Really great video tutorial! Why did you choose scikt learn and not Yolo? How many changes would you have to make to use Yolo?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Do you mean using Yolo for object detection instead of mediapipe + Scikit learn? It can be done. You just need to train it. I did it with mediapipe + Scikit learn only for simplicity, and I think it also results in a more robust classifier. 🙌

  • @tihbohsyednap8644
    @tihbohsyednap8644 Год назад +1

    Hello sir, Kindly solve this error for me ----> ValueError: With n_samples=1, test_size=0.2 and train_size=0.8, the resulting train set will be empty. Adjust any of the aforementioned parameters.

  • @senpaihere6512
    @senpaihere6512 6 месяцев назад

    hehe subscribed, tysm for this it was very helpful

  • @touchwood8404
    @touchwood8404 6 месяцев назад +1

    The mediapipe library is giving error in installation what should I do?

  • @martinsilungwe2725
    @martinsilungwe2725 Год назад

    I have just subscribed,
    Currently working on a similar project, fingers crossed I'm at a right place..😂

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      🤞😀 Good luck with your project, Martin! 🙌

    • @martinsilungwe2725
      @martinsilungwe2725 Год назад

      @@ComputerVisionEngineer Sir i have an error "ValueError: The least populated class in y has only 1 member, which is too few. The minimum number of groups for any class cannot
      be less than 2.
      ", what can be the problem, im trying to classfy all the alphabet letters, your help will be highly appreciated.

    • @sakshi8806
      @sakshi8806 6 дней назад

      ​@@martinsilungwe2725 do you have any solution for it now?

  • @abdallahsamir2707
    @abdallahsamir2707 Год назад +1

    Hello, I have watched your video and found it very informative. However, I was wondering what is the limitation of this project?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +2

      Hey, limitation in terms of possible symbols? I would say any static symbol made with only one hand.

  • @sandanuwan4441
    @sandanuwan4441 6 месяцев назад

    I am new to AI. I just want to know are we using Natural Language, Machine Learning and computer vision.

  • @MoominMoomin-f2b
    @MoominMoomin-f2b Месяц назад

    Hello!! Can you tell me which ML algorithm did you use in this?

  • @MoominMoomin-f2b
    @MoominMoomin-f2b Месяц назад

    Hello! can you please tell me which ML algorithm you used here???

  • @jeet611_
    @jeet611_ 11 месяцев назад

    life saver.

  • @prathamupadhyay1265
    @prathamupadhyay1265 Год назад +2

    How can I get accuracy for the letters predicted?
    Basically I want live accuracy for the letters that are predicted , since if you show any random hand gesture it will always predict some random letter, so it will be much better if you could also show live accuracy .Is it possible can u guide me a little bit through this?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Try using the method 'predict_proba' instead of 'predict'. You wil get a probability vector for all the classes. Taking the largest number will give you the confidence value you are looking for. 💪💪

    • @prathamupadhyay1265
      @prathamupadhyay1265 Год назад

      @@ComputerVisionEngineer Thanks a lot you are amazing !!! 😃

    • @yashanchule9641
      @yashanchule9641 Год назад

      @@prathamupadhyay1265 bhai if u dont mind kya app apke code ki zip file mujhe share kar skte hai, coz im getting many errors and i have tried many steps but kuch ho nahi raha hai. PLZ!!!!!!

    • @yashanchule9641
      @yashanchule9641 Год назад

      plz bhai

    • @054_vishwadhimar4
      @054_vishwadhimar4 Год назад

      @@yashanchule9641 GitHub link is there..or have you tried that too?!

  • @assassinhi4889
    @assassinhi4889 7 месяцев назад +1

    it's showing the error: ValueError: setting an array element with a sequence.
    after loading the dictionary in the model.

    • @mohamedlhachimi2933
      @mohamedlhachimi2933 4 месяца назад +1

      i think guys to solve this problem we had to tell the collect data script to save just frames where he could detect our hands else we will store bad models that will ends with this getting errors like "inhomogeneous shapes" , i actually try to solved this problem by not moving my hand when collecting data and making my model else you can try this code to check your images that are stored
      This script will only print the paths of the images that are deleted due to no hands being detected. It won't display any image windows.
      ##########################################"
      import os
      import cv2
      import mediapipe as mp
      def process_and_show(image_path, mp_drawing):
      mp_hands = mp.solutions.hands
      hands = mp_hands.Hands()

      # Read the image
      image = cv2.imread(image_path)
      image_rgb = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)

      # Detect hands and landmarks
      results = hands.process(image_rgb)

      if not results.multi_hand_landmarks:
      print(f"Deleted image: {image_path}")
      # Delete the image with no hands detected
      os.remove(image_path)
      # Path to your data folder containing subfolders
      data_folder = "data"
      mp_drawing = mp.solutions.drawing_utils
      mp_drawing_styles = mp.solutions.drawing_styles
      # Iterate through subfolders
      for folder_name in os.listdir(data_folder):
      folder_path = os.path.join(data_folder, folder_name)
      if os.path.isdir(folder_path):
      print(f"Checking images in folder: {folder_name}")
      # Iterate through images in the folder
      for filename in os.listdir(folder_path):
      if filename.endswith(".jpg") or filename.endswith(".png"):
      image_path = os.path.join(folder_path, filename)
      process_and_show(image_path, mp_drawing)

  • @Om-id1qr
    @Om-id1qr Год назад +1

    Great tutorial! Can you tell me how can I do this for Indian Sign Language which uses 2 hands?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      I am looking at the Indian sign language alphabet and I see some characters are done with 2 hands and others with 1 hand. In order to do something based on landmarks as we did on this video you would have to train 2 classifiers, one of them taking as input the landmarks of one hand only (as we did on the video) and the other classifier taking as input the landmarks of both hands. Then some logic to apply one classifier or the other one depending on how many hands appear on the frame. Or, you can just follow a different approach and train an image classifier taking the crop of the hand/s. 💪🙌

    • @v5j7bxb
      @v5j7bxb 5 месяцев назад

      Hi ! Have you completed working on this project? Did it worked ?

  • @f1player95
    @f1player95 8 дней назад

    I'm encountering the following error: Exception encountered: Unrecognized keyword arguments passed to DepthwiseConv2D: {'groups': 1}. Can someone help me with this?

  • @texsesyt2902
    @texsesyt2902 Год назад +2

    hello sir i am getting this error
    ValueError: The least populated class in y has only 1 member, which is too few. The minimum number of groups for any class cannot be less than 2.
    x_train, x_test, y_train, y_test = train_test_split(data, labels, test_size=0.2, shuffle=True, stratify=labels)
    i observe that if i remove stratify i donot get error but after that i get
    0.0% of samples were classified correctly !

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Hey, how many different symbols are you trying to classify? How did you collect the data for each symbol?

    • @texsesyt2902
      @texsesyt2902 Год назад

      @@ComputerVisionEngineer I change number_of_classes to 5 and i collect data through opencv by capturing images(by using the method describe in this video)
      Note: python version 3.11.2

    • @texsesyt2902
      @texsesyt2902 Год назад

      total 5 symbols each got 0 to 99 images

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      There is a probably a bug with the data. Take a look at 'labels', how many elements are there for the different classes? Is it an array of integers or is it other data type?

    • @texsesyt2902
      @texsesyt2902 Год назад

      @@ComputerVisionEngineer Now i am getting this error when i make 25 classes(for each alphabet).
      data = np.asarray(data_dict['data'])
      ValueError: setting an array element with a sequence. The requested array has an inhomogeneous shape after 1 dimensions. The detected shape was (2471,) + inhomogeneous part.

  • @kiranmahapatra8716
    @kiranmahapatra8716 Год назад +1

    Sir please help............during training it shows value error..data = np.asarray(data_dict['data'])
    ValueError: setting an array element with a sequence. The requested array has an inhomogeneous shape after 1 dimensions. The detected shape was (199,) + inhomogeneous part......for 3 class

    • @SohamKaranjkar
      @SohamKaranjkar 9 месяцев назад +1

      i got the same error, were you able to solve it?

    • @krzysztofgalek5276
      @krzysztofgalek5276 9 месяцев назад +1

      Did u solve it?

    • @e2mnaturals442
      @e2mnaturals442 8 месяцев назад

      i was able to sort it using padding
      if you want me to explain more, i will be glad to

    • @Elenas1178
      @Elenas1178 7 месяцев назад

      @@e2mnaturals442 please explain

    • @preetirathod5244
      @preetirathod5244 2 месяца назад

      ​@@e2mnaturals442can u please explain it

  • @makiizenin
    @makiizenin Год назад

    Hello sir, I got a one problem. I made the same with you and my code is worked but it only showed at least 5 mins for capturing then the camera will shutdown automatically and got some errors. :((((

  • @CanalIFES
    @CanalIFES Год назад +1

    why do you use and random forest classifier algorithm?
    maybe it is better for it?
    could i try with a pretrained model to get better results?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      No particular reason why I used a Random Forest, I think pretty much any other classifier would have a similar performance in this case.

    • @CanalIFES
      @CanalIFES Год назад

      @@ComputerVisionEngineer Thanks felipe!!

  • @tihbohsyednap8644
    @tihbohsyednap8644 Год назад +1

    Sir kindly help me with this error
    .
    .
    ValueError: The least populated class in y has only 1 member, which is too few. The minimum number of groups for any class cannot be less than 2.

    • @tihbohsyednap8644
      @tihbohsyednap8644 Год назад

      Sir kindly help me with this error. I am working on this project as my final year project and I have to extend it as my major project work.

    • @mohamedlhachimi2933
      @mohamedlhachimi2933 4 месяца назад

      i think guys to solve this problem we had to tell the collect data script to save just frames where he could detect our hands else we will store bad models that will ends with this getting errors like "inhomogeneous shapes" , i actually try to solved this problem by not moving my hand when collecting data and making my model else you can try this code to check your images that are stored
      This script will only print the paths of the images that are deleted due to no hands being detected. It won't display any image windows.
      ##########################################"
      import os
      import cv2
      import mediapipe as mp
      def process_and_show(image_path, mp_drawing):
      mp_hands = mp.solutions.hands
      hands = mp_hands.Hands()

      # Read the image
      image = cv2.imread(image_path)
      image_rgb = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)

      # Detect hands and landmarks
      results = hands.process(image_rgb)

      if not results.multi_hand_landmarks:
      print(f"Deleted image: {image_path}")
      # Delete the image with no hands detected
      os.remove(image_path)
      # Path to your data folder containing subfolders
      data_folder = "data"
      mp_drawing = mp.solutions.drawing_utils
      mp_drawing_styles = mp.solutions.drawing_styles
      # Iterate through subfolders
      for folder_name in os.listdir(data_folder):
      folder_path = os.path.join(data_folder, folder_name)
      if os.path.isdir(folder_path):
      print(f"Checking images in folder: {folder_name}")
      # Iterate through images in the folder
      for filename in os.listdir(folder_path):
      if filename.endswith(".jpg") or filename.endswith(".png"):
      image_path = os.path.join(folder_path, filename)
      process_and_show(image_path, mp_drawing)

  • @miladsayedi59
    @miladsayedi59 5 месяцев назад

    can we make this project with pose detection models like openpose or deeppose? and what is the difference

  • @AkshatManohar
    @AkshatManohar Год назад +1

    Hi,
    I am getting an error that ./data/.DS_Store is not a directory and is not found.

  • @darrellardhanihidayat555
    @darrellardhanihidayat555 9 месяцев назад +1

    Hi sir, i got some error at inference_classifier.py, the errors says:
    Line 36, in
    H, W, _= frame.shape
    AttributeError: ‘NoneType’ object has no attribute ‘shape’
    Thank you for the help🙏🏻

    • @RohanVector
      @RohanVector 8 месяцев назад

      It's fully working for you now?
      Because I cannot able to run the first step please help mee

    • @RohanVector
      @RohanVector 8 месяцев назад

      In collect_img is cv2.imshow(frame) is error bro kindly help me

    • @RohanVector
      @RohanVector 8 месяцев назад

      Error name :size.width>0 && size.height>0 in function 'cv::imshow'

    • @manasayjoseph1075
      @manasayjoseph1075 8 месяцев назад

      can you please show the err
      @@RohanVector

    • @saivaraprasadmandala8558
      @saivaraprasadmandala8558 8 месяцев назад

      Change the line to -> cap = cv2.VideoCapture(0)...
      Previously it was -> cap = cv2.VideoCapture(2)@@RohanVector

  • @yusufcan1304
    @yusufcan1304 4 месяца назад

    Thanks man

  • @jonrexzelvalloyas6294
    @jonrexzelvalloyas6294 Год назад +1

    great project! may i ask what algorithm is used in your sign language?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Hey, thank you! I am using mediapipe as a hand detector and landmark detector and a Random Forest classifier as sign classifier. 🙌

  • @rohitchan007
    @rohitchan007 Год назад

    Thank you for the video, can you also make a video on sign language recognition on a video dataset (Word level american sign language dataset).

  • @adn4779
    @adn4779 8 месяцев назад +1

    @ComputerVisionEngineer ValueError: X has 84 features, but RandomForestClassifier is expecting 42 features as input..I am getting this error when i run the inference_clasifier.py model...What change should i make in the code.....

    • @shwetaevangeline
      @shwetaevangeline 5 месяцев назад

      If you're getting this, that means you're showing something else that isn't in the data. Only show what you've captured. Or else simply increase number of classes and take different pictures from different angles.

    • @mohamedlhachimi2933
      @mohamedlhachimi2933 4 месяца назад

      i think guys to solve this problem we had to tell the collect data script to save just frames where he could detect our hands else we will store bad models that will ends with this getting errors like "inhomogeneous shapes" , i actually try to solved this problem by not moving my hand when collecting data and making my model else you can try this code to check your images that are stored
      This script will only print the paths of the images that are deleted due to no hands being detected. It won't display any image windows.
      ##########################################"
      import os
      import cv2
      import mediapipe as mp
      def process_and_show(image_path, mp_drawing):
      mp_hands = mp.solutions.hands
      hands = mp_hands.Hands()

      # Read the image
      image = cv2.imread(image_path)
      image_rgb = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)

      # Detect hands and landmarks
      results = hands.process(image_rgb)

      if not results.multi_hand_landmarks:
      print(f"Deleted image: {image_path}")
      # Delete the image with no hands detected
      os.remove(image_path)
      # Path to your data folder containing subfolders
      data_folder = "data"
      mp_drawing = mp.solutions.drawing_utils
      mp_drawing_styles = mp.solutions.drawing_styles
      # Iterate through subfolders
      for folder_name in os.listdir(data_folder):
      folder_path = os.path.join(data_folder, folder_name)
      if os.path.isdir(folder_path):
      print(f"Checking images in folder: {folder_name}")
      # Iterate through images in the folder
      for filename in os.listdir(folder_path):
      if filename.endswith(".jpg") or filename.endswith(".png"):
      image_path = os.path.join(folder_path, filename)
      process_and_show(image_path, mp_drawing)

    • @luciferani8279
      @luciferani8279 3 месяца назад

      Do not give 2 hands at the same on your camera

  • @harshasshet6755
    @harshasshet6755 5 месяцев назад

    I am getting plots for every data set size which i have taken is it fine bcs i have plt.savefig function, annotated it so that the plt for every dataset size is saved in main data directory

  • @NarutoTamilan007
    @NarutoTamilan007 2 месяца назад

    Sir what is your python version

  • @septian5761
    @septian5761 5 месяцев назад

    can i ask how can you moved this into mobile / android studio

  • @hamzak2883
    @hamzak2883 Год назад

    First of all i want to thank you for this tutorial. I want actually to make a program for sign language but i am confused about the Dataset and how to process the Data which i will maybe get as Videos or Images. can you maybe give me some advice.

  • @ประหยัดจันอังคาร-ฬ9จ

    Great Video

  • @livelife846
    @livelife846 11 месяцев назад

    Who can I add more sign because it's getting error when I try to add more signs

  • @pawanrajbhar6377
    @pawanrajbhar6377 Год назад +1

    data = np.asarray(data_dict["data"])
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
    ValueError: setting an array element with a sequence. The requested array has an inhomogeneous shape after 1 dimensions. The detected shape was (296,) + inhomogeneous part. can u help me where i going wrong ?

    • @foru1854
      @foru1854 Год назад

      bro did you corrected the error pls can you tell me how you did it??

    • @saivaraprasadmandala8558
      @saivaraprasadmandala8558 8 месяцев назад

      For those who faces the error where it can't convert the 'data' values from dictionary data_dict, just make sure that in photo samples you are giving the full hand because if not, there will be inconsistent data and the lists will not have the same lenght inside the data_dict['data']. Do again the photos retrieve part and all should be fine

    • @saivaraprasadmandala8558
      @saivaraprasadmandala8558 8 месяцев назад

      For those who faces the error where it can't convert the 'data' values from dictionary data_dict, just make sure that in photo samples you are giving the full hand because if not, there will be inconsistent data and the lists will not have the same lenght inside the data_dict['data']. Do again the photos retrieve part and all should be fine@@foru1854

  • @HECTORMAURICIOMORANBOBADILLA
    @HECTORMAURICIOMORANBOBADILLA Год назад

    amazing project, i want to do it but with raspberry pi, some suggestion?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Thank you! I haven't tried to do it an edge device, I don't have any suggestions. 🙌

  • @mahmudsherbhai2325
    @mahmudsherbhai2325 Год назад

    hi I am a 15 year old and i want to do this for my school tech convention. What program are you using to code this

  • @travisfernandes5387
    @travisfernandes5387 6 месяцев назад

    how to make this project on web based like on react or flask

  • @NourashAzmineChowdhury
    @NourashAzmineChowdhury Год назад +1

    Sir i am getting this error:
    [ERROR:0@0.045] global obsensor_uvc_stream_channel.cpp:156 cv::obsensor::getStreamChannelGroup Camera index out of range
    Traceback (most recent call last):
    File "D:\sign-language-detector-python-master\collect_imgs.py", line 25, in
    cv2.imshow('frame', frame)
    cv2.error: OpenCV(4.7.0) D:\a\opencv-python\opencv-python\opencv\modules\highgui\src\window.cpp:971: error: (-215:Assertion failed) size.width>0 && size.height>0 in function 'cv::imshow'
    while running collect_imgs.py can you help solve it. can provide me the model or data set you have used . it will be help for me

  • @ranjanadevi7965
    @ranjanadevi7965 8 месяцев назад

    Hello while executing your codes when i was keeping the number of objects grater than 4 thn trainclassifier was unable to generate model.p file in my device can you help me out to solve this issue

  • @chinmayyxd
    @chinmayyxd Год назад +3

    Hi @computervisionengineer i was just doing the project from one of your youtube video and the error i was getting was "X has 42 features, but RandomForestClassifier is expecting 84 features as input."
    My question is how to modify code so that it recognises both of the hands.
    THANKS!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      If you trained the model using 2 hands, the issue may be that the in inference the model capture only one hand in a frame and that triggers the error. You could add an if statement to only run the inference if 2 hands were detected.

    • @abhisheknegi7608
      @abhisheknegi7608 11 месяцев назад +1

      Can you please write the modified code?

    • @e2mnaturals442
      @e2mnaturals442 8 месяцев назад

      i used padding to sort this out

    • @mohamedlhachimi2933
      @mohamedlhachimi2933 4 месяца назад

      i think guys to solve this problem we had to tell the collect data script to save just frames where he could detect our hands else we will store bad models that will ends with this getting errors like "inhomogeneous shapes" , i actually try to solved this problem by not moving my hand when collecting data and making my model else you can try this code to check your images that are stored
      This script will only print the paths of the images that are deleted due to no hands being detected. It won't display any image windows.
      ##########################################"
      import os
      import cv2
      import mediapipe as mp
      def process_and_show(image_path, mp_drawing):
      mp_hands = mp.solutions.hands
      hands = mp_hands.Hands()

      # Read the image
      image = cv2.imread(image_path)
      image_rgb = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)

      # Detect hands and landmarks
      results = hands.process(image_rgb)

      if not results.multi_hand_landmarks:
      print(f"Deleted image: {image_path}")
      # Delete the image with no hands detected
      os.remove(image_path)
      # Path to your data folder containing subfolders
      data_folder = "data"
      mp_drawing = mp.solutions.drawing_utils
      mp_drawing_styles = mp.solutions.drawing_styles
      # Iterate through subfolders
      for folder_name in os.listdir(data_folder):
      folder_path = os.path.join(data_folder, folder_name)
      if os.path.isdir(folder_path):
      print(f"Checking images in folder: {folder_name}")
      # Iterate through images in the folder
      for filename in os.listdir(folder_path):
      if filename.endswith(".jpg") or filename.endswith(".png"):
      image_path = os.path.join(folder_path, filename)
      process_and_show(image_path, mp_drawing)

  • @foru1854
    @foru1854 Год назад

    hey its does not work for more than 5 sign can show value error about the shape can you please fix it

  • @saivaraprasadmandala8558
    @saivaraprasadmandala8558 8 месяцев назад

    Error:
    Traceback (most recent call last):
    File "h:\Mini Project\Mallikarjun Project\sign-language-detector-python-master\sign-language-detector-python-master\inference_classifier.py", line 7, in
    model_dict = pickle.load(open('./model.p', 'rb'))
    ^^^^^^^^^^^^^^^^^^^^^^^
    FileNotFoundError: [Errno 2] No such file or directory: './model.p'
    Could u help me out in fixing this error sir!!!!.

  • @dinem0023
    @dinem0023 4 месяца назад

    in all hand gesture im getting only L what could be the reason can anyone tell me

  • @Hatakeutd
    @Hatakeutd Год назад

    How will this accept video feed from a phone ?

  • @OsazeOgedegbe
    @OsazeOgedegbe Год назад +1

    Hello!
    I tried to do exactly what you did but using the 26 alphabets. I don't know where I went wrong but the data list when converted to an nparray is giving me this error: ValueError: setting an array element with a sequence. The requested array has an inhomogeneous shape after 1 dimensions. The detected shape was (2581,) + inhomogeneous part. I have so many things but I am utterly stuck. Please do you have any idea on how I can fix this error.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Hey, not sure what could be going on, although it is always a good practice to take projects one step at the time. Try to do it with only 2 or 3 symbols and work your way up. It will make things easier to debug. 😃🙌

    • @OsazeOgedegbe
      @OsazeOgedegbe Год назад

      @@ComputerVisionEngineer Thank you. I took your advice and was able to fix the problem by breaking it down. Turns out the data for 3 letters were not properly captured but I re captured them and the 26 letters are working perfectly!! Thank you.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      @@OsazeOgedegbe Amazing! Happy to hear you solved the problem! 😃

    • @foru1854
      @foru1854 Год назад

      @@OsazeOgedegbe i am actually also facing the same error how can i identify which letter data is not captured correctly? pls can you tell me

    • @OsazeOgedegbe
      @OsazeOgedegbe Год назад

      @@foru1854 What I did was start with the first letter (A) , did carried out all the steps and trained the model. When I saw it worked, I added the second letter and did the two; then the third and that gave me the error so I knew the third had a problem and recaptured. I followed on like that and when I add a new one and I get the error I will know that alphabet needs to be recaptured. Hope that helps

  • @locphan1221
    @locphan1221 Год назад

    thanks

  • @mdemike1781
    @mdemike1781 9 месяцев назад

    If you train it in a specific place ex: your bedroom would this work like with the background of your kitchen or different place?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  9 месяцев назад

      Yes, by the way we are doing it in this tutorial, it should work if you change the background. 🙌

    • @mohamedlhachimi2933
      @mohamedlhachimi2933 4 месяца назад

      @@ComputerVisionEngineer i think guys to solve this problem we had to tell the collect data script to save just frames where he could detect our hands else we will store bad models that will ends with this getting errors like "inhomogeneous shapes" , i actually try to solved this problem by not moving my hand when collecting data and making my model else you can try this code to check your images that are stored
      This script will only print the paths of the images that are deleted due to no hands being detected. It won't display any image windows.
      ##########################################"
      import os
      import cv2
      import mediapipe as mp
      def process_and_show(image_path, mp_drawing):
      mp_hands = mp.solutions.hands
      hands = mp_hands.Hands()

      # Read the image
      image = cv2.imread(image_path)
      image_rgb = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)

      # Detect hands and landmarks
      results = hands.process(image_rgb)

      if not results.multi_hand_landmarks:
      print(f"Deleted image: {image_path}")
      # Delete the image with no hands detected
      os.remove(image_path)
      # Path to your data folder containing subfolders
      data_folder = "data"
      mp_drawing = mp.solutions.drawing_utils
      mp_drawing_styles = mp.solutions.drawing_styles
      # Iterate through subfolders
      for folder_name in os.listdir(data_folder):
      folder_path = os.path.join(data_folder, folder_name)
      if os.path.isdir(folder_path):
      print(f"Checking images in folder: {folder_name}")
      # Iterate through images in the folder
      for filename in os.listdir(folder_path):
      if filename.endswith(".jpg") or filename.endswith(".png"):
      image_path = os.path.join(folder_path, filename)
      process_and_show(image_path, mp_drawing)

  • @harshasshet6755
    @harshasshet6755 4 месяца назад

    I am facing wierd problem actually i have altered your project for all 26 alphabets but whatever I show I am getting wrong alphabets

  • @caio_pohlmann
    @caio_pohlmann Год назад +3

    How could you make the project recognize two hands?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +2

      Hey, take a look at our discord server, a member of our community has shared the code to train using 2 hands 💪🙌

    • @LincolinARanee
      @LincolinARanee 11 месяцев назад

      @@ComputerVisionEngineer Kindly share your discord server