Skip to content

j-hendricks/ASL-Recognition-with-Deep-Learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 

Repository files navigation

ASL-Recognition-with-Deep-Learning

Created CNN that identifies letters from sign language images

Overview

This project shows my results for the DataCamp project titled, "ASL Recognition with Deep Learning". The dataset contains 50x50 images containing hands performing sign-language, and the goal is to train a neural network to determine the letter being relayed in each image. Below are sample images from the dataset. The dataset only contains images of three letters: A, B, and C.

examples

Architecture

The following images display the architecture of the model. There are two convolutional layers, the first with 5 filters and the second with 15. The kernel size was 5, meaning this filters used were 5x5. Two max pooling layers were added in order to highlight the relevant features in the model. The dense layer contains 3 outputs, since there are only three letters being classified: A, B, an C. Only 2 epochs were used during fitting.

architecture

Results

The model achieved 93% accuracy on the test data. This illustrates the power of convlutional layers when combined with max pooling.

results

Here is an example of an image the model failed to interpret:

fail

The model may have had difficult identifying the placement of the thumb and therefore mislabeled the image. To improve this model, image augmentation could be added, and the number of epochs could be increased. In addition, applying dropout layer could improve the model.

About

Created CNN that identifies letter from sign language images

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published