Documente Academic
Documente Profesional
Documente Cultură
1 Need of Project
2 Objective
3 Project Description
4 Hardware
5 Software
6 Advantages
7 Future Scope
NEED OF PROJECT
• Python 3.x
• Tensorflow 1.5
• Keras
• OpenCV 3.4
ALGORITHM
• Step 1: Take picture of the hand to be tested using a camera.
• Step 2: Convert the image into Gray scale.
• Step 3: Convert the Gray scale image into a binary image. Set a
threshold so that the pixels that are above certain intensity
are set to white and those below are set to black.
• Step 4: From the binary image, we generate the coordinates of
the image.
• Step 5: These coordinates are then compared with stored co-
ordinates in the database for the purpose of output
generation using pattern matching technique.
• Step 6: We need to use a pattern matching algorithm for this
purpose.
• Step 7: If the pattern is matched, the corresponding to the
image is displayed.
ADVANTAGES
• Reduces the communication gap between mute people and
normal people
• Eliminates the need of translator or assistant to communicate
with mute people
• Real time functioning
• Portable
• No need of recalibration
• Does not get damaged through use
CONCLUSION
• Our project aims to make communication simpler between deaf and
dumb people by introducing Computer in communication path so
that sign language can be automatically captured, recognized,
translated to text and displayed it on LCD. There are various
methods for sign language conversion. Some of them use wired
electronic glove and others use visual based approach. Electronic
gloves are costly and one person cannot use the glove of other
person. In vision based approach, different techniques are used to
recognize and match the captured gestures with gestures in
database. Converting RGB image to binary and matching it with
database using a comparing algorithm is simple, efficient and robust
technique. This technique is sufficiently accurate to convert sign
language into text.
FUTURE SCOPE
• The system can be further expanded for the alphabet, numbers in
gesture control. The input can be also taken in the1form of videos
and the videos are divided into frames and the frames can be
converted into text. We can also add grammatically structure for
sign language. The system1can be made handy by incorporating it
into a mobile phone.
• System would be able to communicate in both directions i.e. It
will have the capability to translate normal languages to hand
gestures successfully. The image processing part of the system
will also be modified to work with every possible environment. A
challenge will be to recognize signs that involve motion.
REFERENCES
• Hamacher, Volkmar & J, Chalupper & Eggers, Joachim &
Fischer, Eghart & U, Kornagel & H, Puder & U, Rass. (2005).
Signal Processing in High-End Hearing Aids: State of the Art,
Challenges, and Future Trends. EURASIP Journal on Advances
in Signal Processing. 18. 10.1155/ASP.2005.2915.
• Rafael C. Gonzalez, Richard E. Woods. Digital Image
Processing. Pearson (2008).
Thank You!