First of all, thank for AMD to giving this wonderful opportunity to learn this knowledge. I am a beginner to this project. I have only the theoretical knowledge about AI and algorithms. I have created seven module, first module is speech recognition in Tamil language using python, and second thing is computer vision modules I have created the color detection for the Tritanopia defected people and third thing is virtual mouse it helps to use the mouse virtually and it can be also a set of computer vision. forth module is text to speech converter using python for deaf and dumb they can also speak with the help of this text to voice converter once they enter the text, they want it can be converted into audio file the document folder. fifth module is local and private Llama application it supports both Languages Tamil and English, and sixth module is quantization model it used to shrink the size of the model which has.h extension
my project's theme is given below
1. Voice Recognition for the Blind: Enables interaction with devices through spoken commands.
2. Text-to-Voice Conversion for the Mute: Converts written text into spoken words for communication.
3. Voice-to-Text Conversion for the Deaf: Transcribes spoken words into written text in real-time.
4. Optimized with AMD AI Hardware: Utilizes AMD hardware to enhance performance and efficiency.
5. Seamless Accessibility: Empowers individuals with disabilities to engage with technology effectively.
the given below pictures show the implementation of this it shows four options they are given in the picture when it executes you type of letter "F" or "f" through the help of keyboard for speech Recognition and Letter "T" or "t" for Text to voice and "J" or "j" for voice to text conversion. it contains simple usages only
we use python eel module and speech recognition model which is used to create the voice recognition here we use two languages they are Tamil and English for differently abled people. we use auto voice recognition when we execute the app.py python file the voice searches are implement and text search are also implemented, it gets the choice from user to use English language or Tamil to communicate with user and the past executed command are displayed in the Tkinter windows
it simply types the request in the input column and submit the request it executes the command simultaneously
we use the Tamil llama model to implement the generative ai. we use the stream lit framework to implement the user interface.
here we quantize the model into smaller size.
Computer Vision1.) Color detection using python:I had implemented the code for color detection for the Tritanopia disease (Tritanopia (Tritanopia is a rare type of genetic color blindness that affects a person’s ability to distinguish between the colors blue and yellow. It occurs when the blue-sensitive cones (also known as short-wavelength cones or S cones) in the retina are either not functioning or completely missing.) by using the computer vision I had implement the following screenshots. we are using the python module like cv2 for stimulate the camera.
In this model we have use the webcam for to control the mouse keys. it can be able to use our hands for control the primary click.
In this model we have an ability to covert the text to generation is stores to the project folder
it also does the similar thing like get the voice from user and save in txt extension when it executes it get the input from user to select the language.
C:\Users\(your_path)>git clone https://github.com/JAYASIMMA/AMD_Hack.git
clone the repository
C:\Users\(your_path)>pip install -r requirements.txt
after install all the requirements create virtual Environment in python
C:\Users\(your_path)\amd_hack>cd project
C:\Users\(your_path)\projects>python main_app.py
after running this simultaneously running another module Computer visions
C:\Users\(your_path)>cd ..
C:\Users\(your_path)>cd computer_vision
cd color_detection
python main.py
cd ..
cd live_mouse_control_using_hand_gestures
python main.py
cd ..
cd virual_mouse
cd mouse
cd scripts
code .
by using jupyter run > Hand_Gesture_Mouse.ipynb
Generativa AI
Install the ollama
enter to the cmd
ollama pull conceptsintamil/tamil-llama-7b-instruct-v0.2
test the model if it is run correctly
ollama run conceptsintamil/tamil-llama-7b-instruct-v0.2
then
C:\Users\(your_path)>code .
cd ..
cd ..
cd ..
cd ..
cd ollama
pip -m venv venv
cd venv\scripts
activate.bat
cd ..
pip install -r requirements.txt
streamlit run app.py
cd ..
cd quantize
streamlit run app.py
Comments