Consult an Expert
Trademark
Design Registration
Consult an Expert
Trademark
Copyright
Patent
Infringement
Design Registration
More
Consult an Expert
Consult an Expert
Trademark
Design Registration
Login
GESTURE NAVIGATOR: AI-BASED VIRTUAL MOUSE
Extensive patent search conducted by a registered patent agent
Patent search done by experts in under 48hrs
₹999
₹399
Abstract
Information
Inventors
Applicants
Specification
Documents
ORDINARY APPLICATION
Published
Filed on 8 November 2024
Abstract
The "Gesture Navigator: AI-Based Virtual Mouse" is an innovative system that enables natural human-computer interaction through gesture recognition. Utilizing a low-cost USB webcam, the system captures real-time video of user hand movements and processes the frames to interpret gestures. This technology facilitates a range of functionalities, including cursor control, drawing operations, and a virtual keyboard interface, all through intuitive hand movements. Additionally, the system features voice-to-text conversion, allowing users to input text verbally, thereby minimizing reliance on physical devices. With the integration of a virtual assistant, the system provides contextual feedback and guidance, enhancing accessibility and user experience for individuals of all ages, particularly those who may struggle with traditional input methods. This invention represents a significant advancement in gesture recognition technology, making computing more intuitive and accessible. (Accompanied Figure No. 1)
Patent Information
Application ID | 202411086292 |
Invention Field | COMPUTER SCIENCE |
Date of Application | 08/11/2024 |
Publication Number | 47/2024 |
Inventors
Name | Address | Country | Nationality |
---|---|---|---|
Ravikant | Department of CSE, IMS Engineering College, Ghaziabad, Uttar Pradesh, India. | India | India |
Mayank Gupta | Department of CSE, IMS Engineering College, Ghaziabad, Uttar Pradesh, India. | India | India |
Kanak Singh | Department of CSE, IMS Engineering College, Ghaziabad, Uttar Pradesh, India. | India | India |
Ashi Tyagi | Department of CSE, IMS Engineering College, Ghaziabad, Uttar Pradesh, India. | India | India |
Ayush Tiwari | Department of CSE, IMS Engineering College, Ghaziabad, Uttar Pradesh, India. | India | India |
Divyansh Ruhela | Department of CSE, IMS Engineering College, Ghaziabad, Uttar Pradesh, India. | India | India |
Applicants
Name | Address | Country | Nationality |
---|---|---|---|
IMS Engineering College | National Highway 24, Near Dasna, Adhyatmik Nagar, Ghaziabad, Uttar Pradesh- 201015 | India | India |
Specification
Description:The present invention relates to the field of human-computer interaction (HCI) and, more specifically, to gesture recognition technologies. This innovation aims to provide a novel approach for users to interact with computer systems through natural hand movements, eliminating the need for conventional input devices such as keyboards, mice, or touchscreens. This invention also incorporates artificial intelligence (AI) to enhance the accuracy and efficiency of gesture recognition, making it suitable for various applications, including gaming, design, and accessibility solutions for users with disabilities.
BACKGROUND OF THE INVENTION
Human gestures have been a fundamental part of communication for centuries, serving as a universal language that transcends verbal barriers. Common gestures such as thumbs up, waving, and pointing convey messages effortlessly. As technology continues to advance, the limitations of traditional input methods, such as keyboards and mice, become more apparent, particularly for users who may face challenges with dexterity or familiarity with these devices.
Traditional interfaces can be cumbersome, leading to frustration among users who are not adept at using them, including elderly individuals and those with disabilities. Gesture recognition technologies offer a promising solution by allowing users to interact with computers using intuitive hand movements, thereby making technology more accessible and user-friendly. Existing systems, however, often require expensive hardware or complex setups, limiting their adoption.
This invention addresses these issues by providing an affordable and straightforward solution that leverages a low-cost USB webcam and advanced algorithms to interpret gestures, promoting a more natural and engaging interaction with technology.
OBJECTS OF THE INVENTION
An object of the present invention is to create an interface that enables users to interact with computer systems using hand gestures, thus promoting a more intuitive and human-like interaction.
Another object of the present invention is to develop a system that can be easily used by people of all ages and abilities, including the elderly and those with disabilities, by reducing reliance on traditional input devices.
Yet another object of the present invention is to allow users to perform various functions, such as cursor control, drawing, typing, and voice commands, all through simple hand movements.
Another object of the present invention is to utilize a low-cost USB webcam as the primary input device, making the technology affordable and accessible to a broader audience.
Another object of the present invention is to incorporate a virtual assistant that enhances user interaction by providing real-time feedback and guidance, improving the overall experience.
Another object of the present invention is to implement a voice-to-text conversion feature, reducing the need for physical input devices and streamlining the interaction process.
SUMMARY OF THE INVENTION
The Gesture Navigator: AI-Based Virtual Mouse provides a novel solution for human-computer interaction by utilizing a low-cost USB webcam to capture user gestures in real time. The system operates through a four-step process: video capturing, frame processing, region extraction, and feature matching.
Video Capturing: The USB webcam continuously captures video of the user's hand movements, providing a live feed to the processing unit.
Frame Processing: Each frame is analyzed using image processing techniques to isolate and identify hand gestures against the background.
Region Extraction: The system identifies specific regions of interest, focusing on the hands and any relevant movements that correspond to predefined gestures.
Feature Matching: Advanced algorithms are employed to recognize gestures and translate them into corresponding actions, such as moving the cursor, drawing, or activating commands.
The system supports a variety of functions, including cursor navigation, virtual keyboard usage, and drawing capabilities. Additionally, a built-in voice recognition feature allows users to input text verbally, further reducing the need for physical interaction. The integration of a virtual assistant enhances usability by providing contextual feedback, making it easier for users to navigate and utilize the system effectively.
In this respect, before explaining at least one object of the invention in detail, it is to be understood that the invention is not limited in its application to the details of set of rules and to the arrangements of the various models set forth in the following description or illustrated in the drawings. The invention is capable of other objects and of being practiced and carried out in various ways, according to the need of that industry. Also, it is to be understood that the phraseology and terminology employed herein are for the purpose of description and should not be regarded as limiting.
These together with other objects of the invention, along with the various features of novelty which characterize the invention, are pointed out with particularity in the disclosure. For a better understanding of the invention, its operating advantages and the specific objects attained by its uses, reference should be made to the accompanying drawings and descriptive matter in which there are illustrated preferred embodiments of the invention.
BRIEF DESCRIPTION OF DRAWINGS
The advantages and features of the present invention will be understood better with reference to the following detailed description and claims taken in conjunction with the accompanying drawings, wherein like elements are identified with like symbols, and in which:
Figure 1 illustrates the block diagram for the Gesture Navigator: AI-based virtual mouse in accordance with the present invention.
DETAILED DESCRIPTION OF THE INVENTION
An embodiment of this invention, illustrating its features, will now be described in detail. The words "comprising," "having," "containing," and "including," and other forms thereof are intended to be equivalent in meaning and be open-ended in that an item or items following any one of these words is not meant to be an exhaustive listing of such item or items or meant to be limited to only the listed item or items.
The terms "first," "second," and the like, herein do not denote any order, quantity, or importance, but rather are used to distinguish one element from another, and the terms "a" and "an" herein do not denote a limitation of quantity, but rather denote the presence of at least one of the referenced item.
The Gesture Navigator: AI-Based Virtual Mouse integrates advanced technology to create an intuitive and accessible interface for human-computer interaction. The system is composed of several key components and features designed to enhance the user experience, facilitate natural interaction, and promote usability across various applications.
1. Video Capturing
Hardware Configuration: The system employs a low-cost USB webcam, selected for its affordability and accessibility. This camera is positioned to ensure optimal visibility of the user's hand movements while minimizing obstructions from the environment.
Continuous Capture: The webcam operates in real-time, capturing video frames at a high frame rate to ensure smooth tracking of hand gestures. This continuous video feed is crucial for maintaining responsiveness and accuracy in gesture recognition.
Adaptability to Lighting Conditions: The system is designed to function effectively under varying lighting conditions. Algorithms are implemented to adjust the image contrast and brightness dynamically, ensuring clear visibility of the user's hands even in low-light environments.
2. Frame Processing
Image Analysis Techniques: Each video frame captured by the webcam is processed using sophisticated image analysis techniques. These techniques include background subtraction, edge detection, and contour analysis to enhance the visibility of hand gestures.
Hand Detection: The frame processing module utilizes machine learning models trained on large datasets of hand images to identify the presence of hands in the captured frames. The system can recognize hands in different orientations and lighting conditions.
Noise Reduction: Image filtering methods are applied to reduce noise and enhance the quality of the captured images, allowing for more accurate gesture detection. This step is essential for minimizing false positives in gesture recognition.
3. Region Extraction
Segmentation Algorithms: The system employs segmentation algorithms to isolate the user's hands from the background. This process involves creating a mask that highlights the hand regions while suppressing the background noise.
Depth Perception: If a stereo camera setup is utilized, depth perception algorithms can be applied to accurately determine the position of the hands in three-dimensional space, enhancing gesture recognition capabilities.
Real-Time Processing: The region extraction process is optimized for real-time operation, allowing the system to update the recognized hand positions and movements continuously as the user interacts with the computer.
4. Feature Matching
Machine Learning Models: The gesture recognition process relies on machine learning models, such as convolutional neural networks (CNNs), which are specifically designed to recognize complex patterns in images. These models are trained on diverse datasets containing various hand gestures to improve their accuracy.
Gesture Database: A comprehensive database of predefined gestures is maintained within the system, including common gestures like swiping, pinching, pointing, and multi-finger movements. Users can also add custom gestures to personalize their experience.
Contextual Understanding: The system can interpret gestures within context, allowing for more natural interactions. For instance, a swipe gesture could be recognized as a command to scroll up or down, depending on the application in use.
5. Functional Capabilities
Cursor Control:
Movement Detection: Users can control the on-screen cursor by moving their hands in a specific direction. The system translates the movement of the user's hand into corresponding cursor movement, enhancing the user's ability to navigate the interface intuitively.
Speed Adjustment: The sensitivity of cursor movement can be adjusted based on user preferences, allowing for fine-tuned control, which is especially beneficial for tasks requiring precision.
Drawing Operations:
Virtual Drawing Surface: The system provides a virtual canvas on which users can draw by moving their fingers in the air, simulating traditional drawing tools like pens or brushes. This feature is particularly advantageous for artists and designers.
Tool Selection: Users can select different drawing tools (e.g., pencil, brush, eraser) through specific hand gestures, enhancing the interactive drawing experience without requiring physical tools.
Virtual Keyboard:
Dynamic Display: A virtual keyboard can be projected on the screen, allowing users to input text by pointing at the keys. The keyboard can be dynamically resized and repositioned based on user preferences and screen size.
Gesture Typing: Users can perform gesture-based typing by hovering over letters, with the system recognizing their intentions without the need for physical contact.
Voice-to-Text Conversion:
Integrated Voice Recognition: The system incorporates voice recognition technology to allow users to dictate text, which is then converted to written form in real time. This feature supports various languages and dialects, making it accessible to a broader audience.
Command Recognition: Users can also issue voice commands to perform specific actions, such as opening applications or navigating menus, further enhancing hands-free interaction.
6. Integration of Virtual Assistant
Contextual Assistance: The virtual assistant acts as a guide for users, providing real-time feedback on recognized gestures and suggesting actions based on user behavior. This feature helps users understand the system's capabilities and enhances their overall experience.
Tutorial Features: The virtual assistant can offer tutorials and assistance for new users, walking them through the setup process and demonstrating how to perform various gestures and commands.
User Adaptability: The assistant learns from user interactions, adapting its responses and suggestions over time to align with the user's preferences and usage patterns.
7. Applications
Gaming: The Gesture Navigator can enhance the gaming experience by allowing players to control characters or navigate game menus using hand gestures, creating a more immersive experience.
Design and Creativity: Artists and designers can leverage the system to create digital artwork, illustrations, or animations using natural gestures, facilitating a more intuitive creative process.
Educational Tools: The technology can be integrated into educational applications, enabling interactive learning experiences for students and educators through gesture-based interactions.
Assistive Technologies: The Gesture Navigator is particularly beneficial for individuals with disabilities, providing an alternative means of interaction that reduces the barriers associated with traditional input devices.
The foregoing descriptions of specific embodiments of the present invention have been presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the present invention to the precise forms disclosed, and obviously many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described to best explain the principles of the present invention, and its practical application to thereby enable others skilled in the art to best utilize the present invention and various embodiments with various modifications as are suited to the particular use contemplated. It is understood that various omission and substitutions of equivalents are contemplated as circumstance may suggest or render expedient, but such are intended to cover the application or implementation without departing from the spirit or scope of the claims of the present invention.
, Claims:1. A gesture recognition system for human-computer interaction, comprising:
a USB webcam configured to capture real-time video input of a user's hand gestures;
frame processing software for analyzing captured video frames to extract features relevant to gesture recognition;
region extraction algorithms for isolating hand gestures from the background;
machine learning models for interpreting gestures and translating them into corresponding actions;
a virtual assistant for facilitating user interaction and providing feedback based on recognized gestures.
2. A method for facilitating human-computer interaction through gesture recognition, comprising the steps of:
a) capturing real-time video of a user's hand gestures using a USB webcam;
b) processing the captured video frames to enhance image quality and detect hand movements;
c) extracting regions of interest corresponding to hand gestures;
d) matching recognized gestures to predefined actions for cursor control, drawing operations, and virtual keyboard input.
3. The system as claimed in claim 1, wherein the machine learning models employed for feature matching include convolutional neural networks (CNNs) trained on a diverse dataset of hand gestures to improve recognition accuracy.
4. The system as claimed in claim 1, wherein the gesture recognition system is capable of interpreting at least ten distinct hand gestures, allowing for a diverse set of commands.
5. The method as claimed in claim 2, further comprising the step of dynamically adjusting the sensitivity of cursor movement based on user preferences to enhance precision in navigation.
6. The system as claimed in claim 1, wherein the region extraction algorithms utilize segmentation techniques to isolate the user's hands from the background, enabling accurate gesture recognition.
7. The system as claimed in claim 1, further comprising a virtual keyboard displayed on the screen, allowing users to select keys by pointing at them, enabling text input through gesture-based typing.
8. The method as claimed in claim 2, further comprising the step of integrating a voice recognition module to allow users to dictate text and issue commands verbally, minimizing the need for physical interaction.
9. The system as claimed in claim 1, wherein the virtual assistant provides real-time feedback on recognized gestures, helping users navigate the system more effectively and efficiently.
10. The system as claimed in claim 1, further comprising a dynamic drawing surface that allows users to draw by moving their fingers in the air, simulating traditional drawing tools for creative applications.
Documents
Name | Date |
---|---|
202411086292-COMPLETE SPECIFICATION [08-11-2024(online)].pdf | 08/11/2024 |
202411086292-DECLARATION OF INVENTORSHIP (FORM 5) [08-11-2024(online)].pdf | 08/11/2024 |
202411086292-DRAWINGS [08-11-2024(online)].pdf | 08/11/2024 |
202411086292-FORM 1 [08-11-2024(online)].pdf | 08/11/2024 |
202411086292-FORM-9 [08-11-2024(online)].pdf | 08/11/2024 |
202411086292-REQUEST FOR EARLY PUBLICATION(FORM-9) [08-11-2024(online)].pdf | 08/11/2024 |
Talk To Experts
Calculators
Downloads
By continuing past this page, you agree to our Terms of Service,, Cookie Policy, Privacy Policy and Refund Policy © - Uber9 Business Process Services Private Limited. All rights reserved.
Uber9 Business Process Services Private Limited, CIN - U74900TN2014PTC098414, GSTIN - 33AABCU7650C1ZM, Registered Office Address - F-97, Newry Shreya Apartments Anna Nagar East, Chennai, Tamil Nadu 600102, India.
Please note that we are a facilitating platform enabling access to reliable professionals. We are not a law firm and do not provide legal services ourselves. The information on this website is for the purpose of knowledge only and should not be relied upon as legal advice or opinion.