Whole Home Voice Control
Gesture Control for Lights and Audio
Problem
<text>
Solution
The idea is to develop a user programmable gesture-controlled home system, where people can program their own gestures to control lights, audio, emergency commands and SeaPod steering.
The system has to be smart enough to recognise users and know their permissions to commands. For example, the owner might want to give SeaPod steering access to a friend but not for kids.
Also, the system has to be secure so there’s no data leakage possibilities.
Prize
- Get credited as a Project Contributor to the Ocean Builders Project
- Turn this into your own entrepreneurial business venture and we will be your first customers and help bring you media attention and customers
- Get Entrepreneurial Business Coaching to start this as a business
And here are some potential benefits:
- Mass exposure with highly visible project
- Build reputation
- Recognized as an official collaborator/ and/or on Github
- Get noticed
- Product development experience
- Work on projects you are passionate about
- Get your project built and working in the real world
- Participate in interesting work
- Get grants (maybe partner with someone that can help with this or exposure to grant writers)
- Change the world
Industry
There are a few gesture-controlled home systems in the market like Single Cue, that lets you control your home media devices with your finger movements and Fibaro Swipe. Unfortunately their reviews are not so good.
Below, there are listed some open source projects that might be helpful for your R&D:
Teachable Machine is a web-based tool that you can use to train a computer to recognize your own images, sounds, & poses.
Here you can learn how to make a gesture-controlled Arduino robot using PictoBlox AI’s machine learning feature.
In this video, the developer demonstrates how to build a gesture recognition system and use it to control the media player on the computer. OpenCV + Machine Learning + GUI Automation were used to do this.
Here is a project where Alexa is responding to sign language using webcam and TensorFlow.js
Here you can read an article about how an open source Google sign language AI turns hand gestures into speech.
In this tutorial (see here), you can learn how to build an American Sign Language translator using computer vision and a machine learning model.
MediaPipe Hands is a high-fidelity hand and finger tracking solution. It employs machine learning (ML) to infer 21 3D landmarks of a hand from just a single frame. MediaPipe has also other solutions to detect objects like faces, iris, poses and more.
Information
Repository
<text>
License Requirement
Open Source: Can be used for private or commercial projects
Software: GNU General Public License (GNU GPL V3) here
Non-Software: Creative Commons (CC BY-SA 4.0) here
Project Areas
- Software (ML, AI, Computer Visualization, Neural Network)
- More?
Objectives
First part of the project should be a research on what gestures to use for particular commands.
The second part of the project is to develop a gesture control system for lights, audio, emergency and steering.
Project’s Requirements
Stages and deadlines
Project Start |
date |
Team Formed |
date |
Market Research Summary (Report) |
date |
Project Plan Complete |
date |
Preliminary Product Design Complete |
date |
Prototype Development Complete |
date |
Prototype Evaluation Complete |
date |
Product Presentation |
date |
Project Completion |
date |
Project plan should cover the following:
- stages / milestones of a project (not all stages are brought out in a table above)
- activities or tasks in each phase
- task start and end dates
- interdependencies between tasks
Also:
- skills needed
- responsibilities of each team member (identify as many as you can).
Preliminary product design should cover the following:
- production components, raw material
- system block diagram (This diagram specifies each electronic function and how all of the functional components interconnect).
- preliminary Bill of Materials (BOM)
- production cost estimation
- performance
- features
- development feasibility
- manufacturability
Product’s requirements
https://docs.google.com/spreadsheets/d/1X7GzTJoPGgzrKe34GE-U0_3xLz8_-9yF66NEi7J3djc/edit#gid=0
(Will remove URL before publishing)
Basic |
Advanced |
Feature |
Can you control the audio system? |
||
Can you control lights? |
||
Can you control TV? |
||
Can you control SeaPod steering? |
||
Can you control functions for emergencies? |
||
Is it user programmable (users can program their own gestures)? |
||
Can the system recognize users (visual identity)? |
||
Can you select who can have access to particular home control functions? |
||
Can the user create its own gestures? |
||
Is the data sandboxed to protect privacy? |
||
Can it recognize gestures at any angle? |
||
Can it recognize gestures in the dark? |
||
Can it easily understand users' gestures? |
||
Is it user friendly to set up? |
||
Is it more convenient to use than remote control? |
||
Can I use it in any room? |
||
Can I use it in any location of the room? |
Collaboration instructions
It’s up to each team how to break down the project into stages & milestones but it should include at least stages that are brought out in the Project Requirements chapter.
It’s also necessary to open an account in Trello, to manage your project. Please create at least 3 columns that are as follows:
1) To do
2) Doing
3) Done
And give access to your team lead and Ocean Builders manager as well, so that your team lead can keep track of your progress, comment on Trello cards and also have weekly team meetings. Below you can find their contacts:
Team lead:
Name: <text>
Email: <text>
Ocean Builders manager:
Name: <text>
Email: <text>
Project video link:
https://www.dropbox.com/s/j44y0z574mt1ohh/VoiceControl.mp4?dl=0