Whole Home Voice Control
Gesture Control for Lights and Audio
Problem
Typing is not always the best way to get information (e.g. what’s the weather?) or give commands (e.g. turn the kitchen lights off!) in the application. <text>
Solution
The idea is to createdevelop a Voiceuser Assistantprogrammable gesture-controlled home system, where people can program their own gestures to control deviceslights, inaudio, theemergency commands and SeaPod and ask for information that is available in the Ocean Builder’s applications. steering.
The voicesystem assistanthas shouldto be smart enough to recognizerecognise users and beknow abletheir permissions to follow commands, considering users’ settings and permissions.commands. For example, when the userowner asksmight want to turngive onSeaPod thesteering showeraccess thento thea shrower will be turned on and adjusted automatically based on the user's preferences,friend but ifnot thefor user doesn’t have permission to use that shower then the command will be ignored. kids.
Also, the system has to be secure so there’s no data leakage possibilities.
Prize
-Get credited as a Project Contributor to the Ocean Builders ProjectTurn this into your own entrepreneurial business venture and we will be your first customers and help bring you media attention and customers
Get Entrepreneurial Business Coaching to start this as a business
And here are some potential benefits:
- Mass exposure with highly visible project
Build reputation
Recognized as an official collaborator/ and/or on Github
Get noticed
Product development experience
Work on projects you are passionate about
Get your project built and working in the real world
Participate in interesting work
Get grants (maybe partner with someone that can help with this or exposure to grant writers)
Change the world
Industry
Current technological level
You’ve probably heard of voice assistants like Alexa, Siri, Google Assistant, and Cortana. These voice assistants are essentially based on voice recognition, NLP, and synthesis of speech (see picture below).
There are alsoa manyfew gesture-controlled home systems in the market like Single Cue, that lets you control your home media devices with your finger movements and Fibaro Swipe. Unfortunately their reviews are not so good.
Below, there are listed some open source projects like:that might be helpful for your R&D:
Teachable Machine is a web-based tool that you can use to train a computer to recognize your own images, sounds, & poses.
-
MycroftHere you can learn how to make a gesture-controlled Arduino robot using PictoBlox AI’s machine learning feature. - you
- can read an
- article about how
In this video, the developer demonstrates how to build a gesture recognition system and use it to control the media player on the computer. OpenCV + Machine Learning + GUI Automation were used to do this.
OpenAssistantHere is a project where Alexa is responding to sign language using webcam and TensorFlow.js
Many of thesean open source voiceGoogle assistantssign havelanguage comeAI turns hand gestures into existence quite recently and will probably take some time to develop into a more sophisticated solution.speech.
this tutorial (see
Some projects such as MyCroft offer solutions built around Google Home or Alexa. However, certain characteristics of these systems - no data protection and no business vocabulary adaptation - limit them to a B2Cmachine marketlearning that is not (yet) concerned by data sensitivity and criticality issues.model.
There are also some platforms like LinTo, that embraces these challenges from the start in order to be the engine that catapults your professional product.
One of the biggest challenges might be to implement a voice authenticator. Here are some projects to check out to see if any of them could be a fit to integrate with our system:
Here'sMediaPipe Hands is a websitehigh-fidelity abouthand openand sourcefinger projects.tracking Isolution. broughtIt outemploys somemachine morelearning interesting(ML) projectsto below:
https://awesomeopensource.com/project/pyannote/pyannote-audio
https://codeocean.com/capsule/7271435/tree/v1
https://github.com/mravanelli/pytorch-kaldi
https://awesomeopensource.com/project/google/uis-rnn
https://alize.univ-avignon.fr/
Itframe. MediaPipe has aalso Javaother versionsolutions asto well. detect objects like faces, iris, poses and more.
Information
Repository
<text>
License Requirement
Open Source: Can be used for private or commercial projects
Software:Software: GNU General Public License (GNU GPL V3) here
Non-Software: Creative Commons (CC BY-SA 4.0) here
Project Areas
IoT DevelopmentSoftware (sensors,ML,arduinoAI,andComputerraspberryVisualization,pi)Neural Network)Software Development (python)?More?<text>
Objectives
First part of the project should be a research on what gestures to use for particular commands.
Keywords:The <text>second part of the project is to develop a gesture control system for lights, audio, emergency and steering.
Project’s Project requirementsRequirements
Stages and deadlines
Project Start |
date |
Team Formed |
date |
Market Research Summary (Report) |
date |
Project Plan Complete |
date |
Preliminary Product Design Complete |
date |
Prototype Development Complete |
date |
Prototype Evaluation Complete |
date |
Product Presentation |
date |
Project Completion |
date |
Project plan should cover the following:
stages / milestones of a project (not all stages are brought out in a table above)
activities or tasks in each phase
task start and end dates
interdependencies between tasks
Also:
skills needed
responsibilities of each team member (identify as many as you can).
Preliminary product design should cover the following:
- production components, raw material
- system block diagram (This diagram specifies each electronic function and how all of the functional components interconnect).
- preliminary Bill of Materials (BOM)
- production cost estimation
- performance
- features
- development feasibility
- manufacturability
Product’s general requirements
(Will remove URL before publishing)publishing)
Basic |
Advanced |
|
|
||
Can you control lights? |
||
Can you control TV? |
||
Can you control SeaPod steering? |
||
Can you control functions for emergencies? |
||
Is it user programmable (users can program their own gestures)? |
||
Can the system recognize users (visual identity)? |
||
Can you select who can have access to particular home control functions? |
||
Can the user create its own gestures? |
||
Is the data sandboxed to protect privacy? |
||
Can it |
||
Can it recognize gestures in the dark? |
||
Can it easily understand | ||
Is |
||
| than ||
Can I |
||
Can I | ||
Collaboration instructions
It’s up to each team how to break down the project into stages & milestones but it should include at least stages that are brought out in the Project Requirements chapter.
TipsIt’s also necessary to open an account in Trello, to manage your project. Please create at least 3 columns that are as follows:
1) To do
2) Doing
3) Done
And give access to your team lead and Ocean Builders manager as well, so that your team lead can keep track of your progress, comment on Trello cards and also have weekly team meetings. Below you can find sometheir examples of tools to use to build Voice Assistance:
contacts:
gTTS (Google Text-to-Speech) is a speech synthesis library to convert text to speech.
SpeechRecognition is a library for performing speech recognition, with support for several engines and APIs, online and offline
Sphinx is the offline recognition engine called by the SpeechRecognition library.
Packt is a voice recognition library to identify the person who is speaking
Team lead:
VoiceName: Authentication <text>
https://courses.csail.mit.edu/6.857/2016/files/31.pdfEmail:
<text>
NotOcean openBuilders source:manager:
Name: <text>
https://docs.google.com/document/d/1V-cyxivxKFXwYVUO21oleuXcAoXzR8QTKkpI5ug6E0A/editEmail:
<text>
Project video link:
https://www.dropbox.com/s/j44y0z574mt1ohh/VoiceControl.mp4?dl=0