dkhusted / Wally

A tiny robot helper

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Wally

A tiny helper https://www.kaggle.com/code/sayakdasgupta/sign-language-classification-cnn-99-40-accuracy

Table of Contents

General Information

  • The project propose an different approch to smart homes. Instead of using voice commands to get different kinds of information via a traditional smart-home device, like Alexy or Google Home, Wally will use sign language to determine the user request.
  • The reason for this is to inable people with the mute dissability to also have access to an smart home device, and that image processing would be more reliable than audio since bagground noice, such as other people or tv, would not have an effect. The only noice factor would be the light level.

Technologies Used

Microprocessor: Particle Argon, Arduino Nano BLE 33 IDE: Visual Studio Code Intern MCU communication: I2C

Features

Sending 'c' via a serial monitor, like putty, to Arduino will cause it to read a frame from OV7675. Arduino generates only features via Edge Impulses DPS library, no classification as there is not enough space for both classification model and DSP model. Arduino sits as Master on I2C, to be able to talk to OV7675 camera, and sends the generatet features to Argon.

Requirements

Edge Impulse API Arduino Nano 33 BLE Sense Lite Particle Argon 2 x 4.7 kOhm Resistors

Camera

OV7675 (I used it with the Tiny Machine Learning Shield, but you can connect it manually with wires to the Arduino)

Setup

We are going to setup the

Usage

How does one go about using it? Provide various use cases and code examples here.

write-your-code-here

Project Status

Project is: _in_definement

Acknowledgements

Contact

Created by [@dkhusted] - feel free to contact me!

About

A tiny robot helper


Languages

Language:C 77.8%Language:C++ 21.8%Language:Jupyter Notebook 0.3%Language:CMake 0.1%Language:Shell 0.0%