Wraps MediaPipe Handpose inside electron and serves the detected parts via MaxAPI.
Based around Yuichi Yogo's great work porting Electron + Tensorflow models into Node For Max
- Handpose is being used in Arthur Parmentier's Soundpainting Project
- Are you using n4m-handpose in your project? Let me know
- Install npm dependencies by clicking the indicated button. Since Electron's kind of big in size, this make take a while depending on your network environment. When the message object says it's 'completed' you'll never have to do this step again.
- Click on the toggle button to launch n4m-handpose.
- The node.script emits the detected results as a dictionary, which has been parsed into (x,y,z) lists of each finger's tip positions & palmbase
- All landmarks are also aggregated to a list to send to Wekinator if you'd like to perform gesture recognition with the data (sending to port 11000)
- n4m-handposeM4L (Max For Live Device) AMXD coming in 1.1
- Skeleton color can be changed using GUI
- Settings are persisted using electron-store
- Handpose model controls now available in electron window UI
- Skeleton fill and stroke colour available in electron window UI
- Handpose model control settings and skeleton fill and stroke colour save automatically and persist/reload
- Render window on or offscreen using arguments to node.script
- Hand landmarks are scaled and rendered to a jitter window (m4l + maxmsp)