terrykwon / captivate

Code for 'Captivate! Contextual Language Guidance for Parent–Child Interaction' (CHI 2022)

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Captivate! Contextual Language Guidance for Parent–Child Interaction

This repo contains the server-side source code for Captivate!, a system that analyzes real-time video and audio streams to provide language guidance for parents in play situations.

[Paper] [Website]

System overview

The system separately processes video and audio streams, with the goal of extracting attention-related cues and combining them to infer a joint attention distribution. The joint attention distribution provides the basis for the context-relevant phrases that are recommended to the parent in real-time. Below is an overview of the system.

System overview

Code overview

The processing pipline for video and audio is in the modelserver directory. The main process is modelserver.modelserver.run_server.py, wihch starts several parallel processes that fetch and handle real-time data streams from an RTMP video server. Each process writes model inference outputs to a shared queue in the main process, which combines the information to calculate the joint attention distribution.

Process overview

In our study, we used three cameras, so therefore launch three visual processes that perform the same pipeline on each of the video streams. Note that this number can be changed easily.

Finally, communication between the server and client is done in real-time with WebSockets.

Running

Currently the code serves mainly as a reference, as running takes quite a bit of setup (and the client code is not open yet). That said, the Docker image for the modelserver can be built and run with.

./build.sh
./run.sh

Citing

@inproceedings{kwon2022captivate,
  author = {Kwon, Taeahn and Jeong, Minkyung and Ko, Eon-Suk and Lee, Youngki},
  title = {Captivate! Contextual Language Guidance for Parent–Child Interaction},
  year = {2022},
  isbn = {9781450391573},
  publisher = {Association for Computing Machinery},
  address = {New York, NY, USA},
  url = {https://doi.org/10.1145/3491102.3501865},
  doi = {10.1145/3491102.3501865},
  booktitle = {CHI Conference on Human Factors in Computing Systems},
  articleno = {219},
  numpages = {17},
  location = {New Orleans, LA, USA},
  series = {CHI '22}
}

Acknowledgements

We would like to thank and mention the following repositories and their authors for providing the models used in this system.

About

Code for 'Captivate! Contextual Language Guidance for Parent–Child Interaction' (CHI 2022)

License:MIT License


Languages

Language:Python 98.8%Language:Dockerfile 0.8%Language:Shell 0.4%