Skip to content

Code for 'Captivate! Contextual Language Guidance for Parent–Child Interaction' (CHI 2022)

License

Notifications You must be signed in to change notification settings

terrykwon/captivate

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

32 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Captivate! Contextual Language Guidance for Parent–Child Interaction

This repo contains the server-side source code for Captivate!, a system that analyzes real-time video and audio streams to provide language guidance for parents in play situations.

[Paper] [Website]

System overview

The system separately processes video and audio streams, with the goal of extracting attention-related cues and combining them to infer a joint attention distribution. The joint attention distribution provides the basis for the context-relevant phrases that are recommended to the parent in real-time. Below is an overview of the system.

System overview

Code overview

The processing pipline for video and audio is in the modelserver directory. The main process is modelserver.modelserver.run_server.py, wihch starts several parallel processes that fetch and handle real-time data streams from an RTMP video server. Each process writes model inference outputs to a shared queue in the main process, which combines the information to calculate the joint attention distribution.

Process overview

In our study, we used three cameras, so therefore launch three visual processes that perform the same pipeline on each of the video streams. Note that this number can be changed easily.

Finally, communication between the server and client is done in real-time with WebSockets.

Running

Currently the code serves mainly as a reference, as running takes quite a bit of setup (and the client code is not open yet). That said, the Docker image for the modelserver can be built and run with.

./build.sh
./run.sh

Citing

@inproceedings{kwon2022captivate,
  author = {Kwon, Taeahn and Jeong, Minkyung and Ko, Eon-Suk and Lee, Youngki},
  title = {Captivate! Contextual Language Guidance for Parent–Child Interaction},
  year = {2022},
  isbn = {9781450391573},
  publisher = {Association for Computing Machinery},
  address = {New York, NY, USA},
  url = {https://doi.org/10.1145/3491102.3501865},
  doi = {10.1145/3491102.3501865},
  booktitle = {CHI Conference on Human Factors in Computing Systems},
  articleno = {219},
  numpages = {17},
  location = {New Orleans, LA, USA},
  series = {CHI '22}
}

Acknowledgements

We would like to thank and mention the following repositories and their authors for providing the models used in this system.

About

Code for 'Captivate! Contextual Language Guidance for Parent–Child Interaction' (CHI 2022)

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages