Smart beaker based on multimodal fusion and intentional understanding

1Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In the current simulation experiment system, the experimental design of single mode is less interactive and less accurate. In order to solve this problem, this paper proposes an experimental interaction kit based on sound and sensor, and designs a multimodal fusion and intent understanding algorithm. Firstly, the method of multi-sensor signal extraction and speech feature extraction is introduced. Then, based on the results obtained by the two methods, an algorithm based on decision-level fusion is studied, which solves the problem of perception of user's operation intention in virtual chemistry experiments. Finally, the usability of the multimodal intent understanding algorithm proposed in this paper is verified by designing a complete chemical experiment system. Experiments show that the multimodal intent understanding algorithm based on sensor and speech input is due to a single modality in terms of interactivity and accuracy, and the physical interaction suite designed in this paper greatly improves the intelligence and interactivity of the system.

Cite

CITATION STYLE

APA

Dong, D., Feng, Z., & Tian, J. (2020). Smart beaker based on multimodal fusion and intentional understanding. In ACM International Conference Proceeding Series (pp. 206–211). Association for Computing Machinery. https://doi.org/10.1145/3379247.3379257

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free