Audio-Visual Tensor Fusion Network for Piano Player Posture Classification
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Park, So-Hyun | - |
dc.contributor.author | Park, Young-Ho | - |
dc.date.available | 2021-02-22T05:21:46Z | - |
dc.date.issued | 2020-10 | - |
dc.identifier.issn | 2076-3417 | - |
dc.identifier.issn | 2076-3417 | - |
dc.identifier.uri | https://scholarworks.sookmyung.ac.kr/handle/2020.sw.sookmyung/1149 | - |
dc.description.abstract | Playing the piano in the correct position is important because the correct position helps to produce good sound and prevents injuries. Many studies have been conducted in the field of piano playing posture recognition that combines various techniques. Most of these techniques are based on analyzing visual information. However, in the piano education field, it is essential to utilize audio information in addition to visual information due to the deep relationship between posture and sound. In this paper, we propose an audio-visual tensor fusion network (simply, AV-TFN) for piano performance posture classification. Unlike existing studies that used only visual information, the proposed method uses audio information to improve the accuracy in classifying the postures of professional and amateur pianists. For this, we first introduce a dataset called C3Pap (Classic piano performance postures of amateur and professionals) that contains actual piano performance videos in diverse environments. Furthermore, we propose a data structure that represents audio-visual information. The proposed data structure represents audio information on the color scale and visual information on the black and white scale for representing relativeness between them. We call this data structure an audio-visual tensor. Finally, we compare the performance of the proposed method with state-of-the-art approaches: VN (Visual Network), AN (Audio Network), AVN (Audio-Visual Network) with concatenation and attention techniques. The experiment results demonstrate that AV-TFN outperforms existing studies and, thus, can be effectively used in the classification of piano playing postures. | - |
dc.format.extent | 15 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | MDPI | - |
dc.title | Audio-Visual Tensor Fusion Network for Piano Player Posture Classification | - |
dc.type | Article | - |
dc.publisher.location | Switzerland | - |
dc.identifier.doi | 10.3390/app10196857 | - |
dc.identifier.scopusid | 2-s2.0-85092777509 | - |
dc.identifier.wosid | 000586586900001 | - |
dc.identifier.bibliographicCitation | APPLIED SCIENCES-BASEL, v.10, no.19, pp 1 - 15 | - |
dc.citation.title | APPLIED SCIENCES-BASEL | - |
dc.citation.volume | 10 | - |
dc.citation.number | 19 | - |
dc.citation.startPage | 1 | - |
dc.citation.endPage | 15 | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Chemistry | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Materials Science | - |
dc.relation.journalResearchArea | Physics | - |
dc.relation.journalWebOfScienceCategory | Chemistry, Multidisciplinary | - |
dc.relation.journalWebOfScienceCategory | Engineering, Multidisciplinary | - |
dc.relation.journalWebOfScienceCategory | Materials Science, Multidisciplinary | - |
dc.relation.journalWebOfScienceCategory | Physics, Applied | - |
dc.subject.keywordPlus | KINEMATICS | - |
dc.subject.keywordPlus | KINECT | - |
dc.subject.keywordAuthor | piano playing posture | - |
dc.subject.keywordAuthor | audio-visual tensor fusion | - |
dc.subject.keywordAuthor | classification | - |
dc.identifier.url | https://www.mdpi.com/2076-3417/10/19/6857 | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
Sookmyung Women's University. Cheongpa-ro 47-gil 100 (Cheongpa-dong 2ga), Yongsan-gu, Seoul, 04310, Korea02-710-9127
Copyright©Sookmyung Women's University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.