{"created":"2021-03-01T06:34:09.585015+00:00","id":26250,"links":{},"metadata":{"_buckets":{"deposit":"f5edd8ed-962b-4f31-bcc5-1401ea5cdf4a"},"_deposit":{"id":"26250","owners":[],"pid":{"revision_id":0,"type":"depid","value":"26250"},"status":"published"},"_oai":{"id":"oai:nagoya.repo.nii.ac.jp:00026250","sets":["1939:1940:1941"]},"author_link":["78790","78791","78792","78793","78794"],"item_10_biblio_info_6":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicIssueDates":{"bibliographicIssueDate":"2018-01-01","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"1","bibliographicPageEnd":"210","bibliographicPageStart":"199","bibliographicVolumeNumber":"E101A","bibliographic_titles":[{"bibliographic_title":"IEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences","bibliographic_titleLang":"en"}]}]},"item_10_description_4":{"attribute_name":"抄録","attribute_value_mlt":[{"subitem_description":"In this study, toward the development of smartphone-based monitoring system for life logging, we collect over 1,400 hours of data by recording including both the outdoor and indoor daily activities of 19 subjects, under practical conditions with a smartphone and a small camera. We then construct a huge human activity database which consists of an environmental sound signal, triaxial acceleration signals and manually annotated activity tags. Using our constructed database, we evaluate the activity recognition performance of deep neural networks (DNNs), which have achieved great performance in various fields, and apply DNN-based adaptation techniques to improve the performance with only a small amount of subject-specific training data. We experimentally demonstrate that; 1) the use of multi-modal signal, including environmental sound and triaxial acceleration signals with a DNN is effective for the improvement of activity recognition performance, 2) the DNN can discriminate specified activities from a mixture of ambiguous activities, and 3) DNN-based adaptation methods are effective even if only a small amount of subject-specific training data is available.","subitem_description_language":"en","subitem_description_type":"Abstract"}]},"item_10_publisher_32":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"一般社団法人電子情報通信学会","subitem_publisher_language":"ja"}]},"item_10_relation_11":{"attribute_name":"DOI","attribute_value_mlt":[{"subitem_relation_type":"isVersionOf","subitem_relation_type_id":{"subitem_relation_type_id_text":"https://doi.org/10.1587/transfun.E101.A.199","subitem_relation_type_select":"DOI"}}]},"item_10_relation_43":{"attribute_name":"関連情報","attribute_value_mlt":[{"subitem_relation_type":"isVersionOf","subitem_relation_type_id":{"subitem_relation_type_id_text":"http://search.ieice.org/index.html","subitem_relation_type_select":"URI"}}]},"item_10_rights_12":{"attribute_name":"権利","attribute_value_mlt":[{"subitem_rights":"copyright(c)2018 IEICE ","subitem_rights_language":"en"}]},"item_10_select_15":{"attribute_name":"著者版フラグ","attribute_value_mlt":[{"subitem_select_item":"publisher"}]},"item_10_source_id_7":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"0916-8508","subitem_source_identifier_type":"PISSN"}]},"item_1615787544753":{"attribute_name":"出版タイプ","attribute_value_mlt":[{"subitem_version_resource":"http://purl.org/coar/version/c_970fb48d4fbd8a85","subitem_version_type":"VoR"}]},"item_access_right":{"attribute_name":"アクセス権","attribute_value_mlt":[{"subitem_access_right":"open access","subitem_access_right_uri":"http://purl.org/coar/access_right/c_abf2"}]},"item_creator":{"attribute_name":"著者","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"HAYASHI, Tomoki","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"78790","nameIdentifierScheme":"WEKO"}]},{"creatorNames":[{"creatorName":"NISHIDA, Masafumi","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"78791","nameIdentifierScheme":"WEKO"}]},{"creatorNames":[{"creatorName":"KITAOKA, Norihide","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"78792","nameIdentifierScheme":"WEKO"}]},{"creatorNames":[{"creatorName":"TODA, Tomoki","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"78793","nameIdentifierScheme":"WEKO"}]},{"creatorNames":[{"creatorName":"TAKEDA, Kazuya","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"78794","nameIdentifierScheme":"WEKO"}]}]},"item_files":{"attribute_name":"ファイル情報","attribute_type":"file","attribute_value_mlt":[{"accessrole":"open_date","date":[{"dateType":"Available","dateValue":"2018-08-08"}],"displaytype":"detail","filename":"e101-a_1_199.pdf","filesize":[{"value":"4.2 MB"}],"format":"application/pdf","licensetype":"license_note","mimetype":"application/pdf","url":{"label":"e101-a_1_199","objectType":"fulltext","url":"https://nagoya.repo.nii.ac.jp/record/26250/files/e101-a_1_199.pdf"},"version_id":"520b1fa8-6d30-43cc-8e09-ce49b77bdbab"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"human activity recognition","subitem_subject_scheme":"Other"},{"subitem_subject":"activity of daily living","subitem_subject_scheme":"Other"},{"subitem_subject":"database","subitem_subject_scheme":"Other"},{"subitem_subject":"deep neural networks","subitem_subject_scheme":"Other"},{"subitem_subject":"adaptation","subitem_subject_scheme":"Other"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourcetype":"journal article","resourceuri":"http://purl.org/coar/resource_type/c_6501"}]},"item_title":"Daily Activity Recognition with Large-Scaled Real-Life Recording Datasets Based on Deep Neural Network Using Multi-Modal Signals","item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"Daily Activity Recognition with Large-Scaled Real-Life Recording Datasets Based on Deep Neural Network Using Multi-Modal Signals","subitem_title_language":"en"}]},"item_type_id":"10","owner":"1","path":["1941"],"pubdate":{"attribute_name":"PubDate","attribute_value":"2018-08-08"},"publish_date":"2018-08-08","publish_status":"0","recid":"26250","relation_version_is_last":true,"title":["Daily Activity Recognition with Large-Scaled Real-Life Recording Datasets Based on Deep Neural Network Using Multi-Modal Signals"],"weko_creator_id":"1","weko_shared_id":-1},"updated":"2023-01-16T04:29:11.475952+00:00"}