{"created":"2023-08-02T03:58:10.030722+00:00","id":7321,"links":{},"metadata":{"_buckets":{"deposit":"fe8b902b-fac6-404c-8707-53144a766fd7"},"_deposit":{"created_by":10,"id":"7321","owners":[10],"pid":{"revision_id":0,"type":"depid","value":"7321"},"status":"published"},"_oai":{"id":"oai:repository.lib.tottori-u.ac.jp:00007321","sets":["1:10","2:12"]},"author_link":["4786","4538","4671","27213"],"item_10001_biblio_info_7":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicIssueDates":{"bibliographicIssueDate":"2018-09-01","bibliographicIssueDateType":"Issued"},"bibliographicPageEnd":"248","bibliographicPageStart":"241","bibliographicVolumeNumber":"112","bibliographic_titles":[{"bibliographic_title":"PATTERN RECOGNITION LETTERS"},{"bibliographic_title":"PATTERN RECOGNITION LETTERS","bibliographic_titleLang":"en"}]}]},"item_10001_description_5":{"attribute_name":"抄録","attribute_value_mlt":[{"subitem_description":"We discuss how to reveal and use the gaze locations of observers who view pedestrian images for personal attribute classification. Observers look at informative regions when attempting to classify the attributes of pedestrians in images. Thus, we hypothesize that the regions in which observers’ gaze locations are clustered will contain discriminative features for the classifiers of personal attributes. Our method acquires the distribution of gaze locations from several observers while they perform the task of manually classifying each personal attribute. We term this distribution a task-oriented gaze map. To extract discriminative features, we assign large weights to the region with a cluster of gaze locations in the task-oriented gaze map. In our experiments, observers mainly looked at different regions of body parts when classifying each personal attribute. Furthermore, our experiments show that the gaze-based feature extraction method significantly improved the performance of personal attribute classification when combined with a convolutional neural network or metric learning technique.","subitem_description_type":"Other"}]},"item_10001_publisher_8":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"Elsevier"}]},"item_10001_relation_14":{"attribute_name":"DOI","attribute_value_mlt":[{"subitem_relation_type":"isVersionOf","subitem_relation_type_id":{"subitem_relation_type_id_text":"10.1016/j.patrec.2018.08.001","subitem_relation_type_select":"DOI"}}]},"item_10001_relation_16":{"attribute_name":"情報源","attribute_value_mlt":[{"subitem_relation_name":[{"subitem_relation_name_text":"Nishiyama Masashi, Matsumoto Riku, Yoshimura Hiroki, et al. Extracting discriminative features using task-oriented gaze maps measured from observers for personal attribute classification. PATTERN RECOGNITION LETTERS. 2018. 112. 241-248. doi:10.1016/j.patr"}]}]},"item_10001_relation_17":{"attribute_name":"関連サイト","attribute_value_mlt":[{"subitem_relation_name":[{"subitem_relation_name_text":"https://www.sciencedirect.com/science/article/pii/S0167865518303696"}],"subitem_relation_type_id":{"subitem_relation_type_id_text":"https://www.sciencedirect.com/science/article/pii/S0167865518303696","subitem_relation_type_select":"URI"}}]},"item_10001_rights_15":{"attribute_name":"権利","attribute_value_mlt":[{"subitem_rights":"(C) 2018. This manuscript version is made available under the CC-BY-NC-ND 4.0 license https://creativecommons.org/licenses/by-nc-nd/4.0/"}]},"item_10001_source_id_9":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"01678655","subitem_source_identifier_type":"ISSN"}]},"item_10001_text_26":{"attribute_name":"EISSN","attribute_value_mlt":[{"subitem_text_value":"18727344"}]},"item_10001_text_33":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_language":"en","subitem_text_value":"Graduate School of Engineering, Tottori University / Cross-informatics Research Center, Tottori University"},{"subitem_text_language":"en","subitem_text_value":"Graduate School of Engineering, Tottori University"},{"subitem_text_language":"en","subitem_text_value":"Graduate School of Engineering, Tottori University"},{"subitem_text_language":"en","subitem_text_value":"Graduate School of Engineering, Tottori University / Cross-informatics Research Center, Tottori University"}]},"item_10001_version_type_20":{"attribute_name":"著者版フラグ","attribute_value_mlt":[{"subitem_version_type":"AM"}]},"item_creator":{"attribute_name":"著者","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"西山, 正志"},{"creatorName":"ニシヤマ, マサシ","creatorNameLang":"ja-Kana"},{"creatorName":"Nishiyama, Masashi","creatorNameLang":"en"}],"nameIdentifiers":[{},{},{}]},{"creatorNames":[{"creatorName":"吉村, 宏紀"},{"creatorName":"ヨシムラ, ヒロキ","creatorNameLang":"ja-Kana"},{"creatorName":"Yoshimura, Hiroki","creatorNameLang":"en"}],"nameIdentifiers":[{},{},{}]},{"creatorNames":[{"creatorName":"岩井, 儀雄"},{"creatorName":"イワイ, ヨシオ","creatorNameLang":"ja-Kana"},{"creatorName":"Iwai, Yoshio","creatorNameLang":"en"}],"nameIdentifiers":[{},{},{}]},{"creatorNames":[{"creatorName":"Matsumoto, Riku","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_files":{"attribute_name":"ファイル情報","attribute_type":"file","attribute_value_mlt":[{"accessrole":"open_date","date":[{"dateType":"Available","dateValue":"2023-03-17"}],"displaytype":"detail","filename":"prl112_241.pdf","filesize":[{"value":"1.6 MB"}],"format":"application/pdf","licensefree":"(C) 2018. This manuscript version is made available under the CC-BY-NC-ND 4.0 license https://creativecommons.org/licenses/by-nc-nd/4.0/","licensetype":"license_note","mimetype":"application/pdf","url":{"label":"prl112_241.pdf","url":"https://repository.lib.tottori-u.ac.jp/record/7321/files/prl112_241.pdf"},"version_id":"1dd187de-76b6-42f7-afa8-75dd8f43a70a"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"Gaze map","subitem_subject_scheme":"Other"},{"subitem_subject":"Feature extraction","subitem_subject_scheme":"Other"},{"subitem_subject":"Personal attributes","subitem_subject_scheme":"Other"},{"subitem_subject":"Gaze map","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Feature extraction","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Personal attributes","subitem_subject_language":"en","subitem_subject_scheme":"Other"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourcetype":"journal article"}]},"item_title":"Extracting discriminative features using task-oriented gaze maps measured from observers for personal attribute classification","item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"Extracting discriminative features using task-oriented gaze maps measured from observers for personal attribute classification","subitem_title_language":"en"}]},"item_type_id":"10001","owner":"10","path":["12","10"],"pubdate":{"attribute_name":"PubDate","attribute_value":"2022-08-03"},"publish_date":"2022-08-03","publish_status":"0","recid":"7321","relation_version_is_last":true,"title":["Extracting discriminative features using task-oriented gaze maps measured from observers for personal attribute classification"],"weko_creator_id":"10","weko_shared_id":-1},"updated":"2023-11-22T01:51:07.579927+00:00"}