TY - GEN
T1 - Multiple RGB-D camera-based user intent position and object estimation
AU - Hoonkwon, Ki
AU - Oh, Hyun Min
AU - Kim, Min Young
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2018/8/30
Y1 - 2018/8/30
N2 - Human gaze represents the area of interests of the person. By analyzing a time-series of these areas, it is possible to obtain user behavioral pattern that can be used in various fields. Well-known techniques for estimating human gaze are inconvenient because they require a wearable device, or the measurement area is relatively narrow. In this paper, a method to implement gaze estimation system using 3D view tracking with multi RGB-D camera is proposed. Surround 3D cameras are used to extract the region of interest of the user from 3D gaze estimation without wearable device in living space. To implement proposed method, first, 3D space mapping through multiple RGB-D camera calibration is performed. The resulting 3D map is the measurement area, which depends on the number and specifications of the RGB-D cameras used for this purpose. Then, when a person enters the 3D map, the face region is detected using both 2D/3D data, and 3D view tracking is implemented by detecting the gaze vector using the facial feature point and the head data center point extracted from the 3D map. Finally, when the gaze vector line intersects a specific point within the mapping space, the image coordinates corresponding to that point are extracted to implement user Intent position estimation. Applying object detection and classification algorithm to the extracted image can also estimate the intent object at that time.
AB - Human gaze represents the area of interests of the person. By analyzing a time-series of these areas, it is possible to obtain user behavioral pattern that can be used in various fields. Well-known techniques for estimating human gaze are inconvenient because they require a wearable device, or the measurement area is relatively narrow. In this paper, a method to implement gaze estimation system using 3D view tracking with multi RGB-D camera is proposed. Surround 3D cameras are used to extract the region of interest of the user from 3D gaze estimation without wearable device in living space. To implement proposed method, first, 3D space mapping through multiple RGB-D camera calibration is performed. The resulting 3D map is the measurement area, which depends on the number and specifications of the RGB-D cameras used for this purpose. Then, when a person enters the 3D map, the face region is detected using both 2D/3D data, and 3D view tracking is implemented by detecting the gaze vector using the facial feature point and the head data center point extracted from the 3D map. Finally, when the gaze vector line intersects a specific point within the mapping space, the image coordinates corresponding to that point are extracted to implement user Intent position estimation. Applying object detection and classification algorithm to the extracted image can also estimate the intent object at that time.
UR - http://www.scopus.com/inward/record.url?scp=85053906824&partnerID=8YFLogxK
U2 - 10.1109/AIM.2018.8452320
DO - 10.1109/AIM.2018.8452320
M3 - Conference contribution
AN - SCOPUS:85053906824
SN - 9781538618547
T3 - IEEE/ASME International Conference on Advanced Intelligent Mechatronics, AIM
SP - 176
EP - 180
BT - AIM 2018 - IEEE/ASME International Conference on Advanced Intelligent Mechatronics
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2018 IEEE/ASME International Conference on Advanced Intelligent Mechatronics, AIM 2018
Y2 - 9 July 2018 through 12 July 2018
ER -