Chapter
Nov 9, 2020
Construction Research Congress 2020

Estimating the Visual Attention of Construction Workers from Head Pose Using Convolutional Neural Network-Based Multi-Task Learning

Publication: Construction Research Congress 2020: Computer Applications

ABSTRACT

The visual attention of construction workers is an important indicator to assess their situational awareness and infer their intention for reducing construction injuries and improving construction site safety. The eye-tracking technology has been adopted in several studies to directly measure the gaze direction and determine workers’ visual attention. However, eye-trackers are expensive and wearing them may disturb normal operations. Considering the increasing use of surveillance videos and the availability of construction images, it is of great potential to estimate workers’ visual attention from imagery data, which, however, has not been well exploited by existing studies. This paper presents a convolutional neural network (CNN)-based multi-task learning framework to estimate the visual attention of construction workers from head pose using low-resolution images. Visual attention is approximated by head yaw and pitch orientation. The problem is formulated as a multi-task image classification problem, where the first task is head yaw classification, and the second task is head pitch classification. A CNN-based multi-task learning framework is designed to jointly learn two tasks, with shared layers capturing the commonality between tasks, and task-specific layers modeling the uniqueness of individual tasks. Compared to traditional single-task learning mechanism that trains different classifiers for each task, the proposed approach leverages the commonality of relevant tasks and captures the shared representation, which can significantly improve the efficiency and performance. The results suggest the proposed multi-learning framework can achieve an accuracy of 76.5% for head yaw estimation and 88.7% for head pitch estimation, better than the performance obtained using conventional single task learning.

Get full access to this article

View all available purchase options and get full access to this chapter.

REFERENCES

Asteriadis, S., Karpouzis, K., and Kollias, S. (2011). “Robust validation of Visual Focus of Attention using adaptive fusion of head and eye gaze patterns.” Proceedings of the IEEE International Conference on Computer Vision, 414–421.
Cai, J., Zhang, Y., and Cai, H. (2019a). “Two-step long short-term memory method for identifying construction activities through positional and attentional cues.” Automation in Construction, Elsevier, 106, 102886.
Cai, J., Zhang, Y., and Cai, H. (2019b). “Integrating Positional and Attentional Cues for Construction Working Group Identification: A Long Short-Term Memory Based Machine Learning Approach.” Computing in Civil Engineering 2019: Data, Sensing, and Analytics, American Society of Civil Engineers Reston, VA, 35–42.
Chamveha, I., Sugano, Y., Sato, Y., and Sugimoto, A. (2014). “Social Group Discovery from Surveillance Videos: A Data-Driven Approach with Attention-Based Cues.” 121.1-121.11.
Hasanzadeh, S., Esmaeili, B., and Dodd, M. D. (2017). “Impact of Construction Workers’ Hazard Identification Skills on Their Visual Attention.” Journal of Construction Engineering and Management, 143(10), 04017070.
Hasanzadeh, S., Esmaeili, B., and Dodd, M. D. (2018). “Examining the Relationship between Construction Workers’ Visual Attention and Situation Awareness under Fall and Tripping Hazard Conditions: Using Mobile Eye Tracking.” Journal of Construction Engineering and Management, 144(7), 04018060.
Jeelani, I., Albert, A., Han, K., and Azevedo, R. (2018). “Are Visual Search Patterns Predictive of Hazard Recognition Performance? Empirical Investigation Using Eye-Tracking Technology.” Journal of Construction Engineering and Management, 145(1), 04018115.
Krüger, N., Pötzsch, M., and von der Malsburg, C. (1997). “Determination of face position and pose with a learned representation based on labelled graphs.” Image and Vision Computing, 15(8), 665–673.
Murphy-Chutorian, E., and Trivedi, M. M. (2009). “Head pose estimation in computer vision: A survey.” IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(4), 607–626.
OSHA. (2015). “Commonly Used Statistics.” <https://www.osha.gov/oshstats/commonstats.html> (Sep. 26, 2018).
Ozturk, O., Yamasaki, T., and Aizawa, K. (2011). “Estimating human body and head orientation change to detect visual attention direction.” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 410–419.
Patacchiola, M., and Cangelosi, A. (2017). “Head pose estimation in the wild using Convolutional Neural Networks and adaptive gradient methods.” Pattern Recognition, 71, 132–143.
Raza, M., Chen, Z., Rehman, S. U., Wang, P., and Bao, P. (2018). “Appearance based pedestrians’ head pose and body orientation estimation using deep learning.” Neurocomputing, 272, 647–659.
Rehder, E., Kloeden, H., and Stiller, C. (2014). “Head detection and orientation estimation for pedestrian safety.” 2014 17th IEEE International Conference on Intelligent Transportation Systems, ITSC 2014, 2292–2297.
Saleh, K., Hossny, M., and Nahavandi, S. (2017). “Early intent prediction of vulnerable road users from visual attributes using multi-task learning network.” 2017 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2017.
Yan, Y., Ricci, E., Subramanian, R., Liu, G., Lanz, O., and Sebe, N. (2016). “A Multi-Task Learning Framework for Head Pose Estimation under Target Motion.” IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(6), 1070–1083.
YouTube. (2019). “Hospital construction.” <https://www.youtube.com/channel/UCEKwrM78pRv8WRcKvZNtE1w> (Apr. 7, 2019).
Zhang, Y., and Yang, Q. (2017). “A survey on multi-task learning.” arXiv preprint arXiv:1707.08114.

Information & Authors

Information

Published In

Go to Construction Research Congress 2020
Construction Research Congress 2020: Computer Applications
Pages: 116 - 124
Editors: Pingbo Tang, Ph.D., Arizona State University, David Grau, Ph.D., Arizona State University, and Mounir El Asmar, Ph.D., Arizona State University
ISBN (Online): 978-0-7844-8286-5

History

Published online: Nov 9, 2020
Published in print: Nov 9, 2020

Permissions

Request permissions for this article.

Authors

Affiliations

Jiannan Cai [email protected]
Lyles School of Civil Engineering, Purdue Univ., West Lafayette, IN. E-mail: [email protected]
Lyles School of Civil Engineering, Purdue Univ., West Lafayette, IN. E-mail: [email protected]
Lyles School of Civil Engineering, Purdue Univ., West Lafayette, IN. E-mail: [email protected]
Lyles School of Civil Engineering, Purdue Univ., West Lafayette, IN. E-mail: [email protected]

Metrics & Citations

Metrics

Citations

Download citation

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

View Options

Get Access

Access content

Please select your options to get access

Log in/Register Log in via your institution (Shibboleth)
ASCE Members: Please log in to see member pricing

Purchase

Save for later Information on ASCE Library Cards
ASCE Library Cards let you download journal articles, proceedings papers, and available book chapters across the entire ASCE Library platform. ASCE Library Cards remain active for 24 months or until all downloads are used. Note: This content will be debited as one download at time of checkout.

Terms of Use: ASCE Library Cards are for individual, personal use only. Reselling, republishing, or forwarding the materials to libraries or reading rooms is prohibited.
ASCE Library Card (5 downloads)
$105.00
Add to cart
ASCE Library Card (20 downloads)
$280.00
Add to cart
Buy Single Paper
$35.00
Add to cart
Buy E-book
$286.00
Add to cart

Get Access

Access content

Please select your options to get access

Log in/Register Log in via your institution (Shibboleth)
ASCE Members: Please log in to see member pricing

Purchase

Save for later Information on ASCE Library Cards
ASCE Library Cards let you download journal articles, proceedings papers, and available book chapters across the entire ASCE Library platform. ASCE Library Cards remain active for 24 months or until all downloads are used. Note: This content will be debited as one download at time of checkout.

Terms of Use: ASCE Library Cards are for individual, personal use only. Reselling, republishing, or forwarding the materials to libraries or reading rooms is prohibited.
ASCE Library Card (5 downloads)
$105.00
Add to cart
ASCE Library Card (20 downloads)
$280.00
Add to cart
Buy Single Paper
$35.00
Add to cart
Buy E-book
$286.00
Add to cart

Media

Figures

Other

Tables

Share

Share

Copy the content Link

Share with email

Email a colleague

Share