Browse/search for people

Publication - Professor David Bull

    Visual salience and priority estimation for locomotion using a deep convolutional neural network

    Citation

    Anantrasirichai, P, Gilchrist, I & Bull, D, 2017, ‘Visual salience and priority estimation for locomotion using a deep convolutional neural network’. in: 2016 IEEE International Conference on Image Processing (ICIP 2016): Proceedings of a meeting held 25-28 September 2016, Phoenix, Arizona, USA. Institute of Electrical and Electronics Engineers (IEEE), pp. 1599-1603

    Abstract

    This paper presents a novel method of salience and priority estimation for the human visual system during locomotion. This visual information contains dynamic content derived from a moving viewpoint. The priority map, ranking key areas on the image, is created from probabilities of gaze fixations, merged from bottom-up features and top-down control on the locomotion. Two deep convolutional neural networks (CNNs), inspired by models of the primate visual system, are employed to capture local salience features and compute probabilities. The first network operates through the foveal and peripheral areas around the eye positions. The second network obtains the importance of fixated points that have long durations or multiple visits, of which such areas need more times to process or to recheck to ensure smooth locomotion. The results show that our proposed method outperforms the state-of-the-art by up to 30 %, computed from average of four well known metrics for saliency estimation.

    Full details in the University publications repository