We found a match
Your institution may have access to this item. Find your institution then sign in to continue.
- Title
Automatic Salient Object Extraction Based on Locally Adaptive Thresholding to Generate Tactile Graphics.
- Authors
Abdusalomov, Akmalbek; Mukhiddinov, Mukhriddin; Djuraev, Oybek; Khamdamov, Utkir; Whangbo, Taeg Keun
- Abstract
Automatic extraction of salient regions is beneficial for various computer vision applications, such as image segmentation and object recognition. The salient visual information across images is very useful and plays a significant role for the visually impaired in identifying tactile information. In this paper, we introduce a novel saliency cuts method using local adaptive thresholding to obtain four regions from a given saliency map. First, we produced four regions for image segmentation using a saliency map as an input image and local adaptive thresholding. Second, the four regions were used to initialize an iterative version of the GrabCuts algorithm and to produce a robust and high-quality binary mask with a full resolution. Finally, salient objects' outer boundaries and inner edges were detected using the solution from our previous research. Experimental results showed that local adaptive thresholding using integral images can produce a more robust binary mask compared to the results from previous works that make use of global thresholding techniques for salient object segmentation. The proposed method can extract salient objects with a low-quality saliency map, achieving a promising performance compared to existing methods. The proposed method has advantages in extracting salient objects and generating simple, important edges from natural scene images efficiently for delivering visually salient information to the visually impaired.
- Subjects
THRESHOLDING algorithms; COMPUTER vision; OBJECT recognition (Computer vision); IMAGE segmentation; PEOPLE with visual disabilities; APPLICATION software
- Publication
Applied Sciences (2076-3417), 2020, Vol 10, Issue 10, p3350
- ISSN
2076-3417
- Publication type
Article
- DOI
10.3390/app10103350