Lane detection is challenging under varied light conditions (e.g., night, shadow, and dazzling light) because a lane becomes blurred and extracting features becomes more difficult. Some researchers have proposed methods based on multitask learning and contextual information to solve this problem; however, these methods result in additional computing. A data enhancement method based on retinex theory is proposed. This method improves the adaptability of a lane model under varied light conditions. In particular, we design an image enhancement network for calculating the reflectivity of images, modifying their exposure, and then generating images with consistent exposure. These images are fed to the lane detection model for training and detection. Our network consists of two parts: exposure-consistent image generation and lane detection. We validate our method on the CULane dataset, and results show that it can improve lane detection performance, particularly on light-related datasets.
Person re-identification (re-ID) is the task of matching the same individuals across multiple cameras, and its performance is greatly influenced by background clutter. Most re-ID methods remove background clutter using hard manners, such as the use of segmentation algorithms. However, the hard manner may damage the structure information and smoothness of original images. In this work, we propose a unidirectional information-interaction network (UI2N) that consists of a global stream (G-Stream) and a background-graying stream (BGg-Stream). The G-Stream and BGg-Stream carry out unidirectional information interaction such that their features are complementary. We further propose a soft manner with the UI2N to weaken background clutter by background-graying. The soft manner can help the UI2N filter out background interference and retain some informative background cues. Extensive evaluations demonstrate that our method significantly outperforms many state-of-the-art approaches in the challenging Market-1501, DukeMTMC-reID, and CUHK03-NP datasets.
As an instance-level recognition problem, the key to effective vehicle re-identification (Re-ID) is to carefully reason the discriminative and viewpoint-invariant features of vehicle parts at high-level and low-level semantics. However, learning part-based features requires a laborious human annotation of some factors as attributes. To address this issue, we propose a region-aware multi-resolution (RAMR) Re-ID framework that can extract features from a series of local regions without extra manual annotations. Technically, the proposed method improves the discriminative ability of the local features through parallel high-to-low resolution convolutions. We also introduce a position attention module to focus on the prominent regions that can provide effective information. Given that the vehicle Re-ID performance can be affected by background clutters, we use the image obtained through foreground segmentation to extract local features. Results show that using original and foreground images can enhance the Re-ID performance compared with using either the original or foreground images alone. In other words, the original and foreground images complement each other in the vehicle Re-ID process. Finally, we aggregate the global appearance and local features to improve the system performance. Extensive experiments on two publicly available vehicle Re-ID datasets, namely, VeRi-776 and VehicleID, are conducted to validate the effectiveness of each proposed strategy. The findings indicate that the RAMR model achieves significant improvement in comparison with other state-of-the-art methods.
The requirements of distance measurement have increased with the development of auto vehicle driving. Traditional methods for distance estimation require the complex calibration from intrinsic and external parameters of the camera. Recent methods based on the neural network structure mainly measure the relative depth of whole images. We adopt monocular vision with instance segmentation and camera focal length to detect the absolute distance of front cars from in-car cameras. First, we extract the location of the cars from the object detection network. Second, the location of cars is sent to the vehicle classification network and instance segmentation network to obtain the type of the cars and their mask value. Here, we use a model trained by the CompCars dataset to classify car types, and we train a new instance segmentation model using the Cityscapes dataset to obtain each car’s mask. Third, in accordance with the camera imaging principle, the absolute distance of cars in the images is calculated based on the relationship between the size information of different car types and their mask values. The proposed method is examined with the KITTI dataset, and the experiment shows that its results can be close to the ground truth. Moreover, the proposed method uses the instance segmentation network to reduce complexity of the depth estimation process and it can still generate satisfactory results even when cars are partly occluded.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.