Under the influence of air humidity,dust,aerosols,etc.,in real scenes,haze presents an uneven state.In this way,the image quality and contrast will decrease.In this case,It is difficult to detect the target in the ima...Under the influence of air humidity,dust,aerosols,etc.,in real scenes,haze presents an uneven state.In this way,the image quality and contrast will decrease.In this case,It is difficult to detect the target in the image by the universal detection network.Thus,a dual subnet based on multi-task collaborative training(DSMCT)is proposed in this paper.Firstly,in the training phase,the Gated Context Aggregation Network(GCANet)is used as the supervisory network of YOLOX to promote the extraction of clean information in foggy scenes.In the test phase,only the YOLOX branch needs to be activated to ensure the detection speed of the model.Secondly,the deformable convolution module is used to improve GCANet to enhance the model’s ability to capture details of non-homogeneous fog.Finally,the Coordinate Attention mechanism is introduced into the Vision Transformer and the backbone network of YOLOX is redesigned.In this way,the feature extraction ability of the network for deep-level information can be enhanced.The experimental results on artificial fog data set FOG_VOC and real fog data set RTTS show that the map value of DSMCT reached 86.56%and 62.39%,respectively,which was 2.27%and 4.41%higher than the current most advanced detection model.The DSMCT network has high practicality and effectiveness for target detection in real foggy scenes.展开更多
Video salient object detection(VSOD)aims at locating the most attractive objects in a video by exploring the spatial and temporal features.VSOD poses a challenging task in computer vision,as it involves processing com...Video salient object detection(VSOD)aims at locating the most attractive objects in a video by exploring the spatial and temporal features.VSOD poses a challenging task in computer vision,as it involves processing complex spatial data that is also influenced by temporal dynamics.Despite the progress made in existing VSOD models,they still struggle in scenes of great background diversity within and between frames.Additionally,they encounter difficulties related to accumulated noise and high time consumption during the extraction of temporal features over a long-term duration.We propose a multi-stream temporal enhanced network(MSTENet)to address these problems.It investigates saliency cues collaboration in the spatial domain with a multi-stream structure to deal with the great background diversity challenge.A straightforward,yet efficient approach for temporal feature extraction is developed to avoid the accumulative noises and reduce time consumption.The distinction between MSTENet and other VSOD methods stems from its incorporation of both foreground supervision and background supervision,facilitating enhanced extraction of collaborative saliency cues.Another notable differentiation is the innovative integration of spatial and temporal features,wherein the temporal module is integrated into the multi-stream structure,enabling comprehensive spatial-temporal interactions within an end-to-end framework.Extensive experimental results demonstrate that the proposed method achieves state-of-the-art performance on five benchmark datasets while maintaining a real-time speed of 27 fps(Titan XP).Our code and models are available at https://github.com/RuJiaLe/MSTENet.展开更多
基金This work was jointly supported by the Special Fund for Transformation and Upgrade of Jiangsu Industry and Information Industry-Key Core Technologies(Equipment)Key Industrialization Projects in 2022(No.CMHI-2022-RDG-004):“Key Technology Research for Development of Intelligent Wind Power Operation and Maintenance Mothership in Deep Sea”.
文摘Under the influence of air humidity,dust,aerosols,etc.,in real scenes,haze presents an uneven state.In this way,the image quality and contrast will decrease.In this case,It is difficult to detect the target in the image by the universal detection network.Thus,a dual subnet based on multi-task collaborative training(DSMCT)is proposed in this paper.Firstly,in the training phase,the Gated Context Aggregation Network(GCANet)is used as the supervisory network of YOLOX to promote the extraction of clean information in foggy scenes.In the test phase,only the YOLOX branch needs to be activated to ensure the detection speed of the model.Secondly,the deformable convolution module is used to improve GCANet to enhance the model’s ability to capture details of non-homogeneous fog.Finally,the Coordinate Attention mechanism is introduced into the Vision Transformer and the backbone network of YOLOX is redesigned.In this way,the feature extraction ability of the network for deep-level information can be enhanced.The experimental results on artificial fog data set FOG_VOC and real fog data set RTTS show that the map value of DSMCT reached 86.56%and 62.39%,respectively,which was 2.27%and 4.41%higher than the current most advanced detection model.The DSMCT network has high practicality and effectiveness for target detection in real foggy scenes.
基金funded by the Natural Science Foundation China(NSFC)under Grant No.62203192.
文摘Video salient object detection(VSOD)aims at locating the most attractive objects in a video by exploring the spatial and temporal features.VSOD poses a challenging task in computer vision,as it involves processing complex spatial data that is also influenced by temporal dynamics.Despite the progress made in existing VSOD models,they still struggle in scenes of great background diversity within and between frames.Additionally,they encounter difficulties related to accumulated noise and high time consumption during the extraction of temporal features over a long-term duration.We propose a multi-stream temporal enhanced network(MSTENet)to address these problems.It investigates saliency cues collaboration in the spatial domain with a multi-stream structure to deal with the great background diversity challenge.A straightforward,yet efficient approach for temporal feature extraction is developed to avoid the accumulative noises and reduce time consumption.The distinction between MSTENet and other VSOD methods stems from its incorporation of both foreground supervision and background supervision,facilitating enhanced extraction of collaborative saliency cues.Another notable differentiation is the innovative integration of spatial and temporal features,wherein the temporal module is integrated into the multi-stream structure,enabling comprehensive spatial-temporal interactions within an end-to-end framework.Extensive experimental results demonstrate that the proposed method achieves state-of-the-art performance on five benchmark datasets while maintaining a real-time speed of 27 fps(Titan XP).Our code and models are available at https://github.com/RuJiaLe/MSTENet.