期刊文献+
共找到330,772篇文章
< 1 2 250 >
每页显示 20 50 100
An Improved High Precision 3D Semantic Mapping of Indoor Scenes from RGB-D Images
1
作者 Jing Xin Kenan Du +1 位作者 Jiale Feng Mao Shan 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第12期2621-2640,共20页
This paper proposes an improved high-precision 3D semantic mapping method for indoor scenes using RGB-D images.The current semantic mapping algorithms suffer from low semantic annotation accuracy and insufficient real... This paper proposes an improved high-precision 3D semantic mapping method for indoor scenes using RGB-D images.The current semantic mapping algorithms suffer from low semantic annotation accuracy and insufficient real-time performance.To address these issues,we first adopt the Elastic Fusion algorithm to select key frames from indoor environment image sequences captured by the Kinect sensor and construct the indoor environment space model.Then,an indoor RGB-D image semantic segmentation network is proposed,which uses multi-scale feature fusion to quickly and accurately obtain object labeling information at the pixel level of the spatial point cloud model.Finally,Bayesian updating is used to conduct incremental semantic label fusion on the established spatial point cloud model.We also employ dense conditional random fields(CRF)to optimize the 3D semantic map model,resulting in a high-precision spatial semantic map of indoor scenes.Experimental results show that the proposed semantic mapping system can process image sequences collected by RGB-D sensors in real-time and output accurate semantic segmentation results of indoor scene images and the current local spatial semantic map.Finally,it constructs a globally consistent high-precision indoor scenes 3D semantic map. 展开更多
关键词 3D semantic map online reconstruction rgb-d images semantic segmentation indoor mobile robot
下载PDF
Bayesian Saliency Detection for RGB-D Images 被引量:1
2
作者 Songtao Wang Zhen Zhou +1 位作者 Hanbing Qu Bin Li 《自动化学报》 EI CSCD 北大核心 2017年第10期1810-1828,共19页
关键词 贝叶斯定理 检测模型 显著性 图像 期望最大化算法 分布计算 特征映射 高斯分布
下载PDF
Detection of the farmland plow areas using RGB-D images with an improved YOLOv5 model
3
作者 Jiangtao Ji Zhihao Han +2 位作者 Kaixuan Zhao Qianwen Li Shucan Du 《International Journal of Agricultural and Biological Engineering》 SCIE 2024年第3期156-165,共10页
Recognition of the boundaries of farmland plow areas has an important guiding role in the operation of intelligent agricultural equipment.To precisely recognize these boundaries,a detection method for unmanned tractor... Recognition of the boundaries of farmland plow areas has an important guiding role in the operation of intelligent agricultural equipment.To precisely recognize these boundaries,a detection method for unmanned tractor plow areas based on RGB-Depth(RGB-D)cameras was proposed,and the feasibility of the detection method was analyzed.This method applied advanced computer vision technology to the field of agricultural automation.Adopting and improving the YOLOv5-seg object segmentation algorithm,first,the Convolutional Block Attention Module(CBAM)was integrated into Concentrated-Comprehensive Convolution Block(C3)to form C3CBAM,thereby enhancing the ability of the network to extract features from plow areas.The GhostConv module was also utilized to reduce parameter and computational complexity.Second,using the depth image information provided by the RGB-D camera combined with the results recognized by the YOLOv5-seg model,the mask image was processed to extract contour boundaries,align the contours with the depth map,and obtain the boundary distance information of the plowed area.Last,based on farmland information,the calculated average boundary distance was corrected,further improving the accuracy of the distance measurements.The experiment results showed that the YOLOv5-seg object segmentation algorithm achieved a recognition accuracy of 99%for plowed areas and that the ranging accuracy improved with decreasing detection distance.The ranging error at 5.5 m was approximately 0.056 m,and the average detection time per frame is 29 ms,which can meet the real-time operational requirements.The results of this study can provide precise guarantees for the autonomous operation of unmanned plowing units. 展开更多
关键词 plow areas rgb-d camera YOLO object segmentation contour boundary average distance
原文传递
Dendritic Learning-Incorporated Vision Transformer for Image Recognition 被引量:2
4
作者 Zhiming Zhang Zhenyu Lei +2 位作者 Masaaki Omura Hideyuki Hasegawa Shangce Gao 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第2期539-541,共3页
Dear Editor,This letter proposes to integrate dendritic learnable network architecture with Vision Transformer to improve the accuracy of image recognition.In this study,based on the theory of dendritic neurons in neu... Dear Editor,This letter proposes to integrate dendritic learnable network architecture with Vision Transformer to improve the accuracy of image recognition.In this study,based on the theory of dendritic neurons in neuroscience,we design a network that is more practical for engineering to classify visual features.Based on this,we propose a dendritic learning-incorporated vision Transformer(DVT),which out-performs other state-of-the-art methods on three image recognition benchmarks. 展开更多
关键词 image network image
下载PDF
Infrared and Visible Image Fusion Based on Res2Net-Transformer Automatic Encoding and Decoding 被引量:1
5
作者 Chunming Wu Wukai Liu Xin Ma 《Computers, Materials & Continua》 SCIE EI 2024年第4期1441-1461,共21页
A novel image fusion network framework with an autonomous encoder and decoder is suggested to increase thevisual impression of fused images by improving the quality of infrared and visible light picture fusion. The ne... A novel image fusion network framework with an autonomous encoder and decoder is suggested to increase thevisual impression of fused images by improving the quality of infrared and visible light picture fusion. The networkcomprises an encoder module, fusion layer, decoder module, and edge improvementmodule. The encoder moduleutilizes an enhanced Inception module for shallow feature extraction, then combines Res2Net and Transformerto achieve deep-level co-extraction of local and global features from the original picture. An edge enhancementmodule (EEM) is created to extract significant edge features. A modal maximum difference fusion strategy isintroduced to enhance the adaptive representation of information in various regions of the source image, therebyenhancing the contrast of the fused image. The encoder and the EEM module extract features, which are thencombined in the fusion layer to create a fused picture using the decoder. Three datasets were chosen to test thealgorithmproposed in this paper. The results of the experiments demonstrate that the network effectively preservesbackground and detail information in both infrared and visible images, yielding superior outcomes in subjectiveand objective evaluations. 展开更多
关键词 image fusion Res2Net-Transformer infrared image visible image
下载PDF
Using restored two-dimensional X-ray images to reconstruct the three-dimensional magnetopause 被引量:2
6
作者 RongCong Wang JiaQi Wang +3 位作者 DaLin Li TianRan Sun XiaoDong Peng YiHong Guo 《Earth and Planetary Physics》 EI CSCD 2024年第1期133-154,共22页
Astronomical imaging technologies are basic tools for the exploration of the universe,providing basic data for the research of astronomy and space physics.The Soft X-ray Imager(SXI)carried by the Solar wind Magnetosph... Astronomical imaging technologies are basic tools for the exploration of the universe,providing basic data for the research of astronomy and space physics.The Soft X-ray Imager(SXI)carried by the Solar wind Magnetosphere Ionosphere Link Explorer(SMILE)aims to capture two-dimensional(2-D)images of the Earth’s magnetosheath by using soft X-ray imaging.However,the observed 2-D images are affected by many noise factors,destroying the contained information,which is not conducive to the subsequent reconstruction of the three-dimensional(3-D)structure of the magnetopause.The analysis of SXI-simulated observation images shows that such damage cannot be evaluated with traditional restoration models.This makes it difficult to establish the mapping relationship between SXIsimulated observation images and target images by using mathematical models.We propose an image restoration algorithm for SXIsimulated observation images that can recover large-scale structure information on the magnetosphere.The idea is to train a patch estimator by selecting noise–clean patch pairs with the same distribution through the Classification–Expectation Maximization algorithm to achieve the restoration estimation of the SXI-simulated observation image,whose mapping relationship with the target image is established by the patch estimator.The Classification–Expectation Maximization algorithm is used to select multiple patch clusters with the same distribution and then train different patch estimators so as to improve the accuracy of the estimator.Experimental results showed that our image restoration algorithm is superior to other classical image restoration algorithms in the SXI-simulated observation image restoration task,according to the peak signal-to-noise ratio and structural similarity.The restoration results of SXI-simulated observation images are used in the tangent fitting approach and the computed tomography approach toward magnetospheric reconstruction techniques,significantly improving the reconstruction results.Hence,the proposed technology may be feasible for processing SXI-simulated observation images. 展开更多
关键词 Solar wind Magnetosphere Ionosphere Link Explorer(SMILE) soft X-ray imager MAGNETOPAUSE image restoration
下载PDF
The Soft X-ray Imager(SXI)on the SMILE Mission 被引量:4
7
作者 S.Sembay A.L.Alme +83 位作者 D.Agnolon T.Arnold A.Beardmore A.Belén Balado Margeli C.Bicknell C.Bouldin G.Branduardi-Raymont T.Crawford J.P.Breuer T.Buggey G.Butcher R.Canchal J.A.Carter A.Cheney Y.Collado-Vega H.Connor T.Crawford N.Eaton C.Feldman C.Forsyth T.Frantzen G.Galgóczi J.Garcia G.Y.Genov C.Gordillo H-P.Gröbelbauer M.Guedel Y.Guo M.Hailey D.Hall R.Hampson J.Hasiba O.Hetherington A.Holland S-Y.Hsieh M.W.J.Hubbard H.Jeszenszky M.Jones T.Kennedy K.Koch-Mehrin S.Kögl S.Krucker K.D.Kuntz C.Lakin G.Laky O.Lylund A.Martindale J.Miguel Mas Hesse R.Nakamura K.Oksavik N.Østgaard H.Ottacher R.Ottensamer C.Pagani S.Parsons P.Patel J.Pearson G.Peikert F.S.Porter T.Pouliantis B.H.Qureshi W.Raab G.Randal A.M.Read N.M.M.Roque M.E.Rostad C.Runciman S.Sachdev A.Samsonov M.Soman D.Sibeck S.Smit J.Søndergaard R.Speight S.Stavland M.Steller TianRan Sun J.Thornhill W.Thomas K.Ullaland B.Walsh D.Walton C.Wang S.Yang 《Earth and Planetary Physics》 EI CSCD 2024年第1期5-14,共10页
The Soft X-ray Imager(SXI)is part of the scientific payload of the Solar wind Magnetosphere Ionosphere Link Explorer(SMILE)mission.SMILE is a joint science mission between the European Space Agency(ESA)and the Chinese... The Soft X-ray Imager(SXI)is part of the scientific payload of the Solar wind Magnetosphere Ionosphere Link Explorer(SMILE)mission.SMILE is a joint science mission between the European Space Agency(ESA)and the Chinese Academy of Sciences(CAS)and is due for launch in 2025.SXI is a compact X-ray telescope with a wide field-of-view(FOV)capable of encompassing large portions of Earth’s magnetosphere from the vantage point of the SMILE orbit.SXI is sensitive to the soft X-rays produced by the Solar Wind Charge eXchange(SWCX)process produced when heavy ions of solar wind origin interact with neutral particles in Earth’s exosphere.SWCX provides a mechanism for boundary detection within the magnetosphere,such as the position of Earth’s magnetopause,because the solar wind heavy ions have a very low density in regions of closed magnetic field lines.The sensitivity of the SXI is such that it can potentially track movements of the magnetopause on timescales of a few minutes and the orbit of SMILE will enable such movements to be tracked for segments lasting many hours.SXI is led by the University of Leicester in the United Kingdom(UK)with collaborating organisations on hardware,software and science support within the UK,Europe,China and the United States. 展开更多
关键词 Soft X-ray imaging micropore optics large area CCD
下载PDF
Background removal from global auroral images:Data-driven dayglow modeling 被引量:1
8
作者 A.Ohma M.Madelaire +4 位作者 K.M.Laundal J.P.Reistad S.M.Hatch S.Gasparini S.J.Walker 《Earth and Planetary Physics》 EI CSCD 2024年第1期247-257,共11页
Global images of auroras obtained by cameras on spacecraft are a key tool for studying the near-Earth environment.However,the cameras are sensitive not only to auroral emissions produced by precipitating particles,but... Global images of auroras obtained by cameras on spacecraft are a key tool for studying the near-Earth environment.However,the cameras are sensitive not only to auroral emissions produced by precipitating particles,but also to dayglow emissions produced by photoelectrons induced by sunlight.Nightglow emissions and scattered sunlight can contribute to the background signal.To fully utilize such images in space science,background contamination must be removed to isolate the auroral signal.Here we outline a data-driven approach to modeling the background intensity in multiple images by formulating linear inverse problems based on B-splines and spherical harmonics.The approach is robust,flexible,and iteratively deselects outliers,such as auroral emissions.The final model is smooth across the terminator and accounts for slow temporal variations and large-scale asymmetries in the dayglow.We demonstrate the model by using the three far ultraviolet cameras on the Imager for Magnetopause-to-Aurora Global Exploration(IMAGE)mission.The method can be applied to historical missions and is relevant for upcoming missions,such as the Solar wind Magnetosphere Ionosphere Link Explorer(SMILE)mission. 展开更多
关键词 AURORA dayglow modeling global auroral images far ultraviolet images dayglow removal
下载PDF
Method for evaluation of geological strength index of carbonate cliff rocks:Coupled hyperspectral-digital borehole image technique 被引量:1
9
作者 Haiqing Yang Guizhong Huang +3 位作者 Chiwei Chen Yong Yang Qi Wang Xionghui Dai 《Journal of Rock Mechanics and Geotechnical Engineering》 SCIE CSCD 2024年第10期4204-4215,共12页
The deterioration of unstable rock mass raised interest in evaluating rock mass quality.However,the traditional evaluation method for the geological strength index(GSI)primarily emphasizes the rock structure and chara... The deterioration of unstable rock mass raised interest in evaluating rock mass quality.However,the traditional evaluation method for the geological strength index(GSI)primarily emphasizes the rock structure and characteristics of discontinuities.It ignores the influence of mineral composition and shows a deficiency in assessing the integrity coefficient.In this context,hyperspectral imaging and digital panoramic borehole camera technologies are applied to analyze the mineral content and integrity of rock mass.Based on the carbonate mineral content and fissure area ratio,the strength reduction factor and integrity coefficient are calculated to improve the GSI evaluation method.According to the results of mineral classification and fissure identification,the strength reduction factor and integrity coefficient increase with the depth of rock mass.The rock mass GSI calculated by the improved method is mainly concentrated between 40 and 60,which is close to the calculation results of the traditional method.The GSI error rates obtained by the two methods are mostly less than 10%,indicating the rationality of the hyperspectral-digital borehole image coupled evaluation method.Moreover,the sensitivity of the fissure area ratio(Sr)to GSI is greater than that of the strength reduction factor(a),which means the proposed GSI is suitable for rocks with significant fissure development.The improved method reduces the influence of subjective factors and provides a reliable index for the deterioration evaluation of rock mass. 展开更多
关键词 Hyperspectral image Digital panoramic borehole image Geological strength index Carbonate rock mass Quantitative evaluation
下载PDF
SMILE soft X-ray Imager flight model CCD370 pre-flight device characterisation 被引量:1
10
作者 S.Parsons D.J.Hall +4 位作者 O.Hetherington T.W.Buggey T.Arnold M.W.J.Hubbard A.Holland 《Earth and Planetary Physics》 EI CSCD 2024年第1期25-38,共14页
Throughout the SMILE mission the satellite will be bombarded by radiation which gradually damages the focal plane devices and degrades their performance.In order to understand the changes of the CCD370s within the sof... Throughout the SMILE mission the satellite will be bombarded by radiation which gradually damages the focal plane devices and degrades their performance.In order to understand the changes of the CCD370s within the soft X-ray Imager,an initial characterisation of the devices has been carried out to give a baseline performance level.Three CCDs have been characterised,the two flight devices and the flight spa re.This has been carried out at the Open University in a bespo ke cleanroom measure ment facility.The results show that there is a cluster of bright pixels in the flight spa re which increases in size with tempe rature.However at the nominal ope rating tempe rature(-120℃) it is within the procure ment specifications.Overall,the devices meet the specifications when ope rating at -120℃ in 6 × 6 binned frame transfer science mode.The se rial charge transfer inefficiency degrades with temperature in full frame mode.However any charge losses are recovered when binning/frame transfer is implemented. 展开更多
关键词 CCD soft X-ray imager characterisation SMILE
下载PDF
Artificial Intelligence and Computer Vision during Surgery: Discussing Laparoscopic Images with ChatGPT4—Preliminary Results 被引量:1
11
作者 Savvas Hirides Petros Hirides +1 位作者 Kouloufakou Kalliopi Constantinos Hirides 《Surgical Science》 2024年第3期169-181,共13页
Introduction: Ultrafast latest developments in artificial intelligence (ΑΙ) have recently multiplied concerns regarding the future of robotic autonomy in surgery. However, the literature on the topic is still scarce... Introduction: Ultrafast latest developments in artificial intelligence (ΑΙ) have recently multiplied concerns regarding the future of robotic autonomy in surgery. However, the literature on the topic is still scarce. Aim: To test a novel AI commercially available tool for image analysis on a series of laparoscopic scenes. Methods: The research tools included OPENAI CHATGPT 4.0 with its corresponding image recognition plugin which was fed with a list of 100 laparoscopic selected snapshots from common surgical procedures. In order to score reliability of received responses from image-recognition bot, two corresponding scales were developed ranging from 0 - 5. The set of images was divided into two groups: unlabeled (Group A) and labeled (Group B), and according to the type of surgical procedure or image resolution. Results: AI was able to recognize correctly the context of surgical-related images in 97% of its reports. For the labeled surgical pictures, the image-processing bot scored 3.95/5 (79%), whilst for the unlabeled, it scored 2.905/5 (58.1%). Phases of the procedure were commented in detail, after all successful interpretations. With rates 4 - 5/5, the chatbot was able to talk in detail about the indications, contraindications, stages, instrumentation, complications and outcome rates of the operation discussed. Conclusion: Interaction between surgeon and chatbot appears to be an interesting frontend for further research by clinicians in parallel with evolution of its complex underlying infrastructure. In this early phase of using artificial intelligence for image recognition in surgery, no safe conclusions can be drawn by small cohorts with commercially available software. Further development of medically-oriented AI software and clinical world awareness are expected to bring fruitful information on the topic in the years to come. 展开更多
关键词 Artificial Intelligence SURGERY image Recognition Autonomous Surgery
下载PDF
Enhancing Dense Small Object Detection in UAV Images Based on Hybrid Transformer 被引量:1
12
作者 Changfeng Feng Chunping Wang +2 位作者 Dongdong Zhang Renke Kou Qiang Fu 《Computers, Materials & Continua》 SCIE EI 2024年第3期3993-4013,共21页
Transformer-based models have facilitated significant advances in object detection.However,their extensive computational consumption and suboptimal detection of dense small objects curtail their applicability in unman... Transformer-based models have facilitated significant advances in object detection.However,their extensive computational consumption and suboptimal detection of dense small objects curtail their applicability in unmanned aerial vehicle(UAV)imagery.Addressing these limitations,we propose a hybrid transformer-based detector,H-DETR,and enhance it for dense small objects,leading to an accurate and efficient model.Firstly,we introduce a hybrid transformer encoder,which integrates a convolutional neural network-based cross-scale fusion module with the original encoder to handle multi-scale feature sequences more efficiently.Furthermore,we propose two novel strategies to enhance detection performance without incurring additional inference computation.Query filter is designed to cope with the dense clustering inherent in drone-captured images by counteracting similar queries with a training-aware non-maximum suppression.Adversarial denoising learning is a novel enhancement method inspired by adversarial learning,which improves the detection of numerous small targets by counteracting the effects of artificial spatial and semantic noise.Extensive experiments on the VisDrone and UAVDT datasets substantiate the effectiveness of our approach,achieving a significant improvement in accuracy with a reduction in computational complexity.Our method achieves 31.9%and 21.1%AP on the VisDrone and UAVDT datasets,respectively,and has a faster inference speed,making it a competitive model in UAV image object detection. 展开更多
关键词 UAV images TRANSFORMER dense small object detection
下载PDF
Two-Staged Method for Ice Channel Identification Based on Image Segmentation and Corner Point Regression 被引量:1
13
作者 DONG Wen-bo ZHOU Li +2 位作者 DING Shi-feng WANG Ai-ming CAI Jin-yan 《China Ocean Engineering》 SCIE EI CSCD 2024年第2期313-325,共13页
Identification of the ice channel is the basic technology for developing intelligent ships in ice-covered waters,which is important to ensure the safety and economy of navigation.In the Arctic,merchant ships with low ... Identification of the ice channel is the basic technology for developing intelligent ships in ice-covered waters,which is important to ensure the safety and economy of navigation.In the Arctic,merchant ships with low ice class often navigate in channels opened up by icebreakers.Navigation in the ice channel often depends on good maneuverability skills and abundant experience from the captain to a large extent.The ship may get stuck if steered into ice fields off the channel.Under this circumstance,it is very important to study how to identify the boundary lines of ice channels with a reliable method.In this paper,a two-staged ice channel identification method is developed based on image segmentation and corner point regression.The first stage employs the image segmentation method to extract channel regions.In the second stage,an intelligent corner regression network is proposed to extract the channel boundary lines from the channel region.A non-intelligent angle-based filtering and clustering method is proposed and compared with corner point regression network.The training and evaluation of the segmentation method and corner regression network are carried out on the synthetic and real ice channel dataset.The evaluation results show that the accuracy of the method using the corner point regression network in the second stage is achieved as high as 73.33%on the synthetic ice channel dataset and 70.66%on the real ice channel dataset,and the processing speed can reach up to 14.58frames per second. 展开更多
关键词 ice channel ship navigation IDENTIFICATION image segmentation corner point regression
下载PDF
An Intelligent Sensor Data Preprocessing Method for OCT Fundus Image Watermarking Using an RCNN 被引量:1
14
作者 Jialun Lin Qiong Chen 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第2期1549-1561,共13页
Watermarks can provide reliable and secure copyright protection for optical coherence tomography(OCT)fundus images.The effective image segmentation is helpful for promoting OCT image watermarking.However,OCT images ha... Watermarks can provide reliable and secure copyright protection for optical coherence tomography(OCT)fundus images.The effective image segmentation is helpful for promoting OCT image watermarking.However,OCT images have a large amount of low-quality data,which seriously affects the performance of segmentationmethods.Therefore,this paper proposes an effective segmentation method for OCT fundus image watermarking using a rough convolutional neural network(RCNN).First,the rough-set-based feature discretization module is designed to preprocess the input data.Second,a dual attention mechanism for feature channels and spatial regions in the CNN is added to enable the model to adaptively select important information for fusion.Finally,the refinement module for enhancing the extraction power of multi-scale information is added to improve the edge accuracy in segmentation.RCNN is compared with CE-Net and MultiResUNet on 83 gold standard 3D retinal OCT data samples.The average dice similarly coefficient(DSC)obtained by RCNN is 6%higher than that of CE-Net.The average 95 percent Hausdorff distance(95HD)and average symmetric surface distance(ASD)obtained by RCNN are 32.4%and 33.3%lower than those of MultiResUNet,respectively.We also evaluate the effect of feature discretization,as well as analyze the initial learning rate of RCNN and conduct ablation experiments with the four different models.The experimental results indicate that our method can improve the segmentation accuracy of OCT fundus images,providing strong support for its application in medical image watermarking. 展开更多
关键词 Watermarks image segmentation rough convolutional neural network attentionmechanism feature discretization
下载PDF
Image super‐resolution via dynamic network 被引量:1
15
作者 Chunwei Tian Xuanyu Zhang +2 位作者 Qi Zhang Mingming Yang Zhaojie Ju 《CAAI Transactions on Intelligence Technology》 SCIE EI 2024年第4期837-849,共13页
Convolutional neural networks depend on deep network architectures to extract accurate information for image super‐resolution.However,obtained information of these con-volutional neural networks cannot completely exp... Convolutional neural networks depend on deep network architectures to extract accurate information for image super‐resolution.However,obtained information of these con-volutional neural networks cannot completely express predicted high‐quality images for complex scenes.A dynamic network for image super‐resolution(DSRNet)is presented,which contains a residual enhancement block,wide enhancement block,feature refine-ment block and construction block.The residual enhancement block is composed of a residual enhanced architecture to facilitate hierarchical features for image super‐resolution.To enhance robustness of obtained super‐resolution model for complex scenes,a wide enhancement block achieves a dynamic architecture to learn more robust information to enhance applicability of an obtained super‐resolution model for varying scenes.To prevent interference of components in a wide enhancement block,a refine-ment block utilises a stacked architecture to accurately learn obtained features.Also,a residual learning operation is embedded in the refinement block to prevent long‐term dependency problem.Finally,a construction block is responsible for reconstructing high‐quality images.Designed heterogeneous architecture can not only facilitate richer structural information,but also be lightweight,which is suitable for mobile digital devices.Experimental results show that our method is more competitive in terms of performance,recovering time of image super‐resolution and complexity.The code of DSRNet can be obtained at https://github.com/hellloxiaotian/DSRNet. 展开更多
关键词 CNN dynamic network image super‐resolution lightweight network
下载PDF
CAEFusion: A New Convolutional Autoencoder-Based Infrared and Visible Light Image Fusion Algorithm 被引量:1
16
作者 Chun-Ming Wu Mei-Ling Ren +1 位作者 Jin Lei Zi-Mu Jiang 《Computers, Materials & Continua》 SCIE EI 2024年第8期2857-2872,共16页
To address the issues of incomplete information,blurred details,loss of details,and insufficient contrast in infrared and visible image fusion,an image fusion algorithm based on a convolutional autoencoder is proposed... To address the issues of incomplete information,blurred details,loss of details,and insufficient contrast in infrared and visible image fusion,an image fusion algorithm based on a convolutional autoencoder is proposed.The region attention module is meant to extract the background feature map based on the distinct properties of the background feature map and the detail feature map.A multi-scale convolution attention module is suggested to enhance the communication of feature information.At the same time,the feature transformation module is introduced to learn more robust feature representations,aiming to preserve the integrity of image information.This study uses three available datasets from TNO,FLIR,and NIR to perform thorough quantitative and qualitative trials with five additional algorithms.The methods are assessed based on four indicators:information entropy(EN),standard deviation(SD),spatial frequency(SF),and average gradient(AG).Object detection experiments were done on the M3FD dataset to further verify the algorithm’s performance in comparison with five other algorithms.The algorithm’s accuracy was evaluated using the mean average precision at a threshold of 0.5(mAP@0.5)index.Comprehensive experimental findings show that CAEFusion performs well in subjective visual and objective evaluation criteria and has promising potential in downstream object detection tasks. 展开更多
关键词 image fusion deep learning auto-encoder(AE) INFRARED visible light
下载PDF
Efficient Unsupervised Image Stitching Using Attention Mechanism with Deep Homography Estimation 被引量:1
17
作者 Chunbin Qin Xiaotian Ran 《Computers, Materials & Continua》 SCIE EI 2024年第4期1319-1334,共16页
Traditional feature-based image stitching techniques often encounter obstacles when dealing with images lackingunique attributes or suffering from quality degradation. The scarcity of annotated datasets in real-life s... Traditional feature-based image stitching techniques often encounter obstacles when dealing with images lackingunique attributes or suffering from quality degradation. The scarcity of annotated datasets in real-life scenesseverely undermines the reliability of supervised learning methods in image stitching. Furthermore, existing deeplearning architectures designed for image stitching are often too bulky to be deployed on mobile and peripheralcomputing devices. To address these challenges, this study proposes a novel unsupervised image stitching methodbased on the YOLOv8 (You Only Look Once version 8) framework that introduces deep homography networksand attentionmechanisms. Themethodology is partitioned into three distinct stages. The initial stage combines theattention mechanism with a pooling pyramid model to enhance the detection and recognition of compact objectsin images, the task of the deep homography networks module is to estimate the global homography of the inputimages consideringmultiple viewpoints. The second stage involves preliminary stitching of the masks generated inthe initial stage and further enhancement through weighted computation to eliminate common stitching artifacts.The final stage is characterized by adaptive reconstruction and careful refinement of the initial stitching results.Comprehensive experiments acrossmultiple datasets are executed tometiculously assess the proposed model. Ourmethod’s Peak Signal-to-Noise Ratio (PSNR) and Structure Similarity Index Measure (SSIM) improved by 10.6%and 6%. These experimental results confirm the efficacy and utility of the presented model in this paper. 展开更多
关键词 Unsupervised image stitching deep homography estimation YOLOv8 attention mechanism
下载PDF
DeepSVDNet:A Deep Learning-Based Approach for Detecting and Classifying Vision-Threatening Diabetic Retinopathy in Retinal Fundus Images 被引量:1
18
作者 Anas Bilal Azhar Imran +4 位作者 Talha Imtiaz Baig Xiaowen Liu Haixia Long Abdulkareem Alzahrani Muhammad Shafiq 《Computer Systems Science & Engineering》 2024年第2期511-528,共18页
Artificial Intelligence(AI)is being increasingly used for diagnosing Vision-Threatening Diabetic Retinopathy(VTDR),which is a leading cause of visual impairment and blindness worldwide.However,previous automated VTDR ... Artificial Intelligence(AI)is being increasingly used for diagnosing Vision-Threatening Diabetic Retinopathy(VTDR),which is a leading cause of visual impairment and blindness worldwide.However,previous automated VTDR detection methods have mainly relied on manual feature extraction and classification,leading to errors.This paper proposes a novel VTDR detection and classification model that combines different models through majority voting.Our proposed methodology involves preprocessing,data augmentation,feature extraction,and classification stages.We use a hybrid convolutional neural network-singular value decomposition(CNN-SVD)model for feature extraction and selection and an improved SVM-RBF with a Decision Tree(DT)and K-Nearest Neighbor(KNN)for classification.We tested our model on the IDRiD dataset and achieved an accuracy of 98.06%,a sensitivity of 83.67%,and a specificity of 100%for DR detection and evaluation tests,respectively.Our proposed approach outperforms baseline techniques and provides a more robust and accurate method for VTDR detection. 展开更多
关键词 Diabetic retinopathy(DR) fundus images(FIs) support vector machine(SVM) medical image analysis convolutional neural networks(CNN) singular value decomposition(SVD) classification
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部