期刊文献+
共找到32篇文章
< 1 2 >
每页显示 20 50 100
How Big Data and High-performance Computing Drive Brain Science
1
作者 Shanyu Chen Zhipeng He +9 位作者 Xinyin Han Xiaoyu He Ruilin Li Haidong Zhu Dan Zhao Chuangchuang Dai Yu Zhang Zhonghua Lu Xuebin Chi Beifang Niu 《Genomics, Proteomics & Bioinformatics》 SCIE CAS CSCD 2019年第4期381-392,共12页
Brain science accelerates the study of intelligence and behavior,contributes fundamental insights into human cognition,and offers prospective treatments for brain disease.Faced with the challenges posed by imaging tec... Brain science accelerates the study of intelligence and behavior,contributes fundamental insights into human cognition,and offers prospective treatments for brain disease.Faced with the challenges posed by imaging technologies and deep learning computational models,big data and high-performance computing(HPC)play essential roles in studying brain function,brain diseases,and large-scale brain models or connectomes.We review the driving forces behind big data and HPC methods applied to brain science,including deep learning,powerful data analysis capabilities,and computational performance solutions,each of which can be used to improve diagnostic accuracy and research output.This work reinforces predictions that big data and HPC will continue to improve brain science by making ultrahigh-performance analysis possible,by improving data standardization and sharing,and by providing new neuromorphic insights. 展开更多
关键词 Brain science Big data high-performance computing Brain connectomes Deep learning
原文传递
GCSS:a global collaborative scheduling strategy for wide-area high-performance computing
2
作者 Yao SONG Limin XIAO +4 位作者 Liang WANG Guangjun QIN Bing WEI Baicheng YAN Chenhao ZHANG 《Frontiers of Computer Science》 SCIE EI CSCD 2022年第5期1-15,共15页
Wide-area high-performance computing is widely used for large-scale parallel computing applications owing to its high computing and storage resources.However,the geographical distribution of computing and storage reso... Wide-area high-performance computing is widely used for large-scale parallel computing applications owing to its high computing and storage resources.However,the geographical distribution of computing and storage resources makes efficient task distribution and data placement more challenging.To achieve a higher system performance,this study proposes a two-level global collaborative scheduling strategy for wide-area high-performance computing environments.The collaborative scheduling strategy integrates lightweight solution selection,redundant data placement and task stealing mechanisms,optimizing task distribution and data placement to achieve efficient computing in wide-area environments.The experimental results indicate that compared with the state-of-the-art collaborative scheduling algorithm HPS+,the proposed scheduling strategy reduces the makespan by 23.24%,improves computing and storage resource utilization by 8.28%and 21.73%respectively,and achieves similar global data migration costs. 展开更多
关键词 high-performance computing scheduling strategy task scheduling data placement
原文传递
API Development Increases Access to Shared Computing Resources at Boston University
3
作者 George Jones Amanda E. Wakefield +4 位作者 Jeff Triplett Kojo Idrissa James Goebel Dima Kozakov Sandor Vajda 《Journal of Software Engineering and Applications》 2022年第6期197-207,共11页
Within the last few decades, increases in computational resources have contributed enormously to the progress of science and engineering (S & E). To continue making rapid advancements, the S & E community must... Within the last few decades, increases in computational resources have contributed enormously to the progress of science and engineering (S & E). To continue making rapid advancements, the S & E community must be able to access computing resources. One way to provide such resources is through High-Performance Computing (HPC) centers. Many academic research institutions offer their own HPC Centers but struggle to make the computing resources easily accessible and user-friendly. Here we present SHABU, a RESTful Web API framework that enables S & E communities to access resources from Boston University’s Shared Computing Center (SCC). The SHABU requirements are derived from the use cases described in this work. 展开更多
关键词 API Framework Open Source high-performance computing Software Architecture Science and Engineering
下载PDF
Static Analysis Techniques for Fixing Software Defects in MPI-Based Parallel Programs
4
作者 Norah Abdullah Al-Johany Sanaa Abdullah Sharaf +1 位作者 Fathy Elbouraey Eassa Reem Abdulaziz Alnanih 《Computers, Materials & Continua》 SCIE EI 2024年第5期3139-3173,共35页
The Message Passing Interface (MPI) is a widely accepted standard for parallel computing on distributed memorysystems.However, MPI implementations can contain defects that impact the reliability and performance of par... The Message Passing Interface (MPI) is a widely accepted standard for parallel computing on distributed memorysystems.However, MPI implementations can contain defects that impact the reliability and performance of parallelapplications. Detecting and correcting these defects is crucial, yet there is a lack of published models specificallydesigned for correctingMPI defects. To address this, we propose a model for detecting and correcting MPI defects(DC_MPI), which aims to detect and correct defects in various types of MPI communication, including blockingpoint-to-point (BPTP), nonblocking point-to-point (NBPTP), and collective communication (CC). The defectsaddressed by the DC_MPI model include illegal MPI calls, deadlocks (DL), race conditions (RC), and messagemismatches (MM). To assess the effectiveness of the DC_MPI model, we performed experiments on a datasetconsisting of 40 MPI codes. The results indicate that the model achieved a detection rate of 37 out of 40 codes,resulting in an overall detection accuracy of 92.5%. Additionally, the execution duration of the DC_MPI modelranged from 0.81 to 1.36 s. These findings show that the DC_MPI model is useful in detecting and correctingdefects in MPI implementations, thereby enhancing the reliability and performance of parallel applications. TheDC_MPImodel fills an important research gap and provides a valuable tool for improving the quality ofMPI-basedparallel computing systems. 展开更多
关键词 high-performance computing parallel computing software engineering software defect message passing interface DEADLOCK
下载PDF
Parallel Inference for Real-Time Machine Learning Applications
5
作者 Sultan Al Bayyat Ammar Alomran +3 位作者 Mohsen Alshatti Ahmed Almousa Rayyan Almousa Yasir Alguwaifli 《Journal of Computer and Communications》 2024年第1期139-146,共8页
Hyperparameter tuning is a key step in developing high-performing machine learning models, but searching large hyperparameter spaces requires extensive computation using standard sequential methods. This work analyzes... Hyperparameter tuning is a key step in developing high-performing machine learning models, but searching large hyperparameter spaces requires extensive computation using standard sequential methods. This work analyzes the performance gains from parallel versus sequential hyperparameter optimization. Using scikit-learn’s Randomized SearchCV, this project tuned a Random Forest classifier for fake news detection via randomized grid search. Setting n_jobs to -1 enabled full parallelization across CPU cores. Results show the parallel implementation achieved over 5× faster CPU times and 3× faster total run times compared to sequential tuning. However, test accuracy slightly dropped from 99.26% sequentially to 99.15% with parallelism, indicating a trade-off between evaluation efficiency and model performance. Still, the significant computational gains allow more extensive hyperparameter exploration within reasonable timeframes, outweighing the small accuracy decrease. Further analysis could better quantify this trade-off across different models, tuning techniques, tasks, and hardware. 展开更多
关键词 Machine Learning Models computational Efficiency Parallel computing Systems Random Forest Inference Hyperparameter Tuning Python Frameworks (TensorFlow PyTorch Scikit-Learn) high-performance computing
下载PDF
Efficient computation of Hash Hirschberg protein alignment utilizing hyper threading multi-core sharing technology
6
作者 Muhannad Abu-Hashem Adnan Gutub 《CAAI Transactions on Intelligence Technology》 SCIE EI 2022年第2期278-291,共14页
Due to current technology enhancement,molecular databases have exponentially grown requesting faster efficient methods that can handle these amounts of huge data.There-fore,Multi-processing CPUs technology can be used... Due to current technology enhancement,molecular databases have exponentially grown requesting faster efficient methods that can handle these amounts of huge data.There-fore,Multi-processing CPUs technology can be used including physical and logical processors(Hyper Threading)to significantly increase the performance of computations.Accordingly,sequence comparison and pairwise alignment were both found contributing significantly in calculating the resemblance between sequences for constructing optimal alignments.This research used the Hash Table-NGram-Hirschberg(HT-NGH)algo-rithm to represent this pairwise alignment utilizing hashing capabilities.The authors propose using parallel shared memory architecture via Hyper Threading to improve the performance of molecular dataset protein pairwise alignment.The proposed parallel hyper threading method targeted the transformation of the HT-NGH on the datasets decomposition for sequence level efficient utilization within the processing units,that is,reducing idle processing unit situations.The authors combined hyper threading within the multicore architecture processing on shared memory utilization remarking perfor-mance of 24.8%average speed up to 34.4%as the highest boosting rate.The benefit of this work improvement is shown preserving acceptable accuracy,that is,reaching 2.08,2.88,and 3.87 boost-up as well as the efficiency of 1.04,0.96,and 0.97,using 2,3,and 4 cores,respectively,as attractive remarkable results. 展开更多
关键词 computational biology high-performance computing Hyper Threading pairwise sequence alignment parallel design sequence alignment SHARED-MEMORY
下载PDF
Efficient Computation Offloading of IoT-Based Workflows Using Discrete Teaching Learning-Based Optimization
7
作者 Mohamed K.Hussein Mohamed H.Mousa 《Computers, Materials & Continua》 SCIE EI 2022年第11期3685-3703,共19页
As the Internet of Things(IoT)and mobile devices have rapidly proliferated,their computationally intensive applications have developed into complex,concurrent IoT-based workflows involving multiple interdependent task... As the Internet of Things(IoT)and mobile devices have rapidly proliferated,their computationally intensive applications have developed into complex,concurrent IoT-based workflows involving multiple interdependent tasks.By exploiting its low latency and high bandwidth,mobile edge computing(MEC)has emerged to achieve the high-performance computation offloading of these applications to satisfy the quality-of-service requirements of workflows and devices.In this study,we propose an offloading strategy for IoT-based workflows in a high-performance MEC environment.The proposed task-based offloading strategy consists of an optimization problem that includes task dependency,communication costs,workflow constraints,device energy consumption,and the heterogeneous characteristics of the edge environment.In addition,the optimal placement of workflow tasks is optimized using a discrete teaching learning-based optimization(DTLBO)metaheuristic.Extensive experimental evaluations demonstrate that the proposed offloading strategy is effective at minimizing the energy consumption of mobile devices and reducing the execution times of workflows compared to offloading strategies using different metaheuristics,including particle swarm optimization and ant colony optimization. 展开更多
关键词 high-performance computing internet of things(IoT) mobile edge computing(MEC) WORKFLOWS computation offloading teaching learning-based optimization
下载PDF
The use of high-performance and high-throughput computing for the fertilization of digital earth and global change studies
8
作者 Yong Xue Dominic Palmer-Brown Huadong Guo 《International Journal of Digital Earth》 SCIE 2011年第3期185-210,共26页
The study of global climate change seeks to understand:(1)the components of the Earth’s varying environmental system,with a particular focus on climate;(2)how these components interact to determine present conditions... The study of global climate change seeks to understand:(1)the components of the Earth’s varying environmental system,with a particular focus on climate;(2)how these components interact to determine present conditions;(3)the factors driving these components;(4)the history of global change and the projection of future change;and(5)how knowledge about global environmental variability and change can be applied to present-day and future decision-making.This paper addresses the use of high-performance computing and high-throughput computing for a global change study on the Digital Earth(DE)platform.Two aspects of the use of high-performance computing(HPC)/high-throughput computing(HTC)on the DE platform are the processing of data from all sources,especially Earth observation data,and the simulation of global change models.The HPC/HTC is an essential and efficient tool for the processing of vast amounts of global data,especially Earth observation data.The current trend involves running complex global climate models using potentially millions of personal computers to achieve better climate change predictions than would ever be possible using the supercomputers currently available to scientists. 展开更多
关键词 high-performance computing(HPC) high-throughput computing(HTC) digital earth global change climate change Earth observation grid computing
原文传递
Software approaches for resilience of high performance computing systems:a survey
9
作者 Jie JIA Yi LIU +2 位作者 Guozhen ZHANG Yulin GAO Depei QIAN 《Frontiers of Computer Science》 SCIE EI CSCD 2023年第4期43-56,共14页
With the scaling up of high-performance computing systems in recent years,their reliability has been descending continuously.Therefore,system resilience has been regarded as one of the critical challenges for large-sc... With the scaling up of high-performance computing systems in recent years,their reliability has been descending continuously.Therefore,system resilience has been regarded as one of the critical challenges for large-scale HPC systems.Various techniques and systems have been proposed to ensure the correct execution and completion of parallel programs.This paper provides a comprehensive survey of existing software resilience approaches.Firstly,a classification of software resilience approaches is presented;then we introduce major approaches and techniques,including checkpointing,replication,soft error resilience,algorithmbased fault tolerance,fault detection and prediction.In addition,challenges exposed by system-scale and heterogeneous architecture are also discussed. 展开更多
关键词 RESILIENCE high-performance computing fault tolerance CHALLENGE
原文传递
Spatial cloud computing: how can the geospatial sciences use and help shape cloud computing? 被引量:55
10
作者 Chaowei Yang Michael Goodchild +5 位作者 Qunying Huang Doug Nebert Robert Raskin Yan Xu Myra Bambacus Daniel Fay 《International Journal of Digital Earth》 SCIE 2011年第4期305-329,共25页
The geospatial sciences face grand information technology(IT)challenges in the twenty-first century:data intensity,computing intensity,concurrent access intensity and spatiotemporal intensity.These challenges require ... The geospatial sciences face grand information technology(IT)challenges in the twenty-first century:data intensity,computing intensity,concurrent access intensity and spatiotemporal intensity.These challenges require the readiness of a computing infrastructure that can:(1)better support discovery,access and utilization of data and data processing so as to relieve scientists and engineers of IT tasks and focus on scientific discoveries;(2)provide real-time IT resources to enable real-time applications,such as emergency response;(3)deal with access spikes;and(4)provide more reliable and scalable service for massive numbers of concurrent users to advance public knowledge.The emergence of cloud computing provides a potential solution with an elastic,on-demand computing platform to integrateobservation systems,parameter extracting algorithms,phenomena simulations,analytical visualization and decision support,and to provide social impact and user feedbackthe essential elements of the geospatial sciences.We discuss the utilization of cloud computing to support the intensities of geospatial sciences by reporting from our investigations on how cloud computing could enable the geospatial sciences and how spatiotemporal principles,the kernel of the geospatial sciences,could be utilized to ensure the benefits of cloud computing.Four research examples are presented to analyze how to:(1)search,access and utilize geospatial data;(2)configure computing infrastructure to enable the computability of intensive simulation models;(3)disseminate and utilize research results for massive numbers of concurrent users;and(4)adopt spatiotemporal principles to support spatiotemporal intensive applications.The paper concludes with a discussion of opportunities and challenges for spatial cloud computing(SCC). 展开更多
关键词 digital earth Cyber GIS GEODYNAMICS SPACE-TIME high-performance computing geospatial cyberinfrastructure
原文传递
Mobile Multimedia Computing in Cyber-Physical Surveillance Services Through UAV-Borne Video-SAR:A Taxonomy of Intelligent Data Processing for IoMT-Enabled Radar Sensor Networks 被引量:2
11
作者 Mohammad R.Khosravi Sadegh Samadi 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2022年第2期288-302,共15页
This study investigates the different aspects of multimedia computing in Video Synthetic Aperture Radar(Video-SAR)as a new mode of radar imaging for real-time remote sensing and surveillance.This research also conside... This study investigates the different aspects of multimedia computing in Video Synthetic Aperture Radar(Video-SAR)as a new mode of radar imaging for real-time remote sensing and surveillance.This research also considers new suggestions in the systematic design,research taxonomy,and future trends of radar data processing.Despite the conventional modes of SAR imaging,Video-SAR can generate video sequences to obtain online monitoring and green surveillance throughout the day and night(regardless of light sources)in all weathers.First,an introduction to Video-SAR is presented.Then,some specific properties of this imaging mode are reviewed.Particularly,this research covers one of the most important aspects of the Video-SAR systems,namely,the systematic design requirements,and also some new types of visual distortions which are different from the distortions,artifacts and noises observed in the conventional imaging radar.In addition,some topics on the general features and high-performance computing of Video-SAR towards radar communications through Unmanned Aerial Vehicle(UAV)platforms,Internet of Multimedia Things(IoMT),Video-SAR data processing issues,and real-world applications are investigated. 展开更多
关键词 Video Synthetic Aperture Radar(Video-SAR)imaging radar networks radar image processing high-performance computing Internet of Multimedia Things(IoMT) CYBERSECURITY
原文传递
A geospatial hybrid cloud platform based on multi-sourced computing and model resources for geosciences 被引量:2
12
作者 Qunying Huang Jing Li Zhenlong Li 《International Journal of Digital Earth》 SCIE EI 2018年第12期1184-1204,共21页
Cloud computing has been considered as the next-generation computing platform with the potential to address the data and computing challenges in geosciences.However,only a limited number of geoscientists have been ada... Cloud computing has been considered as the next-generation computing platform with the potential to address the data and computing challenges in geosciences.However,only a limited number of geoscientists have been adapting this platform for their scientific research mainly due to two barriers:1)selecting an appropriate cloud platform for a specific application could be challenging,as various cloud services are available and 2)existing general cloud platforms are not designed to support geoscience applications,algorithms and models.To tackle such barriers,this research aims to design a hybrid cloud computing(HCC)platform that can utilize and integrate the computing resources across different organizations to build a unified geospatial cloud computing platform.This platform can manage different types of underlying cloud infrastructure(e.g.,private or public clouds),and enables geoscientists to test and leverage the cloud capabilities through a web interface.Additionally,the platform also provides different geospatial cloud services,such as workflow as a service,on the top of common cloud services(e.g.,infrastructure as a service)provided by general cloud platforms.Therefore,geoscientists can easily create a model workflow by recruiting the needed models for a geospatial application or task on the fly.A HCC prototype is developed and dust storm simulation is used to demonstrate the capability and feasibility of such platform in facilitating geosciences by leveraging across-organization computing and model resources. 展开更多
关键词 Cloud computing Big Data geospatial cloud services workflow as a service(WaaS) geoprocessing as a service(GaaS) model as a service(MaaS) high-performance computing parallel computing
原文传递
High-performance predictor for critical unstable generators based on scalable parallelized neural networks 被引量:2
13
作者 Youbo LIU Yang LIU +3 位作者 Junyong LIU Maozhen LI Zhibo MA Gareth TAYLOR 《Journal of Modern Power Systems and Clean Energy》 SCIE EI 2016年第3期414-426,共13页
A high-performance predictor for critical unstable generators(CUGs) of power systems is presented in this paper. The predictor is driven by the Map Reduce based parallelized neural networks. Specifically, a group of b... A high-performance predictor for critical unstable generators(CUGs) of power systems is presented in this paper. The predictor is driven by the Map Reduce based parallelized neural networks. Specifically, a group of back propagation neural networks(BPNNs), fed by massive response trajectories data, are efficiently organized and concurrently trained in Hadoop to identify dynamic behavior of individual generator. Rather than simply classifying global stability of power systems, the presented approach is able to distinguish unstable generators accurately with a few cycles of synchronized trajectories after fault clearing,enabling more in-depth emergency awareness based on wide-area implementation. In addition, the technique is of rich scalability due to Hadoop framework, which can be deployed in the control centers as a high-performance computing infrastructure for real-time instability alert.Numerical examples are studied using NPCC 48-machines test system and a realistic power system of China. 展开更多
关键词 Transient stability Critical unstable generator(CUG) high-performance computing(HPC) Map Reduce based parallel BPNN Hadoop
原文传递
Parallel optimization of underwater acoustic models:A survey
14
作者 祝子杰 马树青 +3 位作者 朱小谦 蓝强 朴胜春 程玉胜 《Chinese Physics B》 SCIE EI CAS CSCD 2022年第10期189-207,共19页
Underwater acoustic models are effective tools for simulating underwater sound propagation.More than 50 years of research have been conducted on the theory and computational models of sound propagation in the ocean.Un... Underwater acoustic models are effective tools for simulating underwater sound propagation.More than 50 years of research have been conducted on the theory and computational models of sound propagation in the ocean.Unfortunately,underwater sound propagation models were unable to solve practical large-scale three-dimensional problems for many years due to limited computing power and hardware conditions.Since the mid-1980s,research on high performance computing for acoustic propagation models in the field of underwater acoustics has flourished with the emergence of high-performance computing platforms,enabling underwater acoustic propagation models to solve many practical application problems that could not be solved before.In this paper,the contributions of research on high-performance computing for underwater acoustic propagation models since the 1980s are thoroughly reviewed and the possible development directions for the future are outlined. 展开更多
关键词 underwater acoustic models underwater sound propagation high-performance computing modeling three-dimensional problems
原文传递
China’s EarthLab-Forefront of Earth System Simulation Research
15
作者 Zhaoyang CHAI He ZHANG +6 位作者 Mulan ZHANG Xiao TANG Weipeng ZHENG Jiang ZHU Guangqing ZHOU Junji CAO Qingcun ZENG 《Advances in Atmospheric Sciences》 SCIE CAS CSCD 2021年第10期1611-1620,共10页
This article introduces“EarthLab”,a major new Earth system numerical simulation facility developed in China.EarthLab is a numerical simulation system for a physical climate system,an environmental system,an ecologic... This article introduces“EarthLab”,a major new Earth system numerical simulation facility developed in China.EarthLab is a numerical simulation system for a physical climate system,an environmental system,an ecological system,a solid earth system,and a space weather system as a whole with a high-performance scientific computing platform.EarthLab consists of five key elements-namely:a global earth numerical simulation system,a regional high-precision simulation system,a supercomputing support and management system,a database,data assimilation and visualization system,and a high-performance computing system for earth sciences.EarthLab helps to study the atmosphere,hydrosphere,cryosphere,lithosphere,and biosphere,as well as their interactions,to improve the accuracy of predictions by integrating simulations and observations,and to provide a scientific foundation for major issues such as national disaster prevention and mitigation.The construction and operation of EarthLab will involve close cooperation with joint contributions and shared benefits. 展开更多
关键词 EarthLab earth system numerical simulation high-performance computing platform
下载PDF
Application of HPC and big data in post-pandemic times
16
作者 Henry M.Tufo David A.Yuen +3 位作者 Gabriele Morra Matthew G.Knepley Bei Zhang Shi Chen 《Earthquake Research Advances》 CSCD 2021年第3期21-28,共8页
We lay out the ramifications of the 2020 pandemic for all people in geosciences,especially the young,and argue for significant changes on training and career development.We focus primarily on its devastating impact in... We lay out the ramifications of the 2020 pandemic for all people in geosciences,especially the young,and argue for significant changes on training and career development.We focus primarily on its devastating impact in USA and compare with that in other countries especially China.We review the potential effect for the next four years or so on the aspirations of an academic career versus more realistic career goals.We urge people in mid-career about the need to reassess previous goals.We stress the need for students or researchers to acquire technical skills in high-performance computing(HPC),data analytics,artificial intelligence,and/or visualization along with a broad set of technical skills in applied computer science and mathematics.We give advice about hot prospects in several areas that have great potential for advancement in the coming decade,such as visualization,deep learning,quantum computing and information,and cloud computing,all of which lie within the aegis of HPC.Our forecast is that the pandemic will significantly reshape the job landscape and career paths for both young and established researchers and we discuss bluntly the dire situation facing junior people in geosciences in the aftermath of the pandemic around the world until 2024. 展开更多
关键词 GEOPHYSICS high-performance computing Higher education Post-pandemic era
下载PDF
The state-of-the-art in computer generated holography for 3D display 被引量:1
17
作者 David Blinder Tobias Birnbaum +1 位作者 Tomoyoshi Ito Tomoyoshi Shimobaba 《Light(Advanced Manufacturing)》 2022年第3期168-196,共29页
Holographic displays have the promise to be the ultimate 3D display technology,able to account for all visual cues.Recent advances in photonics and electronics gave rise to high-resolution holographic display prototyp... Holographic displays have the promise to be the ultimate 3D display technology,able to account for all visual cues.Recent advances in photonics and electronics gave rise to high-resolution holographic display prototypes,indicating that they may become widely available in the near future.One major challenge in driving those display systems is computational:computer generated holography(CGH)consists of numerically simulating diffraction,which is very computationally intensive.Our goal in this paper is to give a broad overview of the state-of-the-art in CGH.We make a classification of modern CGH algorithms,we describe different algorithmic CGH acceleration techniques,discuss the latest dedicated hardware solutions and indicate how to evaluate the perceptual quality of CGH.We summarize our findings,discuss remaining challenges and make projections on the future of CGH. 展开更多
关键词 Digital holography Holographic display computer generated holography high-performance computing FPGA Deep learning Visual quality assessment
原文传递
OHTMA:an optimized heuristic topology-aware mapping algorithm on the Tianhe-3 exascale supercomputer prototype
18
作者 Yi-shui LI Xin-hai CHEN +5 位作者 Jie LIU Bo YANG Chun-ye GONG Xin-biao GAN Sheng-guo LI Han XU 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2020年第6期939-949,共11页
With the rapid increase of the size of applications and the complexity of the supercomputer architecture,topology-aware process mapping becomes increasingly important.High communication cost has become a dominant cons... With the rapid increase of the size of applications and the complexity of the supercomputer architecture,topology-aware process mapping becomes increasingly important.High communication cost has become a dominant constraint of the performance of applications running on the supercomputer.To avoid a bad mapping strategy which can lead to terrible communication performance,we propose an optimized heuristic topology-aware mapping algorithm(OHTMA).The algorithm attempts to minimize the hop-byte metric that we use to measure the mapping results.OHTMA incorporates a new greedy heuristic method and pair-exchange-based optimization.It reduces the number of long-distance communications and effectively enhances the locality of the communication.Experimental results on the Tianhe-3 exascale supercomputer prototype indicate that OHTMA can significantly reduce the communication costs. 展开更多
关键词 high-performance computing Topology mapping Heuristic algorithm
原文传递
Numerical Study of Geometric Multigrid Methods on CPU–GPU Heterogeneous Computers
19
作者 Chunsheng Feng Shi Shu +1 位作者 Jinchao Xu Chen-Song Zhang 《Advances in Applied Mathematics and Mechanics》 SCIE 2014年第1期1-23,共23页
.The geometric multigrid method(GMG)is one of the most efficient solving techniques for discrete algebraic systems arising from elliptic partial differential equations.GMG utilizes a hierarchy of grids or discretizati... .The geometric multigrid method(GMG)is one of the most efficient solving techniques for discrete algebraic systems arising from elliptic partial differential equations.GMG utilizes a hierarchy of grids or discretizations and reduces the error at a number of frequencies simultaneously.Graphics processing units(GPUs)have recently burst onto the scientific computing scene as a technology that has yielded substantial performance and energy-efficiency improvements.A central challenge in implementing GMG on GPUs,though,is that computational work on coarse levels cannot fully utilize the capacity of a GPU.In this work,we perform numerical studies of GMG on CPU–GPU heterogeneous computers.Furthermore,we compare our implementation with an efficient CPU implementation of GMG and with the most popular fast Poisson solver,Fast Fourier Transform,in the cuFFT library developed by NVIDIA. 展开更多
关键词 high-performance computing CPU–GPU heterogeneous computers multigrid method fast Fourier transform partial differential equations.
下载PDF
Scalable near-repeat and event chain calculations over heterogeneous computer architecture and systems
20
作者 Xinyue Ye Xuan Shi Zhong Chen 《Big Earth Data》 EI 2017年第1期191-203,共13页
As a case of space-time interaction,near-repeat calculation indicates that when an event takes place at a certain location,its immediate geographical surroundings would face an increased risk of experiencing subsequen... As a case of space-time interaction,near-repeat calculation indicates that when an event takes place at a certain location,its immediate geographical surroundings would face an increased risk of experiencing subsequent events within a fairly short period of time.This paper presents an exploratory study that extends the investigation of the near-repeat phenomena to a series of space-time interaction,namely event chain calculation.Existing near-repeat tools can only deal with a limited amount of data due to computation constraints,let alone the event chain analysis.By deploying the modern accelerator technology and hybrid computer systems,this study demonstrates that large-scale near-repeat calculation or event chain analysis can be partially resolved through high-performance computing solutions to advance such a challenging statistical problem in both spatial analysis and crime geography. 展开更多
关键词 Parallel and high-performance computing near-repeat event chain analysis crime analysis
原文传递
上一页 1 2 下一页 到第
使用帮助 返回顶部