An attempt has been made to develop a distributed software infrastructure model for onboard data fusion system simulation, which is also applied to netted radar systems, onboard distributed detection systems and advan...An attempt has been made to develop a distributed software infrastructure model for onboard data fusion system simulation, which is also applied to netted radar systems, onboard distributed detection systems and advanced C3I systems. Two architectures are provided and verified: one is based on pure TCP/IP protocol and C/S model, and implemented with Winsock, the other is based on CORBA (common object request broker architecture). The performance of data fusion simulation system, i.e. reliability, flexibility and scalability, is improved and enhanced by two models. The study of them makes valuable explore on incorporating the distributed computation concepts into radar system simulation techniques.展开更多
In distributed quantum computing(DQC),quantum hardware design mainly focuses on providing as many as possible high-quality inter-chip connections.Meanwhile,quantum software tries its best to reduce the required number...In distributed quantum computing(DQC),quantum hardware design mainly focuses on providing as many as possible high-quality inter-chip connections.Meanwhile,quantum software tries its best to reduce the required number of remote quantum gates between chips.However,this“hardware first,software follows”methodology may not fully exploit the potential of DQC.Inspired by classical software-hardware co-design,this paper explores the design space of application-specific DQC architectures.More specifically,we propose Auto Arch,an automated quantum chip network(QCN)structure design tool.With qubits grouping followed by a customized QCN design,AutoArch can generate a near-optimal DQC architecture suitable for target quantum algorithms.Experimental results show that the DQC architecture generated by Auto Arch can outperform other general QCN architectures when executing target quantum algorithms.展开更多
In LEO(Low Earth Orbit)satellite communication systems,the satellite network is made up of a large number of satellites,the dynamically changing network environment affects the results of distributed computing.In orde...In LEO(Low Earth Orbit)satellite communication systems,the satellite network is made up of a large number of satellites,the dynamically changing network environment affects the results of distributed computing.In order to improve the fault tolerance rate,a novel public blockchain consensus mechanism that applies a distributed computing architecture in a public network is proposed.Redundant calculation of blockchain ensures the credibility of the results;and the transactions with calculation results of a task are stored distributed in sequence in Directed Acyclic Graphs(DAG).The transactions issued by nodes are connected to form a net.The net can quickly provide node reputation evaluation that does not rely on third parties.Simulations show that our proposed blockchain has the following advantages:1.The task processing speed of the blockchain can be close to that of the fastest node in the entire blockchain;2.When the tasks’arrival time intervals and demanded working nodes(WNs)meet certain conditions,the network can tolerate more than 50%of malicious devices;3.No matter the number of nodes in the blockchain is increased or reduced,the network can keep robustness by adjusting the task’s arrival time interval and demanded WNs.展开更多
To security support large-scale intelligent applications,distributed machine learning based on blockchain is an intuitive solution scheme.However,the distributed machine learning is difficult to train due to that the ...To security support large-scale intelligent applications,distributed machine learning based on blockchain is an intuitive solution scheme.However,the distributed machine learning is difficult to train due to that the corresponding optimization solver algorithms converge slowly,which highly demand on computing and memory resources.To overcome the challenges,we propose a distributed computing framework for L-BFGS optimization algorithm based on variance reduction method,which is a lightweight,few additional cost and parallelized scheme for the model training process.To validate the claims,we have conducted several experiments on multiple classical datasets.Results show that our proposed computing framework can steadily accelerate the training process of solver in either local mode or distributed mode.展开更多
In many applications such as computational fluid dynamics and weather prediction, as well as image processing and state of Markov chain etc., the grade of matrix n is often very large, and any serial algorithm cannot ...In many applications such as computational fluid dynamics and weather prediction, as well as image processing and state of Markov chain etc., the grade of matrix n is often very large, and any serial algorithm cannot solve the problems. A distributed cluster-based solution for very large linear equations is discussed, it includes the definitions of notations, partition of matrix, communication mechanism, and a master-slaver algorithm etc., the computing cost is O(n^3/N), the memory cost is O(n^2/N), the I/O cost is O(n^2/N), and the com- munication cost is O(Nn ), here, N is the number of computing nodes or processes. Some tests show that the solution could solve the double type of matrix under 10^6 × 10^6 effectively.展开更多
Many science and engineering applications involve solvinga linear least-squares system formed from some field measurements. In the distributed cyber-physical systems(CPS),each sensor node used for measurement often on...Many science and engineering applications involve solvinga linear least-squares system formed from some field measurements. In the distributed cyber-physical systems(CPS),each sensor node used for measurement often only knowspartial independent rows of the least-squares system. To solve the least-squares all the measurements must be gathered at a centralized location and then perform the computa-tion. Such data collection and computation are inefficient because of bandwidth and time constraints and sometimes areinfeasible because of data privacy concerns. Iterative methods are natural candidates for solving the aforementionedproblem and there are many studies regarding this. However,most of the proposed solutions are related to centralized/parallel computations while only a few have the potential to beapplied in distributed networks. Thus distributed computations are strongly preferred or demanded in many of the realworld applications, e.g. smart-grid, target tracking, etc. Thispaper surveys the representative iterative methods for distributed least-squares in networks.展开更多
In the current noisy intermediate-scale quantum(NISQ)era,a single quantum processing unit(QPU)is insufficient to implement large-scale quantum algorithms;this has driven extensive research into distributed quantum com...In the current noisy intermediate-scale quantum(NISQ)era,a single quantum processing unit(QPU)is insufficient to implement large-scale quantum algorithms;this has driven extensive research into distributed quantum computing(DQC).DQC involves the cooperative operation of multiple QPUs but is concurrently challenged by excessive communication complexity.To address this issue,this paper proposes a quantum circuit partitioning method based on spectral clustering.The approach transforms quantum circuits into weighted graphs and,through computation of the Laplacian matrix and clustering techniques,identifies candidate partition schemes that minimize the total weight of the cut.Additionally,a global gate search tree strategy is introduced to meticulously explore opportunities for merged transfer of global gates,thereby minimizing the transmission cost of distributed quantum circuits and selecting the optimal partition scheme from the candidates.Finally,the proposed method is evaluated through various comparative experiments.The experimental results demonstrate that spectral clustering-based partitioning exhibits robust stability and efficiency in runtime in quantum circuits of different scales.In experiments involving the quantum Fourier transform algorithm and Revlib quantum circuits,the transmission cost achieved by the global gate search tree strategy is significantly optimized.展开更多
The ability of accurate and scalable mobile device recognition is critically important for mobile network operators and ISPs to understand their customers' behaviours and enhance their user experience.In this pape...The ability of accurate and scalable mobile device recognition is critically important for mobile network operators and ISPs to understand their customers' behaviours and enhance their user experience.In this paper,we propose a novel method for mobile device model recognition by using statistical information derived from large amounts of mobile network traffic data.Specifically,we create a Jaccardbased coefficient measure method to identify a proper keyword representing each mobile device model from massive unstructured textual HTTP access logs.To handle the large amount of traffic data generated from large mobile networks,this method is designed as a set of parallel algorithms,and is implemented through the MapReduce framework which is a distributed parallel programming model with proven low-cost and high-efficiency features.Evaluations using real data sets show that our method can accurately recognise mobile client models while meeting the scalability and producer-independency requirements of large mobile network operators.Results show that a 91.5% accuracy rate is achieved for recognising mobile client models from 2 billion records,which is dramatically higher than existing solutions.展开更多
Given the fast growth of intelligent devices, it is expected that a large number of high-stakes artificial intelligence (AI) applications, e. g., drones, autonomous cars, and tac?tile robots, will be deployed at the e...Given the fast growth of intelligent devices, it is expected that a large number of high-stakes artificial intelligence (AI) applications, e. g., drones, autonomous cars, and tac?tile robots, will be deployed at the edge of wireless networks in the near future. Therefore, the intelligent communication networks will be designed to leverage advanced wireless tech?niques and edge computing technologies to support AI-enabled applications at various end devices with limited communication, computation, hardware and energy resources. In this article, we present the principles of efficient deployment of model inference at network edge to provide low-latency and energy-efficient AI services. This includes the wireless distribut?ed computing framework for low-latency device distributed model inference as well as the wireless cooperative transmission strategy for energy-efficient edge cooperative model infer?ence. The communication efficiency of edge inference systems is further improved by build?ing up a smart radio propagation environment via intelligent reflecting surface.展开更多
Virtualization and distributed parallel architecture are typical cloud computing technologies. In the area of virtuatization technology, this article discusses physical resource pooling, resource pool management and u...Virtualization and distributed parallel architecture are typical cloud computing technologies. In the area of virtuatization technology, this article discusses physical resource pooling, resource pool management and use, cluster fault location and maintenance, resource pool grouping, and construction and application of heterogeneous virtualization platforms. In the area of distributed technology, distributed file system and KeyNalue storage engine are discussed. A solution is proposed for the host bottleneck problem, and a standard storage interface is proposed for the distributed file system. A directory-based storage scheme for Key/Value storage engine is also proposed.展开更多
文摘An attempt has been made to develop a distributed software infrastructure model for onboard data fusion system simulation, which is also applied to netted radar systems, onboard distributed detection systems and advanced C3I systems. Two architectures are provided and verified: one is based on pure TCP/IP protocol and C/S model, and implemented with Winsock, the other is based on CORBA (common object request broker architecture). The performance of data fusion simulation system, i.e. reliability, flexibility and scalability, is improved and enhanced by two models. The study of them makes valuable explore on incorporating the distributed computation concepts into radar system simulation techniques.
基金Project supported by the National Key R&D Program of China(Grant No.2023YFA1009403)the National Natural Science Foundation of China(Grant Nos.62072176 and 62472175)the“Digital Silk Road”Shanghai International Joint Lab of Trustworthy Intelligent Software(Grant No.22510750100)。
文摘In distributed quantum computing(DQC),quantum hardware design mainly focuses on providing as many as possible high-quality inter-chip connections.Meanwhile,quantum software tries its best to reduce the required number of remote quantum gates between chips.However,this“hardware first,software follows”methodology may not fully exploit the potential of DQC.Inspired by classical software-hardware co-design,this paper explores the design space of application-specific DQC architectures.More specifically,we propose Auto Arch,an automated quantum chip network(QCN)structure design tool.With qubits grouping followed by a customized QCN design,AutoArch can generate a near-optimal DQC architecture suitable for target quantum algorithms.Experimental results show that the DQC architecture generated by Auto Arch can outperform other general QCN architectures when executing target quantum algorithms.
基金funded in part by the National Natural Science Foundation of China (Grant no. 61772352, 62172061, 61871422)National Key Research and Development Project (Grants nos. 2020YFB1711800 and 2020YFB1707900)+2 种基金the Science and Technology Project of Sichuan Province (Grants no. 2021YFG0152, 2021YFG0025, 2020YFG0479, 2020YFG0322, 2020GFW035, 2020GFW033, 2020YFH0071)the R&D Project of Chengdu City (Grant no. 2019-YF05-01790-GX)the Central Universities of Southwest Minzu University (Grants no. ZYN2022032)
文摘In LEO(Low Earth Orbit)satellite communication systems,the satellite network is made up of a large number of satellites,the dynamically changing network environment affects the results of distributed computing.In order to improve the fault tolerance rate,a novel public blockchain consensus mechanism that applies a distributed computing architecture in a public network is proposed.Redundant calculation of blockchain ensures the credibility of the results;and the transactions with calculation results of a task are stored distributed in sequence in Directed Acyclic Graphs(DAG).The transactions issued by nodes are connected to form a net.The net can quickly provide node reputation evaluation that does not rely on third parties.Simulations show that our proposed blockchain has the following advantages:1.The task processing speed of the blockchain can be close to that of the fastest node in the entire blockchain;2.When the tasks’arrival time intervals and demanded working nodes(WNs)meet certain conditions,the network can tolerate more than 50%of malicious devices;3.No matter the number of nodes in the blockchain is increased or reduced,the network can keep robustness by adjusting the task’s arrival time interval and demanded WNs.
基金partly supported by National Key Basic Research Program of China(2016YFB1000100)partly supported by National Natural Science Foundation of China(NO.61402490)。
文摘To security support large-scale intelligent applications,distributed machine learning based on blockchain is an intuitive solution scheme.However,the distributed machine learning is difficult to train due to that the corresponding optimization solver algorithms converge slowly,which highly demand on computing and memory resources.To overcome the challenges,we propose a distributed computing framework for L-BFGS optimization algorithm based on variance reduction method,which is a lightweight,few additional cost and parallelized scheme for the model training process.To validate the claims,we have conducted several experiments on multiple classical datasets.Results show that our proposed computing framework can steadily accelerate the training process of solver in either local mode or distributed mode.
文摘In many applications such as computational fluid dynamics and weather prediction, as well as image processing and state of Markov chain etc., the grade of matrix n is often very large, and any serial algorithm cannot solve the problems. A distributed cluster-based solution for very large linear equations is discussed, it includes the definitions of notations, partition of matrix, communication mechanism, and a master-slaver algorithm etc., the computing cost is O(n^3/N), the memory cost is O(n^2/N), the I/O cost is O(n^2/N), and the com- munication cost is O(Nn ), here, N is the number of computing nodes or processes. Some tests show that the solution could solve the double type of matrix under 10^6 × 10^6 effectively.
基金partially supported by US NSF under Grant No.NSF-CNS-1066391and No.NSF-CNS-0914371,NSF-CPS-1135814 and NSF-CDI-1125165
文摘Many science and engineering applications involve solvinga linear least-squares system formed from some field measurements. In the distributed cyber-physical systems(CPS),each sensor node used for measurement often only knowspartial independent rows of the least-squares system. To solve the least-squares all the measurements must be gathered at a centralized location and then perform the computa-tion. Such data collection and computation are inefficient because of bandwidth and time constraints and sometimes areinfeasible because of data privacy concerns. Iterative methods are natural candidates for solving the aforementionedproblem and there are many studies regarding this. However,most of the proposed solutions are related to centralized/parallel computations while only a few have the potential to beapplied in distributed networks. Thus distributed computations are strongly preferred or demanded in many of the realworld applications, e.g. smart-grid, target tracking, etc. Thispaper surveys the representative iterative methods for distributed least-squares in networks.
基金supported by the National Natural Science Foundation of China(Grant No.62072259)in part by the Natural Science Foundation of Jiangsu Province(Grant No.BK20221411)+1 种基金the PhD Start-up Fund of Nantong University(Grant No.23B03)the Postgraduate Research&Practice Innovation Program of School of Information Science and Technology,Nantong University(Grant No.NTUSISTPR2405).
文摘In the current noisy intermediate-scale quantum(NISQ)era,a single quantum processing unit(QPU)is insufficient to implement large-scale quantum algorithms;this has driven extensive research into distributed quantum computing(DQC).DQC involves the cooperative operation of multiple QPUs but is concurrently challenged by excessive communication complexity.To address this issue,this paper proposes a quantum circuit partitioning method based on spectral clustering.The approach transforms quantum circuits into weighted graphs and,through computation of the Laplacian matrix and clustering techniques,identifies candidate partition schemes that minimize the total weight of the cut.Additionally,a global gate search tree strategy is introduced to meticulously explore opportunities for merged transfer of global gates,thereby minimizing the transmission cost of distributed quantum circuits and selecting the optimal partition scheme from the candidates.Finally,the proposed method is evaluated through various comparative experiments.The experimental results demonstrate that spectral clustering-based partitioning exhibits robust stability and efficiency in runtime in quantum circuits of different scales.In experiments involving the quantum Fourier transform algorithm and Revlib quantum circuits,the transmission cost achieved by the global gate search tree strategy is significantly optimized.
基金supported in part by the National Natural Science Foundation of China under Grant No.61072061the National Science and Technology Major Projects under Grant No.2012ZX03002008the Fundamental Research Funds for the Central Universities under Grant No.2012RC0121
文摘The ability of accurate and scalable mobile device recognition is critically important for mobile network operators and ISPs to understand their customers' behaviours and enhance their user experience.In this paper,we propose a novel method for mobile device model recognition by using statistical information derived from large amounts of mobile network traffic data.Specifically,we create a Jaccardbased coefficient measure method to identify a proper keyword representing each mobile device model from massive unstructured textual HTTP access logs.To handle the large amount of traffic data generated from large mobile networks,this method is designed as a set of parallel algorithms,and is implemented through the MapReduce framework which is a distributed parallel programming model with proven low-cost and high-efficiency features.Evaluations using real data sets show that our method can accurately recognise mobile client models while meeting the scalability and producer-independency requirements of large mobile network operators.Results show that a 91.5% accuracy rate is achieved for recognising mobile client models from 2 billion records,which is dramatically higher than existing solutions.
文摘Given the fast growth of intelligent devices, it is expected that a large number of high-stakes artificial intelligence (AI) applications, e. g., drones, autonomous cars, and tac?tile robots, will be deployed at the edge of wireless networks in the near future. Therefore, the intelligent communication networks will be designed to leverage advanced wireless tech?niques and edge computing technologies to support AI-enabled applications at various end devices with limited communication, computation, hardware and energy resources. In this article, we present the principles of efficient deployment of model inference at network edge to provide low-latency and energy-efficient AI services. This includes the wireless distribut?ed computing framework for low-latency device distributed model inference as well as the wireless cooperative transmission strategy for energy-efficient edge cooperative model infer?ence. The communication efficiency of edge inference systems is further improved by build?ing up a smart radio propagation environment via intelligent reflecting surface.
文摘Virtualization and distributed parallel architecture are typical cloud computing technologies. In the area of virtuatization technology, this article discusses physical resource pooling, resource pool management and use, cluster fault location and maintenance, resource pool grouping, and construction and application of heterogeneous virtualization platforms. In the area of distributed technology, distributed file system and KeyNalue storage engine are discussed. A solution is proposed for the host bottleneck problem, and a standard storage interface is proposed for the distributed file system. A directory-based storage scheme for Key/Value storage engine is also proposed.