distributed computing Recently Published Documents

Total documents.

  • Latest Documents
  • Most Cited Documents
  • Contributed Authors
  • Related Sources
  • Related Keywords

Reliability of Trust Management Systems in Cloud Computing

Cloud computing is an innovation that conveys administrations like programming, stage, and framework over the web. This computing structure is wide spread and dynamic, which chips away at the compensation per-utilize model and supports virtualization. Distributed computing is expanding quickly among purchasers and has many organizations that offer types of assistance through the web. It gives an adaptable and on-request administration yet at the same time has different security dangers. Its dynamic nature makes it tweaked according to client and supplier’s necessities, subsequently making it an outstanding benefit of distributed computing. However, then again, this additionally makes trust issues and or issues like security, protection, personality, and legitimacy. In this way, the huge test in the cloud climate is selecting a perfect organization. For this, the trust component assumes a critical part, in view of the assessment of QoS and Feedback rating. Nonetheless, different difficulties are as yet present in the trust the board framework for observing and assessing the QoS. This paper talks about the current obstructions present in the trust framework. The objective of this paper is to audit the available trust models. The issues like insufficient trust between the supplier and client have made issues in information sharing likewise tended to here. Besides, it lays the limits and their enhancements to help specialists who mean to investigate this point.

Guest Editorial: Special Section on Parallel and Distributed Computing Techniques for Non-Von Neumann Technologies

Asynchronous rpc interface in distributed computing system, developing an efficient secure query processing algorithm on encrypted databases using data compression.

Abstract Distributed computing includes putting aside the data utilizing outsider storage and being able to get to this information from a place at any time. Due to the advancement of distributed computing and databases, high critical data are put in databases. However, the information is saved in outsourced services like Database as a Service (DaaS), security issues are raised from both server and client-side. Also, query processing on the database by different clients through the time-consuming methods and shared resources environment may cause inefficient data processing and retrieval. Secure and efficient data regaining can be obtained with the help of an efficient data processing algorithm among different clients. This method proposes a well-organized through an Efficient Secure Query Processing Algorithm (ESQPA) for query processing efficiently by utilizing the concepts of data compression before sending the encrypted results from the server to clients. We have addressed security issues through securing the data at the server-side by an encrypted database using CryptDB. Encryption techniques have recently been proposed to present clients with confidentiality in terms of cloud storage. This method allows the queries to be processed using encrypted data without decryption. To analyze the performance of ESQPA, it is compared with the current query processing algorithm in CryptDB. Results have proven the efficiency of storage space is less and it saves up to 63% of its space.

Preparing Distributed Computing Operations for the HL-LHC Era With Operational Intelligence

As a joint effort from various communities involved in the Worldwide LHC Computing Grid, the Operational Intelligence project aims at increasing the level of automation in computing operations and reducing human interventions. The distributed computing systems currently deployed by the LHC experiments have proven to be mature and capable of meeting the experimental goals, by allowing timely delivery of scientific results. However, a substantial number of interventions from software developers, shifters, and operational teams is needed to efficiently manage such heterogenous infrastructures. Under the scope of the Operational Intelligence project, experts from several areas have gathered to propose and work on “smart” solutions. Machine learning, data mining, log analysis, and anomaly detection are only some of the tools we have evaluated for our use cases. In this community study contribution, we report on the development of a suite of operational intelligence services to cover various use cases: workload management, data management, and site operations.

Deep distributed computing to reconstruct extremely large lineage trees

Distributed computing and artificial intelligence, volume 2: special sessions 18th international conference, software engineering, artificial intelligence, networking and parallel/distributed computing, chinese keyword extraction model with distributed computing, on allocation of systematic blocks in coded distributed computing, export citation format, share document.

Subscribe to the PwC Newsletter

Join the community, add a new evaluation result row, distributed computing.

70 papers with code • 0 benchmarks • 1 datasets

Benchmarks Add a Result

Most implemented papers, optuna: a next-generation hyperparameter optimization framework.

We will present the design-techniques that became necessary in the development of the software that meets the above criteria, and demonstrate the power of our new design through experimental results and real world applications.

A System for Massively Parallel Hyperparameter Tuning

recent research paper on distributed computing

Modern learning models are characterized by large hyperparameter spaces and long training times.

FedML: A Research Library and Benchmark for Federated Machine Learning

chaoyanghe/Awesome-Federated-Learning • 27 Jul 2020

Federated learning (FL) is a rapidly growing research field in machine learning.

CoCoA: A General Framework for Communication-Efficient Distributed Optimization

gingsmith/cocoa • 7 Nov 2016

The scale of modern datasets necessitates the development of efficient distributed optimization methods for machine learning.

Billion-scale Network Embedding with Iterative Random Projection

ZW-ZHANG/RandNE • 7 May 2018

Network embedding, which learns low-dimensional vector representation for nodes in the network, has attracted considerable research attention recently.

Orchestral: a lightweight framework for parallel simulations of cell-cell communication

Aratz/orchestral • 28 Jun 2018

By the use of operator-splitting we decouple the simulation of reaction-diffusion kinetics inside the cells from the simulation of molecular cell-cell interactions occurring on the boundaries between cells.

Distributed Bayesian Matrix Decomposition for Big Data Mining and Clustering

zhanglabtools/dbmd • 10 Feb 2020

Such a method should scale up well, model the heterogeneous noise, and address the communication issue in a distributed system.

Computing High Accuracy Power Spectra with Pico

marius311/pypico • 2 Dec 2007

This paper presents the second release of Pico (Parameters for the Impatient COsmologist).

Online Asynchronous Distributed Regression

ryadzenine/dolphin • 16 Jul 2014

Distributed computing offers a high degree of flexibility to accommodate modern learning constraints and the ever increasing size of datasets involved in massive data issues.

MLitB: Machine Learning in the Browser

software-engineering-amsterdam/MLitB • 8 Dec 2014

Beyond an educational resource for ML, the browser has vast potential to not only improve the state-of-the-art in ML research, but also, inexpensively and on a massive scale, to bring sophisticated ML learning and prediction to the public at large.

Distributed Systems and Parallel Computing

No matter how powerful individual computers become, there are still reasons to harness the power of multiple computational units, often spread across large geographic areas. Sometimes this is motivated by the need to collect data from widely dispersed locations (e.g., web pages from servers, or sensors for weather or traffic). Other times it is motivated by the need to perform enormous computations that simply cannot be done by a single CPU.

From our company’s beginning, Google has had to deal with both issues in our pursuit of organizing the world’s information and making it universally accessible and useful. We continue to face many exciting distributed systems and parallel computing challenges in areas such as concurrency control, fault tolerance, algorithmic efficiency, and communication. Some of our research involves answering fundamental theoretical questions, while other researchers and engineers are engaged in the construction of systems to operate at the largest possible scale, thanks to our hybrid research model .

Recent Publications

Some of our teams.

Algorithms & optimization

Graph mining

Network infrastructure

System performance

We're always looking for more talented, passionate people.

Careers

Open Access is an initiative that aims to make scientific research freely available to all. To date our community has made over 100 million downloads. It’s based on principles of collaboration, unobstructed discovery, and, most importantly, scientific progression. As PhD students, we found it difficult to access the research we needed, so we decided to create a new Open Access publisher that levels the playing field for scientists across the world. How? By making research easy to access, and puts the academic needs of the researchers before the business interests of publishers.

We are a community of more than 103,000 authors and editors from 3,291 institutions spanning 160 countries, including Nobel Prize winners and some of the world’s most-cited researchers. Publishing on IntechOpen allows authors to earn citations and find new collaborators, meaning more people see your work not only from your own field of study, but from other related fields too.

Brief introduction to this section that descibes Open Access especially from an IntechOpen perspective

Want to get in touch? Contact our London head office or media team here

Our team is growing all the time, so we’re always on the lookout for smart people who want to help us reshape the world of scientific publishing.

Home > Books > Recent Progress in Parallel and Distributed Computing

Introductory Chapter: The Newest Research in Parallel and Distributed Computing

Submitted: 14 September 2016 Published: 19 July 2017

DOI: 10.5772/intechopen.69201

Cite this chapter

There are two ways to cite this chapter:

From the Edited Volume

Recent Progress in Parallel and Distributed Computing

Edited by Wen-Jyi Hwang

To purchase hard copies of this book, please contact the representative in India: CBS Publishers & Distributors Pvt. Ltd. www.cbspd.com | [email protected]

Chapter metrics overview

1,366 Chapter Downloads

Impact of this chapter

Total Chapter Downloads on intechopen.com

IntechOpen

Total Chapter Views on intechopen.com

Author Information

Wen-jyi hwang *.

  • Department of Computer Science and Information Engineering, National Taiwan Normal University, Taipei, Taiwan

*Address all correspondence to: [email protected]

The parallel and distributed computing is concerned with concurrent use of multiple compute resources to enhance the performance of a distributed and/or computationally intensive application. The compute resources may be a single computer or a number of computers connected by a network. A computer in the system may contain single-core, multi-core and/or many-core processors. The design and implementation of a parallel and distributed system may involve the development, utilization and integration of techniques in computer network, software and hardware. With the advent of networking and computer technology, parallel and distributed computing systems have been widely employed for solving problems in engineering, management, natural sciences and social sciences.

There are six chapters in this book. From Chapters 2 to 6, a wide range of studies in new applications, algorithms, architectures, networks, software implementations and evaluations of this growing field are covered. These studies may be useful to scientists and engineers from various fields of specialization who need the techniques of distributed and parallel computing in their work.

The second chapter of this book considers the applications of distributed computing for social networks. The chapter entitled “A Study on the Node Influential Capability in Social Networks by Incorporating Trust Metrics” by Tong-Ming Lim and Hock Yeow Yap provides useful distributed computing models for the evaluation of node influential capacity in social networks. Two algorithms are presented in this study: Trust-enabled Generic Algorithm Diffusion Model (T-GADM) and Domain-Specified Trust-enabled Generic Algorithm Diffusion Model (DST-GADM). Experimental results confirm the hypothesis that social trust plays an important role in influential propagation. Moreover, it is able to increase the rate of success in influencing other social nodes in a social network.

Another application presented in this book is the smart grid for power engineering. The chapter entitled “A Distributed Computing Architecture for the Large-Scale Integration of Renewable Energy and Distributed Resources in Smart Grids” by Ignacio Aravena, Anthony Papavasiliou and Alex Papalexopoulos analyzes the distributed system for the management of the short-term operations of power systems. They propose optimization algorithms for both the levels of the distribution grid and high voltage grids. Numerical results are also included for illustrating the effectiveness of the algorithms.

This book also contains a chapter covering the programming aspect of parallel and distributed computing. For the study of parallel programming, the general processing units (GPUs) are considered. GPUs have received attention for parallel computing because their many-core capability offers a significant speedup over traditional general purpose processors. In the chapter entitled “GPU Computing Taxonomy” by Abdelrahman Ahmed Mohamed Osman, a new classification mechanism is proposed to facilitate the employment of GPU for solving the single program multiple data problems. Based on the number of hosts and the number of devices, the GPU computing can be separated into four classes. Examples are included to illustrate the features of each class. Efficient coding techniques are also provided.

The final two chapters focus on the software aspects of the distributed and parallel computing. Software tools for the wikinomics-oriented development of scientific applications are presented in the chapter entitled “Distributed Software Development Tools for Distributed Scientific Applications” by Vaidas Giedrimas, Anatoly Petrenko and Leonidas Sakalauskas. The applications are based on service-oriented architectures. Flexibilities are provided so that codes and components deployed can be reused and transformed into a service. Some prototypes are given to demonstrate the effectiveness of the proposed tools.

The chapter entitled “DANP-Evaluation of AHP-DSS” by Wolfgang Ossadnik, Benjamin Föcke and Ralf H. Kaspar evaluates the Analytic Hierarchy Process (AHP)-supporting software for the use of adequate Decision Support Systems (DSS) for the management science. The corresponding software selection, evaluation criteria, evaluation framework, assessments and evaluation results are provided in detail. Issues concerning the evaluation assisted by parallel and distributed computing are also addressed.

These chapters offer comprehensive coverage of parallel and distributed computing from engineering and science perspectives. They may be helpful to further stimulate and promote the research and development in this rapid growing area. It is also hoped that newcomers or researchers from other areas of disciplines desiring to learn more about the parallel and distributed computing will find this book useful.

© 2017 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution 3.0 License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Continue reading from the same book

Published: 19 July 2017

By Yap Hock Yeow and Lim Tong‐Ming

1422 downloads

By Ignacio Aravena, Anthony Papavasiliou and Alex Pap...

1425 downloads

By Abdelrahman Ahmed Mohamed Osman

1839 downloads

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals
  • My Account Login
  • Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • Open access
  • Published: 02 March 2024

Fiber optic computing using distributed feedback

  • Brandon Redding   ORCID: orcid.org/0000-0002-8477-7874 1 ,
  • Joseph B. Murray   ORCID: orcid.org/0000-0002-3965-6280 1 ,
  • Joseph D. Hart   ORCID: orcid.org/0000-0002-9396-0574 1 ,
  • Zheyuan Zhu   ORCID: orcid.org/0000-0001-9992-135X 2 ,
  • Shuo S. Pang 2 &
  • Raktim Sarma 3  

Communications Physics volume  7 , Article number:  75 ( 2024 ) Cite this article

1908 Accesses

55 Altmetric

Metrics details

  • Computational science
  • Optical physics
  • Transformation optics

The widespread adoption of machine learning and other matrix intensive computing algorithms has renewed interest in analog optical computing, which has the potential to perform large-scale matrix multiplications with superior energy scaling and lower latency than digital electronics. However, most optical techniques rely on spatial multiplexing, requiring a large number of modulators and detectors, and are typically restricted to performing a single kernel convolution operation per layer. Here, we introduce a fiber-optic computing architecture based on temporal multiplexing and distributed feedback that performs multiple convolutions on the input data in a single layer. Using Rayleigh backscattering in standard single mode fiber, we show that this technique can efficiently apply a series of random nonlinear projections to the input data, facilitating a variety of computing tasks. The approach enables efficient energy scaling with orders of magnitude lower power consumption than GPUs, while maintaining low latency and high data-throughput.

Similar content being viewed by others

recent research paper on distributed computing

The physics of optical computing

recent research paper on distributed computing

Scalable optical learning operator

recent research paper on distributed computing

Image sensing with multilayer nonlinear optical neural networks

Introduction.

Neural network-based machine learning algorithms have advanced dramatically in the last decade and are now routinely applied to a wide range of applications. However, this increase in performance has been accompanied by rapidly increasing computing demand—particularly in terms of the energy required to train and run these algorithms 1 . This has inspired research in alternative platforms capable of performing the computationally-intensive matrix-vector multiplications (MVMs) and kernel convolution operations at the heart of most machine learning algorithms more efficiently. Among these approaches, optical computing is particularly promising due to its superior energy scaling and potential to overcome memory-access bottlenecks 2 , 3 , 4 , 5 , 6 .

These unique features have led to a series of impressive demonstrations in which photonic computing systems have performed benchmark tasks with comparable accuracy to digital electronic neural networks while consuming orders of magnitude less energy 3 , 7 , 8 . Most of these photonic computing schemes rely on spatial multiplexing in which input data is encoded in parallel on an array of modulators and the MVM output is recorded on an array of photodetectors. This approach has been explored both in free-space and integrated photonic platforms. Free-space platforms typically employ spatial light modulators (SLMs) to encode data and cameras to record the computed output. While this enables very large-scale computing (e.g. input vectors as large as \(N \sim {10}^{6}\)   9 ), the latency is limited by the SLM and camera speeds (in systems where the SLM is used to encode the input data). Recent demonstrations have shown that analog data from the scene can be processed directly by a free-space, all-optical neural network 10 , 11 , overcoming this latency constraint, although this approach is not easily applied to arbitrary and complex datasets (e.g. non-optical datasets such as the SONAR data analyzed in this work). Integrated photonic solutions are both more compact and have the potential for high-speed operation by exploiting state-of-the-art modulators and detectors 12 , 13 . However, processing large matrices on-chip remains a challenge due to the size, heat, and complexity of integrating large numbers of individually addressable modulators. These limitations have inspired recent proposals for frequency 14 , 15 and temporally multiplexed architectures in which a single modulator is used to encode an entire vector 16 , 17 , 18 . These schemes could fill a gap in the photonic computing design space between the relatively slow, but large-scale free-space computing platforms that rely on SLMs to encode the input data and the high-speed but smaller-scale integrated photonic approaches. While initial demonstrations using temporal multiplexing were limited to a single neuron 16 , 19 , spatio-temporal schemes were recently introduced that support larger scale networks 8 , 18 , highlighting the potential for this approach to address the limitations of purely spatial multiplexing.

In addition, most of these photonic approaches are unable to natively perform multiple convolutions on the same input data. Convolutions have been used in a variety of machine learning techniques (e.g. convolutional neural networks) due to their ability to extract hierarchical features in a dataset 20 , 21 . To perform multiple convolutions, photonic computing platforms need to generate multiple copies of the input data, separately apply a kernel transform to each copy, and record each result. For example, this has been achieved using free-space optics by applying a series of convolutions in the Fourier plane of an image using a diffractive optical element designed to map the result of different convolutions to separate positions on a camera 22 . Recently, a time-wavelength multiplexing scheme was proposed to address this challenge 23 . This technique performed convolutions in parallel using optical frequency combs by encoding distinct kernel transforms on different sets of comb teeth. After wavelength de-multiplexing, an array of photodetectors was used to record the output of each kernel transform, enabling impressive throughput at the cost of increased system complexity and limited scalability (the number of kernels operations was limited by the number of individually addressable comb teeth).

In this work, we introduce a temporally multiplexed optical computing platform that performs convolutions using a simple and scalable approach based on distributed feedback in single mode fiber. We first encode the input vectors in the time domain as a pulse-train using a single high-speed modulator. This pulse-train is then injected into an optical fiber where a series of partial reflectors provide distributed feedback, generating a series of delayed copies of the input vector each weighted by the strength of a different reflector. In this demonstration, we rely on Rayleigh backscattering in standard single-mode fiber to provide this distributed feedback. Each Rayleigh scattering center creates a delayed copy of the input vector with random amplitude and phase—corresponding to the weights of a transformation matrix (i.e. a random kernel). The backscattered light is then recorded on a single, high-speed photodetector, performing the accumulation operation and introducing a non-linear transform. As explained in detail below, if the fiber is longer than the equivalent length of the encoded pulse train, then the fiber can perform multiple, distinct kernel operations on the input vector without requiring any additional routing or re-encoding of the input data. In principle, the weights of the transformation matrices (i.e. the kernels) could be inverse designed for specific computational tasks. As proof-of-concept, in this work, we use multiple random transformations to compute a non-linear random projection of the input vector into a higher dimensional space. We show that applying multiple random projections on the same input data can accelerate a variety of computing tasks including both unsupervised learning tasks such as non-linear principal component analysis (PCA) and supervised tasks including support vector machines (SVM) and extreme learning machines (ELM). More generally, this approach offers 5 major benefits: (1) It natively performs multiple convolutions in parallel on the input data. (2) It is scalable and is capable of processing relatively large-scale matrices (we demonstrate matrix operations on vectors with 784 elements) while maintaining relatively high-speed (10 µs per MVM). (3) Since this approach relies on a passive transform to perform the matrix operations, the energy consumption scales as \(O(N)\) , enabling significant reductions in power consumption compared to a graphics processing unit (GPU). (4) The entire system is constructed using commercially available, fiber-coupled components, enabling a robust and compact platform. (5) Since the system operates directly on fiber-coupled, time-series data, this approach could be used to directly analyze data transmitted over fiber, opening up additional applications in remote sensing, RF photonics, and telecommunications.

Operating principle

The basic operating principle is outlined in Fig.  1 . Data is first encoded in the time domain as a series of optical pulses. This pulse train is then injected into an optical fiber where it is partially reflected by a series of Rayleigh scattering centers. This distributed backscattering process randomly mixes the elements in the input vector, resulting in a backscattered speckle pattern that contains a series of random projections of the input vector. The Rayleigh backscattered (RBS) light is then recorded on a high-speed photodetector which performs a non-linear transform on the backscattered electric field 24 , 25 . In this work, we relied entirely on the non-linearity introduced by the detection process, although all-optical non-linearities introduced by light-matter interactions in optical fibers could be considered in the future. The digitized speckle pattern can then be used for a variety of computing tasks including non-linear principal component analysis, support vector machines, or extreme learning machines. More broadly, the Rayleigh backscattering process can form the first layer of an artificial neural network, efficiently expanding the input data to higher dimensional space. Since the specific weights and connections in the first layers of a neural network are not critical in most applications 26 , this process can be used to accelerate one of the most computationally intensive tasks in a neural network. A digital electronics back-end can then be used to complete the neural network. In recent years, random projections have been proposed for a variety of computing tasks in both free-space 9 , 24 , 25 , 27 , 28 and integrated photonic platforms 29 , 30 . While counterintuitive, researchers have shown that random transforms maintain key features in a dataset such as orthogonality while facilitating data analysis tasks such as dimensionality reduction or compressive sensing 31 , 32 .

figure 1

Data is encoded in in the time-domain as a train of optical pulses (represented as blue dots in the equivalent neural network diagram). This pulse train is then injected into an optical fiber where distributed feedback is mediated by Rayleigh backscattering. The distributed feedback produces a series of delayed copies of the original data with random phase and amplitude (the yellow dots in the neural network diagram). Recording the backscattered signal on a photodetector then provides a non-linear transform. Overall, this process results in a random non-linear projection of the original data into a higher dimensional space, facilitating a variety of computing and data analysis tasks including non-linear principal component analysis (PCA), support vector machine (SVM), or extreme learning machine (ELM).

Experimentally, this basic approach was realized using the architecture shown in Fig.  2a . A continuous wave (CW) laser was coupled into an electro-optic modulator (EOM) which was used to encode the input data. An \(N\) -element vector, \({A}_{N}\) , was encoded in the amplitude of a train of \(N\) pulses, as shown in the inset of Fig.  2a . This pulse train was coupled through a circulator into standard single-mode optical fiber where it was partially reflected by a series of Rayleigh scattering centers, creating a series of time-delayed copies of the original vector with random (though fixed) weights. The backscattered field, represented by an \(M\) -element complex vector \(\,{\widetilde{C}}_{M}\) , was directed to a photodetector which performed a non-linear transform, generating photocurrent proportional to \({\left|{\widetilde{C}}_{M}\right|}^{2}\) . We used a narrow-linewidth (<2 kHz) seed laser to ensure that this coherent summation produced a repeatable speckle pattern. Figure  2d, e shows examples of an encoded pulse train and the resulting Rayleigh backscattered speckle pattern, illustrating the dramatic increase in dimensionality provided by the Rayleigh backscattering process. In this case, the input data was a 60-element vector ( \(N=60\) ) representing a SONAR signal (discussed in the SVM section below) and the backscattered pattern was a ~2000-element vector.

figure 2

a The input vector \({A}_{N}\) is encoded in the time domain using an EOM as a train of optical pulses which are injected into the fiber through a circulator. Rayleigh scattering (described by a series of complex reflectivity coefficients, \(\widetilde{{r}_{i}}\) ) provides distributed feedback and the backscattered field, \(\widetilde{{C}_{M}}\) , is then recorded on a photodetector, providing a non-linear response. b The distributed scattering process can be described using a space-time diagram which tracks the position of each pulse as it travels through the fiber. The pulses (shown in different colors for clarity) are partially reflected by Rayleigh scattering centers with varying complex reflectivity as they propagate down the fiber. As a result, the backscattered field contains randomly weighted contributions from each input pulse (i.e. each element in the input vector). c The distributed scattering process can be expressed as the multiplication between a complex transfer matrix \({B}\) and the input vector \(A\) . d An example pulse train representing the SONAR data discussed in the SVM section and e the resulting RBS pattern. The inset in e shows a magnified view of the RBS pattern. EOM electro-optic modulator, SVM support vector machine, ADC analog-to-digital converter, Det photodetector, RBS Rayleigh backscattering.

The random kernel transform introduced by the Rayleigh backscattering process can be visualized using the space-time diagram shown in Fig.  2b . This diagram tracks the position of each pulse over time as it travels to the end of the fiber along a diagonal path in the upper-right direction. As each pulse propagates, it is partially reflected by a series of Rayleigh scattering centers and the reflected light travels back to the beginning of the fiber (along a diagonal path toward the bottom-right of the diagram). The Rayleigh backscattering at a given position in the fiber can be described by a complex reflectance, \({\widetilde{r}}_{k}\) , which is random but fixed. As shown in Fig.  2b , the backscattered light at time \({m}\) includes contributions from each input pulse (once all of the pulses have entered the fiber) weighted by the complex reflectance at different positions in the fiber. While Fig.  2b presents a simplified description of the Rayleigh backscattering process as a series of discrete partial reflectors, in reality, Rayleigh scattering is effectively continuous. However, the temporal correlation width of the Rayleigh backscattered light is set by the bandwidth (or pulse duration) of the encoded data 33 . Thus, for the 5 ns pulses used in this work, the Rayleigh backscattering process can be approximated as a series of discrete complex reflectance coefficients spaced every 0.5 m in the fiber (the round-trip distance covered in 5 ns). Using this approximation, we can express the backscattered light at time \(m\) as the sum of each input vector element scaled by the appropriate reflection coefficient: \({\widetilde{C}}_{m}={\sum }_{n=1}^{N}{A}_{n}{\widetilde{r}}_{m-n+1}\) , where \({\widetilde{r}}_{k}=0\) for \(k < 1\) or \(k > M\) . In other words, the Rayleigh backscattering process performs a series of vector convolution operations, applying different random kernels to the input vector as the pulse train propagates down the fiber. We can also express this transform as a matrix vector multiplication: \({\widetilde{C}}_{M}={\widetilde{B}}_{M\times N}\times {A}_{N}\) , where \({\widetilde{B}}_{M\times N}\) is an \(M\times N\) transfer matrix. As shown in Fig.  2c , the matrix \(\widetilde{B}\) contains the reflection coefficients, \({\widetilde{r}}_{k}\) , arranged so that each row in \(\widetilde{B}\) contains the same elements as the previous row, shifted by one column.

This system becomes particularly interesting if the fiber is longer than the equivalent length of the encoded pulse train (setting \(M > N\) ). In this case the fiber can perform multiple, distinct random kernel operations on the same input vector. This is significant from a neural network perspective since convolution operations are known to help extract hierarchical features in a dataset 20 , 21 . This is mathematically very different from most operations that have been implemented photonically in the past. First, photonic computing platforms exploiting random transforms 9 , 29 , 34 performed global, fully-random projections of the input data, rather than convolutions. Second, most photonic computing platforms, including highly reconfigurable integrated photonic systems 12 , 13 , are limited to applying a single kernel in each layer, rather than performing multiple, distinct kernel transforms on the same input data. A notable exception is the time-wavelength multiplexed approach which leveraged frequency combs to perform multiple convolutions in parallel 23 . However, in addition to the complexity of this approach, the number of kernels was limited by the number of individually addressable comb teeth τ.

In the distributed feedback system, the number of kernel operations can be increased simply by using a longer optical fiber or a faster data encoding rate. In particular, the length of the output vector \({\widetilde{C}}_{M}\) is set by the length of the fiber, \(L\) , and the data encoding rate, \({f}_{0}\) , as M ≈ 2· ([2 L /(c/n)]/ τ  +  N ), where \(c\) is the speed of light, \(n\) is the effective index in the fiber, and \(\tau =1/{f}_{0}\) is the length of the pulses representing the elements of \({A}_{N}\) . The term in the square brackets represents the round-trip time in the fiber while the factor of 2 outside the brackets accounts for using a polarization diversity receiver to record the backscattered light in both polarizations in parallel (see Methods). The additional \(N\) accounts for the length of the input vector and assumes we make use of RBS light that does not include every element in \({A}_{N}\) (i.e. RBS light collected before the entire pulse train enters the fiber and after the pulse train starts to leave the fiber). This expression also assumes that backscattered light is sampled at the data encoding rate of \({f}_{0}\) and that the temporal correlation width of the Rayleigh pattern matches the pulse duration \(\tau\) , which is the case for Rayleigh scattering 33 . By increasing the data encoding rate and the fiber length, this technique could be used to process large scale matrices or perform multiple distinct kernel operations on the same input vector (the number of distinct kernels is set by the ratio \(M/N\) ). For example, this platform could perform MVMs with \(M={10}^{6}\) using an encoding rate of 10 GHz and a 5 km fiber before attenuation becomes significant (note that 0.2 dB/km is typical for telecom fiber at a wavelength of 1550 nm, resulting in a round-trip loss of 2 dB for a 5 km fiber).

Although the temporal multiplexing approach presented here trades-off computing speed for the ability to use a single modulator and detector, the availability of high-speed optical modulators and detectors (e.g. 20 GHz devices are widely available) helps to mitigate this trade-off. The time required to compute a MVM can be expressed as \({\tau }_{{{{{\rm{MVM}}}}}}=N/{f}_{0}+2L/\left(\frac{c}{n}\right)\) , where the first term accounts for the length of the input pulse train and the second term represents the round-trip time in the fiber. In this work, we used a 500 m fiber and a 200 MHz encoding rate, yielding \({\tau }_{{{{{\rm{MVM}}}}}}\approx 10\,\mu s\) for \(N\) approaching 10 3 . Increasing the encoding rate to 20 GHz and using a 5 m fiber could enable a 100× speed-up ( \({\tau }_{{{{{\rm{MVM}}}}}}\approx 100\,{ns}\) ) while performing a MVM with the same matrix dimensions. This analysis implies that using higher frequency encoding is generally beneficial, enabling faster computation for a given matrix size. However, as we will discuss below, the power consumption also increases with the encoding rate and the proper balance will depend on the application.

Non-linear principal component analysis

As an initial demonstration of this technique, we considered a text-book example of non-linear principal component analysis. Non-linear PCA is an unsupervised learning technique that has been used for dimensionality reduction, singular value decomposition, denoising, and regression analysis 35 , 36 . A standard PCA relies on linear transforms to project data onto a new coordinate system that represents the variance in a dataset using as few dimensions as possible. However, relying entirely on linear transforms limits a standard PCA to analyzing data that is linearly separable 37 . A non-linear PCA operates by first applying a non-linear transform to a dataset before performing a standard PCA, facilitating the analysis of a wider range of data types. To illustrate how our platform can be used for non-linear PCA, we first created a dataset consisting of 500 points (defined by their cartesian coordinates \(x,y,z\) ) randomly distributed on 3 concentric spheres with radii of 1, 2, or 3, as shown in Fig.  3a . Ideally, the PCA would decompose the output vector into a single non-zero principal component representing the length of the \(x,y,z\) vector. At minimum, the PCA should result in <3 significant PCs representing a low dimensional space in which the data points can be linearly separated. A standard linear PCA is unable to separate these 3 classes of points, as shown in Fig.  3b which plots the weights of the first 2 PCs for each data point. Moreover, simply expanding the dimensionality without applying a non-linear transform was unable to separate the classes on its own. To illustrate this, we computationally applied a linear random transform to each coordinate by multiplying each coordinate by a 3 × 200 random matrix before applying a PCA. As shown in Fig.  3c , this cascaded transform (linear dimension expansion followed by PCA) is still linear and is unable to separate the classes.

figure 3

a The raw dataset consisted of 500 3-dimensional points randomly distributed on 3 concentric spheres. The colors (blue, red, orange) indicate data points on the 3 different spheres with radii of 1, 2, or 3. b Linear PCA of the raw data fails to separate the 3 classes of points. Plotted are the weights of the first 2 principle components (PC) for each datapoint. c A random linear transform is similarly unable to separate the 3 classes of points. d After the application of a non-linear random projection using Rayleigh backscattering, the 3 classes are clearly separable using a standard PCA. e Result of a PCA applied to 3 speckle grains selected from the RBS pattern, showing that the classes are difficult to separate without expanding the dimensionality. f Result of a PCA applied after re-compressing the RBS speckle pattern into 3 dimensions using a low-pass filter.

To perform a non-linear PCA, we used the Rayleigh backscattering platform to create a non-linear projection of the data onto a higher-dimensional space before applying a PCA. To do this, we injected each data point into the fiber using a train of three 50 ns pulses (i.e. an input vector with \(N=3\) and \(\tau =50\,{{{{{\rm{ns}}}}}}\) ). The EOM was initially biased at zero transmission and the voltage sent to the EOM was set by the amplitude of the \(x,y,z\) coordinates of each point (normalized such that the maximum coordinate of 3 was set to the maximum transmission voltage for the EOM, \({V}_{\pi }\) ). We then recorded the Rayleigh backscattered speckle pattern produced by each point, yielding a \(M=200\) output vector. This process effectively projected each 3-dimensional point into a 200-dimensional speckle pattern.

We then performed a standard linear PCA on the backscattered speckle patterns. As shown in Fig.  3d , the hybrid photonic/electronic non-linear PCA efficiently separated the three classes of points. This separation relied on both expanding the dimensionality and applying a non-linear transform to the original data. To illustrate this, we also attempted to perform a PCA using just 3 of the speckle grains in each backscattered speckle pattern. As shown in Fig.  3e , this was unable to efficiently separate the three classes (the separation depends on the 3 speckle grains selected, but the result shown is typical for most sets of 3 speckle grains). Expanding the dimensionality increased the chance of randomly finding a transform that projected the data into a space where it is highly separable.

While this demonstration showed that non-linear random projections can help identify variations in a dataset, the benefit of using analog optics (in terms of power consumption and speed) was limited since the final PCA was performed on a relatively large dataset with 200 dimensions. Fortunately, after expanding the dataset with the non-linear transform, we could then compress it into a lower dimensional vector in the analog optical domain before performing the final PCA while maintaining many of the benefits. To test this, we re-compressed the 200-dimensional RBS speckle pattern produced by each point to 3-dimensions by averaging 100 points at a time. Experimentally, this could be achieved using a photodetector with a low-pass filtering response. As shown in Fig.  3f , a PCA performed on the re-compressed vectors was still able to efficiently separate the three classes. This simple demonstration illustrates how a non-linear random projection can be used to identify key features in a dataset.

Non-linear support vector machine

The same non-linear random projections can be used to construct a non-linear support vector machine (NL-SVM). SVMs are a supervised learning technique designed to separate different classes in a dataset by identifying the maximum margin hyperplane separating two classes. A non-linear SVM first projects the data into a higher-dimensional space before finding a hyperplane to separate different classes 38 , 39 . The platform proposed in this work is ideally suited for non-linear SVM, since the convolutions introduced by the RBS process can efficiently project an input dataset into a higher-dimensional space to facilitate classification.

To explore the use of our platform for NL-SVM, we selected a benchmark SONAR dataset consisting of SONAR measurements of either rocks or metal cylinders (phantoms for underwater mines) 40 . The dataset consists of 97 measurements of rocks and 111 measurements of cylinders. Each measurement was obtained using a frequency-modulated chirped SONAR and contains a 60-element vector representing the reflected signal as a function of acoustic frequency, as shown in Fig.  4a, b . This dataset has been divided into a training and a validation dataset, each containing 104 measurements (we focused on the Aspect Angle Dependent Test described in 40 ). We first attempted to classify this data by using a standard linear SVM to assign a hyperplane based on the training data and then evaluated how well this hyperplane separates the validation data. A histogram showing the position of the validation data measurements along a 1-dimensional SVM projection is shown in Fig.  4c . Ideally, the two classes would be completely separable in this space; however, we observed significant overlap between the classes and obtained an overall classification accuracy of only 75%. We then used the RBS platform to perform a random non-linear projection on the SONAR data before applying an SVM. In this case, each SONAR measurement (consisting of a 60-element vector) was encoded in a pulse train using 5 ns pulses. The RBS process was used to expand the dimensionality of the SONAR signal from 60 to 2000. An example of an encoded pulse train and the resulting RBS pattern are shown in Fig.  2d, e . As shown in Fig.  4d , the non-linear SVM was much more effective at separating the two classes, obtaining an accuracy of 90.4% on the validation data, comparable to the performance of the neural network reported in 40 . This illustrates the potential for our platform to facilitate data classification by transforming input data into higher dimensional space using random convolutions.

figure 4

The SONAR dataset consisted of 97 measurements of rocks ( a ) and 111 measurements of cylinders ( b ). c Applying a linear SVM directly to the training data resulted in a classification accuracy of 75%. d After using the optical platform to perform a non-linear random projection on the SONAR data, the SVM accuracy increased to 90.4%.

Extreme learning machine

Extreme learning machines are a type of feed-forward neural network in which the weights and connections in the hidden layers are fixed and a single decision layer is trained to complete a task 41 , 42 . ELMs were initially proposed to avoid the computational demands of training every connection in a neural network, but their unique structure is particularly well-suited for optical implementations. Photonic ELMs can use passive photonic structures to apply a complex transform (i.e. the fixed layer in the ELM) and rely on a single electronic decision layer to complete the computing task 28 , 43 . This has enabled photonic computing architectures built around multimode fiber 28 or complex disordered materials 24 where precise control of the transfer matrix would be challenging. Here, we show that our distributed optical feedback platform can be configured as an ELM to perform image classification.

We tested our system on two benchmark tasks: classifying the MNIST Digit database and the MNIST Fashion database 44 . Each dataset consists of 60,000 training images and 10,000 test images in 10 classes (either hand-written digits from 0-9 or 10 types of clothing). Although the distributed feedback system performs vector convolutions, it can also be used to perform two-dimensional convolutions on image data that has been flattened to a one-dimensional vector (see Methods for details) 23 . To do this, we encoded each image as a one-dimensional pulse train by assigning the magnitude of each 5 ns pulse to the intensity of one pixel in the image. We injected pulse trains representing all 70,000 images in series and recorded 70,000 Rayleigh backscattered speckle patterns. We then used a ridge regression algorithm to train a decision layer to classify the 60,000 training images. Finally, we tested the ELM using the backscattered patterns obtained from the 10,000 validation images. This same process was repeated for the digit and fashion databases.

As shown in Fig.  5a, b , the distributed feedback ELM system achieved an accuracy of 96.7% and 85.3% on the digit and fashion databases, respectively. This performance is comparable to the best performing photonic neural networks. For example, on the digit database, a free-space optical extreme learning machine achieved an accuracy of 92% 43 , similar to a diffractive optical neural network which achieved 91.75% 11 . The time-wavelength multiplexed convolutional accelerator reported an accuracy of ~88% 23 . Integrated photonic approaches also report similar accuracies, including 95.3% 13 for a wavelength multiplexed scheme and 93.1% 8 using a VCSEL array. Hybrid optical/electronic approaches have yielded accuracy as high as 98% 45 . Our results are also similar to digital electronic neural networks with similar depth (~96%), while multi-layer neural networks achieve an accuracy above 99% 44 .

figure 5

a , b The photonic ELM classified the MNIST digit database with an accuracy of 96.7% and the Fashion database with an accuracy of 85.3%. c The ELM accuracy increases with the dimensionality of the output vector, \(M\) .

Our approach is able to achieve this accuracy simply by expanding the dimensionality and leveraging the inherent power of convolutions to extract high-level features from a dataset, without explicitly training or controlling the underlying transforms. This demonstration also shows that the same platform, and, in fact, the same optical fiber, can be used for multiple tasks simply by re-training the decision layer, consistent with previous demonstrations of the versatility of photonic ELMs 28 , 43 .

The most important parameter impacting the ELM accuracy is the length of the Rayleigh backscattering pattern (i.e. the length of the output vector, or the extent to which the system expanded the dimensionality of the original data). The entire measured RBS pattern consisted of ~2500 speckle grains (a ~6× increase in dimensionality compared to the 400-pixel MNIST Digit images). To investigate this dependence, we sub-sampled the measured RBS pattern and repeated the training and test procedure using output vectors of varying length. As shown in Fig.  5(c) , the accuracy increases rapidly with the length of the output vector before gradually plateauing. Higher accuracy is achieved for the digit database, which is consistent with previous studies showing that the fashion database is more challenging 44 . Nonetheless, the accuracy increases monotonically with the size of the output dimension in both cases. Fortunately, the distributed feedback platform is well suited for this task and can increase the dimensionality of the output data simply by using a fiber that is longer than the effective length of the pulse train. Moreover, since the backscattering process is entirely passive, increasing the dimensionality in this way has a negligible effect on the power consumption despite increasing the size of the MVM.

Energy consumption

As discussed, the distributed feedback process effectively performs a series of random convolutions on the input vector, which can be expressed as a matrix vector multiplication. This framework allows us to compare the power consumption that could be achieved by our approach to a standard digital electronic processor like a GPU by analyzing the energy per multiply-accumulate (MAC) operation. Our analysis accounts for the energy required to operate the laser, the modulator, and the photodetector as well as the digital to analog converter (DAC) and analog to digital converter (ADC). Assuming shot-noise limited detection, we first estimated the optical power required at the detector to obtain a Rayleigh backscattering pattern with the desired signal-to-noise ratio (SNR), as 46 :

where \(b\) is the effective number of bits (ENOB) with \({{{{{\rm{SNR}}}}}}=6.02\times b+1.76\) in dB, \(q\) is the charge of an electron, \({f}_{0}\) is the measurement bandwidth (and the data encoding rate), and \({{{{{\mathcal{R}}}}}}\) is the responsivity of the detector. We can then estimate the required laser power as

where \({T}_{{{{{\mathrm{mod}}}}}}\) is the transmission through the modulator and \({r}_{{{{{\rm{RBS}}}}}}\) is the average reflectance due to Rayleigh backscattering in the fiber. Since Rayleigh backscattering is a distributed process, the effective Rayleigh backscattering coefficient depends on the duration of the pulse train launched into the fiber and can be estimated as

for Corning SMF-28e+ (as used in this work) 47 , where \(N\) is the length of the input vector and \(\tau\) is the pulse duration for each element in the input vector \(.\) The total electrical power required to operate the laser can then be calculated as \({P}_{{{{{\rm{laser}}}}}}/\eta\) , where \(\eta\) is the wall-plug efficiency of the laser.

The electrical power consumed by the modulator can be expressed as 48

where \(N\) is the length of the input vector, \(M\) is the length of the output vector (i.e. the Rayleigh backscattering pattern), \({C}_{{{{{\mathrm{mod}}}}}}\) is the capacitance of the modulator, and \({V}_{\pi }\) is the peak-to-peak driving voltage of the modulator. The ratio of \(N/M\) denotes the operating duty-cycle of the modulator and we assume \(M\ge N\) , which corresponds to expanding the dimensionality of the input vector. The electrical power consumed by the photodetector can be expressed as

where \({V}_{{{{{\rm{bias}}}}}}\) is the bias voltage applied to the detector. State-of-the-art DACs and ADCs require ~0.5 pJ/use and ~1 pJ/use, respectively 49 , 50 , 51 . The DAC is used each time we switch the modulator, which occurs \(\left(N/M\right){f}_{0}\) times per second, whereas the ADC operates continuously at a rate of \({f}_{0}\) , resulting in power consumption of

The total power can then be expressed as \({P}_{{{{{\rm{total}}}}}}={P}_{{{{{\rm{laser}}}}}}/\eta +{P}_{{{{{\mathrm{mod}}}}}}+{P}_{\det }+{P}_{{{{{\rm{DAC}}}}}}+{P}_{{{{{\rm{ADC}}}}}}\) . In order to estimate the energy per MAC, we can then calculate the total number of MACs per second as \(N\cdot M/{\tau }_{{{{{\rm{MVM}}}}}}\) , where \({\tau }_{{{{{\rm{MVM}}}}}}\) is the time required to complete a matrix vector multiplication. Since a MVM is completed in the time required to record a Rayleigh backscattering pattern with \(M\) outputs, \({\tau }_{{{{{\rm{MVM}}}}}}=M\cdot \tau\) . Thus, the number of MACs per second is \(N/\tau\) or \(N{f}_{0}\) and the energy per MAC can be estimated as \({P}_{{{{{\rm{total}}}}}}/\left(N{f}_{0}\right)\) . Based on this analysis, we find that the energy per MAC required to power the laser scales as \({f}_{0}/{N}^{2}\) whereas the energy per MAC required to power the modulator, detector, DAC, and ADC scales as \(1/N\) and is independent of modulation frequency. This difference in scaling results from the dependence of \({r}_{{RBS}}\) on \(N\) and \(\tau\) in Eq.  3 . Finally, this analysis ignored the use of a polarization diversity receiver for simplicity. However, using polarization diversity does not change the energy per MAC. For example, using polarization diversity would require twice the laser power to achieve the same SNR, since the RBS light is split between two detectors, but enables twice the MACs per second. Instead, using a polarization diversity receiver doubles the data throughput (for the same modulation and detection speeds) without affecting the power efficiency.

This model allowed us to quantitatively evaluate the energy/MAC which could be achieved using the optical distributed feedback scheme. Here, we assumed a required ENOB of \(b=6\) , which corresponds to a SNR of 38 dB, considerably higher than the 23.6 dB SNR of the RBS patterns measured experimentally in this work (see Methods). We set \(M=10\times N\) , implying that the system was used to expand the dimensionality of the input vector by a factor of 10. We then assumed typical values for the remaining parameters: a laser with wall-plug efficiency of \(\eta =0.2\) , a modulator with \({C}_{{{{{\mathrm{mod}}}}}}=1{fF}\) and \({V}_{\pi }=1{V}\) , and a detector with \({V}_{{{{{\rm{bias}}}}}}=3V\) and \({{{{{\mathcal{R}}}}}}=1{A}/W\) 52 , 53 \(.\) We first estimated the energy per MAC as a function of the input vector size \(N\) using a modulation frequency of 200 MHz, matching our experimental conditions. As shown in Fig.  6a , the energy/MAC is dominated by the laser under these conditions. For comparison, Fig.  6 also shows the power consumption typical of state-of-the-art GPUs (~1 pJ/MAC), which is independent of vector size 30 . At a vector size of \(N=784\) , corresponding to the size of the \(28\times 28\) pixel images in the fashion database, the RBS system provides a \(30\times\) reduction in power consumption compared to a GPU. This improvement increases dramatically for larger vector sizes due to the \(1/{N}^{2}\) dependency of the laser energy. Reducing the required ENOB would also significantly reduce the power consumption by reducing the required laser power.

figure 6

a The energy/MAC required by the laser, detector, modulator, DAC, and ADC as a function of input vector size at an encoding rate of 200 MHz and assuming an ENOB of 6. b The total energy/MAC assuming an ENOB of 6 for varying encoding rates, \({f}_{0}\) , using either standard fiber (i.e. relying on Rayleigh backscattering, RBS) or using enhanced Rayleigh backscattering fiber with 15 dB higher reflectance. GPU graphics processing unit, ADC analog-to-digital convert, DAC digital-to-analog converter, RBS Rayleigh backscattering, MAC multiply-accumulate, ENOB effective number of bits.

For many applications, using higher modulation frequencies could be attractive to enable low latency computing and higher data throughput. For example, increasing the modulation rate to 10 GHz would reduce the latency from 10 \(\mu s\) to 200 ns for the same size MVMs presented in the ELM Section. In terms of computing power, operating at 10 GHz would enable this system to perform > \({10}^{12}\) MACs/second assuming an input vector size of \(N={10}^{3}\) , approaching the throughput of time-wavelength multiplexed schemes which have achieved 11 \(\times {10}^{12}\) operations per second 23 . The expected energy/MAC at an encoding rate of 10 GHz is shown in orange in Fig.  6(b) (the laser and ADC are still the dominant sources of energy consumption). In this system, the energy/MAC increases with encoding rate to compensate for the reduced backscattering at higher encoding rates (since \({r}_{{RBS}}\) depends on \(\tau\) ). Nonetheless, the RBS system still outperforms a GPU for input vectors with \(N > {10}^{3}\) at an encoding rate of 10 GHz. In order to reduce the energy consumption, we could use enhanced backscattering fiber, which is commercially available and provides ~15 dB higher backscattering than standard single-mode fiber 54 . Figure  6(b) also shows the energy/MAC required using enhanced backscattering fiber with a data encoding rate of 10 GHz, which approaches the energy/MAC achieved at 200 MHz using standard fiber. Combining enhanced scattering fiber with a 200 MHz encoding rate could enable even lower power consumption, outperforming a GPU for vectors with \(N > 25\) . Although this type of fiber was initially developed for fiber optic sensing, it is ideally suited for this type of optical computing platform and could enable a significant reduction in power consumption. A distributed feedback fiber employing ultra-weak fiber Bragg gratings 55 or point reflectors 56 could also be used to achieve higher reflectance and thus lower power consumption. These technologies could also enable customized kernel transforms by tailoring the position and strength of each reflector.

One challenge not addressed in this initial work is the environmental stability of the fiber. The PCA, SVM, and ELM algorithms investigated here each assume that a given launch pattern will always produce the same Rayleigh backscattering pattern (albeit with a finite SNR). However, the Rayleigh backscattering pattern depends on temperature and strain in addition to the launch pattern (Rayleigh backscattering in optical fiber is often used for distributed temperature and strain sensing 57 ). As an example, for the \(N=400\) pixel images in the digit database encoded at 200 MHz, the Rayleigh backscattering pattern would decorrelate if the temperature of the fiber drifted by 0.4 \(m^\circ C\) (see Methods for details). While this level of temperature stabilization is possible, single-mode fiber is a very controlled platform with limited degrees of freedom and relatively simple calibration techniques have been proposed in other applications leveraging Rayleigh backscattering which might be applicable here 58 . Note that in the experiments reported in this work, the fiber was enclosed in an aluminum box, but was not temperature stabilized. Instead, the training and validation datasets were recorded during a short measurement window while the fiber environment remained stable. For example, the ELM experiments required 0.7 s to record the RBS pattern from all 70,000 test images and we found that the RBS pattern did not drift significantly on this time-scale. Similarly, laser frequency drift could also impact the repeatability of the Rayleigh backscattering pattern. The sensitivity to laser frequency depends on the inverse of the pulse-train duration. For example, the digit database images were encoded using 2 \(\mu s\) pulse trains, requiring frequency stability better than ~500 kHz. In this work, we used a narrow-linewidth (~kHz) laser which was sufficiently stable on the time-scale of these measurements. However, active frequency stabilization may be required for longer term operation. Alternately, faster encoding or smaller datasets would relax the frequency stability requirement.

While this work introduced a basic architecture for performing optical computing with distributed feedback, there are several areas for improvement and further investigation. First, in this work, we relied entirely on the photodetection process to introduce the required non-linearity. Introducing non-linearities induced by light-matter interaction in the optical domain could enable more sophisticated, multi-layer networks capable of addressing more challenging computing tasks 59 . There are a few promising approaches which could be pursued to introduce intra-network non-linearities in this platform. For example, an analog optical-electrical-optical conversion could be introduced, following the approach used in single-node fiber-optic reservoir computers 19 and recently explored in the context of integrated photonic neural networks 59 , 60 . Alternately, the long interaction lengths and high-power density in single mode fiber could be leveraged to introduce all-optical non-linearities such as four-wave mixing, modulation instability, or stimulated Brillouin scattering 61 . Second, the temporal multiplexing scheme proposed here sacrifices computing speed for simplicity by enabling the use of a single modulator to encode the entire input vector. In the future, some degree of spatial multiplexing could also be included to balance this trade-off. On the detection side, this is particularly straight-forward and was already implemented in part using a polarization diversity receiver. In the future, recording the backscattered pattern from a few-mode fiber or multiple discrete fibers could increase the output vector size without compromising the computing speed. Some degree of parallelization could also be explored on the data encoding side, e.g., by using multiple modulators and coupling light into a few-mode fiber. Third, while we focused on random projections in this work, the same basic combination of temporal encoding and distributed feedback could be used to perform other operations. For example, instead of relying on Rayleigh backscattering, a series of carefully positioned partial reflectors (such as weak fiber Bragg gratings 55 or point reflectors 56 ) could be used to implement distinct kernel transforms at different positions along the fiber (the reflectivity at each position in the fiber could then be set to encode a desired kernel weight). Alternately, inverse design principles could be used to optimize the reflector geometry for a desired kernel response. These reflectors could potentially be reconfigurable (e.g. by straining a weak fiber Bragg grating to adjust its reflectivity at the seed laser wavelength), enabling reconfigurable kernels.

In summary, this work introduced an optical computing platform based on temporal multiplexing and distributed feedback that performs random convolutions using a passive optical fiber. We showed that Rayleigh backscattering in single mode fiber can be used to perform non-linear random kernel transforms on arbitrary input data to facilitate a variety of computing tasks, including non-linear principal component analysis, support vector machines, or extreme learning machines. This approach enables large scale MVMs with \(O\left(N\right)\) energy scaling using a single modulator and photodetector. The entire system can be constructed using off-the-shelf fiber-coupled components, providing a compact and accessible approach to analog optical computing. Finally, since this approach operates on temporally encoded light in standard single-mode fiber, it could potentially be applied directly to optical data transmitted over fiber enabling applications in remote sensing and RF photonics.

Experimental details

A narrow-linewidth laser (<2 kHz, RIO Orion, Grade 4) was used as the seed laser. In the experiments reported in this work, an Erbium-doped fiber amplifier (EDFA) was inserted after the EOM in Fig.  2 to increase the launch power and compensate for the use of a relatively low power (10 mW) seed laser and relatively high insertion loss EOM (4.5 dB). In the future, a higher power laser could be used to avoid needing the EDFA (depending on the length of the pulse train and encoding rate, peak launch power in the range of 10-100 mW is sufficient to achieve an \({enob}\) of 6 based on Eqs.  1 – 3 ). In this work, we also used a second EDFA followed by a wavelength division multiplexing filter after the circulator to amplify the Rayleigh backscattered pattern before detection. This EDFA was used to minimize the effect of photodetector noise. In the future, using a lower-noise detector would preclude the need for the second EDFA. Note that the EDFA operated in the small signal gain regime and did not introduce a non-linear response to the RBS pattern. A polarization diversity receiver (a polarizing beam splitter combined with two detectors) was used to record the RBS pattern in orthogonal polarizations.

In the ELM experiments, the 2-dimensional images were flattened into 1-dimensional vectors. As discussed in 23 , vector convolutions can be used to perform a convolution on 2-dimensional image data, although there is an overhead cost (not all output measurements are used) and the stride is inherently asymmetric (a symmetric stride could be obtained through sequential measurements by modifying the encoding strategy). In the image recognition tasks described here, we used all of the output samples, which included measurements representing standard 2D convolutions as well as mixtures that convolved different combinations of input pixels and would not be computed in a standard convolution. We also did not attempt to correct for the asymmetric stride introduced by the vector convolution process.

The digitized RBS speckle patterns were recorded at 1 GS/s. This is slightly faster than the temporal correlation width, which is set by the data encoding rate (i.e. 200 MHz for the SVM or ELM test and 20 MHz for the PCA test) 33 . In this work, we processed slightly oversampled Rayleigh backscattering patterns, but the dimensionality listed as \(M\) in each section corresponded to the pattern length divided by the correlation width of 5 ns or 50 ns.

We calculated the SNR of the RBS patterns recorded in this work as the ratio of the average power of the RBS pattern to the standard deviation in the background (recorded with the EOM blocked). We found an SNR of 23.6 dB, which corresponds to an \({enob}\) of 3.6. Fortunately, neural networks are quite robust to low-precision computing 62 , enabling the excellent performance achieved in the benchmark tasks explored in this work despite this modest precision.

Temperature Sensitivity

The temperature sensitivity of the Rayleigh backscattering pattern depends on the length of the incident pulse train and can be estimated as ( f 0 ⁄ N )⁄[1.2 GHz⁄°C], where the \(\left({f}_{0}/N\right)\) term represents the transform limit of the input pulse train (which has a duration of \(N\tau =N/{f}_{0}\) ) and the term in square brackets represents the shift in the Rayleigh spectrum with temperature 63 . This expression calculates the temperature shift required for the Rayleigh backscattering spectrum to shift by the transform limit of the pulse train, which would result in a decorrelated speckle pattern.

Data availability

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Patterson, D. et al. Carbon emissions and large neural network training. Preprint at arXiv https://doi.org/10.48550/arXiv.2104.10350 (2021).

Zhou, H. et al. Photonic matrix multiplication lights up photonic accelerator and beyond. Light Sci. Appl. 11 , 30 (2022).

Article   ADS   PubMed   PubMed Central   Google Scholar  

Wetzstein, G. et al. Inference in artificial intelligence with deep optics and photonics. Nature 588 , 39–47 (2020).

Article   ADS   CAS   PubMed   Google Scholar  

Solli, D. R. & Jalali, B. Analog optical computing. Nat. Photonics 9 , 704–706 (2015).

Article   ADS   CAS   Google Scholar  

Wu, J. et al. Analog optical computing for artificial intelligence. Engineering 10 , 133–145 (2022).

Article   Google Scholar  

Shastri, B. J. et al. Photonics for artificial intelligence and neuromorphic computing. Nat. Photonics 15 , 102–114 (2021).

El Srouji, L. et al. Photonic and optoelectronic neuromorphic computing. APL Photonics 7 , 051101 (2022).

Chen, Z. et al. Deep learning with coherent VCSEL neural networks. Nat. Photonics 17 , 723–730 (2023).

Rafayelyan, M., Dong, J., Tan, Y., Krzakala, F. & Gigan, S. Large-scale optical reservoir computing for spatiotemporal chaotic systems prediction. Phys. Rev. X 10 , 41037 (2020).

CAS   Google Scholar  

Wang, T. et al. Image sensing with multilayer nonlinear optical neural networks. Nat. Photonics 17 , 408 (2023).

Lin, X. et al. All-optical machine learning using diffractive deep neural networks. Science 361 , 1004 (2018).

Article   ADS   MathSciNet   CAS   PubMed   Google Scholar  

Shen, Y. et al. Deep learning with coherent nanophotonic circuits. Nat. Photonics 11 , 441–446 (2017).

Feldmann, J. et al. Parallel convolutional processing using an integrated photonic tensor core. Nature 589 , 52–58 (2021).

Fan, L. et al. Multidimensional convolution operation with synthetic frequency dimensions in photonics. Phys. Rev. Appl. 18 , 034088 (2022).

Fan, L., Wang, K., Wang, H., Dutt, A. & Fan, S. Experimental realization of convolution processing in photonic synthetic frequency dimensions. Sci. Adv. 9 , 4956 (2023).

Appeltant, L. et al. Information processing using a single dynamical node as complex system. Nat. Commun. 2 , 468 (2011).

Stelzer, F., Röhm, A., Vicente, R., Fischer, I. & Yanchuk, S. Deep neural networks using a single neuron: folded-in-time architecture using feedback-modulated delay loops. Nat. Commun. 12 , 5164 (2021).

Article   ADS   CAS   PubMed   PubMed Central   Google Scholar  

Hamerly, R., Bernstein, L., Sludds, A., Soljačić, M. & Englund, D. Large-scale optical neural networks based on photoelectric multiplication. Phys. Rev. X 9 , 021032 (2019).

Larger, L. et al. High-speed photonic reservoir computing using a time-delay-based architecture: million words per second classification. Phys. Rev. X 7 , 011015 (2017).

Google Scholar  

Lawrence, S., Giles, C. L., Tsoi, A. C. & Back, A. D. Face recognition: a convolutional neural-network approach. IEEE Trans. Neural Networks 8 , 98–113 (1997).

Article   CAS   PubMed   Google Scholar  

Lecun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature 521 , 436–444 (2015).

Chang, J., Sitzmann, V., Dun, X., Heidrich, W. & Wetzstein, G. Hybrid optical-electronic convolutional neural networks with optimized diffractive optics for image classification. Sci. Rep. 8 , 12324 (2018).

Xu, X. et al. 11 TOPS photonic convolutional accelerator for optical neural networks. Nature 589 , 44–51 (2021).

Ando, T., Horisaki, R. & Tanida, J. Speckle-learning-based object recognition through scattering media. Opt. Express 23 , 33902–33910 (2015).

Saade, A. et al. Random projections through multiple optical scattering: approximating kernels at the speed of light. in 2016 IEEE Int. Conf. Acoust. Speech Signal Process . 6215 (2016).

Havasi, M., Peharz, R. & Hernandez-Lobato, J. M. Minimal random code learning: getting bits back from compressed model parameters. Preprint at arXiv https://doi.org/10.48550/arXiv.1810.00440 (2018).

Dong, J., Rafayelyan, M., Krzakala, F. & Gigan, S. Optical reservoir computing using multiple light scattering for chaotic systems prediction. IEEE J. Sel. Top. Quantum Electron. 26 , 7701012 (2020).

Article   CAS   Google Scholar  

Teğin, U., Yıldırım, M., Oğuz, İ., Moser, C. & Psaltis, D. Scalable optical learning operator. Nat. Comput. Sci 1 , 542–549 (2021).

Article   PubMed   Google Scholar  

Wendland, D. et al. Coherent dimension reduction with integrated photonic circuits exploiting tailored disorder. J. Opt. Soc. Am. B 40 , B35–B40 (2023).

Wang, X. et al. Integrated photonic encoder for terapixel image processing. Preprint at arXiv https://doi.org/10.48550/arXiv.2306.04554 (2023).

Johnson, W. B. & Lindenstrauss, J. Extensions of Lipschitz mappings into a hilbert space. Contemp. Math 26 , 189–206 (1984).

Article   MathSciNet   Google Scholar  

Gigan, S. Imaging and computing with disorder. Nat. Phys. 18 , 980–985 (2022).

Mermelstein, M. D., Posey, R., Johnson, G. A. & Vohra, S. T. Rayleigh scattering optical frequency correlation in a single-mode optical fiber. Opt. Lett. 26 , 58–60 (2001).

Sarma, R., Yamilov, A., Neupane, P., Shapiro, B. & Cao, H. Probing long-range intensity correlations inside disordered photonic nanostructures. Phys. Rev. B - Condens. Matter Mater. Phys. 90 , 014203 (2014).

M. Scholz and R. Vigario, “Nonlinear PCA: A new hierarchical approach,” Proc. 10th Eur. Symp. Artif. Neural Networks , 439–444 (2002).

Mika, S. et al. Kernel PCA and de-noising in feature spaces. Adv. Neural Inf. Process. Syst . 11 , 536–542 (1999).

Lever, J., Krzywinski, M. & Altman, N. Principal component analysis. Nat. Methods 14 , 641–642 (2017).

Cortes, C. & Vapnik, V. Support-vector networks. Mach. Learn. 20 , 273–297 (1995).

Noble, W. S. What is a support vector machine?. Comput. Biol 24 , 1565–1567 (2006).

Gorman, R. P. & Sejnowski, T. J. Analysis of hidden units in a layered network trained to classify sonar targets. Neural Networks 1 , 75–89 (1988).

Huang, G., Zhu, Q. & Siew, C. Extreme learning machine: theory and applications. Neurocomputing 70 , 489–501 (2006).

Bin Huang, G., Zhou, H., Ding, X. & Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B Cybern. 42 , 513–529 (2012).

Pierangeli, D., Marcucci, G. & Conti, C. Photonic extreme learning machine by free-space optical propagation. Photonics Res. 9 , 1446–1454 (2021).

LeCun, Y., Bottou, L., Bengio, Y. & Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 86 , 2278–2323 (1998).

Miscuglio, M. et al. Massively parallel amplitude-only Fourier neural network. Optica 7 , 1812 (2020).

Article   ADS   Google Scholar  

Valley, G. C. Photonic analog-to-digital converters. Opt. Express 15 , 1955–1982 (2007).

Article   ADS   PubMed   Google Scholar  

Corning Incorporated. https://www.corning.com/media/worldwide/coc/documents/Fiber/PI-1463-AEN.pdf .

Miller, D. A. B. Energy consumption in optical modulators for interconnects. Opt. Express 20 , A293–A308 (2012).

Wada, O. et al. 5 GHz-band CMOS direct digital RF modulator using current-mode DAC. Asia-Pacific Microw. Conf. Proc. APMC , 1118–1120 (IEEE, 2012).

Caragiulo, B. M. P., Daigle, C. https://github.com/pietro-caragiulo/survey-DAC . DAC Perform. Surv. 1996–2020 .

Murmann, B. https://github.com/bmurmann/ADC-survey . ADC Perform. Surv. 1997–2023 .

Nozaki, K. et al. Femtofarad optoelectronic integration demonsrtating energy-saving signal conversion and nonlinear functions. Nat. Photonics 13 , 454–459 (2019).

Li, G. et al. 25Gb/s 1V-driving CMOS ring modulator with integrated thermal tuning. Opt. Express 19 , 20435–20443 (2011).

V. A. Handerek, et al. Improved optical power budget in distributed acoustic sensing using enhanced scattering optical fibre. 26th Int. Conf. Opt. Fiber Sensors , TuC5 (2018).

Guo, H., Liu, F., Yuan, Y., Yu, H. & Yang, M. Ultra-weak FBG and its refractive index distribution in the drawing optical fiber. Opt. Express 23 , 4829–4838 (2015).

Redding, B. et al. Low-noise distributed acoustic sensing using enhanced backscattering fiber with ultra-low-loss point reflectors. Opt. Express 28 , 14638–14647 (2020).

Masoudi, A. & Newson, T. P. Contributed review: distributed optical fibre dynamic strain sensing. Rev. Sci. Instrum. 87 , 011501 (2016).

Murray, M. J., Murray, J. B., Schermer, R. T., Mckinney, J. D. & Redding, B. High-speed RF spectral analysis using a Rayleigh backscattering speckle spectrometer. Opt. Express 31 , 20651–20664 (2023).

Pour Fard, M. M. et al. Experimental realization of arbitrary activation functions for optical neural networks. Opt. Express 28 , 12138–12148 (2020).

Ashtiani, F., Geers, A. J. & Aflatouni, F. An -on-chip photonic deep neural network for image classification. Nature 606 , 501–506 (2022).

Alem, M., Soto, M. A. & Thévenaz, L. Analytical model and experimental verification of the critical power for modulation instability in optical fibers. Opt. Express 23 , 29514–29532 (2015).

Sze, V., Chen, Y. H., Yang, T. J. & Emer, J. S. Efficient processing of deep neural networks: a tutorial and survey. Proc. IEEE 105 , 2295–2329 (2017).

Murray, M. J., Murray, J. B., Ogden, H. M. & Redding, B. Dynamic temperature-strain discrimination using a hybrid distributed fiber sensor based on Brillouin and Rayleigh scattering. Opt. Express 31 , 287–300 (2023).

Download references

Acknowledgements

B.R. and J.M. acknowledge support from the U.S. Naval Research Lab (6.1 Base Program). R.S. acknowledges the support by the Laboratory Directed Research and Development program at Sandia National Laboratories, a multimission laboratory managed and operated by National Technology and Engineering Solutions of Sandia, LLC, a wholly owned subsidiary of Honeywell International, Inc., for the U.S. Department of Energy’s National Nuclear Security Administration under contract DE-NA-003525. This paper describes objective technical results and analysis. Any subjective views or opinions that might be expressed in the paper do not necessarily represent the views of the U.S. Department of Energy or the United States Government.

Author information

Authors and affiliations.

US Naval Research Laboratory, Washington, DC, USA

Brandon Redding, Joseph B. Murray & Joseph D. Hart

CREOL, The College of Optics and Photonics, University of Central Florida, Orlando, FL, USA

Zheyuan Zhu & Shuo S. Pang

Center for Integrated Nanotechnologies, Sandia National Laboratories, Albuquerque, NM, USA

Raktim Sarma

You can also search for this author in PubMed   Google Scholar

Contributions

B.R., J.B.M, J.D.H., S.S.P., Z.Z., and R.S. conceived of the experiments. B.R. and J.B.M. performed the experiment. B.R., J.B.M, J.D.H., S.S.P., and R.S. analyzed the data. B.R., Z.Z., and S.S.P. performed the energy consumption analysis. B.R. wrote the paper with contributions from all authors.

Corresponding author

Correspondence to Brandon Redding .

Ethics declarations

Competing interests.

The authors declare no competing interests.

Peer review

Peer review information.

Communications Physics thanks Christophe Moser and the other, anonymous, reviewer(s) for their contribution to the peer review of this work.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Cite this article.

Redding, B., Murray, J.B., Hart, J.D. et al. Fiber optic computing using distributed feedback. Commun Phys 7 , 75 (2024). https://doi.org/10.1038/s42005-024-01549-1

Download citation

Received : 19 September 2023

Accepted : 05 February 2024

Published : 02 March 2024

DOI : https://doi.org/10.1038/s42005-024-01549-1

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

By submitting a comment you agree to abide by our Terms and Community Guidelines . If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.

recent research paper on distributed computing

A Research Review of Distributed Computing System

  • Conference paper
  • First Online: 23 August 2018
  • Cite this conference paper

recent research paper on distributed computing

  • Wang Xingang 16  

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 752))

1155 Accesses

1 Citations

With the rapid development of Internet technology, global data are growing at an exponential rate, and human computing has undergone profound changes. The early stand-alone mode has not met people’s needs, and the network-based collaborative distributed gained more and more attention and favor. Distributed computing system is one of the hottest Internet research directions in the era of big data era. It has the characteristics of high efficiency, high capacity, dynamic processing, and so on. It shows great application value in the commercial field and scientific research field of society. Based on the development and present situation of distributed computing system, this paper reviews and summarizes two popular key technologies: grid technology and cloud computing technology and expounds the differences between the two. At the same time, the design principles and system performance of Hadoop, Storm, Spark, and other typical distributed computing platforms are compared and analyzed in detail, which is of theoretical significance.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
  • Available as EPUB and PDF
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Coulouris G, Dollimore J, Kindberg T. Distributed systems: concepts and design—4th ed. [J]. Programmable Controllers, 2011, 18(95):182–231.

Google Scholar  

Mao zhen li, Mark Baker. The Grid core technologies, 2006: 254–262.

ZHANG Jian-xun, GU Zhi-min, ZHENG Chao. Survey of research progress on cloud computing [J]. Application Research of Computer. 2010, 27(02):429–433.

Werner Vogels. Web services are not distributed objects. Internet Computing, IEEE, 2003, 7(6): P59–66.

XIA Jing-bo, WEI Ze-kun, FU Kai, CHEN Zhen. Review of research and application on Hadoop on cloud computing [J]. Computer Science, 2016, 43(11):6–11.

Apache Software Foundation. Hadoop [EB/OL]. http://hadoop.apache.org/ .

Anderson Q. Storm Real-Time Processing Cookbook [M]. Packt Publishing, 2013.

Apache spark [EB/OL]. http://spark.apache.org/ .

Zhang Wenfeng. The principle and design of distributed computing platform based on MapReduce [D]. Huazhong University of Science and Technology, 2010.

Zhou Xiao-feng, Wang Zhi-jian. Overview of distributed computing technology [J]. Computer Era, 2004(12):3–5.

GE Peng. A brief overview on distributed computing technology [J]. Micro Electronics & Computer, 2012, 29(5):201–204.

Foster I, Kesselman C, Tuecke S. The Anatomy of the Grid: Enabling Scalable Virtual Organizations [J]. International Journal of High Performance Computing Applications, 2001, 15(1):6.

Foster I. What is the Grid? A Three Point Checklist [J]. Grid Today, 2002, 1:32–36.

Nist S P. A NIST definition of Cloud computing [J]. Communications of the Acm, 2015, 53(6):50–50.

Hayes B. Cloud computing [J]. Communications of the Acm, 2008, 51(7):9–11.

Apache Software Foundation. Apache Nutch [EB/OL]. http://nutch.apache.org/ .

YANG Junjie, LIAO Zhuofan, FENG Chaochao. Survey on big data storage framework and algorithm [J]. Journal of Computer Applications. 2016, 36(9):2465–2471.

Dean J, Ghemawat S. MapReduce: simplified data processing on large clusters [C]. Conference on Symposium on Operating Systems Design & Implementation. USENIX Association, 2004:10–10.

Borthakur D. The Hadoop Distributed File System: Architecture and Design [J]. Hadoop Project Website, 2007, 11(11):1 – 10.

Konar M, Konar M, Konar M, et al. Apache Hadoop YARN: yet another resource negotiator [C]. Symposium on Cloud Computing. ACM, 2013:5.

Zaharia M, Chowdhury M, Franklin M J, et al. Spark: cluster computing with working sets [C]. Usenix Conference on Hot Topics in Cloud Computing. USENIX Association, 2010:10–10.

HU Jun1, HU Xian-De1, CHEN Jia-Xing. Big Data Hybrid Computing Mode Based on Spark [J]. Computer Systems & Applications, 2015, 24(4):214–218.

Apache Storm [EB/OL]. http://storm.apache.org/ .

LI Chuan, E Hai-hong, SONG Mei-na. Research & Application of real-time compute framework based on Storm [J]. Software, 2014(10):16–20.

Github Inc. Storm Wiki [EB/OL]. [2014-11-02]. https://github.com/apache/storm .

Malewicz G, Austern M H, Bik A J C, et al. Pregel: a system for large-scale graph processing [C]. ACM SIGMOD International Conference on Management of Data. ACM, 2010:135–146.

Melnik S, Gubarev A, Long J J, et al. Dremel: Interactive Analysis of Web-Scale Datasets [J]. Communications of the Acm, 2011, 3(12):114–123.

Isard M, Budiu M, Yu Y, et al. Dryad: distributed data-parallel programs from sequential building blocks [J]. Acm Sigops Operating Systems Review, 2007, 41(3):59–72.

Download references

Author information

Authors and affiliations.

Administration Center for the DTH Service in China, SAPPRFT, No. 2, FuXingMenWai Street, Xicheng District, Beijing, 100045, China

Wang Xingang

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Wang Xingang .

Editor information

Editors and affiliations.

Department of Computer Science and Engineering, Faculty of Engineering and Technology, SOA University, Bhubaneswar, India

Srikanta Patnaik

Victoria Business School, Victoria University of Wellington, Wellington, New Zealand

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Singapore Pte Ltd.

About this paper

Cite this paper.

Xingang, W. (2019). A Research Review of Distributed Computing System. In: Patnaik, S., Jain, V. (eds) Recent Developments in Intelligent Computing, Communication and Devices. Advances in Intelligent Systems and Computing, vol 752. Springer, Singapore. https://doi.org/10.1007/978-981-10-8944-2_42

Download citation

DOI : https://doi.org/10.1007/978-981-10-8944-2_42

Published : 23 August 2018

Publisher Name : Springer, Singapore

Print ISBN : 978-981-10-8943-5

Online ISBN : 978-981-10-8944-2

eBook Packages : Intelligent Technologies and Robotics Intelligent Technologies and Robotics (R0)

Share this paper

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Publish with us

Policies and ethics

  • Find a journal
  • Track your research

Introduction—Parallel and Distributed Computing

Ieee account.

  • Change Username/Password
  • Update Address

Purchase Details

  • Payment Options
  • Order History
  • View Purchased Documents

Profile Information

  • Communications Preferences
  • Profession and Education
  • Technical Interests
  • US & Canada: +1 800 678 4333
  • Worldwide: +1 732 981 0060
  • Contact & Support
  • About IEEE Xplore
  • Accessibility
  • Terms of Use
  • Nondiscrimination Policy
  • Privacy & Opting Out of Cookies

A not-for-profit organization, IEEE is the world's largest technical professional organization dedicated to advancing technology for the benefit of humanity. © Copyright 2024 IEEE - All rights reserved. Use of this web site signifies your agreement to the terms and conditions.

ORIGINAL RESEARCH article

This article is part of the research topic.

Human Factors and Cognitive Ergonomics in Advanced Industrial Human-Robot Interaction

SOCIALLY ADAPTIVE COGNITIVE ARCHITECTURE FOR HUMAN-ROBOT COLLABORATION IN INDUSTRIAL SETTINGS Provisionally Accepted

  • 1 Donders Institute for Brain, Cognition and Behaviour, Radboud University, Netherlands

The final, formatted version of the article will be published soon.

This paper introduces DAC-HRC, a novel cognitive architecture designed to optimize human-robot collaboration (HRC) in industrial settings, particularly within the context of Industry 4.0. The architecture is grounded in the Distributed Adaptive Control theory and the principles of joint intentionality and interdependence, which are key to effective HRC. Joint intentionality refers to the shared goals and mutual understanding between a human and a robot, while interdependence emphasizes the reliance on each other's capabilities to complete tasks. DAC-HRC is applied to a hybrid recycling plant for the disassembly and recycling of Waste Electrical and Electronic Equipment (WEEE) devices. The architecture incorporates several cognitive modules operating at different timescales and abstraction levels, fostering adaptive collaboration that is personalized to each human user. The effectiveness of DAC-HRC is demonstrated through several pilot studies, showcasing functionalities such as turn-taking interaction, personalized error-handling mechanisms, adaptive safety measures, and gesture-based communication. These features enhance human-robot collaboration in the recycling plant by promoting real-time robot adaptation to human needs and preferences. The DAC-HRC architecture aims to contribute to the development of a new HRC paradigm by paving the way for more seamless and efficient collaboration in Industry 4.0 by relying on socially adept cognitive architectures.

Keywords: Cognitive Architecture, social robotics, Human-robot collaboration, Industry 4.0, Distributed adaptive control

Received: 27 Jun 2023; Accepted: 14 May 2024.

Copyright: © 2024 Freire, Guerrero-Rosado, Amil and Verschure. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY) . The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Mr. Ismael T. Freire, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands Mr. Oscar Guerrero-Rosado, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands Prof. Paul F. Verschure, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands

People also looked at

IMAGES

  1. Distributed Computing Research Free Essay Example

    recent research paper on distributed computing

  2. (PDF) Special Issue of Journal of Parallel and Distributed Computing

    recent research paper on distributed computing

  3. Distributed Computing Systems group assignment

    recent research paper on distributed computing

  4. (PDF) Distributed Computing: An Overview

    recent research paper on distributed computing

  5. Distributed Computing Report

    recent research paper on distributed computing

  6. (PDF) Research Paper on Cloud Computing

    recent research paper on distributed computing

VIDEO

  1. Distributed Computing Lecture 8: Architectures P2

  2. Distributed Computing-Class-1 (MS: 5203)

  3. CS 436: Distributed Computer Systems

  4. 11th standard question paper distributed to 12th class students

  5. PAPER DISTRIBUTED BUSSINESS JUST PART TIME JOB

  6. evolution of distributed computing

COMMENTS

  1. distributed computing Latest Research Papers

    Cloud computing is an innovation that conveys administrations like programming, stage, and framework over the web. This computing structure is wide spread and dynamic, which chips away at the compensation per-utilize model and supports virtualization. Distributed computing is expanding quickly among purchasers and has many organizations that ...

  2. New Advances in Distributed Computing and Its Applications

    Topics of interest for the Special Issue include, but are not limited to, the following: Advanced distributed computing architectures, such as edge computing, container-based microservices, and blockchain; New optimizations of distributed computing in big data processing and artificial intelligence; Innovations and challenges in the synergy ...

  3. The evolution of distributed computing systems: from fundamental to new

    Distributed systems have been an active field of research for over 60 years, and has played a crucial role in computer science, enabling the invention of the Internet that underpins all facets of modern life. Through technological advancements and their changing role in society, distributed systems have undergone a perpetual evolution, with each change resulting in the formation of a new ...

  4. Journal of Parallel and Distributed Computing

    The Journal of Parallel and Distributed Computing publishes original research papers and timely review articles on the theory, design, evaluation, and use of parallel and/or distributed computing systems. The journal also features special issues on these topics; again covering the full range from the design to the use of our targeted systems.

  5. Advances in parallel and distributed computing and its applications

    Furthermore, parallel and distributed computing has emerged in recent advances of many hotspot research directions including artificial intelligence, machine learning, Internet of Things, bioinformatics, digital medicine, cybersecurity, and social computing, resulting in numerous ground-breading discoveries that are changing our society and life.

  6. 130073 PDFs

    Explore the latest full-text research PDFs, articles, conference papers, preprints and more on DISTRIBUTED COMPUTING. Find methods information, sources, references or conduct a literature review ...

  7. Home

    Overview. Distributed Computing is a peer-reviewed journal that serves as a forum for significant contributions to the theory and practical aspects of distributed systems. Covers topics from design and analysis of distributed algorithms to architectures and protocols for communication networks. Includes discussions on synchronization protocols ...

  8. Exploring the Potential of Distributed Computing Continuum Systems

    Computing paradigms have evolved significantly in recent decades, moving from large room-sized resources (processors and memory) to incredibly small computing nodes. Recently, the power of computing has attracted almost all current application fields. Currently, distributed computing continuum systems (DCCSs) are unleashing the era of a computing paradigm that unifies various computing ...

  9. PDF The evolution of distributed computing systems: from fundamental to new

    The objective of this work is to study and evaluate the key factors that have influenced and driven the evolution of distributed system para-digms, from early mainframes, inception of the global inter-network, and to present contemporary systems such as edge computing, Fog computing and IoT. Our analy-sis highlights assumptions that have driven ...

  10. A survey on the Distributed Computing stack

    In this paper, we review the background and the state of the art of the Distributed Computing software stack. We aim to provide the readers with a comprehensive overview of this area by supplying a detailed big-picture of the latest technologies. First, we introduce the general background of Distributed Computing and propose a layered top ...

  11. Distributed Computing

    Stay informed on the latest trending ML papers with code, research developments, libraries, methods, and datasets. ... Distributed computing offers a high degree of flexibility to accommodate modern learning constraints and the ever increasing size of datasets involved in massive data issues. 1.

  12. PDF Distributed Computing With the Cloud

    single vector from any processing node (and hence computing the sum) requires Ω(n/ √ n) = Ω(√ n) rounds - due to the limited bandwidth to the cloud. But using both cloud links and local links (Fig. 1, right), the sum can be computed in Θ(˜ 4 √ n) rounds, as we show in this paper.

  13. The Analysis of Distributed Computing Systems with Machine Learning

    The distributed computing system is a hot research field. Deep learning, the Internet of Things, and other technologies are rapidly advancing, necessitating better levels of computation, storage, and communication efficiency for many jobs and applications. The advanced distributed computing model, represented by cloud computing and edge computing, has attracted the attention of researchers ...

  14. PDF Future Directions for Parallel and Distributed Computing

    However, the field of Parallel and Distributed Computing faces a number of existing and emerging challenges that must be overcome in order to meet the rapidly increasing computational requirements of new and larger highly distributed applications. Existing parallel and distributed computing abstractions impose substantial complexity on the

  15. Distributed Systems and Parallel Computing

    From our company's beginning, Google has had to deal with both issues in our pursuit of organizing the world's information and making it universally accessible and useful. We continue to face many exciting distributed systems and parallel computing challenges in areas such as concurrency control, fault tolerance, algorithmic efficiency, and ...

  16. Algorithms

    Dear Colleagues, We invite you to submit your latest research on the broad area of distributed computing to this Special Issue, "Distributed Computing Theory, Systems, Algorithms, and Data Structures". The goal of this Special Issue is to improve understanding of the principles and practices underlying distributed computing.

  17. On Distributed Computing Continuum Systems

    This article presents our vision on the need of developing new managing technologies to harness distributed "computing continuum" systems. These systems are concurrently executed in multiple computing tiers: Cloud, Fog, Edge and IoT. This simple idea develops manifold challenges due to the inherent complexity inherited from the underlying infrastructures of these systems. This makes ...

  18. Introductory Chapter: The Newest Research in Parallel and Distributed

    With the advent of networking and computer technology, parallel and distributed computing systems have been widely employed for solving problems in engineering, management, natural sciences and social sciences. There are six chapters in this book. From Chapters 2 to 6, a wide range of studies in new applications, algorithms, architectures ...

  19. Fiber optic computing using distributed feedback

    In this work, we introduce a temporally multiplexed optical computing platform that performs convolutions using a simple and scalable approach based on distributed feedback in single mode fiber ...

  20. Next generation cloud computing: New trends and research directions

    Next generation cloud computing systems are aimed at becoming more ambient, pervasive and ubiquitous given the emerging trends of distributed, heterogeneous and ad hoc cloud infrastructure and associated computing architectures. This will impact at least the following four areas considered in this paper.

  21. (PDF) Distributed Computing: An Overview

    The differences in calculation distribution and parallel computing, along with terminology, assignment of tasks, performance parameters, benefits and range of distributed computing, and parallel ...

  22. A Research Review of Distributed Computing System

    The features of distributed computing system include the following four aspects [ 9 ]: (1) Fault tolerance: the system has the ability to tolerate stand-alone failures; i.e., it has fault-tolerant mechanism; (2) Independence: the working nodes do not affect each other, and only know part of the input data. (3)

  23. Introduction—Parallel and Distributed Computing

    THE research domains of parallel and distributed computing have a significant overlap. With the advent of general-purpose multiprocessors, this overlap is bound to increase. This Special Issue attempts to draw together several papers from both of these separate research domains to illustrate commonalty and to encourage greater interaction among researchers in the two communities.

  24. Intelligent Computing: The Latest Advances, Challenges, and Future

    Computing is a critical driving force in the development of human civilization. In recent years, we have witnessed the emergence of intelligent computing, a new computing paradigm that is reshaping traditional computing and promoting digital revolution in the era of big data, artificial intelligence, and internet of things with new computing theories, architectures, methods, systems, and ...

  25. Research of PM-ADC scheme for optimized computational CFD analysis in

    This paper proposes a CFD scheme called the Parameterized Modeling and Automated Distributed Computing Approach (PM-ADC) to address the challenges. This approach includes a fuel assembly model tailored for the fast reactor core and incorporates parameterized geometric modeling coupled with a system to automate the distribution of parallel ...

  26. Frontiers

    This paper introduces DAC-HRC, a novel cognitive architecture designed to optimize human-robot collaboration (HRC) in industrial settings, particularly within the context of Industry 4.0. The architecture is grounded in the Distributed Adaptive Control theory and the principles of joint intentionality and interdependence, which are key to effective HRC. Joint intentionality refers to the ...

  27. A century of statistical Ecology

    Ecology is a leading journal publishing original research and synthesis papers on all aspects of ecology, with particular emphasis on cutting-edge research and new concepts. Abstract As data and computing power have surged in recent decades, statistical modeling has become an important tool for understanding ecological patterns and processes.