Quantum Neural Network

Quantum neural networks are computational neural network models which are based on the principles of quantum mechanics.

The first ideas on quantum neural computation were published independently in 1995 by Subhash Kak and Ron Chrisley, engaging with the theory of quantum mind, which posits that quantum effects play a role in cognitive function. However, typical research in quantum neural networks involves combining classical artificial neural network models (which are widely used in machine learning for the important task of pattern recognition) with the advantages of quantum information in order to develop more efficient algorithms. One important motivation for these investigations is the difficulty to train classical neural networks, especially in big data applications. The hope is that features of quantum computing such as quantum parallelism or the effects of interference and entanglement can be used as resources. Since the technological implementation of a quantum computer is still in a premature stage, such quantum neural network models are mostly theoretical proposals that await their full implementation in physical experiments.

Quantum Neural Network
Sample model of a feed forward neural network. For a deep learning network, increase the number of hidden layers.

Most Quantum neural networks are developed as feed-forward networks. Similar to their classical counterparts, this structure intakes input from one layer of qubits, and passes that input onto another layer of qubits. This layer of qubits evaluates this information and passes on the output to the next layer. Eventually the path leads to the final layer of qubits. The layers do not have to be of the same width, meaning they don't have to have the same number of qubits as the layer before or after it. This structure is trained on which path to take similar to classical artificial neural networks. This is discussed in a lower section. Quantum neural networks refer to three different categories: Quantum computer with classical data, classical computer with quantum data, and quantum computer with quantum data.

Examples

Quantum neural network research is still in its infancy, and a conglomeration of proposals and ideas of varying scope and mathematical rigor have been put forward. Most of them are based on the idea of replacing classical binary or McCulloch-Pitts neurons with a qubit (which can be called a “quron”), resulting in neural units that can be in a superposition of the state ‘firing’ and ‘resting’.

Quantum perceptrons

A lot of proposals attempt to find a quantum equivalent for the perceptron unit from which neural nets are constructed. A problem is that nonlinear activation functions do not immediately correspond to the mathematical structure of quantum theory, since a quantum evolution is described by linear operations and leads to probabilistic observation. Ideas to imitate the perceptron activation function with a quantum mechanical formalism reach from special measurements to postulating non-linear quantum operators (a mathematical framework that is disputed). A direct implementation of the activation function using the circuit-based model of quantum computation has recently been proposed by Schuld, Sinayskiy and Petruccione based on the quantum phase estimation algorithm.

Quantum networks

At a larger scale, researchers have attempted to generalize neural networks to the quantum setting. One way of constructing a quantum neuron is to first generalise classical neurons and then generalising them further to make unitary gates. Interactions between neurons can be controlled quantumly, with unitary gates, or classically, via measurement of the network states. This high-level theoretical technique can be applied broadly, by taking different types of networks and different implementations of quantum neurons, such as photonically implemented neurons and quantum reservoir processor (quantum version of reservoir computing). Most learning algorithms follow the classical model of training an artificial neural network to learn the input-output function of a given training set and use classical feedback loops to update parameters of the quantum system until they converge to an optimal configuration. Learning as a parameter optimisation problem has also been approached by adiabatic models of quantum computing.

Quantum neural networks can be applied to algorithmic design: given qubits with tunable mutual interactions, one can attempt to learn interactions following the classical backpropagation rule from a training set of desired input-output relations, taken to be the desired output algorithm's behavior. The quantum network thus ‘learns’ an algorithm.

Quantum associative memory

The first quantum associative memory algorithm was introduced by Dan Ventura and Tony Martinez in 1999. The authors do not attempt to translate the structure of artificial neural network models into quantum theory, but propose an algorithm for a circuit-based quantum computer that simulates associative memory. The memory states (in Hopfield neural networks saved in the weights of the neural connections) are written into a superposition, and a Grover-like quantum search algorithm retrieves the memory state closest to a given input. As such, this is not a fully content-addressable memory, since only incomplete patterns can be retrieved.

The first truly content-addressable quantum memory, which can retrieve patterns also from corrupted inputs, was proposed by Carlo A. Trugenberger. Both memories can store an exponential (in terms of n qubits) number of patterns but can be used only once due to the no-cloning theorem and their destruction upon measurement.

Trugenberger, however, has shown that his proababilistic model of quantum associative memory can be efficiently implemented and re-used multiples times for any polynomial number of stored patterns, a large advantage with respect to classical associative memories.

Classical neural networks inspired by quantum theory

A substantial amount of interest has been given to a “quantum-inspired” model that uses ideas from quantum theory to implement a neural network based on fuzzy logic.

Training

Quantum Neural Networks can be theoretically trained similarly to training classical/artificial neural networks. A key difference lies in communication between the layers of a neural networks. For classical neural networks, at the end of a given operation, the current perceptron copies its output to the next layer of perceptron(s) in the network. However, in a quantum neural network, where each perceptron is a qubit, this would violate the no-cloning theorem. A proposed generalized solution to this is to replace the classical fan-out method with an arbitrary unitary that spreads out, but does not copy, the output of one qubit to the next layer of qubits. Using this fan-out Unitary (Quantum Neural Network ) with a dummy state qubit in a known state (Ex. Quantum Neural Network  in the computational basis), also known as an Ancilla bit, the information from the qubit can be transferred to the next layer of qubits. This process adheres to the quantum operation requirement of reversibility.

Using this quantum feed-forward network, deep neural networks can be executed and trained efficiently. A deep neural network is essentially a network with many hidden-layers, as seen in the sample model neural network above. Since the Quantum neural network being discussed uses fan-out Unitary operators, and each operator only acts on its respective input, only two layers are used at any given time. In other words, no Unitary operator is acting on the entire network at any given time, meaning the number of qubits required for a given step depends on the number of inputs in a given layer. Since Quantum Computers are notorious for their ability to run multiple iterations in a short period of time, the efficiency of a quantum neural network is solely dependent on the number of qubits in any given layer, and not on the depth of the network.

Cost functions

To determine the effectiveness of a neural network, a cost function is used, which essentially measures the proximity of the network's output to the expected or desired output. In a Classical Neural Network, the weights (Quantum Neural Network ) and biases (Quantum Neural Network ) at each step determine the outcome of the cost function Quantum Neural Network . When training a Classical Neural network, the weights and biases are adjusted after each iteration, and given equation 1 below, where Quantum Neural Network  is the desired output and Quantum Neural Network  is the actual output, the cost function is optimized when Quantum Neural Network = 0. For a quantum neural network, the cost function is determined by measuring the fidelity of the outcome state (Quantum Neural Network ) with the desired outcome state (Quantum Neural Network ), seen in Equation 2 below. In this case, the Unitary operators are adjusted after each iteration, and the cost function is optimized when C = 1.

Equation 1 Quantum Neural Network  
Equation 2 Quantum Neural Network  

See also

References

Tags:

Quantum Neural Network ExamplesQuantum Neural Network TrainingQuantum Neural NetworkArtificial neural networkBig dataNeural network (machine learning)Quantum computingQuantum entanglementQuantum informationQuantum interferenceQuantum mechanicsQuantum mindQuantum parallelismSubhash Kak

🔥 Trending searches on Wiki English:

List of largest citiesRyan GarciaTyler HerroMurder of Lauren GiddingsMS DhoniAnyone but You2024 Indian general election in BiharDark web2024–25 UEFA Champions League2024 in filmThe SimpsonsMin Hee-jinTurks and Caicos Islands2024 World Snooker ChampionshipDune (2021 film)Planet of the ApesRajiv Gandhi International Cricket StadiumM. Night Shyamalan2022 NFL draftMichael DreebenKaty PerryKurt RussellSimon CowellGoogleDhruv RatheeSwapnil SinghFallout (video game)AustraliaIndian Super LeagueHybe CorporationRed Eye (British TV series)Harvey WeinsteinRule 34Jerry SeinfeldSam PitrodaFeyenoordMillie Bobby Brown2024 NFL draft2024 NBA playoffsAparna DasPhilippinesDonald TrumpLady GagaReal Madrid CFStephen CurryNelson MandelaNikola JokićElon MuskRussian invasion of UkraineLa LigaBob MarleyCaitlin ClarkOrlando BloomEarthFallout 4Czech RepublicTaylor Swift albums discographyLok SabhaMeta PlatformsHenry VIIIAshley JuddList of prime ministers of IndiaAndrew Scott (actor)Shah Rukh KhanSurvivor 46BrooklynOttoman EmpireMichael AvenattiQueen VictoriaSandra OhRusso-Ukrainian WarBhimaa2024 Indian general electionSex positionAshlyn HarrisWinston ChurchillSachin TendulkarCanelo ÁlvarezBaldwin IV of Jerusalem🡆 More