Machine Learning Algorithms In Quantum Computing: A Survey

1y ago
4 Views
4 Downloads
914.01 KB
8 Pages
Last View : 1m ago
Last Download : 3m ago
Upload by : Francisco Tran
Transcription

Machine Learning Algorithms in QuantumComputing: A SurveySomayeh Bakhtiari RamezaniComputer Science and Engineering Dept.Mississippi State UniversityStarkville, USsb3182@msstate.eduAlexander SommersComputer Science and Engineering Dept.Mississippi State UniversityStarkville, USams1988@msstate.eduShahram RahimiComputer Science and Engineering Dept.Mississippi State UniversityStarkville, USsr2002@msstate.eduAmin AmirlatifiDave C. Swalm School of Chemical Eng.Mississippi State UniversityStarkville, USaa2340@msstate.eduAbstract —Machine Learning (ML) aims at designing modelsthat learn from previous experience, without being explicitlyformulated. Applications of machine learning are inexhaustible,including recognizing patterns, predicting future trends andmaking decisions, and they are capable of handling sizablequantities of multi-dimensional data in the form of large vectorsand tensors. To perform these operations on classical computers,however, requires vast time and computational resources. Unlikethe classical computers that rely on computations using binarybits, Quantum Computers (QC) benefit from qubits which canhold combinations of 0 and 1 at the same time via superpositionand entanglement. This makes QCs powerful at handling and postprocessing large tensors, making them a prime target forimplementing ML algorithms. While several models used for MLon QCs are based on concepts from their classical computingcounterparts, utilization of the QC’s potential has made them thesuperior of the two. This paper presents an overview of the currentstate of knowledge in application of ML on QC, and evaluates thespeed up, and complexity advantages of using quantum antumis an abstract concept of existence, or lack of, connectivitybetween two elements in the processing unit, or magneticorientation of memory sectors.Similar to the classical computers, the quantum computersuse a “quantum bit” or “qubit” [5] which is the probability of“spin up” and “spin down” of an electron after going through amagnetic field. The spin can be thought as an equivalent of thevalue of a bit in classical computing.A. Quantum ComputationQuantum mechanics deals with an infinite dimensionalHilbert space ( ℋ ) and as a result, requires an infinitelydimensional vector notation. The state of qubit is shown by atwo dimensional vector in Hilbert space [5], using Dirac’s BraKet notation, which was created by Paul Dirac in 1939 [6]. Inthis notation, 1 (read “Ket one”) is used for the qubit “on” or“spin up” state, and 0 (read “Ket zero”) is used for qubit “off”or “spin down”.Computing,INTRODUCTION AND BACKGROUNDThe concept of Quantum Computation (QC) was originallypresented in 1982 by Richard Feynman [1] following anobservation of the exponential complexity involved in modelingthe behavior of a quantum system with the existing knowledgeof classical computing. Unlike classical computers where thestate of any unit of computation is deterministically prescribedto be either a zero or one, the state of the unit of computation ina quantum computer could be zero, one or anything in between.This unique feature of QC, results in their ability to pursueseveral parallel paths simultaneously in a single calculation unit,a concept that is not physically possible in classical machines,and would rather require several passes to be achieved, hencecausing higher orders of complexity [2]–[4].In classical computing, data storage, handling, andcalculations are established on a binary basis called “bit”, which978-1-7281-6926-2/20/ 31.00 2020 IEEEHarish Kumar ManchukondaComputer Science and Engineering Dept.Mississippi State UniversityStarkville, UShm1089@msstate.edu 0 10, 1 01(1)1As an example, Ket zeroshows that the qubit is at a0spin down state. Here, the first element represents theprobability of spin down, and the second element shows theprobability of spin up.In quantum physics the notation is used to denote thearbitrary state of qubit through superposition. In superposedqubit, the state of a qubit is the combination of being 0 and atthe same time being 1. 0 1, ℂ(2)where 1 . The complex numbers , areamplitudes of the basic states of 1 and 0 [7]. The continuouslinear combination of these two states could be one of anypossible points on a Bloch sphere, presented in Figure 1.

Additionally, in Dirac’s notation, the Bra is used for innerproduct operations on two vectors or qubits, which is aHermitian conjugate of a Ket. For example, (read bra of )is a complex conjugate vector of Ket , where the sign of everyimaginary part is changed, and every Ket is changed into brasand vice versa. Similarly, the inner product of these two vectors ( ),(5)Such an operation suggests that by applying the innerproduct operator matrix on the state matrix that is comprising ofall input qubits, the operation is applied to the whole system inone step; this is known as quantum parallelism [11]. Sincemeasurement of an unknown state of the system will result in acollapse of the values, the challenge of using quantumparallelism lays in taking advantage of parallelism beforemeasuring the system [13].One of the strengths of quantum computers lays in theirability to process large amounts of high dimensional vectors inpolynomial order [3], which makes them a prime target formachine learning applications. In the present paper, we willinvestigate current studies on machine learning algorithms inquantum computing. is a scaler [8].Fig 1. Qubit sate in the Bloch Sphere [9]Quantum mechanics laws dictate that measurement of thevalue of a qubit will collapse the arbitrary state of the qubitto one of the two ground states of 0 or 1 [9]. The probabilityof the final state can be calculated by equation 3 [10]:() , ,(3),(4)wherePr( ) ,In equation 3, which is generalized for multi qubit systems,Pr( ) is the probability of the system, as a whole, to be in the state, once the measurement is performed. In other words, afterthe measurement, the state of all qubits will change to 0 or 1 .As an example, in a two-qubit system, the final state could be 00 , 01 , 10 or 11 .In a multi qubit system, when all qubits are at superposition,the value of one qubit could be connected and intertwined to thevalue of another qubit; in this case, the change or measurementof one qubit will reveal the value of the other qubit; additionally,the square magnitude of probability of all qubits together will be1. This concept is called entanglement [11]. Regardless of thephysical distance of the two entangled quantum systems, achange in either quantum system, can change the other quantumsystem, even far from the first one [12]. As an example, in anentangled two-qubit system, called the Bell state, uponmeasurement of the first qubit, the second qubit collapses to thestate of the first qubit, thus the final state could be either 00 or 11 [10].An operator U is defined as a matrix [10], such that it canapply function ( ) to the system as a whole. Assuming that thequbits in the system are entangled, the computed result ofapplying Uf on all qubits of is expressed as [13]:B. Grover’s AlgorithmA key component of almost all NP-complete algorithms issearching for a particular key or entity in the input set, and lackof an efficient search algorithm leads to their drearyperformance. Grover algorithm is a quantum specific algorithmthat quadratically increases the search performance in arandomly ordered dataset, thus surpassing any classicalcounterparts. While classical algorithms can take 0.5N or worseto find the match in a dataset with N members, the Grover’salgorithm can accomplish the same task on a quantum computerin ( ) steps [14].C. Quantum Machine LearningMachine Learning (ML) tasks often involve analysis andclassification of large number of vectors in multi-dimensionalspaces and the trend in ML, already a set of practices which relyon a sufficient abundance and expressiveness (read complexity)of data is to apply it to "big data". Quantum Machine Learning(QML) is a growing field that brings quantum computing andmachine learning together. The main premise of QML is to usethe inherent advantages of quantum computing (includingsuperposition, entanglement, and quantum parallelism), toimprove the performance of classical machine learningalgorithms.Two of the biggest challenges that QC faces, are limitednumber of qubits available, and the high level of noise that isexperienced when using them. Quantum computers need toexpand in terms of available qubits, and fault tolerance;however, it is anticipated that the Noisy Intermediate-ScaleQuantum (NISQ) [15] computers that will become available innear future, are capable of performing computations thatshowcase the inherent capabilities of quantum computers.In section II we briefly explain the current algorithms thatare proposed for NISQ era quantum computers. Majority of theproposed algorithms that can be run on near-term devices areheuristic, i.e. we do not know their performance beforehand andcannot prove their speed up. Algorithms that have provableperformance advantages over classical computers require vastlylarger, high fidelity and fault tolerant quantum computers.Section III offers a comparison between different quantummachine learning algorithms, in terms of their relative speedup,use of Grover algorithm, and existence of an implementation.

II.QUANTUM MACHINE LEARNING ALGORITHMSMajority of quantum machine learning algorithms are basedon the concepts and architectures borrowed from classicalalgorithms, while utilizing quantum physical concepts such asentanglement and supper position. In some cases, hybridsystems which are combinations of classical and quantumcomputers, are used to overcome limitations in the NISQ eraquantum computers. Examples include, the use of QuantumApproximate Optimization Algorithm (QAOA) [16] whereparameters from classical solution to a problem are used as theinitial guess and starting point for larger problems, or variationalquantum eigen solvers [17], [18] . For the purpose ofclassification, this paper looks at the hierarchy of quantummachine learning algorithms as Supervised, Semi-Supervised,and Un-supervised.A. Supervised AlgorithmsQuantum Neural Network (QNN) is the prime example ofsupervised quantum algorithm. As an early approach in thisfield, Narayanan and Menneer [19], showed the theoretical formof a QNN architecture and how the components of such a systemwould perform compared to classical counter parts. Althoughtheir work is high level description of the components, needed,it laid the foundation for future implementations of QNN.Artificial Neural Networks (ANNs) are based onaggregation of non-linear functions applied to neurons that arelaid out in layers and sequences. The linear nature of quantummechanics, however, makes the implementation of non-linearactivation functions challenging. Cao et al. [20] proposed aconcept for the building block of QNNs. They have presented aquantum neuron concept, based on a quantum circuit thatnatively simulates threshold activation of neurons, and canreplicate the response from a wide range of ANN settings. Theirproposed model honors inherent quantum advantages, i.e.superpositions of inputs, quantum coherence and entanglement,and can be used to construct several classical networkarrangements, including supervised network, unsupervisednetwork and reinforcement learning.Schuld et al. [21] have presented a general procedure formodeling quantum perceptron that simulates the step-activationfunction, which is a key component of ANNs. Their approachneeds O(n) qubits, with n being the size of the input, and givingan efficiency that is equivalent to classical models.Implementation of such step activation functions is the key toneural networks that can be trained on quantum computers,which can potentially benefit from superposition-basedlearning, thus training in quantum parallel space.Ricks and Ventura [22] proposed an algorithm for trainingquantum based neural networks, capable of producing highaccuracy solutions. Their approach uses a minimum set ofentangled qubits to cover the training data set and can returnresults that encompass a predetermined portion of the trainingdata. The downside of their approach is that complexity willincrease exponentially, as the problem size increases, leading toextremely complicated scenarios, regardless of the reduction incomplexity that is offered through the use of quantum machines.As a possible solution to this problem, they have proposed amodified random version of the algorithm that is less complex,unlike the full model that used composite weight vector for thewhole network. Complexity reduction of the randomizedalgorithm makes it suitable for large problems with many nodesand connections between them.Panella and Martinelli [23] proposed a technique that usesBoolean functions and proper implementation of nonlinearquantum circuits to transform non-linear data into a linear form.Nonlinear quantum operators are used to overcome the lack ofexhaustive search in the network. Their approach had thepotential of overcoming the aforementioned hurdles faced inQNN. Similar to other implementations, however, theirtechnique included evaluating every possible configuration tofind the optimized network configuration, but the mainadvantage of their technique was the combination of linear andnonlinear components in the implementation of quantum circuit.The linear nature of unitary input-output gates, counteractsthe efficiency of QC machines in handling linear inputs [24]–[26]. Panella and Martinelli have also presented a model forQuantum Neuro-Fuzzy Networks (QNFN) based on BinaryNeuro-Fuzzy Networks (BNFN) with a tailored binarymembership function that benefits from inherent quantumcomputing features [25]. Their model employs a quantumOracle to get superposition as a way of ensuring parallelism, andto mark the optimal solutions with a dedicated qubit. Likewise,entanglement is used to tie the superposed solutions to theirrespective network performances. The exponential speedupgained by the nonlinear nature of quantum algorithm makesQNFN capable of performing an exhaustive search for theoptimal solution from a staggering number of different QNFNsin superposition, that is not feasible in classical models. Such anapproach is optimized where parameter values are considered,but it is not optimal from a complexity or number of rulesperspective.Similar to classical computers, training and learning of aQNN relies on existence of a quantum memory. Achieving aQuantum Random Access Memory (QRAM) that is independentof the classical counter parts has been an intersection of quantumcomputing and neural networks.Ventura and Martinez [27] introduced a quantum associativememory neural network architecture through wave functionsand operators, that compared to a classical Hopfield network,would exponentially improve storage capacity. This techniqueuses patterns as quantum operators applied on two state (halfspin) quantum systems but lacks efficiency in recalling patterns.Trugenberger [28] presented a model for quantummemories, where capacity increased exponentially by thenumber of qubits and used a probabilistic approach for memoryretrieval. He later presented [29] a model for quantumassociative memory to overcome the capacity shortage ofassociative memory in classical models, by benefiting from thefact that quantum memory, free from spurious memories. Heused an n-qubit quantum superposition state to store n-bit binarypatterns, but information retrieval in this model was not exactand relied on a probabilistic approach of finding a pattern in thequantum state memory that most closely resembled the input.Bisio, et al. [30], showed an implementation of the state of aquantum memory through learning to store and retrieve anunknown unitary transformation value from a number of

training examples. Similarly, [31] used unitary transformation toestablish quantum coherence between various instances of aquantum system and conduct quantum Principal ComponentAnalysis (PCA), leading to exponential speed up indetermination of eigenvectors of large systems.Oliveira [32] introduced a quantized equivalent for RAMbased Neural Networks (RbNNs) that was debuted as quantumRAM based Neural Networks (q-RbNN). The main advantageof q-RbNNs is their ability to use the classical learningalgorithms, while benefiting from the physical advantages ofquantum-based machines. Silva, et. al [33]–[35] later expandedthis idea and presented a model for neural networks based onweightless neural nodes, to serve as a Quantum Random AccessMemory (QRAM) that concurrently uses the ensemble oftraining patterns in superposition. Through modification ofGrover search algorithm [36] and applying it on a probabilisticquantum memory, they have been able to run forward andbackward propagation schemes.Silva, et. al [13], presented a supervised learning algorithmfor QNN that is optimized for quantum learning and is alsocapable of operating on classical models. Their approachincludes a quantum Weightless Neural Network (qWNN) basedon the quantization of the QRAM, capable of receiving alltraining set patterns concurrently in superposition. They havealso presented a Superposition-based Learning Algorithm(SLA) for supervised learning. The SLA determines capabilityof the QRAM to accept qubits while in superposition mode. Italso employs a probabilistic approach in determining theoptimal configuration for the network, out of a cohort ofconfigurations that are received in superposition.Kerenidis et al. [37] have offered a shallow circuit algorithmfor deep Quantum Convolutional Neural Networks (QCNN).Their proposed algorithm is capable of handling non linearitiesand pooling operations, thus mimicking the behavior ofclassical CNN, with potential for offering more sophisticatedkernels, and handling larger or deeper inputs. A unique featureof their approach lies in a new quantum tomography where themost significant data is extracted with higher likelihood, thusdecreasing the complexity of the system.Similarly, Cong et al. [38] have proposed a QNNarchitecture for QCNN that uses O(log(N)) on NISQ devices. Inaddition to multi-scale entanglement renormalization ansatz,their approach uses quantum error correction to identifyquantum states for one dimensional topological phases that areprotected by the symmetry. They used sample complexity todemonstrate the performance of their Quantum PhaseRecognition (QPR), over String Order Parameters (SOP), whereQCNN is shown to outperform conventional methods byrequiring much less repetitions. Additionally, they demonstratea quantum error correction system based on QCNN that isoptimized for a particular error model. It is shown that theirapproach for concurrent optimization of encoding and decodingprocedures, results in a scheme that is superior over otherquantum codes.Kerenidis and Luongo [39] presented a quantum classifierbased on quantum equivalent of Slow Feature Analysis (QSFA)for dimensionality reduction, and Quantum Frobenius Distance(QFD) as an algorithm for classification. Their simulatedanalysis on MNIST dataset shows that such a classifier wouldbe 98.5% accurate in differentiating handwritten digits. Theirproposed dimensionality reduction algorithm has the potentialof direct application on other algorithms that produce quantumdata, thus eliminating the need for QRAM, and has apolylogarithmic execution order in the dimension and numberof inputs. While this approach is not necessarily a faster way ofclassifying data, it is assumed to be more accurate, and cansimplify implementation of quantum-based techniques, byeliminating the need for QRAM.Recommender systems use ratings given by a group of users(M) to several products purchased in the past (N), torecommend new products that may align well with preferencesof a specific user. Such systems can be considered as supervisedalgorithms, if they provide new suggestions. In these systems,it is assumed that users belong to one of K user types, where Kis the rank of matrix. Kerenidis and Prakash [40] haveproposed an algorithm for recommender systems thateliminates the need to access all information at once, and justfocuses on sampling parts of the matrix that are relevant to therow of interest. This has been done through mapping of a vectorto the row space of the matrix. This approach was exponentiallyfaster than known classical algorithms at the time, with an orderof O(poly(K)polylog(M.N)). This algorithm, however, inspireda new classical algorithm that is much faster than previousclassical algorithms [41] and is only polynomially slower thanthe Kerenidis and Prakash quantum algorithm.Improvements in Support Vector Machines (SVMs), whichare used for regression and pattern recognition, have led them tobeing used in situations where Quadratic-Programming (QP),which allowed for the application of algorithms to training theSVMs [42], is no longer possible. These QP-algorithms enabledefficient training ([42], [43]) but advances in the use ofRademacher estimates to reduce the complexity of the modelsdisallow the use of quadratic programming optimization and thiscan turn optimization into a task with NP-complete complexity[42]. Anguita, et. al. [42] explored the possibility of using QCbased optimization in these cases and compared theirperformance to those of a Monte Carlo classical method. Theauthors point out proof that QC can break the NP-completenessbarrier is not given yet; still, benefits are evident when applyingQC, especially as the problem complexity increases, presumablybecause conventional computing becomes less useful due to theimpossibility of its runtime being acceptable.The premise of quantum computing, with its ability to solvecomplex problems more rapidly, could solve the complexoptimization problem of SVM, more efficiently. As an examplebacking this claim, Kerenidis, et al. [44], have used InteriorPoint Method (IPM) method for Second Order ConicProgramming (SOCP) to train SVM binary classifier on realworld data and have shown that the algorithm works moreeasily, than in theory. Their results for the performancecomparison between traditional IPM for SOCPs and quantumIPM for SOCPs show that the quantized instance convergedaround the same iteration, but the runtime of quantum SOCPsare much better than traditional IPM.Farhi and Gutmann [45] presented a performancecomparison between classical decision tree traverse, and a

similar walk using quantum interference. They show thatquantum interference can result in exponential speed up intraversing a class of trees, rather than randomly walking throughthem, as is the case for the classical algorithm. They use a singletime dependent Hamiltonian to create a quantum state that isinitially centered around the root of a decision tree but evolvesthrough its nodes. The Hamiltonian calculation is further spedup by determining energy-dependent transmission coefficients.Building upon the PCA method presented in their earlierwork [31], Rebentrost, et al. [46], used non-sparse matrixsimulation to perform PCA and inversion of the training kernelmatrix, thus reaching exponential speed up with the size of inputvectors and number of available training examples for quantizedSVM.B. Unsupervised AlgorithmsDimensionality reduction and clustering algorithms areprimary examples of unsupervised quantum machine learningalgorithms. One use case of quantum clustering algorithms canbe in privacy enhancement, since the database holding thevectors to be clustered needs to be queried less in frequency andvolume by virtue of the lower number of calls needed by thequantum algorithms used. This exposes less of the database'sdata to the user of the algorithm. Classical algorithms takepolynomial time in vector number and dimension to solve theseproblems, while QML algorithms can take logarithmic time inboth, thus achieving exponential speed-up. The amount ofinformation O(log MN) that needs to be accessed for QMLoperations, compared to the size of information O(MN) beingapplied to classical machine learning, means that privacy of thatdata is also increased by the use of QML over classical ML.Aïmeur, et. al. [47], [48] have described three quantumalgorithms that could be substituted for components of classicalalgorithms and achieve exponential speedup in clustering, overclassical algorithms; their quantization presupposes existence ofa black-box quantum circuit which acts as a distance oracle,giving the distance between vector inputs. While such anassumption may not be valid, or at least readily available, theirsubroutines can be used, respectively: (1) To find the two mostdistant points from one another in a vector dataset, (2) To findthe n closest points to another, specified point, all in a vectordataset, (3) To produce neighborhood graphs of vector datasets,all in times superior to classical counterparts. With thesecapabilities, their proposed algorithms can be used, respectivelyfor quantizing: (1) Divisive clustering, (2) K-medians clustering,(3) Unsupervised learning algorithms. These subroutines arebased on Grover's algorithm and use Grover iterations to isolatedesired outputs from the results of computing with superpositioned inputs.Horn [49], [50] has proposed a general solution to theunsupervised learning problem of clustering using quantummechanics. Here, the space of the points is used to create a scalespace probability function, which in turn utilizes a Parzenwindow estimator (a means of estimating an unknowncontinuous function using a sample of its output [51]) to estimatethe probability distribution that would produce the witnesseddata. The minima of this potential function are then used todetermine cluster centers.Support Vector Clustering (SVC) correlates data-points withstates in Hilbert space. These states are represented by Gaussianwave functions and can allow for the weighting of certain pointsto give them more emphasis, presumably as cluster centerpossibilities. This is valuable if one has a method, such as SVC,which can be unduly influenced by outliers. With this additionalinformation computations could be weighed against beinginfluenced by these outlier points when determining clustercenters. The one controlled parameter proposed in this methodcontrols the width of the clusters that are being sought. Both itand the scale-space method search for cluster centers rather thanboundaries but the Horn’s method produces minima indicatingthe centers which are more defined (i.e. deep and robust) thanthe maxima the alternative approach produces. The complexityof this method is order N squared, independent ofdimensionality. Horn’s method, which easily identifies clustercenters, is part of a hybrid approach which would then work onthe other aspects of cluster identification.Lloyd, et al. [3] presented QML algorithms for assigningpoints to clusters and for finding clusters. They have presenteda quantum computer algorithm that can assign vectors of Nfeatures/dimensions to one of M clusters in O(log (M.N)) time,compared to the best classical algorithms run time ofO(poly(M.N)). In their approach for k-means clustering, aquantized version of Lloyd’s algorithm [52] is presented inO(k.log (k.M.N)) time where M is the number of vectors toclassify into k clusters and N is the number of features per vector(the dimensionality of the space all this is going on in). Wiebe,et al. [53], have presented algorithms for calculating nearestneighbors and k-means based on the Euclidean distance of thepoints, and amplitude estimation that precludes the need formeasurement. They show that use of these techniques can resultin polynomial reduction in complexity, compared to classicalMonte Carlo simulations.Similarly, Kerenidis, et al. [54], have presented a quantizedvariant of k-means algorithm (Q-means) which showscomparable performance and convergence to classical k-means,and uses k-means with Grover-type quadratic speed up. Suchalgorithms can provide substantial saving compared to theclassical algorithms particularly for the case of large datasets.Suzuki, et al. [55], have presented a kernel-based quantumclassifier where Pauli decomposition is used for evaluation andcreation of the feature map, using the real-value representations.They have presented a general formula that calculates the lowerlimit of the accuracy, which helps in finding the selected featuremap, and separating it linearly.C. Semi-Supervised AlgorithmsSemi-supervised quantum algorithms can be viewed asoptimization algorithms, quantum enhanced ReinforcementLearning (RL), and generative models.We can also consider QC for solving systems of linearequations or determination of Stochastic Gradient Descent(SGD), where gradient itself is affine, and doing so with lessdemand for quantum memory. Such algorithms, which arewidely used in ML, are examples of semi-supervised algorithms.To this end, Kerenidis and Prakash [56] have used the HHLalgorithm [57] to solve systems of linear equations.

Although researchers have examined ways that ArtificialIntelligence (AI) can benefit from Quantum InformationProcessing (QIP) in supervised and unsupervised learning ratherexhaustively, RL still remains as an area that needs to beexplored. Learning from experience is a defining signature of anintelligent agent, artificial or otherwise. The volume andcomplexity of information needed to describe real-life situationsin which AI might find itself, still often presents the agent withso much data to compute on, as to make the learning take toolong to be useful.To utilize superposition principle and quantum parallelism,Dong, et al. [58], proposed combining RL with quantum theory,and introduced Quantum Reinforcement Learning (QRL). Theyobserved that quantum parallelism and probability amplitudecan help to speed up the learning.Dunjk, et al. [59], have produced a schema which can

C. Quantum Machine Learning Machine Learning (ML) tasks often involve analysis and classification of large number of vectors in multi-dimensional spaces and the trend in ML, already a set of practices which rely on a sufficient abundance and expressiveness (read complexity) of data is to apply it to "big data". Quantum Machine Learning

Related Documents:

1.3.7 Example: quantum teleportation 26 1.4 Quantum algorithms 28 1.4.1 Classical computations on a quantum computer 29 1.4.2 Quantum parallelism 30 1.4.3 Deutsch's algorithm 32 1.4.4 The Deutsch-Jozsa algorithm 34 1.4.5 Quantum algorithms summarized 36 1.5 Experimental quantum information processing 42 1.5.1 The Stern-Gerlach experiment 43

Machine Learning CC CQ QC QQ QML Quantum algorithms feed with classical or quantum data - Supervised Learning - Unsupervised Learning - Reinforcement Learning Algorithm Classical Quantum Data m l Noisy Intermediate-Scale Quantum (NISQ) algorithms, K. Bharti, ACL, T.H. Kyaw, et. al., arXiv:2101.08448 (2021)

According to the quantum model, an electron can be given a name with the use of quantum numbers. Four types of quantum numbers are used in this; Principle quantum number, n Angular momentum quantum number, I Magnetic quantum number, m l Spin quantum number, m s The principle quantum

1. Quantum bits In quantum computing, a qubit or quantum bit is the basic unit of quantum information—the quantum version of the classical binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical system, one of the simplest quantum systems displaying the peculiarity of quantum mechanics.

mechanics. Rather than creating new quantum machine learning algorithms, let us now try to think if we can change only parts of existing classical machine learning algorithms to quantum ones. Machine learning and deep learning use linear algebra routines to manipulate and analyse data to learn from it.

the quantum operations which form basic building blocks of quantum circuits are known as quantum gates. Quantum algorithms typically describe a quantum circuit de ning the evolution of multiple qubits using basic quantum gates. Compiler Implications: This theoretical background guides the design of an e ective quantum compiler. Some of

Keywords—machine learning, quantum computing, deep learning, quantum machine intelligence I. INTRODUCTION There is currently considerable interest in quantum computing. In the U.S. in January 2019, the National Quantum Initiative Act authorized 1.2 billion investment over the next 5-10 years (Rep Smith, 2019). A quantum computing race is ongoing

Our study highlights the potential usefulness of RL for applications in out-of-equilibrium quantum physics. Jens Eisert Reinforcement learning decoders for fault-tolerant quantum computation and other perspectives of quantum machine learning Quantum machine learning comes in many facets: It either makes use of a methodology of machine