High-Dimensional Quantum Feature Mapping

High-Dimensional Quantum Feature Mapping

1 Quantum Principal Component Analysis

Quantum Principal Component Analysis qPCA identifies large eigenvalues of unknown density matrices utilizing corresponding eigenvectors in O(logd). Where principal component analysis analyzes positive semi-definite Hermitian matrices by decomposing eigenvectors in relation to the largest eigenvalues in the matrix for dimensionality reduction. Improved computational complexity will hopefully allow new methods for state discrimination and cluster assignment for variational and kernel classification machine learning algorithms.

Classical Principal Component Analysis PCA is greatly used in signal processing and machine learning for dimension reduction with a time complexity of O(N3), where N is the dimension of the data. The current issue with this form of dimension reduction is that when the data is large, the classical PCA ends up becoming non-tractable.

A possible solution for dimensional reduction of large data is utilizing Quantum computing parallelism, where the qPCA algorithm can then run with a time complexity of O(Nploy(logN)). The outputs of qPCA are the quantum states containing all eigenvalues and eigenvectors of the principal components of the data for sampling.

Consider a dataset {xi}i=1N with N data points where each data point is a D-dimensional column vector xi= (xi1,xi2,,xiD)TRD. Through principal component analysis we can lower the dimensional space to a N×D matrix X=(x1,,xN)T while maximally preserving data variance. Where an eigendecomposition for a single matrix X is:

X=j=1Dσj|ujvj|

such that {σjR0}j=1D are the singular eigenvalues of the matrix in descending order and {|ujRN}j=1D and {|vjRD}j=1D are, respectively, the left and right singular eigenvectors or principal components. For the rest of this section, we will look at why we need qPCA, how we can achieve it, and what is the procedure for the underlying algorithm.

1.1 High-Dimensional Quantum Data Reduction Algorithm

Here, we will analyze matrix dimensionality reduction using qPCA. Let ξ be the precision parameter and p denote variance retained after dimensionality reduction. Given efficient quantum access to the matrix A=UΣVT = irσiuiviTRn×m, let the top k right singular vectors V¯(k)Rm×k, where V(k)V¯(k) ξp2. There exists a quantum algorithm that creates |Y¯ = 1YFinyi,|i|yi, proportional to the projection of A in the PCA subspace. Where the algorithmic lower-bound probability 11/poly(m) and with error |Y|Y¯ ξ with the time complexity O~(1/p). If, instead, we estimate the value of Y¯F to a realtive error of η then the time complexity becomes O~(1pη)

It is important to note what data is PCA-representable and qPCA-representable. First, we begin by defining PCA-representable data. Let a set of n data points be defined m coordinates and represented by the matrix A, which was given earlier as irσiuiviTRn×m. Matrix or data A is PCA-representable if there exists p[12,1], ε[0,1/2], β[pε,p+ε], and α[0,1] such that:

  • kO(1) where the variance retained after dimensionality reduction p=ikσi2imσi2.
  • For at least αn points ai, it is maintained that yiaiβ, where yi = ik|aivj|2ai.

Note, this allows for algorithmic lower-bound probability of 11/poly(m) we saw before.

Let's consider qPCA-representable case. Given ai is a row of ARn×d, the time complexity is aiy¯i = 1β = O(1) with the a probability greater than α.

Now that we have defined the parameters, data assumptions, and goal for our qPCA algorithm, let's walk through the algorithm itself. The abstract concept is that the algorithm transforms the quantum state |ψs that holds the original data in quantum parallel to another quantum state |ψe, that will store the new low-dimensional data points:

|ψs:=i=1N|ixi||X||F |ψe:=i=1N|iyiYF

such that, XF and YF are the Frobenius norms of X and Y. Let:

xi=j=1Dxij|j=||xi|||xi

yi=j=1dyij|j=||yi|||yi

For the implementation of the quantum algorithm, first consider the data set {xi}i=1N, or matrix X, is stored in a quantum random access memory qRAM, such that qRAM takes |i|0|0|i|ai||ai|. Next, using quantum access to the vectors and norms, we constuct the state i|ai||ei|ai, where the density matrix for the first register is exactly X. Now, we must create n=O(t2ϵ1) copies of X, in order to have an accuracy ϵ in time O(nlogd) by completing eiXt implementations of this process.

An important aspect to keep in mind about this algorithm is that the density matrix exponentiation is most effective when some of the eigenvalues are large. This is because we are utilizing quantum parallelism to amplify deviations in the probability amplitudes in order to reveal eigenvectors corresponding to the large eigenvalues of the unknown state. If all eigenvalues are of size O(1/d), then the time complexity increases to t=(d) to generate a transformation such that it rotates the input state σ to an orthogonal state.

1.1.1 Extracting Quantum Principal Components

Let's obtain the first d principal components in the quantum state form |v1,|v2...|vd, where the variance proportion is:

λj:=σj2j=1Dσj2, j=1,2,,D

From the singular value decomposition form of matrix X, we can rewrite |ψs as:

|ψs=j=1Dλj|uj|vj

The state of the second quantum register in QRAM is ρ=Tr1(|ψs)=j=1Dλj|vjvj|, equivalent to XTX/Tr(XTX).

To continue.

2 Efficient Discrete Feature Encoding

To continue.

Cloud A.I.

Hi, I'm a Cloud A.I. coded by Alejandro, the creator of this website. Ask me things like, "What is Hilbert Space", "Go to the math page", or "Turn on dark mode". For more ask me, "What can you do?".