Skip to main content

Fast Quantum Algorithm for Learning

 -By Hayata Yamasaki,  Sathyawageeswar Subramanian,  Sho Sonoda,  Masato Koashi


Paper Link

image Courtesy: Fossguru.com


Abstract

Kernel methods augmented with random features give scalable algorithms for learning from big data. But it has been computationally hard to sample random features according to a probability distribution that is optimized for the data, so as to minimize the required number of features for achieving the learning to the desired accuracy. Here, we develop a quantum algorithm for sampling from this optimized distribution over features, in runtime O(D) that is linear in the dimension D of the input data. Our algorithm achieves an exponential speedup in D compared to any known classical algorithm for this sampling task. In contrast to existing quantum machine learning algorithms, our algorithm circumvents sparsity and low-rank assumptions and thus has wide applicability. We also show that the sampled features can be combined with regression by stochastic gradient descent to achieve the learning without canceling out our exponential speedup. Our algorithm-based on sampling optimized random features leads to an accelerated framework for machine learning that takes advantage of quantum computers. 


Random features provide a powerful technique for scaling up kernel methods applicable to various machine learning tasks, such as ridge regression, kernel learning, and principle component analysis. Recently, Bach has shown an optimized probability distribution of random features, and sampling features from this optimized distribution would drastically improve runtime of learning algorithms based on random features. However, this sampling task has been computationally “hard in practice” due to inversion of a high-dimensional operator. In contrast, the power of quantum computers to process data in quantum superposition attracts growing attention towards accelerating learning tasks, opening a new field of quantum machine learning (QML). QML accelerates a supervised learning task, by constructing an efficient quantum algorithm for sampling from this optimized distribution.

Problem: 

These conventional methods for obtaining random features from the data-independent distribution dτ (v) require a large number M of features to learn the function f, which slows down the decision of all M features and the regression over M coefficients. Aiming at minimizing M required for the learning, rather than sampling from dτ (v), we will sample features from a probability distribution that puts greater weight on important features that are optimized for the data, via a probability density function q(v) for dτ (v). 


Solution

  •  (Theorem 1) We construct a new quantum algorithm for sampling an optimized random feature from q ∗  (v)dτ (v) in as fast as linear runtime O(D) in the data dimension D. The best existing classical algorithm by Bach for sampling each single feature from this data-optimized distribution q ∗  (v)dτ (v) requires exponential runtime O(exp(D)). In contrast, our quantum algorithm can sample each single feature from q ∗  (v)dτ (v) with runtime O(D), which is as fast as the conventional algorithms such as. These conventional algorithms perform an easier task, i.e. sampling from a data-independent distribution dτ (v). Advantageously over the conventional algorithms sampling from dτ (v), quantum algorithm sampling is used from q ∗  (v)dτ (v) to achieve learning with a significantly small number M of features, which is proven to be minimal up to a logarithmic gap. Remarkably, we achieve this without assuming sparsity or low rank of relevant operators.

  • To construct this quantum algorithm, we circumvent the difficulty of infinite dimension by formulating a discrete approximation of the problem of sampling a real-valued feature from q ∗  (v)dτ (v). This approximation is equivalent to using fixed-point number representation with rescaling.

  • (Theorem 2) The combined M features sampled by the algorithm with regression by stochastic gradient descent to achieve supervised learning in time O(MD), i.e. without canceling out our exponential speedup. This M is minimal up to a logarithmic gap, since we use optimized random features. Thus, by improving the computational bottleneck faced by classical algorithms for sampling optimized random features, we provide a promising framework of quantum machine learning that leverages our O(D) sampling algorithm to achieve the optimal M among all algorithms using random features.

Conclusion

 a quantum algorithm is constructed for sampling an optimized random feature within a linear time O(D) in data dimension D, achieving an exponential speedup in D compared to the existing classical algorithm of Bach for this sampling task. Combining M features sampled by this quantum algorithm with stochastic gradient descent, we can achieve supervised learning in time O(MD), where this M is expected to be nearly minimal since we use the optimized random features. As for future work, it is open to prove the hardness of sampling an optimized random feature for any possible classical algorithm under complexity-theoretical assumptions. It is also interesting to investigate whether we can reduce the runtime to O(M log D).  but using the optimized random features to achieve minimal M. Since our quantum algorithm does not impose sparsity or low-rank assumptions, our results open a route to a widely applicable framework of kernel-based quantum machine learning with an exponential speedup.



Comments

Popular posts from this blog

ABOD and its PyOD python module

Angle based detection By  Hans-Peter Kriegel, Matthias Schubert, Arthur Zimek  Ludwig-Maximilians-Universität München  Oettingenstr. 67, 80538 München, Germany Ref Link PyOD By  Yue Zhao   Zain Nasrullah   Department of Computer Science, University of Toronto, Toronto, ON M5S 2E4, Canada  Zheng Li jk  Northeastern University Toronto, Toronto, ON M5X 1E2, Canada I am combining two papers to summarize Anomaly detection. First one is Angle Based Outlier Detection (ABOD) and other one is python module that  uses ABOD along with over 20 other apis (PyOD) . This is third part in the series of Anomaly detection. First article exhibits survey that covered length and breadth of subject, Second article highlighted on data preparation and pre-processing.  Angle Based Outlier Detection. Angles are more stable than distances in high dimensional spaces for example the popularity of cosine-based similarity measures for text data. Object o is an out

TableSense: Spreadsheet Table Detection with Convolutional Neural Networks

 - By Haoyu Dong, Shijie Liu, Shi Han, Zhouyu Fu, Dongmei Zhang Microsoft Research, Beijing 100080, China. Beihang University, Beijing 100191, China Paper Link Abstract Spreadsheet table detection is the task of detecting all tables on a given sheet and locating their respective ranges. Automatic table detection is a key enabling technique and an initial step in spreadsheet data intelligence. However, the detection task is challenged by the diversity of table structures and table layouts on the spreadsheet. Considering the analogy between a cell matrix as spreadsheet and a pixel matrix as image, and encouraged by the successful application of Convolutional Neural Networks (CNN) in computer vision, we have developed TableSense, a novel end-to-end framework for spreadsheet table detection. First, we devise an effective cell featurization scheme to better leverage the rich information in each cell; second, we develop an enhanced convolutional neural network model for tab

DEEP LEARNING FOR ANOMALY DETECTION: A SURVEY

-By  Raghavendra Chalapathy  University of Sydney,  Capital Markets Co-operative Research Centre (CMCRC)  Sanjay Chawla  Qatar Computing Research Institute (QCRI),  HBKU  Paper Link Anomaly detection also known as outlier detection is the identification of rare items, events or observations which raise suspicions by differing significantly from the majority of the data. Typically the anomalous items will translate to some kind of problem such as bank fraud, a structural defect, medical problems or errors in a text. Anomalies are also referred to as outliers, novelties, noise, deviations and exceptions Hawkins defines an outlier as an observation that deviates so significantly from other observations as to arouse suspicion that it was generated by a different mechanism. Aim of this paper is two-fold, First is a structured and comprehensive overview of research methods in deep learning-based anomaly detection. Furthermore the adoption of these methods