Skip to main content

Fast Quantum Algorithm for Learning

 -By Hayata Yamasaki,  Sathyawageeswar Subramanian,  Sho Sonoda,  Masato Koashi


Paper Link

image Courtesy: Fossguru.com


Abstract

Kernel methods augmented with random features give scalable algorithms for learning from big data. But it has been computationally hard to sample random features according to a probability distribution that is optimized for the data, so as to minimize the required number of features for achieving the learning to the desired accuracy. Here, we develop a quantum algorithm for sampling from this optimized distribution over features, in runtime O(D) that is linear in the dimension D of the input data. Our algorithm achieves an exponential speedup in D compared to any known classical algorithm for this sampling task. In contrast to existing quantum machine learning algorithms, our algorithm circumvents sparsity and low-rank assumptions and thus has wide applicability. We also show that the sampled features can be combined with regression by stochastic gradient descent to achieve the learning without canceling out our exponential speedup. Our algorithm-based on sampling optimized random features leads to an accelerated framework for machine learning that takes advantage of quantum computers. 


Random features provide a powerful technique for scaling up kernel methods applicable to various machine learning tasks, such as ridge regression, kernel learning, and principle component analysis. Recently, Bach has shown an optimized probability distribution of random features, and sampling features from this optimized distribution would drastically improve runtime of learning algorithms based on random features. However, this sampling task has been computationally “hard in practice” due to inversion of a high-dimensional operator. In contrast, the power of quantum computers to process data in quantum superposition attracts growing attention towards accelerating learning tasks, opening a new field of quantum machine learning (QML). QML accelerates a supervised learning task, by constructing an efficient quantum algorithm for sampling from this optimized distribution.

Problem: 

These conventional methods for obtaining random features from the data-independent distribution dτ (v) require a large number M of features to learn the function f, which slows down the decision of all M features and the regression over M coefficients. Aiming at minimizing M required for the learning, rather than sampling from dτ (v), we will sample features from a probability distribution that puts greater weight on important features that are optimized for the data, via a probability density function q(v) for dτ (v). 


Solution

  •  (Theorem 1) We construct a new quantum algorithm for sampling an optimized random feature from q ∗  (v)dτ (v) in as fast as linear runtime O(D) in the data dimension D. The best existing classical algorithm by Bach for sampling each single feature from this data-optimized distribution q ∗  (v)dτ (v) requires exponential runtime O(exp(D)). In contrast, our quantum algorithm can sample each single feature from q ∗  (v)dτ (v) with runtime O(D), which is as fast as the conventional algorithms such as. These conventional algorithms perform an easier task, i.e. sampling from a data-independent distribution dτ (v). Advantageously over the conventional algorithms sampling from dτ (v), quantum algorithm sampling is used from q ∗  (v)dτ (v) to achieve learning with a significantly small number M of features, which is proven to be minimal up to a logarithmic gap. Remarkably, we achieve this without assuming sparsity or low rank of relevant operators.

  • To construct this quantum algorithm, we circumvent the difficulty of infinite dimension by formulating a discrete approximation of the problem of sampling a real-valued feature from q ∗  (v)dτ (v). This approximation is equivalent to using fixed-point number representation with rescaling.

  • (Theorem 2) The combined M features sampled by the algorithm with regression by stochastic gradient descent to achieve supervised learning in time O(MD), i.e. without canceling out our exponential speedup. This M is minimal up to a logarithmic gap, since we use optimized random features. Thus, by improving the computational bottleneck faced by classical algorithms for sampling optimized random features, we provide a promising framework of quantum machine learning that leverages our O(D) sampling algorithm to achieve the optimal M among all algorithms using random features.

Conclusion

 a quantum algorithm is constructed for sampling an optimized random feature within a linear time O(D) in data dimension D, achieving an exponential speedup in D compared to the existing classical algorithm of Bach for this sampling task. Combining M features sampled by this quantum algorithm with stochastic gradient descent, we can achieve supervised learning in time O(MD), where this M is expected to be nearly minimal since we use the optimized random features. As for future work, it is open to prove the hardness of sampling an optimized random feature for any possible classical algorithm under complexity-theoretical assumptions. It is also interesting to investigate whether we can reduce the runtime to O(M log D).  but using the optimized random features to achieve minimal M. Since our quantum algorithm does not impose sparsity or low-rank assumptions, our results open a route to a widely applicable framework of kernel-based quantum machine learning with an exponential speedup.



Comments

Popular posts from this blog

ABOD and its PyOD python module

Angle based detection By  Hans-Peter Kriegel, Matthias Schubert, Arthur Zimek  Ludwig-Maximilians-Universität München  Oettingenstr. 67, 80538 München, Germany Ref Link PyOD By  Yue Zhao   Zain Nasrullah   Department of Computer Science, University of Toronto, Toronto, ON M5S 2E4, Canada  Zheng Li jk  Northeastern University Toronto, Toronto, ON M5X 1E2, Canada I am combining two papers to summarize Anomaly detection. First one is Angle Based Outlier Detection (ABOD) and other one is python module that  uses ABOD along with over 20 other apis (PyOD) . This is third part in the series of Anomaly detection. First article exhibits survey that covered length and breadth of subject, Second article highlighted on data preparation and pre-processing.  Angle Based Outlier Detection. Angles are more stable than distances in high dimensional spaces for example the popularity of cosine-based sim...

Ownership at Large

 Open Problems and Challenges in Ownership Management -By John Ahlgren, Maria Eugenia Berezin, Kinga Bojarczuk, Elena Dulskyte, Inna Dvortsova, Johann George, Natalija Gucevska, Mark Harman, Shan He, Ralf Lämmel, Erik Meijer, Silvia Sapora, and Justin Spahr-Summers Facebook Inc.  Software-intensive organizations rely on large numbers of software assets of different types, e.g., source-code files, tables in the data warehouse, and software configurations. Who is the most suitable owner of a given asset changes over time, e.g., due to reorganization and individual function changes. New forms of automation can help suggest more suitable owners for any given asset at a given point in time. By such efforts on ownership health, accountability of ownership is increased. The problem of finding the most suitable owners for an asset is essentially a program comprehension problem: how do we automatically determine who would be best placed to understand, maintain, ev...

Hybrid Approach to Automation, RPA and Machine Learning

- By Wiesław Kopec´, Kinga Skorupska, Piotr Gago, Krzysztof Marasek  Polish-Japanese Academy of Information Technology Paper Link Courtesy DZone   Abstract One of the more prominent trends within Industry 4.0 is the drive to employ Robotic Process Automation (RPA), especially as one of the elements of the Lean approach.     The full implementation of RPA is riddled with challenges relating both to the reality of everyday business operations, from SMEs to SSCs and beyond, and the social effects of the changing job market. To successfully address these points there is a need to develop a solution that would adjust to the existing business operations and at the same time lower the negative social impact of the automation process. To achieve these goals we propose a hybrid, human-centred approach to the development of software robots. This design and  implementation method combines the Living Lab approach with empowerment through part...