Quantum machine learning can work even with noisy data

Machine Learning


Imperfections in both classical data and quantum hardware impact the performance of quantum machine learning models. Bhavna Bose of SVKM’s NMIMS Mukesh Patel School of Technology and Management and Muhammad Faryad of the Lahore University of Management have shown that noise from flawed classical data can exacerbate the effects of quantum decoherence. Their exhaustive research utilizing the Titanic dataset and various noise models within the Qiskit Aer simulator revealed that classification accuracy is significantly reduced in the presence of both classical and quantum noise. These findings highlight the critical need to consider the combination of classical and quantum noise when developing and evaluating quantum machine learning pipelines for near-term applications.

Simulating data corruption and quantum noise for powerful machine learning

A carefully constructed methodology to simulate realistic noise conditions within a quantum machine learning pipeline formed the basis of this research. Classical data is initially corrupted by several models such as speckle, impulse, quantization, and feature dropout, representing common real-world imperfections encountered during data acquisition and preprocessing. Speckle noise, which mimics sensor inaccuracies, introduces multiplicative noise, while impulse noise simulates sudden, temporary errors like bit flips. Quantization represents the loss of precision when converting continuous data to discrete values, and feature dropout randomly removes input features to simulate missing data scenarios. This flawed data was passed through a ZZ feature map, a technique that converts classical information into a quantum form that quantum computers can understand. ZZ feature maps encode classical data into quantum states by applying a rotation about the qubit’s Z axis, effectively creating a superposition of states representing the input features. This is an important step because initial errors can be translated into the quantum domain and their effects can be amplified during subsequent quantum calculations.

The Titanic dataset, a commonly used benchmark in machine learning, was used to benchmark variational quantum classifiers against multiple noise types. This dataset contains information about Titanic passengers and is used to predict survival based on characteristics such as age, gender, and class. After ZZ feature map transformation, the data was affected by quantum noise channels such as depolarization, amplitude, phase attenuation, Pauli, and readout errors. Depolarization noise randomly applies identity operations to a qubit, effectively erasing its quantum state. Amplitude and phase attenuation represent energy loss from the qubit, causing decoherence. Pauli errors cause bit-flip or phase-flip errors, which result in readout errors while measuring the state of a qubit. The Qiskit Aer simulator can now be used to model these quantum effects and closely examine how these errors propagate through the system and impact classification results. This simulator allows researchers to control the levels of each noise type and systematically investigate their combined effects. By simulating these errors, the researchers were able to assess the robustness of their variational quantum classifier without having access to actual noisy quantum hardware.

The combination of classical and quantum noise severely impairs variational quantum classification

Classification accuracy utilizing the Titanic dataset decreased to 78.5% when both classical and quantum noise were applied, a drop of more than 15 percent compared to a simulation without noise. This threshold represents a critical point at which the ability of a variational quantum classifier to reliably distinguish between classes is severely compromised. Previously, it was considered impossible to achieve accurate classification under such complex noise conditions. Classical input noise worsens quantum decoherence and creates unstable training dynamics, significantly reducing the potential benefits of short-term quantum machine learning. Variational quantum classifiers rely on iterative optimization of quantum circuit parameters, and noise can disrupt this process, resulting in suboptimal solutions and reduced accuracy.

By modeling realistic noise at multiple levels, from data acquisition to quantum circuit execution, we obtain a more practical assessment of QML performance than previous work. Before quantum processing, we applied four different classical noise types (speckle, impulse, quantization, and feature dropout) to the Titanic dataset and found that even moderate classical corruption had a significant impact on performance. Specifically, a 10% level of impulse noise reduced the classification accuracy to 72.3%, demonstrating clear sensitivity to the quality of the input data beyond quantum errors. This emphasizes that the quality of the initial data is paramount before considering the limitations of quantum hardware. A simulation that incorporated both Pauli error and amplitude attenuation at the quantum circuit level, along with speckle noise from classical data, resulted in an overall accuracy of only 69.8%. This highlights combinations of degradation that were not observed when the noise sources were tested alone. The synergistic effects of complex noises suggest that mitigating just one type of noise is not sufficient. A holistic approach is required. Although these results indicate significant performance degradation under realistic conditions, extending these results to larger, more complex problems and diverse hardware platforms remains an important next step. Investigating the impact of these noise combinations on datasets with higher dimensions and more complex feature relationships is important to assess the generalizability of these results.

Classical data incompleteness limits the performance of variational quantum classifiers

Establishing reliable quantum machine learning requires more than simply improving the quantum hardware itself. Researchers at SVKM’s NMIMS Mukesh Patel School of Technology and Management, led by Bhavna Bose, have convincingly demonstrated that even the earliest qubits are vulnerable to errors caused by the classical data used to train the algorithms. Variational quantum classifiers, a core component of this emerging field, are particularly sensitive. A variational quantum classifier is a hybrid quantum-classical algorithm that uses a quantum circuit to prepare quantum states and a classical optimizer to tune the circuit parameters to minimize the cost function. Bose acknowledged that their analysis relied on a single, well-known data set: the Titanic passenger list, raising questions about whether these findings apply to more complex real-world problems. Further research is needed to assess the robustness of these findings across a wider range of datasets and problem areas.

Classical data imperfections worsen quantum decoherence within variational quantum classifiers, impacting training stability and reducing classification accuracy. By systematically combining realistic classical noise and simulated quantum hardware errors, we demonstrated synergistic performance degradation, highlighting the importance of data quality alongside hardware improvements for reliable machine learning. This study highlights the importance of addressing classical data imperfections to realize the full potential of quantum machine learning algorithms. Further research into robust data preprocessing techniques and noise mitigation strategies, such as error correcting codes and noise filtering algorithms, is essential to build practical and reliable QML systems. Considering the use of data augmentation techniques to increase model robustness to noisy data is also a promising avenue for future research. Ultimately, unlocking the full potential of quantum machine learning will require a comprehensive approach that addresses both classical and quantum noise sources.

In this study, we demonstrated that classical data imperfections can exacerbate the effects of quantum decoherence in variational quantum classifiers. This is important because it shows that the quality of the input data is as important as the quantum hardware itself to achieve reliable machine learning results. Using the Titanic dataset, the researchers showed that combining classical noise with simulated quantum errors reduces classification accuracy and makes training unstable. The authors suggest that future research should focus on improving data preprocessing and noise reduction strategies to build more robust quantum machine learning systems.



Source link