Recognition: no theorem link
Soft-Quantum Algorithms
Pith reviewed 2026-05-10 18:29 UTC · model grok-4.3
The pith
Direct matrix optimization with unitarity regularization trains variational quantum circuits faster than gate decomposition.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Quantum operations on pure states can be trained directly as matrices by minimizing a loss that includes one regularization term to enforce near-unitarity; a subsequent circuit-alignment procedure recovers a gate-based variational circuit from the trained soft-unitary matrix. On a five-qubit supervised classification task with 1000 data points the procedure yields lower binary cross-entropy loss in under four minutes, versus more than two hours for conventional circuit training. The same soft-unitary matrices can be embedded inside a hybrid quantum-classical network that outperforms a purely classical baseline on a cartpole reinforcement-learning task.
What carries the argument
The soft-unitary training procedure: direct optimization of matrix elements under a single unitarity-regularization term, followed by circuit alignment to recover an equivalent gate decomposition.
If this is right
- For few-qubit problems with large data sets, the matrix-first route bypasses gate decomposition during the expensive optimization phase.
- The resulting trained circuits can be deployed on actual quantum hardware after the alignment step.
- Soft-unitary matrices can be inserted into hybrid quantum-classical architectures, as demonstrated by the cartpole reinforcement-learning agent that exceeds a classical baseline of similar size.
- Training time scales with matrix size rather than with the number of gates, offering a practical speed-up when datasets are large.
Where Pith is reading between the lines
- Because the method works on classical simulators, it could let researchers test variational quantum models on data sets an order of magnitude larger than those currently feasible with gate-level training.
- The approach implicitly decouples the search for good unitary operators from the choice of gate set, which may allow systematic comparison of different ansatz families once the matrices are obtained.
- If the regularization strength can be tuned automatically, the same pipeline might be applied to deeper circuits where direct gate training becomes prohibitive.
Load-bearing premise
A single regularization term added to the loss is sufficient to keep the learned matrices close enough to unitary that the subsequent alignment step produces a usable circuit without large fidelity loss.
What would settle it
On the same five-qubit, 1000-point classification task, run the two-step procedure and measure whether the final aligned circuit achieves binary cross-entropy loss comparable to or lower than direct gate training; if the aligned circuit instead shows markedly higher loss or requires comparable wall-clock time, the central claim does not hold.
Figures
read the original abstract
Quantum operations on pure states can be fully represented by unitary matrices. Variational quantum circuits, also known as quantum neural networks, embed data and trainable parameters into gate-based operations and optimize the parameters via gradient descent. The high cost of training and low fidelity of current quantum devices, however, restricts much of quantum machine learning to classical simulation. For few-qubit problems with large datasets, training the matrix elements directly, as is done with weight matrices in classical neural networks, can be faster than decomposing data and parameters into gates. We propose a method that trains matrices directly while maintaining unitarity through a single regularization term added to the loss function. A second training step, circuit alignment, then recovers a gate-based architecture from the resulting soft-unitary. On a five-qubit supervised classification task with 1000 datapoints, this two-step process produces a trained variational circuit in under four minutes, compared to over two hours for direct circuit training, while achieving lower binary cross-entropy loss. In a second experiment, soft-unitaries are embedded in a hybrid quantum-classical network for a reinforcement learning cartpole task, where the hybrid agent outperforms a purely classical baseline of comparable size.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper proposes training variational quantum circuits by directly optimizing matrix elements as 'soft-unitaries' via gradient descent on a composite loss that includes a single regularization term to enforce approximate unitarity, followed by a circuit-alignment step that recovers an equivalent gate-based variational circuit. On a 5-qubit supervised classification task with 1000 datapoints, the method reportedly trains in under 4 minutes (vs. >2 hours for direct circuit training) while achieving lower binary cross-entropy loss; a second experiment embeds soft-unitaries in a hybrid quantum-classical agent that outperforms a classical baseline on cartpole reinforcement learning.
Significance. If the regularization reliably keeps the learned matrices sufficiently close to the unitary manifold for the alignment step to recover high-fidelity circuits, the approach could provide a practical route to faster classical simulation and training of small-qubit variational quantum models on large datasets by borrowing techniques from classical neural-network training. The reported empirical speed-up and loss improvement constitute a concrete, falsifiable demonstration that merits attention, though its broader significance depends on quantitative validation of unitarity preservation and generalization beyond the 5-qubit regime.
major comments (3)
- [Method and regularization description] The central claim that a single additive regularization term suffices to produce soft-unitaries close enough to the unitary group for circuit alignment to preserve performance is load-bearing, yet the manuscript supplies no post-training measurement of ||U†U − I||_F, spectral deviation, or average gate fidelity, nor any scaling analysis with qubit number or dataset size. Gradient descent on the composite loss can therefore trade classification accuracy against unitarity without the reader being able to assess the resulting fidelity loss in the aligned circuit.
- [Experimental results on the five-qubit classification task] The timing and loss comparisons (under 4 min vs. >2 h, lower BCE) are presented without error bars, details on the exact form and coefficient of the regularization term, the circuit-alignment algorithm, or the precise gate decomposition used in the baseline. This makes it impossible to determine whether the reported advantage is robust or sensitive to hyper-parameter choices and implementation details.
- [Reinforcement-learning experiment] In the hybrid RL experiment, the manner in which soft-unitary matrices are embedded and simulated within the quantum component is not specified quantitatively (e.g., whether they are projected back onto unitaries or how non-unitary deviations propagate through the policy network), leaving open whether the performance gain truly stems from the quantum part or from the hybrid architecture itself.
minor comments (2)
- The term 'soft-unitary' is introduced without a precise mathematical definition (e.g., a bound on the deviation from U†U = I) or comparison to related concepts such as approximate unitaries or Stiefel-manifold optimization already present in the quantum-machine-learning literature.
- Notation for the loss function, regularization coefficient, and circuit-alignment procedure should be introduced with explicit equations rather than descriptive prose to improve reproducibility.
Simulated Author's Rebuttal
We thank the referee for their thorough review and constructive comments on our manuscript. We address each major comment below and have revised the manuscript to incorporate the suggested clarifications and additions.
read point-by-point responses
-
Referee: [Method and regularization description] The central claim that a single additive regularization term suffices to produce soft-unitaries close enough to the unitary group for circuit alignment to preserve performance is load-bearing, yet the manuscript supplies no post-training measurement of ||U†U − I||_F, spectral deviation, or average gate fidelity, nor any scaling analysis with qubit number or dataset size. Gradient descent on the composite loss can therefore trade classification accuracy against unitarity without the reader being able to assess the resulting fidelity loss in the aligned circuit.
Authors: We agree that explicit post-training unitarity metrics were not reported. In the revised manuscript we will include measurements of ||U†U − I||_F, the spectral norm deviation, and average gate fidelity for the trained soft-unitaries on the 5-qubit task. We will also add a short scaling study with qubit number (up to 6–7 qubits) and dataset size to quantify how the regularization coefficient affects the trade-off. These additions will allow readers to directly evaluate the fidelity preserved by the alignment step. revision: yes
-
Referee: [Experimental results on the five-qubit classification task] The timing and loss comparisons (under 4 min vs. >2 h, lower BCE) are presented without error bars, details on the exact form and coefficient of the regularization term, the circuit-alignment algorithm, or the precise gate decomposition used in the baseline. This makes it impossible to determine whether the reported advantage is robust or sensitive to hyper-parameter choices and implementation details.
Authors: We acknowledge the lack of these implementation details and statistical measures. The revised version will report error bars computed over 10 independent runs, specify the regularization term as λ‖U†U − I‖_F² with the exact λ value used, provide pseudocode and a description of the circuit-alignment procedure, and detail the gate set and decomposition method employed for the direct circuit baseline. These changes will enable reproducibility and assessment of robustness. revision: yes
-
Referee: [Reinforcement-learning experiment] In the hybrid RL experiment, the manner in which soft-unitary matrices are embedded and simulated within the quantum component is not specified quantitatively (e.g., whether they are projected back onto unitaries or how non-unitary deviations propagate through the policy network), leaving open whether the performance gain truly stems from the quantum part or from the hybrid architecture itself.
Authors: We thank the referee for highlighting this ambiguity. In the revision we will add a quantitative description of the embedding: the soft-unitary matrices are used directly in the quantum simulation layer without projection, with any non-unitary component propagating through the density-matrix evolution; we will also report the observed deviation norm during training and include an ablation comparing the hybrid agent against a version that forces unitarity via projection. This will clarify the contribution of the soft-unitary quantum component. revision: yes
Circularity Check
No circularity: proposal is self-contained empirical method
full rationale
The paper introduces a two-step procedure—direct matrix training with an added regularization term to enforce approximate unitarity, followed by a separate circuit-alignment step to recover a gate decomposition. No equation or claim reduces the final trained circuit or reported speed/accuracy gains to a fitted parameter renamed as a prediction, nor to a self-citation chain, uniqueness theorem, or ansatz imported from prior work by the same authors. The regularization term is presented as a direct proposal rather than derived from the target result, and performance numbers are obtained from explicit experiments on concrete tasks. The derivation chain therefore remains independent of its own outputs.
Axiom & Free-Parameter Ledger
free parameters (1)
- regularization coefficient
axioms (1)
- standard math Quantum operations on pure states can be represented by unitary matrices
invented entities (1)
-
soft-unitary matrix
no independent evidence
Reference graph
Works this paper leans on
-
[1]
Peruzzo, J
A. Peruzzo, J. McClean, P. Shadbolt, M.-H. Yung, X.-Q. Zhou, P. J. Love, A. Aspuru-Guzik, and J. L. O’brien, A variational eigenvalue solver on a photonic quantum processor, Nature Communications5, 4213 (2014). 6
2014
-
[2]
Mitarai, M
K. Mitarai, M. Negoro, M. Kitagawa, and K. Fujii, Quan- tum circuit learning, Physical Review A98, 032309 (2018)
2018
-
[3]
Benedetti, E
M. Benedetti, E. Lloyd, S. Sack, and M. Fiorentini, Pa- rameterized quantum circuits as machine learning models, Quantum Science and Technology4, 043001 (2019)
2019
-
[4]
Sagingalieva, M
A. Sagingalieva, M. Kordzanganeh, N. Kenbayev, D. Kosichkina, T. Tomashuk, and A. Melnikov, Hybrid quantum neural network for drug response prediction, Cancers15, 2705 (2023)
2023
-
[5]
Cerezo, A
M. Cerezo, A. Arrasmith, R. Babbush, S. C. Benjamin, S. Endo, K. Fujii, J. R. McClean, K. Mitarai, X. Yuan, L. Cincio, and P. J. Coles, Variational quantum algo- rithms, Nature Reviews Physics3, 625 (2021)
2021
-
[6]
Melnikov, M
A. Melnikov, M. Kordzanganeh, A. Alodjants, and R.-K. Lee, Quantum machine learning: from physics to software engineering, Advances in Physics: X8, 2165452 (2023)
2023
-
[7]
Preskill, Quantum computing in the NISQ era and beyond, Quantum2, 79 (2018)
J. Preskill, Quantum computing in the NISQ era and beyond, Quantum2, 79 (2018)
2018
-
[8]
Bharti, A
K. Bharti, A. Cervera-Lierta, T. H. Kyaw, T. Haug, S. Alperin-Lea, A. Anand, M. Degroote, H. Heimonen, J. S. Kottmann, T. Menke,et al., Noisy intermediate- scale quantum algorithms, Reviews of Modern Physics 94, 015004 (2022)
2022
-
[9]
Kordzanganeh, M
M. Kordzanganeh, M. Buchberger, B. Kyriacou, M. Po- volotskii, W. Fischer, A. Kurkin, W. Somogyi, A. Sagin- galieva, M. Pflitsch, and A. Melnikov, Benchmarking sim- ulated and physical quantum processing units using quan- tum and hybrid algorithms, Advanced Quantum Tech- nologies6, 2300043 (2023)
2023
- [10]
- [11]
-
[12]
Kandala, A
A. Kandala, A. Mezzacapo, K. Temme, M. Takita, M. Brink, J. M. Chow, and J. M. Gambetta, Hardware- efficient variational quantum eigensolver for small molecules and quantum magnets, Nature549, 242 (2017)
2017
-
[13]
B. M´ at´ e, B. L. Saux, and M. Henderson, Beyond Ans¨ atze: Learning quantum circuits as unitary operators, arXiv preprint arXiv:2203.00601 (2022)
-
[14]
Arjovsky, A
M. Arjovsky, A. Shah, and Y. Bengio, Unitary evolution recurrent neural networks, inInternational Conference on Machine Learning(PMLR, 2016) pp. 1120–1128
2016
-
[15]
Lezcano-Casado and D
M. Lezcano-Casado and D. Mart´ ınez-Rubio, Cheap or- thogonal constraints in neural networks: A simple parametrization of the orthogonal and unitary group, in International Conference on Machine Learning(PMLR,
-
[16]
Kuzmin, W
V. Kuzmin, W. Somogyi, E. Pankovets, and A. Melnikov, Method for noise-induced regularization in quantum neu- ral networks, Advanced Quantum Technologies8, e00603 (2025)
2025
-
[17]
J. R. McClean, S. Boixo, V. N. Smelyanskiy, R. Bab- bush, and H. Neven, Barren plateaus in quantum neural network training landscapes, Nature Communications9, 4812 (2018)
2018
-
[18]
Cerezo, A
M. Cerezo, A. Sone, T. Volkoff, L. Cincio, and P. J. Coles, Cost function dependent barren plateaus in shallow parametrized quantum circuits, Nature Communications 12, 1791 (2021)
2021
-
[19]
Ragone, B
M. Ragone, B. N. Bakalov, F. Sauvage, A. F. Kemper, C. O. Marrero, M. Larocca, and M. Cerezo, A Lie algebraic theory of barren plateaus for deep parameterized quantum circuits, Nature Communications15, 7172 (2024)
2024
-
[20]
Holmes, K
Z. Holmes, K. Sharma, M. Cerezo, and P. J. Coles, Con- necting ansatz expressibility to gradient magnitudes and barren plateaus, PRX Quantum3, 010313 (2022)
2022
-
[21]
S. Sim, P. D. Johnson, and A. Aspuru-Guzik, Express- ibility and entangling capability of parameterized quan- tum circuits for hybrid quantum-classical algorithms, Ad- vanced Quantum Technologies2, 1900070 (2019)
2019
-
[22]
V. Patapovich, M. Periyasamy, M. Kordzanganeh, and A. Melnikov, Superposed parameterised quantum circuits, arXiv preprint arXiv:2506.08749 (2025)
-
[23]
M. A. Nielsen and I. L. Chuang,Quantum computation and quantum information(Cambridge University Press, 2010)
2010
-
[24]
V. V. Shende, S. S. Bullock, and I. L. Markov, Synthesis of quantum logic circuits, inProceedings of the 2005 Asia and South Pacific Design Automation Conference(2005) pp. 272–275
2005
-
[25]
Khatri, R
S. Khatri, R. LaRose, A. Poremba, L. Cincio, A. T. Sorn- borger, and P. J. Coles, Quantum-assisted quantum com- piling, Quantum3, 140 (2019)
2019
-
[26]
Schuld, R
M. Schuld, R. Sweke, and J. J. Meyer, Effect of data encoding on the expressive power of variational quantum- machine-learning models, Physical Review A103, 032430 (2021)
2021
-
[27]
Kordzanganeh, P
M. Kordzanganeh, P. Sekatski, L. Fedichkin, and A. Mel- nikov, An exponentially-growing family of universal quan- tum circuits, Machine Learning: Science and Technology 4, 035036 (2023)
2023
-
[28]
V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller, Playing Atari with deep reinforcement learning, arXiv preprint arXiv:1312.5602 (2013)
work page internal anchor Pith review arXiv 2013
-
[29]
Jerbi, C
S. Jerbi, C. Gyurik, S. C. Marshall, H. J. Briegel, and V. Dunjko, Parametrized quantum policies for reinforce- ment learning, inAdvances in Neural Information Pro- cessing Systems, Vol. 34 (2021) pp. 28362–28375
2021
-
[30]
Skolik, S
A. Skolik, S. Jerbi, and V. Dunjko, Quantum agents in the Gym: a variational quantum algorithm for deep Q- learning, Quantum6, 720 (2022)
2022
-
[31]
Lockwood and M
O. Lockwood and M. Si, Playing Atari with hybrid quantum-classical reinforcement learning, inNeurIPS 2020 Workshop on Pre-registration in Machine Learn- ing(PMLR, 2021) pp. 285–301
2020
-
[32]
Kordzanganeh, D
M. Kordzanganeh, D. Kosichkina, and A. Melnikov, Par- allel hybrid networks: an interplay between quantum and classical neural networks, Intelligent Computing2, 0028 (2023)
2023
-
[33]
Kurkin, J
A. Kurkin, J. Hegemann, M. Kordzanganeh, and A. Mel- nikov, Forecasting steam mass flow in power plants using the parallel hybrid network, Engineering Applications of Artificial Intelligence160, 111912 (2025)
2025
-
[34]
Z. Li, N. Kovachki, K. Azizzadenesheli, B. Liu, K. Bhat- tacharya, A. Stuart, and A. Anandkumar, Fourier neural operator for parametric partial differential equations, in International Conference on Learning Representations (2021)
2021
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.