Table of Contents
- Introduction
- What Is Overfitting in Machine Learning?
- Manifestation of Overfitting in Quantum Models
- Sources of Overfitting in Quantum Machine Learning
- Variational Quantum Circuits and Model Complexity
- Role of Circuit Depth in Overfitting
- The Curse of Expressivity in QML
- Quantum Generalization Theory: Early Insights
- Indicators of Overfitting in Quantum Workflows
- Evaluating Generalization on Quantum Devices
- Regularization Techniques for Quantum Models
- Parameter Norm Penalties (L2 Regularization)
- Early Stopping in Quantum Training
- Circuit Pruning and Parameter Dropout
- Noise-Injection as a Regularizer
- Ensemble Learning in Quantum Circuits
- Cross-Validation in Quantum Learning
- Hybrid Regularization Strategies
- Research on Generalization Bounds in QML
- Conclusion
1. Introduction
Overfitting occurs when a model performs well on training data but fails to generalize to unseen examples. In quantum machine learning (QML), overfitting arises due to overparameterized variational circuits, excessive entanglement, or insufficient training samples.
2. What Is Overfitting in Machine Learning?
- High accuracy on training set
- Poor performance on test/validation data
- Model “memorizes” instead of learning patterns
3. Manifestation of Overfitting in Quantum Models
- Low training loss, high validation loss
- Quantum classifiers that learn noise in training measurements
- Circuit configurations that are too expressive
4. Sources of Overfitting in Quantum Machine Learning
- Too many variational parameters
- Deep quantum circuits on small datasets
- Poor encoding strategies
- Insufficient shot counts (measurement noise)
5. Variational Quantum Circuits and Model Complexity
- Like deep neural networks, VQCs can be overparameterized
- More parameters → more capacity to memorize noise
6. Role of Circuit Depth in Overfitting
- Deeper circuits often provide greater expressivity
- But increase risk of overfitting, especially on small data
7. The Curse of Expressivity in QML
- Highly expressive circuits can represent arbitrary functions
- Without regularization, this leads to poor generalization
8. Quantum Generalization Theory: Early Insights
- Still a developing field
- Concepts like VC-dimension, Rademacher complexity being adapted for QML
- Fidelity-based generalization bounds under study
9. Indicators of Overfitting in Quantum Workflows
- Loss curves diverging after initial convergence
- Overly sensitive circuit outputs to small input changes
- Highly unstable gradients
10. Evaluating Generalization on Quantum Devices
- Use separate validation set with fixed shot budget
- Monitor variance across multiple runs
11. Regularization Techniques for Quantum Models
- L2 weight decay
- Circuit structure restrictions
- Parameter sparsity
12. Parameter Norm Penalties (L2 Regularization)
Add to cost function:
\[
\mathcal{L}_{ ext{reg}} = \mathcal{L} + \lambda \sum_i heta_i^2
\]
Where \( heta_i \) are circuit parameters.
13. Early Stopping in Quantum Training
- Stop optimization when validation loss increases
- Helps prevent convergence to overfitted minima
14. Circuit Pruning and Parameter Dropout
- Disable certain gates randomly during training
- Reduce circuit size post-training by removing low-contribution parameters
15. Noise-Injection as a Regularizer
- Add controlled noise to circuit outputs or parameters
- Mimics classical dropout
16. Ensemble Learning in Quantum Circuits
- Average predictions from multiple small VQCs
- Reduces variance and improves robustness
17. Cross-Validation in Quantum Learning
- k-fold or leave-one-out strategies adapted for QML
- Evaluate circuit generality over multiple splits
18. Hybrid Regularization Strategies
- Combine classical regularizers with quantum-specific constraints
- e.g., depth limits + weight decay + early stopping
19. Research on Generalization Bounds in QML
- Ongoing research in quantum PAC learning
- Fidelity-based loss bounds, entropic capacity metrics
- Open question: What governs the learnability of QNNs?
20. Conclusion
Quantum models, like their classical counterparts, are prone to overfitting. Regularization techniques such as early stopping, L2 penalties, circuit pruning, and hybrid strategies can enhance generalization, ensuring quantum systems not only learn but also perform reliably on unseen data.