Table of Contents
- Introduction
- The Challenge of Expressivity in Quantum Circuits
- What Is Data Re-uploading?
- Motivation Behind Data Re-uploading
- Mathematical Foundation of Re-uploading
- Circuit Architecture with Re-uploading
- Implementation Techniques
- Periodic vs Adaptive Re-uploading
- Comparison with Classical Deep Networks
- Advantages of Data Re-uploading
- Trade-Offs: Depth vs Expressivity
- Examples in PennyLane
- Examples in Qiskit
- QNN Performance with Re-uploading
- Noise Considerations and Depth Limitation
- Re-uploading in Hybrid Quantum-Classical Models
- Relation to Universal Approximation Theorems
- Visualization of Feature Space Expansion
- Empirical Benchmarks and Research Results
- Conclusion
1. Introduction
Data re-uploading is a strategy used in quantum machine learning (QML) to enhance the expressive power of parameterized quantum circuits by embedding classical data multiple times across layers of the quantum circuit.
2. The Challenge of Expressivity in Quantum Circuits
- Single-layer embeddings are limited by circuit depth
- Quantum feature maps might not separate complex data sufficiently
- NISQ devices impose constraints on width/depth
3. What Is Data Re-uploading?
- Repeatedly encoding input data at multiple layers in a variational circuit
- Interleaved with learnable quantum operations
- Analogous to depth in classical neural networks
4. Motivation Behind Data Re-uploading
- Improves model expressiveness without needing additional qubits
- Allows quantum circuits to approximate nonlinear functions
- Inspired by residual connections and multilayer networks
5. Mathematical Foundation of Re-uploading
A circuit with re-uploading can be written as:
\[
U(x, heta) = \prod_{l=1}^L U_l(x, heta_l)
\]
Each \( U_l \) includes both data encoding and parameterized unitaries.
6. Circuit Architecture with Re-uploading
- Each block consists of:
- Data embedding gate (e.g., RX(x_i))
- Parameterized gate (e.g., RZ(θ_i))
- Stack several such blocks
7. Implementation Techniques
- Use angle encoding repeatedly in successive layers
- PennyLane:
qml.AngleEmbedding(data, wires, rotation='Y')
inside loop - Qiskit: Apply data-driven RX/RY multiple times
8. Periodic vs Adaptive Re-uploading
- Periodic: repeat same data in each layer
- Adaptive: use transformed data in deeper layers
9. Comparison with Classical Deep Networks
- Each re-uploading layer plays the role of a neural net layer
- Expressivity grows with number of re-upload layers
10. Advantages of Data Re-uploading
- Increases nonlinear decision boundaries
- Easy to implement on hardware
- Compatible with all variational models
11. Trade-Offs: Depth vs Expressivity
- More layers → more expressivity
- But also → increased noise and training difficulty
12. Examples in PennyLane
def circuit(x, weights):
for i in range(len(weights)):
qml.AngleEmbedding(x, wires=[0, 1])
qml.RY(weights[i][0], wires=0)
qml.RZ(weights[i][1], wires=1)
13. Examples in Qiskit
for layer in range(depth):
for i in range(num_qubits):
qc.ry(x[i], i)
qc.rz(params[layer][i], i)
14. QNN Performance with Re-uploading
- Demonstrated improvement on toy datasets
- Better accuracy on binary classification tasks
- Comparable to classical neural nets for small N
15. Noise Considerations and Depth Limitation
- Re-uploading increases circuit depth
- Apply transpilation and noise mitigation
- Use simulators to test different depths
16. Re-uploading in Hybrid Quantum-Classical Models
- Combine quantum layers with classical dense layers
- Re-uploading improves interface between quantum and classical data
17. Relation to Universal Approximation Theorems
- Data re-uploading contributes to universality
- Like multilayer perceptrons, quantum circuits with re-uploading can approximate any bounded continuous function
18. Visualization of Feature Space Expansion
- Project intermediate Bloch vectors
- Visualize learned class separability over layers
19. Empirical Benchmarks and Research Results
- Schuld et al. (2021): proved re-uploading enhances model capacity
- Used in QNNs, QSVR, QGANs with improved fidelity
20. Conclusion
Data re-uploading is a powerful, hardware-compatible method to increase the representational capacity of quantum machine learning circuits. By repeatedly encoding inputs, QML models can approximate complex functions, making them more viable for real-world data-driven applications.