Skip to content

ChiShengChen/QASA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

QASA: Quantum Adaptive Self-Attention for Quantum Transformer Models

arXiv
Python PyTorch PennyLane

Overview

QASA (Quantum Adaptive Self-Attention) is an innovative hybrid quantum-classical model designed to improve time series prediction tasks. By leveraging the power of quantum computing principles, QASA demonstrates superior predictive capabilities compared to traditional classical models, particularly for complex time-varying systems such as damped oscillators.

image

📢 News / Updates

  • 🔥 [2025-05-21]: Add Benchmark files.
  • [2025-04-13]: Released QASA v2 with improved stability, LayerNorm, Kaiming init, and dropout support.

Key Advantages of QASA

  • Enhanced Prediction Accuracy: QASA's hybrid architecture combines the representational power of quantum circuits with classical transformer models to achieve improved prediction accuracy over purely classical approaches.

  • Efficient Parameter Usage: The quantum components enable more efficient parameter utilization, requiring fewer trainable parameters to achieve comparable or better results than classical counterparts.

  • Robust Feature Extraction: The quantum layers provide unique feature extraction capabilities that can capture complex non-linear patterns in time series data.

  • Adaptability: QASA is particularly effective at modeling systems with complex dynamics, such as the damped oscillator demonstrated in this repository.

Project Structure

The repository contains three main implementations for comparison:

  1. QASA (Quantum-Assisted Model): quantum_v4_damped.py

    • A hybrid model that integrates quantum circuits within a transformer architecture
    • Leverages quantum principles for enhanced feature representation
    • Implements parameterized quantum circuits using PennyLane
  2. Classical Transformer V4: transformer_classicalv4_damped.py

    • A purely classical transformer model with similar architecture to QASA
    • Used as a direct comparison to demonstrate quantum advantage
  3. Baseline Transformer: transformer_attention_damped.py

    • A standard transformer implementation with attention mechanisms
    • Serves as the baseline for performance benchmarking

Technical Details

QASA Architecture

  • Hybrid Design: Combines quantum circuit layers with classical transformer components
  • Quantum Processing: Incorporates parameterized quantum circuits with 8 qubits and 4 layers
  • Circuit Interface: Implemented using PennyLane with PyTorch integration

Data

The models are trained on simulated damped oscillator data, representing a common physical system with complex dynamics:

x(t) = A * exp(-γt) * cos(ωt + φ)

where:

  • A: Amplitude
  • γ: Damping coefficient
  • ω: Angular frequency
  • φ: Initial phase

Results

QASA demonstrates several advantages over the classical implementations:

  1. Prediction Accuracy: Lower MSE and MAE values on test data compared to classical models
  2. Convergence Rate: Faster convergence to optimal solutions during training
  3. Pattern Recognition: Better ability to capture the underlying patterns of damped oscillation

QASA v1 vs QASA v2 Comparison

Component QASA v1 (quantum_v4_damped.py) QASA v2 (qasa_v2_damped.py) Improvement Description
Normalization BatchNorm1d in QuantumLayer LayerNorm in QuantumLayer LayerNorm is more stable for Transformer-style models, especially with small batch sizes.
Weight Initialization None Kaiming initialization for all linear layers Ensures better training stability and faster convergence.
Dropout Not used Added in embedding and FFN layers Helps reduce overfitting and improves generalization.
QuantumLayer Skip Connection Direct residual connection, no shape check Skip connection with shape check and warning if mismatched Prevents silent errors when dimensions mismatch.
Optimizer AdamW, lr = 1e-4, no weight decay AdamW, lr = 5e-5, with weight_decay=1e-4 Lower learning rate and weight decay improve convergence and regularization.
Scheduler CosineAnnealingLR CosineAnnealingLR Same scheduling method in both.
Embedding Layer Linear + LayerNorm Linear + LayerNorm + Dropout Additional dropout improves training robustness.
QuantumLayer Input TanhBatchNormTorchLayer TanhLayerNormTorchLayer Modernized normalization strategy.
QuantumLayer Output x + output x + output (if dimensions match), else return output with warning Improves logical safety of skip connections.
FFN Layers in Encoder No dropout Dropout added Improves regularization.
CSV Logger Present Present Same functionality.
Plotting and Logging Enabled Enabled Same functionality.
Model Flexibility Fixed structure Supports dropout_rate argument and better modularity Easier to tune and extend.

Getting Started

Prerequisites

  • Python 3.8+
  • PyTorch 1.9+
  • PennyLane 0.20+
  • PyTorch Lightning 1.5+

Installation

# Clone the repository
git clone https://github.com/ChiShengChen/QASA.git
cd QASA

# Install dependencies
pip install -r requirements.txt

Training the Models

Each model can be trained using the respective Python scripts:

# Train the QASA model
python quantum_v4_damped.py

# Train the Classical V4 model
python transformer_classicalv4_damped.py

# Train the baseline transformer model
python transformer_attention_damped.py

Conclusion

QASA represents a significant step forward in leveraging quantum computing principles for practical machine learning applications. The hybrid approach demonstrates that even with today's quantum computing capabilities, we can achieve meaningful improvements over classical methods in predicting complex time series data.

The results suggest that as quantum hardware continues to advance, hybrid quantum-classical models like QASA will likely offer even greater advantages in accuracy, efficiency, and capability for a wide range of predictive modeling tasks.

Citation

If you use this code for your research, please cite:

@article{chen2025qasa,
  title={Quantum Adaptive Self-Attention for Quantum Transformer Models},
  author={Chen, Chi-Sheng and En-Jui Kuo},
  journal={arXiv preprint arXiv:2504.05336},
  year={2025}
}

About

QASA: Quantum Adaptive Self-Attention for Quantum Transformer Models.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages