Final Year Project (PFE) | Engineer’s Degree Dissertation This repository presents an award-winning Intelligent Digital Twin designed to solve the energy inefficiency of urban parking guidance systems using Deep Reinforcement Learning.
This research was recognized for its innovation and impact:

Modern smart parkings use LED-based guidance systems that consume massive amounts of energy. This project introduces a proactive Digital Twin orchestrated by a Reinforcement Learning (RL) agent.
I developed a standalone web-based simulator to demonstrate the energy logic of the 4 different scenarios. No installation is required.
The project features a decoupled architecture ensuring real-time synchronization between the virtual model and the physical simulation.

The “Quantum Parking Matrix” is a high-fidelity web dashboard (HTML/CSS/JS) that visualizes 150 parking spots and AI decisions in real-time via MQTT.

The agent (Deep Q-Network) learns to assign vehicles to specific floors to minimize energy-intensive states while maintaining service quality.
| Learning Convergence | Training KPIs Evolution |
|:—:|:—:|
|
|
|
| Stable convergence achieved after 80 episodes | Drastic reduction in variance and failures |
The effectiveness of the Intelligent Digital Twin was evaluated through a rigorous 24-hour simulation, comparing three distinct management strategies.
To mirror real-world conditions, we modeled a stochastic diurnal traffic cycle representing a typical urban business area, featuring morning and evening peak hours.

The agent uses an Epsilon-Decay strategy to manage the exploration-exploitation trade-off. It starts by exploring the parking environment and gradually transitions to exploiting its learned knowledge for optimal vehicle assignment.

Our RL-based strategy (blue) was compared against the “Always-On” baseline (red) and the “All-Green” heuristic (green).
| Instantaneous Power Demand (W) | Cumulative Energy Consumption (Wh) |
|---|---|
![]() |
![]() |
| Real-time adaptation to traffic | Final 81.8% efficiency gain |
Key Finding: While the heuristic approach reduces waste, only the Reinforcement Learning agent manages to keep the infrastructure in “standby mode” during low-traffic periods, leading to a massive 81.8% energy saving.
/docs: Full dissertation report and Best Paper Award documentation./rl-engine: Python scripts and Jupyter Notebooks for agent training and traffic simulation.
q_table_final.pkl: The pre-trained brain of the system./ui-digital-twin: Source code of the interactive Web Dashboard./screenshots: Full gallery of research results and UI previews.Developed by Lynda OUALLAM - Academic Year 2024/2025.
Supervised by Dr. Ahcene BOUNCEUR & Pr. Faïçal AZOUAOU.