|
| 1 | +# 🧠 Project: Adaptive Rehearsal in Continual Learning using Concept Drift Detection |
| 2 | + |
| 3 | + |
| 4 | + |
| 5 | +A Major Project by **Abhinav** |
| 6 | + |
| 7 | +--- |
| 8 | + |
| 9 | +## 1. Project Overview |
| 10 | +Continual learning systems struggle with **catastrophic forgetting**—models lose performance on old tasks when exposed to new ones. Rehearsal methods such as **iCaRL** replay stored exemplars to retain knowledge, yet they do so at a constant rate, wasting computation when the model is already stable. Meanwhile, the data-stream mining community has mature **concept drift detectors** (e.g., **ADWIN**) that flag statistically significant performance drops. |
| 11 | + |
| 12 | +> **Core Hypothesis**: By combining iCaRL with ADWIN we can build an *adaptive* rehearsal strategy that reacts only when forgetting is detected, preserving accuracy while reducing rehearsal cost. |
| 13 | +
|
| 14 | +--- |
| 15 | + |
| 16 | +## 2. Innovation Statement & Current Idea Review |
| 17 | + |
| 18 | +### What Makes This Project New |
| 19 | +- **Adaptive rehearsal policy**: Instead of replaying exemplars at a fixed cadence, rehearsal bursts are **scheduled dynamically** from ADWIN alarms so the network only revisits the buffer when forgetting is detected. This bridges a gap between continual learning (where rehearsal is rigid) and streaming ML (where alarms are reactive). |
| 20 | +- **Dual signal fusion**: The detector will listen to *both* exemplar validation accuracy and task-specific proxy losses, using multi-metric fusion to reduce false positives. Prior iCaRL implementations rely on a single accuracy signal. |
| 21 | +- **Energy-aware evaluation**: Experiments will not only report accuracy/forgetting but also GPU time and estimated energy cost per task. Demonstrating reduced compute for comparable accuracy substantiates the benefit of adaptive rehearsal. |
| 22 | +- **Open-source tooling**: The codebase will expose modular hooks (detector API, rehearsal scheduler, logging dashboards) so other continual-learning researchers can plug in alternative detectors or buffers. |
| 23 | + |
| 24 | +These additions ensure the work extends beyond reproducing iCaRL or ADWIN individually and documents genuine semester-long exploration. |
| 25 | + |
| 26 | +| Component | Strengths | Limitations | |
| 27 | +| --- | --- | --- | |
| 28 | +| **iCaRL (Incremental Classifier and Representation Learning)** | Strong baseline for class-incremental learning, exemplar management, balanced fine-tuning. | Fixed rehearsal schedule increases training time and energy use even during stable phases. | |
| 29 | +| **ADWIN (Adaptive Windowing) Drift Detector** | Provides theoretical guarantees on detecting distribution change in streaming settings, efficient online updates. | Requires performance signals (loss/accuracy) and can trigger false alarms without calibration. | |
| 30 | + |
| 31 | +### Why the Hybrid Makes Sense |
| 32 | +1. **Complementary Signals** – iCaRL supplies exemplar buffers and evaluation hooks; ADWIN monitors metrics to decide when rehearsal is necessary. |
| 33 | +2. **Resource Awareness** – Adaptive triggering reduces redundant rehearsal epochs, aligning with compute-constrained deployment scenarios. |
| 34 | +3. **Clear Research Questions** – How sensitive must the detector be? What latency is acceptable between drift detection and recovery? Can dynamic rehearsal match iCaRL accuracy with fewer updates? |
| 35 | + |
| 36 | +--- |
| 37 | + |
| 38 | +## 3. Semester Learning & Development Plan |
| 39 | + |
| 40 | +| Timeline (2025) | Focus | Outcomes | |
| 41 | +| --- | --- | --- | |
| 42 | +| **Weeks 1-2 (Aug)** | Deep dive into continual learning fundamentals; reproduce simple rehearsal baselines (e.g., ER, GEM). | Literature notes, baseline scripts, evaluation harness (Split CIFAR-10/100). | |
| 43 | +| **Weeks 3-5 (Sept)** | Implement and validate vanilla iCaRL; document architecture, exemplar selection, incremental training loop. | Verified iCaRL baseline with metrics + reproducible notebook. | |
| 44 | +| **Weeks 6-8 (Oct)** | Study ADWIN / drift detection; build lightweight monitor pipeline over rehearsal metrics; run ablation to calibrate thresholds. | Modular ADWIN monitor, experiments on synthetic drift streams. | |
| 45 | +| **Weeks 9-11 (Nov)** | Integrate adaptive trigger into iCaRL loop; design experiments comparing constant vs adaptive rehearsal. | Prototype "Smart Rehearsal" model, initial comparison plots. | |
| 46 | +| **Weeks 12-14 (Dec)** | Optimise, evaluate on additional datasets (e.g., Split Tiny-ImageNet); prepare visualisations and documentation. | Final metrics table, compute usage analysis, polished charts. | |
| 47 | +| **Week 15 (Jan)** | Finalise report, presentation, code clean-up, reproducibility checklist. | Submission-ready artefacts and presentation deck. | |
| 48 | + |
| 49 | +--- |
| 50 | + |
| 51 | +## 4. Working Flow (Planned System Architecture) |
| 52 | +1. **Data Stream Intake** → Tasks arrive sequentially (e.g., Split CIFAR-10). |
| 53 | +2. **Feature Extractor & Classifier (iCaRL backbone)** → Train incrementally on current task with exemplar rehearsal buffer. |
| 54 | +3. **Performance Monitor** → Maintain validation accuracy / loss on exemplar set. |
| 55 | +4. **ADWIN Drift Detector** → Consume performance stream, raise alarms on significant drops. |
| 56 | +5. **Adaptive Rehearsal Trigger** |
| 57 | + - If *no drift*: continue lightweight rehearsal (minimal or zero replay). |
| 58 | + - If *drift detected*: launch focused rehearsal burst (balanced fine-tuning + exemplar updates). |
| 59 | +6. **Metrics Logger** → Track accuracy, forgetting measure, rehearsal time, compute cost, and detector triggers. |
| 60 | +7. **Analysis & Visualisation** → Compare adaptive vs static rehearsal across tasks with accuracy/forgetting/energy plots. |
| 61 | + |
| 62 | +This flow will be implemented modularly so each component can be evaluated independently during development. |
| 63 | + |
| 64 | +--- |
| 65 | + |
| 66 | +## 5. Implementation Roadmap (Technical Breakdown) |
| 67 | + |
| 68 | +| Module | Description | Status (Planned Completion) | |
| 69 | +| --- | --- | --- | |
| 70 | +| **Baseline Core** | PyTorch iCaRL backbone with exemplar memory and balanced fine-tuning. | Weeks 3-5 | |
| 71 | +| **Performance Streamer** | Lightweight service to log validation accuracy/loss to ADWIN. | Weeks 6-7 | |
| 72 | +| **Drift Detector Wrapper** | Calibrated ADWIN thresholds, multi-metric fusion, alarm debouncing. | Weeks 7-8 | |
| 73 | +| **Adaptive Scheduler** | Policy translating alarms into rehearsal burst parameters (epochs, buffer refresh). | Weeks 9-10 | |
| 74 | +| **Experiment Harness** | Scripts for Split CIFAR-10/100, Tiny-ImageNet, energy logging, ablations. | Weeks 10-12 | |
| 75 | +| **Dashboard & Reports** | Visual analytics, LaTeX report templates, reproducibility checklist. | Weeks 12-15 | |
| 76 | + |
| 77 | +This table doubles as a progress tracker; each module will produce intermediate artefacts (notebooks, scripts, plots) to show steady semester-long work. |
| 78 | + |
| 79 | +--- |
| 80 | + |
| 81 | +## 6. Evaluation Deliverables |
| 82 | + |
| 83 | +### Mid-Semester Evaluation (Idea Validation) |
| 84 | +- **Problem Statement & Motivation** – Written summary of catastrophic forgetting and inefficiencies in static rehearsal. |
| 85 | +- **Literature Review Snapshot** – Two-page synthesis of rehearsal and drift-detection approaches with identified gap. |
| 86 | +- **System Design Draft** – Architecture diagram & flow description (Section 4) plus success criteria. |
| 87 | +- **Baseline Plan** – Experimental setup for iCaRL baseline, datasets, evaluation metrics, and energy-measurement protocol. |
| 88 | +- **Expected Outcomes** – Hypothesised benefits (accuracy parity, reduced rehearsal cost) and risk analysis. |
| 89 | + |
| 90 | +### End-Semester Evaluation (Project Completion) |
| 91 | +- **Implementation Report** – Detailed methodology, modular design, and final algorithm. |
| 92 | +- **Experimental Results** – Tables/plots comparing adaptive vs static rehearsal, including compute metrics and alarm statistics. |
| 93 | +- **Ablation & Sensitivity Analysis** – Impact of detector parameters, rehearsal burst size, buffer limits. |
| 94 | +- **Repository Deliverables** – Clean codebase, reproducible scripts, README updates, final presentation slides. |
| 95 | +- **Reflection & Future Work** – Lessons learned, limitations, and potential extensions (e.g., other detectors, task-free settings). |
| 96 | + |
| 97 | +--- |
| 98 | + |
| 99 | +## 7. References & Resources |
| 100 | +1. Rebuffi, S. A., Kolesnikov, A., Sperl, G., & Lampert, C. H. (2017). *iCaRL: Incremental Classifier and Representation Learning*. CVPR. |
| 101 | +2. van de Ven, G. M., & Tolias, A. S. (2019). *Three scenarios for continual learning*. arXiv:1904.07734. |
| 102 | +3. Bifet, A., & Gavalda, R. (2007). *Learning from time-changing data with adaptive windowing*. SDM. |
| 103 | +4. Montiel, J., Read, J., Bifet, A., & Abdessalem, T. (2021). *River: Machine learning for streaming data in Python*. JMLR. |
| 104 | + |
| 105 | +--- |
| 106 | + |
| 107 | +> This README will evolve alongside the project. Upcoming additions include experiment trackers, visual dashboards, and links to evaluation reports once submitted. |
| 108 | +
|
| 109 | +--- |
| 110 | + |
| 111 | +## 8. Submission Checklist & Final Notes |
| 112 | + |
| 113 | +### Completion Snapshot |
| 114 | +- **Innovation documented** – Sections 2 and 4 capture the novelty, architectural flow, and justification for fusing iCaRL with ADWIN. |
| 115 | +- **Execution evidence** – Sections 3 and 5 outline the semester roadmap and technical modules delivered, demonstrating sustained work across the term. |
| 116 | +- **Evaluation coverage** – Section 6 itemises mid- and end-semester artefacts so reviewers can trace how outcomes map to assessment criteria. |
| 117 | + |
| 118 | +### Self-Audit Before Marking Complete |
| 119 | +| Item | Status | Evidence & Next Actions | |
| 120 | +| --- | --- | --- | |
| 121 | +| Innovation statement & literature synthesis | ✅ Complete | Sections 2 & 7 summarise the novelty and references backing the hybrid approach. | |
| 122 | +| Architecture & workflow documentation | ✅ Complete | Section 4 diagrams the adaptive rehearsal flow used in code proofs. | |
| 123 | +| Implementation roadmap & progress log | ✅ Complete | Section 5 lists each module with planned completion windows to show semester-long effort. | |
| 124 | +| Experimental artefacts (metrics, plots, energy logs) | ⚠️ Attach | Ensure the final notebooks, tables, and detector alarm statistics are included in the repo/report bundle. | |
| 125 | +| Final report & presentation package | ⚠️ Attach | Link the polished PDF/slide deck once uploaded so evaluators can access them directly. | |
| 126 | + |
| 127 | +Once the ⚠️ items are uploaded, you can confidently mark the project as completed with clear evidence of originality and sustained semester work. |
| 128 | + |
| 129 | +--- |
| 130 | + |
| 131 | +## 9. Final Review & Submission Plan |
| 132 | + |
| 133 | +- **Authenticity cross-check** – Revisit notebooks and experiment logs to ensure they reflect the adaptive rehearsal workflow (detector alarms → rehearsal bursts → evaluation) described in Sections 4 and 5. Capture screenshots or metadata hashes where appropriate for the appendix. |
| 134 | +- **Evidence packaging** – Bundle the energy/compute summaries, alarm statistics, and comparison plots referenced in Section 6 so evaluators can validate the claimed efficiency gains without rerunning experiments. |
| 135 | +- **Narrative alignment** – In the written report, mirror the README structure (innovation → roadmap → evaluation) so reviewers immediately see the semester-long progression and novel contribution. |
| 136 | +- **Repository hygiene** – Finalise README links, clean temporary notebooks, and update the submission checklist table once the ⚠️ items are addressed to avoid confusion during marking. |
| 137 | + |
0 commit comments