Skip to content

cherikakaushal/when-systems-break

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

32 Commits
 
 
 
 
 
 

Repository files navigation

when-systems-break

A research-style project exploring how machine learning systems behave under real-world imperfect data conditions.


Overview

Real-world data is rarely clean or complete.

This project focuses on understanding how machine learning models behave when:

  • data is noisy
  • data is missing
  • important features are removed
  • conditions change gradually

The goal is not just to optimize accuracy, but to study how systems respond when things go wrong.


Blogs


Experiments

  1. Noise Injection
    Introduced randomness to observe impact on accuracy

  2. Missing Data Simulation
    Tested how incomplete inputs affect model performance

  3. Feature Importance Analysis
    Identified which features influence predictions most

  4. Model Comparison
    Compared how different models respond to imperfect data

  5. Feature Removal Sensitivity
    Removed key features to observe system degradation

  6. Robustness Curve
    Measured how accuracy changes as noise increases


Key Insights

  • Not all imperfections affect systems equally
  • Removing critical features causes sharper failure than random noise
  • Models can appear stable while becoming internally unreliable
  • Performance degradation is gradual, not always immediate

Visualizations

Feature Importance

Feature Importance

Noise vs Accuracy

Noise Curve


Tech Stack

  • Python
  • NumPy, Pandas
  • Scikit-learn
  • Matplotlib

Goal

To move beyond “building models” and toward understanding how systems behave when things go wrong.

About

Exploring how machine learning systems behave under noisy, missing, and imperfect data.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages