Skip to content

Latest commit

 

History

History
42 lines (25 loc) · 1.69 KB

README.md

File metadata and controls

42 lines (25 loc) · 1.69 KB

Responsible AI

We want to prototype a method for removing discriminating biases in data. First the bias has to be detected and measured. Then the data should be adjusted acordingly. E.g via undersampling/oversampling. In this way, we want to create unbiased data with which a fair AI can be trained.

Plan:

  1. find nice datasets
  2. make prediction (e.g predict pay)
  3. statistik dies das -> show bias in prediction (e.g towards woman)
  4. based on statistics adjust training set
  5. retrain
  6. profit.

Possible Data Sets:

Salary based on gender, experience, education:

Detailed Salary, Gender by name:

Other: https://www.kaggle.com/datasets/kaggle/sf-salaries/discussion/45482 way to evaluate pay gap in ethnics

https://www.ons.gov.uk/peoplepopulationandcommunity/culturalidentity/ethnicity/datasets/2011censusanalysisethnicityandthelabourmarket economic activity by gender and ethnicity

https://www.ons.gov.uk/peoplepopulationandcommunity/culturalidentity/ethnicity/datasets/ethnicgroupbyeconomicactivitystatusandoccupationenglandandwalescensus2021

https://www.ons.gov.uk/peoplepopulationandcommunity/culturalidentity/ethnicity/datasets/ethnicgroupbyhighestlevelqualificationenglandandwalescensus2021

https://www-genesis.destatis.de/genesis/online?operation=abruftabelleBearbeiten&levelindex=2&levelid=1684845403062&auswahloperation=abruftabelleAuspraegungAuswaehlen&auswahlverzeichnis=ordnungsstruktur&auswahlziel=werteabruf&code=62111-0005&auswahltext=&nummer=7&variable=7&name=ALT035&werteabruf=Werteabruf#abreadcrumb