Why we participate

Credit scoring and other black box models in finance shape our life tremendously. They are the reason one might not receive a credit card or a loan or might not be considered a business partner. Since both, models and data are often owned by private companies, the ones who are affected by them have no insight and no handle to find biases within these. We want to change that. We deeply believe that each powerful tool needs to be controlled thoroughly and that people must have the opportunity to understand and dispute its negative impacts.

What we offer

We want to empower people to understand black box models and to find their underlying biases. Therefore we offer a public dataset example so that users can build trust to our approach.

How it works

We collect data where we know the in- and output. The output is generated by the black box model. Since we don't have access to this model (and we never will have, due to it's being under private ownership) we train our own model, which shall approximate the black box model as good as possible. Once we are close enough to the black box model we use advanced methods from game theory developed by Nobel laureate Adam Shapely to explain it's outcomes. We can therefore explain how important each feature is and how changes in the inputs affects changes in the output, so that one can find out whether one would still be declined a loan if having a different gender, race or a higher education degree.

What is the road ahead

Our vision doesn't stop using publicly available data. We believe that true ownership of data is to realize what can be done with it and how it compares to personal data of others. Therefore we want to offer the chance to donate anonymous data so that we can share insights on more models.

What it does

We want to explain why machine learning models behave the way they do and show everybody affected by them what influence small changes in the features can have to the output produced Therefore we used some public data (the adult dataset) to train a model and use a game theoretical approach to explain its outcome

How We built it

We initially built it using Python and Flask however, we found that using Streamlit gives us more flexibility and time to develop and roll out our App was quicker with Streamlit. So we built our initial prototype with Streamlit and deployed it with Microsoft Azure

Challenges We ran into

The speed of running our models & our App wasn't too great, so we looked at various ways that we can improve the performance of our App

Accomplishments that we're proud of

We built our custom dashboard after overcoming many technical obstacles and difficulties along the way We also learnt how to build a super-cool dashboard using Streamlit which worked really well for us #thankyoustreamlit

What We learned

We learned a lot about working remotely and between different timezones....and also how easy it was to build a dashboard with Streamlit

What's next for Bias Analyzer

Bias Analyzer still has a long way to go We have more datasets that we can add to improve our models. Users can donate datasets to us so we can improve our models as well at the outputs that we produce The performance of our dashboard still isn't that great, so we've got a bit of work in terms of enhancing our performance with larger volumes of data We can also add in more metadata for our datasets

Share this project:

Updates