Top Banner
When Recommendation Systems Go Bad Evan Estola 12/10/15
38

When recommendation systems go bad

Apr 16, 2017

Download

Technology

Evan Estola
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: When recommendation systems go bad

When Recommendation Systems Go Bad

Evan Estola12/10/15

Page 2: When recommendation systems go bad

About Me

● Evan Estola

● Senior Machine Learning Engineer @ Meetup

[email protected]

● @estola

Page 3: When recommendation systems go bad

Agenda

1. Meetup

2. Recommendation Systems

3. What can go wrong?

4. How to prevent it

Page 4: When recommendation systems go bad
Page 5: When recommendation systems go bad
Page 6: When recommendation systems go bad
Page 7: When recommendation systems go bad
Page 8: When recommendation systems go bad

We want a world full of real, local community.Women’s Veterans Meetup, San Antonio, TX

Page 9: When recommendation systems go bad
Page 10: When recommendation systems go bad
Page 11: When recommendation systems go bad
Page 12: When recommendation systems go bad
Page 13: When recommendation systems go bad

Recommendation Systems: Collaborative Filtering

Page 15: When recommendation systems go bad

You just wanted a kitchen scale, now Amazon thinks you’re a drug dealer

Page 16: When recommendation systems go bad

Recommendation Systems: Rating Prediction

● Netflix prize

● How many stars would user X give movie Y

● Boring

Page 17: When recommendation systems go bad

Recommendation Systems: Learning To Rank

● Active area of research

● Use ML model to solve a ranking problem

● Pointwise: Logistic Regression on binary label, use output for ranking

● Listwise: Optimize entire list

● Performance Metrics

○ Mean Average Precision

○ P@K

○ Discounted Cumulative Gain

Page 18: When recommendation systems go bad
Page 19: When recommendation systems go bad

Data Science impacts

lives

● News you’re exposed to

● Friend’s Activity/Facebook feed

● Job openings you find/get

● If you can get a loan

● Route you take to get somewhere

● If you can get a ride

● Movies you watch on Netflix

● Products you buy on Amazon

● Price you pay for things

● If a product is available to you at all

● Apps you download

● Ads you see

Page 20: When recommendation systems go bad
Page 22: When recommendation systems go bad
Page 23: When recommendation systems go bad

Ego

● Member/customer/user first

● Focus on building the best product,

not on being the most clever data

scientist

● Much harder to spin a positive user

story than a story about how smart

you are

Page 24: When recommendation systems go bad

● Fake profiles, track ads

● Career coaching for “200k+”

Executive jobs Ad

● Male group: 1852 impressions

● Female group: 318

Page 25: When recommendation systems go bad

● “Black-sounding” names 25% more

likely to be served ad suggesting

criminal record

Page 26: When recommendation systems go bad

Ethics

We have accepted that Machine Learning

can seem creepy, how do we prevent it

from becoming immoral?

We have an ethical obligation to not

teach machines to be prejudiced.

Page 27: When recommendation systems go bad

Data Ethics

Awareness

● Tell your friends

● Tell your coworkers

● Tell your boss

Page 28: When recommendation systems go bad

Awareness

● Start a conversation

○ Identify potentially marginalized user groups

○ Have an ethics strategy for evaluating whether to include

sensitive features

Page 29: When recommendation systems go bad
Page 30: When recommendation systems go bad

Interpretable Models

● For simple problems, simple

solutions are often worth a small

concession in performance

● Inspectable models make it easier

to debug problems in data

collection, feature engineering etc.

● Only include features that work the

way you want

● Don’t include feature interactions

that you don’t want

Page 31: When recommendation systems go bad

Logistic Regression

StraightDistanceFeature(-0.0311f),

ChapterZipScore(0.0250f),

RsvpCountFeature(0.0207f),

AgeUnmatchFeature(-1.5876f),

GenderUnmatchFeature(-3.0459f),

StateMatchFeature(0.4931f),

CountryMatchFeature(0.5735f),

FacebookFriendsFeature(1.9617f),

SecondDegreeFacebookFriendsFeature(0.1594f),

ApproxAgeUnmatchFeature(-0.2986f),

SensitiveUnmatchFeature(-0.1937f),

KeywordTopicScoreFeatureNoSuppressed(4.2432f),

TopicScoreBucketFeatureNoSuppressed(1.4469f,0.257f,10f),

TopicScoreBucketFeatureSuppressed(0.2595f,0.099f,10f),

ExtendedTopicsBucketFeatureNoSuppressed(1.6203f,1.091f,10f),

ChapterRelatedTopicsBucketFeatureNoSuppressed(0.1702f,0.252f,0.641f),

ChapterRelatedTopicsBucketFeatureNoSuppressed(0.4983f,0.641f,10f),

DoneChapterTopicsFeatureNoSuppressed(3.3367f)

Page 32: When recommendation systems go bad

Feature Engineering and Interactions

● Good Feature:○ Join! You’re interested in Tech x Meetup is about Tech

● Good Feature: ○ Don’t join! Group is intended only for Women x You are a Man

● Bad Feature:○ Don’t join! Group is mostly Men x You are a Woman

● Horrible Feature:○ Don’t join! Meetup is about Tech x You are a Woman

Meetup is not interested in propagating gender stereotypes

Page 33: When recommendation systems go bad

Ensemble Models and

Data segregation

Ensemble Models: Combine outputs of

several classifiers for increased accuracy

If you have features that are useful but

you’re worried about interaction (and

your model does it automatically) use

ensemble modeling to restrict the

features to separate models.

Page 34: When recommendation systems go bad

Ensemble Model, Data Segregation

Data:*InterestsSearchesFriendsLocation

Data:*GenderFriendsLocation

Data:Model1 PredictionModel2 Prediction

Model1 Prediction

Model2 Prediction

Final Prediction

Page 36: When recommendation systems go bad

Diversity controlled test data

● Make sure product works for

everybody

● Generate test data and evaluate

your model against it to confirm no

encapsulated prejudice

Page 37: When recommendation systems go bad

Diversity Controlled Testing

● CMU - AdFisher

○ Crawls ads with simulated user profiles

● Same technique can work to find bias in your own models!

○ Generate Test Data

■ Randomize sensitive feature in real data set

○ Run Model

■ Evaluate for unacceptable biased treatment

● Must identify what features are sensitive and what outcomes are

unwanted

Page 38: When recommendation systems go bad

You know racist computers are a bad idea

Don’t let your company invent racist computers