Top Banner
Ryen W. White Microsoft Research [email protected] m Steven M. Drucker Microsoft Live Labs [email protected] m
23

Ryen W. White Microsoft Research [email protected] Steven M. Drucker Microsoft Live Labs [email protected].

Mar 27, 2015

Download

Documents

Diego Skinner
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Ryen W. WhiteMicrosoft Research

[email protected]

Steven M. DruckerMicrosoft Live Labs

[email protected]

Page 2: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Jack searches for “digital cameras”. He knows what he wants (he’s did this before) and goes straight to a particular page.Jill searches for “digital cameras”. She is unsure of what she’s looking for, and wants to explore the options.

Both type “digital cameras” into a search engine…

Page 3: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Same interface support for Jack and Jill regardless of prior experience or taskNo support for decisions beyond this page

Jack sees: Jill sees:

Page 4: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Search engines adopt a “one-size-fits-all” approach to interface design

Users benefit from familiarityCost to user-interface designers minimizedLimited support for next steps

Important to understand what users are doing beyond the result page, and in what ways “one-size-fits-all” can be enhanced

Page 5: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Approx. 2500 consenting usersInstrumented client-side logging of URLs visited, timestamps, referral information, etc.20 weeks (Dec 05 – April 06)

Analysis focused on:Interaction patterns (e.g., SBBBSBSbBbBBB)Features of interaction (e.g., time spent)Domains visited

Page 6: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Our analysis based on browse trailsOrdered series of page views from opening Internet Explorer until closing browserExample trail as Web Behavior Graph:

S2

S3 S4

X

S7 S8

S9S7

S7

S10

S1 S5

S1

hotmail.com

S2

S6

Non-result page

Search engine result page

Page 7: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Search trails situated within browse trails:

Initiated with a query to top-5 search engineCan contain multiple queriesTerminate with:

Session timeout Visit homepageType URLCheck Web-based email or logon to online service

S2

S3 S4

X

S7 S8

S9S7

S7

S10

S1 S5

S1

hotmail.com

S2

S6

Page 8: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

< 30% of interaction with search engines> 70% of interaction is forward motionTakeaway: Post SE interaction important

SearchengineInteractions

Interactionsbeyond thesearch engine

Page 9: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

We studied all search interactions (w/ search engine and post-engine) to better understand:

User Interaction VariabilityExtent of differences within and between users

Query Interaction VariabilityExtent of differences within and between queries

Page 10: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Differences in:Interaction patternsFeatures of the interactionDomains visited

Within each user How consistent is user X?

Between all usersHow consistent are all users together?

Page 11: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

1. Represent all users’ trails as strings

2. For each user compute Edit Distance from each trail to every other trail

S2

S3 S4

X

S7 S8

S9S7

S7

S10

S1 S5

S1

EmailS B B B Sb S

S = searchB = browseb = back

S6

S2

Page 12: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

3. Average Edit Distance from each trail to other trails, e.g.,

4. Trail with smallest avg. distance most representative of user interaction patterns

S S B B

S B B S

B B B B

Trail 1

Trail 2

Trail 3

S

S

b S

b B b

S

ED(1,2) = 4

ED(2,1) = 4

ED(3,1) = 4

ED(1,3) = 4

ED(2,3) = 5

ED(3,2) = 5

Average

4

4.5

4.5

Page 13: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

5. Avg. Edit Distance of representative trail:

Low = user interaction patterns consistentHigh = user interaction patterns variable

Interaction variance“Navigators” “Explorers”

# users

3.2 94.4

Average = 20.1Median = 16

8010

Boundaries fuzzy

Page 14: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Consistent patterns (most trails same), e.g.,

Most users interact like this sometimes – Navigators interact like this most of the time

Few deviations or regressionsTackled problems sequentiallyMore likely to revisit domains

Few deviations/regressionsSearched sequentiallyLikely to revisit domains

Cleary defined subtaskse.g., 1. Comparison2. Review

S1 S2 S3 S4

S2

S2

S5 S6

S7 S9S8Sub-trail 2:“Review”

Sub-trail 1:“Compare”

dpreview.com amazon.com

digital cameras

amazon

Sub-trail 1:“Compare”

Page 15: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Variable patterns (most trails different), e.g.,

Almost all of their trails different

Explorers:Trails branched frequentlySubmitted many queriesVisited many new domains

canon.com

amazon.com

S1 S3 S4

S3

S5

dpreview.com

howstuffworks.com

S2

S2 S6

S5 S8

S6 S9

S1 S10 S11

S10 S12 S13 S14

amazon

pmai.org

digitalcamera-hq.com

digital cameras

digital camera canon

canon lenses

S7

S6

canon lenses

Page 16: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Studied features of the trailsTime spent, Num. queries, Num. steps, Branchiness, Num. revisits, Avg. branch len.

Factor analysis revealed three factors that captured 80.6% of variance between users

Forward and backward motion (52.5%)“Branchiness” (i.e., how many sub-trails?) (17.4%)Time (10.7%)

Factors can be used to differentiate users

Page 17: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Proportion of domains visited that were unique, computed as:Num unique domains / Num of

domains17% had variance of .1 or less

Most of the domains visited were revisits

2% had variance of .9 or moreMost of the domains visited were unique

Roughly same users at extremities as with interaction variance (≥ 86% overlap)

Page 18: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

“Navigators” and “Explorers” extreme casesAll users exhibit extreme behavior at timesLearn from Navigators and ExplorersDecide what interface support they needOffer this support as optional functionality to all users in a search “toolkit”Default search interface does not changeMore on this later…

Page 19: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Focus on queries rather than usersIf interaction variable we may need:

Tailored search interfaces for different queriesQuery segmentation and tailored ranking

385 queries with sufficient interaction data

Submitted at least 15 times by at least 15 unique participantsDistribution of informational / navigational matched that of much larger query logs

Page 20: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Same analysis as earlier, but with queriesLow variance (based on ED):

Queries generally navigational (e.g., “msn”)

High variance:Undirected, exploratory searchesSearches where people’s tastes differ (e.g., travel, art)

Nav. and Explor. query behavior similar to Nav. and Explor. user behavior

Page 21: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

TeleportationThey follow short directed search trails

Jump users direct to targets, offer shortcuts

Personal Search HistoriesThey conduct the same search repeatedly

Present previous searches on search engine

Interaction HubsThey rely on important pages within domains

Surface these domains as branching points

Page 22: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Guided Tours and Domain IndicesThey visit multiple domains

Offer list of “must see” domains for query topic

Predictive RetrievalThey want serendipity

Automatically retrieve novel information

Support for Rapid RevisitationThey use “back” and visit previous pages a lot

Mechanisms to return them to branching points

Page 23: Ryen W. White Microsoft Research ryenw@microsoft.com Steven M. Drucker Microsoft Live Labs sdrucker@microsoft.com.

Conducted a longitudinal study of Web search behavior involving 2500 usersFound differences in interaction flow within and between users and within and between queriesIdentified two types of user with extremely consistent / variable interaction patternsLearned how to support these users that can be used to help everyone