/
Improving relevance prediction by addressing biases and Improving relevance prediction by addressing biases and

Improving relevance prediction by addressing biases and - PowerPoint Presentation

aaron
aaron . @aaron
Follow
412 views
Uploaded On 2017-06-18

Improving relevance prediction by addressing biases and - PPT Presentation

sparsity in web search click data Qi Guo Dmitry Lagun Denis Savenkov Qiaoling Liu qguo3 dlagundenissavenkov qiaolingliu emoryedu Mathematics amp Computer ID: 560777

query relevance bias features relevance query features bias click prediction url session position clicks dwell problems shows queries time

Share:

Link:

Embed:

Download Presentation from below link

Download Presentation The PPT/PDF document "Improving relevance prediction by addres..." is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

Improving relevance prediction by addressing biases and sparsity in web search click data

Qi Guo, Dmitry Lagun, Denis Savenkov, Qiaoling Liu[qguo3,dlagun,denis.savenkov,qiaoling.liu]@emory.eduMathematics & Computer Science, Emory UniversitySlide2

Relevance Prediction ChallengeSlide3

Web Search Click DataSlide4

Relevance prediction problemsPosition-biasPerception-bias

Query-biasSession-biasSparsitySlide5

Relevance prediction problems: position-biasCTR is a good indicator of document relevance

search results are not independentdifferent positions – different attention [Joachims+07]

Normal Position

Percentage

Reversed Impression

PercentageSlide6

Relevance prediction problems: perception-biasUser decides to click or to skip based on snippets

“Perceived” relevance may be inconsistent with “intrinsic” relevanceSlide7

Relevance prediction problems: query-biasqueries are different

Ctr for difficult queries might not be trustworthyFor infrequent queries we might not have enough dataNavigational vs informationalDifferent queries – different time to get the answerQueries:P versus NPhow to get rid of acneWhat is the capital of Honduras

grand hyatt seattle zip code

Why

am I still

single

why is hemp illegalSlide8

Relevance prediction problems: session-biasUsers are different

Query ≠ Intent30s dwell time might not indicate relevance for some types of users [Buscher et al. 2012]Slide9

Relevance prediction problems: sparsity

1 show – 1 clicks means relevant document?What about 1 show – 0 clicks, non-relevant?For tail queries (non-frequent doc-query-region) we might not have enough clicks/shows to make robust relevance predictionSlide10

Click ModelsUser browsing probability models

DBN, CCM, UBM, DCM, SUM, PCC Don’t work well for infrequent queriesHard to incorporate different kind of featuresSlide11

Our approachClick Models are goodBut we have different types of information we want to combine in our model

Let’s use Machine LearningML algorithms:AUCRankGradient Boosted Decision Trees (pGBRT implementation) – regression problemSlide12

DatasetYandex Relevance Prediction Challenge data:

Unique queries: 30,717,251Unique urls: 117,093,258Sessions: 43,977,8594 Regions:Probably: Russia, Ukraine, Belarus & KazakhstanQuality measureAUC - Area Under CurvePublic and hidden test subsetsHidden subset labels aren’t currently availableSlide13

Features: position-biasp

er position CTR“Click-SkipAbove” and similar behavior patternsDBN (Dynamic Bayesian Network)“Corrected” shows: shows with clicks on the current position or below (cascade hypothesis)Slide14

Features: perception-bias

Post-click behaviorAverage/median/min/max/std dwell-timeSat[Dissat] ctr (clicks with dwell >[<] threshold)Last click ctr (in query/session)Time before clickSlide15

Features: query-biasQuery features

: ctr, no click shows, average click position, etc.Url features normalization:>average query dwell time# clicks before click on the given urlThe only click in query/showsUrl dwell/total dwellSlide16

Features: session-bias

Url features normalization>average session dwell time#clicks in session#longest clicks in session/clicksdwell/session durationSlide17

Features: sparsity

Pseudo-counts for sparsityPrior information: original ranking (average show position; shows on i-th pos / shows)Back-offs (more data – less precise): url-query-regionurl-queryurl-regionurl

query-regionquerySlide18

Parameter tuning

Later experiments:

5-fold CV

Tree height h=3

Iterations: ~250

Learning rate: 0.1Slide19

Results (5-fold CV)

Baselines:Original ranking (average show position): 0.6126Ctr: 0.6212Models:

AUC-Rank: 0.6337

AUC-Rank + Regression: 0.6495

Gradient Boosted Regression Trees: 0.6574Slide20

Results (5-fold CV)

session and perception-bias features are the most important relevance signalsQuery-bias features don’t work well by itself but provide important information to other feature groupsSlide21

Results (5-fold CV)

q

uery-

url

level features are the best trade-off between precision and

sparsity

r

egion-

url

features have both problems: sparse and not preciseSlide22

Feature importanceSlide23

ConclusionsSparsity: Back-off strategy to address data

sparsity = +3.1% AUC improvementPerception-bias: dwell-time is the most important relevance signal (who would’ve guessed )Session-bias: session-level normalization helps to improve relevance prediction qualityQuery-bias: query-level information gives an important additional information that helps predict relevancePosition-bias features are usefulSlide24

THANK YOUThanks to the organizers for such an interesting challenge & open dataset!

Thank you for listening!P.S. Do not overfit 