Towards Context-Aware Search by Learning a Very Large Variable Length Hidden Markov Model from Search Logs

Presented at: 18th International World Wide Web Conference (WWW2009)

by Huanhuan Cao, Daxin Jiang, Jian Pei, Enhong Chen, Hang Li

Webpage: http://www2009.eprints.org/20/1/p191.pdf

Capturing the context of a user's query from the previous queries and clicks in the same session may help understand the user's information need. A context-aware approach to document re-ranking, query suggestion, and URL recommendation may improve users' search experience substantially. In this paper, we propose a general approach to context-aware search. To capture contexts of queries, we learn a variable length Hidden Markov Model (vlHMM) from search sessions extracted from log data. Although the mathematical model is intuitive, how to learn a large vlHMM with millions of states from hundreds of millions of search sessions poses a grand challenge. We develop a strategy for parameter initialization in vlHMM learning which can greatly reduce the number of parameters to be estimated in practice. We also devise a method for distributed vlHMM learning under the map-reduce model. We test our approach on a real data set consisting of 1.8 billion queries, 2.6 billion clicks, and 840 million search sessions, and evaluate the effectiveness of the vlHMM learned from the real data on three search applications: document re-ranking, query suggestion, and URL recommendation. The experimental results show that our approach is both effective and efficient.

Keywords: Data Mining


Resource URI on the dog food server: http://data.semanticweb.org/conference/www/2009/paper/20


Explore this resource elsewhere: