期刊文献+

搜索引擎检索系统质量评估 被引量:10

On Retrieval System Evaluation of Search Engines
下载PDF
导出
摘要 搜索引擎检索系统的质量评估对传统信息检索系统评估带来了新的研究问题·利用Tiangwang搜索引擎查询日志,按类别构造评估查询集,用人工判别相关性的方法对3个搜索引擎进行了检索质量评估·实验用InfoMall系统提供的历史网页服务消除不同搜索引擎搜集系统收集网页集合的差异,得到如下结论:①评测员之间的差异很大,但评估实验结果保持稳定;②使用连续型的相关度评分以及对应的评估指标比二元相关度评分及指标具有更好的区分能力;③使用50左右规模的查询集合和DCG这样的连续型评估指标可以有效进行评估实验· Evaluation of Web search brings challenges into the traditional evaluation methods of information retrieval systems. In this paper, the query set with different user's information categories is constructed by analyzing the query log of Tianwang search engine. In the evaluation experiments for three popular search engines, the differences of indexed document sets are reduced by filtering the query results on the InfoMall Web archive. Experiments show that: ①Significant differences are found in voluntary assessors, but the results of evaluation keep stable, ②Continuous relevant scores and corresponding measures have better distinction capability than the binary ones, and ③Query set with size of 50 is enough for the evaluation measure DCG in the Web search evaluation.
作者 彭波 闫宏飞
出处 《计算机研究与发展》 EI CSCD 北大核心 2005年第10期1706-1711,共6页 Journal of Computer Research and Development
基金 国家自然科学基金重点项目(60435020) 教育部博士点基金项目(20030001076)
关键词 搜索引擎 信息检索 评估 search engine information retrieval evaluation
  • 相关文献

参考文献16

  • 1T. Saracevic. Evaluation of evaluation in information retrieval.In: Proc. 18th Annual Int'l ACM SIGIR Conf. Research and Development in Information Retrieval, SIGIR Forum (ACM Special Interest Group on Information Retrieval). New York:ACM Press, 1995. 137~146.
  • 2J. Zobel. How reliable are the results of large-scale information retrieval experiments? In: Proc. 1998 21st Annual Int'l ACM SIGIR Conf. Research and Development in Information Retrieval(SIGIR'98). New York: ACM Press, 1998. 307~314.
  • 3E.M. Voorhees, C. Buckley. The effect of topic set size on retrieval experiment error. In: Proc. 25th Annual Int'l ACM SIGIR Conf. Research and Development in Information Retrieval,SIGIR Forum (ACM Special Interest Group on Information Retrieval). New York: ACM Press, 2002. 316~323.
  • 4C. Buckley, E. M. Voorhees. Evaluating evaluation measure stability. In: Proc. 23rd Int'l ACM SIGIR Conf. on Research and Development in Infornation Retrieval (SIGIR 2000) . New York: ACM Press, 2000. 33~40.
  • 5E.M. Voorhees. Evaluation by highly relevant documents. In:Proc. 24th Annual Int'l ACM SIGIR Conf. Research and Development in Information Retrieval, SIGIR Forum (ACM Special Interest Group on Information Retrieval). New York:ACM Press, 2001. 74~82.
  • 6G.V. Cormack, C. R. Palmer, C. L. A. Clarke. Efficient construction of large test collections. In: Proc. 1998 21st Annual Int'l ACM SIGIR Conf. Research and Development in Information Retrieval ( SIGIR' 98 ) . New York: ACM Press,1998. 282~289.
  • 7D. Hawking, N. Craswell, P. Thistlewaite, et al. Results and challenges in Web search evaluation. Computer Networks, 1999,31(11): 1321~1330.
  • 8D. Hawking, N. Craswell, P. Bailey, et al, Measuring search engine quality. Information Retrieval, 2001, 4(1): 33~59.
  • 9A. Singhal, M. Kaszkiel. A case study in web search using TREC algorithms. In: Proc. 10th International Conference on World Wide Web. New York: ACM Press, 2001. 708~716.
  • 10N. Craswell, D. Hawking, R. Wilkinson, et al. Overview of the TREC-2003 Web Track. TREC 2003, Gaithersburg, Maryland,USA, 2003.

同被引文献123

引证文献10

二级引证文献23

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部