共查询到20条相似文献,搜索用时 718 毫秒
1.
针对高校图书馆场景存在的无显式反馈、借阅数据稀疏和传统推荐算法效果不好问题,提出基于时间上下文优化协同过滤的推荐算法,包含读者阅读行为评分、时间上下文和内容兴趣变迁3个要素。在数据准备阶段,通过制定评分转化规则、设计标准化函数来构建一种基于用户行为操作的兴趣评分模型,以解决用户评分缺失问题;在推荐召回阶段,提出一种非线性的时间衰减模型来对评价矩阵进行优化,以提高推荐效果;在推荐排序阶段,提出一种兴趣捕捉模型对召回结果按照图书类别进行精排序,以缓解数据稀疏问题并进一步提高推荐效果。实验结果表明,文章提出的优化算法在Top5的F值较未经优化的协同过滤提升增幅达141%。 相似文献
2.
[目的/意义]在对MNCS和百分位数两种指标机制进行阐述、对比和分析的基础上,对百分位数指标的计算框架进行改进,提出一种动态权重的百分位数指标用于学术影响力的评价。[方法/过程]以ESI学科为评价对象,分别选取同一研究实体下的不同学科和不同研究实体下的同一学科作为两个实例进行实证研究。[结果/结论]实证结果表明这种动态权重的百分位数指标与MNCS和百分位数指标相比更能展现评价对象学术影响力的细节。 相似文献
3.
Most ranking algorithms are based on the optimization of some loss functions, such as the pairwise loss. However, these loss
functions are often different from the criteria that are adopted to measure the quality of the web page ranking results. To
overcome this problem, we propose an algorithm which aims at directly optimizing popular measures such as the Normalized Discounted
Cumulative Gain and the Average Precision. The basic idea is to minimize a smooth approximation of these measures with gradient
descent. Crucial to this kind of approach is the choice of the smoothing factor. We provide various theoretical analysis on
that choice and propose an annealing algorithm to iteratively minimize a less and less smoothed approximation of the measure of interest. Results on the Letor
benchmark datasets show that the proposed algorithm achieves state-of-the-art performances. 相似文献
4.
Web search algorithms that rank Web pages by examining the link structure of the Web are attractive from both theoretical and practical aspects. Todays prevailing link-based ranking algorithms rank Web pages by using the dominant eigenvector of certain matrices—like the co-citation matrix or variations thereof. Recent analyses of ranking algorithms have focused attention on the case where the corresponding matrices are irreducible, thus avoiding singularities of reducible matrices. Consequently, rank analysis has been concentrated on authority connected graphs, which are graphs whose co-citation matrix is irreducible (after deleting zero rows and columns). Such graphs conceptually correspond to thematically related collections, in which most pages pertain to a single, dominant topic of interest.A link-based search algorithm A is rank-stable if minor changes in the link structure of the input graph, which is usually a subgraph of the Web, do not affect the ranking it produces; algorithms A,B are rank-similar if they produce similar rankings. These concepts were introduced and studied recently for various existing search algorithms.This paper studies the rank-stability and rank-similarity of three link-based ranking algorithms—PageRank, HITS and SALSA—in authority connected graphs. For this class of graphs, we show that neither HITS nor PageRank is rank stable. We then show that HITS and PageRank are not rank similar on this class, nor is any of them rank similar to SALSA.This research was supported by the Fund for the Promotion of Research at the Technion, and by the Barnard Elkin Chair in Computer Science. 相似文献
5.
LETOR: A benchmark collection for research on learning to rank for information retrieval 总被引:2,自引:0,他引:2
LETOR is a benchmark collection for the research on learning to rank for information retrieval, released by Microsoft Research
Asia. In this paper, we describe the details of the LETOR collection and show how it can be used in different kinds of researches.
Specifically, we describe how the document corpora and query sets in LETOR are selected, how the documents are sampled, how
the learning features and meta information are extracted, and how the datasets are partitioned for comprehensive evaluation.
We then compare several state-of-the-art learning to rank algorithms on LETOR, report their ranking performances, and make
discussions on the results. After that, we discuss possible new research topics that can be supported by LETOR, in addition
to algorithm comparison. We hope that this paper can help people to gain deeper understanding of LETOR, and enable more interesting
research projects on learning to rank and related topics. 相似文献
6.
Jovan Pehcevski James A. Thom Anne-Marie Vercoustre Vladimir Naumovski 《Information Retrieval》2010,13(5):568-600
Entity ranking has recently emerged as a research field that aims at retrieving entities as answers to a query. Unlike entity
extraction where the goal is to tag names of entities in documents, entity ranking is primarily focused on returning a ranked
list of relevant entity names for the query. Many approaches to entity ranking have been proposed, and most of them were evaluated
on the INEX Wikipedia test collection. In this paper, we describe a system we developed for ranking Wikipedia entities in
answer to a query. The entity ranking approach implemented in our system utilises the known categories, the link structure
of Wikipedia, as well as the link co-occurrences with the entity examples (when provided) to retrieve relevant entities as
answers to the query. We also extend our entity ranking approach by utilising the knowledge of predicted classes of topic
difficulty. To predict the topic difficulty, we generate a classifier that uses features extracted from an INEX topic definition
to classify the topic into an experimentally pre-determined class. This knowledge is then utilised to dynamically set the
optimal values for the retrieval parameters of our entity ranking system. Our experiments demonstrate that the use of categories
and the link structure of Wikipedia can significantly improve entity ranking effectiveness, and that topic difficulty prediction
is a promising approach that could also be exploited to further improve the entity ranking performance. 相似文献
7.
Web search engines are increasingly deploying many features, combined using learning to rank techniques. However, various practical questions remain concerning the manner in which learning to rank should be deployed. For instance, a sample of documents with sufficient recall is used, such that re-ranking of the sample by the learned model brings the relevant documents to the top. However, the properties of the document sample such as when to stop ranking—i.e. its minimum effective size—remain unstudied. Similarly, effective listwise learning to rank techniques minimise a loss function corresponding to a standard information retrieval evaluation measure. However, the appropriate choice of how to calculate the loss function—i.e. the choice of the learning evaluation measure and the rank depth at which this measure should be calculated—are as yet unclear. In this paper, we address all of these issues by formulating various hypotheses and research questions, before performing exhaustive experiments using multiple learning to rank techniques and different types of information needs on the ClueWeb09 and LETOR corpora. Among many conclusions, we find, for instance, that the smallest effective sample for a given query set is dependent on the type of information need of the queries, the document representation used during sampling and the test evaluation measure. As the sample size is varied, the selected features markedly change—for instance, we find that the link analysis features are favoured for smaller document samples. Moreover, despite reflecting a more realistic user model, the recently proposed ERR measure is not as effective as the traditional NDCG as a learning loss function. Overall, our comprehensive experiments provide the first empirical derivation of best practices for learning to rank deployments. 相似文献
8.
Direct optimization of evaluation measures has become an important branch of learning to rank for information retrieval (IR).
Since IR evaluation measures are difficult to optimize due to their non-continuity and non-differentiability, most direct
optimization methods optimize some surrogate functions instead, which we call surrogate measures. A critical issue regarding
these methods is whether the optimization of the surrogate measures can really lead to the optimization of the original IR
evaluation measures. In this work, we perform formal analysis on this issue. We propose a concept named “tendency correlation”
to describe the relationship between a surrogate measure and its corresponding IR evaluation measure. We show that when a
surrogate measure has arbitrarily strong tendency correlation with an IR evaluation measure, the optimization of it will lead
to the effective optimization of the original IR evaluation measure. Then, we analyze the tendency correlations of the surrogate
measures optimized in a number of direct optimization methods. We prove that the surrogate measures in SoftRank and ApproxRank
can have arbitrarily strong tendency correlation with the original IR evaluation measures, regardless of the data distribution,
when some parameters are appropriately set. However, the surrogate measures in SVM
MAP
, DORM
NDCG
, PermuRank
MAP
, and SVM
NDCG
cannot have arbitrarily strong tendency correlation with the original IR evaluation measures on certain distributions of
data. Therefore SoftRank and ApproxRank are theoretically sounder than SVM
MAP
, DORM
NDCG
, PermuRank
MAP
, and SVM
NDCG
, and are expected to result in better ranking performances. Our theoretical findings can explain the experimental results
observed on public benchmark datasets. 相似文献
9.
Query Expansion is commonly used in Information Retrieval to overcome vocabulary mismatch issues, such as synonymy between
the original query terms and a relevant document. In general, query expansion experiments exhibit mixed results. Overall TREC
Genomics Track results are also mixed; however, results from the top performing systems provide strong evidence supporting
the need for expansion. In this paper, we examine the conditions necessary for optimal query expansion performance with respect
to two system design issues: IR framework and knowledge source used for expansion. We present a query expansion framework
that improves Okapi baseline passage MAP performance by 185%. Using this framework, we compare and contrast the effectiveness
of a variety of biomedical knowledge sources used by TREC 2006 Genomics Track participants for expansion. Based on the outcome
of these experiments, we discuss the success factors required for effective query expansion with respect to various sources
of term expansion, such as corpus-based cooccurrence statistics, pseudo-relevance feedback methods, and domain-specific and
domain-independent ontologies and databases. Our results show that choice of document ranking algorithm is the most important
factor affecting retrieval performance on this dataset. In addition, when an appropriate ranking algorithm is used, we find
that query expansion with domain-specific knowledge sources provides an equally substantive gain in performance over a baseline
system.
相似文献
Nicola StokesEmail: Email: |
10.
《Journal of Informetrics》2023,17(3):101422
Several studies have reported on metrics for measuring the influence of scientific topics from different perspectives; however, current ranking methods ignore the reinforcing effect of other academic entities on topic influence. In this paper, we developed an effective topic ranking model, 4EFRRank, by modeling the influence transfer mechanism among all academic entities in a complex academic network using a four-layer network design that incorporates the strengthening effect of multiple entities on topic influence. The PageRank algorithm is utilized to calculate the initial influence of topics, papers, authors, and journals in a homogeneous network, whereas the HITS algorithm is utilized to express the mutual reinforcement between topics, papers, authors, and journals in a heterogeneous network, iteratively calculating the final topic influence value. Based on a specific interdisciplinary domain, social media data, we applied the 4ERRank model to the 19,527 topics included in the criteria. The experimental results demonstrate that the 4ERRank model can successfully synthesize the performance of classic co-word metrics and effectively reflect high citation topics. This study enriches the methodology for assessing topic impact and contributes to the development of future topic-based retrieval and prediction tasks. 相似文献
11.
Knowledge transfer for cross domain learning to rank 总被引:1,自引:1,他引:0
Depin Chen Yan Xiong Jun Yan Gui-Rong Xue Gang Wang Zheng Chen 《Information Retrieval》2010,13(3):236-253
Recently, learning to rank technology is attracting increasing attention from both academia and industry in the areas of machine
learning and information retrieval. A number of algorithms have been proposed to rank documents according to the user-given
query using a human-labeled training dataset. A basic assumption behind general learning to rank algorithms is that the training
and test data are drawn from the same data distribution. However, this assumption does not always hold true in real world
applications. For example, it can be violated when the labeled training data become outdated or originally come from another
domain different from its counterpart of test data. Such situations bring a new problem, which we define as cross domain learning
to rank. In this paper, we aim at improving the learning of a ranking model in target domain by leveraging knowledge from
the outdated or out-of-domain data (both are referred to as source domain data). We first give a formal definition of the
cross domain learning to rank problem. Following this, two novel methods are proposed to conduct knowledge transfer at feature
level and instance level, respectively. These two methods both utilize Ranking SVM as the basic learner. In the experiments,
we evaluate these two methods using data from benchmark datasets for document retrieval. The results show that the feature-level
transfer method performs better with steady improvements over baseline approaches across different datasets, while the instance-level
transfer method comes out with varying performance depending on the dataset used. 相似文献
12.
We investigate temporal factors in assessing the authoritativeness of web pages. We present three different metrics related
to time: age, event, and trend. These metrics measure recentness, special event occurrence, and trend in revisions, respectively.
An experimental dataset is created by crawling selected web pages for a period of several months. This data is used to compare
page rankings by human users with rankings computed by the standard PageRank algorithm (which does not include temporal factors)
and three algorithms that incorporate temporal factors, including the Time-Weighted PageRank (TWPR) algorithm introduced here. Analysis of the rankings shows that all three temporal-aware algorithms produce rankings more
like those of human users than does the PageRank algorithm. Of these, the TWPR algorithm produces rankings most similar to human users’, indicating that all three temporal factors are relevant in page
ranking. In addition, analysis of parameter values used to weight the three temporal factors reveals that age factor has the
most impact on page rankings, while trend and event factors have the second and the least impact. Proper weighting of the
three factors in TWPR algorithm provides the best ranking results. 相似文献
13.
用AUC评估分类器的预测性能 总被引:1,自引:0,他引:1
准确率一直被作为分类器预测性能的主要评估标准,但是它存在着诸多的缺点和不足。本文将准确率与AUC(the area under the Receiver Operating Characteristic curve)进行了理论上的对比分析,并分别使用AUC和准确率对3种分类学习算法在15个两类数据集上进行了评估。综合理论和实验两个方面的结果,显示了AUC不但优于而且应该替代准确率,成为更好的分类器性能的评估度量。同时,用AUC对3种分类学习算法的重新评估,进一步证实了基于贝叶斯定理的NaiveBayes和TAN-CMI分类算法优于决策树分类算法C4.5。 相似文献
14.
15.
[目的/意义] 科研评价中指标权重计算的合理性将直接影响科研机构评价结果的客观性和准确性,本文提出利用信息论方法来计算指标权重,为指标权重计算提供一种新的思路。[方法/过程] 基于DBpedia数据集,利用信息论方法计算出的指标权重和上海交通大学世界大学排行榜已有的指标权重,同时对榜单前100名大学机构进行排名,并将排名结果进行对比分析。[结果/结论] 实验发现利用本文提出的权重计算方法得出的机构得分结果与上海交通大学已有指标权重的得分结果皮尔逊相关性为0.980,斯皮尔曼相关性为0.939,并且其排名顺序和上海交通大学给出的排名顺序皮尔逊相关性和斯皮尔曼相关性均为0.939。以上两个排名结果的得分相关性和排名相关性极强,证明本研究中关联数据的权重计算方法的有效性。 相似文献
16.
The deployment of Web 2.0 technologies has led to rapid growth of various opinions and reviews on the web, such as reviews
on products and opinions about people. Such content can be very useful to help people find interesting entities like products,
businesses and people based on their individual preferences or tradeoffs. Most existing work on leveraging opinionated content
has focused on integrating and summarizing opinions on entities to help users better digest all the opinions. In this paper,
we propose a different way of leveraging opinionated content, by directly ranking entities based on a user’s preferences.
Our idea is to represent each entity with the text of all the reviews of that entity. Given a user’s keyword query that expresses
the desired features of an entity, we can then rank all the candidate entities based on how well opinions on these entities
match the user’s preferences. We study several methods for solving this problem, including both standard text retrieval models
and some extensions of these models. Experiment results on ranking entities based on opinions in two different domains (hotels
and cars) show that the proposed extensions are effective and lead to improvement of ranking accuracy over the standard text
retrieval models for this task. 相似文献
17.
The objective assessment of the prestige of an academic institution is a difficult and hotly debated task. In the last few years, different types of university rankings have been proposed to quantify it, yet the debate on what rankings are exactly measuring is enduring.To address the issue we have measured a quantitative and reliable proxy of the academic reputation of a given institution and compared our findings with well-established impact indicators and academic rankings. Specifically, we study citation patterns among universities in five different Web of Science Subject Categories and use the PageRank algorithm on the five resulting citation networks. The rationale behind our work is that scientific citations are driven by the reputation of the reference so that the PageRank algorithm is expected to yield a rank which reflects the reputation of an academic institution in a specific field. Given the volume of the data analysed, our findings are statistically sound and less prone to bias, than, for instance, ad–hoc surveys often employed by ranking bodies in order to attain similar outcomes. The approach proposed in our paper may contribute to enhance ranking methodologies, by reconciling the qualitative evaluation of academic prestige with its quantitative measurements via publication impact. 相似文献
18.
Balázs R. Sziklai 《Journal of Informetrics》2021,15(2):101133
We present a novel algorithm to rank smaller academic entities such as university departments or research groups within a research discipline. The Weighted Top Candidate (WTC) algorithm is a generalisation of an expert identification method. The axiomatic characterisation of WTC shows why it is especially suitable for scientometric purposes. The key axiom is stability – the selected institutions support each other's membership. The WTC algorithm, upon receiving an institution citation matrix, produces a list of institutions that can be deemed experts of the field. With a parameter we can adjust how exclusive our list should be. By completely relaxing the parameter, we obtain the largest stable set – academic entities that can qualify as experts under the mildest conditions. With a strict setup, we obtain a short list of the absolute elite. We demonstrate the algorithm on a citation database compiled from game theoretic literature published between 2008–2017. By plotting the size of the stable sets with respect to exclusiveness, we can obtain an overview of the competitiveness of the field. The diagram hints at how difficult it is for an institution to improve its position. 相似文献
19.
Search effectiveness metrics are used to evaluate the quality of the answer lists returned by search services, usually based
on a set of relevance judgments. One plausible way of calculating an effectiveness score for a system run is to compute the
inner-product of the run’s relevance vector and a “utility” vector, where the ith element in the utility vector represents the relative benefit obtained by the user of the system if they encounter a relevant
document at depth i in the ranking. This paper uses such a framework to examine the user behavior patterns—and hence utility weightings—that
can be inferred from a web query log. We describe a process for extrapolating user observations from query log clickthroughs,
and employ this user model to measure the quality of effectiveness weighting distributions. Our results show that for measures
with static distributions (that is, utility weighting schemes for which the weight vector is independent of the relevance
vector), the geometric weighting model employed in the rank-biased precision effectiveness metric offers the closest fit to
the user observation model. In addition, using past TREC data as to indicate likelihood of relevance, we also show that the
distributions employed in the BPref and MRR metrics are the best fit out of the measures for which static distributions do
not exist. 相似文献
20.
Query suggestions have become pervasive in modern web search, as a mechanism to guide users towards a better representation of their information need. In this article, we propose a ranking approach for producing effective query suggestions. In particular, we devise a structured representation of candidate suggestions mined from a query log that leverages evidence from other queries with a common session or a common click. This enriched representation not only helps overcome data sparsity for long-tail queries, but also leads to multiple ranking criteria, which we integrate as features for learning to rank query suggestions. To validate our approach, we build upon existing efforts for web search evaluation and propose a novel framework for the quantitative assessment of query suggestion effectiveness. Thorough experiments using publicly available data from the TREC Web track show that our approach provides effective suggestions for adhoc and diversity search. 相似文献