Skip to content

Latest commit

 

History

History
51 lines (51 loc) · 1.91 KB

2008-07-09-gruber08a.md

File metadata and controls

51 lines (51 loc) · 1.91 KB
abstract title year layout series publisher issn id month tex_title firstpage lastpage page order cycles bibtex_author author date note address container-title volume genre issued pdf extras
Latent topic models have been successfully applied as an unsupervised topic discovery technique in large document collections. With the proliferation of hypertext document collection such as the Internet, there has also been great interest in extending these approaches to hypertext [6, 9]. These approaches typically model links in an analogous fashion to how they model words - the document-link co-occurrence matrix is modeled in the same way that the document-word co-occurrence matrix is modeled in standard topic models.In this paper we present a probabilistic generative model for hypertext document collections that explicitly models the generation of links. Specifically, links from a word w to a document d depend directly on how frequent the topic of w is in d, in addition to the in-degree of d. We show how to perform EM learning on this model efficiently. By not modeling links as analogous to words, we end up using far fewer free parameters and obtain better link prediction results.
Latent topic models for hypertext
2008
inproceedings
Proceedings of Machine Learning Research
PMLR
2640-3498
gruber08a
0
Latent topic models for hypertext
230
239
230-239
230
false
Gruber, Amit and Rosen-Zvi, Michal and Weiss, Yair
given family
Amit
Gruber
given family
Michal
Rosen-Zvi
given family
Yair
Weiss
2008-07-09
Reissued by PMLR on 09 October 2024.
Proceedings of the 24th Conference on Uncertainty in Artificial Intelligence
R6
inproceedings
date-parts
2008
7
9