M. Steyvers, and T. Griffiths. Latent Semantic Analysis: A Road to Meaning, Laurence Erlbaum, (2005)
Abstract
Many chapters in this book illustrate that applying a statistical method such as Latent Semantic Analysis (LSA;
Landauer & Dumais, 1997; Landauer, Foltz, & Laham, 1998) to large databases can yield insight into human
cognition. The LSA approach makes three claims: that semantic information can be derived from a word-document
co-occurrence matrix; that dimensionality reduction is an essential part of this derivation; and that words and
documents can be represented as points in Euclidean space. In this chapter, we pursue an approach that is consistent
with the first two of these claims, but differs in the third, describing a class of statistical models in which the
semantic properties of words and documents are expressed in terms of probabilistic topics.
Overview about topic/author inference
---
More info on http://oz.ss.uci.edu/237/
e.g. slides http://oz.ss.uci.edu/237/lectures/topics.pdf
slides on LSA
http://lsa.colorado.edu/~quesadaj/pdf/LSATutorial.pdf
%0 Book Section
%1 citeulike:383010
%A Steyvers, M.
%A Griffiths, T.
%B Latent Semantic Analysis: A Road to Meaning
%D 2005
%E Landauer, T.
%E Mcnamara, D.
%E Dennis, S.
%E Kintsch, W.
%I Laurence Erlbaum
%K topicinference aaa socialnets
%T Probabilistic topic models
%U http://psiexp.ss.uci.edu/research/papers/SteyversGriffithsLSABookFormatted.pdf
%X Many chapters in this book illustrate that applying a statistical method such as Latent Semantic Analysis (LSA;
Landauer & Dumais, 1997; Landauer, Foltz, & Laham, 1998) to large databases can yield insight into human
cognition. The LSA approach makes three claims: that semantic information can be derived from a word-document
co-occurrence matrix; that dimensionality reduction is an essential part of this derivation; and that words and
documents can be represented as points in Euclidean space. In this chapter, we pursue an approach that is consistent
with the first two of these claims, but differs in the third, describing a class of statistical models in which the
semantic properties of words and documents are expressed in terms of probabilistic topics.
@incollection{citeulike:383010,
abstract = {Many chapters in this book illustrate that applying a statistical method such as Latent Semantic Analysis (LSA;
Landauer \& Dumais, 1997; Landauer, Foltz, \& Laham, 1998) to large databases can yield insight into human
cognition. The LSA approach makes three claims: that semantic information can be derived from a word-document
co-occurrence matrix; that dimensionality reduction is an essential part of this derivation; and that words and
documents can be represented as points in Euclidean space. In this chapter, we pursue an approach that is consistent
with the first two of these claims, but differs in the third, describing a class of statistical models in which the
semantic properties of words and documents are expressed in terms of probabilistic topics.},
added-at = {2006-06-16T10:34:37.000+0200},
author = {Steyvers, M. and Griffiths, T.},
biburl = {https://www.bibsonomy.org/bibtex/216802fb465ca95ac77434bf73c6b271b/ldietz},
booktitle = {Latent Semantic Analysis: A Road to Meaning},
citeulike-article-id = {383010},
comment = {Overview about topic/author inference
---
More info on http://oz.ss.uci.edu/237/
e.g. slides http://oz.ss.uci.edu/237/lectures/topics.pdf
slides on LSA
http://lsa.colorado.edu/~quesadaj/pdf/LSATutorial.pdf},
editor = {Landauer, T. and Mcnamara, D. and Dennis, S. and Kintsch, W.},
interhash = {c2bf94c6b77cd997be7760fc59609a0e},
intrahash = {16802fb465ca95ac77434bf73c6b271b},
keywords = {topicinference aaa socialnets},
priority = {0},
publisher = {Laurence Erlbaum},
timestamp = {2006-06-16T10:34:37.000+0200},
title = {Probabilistic topic models},
url = {http://psiexp.ss.uci.edu/research/papers/SteyversGriffithsLSABookFormatted.pdf},
year = 2005
}