U.S. flag

An official website of the United States government

Display Settings:

Items per page

PMC Full-Text Search Results

Items: 6

1.
Fig. 3.

Fig. 3. From: Finding scientific topics.

Model selection results, showing the log-likelihood of the data for different settings of the number of topics, T. The estimated standard errors for each point were smaller than the plot symbols.

Thomas L. Griffiths, et al. Proc Natl Acad Sci U S A. 2004 Apr 6;101(Suppl 1):5228-5235.
2.
Fig. 5.

Fig. 5. From: Finding scientific topics.

The plots show the dynamics of the three hottest and three coldest topics from 1991 to 2001, defined as those topics that showed the strongest positive and negative linear trends. The 12 most probable words in those topics are shown below the plots.

Thomas L. Griffiths, et al. Proc Natl Acad Sci U S A. 2004 Apr 6;101(Suppl 1):5228-5235.
3.
Fig. 6.

Fig. 6. From: Finding scientific topics.

A PNAS abstract () tagged according to topic assignment. The superscripts indicate the topics to which individual words were assigned in a single sample, whereas the contrast level reflects the probability of a word being assigned to the most prevalent topic in the abstract, computed across samples.

Thomas L. Griffiths, et al. Proc Natl Acad Sci U S A. 2004 Apr 6;101(Suppl 1):5228-5235.
4.
Fig. 4.

Fig. 4. From: Finding scientific topics.

(Upper) Mean values of θ at each of the diagnostic topics for all 33 PNAS minor categories, computed by using all abstracts published in 2001. Higher probabilities are indicated with darker cells. (Lower) The five most probable words in the topics themselves listed in the same order as on the horizontal axis in Upper.

Thomas L. Griffiths, et al. Proc Natl Acad Sci U S A. 2004 Apr 6;101(Suppl 1):5228-5235.
5.
Fig. 1.

Fig. 1. From: Finding scientific topics.

(a) Graphical representation of 10 topics, combined to produce “documents” like those shown in b, where each image is the result of 100 samples from a unique mixture of these topics. (c) Performance of three algorithms on this dataset: variational Bayes (VB), expectation propagation (EP), and Gibbs sampling. Lower perplexity indicates better performance, with chance being a perplexity of 25. Estimates of the standard errors are smaller than the plot symbols, which mark 1, 5, 10, 20, 50, 100, 150, 200, 300, and 500 iterations.

Thomas L. Griffiths, et al. Proc Natl Acad Sci U S A. 2004 Apr 6;101(Suppl 1):5228-5235.
6.
Fig. 2.

Fig. 2. From: Finding scientific topics.

Results of running the Gibbs sampling algorithm. The log-likelihood, shown on the left, stabilizes after a few hundred iterations. Traces of the log-likelihood are shown for all four runs, illustrating the consistency in values across runs. Each row of images on the right shows the estimates of the topics after a certain number of iterations within a single run, matching the points indicated on the left. These points correspond to 1, 2, 5, 10, 20, 50, 100, 150, 200, 300, and 500 iterations. The topics expressed in the data gradually emerge as the Markov chain approaches the posterior distribution.

Thomas L. Griffiths, et al. Proc Natl Acad Sci U S A. 2004 Apr 6;101(Suppl 1):5228-5235.

Display Settings:

Items per page

Supplemental Content

Recent activity

Your browsing activity is empty.

Activity recording is turned off.

Turn recording back on

See more...
Support Center