leonjwu / Paper-Configurations

Academic Paper Recommender Systems using Multidimensional Scaling

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Recommender System for Academic Papers using Multidimensional Scaling in R


Overview

  • Goal: Reveal configurations of academic papers so that a user can find targeted lists of research materials
  • Dissimilarity measures
  • Data scraping using unix and Google Scholar
  • Classical and Ordinal Scaling, analysis of eigenvalues
  • K-Means Clustering of final configurations for labelling
Word Counts in Papers Heatmap of Word Frequencies for Each Paper

1 Choice of Data

The aim of this analysis is to reveal configurations of academic papers so that a user can find targeted lists of research materials related to a certain paper or topic, without relying on manually validating papers and using complex searching methods within databases of academic papers. In this way, the user can visualize distances between papers, chose the relevant ones and perhaps observe and explore clusterings of papers that are about similar topics. This is useful since you can understand better how different topics and individual papers link to one another in a more informative and exploratory way. The method of collecting the data is as follows:

  • Scrape statistics lecture notes to acquire a list of p≈ 50 words (ignoring common English words) that will be used to describe each paper. See the wordcloud generated from the lecture notes (Figure 1) to get an idea of these words.
  • Use keywords and Google Scholar (See Github link for details) to collect papers relating to the subject of statistics
  • Count the number of occurunces of each of the words in each of them ≈ 50 papers.

This result is annxpmatrix. This can easily be expanded to include more words and papers, and a small example (n, p ≈ 50 ) is used in order to better illustrate the methods. A heatmap representing the data used can be found in Figure 2.

Figure 1: High-frequency words in Statistics Lecture Notes 

Image

Figure 2: Heatmap word frequencies for each paper

Image

2 Distances / Dissimilarities Between Papers

A distance matrix between thenpapers is to be computed. Different metrics to determine these distances can be used to acheive this. In this case, papers with similar (standardized) frequencies of words could be considered "close" to eachother. A natural choice is to use the Euclidean distance (L2 norm). Other non-euclidean metrics can be used, and may be better for some applications. In this case, the Manhattan distance (L1 norm) could be useful in that if there is a paper which mentions the word "Python" a disproportionate number of times, say, then this paper would have a very large distance between other papers when using the L2 norm. Using the L1 norm reduces this weighting of extreme values which could be desirable.

3 Multi-Dimensional Scaling

Classical Scaling

Classical scaling aims to create a configuration of the data from the distances between points. From Figure 3 it can be seen that there only 3 large eigenvalues for the Euclidean distance matrix. This means that the data can be projected onto 3 (possibly 2) dimensions without losing too much information. For the Manhattan distance matrix, there are perhaps 3-7 significantly large eigenvalues (Figure 4). However, projection onto 2 dimensions would likely yield useful results still. A 2-dimensional plot is also extremely informative to human users and so this is the dimensionality that is chosen. The negative eigenvalues in Figure 4 arise due to the fact the distances between papers are non-euclidean. The configurations acquired from applying classical scaling are shown in Figures 5 and 6 for the Eu- clidean and Manhattan distances respectively. (A reduced number of papers is displayed in order to increase interpretability.)

Figure 3: Eigenvalues from the Euclidean Distance Matrix

Image

Figure 4: Eigenvalues from the Manhatten Distance
Matrix

Image

Figure 5: Configuration from Classical scaling (Euclidean)

Image

Figure 6: Configuration from Classical scaling (Manhattan)

Image

Ordinal Scaling

Ordinal scaling also aims to create a configuration of the data, but minimizes a penalty function that measures the "stress" of the configuration. Using a random start, the resulting configuration of the ordinal scaling results in a lower stress than that of the classical scaling solution. Figure 7 shows the configuration from the ordinal scaling of the Manhattan distances matrix.

Figure 7: Configuration from Ordinal scaling (Manhattan) with lower stress

Image

4 Comparison of Configurations

Procrustes Analysis can be used to compare configurations. Figure 8 shows the aligned configurations for the classical and ordinal configurations from the Manhattan distances. The configurations have some similarities, especially around the edges, but are distinctly different from eachother. The ordinal scaled configuration will be chosen since its stress was lower as previously discussed.

Figure 8: Classical and Ordinal Configurations aligned
(Manhatten)

Image

5 K-Means Clustering and Labelling

A K-Means clusterings is applied to the data to obtain 4 clusters. In order to label these clusters, the 3 nearest neighbours were computed for each centroid. The most common word (standardized) was chosen to be the label for that centroid. The final configuration is plotted using the ordinal scaling with the labelled centroids in Figure 9.

Figure 9: Final Labelled Configurations using K-Means

Image

6 Conclusion

Using this method, any academic search term can be queried, resulting in a 2D configuration that allows the user to visualize inspect and collect papers that are most similar to search terms in a fast and automated way. In the case of the example configuration, the user can choose to explore papers that mention "inference" often, and find similar papers to the statistics lectures notes (located at position "12") This is extremely useful and time saving when collecting research material!

Extensions

An interactive version of this analysis could be developed, so that general search terms and a larger number of papers and words can be used to construct more complex configurations. The labelling of clusters is also far from optimal; for example the number of clusters was arbitrary and can be decided in a more systematic way using silhouette plots, say.

How to fetch PDFs from Google Scholar

The papers used can be acquired using the following unix command:

wget -e robots=off -H --user-agent="Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.0.3) Gecko/2008092416 #Firefox/3.0.3" -r -l 1 -nd -A pdf http://scholar.google.com/scholar?q=filetype%3Apdf+statistics&btnG=&hl=en&as_sdt=0%2C23

The search terms that I used were:

  • +statistical+inference
  • +statistics
  • +machine+learning
  • +data+science
  • +optimization
  • +modelling
  • +statistical+learning
  • +machine+learning

About

Academic Paper Recommender Systems using Multidimensional Scaling


Languages

Language:Python 100.0%