# jemdoc: menu{MENU}{index.html}, showsource = Discovering Word Associations in News Media via Feature Selection and Sparse Classification - *Download:* [Pubs/MIR2010.pdf .pdf] - *Authors:* B. Gawalt, J. Jia, L. Miratrix, L. El Ghaoui, B. Yu, S. Clavier. - *Status:* In /Proc. [http://riemann.ist.psu.edu/mir2010/ 11th ACM SIGMM International Conference on Multimedia Information Retrieval]/, 2010. - *Abstract:* Our perceptions of the world are largely shaped by news media. Understanding how media portray certain words and terms is a critical step towards assessing media's influence on those perceptions. In this paper we analyze the ``image" of a given query word in a given corpus of text news by producing a short list of other words with which this query is strongly associated. We use a number of feature selection schemes for text classification to help in this task. We apply these classification techniques using indicators of the query word's appearance in each document used as the document ``labels" and the indicators for all other words as document predictors/features. The features selected by any scheme is then considered the list of words comprising the query word's ``image". To be easily understandable, a list should be extremely short with respect to the dictionary of terms present in the corpus. The approach thus requires aggressive feature (word) selection in order to single out at most a few tens of terms in a universe of hundreds of thousands or more. In addition, a word imaging scheme should scale well with the size of data (number and size of documents, size of dictionary). We produce one scheme for feature selection through a sparse classification model. A standard classification algorithm assigns one weight per term in the predictor dictionary, in order to maximize the capacity to successfully predict the labels of document units. By imposing a sparsity constraint on the weight vector, we single out the few words that are most able to predict the presence or absence of a query word in any document. This paper compares this and several other schemes that are potentially well suited to the task of word imaging, each method presenting a different manner of feature selection. We present two evaluations of these schemes. One evaluates the predictive classification performance of a logistic regression model trained over the corpus using only a scheme's selected features. The other is based on the judgement of human readers: a pair of word lists generated by different schemes operating on identical queries are presented to a human subject alongside a trio of document units (paragraphs) containing the query word. This subject then chooses the list which in his/her estimation is the best summary of the document units. We apply these schemes to study the images of frequently covered countries and regions in recent news articles from the International section of the New York Times. Our preliminary experiments indicate that, while most methods perform similarly well on our data in terms of predictive performance, human-based evaluations appear to favor features selected by training of sparse logistic regression, a penalized variant of logistic regression that encourages sparse classifiers. This indicates that classification metrics based on pure predictive performance, while useful as a indicator for preselecting algorithms, are not enough to predict human assessment of word association algorithms. #- *Related entries:* - *Bibtex reference:* ~~~ {}{} @conference{GJMEYC:10, Author = {B. Gawalt and J. Jia and L. Miratrix and L. {El Ghaoui} and B. Yu and S. Clavier}, Title = {Discovering Word Associations in News Media via Feature Selection and Sparse Classification}, Booktitle = {Proc. 11th ACM SIGMM International Conference on Multimedia Information Retrieval}, Year = 2010 } ~~~