% This is a sample LaTeX 2e input file.
% (Gene Pozniak - ASA, 7/28/97)
%
% A '%' character causes TeX to ignore all remaining text on the line,
% and is used for comments like this one.
\documentclass[times,11pt]{article}
% choose options for [] as required from the list
% in the Reference Guide, Sect. 2.2
\usepackage{makeidx} % allows index generation
\usepackage{graphicx} % standrd LaTeX graphics tool
% when including figure files
\usepackage{multicol} % used for the two-column index
% etc.
% see the list of further useful packages
% in the Reference Guide, Sects. 2.3, 3.1-3.3
\makeindex % used for the subject index
% please use the style sprmidx.sty with
% your makeindex program
%\usepackage{amssymb}
%\usepackage{amsmath}
%\usepackage{epsfig}
%\usepackage{latexsym}
\usepackage{amsfonts}
\usepackage{amsmath}
%
% What follows are some hacks that enable one to obtain special math
% notations that are not commonly available in LaTeX. The standard
% math notation is covered in many places; e.g., Leslie Lamport's
% *LaTeX: A Document Preparation System*, Addison-Wesley, 1986.
%
\newcommand{\Pb}{\hbox{{I}\kern-.1667em\hbox{P}}}
%produces blackboard bold probability sign; use \Pb
%
\newcommand{\Ex}{\hbox{{I}\kern-.1667em\hbox{E}}}
%produces blackboard bold expectation sign; use \Ex
\newcommand{\Real}{\hbox{{I}\kern-.1667em\hbox{R}}}
%produces blackboard bold real numbers sign; use \Real
%
\newcommand{\sReal}{\hbox{\scriptsize {I}\kern-.1667em\hbox{R}}}
%produces subscript-sized blackboard bold real numbers sign; use \sReal
%
\newcommand{\vect}[1]{\mbox{\boldmath $ #1$}}
%produces bold anything (such as Greek letters); use \vect{anything}
%
\newcommand{\combos}[2]{\left(\mbox{\small $\begin{array}{c} {#1}\\{#2}
\end{array}$} \right)}
%produces n choose r sign; use \combos{n}{r}
%
\newcommand{\var}{{\rm var}\,}
%
\newcommand{\abs}{{\bf Abs: }}
\newcommand{\gp}{\vspace{.32 in}}
\begin{document}
\begin{center}
{\large \bf CSNA-2007 Abstracts}
(Ordered alphabetically by first author's name)
\end{center}
The 2007 Meeting of the Classification Society of North America is
Sponsored by National Center for Supercomputing Applications, the UIUC
College of Education, Department of Computer Science, and the Graduate
School of Library and Information Science. Special thanks also to CRC
Press and to Wolfram Research.
\gp
\textbf{How are bullies and victims embedded in the classroom peer
network? : Factors linked to the heterogeneity of their structural
embeddedness}
Hai-Jeong Ahn, Claire Garandeau
University of Illinois at Urbana Champaign
\abs
We examined the structural embeddedness of bullies and victims in the
classroom peer network. 3rd and 4th graders (N = 680) completed
within-class peer nominations for bullies, victims, and behavioral
profiles (i.e. aggressive and prosocial behavior). 147 bullies (78
boys) and 150 victims (77 boys) were identified after controlling for
gender and class. Classroom social network structure was determined by
cohesive blocking procedure based on affiliation relationships
(i.e. hang around together). Classroom social network embeddedness
varied across classes. The structural embeddedness of bullies and
victims was heterogeneous. 35\% of bullies and 32\% of victims are
deeply embedded in the classroom social network, but 22\% of bullies
and 30\% of victims are placed outside of the social network. Boy
bullies are more deeply embedded in the peer network than girl
bullies, but no gender differences are found for victims. Black
bullies and victims are more deeply embedded in the classro! om!
social network structure than non-black bullies and victims. We
analyze classroom racial composition and behavioral profiles of
bullies and victims in order to clarify the heterogeneity of their
structural embeddedness.
\gp \newpage
\textbf{Dimension Reduction Using a Hybrid of PARAMAP and Isomap Procedures}
Ulas Akkucuk, Bogazici University, Istanbul, Turkey
J. Douglas Carroll, Rutgers Business School
\abs
Dimensionality reduction aims to represent higher dimensional data by
a lower-dimensional structure. A well-known approach by Carroll,
Parametric Mapping (abbreviated PARAMAP) relies on iterative
minimization of a loss function measuring the smoothness or
``continuity'' of the mapping from the lower dimensional representation
to the original data. This approach was primarily only of theoretical
interest at the time it was developed, since it was computationally
expensive and prone to local optima. The approach was resuscitated
recently with important algorithmic modifications and we will call
this new approach ``PARAMAP-2''. In the earlier PARAMAP approach, when
given data sets that either were ``large'', had added error, irregular
spacing, or all of these features combined, the method tended
essentially always to converge to a clear local optimum, and was
incapable of obtaining the globally optimal solution known to
exist. Using the PARAMAP-2 approach we were able to obtain solutions
we were certain were globally optimal in reasonable computational
time. In this paper we discuss use of a variant of a method called
Isomap to obtain a starting framework, and then adding new points in
batches based on their proximity to landmark points in this initial
framework using the PARAMAP-2 algorithm. Since Isomap is faster and
less prone to local optimum problems than PARAMAP, and the iterative
process involved in adding new points to the configuration should be
much less time consuming we believe the resulting method should be
much better suited to dealing with large sets of realistically based
data, and more inclined to obtain a satisfactory solution in roughly
order $n^2$ time. One initial test of a version of this new ``PARAMAP-3''
algorithm has been implemented, with very promising results. In this
paper we will explain the methods of obtaining the framework and the
iterative procedure of mapping in the rest of the points,
demonstrating this methodology on the well-known nonlinear manifolds
such as the sphere and four dimensional torus.
\gp \newpage
\textbf{On Similarity Indices and Correction for Chance Agreement}
Ahmed N. Albatineh, Nova Southeastern University
Magdalena Niewiadomska-Bugaj, Western Michigan University
Daniel P. Mihalko, Western Michigan University
\abs
Similarity indices can be used to compare partitions
(clusterings) of a data set. Many such indices were introduced in the
literature over the years. We are showing that out of 28 indices we
were able to track, there are 22 different ones. Even though their
values differ for the same clusterings compared, after correcting for
agreement attributed to chance only, their values become similar and
some of them even become equivalent. Consequently, the problem of
choice of the index to be used for comparing different clusterings
become less important.
\gp
\textbf{An ordinal impurity function for classification trees when
predicting an ordinal response}
Kellie J. Archer, Department of Biostatistics, Virginia Commonwealth University
\abs
Ensemble methods have been demonstrated to be competitive with other
machine learning approaches for classification and have been described
for nominal, continuous, and survival responses. However, in a large
number of biomedical applications, the class to be predicted may be
inherently ordinal. Examples of ordinal responses include TNM stage
(I, II, III, IV) and drug toxicity (none, mild, moderate,
severe). While nominal response methods may be applied to ordinal
response data, in so doing some information is lost that may improve
the predictive performance of the classifier. This study examined the
effectiveness of using both an ordinal impurity function for
classification tree growing and bootstrap aggregation. Results using
the ordinal impurity function are compared to those obtained using the
Gini impurity function, Gini impurity with a linear loss, and Gini
impurity with quadratic loss on both simulated and benchmark datasets.
\gp \newpage
\textbf{Influence analysis in Depth Transvariation based Classification}
Nedret Billor, Asheber Abebe, Asuman Turkmen, and Sai Nudurapati,
Department of Mathematics and Statistics, Auburn University, AL
\abs
Depth transvariation based classification is a new
nonparametric classification technique based on classifying a
multivariate data point through maximizing the estimated
transvariation probability of statistical depths.
In this paper we will study the influence of observations
on the misclassification error rate in depth transvariation based
discriminant analysis. We assess the partial influence of the error
rate , which allows us to quantify the effect of observations in the
training sample, based on the performance of the depth
transvariation based classification rule.
We use some simulated data sets as well as some real
examples to evaluate the robustness of error rate based on the depth
transvariation classification rule.
\gp
\textbf{Challenges in the classification of spatial data}
Alexander Brenning, Department of Geography, University of Waterloo,
Ontario, Canada
\abs
While the amount of spatial data available within Geographical Information
Systems is exploding, research on the construction and assessment of
classification techniques for spatial data is still very limited. In the
case of gridded spatial data such as remote-sensing data or landslide
inventories, one common characteristic that neighboring raster cells have
to be considered pseudo-replications representing virtually identical
observations. This phenomenon, and more generally the presence of spatial
autocorrelation, has important implications: (1) Successful classifiers
must avoid over-fitting in the presence of spatial autocorrelation, and
(2) appropriate techniques for the estimation of misclassification error
rates, such as spatial cross-validation or a spatial bootstrap, are needed
for error assessment. Results of classification and estimation techniques
in two different applications are presented: The two-class problem of
landslide susceptibility mapping, and the multi-class problem of crop
identification from multi-temporal remote-sensing data.
\gp \newpage
\textbf{Title: Classification of Massive, Structured Data: Research
Progress @ Data Mining Group.CS.UIUC}
Deng Cai, University of Illinois
\abs
During the past decades, numerous classification algorithms have been
proposed. Many of them take the numerical feature vectors as input and
predict the label of a sample. However, much real-world data is not
represented as numerical vectors, but as more complicated structures,
such as sequences, trees, or graphs. To solve these real-world
problems, we need to develop more flexible classification based
approaches. Our group aims at developing algorithms that can well
handle the real-world problems. Specifically, we are interested in and
developing algorithms for the following problems:
\begin{description}
\item[Frequent pattern based classification] The frequent pattern
(sub-structure) in complicated structured data can naturally be used
as features. How to efficiently mining these sub-structures? How to
select the most powerful substructures with respective classification
performance? We conducted a systematic exploration of frequent
pattern-based classification, and provide solid reasons supporting
this methodology.
\item[Name entity distinction] Different people or objects may share identical names in the real world, which
causes confusion in many applications. It is a nontrivial task to distinguish
those objects, especially when there is only very limited information
associated with each of them. Although linkages among objects provide useful
information, such information is often intertwined and inconsistent with the
identities of objects. Moreover, different types of linkages carry different
semantic meanings and have different levels of pertinence. We developed a
general object distinction methodology called DISTINCT for analysis based on
supervised composition of heterogeneous links.
\item[Classification on Stream]In recent years, there have been some interesting studies on predictive
modeling in data streams. However, most such studies assume relatively
balanced and stable data streams but cannot handle well rather skewed (e.g.,
few positives but lots of negatives) and stochastic distributions, which are
typical in many data stream applications. We proposed a new approach to mine
data streams by estimating reliable posterior probabilities using an ensemble
of models to match the distribution over under-samples of negatives and
repeated samples of positives.
\item[Trajectory classification]With the emerging GPS and RFID technologies, modeling trajectories on road
networks becomes more and more important for transportation and traffic
planning. We studied methods for classifying trajectories on road networks. By
analyzing the behavior of trajectories on road networks, we observe that, in
addition to the locations where vehicles have visited, the order of these
visited locations is crucial for improving classification accuracy. Based on
our analysis, we contend that frequent sequential patterns are the best
feature candidates since they preserve this order information. Furthermore,
when mining sequential patterns, we propose to confine the length of
sequential patterns to ensure high efficiency. Our comparative study over a
broad range of classification approaches demonstrates that our frequent
pattern-based classification method significantly improves accuracy over a
na\"{\i}ve method in some synthetic trajectory data.
\item[Automatic Software Debug] Automated localization of software
bugs is one of the essential issues in debugging aids. What can be the
effective features for automatic software debugging? What statistical
model is suitable for this problem? We have proposed a new statistical
model-based approach, called SOBER, which localizes software bugs
without any prior knowledge of program semantics.
\item[Discriminant Analysis for Large Scale High Dimensional Data]
Many real-world data are with high dimensionality, so as the generated
features (frequent patterns) of structured data. When one performs
discriminant analysis on such data set, the computational cost is
always a bottleneck. Specifically, the most well applied algorithm,
Linear Discriminant Analysis (LDA), has cubic-time complexity with
respect to min(m,n), where m is the number of samples and n is the
number of features. When both m and n are large, it is infeasible to
apply LDA. We developed a novel algorithm for discriminant analysis,
called Spectral Regression Discriminant Analysis (SRDA). SRDA has
linear-time complexity with respect to both m and n.
\end{description}
\gp \newpage
\textbf{Modularity Clustering for Thematic Document Clustering}
Brant Chee and Bruce Schatz
University Of Illinois, Institute for Genomic Biology
\abs
We present a modified physics algorithm that is single link in nature
that takes advantage of the inherent scale free and small world
characteristics of word graphs in order to create semantic concept
clusters of words. The word clusters are then used to segment large
number of documents into overlapping thematic document clusters. Our
method differs from existing methods such as Carrot 2 or latent
semantic indexing in that the running time is roughly linear (O(n log2
n) in the number of terms). We have demonstrated results on large
numbers of documents (>100K) and faster running time than complete
link algorithms while the clusters are more coherent and have higher
utility than faster methods such as K-means.
Our algorithm takes advantage of the inherent clustering that exists
in a small world graph whereas, the original physics algorithm makes
assumptions about a log distribution in the number of elements in the
in a cluster. We make no such assumptions and explicitly remove
restrictions so that the number of elements in a cluster follows a
more normal distribution providing better thematic document mappings.
\gp % % \newpage
\textbf{Connections between K-Means Cluster Analysis and Restricted
Latent Class Models}
Chia-Yi Chiu, Department of Educational Psychology, University of Illinois
Jeff Douglas, Department of Statistics, University of Illinois
\abs
Restricted latent class models have seen many applications in
psychometrics in recent years. One application has been in multiple
classification latent class models where presence or absence of each
of several latent attributes is under diagnosis. Maximum likelihood
estimation or Bayesian computation, such as Markov chain Monte
Carlo, are usually used to estimate parameters of these latent class
models. K-means cluster analysis, when informed by the assumed
structure of the latent class model, can be used as an alternative.
By selecting an appropriate multidimensional statistic as an input
of K-means, K-means results in almost identical
classification as the conventional estimation methods can do, but
does not need to access the complex models. This study will
demonstrate how to select a statistic for K-means with latent class
models and compares the results of the K-means clustering method
with the conventional estimation method. An application
to skills diagnosis with education testing data is presented.
\gp \newpage
\textbf{Evaluation of Hierarchies based on the Longest Common Prefix, or
Baire, Metric}
Pedro Contreras and Fionn Murtagh
Department of Computer Science
Royal Hollow, University of London
\abs
The Baire space has a metric that can be defined from the
longest common prefix of two strings. Consider two floating point
numbers with the first $p$ digits identical. Then what we call
their Baire distance is $2^{-p}$. This distance is an ultrametric.
It follows that a hierarchy can be used to represent the relationships
associated with this distance. We address the issue of whether such a
hierarchy (let us call it a Baire hierarchy, because it is
clearly a different hierarchy compared to one resulting from any
of the commonly used agglomerative hierarchical clustering
algorithms) is advantageous, computationally, for clustering
large, high dimensional data sets; and also how useful it is,
in particular compared to k-means.
\gp % % \newpage
\textbf{Putting Some 'WoW' into Modeling Longitudinal Networks}
Bethany Dohleman, Department of Psychology, University of Illinois at
Urbana-Champaign
Harold D. Green Jr., Science of Networks in Communities (SONIC),
National Center for Supercomputing Applications, University of
Illinois at Urbana-Champaign
Dmitri Williams, Department of Speech Communication, University of
Illinois at Urbana-Champaign
Noshir Contractor, Department of Speech Communication \& Science of
Networks in Communities (SONIC), National Center for Supercomputing
Applications, University of Illinois at Urbana-Champaign
\abs
This study concerns co-evolution of communication networks and
perceived expertise among members of a Massively Multiplayer Online
Role-Playing Game called World of Warcraft (WoW). We explored
multi-theoretical multilevel hypotheses about what motivates an
individual to create expertise-seeking advice ties from other players
within their virtual communities, called guilds.
We used SIENA to test how co-evolutionary dynamics vary for guilds
that have audio and text communication versus only text communication.
Nine guilds received a Voice over Internet Protocol (VoIP) headset,
allowing them to communicate with each other using audio and text
means. Changes in network structure and member attributes in these
treatment guilds are compared with changes in seven guilds that relied
only on text means.
Results reveal that mechanisms associated with the co-evolution of
network structure and perceived expertise levels varied between
treatment and control guilds. For example, experts in VoIP guilds are
more likely to identify and approach other experts for advice. Other
results reveal that for VoIP guilds, rates of communication are larger
in the initial time interval than in other time intervals, a pattern
not found in control guilds. Comparing patterns of co-evolution for
control and treatment guilds demonstrates the application of new
statistical tools for comparing parameters.
\gp % \newpage
\textbf{One-sided Elasticities and Technical Efficiency in
Multi-output Production: A Theoretical Framework}
Petros Hadjicostas, Department of Mathematics and Statistics,
Texas Tech University
(Joint work with Andreas C. Soteriou)
\abs
One of the concepts that have sparked considerable interest in the
theory of production and efficiency is that of returns to scale
(RTS). Economics researchers typically define RTS using the notion of
elasticity. Considerable research activity on RTS has also been
observed by management science researchers, who utilize the
methodology of Data Envelopment Analysis (DEA) to gain insights on
RTS. In this talk, we present a theoretical framework (developed
jointly with Andreas Soteriou) that integrates existing economics and
management science literature on RTS, and provides a foundation for
research work in this area. Our framework defines, discusses, and
proposes an approach to measure input- and output-oriented
elasticities, and one-sided RTS. We demonstrate how the work done in
DEA is a special case of our framework, and discuss the conditions
under which the resulting two left-hand, and the two right-hand
elasticities can be equal. The results of this work have been
published in 2006 in the European Journal of Operational Society.
Current and future research directions are also discussed. For
example, in a recent working paper, Hadjicostas and Soteriou explore
properties of different orders of one-sided elasticities in
multi-input multi-output production using the aforementioned
theoretical framework. A special case of the theory in the paper is
the Banker-Morey (1986) DEA model for data that include both
discretionary and non-discretionary inputs and outputs.
\gp \newpage
\textbf{A Useful Combinatorial Lemma and Some Applications}
Bernard Harris
\abs
About 50 years ago, Leo Katz introduced a formula that he used to
determine the probability that a random mapping is connected. This
formula is revisited and generalized. Several applications of the
original result and the generalization are discussed.
\gp % \newpage
\textbf{Identification of Multiple Functional Peaks resulting from a
Common Peak Shape Function}
Matt Hersh, Kert Viele, and Robin Cooper
Departments of Statistics and Biology, University of Kentucky
\abs
In synaptic transmission data, we observe electrical currents across
time. These electrical current traces have peaks resulting from the
release of transmitter from a vesicle across the synaptic cleft. If
multiple vesicles release transmitter, one may observe multiple peaks
within the electrical current trace. Previous work indicates that the
electrical current from a single vesicle release, while varying in
peak intensity and width of the synaptic potential, follow a similar
shape. Thus, we analyse our electrical currents under the assumption
that each function is the sum of equally shaped (though differingly
scaled) peak functions. This has similarities to a mixture model,
although instead of classifying individual observations into groups we
only have the functional data. Our goals are to simultaneously
estimate the underlying shape function and to classify each trace as
to the number of peaks it contains. We use a mixture of functions to
accomplish this task.
\gp % \newpage
\textbf{Intrusion Detection and Response using Effective Data Mining
Techniques: Classification, Clustering and Data Analysis in
Intrusion Information Retrieval}
Dr. Emmanuel Hooper, Information Security Group, University of London
\abs
There are major challenges in intrusion detection and data mining,
specifically, classification, clustering and data analysis of
intrusion and alert datasets. Information retrieval in Intrusion
datasets presents the problems of relevant feature attributes for
effective detection and response to intrusions/attacks and
alerts. Furthermore, the problem of false positives and detection
capability of intrusion detection systems (IDSs) are major
issues. This paper examines classification and clustering techniques
for effective detection and response strategies to network
infrastructure attacks. The approaches are two-fold: first,
Classification of known attacks and alerts, and secondly, clustering
of unknown attacks and alerts. The classification approach involves a
hybrid of Bayesian Classification and Discriminant Analysis for known
attacks and alerts. The clustering approach involves a hybrid of
Ward's agglomerative algorithm and Pearson's Association Correlation
along with Chi-square Analysis for unknown alerts and
attacks. Abstract subcategories of the feature attributes are used to
identify unknown attacks or benign alerts. Then appropriate responses
are sent to mitigate the impact of the attacks and filter benign
alerts from the IDS monitor. These strategies improve the performance
of the IDS and enhance responses to various subcategories of false
positives and complex attacks.
\gp % \newpage
\textbf{Text Classification with Customized Word Lists: Delta-Lz and Delta-Oz}
David L. Hoover, New York University
\abs
J. F. Burrows recently introduced Delta, a new measure of textual
difference for authorship attribution, and I have introduced five
variants that improve upon Delta's already impressive results. Further
testing confirms that two of the variants, Delta-Lz and Delta-Oz, are
especially effective. My presentation reports the results of further
inter-authorial and intra-authorial tests and investigates their
implications. Both measures compare test and authorial text samples
using individually determined and unique subsets of the word-frequency
list for all test and authorial samples combined. Delta-Lz focuses
only on words with test-sample frequencies quite different from the
mean for the authorial samples: each test sample determines its own
word-list. Delta-Oz compares each word's frequency in each test and
authorial sample to its mean frequency in all the authorial
samples. It compares only the frequencies of words with test-sample
and authorial-sample frequencies that differ in opposite directions
from the mean: each comparison between a test sample and an authorial
sample determines its own word-list. Closely examining the words
selected by Delta-Lz and Delta-Oz allows for a more accurate and
informative characterization of the texts, and comparing the various
lists should reveal why and how these methods are so effective.
\gp \newpage
\textbf{Constructing A Music Mood Taxonomy By Clustering}
Xiao Hu and J. Stephen Downie
GSLIS, University of Illinois at Urbana-Champaign
\abs
Classifying music pieces by the moods they express has attracted
researchers' interests and efforts, but a standardized mood taxonomy
is still in need for cross algorithm comparison and evaluation. This
research strives to construct a reasonable music mood taxonomy for a
common testbed of music mood classification in this year's Music
Information Retrieval Evaluation eXchange (MIREX). We clustered the
40 most popular mood labels on a leading music website, AllMusicGuild
(AMG) where each mood label is associated with a list of
representative ``Top Albums'' and a list of ``Top Songs''. As these top
albums and songs are often shared by different mood labels, they can
be exploited to group the mood labels into several super-classes,
which then would be a good candidate for a standardized mood taxonomy.
A co-occurrence matrix was formed for shared albums and shared songs
respectively. Pearson's correlation was calculated for each pair of
rows (or columns) as similarity measures between any two mood
labels. Then a hierarchical clustering procedure using Ward's link was
applied to such similarity measures. Comparing the clusters resulted
from mood-album co-occurrences and mood-song co-occurrences, we found
29 mood labels were consistently grouped into 5 different clusters at
about 1.5 distance level.
\gp % \newpage
\textbf{Avoiding Degeneracy in Multidimensional Unfolding by Combinatorial
Optimization}
Hans-Friedrich K\"{o}hn,
Department of Psychology, University of Illinois, Urbana-Champaign
\abs The recently introduced PREFSCAL algorithm (implemented in SPSS
Categories) for fitting distance-based nonmetric unfolding models
avoids degenerate solutions through augmenting the loss function by a
penalty term to maintain a sufficient level of variability among the
transformed proximities or pseudo-distances that prevents the distance
estimates and associated object coordinates from collapsing into a
single or a small number of locations.
We demonstrate that the same result can be obtained through a discrete
(combinatorial) optimization strategy that does not involve any
penalty functions for minimizing the loss function. In fact, PREFSCAL
and combinatorial unfolding yield almost indistinguishable solutions
for the test data sets employed.
\gp % \newpage
\textbf{The Astronomical Information Network}
Michael J. Kurtz
Harvard-Smithsonian Center for Astrophysics
\abs
Astronomical Objects (stars, galaxies, ...) are bound together by a
complex and often surprising assortment of shared ond/or similar
interactions and histories. As a purely observational science it is
the task of astronomy to disentangle the vast network of objects with
shared or similar properties and discover the underlying causal
relationships which govern our universe.
As an example some galaxies are blue, and exibit spectral features
typical of a hot, ionized gas; other galaxies are red, and show none
of these gaseous features. These red galaxies tend, very strongly, to
cluster together in space, while the blue galaxies do not. Edwin
Hubble first noticed this 75 years ago. What causes this effect, were
the galaxies formed this way, or did they become this way over the
history of the universe?
Astronomy research exists now within a large and growing man-made
network of tightly interconnected data sources and services. Based on
a culture of freely shared information and shared goals astronomers
are building a Virtual Observatory, using the internet to bring the
totality of astronomical information to anyone, anywhere in the world.
Astronomy research also exists within more abstract networks of
thought and behavior. The structure of astronomy research can be seen
in the co-citation network of astronomy research articles, but it can
also be seen in the co-reader network of those articles, and the
co-keyword network of those same articles again. Are these structures
the same?
\gp % \newpage
\textbf{Logistic Regression using Fractional Imputation for Missing Data}
Michael D. Larsen,
Department of Statistics and Center for Survey Statistics and Methodology,
Iowa State University
\abs
Imputation is used to fill in missing values so that analyses based on
complete data methods can be completed. Random imputation methods can
add imputation variance to the results. Not accounting for the fact
that some records are imputed can lead to understatement of
uncertainty in conclusions. A fractional imputation method for a
missing outcome variable in logistic regression are proposed and
implemented. The purpose of the methods is to reduce imputation
variance while allowing accurate estimation of uncertainty. The
method is applied to data from a longitudinal study of families in
Iowa.
\gp % \newpage
\textbf{Non-Parametric Modeling of Partially Ranked Data}
Guy Lebanon, Purdue University
\abs Statistical models on full and partial ranking of n items are
often of limited practical use for large n due to computational
consideration. We explore the use of non-parametric models for
partially ranked data and derive efficient procedures for their use
for large n. The derivations are largely possible through
combinatorial and algebraic manipulations based on the lattice of
partial rankings. In particular, we demonstrate for the first time a
non-parametric coherent and consistent model capable of efficiently
aggregating partially ranked data of different types.
\gp % \newpage
\textbf{Analysis of Information Features in Natural Language Queries
Seeking Music}
Jin Ha Lee, University of Illinois at Urbana-Champaign
\abs The deficiency of a formal taxonomy for representing
real-life queries is a major barrier in appropriate design and
evaluation of Music Information Retrieval (MIR)/Music Digital
Libraries (MDL) systems. To address this issue, prior studies have
analyzed real-life examples of natural language queries describing
users' sought music and identified the general types of information
features provided by users in their queries. However, the information
regarding the empirical association among query features that would
allow us to establish a meaningful classification scheme for these
features is still lacking. This study aims to discover associative
patterns in the kinds of information features provided in music
queries by an empirical investigation. Real-life queries were
collected from an online reference website and the categories of
different information features were established by an iterative coding
process. Different clustering approaches will be explored to infer
associations from patterns of co-occurrences among these features and
establish a classification scheme.
\gp % \newpage
\textbf{Large Scale Functional Data Clustering}
Ping Ma, Department of Statistics, University of Illinois at Urbana-Champaign
\abs
Large scale functional data rise from many scientific
investigations. Identifying cluster information is a crucial first
step to navigate further scientific investigation.
Motivated by analyzing temporal gene expression data, we propose a
novel functional data clustering method based on a mixture
smoothing-spline model. For each cluster, we model its mean profile
using a smoothing spline and describe its individual gene's variation
by a parametric random effect. We present an EM algorithm to find the
maximum a posteriori. Our method automatically takes care of the
missing data and infers the number of clusters in the data. Emprical
studies suggest that the proposed method outperforms the existing
methods.
\gp % \newpage
\textbf{How Many Clusters?}
Peter McCullagh, University of Chicago
Jie Yang, University of Illinois at Chicago
\abs The title poses a deceptively simple question that must be
addressed by any statistical model or computational algorithm for the
clustering of points. Two distinct interpretations are possible, one
connected with the number of clusters in the sample and one with the
number in the population. Under suitable conditions, these questions
may have essentially the same answer, but it is logically possible for
one answer to be finite and the other infinite. This paper
reformulates the standard Dirichlet allocation model as a cluster
process in such a way that these and related questions can be
addressed directly. Our conclusion is that the data are sometimes
informative for clustering points in the sample, but they seldom
contain much information about parameters such as the number of
clusters in the population.
\gp % \newpage
\textbf{Visualizing Clusters With a Density-Based Similarity Measure}
Rebecca Nugent, Department of Statistics, Carnegie Mellon University
Werner Stuetzle, Department of Statistics, University of Washington
Xiaoyi Fei, Department of Statistics, Carnegie Mellon University
\abs
The goal of clustering is to identify distinct groups in a dataset and
assign a group label to each observation. To cast clustering as a
statistical problem, we regard the data as a sample from an unknown density
p(x). To generate clusters, we estimate the properties of p(x) either with
parametric (model-based) or nonparametric methods. In contrast, the
algorithmic approach to clustering (linkage methods, spectral clustering)
applies an algorithm, often based on a distance measure, to data in m-
dimensional space.
Many commonly used clustering methods employ functions of Euclidean
distance between observations to determine groupings. Spherical groups are
easily identified, curvilinear groups less so. We first motivate the use
of a density-based similarity measure and briefly introduce generalized
single linkage, a graph-based clustering approach. We describe a
refinement algorithm used to bound the measure and then explore the
performance of this measure in clustering and visualization methods.
\gp % \newpage
\textbf{Discrimination and Classification of Non-Stationary
Brain Signals Using Higher Order Spectral Analysis}
Hernando Ombao, University of Illinois at Urbana-Champaign
Ringo Ho, Nanyang Technical University
\abs
We consider a data set that consists of MEG (magnetoencephalogram)
signals recorded from healthy controls and schizophrenic patients. Our
neuroscience collaborators are interested in identifying features that
can separate the two groups with the hope that these physiological
measures may be used in conjunction with behavioral measures for
patient diagnosis. In this talk, we will develop an automatic
procedure for time-frequency spectral feature selection via localized
transforms. Moreover, given the high degree of complexity of brain
signals, we will consider the time- evolutionary spectrum of
non-linear transforms as potential features for classification and
discrimination.
\gp % \newpage
\textbf{An evaluation of social cognitive mapping procedures for
identifying middle childhood social networks}
Philip C. Rodkin and Hai-Jeong Ahn
University of Illinois at Urbana-Champaign
\abs This study compares middle childhood social networks
constructed from children's reports of: (a) their own
affiliations and the affiliations of their peers (multi-informant
affiliations); (b) their own affiliations (self-reported
affiliations), and; (c) their own friendships. The sample
consisted of 390 fourth and fifth graders surveyed in fall and
spring. Multi-informant affiliation networks yielded larger peer
groups and fewer isolates than networks constructed from
friendship self-reports. Multi-informant affiliation networks
were more stable from fall to spring and more robust to
variations in statistical algorithms than friendship networks.
Multi-informant affiliation and friendship networks had poor
agreement with one another, particularly in their placement of
unpopular children. Multi-informant affiliation groups had
greater homophily on aggression than friendship groups and
accorded higher social centrality to aggressive
behavior. Self-reported affiliation networks were intermediate
between multi-informant affiliations and self-reported
friendships. Discussion focuses on incorporating multi-informancy
into emerging sociometric technologies.
\gp % \newpage
\textbf{Racially integrated social networks among African- and
European-American elementary children across differing classroom
contexts}
Philip C. Rodkin and Travis Wilson
University of Illinois at Urbana-Champaign
\abs
The racial composition of school classrooms, and the possible benefits
of diverse classrooms, receives much attention in debates about school
choice. We examine social integration between African- and
European-American children across three elementary classroom contexts:
11 classrooms that are majority White (65\% W, 35\% B), 11 that are
majority Black (65\% B, 35\% W), and 11 that are multicultural, with
equal proportions of Whites and Blacks and 30\% Hispanics and
Asians. Participants were 680 3rd and 4th graders. Social integration
was assessed by using a compositionally invariant ratio of same-race
to cross-race nominations of who children: (a) affiliated with, (b)
were friends with, (c) liked most, and (d) liked least. We expected
interactions where the salience of classroom context to social
integration would vary for European- and African-American
children. Using multilevel modeling, we consistently found
interactions between individual race and classroom racial majority. In
majority white classrooms, African-American children had more
segregated social relations and European- and African-American
children mutually nominated one another as disliked. Along with
substantive implications, discussion will focus on analytic concerns
of how to classify children into groups, and how integrate metrics
from research in child psychology and network science.
\gp % \newpage
\textbf{Objective Measurement of Fatigue in HIV/AIDS Using Actigraphy and
Functional Data Analysis}
William Shannon
Washington University School of Medicine
\abs
Behavioral changes associated with chronic HIV infection include
lethargy or fatigue defined as the inability to continue functioning
at a prescribed work rate in the presence of an increased perception
of effort. While the mechanism of this fatigue is uncertain, it may
relate to the intrinsic brain infection by HIV that also is the cause
of AIDS dementia complex. Fatigue is described by HIV patients as
``painful'' and is one of the most debilitating symptoms that limit
their quality of life. In addition to fatigue, sleep disturbances
also affect patients with HIV/AIDS, and likely contribute to the
severity of their daytime fatigue. Sleep dysfunction, and in
particular insomnia, are associated with CNS infection by HIV many
years before the onset of AIDS.
The diagnosis of ADC is based on neurological history, testing, and
examination and is only diagnosed when the patient's day-to-day
activities have already become severely degraded. The need for an
objective and early method for detecting neurologic impairment due to
HIV is vital for the proper treatment and management of HIV/AIDS. We
propose the use of actigraphy analyzed by functional data analysis
methods as an objective and early method for detecting neurologic
impairment due to HIV
\gp % \newpage
\textbf{Order-Constrained Solutions in $K$-Means Clustering: Even
Better Than Being Globally Optimal}
Douglas Steinley, University of Missouri
Lawrence Hubert, University of Illinois
\abs
An order-constrained K-means cluster analysis strategy
is proposed and implemented through an auxiliary quadratic
assignment optimization heuristic that identifies an initial
object order. A subsequent dynamic programming recursion is
applied to optimally subdivide the object set subject to the order
constraint. It is shown that although the usual K-means
sum-of-squared-error criterion is not guaranteed to be minimal, a
true underlying cluster structure may be more accurately
recovered. Also, substantive interpretability seems generally
improved when constrained solutions are considered. We illustrate
the procedure with several data sets from the literature.
\gp % \newpage
\textbf{A Scalable Approach to Clustering Massive Audio Catalogs}
David K. Tcheng, NCSA/ALG, One Llama Media Inc.
\abs
Our goal is to create a general purpose audio clustering algorithm that scales
well to the largest of audio data sets. Currently we are processing a 600
hour audio catalog of continuous (24/7) recordings from wireless microphones
mounted on birds (territorial cardinals). In addition audio recordings, we
have corresponding recordings of the birds X/Y position over time as estimated
by radio location. Our first task is to discover the syllables and grammar
that fully describe the cardinal utterances. Next we seek to understand bird
behavior -- how the utterances of a specific bird is influenced by the position
of the bird, time of day, season, weather, and the sounds of other birds and
animals in proximity. Our approach begins by creating a continuous
spectrogram of the audio waves using a bank of band pass filters. Next we
segment the spectrogram using a sliding window of fixed length (e.g., one
second). A hash table based clustering algorithm is used to find frequently
occurring patterns. A simulated annealing based clustering algorithm is used
to arrange the audio spectra segments into a 2-d map which maximizes the
relationship between map distance and audio spectra similarity.
\gp % \newpage
\textbf{Author name disambiguation in MEDLINE: results from first-pass
clustering}
Vetle I. Torvik and Neil R. Smalheiser
University of Illinois at Chicago
\abs
We have previously described (Torvik et al., JASIST, 2005) a method for
automatically generating training sets and estimating the probability that
a pair of papers in MEDLINE sharing a last and first name initial are
authored by the same individual. The probability estimates were based on
shared title words, journal name, co-author names, medical subject
headings, language, and affiliation, as well as distinctive features of the
name itself (i.e., presence of middle initial, suffix, and prevalence in
MEDLINE). This project, which we call ``Author-ity'', was subsequently funded
by an NIH grant to create a database of all papers in MEDLINE clustered by
predicted author-individuals. We have recently completed a first-pass
clustering of the 2006 baseline version of MEDLINE (> 15 million papers)
resulting in > 5 million predicted author-individuals. In this talk, we
will summarize our work to date: a) the basic pairwise model with recently
added and modified predictive features (first name variants, email address,
last name specific affiliation word stoplists); b) new automatic methods of
generating large training sets; c) methods for estimating the prior
probability for given a population of papers to be compared; d) a weighted
least squares algorithm for correcting triplet violations of the form
$p_{ij} + p_{ik} \ge 1 + p_{jk}$,
e) a database with predicted author-individuals resulting
from simple agglomerative clustering; and f) some preliminary data for
future research directions.
\gp \newpage
\textbf{Robust Partial Logistic Regression (RoPLR)}
Asuman S. Turkmen and Nedret Billor
Department of Mathematics and Statistics,
Auburn University
\abs
Partial least squares (PLS) is a statistical technique to summarize
high dimensional and correlated predictor variables into low
dimensional, uncorrelated variables which have the best predictive
power. High dimensionality and collinearity make the application of
the available classification methods difficult and even impossible for
some cases. An analyst can reduce the dimension by constructing new
components employing PLS technique and then apply a classification
method on the constructed PLS components. Even though PLS was
originally designed for continuous response variables, only for last
six years, it has become a widely used statistical dimension reduction
technique for classification. The classical PLS method is known to be
very sensitive to outlying observations that usually exist in
experimental data. Therefore several robust PLS methods have been
proposed when the response variable is continuous. However, to our
knowledge, there has been no study on robustness of PLS methods for
dimension reduction in classification. In this study, the effect of
outliers on existing PLS classification methods is investigated. We
also propose a new PLS algorithm based on robust logistic regression
(RoPLR) for classification problems. Real and simulated data sets are
used to demonstrate the performance of RoPLR method.
\gp % \newpage
\textbf{Classification of Self-Modeling Regressions}
Rhonda VanDyke, Kert Viele, and Robin Cooper
Departments of Statistics and Biology, University of Kentucky
\abs
A set of self-modeling functions is defined by the entire set of
functions being related through affine transformations of the $x$ and
$y$ axes to a common function $g(t)$. We expand this definition to
include the possibility that a set of functions contains two
underlying sets of self-modeling functions, the first related through
a common shape $g_1(t)$ and the second related through a separate shape
function $g_2(t)$. Our goal is to take data consisting of a set of
functions, estimate the two underlying shape functions, and to
classify each function as belonging to either the first or second
group of self-modeling functions. We estimate the underlying shape
functions through Bayesian Adaptive Regression Splines (BARS). We
illustrate the methodology through Synaptic Transmission data, where
the functions measure electrical current across time and the two
self-modeling groups of functions are hypothesized to result from
different vesicles within synapses releasing transmitter in
qualitatively different manners.
\gp % \newpage
\textbf{Beta-Distributed Generalized Linear Mixed Models}
Jay Verkuilen, Department of
Psychology, University of Illinois Urbana-Champaign
Michael Smithson, School of Psychology, The Australian
National University.
\abs
Smithson and Verkuilen (2006) proposed the use of beta regression for
interval data with known lower and upper bounds. Variables of this
form abound in the behavioral sciences and are frequently not
well-modeled by Gaussian error distributions. For instance, confidence
ratings, judged probabilities, or leverage ratios of firms are all in
the unit interval and other bounded interval variables can be rescaled
to the unit interval without loss of generality. These variables
frequently exhibit marked skew or floor/ceiling effects. Unlike the
Gaussian, the beta distribution better accommodates empirical
distributions which may be L- or J-shaped, long tailed, or
symmetric. Building on our prior work, we present a regression model
that assumes, conditional on fixed predictors and random effects, the
dependent variable is beta distributed. Estimation by marginal maximum
likelihood or Bayesian MCMC works quite well in practice. The model is
illustrated with repeated measures data exhibiting between-subject
heteroscedasticity taken from an experiment considering the
"LakeWobegon effect".
\gp % \newpage
\textbf{Multilevel Latent Markov Models for Discrete Longitudinal Data}
Hsiu-Ting Yu, Department of Psychology, University of Illinois at
Urbana-Champaign
\abs
Multilevel longitudinal data are clustered both temporally and
spatially. The longitudinal or repeated measures within subjects
aspect of the data extends horizontally along a time dimension. The
multilevel or hierarchically nested structure extends vertically on a
spatial dimension. Since both types of clustered structures can induce
dependency into the data, both aspects need to take into account when
modeling the data. Many developments for multilevel and longitudinal
data have focused on continuous response or outcome variables;
however, less attention has been paid to the data with discrete
manifest and latent variables. In this talk, I will review the
classical latent class models for discrete data and discuss the
current methods of the extension on each of the clustering data
structures. Multilevel Latent Markov Models are proposed to unify the
two types of dependency due to clustering in a single model. The
proposed models are hybrids of random-effects and conditional models,
where conditional model is adopted to model the change between two
occasions; and random-effects modeling approach is utilized to account
for the effects of nested structure. A data set from Educational
Longitudinal Study of 2002 is used to illustrate various models for
discrete longitudinal data with multilevel data structure.
\end{document}