---- Neurocomputing special issue on Visual Analytics using Multidimensional 
Projections ----
---- Extended submission deadline: DECEMBER 1ST, 2013 ----


Following the successful workshop on Visual Analytics using Multidimensional 
Projections (VAMP) at EuroVis 2013 
(http://homepage.tudelft.nl/19j49/eurovis2013/), we solicit submissions to a 
special issue of the Neurocomputing journal on Visual Analytics using 
Multidimensional Projections (VAMP).

Submissions are due December 1st, 2013 via the Neurocomputing submission system 
(please select "SI: VAMP 2013" when you reach the “Article Type” step in the 
submission process). Each paper will be reviewed by experts from both Machine 
Learning and Visualization communities.

Dimensionality reduction is an active area in machine learning. New techniques 
have been proposed for more than 50 years, for instance, principal component 
analysis, classical scaling, isomap, probabilistic latent trait models, 
stochastic neighbor embedding, and neighborhood retrieval visualization. These 
techniques facilitate the visualization of high-dimensional data by 
representing data instances as points in a two-dimensional space in such a way 
that similar instances are modeled by nearby points and dissimilar instances 
are modeled by distant points.

Although many papers on these so-called “embedding” techniques are published 
every year, which all aim to improve visual representations of high-dimensional 
data, it appears that these techniques have not gained popularity in the 
information-visualization community due to the inherent complexity of their 
interpretation.

At the cross-section of information visualization, machine learning, and graph 
drawing, the special issue intends to focus on issues that embedding techniques 
should address to bridge the gap with the information-visualization community. 
A non-exhaustive list of such issues is given below:

⎯ Stability: Nonlinear embedding techniques are more efficient at preserving 
similarities than linear ones. However, non-linearities generate local optima 
as a result of which different initializations lead to different 
representations of the same data. The differences between these embeddings of 
the same data create confusion for the analyst, who is unable to grasp the 
common facts across the different visualizations. How can we design efficient 
and stable nonlinear embeddings?

⎯ Embedding of dynamic data: Embedding usually projects all the data at once; 
when new data arrive, how can we embed these data without modifying the current 
embedding too much?

⎯ Multiple methods: Each embedding algorithm necessarily comes with its own set 
of built-in underlying assumptions, and knowledge of these assumptions is often 
helpful in making sense of the visual output. How can we design black-box 
visualization methods that demand less understanding of underlying assumptions 
from the side of the analyst?

⎯ Evaluation and subjectivity: Visual interpretation is inherently subjective. 
How can we help analysts to verify whether an eye-catching pattern is 
real/essential or whether it just happens to be an artifact?

⎯ Inference and interactions: Nonlinear embedding techniques produce points 
clouds in which the axes have no meaning and pairwise distances are 
approximations which may have many artifacts. What kinds of analytical tasks 
can be performed with such embeddings? How can we better convey the meaning of 
the embeddings to analysts?

⎯ Feedback: The human eye is excellent at visual analysis, and can identify 
regularities and anomalous data even without having to define an algorithm. How 
can we make use of this ability to enhance the predictive performance of 
machine learning and embedding techniques?

⎯ Input data: Currently, the input data in embedding techniques typically 
comprises high-dimensional feature vectors or pairwise distance between 
objects. However, this is not always the kind of data that analysts encounter 
in practice. How can embeddings be constructed based on partial similarity 
rankings, associations or co-occurences of objects, heterogeneous data, data 
with missing values, relations between objects, structured objects, etc.?

⎯ Optimizing embeddings for visual analysis: nonlinear embeddings are found by 
optimizing mathematical goodness-of-fit measures. Instead of using 
off-the-shelf embedding methods, can the measures and methods be designed so 
that the optimized embeddings will be good for carrying out concrete low-level 
or high-level analysis tasks from the visualization?

The special issue aims to attract contributions on these or related topics. 

In case of any questions about this special issue, please contact the guest 
editors: Laurens van der Maaten <[email protected]> and Michaël 
Aupetit <[email protected]>.
_______________________________________________
uai mailing list
[email protected]
https://secure.engr.oregonstate.edu/mailman/listinfo/uai

Reply via email to