WSOM+ 2024
15th Workshop on Self-Organizing Maps,
Learning Vector Quantization & Beyond
10 - 12 July 2024 | Mittweida, Germany
Conference Agenda
Overview and details of the sessions of this conference. Please select a date or location to show only sessions at that day or location. Please select a single session for detailed view (with abstracts and downloads if available).
Please note that all times are shown in the time zone of the conference. The current conference time is: 2nd May 2025, 03:33:37pm CEST
|
Session Overview |
Date: Wednesday, 10/July/2024 | |
9:00am - 10:00am | Welcome and Coffee Location: 39-001 |
10:00am - 10:15am | Opening Location: 39-001 |
10:15am - 11:55am | Prototype-Based Supervised & Unsupervised Learning Location: 39-001 Session Chair: Alexander R.T. Gepperth |
|
New Cloth unto an Old Garment: SOM for Regeneration Learning 1PETROBRAS, Brazil; 2Graduate Program in Teleinformatics Engineering, Center of Technology, Federal University of Ceará (UFC), Fortaleza - CE, Brazil; 3Department of Electrical Engineering, Pontifical Catholic University of Rio de Janeiro (PUC-Rio), Rio de Janeiro - RJ, Brazil A recent paradigm called Regeneration Learning addresses generative problems where the target data (e.g., images) is more complex than the available input source. While current cross-modal representation and regeneration learning rely on supervised deep learning models, this paper aims to revisit the adequacy of unsupervised models in this field. In this regard, we propose a new unsupervised approach that utilizes the SOM as a heteroassociative memory model to learn cross-modal representations in a topologically coherent map. This approach enables bidirectional predictive/regenerative mapping between domains. We evaluate the potential of this method on an unsolved (so far!) practical problem in petroleum geoscience. Unsupervised Learning-based Data Collection Planning with Dubins Vehicle and Constrained Data Retrieving Time Czech Technical University in Prague, Czech Republic In remote data collection from sampling stations, a vehicle must be within sufficient distance from the particular station for a predefined minimal time to retrieve all the required data from the site. The planning task is to find a cost-efficient data collection trajectory, allowing the data collection vehicle to retrieve data from all sensing sites. Having a fixed-wing aerial vehicle flying with a constant forward velocity, the problem is to determine the shortest feasible path that visits every sensing site and the vehicle is within a reliable communication distance from the station in a sufficient period. We propose to formulate the planning problem as a~variant of the Close Enough Dubins Traveling Salesman Problem with Time Constraints (CEDTSP-TC) that is heuristically solved by unsupervised learning of the Growing Self-Organizing Array (GSOA) modified to address the minimal required time for the vehicle to be within the communication range of the station. The proposed method is compared with a baseline based on a sampling-based decoupled approach. The presented results support the feasibility of both proposed solvers on random instances and show that the GSOA-based approach outperforms the decoupled approach or provides similar results. Hyperbox Learning Vector Quantization Based on Min-Max-Neurons University of Applied Sciences Mittweida, Germany In this paper we propose the application of min-max-neurons for the use in generalized learning vector quantization (GLVQ) models, which correspond to min-max-prototypes. These prototypes can be identified with hyperboxes in the data space. Keeping the general GLVQ cost function, we redefine the Hebb-responsibilities for min-max-prototypes and derive consistent learning rules for stochastic gradient descent learning. We demonstrate that the resulting hyperbox-based GLVQ is capable to solve several classification tasks in robust manner, which can be dedicated to the use of robust min-max-prototypes. Finally, we give suggestions for future research for GLVQ based on min-max-prototypes. Sparse Clustering with K-means - Which Penalties and for Which Data? 1Institut de Mathématique de Bordeaux, France; 2Université Paris 1 Panthéon Sorbonne, France; 3Université Paris Dauphine PSL, France While high dimensionality and the selection of meaningful features is usually a burden in machine learning, it is even more so in the case of unsupervised learning and particularly in clustering. The presence of uninformative features, sometimes correlated, may bias significantly the results of distance-based methods such as k-means for instance. Since the seminal work of Witten et al. (2010), different versions of sparse k-means have been introduced, building on the idea of adding some penalty terms in the loss function and resulting into automatic feature selection and/or weighting. This paper investigates the connections between some of these methods, and particularly the differences induced by the choices of the penalty terms. It also focuses on the case of mixed data, and how they may be handled by the sparse k-means approaches. Eventually, it presents the algorithms and model selection tools made available through a recently implemented R package, vimpclust. |
12:00pm - 1:00pm | Lunch Location: 39-001 |
1:00pm - 2:00pm | Invited speaker Location: 39-001 Session Chair: Michael Biehl |
|
Is t-SNE Becoming the New Self-organizing Map? Similarities and Differences FNRS Research Director Head of Molecular Imaging, Radiotherapy, and Oncology Université catholique de Louvain, Belgium Self-organizing maps (SOMs) have many advantages as a tool for exploratory data analysis. Combining vector quantization and topological relationships that are defined in a low-dimensional space, they can run on big data sets and are mostly immune to the curse of dimensionality in the data space. SOMs are used mainly for dimensionality reduction and marginally for clustering; however, SOMs also suffer from some shortcomings. Vector quantization makes them unable to embed all data points, only prototypes or centroid are mapped. Being defined as a regular grid in the low-dimensional space, dimensionality reduction and clustering with SOMs are indirect, as compared to methods of direct embedding like multi-dimensional scaling. Since 2008, t-SNE (t-distributed stochastic neighbor embedding) has raised growing interest, first in the machine learning community and now outside of it, with many applications in cell biology, for instance. Primarily used as a 2D embedding and visualization technique, t-SNE is more and more used as a clustering technique, which is capable of identifying meaningful clusters that classical clustering tools struggle to see. Quite counter-intuitively, t-SNE often better separates clusters in low-dimensional embeddings than clustering tools would do in the high-dimensional data space. To understand this paradox, several mechanisms of t-SNE can be framed as a distance transformation with the possibility (i) to denoise distances in high-dimensional spaces and (ii) to impose a strong inductive bias on them, which magnifies inter-cluster gaps. Despite these strengths, t-SNE is not free of drawbacks, which we quickly review to sketch perspectives of future developments. |
2:00pm - 2:30pm | Coffeebreak Location: 39-001 |
2:30pm - 3:45pm | Visualization Location: 39-001 Session Chair: Verleysen Michel |
|
Pursuing the Perfect Projection: A Projection Pursuit Framework for Deep Learning Otto von Guericke University, Germany Generalizing Self-Organizing Maps: Large-Scale Training of Gaussian Mixture Models and Applications in Data Science HAW Fulda, Germany A Self-Organizing UMAP For Clustering The University of Texas at Austin, United States of America |
4:00pm - 5:15pm | Bioinformatics Location: 39-001 Session Chair: Guilherme A. Barreto |
|
Knowledge Integration in Vector Quantization Models and Corresponding Structured Covariance Estimation Hochschule Mittweida, Germany Exploring Data Distributions In Machine Learning Models With SOMs Universitat Politècnica de Catalunya, Spain Interpretable Machine Learning In Endocrinology: A Diagnostic Tool In Primary Aldosteronism 1Bernoulli Intitute for Mathematics, Computer Science and Artificial Intelligence, University of Groningen, The Netherlands; 2Centre for Systems Modelling and Quantitative Biomedicine, University of Birmingham, UK; 3NIHR Birmingham Biomedical Research Centre, University of Birmingham and University Hospitals Birmingham NHS Foundation Trust, Birmingham, UK; 4Institute of Applied Health Research, University of Birmingham, UK; 5Institute of Metabolism and Systems Research, College of Medical and Dental Sciences, University of Birmingham, UK; 6Centre for Endocrinology, Diabetes and Metabolism, Birmingham Health Partners, Birmingham, UK; 7Medical Research Council Laboratory of Medical Sciences, London, UK; 8Institute of Clinical Sciences, Faculty of Medicine, Imperial College London, UK |
Date: Thursday, 11/July/2024 | |
9:00am - 11:00am | MIWOCI Workshop: Poster Spotlight and Poster Session Location: 39-001 Session Chair: Tina Geweniger |
|
IRMA on Steroids: Improved Robustness and Interpretability of Feature Relevances University of Groningen Rediscovering Chaos? Analysis of GPU Computing Effects in Graph-coupled NeuralODEs Friedrich-Alexander University Erlangen-Nürnberg, Germany Coping with Drift in Hyperspectral Sensor Data Bielefeld University, Germany A Measure Theoretic Approach to Concept Drift in Infinite Data Streams Bielefeld University, Germany Probabilistic Learning Vector Quantization Based on Cross-Entropy-Loss and Integration of Class Relation Knowledge Hochschule Mittweida, Germany Online Learning Dynamics in Layered Neural Networks with Arbitrary Activation Functions RUG, Groningen, The Netherland GMLVQ for fMRI Analysis in the Context of Movement Disorders RUG, Groningen, The Netherland Towards Explainable Rejects for Prototype-Based Classifiers. Universität Bielefeld Phase Transitions of Soft Committee Machines with Arbitrary Activation Function University of Groningen, The Netherlands Runtime-Processing for Microgravity Investigations Directly Attached to the Experiment Hochschule Mittweida, Germany |
9:00am - 11:00am | Poster Spotlights and Poster Session Location: 39-001 Session Chair: Tina Geweniger |
|
Probabilistic Models with Invariance HAW Fulda, Germany Optimizing YOLOv5 for Green AI: A Study on Model Pruning and Lightweight Networks Technical University of Applied Science Wuerzburg-Schweinfurt, Germany Process Phase Monitoring in Industrial Manufacturing Processes with a Hybrid Unsupervised Learning Strategy Fraunhofer-Institut für Optronik, Systemtechnik und Bildauswertung IOSB, Germany Knowledge Management in SMEs: Applying Link Prediction for Assisted Decision Making 1Institute for Applied Informatics at Leipzig University, Leipzig 04109, Germany; 2Chemnitz University of Technology, Institute for Management and Factory Systems, 09125 Chemnitz, Germany |
11:00am - 11:45am | Explainable and Interpretable Models I Location: 39-001 Session Chair: Sascha Saralajew |
|
Precision and Recall Reject Curves Honda Research Institute Europe, Germany K Minimum Enclosing Balls For Outlier Detection 1Saxon Institute for Computational Intelligence and Machine Learning, University of Applied Sciences, Mittweida, Germany; 2Faculty of Technology, Bielefeld University; 3Educational Technology Lab, German Research Center for Artificial Intelligence |
11:45am - 12:00pm | Group Photo |
12:00pm - 1:00pm | Lunch Location: 39-001 |
12:00pm - 1:00pm | Steering Commitee |
1:00pm - 2:00pm | Invited speaker II Location: 39-001 Session Chair: Frank-Michael Schleif |
|
The Beauty of Prototype Based Learning School of Computer Science University of Birmingham, UK There is a long history of formulation of machine learning models that have at their core "representative prototypes" living in the data space (examples include variants of Self Organizing Maps in the unsupervised setting or variants of Learning Vector Quantization in the supervised one). The prototype based formulations make such models intuitive and naturally amenable to interpretation. I will present a series of generalizations and extensions of this idea, arguing that prototype based learning still deserves attention in the current era of deep learning. In particular, I will present extensions of prototype based learning to model spaces and Riemannian manifolds. I will also show how prototype based models can be naturally extended to the setting of ordinal regression and learning with privileged information. |
2:15pm - 3:15pm | Statistical Methods and Learning Location: 39-001 Session Chair: Lydia Fischer |
|
Setting Vector Quantizer Resolution via Density Estimation Theory The University of Texas at Austin, United States of America Practical Approaches to Approximate Dominant Eigenvalues in Large Matrices Technical University of Applied Science Wuerzburg-Schweinfurt, Germany Enhancing LDA Method by the Use of Feature Maximization Université de Strasbourg, France |
3:30pm - 10:00pm | Social Dinner at Castle Rochlitz |
Date: Friday, 12/July/2024 | |
9:00am - 10:00am | Invited speaker III Location: 39-001 Session Chair: Sven Hellbach |
|
Explaining Neural Networks - Deep and Shallow Machine Learning Group, CITEC Bielefeld University, Germany Variable importance determination refers to the challenge to identify the most relevant input dimensions or features for a given learning task and quantify their relevance, either with respect to a local decision or a global model. Feature relevance determination constitutes a foundation for feature selection, and it enables an intuitive insight into the rational of model decisions. Indeed, it constitutes one of the oldest and most prominent explanation technologies for machine learning models with relevance for both, deep and shallow networks. A huge number of measures have been proposed such as mutual information, permutation feature importance, deep lift, LIME, GMLVQ, or Shapley values, to name just a few. Within the talk, I will address recent extensions of feature relevance determination, which occur as machine learning models are increasingly used in everyday life. Here, models face an open environment, possibly changing dynamics, and the necessity of model adaptation to account for changes of the underlying distribution. At present, feature relevance determination almost solely focusses on static scenarios and batch training. In the talk, I will target the question of how to efficiently and effectively accompany a model which learns incrementally by feature relevance determination methods(1,2). As a second challenge, features are often not mutually independent, and the relevance of groups rather than single features should be judged. While mathematical models such as Shapley values take feature correlations into account for individual additive feature relevance terms, it is unclear how to efficiently and effectively extend those to groups of features. In the talk, I will discuss novel methods for the efficient computation of feature interaction indices (3,4). (1) F Fumagalli, M Muschalik, E Hüllermeier, B Hammer: Incremental Permutation Feature Importance (iPFI): Towards Online Explanations on Data Streams. Mach. Learn. 112(12): 4863-4903, 2023. (2) M Muschalik, F Fumagalli, B Hammer, E Hüllermeier: iSAGE: An Incremental Version of SAGE for Online Explanation on Data Streams. ECML/PKDD (3) 2023: 428-445.(3) F Fumagalli, M Muschalik, P Kolpaczki, E Hüllermeier, B Hammer: SHAP-IQ: Unified Approximation of Any-Order Shapley Interactions. NeurIPS 2023.(4) M Muschalik, F Fumagalli, B Hammer, E Hüllermeier: Beyond TreeSHAP: Efficient Computation of Any-Order Shapley Interactions for Tree Ensembles AAAI 2024. |
10:00am - 10:30am | Coffeebreak Location: 39-001 |
10:30am - 11:15am | Explainable and Interpretable Models II Location: 39-001 Session Chair: Dietlind Zühlke |
|
FairGLVQ: Fairness in Partition-Based Classification Bielefeld University, Germany About Interpretable Learning Rules For Vector Quantizers - A Methodological Approach University of Applied Sciences Mittweida, Germany |
11:15am - 11:45am | Closing Location: 39-001 |
Contact and Legal Notice · Contact Address: Privacy Statement · Conference: WSOM+ 2024 |
Conference Software: ConfTool Pro 2.8.105 © 2001–2025 by Dr. H. Weinreich, Hamburg, Germany |