Scientific Works of Johannes Hirth.
2024
- 1.Draude, C., Engert, S., Hess, T., Hirth, J., Horn, V., Kropf, J., Lamla, J., Stumme, G., Uhlmann, M., Zwingmann, N.: Verrechnung – Design – Kultivierung: Instrumentenkasten für die Gestaltung fairer Geschäftsmodelle durch Ko-Valuation, https://plattform-privatheit.de/p-prv-wAssets/Assets/Veroeffentlichungen_WhitePaper_PolicyPaper/whitepaper/WP_2024_FAIRDIENSTE_1.0.pdf, (2024). https://doi.org/10.24406/publica-2497.
@misc{claude2024verrechnung,
address = {Karlsruhe},
author = {Draude, Claude and Engert, Simon and Hess, Thomas and Hirth, Johannes and Horn, Viktoria and Kropf, Jonathan and Lamla, Jörn and Stumme, Gerd and Uhlmann, Markus and Zwingmann, Nina},
edition = 1,
editor = {Friedewald, Michael and Roßnagel, Alexander and Geminn, Christian and Karaboga, Murat},
howpublished = {White Paper},
keywords = {itegpub},
month = {03},
publisher = {Fraunhofer-Institut für System- und Innovationsforschung ISI},
series = {Plattform Privatheit},
title = {Verrechnung – Design – Kultivierung: Instrumentenkasten für die Gestaltung fairer Geschäftsmodelle durch Ko-Valuation},
year = 2024
}%0 Generic
%1 claude2024verrechnung
%A Draude, Claude
%A Engert, Simon
%A Hess, Thomas
%A Hirth, Johannes
%A Horn, Viktoria
%A Kropf, Jonathan
%A Lamla, Jörn
%A Stumme, Gerd
%A Uhlmann, Markus
%A Zwingmann, Nina
%B Plattform Privatheit
%C Karlsruhe
%D 2024
%E Friedewald, Michael
%E Roßnagel, Alexander
%E Geminn, Christian
%E Karaboga, Murat
%I Fraunhofer-Institut für System- und Innovationsforschung ISI
%R 10.24406/publica-2497
%T Verrechnung – Design – Kultivierung: Instrumentenkasten für die Gestaltung fairer Geschäftsmodelle durch Ko-Valuation
%U https://plattform-privatheit.de/p-prv-wAssets/Assets/Veroeffentlichungen_WhitePaper_PolicyPaper/whitepaper/WP_2024_FAIRDIENSTE_1.0.pdf
%7 1 - 1.Hirth, J.: Conceptual Data Scaling in Machine Learning, (2024). https://doi.org/10.17170/kobra-2024100910940.Information that is intended for human interpretation is frequently represented in a structured manner. This allows for a navigation between individual pieces to find, connect or combine information to gain new insights. Within a structure, we derive knowledge from inference of hierarchical or logical relations between data objects. For unstructured data there are numerous methods to define a data schema based on user interpretations. Afterward, data objects can be aggregated to derive (hierarchical) structures based on common properties. There are four main challenges with respect to the explainability of the derived structures. First, formal procedures are needed to infer knowledge about the data set, or parts of it, from hierarchical structures. Second, what does knowledge inferred from a structure imply for the data set it was derived from? Third, structures may be incomprehensibly large for human interpretation. Methods are needed to reduce structures to smaller representations in a consistent, comprehensible manner that provides control over possibly introduced error. Forth, the original data set does not need to have interpretable features and thus only allow for the inference of structural properties. In order to extract information based on real world properties, we need methods that are able to add such properties. With the presented work, we address these challenges using and extending the rich tool-set of Formal Concept Analysis. Here, data objects are aggregated to closed sets called formal concepts based on (unary) symbolic attributes that they have in common. The process of deriving symbolic attributes is called conceptual scaling and depends on the interpretation of the data by the analyst. The resulting hierarchical structure of concepts is called concept lattice. To infer knowledge from the concept lattice structures we introduce new methods based on sub-structures that are of standardized shape, called ordinal motifs. This novel method allows us to explain the structure of a concept lattice based on geometric aspects. Throughout our work, we focus on data representations from multiple state-of-the-art machine learning algorithms. In all cases, we elaborate extensively on how to interpret these models through derived concept lattices and develop scaling procedures specific to each algorithm. Some of the considered models are black-box models whose internal data representations are numeric with no clear real world semantics. For these, we present a method to link background knowledge to the concept lattice structure. To reduce the complexity of concept lattices we provide a new theoretical framework that allows us to generate (small) views on a concept lattice. These enable more selective and comprehensibly sized explanations for data parts that are of interest. In addition to that, we introduce methods to combine and subtract views from each other, and to identify missing or incorrect parts.
@phdthesis{doi:10.17170/kobra-2024100910940,
abstract = {Information that is intended for human interpretation is frequently represented in a structured manner. This allows for a navigation between individual pieces to find, connect or combine information to gain new insights. Within a structure, we derive knowledge from inference of hierarchical or logical relations between data objects. For unstructured data there are numerous methods to define a data schema based on user interpretations. Afterward, data objects can be aggregated to derive (hierarchical) structures based on common properties. There are four main challenges with respect to the explainability of the derived structures. First, formal procedures are needed to infer knowledge about the data set, or parts of it, from hierarchical structures. Second, what does knowledge inferred from a structure imply for the data set it was derived from? Third, structures may be incomprehensibly large for human interpretation. Methods are needed to reduce structures to smaller representations in a consistent, comprehensible manner that provides control over possibly introduced error. Forth, the original data set does not need to have interpretable features and thus only allow for the inference of structural properties. In order to extract information based on real world properties, we need methods that are able to add such properties. With the presented work, we address these challenges using and extending the rich tool-set of Formal Concept Analysis. Here, data objects are aggregated to closed sets called formal concepts based on (unary) symbolic attributes that they have in common. The process of deriving symbolic attributes is called conceptual scaling and depends on the interpretation of the data by the analyst. The resulting hierarchical structure of concepts is called concept lattice. To infer knowledge from the concept lattice structures we introduce new methods based on sub-structures that are of standardized shape, called ordinal motifs. This novel method allows us to explain the structure of a concept lattice based on geometric aspects. Throughout our work, we focus on data representations from multiple state-of-the-art machine learning algorithms. In all cases, we elaborate extensively on how to interpret these models through derived concept lattices and develop scaling procedures specific to each algorithm. Some of the considered models are black-box models whose internal data representations are numeric with no clear real world semantics. For these, we present a method to link background knowledge to the concept lattice structure. To reduce the complexity of concept lattices we provide a new theoretical framework that allows us to generate (small) views on a concept lattice. These enable more selective and comprehensibly sized explanations for data parts that are of interest. In addition to that, we introduce methods to combine and subtract views from each other, and to identify missing or incorrect parts.},
author = {Hirth, Johannes},
keywords = {Knowldege~Representation},
school = {Kassel, Universität Kassel, Fachbereich Elektrotechnik/Informatik},
title = {Conceptual Data Scaling in Machine Learning},
year = 2024
}%0 Thesis
%1 doi:10.17170/kobra-2024100910940
%A Hirth, Johannes
%D 2024
%R 10.17170/kobra-2024100910940
%T Conceptual Data Scaling in Machine Learning
%X Information that is intended for human interpretation is frequently represented in a structured manner. This allows for a navigation between individual pieces to find, connect or combine information to gain new insights. Within a structure, we derive knowledge from inference of hierarchical or logical relations between data objects. For unstructured data there are numerous methods to define a data schema based on user interpretations. Afterward, data objects can be aggregated to derive (hierarchical) structures based on common properties. There are four main challenges with respect to the explainability of the derived structures. First, formal procedures are needed to infer knowledge about the data set, or parts of it, from hierarchical structures. Second, what does knowledge inferred from a structure imply for the data set it was derived from? Third, structures may be incomprehensibly large for human interpretation. Methods are needed to reduce structures to smaller representations in a consistent, comprehensible manner that provides control over possibly introduced error. Forth, the original data set does not need to have interpretable features and thus only allow for the inference of structural properties. In order to extract information based on real world properties, we need methods that are able to add such properties. With the presented work, we address these challenges using and extending the rich tool-set of Formal Concept Analysis. Here, data objects are aggregated to closed sets called formal concepts based on (unary) symbolic attributes that they have in common. The process of deriving symbolic attributes is called conceptual scaling and depends on the interpretation of the data by the analyst. The resulting hierarchical structure of concepts is called concept lattice. To infer knowledge from the concept lattice structures we introduce new methods based on sub-structures that are of standardized shape, called ordinal motifs. This novel method allows us to explain the structure of a concept lattice based on geometric aspects. Throughout our work, we focus on data representations from multiple state-of-the-art machine learning algorithms. In all cases, we elaborate extensively on how to interpret these models through derived concept lattices and develop scaling procedures specific to each algorithm. Some of the considered models are black-box models whose internal data representations are numeric with no clear real world semantics. For these, we present a method to link background knowledge to the concept lattice structure. To reduce the complexity of concept lattices we provide a new theoretical framework that allows us to generate (small) views on a concept lattice. These enable more selective and comprehensibly sized explanations for data parts that are of interest. In addition to that, we introduce methods to combine and subtract views from each other, and to identify missing or incorrect parts. - 1.Horn, V., Hirth, J., Holfeld, J., Behmenburg, J.H., Draude, C., Stumme, G.: Disclosing Diverse Perspectives of News Articles for Navigating between Online Journalism Content. In: Nordic Conference on Human-Computer Interaction. Association for Computing Machinery, Uppsala, Sweden (2024). https://doi.org/10.1145/3679318.3685414.Today, exposure to journalistic online content is predominantly controlled by news recommender systems, which often suggest content that matches user’s interests or is selected according to non-transparent recommendation criteria. To circumvent resulting trade-offs like polarisation or fragmentation whilst ensuring user’s autonomy, we explore how different perspectives within online news can be disclosed instead for guiding navigation. To do so, we developed an interactive prototype that displays article titles in correspondence to their argumentative orientation. In order to investigate how the usage of our novel navigation structure impacts the choice of news articles and user experience, we conducted an exploratory user study assessing the impact of the design parameters chosen. Implications are drawn from the study results and the development of the interactive prototype for the exposure to diversity in the context of navigating news content online.
@inproceedings{hci-lattice,
abstract = {Today, exposure to journalistic online content is predominantly controlled by news recommender systems, which often suggest content that matches user’s interests or is selected according to non-transparent recommendation criteria. To circumvent resulting trade-offs like polarisation or fragmentation whilst ensuring user’s autonomy, we explore how different perspectives within online news can be disclosed instead for guiding navigation. To do so, we developed an interactive prototype that displays article titles in correspondence to their argumentative orientation. In order to investigate how the usage of our novel navigation structure impacts the choice of news articles and user experience, we conducted an exploratory user study assessing the impact of the design parameters chosen. Implications are drawn from the study results and the development of the interactive prototype for the exposure to diversity in the context of navigating news content online.},
address = {New York, NY, USA},
author = {Horn, Viktoria and Hirth, Johannes and Holfeld, Julian and Behmenburg, Jens Hendrik and Draude, Claude and Stumme, Gerd},
booktitle = {Nordic Conference on Human-Computer Interaction},
keywords = {Formal-Journalism-Navigation},
publisher = {Association for Computing Machinery},
series = {NordiCHI 2024},
title = {Disclosing Diverse Perspectives of News Articles for Navigating between Online Journalism Content},
year = 2024
}%0 Conference Paper
%1 hci-lattice
%A Horn, Viktoria
%A Hirth, Johannes
%A Holfeld, Julian
%A Behmenburg, Jens Hendrik
%A Draude, Claude
%A Stumme, Gerd
%B Nordic Conference on Human-Computer Interaction
%C New York, NY, USA
%D 2024
%I Association for Computing Machinery
%R 10.1145/3679318.3685414
%T Disclosing Diverse Perspectives of News Articles for Navigating between Online Journalism Content
%U https://doi.org/10.1145/3679318.3685414
%X Today, exposure to journalistic online content is predominantly controlled by news recommender systems, which often suggest content that matches user’s interests or is selected according to non-transparent recommendation criteria. To circumvent resulting trade-offs like polarisation or fragmentation whilst ensuring user’s autonomy, we explore how different perspectives within online news can be disclosed instead for guiding navigation. To do so, we developed an interactive prototype that displays article titles in correspondence to their argumentative orientation. In order to investigate how the usage of our novel navigation structure impacts the choice of news articles and user experience, we conducted an exploratory user study assessing the impact of the design parameters chosen. Implications are drawn from the study results and the development of the interactive prototype for the exposure to diversity in the context of navigating news content online.
%@ 9798400709661 - 1.Hirth, J., Hanika, T.: The Geometric Structure of Topic Models, (2024). https://doi.org/10.48550/arxiv.2403.03607.
@misc{hirth2024geometric,
author = {Hirth, Johannes and Hanika, Tom},
keywords = {selected},
publisher = {arXiv},
title = {The Geometric Structure of Topic Models},
year = 2024
}%0 Generic
%1 hirth2024geometric
%A Hirth, Johannes
%A Hanika, Tom
%D 2024
%I arXiv
%R 10.48550/arxiv.2403.03607
%T The Geometric Structure of Topic Models - 1.Draude, C., Dürrschnabel, D., Hirth, J., Horn, V., Kropf, J., Lamla, J., Stumme, G., Uhlmann, M.: Conceptual Mapping of Controversies. In: Cabrera, I.P., Ferré, S., and Obiedkov, S. (eds.) Conceptual Knowledge Structures. pp. 201–216. Springer Nature Switzerland, Cham (2024).With our work, we contribute towards a qualitative analysis of the discourse on controversies in online news media. For this, we employ Formal Concept Analysis and the economics of conventions to derive conceptual controversy maps. In our experiments, we analyze two maps from different news journals with methods from ordinal data science. We show how these methods can be used to assess the diversity, complexity and potential bias of controversies. In addition to that, we discuss how the diagrams of concept lattices can be used to navigate between news articles.
@inproceedings{10.1007/978-3-031-67868-4_14,
abstract = {With our work, we contribute towards a qualitative analysis of the discourse on controversies in online news media. For this, we employ Formal Concept Analysis and the economics of conventions to derive conceptual controversy maps. In our experiments, we analyze two maps from different news journals with methods from ordinal data science. We show how these methods can be used to assess the diversity, complexity and potential bias of controversies. In addition to that, we discuss how the diagrams of concept lattices can be used to navigate between news articles.},
address = {Cham},
author = {Draude, Claude and Dürrschnabel, Dominik and Hirth, Johannes and Horn, Viktoria and Kropf, Jonathan and Lamla, J{ö}rn and Stumme, Gerd and Uhlmann, Markus},
booktitle = {Conceptual Knowledge Structures},
editor = {Cabrera, Inma P. and Ferré, Sébastien and Obiedkov, Sergei},
keywords = {itegpub},
pages = {201--216},
publisher = {Springer Nature Switzerland},
title = {Conceptual Mapping of Controversies},
year = 2024
}%0 Conference Paper
%1 10.1007/978-3-031-67868-4_14
%A Draude, Claude
%A Dürrschnabel, Dominik
%A Hirth, Johannes
%A Horn, Viktoria
%A Kropf, Jonathan
%A Lamla, J{ö}rn
%A Stumme, Gerd
%A Uhlmann, Markus
%B Conceptual Knowledge Structures
%C Cham
%D 2024
%E Cabrera, Inma P.
%E Ferré, Sébastien
%E Obiedkov, Sergei
%I Springer Nature Switzerland
%P 201--216
%T Conceptual Mapping of Controversies
%X With our work, we contribute towards a qualitative analysis of the discourse on controversies in online news media. For this, we employ Formal Concept Analysis and the economics of conventions to derive conceptual controversy maps. In our experiments, we analyze two maps from different news journals with methods from ordinal data science. We show how these methods can be used to assess the diversity, complexity and potential bias of controversies. In addition to that, we discuss how the diagrams of concept lattices can be used to navigate between news articles.
%@ 978-3-031-67868-4 - 1.Abdulla, M., Hirth, J., Stumme, G.: The Birkhoff Completion of Finite Lattices. In: Cabrera, I.P., Ferré, S., and Obiedkov, S. (eds.) Conceptual Knowledge Structures. pp. 20–35. Springer Nature Switzerland, Cham (2024).We introduce the Birkhoff completion as the smallest distributive lattice in which a given finite lattice can be embedded as semi-lattice. We discuss its relationship to implicational theories, in particular to R. Wille's simply-implicational theories. By an example, we show how the Birkhoff completion can be used as a tool for ordinal data science.
@inproceedings{10.1007/978-3-031-67868-4_2,
abstract = {We introduce the Birkhoff completion as the smallest distributive lattice in which a given finite lattice can be embedded as semi-lattice. We discuss its relationship to implicational theories, in particular to R. Wille's simply-implicational theories. By an example, we show how the Birkhoff completion can be used as a tool for ordinal data science.},
address = {Cham},
author = {Abdulla, Mohammad and Hirth, Johannes and Stumme, Gerd},
booktitle = {Conceptual Knowledge Structures},
editor = {Cabrera, Inma P. and Ferré, Sébastien and Obiedkov, Sergei},
keywords = {itegpub},
pages = {20--35},
publisher = {Springer Nature Switzerland},
title = {The Birkhoff Completion of Finite Lattices},
year = 2024
}%0 Conference Paper
%1 10.1007/978-3-031-67868-4_2
%A Abdulla, Mohammad
%A Hirth, Johannes
%A Stumme, Gerd
%B Conceptual Knowledge Structures
%C Cham
%D 2024
%E Cabrera, Inma P.
%E Ferré, Sébastien
%E Obiedkov, Sergei
%I Springer Nature Switzerland
%P 20--35
%T The Birkhoff Completion of Finite Lattices
%X We introduce the Birkhoff completion as the smallest distributive lattice in which a given finite lattice can be embedded as semi-lattice. We discuss its relationship to implicational theories, in particular to R. Wille's simply-implicational theories. By an example, we show how the Birkhoff completion can be used as a tool for ordinal data science.
%@ 978-3-031-67868-4
2023
- 1.Hanika, T., Hirth, J.: Conceptual Views on Tree Ensemble Classifiers. International Journal of Approximate Reasoning. 108930 (2023). https://doi.org/https://doi.org/10.1016/j.ijar.2023.108930.Random Forests and related tree-based methods are popular for supervised learning from table based data. Apart from their ease of parallelization, their classification performance is also superior. However, this performance, especially parallelizability, is offset by the loss of explainability. Statistical methods are often used to compensate for this disadvantage. Yet, their ability for local explanations, and in particular for global explanations, is limited. In the present work we propose an algebraic method, rooted in lattice theory, for the (global) explanation of tree ensembles. In detail, we introduce two novel conceptual views on tree ensemble classifiers and demonstrate their explanatory capabilities on Random Forests that were trained with standard parameters.
@article{https://doi.org/10.48550/arxiv.2302.05270,
abstract = {Random Forests and related tree-based methods are popular for supervised learning from table based data. Apart from their ease of parallelization, their classification performance is also superior. However, this performance, especially parallelizability, is offset by the loss of explainability. Statistical methods are often used to compensate for this disadvantage. Yet, their ability for local explanations, and in particular for global explanations, is limited. In the present work we propose an algebraic method, rooted in lattice theory, for the (global) explanation of tree ensembles. In detail, we introduce two novel conceptual views on tree ensemble classifiers and demonstrate their explanatory capabilities on Random Forests that were trained with standard parameters.},
author = {Hanika, Tom and Hirth, Johannes},
journal = {International Journal of Approximate Reasoning},
keywords = {itegpub},
pages = 108930,
title = {Conceptual Views on Tree Ensemble Classifiers},
year = 2023
}%0 Journal Article
%1 https://doi.org/10.48550/arxiv.2302.05270
%A Hanika, Tom
%A Hirth, Johannes
%D 2023
%J International Journal of Approximate Reasoning
%P 108930
%R https://doi.org/10.1016/j.ijar.2023.108930
%T Conceptual Views on Tree Ensemble Classifiers
%U https://www.sciencedirect.com/science/article/pii/S0888613X23000610
%X Random Forests and related tree-based methods are popular for supervised learning from table based data. Apart from their ease of parallelization, their classification performance is also superior. However, this performance, especially parallelizability, is offset by the loss of explainability. Statistical methods are often used to compensate for this disadvantage. Yet, their ability for local explanations, and in particular for global explanations, is limited. In the present work we propose an algebraic method, rooted in lattice theory, for the (global) explanation of tree ensembles. In detail, we introduce two novel conceptual views on tree ensemble classifiers and demonstrate their explanatory capabilities on Random Forests that were trained with standard parameters. - 1.Ganter, B., Hanika, T., Hirth, J.: Scaling Dimension. In: Dürrschnabel, D. and López-Rodríguez, D. (eds.) Formal Concept Analysis - 17th International Conference, ICFCA 2023, Kassel, Germany, July 17-21, 2023, Proceedings. pp. 64–77. Springer (2023). https://doi.org/10.1007/978-3-031-35949-1_5.
@inproceedings{DBLP:conf/icfca/GanterHH23,
author = {Ganter, Bernhard and Hanika, Tom and Hirth, Johannes},
booktitle = {Formal Concept Analysis - 17th International Conference, ICFCA 2023, Kassel, Germany, July 17-21, 2023, Proceedings},
editor = {Dürrschnabel, Dominik and López-Rodríguez, Domingo},
keywords = {itegpub},
pages = {64--77},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
title = {Scaling Dimension},
volume = 13934,
year = 2023
}%0 Conference Paper
%1 DBLP:conf/icfca/GanterHH23
%A Ganter, Bernhard
%A Hanika, Tom
%A Hirth, Johannes
%B Formal Concept Analysis - 17th International Conference, ICFCA 2023, Kassel, Germany, July 17-21, 2023, Proceedings
%D 2023
%E Dürrschnabel, Dominik
%E López-Rodríguez, Domingo
%I Springer
%P 64--77
%R 10.1007/978-3-031-35949-1_5
%T Scaling Dimension
%U https://doi.org/10.1007/978-3-031-35949-1_5
%V 13934 - 1.Hirth, J., Horn, V., Stumme, G., Hanika, T.: Automatic Textual Explanations of Concept Lattices. In: Ojeda-Aciego, M., Sauerwald, K., and Jäschke, R. (eds.) Graph-Based Representation and Reasoning. pp. 138–152. Springer Nature Switzerland, Cham (2023).Lattices and their order diagrams are an essential tool for communicating knowledge and insights about data. This is in particular true when applying Formal Concept Analysis. Such representations, however, are difficult to comprehend by untrained users and in general in cases where lattices are large. We tackle this problem by automatically generating textual explanations for lattices using standard scales. Our method is based on the general notion of ordinal motifs in lattices for the special case of standard scales. We show the computational complexity of identifying a small number of standard scales that cover most of the lattice structure. For these, we provide textual explanation templates, which can be applied to any occurrence of a scale in any data domain. These templates are derived using principles from human-computer interaction and allow for a comprehensive textual explanation of lattices. We demonstrate our approach on the spices planner data set, which is a medium sized formal context comprised of fifty-six meals (objects) and thirty-seven spices (attributes). The resulting 531 formal concepts can be covered by means of about 100 standard scales.
@inproceedings{10.1007/978-3-031-40960-8_12,
abstract = {Lattices and their order diagrams are an essential tool for communicating knowledge and insights about data. This is in particular true when applying Formal Concept Analysis. Such representations, however, are difficult to comprehend by untrained users and in general in cases where lattices are large. We tackle this problem by automatically generating textual explanations for lattices using standard scales. Our method is based on the general notion of ordinal motifs in lattices for the special case of standard scales. We show the computational complexity of identifying a small number of standard scales that cover most of the lattice structure. For these, we provide textual explanation templates, which can be applied to any occurrence of a scale in any data domain. These templates are derived using principles from human-computer interaction and allow for a comprehensive textual explanation of lattices. We demonstrate our approach on the spices planner data set, which is a medium sized formal context comprised of fifty-six meals (objects) and thirty-seven spices (attributes). The resulting 531 formal concepts can be covered by means of about 100 standard scales.},
address = {Cham},
author = {Hirth, Johannes and Horn, Viktoria and Stumme, Gerd and Hanika, Tom},
booktitle = {Graph-Based Representation and Reasoning},
editor = {Ojeda-Aciego, Manuel and Sauerwald, Kai and Jäschke, Robert},
keywords = {itegpub},
pages = {138--152},
publisher = {Springer Nature Switzerland},
title = {Automatic Textual Explanations of Concept Lattices},
year = 2023
}%0 Conference Paper
%1 10.1007/978-3-031-40960-8_12
%A Hirth, Johannes
%A Horn, Viktoria
%A Stumme, Gerd
%A Hanika, Tom
%B Graph-Based Representation and Reasoning
%C Cham
%D 2023
%E Ojeda-Aciego, Manuel
%E Sauerwald, Kai
%E Jäschke, Robert
%I Springer Nature Switzerland
%P 138--152
%T Automatic Textual Explanations of Concept Lattices
%X Lattices and their order diagrams are an essential tool for communicating knowledge and insights about data. This is in particular true when applying Formal Concept Analysis. Such representations, however, are difficult to comprehend by untrained users and in general in cases where lattices are large. We tackle this problem by automatically generating textual explanations for lattices using standard scales. Our method is based on the general notion of ordinal motifs in lattices for the special case of standard scales. We show the computational complexity of identifying a small number of standard scales that cover most of the lattice structure. For these, we provide textual explanation templates, which can be applied to any occurrence of a scale in any data domain. These templates are derived using principles from human-computer interaction and allow for a comprehensive textual explanation of lattices. We demonstrate our approach on the spices planner data set, which is a medium sized formal context comprised of fifty-six meals (objects) and thirty-seven spices (attributes). The resulting 531 formal concepts can be covered by means of about 100 standard scales.
%@ 978-3-031-40960-8 - 1.Hirth, J., Horn, V., Stumme, G., Hanika, T.: Ordinal Motifs in Lattices. Information Sciences. 120009 (2023). https://doi.org/https://doi.org/10.1016/j.ins.2023.120009.Lattices are a commonly used structure for the representation and analysis of relational and ontological knowledge. In particular, the analysis of these requires a decomposition of a large and high-dimensional lattice into a set of understandably large parts. With the present work we propose /ordinal motifs/ as analytical units of meaning. We study these ordinal substructures (or standard scales) through (full) scale-measures of formal contexts from the field of formal concept analysis. We show that the underlying decision problems are NP-complete and provide results on how one can incrementally identify ordinal motifs to save computational effort. Accompanying our theoretical results, we demonstrate how ordinal motifs can be leveraged to retrieve basic meaning from a medium sized ordinal data set.
@article{hirth2023ordinal,
abstract = {Lattices are a commonly used structure for the representation and analysis of relational and ontological knowledge. In particular, the analysis of these requires a decomposition of a large and high-dimensional lattice into a set of understandably large parts. With the present work we propose /ordinal motifs/ as analytical units of meaning. We study these ordinal substructures (or standard scales) through (full) scale-measures of formal contexts from the field of formal concept analysis. We show that the underlying decision problems are NP-complete and provide results on how one can incrementally identify ordinal motifs to save computational effort. Accompanying our theoretical results, we demonstrate how ordinal motifs can be leveraged to retrieve basic meaning from a medium sized ordinal data set.},
author = {Hirth, Johannes and Horn, Viktoria and Stumme, Gerd and Hanika, Tom},
journal = {Information Sciences},
keywords = {selected},
pages = 120009,
title = {Ordinal Motifs in Lattices},
year = 2023
}%0 Journal Article
%1 hirth2023ordinal
%A Hirth, Johannes
%A Horn, Viktoria
%A Stumme, Gerd
%A Hanika, Tom
%D 2023
%J Information Sciences
%P 120009
%R https://doi.org/10.1016/j.ins.2023.120009
%T Ordinal Motifs in Lattices
%U https://www.sciencedirect.com/science/article/pii/S0020025523015943
%X Lattices are a commonly used structure for the representation and analysis of relational and ontological knowledge. In particular, the analysis of these requires a decomposition of a large and high-dimensional lattice into a set of understandably large parts. With the present work we propose /ordinal motifs/ as analytical units of meaning. We study these ordinal substructures (or standard scales) through (full) scale-measures of formal contexts from the field of formal concept analysis. We show that the underlying decision problems are NP-complete and provide results on how one can incrementally identify ordinal motifs to save computational effort. Accompanying our theoretical results, we demonstrate how ordinal motifs can be leveraged to retrieve basic meaning from a medium sized ordinal data set.
2022
- 1.Schäfermeier, B., Hirth, J., Hanika, T.: Research Topic Flows in Co-Authorship Networks. Scientometrics. (2022). https://doi.org/10.1007/s11192-022-04529-w.In scientometrics, scientific collaboration is often analyzed by means of co-authorships. An aspect which is often overlooked and more difficult to quantify is the flow of expertise between authors from different research topics, which is an important part of scientific progress. With the Topic Flow Network (TFN) we propose a graph structure for the analysis of research topic flows between scientific authors and their respective research fields. Based on a multi-graph and a topic model, our proposed network structure accounts for intratopic as well as intertopic flows. Our method requires for the construction of a TFN solely a corpus of publications (i.e., author and abstract information). From this, research topics are discovered automatically through non-negative matrix factorization. The thereof derived TFN allows for the application of social network analysis techniques, such as common metrics and community detection. Most importantly, it allows for the analysis of intertopic flows on a large, macroscopic scale, i.e., between research topic, as well as on a microscopic scale, i.e., between certain sets of authors. We demonstrate the utility of TFNs by applying our method to two comprehensive corpora of altogether 20 Mio. publications spanning more than 60 years of research in the fields computer science and mathematics. Our results give evidence that TFNs are suitable, e.g., for the analysis of topical communities, the discovery of important authors in different fields, and, most notably, the analysis of intertopic flows, i.e., the transfer of topical expertise. Besides that, our method opens new directions for future research, such as the investigation of influence relationships between research fields.
@article{schafermeier2022research,
abstract = {In scientometrics, scientific collaboration is often analyzed by means of co-authorships. An aspect which is often overlooked and more difficult to quantify is the flow of expertise between authors from different research topics, which is an important part of scientific progress. With the Topic Flow Network (TFN) we propose a graph structure for the analysis of research topic flows between scientific authors and their respective research fields. Based on a multi-graph and a topic model, our proposed network structure accounts for intratopic as well as intertopic flows. Our method requires for the construction of a TFN solely a corpus of publications (i.e., author and abstract information). From this, research topics are discovered automatically through non-negative matrix factorization. The thereof derived TFN allows for the application of social network analysis techniques, such as common metrics and community detection. Most importantly, it allows for the analysis of intertopic flows on a large, macroscopic scale, i.e., between research topic, as well as on a microscopic scale, i.e., between certain sets of authors. We demonstrate the utility of TFNs by applying our method to two comprehensive corpora of altogether 20 Mio. publications spanning more than 60 years of research in the fields computer science and mathematics. Our results give evidence that TFNs are suitable, e.g., for the analysis of topical communities, the discovery of important authors in different fields, and, most notably, the analysis of intertopic flows, i.e., the transfer of topical expertise. Besides that, our method opens new directions for future research, such as the investigation of influence relationships between research fields.},
author = {Schäfermeier, Bastian and Hirth, Johannes and Hanika, Tom},
journal = {Scientometrics},
keywords = {selected},
month = 10,
title = {Research Topic Flows in Co-Authorship Networks},
year = 2022
}%0 Journal Article
%1 schafermeier2022research
%A Schäfermeier, Bastian
%A Hirth, Johannes
%A Hanika, Tom
%D 2022
%J Scientometrics
%R 10.1007/s11192-022-04529-w
%T Research Topic Flows in Co-Authorship Networks
%X In scientometrics, scientific collaboration is often analyzed by means of co-authorships. An aspect which is often overlooked and more difficult to quantify is the flow of expertise between authors from different research topics, which is an important part of scientific progress. With the Topic Flow Network (TFN) we propose a graph structure for the analysis of research topic flows between scientific authors and their respective research fields. Based on a multi-graph and a topic model, our proposed network structure accounts for intratopic as well as intertopic flows. Our method requires for the construction of a TFN solely a corpus of publications (i.e., author and abstract information). From this, research topics are discovered automatically through non-negative matrix factorization. The thereof derived TFN allows for the application of social network analysis techniques, such as common metrics and community detection. Most importantly, it allows for the analysis of intertopic flows on a large, macroscopic scale, i.e., between research topic, as well as on a microscopic scale, i.e., between certain sets of authors. We demonstrate the utility of TFNs by applying our method to two comprehensive corpora of altogether 20 Mio. publications spanning more than 60 years of research in the fields computer science and mathematics. Our results give evidence that TFNs are suitable, e.g., for the analysis of topical communities, the discovery of important authors in different fields, and, most notably, the analysis of intertopic flows, i.e., the transfer of topical expertise. Besides that, our method opens new directions for future research, such as the investigation of influence relationships between research fields. - 1.Hanika, T., Hirth, J.: Knowledge cores in large formal contexts. Annals of Mathematics and Artificial Intelligence. (2022). https://doi.org/10.1007/s10472-022-09790-6.Knowledge computation tasks, such as computing a base of valid implications, are often infeasible for large data sets. This is in particular true when deriving canonical bases in formal concept analysis (FCA). Therefore, it is necessary to find techniques that on the one hand reduce the data set size, but on the other hand preserve enough structure to extract useful knowledge. Many successful methods are based on random processes to reduce the size of the investigated data set. This, however, makes them hardly interpretable with respect to the discovered knowledge. Other approaches restrict themselves to highly supported subsets and omit rare and (maybe) interesting patterns. An essentially different approach is used in network science, called k-cores. These cores are able to reflect rare patterns, as long as they are well connected within the data set. In this work, we study k-cores in the realm of FCA by exploiting the natural correspondence of bi-partite graphs and formal contexts. This structurally motivated approach leads to a comprehensible extraction of knowledge cores from large formal contexts.
@article{Hanika2022,
abstract = {Knowledge computation tasks, such as computing a base of valid implications, are often infeasible for large data sets. This is in particular true when deriving canonical bases in formal concept analysis (FCA). Therefore, it is necessary to find techniques that on the one hand reduce the data set size, but on the other hand preserve enough structure to extract useful knowledge. Many successful methods are based on random processes to reduce the size of the investigated data set. This, however, makes them hardly interpretable with respect to the discovered knowledge. Other approaches restrict themselves to highly supported subsets and omit rare and (maybe) interesting patterns. An essentially different approach is used in network science, called k-cores. These cores are able to reflect rare patterns, as long as they are well connected within the data set. In this work, we study k-cores in the realm of FCA by exploiting the natural correspondence of bi-partite graphs and formal contexts. This structurally motivated approach leads to a comprehensible extraction of knowledge cores from large formal contexts.},
author = {Hanika, Tom and Hirth, Johannes},
journal = {Annals of Mathematics and Artificial Intelligence},
keywords = {itegpub},
month = {04},
title = {Knowledge cores in large formal contexts},
year = 2022
}%0 Journal Article
%1 Hanika2022
%A Hanika, Tom
%A Hirth, Johannes
%D 2022
%J Annals of Mathematics and Artificial Intelligence
%R 10.1007/s10472-022-09790-6
%T Knowledge cores in large formal contexts
%U https://doi.org/10.1007/s10472-022-09790-6
%X Knowledge computation tasks, such as computing a base of valid implications, are often infeasible for large data sets. This is in particular true when deriving canonical bases in formal concept analysis (FCA). Therefore, it is necessary to find techniques that on the one hand reduce the data set size, but on the other hand preserve enough structure to extract useful knowledge. Many successful methods are based on random processes to reduce the size of the investigated data set. This, however, makes them hardly interpretable with respect to the discovered knowledge. Other approaches restrict themselves to highly supported subsets and omit rare and (maybe) interesting patterns. An essentially different approach is used in network science, called k-cores. These cores are able to reflect rare patterns, as long as they are well connected within the data set. In this work, we study k-cores in the realm of FCA by exploiting the natural correspondence of bi-partite graphs and formal contexts. This structurally motivated approach leads to a comprehensible extraction of knowledge cores from large formal contexts. - 1.Hirth, J., Hanika, T.: Formal Conceptual Views in Neural Networks, https://arxiv.org/abs/2209.13517, (2022). https://doi.org/10.48550/ARXIV.2209.13517.
@misc{https://doi.org/10.48550/arxiv.2209.13517,
author = {Hirth, Johannes and Hanika, Tom},
keywords = {sai},
publisher = {arXiv},
title = {Formal Conceptual Views in Neural Networks},
year = 2022
}%0 Generic
%1 https://doi.org/10.48550/arxiv.2209.13517
%A Hirth, Johannes
%A Hanika, Tom
%D 2022
%I arXiv
%R 10.48550/ARXIV.2209.13517
%T Formal Conceptual Views in Neural Networks
%U https://arxiv.org/abs/2209.13517 - 1.Hanika, T., Hirth, J.: On the lattice of conceptual measurements. Information Sciences. 613, 453–468 (2022). https://doi.org/https://doi.org/10.1016/j.ins.2022.09.005.We present a novel approach for data set scaling based on scale-measures from formal concept analysis, i.e., continuous maps between closure systems, for which we derive a canonical representation. Moreover, we prove that scale-measures can be lattice ordered using the canonical representation. This enables exploring the set of scale-measures by the use of meet and join operations. Furthermore we show that the lattice of scale-measures is isomorphic to the lattice of sub-closure systems that arises from the original data. Finally, we provide another representation of scale-measures using propositional logic in terms of data set features. Our theoretical findings are discussed by means of examples.
@article{hanika2020lattice,
abstract = {We present a novel approach for data set scaling based on scale-measures from formal concept analysis, i.e., continuous maps between closure systems, for which we derive a canonical representation. Moreover, we prove that scale-measures can be lattice ordered using the canonical representation. This enables exploring the set of scale-measures by the use of meet and join operations. Furthermore we show that the lattice of scale-measures is isomorphic to the lattice of sub-closure systems that arises from the original data. Finally, we provide another representation of scale-measures using propositional logic in terms of data set features. Our theoretical findings are discussed by means of examples.},
author = {Hanika, Tom and Hirth, Johannes},
journal = {Information Sciences},
keywords = {selected},
pages = {453-468},
title = {On the lattice of conceptual measurements},
volume = 613,
year = 2022
}%0 Journal Article
%1 hanika2020lattice
%A Hanika, Tom
%A Hirth, Johannes
%D 2022
%J Information Sciences
%P 453-468
%R https://doi.org/10.1016/j.ins.2022.09.005
%T On the lattice of conceptual measurements
%U https://www.sciencedirect.com/science/article/pii/S0020025522010489
%V 613
%X We present a novel approach for data set scaling based on scale-measures from formal concept analysis, i.e., continuous maps between closure systems, for which we derive a canonical representation. Moreover, we prove that scale-measures can be lattice ordered using the canonical representation. This enables exploring the set of scale-measures by the use of meet and join operations. Furthermore we show that the lattice of scale-measures is isomorphic to the lattice of sub-closure systems that arises from the original data. Finally, we provide another representation of scale-measures using propositional logic in terms of data set features. Our theoretical findings are discussed by means of examples.
2021
- 1.Hanika, T., Hirth, J.: Exploring Scale-Measures of Data Sets. In: Braud, A., Buzmakov, A., Hanika, T., and Ber, F.L. (eds.) Formal Concept Analysis - 16th International Conference, ICFCA 2021, Strasbourg, France, June 29 - July 2, 2021, Proceedings. pp. 261–269. Springer (2021). https://doi.org/10.1007/978-3-030-77867-5_17.
@inproceedings{DBLP:conf/icfca/HanikaH21,
author = {Hanika, Tom and Hirth, Johannes},
booktitle = {Formal Concept Analysis - 16th International Conference, ICFCA 2021, Strasbourg, France, June 29 - July 2, 2021, Proceedings},
editor = {Braud, Agn{{è}}s and Buzmakov, Aleksey and Hanika, Tom and Ber, Florence Le},
keywords = {closure},
pages = {261--269},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
title = {Exploring Scale-Measures of Data Sets},
volume = 12733,
year = 2021
}%0 Conference Paper
%1 DBLP:conf/icfca/HanikaH21
%A Hanika, Tom
%A Hirth, Johannes
%B Formal Concept Analysis - 16th International Conference, ICFCA 2021, Strasbourg, France, June 29 - July 2, 2021, Proceedings
%D 2021
%E Braud, Agn{{è}}s
%E Buzmakov, Aleksey
%E Hanika, Tom
%E Ber, Florence Le
%I Springer
%P 261--269
%R 10.1007/978-3-030-77867-5_17
%T Exploring Scale-Measures of Data Sets
%U https://doi.org/10.1007/978-3-030-77867-5_17
%V 12733 - 1.Hanika, T., Hirth, J.: Quantifying the Conceptual Error in Dimensionality Reduction. In: Braun, T., Gehrke, M., Hanika, T., and Hernandez, N. (eds.) Graph-Based Representation and Reasoning - 26th International Conference on Conceptual Structures, ICCS 2021, Virtual Event, September 20-22, 2021, Proceedings. pp. 105–118. Springer (2021). https://doi.org/10.1007/978-3-030-86982-3_8.
@inproceedings{DBLP:conf/iccs/HanikaH21,
author = {Hanika, Tom and Hirth, Johannes},
booktitle = {Graph-Based Representation and Reasoning - 26th International Conference on Conceptual Structures, ICCS 2021, Virtual Event, September 20-22, 2021, Proceedings},
editor = {Braun, Tanya and Gehrke, Marcel and Hanika, Tom and Hernandez, Nathalie},
keywords = {selected},
pages = {105--118},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
title = {Quantifying the Conceptual Error in Dimensionality Reduction},
volume = 12879,
year = 2021
}%0 Conference Paper
%1 DBLP:conf/iccs/HanikaH21
%A Hanika, Tom
%A Hirth, Johannes
%B Graph-Based Representation and Reasoning - 26th International Conference on Conceptual Structures, ICCS 2021, Virtual Event, September 20-22, 2021, Proceedings
%D 2021
%E Braun, Tanya
%E Gehrke, Marcel
%E Hanika, Tom
%E Hernandez, Nathalie
%I Springer
%P 105--118
%R 10.1007/978-3-030-86982-3_8
%T Quantifying the Conceptual Error in Dimensionality Reduction
%U https://doi.org/10.1007/978-3-030-86982-3_8
%V 12879
2019
- 1.Hanika, T., Hirth, J.: Conexp-Clj - A Research Tool for FCA. In: Cristea, D., Ber, F.L., Missaoui, R., Kwuida, L., and Sertkaya, B. (eds.) ICFCA (Supplements). pp. 70–75. CEUR-WS.org (2019).
@inproceedings{conf/icfca/HanikaH19,
author = {Hanika, Tom and Hirth, Johannes},
booktitle = {ICFCA (Supplements)},
crossref = {conf/icfca/2019suppl},
editor = {Cristea, Diana and Ber, Florence Le and Missaoui, Rokia and Kwuida, Léonard and Sertkaya, Baris},
keywords = {itegpub},
pages = {70-75},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
title = {Conexp-Clj - A Research Tool for FCA.},
volume = 2378,
year = 2019
}%0 Conference Paper
%1 conf/icfca/HanikaH19
%A Hanika, Tom
%A Hirth, Johannes
%B ICFCA (Supplements)
%D 2019
%E Cristea, Diana
%E Ber, Florence Le
%E Missaoui, Rokia
%E Kwuida, Léonard
%E Sertkaya, Baris
%I CEUR-WS.org
%P 70-75
%T Conexp-Clj - A Research Tool for FCA.
%U http://dblp.uni-trier.de/db/conf/icfca/icfca2019suppl.html#HanikaH19
%V 2378