[
    {
        "id": "9345",
        "title": "Learning Accurate and Interpretable Decision Trees (Extended Abstract)",
        "authors": "Maria-Florina Balcan, Dravyansh Sharma",
        "abstract": "Decision trees are a popular tool in machine learning and yield easy-to-understand models. Several techniques have been proposed in the literature for learning a decision tree classifier, with different techniques working well for data from different domains. In this work, we develop a data-driven approach to design decision tree learning algorithms given repeated access to data from the same domain. We study multiple formulations covering different aspects and popular techniques for learning decision trees. We propose novel parameterized classes of node splitting criteria in top-down algorithms, which interpolate between popularly used entropy and Gini impurity based criteria, and provide theoretical bounds on the number of samples needed to learn the splitting function appropriate for the data at hand. We also study the sample complexity of tuning prior parameters in Bayesian decision tree learning, and extend our results to decision tree regression. We further consider the problem of tuning hyperparameters in pruning the decision tree for classical pruning algorithms including min-cost complexity pruning. We also study the interpretability of the learned decision trees and introduce a data-driven approach for optimizing the explainability versus accuracy trade-off using decision trees. Finally, we demonstrate the significance of our approach on real world datasets by learning data-specific decision trees which are simultaneously more accurate and interpretable.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "11:30",
        "session": "Humans and AI: Interpretable Models",
        "poster_positions": "From board n102 to board n106"
    },
    {
        "id": "9357",
        "title": "How to Teach Programming in the AI Era? Using LLMs as a Teachable Agent for Debugging (Extended Abstract)",
        "authors": "Qianou Ma, Hua Shen, Ken Koedinger, Tongshuang Wu",
        "abstract": "Large Language Models (LLMs) excel at generating content at impeccable speeds. However, they are imperfect and still make various mistakes. In Computer Science education, as LLMs are widely recognized as \"AI pair programmers,\" it becomes increasingly important to train students on evaluating and debugging LLM-generated codes. In this work, we introduce HypoCompass, a novel system to facilitate deliberate practice on debugging, where human novices play the role of Teaching Assistants and help LLM-powered teachable agents debug code.\r\nWe enable effective task delegation between students and LLMs in this learning-by-teaching environment: students focus on hypothesizing the cause of code errors, while adjacent skills like code completion are offloaded to LLM-agents. Our evaluations demonstrate that HypoCompass generates high-quality training materials (e.g., bugs and fixes), outperforming human counterparts fourfold in efficiency, and significantly improves student performance on debugging by 12% in the pre-to-post test.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "10:00",
        "session": "Humans and AI",
        "poster_positions": "From board n97 to board n101"
    },
    {
        "id": "9360",
        "title": "Data Void Exploits: Tracking & Mitigation Strategies (Extended Abstract)",
        "authors": "Miro Mannino, Junior Garcia, Reem Hazim, Azza Abouzied, Paolo Papotti",
        "abstract": "In the evolving landscape of online information, disinformation is a growing concern. A concept central to this challenge is the \"data void\", a situation where there is a lack of relevant information online regarding certain search terms. This creates an opportunity for misleading or false narratives to fill the gap, often influencing public perception. In this work, we present methods to track and mitigate data voids in Web search settings.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "11:30",
        "session": "Knowledge Representation and Reasoning (3\/4)",
        "poster_positions": "From board n78 to board n82"
    },
    {
        "id": "9344",
        "title": "Incentives for Early Arrival in Cooperative Games (Extended Abstract)",
        "authors": "Yaoxin Ge, Yao Zhang, Dengji Zhao, Zhihao Gavin Tang, Hu Fu, Pinyan Lu",
        "abstract": "We study cooperative games where players join sequentially, and the value generated by those who have joined at any point must be irrevocably divided among these players. We introduce two desiderata for the value division mechanism: that the players should have incentives to join as early as possible, and that the division should be considered fair. For the latter, we require that each player's expected share in the mechanism should equal her Shapley value if the players' arrival order is uniformly at random. \r\n\r\nWhen the value generation function is submodular, allocating the marginal value to the player satisfies these properties. This is no longer true for more general functions. Our main technical contribution is a complete characterization of 0-1 value games for which desired mechanisms exist. We show that a natural mechanism, Rewarding First Critical Player (RFC), is complete, in that a 0-1 value function admits a mechanism with the properties above if and only if RFC satisfies them; we analytically characterize all such value functions. Moreover, we give an algorithm that decomposes, in an online fashion, any value function into 0-1 value functions, on each of which RFC can be run. In this way, we design an extension of RFC for general monotone games, and the properties are proved to be maintained.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "Game Theory and Economic Paradigms"
    },
    {
        "id": "9347",
        "title": "Efficient Rectification of Neuro-Symbolic Reasoning Inconsistencies by Abductive Reflection (Extended Abstract)",
        "authors": "Wen-Chao Hu, Wang-Zhou Dai, Yuan Jiang, Zhi-Hua Zhou",
        "abstract": "Neuro-Symbolic (NeSy) AI could be regarded as an analogy to human dual-process cognition, modeling the intuitive System 1 with neural networks and the algorithmic System 2 with symbolic reasoning. However, for complex learning targets, NeSy systems often generate outputs inconsistent with domain knowledge. Inspired by the human Cognitive Reflection, which promptly detects errors in our intuitive response and revises them by invoking the System 2 reasoning, we propose to improve NeSy systems by introducing Abductive Reflection (ABL-Refl) based on the Abductive Learning (ABL) framework. ABL-Refl leverages domain knowledge to abduce a reflection vector during training, which can then flag potential errors in the neural network outputs and invoke abduction to rectify them and generate consistent outputs during inference. Experiments show that ABL-Refl outperforms state-of-the-art NeSy methods, achieving excellent accuracy with fewer training resources and enhanced efficiency.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "14:45",
        "session": "Machine Learning (3\/3)"
    },
    {
        "id": "9349",
        "title": "Domain Prompt Learning with Quaternion Networks (Extended Abstract)",
        "authors": "Qinglong Cao, Zhengqin Xu, Yuntian Chen, Chao Ma, Xiaokang Yang",
        "abstract": "Foundational vision-language models (VLMs) like CLIP have revolutionized image recognition, but adapting them to specialized domains with limited data remains challenging. We propose Domain Prompt Learning with Quaternion Networks (DPLQ), which leverages domain-specific foundation models and quaternion-based prompt tuning to effectively transfer recognition capabilities. Our method achieves state-of-the-art results in remote sensing and medical imaging tasks. This extended abstract highlights the key contributions and performance of DPLQ.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "Multidisciplinary Topics and Applications"
    },
    {
        "id": "9353",
        "title": "Scalable Graph Classification via Random Walk Fingerprints (Extended Abstract)",
        "authors": "Peiyan Li, Honglian Wang, Christian Böhm",
        "abstract": "We design a lightweight structural feature extraction technique for graph classification. It leverages node subsets and connection strength reflected by random-walk-based heuristics, presenting a scalable, unsupervised, and easily interpretable alternative. We provide theoretical insights into our technical design and establish a relation between the extracted structural features and the graph spectrum. We show our method achieves high levels of computational efficiency while maintaining robust classification accuracy.",
        "location": "Guangzhou",
        "day": "August 29th",
        "hour": "14:30",
        "session": "DM: Mining graphs (1\/3)"
    },
    {
        "id": "9342",
        "title": "Never Train from Scratch: Fair Comparison of Long-Sequence Models Requires Data-Driven Priors (Extended Abstract)",
        "authors": "Ido Amos, Jonathan Berant, Ankit Gupta",
        "abstract": "This paper is an extended abstract of our ICLR 2024 Outstanding Paper Award work. Modeling long-range dependencies across sequences is a longstanding goal in machine learning. While state space models reportedly outperform Transformers on benchmarks like Long Range Arena, we show that random initialization significantly overestimates architectural differences. Pretraining with standard denoising objectives on downstream task data leads to dramatic gains across architectures and minimal performance gaps between Transformers and state space models (SSMs). We demonstrate that properly pretrained vanilla Transformers match S4 performance on Long Range Arena and improve SSM results on PathX-256 by 20 absolute points. Our analysis shows previously-proposed structured parameterizations for SSMs become largely redundant with pretraining. When evaluating architectures on supervised tasks, incorporating data-driven priors via pretraining is essential for reliable performance estimation.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "15:00",
        "session": "ML: Large Language Models",
        "poster_positions": "From board n15 to board n21"
    },
    {
        "id": "9354",
        "title": "FairCognizer: A Model for Accurate Predictions with Inherent Fairness Evaluation (Extended Abstract)",
        "authors": "Adda Akram Bendoukha, Nesrine Kaaniche, Aymen Boudguiga, Renaud Sirdey",
        "abstract": "Algorithmic fairness is a critical challenge in building trustworthy Machine Learning (ML)  models. ML classifiers strive to make predictions that closely match real-world observations (ground truth). However, if the ground truth data itself reflects biases against certain sub-populations, a dilemma arises: prioritize fairness and potentially reduce accuracy, or emphasize accuracy at the expense of fairness.\r\nThis work proposes a novel training framework that goes beyond achieving high accuracy. Our framework trains a classifier to not only deliver optimal predictions but also to identify potential fairness risks associated with each prediction. \r\nTo do so, we specify a dual-labeling strategy where the second label contains a per-prediction fairness evaluation, referred to as an unfairness risk evaluation. In addition, we identify a subset of samples as highly vulnerable to group-unfair classifiers.\r\nOur experiments demonstrate that our classifiers attain optimal accuracy levels on both the Adult-Census-Income and Compas-Recidivism datasets. Moreover, they identify unfair predictions with nearly 75% accuracy at the cost of expanding the size of the classifier by 45%.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "11:30",
        "session": "AI Ethics, Trust, Fairness (3\/3)",
        "poster_positions": "From board n39 to board n41"
    },
    {
        "id": "9343",
        "title": "Mechanism Design for Large Language Models (Extended Abstract)",
        "authors": "Paul Dütting, Vahab Mirrokni, Renato Paes Leme, Haifeng Xu, Song Zuo",
        "abstract": "We investigate auction mechanisms for AI-generated content, focusing on applications like ad creative generation. In our model, agents' preferences over stochastically generated content are encoded as large language models (LLMs). \r\n\r\nWe propose an auction format that operates on a token-by-token basis, and allows LLM agents to influence content creation through single dimensional bids. We formulate two desirable incentive properties and prove their equivalence to a monotonicity condition on output aggregation. This equivalence enables a second-price rule design, even absent explicit agent valuation functions. Our design is supported by demonstrations on a publicly available LLM.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "Game Theory and Economic Paradigms"
    },
    {
        "id": "9351",
        "title": "Shapley Value Computation in Ontology-Mediated Query Answering (Extended Abstract)",
        "authors": "Meghyn Bienvenu, Diego Figueira, Pierre Lafourcade",
        "abstract": "In this work, we explore the use of the Shapley value in ontology-mediated query answering (OMQA) and provide a detailed complexity analysis of Shapley value computation (SVC) in the OMQA setting. In particular, we establish a FP\/#P-hard dichotomy for SVC for ontology-mediated queries (T,q) composed of an ontology T formulated in the description logic ELHI-bot and a connected constant-free homomorphism-closed query q. We further strengthen the #P-hardness side of the dichotomy to cover possibly disconnected queries with constants. Our results exploit recently discovered connections between SVC and probabilistic query evaluation and allow us to generalize existing results on probabilistic OMQA.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "10:00",
        "session": "KR: ontologies",
        "poster_positions": "From board n20 to board n23"
    },
    {
        "id": "9361",
        "title": "A Relaxed Symmetric Non-negative Matrix Factorization Approach for Community Discovery (Extended Abstract)",
        "authors": "Zhigang Liu, Hao Yan, Yurong Zhong, Weiling Li",
        "abstract": "Community discovery is a prominent issue in com-plex network analysis. Symmetric non-negative matrix factorization (SNMF) is frequently adopted to tackle this issue. The use of a single feature matrix can depict network symmetry, but it limits its ability to learn node representations. To break this limitation, we present a novel Relaxed Symmetric NMF (RSN) approach to boost an SNMF-based community detector. It works by 1) expanding the representational space and its degrees of freedom with multiple feature factors; 2) integrating the well-designed equality-constraints to make the model well-aware of the network’s intrinsic symmetry; 3) employing graph regularization to pre-serve the local geometric invariance of the network structure; and 4) separating constraints from decision variables for efficient optimization via the principle of alternating-direction-method of multi-pliers. RSN’s effectiveness is verified through empirical studies on six real social networks, show-casing superior precision in community discovery over existing models and baselines.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "09:40",
        "session": "Knowledge Representation and Reasoning"
    },
    {
        "id": "9348",
        "title": "CAM-Based Methods Can See through Walls (Extended Abstract)",
        "authors": "Magamed Taimeskhanov, Ronan Sicre, Damien Garreau",
        "abstract": "CAM-based methods are widely-used post-hoc interpretability methods that produce a saliency map to explain the decision of an image classification model. The saliency map highlights the important areas of the image relevant to the prediction. In this paper, we show that most of these methods can incorrectly attribute an important score to parts of the image that the model cannot see. We show that this phenomenon occurs both theoretically and experimentally. On the theory side, we analyze the behavior of GradCAM on a simple masked CNN model at initialization. Experimentally, we train a VGG-like model constrained to not use the lower part of the image and  nevertheless observe positive scores in the unseen part of the image. This behavior is evaluated quantitatively on two new datasets. We believe that this is problematic, potentially leading to mis-interpretation of the model's behavior.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "11:30",
        "session": "Humans and AI: Interpretable Models",
        "poster_positions": "From board n102 to board n106"
    },
    {
        "id": "9359",
        "title": "SEE: Spherical Embedding Expansion for Improving Deep Metric Learning (Extended Abstract)",
        "authors": "Binh M. Le, Simon S. Woo",
        "abstract": "The primary goal of deep metric learning is to construct a comprehensive embedding space that can effectively represent samples originating from both intra- and inter-classes. Although extensive prior work has explored diverse metric functions and innovative training strategies, much of this work relies on default training data. Consequently, the potential variations inherent within this data remain largely unexplored, constraining the model's robustness to unseen images. In this context, we introduce the  Spherical Embedding Expansion (SEE ) method. SEE aims to uncover the latent semantic variations in training data. Especially, our method augments the embedding space with synthetic representations based on Max-Mahalanobis distribution (MMD) centers, which maximize the dispersion of these synthetic features without increasing computational costs.We evaluated the efficacy of SEE on four renowned standard benchmarks for the image retrieval task. The results demonstrate that SEE consistently enhances the performance of conventional methods when integrated with them, setting a new benchmark for deep metric learning performance across all settings.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "CV: Learning"
    },
    {
        "id": "9355",
        "title": "Meaning Holism and Indeterminacy of Reference in Ontologies (Extended Abstract)",
        "authors": "Adrien Barton, Paul Fabry, Jean-François Ethier",
        "abstract": "According to meaning holism, the meanings of all the words in a language are interdependent. If this was true, then the very practice of building largely interconnected set of ontologies would be threatened. We examine here the extent of the severity of meaning holism for ontology engineering, based on a definition of the meaning of a class term in an ontology, with regard to the classical analytic\/synthetic distinction. We show that meaning holism is not as pervasive in ontologies as traditionally assumed in philosophy of language when interpreting the meaning of a class term as a collection of statements expressing necessary conditions on this term. Still, meaning holism presents substantial challenges for ontology engineering and requires mitigation strategies. We also investigate the related phenomenon of indeterminacy of reference and show how anchoring formal ontologies in natural language can mitigate this problem, even if not fully control it.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "10:00",
        "session": "KR: ontologies",
        "poster_positions": "From board n20 to board n23"
    },
    {
        "id": "9358",
        "title": "Interpreting Pretrained Language Models via Concept Bottlenecks (Extended Abstract)",
        "authors": "Zhen Tan, Lu Cheng, Song Wang, Yuan Bo, Jundong Li, Huan Liu",
        "abstract": "Pretrained language models (PLMs) achieve state-of-the-art results but often function as ``black boxes'', hindering interpretability and responsible deployment. While methods like attention analysis exist, they often lack clarity and intuitiveness. We propose interpreting PLMs through high-level, human-understandable concepts using Concept Bottleneck Models (CBMs). This extended abstract introduces C3M (ChatGPT-guided Concept augmentation with Concept-level Mixup), a novel framework for training Concept-Bottleneck-Enabled PLMs (CBE-PLMs). C3M leverages Large Language Models (LLMs) like ChatGPT to augment concept sets and generate noisy concept labels, combined with a concept-level MixUp mechanism to enhance robustness and effectively learn from both human-annotated and machine-generated concepts. Empirical results show our approach provides intuitive explanations, aids model diagnosis via test-time intervention, and improves the interpretability-utility trade-off, even with limited or noisy concept annotations. This is an concise version of [Tan et al., 2024b], recipient of the Best Paper Award at PAKDD 2024. Code and data are released at https:\/\/github.com\/Zhen-Tan-dmml\/CBM_NLP.git.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "11:30",
        "session": "Humans and AI: Interpretable Models",
        "poster_positions": "From board n102 to board n106"
    },
    {
        "id": "9350",
        "title": "Contractions Based on Optimal Repairs (Extended Abstract)",
        "authors": "Franz Baader, Renata Wassermann",
        "abstract": "Removing unwanted consequences from a knowledge base has been investigated in belief change under the name contraction and is called repair in ontology engineering. Simple repair and contraction approaches based on removing statements from the knowledge base (respectively called belief base contractions and classical repairs) have the disadvantage that they are syntax-dependent and may remove more consequences than necessary. Belief set contractions do not have these problems, but may result in belief sets that have no finite representation. Similarly, optimal repairs, which are syntax-independent and maximize the retained consequences, may not exist. Our KR 2024 paper leverage advances in characterizing and computing optimal repairs of ontologies based on the description logics EL to obtain contraction operations that combine the advantages of belief set and belief base contractions. It introduces this new approach in a very general setting, and proves a characterization theorem that relates the obtained contractions with well-known rationality postulates. Then, it describes a variety of interesting instances, not only in the standard repair\/contraction setting where one wants to get rid of a consequence, but also in other settings such as variants of forgetting in propositional and description logic.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "11:30",
        "session": "Knowledge Representation and Reasoning (1\/4)",
        "poster_positions": "From board n25 to board n29"
    },
    {
        "id": "9356",
        "title": "Explanatory Capabilities of Large Language Models in Prescriptive Process Monitoring (Extended Abstract)",
        "authors": "Kateryna Kubrak, Lana Botchorishvili, Fredrik Milani, Alexander Nolte, Marlon Dumas",
        "abstract": "Prescriptive Process Monitoring (PrPM) systems recommend interventions in ongoing business process cases to improve performance. However, performance gains only materialize if users follow the recommendations. Prior research has shown that users are more likely to follow recommendations when they understand them. In this paper, we explore the use of Large Language Models (LLMs) to generate explanations for PrPM recommendations. We developed a prompting method based on typical user questions and integrated it into an existing PrPM system. Our evaluation indicates that LLMs can help users of PrPM systems to better understand the recommendations. The results indicate that LLMs can help users of PrPM systems to better understand the recommendations, and to produce recommendations that have sufficient detail and fulfill their expectations. However, the explanations fall short in addressing the underlying \"why\" and do not always support users in assessing the trustworthiness of the recommendations.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "10:00",
        "session": "LLM applications",
        "poster_positions": "From board n11 to board n14"
    },
    {
        "id": "9352",
        "title": "Decoupled Search for the Masses: A Novel Task Transformation for Classical Planning (Extended Abstract)",
        "authors": "David Speck, Daniel Gnad",
        "abstract": "Classical planning provides a framework for solving sequential decision-making problems, i.e., finding a sequence of actions that transforms the current state of the world into a state that satisfies a desired goal condition. Planning tasks are modeled in a logic that describes the environment and its dynamics. It is well known that the specific problem formulation can significantly affect the performance of planning systems solving problems like the Rubik's Cube or finding algorithms for matrix multiplication. In this work, we propose a domain-general problem reformulation that embodies decoupled search, a search-reduction technique from classical planning and model checking. Decoupled search decomposes a given problem to exploit its structure, achieving exponential reductions over other search techniques. We show that decoupled search can be captured exactly as a task reformulation and that, on many benchmark domains, it performs as good and sometimes even better than a native decoupled-search implementation.",
        "location": "Montreal",
        "day": "August 20th",
        "hour": "10:00",
        "session": "Planning and Scheduling (3\/5)",
        "poster_positions": "From board n50 to board n54"
    }
]