[
    {
        "id": "947",
        "title": "Deep Reinforcement Learning for Efficient and Fair Allocation of Healthcare Resources",
        "authors": "Yikuan Li, Chengsheng Mao, Kaixuan Huang, Hanyin Wang, Zheng Yu, Mengdi Wang, Yuan Luo",
        "abstract": "The scarcity of health care resources, such as ventilators, often leads to the unavoidable consequence of rationing, particularly during public health emergencies or in resource-constrained settings like pandemics. The absence of a universally accepted standard for resource allocation protocols results in governments relying on varying criteria and heuristic-based approaches, often yielding suboptimal and inequitable outcomes. This study addresses the societal challenge of fair and effective critical care resource allocation by leveraging deep reinforcement learning to optimize policy decisions. We propose a transformer-based deep Q-network that integrates individual patient disease progression and interaction effects among patients to enhance allocation decisions. Our method aims to improve both fairness and overall patient outcomes. Experiments using metrics such as normalized survival rates and interracial allocation rate differences demonstrate that our approach significantly reduces excess deaths and achieves more equitable resource allocation compared to severity- and comorbidity-based protocols currently in use. Our findings highlight the potential of deep reinforcement learning to address critical health care challenges.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "10:00",
        "session": "AI for Social Good (4\/8)",
        "poster_positions": "From board n118 to board n120"
    },
    {
        "id": "1652",
        "title": "MutationGuard: A Graph and Temporal-Spatial Neural Method for Detecting Mutation Telecommunication Fraud",
        "authors": "Haitao Bai, Pinghui Wang, Ruofei Zhang, Ziyang Zhou, Juxiang Zeng, Yulou Su, Li Xing, Zhou Su, Chen Zhang, Lizhen Cui, Jun Hao, Wei Wang",
        "abstract": "Telecommunication fraud refers to deceptive activities in the field of communication services. This research focuses on a category of fraud identified as ''mutation telecommunication fraud\". There is currently a lack of research on mutation telecommunication fraud detection, allowing this type of fraud to persist uncaught. We identify that detecting mutation fraud requires capturing multi-source patterns, including user communication graphs and temporal-spatial Voice of Call (VOC) features. Specifically, we introduce MutationGuard, which leverages Graph Neural Networks (GNN) to capture changes in user communication graphs. For VOC records, we map call start times onto a 3D cylindrical surface, thereby representing each VOC record in spatial coordinates and utilizing proposed LFFE and TCFE modules to capture local fraud behaviors and temporal behavior changes. The proposed neural modeling approach that facilitates multi-source information fusion constitutes a significant advancement in detecting mutation fraud.\r\nExperiment results reveal a significant improvement in the AUC score by 1.52% and the F1 score by 1.36% on the proposed telecommunication fraud dataset. Particularly, our method shows a significant improvement of 13.93% in accuracy on mutation fraud data. We also validate the effectiveness of our method on the publicly available Sichuan Telecommunication Fraud dataset.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "2373",
        "title": "OpenCarbon: A Contrastive Learning-based Cross-Modality Neural Approach for High-Resolution Carbon Emission Prediction Using Open Data",
        "authors": "Jinwei Zeng, Yu Liu, Guozhen Zhang, Jingtao Ding, Yuming Lin, Jian Yuan, Yong Li",
        "abstract": "Accurately estimating high-resolution carbon emissions is crucial for effective emission governance and mitigation planning. While conventional methods for precise carbon accounting are hindered by substantial data collection efforts, the rise of open data and advanced learning techniques offers a promising solution. Once an open data-based prediction model is developed and trained, it can easily infer emissions for new areas based on available open data. To address this, we incorporate two modalities of open data, satellite images and point-of-interest (POI) data, to predict high-resolution urban carbon emissions, with satellite images providing macroscopic and static and POI data offering fine-grained and relatively dynamic functionality information. However, estimating high-resolution carbon emissions presents two significant challenges: the intertwined and implicit effects of various functionalities on carbon emissions, and the complex spatial contiguity correlations that give rise to the agglomeration effect. Our model, OpenCarbon, features two major designs that target the challenges: a cross-modality information extraction and fusion module to extract complementary functionality information from two modules and model their interactions, and a neighborhood-informed aggregation module to capture the spatial contiguity correlations. Extensive experiments demonstrate our model's superiority, with a significant performance gain of 26.6% on R2. Further generalizability tests and case studies also show OpenCarbon's capacity to capture the intrinsic relation between urban functionalities and carbon emissions, validating its potential to empower efficient carbon governance and targeted carbon mitigation planning. Codes and data are available: https:\/\/github.com\/JinweiZzz\/OpenCarbon.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "583",
        "title": "Deconfounding Multi-Cause Latent Confounders: A Factor-Model Approach to Climate Model Bias Correction",
        "authors": "Wentao Gao, Jiuyong Li, Debo Cheng, Lin Liu, Jixue Liu, Thuc Le, Xiaojing Du, Xiongren Chen, Yun Chen, Yanchang Zhao",
        "abstract": "Global Climate Models (GCMs) are crucial for predicting future climate changes by simulating the Earth systems. However, GCM outputs exhibit systematic biases due to model uncertainties, parameterization simplifications, and inadequate representation of complex climate phenomena. Traditional bias correction methods, which rely on historical observation data and statistical techniques, often neglect unobserved confounders, leading to biased results. This paper proposes a novel bias correction approach to utilize both GCM and observational data to learn a factor model that captures multi-cause latent confounders. Inspired by recent advances in causality based time series deconfounding, our method first constructs a factor model to learn latent confounders from historical data and then applies them to enhance the bias correction process using advanced time series forecasting models. The experimental results demonstrate significant improvements in the accuracy of precipitation outputs. By addressing unobserved confounders, our approach offers a robust and theoretically grounded solution for climate model bias correction.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "5645",
        "title": "Reinforcement Learning for Hybrid Charging Stations Planning and Operation Considering Fixed and Mobile Chargers",
        "authors": "Yanchen Zhu, Honghui Zou, Chufan Liu, Yuyu Luo, Yuankai Wu, Yuxuan Liang",
        "abstract": "efficient and adaptable charging infrastructure. Fixed-location charging stations often suffer from underutilization or congestion due to fluctuating demand, while mobile chargers offer flexibility by relocating as needed. This paper studies the optimal planning and operation of hybrid charging infrastructures that combine both fixed and mobile chargers within urban road networks. We formulate the Hybrid Charging Station Planning and Operation (HCSPO) problem, jointly optimizing the placement of fixed stations and the scheduling of mobile chargers. A charging demand prediction model based on Model Predictive Control (MPC) supports dynamic decision-making. To solve the HCSPO problem, we propose a deep reinforcement learning approach enhanced with heuristic scheduling. Experiments on real-world urban scenarios show that our method improves infrastructure availability—achieving up to 244.4% increase in coverage—and reduces user inconvenience with up to 79.8% shorter waiting times, compared to existing solutions.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "1005",
        "title": "Beyond Patterns: Harnessing Causal Logic for Autonomous Driving Trajectory Prediction",
        "authors": "Bonan Wang, Haicheng Liao, Chengyue Wang, Bin Rao, Yanchen Guan, Guyang Yu, Jiaxun Zhang, Songning Lai, Chengzhong Xu, Zhenning Li",
        "abstract": "Accurate trajectory prediction has long been a major challenge for autonomous driving (AD). Traditional data-driven models predominantly rely on statistical correlations, often overlooking the causal relationships that govern traffic behavior. In this paper, we introduce a novel trajectory prediction framework that leverages causal inference to enhance predictive robustness,  generalization, and accuracy. By decomposing the environment into spatial and temporal components, our approach identifies and mitigates spurious correlations, uncovering genuine causal relationships. We also employ a progressive fusion strategy to integrate multimodal information, simulating human-like reasoning processes and enabling real-time inference. Evaluations on five real-world datasets—ApolloScape, nuScenes, NGSIM, HighD, and MoCAD—demonstrate our model's superiority over existing state-of-the-art (SOTA) methods, with improvements in key metrics such as RMSE and FDE. Our findings highlight the potential of causal reasoning to transform trajectory prediction, paving the way for robust AD systems.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "8917",
        "title": "Enhancing Online Climate Discourse: A Two-Stage Framework for Climate Content Categorization and Moderation",
        "authors": "Apoorva Upadhyaya, Wolfgang Nejdl, Marco Fisichella",
        "abstract": "Climate change is one of the most pressing global challenges that requires urgent adaptation and resilience efforts, highlighting the need for both scientific solutions and effective communication. In the digital age, online content plays a key role in shaping climate narratives. Therefore, previous research has mainly focused on public perception or categorized content by topics such as impacts, mitigation, policy, etc. Despite these efforts, identifying discussions that address climate change adaptation is crucial for monitoring resilience and assessing public sentiment, while recognizing denial narratives helps combat misinformation. Moreover, the public's exposure to online climate content can either lead to or hinder climate action, emphasizing the need for climate content moderation. To address these issues, we propose a novel multi-stage framework where stage 1 categorizes climate-related content into adaptation, resilience, and denial while stage 2 moderates content by enhancing or intervening based on its alignment with climate goals. We present a novel dataset by manually annotating publicly available tweets and news articles into different climate categories with the help of a taxonomy developed by domain experts. Extensive experiments with benchmark climate and other domain datasets validate the efficacy of our prediction stage, while human and external evaluations confirm the relevance of our moderation stage.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "10:00",
        "session": "AI for Social Good (7\/8)",
        "poster_positions": "From board n46 to board n48"
    },
    {
        "id": "8707",
        "title": "QBR – A Question-Bank-Based Approach to Fine-Grained Legal Knowledge Retrieval for the General Public",
        "authors": "Mingruo Yuan, Ben Kao, Tien-Hsuan Wu",
        "abstract": "Retrieval of legal knowledge by the general public is a challenging problem due to the technicality of the professional knowledge and the lack of fundamental understanding by laypersons on the subject. Traditional information retrieval techniques assume that users are capable of formulating succinct and precise queries for effective document retrieval. In practice, however, the wide gap between the highly technical contents and untrained users makes legal knowledge retrieval very difficult. We propose a methodology, called QBR, which employs a Questions Bank (QB) as an effective medium for bridging the knowledge gap. We show how the QB is used to derive training samples to enhance the embedding of knowledge units within documents, which leads to effective fine-grained knowledge retrieval. We discuss and evaluate through experiments various advantages of QBR over traditional methods. These include more accurate, efficient, and explainable document retrieval, better comprehension of retrieval results, and highly effective fine-grained knowledge retrieval. We also present some case studies and show that QBR achieves social impact by assisting citizens to resolve everyday legal concerns.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "15:00",
        "session": "AI for Social Good (2\/8)",
        "poster_positions": "From board n84 to board n89"
    },
    {
        "id": "8927",
        "title": "Early Detection of Patient Deterioration from Real-Time Wearable Monitoring System",
        "authors": "Lo Pang-Yun Ting, Hong-Pei Chen, An-Shan Liu, Chun-Yin Yeh, Po-Lin Chen, Kun-Ta Chuang",
        "abstract": "Early detection of patient deterioration is crucial for reducing mortality rates. Heart rate data has shown promise in assessing patient health, and wearable devices offer a cost-effective solution for real-time monitoring. However, extracting meaningful insights from diverse heart rate data and handling missing values in wearable device data remain key challenges.  To address these challenges, we propose TARL, an innovative approach that models the structural relationships of representative subsequences, known as shapelets, in heart rate time series. TARL creates a shapelet-transition knowledge graph to model shapelet dynamics in heart rate time series, indicating illness progression and potential future changes. We further introduce a transition-aware knowledge embedding to reinforce relationships among shapelets and quantify the impact of missing values, enabling the formulation of comprehensive heart rate representations. These representations capture explanatory structures and predict future heart rate trends, aiding early illness detection. We collaborate with physicians and nurses to gather ICU patient heart rate data from wearables and diagnostic metrics to assess illness severity and evaluate deterioration. Experiments on real-world ICU data demonstrate that TARL achieves both high reliability and early detection. A case study further showcases TARL's explainable detection process, highlighting its potential as an AI-driven tool to assist clinicians in recognizing early signs of patient deterioration.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "11:30",
        "session": "AI for Social Good (8\/8)",
        "poster_positions": "From board n49 to board n52"
    },
    {
        "id": "9225",
        "title": "An Ethical Dataset from Real-World Interactions Between Users and Large Language Models",
        "authors": "Masahiro Kaneko, Danushka Bollegala, Timothy Baldwin",
        "abstract": "Recent studies have demonstrated that Large Language Models (LLMs) have ethical-related problems such as social biases, lack of moral reasoning, and generation of offensive content.\r\n    The existing evaluation metrics and methods to address these ethical challenges use datasets intentionally created by instructing humans to create instances including ethical problems.\r\n    Therefore, the data does not sufficiently include comprehensive prompts that users actually provide when using LLM services in everyday contexts and outputs that LLMs generate.\r\n    There may be different tendencies between unethical instances intentionally created by humans and actual user interactions with LLM services, which could result in a lack of comprehensive evaluation.\r\n    To investigate the difference, we create Eagle datasets extracted from actual interactions between ChatGPT and users that exhibit social biases, opinion biases, toxicity, and immoral problems.\r\n    Our experiments show that Eagle captures complementary aspects, not covered by existing datasets proposed for evaluation and mitigation.\r\n    We argue that using both existing and proposed datasets leads to a more comprehensive assessment of the ethics.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "15:00",
        "session": "AI for Social Good (6\/8)",
        "poster_positions": "From board n112 to board n117"
    },
    {
        "id": "8718",
        "title": "Towards the Terminator Economy: Assessing Job Exposure to AI Through LLMs",
        "authors": "Emilio Colombo, Fabio Mercorio, Mario Mezzanzanica, Antonio Serino",
        "abstract": "AI and related technologies are reshaping jobs and tasks, either by automating or augmenting human skills in the workplace. Many researchers have been working on estimating if and to what extent jobs and tasks are exposed to the risk of being automatized by AI-related technologies. Our work tackles this issue through a data-driven approach by:\r\n(i) developing a reproducible framework that uses cutting-edge open-source large language models to assess the current capabilities of AI and robotics in performing job-related tasks;\r\n(ii) formalizing and computing a measure of AI exposure by occupation, the Task Exposure to AI (TEAI) index, and a measure of Task Replacement by AI (TRAI) index, both validated through a human user evaluation and compared with the state-of-the-art.\r\n\r\nOur results show that the TEAI index is positively correlated with cognitive, problem-solving, and management skills, while it is negatively correlated with social skills. Results also suggest that about one-third of U.S. employment is highly exposed to AI, primarily in high-skill jobs requiring a graduate or postgraduate level of education. We also find that AI exposure is positively associated with employment and wage growth from 2003 to 2023, suggesting that AI has had an overall positive effect on productivity.\r\n\r\nConsidering specifically the TRAI index, we find that even in high-skill occupations, AI exhibits high variability in task substitution, suggesting that AI and humans complement each other within the same occupation, while the allocation of tasks within occupations is likely to change.\r\n\r\nAll results, models, and code are freely available online to allow the community to reproduce our results, compare outcomes, and use our work as a benchmark to monitor AI’s progress over time.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "15:00",
        "session": "AI for Social Good (6\/8)",
        "poster_positions": "From board n112 to board n117"
    },
    {
        "id": "8956",
        "title": "LogiDebrief: A Signal-Temporal Logic Based Automated Debriefing Approach with Large Language Models Integration",
        "authors": "Zirong Chen, Ziyan An, Jennifer Reynolds, Kristin Mullen, Stephen Maritini, Meiyi Ma",
        "abstract": "Emergency response services are critical to public safety, with 9-1-1 call-takers playing a key role in ensuring timely and effective emergency operations. To ensure call-taking performance consistency, quality assurance is implemented to evaluate and refine call-takers' skillsets. However, traditional human-led evaluations struggle with high call volumes, leading to low coverage and delayed assessments.  We introduce LogiDebrief, an AI-driven framework that automates traditional 9-1-1 call debriefing by integrating Signal-Temporal Logic (STL) with Large Language Models (LLMs) for fully-covered rigorous performance evaluation. LogiDebrief formalizes call-taking requirements as logical specifications, enabling systematic assessment of 9-1-1 calls against procedural guidelines. It employs a three-step verification process: (1) contextual understanding to identify responder types, incident classifications, and critical conditions; (2) STL-based runtime checking with LLM integration to ensure compliance; and (3) automated aggregation of results into quality assurance reports.  Beyond its technical contributions, LogiDebrief has demonstrated real-world impact. Successfully deployed at Metro Nashville Department of Emergency Communications, it has assisted in debriefing 1,701 real-world calls, saving 311.85 hours of active engagement. Empirical evaluation with real-world data confirms its accuracy, while a case study and extensive user study highlight its effectiveness in enhancing call-taking performance.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "15:00",
        "session": "AI for Social Good (2\/8)",
        "poster_positions": "From board n84 to board n89"
    },
    {
        "id": "9040",
        "title": "CoDiCast: Conditional Diffusion Model for Global Weather Forecasting with Uncertainty Quantification",
        "authors": "Jimeng Shi, Bowen Jin, Jiawei Han, Sundararaman Gopalakrishnan, Giri Narasimhan",
        "abstract": "Accurate weather forecasting is critical for science and society. However, existing methods have not achieved the combination of high accuracy, low uncertainty, and high computational efficiency simultaneously. On one hand, traditional numerical weather prediction (NWP) models are computationally intensive because of their complexity. On the other hand, most machine learning-based weather prediction (MLWP) approaches offer efficiency and accuracy but remain deterministic, lacking the ability to capture forecast uncertainty. To tackle these challenges, we propose a conditional diffusion model, CoDiCast, to generate global weather prediction, integrating accuracy and uncertainty quantification at a modest computational cost. The key idea behind the prediction task is to generate realistic weather scenarios at a future time point, conditioned on observations from the recent past. Due to the probabilistic nature of diffusion models, they can be properly applied to capture the uncertainty of weather predictions. Therefore, we accomplish uncertainty quantifications by repeatedly sampling from stochastic Gaussian noise for each initial weather state and running the denoising process multiple times. Experimental results demonstrate that CoDiCast outperforms several existing MLWP methods in accuracy, and is faster than NWP models in inference speed. Our model can generate 6-day global weather forecasts, at 6-hour steps and 5.625-degree latitude-longitude resolutions, for over 5 variables, in about 12 minutes on a commodity A100 GPU machine with 80GB memory. The source code is available at https:\/\/github.com\/JimengShi\/CoDiCast.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "11:30",
        "session": "AI for Social Good (5\/8)",
        "poster_positions": "From board n121 to board n123"
    },
    {
        "id": "8400",
        "title": "Detection and Geographic Localization of Natural Objects in the Wild: A Case Study on Palms",
        "authors": "Kangning Cui, Rongkun Zhu, Manqi Wang, Wei Tang, Gregory D. Larsen, Victor P. Pauca, Sarra Alqahtani, Fan Yang, David Segurado, David A. Lutz, Jean-Michel Morel, Miles R. Silman",
        "abstract": "Palms are ecologically and economically indicators of tropical forest health, biodiversity, and human impact that support local economies and global forest product supply chains. While palm detection in plantations is well-studied, efforts to map naturally occurring palms in dense forests remain limited by overlapping crowns, uneven shading, and heterogeneous landscapes. We develop PRISM (Processing, Inference, Segmentation, and Mapping), a flexible pipeline for detecting and localizing palms in dense tropical forests using large orthomosaic images. Orthomosaics are created from thousands of aerial images and spanning several to hundreds of gigabytes. Our contributions are threefold. First, we construct a large UAV-derived orthomosaic dataset collected across 21 ecologically diverse sites in western Ecuador, annotated with 8,830 bounding boxes and 5,026 palm center points. Second, we evaluate multiple state-of-the-art object detectors based on efficiency and performance, integrating zero-shot SAM~2 as the segmentation backbone, and refining the results for precise geographic mapping. Third, we apply calibration methods to align confidence scores with IoU and explore saliency maps for feature explainability. Though optimized for palms, PRISM is adaptable for identifying other natural objects, such as eastern white pines. Future work will explore transfer learning for lower-resolution datasets (0.5–1m). Data and code can be found at github.com\/Zippppo\/PRISM.",
        "location": "Montreal",
        "day": "August 20th",
        "hour": "14:00",
        "session": "AI for Social Good (3\/8)",
        "poster_positions": "From board n98 to board n104"
    },
    {
        "id": "8760",
        "title": "Exploring Equity of Climate Policies Using Multi-Agent Multi-Objective Reinforcement Learning",
        "authors": "Palok Biswas, Zuzanna Osika, Isidoro Tamassia, Adit Whorra, Jazmin Zatarain-Salazar, Jan Kwakkel, Frans A. Oliehoek, Pradeep K. Murukannaiah",
        "abstract": "Addressing climate change requires coordinated policy efforts of nations worldwide. These efforts are informed by scientific reports, which rely in part on Integrated Assessment Models (IAMs), prominent tools used to assess the economic impacts of climate policies. However, traditional IAMs optimize policies based on a single objective, limiting their ability to capture the trade-offs among economic growth, temperature goals, and climate justice. As a result, policy recommendations have been criticized for perpetuating inequalities, fueling disagreements during policy negotiations. We introduce JUSTICE, the first framework integrating IAM with Multi-Objective Multi-Agent Reinforcement Learning (MOMARL). By incorporating multiple objectives, JUSTICE generates policy recommendations that shed light on equity while balancing climate and economic goals. Further, using multiple agents can provide a realistic representation of the interactions among the diverse policy actors. We identify equitable Pareto-optimal policies using our framework, which facilitates deliberative decision-making by presenting policymakers with the inherent trade-offs in climate and economic policy.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "11:30",
        "session": "AI for Social Good (1\/8)",
        "poster_positions": "From board n90 to board n93"
    },
    {
        "id": "8411",
        "title": "DeepShade: Enable Shade Simulation by Text-conditioned Image Generation",
        "authors": "Longchao Da, Xiangrui Liu, Mithun Shivakoti, Thirulogasankar Pranav Kutralingam, Yezhou Yang, Hua Wei",
        "abstract": "Heatwaves pose a significant threat to public health, especially as global warming intensifies. However, current routing systems (e.g., online maps) fail to incorporate shade information due to the difficulty of estimating shades directly from noisy satellite imagery and the limited availability of training data for generative models. In this paper, we address these challenges through two main contributions. First, we build an extensive dataset covering diverse longitude-latitude regions, varying levels of building density, and different urban layouts. Leveraging Blender-based 3D simulations alongside building outlines, we capture building shadows under various solar zenith angles throughout the year and at different times of day. These simulated shadows are aligned with satellite images, providing a rich resource for learning shade patterns. Second, we propose the DeepShade, a diffusion-based model designed to learn and synthesize shade variations over time. It emphasizes the nuance of edge features by jointly considering RGB with the Canny edge layer, and incorporates contrastive learning to capture the temporal change rules of shade. Then, by conditioning on textual descriptions of known conditions (e.g., time of day, solar angles), our framework provides improved performance in generating shade images. We demonstrate the utility of our approach by using our shade predictions to calculate shade ratios for real-world route planning in Tempe, Arizona.  We believe this work will benefit society by providing a reference for urban planning in extreme heat weather and its potential practical applications in the environment.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "11:30",
        "session": "AI for Social Good (8\/8)",
        "poster_positions": "From board n49 to board n52"
    },
    {
        "id": "1853",
        "title": "Faster Annotation for Elevation-Guided Flood Extent Mapping by Consistency-Enhanced Active Learning",
        "authors": "Saugat Adhikari, Da Yan, Tianyang Wang, Landon Dyken, Sidharth Kumar, Lyuheng Yuan, Akhlaque Ahmad, Jiao Han, Yang Zhou, Steve Petruzza",
        "abstract": "Flood extent mapping is crucial for disaster response and damage assessment. While Earth imagery and terrain data (in the form of DEM) are now readily available, there are few flood annotation data for training machine learning models, which hinders the automated mapping of flooded areas. We propose ALFA, an interactive active-learning-based approach to minimize the annotators' efforts when preparing the ground-truth flood map in a satellite image. ALFA calibrates the prediction consistency of a segmentation model (1) across training cycles and (2) for various data augmentations. The two consistencies are integrated into the design of both the acquisition function and the loss function to enhance the robustness of active learning with limited annotation inputs. ALFA recommends those superpixels that the underlying model is most uncertain about, and users can annotate their pixels with minimal clicks with the help of elevation guidance. Extensive experiments on various regions hit by flooding show that we can improve the annotation time from hours to around 20 minutes. ALFA is open sourced at https:\/\/github.com\/saugatadhikari\/alfa.",
        "location": "Montreal",
        "day": "August 20th",
        "hour": "14:00",
        "session": "AI for Social Good (3\/8)",
        "poster_positions": "From board n98 to board n104"
    },
    {
        "id": "8422",
        "title": "Exploring Multimodal Foundation AI and Expert-in-the-Loop for Sustainable Management of Wild Salmon Fisheries in Indigenous Rivers",
        "authors": "Chi Xu, Yili Jin, Sami Ma, Rongsheng Qian, Hao Fang, Jiangchuan Liu, Xue Liu, Edith C.H. Ngai, William I. Atlas, Katrina M. Connors, Mark A. Spoljaric",
        "abstract": "Wild salmon are essential to the ecological, economic, and cultural sustainability of the North Pacific Rim. Yet climate variability, habitat loss, and data limitations in remote ecosystems that lack basic infrastructure support pose significant challenges to effective fisheries management. This project explores the integration of multimodal foundation AI and expert-in-the-loop frameworks to enhance wild salmon monitoring and sustainable fisheries management in Indigenous rivers across Pacific Northwest. By leveraging video and sonar-based monitoring, we develop AI-powered tools for automated species identification, counting, and length measurement, reducing manual effort, expediting delivery of results, and improving decision-making accuracy. Expert validation and active learning frameworks ensure ecological relevance while reducing annotation burdens. To address unique technical and societal challenges, we bring together a cross-domain, interdisciplinary team of university researchers, fisheries biologists, Indigenous stewardship practitioners, government agencies, and conservation organizations. Through these collaborations, our research fosters ethical AI co-development, open data sharing, and culturally informed fisheries management.",
        "location": "Montreal",
        "day": "August 20th",
        "hour": "14:00",
        "session": "AI for Social Good (3\/8)",
        "poster_positions": "From board n98 to board n104"
    },
    {
        "id": "9049",
        "title": "IGraSS: Learning to Identify Infrastructure Networks from Satellite Imagery by Iterative Graph-constrained Semantic Segmentation",
        "authors": "Oishee Bintey Hoque, Abhijin Adiga, Aniruddha Adiga, Siddharth Chaudhary, Madhav V. Marathe, S.S. Ravi, Kirti Rajagopalan, Amanda Wilson, Samarth Swarup",
        "abstract": "Accurate canal network mapping is essential for water management, including irrigation planning and infrastructure maintenance. State-of-the-art semantic segmentation models for infrastructure mapping, such as roads, rely on large, well-annotated remote sensing datasets. However, incomplete or inadequate ground truth can hinder these learning approaches. Many infrastructure networks have graph-level properties such as reachability to a source (like canals) or connectivity (roads) that can be leveraged to improve these existing ground truth. This paper develops a novel iterative framework IGraSS, combining a semantic segmentation module—incorporating RGB and additional modalities (NDWI, DEM)—with a graph-based ground-truth refinement module. The segmentation module processes satellite imagery patches, while the refinement module operates on the entire data viewing the infrastructure network as a graph. Experiments show that IGraSS reduces unreachable canal segments from ~18% to ~3%, and training with refined ground truth significantly improves canal identification. IGraSS serves as a robust framework for both refining noisy ground truth and mapping canal networks from remote sensing imagery. We also demonstrate the effectiveness and generalizability of IGraSS using road networks as an example, applying a different graph-theoretic constraint to complete road networks.",
        "location": "Montreal",
        "day": "August 20th",
        "hour": "14:00",
        "session": "AI for Social Good (3\/8)",
        "poster_positions": "From board n98 to board n104"
    },
    {
        "id": "8747",
        "title": "ECG2TOK: ECG Pre-Training with Self-Distillation Semantic Tokenizers",
        "authors": "Xiaoyan Yuan, Wei Wang, Han Liu, Jian Chen, Xiping Hu",
        "abstract": "Self-supervised learning (SSL) has garnered increasing attention in electrocardiogram (ECG) analysis for its effectiveness in resource-limited settings. Existing state-of-the-art SSL methods rely on time-frequency detail reconstruction, but due to the inherent redundancy of ECG signals and individual variability, these approaches often yield suboptimal performance. In contrast, discrete label prediction becomes a superior pre-training objective by encouraging models to efficiently abstract ECG high-level semantics. However, the continuity and significant variability of ECG signals pose a challenge in generating semantically discrete labels. To address this issue, we propose an ECG pretraining framework with a self-distillation semantic tokenizer (ECG2TOK), which maps continuous ECG signals into discrete labels for self-supervised training. Specifically, the tokenizer extracts semantically aware embeddings of ECG by self-distillation and performs online clustering to generate semantically rich discrete labels. Subsequently, the SSL model is trained in conjunction with masking strategies and discrete label prediction to facilitate the abstraction of high-level semantic representations. We evaluate ECG2TOK in six downstream tasks, demonstrating that ECG2TOK efficiently achieves state-of-the-art performance and up to a 30.73% AUC increase in low-resource scenarios. Moreover, visualization experiments demonstrate that the discrete labels generated by ECG2TOK exhibit consistent semantics closely associated with clinical features. Our code is available on https:\/\/github.com\/YXYanova\/ECG2TOK.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "8930",
        "title": "Bidirectional Human–AI Collaboration for Equitable Student Performance Prediction via Deep Uncertainty Learning",
        "authors": "Ruohan Zong, Yang Zhang, Lanyu Shang, Frank Stinar, Nigel Bosch, Dong Wang",
        "abstract": "This paper studies a bidirectional human-AI collaborative student performance prediction problem to enhance equitable online education, aligning with the United Nations' Sustainable Development Goal (SDG) of ensuring inclusive and equitable quality education for all. The goal is to leverage collaborative intelligence to generate accurate and fair student outcome predictions from behavioral data, ensuring equitable estimation for underrepresented populations. Current fair AI solutions often fail to mitigate demographic bias in the absence of student demographic data, while human-AI collaborative approaches frequently overlook human cognitive biases, leading to inaccurate predictions. We develop CollabDebias, a novel bidirectional human-AI collaborative framework that utilizes the complementary strengths of AI and humans to mitigate the AI demographic bias and human cognitive bias. To address AI demographic bias, we propose an uncertainty learning-based bias identification method and a reliability-aware human-AI integration approach. To reduce human cognitive bias, we design uncertainty-aware visualization of AI decision area and attention mechanism. Experimental results on an online course demonstrate CollabDebias's effectiveness in improving student performance prediction accuracy and fairness.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "15:00",
        "session": "AI for Social Good (6\/8)",
        "poster_positions": "From board n112 to board n117"
    },
    {
        "id": "8435",
        "title": "HARMONY: A Privacy-preserving and Sensor-agnostic Tele-monitoring system",
        "authors": "Qipeng Xie, Hao Guo, Weizheng Wang, Yongzhi Huang, Linshan Jiang, Jiafei Wu, Shuxin Zhong, Lu Wang, Kaishun Wu",
        "abstract": "Global aging necessitates tele-monitoring systems to provide real-time tracking and timely assistance for older adults living independently. While pervasive wireless devices (e.g., CSI, IMU, UWB) enable cost-effective, non-intrusive monitoring, existing systems lack flexibility, limiting their adaptability to different environments. In this work, we posit that the motion dynamics of human movement are invariant across sensing modalities, inspiring the design of HARMONY—a privacy-preserving, sensor-agnostic system that supports multi-modal inputs and diverse tele-monitoring tasks. HARMONY incorporates Modality-agnostic Data Processing to uniformly encrypt multi-modal signals and Task-specific Activity Recognition for seamless tasks adaptation. A novel Encrypted-processing Engine then significantly accelerates computations on encrypted data by optimizing matrix and convolution operations. Evaluations across five different sensing modalities show that HARMONY consistently achieves high accuracy while delivering 3.5 × to 130 × speedups over state-of-the-art baselines. Our results demonstrate that HARMONY is a practical, scalable, and privacy-centric prototype for next-generation remote healthcare.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "1058",
        "title": "Weather Foundation Model Enhanced Decentralized Photovoltaic Power Forecasting Through Spatio-temporal Knowledge Distillation",
        "authors": "Fang He, Jiaqi Fan, Yang Deng, Xiaoyang Zhang, Ka Tai Lau, Dan Wang",
        "abstract": "The solar photovoltaic power forecasting (SPPF) of a PV system is vital for the downstream power estimation. While approaches for recent decentralized PV systems require customized models for each PV installation, this method is labor-intensive and not scalable. Therefore, developing a general SPPF model for a decentralized PV system is essential. The primary challenge in developing such a model is accounting for regional weather variations. Recent advancements in weather foundation models (WFMs) offer a promising opportunity, providing accurate forecasts with reduced computational demands. However, integrating WFMs into SPPF models remains challenging due to the complexity of WFMs. This paper introduces a novel approach, spatio-temporal knowledge distillation (STKD), to efficiently adapt WFMs for SPPF. The proposed STKD-PV models leverage regional weather and PV power data to forecast power generation from six hours to a day ahead. Globally evaluated across six datasets, STKD-PV models demonstrate superior performance compared to state-of-the-art (SOTA) time-series models and fine-tuned WFMs, achieving significant improvements in forecasting accuracy. This study marks the first application of knowledge distillation from WFMs to SPPF, offering a scalable and cost-effective solution for decentralized PV systems.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "8923",
        "title": "Expanding Connected Components from Alternative Terminals: Global Optimization for Freshwater Fishes Under the UN's 30x30 Conservation Goal",
        "authors": "Yue Mao, Zhongdi Qu, Imanol Miqueleiz, Aaron Ferber, Sami Wolf, Marc Grimson, Sebastian Heilpern, Felipe S. Pacheco, Alexander S. Flecker, Peter B. McIntyre, Carla P. Gomes",
        "abstract": "Climate change and biodiversity loss are among humanity’s most pressing challenges. In 2022, under the auspices of the United Nations, over 190 countries reached a historic agreement to address the alarming loss of biodiversity and restore natural ecosystems. Target 3, often referred to as ``30x30'', seeks to effectively protect and manage 30% of the world’s terrestrial, inland water, coastal, and marine areas by 2030. In this work, we address the UN 30x30 target in the context of global freshwater fish conservation. Freshwater ecosystems are disproportionately unprotected, and their biota are declining at an alarming rate. Our goal is to select new protected areas that protect freshwater fish species as much as possible without exceeding total coverage of 30% of land area. To support this goal, we introduce the Expansion of Connected Components from Alternative Terminals Problem, a graph-based optimization problem that captures ecological priorities and connectivity constraints. We analyze its computational complexity, propose novel integer programming formulations, and develop scalable solution methods. We further evaluate its typical-case complexity under diverse settings and demonstrate that our approach scales to a global real-world scope, encompassing approximately 200,000 freshwater basins and 13,000 species, paving the way for implementing the 30x30 target on a worldwide scale.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "11:30",
        "session": "AI for Social Good (5\/8)",
        "poster_positions": "From board n121 to board n123"
    },
    {
        "id": "8946",
        "title": "Perspectives in Play: A Multi-Perspective Approach for More Inclusive NLP Systems",
        "authors": "Benedetta Muscato, Lucia Passaro, Gizem Gezici, Fosca Giannotti",
        "abstract": "In the realm of Natural Language Processing (NLP), common approaches for handling human disagreement consist of aggregating annotators' viewpoints to establish a single ground truth. However, prior studies show that disregarding individual opinions can lead to the side-effect of under-representing minority perspectives, especially in subjective tasks, where annotators may systematically disagree because of their preferences. Recognizing that labels reflect the diverse backgrounds, life experiences, and values of individuals, this study proposes a new multi-perspective approach using soft labels to encourage the development of the next generation of perspective-aware models—more  inclusive and pluralistic. We conduct an extensive analysis across diverse subjective text classification tasks including hate speech, irony, abusive language, and stance detection, to highlight the importance of capturing human disagreements, often overlooked by traditional aggregation methods. Results show that the multi-perspective approach not only better approximates human label distributions, as measured by Jensen-Shannon Divergence (JSD), but also achieves superior classification performance (higher F1-scores), outperforming traditional approaches. However, our approach exhibits lower confidence in tasks like irony and stance detection, likely due to the inherent subjectivity present in the texts. Lastly, leveraging Explainable AI (XAI), we explore model uncertainty and uncover meaningful insights into model predictions. All implementation details are available at our github repo.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "10:00",
        "session": "AI for Social Good (7\/8)",
        "poster_positions": "From board n46 to board n48"
    },
    {
        "id": "9202",
        "title": "Moral Compass: A Data-Driven Benchmark for Ethical Cognition in AI",
        "authors": "Aisha Aijaz, Arnav Batra, Aryaan Bazaz, Srinath Srinivasa, Raghava Mutharaju, Manohar Kumar",
        "abstract": "We propose the Moral Compass benchmark, a point of reference for incorporating ethical cognition in AI. It has four key contributions. A Moral Decision Dataset (MDD) that captures cases with ethical ambiguity, along with parameters that aid moral decision-making. It is created using a methodology that leverages the use of Large Language Models (LLMs) and seed data from real-world sources which are processed, summarized, and augmented. We also introduce a Moral Decision Knowledge Graph (MDKG) that is created using feature mappings of the relational dataset MDD to facilitate efficient querying. To demonstrate the validity and robustness of this dataset, we introduce an Ethics Scoring Algorithm (ESA) that makes use of the parameters defined in the dataset to calculate ethical scores for isolated actions. Furthermore, ESA is extended by the novel concept of context-sensitive thresholding (CST) to discretize grey areas to resolve ethical dilemmas with explainable results. This work aims to facilitate ethical cognition in AI systems that are deployed in various important sections of society through a clear methodology, modular development, and broad applicability.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "8690",
        "title": "BGM: Demand Prediction for Expanding Bike-Sharing Systems with Dynamic Graph Modeling",
        "authors": "Yixuan Zhao, Hongkai Wen, Xingchen Zhang, Man Luo",
        "abstract": "Accurate demand prediction is crucial for the equitable and sustainable expansion of bike-sharing systems, which help reduce urban congestion, promote low-carbon mobility, and improve transportation access in underserved areas. However, expanding these systems presents societal challenges, particularly in ensuring fair resource distribution and operational efficiency. A major hurdle is the difficulty of demand prediction at new stations, which lack historical usage data and are heavily influenced by the existing network. Additionally, new stations dynamically reshape demand patterns across time and space, complicating efforts to balance supply and accessibility in evolving urban environments. Existing methods model relationships between new and existing stations but often assume static patterns, overlooking how new stations reshape demand dynamics over time and space. To tackle these challenges, we propose a novel demand prediction framework for expanding bike-sharing systems, namely BGM, which leverages dynamic graph modeling to capture the evolving inter-station correlations while accounting for spatial and temporal heterogeneity. Specifically, we develop a knowledge transfer approach that studies the embeddings transformation across existing and new stations through a learnable orthogonal mapping matrix. We further design a gated selecting vector-based feature fusion mechanism to integrate the transferred embeddings and the intrinsic features of stations for precise predictions. Experiments on real-world bike-sharing data demonstrate that BGM outperforms existing methods.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "8675",
        "title": "Uncertainty-aware Predict-Then-Optimize Framework for Equitable Post-Disaster Power Restoration",
        "authors": "Lin Jiang, Dahai Yu, Rongchao Xu, Tian Tang, Guang Wang",
        "abstract": "The increasing frequency of extreme weather events, such as hurricanes, highlights the urgent need for efficient and equitable power system restoration. Many electricity providers make restoration decisions primarily based on the volume of power restoration requests from each region. However, our data-driven analysis reveals significant disparities in request submission volume, as disadvantaged communities tend to submit fewer restoration requests. This disparity makes the current restoration solution inequitable, leaving these communities vulnerable to extended power outages. To address this, we aim to propose an equity-aware power restoration strategy that balances both restoration efficiency and equity across communities. However, achieving this goal is challenging for two reasons: the difficulty of predicting repair durations under dataset heteroscedasticity, and the tendency of reinforcement learning agents to favor low-uncertainty actions, which potentially undermine equity. To overcome these challenges, we design a predict-then-optimize framework called EPOPR with two key components: (1) Equity-Conformalized Quantile Regression for uncertainty-aware repair duration prediction, and (2) Spatial-Temporal Attentional RL that adapts to varying uncertainty levels across regions for equitable decision-making. Experimental results show that our EPOPR effectively reduces the average power outage duration by 3.60% and decreases inequity between different communities by 14.19% compared to state-of-the-art baselines.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "8459",
        "title": "Classifying and Tracking International Aid Contribution Towards SDGs",
        "authors": "Sungwon Park, Dongjoon Lee, Kyeongjin Ahn, Yubin Choi, Junho Lee, Meeyoung Cha, Kyung Ryul Park",
        "abstract": "International aid is a critical mechanism for promoting economic growth and well-being in developing nations, supporting progress toward the Sustainable Development Goals (SDGs). However, tracking aid contributions remains challenging due to labor-intensive data management, incomplete records, and the heterogeneous nature of aid data. Recognizing the urgency of this challenge, we partnered with government agencies to develop an AI model that complements manual classification and mitigates human bias in subjective interpretation. By integrating SDG-specific semantics and leveraging prior knowledge from language models, our approach enhances classification accuracy and accommodates the diversity of aid projects. When applied to a comprehensive dataset spanning multiple years, our model can reveal hidden trends in the temporal evolution of international development cooperation. Expert interviews further suggest how these insights can empower policymakers with data-driven decision-making tools, ultimately improving aid effectiveness and supporting progress toward SDGs.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "15:00",
        "session": "AI for Social Good (6\/8)",
        "poster_positions": "From board n112 to board n117"
    },
    {
        "id": "8912",
        "title": "Agent-based Modeling Meets the Capability Approach for Human Development: Simulating Homelessness Policy-making",
        "authors": "Alba Aguilera, Nardine Osman, Georgina Curto",
        "abstract": "The global rise in homelessness calls for urgent and alternative policy solutions. Non-profits and governmental organizations alert about the many challenges faced by people experiencing homelessness (PEH), which include not only the lack of shelter but also the lack of opportunities for personal development. In this context, the capability approach (CA), which underpins the United Nations Sustainable Development Goals (SDGs), provides a comprehensive framework to assess inequity in terms of real opportunities. This paper explores how the CA can be combined with agent-based modelling and reinforcement learning. The goals are: (1) implementing the CA as a Markov Decision Process (MDP), (2) building on such MDP to develop a rich decision-making model that accounts for more complex motivators of behaviour, such as values and needs, and (3) developing an agent-based simulation framework that allows to assess alternative policies aiming to expand or restore people's capabilities. The framework is developed in a real case study of health inequity and homelessness, working in collaboration with stakeholders, non-profits and domain experts. The ultimate goal of the project is to develop a novel agent-based simulation framework, rooted in the CA, which can be replicated in a diversity of social challenges to assess policies in a non-invasive way.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "11:30",
        "session": "AI for Social Good (1\/8)",
        "poster_positions": "From board n90 to board n93"
    },
    {
        "id": "8743",
        "title": "SMILE: A Scale-aware Multiple Instance Learning Method for Multicenter STAS  Lung Cancer Histopathology Diagnosis",
        "authors": "Liangrui Pan, Xiaoyu Li, Yutao Dou, Qiya Song, Jiadi Luo, Qingchun Liang, Shaoliang Peng",
        "abstract": "Spread through air spaces (STAS) represents a newly identified aggressive pattern in lung cancer, which is known to be associated with adverse prognostic factors and complex pathological features. Pathologists currently rely on time-consuming manual assessments, which are highly subjective and prone to variation. This highlights the urgent need for automated and precise diagnostic solutions. 2,970 lung cancer tissue slides are comprised from multiple centers, re-diagnosed them, and constructed and publicly released three lung cancer STAS datasets: STAS-CSU (hospital), STAS-TCGA, and STAS-CPTAC. All STAS datasets provide corresponding pathological feature diagnoses and related clinical data. To address the bias, sparse and heterogeneous nature of STAS, we propose an scale-aware multiple instance learning(SMILE) method for STAS diagnosis of lung cancer. By introducing a scale-adaptive attention mechanism, the SMILE can adaptively adjust high-attention instances, reducing over-reliance on local regions and promoting consistent detection of STAS lesions. Extensive experiments show that SMILE achieved competitive diagnostic results on STAS-CSU, diagnosing 251 and 319 STAS samples in CPTAC and TCGA, respectively, surpassing clinical average AUC. The 11 open baseline results are the first to be established for STAS research, laying the foundation for the future expansion, interpretability, and clinical integration of computational pathology technologies. The datasets and code are available at https:\/\/github.com\/panliangrui\/IJCAI25.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "8825",
        "title": "City-Level Foreign Direct Investment Prediction with Tabular Learning on Judicial Data",
        "authors": "Tianxing Wu, Lizhe Cao, Shuang Wang, Jiming Wang, Shutong Zhu, Yerong Wu, Yuqing Feng",
        "abstract": "To advance the United Nations Sustainable Development Goal on promoting sustained, inclusive, and sustainable economic growth, foreign direct investment (FDI) plays a crucial role in catalyzing economic expansion and fostering innovation. Precise city-level FDI prediction is quite important for local government and is commonly studied based on economic data (e.g., GDP). However, such economic data could be prone to manipulation, making predictions less reliable. To address this issue, we try to leverage large-scale judicial data which reflects judicial performance influencing local investment security and returns, for city-level FDI prediction. Based on this, we first build an index system for the evaluation of judicial performance over twelve million publicly available adjudication documents according to which a tabular dataset is reformulated. We then propose a new Tabular Learning method on Judicial Data (TLJD) for city-level FDI prediction. TLJD integrates row data and column data in our built tabular dataset for judicial performance indicator encoding, and utilizes a mixture of experts model to adjust the weights of different indicators considering regional variations. To validate the effectiveness of TLJD, we design cross-city and cross-time tasks for city-level FDI predictions. Extensive experiments on both tasks demonstrate the superiority of TLJD (reach to at\r\nleast 0.92 R2) over the other ten state-of-the-art baselines in different evaluation metrics.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "8601",
        "title": "Resolving Conflicting Evidence in Automated Fact-Checking: A Study on Retrieval-Augmented LLMs",
        "authors": "Ziyu Ge, Yuhao Wu, Daniel Wai Kit Chin, Roy Ka-Wei Lee, Rui Cao",
        "abstract": "Large Language Models (LLMs) augmented with retrieval mechanisms have demonstrated significant potential in fact-checking tasks by integrating external knowledge. However, their reliability decreases when confronted with conflicting evidence from sources of varying credibility. This paper presents the first systematic evaluation of Retrieval-Augmented Generation (RAG) models for fact-checking in the presence of conflicting evidence. To support this study, we introduce CONFACT (Conflicting Evidence for Fact-Checking), a novel dataset comprising questions paired with conflicting information from various sources. Extensive experiments reveal critical vulnerabilities in state-of-the-art RAG methods, particularly in resolving conflicts stemming from differences in media source credibility. To address these challenges, we investigate strategies to integrate media background information into both the retrieval and generation stages. Our results show that effectively incorporating source credibility significantly enhances the ability of RAG models to resolve conflicting evidence and improve fact-checking performance.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "8714",
        "title": "Direct Estimation of Attenuation Information from Sinograms for Positron Emission Tomography Reconstruction",
        "authors": "Prabath Hetti Mudiyanselage, Ruwan Tennakoon, John Thangarajah, Robert Ware, Jason Callahan",
        "abstract": "Positron Emission Tomography (PET) is a powerful imaging modality for assessing biochemical processes within the body. However, accurate image reconstruction is challenged by photon attenuation, particularly in dense structures such as bones, leading to quantification errors and reduced diagnostic confidence. Computed Tomography (CT) based attenuation correction is the standard approach but introduces additional radiation exposure, longer imaging times, and patient inconvenience, as well as potential registration errors, motion artifacts, and energy scaling inaccuracies.  \r\nIn this study, we propose a 3D U-Net based deep learning framework that directly estimates attenuation information from PET sinograms, eliminating the need for additional imaging modalities. Our approach integrates PET physics and employs custom skip connections to enhance cross-domain learning. We evaluate our model on a simulated brain dataset derived from real patient templates, achieving a Dice coefficient of 0.650 and an accuracy of 0.486 for bone structures. The clinical applicability of our method is further assessed by reconstructing PET images with the generated attenuation maps, yielding an MSE of 0.007 and an SSIM of 0.956, demonstrating strong structural consistency with CT-based attenuation correction. These results highlight the feasibility of performing PET image attenuation correction using PET sinograms alone, offering a promising alternative that reduces imaging time, radiation exposure, and patient burden while enabling faster and more efficient PET reconstruction.",
        "location": "Montreal",
        "day": "August 20th",
        "hour": "14:00",
        "session": "AI for Social Good (3\/8)",
        "poster_positions": "From board n98 to board n104"
    },
    {
        "id": "7716",
        "title": "Generative Agents for Multimodal Controversy Detection",
        "authors": "Tianjiao Xu, Jinfei Gao, Keyi Kong, Jianhua Yin, Tian Gan, Liqiang Nie",
        "abstract": "Multimodal controversy detection, which involves determining whether a given video and its associated comments are controversial, plays a pivotal role in risk management on social video platforms. Existing methods typically provide only classification results, failing to identify what aspects are controversial and why, thereby lacking detailed explanations. To address this limitation, we propose a novel Agent-based Multimodal Controversy Detection architecture, termed AgentMCD. This architecture leverages Large Language Models (LLMs) as generative agents to simulate human behavior and improve explainability. AgentMCD employs a multi-aspect reasoning process, where multiple judges conduct evaluations from diverse perspectives to derive a final decision. Furthermore, a multi-agent simulation process is incorporated, wherein agents act as audiences, offering opinions and engaging in free discussions after watching videos. This hybrid framework enables comprehensive controversy evaluation and significantly enhances explainability. Experiments conducted on the MMCD dataset demonstrate that our proposed architecture outperforms existing LLM-based baselines in both high-resource and low-resource comment scenarios, while maintaining superior explainability.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "9177",
        "title": "Detecting Illicit Massage Businesses by Leveraging Graph Machine Learning",
        "authors": "Vasuki Garg, Osman Y. Özaltın, Maria E. Mayorga, Sherrie Bosisto",
        "abstract": "Thousands of Illicit Massage Businesses (IMBs) are estimated to be operating in the United States by disguising themselves as legitimate establishments while exploiting trafficked workers, harming both the victims and the massage industry. The increasing digital presence of these illicit businesses presents an opportunity for detection, a crucial task for law enforcement and social service agencies aiming to disrupt their operations. Our research leverages user-generated business reviews from Yelp.com, enriched with data from multiple sources, including RubMaps.ch, U.S. Census records, GIS data, and licensing information. We present a feasibility study of developing a graph convolutional network (GCN) for a novel application and exploring its benefits and drawbacks in identifying IMBs. The novelty of our approach lies in its ability to link and analyze businesses, reviews, and reviewers within a heterogeneous network and employ a relational GCN to capture their complex relationships.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "15:00",
        "session": "AI for Social Good (6\/8)",
        "poster_positions": "From board n112 to board n117"
    },
    {
        "id": "9198",
        "title": "DECASTE: Unveiling Caste Stereotypes in Large Language Models Through Multi-Dimensional Bias Analysis",
        "authors": "Prashanth Vijayaraghavan, Soroush Vosoughi, Lamogha Chiazor, Raya Horesh, Rogerio Abreu de Paula, Ehsan Degan, Vandana Mukherjee",
        "abstract": "Recent advancements in large language models (LLMs) have revolutionized natural language processing (NLP) and expanded their applications across diverse domains. However, despite their impressive capabilities, LLMs have been shown to reflect and perpetuate harmful societal biases, including those based on ethnicity, gender, and religion. A critical and underexplored issue is the reinforcement of caste-based biases, particularly towards India's marginalized caste groups such as Dalits and Shudras. In this paper, we address this gap by proposing DECASTE, a novel, multi-dimensional framework designed to detect and assess both implicit and explicit caste biases in LLMs. Our approach evaluates caste fairness across four dimensions: socio-cultural, economic, educational, and political, using a range of customized prompting strategies. By benchmarking several state-of-the-art LLMs, we reveal that these models systematically reinforce caste biases, with significant disparities observed in the treatment of oppressed versus dominant caste groups. For example, bias scores are notably elevated when comparing Dalits and Shudras with dominant caste groups, reflecting societal prejudices that persist in model outputs. These results expose the subtle yet pervasive caste biases in LLMs and emphasize the need for more comprehensive and inclusive bias evaluation methodologies that assess the potential risks of deploying such models in real-world contexts.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "10:00",
        "session": "AI for Social Good (4\/8)",
        "poster_positions": "From board n118 to board n120"
    },
    {
        "id": "8659",
        "title": "Mat-Instructions: A Large-Scale Inorganic Material Instruction Dataset for Large Language Models",
        "authors": "Ke Liu, Shangde Gao, Yichao Fu, Xiaoliang Wu, Shuo Tong, Ajitha Rajan, Hao Xu",
        "abstract": "Recent advancements in large language models (LLMs) have revolutionized research discovery across various scientific disciplines, including materials science. The discovery of novel materials, particularly crystal materials, is essential for achieving sustainable development goals (SDGs), as they drive breakthroughs in climate change mitigation, clean and affordable energy, and the promotion of industrial innovation. However, unlocking the full potential of LLMs in materials research remains challenging due to the lack of high-quality, diverse, and instruction-based datasets. Such datasets are crucial for guiding these models in understanding and predicting the structure, property, and function of materials across various tasks. To address this limitation, we introduce Mat-Instruction, a large-scale inorganic material instruction dataset, specifically designed to unlock the potential of LLMs in materials science. Extensive experiments on fine-tuning LLaMA with our Mat-Instruction dataset demonstrate its effectiveness in advancing progress for materials science. The code and dataset are available at https:\/\/github.com\/zjuKeLiu\/Mat-Instructions",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "2260",
        "title": "Automating Intervention Discovery from Scientific Literature: A Progressive Ontology Prompting and Dual-LLM Framework",
        "authors": "Yuting Hu, Dancheng Liu, Qingyun Wang, Charles Yu, Chenhui Xu, Qingxiao Zheng, Heng Ji, Jinjun Xiong",
        "abstract": "Identifying effective interventions from the scientific literature is challenging due to the high volume of publications, specialized terminology, and inconsistent reporting formats, making manual curation laborious and prone to oversight. To address this challenge, this paper proposes a novel framework leveraging large language models (LLMs), which integrates a progressive ontology prompting (POP) algorithm with a dual-agent system, named LLM-Duo. On the one hand, the POP algorithm conducts a prioritized breadth-first search (BFS) across a predefined ontology, generating structured prompt templates and action sequences to guide the automatic annotation process. On the other hand, the LLM-Duo system features two specialized LLM agents, an explorer and an evaluator, working collaboratively and adversarially to continuously refine annotation quality. We showcase the real-world applicability of our framework through a case study focused on speech-language intervention discovery. Experimental results show that our approach surpasses advanced baselines, achieving more accurate and comprehensive annotations through a fully automated process. Our approach successfully identified 2,421 interventions from a corpus of 64,177 research articles in the speech-language pathology domain, culminating in the creation of a publicly accessible intervention knowledge base with great potential to benefit the speech-language pathology community.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "15:00",
        "session": "AI for Social Good (2\/8)",
        "poster_positions": "From board n84 to board n89"
    },
    {
        "id": "9048",
        "title": "Towards a Bipartisan Understanding of Peace and Vicarious Interactions",
        "authors": "Arka Dutta, Syed Mohammad Sualeh Ali, Usman Naseem, Ashiqur R. KhudaBukhsh",
        "abstract": "Human input plays a critical role in modern AI systems. As machines take on increasingly nuanced tasks, it becomes essential for the community to embrace subjectivity and diverse perspectives. However, research on sensitive topics often fails to incorporate diverse and balanced perspectives. This paper makes a key contribution to participatory AI design in the context of conflicts between nuclear adversaries (India and Pakistan); where disagreement between stakeholders is anticipated. The paper explores the notion of hope speech detection -- detecting de-escalating content in the context of nuclear adversaries on the brink of war -- through the lens of participatory AI design and vicarious interactions. We release a dataset of 10,081 social web posts annotated by raters from India and Pakistan and examine the bipartisan nature of the language of de-escalation. Our study reveals that vicarious perspectives can be useful for modeling out-group preferences.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "15:00",
        "session": "AI for Social Good (2\/8)",
        "poster_positions": "From board n84 to board n89"
    },
    {
        "id": "8662",
        "title": "Towards the 30 by 30 Kunming-Montreal Global Biodiversity Framework Target: Optimising Graph Connectivity in Constraint-Based Spatial Planning",
        "authors": "Sulian Le Bozec-Chiffoleau, Dimitri Justeau-Allaire, Xavier Lorca, Charles Prud'homme, Gilles Simonin, Philippe Vismara, Philippe Birnbaum, Nicolas Rinck, Nicolas Beldiceanu",
        "abstract": "The Kunming-Montreal Global Biodiversity Framework aims to protect 30% of terrestrial, inland water, marine, and coastal ecosystems worldwide, and ensuring that at least 30% of these areas are under effective restoration by 2030.\r\nMaintaining and restoring ecological connectivity between natural habitats and protected areas is a key feature of this target.\r\nAchieving it will require effective and inclusive spatial planning supported by appropriate decision-support tools.\r\nMost spatial planning models address budget as an objective and connectivity as a constraint, formulating problems with Steiner trees.\r\nIn many real-world cases, such as landscape-scale restoration planning, this formulation is inappropriate when environmental managers seek to optimise connectivity under a budget constraint.\r\nThis problem was previously addressed with Constraint Programming (CP) and graph variables, but the current approach is severely limited in terms of spatial resolution.\r\nIn this article, we formalise this problem as the budget-constrained graph connectivity optimisation problem. Based on a real case study: the restoration of forest connectivity in New Caledonia, we illustrate why ``naive'' CP approaches are inefficient.\r\nIn response, we provide a preprocessing method based on Hanan grids which preserves the existence of at least one optimal solution.\r\nFinally, we assess the efficiency of our approach in the New Caledonian case study.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "11:30",
        "session": "AI for Social Good (5\/8)",
        "poster_positions": "From board n121 to board n123"
    },
    {
        "id": "8944",
        "title": "Recommender Systems for Democracy: Toward Adversarial Robustness in Voting Advice Applications",
        "authors": "Frédéric Berdoz, Dustin Brunner, Yann Vonlanthen, Roger Wattenhofer",
        "abstract": "Voting advice applications (VAAs) help millions of voters understand which political parties or candidates best align with their views. This paper explores the potential risks these applications pose to the democratic process when targeted by adversarial entities. In particular, we expose 11 manipulation strategies and measure their impact using data from Switzerland’s primary VAA, Smartvote, collected during the last two national elections. We find that altering application parameters, such as the matching method, can shift a party’s recommendation frequency by up to 105%. Cherry-picking questionnaire items can increase party recommendation frequency by over 261%, while subtle changes to parties’ or candidates’ responses can lead to a 248% increase. To address these vulnerabilities, we propose adversarial robustness properties VAAs should satisfy, introduce empirical metrics for assessing the resilience of various matching methods, and suggest possible avenues for research toward mitigating the effect of manipulation. Our framework is key to ensuring secure and reliable AI-based VAAs poised to emerge in the near future.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "15:00",
        "session": "AI for Social Good (6\/8)",
        "poster_positions": "From board n112 to board n117"
    },
    {
        "id": "8413",
        "title": "ContextAware: A Multi-Agent Framework for Detecting Harmful Image-Based Comments on Social Media",
        "authors": "Zheng Wei, Mingchen Li, Pu Zhang, Xinyu Liu, Huamin Qu, Pan Hui",
        "abstract": "Detecting hidden stigmatization in social media poses significant challenges due to semantic misalignments between textual and visual modalities, as well as the subtlety of implicit stigmatization. Traditional approaches often fail to capture these complexities in real-world, multimodal content. To address this gap, we introduce ContextAware, an agent-based framework that leverages specialized modules to collaboratively process and analyze images, textual context, and social interactions. Our approach begins by clustering image embeddings to identify recurring content, activating high-likes agents for deeper analysis of images receiving substantial user engagement, while comprehensive agents handle lower-engagement images. By integrating case-based learning, textual sentiment, and vision-language models (VLMs), ContextAware refines its detection of harmful content. We evaluate ContextAware on a self-collected Douyin dataset focused on interracial relationships, comprising 871 short videos and 885,502 comments—of which a notable portion are image-based. Experimental results show that ContextAware not only outperforms state-of-the-art methods in accuracy and F1 score but also effectively detects implicit stigmatization within the highly contextual environment of social media. Our findings underscore the importance of agent-based architectures and multimodal alignment in capturing nuanced, culturally specific forms of harmful content.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "8559",
        "title": "AI-Assisted Triage and Decision Support of Head and Neck Cancer Screening and Diagnosis in Low-Resourced Settings",
        "authors": "Min Hun Lee, Sean Shao Wei Lam, Shaun Xin Hong Liew, Michael Dorosan, Nicholas Graves, Jonas Karlström, Hiang Khoon Tan, Walter Tsong Lee",
        "abstract": "The mortality burden of head and neck cancer (HNC) is increasing globally and disproportionately affects people in low-and middle-income countries with limited medical workforce. To address this issue, artificial intelligence (AI) algorithms are increasingly being explored to process medical imaging data, demonstrating competitive performance. However, the clinical adoption of AI remains challenging as clinicians struggle to understand how complex AI works and trust it to use in practice. In addition, AI may not perform well on varying data qualities of endoscopy videos for HNC screening and diagnosis from multiple sites. \r\n\r\nIn this project, our international and interdisciplinary team will collaborate with clinicians from multiple sites (e.g. Singapore, the U.S., and Bangladesh) to collect a diverse, multi-site dataset. In addition, we aim to design and develop computational techniques and practices to improve collaborations between clinicians and AI for the triage and diagnosis of HNC. Specifically, these techniques include a YOLOv5-based glottis detector, a classifier of patient's status using clinical endoscopy videos, uncertainty quantification techniques, and interactive Vision Language Model-based AI explanations, which will enable clinicians to understand AI outputs and provide their inputs to improve AI. After developing our system, we will evaluate the effectiveness of these computational techniques in enabling AI-assisted point-of-care triage and decision-support for HNC, particularly in resource-limited settings.",
        "location": "Montreal",
        "day": "August 20th",
        "hour": "14:00",
        "session": "AI for Social Good (3\/8)",
        "poster_positions": "From board n98 to board n104"
    },
    {
        "id": "9039",
        "title": "An Interactive Game-based Multi-Agent AI System for Children’s Social and Emotional Development",
        "authors": "Shreya Banerjee, Soheil Saneei, Lisa Pham, Elliott Alexander Beaton, Henry Fordjour Ansah, Ben Samuel, Jenny Spicer, Huda Hammad, Amanda Stage",
        "abstract": "The earliest years of life, from birth to elementary school, are the most critical time for children's social and emotional development. Recently, schools and workplaces have become increasingly concerned with cultivating social and emotional skills, especially with the decline of face-to-face interaction and the pervasive influence of modern communications technology. This paper proposes a game in development that navigates this challenge and aims to facilitate skill development in children ages 3-10 by having them identify, understand, feel their emotions, and regulate their actions. Using an established framework for social and emotional learning, it employs multi-agent artificial intelligence-based interactive gaming that generates dynamic scenarios and adjusts learning experiences based on individual child's or group's needs over time. We discuss the various modes of this game and its target frameworks, along with ways to evaluate effectiveness in facilitating social and emotional skill development.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "11:30",
        "session": "AI for Social Good (1\/8)",
        "poster_positions": "From board n90 to board n93"
    },
    {
        "id": "9172",
        "title": "Knowledge-Informed Deep Learning for Irrigation Type Mapping from Remote Sensing",
        "authors": "Oishee Bintey Hoque, Nibir Chandra Mandal, Abhijin Adiga, Samarth Swarup, Sayjro Kossi Nouwakpo, Amanda Wilson, Madhav Marathe",
        "abstract": "Accurate mapping of irrigation methods is crucial for sustainable agricultural practices and food systems. However, existing models that rely solely on spectral features from satellite imagery are ineffective due to the complexity of agricultural landscapes and limited training data, making this a challenging problem. We present Knowledge-Informed Irrigation Mapping (KIIM), a novel Swin-Transformer based approach that uses (i) a specialized projection matrix to encode crop to irrigation probability, (ii) a spatial attention map to identify agricultural lands from non-agricultural lands, (iii) bi-directional cross-attention to focus complementary information from different modalities, and (iv) a weighted ensemble for combining predictions from images and crop information. Our experimentation on five states in the US shows up to 22.9% (IoU) improvement over baseline with a 71.4% (IoU) improvement for hard-to-classify drip irrigation. In addition, we propose a two-phase transfer learning approach to enhance cross-state irrigation mapping, achieving a 51% IoU boost in a state with limited labeled data. The ability to achieve baseline performance with only 40% of the training data highlights its efficiency, reducing the dependency on extensive manual labeling efforts and making large-scale, automated irrigation mapping more feasible and cost-effective. Code: https:\/\/github.com\/Nibir088\/KIIM",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "11:30",
        "session": "AI for Social Good (8\/8)",
        "poster_positions": "From board n49 to board n52"
    },
    {
        "id": "9041",
        "title": "What is Behind Homelessness Bias? Using LLMs and NLP to Mitigate Homelessness by Acting on Social Stigma",
        "authors": "Jonathan A. Karr Jr., Emory Smith, Matthew Hauenstein, Georgina Curto, Nitesh V. Chawla",
        "abstract": "Bias towards people experiencing homelessness (PEH) is prevalent in online spaces. This project will leverage natural language processing (NLP) and large language models (LLMs) to identify, classify, and measure bias using geolocalized data collected from X (formerly Twitter), Reddit, meeting minutes, and news media across the United States. While public opinion often refers to addictions, criminality, and high levels of welfare spending to justify bias against PEH, we will conduct a comparative study to determine whether racial fractionalization is associated with homelessness bias. The results of the study aim to provide a new path to alleviate homelessness by unveiling the intersectional bias that affects PEH and minority racial groups. During the course of the project, we will deliver a lexicon, compile an annotated database for homelessness and homelessness-racism intersectional (HRI) bias, evaluate LLMs as classifiers of homelessness and HRI bias, develop homelessness and HRI bias metrics, and audit existing LLMs on HRI. In collaboration with non-profits and the city council of South Bend, Indiana, USA, our ultimate goal is to contribute to homelessness alleviation by counteracting social stigma, restoring the dignity and well-being of the persons affected.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "15:00",
        "session": "AI for Social Good (2\/8)",
        "poster_positions": "From board n84 to board n89"
    },
    {
        "id": "8772",
        "title": "Hazard Function Guided Agent-Based Models: A Case Study of Return Migration from Poland to Ukraine",
        "authors": "Zakaria Mehrab, S.S. Ravi, Logan Stundal, Samarth Swarup, Srini Venkatramanan, Bryan Lewis, Henning Mortveit, David Leblang, Madhav V. Marathe",
        "abstract": "The Russian invasion of Ukraine in February 2022 has led to the largest forced migration crisis in Europe since World War II, with millions displaced both internally and internationally. Among the displaced, approximately 4.2 million individuals have returned, highlighting the significance of return migration as a critical phase in the migration continuum. Existing studies on return migration are limited in scope, relying on survey-based approaches that suffer from demographic bias, lack of validation against ground truth, and inability to account for uncertainty. We propose a novel computational framework for modeling the return of conflict-induced migrants, using agent-based models (ABMs) and their surrogates. These models are grounded in hazard functions and account for sociopolitical contexts. Our proposed ABMs outperform baseline methods in estimating return migration from Poland to Ukraine by at least 42% and by as much as 57% in terms of normalized root mean squared error (NRMSE). Further, to illustrate the utility of such models for policymakers, we conduct two case studies that estimate the duration of displacement and characterize the demographic breakdown among the returnees.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "11:30",
        "session": "AI for Social Good (1\/8)",
        "poster_positions": "From board n90 to board n93"
    },
    {
        "id": "9173",
        "title": "Leveraging Artificial Intelligence to Bridge Gaps in Pediatric Oncology Care for Marginalized Spanish-Speaking Communities",
        "authors": "Grigorii Khvatskii, Angelica Garcia Martinez, Deng Pan, Matthew Belcher, Gerónimo Medrano Loera, Dayana Pineda Pérez, Juan Emmanuel Ferrari Muñoz-Ledo, Horacio Márquez-González, Nuno Moniz, Nitesh V. Chawla",
        "abstract": "In low-and middle-income countries (LMICs) pediatric cancer patients and their caregivers often suffer from effects of underfunded, fragmented and outdated healthcare systems. One of these effects is a breakdown of communication between hospital staff and caregivers, which is felt stronger among vulnerable populations. Our proposed solution integrates Large Language Models (LLM) and Automatic Speech Recognition (ASR) technologies to enhance communication between caregivers and healthcare providers while integrating community feedback. We combine cutting-edge technology with existing hospital infrastructure to allow for easy deployment and testing. The system will improve access to health, nutrition, and parental care programs, prioritizing caregiver engagement and real-time interaction. Ultimately, our system will pave the way to more equitable access to medical care, and address structural barriers affecting marginalized communities.",
        "location": "Montreal",
        "day": "August 19th",
        "hour": "15:00",
        "session": "AI for Social Good (2\/8)",
        "poster_positions": "From board n84 to board n89"
    },
    {
        "id": "8896",
        "title": "REVEAL: Multi-turn Evaluation of Image-Input Harms for Vision LLMs",
        "authors": "Madhur Jindal, Saurabh Deshpande",
        "abstract": "Vision Large Language Models (VLLMs) represent a significant advancement in artificial intelligence by integrating image-processing capabilities with textual understanding, thereby enhancing user interactions and expanding application domains. However, their increased complexity introduces novel safety and ethical challenges, particularly in multi-modal and multi-turn conversations. Traditional safety evaluation frameworks, designed for text-based, single-turn interactions, are inadequate for addressing these complexities. To bridge this gap, we introduce the REVEAL (Responsible Evaluation of Vision-Enabled AI LLMs) Framework, a scalable and automated pipeline for evaluating image-input harms in VLLMs. REVEAL includes automated image mining, synthetic adversarial data generation, multi-turn conversational expansion using crescendo attack strategies, and comprehensive harm assessment through evaluators like GPT-4o.\r\n\r\nWe extensively evaluated five state-of-the-art VLLMs, GPT-4o, Llama-3.2, Qwen2-VL, Phi3.5V, and Pixtral, across three important harm categories: sexual harm, violence, and misinformation. Our findings reveal that multi-turn interactions result in significantly higher defect rates compared to single-turn evaluations, highlighting deeper vulnerabilities in VLLMs. Notably, GPT-4o demonstrated the most balanced performance as measured by our Safety-Usability Index (SUI) followed closely by Pixtral. Additionally, misinformation emerged as a critical area requiring enhanced contextual defenses. Llama-3.2 exhibited the highest MT defect rate (16.55%) while Qwen2-VL showed the highest MT refusal rate (19.1%).",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "8968",
        "title": "AI Diagnostic Assistant (AIDA): A Predictive Model for Diagnoses from Health Records in Clinical Decision Support Systems",
        "authors": "Dmitriy Umerenkov, Alexandr Nesterov, Vladimir Shaposhnikov, Ruslan Abramov, Nikolay Romanenko, Vladimir Kokh, Marina Kirina, Anton Abrosimov, Dmitry V. Dylov, Ivan Oseledets",
        "abstract": "Clinical Decision Support Systems (CDSS) play an increasingly important role in medical diagnostics. We present AI Diagnostic Assistant (AIDA), a real-time predictive model designed to assist doctors in interpreting patient conditions. AIDA analyzes electronic health records (EHR), including medical history, laboratory results, and complaints, to suggest potential diagnoses from 95 common conditions before the doctor makes the final decision. The model acts as a verification and backup tool, ensuring that no critical details are overlooked. Trained on 1.5 million patient records and validated on a dataset curated by a panel of experts, AIDA proves trustworthy as a diagnosis-making assistant (87.7% accuracy compared to 91.7% accuracy among doctors).\r\n\r\nIntegrated into a megapolis-wide CDSS, AIDA has assisted doctors in over 3 million real-world diagnoses to date.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "4538",
        "title": "LLM-based Collaborative Agents with Pedagogy-guided Interaction Modeling for Timely Instructive Feedback Generation in Task-oriented Group Discussions",
        "authors": "Qihao Yang, Yu Yang, Sixu An, Tianyong Hao, Guandong Xu",
        "abstract": "Large language models (LLMs) fundamentally reshape learning and teaching models, shifting tutoring systems from supporting individual learning to facilitating collaborative learning (CL) like task-oriented group discussions. However, existing AI tutors struggle to guide CL, as they seldom model the interactions between AI tutors and students. Therefore, they cannot scaffold students to complete tasks collaboratively, which impairs learning outcomes and pedagogy adaptability. Additionally, existing AI tutors fail to make use of CL theories to generate instructive feedback, which leads to undesirable interactions such as over-instruction and limits students' autonomy. In this paper, we propose an LLM-based collaborative agent that innovatively leverages pedagogical strategies to sense discussion stages, detect learning issues, identify the timing of intervention, and generate instructive feedback. To develop the agent, we first design a prompting strategy based on a CL theory, that is, the Community of Inquiry, to cultivate the agent to understand the discussion status. Second, a multi-agent interaction framework is proposed to simulate the collaborative learning behavior between AI tutors and students. Meanwhile, a synthetic task-oriented group discussion dataset, namely CLTeach, is generated, which consists of 27k manually-verified multi-party dialogues with fine-grained annotations of instructive feedback and explanations. Lastly, we use CLTeach to fine-tune the LLM agent, ultimately enabling it to generate instructive feedback at the right time to support students in CL. Extensive experiments demonstrate that our agent achieves state-of-the-art performance in feedback generation and has the potential to mimic human teachers effectively.",
        "location": "Guangzhou",
        "day": "August 31st",
        "hour": "11:00",
        "session": "AI and social good (2\/2)"
    },
    {
        "id": "9097",
        "title": "SAHAY: Multimodal, Privacy-Preserving AI for Suicide Risk Detection and Intervention in India",
        "authors": "Salam Michael Singh, Manik Inder Singh Sethi, Suresh Bada Math, Tanmoy Chakraborty",
        "abstract": "Suicide accounts for one of the leading causes of death in India, with over 164,033 deaths reported in 2021. Despite increased awareness, the gap between the need for consistent treatment and actual accessibility remains a challenge due to limited mental health infrastructure, the stigma surrounding mental illness in society, and the lack of real-time detection mechanisms. Traditional suicide risk assessments often miss early signs of distress, which rely heavily on clinical evaluations and self-reporting. Although AI-based monitoring seems promising, currently available models focus only on risk prediction without intervention and treatment, leaving a critical gap in tackling crisis management. In this proposal, we strive to design SAHAY, the first-of-its-kind AI-based, suicide prevention framework that seamlessly couples prediction with prevention and treatment access. Leveraging multimodal data, including the social media text and Electronic Health Records (EHR) and Ecological Momentary Assessments (EMA) such as wearable physiological data, SAHAY aims to assess suicide risk dynamically. Unlike existing models, SAHAY is culturally adaptive, multilingual and seamlessly integrates with India’s TeleMANAS mental health support system, to provide structured AI-human collaboration for long-term care and crisis interventions. It will be an adaptable, scalable, modular, and plug-and-play solution based on the Digital Public Infrastructure principle. Additionally, we intend to incorporate AI-driven geo-spatial crisis mapping to identify suicide hotspots in underserved regions. By combining real-time multimodal risk detection, professional mental health intervention, and geo-spatial outreach, SAHAY represents a scalable, adaptable, and end-to-end suicide prevention system. The design of SAHAY aligns with UN Sustainable Development Goals (SDGs) 3, 4, 5, 10, and 17, promoting inclusive, accessible, and data-driven mental healthcare.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "10:00",
        "session": "AI for Social Good (7\/8)",
        "poster_positions": "From board n46 to board n48"
    },
    {
        "id": "9079",
        "title": "What is Beneath Misogyny: Misogynous Memes Classification and Explanation",
        "authors": "Kushal Kanwar, Dushyant Singh Chauhan, Gopendra Vikram Singh, Asif Ekbal",
        "abstract": "Memes are popular in the modern world and are distributed primarily for entertainment. However, harmful ideologies such as misogyny can be propagated through innocent-looking memes. The detection and understanding of why a meme is misogynous is a research challenge due to its multimodal nature (image and text) and its nuanced manifestations across different societal contexts. We introduce a novel multimodal approach, namely, MM-Misogyny to detect, categorize, and explain misogynistic content in memes. MM-Misogyny processes text and image modalities separately and unifies them into a multimodal context through a cross-attention mechanism. The resulting multimodal context is then easily processed for labeling, categorization, and explanation via a classifier and Large Language Model (LLM). The evaluation of the proposed model is performed on a newly curated dataset (What’s Beneath Misogynous Stereotyping (WBMS)) created by collecting misogynous memes from cyberspace and categorizing them into four categories, namely, Kitchen, Leadership, Working, and Shopping. The model not only detects and classifies misogyny, but also provides a granular understanding of how misogyny operates in operates in domains of life. The results demonstrate the superiority of our approach compared to existing methods. The code and dataset are available at https:\/\/github.com\/Misogyny.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    },
    {
        "id": "9133",
        "title": "SHIELD: A Self-supervised, Silicosis-focused Hierarchical Imaging Framework for Occupational Lung Disease Diagnosis",
        "authors": "Yasmeena Akhter, Rishabh Ranjan, Richa Singh, Mayank Vatsa",
        "abstract": "Silicosis is an irreversible lung disease caused by silica dust exposure in industrial settings. Early detection is crucial, but automatic diagnostic methods are hindered by limited data availability. We propose SHIELD - a self-supervised, Silicosis-focused Hierarchical Imaging framework for early occupational Lung disease Diagnosis. Our method leverages a multi-resolution jigsaw puzzle pretext task on CXR images to extract and preserve features for lung region analysis. By employing a pyramidal strategy to generate pretrained models at various resolutions, followed by fine-tuning and a two-level ensembling across diverse deep learning architectures, SHIELD achieves enhanced diagnostic accuracy. We validate our approach on a publicly collected CXR dataset of 3044 samples from public health centers in India. SHIELD achieves 72% accuracy, demonstrating up to 20% improvement over baseline approaches. This work advances medical image analysis and supports UN Sustainable Development Goal 3 by providing cost-effective early screening in resource-limited settings.",
        "location": "Montreal",
        "day": "August 20th",
        "hour": "14:00",
        "session": "AI for Social Good (3\/8)",
        "poster_positions": "From board n98 to board n104"
    },
    {
        "id": "9119",
        "title": "MCloudNet: An Ultra-Short-Term Photovoltaic Power Forecasting Framework With Multi-Layer Cloud Coverage",
        "authors": "Meng Wan, Tiantian Liu, Yuxuan Bi, Jue Wang, Hui Cui, Rongqiang Cao, Jiaxiang Wang, Peng Shi, Ningming Nie, Yangang Wang",
        "abstract": "Over 4.15 million low-income households across nearly 60,000 villages in China benefit from photovoltaic (PV) poverty alleviation power stations. However, weak infrastructure and limited capabilities make these systems vulnerable to fluctuations. One of the United Nations' Sustainable Development Goals (SDG 7) seeks to ensure access to affordable and reliable energy for all, especially in underdeveloped regions. This paper proposes MCloudNet, a multi-modal framework designed to improve ultra-short-term PV prediction in data-scarce, cloud-dynamic environments. MCloudNet explicitly models multi-layer cloud structures from satellite imagery and fuses them with time-series meteorological data to enhance prediction accuracy and interpretability. A province-level dispatch system with MCloudNet has been deployed in Hebei, supporting scheduling across rural PV stations. Experiments conducted in counties such as Shexian and Luxi highlight the framework's effectiveness for use in underdeveloped micro-grids. Operational results show that the system has reduced over 60 million kWh of solar curtailment and generated 24 million CNY in economic value, benefiting approximately 50,000 rural households. By minimizing power fluctuations and improving rural energy scheduling, MCloudNet supports essential services such as lighting, medical facilities, and communications. The source code is available at: https:\/\/github.com\/AI4SClab\/MCloudNet.",
        "location": "Montreal",
        "day": "August 22nd",
        "hour": "11:30",
        "session": "AI for Social Good (8\/8)",
        "poster_positions": "From board n49 to board n52"
    },
    {
        "id": "9008",
        "title": "Sustainable Wearables for Health Applications and Beyond via Uncertainty-Aware Energy Management",
        "authors": "Dina Hussein, Chibuike E. Ugwu, Ganapati Bhat, Janardhan Rao Doppa",
        "abstract": "Achieving good health and well-being through lower mortality rates of non-communicable diseases and early warning of health risks are key goals of United Nations (UN). Wearable internet of things (IoT) are one of the most promising technology to achieve these goals through their ubiquitous monitoring of key health indicators and in-situ data processing. However, small form-factor of wearable devices constrains the battery capacity, thus requiring frequent recharging or battery replacements, which lowers their adoption rate and benefits. Augmentation of battery energy by scavenging ambient sources, such as light, is a promising solution to improve operating lifetime of IoT devices. However, ambient energy sources are highly uncertain, making energy management (EM) challenging. To handle these challenges, this paper presents a novel uncertainty-aware EM approach. First, we develop a conformal prediction-based method for future energy harvest (EH) that provides small uncertainty regions with provable coverage guarantees (true output vector is within the region). The EH uncertainty regions are then leveraged in an EM algorithm that uses overhead-aware sampling to evaluate the quality of multiple decisions with varying EH before making a decision using a lightweight machine learning model. Experiments on two diverse real-world datasets with 10 users show that conformal prediction achieves more than 90% coverage with tight prediction intervals; and the EM algorithm produces decisions that are, on average, within 2 Joules of an optimal Oracle.",
        "location": "Montreal",
        "day": "August 21st",
        "hour": "10:00",
        "session": "AI for Social Good (4\/8)",
        "poster_positions": "From board n118 to board n120"
    },
    {
        "id": "9121",
        "title": "Denoised Attention and Question-Augmented Representations for Knowledge Tracing",
        "authors": "Jiwei Deng, Youheng Bai, Mingliang Hou, Teng Guo, Zitao Liu, Weiqi Luo",
        "abstract": "Knowledge tracing (KT) is an essential task in online education systems. It aims to predict the future performance of students based on their historical learning interaction data. Despite significant advancements in attention-based KT models, they still face some limitations: inaccurate input representation and excessive student forgetting modeling. These limitations often lead to the attention noise problem: the model assigns non-negligible attention weight to some information that is cognitively irrelevant in nature, thereby generating interference signals. To address this problem, we propose a novel KT model, i.e., DenoiseKT. DenoiseKT effectively models the difficulty of the questions and utilizes graph neural network to capture the complex relationship between questions, thereby refining the representations of input features. Additionally, the denoised attention mechanism introduces a weight factor to reduce the model's attention weight distribution on irrelevant information. We extensively compare DenoiseKT with 22 state-of-the-art KT models on 4 widely-used public datasets. Experimental results show that DenoiseKT can effectively solve the attention noise problem and outperform other models. The source code of DenoiseKT is available at https:\/\/pykt.org.",
        "location": "Guangzhou",
        "day": "August 30th",
        "hour": "15:00",
        "session": "AI and social good (1\/2)"
    }
]