本篇博文主要展示 2024-12-18 从Arxiv.org论文网站获取的最新论文列表,自动更新,按照NLP、CV、ML、AI、IR五个大方向区分,若需要邮件定时接收,请在评论区留下你的邮箱号。

说明:每日论文数据从Arxiv.org获取,每天早上12:00左右定时自动更新。

友情提示: 如何您需要邮箱接收每日论文数据,请在评论处留下你的邮箱。

目录

概览 (2024-12-18)

今日共更新571篇论文,其中:

  • 自然语言处理126篇(Computation and Language (cs.CL))
  • 人工智能186篇(Artificial Intelligence (cs.AI))
  • 计算机视觉152篇(Computer Vision and Pattern Recognition (cs.CV))
  • 机器学习186篇(Machine Learning (cs.LG))

自然语言处理

[NLP-0] DnDScore: Decontextualization and Decomposition for Factuality Verification in Long-Form Text Generation

【速读】: 该论文试图解决大语言模型 (LLM) 生成内容验证中的分解与去上下文化之间的矛盾问题。分解策略将声明拆分为独立验证的原子事实,而去上下文化则通过增加相关信息来确保声明可以在原始上下文之外进行验证。这两种策略的冲突在于,分解追求原子事实的隔离,而去上下文化则引入了额外的信息,导致验证步骤中难以确定应验证哪一部分信息。论文的关键解决方案是提出了DnDScore方法,该方法在验证子声明时考虑了上下文信息,从而在分解和去上下文化之间实现了平衡,提高了验证的准确性。

链接: https://arxiv.org/abs/2412.13175
作者: Miriam Wanner,Benjamin Van Durme,Mark Dredze
机构: Johns Hopkins University (约翰霍普金斯大学)
关键词: Large Language Model, Language Model, Large Language, generations decomposes claims, generations decomposes
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:The decompose-then-verify strategy for verification of Large Language Model (LLM) generations decomposes claims that are then independently verified. Decontextualization augments text (claims) to ensure it can be verified outside of the original context, enabling reliable verification. While decomposition and decontextualization have been explored independently, their interactions in a complete system have not been investigated. Their conflicting purposes can create tensions: decomposition isolates atomic facts while decontextualization inserts relevant information. Furthermore, a decontextualized subclaim presents a challenge to the verification step: what part of the augmented text should be verified as it now contains multiple atomic facts? We conduct an evaluation of different decomposition, decontextualization, and verification strategies and find that the choice of strategy matters in the resulting factuality scores. Additionally, we introduce DnDScore, a decontextualization aware verification method which validates subclaims in the context of contextual information.
zh

[NLP-1] Compressed Chain of Thought: Efficient Reasoning Through Dense Representations

【速读】: 该论文试图解决在推理任务中,使用链式思维解码(Chain-of-Thought, CoT)时生成延迟高的问题。解决方案的关键在于提出了压缩链式思维(Compressed Chain-of-Thought, CCoT)框架,通过生成内容丰富且连续的可变长度沉思标记(contemplation tokens)来替代传统的固定长度离散嵌入。这些沉思标记是显式推理链的压缩表示,能够应用于现成的解码器语言模型,并通过控制生成的沉思标记数量来动态调整推理性能,从而在提高准确性的同时减少生成延迟。

链接: https://arxiv.org/abs/2412.13171
作者: Jeffrey Cheng,Benjamin Van Durme
机构: 未知
关键词: high generation latency, improve reasoning performance, contemplation tokens, cost of high, high generation
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Chain-of-thought (CoT) decoding enables language models to improve reasoning performance at the cost of high generation latency in decoding. Recent proposals have explored variants of contemplation tokens, a term we introduce that refers to special tokens used during inference to allow for extra computation. Prior work has considered fixed-length sequences drawn from a discrete set of embeddings as contemplation tokens. Here we propose Compressed Chain-of-Thought (CCoT), a framework to generate contentful and continuous contemplation tokens of variable sequence length. The generated contemplation tokens are compressed representations of explicit reasoning chains, and our method can be applied to off-the-shelf decoder language models. Through experiments, we illustrate how CCoT enables additional reasoning over dense contentful representations to achieve corresponding improvements in accuracy. Moreover, the reasoning improvements can be adaptively modified on demand by controlling the number of contemplation tokens generated.
zh

[NLP-2] Algorithmic Fidelity of Large Language Models in Generating Synthetic German Public Opinions: A Case Study

【速读】: 该论文试图解决大语言模型(LLMs)在模拟公众意见时的算法忠实性问题,即模型能否准确反映人类参与者的社会文化背景和细微意见差异。解决方案的关键在于通过将人口统计特征融入角色提示(persona prompts),使LLMs生成能够代表德国不同亚群体的合成公众意见。研究结果表明,Llama模型在代表亚群体方面表现优于其他LLMs,尤其是在意见多样性较低的群体中。此外,模型在模拟左翼政党支持者的意见时表现更好,而与右翼政党AfD的支持者匹配度最低。论文强调了在提示中包含或排除特定变量对模型预测的显著影响,并指出通过调整LLMs以减少政治偏见并增强代表性的稳健性,可以更有效地模拟多样化的公众意见。

链接: https://arxiv.org/abs/2412.13169
作者: Bolei Ma,Berk Yoztyurk,Anna-Carolina Haensch,Xinpeng Wang,Markus Herklotz,Frauke Kreuter,Barbara Plank,Matthias Assenmacher
机构: LMU Munich(慕尼黑大学); Munich Center for Machine Learning(慕尼黑机器学习中心); University of Maryland, College Park(马里兰大学帕克分校)
关键词: large language models, German Longitudinal Election, Longitudinal Election Studies, investigate public opinions, recent research
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:In recent research, large language models (LLMs) have been increasingly used to investigate public opinions. This study investigates the algorithmic fidelity of LLMs, i.e., the ability to replicate the socio-cultural context and nuanced opinions of human participants. Using open-ended survey data from the German Longitudinal Election Studies (GLES), we prompt different LLMs to generate synthetic public opinions reflective of German subpopulations by incorporating demographic features into the persona prompts. Our results show that Llama performs better than other LLMs at representing subpopulations, particularly when there is lower opinion diversity within those groups. Our findings further reveal that the LLM performs better for supporters of left-leaning parties like The Greens and The Left compared to other parties, and matches the least with the right-party AfD. Additionally, the inclusion or exclusion of specific variables in the prompts can significantly impact the models’ predictions. These findings underscore the importance of aligning LLMs to more effectively model diverse public opinions while minimizing political biases and enhancing robustness in representativeness.
zh

[NLP-3] BanglishRev: A Large-Scale Bangla-English and Code-mixed Dataset of Product Reviews in E-Commerce

【速读】: 该论文旨在解决针对孟加拉语、英语及其混合语言(Banglish)的电子商务产品评论数据集的缺乏问题,并验证其在情感分析任务中的有效性。解决方案的关键在于构建了目前最大的BanglishRev数据集,包含1.74百万条评论和3.2百万条评分信息,涵盖128,000种产品。通过训练BanglishBERT模型,并利用评论评分作为情感标签(评分大于3为正面,小于等于3为负面),模型在手动标注的混合语言评论数据集上取得了94%的准确率和0.94的F1分数,证明了该数据集在情感分析任务中的高效性。

链接: https://arxiv.org/abs/2412.13161
作者: Mohammad Nazmush Shamael,Sabila Nawshin,Swakkhar Shatabda,Salekul Islam
机构: United International University; Indiana University Bloomington; BRAC University; North South University
关键词: Bengali words written, largest e-commerce product, e-commerce reviews written, Bengali words, product review dataset
类目: Computation and Language (cs.CL); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:This work presents the BanglishRev Dataset, the largest e-commerce product review dataset to date for reviews written in Bengali, English, a mixture of both and Banglish, Bengali words written with English alphabets. The dataset comprises of 1.74 million written reviews from 3.2 million ratings information collected from a total of 128k products being sold in online e-commerce platforms targeting the Bengali population. It includes an extensive array of related metadata for each of the reviews including the rating given by the reviewer, date the review was posted and date of purchase, number of likes, dislikes, response from the seller, images associated with the review etc. With sentiment analysis being the most prominent usage of review datasets, experimentation with a binary sentiment analysis model with the review rating serving as an indicator of positive or negative sentiment was conducted to evaluate the effectiveness of the large amount of data presented in BanglishRev for sentiment analysis tasks. A BanglishBERT model is trained on the data from BanglishRev with reviews being considered labeled positive if the rating is greater than 3 and negative if the rating is less than or equal to 3. The model is evaluated by being testing against a previously published manually annotated dataset for e-commerce reviews written in a mixture of Bangla, English and Banglish. The experimental model achieved an exceptional accuracy of 94% and F1 score of 0.94, demonstrating the dataset’s efficacy for sentiment analysis. Some of the intriguing patterns and observations seen within the dataset and future research directions where the dataset can be utilized is also discussed and explored. The dataset can be accessed through this https URL.
zh

[NLP-4] Are Your LLM s Capable of Stable Reasoning?

【速读】: 该论文试图解决当前大型语言模型 (Large Language Models, LLMs) 在复杂推理任务中表现与实际应用之间存在的显著差距问题。解决方案的关键在于提出了两个创新点:首先,引入了一种新的评估指标 G-Pass@k,该指标通过连续评估模型在多次采样中的表现,量化了模型的峰值性能潜力及其稳定性;其次,开发了动态基准 LiveMathBench,该基准包含具有挑战性的现代数学问题,旨在减少评估过程中的数据泄露风险。通过在最新的 LLMs 上使用 G-Pass@k 进行广泛实验,论文提供了关于模型最大能力和操作一致性的全面见解,揭示了 LLMs 在“现实”推理能力方面仍有显著改进空间,强调了更强大评估方法的必要性。

链接: https://arxiv.org/abs/2412.13147
作者: Junnan Liu,Hongwei Liu,Linchen Xiao,Ziyi Wang,Kuikun Liu,Songyang Gao,Wenwei Zhang,Songyang Zhang,Kai Chen
机构: Shanghai AI Laboratory(上海人工智能实验室)
关键词: Large Language Models, Large Language, demonstrated remarkable progress, advancement of Large, complex reasoning tasks
类目: Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注: Preprint

点击查看摘要

Abstract:The rapid advancement of Large Language Models (LLMs) has demonstrated remarkable progress in complex reasoning tasks. However, a significant discrepancy persists between benchmark performances and real-world applications. We identify this gap as primarily stemming from current evaluation protocols and metrics, which inadequately capture the full spectrum of LLM capabilities, particularly in complex reasoning tasks where both accuracy and consistency are crucial. This work makes two key contributions. First, we introduce G-Pass@k, a novel evaluation metric that provides a continuous assessment of model performance across multiple sampling attempts, quantifying both the model’s peak performance potential and its stability. Second, we present LiveMathBench, a dynamic benchmark comprising challenging, contemporary mathematical problems designed to minimize data leakage risks during evaluation. Through extensive experiments using G-Pass@k on state-of-the-art LLMs with LiveMathBench, we provide comprehensive insights into both their maximum capabilities and operational consistency. Our findings reveal substantial room for improvement in LLMs’ “realistic” reasoning capabilities, highlighting the need for more robust evaluation methods. The benchmark and detailed results are available at: this https URL.
zh

[NLP-5] Syntactic Transfer to Kyrgyz Using the Treebank Translation Method WWW

【速读】: 该论文试图解决吉尔吉斯语(Kyrgyz)作为低资源语言在构建高质量句法语料库(syntactic corpora)时面临的挑战。解决方案的关键在于提出了一种基于树库翻译方法(treebank translation method)的工具,能够将土耳其语(Turkish)的句法标注迁移到吉尔吉斯语中。通过使用TueCL树库进行评估,结果表明该方法在句法标注准确性上优于仅基于吉尔吉斯语KTMU树库训练的单语模型。此外,研究还引入了一种评估手动标注复杂度的方法,以进一步优化标注过程。

链接: https://arxiv.org/abs/2412.13146
作者: Anton Alekseev,Alina Tillabaeva,Gulnara Dzh. Kabaeva,Sergey I. Nikolenko
机构: Steklov Mathematical Institute at St. Petersburg(斯捷克洛夫数学研究所圣彼得堡分所); St. Petersburg University(圣彼得堡大学); Kazan (Volga Region) Federal University(喀山(伏尔加地区)联邦大学); Kyrgyz State Technical University n. a. I. Razzakov(吉尔吉斯国立技术大学以I. Razzakov命名); ITMO University(圣彼得堡国立信息技术、机械与光学大学)
关键词: requires significant effort, high-quality syntactic corpora, create high-quality syntactic, Kyrgyz language, low-resource language
类目: Computation and Language (cs.CL)
备注: To be published in the Journal of Math. Sciences. Zapiski version (in Russian): this http URL

点击查看摘要

Abstract:The Kyrgyz language, as a low-resource language, requires significant effort to create high-quality syntactic corpora. This study proposes an approach to simplify the development process of a syntactic corpus for Kyrgyz. We present a tool for transferring syntactic annotations from Turkish to Kyrgyz based on a treebank translation method. The effectiveness of the proposed tool was evaluated using the TueCL treebank. The results demonstrate that this approach achieves higher syntactic annotation accuracy compared to a monolingual model trained on the Kyrgyz KTMU treebank. Additionally, the study introduces a method for assessing the complexity of manual annotation for the resulting syntactic trees, contributing to further optimization of the annotation process.
zh

[NLP-6] Improving Explainability of Sentence-level Metrics via Edit-level Attribution for Grammatical Error Correction

【速读】: 该论文试图解决语法错误纠正 (Grammatical Error Correction, GEC) 评估指标中缺乏可解释性的问题。现有的许多评估指标,尤其是无参考指标,无法解释其评分依据,这阻碍了研究人员分析GEC模型的优缺点,并限制了为用户提供详细反馈的能力。论文的关键解决方案是提出将句子级别的评分归因于单个编辑操作,通过使用合作博弈论中的Shapley值来计算每个编辑的贡献,从而提供对特定修正如何影响整体性能的深入见解。实验结果表明,该方法在不同编辑粒度下与现有句子级别指标具有高度一致性,并与人工评估结果达到约70%的吻合度,同时揭示了评估指标中存在的偏差,如忽略拼写编辑的趋势。

链接: https://arxiv.org/abs/2412.13110
作者: Takumi Goto,Justin Vasselli,Taro Watanabe
机构: Nara Institute of Science and Technology (奈良先端科学技术大学院大学)
关键词: Grammatical Error Correction, Grammatical Error, proposed for Grammatical, Error Correction, GEC models
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Various evaluation metrics have been proposed for Grammatical Error Correction (GEC), but many, particularly reference-free metrics, lack explainability. This lack of explainability hinders researchers from analyzing the strengths and weaknesses of GEC models and limits the ability to provide detailed feedback for users. To address this issue, we propose attributing sentence-level scores to individual edits, providing insight into how specific corrections contribute to the overall performance. For the attribution method, we use Shapley values, from cooperative game theory, to compute the contribution of each edit. Experiments with existing sentence-level metrics demonstrate high consistency across different edit granularities and show approximately 70% alignment with human evaluations. In addition, we analyze biases in the metrics based on the attribution results, revealing trends such as the tendency to ignore orthographic edits. Our implementation is available at \urlthis https URL.
zh

[NLP-7] AI PERSONA: Towards Life-long Personalization of LLM s

【速读】: 该论文试图解决大型语言模型(LLM)在长期使用中如何持续适应用户个性化需求的问题。解决方案的关键在于提出了一种简单、通用、有效且可扩展的框架,用于实现LLM系统的终身个性化(life-long personalization)。该框架使语言模型能够不断适应每个用户的多样化且不断变化的需求,提供最新的个性化帮助。此外,论文还引入了合成现实基准和鲁棒评估指标的方法,以促进未来在LLM个性化方面的研究,并计划公开所有代码和数据以支持这一领域的进一步发展。

链接: https://arxiv.org/abs/2412.13103
作者: Tiannan Wang,Meiling Tao,Ruoyu Fang,Huilin Wang,Shuai Wang,Yuchen Eleanor Jiang,Wangchunshu Zhou
机构: OPPO AI Center(OPPO AI中心); Guangdong University of Technology(广东工业大学); University of Illinois at Urbana-Champaign(伊利诺伊大学厄巴纳-香槟分校); Beihang University(北京航空航天大学); Tsinghua University(清华大学)
关键词: large language models, LLM systems, enable LLM systems, language agents, LLM
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: Work in progress

点击查看摘要

Abstract:In this work, we introduce the task of life-long personalization of large language models. While recent mainstream efforts in the LLM community mainly focus on scaling data and compute for improved capabilities of LLMs, we argue that it is also very important to enable LLM systems, or language agents, to continuously adapt to the diverse and ever-changing profiles of every distinct user and provide up-to-date personalized assistance. We provide a clear task formulation and introduce a simple, general, effective, and scalable framework for life-long personalization of LLM systems and language agents. To facilitate future research on LLM personalization, we also introduce methods to synthesize realistic benchmarks and robust evaluation metrics. We will release all codes and data for building and benchmarking life-long personalized LLM systems.
zh

[NLP-8] AIR-Bench: Automated Heterogeneous Information Retrieval Benchmark

【速读】: 该论文试图解决现有信息检索(IR)模型评估基准在应对新兴领域时成本高、效率低的问题。解决方案的关键在于提出了自动化异构信息检索基准(AIR-Bench),其核心特征包括:1) 自动化,通过大型语言模型(LLMs)自动生成测试数据,无需人工干预;2) 异构性,测试数据涵盖多样化的任务、领域和语言;3) 动态性,不断扩展涵盖的领域和语言,以提供更全面的评估基准。通过可靠且稳健的数据生成管道,AIR-Bench能够基于真实语料库自动创建多样化且高质量的评估数据集,确保其与人工标注数据的高度一致性,从而成为评估IR模型的可靠基准。

链接: https://arxiv.org/abs/2412.13102
作者: Jianlyu Chen,Nan Wang,Chaofan Li,Bo Wang,Shitao Xiao,Han Xiao,Hao Liao,Defu Lian,Zheng Liu
机构: University of Science and Technology of China; Beijing Academy of Artificial Intelligence; Jina AI; Beijing University of Posts and Telecommunications; Shenzhen University; State Key Laboratory of Cognitive Intelligence
关键词: Heterogeneous Information Retrieval, Automated Heterogeneous Information, information retrieval, Information Retrieval Benchmark, AIR-Bench
类目: Information Retrieval (cs.IR); Computation and Language (cs.CL)
备注: 31 pages, 6 figures

点击查看摘要

Abstract:Evaluation plays a crucial role in the advancement of information retrieval (IR) models. However, current benchmarks, which are based on predefined domains and human-labeled data, face limitations in addressing evaluation needs for emerging domains both cost-effectively and efficiently. To address this challenge, we propose the Automated Heterogeneous Information Retrieval Benchmark (AIR-Bench). AIR-Bench is distinguished by three key features: 1) Automated. The testing data in AIR-Bench is automatically generated by large language models (LLMs) without human intervention. 2) Heterogeneous. The testing data in AIR-Bench is generated with respect to diverse tasks, domains and languages. 3) Dynamic. The domains and languages covered by AIR-Bench are constantly augmented to provide an increasingly comprehensive evaluation benchmark for community developers. We develop a reliable and robust data generation pipeline to automatically create diverse and high-quality evaluation datasets based on real-world corpora. Our findings demonstrate that the generated testing data in AIR-Bench aligns well with human-labeled testing data, making AIR-Bench a dependable benchmark for evaluating IR models. The resources in AIR-Bench are publicly available at this https URL.
zh

[NLP-9] Uchaguzi-2022: A Dataset of Citizen Reports on the 2022 Kenyan Election COLING2025

【速读】: 该论文试图解决大规模众包信息(crowdsourced information)的系统化组织问题,特别是通过属性分类和地理标记来确保从这些数据中提取准确且有意义的洞察,以便政策制定者能够利用这些信息推动积极变革。解决方案的关键在于利用语言模型(language models)来实现报告的规模化分类和地理标记,从而减少传统上需要大量人工标注的工作量,并展示了其在社会公益AI(AI for Social Good)领域的应用潜力。

链接: https://arxiv.org/abs/2412.13098
作者: Roberto Mondini,Neema Kotonya,Robert L. Logan IV,Elizabeth M Olson,Angela Oduor Lungati,Daniel Duke Odongo,Tim Ombasa,Hemank Lamba,Aoife Cahill,Joel R. Tetreault,Alejandro Jaimes
机构: Dataminr Inc.(Dataminr公司); Ushahidi(Ushahidi)
关键词: Online reporting platforms, Online reporting, local communities, reporting platforms, platforms have enabled
类目: Computation and Language (cs.CL); Social and Information Networks (cs.SI)
备注: COLING 2025

点击查看摘要

Abstract:Online reporting platforms have enabled citizens around the world to collectively share their opinions and report in real time on events impacting their local communities. Systematically organizing (e.g., categorizing by attributes) and geotagging large amounts of crowdsourced information is crucial to ensuring that accurate and meaningful insights can be drawn from this data and used by policy makers to bring about positive change. These tasks, however, typically require extensive manual annotation efforts. In this paper we present Uchaguzi-2022, a dataset of 14k categorized and geotagged citizen reports related to the 2022 Kenyan General Election containing mentions of election-related issues such as official misconduct, vote count irregularities, and acts of violence. We use this dataset to investigate whether language models can assist in scalably categorizing and geotagging reports, thus highlighting its potential application in the AI for Social Good space.
zh

[NLP-10] LMUnit: Fine-grained Evaluation with Natural Language Unit Tests

【速读】: 该论文试图解决语言模型在关键工作流程中的行为评估问题,传统的人工评估成本高且结果不一致,而自动化指标则过于粗糙且难以解释。论文提出的解决方案是引入自然语言单元测试(natural language unit tests),通过将响应质量分解为明确的、可测试的标准,并结合统一的评分模型LMUnit,该模型整合了多目标训练、直接评分和自然语言解释。通过控制性的人类研究,该方法显著提高了评估者间的一致性,并支持更有效的语言模型开发流程。LMUnit在评估基准(如FLASK和BigGenBench)上达到了最先进的性能,并在RewardBench上取得了竞争性结果,验证了所提出范式和评分模型的有效性。

链接: https://arxiv.org/abs/2412.13091
作者: Jon Saad-Falcon,Rajan Vivek,William Berrios,Nandita Shankar Naik,Matija Franklin,Bertie Vidgen,Amanpreet Singh,Douwe Kiela,Shikib Mehri
机构: Contextual AI; Stanford University (斯坦福大学)
关键词: automated metrics provide, assessing their behavior, fundamental challenge, costly and noisy, provide only coarse
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:As language models become integral to critical workflows, assessing their behavior remains a fundamental challenge – human evaluation is costly and noisy, while automated metrics provide only coarse, difficult-to-interpret signals. We introduce natural language unit tests, a paradigm that decomposes response quality into explicit, testable criteria, along with a unified scoring model, LMUnit, which combines multi-objective training across preferences, direct ratings, and natural language rationales. Through controlled human studies, we show this paradigm significantly improves inter-annotator agreement and enables more effective LLM development workflows. LMUnit achieves state-of-the-art performance on evaluation benchmarks (FLASK, BigGenBench) and competitive results on RewardBench. These results validate both our proposed paradigm and scoring model, suggesting a promising path forward for language model evaluation and development.
zh

[NLP-11] CLASP: Contrastive Language-Speech Pretraining for Multilingual Multimodal Information Retrieval ECIR2025

【速读】: 该论文试图解决多语言、多模态的音频-文本信息检索问题,关键解决方案在于引入了一种名为CLASP (Contrastive Language-Speech Pretraining) 的多语言、多模态表示方法。CLASP通过结合语音内容和文本数据的协同作用,利用新引入的语音-文本数据集进行训练,该数据集涵盖了15个不同类别。其音频部分结合了音频频谱图与预训练的自监督语音模型,而语言编码部分则采用了一个在100多种语言上预训练的句子编码器。这种统一的轻量级模型有效地弥合了不同模态和语言之间的差距,显著提升了多语言和多模态数据的处理和检索能力。

链接: https://arxiv.org/abs/2412.13071
作者: Mohammad Mahdi Abootorabi,Ehsaneddin Asgari
机构: Qatar Computing Research Institute (QCRI); Hamad Bin Khalifa University (HBKU)
关键词: Contrastive Language-Speech Pretraining, Contrastive Language-Speech, Language-Speech Pretraining, study introduces CLASP, multimodal representation tailored
类目: Computation and Language (cs.CL); Information Retrieval (cs.IR); Sound (cs.SD); Audio and Speech Processing (eess.AS)
备注: accepted at ECIR 2025

点击查看摘要

Abstract:This study introduces CLASP (Contrastive Language-Speech Pretraining), a multilingual, multimodal representation tailored for audio-text information retrieval. CLASP leverages the synergy between spoken content and textual data. During training, we utilize our newly introduced speech-text dataset, which encompasses 15 diverse categories ranging from fiction to religion. CLASP’s audio component integrates audio spectrograms with a pre-trained self-supervised speech model, while its language encoding counterpart employs a sentence encoder pre-trained on over 100 languages. This unified lightweight model bridges the gap between various modalities and languages, enhancing its effectiveness in handling and retrieving multilingual and multimodal data. Our evaluations across multiple languages demonstrate that CLASP establishes new benchmarks in HITS@1, MRR, and meanR metrics, outperforming traditional ASR-based retrieval approaches in specific scenarios.
zh

[NLP-12] Modality-Inconsistent Continual Learning of Multimodal Large Language Models

【速读】: 该论文试图解决多模态大语言模型(Multimodal Large Language Models, MLLMs)在持续学习(continual learning)中面临的问题,特别是当任务涉及不一致的模态(如图像、音频或视频)和不同任务类型(如描述或问答)时,模型容易遭受灾难性遗忘(catastrophic forgetting)的问题。解决方案的关键在于提出了一种名为MoInCL的方法,该方法通过引入伪目标生成模块(Pseudo Targets Generation Module)来缓解由于任务类型变化导致的遗忘,并采用基于指令的知识蒸馏(Instruction-based Knowledge Distillation)来在新模态引入时保留模型对先前学习模态的处理能力。

链接: https://arxiv.org/abs/2412.13050
作者: Weiguo Pian,Shijian Deng,Shentong Mo,Yunhui Guo,Yapeng Tian
机构: The University of Texas at Dallas(德克萨斯大学达拉斯分校); Carnegie Mellon University(卡内基梅隆大学)
关键词: Multimodal Large Language, Large Language Models, Multimodal Large, Large Language, scenario for Multimodal
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS)
备注:

点击查看摘要

Abstract:In this paper, we introduce Modality-Inconsistent Continual Learning (MICL), a new continual learning scenario for Multimodal Large Language Models (MLLMs) that involves tasks with inconsistent modalities (image, audio, or video) and varying task types (captioning or question-answering). Unlike existing vision-only or modality-incremental settings, MICL combines modality and task type shifts, both of which drive catastrophic forgetting. To address these challenges, we propose MoInCL, which employs a Pseudo Targets Generation Module to mitigate forgetting caused by task type shifts in previously seen modalities. It also incorporates Instruction-based Knowledge Distillation to preserve the model’s ability to handle previously learned modalities when new ones are introduced. We benchmark MICL using a total of six tasks and conduct experiments to validate the effectiveness of our proposed MoInCL. The experimental results highlight the superiority of MoInCL, showing significant improvements over representative and state-of-the-art continual learning baselines.
zh

[NLP-13] Harnessing Event Sensory Data for Error Pattern Prediction in Vehicles: A Language Model Approach AAAI2025

【速读】: 该论文试图解决的问题是通过处理来自车辆的多变量事件流,预测未来可能发生的故障模式及其发生时间。解决方案的关键在于利用事件数据的时间动态性和上下文关系,通过两个因果Transformer模型(causal Transformers)进行建模,从而在故障发生前进行预测。具体来说,论文提出了CarFormer模型,该模型通过一种新的自监督学习策略进行训练,以及EPredictor模型,这是一个自回归Transformer解码器模型,能够预测在某个错误代码出现后,最可能发生的错误模式及其发生时间。尽管面临事件类型的高基数性、不平衡的频率分布和有限的标注数据等挑战,实验结果表明该模型在预测错误模式和发生时间方面表现出色,能够实现80%的F1分数和平均绝对误差为58.4 ± 13.2小时的预测精度,从而支持可靠的预测性维护并提升车辆安全性。

链接: https://arxiv.org/abs/2412.13041
作者: Hugo Math,Rainer Lienhart,Robin Schön
机构: 未知
关键词: processing natural languages, processing multivariate event, multivariate event streams, textit, processing natural
类目: Computation and Language (cs.CL); Machine Learning (cs.LG)
备注: 10 pages, 8 figures, accepted to AAAI 2025

点击查看摘要

Abstract:In this paper, we draw an analogy between processing natural languages and processing multivariate event streams from vehicles in order to predict \textitwhen and \textitwhat error pattern is most likely to occur in the future for a given car. Our approach leverages the temporal dynamics and contextual relationships of our event data from a fleet of cars. Event data is composed of discrete values of error codes as well as continuous values such as time and mileage. Modelled by two causal Transformers, we can anticipate vehicle failures and malfunctions before they happen. Thus, we introduce \textitCarFormer , a Transformer model trained via a new self-supervised learning strategy, and \textitEPredictor , an autoregressive Transformer decoder model capable of predicting \textitwhen and \textitwhat error pattern will most likely occur after some error code apparition. Despite the challenges of high cardinality of event types, their unbalanced frequency of appearance and limited labelled data, our experimental results demonstrate the excellent predictive ability of our novel model. Specifically, with sequences of 160 error codes on average, our model is able with only half of the error codes to achieve 80% F1 score for predicting \textitwhat error pattern will occur and achieves an average absolute error of 58.4 \pm 13.2 h \textitwhen forecasting the time of occurrence, thus enabling confident predictive maintenance and enhancing vehicle safety.
zh

[NLP-14] NAVCON: A Cognitively Inspired and Linguistically Grounded Corpus for Vision and Language Navigation

【速读】: 该论文旨在解决视觉语言导航 (Vision-Language Navigation, VLN) 领域中缺乏大规模标注数据的问题。解决方案的关键在于提出了NAVCON语料库,该语料库基于R2R和RxR数据集构建,包含了四个核心的认知驱动和语言学基础的导航概念,并通过算法生成了大规模的银标注 (silver annotations),这些标注反映了自然语言指令中这些概念的实现。NAVCON不仅提供了约30,000条指令的236,316个概念标注,还包含了与指令对应的2.7百万张对齐图像,展示了执行指令时代理的视觉信息。通过人类评估和模型训练实验,验证了银标注的质量和实用性,并展示了GPT-4在少量样本学习中的有效性。

链接: https://arxiv.org/abs/2412.13026
作者: Karan Wanchoo,Xiaoye Zuo,Hannah Gonzalez,Soham Dan,Georgios Georgakis,Dan Roth,Kostas Daniilidis,Eleni Miltsakaki
机构: University of Pennsylvania; Microsoft; NASA
关键词: large-scale annotated Vision-Language, annotated Vision-Language Navigation, corpus built, popular datasets, built on top
类目: Computation and Language (cs.CL); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:We present NAVCON, a large-scale annotated Vision-Language Navigation (VLN) corpus built on top of two popular datasets (R2R and RxR). The paper introduces four core, cognitively motivated and linguistically grounded, navigation concepts and an algorithm for generating large-scale silver annotations of naturally occurring linguistic realizations of these concepts in navigation instructions. We pair the annotated instructions with video clips of an agent acting on these instructions. NAVCON contains 236, 316 concept annotations for approximately 30, 0000 instructions and 2.7 million aligned images (from approximately 19, 000 instructions) showing what the agent sees when executing an instruction. To our knowledge, this is the first comprehensive resource of navigation concepts. We evaluated the quality of the silver annotations by conducting human evaluation studies on NAVCON samples. As further validation of the quality and usefulness of the resource, we trained a model for detecting navigation concepts and their linguistic realizations in unseen instructions. Additionally, we show that few-shot learning with GPT-4o performs well on this task using large-scale silver annotations of NAVCON.
zh

[NLP-15] OmniEval: An Omnidirectional and Automatic RAG Evaluation Benchmark in Financial Domain

【速读】: 该论文试图解决在垂直领域(如金融领域)中,大型语言模型(LLMs)可能缺乏特定领域知识的问题。解决方案的关键在于引入了一个全方位且自动化的检索增强生成(RAG)基准测试,称为OmniEval。该基准测试通过多维度的评估框架,包括基于矩阵的RAG场景评估系统、多维度的评估数据生成方法、多阶段的评估系统以及基于规则和LLM的鲁棒评估指标,全面评估RAG系统在不同任务和主题上的表现,从而揭示RAG模型在垂直领域中的改进潜力。

链接: https://arxiv.org/abs/2412.13018
作者: Shuting Wang,Jiejun Tan,Zhicheng Dou,Ji-Rong Wen
机构: Gaoling School of Artificial Intelligence, Renmin University of China (高瓴人工智能学院,中国人民大学)
关键词: Large Language Models, Large Language, lack domain-specific knowledge, application of Large, gained extensive attention
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:As a typical and practical application of Large Language Models (LLMs), Retrieval-Augmented Generation (RAG) techniques have gained extensive attention, particularly in vertical domains where LLMs may lack domain-specific knowledge. In this paper, we introduce an omnidirectional and automatic RAG benchmark, OmniEval, in the financial domain. Our benchmark is characterized by its multi-dimensional evaluation framework, including (1) a matrix-based RAG scenario evaluation system that categorizes queries into five task classes and 16 financial topics, leading to a structured assessment of diverse query scenarios; (2) a multi-dimensional evaluation data generation approach, which combines GPT-4-based automatic generation and human annotation, achieving an 87.47% acceptance ratio in human evaluations on generated instances; (3) a multi-stage evaluation system that evaluates both retrieval and generation performance, result in a comprehensive evaluation on the RAG pipeline; and (4) robust evaluation metrics derived from rule-based and LLM-based ones, enhancing the reliability of assessments through manual annotations and supervised fine-tuning of an LLM evaluator. Our experiments demonstrate the comprehensiveness of OmniEval, which includes extensive test datasets and highlights the performance variations of RAG systems across diverse topics and tasks, revealing significant opportunities for RAG models to improve their capabilities in vertical domains. We open source the code of our benchmark in \hrefthis https URLthis https URL.
zh

[NLP-16] RCLMuFN: Relational Context Learning and Multiplex Fusion Network for Multimodal Sarcasm Detection

【速读】: 该论文试图解决多模态讽刺检测中的两个关键问题:一是现有方法未能充分学习文本与图像之间的关系上下文 (relational context),而这是理解讽刺意义的重要依据;二是讽刺的含义随上下文变化而动态变化,现有方法在捕捉这种动态变化时存在局限,导致模型的泛化能力不足。为解决这些问题,论文提出了一种关系上下文学习与多重融合网络 (RCLMuFN),其关键在于:首先,通过四个特征提取器全面提取文本和图像的特征,挖掘潜在特征;其次,利用关系上下文学习模块捕捉文本与图像之间的动态交互;最后,通过多重特征融合模块深度整合多模态特征,提升模型的泛化能力。

链接: https://arxiv.org/abs/2412.13008
作者: Tongguan Wang,Junkai Li,Guixin Su,Yongcheng Zhang,Dongyu Su,Yuxue Hu,Ying Sha
机构: College of Informatics, Huazhong Agricultural University(信息学院,华中农业大学); College of Informatics, Huazhong Agricultural University(信息学院,华中农业大学); College of Informatics, Huazhong Agricultural University(信息学院,华中农业大学); College of Informatics, Huazhong Agricultural University(信息学院,华中农业大学); College of Informatics, Huazhong Agricultural University(信息学院,华中农业大学); College of Informatics, Huazhong Agricultural University(信息学院,华中农业大学); College of Informatics, Huazhong Agricultural University(信息学院,华中农业大学)
关键词: speaker true intent, typically conveys emotions, Sarcasm typically conveys, multimodal sarcasm detection, sarcasm detection
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Sarcasm typically conveys emotions of contempt or criticism by expressing a meaning that is contrary to the speaker’s true intent. Accurate detection of sarcasm aids in identifying and filtering undesirable information on the Internet, thereby reducing malicious defamation and rumor-mongering. Nonetheless, the task of automatic sarcasm detection remains highly challenging for machines, as it critically depends on intricate factors such as relational context. Most existing multimodal sarcasm detection methods focus on introducing graph structures to establish entity relationships between text and images while neglecting to learn the relational context between text and images, which is crucial evidence for understanding the meaning of sarcasm. In addition, the meaning of sarcasm changes with the evolution of different contexts, but existing methods may not be accurate in modeling such dynamic changes, limiting the generalization ability of the models. To address the above issues, we propose a relational context learning and multiplex fusion network (RCLMuFN) for multimodal sarcasm detection. Firstly, we employ four feature extractors to comprehensively extract features from raw text and images, aiming to excavate potential features that may have been previously overlooked. Secondly, we utilize the relational context learning module to learn the contextual information of text and images and capture the dynamic properties through shallow and deep interactions. Finally, we employ a multiplex feature fusion module to enhance the generalization of the model by penetratingly integrating multimodal features derived from various interaction contexts. Extensive experiments on two multimodal sarcasm detection datasets show that our proposed method achieves state-of-the-art performance.
zh

[NLP-17] Enabling Low-Resource Language Retrieval: Establishing Baselines for Urdu MS MARCO ECIR2025

【速读】: 该论文试图解决低资源语言(low-resource languages)在信息检索(Information Retrieval, IR)领域中的代表性和可用性问题,特别是针对乌尔都语(Urdu)。解决方案的关键在于通过机器翻译将MS MARCO数据集转换为乌尔都语,创建了首个大规模乌尔都语IR数据集,并采用多语言IR方法(mMARCO)进行微调。研究结果表明,经过微调的模型(Urdu-mT5-mMARCO)在Mean Reciprocal Rank (MRR@10)和Recall@10指标上分别达到0.247和0.439,显著提升了零样本学习(zero-shot learning)的效果,展示了扩展乌尔都语用户信息检索访问的潜力。该研究不仅推动了多语言IR研究,还强调了包容性IR技术的伦理和社会重要性。

链接: https://arxiv.org/abs/2412.12997
作者: Umer Butt,Stalin Veranasi,Günter Neumann
机构: 未知
关键词: field increasingly recognizes, Information Retrieval, field increasingly, low-resource languages remains, increasingly recognizes
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Information Retrieval (cs.IR)
备注: 6 pages, ECIR 2025, conference submission version

点击查看摘要

Abstract:As the Information Retrieval (IR) field increasingly recognizes the importance of inclusivity, addressing the needs of low-resource languages remains a significant challenge. This paper introduces the first large-scale Urdu IR dataset, created by translating the MS MARCO dataset through machine translation. We establish baseline results through zero-shot learning for IR in Urdu and subsequently apply the mMARCO multilingual IR methodology to this newly translated dataset. Our findings demonstrate that the fine-tuned model (Urdu-mT5-mMARCO) achieves a Mean Reciprocal Rank (MRR@10) of 0.247 and a Recall@10 of 0.439, representing significant improvements over zero-shot results and showing the potential for expanding IR access for Urdu speakers. By bridging access gaps for speakers of low-resource languages, this work not only advances multilingual IR research but also emphasizes the ethical and societal importance of inclusive IR technologies. This work provides valuable insights into the challenges and solutions for improving language representation and lays the groundwork for future research, especially in South Asian languages, which can benefit from the adaptable methods used in this study.
zh

[NLP-18] Unlocking LLM s: Addressing Scarce Data and Bias Challenges in Mental Health

【速读】: 该论文试图解决大型语言模型(LLMs)在医疗分析中面临的挑战,如幻觉(hallucinations)、鹦鹉学舌(parroting)和偏见表现(bias manifestation),特别是在复杂、敏感和低资源领域。解决方案的关键在于引入IC-AnnoMI,这是一个基于专家注释的动机性访谈(Motivational Interviewing, MI)数据集,通过利用LLMs(特别是ChatGPT)生成上下文相关的对话,并结合精心设计的提示(targeted prompts),考虑治疗风格(如共情、反思)、语境相关性和语义变化。随后,专家严格遵循动机性访谈技能编码(Motivational Interviewing Skills Code, MISC)对对话进行注释,重点关注心理和语言维度。通过建模新的分类任务,论文全面评估了IC-AnnoMI数据集和ChatGPT的情感推理能力,并探讨了渐进提示策略和增强数据对缓解偏见的影响。

链接: https://arxiv.org/abs/2412.12981
作者: Vivek Kumar,Eirini Ntoutsi,Pushpraj Singh Rajawat,Giacomo Medda,Diego Reforgiato Recupero
机构: Research Institute CODE, University of the Bundeswehr, Munich, Germany; Barkatullah University, Bhopal, India; University of Cagliari, Cagliari, Italy
关键词: Large language models, shown promising capabilities, Large language, language models, bias manifestation
类目: Computation and Language (cs.CL)
备注: International Conference on Natural Language Processing and Artificial Intelligence for Cyber Security (NLPAICS) 2024

点击查看摘要

Abstract:Large language models (LLMs) have shown promising capabilities in healthcare analysis but face several challenges like hallucinations, parroting, and bias manifestation. These challenges are exacerbated in complex, sensitive, and low-resource domains. Therefore, in this work we introduce IC-AnnoMI, an expert-annotated motivational interviewing (MI) dataset built upon AnnoMI by generating in-context conversational dialogues leveraging LLMs, particularly ChatGPT. IC-AnnoMI employs targeted prompts accurately engineered through cues and tailored information, taking into account therapy style (empathy, reflection), contextual relevance, and false semantic change. Subsequently, the dialogues are annotated by experts, strictly adhering to the Motivational Interviewing Skills Code (MISC), focusing on both the psychological and linguistic dimensions of MI dialogues. We comprehensively evaluate the IC-AnnoMI dataset and ChatGPT’s emotional reasoning ability and understanding of domain intricacies by modeling novel classification tasks employing several classical machine learning and current state-of-the-art transformer approaches. Finally, we discuss the effects of progressive prompting strategies and the impact of augmented data in mitigating the biases manifested in IC-AnnoM. Our contributions provide the MI community with not only a comprehensive dataset but also valuable insights for using LLMs in empathetic text generation for conversational therapy in supervised settings.
zh

[NLP-19] Adaptations of AI models for querying the LandMatrix database in natural language

【速读】: 该论文旨在解决大规模土地收购数据在公共政策中未被充分利用的问题,主要原因是这些数据的访问和利用复杂,需要技术专长和对数据库架构的深入理解。论文的关键解决方案是通过使用大型语言模型 (LLMs) 及其适应技术(如提示工程 (Prompt Engineering)、检索增强生成 (RAG) 和代理 (Agents))来简化对不同数据库系统(如 GraphQL 和 REST 查询)的数据访问。通过这些方法,论文展示了如何有效查询和利用 Land Matrix 的数据,从而为农业、采掘和能源等领域的政策制定提供支持。

链接: https://arxiv.org/abs/2412.12961
作者: Fatiha Ait Kbir,Jérémy Bourgoin,Rémy Decoupes,Marie Gradeler,Roberto Interdonato
机构: INRAE, F-34398 Montpellier, France; CIRAD, F-34398 Montpellier, France; TETIS, Univ. Montpellier, AgroParisTech, CIRAD, CNRS, INRAE, Montpellier 34090, France; International Land Coalition, Rome, Italy
关键词: global observatory aim, Land Matrix initiative, large-scale land acquisitions, provide reliable data, Land Matrix
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:The Land Matrix initiative (this https URL) and its global observatory aim to provide reliable data on large-scale land acquisitions to inform debates and actions in sectors such as agriculture, extraction, or energy in low- and middle-income countries. Although these data are recognized in the academic world, they remain underutilized in public policy, mainly due to the complexity of access and exploitation, which requires technical expertise and a good understanding of the database schema. The objective of this work is to simplify access to data from different database systems. The methods proposed in this article are evaluated using data from the Land Matrix. This work presents various comparisons of Large Language Models (LLMs) as well as combinations of LLM adaptations (Prompt Engineering, RAG, Agents) to query different database systems (GraphQL and REST queries). The experiments are reproducible, and a demonstration is available online: this https URL. Subjects: Computation and Language (cs.CL) Cite as: arXiv:2412.12961 [cs.CL] (or arXiv:2412.12961v1 [cs.CL] for this version) https://doi.org/10.48550/arXiv.2412.12961 Focus to learn more arXiv-issued DOI via DataCite (pending registration)
zh

[NLP-20] SnakModel: Lessons Learned from Training an Open Danish Large Language Model

【速读】: 该论文试图解决为资源有限的小语种(如丹麦语)构建大型语言模型 (LLM) 的挑战,并探索在这一过程中早期建模和训练决策对下游性能的影响。解决方案的关键在于:1) 创建一个严格筛选的多源丹麦语文本语料库;2) 对 Llama2-7B 模型进行持续预训练和指令微调,分析训练过程中的动态变化及超参数调整;3) 在八个语言和文化特定的任务上进行评估。通过这些步骤,SnakModel 在多个 Llama2-7B 模型中表现最佳,并为类似资源受限的语言提供了可行的训练指南。

链接: https://arxiv.org/abs/2412.12956
作者: Mike Zhang,Max Müller-Eberstein,Elisa Bassignana,Rob van der Goot
机构: Aalborg University, Denmark(奥尔堡大学,丹麦); IT University of Copenhagen, Denmark(哥本哈根信息技术大学,丹麦); Pioneer Center for Artificial Intelligence, Denmark(丹麦人工智能先锋中心)
关键词: Danish large language, large language model, Danish Natural Language, Danish words, Danish instructions
类目: Computation and Language (cs.CL)
备注: Accepted at NoDaLiDa 2025 (oral)

点击查看摘要

Abstract:We present SnakModel, a Danish large language model (LLM) based on Llama2-7B, which we continuously pre-train on 13.6B Danish words, and further tune on 3.7M Danish instructions. As best practices for creating LLMs for smaller language communities have yet to be established, we examine the effects of early modeling and training decisions on downstream performance throughout the entire training pipeline, including (1) the creation of a strictly curated corpus of Danish text from diverse sources; (2) the language modeling and instruction-tuning training process itself, including the analysis of intermediate training dynamics, and ablations across different hyperparameters; (3) an evaluation on eight language and culturally-specific tasks. Across these experiments SnakModel achieves the highest overall performance, outperforming multiple contemporary Llama2-7B-based models. By making SnakModel, the majority of our pre-training corpus, and the associated code available under open licenses, we hope to foster further research and development in Danish Natural Language Processing, and establish training guidelines for languages with similar resource constraints.
zh

[NLP-21] Learning from Noisy Labels via Self-Taught On-the-Fly Meta Loss Rescaling AAAI’25

【速读】: 该论文试图解决在训练机器学习模型时,由于标签数据质量不高(包含错误和模糊标签)而影响模型性能的问题。解决方案的关键在于提出了一种无监督的实时元损失重调整方法 (unsupervised on-the-fly meta loss rescaling),通过在训练过程中动态调整样本权重来减轻错误或模糊标签的负面影响。该方法无需依赖干净的种子数据,而是利用正在训练的模型自身的特征来实时学习重调整函数,并通过一种新颖的元学习设置从噪声训练语料中直接采样验证数据进行元更新。这种方法在各种自然语言处理任务中表现出一致的性能提升,并且在面对噪声和模糊标签的对话建模任务中显示出鲁棒性,能够处理类别不平衡问题并防止过拟合到噪声标签。

链接: https://arxiv.org/abs/2412.12955
作者: Michael Heck,Christian Geishauser,Nurul Lubis,Carel van Niekerk,Shutong Feng,Hsien-Chin Lin,Benjamin Matthias Ruppik,Renato Vukovic,Milica Gašić
机构: 未知
关键词: training effective machine, Correct labels, effective machine learning, labeled data, data
类目: Computation and Language (cs.CL)
备注: 10 pages, 3 figures, accepted at AAAI’25

点击查看摘要

Abstract:Correct labels are indispensable for training effective machine learning models. However, creating high-quality labels is expensive, and even professionally labeled data contains errors and ambiguities. Filtering and denoising can be applied to curate labeled data prior to training, at the cost of additional processing and loss of information. An alternative is on-the-fly sample reweighting during the training process to decrease the negative impact of incorrect or ambiguous labels, but this typically requires clean seed data. In this work we propose unsupervised on-the-fly meta loss rescaling to reweight training samples. Crucially, we rely only on features provided by the model being trained, to learn a rescaling function in real time without knowledge of the true clean data distribution. We achieve this via a novel meta learning setup that samples validation data for the meta update directly from the noisy training corpus by employing the rescaling function being trained. Our proposed method consistently improves performance across various NLP tasks with minimal computational overhead. Further, we are among the first to attempt on-the-fly training data reweighting on the challenging task of dialogue modeling, where noisy and ambiguous labels are common. Our strategy is robust in the face of noisy and clean data, handles class imbalance, and prevents overfitting to noisy labels. Our self-taught loss rescaling improves as the model trains, showing the ability to keep learning from the model’s own signals. As training progresses, the impact of correctly labeled data is scaled up, while the impact of wrongly labeled data is suppressed.
zh

[NLP-22] Recipient Profiling: Predicting Characteristics from Messages

【速读】: 该论文试图解决接收者画像 (Recipient Profiling) 问题,即在文本交流中,除了作者的敏感信息外,接收者也可能面临隐私泄露的风险。解决方案的关键在于提出并验证了一种新的任务,即通过分析文本内容来推断接收者的特征。研究通过实验证明了该任务在多个公开数据集上的可行性,并展示了模型在跨数据集迁移时的性能表现,尽管存在一定的准确性损失。

链接: https://arxiv.org/abs/2412.12954
作者: Martin Borquez,Mikaela Keller,Michael Perrot,Damien Sileo
机构: Univ. Lille(里尔大学); Inria(法国国家信息与自动化研究所); CNRS(法国国家科学研究中心); Centrale Lille(里尔中央理工学院); UMR 9189 - CRIStAL(复杂系统与应用实验室)
关键词: inadvertently reveal sensitive, reveal sensitive information, Author Profiling, gender or age, inadvertently reveal
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:It has been shown in the field of Author Profiling that texts may inadvertently reveal sensitive information about their authors, such as gender or age. This raises important privacy concerns that have been extensively addressed in the literature, in particular with the development of methods to hide such information. We argue that, when these texts are in fact messages exchanged between individuals, this is not the end of the story. Indeed, in this case, a second party, the intended recipient, is also involved and should be considered. In this work, we investigate the potential privacy leaks affecting them, that is we propose and address the problem of Recipient Profiling. We provide empirical evidence that such a task is feasible on several publicly accessible datasets (this https URL). Furthermore, we show that the learned models can be transferred to other datasets, albeit with a loss in accuracy.
zh

[NLP-23] MOPO: Multi-Objective Prompt Optimization for Affective Text Generation COLING2025

【速读】: 该论文试图解决在不同领域和情境下生成具有情感色彩文本时,如何选择合适的情感表达方式的问题。解决方案的关键在于引入了一种多目标提示优化方法 (Multi-Objective Prompt Optimization, MOPO),该方法通过同时优化多个目标(由不同领域的情感分类器确定)来生成一系列提示,每个提示对多个目标的权重不同。用户可以根据具体情境选择最合适的提示,从而实现跨领域的情感表达优化。与单一目标优化相比,MOPO在多个目标上提升了性能(最多15个百分点),并且在计算需求上更为高效,避免了为每个目标单独进行优化。

链接: https://arxiv.org/abs/2412.12948
作者: Yarik Menchaca Resendiz,Roman Klinger
机构: Institut für Maschinelle Sprachverarbeitung, University of Stuttgart, Germany(斯图加特大学机器语言处理研究所); Fundamentals of Natural Language Processing, University of Bamberg, Germany(德国班贝格大学自然语言处理基础)
关键词: MOPO, expressed depends, multiple objectives, objectives, Optimization
类目: Computation and Language (cs.CL)
备注: accepted to COLING 2025

点击查看摘要

Abstract:How emotions are expressed depends on the context and domain. On X (formerly Twitter), for instance, an author might simply use the hashtag #anger, while in a news headline, emotions are typically written in a more polite, indirect manner. To enable conditional text generation models to create emotionally connotated texts that fit a domain, users need to have access to a parameter that allows them to choose the appropriate way to express an emotion. To achieve this, we introduce MOPO, a Multi-Objective Prompt Optimization methodology. MOPO optimizes prompts according to multiple objectives (which correspond here to the output probabilities assigned by emotion classifiers trained for different domains). In contrast to single objective optimization, MOPO outputs a set of prompts, each with a different weighting of the multiple objectives. Users can then choose the most appropriate prompt for their context. We evaluate MOPO using three objectives, determined by various domain-specific emotion classifiers. MOPO improves performance by up to 15 pp across all objectives with a minimal loss (1-2 pp) for any single objective compared to single-objective optimization. These minor performance losses are offset by a broader generalization across multiple objectives - which is not possible with single-objective optimization. Additionally, MOPO reduces computational requirements by simultaneously optimizing for multiple objectives, eliminating separate optimization procedures for each objective.
zh

[NLP-24] Improving Fine-grained Visual Understanding in VLMs through Text-Only Training AAAI25

【速读】: 该论文试图解决传统视觉-语言模型 (Visual-Language Models, VLMs) 在训练过程中对图像-文本配对数据的高资源需求问题。解决方案的关键在于探索仅通过文本训练来增强 VLMs 的细粒度视觉理解能力。研究受人类通过丰富的文本描述引导视觉识别的启发,提出利用基于文本的表示来提升 VLMs 的视觉识别能力。实验结果表明,文本训练在细粒度物种分类和文化视觉理解任务中与传统的图像-文本训练效果相当,同时显著降低了计算成本,为资源受限环境下的 VLM 能力提升提供了一条更高效且经济的路径。

链接: https://arxiv.org/abs/2412.12940
作者: Dasol Choi,Guijin Son,Soo Yong Kim,Gio Paik,Seunghyeok Hong
机构: 未知
关键词: Visual-Language Models, powerful tool, tool for bridging, bridging the gap, Models
类目: Computation and Language (cs.CL)
备注: AAAI25 workshop accepted

点击查看摘要

Abstract:Visual-Language Models (VLMs) have become a powerful tool for bridging the gap between visual and linguistic understanding. However, the conventional learning approaches for VLMs often suffer from limitations, such as the high resource requirements of collecting and training image-text paired data. Recent research has suggested that language understanding plays a crucial role in the performance of VLMs, potentially indicating that text-only training could be a viable approach. In this work, we investigate the feasibility of enhancing fine-grained visual understanding in VLMs through text-only training. Inspired by how humans develop visual concept understanding, where rich textual descriptions can guide visual recognition, we hypothesize that VLMs can also benefit from leveraging text-based representations to improve their visual recognition abilities. We conduct comprehensive experiments on two distinct domains: fine-grained species classification and cultural visual understanding tasks. Our findings demonstrate that text-only training can be comparable to conventional image-text training while significantly reducing computational costs. This suggests a more efficient and cost-effective pathway for advancing VLM capabilities, particularly valuable in resource-constrained environments.
zh

[NLP-25] ruthful Text Sanitization Guided by Inference Attacks

【速读】: 该论文试图解决文本脱敏(text sanitization)中隐私保护(privacy protection)与内容保留(utility preservation)之间的平衡问题。解决方案的关键在于提出了一种基于泛化(generalizations)的自动化文本脱敏策略,利用指令微调的大语言模型(instruction-tuned large language models, LLMs)进行两阶段处理。首先,LLM生成保持真实性的替换候选词并根据其抽象程度排序;其次,通过LLM进行推理攻击评估这些候选词的隐私保护能力,最终选择既能抵抗攻击又能保留最多原始信息的最优替换词。该方法通过引入新的自动化评估指标,实现了对隐私保护和内容保留的自动评估,并在实验中展示了其在提升内容保留度的同时,仅略微增加了重新识别的风险。

链接: https://arxiv.org/abs/2412.12928
作者: Ildikó Pilán,Benet Manzanares-Salor,David Sánchez,Pierre Lison
机构: Benet Manzanares-Salor111Corresponding author. Email: benet.manzanares@urv.cat
关键词: longer disclose personal, disclose personal information, text sanitization, personal information, original text spans
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:The purpose of text sanitization is to rewrite those text spans in a document that may directly or indirectly identify an individual, to ensure they no longer disclose personal information. Text sanitization must strike a balance between preventing the leakage of personal information (privacy protection) while also retaining as much of the document’s original content as possible (utility preservation). We present an automated text sanitization strategy based on generalizations, which are more abstract (but still informative) terms that subsume the semantic content of the original text spans. The approach relies on instruction-tuned large language models (LLMs) and is divided into two stages. The LLM is first applied to obtain truth-preserving replacement candidates and rank them according to their abstraction level. Those candidates are then evaluated for their ability to protect privacy by conducting inference attacks with the LLM. Finally, the system selects the most informative replacement shown to be resistant to those attacks. As a consequence of this two-stage process, the chosen replacements effectively balance utility and privacy. We also present novel metrics to automatically evaluate these two aspects without the need to manually annotate data. Empirical results on the Text Anonymization Benchmark show that the proposed approach leads to enhanced utility, with only a marginal increase in the risk of re-identifying protected individuals compared to fully suppressing the original information. Furthermore, the selected replacements are shown to be more truth-preserving and abstractive than previous methods.
zh

[NLP-26] An Agent ic Approach to Automatic Creation of PID Diagrams from Natural Language Descriptions AAAI’25

【速读】: 该论文试图解决工程和流程工业中Piping and Instrumentation Diagrams (PIDs)手动创建过程中存在的劳动密集、易出错以及缺乏有效的错误检测和纠正机制的问题。解决方案的关键在于引入一种基于生成式AI(Generative AI)的新型辅助工具,特别是利用大型语言模型(Large Language Models, LLMs)和视觉-语言模型(Vision-Language Models, VLMs),通过多步骤的智能工作流程,直接从自然语言描述中自动化生成PIDs。该方法提供了一种结构化和迭代式的图表创建方式,显著提升了生成过程的合理性和完整性,相较于传统的零样本和少样本生成方法,取得了更好的效果。

链接: https://arxiv.org/abs/2412.12898
作者: Shreeyash Gowaikar,Srinivasan Iyengar,Sameer Segal,Shivkumar Kalyanaraman
机构: 1. IBM Research(IBM研究院); 2. Rensselaer Polytechnic Institute(伦斯勒理工学院)
关键词: Piping and Instrumentation, Large Language Models, Instrumentation Diagrams, natural language, natural language descriptions
类目: Machine Learning (cs.LG); Computational Engineering, Finance, and Science (cs.CE); Computation and Language (cs.CL); Multiagent Systems (cs.MA)
备注: Accepted at the AAAI’25 Workshop on AI to Accelerate Science and Engineering (AI2ASE)

点击查看摘要

Abstract:The Piping and Instrumentation Diagrams (PIDs) are foundational to the design, construction, and operation of workflows in the engineering and process industries. However, their manual creation is often labor-intensive, error-prone, and lacks robust mechanisms for error detection and correction. While recent advancements in Generative AI, particularly Large Language Models (LLMs) and Vision-Language Models (VLMs), have demonstrated significant potential across various domains, their application in automating generation of engineering workflows remains underexplored. In this work, we introduce a novel copilot for automating the generation of PIDs from natural language descriptions. Leveraging a multi-step agentic workflow, our copilot provides a structured and iterative approach to diagram creation directly from Natural Language prompts. We demonstrate the feasibility of the generation process by evaluating the soundness and completeness of the workflow, and show improved results compared to vanilla zero-shot and few-shot generation approaches.
zh

[NLP-27] Question: How do Large Language Models perform on the Question Answering tasks? Answer:

【速读】: 该论文旨在解决在无需任务特定训练的情况下,比较小型微调模型与现成指令遵循的大型语言模型(LLMs)在斯坦福问答数据集2.0(SQuAD2)上的表现,特别是在使用单次推理提示技术时的性能。关键解决方案在于提出一种无需双次推理的提示风格,以节省计算时间和资源,并探讨这些模型在不同但相似的问答数据集上的泛化能力。研究结果表明,小型微调模型在特定任务上优于当前最先进的LLMs,但在分布外测试中,最新的SOTA模型能够缩小差距,甚至在五个测试数据集中有三个表现优于微调模型。

链接: https://arxiv.org/abs/2412.12893
作者: Kevin Fischer,Darren Fürst,Sebastian Steindl,Jakob Lindner,Ulrich Schäfer
机构: 未知
关键词: Large Language Models, Large Language, showing promising results, zero-shot prompting techniques, Stanford Question Answering
类目: Computation and Language (cs.CL)
备注: Accepted at SAI Computing Conference 2025

点击查看摘要

Abstract:Large Language Models (LLMs) have been showing promising results for various NLP-tasks without the explicit need to be trained for these tasks by using few-shot or zero-shot prompting techniques. A common NLP-task is question-answering (QA). In this study, we propose a comprehensive performance comparison between smaller fine-tuned models and out-of-the-box instruction-following LLMs on the Stanford Question Answering Dataset 2.0 (SQuAD2), specifically when using a single-inference prompting technique. Since the dataset contains unanswerable questions, previous work used a double inference method. We propose a prompting style which aims to elicit the same ability without the need for double inference, saving compute time and resources. Furthermore, we investigate their generalization capabilities by comparing their performance on similar but different QA datasets, without fine-tuning neither model, emulating real-world uses where the context and questions asked may differ from the original training distribution, for example swapping Wikipedia for news articles. Our results show that smaller, fine-tuned models outperform current State-Of-The-Art (SOTA) LLMs on the fine-tuned task, but recent SOTA models are able to close this gap on the out-of-distribution test and even outperform the fine-tuned models on 3 of the 5 tested QA datasets. Comments: Accepted at SAI Computing Conference 2025 Subjects: Computation and Language (cs.CL) Cite as: arXiv:2412.12893 [cs.CL] (or arXiv:2412.12893v1 [cs.CL] for this version) https://doi.org/10.48550/arXiv.2412.12893 Focus to learn more arXiv-issued DOI via DataCite (pending registration)
zh

[NLP-28] RAG-Star: Enhancing Deliberative Reasoning with Retrieval Augmented Verification and Refinement

【速读】: 该论文试图解决现有大型语言模型(LLMs)在处理复杂推理任务时的局限性问题。尽管链式思维(chain-of-thought)和基于树的搜索方法在简单任务中表现出色,但它们主要依赖于LLMs的内部知识,难以应对涉及更多推理步骤的复杂任务。论文提出的解决方案是RAG-Star,一种新颖的检索增强生成(RAG)方法,通过整合检索到的信息来指导基于树的深思熟虑推理过程。关键在于利用蒙特卡洛树搜索(Monte Carlo Tree Search)迭代规划中间子查询和答案,并结合检索增强验证(retrieval-augmented verification),通过查询和答案感知的奖励建模提供反馈,从而有效整合内部和外部知识,显著提升推理能力。

链接: https://arxiv.org/abs/2412.12881
作者: Jinhao Jiang,Jiayi Chen,Junyi Li,Ruiyang Ren,Shijie Wang,Wayne Xin Zhao,Yang Song,Tao Zhang
机构: Gaoling School of Artificial Intelligence, Renmin University of China; Wuhan University of Science and Technology; Northeastern University at Qinhuangdao; Department of Computer Science, National University of Singapore; BOSS Zhipin, Beijing, China
关键词: Existing large language, large language models, show exceptional problem-solving, exceptional problem-solving capabilities, Existing large
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: LLM;RAG;MCTS

点击查看摘要

Abstract:Existing large language models (LLMs) show exceptional problem-solving capabilities but might struggle with complex reasoning tasks. Despite the successes of chain-of-thought and tree-based search methods, they mainly depend on the internal knowledge of LLMs to search over intermediate reasoning steps, limited to dealing with simple tasks involving fewer reasoning steps. In this paper, we propose \textbfRAG-Star, a novel RAG approach that integrates the retrieved information to guide the tree-based deliberative reasoning process that relies on the inherent knowledge of LLMs. By leveraging Monte Carlo Tree Search, RAG-Star iteratively plans intermediate sub-queries and answers for reasoning based on the LLM itself. To consolidate internal and external knowledge, we propose an retrieval-augmented verification that utilizes query- and answer-aware reward modeling to provide feedback for the inherent reasoning of LLMs. Our experiments involving Llama-3.1-8B-Instruct and GPT-4o demonstrate that RAG-Star significantly outperforms previous RAG and reasoning methods.
zh

[NLP-29] Preference-Oriented Supervised Fine-Tuning: Favoring Target Model Over Aligned Large Language Models AAAI2025

【速读】: 该论文试图解决传统监督微调(Supervised Fine-Tuning, SFT)方法在数据质量无法保证的情况下,如何有效提升预训练大语言模型(LLM)的指令遵循能力的问题。解决方案的关键在于引入了一种新的偏好导向的监督微调方法,即PoFT(Preference-oriented Supervised Fine-Tuning)。PoFT的核心思想是通过在相同的SFT数据上,使目标模型相对于已对齐的LLM具有更高的预测概率,从而将数据质量评估信息(即已对齐LLM的预测概率)融入训练过程中,提升模型的指令遵循能力。实验结果表明,PoFT在不同训练数据集和基础模型上均能稳定且一致地优于传统的SFT方法,并且可以与现有的SFT数据过滤方法结合以进一步提升性能。

链接: https://arxiv.org/abs/2412.12865
作者: Yuchen Fan,Yuzhong Hong,Qiushi Wang,Junwei Bao,Hongfei Jiang,Yang Song
机构: 未知
关键词: pre-trained Large language, Large language model, SFT, pre-trained Large, endowing a pre-trained
类目: Computation and Language (cs.CL)
备注: AAAI2025, 12 pages, 9 figures

点击查看摘要

Abstract:Alignment, endowing a pre-trained Large language model (LLM) with the ability to follow instructions, is crucial for its real-world applications. Conventional supervised fine-tuning (SFT) methods formalize it as causal language modeling typically with a cross-entropy objective, requiring a large amount of high-quality instruction-response pairs. However, the quality of widely used SFT datasets can not be guaranteed due to the high cost and intensive labor for the creation and maintenance in practice. To overcome the limitations associated with the quality of SFT datasets, we introduce a novel \textbfpreference-\textbforiented supervised \textbffine-\textbftuning approach, namely PoFT. The intuition is to boost SFT by imposing a particular preference: \textitfavoring the target model over aligned LLMs on the same SFT data. This preference encourages the target model to predict a higher likelihood than that predicted by the aligned LLMs, incorporating assessment information on data quality (i.e., predicted likelihood by the aligned LLMs) into the training process. Extensive experiments are conducted, and the results validate the effectiveness of the proposed method. PoFT achieves stable and consistent improvements over the SFT baselines across different training datasets and base models. Moreover, we prove that PoFT can be integrated with existing SFT data filtering methods to achieve better performance, and further improved by following preference optimization procedures, such as DPO.
zh

[NLP-30] DISC: Plug-and-Play Decoding Intervention with Similarity of Characters for Chinese Spelling Check

【速读】: 该论文试图解决中文拼写检查 (Chinese Spelling Check, CSC) 任务中错误字符通常与正确字符在语音或字形上相似的问题。传统方法依赖于混淆集 (confusion sets),但存在两个主要问题:难以确定哪些字符对应包含在混淆集中,以及缺乏概率来区分混淆集中的项目。论文提出的解决方案是一个轻量级的即插即用模块,称为DISC(即基于字符相似度的解码干预),该模块在推理阶段测量字符间的语音和字形相似度,并将其信息整合到现有CSC模型中,而无需额外的训练成本。该方法可以轻松集成到如ReaLiSe、SCOPE和ReLM等多种现有CSC模型中,并通过在三个CSC基准测试中的实验,显著提升了模型性能,接近甚至超越了当前最先进的模型。

链接: https://arxiv.org/abs/2412.12863
作者: Ziheng Qiao,Houquan Zhou,Yumeng Liu,Zhenghua Li,Min Zhang,Bo Zhang,Chen Li,Ji Zhang,Fei Huang
机构: Soochow University(苏州大学); DAMO Academy, Alibaba Group(阿里巴巴集团达摩院)
关键词: Chinese spelling check, Chinese spelling, spelling check, key characteristic, Chinese
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:One key characteristic of the Chinese spelling check (CSC) task is that incorrect characters are usually similar to the correct ones in either phonetics or glyph. To accommodate this, previous works usually leverage confusion sets, which suffer from two problems, i.e., difficulty in determining which character pairs to include and lack of probabilities to distinguish items in the set. In this paper, we propose a light-weight plug-and-play DISC (i.e., decoding intervention with similarity of characters) module for CSC this http URL measures phonetic and glyph similarities between characters and incorporates this similarity information only during the inference phase. This method can be easily integrated into various existing CSC models, such as ReaLiSe, SCOPE, and ReLM, without additional training costs. Experiments on three CSC benchmarks demonstrate that our proposed method significantly improves model performance, approaching and even surpassing the current state-of-the-art models.
zh

[NLP-31] Selective Shot Learning for Code Explanation

【速读】: 该论文试图解决在代码解释任务中,现有选择性少样本学习(Selective Shot Learning, SSL)方法在开源代码语言模型(Code-LLMs)上的应用和评估不足的问题,特别是这些方法未充分考虑编程语言语法。解决方案的关键在于提出了一种新的SSL方法(SSL_ner),该方法利用实体信息进行少样本示例的选择,并通过系统性基准测试展示了其在两个数据集上优于现有最先进方法的效果。这是首次对开源Code-LLMs进行系统性基准测试,并评估了不同少样本示例选择方法在代码解释任务中的性能。

链接: https://arxiv.org/abs/2412.12852
作者: Paheli Bhattacharya,Rishabh Gupta
机构: Bosch Research and Technology Centre, India(博世研发技术中心,印度)
关键词: software engineering domain, code functionality efficiently, grasping code functionality, Code explanation plays, Code explanation
类目: oftware Engineering (cs.SE); Computation and Language (cs.CL); Information Retrieval (cs.IR)
备注:

点击查看摘要

Abstract:Code explanation plays a crucial role in the software engineering domain, aiding developers in grasping code functionality efficiently. Recent work shows that the performance of LLMs for code explanation improves in a few-shot setting, especially when the few-shot examples are selected intelligently. State-of-the-art approaches for such Selective Shot Learning (SSL) include token-based and embedding-based methods. However, these SSL approaches have been evaluated on proprietary LLMs, without much exploration on open-source Code-LLMs. Additionally, these methods lack consideration for programming language syntax. To bridge these gaps, we present a comparative study and propose a novel SSL method (SSL_ner) that utilizes entity information for few-shot example selection. We present several insights and show the effectiveness of SSL_ner approach over state-of-the-art methods across two datasets. To the best of our knowledge, this is the first systematic benchmarking of open-source Code-LLMs while assessing the performances of the various few-shot examples selection approaches for the code explanation task.
zh

[NLP-32] Benchmarking and Understanding Compositional Relational Reasoning of LLM s AAAI-25 AAAI

【速读】: 该论文试图解决的问题是现有的大语言模型 (LLMs) 在组合关系推理 (Compositional Relational Reasoning, CRR) 任务中的表现和能力。解决方案的关键在于提出了一个新的合成基准测试——广义联想回忆 (Generalized Associative Recall, GAR),并通过归因修补 (attribution patching) 方法揭示了模型在解决GAR任务时所依赖的核心电路和注意力头 (attention heads)。特别是,研究识别了两类注意力头,它们分别代表了GAR任务中的“真”和“假”抽象概念,这些头在不同模型和任务中的正确激活对CRR能力至关重要。

链接: https://arxiv.org/abs/2412.12841
作者: Ruikang Ni,Da Xiao,Qingye Meng,Xiangyu Li,Shihui Zheng,Hongliang Liang
机构: ColorfulClouds Tech.
关键词: Compositional relational reasoning, transformer large language, Generalized Associative Recall, Compositional relational, existing transformer large
类目: Computation and Language (cs.CL); Machine Learning (cs.LG)
备注: Accepted to the 39th Annual AAAI Conference on Artificial Intelligence (AAAI-25)

点击查看摘要

Abstract:Compositional relational reasoning (CRR) is a hallmark of human intelligence, but we lack a clear understanding of whether and how existing transformer large language models (LLMs) can solve CRR tasks. To enable systematic exploration of the CRR capability of LLMs, we first propose a new synthetic benchmark called Generalized Associative Recall (GAR) by integrating and generalizing the essence of several tasks in mechanistic interpretability (MI) study in a unified framework. Evaluation shows that GAR is challenging enough for existing LLMs, revealing their fundamental deficiency in CRR. Meanwhile, it is easy enough for systematic MI study. Then, to understand how LLMs solve GAR tasks, we use attribution patching to discover the core circuits reused by Vicuna-33B across different tasks and a set of vital attention heads. Intervention experiments show that the correct functioning of these heads significantly impacts task performance. Especially, we identify two classes of heads whose activations represent the abstract notion of true and false in GAR tasks respectively. They play a fundamental role in CRR across various models and tasks. The dataset and code are available at this https URL.
zh

[NLP-33] DSGram: Dynamic Weighting Sub-Metrics for Grammatical Error Correction in the Era of Large Language Models AAAI-25

【速读】: 该论文试图解决语法错误纠正 (Grammatical Error Correction, GEC) 模型性能评估中的挑战,特别是基于大语言模型 (LLM) 的GEC系统生成的修正与提供的黄金参考之间存在差异,从而削弱了传统基于参考的评估指标的可靠性。解决方案的关键在于提出了一个名为DSGram的新评估框架,该框架整合了语义连贯性 (Semantic Coherence)、编辑级别 (Edit Level) 和流畅性 (Fluency),并采用动态加权机制。通过结合层次分析法 (Analytic Hierarchy Process, AHP) 和大语言模型来确定各评估标准的相对重要性,并利用包含人工标注和LLM模拟句子的数据集来验证算法和优化更具成本效益的模型。实验结果表明,该方法显著提升了GEC模型评估的有效性。

链接: https://arxiv.org/abs/2412.12832
作者: Jinxiang Xie,Yilin Li,Xunjian Yin,Xiaojun Wan
机构: Peking University (北京大学)
关键词: Grammatical Error Correction, Grammatical Error, provided gold references, Error Correction, based GEC systems
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: Extended version of a paper to appear in AAAI-25

点击查看摘要

Abstract:Evaluating the performance of Grammatical Error Correction (GEC) models has become increasingly challenging, as large language model (LLM)-based GEC systems often produce corrections that diverge from provided gold references. This discrepancy undermines the reliability of traditional reference-based evaluation metrics. In this study, we propose a novel evaluation framework for GEC models, DSGram, integrating Semantic Coherence, Edit Level, and Fluency, and utilizing a dynamic weighting mechanism. Our framework employs the Analytic Hierarchy Process (AHP) in conjunction with large language models to ascertain the relative importance of various evaluation criteria. Additionally, we develop a dataset incorporating human annotations and LLM-simulated sentences to validate our algorithms and fine-tune more cost-effective models. Experimental results indicate that our proposed approach enhances the effectiveness of GEC model evaluations.
zh

[NLP-34] Detecting Emotional Incongruity of Sarcasm by Commonsense Reasoning

【速读】: 该论文试图解决讽刺检测(sarcasm detection)中现有方法在复杂现实场景下缺乏常识推理能力的问题。解决方案的关键在于提出了一种基于常识增强的讽刺检测框架,称为EICR。具体来说,EICR通过检索增强的大型语言模型补充缺失的常识背景知识,构建依赖图并通过图优化获取最佳拓扑结构,引入自适应推理框架以提取情感不一致的子图,并采用对抗对比学习来增强检测器的鲁棒性。

链接: https://arxiv.org/abs/2412.12808
作者: Ziqi Qiu,Jianxing Yu,Yufeng Zhang,Hanjiang Lai,Yanghui Rao,Qinliang Su,Jian Yin
机构: School of Artificial Intelligence, Sun Yat-sen University, Zhuhai, 519082, China; School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, 510006, China; Pazhou Lab, Guangzhou, 510330, China
关键词: negative sentiment opposite, statements convey criticism, convey criticism, literal meaning, paper focuses
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:This paper focuses on sarcasm detection, which aims to identify whether given statements convey criticism, mockery, or other negative sentiment opposite to the literal meaning. To detect sarcasm, humans often require a comprehensive understanding of the semantics in the statement and even resort to external commonsense to infer the fine-grained incongruity. However, existing methods lack commonsense inferential ability when they face complex real-world scenarios, leading to unsatisfactory performance. To address this problem, we propose a novel framework for sarcasm detection, which conducts incongruity reasoning based on commonsense augmentation, called EICR. Concretely, we first employ retrieval-augmented large language models to supplement the missing but indispensable commonsense background knowledge. To capture complex contextual associations, we construct a dependency graph and obtain the optimized topology via graph refinement. We further introduce an adaptive reasoning skeleton that integrates prior rules to extract sentiment-inconsistent subgraphs explicitly. To eliminate the possible spurious relations between words and labels, we employ adversarial contrastive learning to enhance the robustness of the detector. Experiments conducted on five datasets demonstrate the effectiveness of EICR.
zh

[NLP-35] Cross-Dialect Information Retrieval: Information Access in Low-Resource and High-Variance Languages COLING2025

【速读】: 该论文试图解决跨方言检索(Cross-Dialect Retrieval, CDIR)中的挑战,特别是由于资源有限和非标准化语言的高变异性带来的问题。解决方案的关键在于引入首个德语方言检索数据集WikiDIR,并通过实验证明词汇方法在处理方言高词汇变异时的不足。论文进一步指出,常用的零样本跨语言迁移方法在极低资源环境下效果不佳,因此需要资源节约型且针对方言的检索模型。最后,论文展示了文档翻译在减少方言检索中的方言差异方面的有效性。

链接: https://arxiv.org/abs/2412.12806
作者: Robert Litschko,Oliver Kraus,Verena Blaschke,Barbara Plank
机构: MaiNLP, Center for Information and Language Processing, LMU Munich, Germany; Munich Center for Machine Learning (MCML), Munich, Germany
关键词: culture-specific knowledge, large amount, amount of local, local and culture-specific, German dialects
类目: Computation and Language (cs.CL); Information Retrieval (cs.IR)
备注: Accepted at COLING 2025

点击查看摘要

Abstract:A large amount of local and culture-specific knowledge (e.g., people, traditions, food) can only be found in documents written in dialects. While there has been extensive research conducted on cross-lingual information retrieval (CLIR), the field of cross-dialect retrieval (CDIR) has received limited attention. Dialect retrieval poses unique challenges due to the limited availability of resources to train retrieval models and the high variability in non-standardized languages. We study these challenges on the example of German dialects and introduce the first German dialect retrieval dataset, dubbed WikiDIR, which consists of seven German dialects extracted from Wikipedia. Using WikiDIR, we demonstrate the weakness of lexical methods in dealing with high lexical variation in dialects. We further show that commonly used zero-shot cross-lingual transfer approach with multilingual encoders do not transfer well to extremely low-resource setups, motivating the need for resource-lean and dialect-specific retrieval models. We finally demonstrate that (document) translation is an effective way to reduce the dialect gap in CDIR.
zh

[NLP-36] Is it the end of (generative) linguistics as we know it?

【速读】: 该论文试图解决生成语言学(Generative Linguistics)在面对Piantadosi的批评时所面临的挑战,特别是关于刺激贫乏假设(Poverty of Stimulus, PoS)的证据问题以及主流最简主义(Minimalism)中简单性概念的争议。解决方案的关键在于通过三种理想化的视角——计算、理论和实验——来重新审视这些基本问题,并提出生成语言学需要进行重大更新,包括(i)对基础直觉进行更精确、一致和完整的公式化,以及(ii)建立和利用标准化的关键实证数据集来评估理论的充分性。论文强调,忽视形式化视角会导致计算和实验方法的重大缺陷,无法轻易实现描述性或解释性充分性。

链接: https://arxiv.org/abs/2412.12797
作者: Cristiano Chesi
机构: 未知
关键词: written by Steven, Steven Piantadosi, LingBuzz platform, significant debate, debate has emerged
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:A significant debate has emerged in response to a paper written by Steven Piantadosi (Piantadosi, 2023) and uploaded to the LingBuzz platform, the open archive for generative linguistics. Piantadosi’s dismissal of Chomsky’s approach is ruthless, but generative linguists deserve it. In this paper, I will adopt three idealized perspectives – computational, theoretical, and experimental – to focus on two fundamental issues that lend partial support to Piantadosi’s critique: (a) the evidence challenging the Poverty of Stimulus (PoS) hypothesis and (b) the notion of simplicity as conceived within mainstream Minimalism. In conclusion, I argue that, to reclaim a central role in language studies, generative linguistics – representing a prototypical theoretical perspective on language – needs a serious update leading to (i) more precise, consistent, and complete formalizations of foundational intuitions and (ii) the establishment and utilization of a standardized dataset of crucial empirical evidence to evaluate the theory’s adequacy. On the other hand, ignoring the formal perspective leads to major drawbacks in both computational and experimental approaches. Neither descriptive nor explanatory adequacy can be easily achieved without the precise formulation of general principles that can be challenged empirically.
zh

[NLP-37] A Survey of Calibration Process for Black-Box LLM s

【速读】: 该论文试图解决黑箱大语言模型 (Black-Box LLMs) 的校准问题,特别是在仅通过 API 交互的情况下如何准确评估其输出可靠性。解决方案的关键在于定义了校准过程的两个核心步骤:置信度估计 (Confidence Estimation) 和校准 (Calibration)。论文系统性地回顾了适用于黑箱设置的校准方法,并探讨了在这些步骤中实施的独特挑战和联系。此外,论文还探讨了校准过程在黑箱 LLMs 中的典型应用,并指出了未来研究的方向,以增强模型的可靠性和人机对齐。

链接: https://arxiv.org/abs/2412.12767
作者: Liangru Xie,Hui Liu,Jingying Zeng,Xianfeng Tang,Yan Han,Chen Luo,Jing Huang,Zhen Li,Suhang Wang,Qi He
机构: Amazon(亚马逊); Penn State University (宾夕法尼亚州立大学)
关键词: Large Language Models, Large Language, Language Models, demonstrate remarkable performance, output reliability remains
类目: Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Large Language Models (LLMs) demonstrate remarkable performance in semantic understanding and generation, yet accurately assessing their output reliability remains a significant challenge. While numerous studies have explored calibration techniques, they primarily focus on White-Box LLMs with accessible parameters. Black-Box LLMs, despite their superior performance, pose heightened requirements for calibration techniques due to their API-only interaction constraints. Although recent researches have achieved breakthroughs in black-box LLMs calibration, a systematic survey of these methodologies is still lacking. To bridge this gap, we presents the first comprehensive survey on calibration techniques for black-box LLMs. We first define the Calibration Process of LLMs as comprising two interrelated key steps: Confidence Estimation and Calibration. Second, we conduct a systematic review of applicable methods within black-box settings, and provide insights on the unique challenges and connections in implementing these key steps. Furthermore, we explore typical applications of Calibration Process in black-box LLMs and outline promising future research directions, providing new perspectives for enhancing reliability and human-machine alignment. This is our GitHub link: this https URL
zh

[NLP-38] Revealing the impact of synthetic native samples and multi-tasking strategies in Hindi-English code-mixed humour and sarcasm detection

【速读】: 该论文试图解决印地语-英语代码混合场景下的幽默和讽刺检测问题。解决方案的关键在于三种策略的实验:(i) 原生样本混合 (native sample mixing),通过在代码混合训练集中加入单语任务样本提升检测效果;(ii) 多任务学习 (Multi-task Learning, MTL),利用语义相关的任务(如仇恨检测)的原生和代码混合样本进行训练,显著提升幽默和讽刺检测的F1分数;(iii) 提示超大型多语言语言模型 (very large multilingual language models, VMLMs),通过少样本上下文提示进行评估,但效果未能超越前两种方法。

链接: https://arxiv.org/abs/2412.12761
作者: Debajyoti Mazumder,Aakash Kumar,Jasabanta Patro
机构: Indian Institute of Science Education and Research, Bhopal (印度科学教育与研究学院,博帕尔); Intelligent Systems’ Lab (ISL) (智能系统实验室); Department of Data Science and Engineering (数据科学与工程系)
关键词: reported our experiments, native sample mixing, sample mixing, MTL, code-mixed
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: 26 pages; under review

点击查看摘要

Abstract:In this paper, we reported our experiments with various strategies to improve code-mixed humour and sarcasm detection. We did all of our experiments for Hindi-English code-mixed scenario, as we have the linguistic expertise for the same. We experimented with three approaches, namely (i) native sample mixing, (ii) multi-task learning (MTL), and (iii) prompting very large multilingual language models (VMLMs). In native sample mixing, we added monolingual task samples in code-mixed training sets. In MTL learning, we relied on native and code-mixed samples of a semantically related task (hate detection in our case). Finally, in our third approach, we evaluated the efficacy of VMLMs via few-shot context prompting. Some interesting findings we got are (i) adding native samples improved humor (raising the F1-score up to 6.76%) and sarcasm (raising the F1-score up to 8.64%) detection, (ii) training MLMs in an MTL framework boosted performance for both humour (raising the F1-score up to 10.67%) and sarcasm (increment up to 12.35% in F1-score) detection, and (iii) prompting VMLMs couldn’t outperform the other approaches. Finally, our ablation studies and error analysis discovered the cases where our model is yet to improve. We provided our code for reproducibility.
zh

[NLP-39] Your Next State-of-the-Art Could Come from Another Domain: A Cross-Domain Analysis of Hierarchical Text Classification

【速读】: 该论文试图解决层次标签文本分类任务中缺乏跨领域全面理解和评估的问题。解决方案的关键在于提出了一个统一的框架,将不同领域中的现有方法纳入一个共同的结构中,并通过跨领域的实证分析来验证这些方法的有效性。该框架不仅促进了研究的可比性,还通过跨领域技术应用实现了新的最先进结果。

链接: https://arxiv.org/abs/2412.12744
作者: Nan Li,Bo Kang,Tijl De Bie
机构: Ghent University (根特大学)
关键词: natural language processing, language processing, European legal texts, classification with hierarchical, hierarchical labels
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:Text classification with hierarchical labels is a prevalent and challenging task in natural language processing. Examples include assigning ICD codes to patient records, tagging patents into IPC classes, assigning EUROVOC descriptors to European legal texts, and more. Despite its widespread applications, a comprehensive understanding of state-of-the-art methods across different domains has been lacking. In this paper, we provide the first comprehensive cross-domain overview with empirical analysis of state-of-the-art methods. We propose a unified framework that positions each method within a common structure to facilitate research. Our empirical analysis yields key insights and guidelines, confirming the necessity of learning across different research areas to design effective methods. Notably, under our unified evaluation pipeline, we achieved new state-of-the-art results by applying techniques beyond their original domains.
zh

[NLP-40] GIRAFFE: Design Choices for Extending the Context Length of Visual Language Models

【速读】: 该论文试图解决视觉语言模型 (Visual Language Models, VLMs) 在处理多图像和高分辨率视频等需要长程建模的应用场景中性能不足的问题。解决方案的关键在于通过系统性的方法提升VLMs在长上下文场景中的表现,同时保持其在短上下文场景中的能力。具体措施包括:(1) 通过分析数据源和长度分布构建ETVLM数据集,以平衡不同场景下的性能;(2) 提出增强的位置扩展方法M-RoPE++,并仅使用混合数据对骨干网络进行指令微调;(3) 提出混合分辨率训练策略,以更好地利用扩展的上下文窗口。基于Qwen-VL系列模型,论文提出了Giraffe模型,将其上下文长度扩展至128K,并在多个长上下文VLM基准测试中取得了与商业模型GPT-4V相竞争的性能。

链接: https://arxiv.org/abs/2412.12735
作者: Mukai Li,Lei Li,Shansan Gong,Qi Liu
机构: The University of Hong Kong (香港大学)
关键词: Visual Language Models, Visual Language, demonstrate impressive capabilities, processing multimodal inputs, require handling multiple
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注: Working in progress

点击查看摘要

Abstract:Visual Language Models (VLMs) demonstrate impressive capabilities in processing multimodal inputs, yet applications such as visual agents, which require handling multiple images and high-resolution videos, demand enhanced long-range modeling. Moreover, existing open-source VLMs lack systematic exploration into extending their context length, and commercial models often provide limited details. To tackle this, we aim to establish an effective solution that enhances long context performance of VLMs while preserving their capacities in short context scenarios. Towards this goal, we make the best design choice through extensive experiment settings from data curation to context window extending and utilizing: (1) we analyze data sources and length distributions to construct ETVLM - a data recipe to balance the performance across scenarios; (2) we examine existing position extending methods, identify their limitations and propose M-RoPE++ as an enhanced approach; we also choose to solely instruction-tune the backbone with mixed-source data; (3) we discuss how to better utilize extended context windows and propose hybrid-resolution training. Built on the Qwen-VL series model, we propose Giraffe, which is effectively extended to 128K lengths. Evaluated on extensive long context VLM benchmarks such as VideoMME and Viusal Haystacks, our Giraffe achieves state-of-the-art performance among similarly sized open-source long VLMs and is competitive with commercial model GPT-4V. We will open-source the code, data, and models.
zh

[NLP-41] EventFull: Complete and Consistent Event Relation Annotation

【速读】: 该论文试图解决事件关系检测任务中数据标注的系统性和完整性问题。由于事件关系标注需要考虑事件对之间的所有可能关系,导致标注成本高且复杂。论文提出的解决方案是引入 EventFull,这是一个支持时间关系、因果关系和共指关系一致、完整且高效标注的工具。其关键在于通过统一的协同过程实现标注,从而加速标注流程并提高标注者间的一致性。

链接: https://arxiv.org/abs/2412.12733
作者: Alon Eirew,Eviatar Nachshoni,Aviv Slobodkin,Ido Dagan
机构: Bar-Ilan University (巴伊兰大学)
关键词: fundamental NLP task, NLP task, fundamental NLP, modeling requires datasets, requires datasets annotated
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Event relation detection is a fundamental NLP task, leveraged in many downstream applications, whose modeling requires datasets annotated with event relations of various types. However, systematic and complete annotation of these relations is costly and challenging, due to the quadratic number of event pairs that need to be considered. Consequently, many current event relation datasets lack systematicity and completeness. In response, we introduce \textitEventFull, the first tool that supports consistent, complete and efficient annotation of temporal, causal and coreference relations via a unified and synergetic process. A pilot study demonstrates that EventFull accelerates and simplifies the annotation process while yielding high inter-annotator agreement.
zh

[NLP-42] SentiQNF: A Novel Approach to Sentiment Analysis Using Quantum Algorithms and Neuro-Fuzzy Systems

【速读】: 该论文试图解决传统情感分析方法在处理大规模数据、高维数据、噪声和异常值时的局限性,以及对输入变化不敏感的问题。解决方案的关键在于提出了一种新颖的混合方法——量子模糊神经网络 (Quantum Fuzzy Neural Network, QFNN),该方法结合了量子特性并引入模糊层,以克服传统情感分析算法的不足。QFNN在处理噪声和异常值时表现出更强的鲁棒性,并能有效处理高维数据,从而在情感分类任务中实现了更高的准确性和效率。

链接: https://arxiv.org/abs/2412.12731
作者: Kshitij Dave,Nouhaila Innan,Bikash K. Behera,Zahid Mumtaz,Saif Al-Kuwari,Ahmed Farouk
机构: Institute of Advanced Research(高级研究所); eBRAIN Lab, Division of Engineering, New York University Abu Dhabi (NYUAD)(纽约大学阿布扎比分校工程系eBRAIN实验室); Center for Quantum and Topological Systems (CQTS), NYUAD Research Institute, NYUAD(纽约大学阿布扎比分校量子与拓扑系统研究中心); Bikash’s Quantum (OPC) Pvt. Ltd.(比卡什量子私人有限公司); London School of Economics and Political Science(伦敦政治经济学院); Qatar Center for Quantum Computing, College of Science and Engineering, Hamad Bin Khalifa University, Qatar Foundation(哈马德·本·哈利法大学卡塔尔量子计算中心,科学与工程学院); Department of Computer Science, Faculty of Computers and Artificial Intelligence, Hurghada University(赫尔格达大学计算机科学与人工智能系)
关键词: Sentiment analysis, essential component, component of natural, emotional tones, Sentiment
类目: Computation and Language (cs.CL); Quantum Physics (quant-ph)
备注:

点击查看摘要

Abstract:Sentiment analysis is an essential component of natural language processing, used to analyze sentiments, attitudes, and emotional tones in various contexts. It provides valuable insights into public opinion, customer feedback, and user experiences. Researchers have developed various classical machine learning and neuro-fuzzy approaches to address the exponential growth of data and the complexity of language structures in sentiment analysis. However, these approaches often fail to determine the optimal number of clusters, interpret results accurately, handle noise or outliers efficiently, and scale effectively to high-dimensional data. Additionally, they are frequently insensitive to input variations. In this paper, we propose a novel hybrid approach for sentiment analysis called the Quantum Fuzzy Neural Network (QFNN), which leverages quantum properties and incorporates a fuzzy layer to overcome the limitations of classical sentiment analysis algorithms. In this study, we test the proposed approach on two Twitter datasets: the Coronavirus Tweets Dataset (CVTD) and the General Sentimental Tweets Dataset (GSTD), and compare it with classical and hybrid algorithms. The results demonstrate that QFNN outperforms all classical, quantum, and hybrid algorithms, achieving 100% and 90% accuracy in the case of CVTD and GSTD, respectively. Furthermore, QFNN demonstrates its robustness against six different noise models, providing the potential to tackle the computational complexity associated with sentiment analysis on a large scale in a noisy environment. The proposed approach expedites sentiment data processing and precisely analyses different forms of textual data, thereby enhancing sentiment classification and insights associated with sentiment analysis.
zh

[NLP-43] Enhancing Naturalness in LLM -Generated Utterances through Disfluency Insertion

【速读】: 该论文试图解决大语言模型 (Large Language Models, LLMs) 生成的语音中缺乏自然的不流畅性 (disfluencies) 的问题,这会降低合成语音的自然感,影响对话代理模仿人类行为的效果。解决方案的关键在于通过两种方法来引入不流畅性:(1) 使用低秩适应 (Low-Rank Adaptation, LoRA) 对 LLM 进行微调,使其在生成语句时包含各种类型的不流畅性;(2) 使用支持生成不流畅性等语音现象的文本到语音模型 (text-to-speech model) 来合成这些语句。通过用户研究,论文证明了插入不流畅性显著提高了生成语音的感知自然度 (perceived spontaneity),但同时也略微降低了可理解性 (intelligibility)。

链接: https://arxiv.org/abs/2412.12710
作者: Syed Zohaib Hassan,Pierre Lison,Pål Halvorsen
机构: SimulaMet; Norwegian Computing Center
关键词: Large Language Models, Large Language, outputs of Large, Language Models, spontaneous human speech
类目: Computation and Language (cs.CL)
备注: 4 pages short paper, references and appendix are additional

点击查看摘要

Abstract:Disfluencies are a natural feature of spontaneous human speech but are typically absent from the outputs of Large Language Models (LLMs). This absence can diminish the perceived naturalness of synthesized speech, which is an important criteria when building conversational agents that aim to mimick human behaviours. We show how the insertion of disfluencies can alleviate this shortcoming. The proposed approach involves (1) fine-tuning an LLM with Low-Rank Adaptation (LoRA) to incorporate various types of disfluencies into LLM-generated utterances and (2) synthesizing those utterances using a text-to-speech model that supports the generation of speech phenomena such as disfluencies. We evaluated the quality of the generated speech across two metrics: intelligibility and perceived spontaneity. We demonstrate through a user study that the insertion of disfluencies significantly increase the perceived spontaneity of the generated speech. This increase came, however, along with a slight reduction in intelligibility.
zh

[NLP-44] More Tokens Lower Precision: Towards the Optimal Token-Precision Trade-off in KV Cache Compression

【速读】: 该论文试图解决大语言模型 (LLMs) 在处理长上下文窗口时,KV缓存内存使用成为推理过程中关键瓶颈的问题。解决方案的关键在于全面研究KV缓存压缩中的token-precision权衡,提出了一种结合token剪枝 (KV pruning) 和量化 (KV quantization) 的量化剪枝方法 (quantized pruning)。实验表明,通过以较低精度存储更多token,量化剪枝显著提升了LLMs在长上下文场景下的性能,并在检索相关任务中表现出显著的稳定性和一致性,适用于不同剪枝方法、量化策略和模型规模。

链接: https://arxiv.org/abs/2412.12706
作者: Jiebin Zhang,Dawei Zhu,Yifan Song,Wenhao Wu,Chuqiao Kuang,Xiaoguang Li,Lifeng Shang,Qun Liu,Sujian Li
机构: Peking University(北京大学); Huawei Technologies(华为技术)
关键词: process increasing context, increasing context windows, large language models, process increasing, context windows
类目: Computation and Language (cs.CL)
备注: 13pages,7 figures

点击查看摘要

Abstract:As large language models (LLMs) process increasing context windows, the memory usage of KV cache has become a critical bottleneck during inference. The mainstream KV compression methods, including KV pruning and KV quantization, primarily focus on either token or precision dimension and seldom explore the efficiency of their combination. In this paper, we comprehensively investigate the token-precision trade-off in KV cache compression. Experiments demonstrate that storing more tokens in the KV cache with lower precision, i.e., quantized pruning, can significantly enhance the long-context performance of LLMs. Furthermore, in-depth analysis regarding token-precision trade-off from a series of key aspects exhibit that, quantized pruning achieves substantial improvements in retrieval-related tasks and consistently performs well across varying input lengths. Moreover, quantized pruning demonstrates notable stability across different KV pruning methods, quantization strategies, and model scales. These findings provide valuable insights into the token-precision trade-off in KV cache compression. We plan to release our code in the near future.
zh

[NLP-45] rigger3: Refining Query Correction via Adaptive Model Selector

【速读】: 该论文试图解决搜索场景中由于拼写错误、语音错误或知识空白导致的错误查询问题,关键在于提出了一种名为 Trigger³ 的大模型与小模型协作框架。该框架通过结合传统纠错模型和大型语言模型 (LLM),能够根据查询内容和纠错结果自适应选择合适的纠错方法。具体流程包括:首先使用纠错触发器过滤正确查询,对错误查询由传统纠错模型处理;若失败,则激活LLM触发器调用LLM进行纠错;最后,对于无法纠正的查询,回退触发器决定返回原始查询。实验表明,Trigger³ 在保持效率的同时,显著优于现有的纠错基线模型。

链接: https://arxiv.org/abs/2412.12701
作者: Kepu Zhang,Zhongxiang Sun,Xiao Zhang,Xiaoxue Zang,Kai Zheng,Yang Song,Jun Xu
机构: Renmin University of China (中国人民大学); Engineering Research Center of Next-Generation Intelligent Search and Recommendation, Ministry of Education (教育部下一代智能搜索与推荐工程研究中心); Kuaishou
关键词: correction, erroneous queries due, voice errors, traditional correction model, user experience
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:In search scenarios, user experience can be hindered by erroneous queries due to typos, voice errors, or knowledge gaps. Therefore, query correction is crucial for search engines. Current correction models, usually small models trained on specific data, often struggle with queries beyond their training scope or those requiring contextual understanding. While the advent of Large Language Models (LLMs) offers a potential solution, they are still limited by their pre-training data and inference cost, particularly for complex queries, making them not always effective for query correction. To tackle these, we propose Trigger ^3 , a large-small model collaboration framework that integrates the traditional correction model and LLM for query correction, capable of adaptively choosing the appropriate correction method based on the query and the correction results from the traditional correction model and LLM. Trigger ^3 first employs a correction trigger to filter out correct queries. Incorrect queries are then corrected by the traditional correction model. If this fails, an LLM trigger is activated to call the LLM for correction. Finally, for queries that no model can correct, a fallback trigger decides to return the original query. Extensive experiments demonstrate Trigger ^3 outperforms correction baselines while maintaining efficiency.
zh

[NLP-46] XTransplant: A Probe into the Upper Bound Performance of Multilingual Capability and Culture Adaptability in LLM s via Mutual Cross-lingual Feed-forward Transplantation

【速读】: 该论文试图解决当前大型语言模型(LLMs)在多语言能力和文化适应性方面的不平衡问题,主要源于其以英语为中心的预训练数据。解决方案的关键在于提出了一种名为XTransplant的探测方法,通过在推理阶段进行跨语言前馈移植(cross-lingual feed-forward transplantation)来探索跨语言的潜在交互,从而使模型能够同时利用英语和非英语语言的优势。实验证明,XTransplant能够显著提升LLMs的多语言能力和文化适应性,并揭示了当前LLMs多语言潜力未被充分利用的问题。此外,该方法还启发了一种离线扩展推理策略,进一步在多语言和文化感知任务中表现出一致的性能提升,甚至在某些情况下超越了多语言监督微调的效果。

链接: https://arxiv.org/abs/2412.12686
作者: Yangfan Ye,Xiaocheng Feng,Xiachong Feng,Libo Qin,Yichong Huang,Lei Huang,Weitao Ma,Zhirui Zhang,Yunfei Lu,Xiaohui Yan,Duyu Tang,Dandan Tu,Bing Qin
机构: 未知
关键词: English-centric pretraining data, English-centric pretraining, Current large language, large language models, largely due
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Current large language models (LLMs) often exhibit imbalances in multilingual capabilities and cultural adaptability, largely due to their English-centric pretraining data. To address this imbalance, we propose a probing method named XTransplant that explores cross-lingual latent interactions via cross-lingual feed-forward transplantation during inference stage, with the hope of enabling the model to leverage the strengths of both English and non-English languages. Through extensive pilot experiments, we empirically prove that both the multilingual capabilities and cultural adaptability of LLMs hold the potential to be significantly improved by XTransplant, respectively from En - non-En and non-En - En, highlighting the underutilization of current LLMs’ multilingual potential. And the patterns observed in these pilot experiments further motivate an offline scaling inference strategy, which demonstrates consistent performance improvements in multilingual and culture-aware tasks, sometimes even surpassing multilingual supervised fine-tuning. And we do hope our further analysis and discussion could help gain deeper insights into XTransplant mechanism.
zh

[NLP-47] Detecting Document-level Paraphrased Machine Generated Content: Mimicking Human Writing Style and Involving Discourse Features

【速读】: 该论文旨在解决现有机器生成内容(MGC)检测器在识别经过改写或较长文本时存在的局限性,特别是它们依赖于表面信息而忽视隐含和结构特征的问题。解决方案的关键在于引入新的方法和数据集,包括开发了改写的长篇问答(paraLFQA)和改写的写作提示(paraWP)数据集,并提出了两种新型模型:MhBART和DTransformer。MhBART通过引入差异分数机制来模拟人类写作风格,从而识别欺骗性句式;DTransformer则通过结合篇章分析(PDTB)预处理来编码结构特征,以更好地捕捉长文本的文档级结构。这些方法在检测高度相似的改写文本和长文本方面显著提升了性能,分别在paraLFQA和paraWP数据集上实现了15.5%和4%的绝对性能提升,并在M4数据集上超越了现有最先进(SOTA)方法1.5%。

链接: https://arxiv.org/abs/2412.12679
作者: Yupei Li,Manuel Milling,Lucia Specia,Björn W. Schuller
机构: Imperial College London(伦敦帝国学院); Technical University of Munich(慕尼黑工业大学)
关键词: Machine-Generated Content, Large Language Models, APIs for Large, Large Language, spread of misinformation
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:The availability of high-quality APIs for Large Language Models (LLMs) has facilitated the widespread creation of Machine-Generated Content (MGC), posing challenges such as academic plagiarism and the spread of misinformation. Existing MGC detectors often focus solely on surface-level information, overlooking implicit and structural features. This makes them susceptible to deception by surface-level sentence patterns, particularly for longer texts and in texts that have been subsequently paraphrased. To overcome these challenges, we introduce novel methodologies and datasets. Besides the publicly available dataset Plagbench, we developed the paraphrased Long-Form Question and Answer (paraLFQA) and paraphrased Writing Prompts (paraWP) datasets using GPT and DIPPER, a discourse paraphrasing tool, by extending artifacts from their original versions. To address the challenge of detecting highly similar paraphrased texts, we propose MhBART, an encoder-decoder model designed to emulate human writing style while incorporating a novel difference score mechanism. This model outperforms strong classifier baselines and identifies deceptive sentence patterns. To better capture the structure of longer texts at document level, we propose DTransformer, a model that integrates discourse analysis through PDTB preprocessing to encode structural features. It results in substantial performance gains across both datasets – 15.5% absolute improvement on paraLFQA, 4% absolute improvement on paraWP, and 1.5% absolute improvement on M4 compared to SOTA approaches. Subjects: Computation and Language (cs.CL) Cite as: arXiv:2412.12679 [cs.CL] (or arXiv:2412.12679v1 [cs.CL] for this version) https://doi.org/10.48550/arXiv.2412.12679 Focus to learn more arXiv-issued DOI via DataCite (pending registration)
zh

[NLP-48] rain More Parameters But Mind Their Placement: Insights into Language Adaptation with PEFT

【速读】: 该论文试图解决小型语言模型(LLMs)在中等资源语言中面临的语言特定知识不足的问题,尤其是在使用机器翻译数据难以解决的情况下。解决方案的关键在于通过使用非结构化文本语料库对模型进行专门化,同时避免在适应过程中干扰模型处理较长上下文的能力。研究通过参数高效微调(PEFT)方法的消融实验发现,增加可训练参数的数量能显著提升语言适应的性能和鲁棒性。具体来说,放置在前馈层的LoRAs和瓶颈适配器在有足够参数的情况下表现出色,而前缀调优和(IA)3则不适用。尽管在零样本摘要任务中适应后的模型表现一致,但在处理较长上下文时仍存在挑战,这一问题可以通过仅适应最终层来缓解。

链接: https://arxiv.org/abs/2412.12674
作者: Jenny Kunz
机构: Linköping University (林雪平大学)
关键词: face significant challenges, Smaller LLMs, language-specific knowledge, machine-translated data, face significant
类目: Computation and Language (cs.CL)
备注: To appear at NoDaLiDa 2025

点击查看摘要

Abstract:Smaller LLMs still face significant challenges even in medium-resourced languages, particularly when it comes to language-specific knowledge – a problem not easily resolved with machine-translated data. In this case study on Icelandic, we aim to enhance the generation performance of an LLM by specialising it using unstructured text corpora. A key focus is on preventing interference with the models’ capabilities of handling longer context during this adaptation. Through ablation studies using various parameter-efficient fine-tuning (PEFT) methods and setups, we find that increasing the number of trainable parameters leads to better and more robust language adaptation. LoRAs placed in the feed-forward layers and bottleneck adapters show promising results with sufficient parameters, while prefix tuning and (IA)3 are not suitable. Although improvements are consistent in 0-shot summarisation, some adapted models struggle with longer context lengths, an issue that can be mitigated by adapting only the final layers.
zh

[NLP-49] MedMax: Mixed-Modal Instruction Tuning for Training Biomedical Assistants

【速读】: 该论文试图解决现有混合模态生成模型在生物医学领域应用中的数据可用性有限、领域覆盖狭窄以及数据来源受限等问题。解决方案的关键在于提出了MedMax,这是一个大规模的多模态生物医学指令调优数据集,包含1.47百万个实例,涵盖了多模态内容生成、生物医学图像描述与生成、视觉聊天和报告理解等多种任务,跨越放射学和组织病理学等多个医学领域。通过在MedMax数据集上微调混合模态基础模型,论文展示了显著的性能提升,例如在12个下游生物医学视觉问答任务中,相较于Chameleon模型提升了26%,相较于GPT-4o提升了18.3%。此外,论文还引入了一个统一的生物医学任务评估套件,为下一代混合模态生物医学AI助手的开发提供了强有力的框架。

链接: https://arxiv.org/abs/2412.12661
作者: Hritik Bansal,Daniel Israel,Siyan Zhao,Shufan Li,Tung Nguyen,Aditya Grover
机构: University of California Los Angeles (加州大学洛杉矶分校)
关键词: enabled flexible integration, Recent advancements, enabled flexible, flexible integration, integration of information
类目: Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Computer Vision and Pattern Recognition (cs.CV)
备注: 12 figures, 15 tables

点击查看摘要

Abstract:Recent advancements in mixed-modal generative models have enabled flexible integration of information across image-text content. These models have opened new avenues for developing unified biomedical assistants capable of analyzing biomedical images, answering complex questions about them, and predicting the impact of medical procedures on a patient’s health. However, existing resources face challenges such as limited data availability, narrow domain coverage, and restricted sources (e.g., medical papers). To address these gaps, we present MedMax, the first large-scale multimodal biomedical instruction-tuning dataset for mixed-modal foundation models. With 1.47 million instances, MedMax encompasses a diverse range of tasks, including multimodal content generation (interleaved image-text data), biomedical image captioning and generation, visual chatting, and report understanding. These tasks span diverse medical domains such as radiology and histopathology. Subsequently, we fine-tune a mixed-modal foundation model on the MedMax dataset, achieving significant performance improvements: a 26% gain over the Chameleon model and an 18.3% improvement over GPT-4o across 12 downstream biomedical visual question-answering tasks. Additionally, we introduce a unified evaluation suite for biomedical tasks, providing a robust framework to guide the development of next-generation mixed-modal biomedical AI assistants.
zh

[NLP-50] ClustEm4Ano: Clustering Text Embeddings of Nominal Textual Attributes for Microdata Anonymization

【速读】: 该论文旨在解决对名义文本表格数据进行基于泛化和抑制的匿名化问题,提出了一个名为 ClustEm4Ano 的匿名化管道。其关键解决方案在于自动生成值泛化层次结构 (Value Generalization Hierarchies, VGHs),并通过嵌入技术利用迭代聚类生成语义上接近的值泛化。该管道结合了 KMeans 和层次凝聚聚类 (Hierarchical Agglomerative Clustering) 在 13 种预定义的文本嵌入上进行实验,相较于手动构建的 VGHs,实验结果表明其在下游效能(特别是在小 k-匿名性 (2 ≤ k ≤ 30) 的情况下)上表现更优,从而提升了匿名化数据集的质量。

链接: https://arxiv.org/abs/2412.12649
作者: Robert Aufschläger,Sebastian Wilhelm,Michael Heigl,Martin Schramm
机构: 未知
关键词: textual tabular data, nominal textual tabular, work introduces, tabular data, nominal textual
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: 16 pages, 5 figures, accepted for presentation at IDEAS: 2024 28th International Symposium on Database Engineered Applications, Bayonne, France, August 26-29, 2024

点击查看摘要

Abstract:This work introduces ClustEm4Ano, an anonymization pipeline that can be used for generalization and suppression-based anonymization of nominal textual tabular data. It automatically generates value generalization hierarchies (VGHs) that, in turn, can be used to generalize attributes in quasi-identifiers. The pipeline leverages embeddings to generate semantically close value generalizations through iterative clustering. We applied KMeans and Hierarchical Agglomerative Clustering on 13 different predefined text embeddings (both open and closed-source (via APIs)). Our approach is experimentally tested on a well-known benchmark dataset for anonymization: The UCI Machine Learning Repository’s Adult dataset. ClustEm4Ano supports anonymization procedures by offering more possibilities compared to using arbitrarily chosen VGHs. Experiments demonstrate that these VGHs can outperform manually constructed ones in terms of downstream efficacy (especially for small k -anonymity ( 2 \leq k \leq 30 )) and therefore can foster the quality of anonymized datasets. Our implementation is made public.
zh

[NLP-51] PrOp: Interactive Prompt Optimization for Large Language Models with a Human in the Loop

【速读】: 该论文试图解决手动提示工程(prompt engineering)中依赖于提示作者技能的问题,并提出了一种交互式提示优化系统(Interactive Prompt Optimization system, iPrOp)。该系统的关键在于通过人机交互的方式,允许用户在优化过程中评估和选择提示变体,结合大语言模型的预测、解释和性能指标,从而根据个人需求进一步精炼提示。这种方法不仅帮助非技术领域的专家生成适合其特定任务或领域的最佳提示,还支持研究影响提示优化性能的内在参数。

链接: https://arxiv.org/abs/2412.12644
作者: Jiahui Li,Roman Klinger
机构: University of Bamberg (班贝格大学)
关键词: made significant contributions, Automatic prompt optimization, prompt optimization, Prompt, Interactive Prompt Optimization
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Prompt engineering has made significant contributions to the era of large language models, yet its effectiveness depends on the skills of a prompt author. Automatic prompt optimization can support the prompt development process, but requires annotated data. This paper introduces \textitiPrOp , a novel Interactive Prompt Optimization system, to bridge manual prompt engineering and automatic prompt optimization. With human intervention in the optimization loop, \textitiPrOp offers users the flexibility to assess evolving prompts. We present users with prompt variations, selected instances, large language model predictions accompanied by corresponding explanations, and performance metrics derived from a subset of the training data. This approach empowers users to choose and further refine the provided prompts based on their individual preferences and needs. This system not only assists non-technical domain experts in generating optimal prompts tailored to their specific tasks or domains, but also enables to study the intrinsic parameters that influence the performance of prompt optimization. Our evaluation shows that our system has the capability to generate improved prompts, leading to enhanced task performance.
zh

[NLP-52] LLM -based Discriminative Reasoning for Knowledge Graph Question Answering

【速读】: 该论文试图解决基于生成式预训练Transformer的大语言模型(LLMs)在知识图谱问答(KGQA)任务中由于生成式范式导致的幻觉行为问题,即模型可能产生不基于实际图谱的子图规划或推理结果。解决方案的关键在于提出了一种基于LLM的判别推理(LDR)方法,通过显式建模子图检索和答案推断过程,采用判别策略来增强LLMs检索与问题相关子图的能力,并缓解生成式范式带来的无根据推理问题。实验结果表明,该方法在WebQSP和CWQ两个广泛使用的基准测试中达到了最先进的性能。

链接: https://arxiv.org/abs/2412.12643
作者: Mufan Xu,Kehai Chen,Xuefeng Bai,Muyun Yang,Tiejun Zhao,Min Zhang
机构: Harbin Institute of Technology (哈尔滨工业大学)
关键词: generative pre-trained Transformer, Large language models, knowledge graph question-answering, Large language, pre-trained Transformer
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Large language models (LLMs) based on generative pre-trained Transformer have achieved remarkable performance on knowledge graph question-answering (KGQA) tasks. However, LLMs often produce ungrounded subgraph planning or reasoning results in KGQA due to the hallucinatory behavior brought by the generative paradigm, which may hinder the advancement of the LLM-based KGQA model. To deal with the issue, we propose a novel LLM-based Discriminative Reasoning (LDR) method to explicitly model the subgraph retrieval and answer inference process. By adopting discriminative strategies, the proposed LDR method not only enhances the capability of LLMs to retrieve question-related subgraphs but also alleviates the issue of ungrounded reasoning brought by the generative paradigm of LLMs. Experimental results show that the proposed approach outperforms multiple strong comparison methods, along with achieving state-of-the-art performance on two widely used WebQSP and CWQ benchmarks.
zh

[NLP-53] Falcon: Faster and Parallel Inference of Large Language Models through Enhanced Semi-Autoregressive Drafting and Custom-Designed Decoding Tree AAAI2025

【速读】: 该论文试图解决在推测性解码(speculative decoding)中,如何在最小化草稿延迟(drafting latency)和提高推测准确性(speculation accuracy)之间取得最佳平衡,以提升大语言模型(Large Language Models)的推理速度。解决方案的关键在于引入Falcon框架,该框架采用半自回归推测性解码(semi-autoregressive speculative decoding),并通过以下创新技术增强解码效率和质量:1) 结合顺序 glancing 蒸馏技术(Coupled Sequential Glancing Distillation),强化同一块内 token 间的依赖关系,提高推测准确性;2) 设计自定义解码树(Custom-Designed Decoding Tree),允许在一次前向传递中生成多个 token,并支持多次前向传递,从而显著提高草稿 token 数量和整体接受率。这些技术使得Falcon在保持紧凑解码器架构(仅相当于两个 Transformer 层)的同时,实现了无损加速比(2.91x 到 3.51x),超越了现有的推测性解码方法。

链接: https://arxiv.org/abs/2412.12639
作者: Xiangxiang Gao,Weisheng Xie,Yiwei Xiang,Feng Ji
机构: 未知
关键词: Large Language Models, Language Models remains, minimal drafting latency, Large Language, Striking an optimal
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: AAAI 2025 Accepted

点击查看摘要

Abstract:Striking an optimal balance between minimal drafting latency and high speculation accuracy to enhance the inference speed of Large Language Models remains a significant challenge in speculative decoding. In this paper, we introduce Falcon, an innovative semi-autoregressive speculative decoding framework fashioned to augment both the drafter’s parallelism and output quality. Falcon incorporates the Coupled Sequential Glancing Distillation technique, which fortifies inter-token dependencies within the same block, leading to increased speculation accuracy. We offer a comprehensive theoretical analysis to illuminate the underlying mechanisms. Additionally, we introduce a Custom-Designed Decoding Tree, which permits the drafter to generate multiple tokens in a single forward pass and accommodates multiple forward passes as needed, thereby boosting the number of drafted tokens and significantly improving the overall acceptance rate. Comprehensive evaluations on benchmark datasets such as MT-Bench, HumanEval, and GSM8K demonstrate Falcon’s superior acceleration capabilities. The framework achieves a lossless speedup ratio ranging from 2.91x to 3.51x when tested on the Vicuna and LLaMA2-Chat model series. These results outstrip existing speculative decoding methods for LLMs, including Eagle, Medusa, Lookahead, SPS, and PLD, while maintaining a compact drafter architecture equivalent to merely two Transformer layers.
zh

[NLP-54] What External Knowledge is Preferred by LLM s? Characterizing and Exploring Chain of Evidence in Imperfect Context

【速读】: 该论文试图解决在大语言模型 (LLMs) 中引入外部知识时,由于外部知识的不完美性(包含无关信息或错误信息)导致的模型响应可靠性下降的问题。解决方案的关键在于提出了一种基于证据链 (Chain of Evidence, CoE) 的自动化判别方法,该方法要求所选外部知识不仅与问题相关,而且知识片段之间相互支持。通过这种方法,论文展示了 CoE 在提升生成准确性、增强答案忠实度、提高对知识冲突的鲁棒性以及在检索增强生成 (RAG) 场景中的实用性。

链接: https://arxiv.org/abs/2412.12632
作者: Zhiyuan Chang,Mingyang Li,Xiaojun Jia,Junjie Wang,Yuekai Huang,Qing Wang,Yihao Huang,Yang Liu
机构: State Key Laboratory of Intelligent Game, Beijing, China; Science and Technology on Integrated Information System Laboratory, Institute of Software Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences; Nanyang Technological University
关键词: large language models, Incorporating external knowledge, Incorporating external, mitigate outdated knowledge, language models
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: 12 pages, 4 figures

点击查看摘要

Abstract:Incorporating external knowledge into large language models (LLMs) has emerged as a promising approach to mitigate outdated knowledge and hallucination in LLMs. However, external knowledge is often imperfect. In addition to useful knowledge, external knowledge is rich in irrelevant or misinformation in the context that can impair the reliability of LLM responses. This paper focuses on LLMs’ preferred external knowledge in imperfect contexts when handling multi-hop QA. Inspired by criminal procedural law’s Chain of Evidence (CoE), we characterize that knowledge preferred by LLMs should maintain both relevance to the question and mutual support among knowledge pieces. Accordingly, we propose an automated CoE discrimination approach and explore LLMs’ preferences from their effectiveness, faithfulness and robustness, as well as CoE’s usability in a naive Retrieval-Augmented Generation (RAG) case. The evaluation on five LLMs reveals that CoE enhances LLMs through more accurate generation, stronger answer faithfulness, better robustness against knowledge conflict, and improved performance in a popular RAG case.
zh

[NLP-55] Make Imagination Clearer! Stable Diffusion-based Visual Imagination for Multimodal Machine Translation

【速读】: 该论文试图解决机器翻译 (MT) 中视觉信息利用的瓶颈问题,即现有方法依赖于大量带有手动图像标注的双语平行句对。解决方案的关键在于引入基于稳定扩散的想象网络 (stable diffusion-based imagination network) 到多模态大语言模型 (MLLM) 中,通过生成与源句子对应的图像来增强多模态机器翻译。特别地,通过构建启发式人类反馈与强化学习相结合的方法,确保生成的图像与源句子的一致性,而无需图像标注的监督,从而突破了视觉信息在机器翻译中的使用限制。此外,该方法还使得想象视觉信息能够集成到大规模纯文本机器翻译中,实验结果表明该模型在多模态和纯文本机器翻译任务中均显著优于现有方法。

链接: https://arxiv.org/abs/2412.12627
作者: Andong Chen,Yuchen Song,Kehai Chen,Muyun Yang,Tiejun Zhao,Min Zhang
机构: School of Computer Science and Technology, Harbin Institute of Technology, China
关键词: enhancing machine translation, effectiveness heavily relies, bilingual parallel sentence, parallel sentence pairs, manual image annotations
类目: Computation and Language (cs.CL)
备注: Work in progress

点击查看摘要

Abstract:Visual information has been introduced for enhancing machine translation (MT), and its effectiveness heavily relies on the availability of large amounts of bilingual parallel sentence pairs with manual image annotations. In this paper, we introduce a stable diffusion-based imagination network into a multimodal large language model (MLLM) to explicitly generate an image for each source sentence, thereby advancing the multimodel MT. Particularly, we build heuristic human feedback with reinforcement learning to ensure the consistency of the generated image with the source sentence without the supervision of image annotation, which breaks the bottleneck of using visual information in MT. Furthermore, the proposed method enables imaginative visual information to be integrated into large-scale text-only MT in addition to multimodal MT. Experimental results show that our model significantly outperforms existing multimodal MT and text-only MT, especially achieving an average improvement of more than 14 BLEU points on Multi30K multimodal MT benchmarks.
zh

[NLP-56] Jailbreaking? One Step Is Enough!

【速读】: 该论文试图解决大型语言模型 (Large Language Models, LLMs) 在面对jailbreak攻击时的脆弱性问题,特别是当前攻击与防御之间的独立对抗过程导致频繁的攻击迭代和针对不同模型的攻击策略重设计。解决方案的关键在于提出了逆向嵌入防御攻击 (Reverse Embedded Defense Attack, REDA) 机制,通过伪装攻击意图为“防御”意图,引导模型将有害内容嵌入其防御措施中,从而使有害内容被降级为次要角色,并让模型误以为自己在执行防御任务。REDA利用上下文学习 (In-context Learning, ICL) 结合少量攻击示例来增强模型在“防御”意图上的信心和指导,实现了跨模型攻击,无需为不同模型重新设计攻击策略,并在一次迭代中成功实现jailbreak,性能优于现有方法。

链接: https://arxiv.org/abs/2412.12621
作者: Weixiong Zheng,Peijian Zeng,Yiwei Li,Hongyan Wu,Nankai Lin,Junhao Chen,Aimin Yang,Yongmei Zhou
机构: Guangdong University of Technology(广东工业大学); Lingnan Normal University(岭南师范学院); National University of Defense Technology(国防科技大学); Guangdong University of Foreign Studies(广东外语外贸大学)
关键词: Large language models, Large language, adversaries manipulate prompts, generate harmful outputs, remain vulnerable
类目: Computation and Language (cs.CL)
备注: 17 pages

点击查看摘要

Abstract:Large language models (LLMs) excel in various tasks but remain vulnerable to jailbreak attacks, where adversaries manipulate prompts to generate harmful outputs. Examining jailbreak prompts helps uncover the shortcomings of LLMs. However, current jailbreak methods and the target model’s defenses are engaged in an independent and adversarial process, resulting in the need for frequent attack iterations and redesigning attacks for different models. To address these gaps, we propose a Reverse Embedded Defense Attack (REDA) mechanism that disguises the attack intention as the “defense”. intention against harmful content. Specifically, REDA starts from the target response, guiding the model to embed harmful content within its defensive measures, thereby relegating harmful content to a secondary role and making the model believe it is performing a defensive task. The attacking model considers that it is guiding the target model to deal with harmful content, while the target model thinks it is performing a defensive task, creating an illusion of cooperation between the two. Additionally, to enhance the model’s confidence and guidance in “defensive” intentions, we adopt in-context learning (ICL) with a small number of attack examples and construct a corresponding dataset of attack examples. Extensive evaluations demonstrate that the REDA method enables cross-model attacks without the need to redesign attack strategies for different models, enables successful jailbreak in one iteration, and outperforms existing methods on both open-source and closed-source models.
zh

[NLP-57] SynthCypher: A Fully Synthetic Data Generation Framework for Text-to-Cypher Querying in Knowledge Graphs

【速读】: 该论文试图解决图数据库查询语言Cypher的生成问题,即从自然语言生成Cypher查询(Text2Cypher),这一领域相对未被充分研究。解决方案的关键在于引入了一个名为SynthCypher的全自动合成数据生成管道,采用了一种新颖的LLM监督生成-验证框架(LLM-Supervised Generation-Verification framework),确保生成的Cypher查询在语法和语义上正确,并适用于不同领域和复杂度的查询。通过这一管道生成的SynthCypher数据集,包含29.8k个Text2Cypher实例,用于微调开源大语言模型(LLMs),如LLaMa-3.1-8B、Mistral-7B和QWEN-7B,显著提升了Text2Cypher任务的性能,分别在Text2Cypher测试集和适配图数据库的SPIDER基准上提升了40%和30%。

链接: https://arxiv.org/abs/2412.12612
作者: Aman Tiwari,Shiva Krishna Reddy Malay,Vikas Yadav,Masoud Hashemi,Sathwik Tejaswi Madhusudhan
机构: ServiceNow
关键词: enabling graph-based analytics, graph databases, plays a critical, critical role, role in enabling
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Information Retrieval (cs.IR); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:Cypher, the query language for Neo4j graph databases, plays a critical role in enabling graph-based analytics and data exploration. While substantial research has been dedicated to natural language to SQL query generation (Text2SQL), the analogous problem for graph databases referred to as Text2Cypher remains underexplored. In this work, we introduce SynthCypher, a fully synthetic and automated data generation pipeline designed to address this gap. SynthCypher employs a novel LLMSupervised Generation-Verification framework, ensuring syntactically and semantically correct Cypher queries across diverse domains and query complexities. Using this pipeline, we create SynthCypher Dataset, a large-scale benchmark containing 29.8k Text2Cypher instances. Fine-tuning open-source large language models (LLMs), including LLaMa-3.1- 8B, Mistral-7B, and QWEN-7B, on SynthCypher yields significant performance improvements of up to 40% on the Text2Cypher test set and 30% on the SPIDER benchmark adapted for graph databases. This work demonstrates that high-quality synthetic data can effectively advance the state-of-the-art in Text2Cypher tasks.
zh

[NLP-58] MultiLingPoT: Enhancing Mathematical Reasoning with Multilingual Program Fine-tuning

【速读】: 该论文试图解决当前程序化思维(Program-of-Thought, PoT)研究中仅关注单一编程语言的问题,忽略了不同编程语言在解决特定问题上的差异。解决方案的关键在于提出了多语言程序推理方法(MultiLingPoT),通过在多语言数据上进行微调,使模型能够使用多种编程语言来回答问题。此外,该方法采用了前后混合策略,帮助模型为每个问题选择最合适的编程语言。实验结果表明,MultiLingPoT的训练使每个程序的数学推理能力提升了约2.5%,并且在适当混合语言的情况下,性能可进一步提升,相比单一语言的PoT方法提高了6%。

链接: https://arxiv.org/abs/2412.12609
作者: Nianqi Li,Zujie Liang,Siyu Yuan,Jiaqing Liang,Feng Wei,Yanghua Xiao
机构: 未知
关键词: programming languages, solve mathematical problems, intermediate step, LLMs to solve, programming
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Program-of-Thought (PoT), which aims to use programming language instead of natural language as an intermediate step in reasoning, is an important way for LLMs to solve mathematical problems. Since different programming languages excel in different areas, it is natural to use the most suitable language for solving specific problems. However, current PoT research only focuses on single language PoT, ignoring the differences between different programming languages. Therefore, this paper proposes an multilingual program reasoning method, MultiLingPoT. This method allows the model to answer questions using multiple programming languages by fine-tuning on multilingual data. Additionally, prior and posterior hybrid methods are used to help the model select the most suitable language for each problem. Our experimental results show that the training of MultiLingPoT improves each program’s mathematical reasoning by about 2.5%. Moreover, with proper mixing, the performance of MultiLingPoT can be further improved, achieving a 6% increase compared to the single-language PoT with the data this http URL of this paper can be found at this https URL.
zh

[NLP-59] Multi-Dimensional Insights: Benchmarking Real-World Personalization in Large Multimodal Models

【速读】: 该论文试图解决现有大型多模态模型 (Large Multimodal Models, LMMs) 基准测试无法全面、客观、准确评估模型在真实场景中满足人类多样化需求的问题。解决方案的关键在于提出了多维度洞察 (Multi-Dimensional Insights, MDI) 基准测试,该基准包含超过500张覆盖六种常见生活场景的图片,并通过两种类型的问题(简单问题和复杂问题)以及按年龄分层的评估方式,详细评估模型在不同年龄段用户需求和偏好下的表现。这一设计使得MDI-Benchmark能够更全面地衡量LMMs在实际应用中的个性化适应能力。

链接: https://arxiv.org/abs/2412.12606
作者: YiFan Zhang,Shanglin Lei,Runqi Qiao,Zhuoma GongQue,Xiaoshuai Song,Guanting Dong,Qiuna Tan,Zhe Wei,Peiqing Yang,Ye Tian,Yadong Xue,Xiaofei Wang,Honggang Zhang
机构: Beijing University of Posts and Telecommunications(北京邮电大学); Huazhong University of Science and Technology(华中科技大学)
关键词: rapidly developing field, large multimodal models, rapidly developing, developing field, field of large
类目: Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Computer Vision and Pattern Recognition (cs.CV)
备注: 33 pages, 33 figures, Work in progress

点击查看摘要

Abstract:The rapidly developing field of large multimodal models (LMMs) has led to the emergence of diverse models with remarkable capabilities. However, existing benchmarks fail to comprehensively, objectively and accurately evaluate whether LMMs align with the diverse needs of humans in real-world scenarios. To bridge this gap, we propose the Multi-Dimensional Insights (MDI) benchmark, which includes over 500 images covering six common scenarios of human life. Notably, the MDI-Benchmark offers two significant advantages over existing evaluations: (1) Each image is accompanied by two types of questions: simple questions to assess the model’s understanding of the image, and complex questions to evaluate the model’s ability to analyze and reason beyond basic content. (2) Recognizing that people of different age groups have varying needs and perspectives when faced with the same scenario, our benchmark stratifies questions into three age categories: young people, middle-aged people, and older people. This design allows for a detailed assessment of LMMs’ capabilities in meeting the preferences and needs of different age groups. With MDI-Benchmark, the strong model like GPT-4o achieve 79% accuracy on age-related tasks, indicating that existing LMMs still have considerable room for improvement in addressing real-world applications. Looking ahead, we anticipate that the MDI-Benchmark will open new pathways for aligning real-world personalization in LMMs. The MDI-Benchmark data and evaluation code are available at this https URL
zh

[NLP-60] LLM s are Also Effective Embedding Models: An In-depth Overview

【速读】: 该论文试图解决如何将大型语言模型(LLMs)有效地应用于嵌入模型的问题,并探讨了从传统的编码器模型(如ELMo和BERT)向解码器模型(如GPT、LLaMA和Mistral)的范式转变。解决方案的关键在于两种主要策略:1) 直接提示(Direct prompting),通过设计有效的提示来生成竞争性的嵌入;2) 数据中心化调优(Data-centric tuning),涵盖模型架构、训练目标和数据构建等多个方面,以优化嵌入模型的性能。此外,论文还讨论了处理长文本、多语言和跨模态数据的高级方法,以及影响嵌入模型选择的各种因素,如性能与效率的权衡、密集与稀疏嵌入、池化策略和扩展规律。最后,论文指出了在将LLMs应用于嵌入模型时面临的挑战和局限性,为研究人员和实践者提供了全面的框架和未来研究方向。

链接: https://arxiv.org/abs/2412.12591
作者: Chongyang Tao,Tao Shen,Shen Gao,Junshuo Zhang,Zhen Li,Zhengwei Tao,Shuai Ma
机构: Beihang University(北京航空航天大学); University of Technology Sydney(悉尼科技大学); University of Electronic Science and Technology of China(中国电子科技大学); Peking University(北京大学)
关键词: Large language models, natural language processing, revolutionized natural language, Large language, natural language
类目: Computation and Language (cs.CL)
备注: 32 pages

点击查看摘要

Abstract:Large language models (LLMs) have revolutionized natural language processing by achieving state-of-the-art performance across various tasks. Recently, their effectiveness as embedding models has gained attention, marking a paradigm shift from traditional encoder-only models like ELMo and BERT to decoder-only, large-scale LLMs such as GPT, LLaMA, and Mistral. This survey provides an in-depth overview of this transition, beginning with foundational techniques before the LLM era, followed by LLM-based embedding models through two main strategies to derive embeddings from LLMs. 1) Direct prompting: We mainly discuss the prompt designs and the underlying rationale for deriving competitive embeddings. 2) Data-centric tuning: We cover extensive aspects that affect tuning an embedding model, including model architecture, training objectives, data constructions, etc. Upon the above, we also cover advanced methods, such as handling longer texts, and multilingual and cross-modal data. Furthermore, we discuss factors affecting choices of embedding models, such as performance/efficiency comparisons, dense vs sparse embeddings, pooling strategies, and scaling law. Lastly, the survey highlights the limitations and challenges in adapting LLMs for embeddings, including cross-task embedding quality, trade-offs between efficiency and accuracy, low-resource, long-context, data bias, robustness, etc. This survey serves as a valuable resource for researchers and practitioners by synthesizing current advancements, highlighting key challenges, and offering a comprehensive framework for future work aimed at enhancing the effectiveness and efficiency of LLMs as embedding models.
zh

[NLP-61] PerSphere: A Comprehensive Framework for Multi-Faceted Perspective Retrieval and Summarization

【速读】: 该论文试图解决在线平台和推荐算法导致的“回音壁”现象,即用户被局限在单一视角的信息茧房中,从而产生偏见性理解的问题。解决方案的关键在于引入PerSphere基准,通过多方面的视角检索和总结来打破这些信息茧房。具体而言,PerSphere针对每个查询提供两个对立的主张,每个主张由来自一个或多个文档的不同且不重叠的视角支持。解决方案的核心是一个两步的端到端流程,包括全面的文档检索和多方面的总结,并通过一组评估指标来衡量检索和总结的全面性。实验表明,现有模型在处理这一复杂任务时表现不佳,主要挑战在于长文本和视角提取。为此,论文提出了一种简单但有效的多代理总结系统,为提升PerSphere上的性能提供了有前景的解决方案。

链接: https://arxiv.org/abs/2412.12588
作者: Yun Luo,Yingjie Li,Xiangkun Hu,Qinglin Qi,Fang Guo,Qipeng Guo,Zheng Zhang,Yue Zhang
机构: Westlake University(西湖大学); Amazon AWS AI(亚马逊AWS AI); Sichuan University(四川大学); Shanghai AI Lab(上海人工智能实验室); New York University Shanghai(上海纽约大学); Westlake Institute for Advanced Study(西湖高等研究院)
关键词: recommendation algorithms evolve, algorithms evolve, people are increasingly, echo chambers, leading to biased
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:As online platforms and recommendation algorithms evolve, people are increasingly trapped in echo chambers, leading to biased understandings of various issues. To combat this issue, we have introduced PerSphere, a benchmark designed to facilitate multi-faceted perspective retrieval and summarization, thus breaking free from these information silos. For each query within PerSphere, there are two opposing claims, each supported by distinct, non-overlapping perspectives drawn from one or more documents. Our goal is to accurately summarize these documents, aligning the summaries with the respective claims and their underlying perspectives. This task is structured as a two-step end-to-end pipeline that includes comprehensive document retrieval and multi-faceted summarization. Furthermore, we propose a set of metrics to evaluate the comprehensiveness of the retrieval and summarization content. Experimental results on various counterparts for the pipeline show that recent models struggle with such a complex task. Analysis shows that the main challenge lies in long context and perspective extraction, and we propose a simple but effective multi-agent summarization system, offering a promising solution to enhance performance on PerSphere.
zh

[NLP-62] Process-Supervised Reward Models for Clinical Note Generation: A Scalable Approach Guided by Domain Expertise

【速读】: 该论文试图解决在临床笔记生成任务中,如何通过过程监督奖励模型 (Process-supervised Reward Models, PRMs) 提供细粒度的奖励信号,以提高生成式大语言模型 (LLM) 的输出质量。解决方案的关键在于:1) 基于临床专家的指导,精心设计临床笔记的步骤定义;2) 利用 Gemini-Pro 1.5 大规模自动生成过程监督数据;3) 在 LLaMA-3.1 8B 指令模型上训练 PRM,并通过实验验证其在选择金标准样本和医生偏好笔记方面的优越性能。此外,论文还通过消融实验优化损失函数和数据选择策略,并通过医生阅读研究探索下游 Best-of-N 性能的预测因素。

链接: https://arxiv.org/abs/2412.12583
作者: Hanyin Wang,Qiping Xu,Bolun Liu,Guleid Hussein,Hariprasad Korsapati,Mohamad El Labban,Kingsley Iheasirim,Mohamed Hassan,Gokhan Anil,Brian Bartlett,Jimeng Sun
机构: Mayo Clinic Health System(梅奥诊所健康系统); University of Illinois Urbana-Champaign(伊利诺伊大学厄巴纳-香槟分校)
关键词: verify large language, achieved significant success, Process-supervised reward models, large language model, Process-supervised reward
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Process-supervised reward models (PRMs), which verify large language model (LLM) outputs step-by-step, have achieved significant success in mathematical and coding problems. However, their application to other domains remains largely unexplored. In this work, we train a PRM to provide step-level reward signals for clinical notes generated by LLMs from patient-doctor dialogues. Guided by real-world clinician expertise, we carefully designed step definitions for clinical notes and utilized Gemini-Pro 1.5 to automatically generate process supervision data at scale. Our proposed PRM, trained on the LLaMA-3.1 8B instruct model, demonstrated superior performance compared to Gemini-Pro 1.5 and an outcome-supervised reward model (ORM) across two key evaluations: (1) the accuracy of selecting gold-reference samples from error-containing samples, achieving 98.8% (versus 61.3% for ORM and 93.8% for Gemini-Pro 1.5), and (2) the accuracy of selecting physician-preferred notes, achieving 56.2% (compared to 51.2% for ORM and 50.0% for Gemini-Pro 1.5). Additionally, we conducted ablation studies to determine optimal loss functions and data selection strategies, along with physician reader studies to explore predictors of downstream Best-of-N performance. Our promising results suggest the potential of PRMs to extend beyond the clinical domain, offering a scalable and effective solution for diverse generative tasks.
zh

[NLP-63] Quantifying Lexical Semantic Shift via Unbalanced Optimal Transport

【速读】: 该论文试图解决词汇语义变化检测中现有方法在个体使用实例层面上的变化洞察不足的问题。解决方案的关键在于应用非平衡最优传输 (Unbalanced Optimal Transport, UOT) 到上下文化的词嵌入集,通过捕获使用实例之间的对齐差异来量化语义变化。具体来说,论文提出了“词义使用变化” (Sense Usage Shift, SUS) 这一度量,用于量化每个使用实例中词义使用频率的变化,从而在统一框架下解决实例级语义变化量化、词级语义变化幅度测量以及词义扩展或缩窄等问题。

链接: https://arxiv.org/abs/2412.12569
作者: Ryo Kishino,Hiroaki Yamagiwa,Ryo Nagata,Sho Yokoi,Hidetoshi Shimodaira
机构: Kyoto University(京都大学); Konan University(关南大学); Tohoku University(东北大学); RIKEN(理化学研究所)
关键词: Lexical semantic change, Lexical semantic, Unbalanced Optimal Transport, semantic change, aims to identify
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Lexical semantic change detection aims to identify shifts in word meanings over time. While existing methods using embeddings from a diachronic corpus pair estimate the degree of change for target words, they offer limited insight into changes at the level of individual usage instances. To address this, we apply Unbalanced Optimal Transport (UOT) to sets of contextualized word embeddings, capturing semantic change through the excess and deficit in the alignment between usage instances. In particular, we propose Sense Usage Shift (SUS), a measure that quantifies changes in the usage frequency of a word sense at each usage instance. By leveraging SUS, we demonstrate that several challenges in semantic change detection can be addressed in a unified manner, including quantifying instance-level semantic change and word-level tasks such as measuring the magnitude of semantic change and the broadening or narrowing of meaning.
zh

[NLP-64] FCMR: Robust Evaluation of Financial Cross-Modal Multi-Hop Reasoning

【速读】: 该论文试图解决多模态大语言模型 (Multimodal Large Language Models, MLLMs) 在跨模态多跳推理 (multi-hop reasoning) 任务中的评估不足问题。现有基准如 MMQA 存在数据污染和缺乏复杂查询的挑战,导致难以准确评估模型性能。论文提出的解决方案是引入金融跨模态多跳推理基准 (Financial Cross-Modal Multi-Hop Reasoning, FCMR),通过结合金融领域的文本报告、表格和图表信息,设计了从简单到复杂的三个难度级别(Easy、Medium、Hard),特别是 Hard 级别要求精确的跨模态三跳推理,以全面评估模型的推理能力。实验表明,即使是最先进的 MLLMs 也难以应对这一挑战,揭示了信息检索阶段的关键瓶颈。

链接: https://arxiv.org/abs/2412.12567
作者: Seunghee Kim,Changhyeon Kim,Taeuk Kim
机构: Hanyang University (汉阳大学)
关键词: Real-world decision-making, multiple modalities, reasoning, Real-world, multi-hop reasoning
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Real-world decision-making often requires integrating and reasoning over information from multiple modalities. While recent multimodal large language models (MLLMs) have shown promise in such tasks, their ability to perform multi-hop reasoning across diverse sources remains insufficiently evaluated. Existing benchmarks, such as MMQA, face challenges due to (1) data contamination and (2) a lack of complex queries that necessitate operations across more than two modalities, hindering accurate performance assessment. To address this, we present Financial Cross-Modal Multi-Hop Reasoning (FCMR), a benchmark created to analyze the reasoning capabilities of MLLMs by urging them to combine information from textual reports, tables, and charts within the financial domain. FCMR is categorized into three difficulty levels-Easy, Medium, and Hard-facilitating a step-by-step evaluation. In particular, problems at the Hard level require precise cross-modal three-hop reasoning and are designed to prevent the disregard of any modality. Experiments on this new benchmark reveal that even state-of-the-art MLLMs struggle, with the best-performing model (Claude 3.5 Sonnet) achieving only 30.4% accuracy on the most challenging tier. We also conduct analysis to provide insights into the inner workings of the models, including the discovery of a critical bottleneck in the information retrieval phase.
zh

[NLP-65] Evaluating Zero-Shot Multilingual Aspect-Based Sentiment Analysis with Large Language Models

【速读】: 该论文试图解决在多语言环境下进行基于方面的情感分析 (Aspect-based Sentiment Analysis, ABSA) 的问题,特别是探索大型语言模型 (Large Language Models, LLMs) 在零样本 (zero-shot) 条件下处理这一任务的潜力。解决方案的关键在于评估一系列 LLMs 在多语言 ABSA 任务中的表现,并研究不同的提示策略(如普通零样本、思维链 (Chain-of-Thought, CoT)、自我改进、自我辩论和自我一致性)对模型性能的影响。研究发现,尽管 LLMs 在处理多语言 ABSA 方面显示出潜力,但它们通常仍不及经过微调的专门模型。特别是,简单的零样本提示在资源丰富的语言(如英语)中往往优于更复杂的策略。这些结果强调了进一步改进基于 LLM 的方法以有效应对多语言 ABSA 任务的必要性。

链接: https://arxiv.org/abs/2412.12564
作者: Chengyan Wu,Bolei Ma,Zheyu Zhang,Ningyuan Deng,Yanqing He,Yun Xue
机构: South China Normal University (华南师范大学); Ludwig Maximilian University of Munich (慕尼黑大学); Technical University of Munich (慕尼黑工业大学); Institute of Software, Chinese Academy of Sciences (中国科学院软件研究所)
关键词: Aspect-based sentiment analysis, attracted increasing attention, Aspect-based sentiment, sequence labeling task, sentiment analysis
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Aspect-based sentiment analysis (ABSA), a sequence labeling task, has attracted increasing attention in multilingual contexts. While previous research has focused largely on fine-tuning or training models specifically for ABSA, we evaluate large language models (LLMs) under zero-shot conditions to explore their potential to tackle this challenge with minimal task-specific adaptation. We conduct a comprehensive empirical evaluation of a series of LLMs on multilingual ABSA tasks, investigating various prompting strategies, including vanilla zero-shot, chain-of-thought (CoT), self-improvement, self-debate, and self-consistency, across nine different models. Results indicate that while LLMs show promise in handling multilingual ABSA, they generally fall short of fine-tuned, task-specific models. Notably, simpler zero-shot prompts often outperform more complex strategies, especially in high-resource languages like English. These findings underscore the need for further refinement of LLM-based approaches to effectively address ABSA task across diverse languages.
zh

[NLP-66] ask-Agnostic Language Model Watermarking via High Entropy Passthrough Layers AAAI2025

【速读】: 该论文试图解决在大规模语言模型预训练成本高昂的时代,如何确保模型所有者的知识产权并负责任地部署模型的问题。解决方案的关键是通过添加passthrough layers(直通层)来实现模型水印技术,这些层在现有预训练网络中引入,并通过自监督损失进行训练,使得模型在接收到特定私有密钥时生成高熵输出,而在其他情况下正常运行。该方法与现有水印技术不同,具有完全的任务无关性,可应用于分类和序列到序列任务,且无需访问下游微调数据集。实验表明,该方法在大多数情况下实现了接近完美的水印提取精度和低误报率,同时保持了原始模型的性能,并且对下游微调、微调剪枝和层移除攻击具有鲁棒性,训练时间仅为原始模型训练时间的一小部分。

链接: https://arxiv.org/abs/2412.12563
作者: Vaden Masrani,Mohammad Akbari,David Ming Xuan Yue,Ahmad Rezaei,Yong Zhang
机构: 未知
关键词: large language models, ensuring the intellectual, responsibly deployed, increasingly important, era of costly
类目: Computation and Language (cs.CL)
备注: Accepted to AAAI2025

点击查看摘要

Abstract:In the era of costly pre-training of large language models, ensuring the intellectual property rights of model owners, and insuring that said models are responsibly deployed, is becoming increasingly important. To this end, we propose model watermarking via passthrough layers, which are added to existing pre-trained networks and trained using a self-supervised loss such that the model produces high-entropy output when prompted with a unique private key, and acts normally otherwise. Unlike existing model watermarking methods, our method is fully task-agnostic, and can be applied to both classification and sequence-to-sequence tasks without requiring advanced access to downstream fine-tuning datasets. We evaluate the proposed passthrough layers on a wide range of downstream tasks, and show experimentally our watermarking method achieves a near-perfect watermark extraction accuracy and false-positive rate in most cases without damaging original model performance. Additionally, we show our method is robust to both downstream fine-tuning, fine-pruning, and layer removal attacks, and can be trained in a fraction of the time required to train the original model. Code is available in the paper.
zh

[NLP-67] EXIT: Context-Aware Extractive Compression for Enhancing Retrieval-Augmented Generation

【速读】: 该论文试图解决在问答系统中,基于检索增强生成 (RAG) 的模型在检索相关文档时可能出现的效率和准确性问题。当前的 RAG 系统在检索模型未能正确排序最相关文档时,往往通过增加上下文信息来弥补,但这会导致延迟增加和准确性下降。论文提出的解决方案是引入一种名为 EXIT 的抽取式上下文压缩框架,其关键在于通过分类检索文档中的句子并保留其上下文依赖关系,实现可并行化的、自适应于查询复杂度和检索质量的上下文感知抽取。这种方法不仅提高了问答任务的准确性,还显著减少了推理时间和令牌数量,从而在效率和效果上均优于现有压缩方法和未压缩基线。

链接: https://arxiv.org/abs/2412.12559
作者: Taeho Hwang,Sukmin Cho,Soyeong Jeong,Hoyun Song,SeungYoon Han,Jong C. Park
机构: School of Computing, Korea Advanced Institute of Science and Technology (计算学院,韩国科学技术院); Graduate School of AI, Korea Advanced Institute of Science and Technology (人工智能研究生院,韩国科学技术院)
关键词: context compression framework, question answering, framework that enhances, Current RAG systems, RAG
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Information Retrieval (cs.IR)
备注: Under Review

点击查看摘要

Abstract:We introduce EXIT, an extractive context compression framework that enhances both the effectiveness and efficiency of retrieval-augmented generation (RAG) in question answering (QA). Current RAG systems often struggle when retrieval models fail to rank the most relevant documents, leading to the inclusion of more context at the expense of latency and accuracy. While abstractive compression methods can drastically reduce token counts, their token-by-token generation process significantly increases end-to-end latency. Conversely, existing extractive methods reduce latency but rely on independent, non-adaptive sentence selection, failing to fully utilize contextual information. EXIT addresses these limitations by classifying sentences from retrieved documents - while preserving their contextual dependencies - enabling parallelizable, context-aware extraction that adapts to query complexity and retrieval quality. Our evaluations on both single-hop and multi-hop QA tasks show that EXIT consistently surpasses existing compression methods and even uncompressed baselines in QA accuracy, while also delivering substantial reductions in inference time and token count. By improving both effectiveness and efficiency, EXIT provides a promising direction for developing scalable, high-quality QA solutions in RAG pipelines. Our code is available at this https URL
zh

[NLP-68] LLM CL-GEC: Advancing Grammatical Error Correction with LLM -Driven Curriculum Learning

【速读】: 该论文试图解决大规模语言模型 (LLMs) 在特定领域(如语法错误纠正 (GEC))中可能不如专门模型的问题。解决方案的关键在于提出了一种基于 LLM 的课程学习 (LLM-based curriculum learning) 方法,该方法利用 LLMs 强大的语义理解和判别能力来评估 GEC 训练数据的复杂度,并通过从易到难的顺序选择课程,迭代训练和优化预训练的 T5 和 LLaMA 系列模型。这种方法显著提升了在多个英语 GEC 基准测试(如 CoNLL14、BEA19 测试集和开发集)中的性能,超越了基线模型和传统课程学习方法。

链接: https://arxiv.org/abs/2412.12541
作者: Tao Fang,Derek F. Wong,Lusheng Zhang,Keyan Jin,Qiang Zhang,Tianjiao Li,Jinlong Hou,Lidia S. Chao
机构: NLP2CT Lab, Department of Computer and Information Science, University of Macau(NLP2CT实验室,计算机与信息科学系,澳门大学); Bilibili Inc., LLM Team(哔哩哔哩公司,LLM团队); Macao Polytechnic University(澳门理工学院); Xi’an Jiaotong University(西安交通大学)
关键词: natural language processing, grammatical error correction, specific natural language, demonstrated remarkable capabilities, lack proficiency compared
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: Derek F. Wong is the corresponding author. The preprint version consists of 15 Pages, 5 Figures, 5 Tables, and 3 Appendices

点击查看摘要

Abstract:While large-scale language models (LLMs) have demonstrated remarkable capabilities in specific natural language processing (NLP) tasks, they may still lack proficiency compared to specialized models in certain domains, such as grammatical error correction (GEC). Drawing inspiration from the concept of curriculum learning, we have delved into refining LLMs into proficient GEC experts by devising effective curriculum learning (CL) strategies. In this paper, we introduce a novel approach, termed LLM-based curriculum learning, which capitalizes on the robust semantic comprehension and discriminative prowess inherent in LLMs to gauge the complexity of GEC training data. Unlike traditional curriculum learning techniques, our method closely mirrors human expert-designed curriculums. Leveraging the proposed LLM-based CL method, we sequentially select varying levels of curriculums ranging from easy to hard, and iteratively train and refine using the pretrianed T5 and LLaMA series models. Through rigorous testing and analysis across diverse benchmark assessments in English GEC, including the CoNLL14 test, BEA19 test, and BEA19 development sets, our approach showcases a significant performance boost over baseline models and conventional curriculum learning methodologies.
zh

[NLP-69] When to Speak When to Abstain: Contrastive Decoding with Abstention

【速读】: 该论文试图解决大语言模型(LLMs)在缺乏相关知识时可能产生的幻觉(hallucination)问题,特别是在高风险应用中可能导致的可靠性降低和潜在风险。解决方案的关键是引入了一种无需训练的解码方法——对比解码与放弃机制(Contrastive Decoding with Abstention, CDA)。CDA通过评估每条知识与给定查询的相关性,动态决定优先使用哪些知识或完全忽略哪些知识,从而在知识可用时生成准确响应,在知识缺失时选择放弃生成,以提高模型的可靠性和用户信任。

链接: https://arxiv.org/abs/2412.12527
作者: Hyuhng Joon Kim,Youna Kim,Sang-goo Lee,Taeuk Kim
机构: Seoul National University(首尔国立大学); IntelliSys, Korea(韩国IntelliSys); Hanyang University(汉阳大学)
关键词: Large Language Models, Large Language, Language Models, demonstrate exceptional performance, exceptional performance
类目: Computation and Language (cs.CL)
备注: under-review

点击查看摘要

Abstract:Large Language Models (LLMs) demonstrate exceptional performance across diverse tasks by leveraging both pre-trained knowledge (i.e., parametric knowledge) and external knowledge (i.e., contextual knowledge). While substantial efforts have been made to leverage both forms of knowledge, scenarios in which the model lacks any relevant knowledge remain underexplored. Such limitations can result in issues like hallucination, causing reduced reliability and potential risks in high-stakes applications. To address such limitations, this paper extends the task scope to encompass cases where the user’s request cannot be fulfilled due to the lack of relevant knowledge. To this end, we introduce Contrastive Decoding with Abstention (CDA), a training-free decoding method that empowers LLMs to generate responses when relevant knowledge is available and to abstain otherwise. CDA evaluates the relevance of each knowledge for a given query, adaptively determining which knowledge to prioritize or which to completely ignore. Extensive experiments with four LLMs on three question-answering datasets demonstrate that CDA can effectively perform accurate generation and abstention simultaneously. These findings highlight CDA’s potential to broaden the applicability of LLMs, enhancing reliability and preserving user trust.
zh

[NLP-70] Solid-SQL: Enhanced Schema-linking based In-context Learning for Robust Text-to-SQL COLING2025

【速读】: 该论文试图解决大语言模型(LLMs)驱动的文本到SQL系统在面对对抗性扰动时鲁棒性不足的问题。解决方案的关键在于提出了一种名为Solid-SQL的鲁棒文本到SQL解决方案,该方案通过两个主要策略提升系统的鲁棒性:首先,在预处理阶段训练了一个增强的架构链接模型(schema-linking model),并结合基于LLM的数据增强技术;其次,设计了一种基于结构相似性的两轮示例检索策略,用于上下文学习。这些改进使得Solid-SQL在标准和对抗性数据集上均达到了最先进的SQL执行准确率,并在扰动数据集上相较于基线方法平均提升了11.6%的性能。

链接: https://arxiv.org/abs/2412.12522
作者: Geling Liu,Yunzhi Tan,Ruichao Zhong,Yuanzhen Xie,Lingchen Zhao,Qian Wang,Bo Hu,Zang Li
机构: School of Cyber Science and Engineering, Wuhan University, China; Big Data and AI Platform Department, Tencent, China
关键词: large language models, large language, significantly improved, improved the performance, Recently
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: Accepted at COLING 2025 Main

点击查看摘要

Abstract:Recently, large language models (LLMs) have significantly improved the performance of text-to-SQL systems. Nevertheless, many state-of-the-art (SOTA) approaches have overlooked the critical aspect of system robustness. Our experiments reveal that while LLM-driven methods excel on standard datasets, their accuracy is notably compromised when faced with adversarial perturbations. To address this challenge, we propose a robust text-to-SQL solution, called Solid-SQL, designed to integrate with various LLMs. We focus on the pre-processing stage, training a robust schema-linking model enhanced by LLM-based data augmentation. Additionally, we design a two-round, structural similarity-based example retrieval strategy for in-context learning. Our method achieves SOTA SQL execution accuracy levels of 82.1% and 58.9% on the general Spider and Bird benchmarks, respectively. Furthermore, experimental results show that Solid-SQL delivers an average improvement of 11.6% compared to baselines on the perturbed Spider-Syn, Spider-Realistic, and Dr. Spider benchmarks.
zh

[NLP-71] Can Large Language Models Understand You Better? An MBTI Personality Detection Dataset Aligned with Population Traits COLING2025

【速读】: 该论文试图解决MBTI人格检测任务中的两个主要问题:(1) 现有数据集中自我报告标签导致的错误标注问题,占比达29.58%;(2) 硬标签无法捕捉人口性格分布的完整范围。解决方案的关键在于构建了MBTIBench,这是首个由心理学家指导的手动标注的高质量MBTI人格检测数据集,并采用软标签来估计样本的极性倾向。通过这种方式,MBTIBench不仅解决了错误标注问题,还确认了更多人群具有非极端性格特质,从而为其他心理任务提供了比硬标签更具优势的软标签。

链接: https://arxiv.org/abs/2412.12510
作者: Bohan Li,Jiannan Guan,Longxu Dou,Yunlong Feng,Dingzirui Wang,Yang Xu,Enbo Wang,Qiguang Chen,Bichen Wang,Xiao Xu,Yimeng Zhang,Libo Qin,Yanyan Zhao,Qingfu Zhu,Wanxiang Che
机构: Harbin Institute of Technology(哈尔滨工业大学); Individual Researcher(个人研究者); Central South University(中南大学)
关键词: Myers-Briggs Type Indicator, Type Indicator, theories reflecting individual, reflecting individual differences, MBTI personality detection
类目: Computation and Language (cs.CL); Computers and Society (cs.CY)
备注: Accepted by COLING 2025. 28 papges, 20 figures, 10 tables

点击查看摘要

Abstract:The Myers-Briggs Type Indicator (MBTI) is one of the most influential personality theories reflecting individual differences in thinking, feeling, and behaving. MBTI personality detection has garnered considerable research interest and has evolved significantly over the years. However, this task tends to be overly optimistic, as it currently does not align well with the natural distribution of population personality traits. Specifically, (1) the self-reported labels in existing datasets result in incorrect labeling issues, and (2) the hard labels fail to capture the full range of population personality distributions. In this paper, we optimize the task by constructing MBTIBench, the first manually annotated high-quality MBTI personality detection dataset with soft labels, under the guidance of psychologists. As for the first challenge, MBTIBench effectively solves the incorrect labeling issues, which account for 29.58% of the data. As for the second challenge, we estimate soft labels by deriving the polarity tendency of samples. The obtained soft labels confirm that there are more people with non-extreme personality traits. Experimental results not only highlight the polarized predictions and biases in LLMs as key directions for future research, but also confirm that soft labels can provide more benefits to other psychological tasks than hard labels. The code and data are available at this https URL.
zh

[NLP-72] Can You Trust LLM Judgments? Reliability of LLM -as-a-Judge

【速读】: 该论文试图解决大型语言模型(LLMs)输出可靠性问题,特别是由于其随机性导致的不可靠性。解决方案的关键在于引入一种新的评估框架,利用McDonald’s omega来严格评估LLM作为判断者的可靠性。论文通过在标准单轮和多轮基准上评估LLM对其他LLM输出的判断,同时研究温度参数对可靠性的影响,展示了固定随机性的局限性,并强调了考虑多样本的重要性。这一研究强调了对LLM可靠性进行细致理解以及避免过度依赖单次评估的必要性,为构建更可信和可靠的LLM系统提供了重要步骤。

链接: https://arxiv.org/abs/2412.12509
作者: Kayla Schroeder,Zach Wood-Doughty
机构: Northwestern University (西北大学)
关键词: Large Language Models, Large Language, stochastic nature poses, nature poses challenges, Language Models
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Large Language Models (LLMs) have become increasingly powerful and ubiquitous, but their stochastic nature poses challenges to the reliability of their outputs. While deterministic settings can improve consistency, they do not guarantee reliability, as a single sample from the model’s probability distribution can still be misleading. Building upon the concept of LLM-as-a-judge, we introduce a novel framework for rigorously evaluating the reliability of LLM judgments, leveraging McDonald’s omega. We evaluate the reliability of LLMs when judging the outputs of other LLMs on standard single-turn and multi-turn benchmarks, simultaneously investigating the impact of temperature on reliability. By analyzing these results, we demonstrate the limitations of fixed randomness and the importance of considering multiple samples, which we show has significant implications for downstream applications. Our findings highlight the need for a nuanced understanding of LLM reliability and the potential risks associated with over-reliance on single-shot evaluations. This work provides a crucial step towards building more trustworthy and reliable LLM-based systems and applications.
zh

[NLP-73] DocFusion: A Unified Framework for Document Parsing Tasks

【速读】: 该论文试图解决现有文档解析方法中因集成多个独立模型而导致的高复杂性和维护开销问题。解决方案的关键在于提出了DocFusion,一个仅包含0.28B参数的轻量级生成模型。DocFusion通过统一任务表示和改进的目标函数实现协同训练,利用识别任务之间的相互促进作用,显著提升了检测性能,并在四个关键任务上达到了最先进的性能(SOTA)。

链接: https://arxiv.org/abs/2412.12505
作者: Mingxu Chai,Ziyu Shen,Chong Zhang,Yue Zhang,Xiao Wang,Shihan Dou,Jihua Kang,Jiazheng Zhang,Qi Zhang
机构: Fudan University(复旦大学); Shanghai Innovation Institute(上海创新研究院)
关键词: analyzing complex document, complex document structures, extracting fine-grained information, supporting numerous downstream, numerous downstream applications
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Document parsing is essential for analyzing complex document structures and extracting fine-grained information, supporting numerous downstream applications. However, existing methods often require integrating multiple independent models to handle various parsing tasks, leading to high complexity and maintenance overhead. To address this, we propose DocFusion, a lightweight generative model with only 0.28B parameters. It unifies task representations and achieves collaborative training through an improved objective function. Experiments reveal and leverage the mutually beneficial interaction among recognition tasks, and integrating recognition data significantly enhances detection performance. The final results demonstrate that DocFusion achieves state-of-the-art (SOTA) performance across four key tasks.
zh

[NLP-74] Unleashing the Potential of Model Bias for Generalized Category Discovery AAAI2025

【速读】: 该论文试图解决广义类别发现 (Generalized Category Discovery) 中的两个主要问题:模型在预训练阶段对已知类别的偏差 (model bias) 以及对新类别缺乏精确监督导致的类别混淆。解决方案的关键在于提出了一种名为自去偏校准 (Self-Debiasing Calibration, SDC) 的新框架,该框架通过利用偏差模型的输出来实现两个目标:一是准确建模类别偏差,并用于校准当前训练模型的输出,从而减少对已知类别的偏差;二是通过相似类别间的知识迁移,帮助区分不同的新类别。SDC通过动态调整当前训练模型的输出logits,生成更少偏差的输出和更准确的伪标签,从而有效缓解新类别的类别混淆问题。

链接: https://arxiv.org/abs/2412.12501
作者: Wenbin An,Haonan Lin,Jiahao Nie,Feng Tian,Wenkai Shi,Yaqiang Wu,Qianying Wang,Ping Chen
机构: 未知
关键词: Generalized Category Discovery, Generalized Category, Category Discovery, categories, Category
类目: Machine Learning (cs.LG); Computation and Language (cs.CL); Computer Vision and Pattern Recognition (cs.CV)
备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Generalized Category Discovery is a significant and complex task that aims to identify both known and undefined novel categories from a set of unlabeled data, leveraging another labeled dataset containing only known categories. The primary challenges stem from model bias induced by pre-training on only known categories and the lack of precise supervision for novel ones, leading to category bias towards known categories and category confusion among different novel categories, which hinders models’ ability to identify novel categories effectively. To address these challenges, we propose a novel framework named Self-Debiasing Calibration (SDC). Unlike prior methods that regard model bias towards known categories as an obstacle to novel category identification, SDC provides a novel insight into unleashing the potential of the bias to facilitate novel category learning. Specifically, the output of the biased model serves two key purposes. First, it provides an accurate modeling of category bias, which can be utilized to measure the degree of bias and debias the output of the current training model. Second, it offers valuable insights for distinguishing different novel categories by transferring knowledge between similar categories. Based on these insights, SDC dynamically adjusts the output logits of the current training model using the output of the biased model. This approach produces less biased logits to effectively address the issue of category bias towards known categories, and generates more accurate pseudo labels for unlabeled data, thereby mitigating category confusion for novel categories. Experiments on three benchmark datasets show that SDC outperforms SOTA methods, especially in the identification of novel categories. Our code and data are available at \urlthis https URL.
zh

[NLP-75] Beyond Data Quantity: Key Factors Driving Performance in Multilingual Language Models COLING2025

【速读】: 该论文旨在解决多语言语言模型 (Multilingual Language Models, MLLMs) 在不同语言间表现差异的问题,尤其是由于资源可用性和语言特性差异导致的性能不均衡。研究的关键在于识别影响 MLLM 性能的关键因素,除了已知的预训练数据比例和模型大小外,还发现了两个重要因素:词元相似性 (token similarity)国家相似性 (country similarity)。词元相似性有助于跨语言迁移,而国家相似性则强调了共享文化和语言背景的重要性。这些发现为开发更公平和有效的多语言语言模型,特别是对资源匮乏的语言,提供了重要的指导。

链接: https://arxiv.org/abs/2412.12500
作者: Sina Bagheri Nezhad,Ameeta Agrawal,Rhitabrat Pokharel
机构: Portland State University, USA (波特兰州立大学,美国)
关键词: show performance disparities, performance disparities due, Multilingual language models, crucial for handling, handling text
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: Accepted at The First Workshop on Language Models for Low-Resource Languages @ COLING 2025

点击查看摘要

Abstract:Multilingual language models (MLLMs) are crucial for handling text across various languages, yet they often show performance disparities due to differences in resource availability and linguistic characteristics. While the impact of pre-train data percentage and model size on performance is well-known, our study reveals additional critical factors that significantly influence MLLM effectiveness. Analyzing a wide range of features, including geographical, linguistic, and resource-related aspects, we focus on the SIB-200 dataset for classification and the Flores-200 dataset for machine translation, using regression models and SHAP values across 204 languages. Our findings identify token similarity and country similarity as pivotal factors, alongside pre-train data and model size, in enhancing model performance. Token similarity facilitates cross-lingual transfer, while country similarity highlights the importance of shared cultural and linguistic contexts. These insights offer valuable guidance for developing more equitable and effective multilingual language models, particularly for underrepresented languages.
zh

[NLP-76] LinguaLIFT: An Effective Two-stage Instruction Tuning Framework for Low-Resource Language Tasks

【速读】: 该论文试图解决多语言任务中高资源语言与低资源语言之间的性能差距问题,尤其是在预训练语料库中语言不平衡的情况下。解决方案的关键是提出了一种两阶段的指令微调框架,称为LinguaLIFT。首先,通过在大型语言模型(LLM)中集成一个额外的语言对齐层,并使用代码切换的微调方法来增强多语言对齐。其次,在冻结语言对齐层的同时,仅使用英语指令数据对LLM进行微调,从而实现从英语任务到低资源语言任务的能力迁移。此外,论文还引入了多语言数学世界问题(MMWP)基准,用于全面评估多语言推理能力。

链接: https://arxiv.org/abs/2412.12499
作者: Hongbin Zhang,Kehai Chen,Xuefeng Bai,Yang Xiang,Min Zhang
机构: Institute of Computing and Intelligence, Harbin Institute of Technology, Shenzhen, China(计算与智能研究所,哈尔滨工业大学,深圳,中国); Peng Cheng Laboratory, Shenzhen, China(鹏城实验室,深圳,中国)
关键词: Large language models, Large language, demonstrated impressive multilingual, impressive multilingual understanding, driven by extensive
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Large language models (LLMs) have demonstrated impressive multilingual understanding and reasoning capabilities, driven by extensive pre-training multilingual corpora and fine-tuning instruction data. However, a performance gap persists between high-resource and low-resource language tasks due to language imbalance in the pre-training corpus, even using more low-resource data during fine-tuning. To alleviate this issue, we propose LinguaLIFT, a two-stage instruction tuning framework for advancing low-resource language tasks. An additional language alignment layer is first integrated into the LLM to adapt a pre-trained multilingual encoder, thereby enhancing multilingual alignment through code-switched fine-tuning. The second stage fine-tunes LLM with English-only instruction data while freezing the language alignment layer, allowing LLM to transfer task-specific capabilities from English to low-resource language tasks. Additionally, we introduce the Multilingual Math World Problem (MMWP) benchmark, which spans 21 low-resource, 17 medium-resource, and 10 high-resource languages, enabling comprehensive evaluation of multilingual reasoning. Experimental results show that LinguaLIFT outperforms several competitive baselines across MMWP and other widely used benchmarks.
zh

[NLP-77] NLSR: Neuron-Level Safety Realignment of Large Language Models Against Harmful Fine-Tuning

【速读】: 该论文试图解决大语言模型 (LLMs) 在微调过程中因少量恶意数据上传而导致的模型安全问题,即微调攻击 (fine-tuning attacks)。解决方案的关键是提出了一种无需训练的框架——神经元级安全重对齐 (Neuron-Level Safety Realignment, NLSR)。该框架通过构建一个安全参考模型来放大神经元中的安全相关特征,并利用该参考模型识别安全关键神经元作为补丁。随后,通过移植这些预先准备的补丁,仅恢复那些在微调后表现出显著相似性差异的神经元,从而在最小化对模型任务精度影响的前提下,显著提升微调后模型的安全性。

链接: https://arxiv.org/abs/2412.12497
作者: Xin Yi,Shunfan Zheng,Linlin Wang,Gerard de Melo,Xiaoling Wang,Liang He
机构: 未知
关键词: large language models, textbf, vulnerability in large, large language, model
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:The emergence of finetuning-as-a-service has revealed a new vulnerability in large language models (LLMs). A mere handful of malicious data uploaded by users can subtly manipulate the finetuning process, resulting in an alignment-broken model. Existing methods to counteract fine-tuning attacks typically require substantial computational resources. Even with parameter-efficient techniques like LoRA, gradient updates remain essential. To address these challenges, we propose \textbfNeuron-\textbfLevel \textbfSafety \textbfRealignment (\textbfNLSR), a training-free framework that restores the safety of LLMs based on the similarity difference of safety-critical neurons before and after fine-tuning. The core of our framework is first to construct a safety reference model from an initially aligned model to amplify safety-related features in neurons. We then utilize this reference model to identify safety-critical neurons, which we prepare as patches. Finally, we selectively restore only those neurons that exhibit significant similarity differences by transplanting these prepared patches, thereby minimally altering the fine-tuned model. Extensive experiments demonstrate significant safety enhancements in fine-tuned models across multiple downstream tasks, while greatly maintaining task-level accuracy. Our findings suggest regions of some safety-critical neurons show noticeable differences after fine-tuning, which can be effectively corrected by transplanting neurons from the reference model without requiring additional training. The code will be available at \urlthis https URL
zh

[NLP-78] Boosting Long-Context Information Seeking via Query-Guided Activation Refilling

【速读】: 该论文试图解决大型语言模型(LLMs)在处理长上下文时面临的挑战,主要问题包括上下文窗口限制和大规模键值(KV)激活带来的计算负担,这严重影响了效率。解决方案的关键是提出了一种名为查询引导的激活重填(ACRE)的方法,通过构建双层KV缓存(Bi-layer KV Cache)来处理长上下文信息检索任务。其中,第一层缓存(L1)紧凑地捕捉全局信息,第二层缓存(L2)提供详细的局部信息。ACRE通过建立两层缓存之间的代理关系,允许输入查询关注L1缓存,并根据查询需求动态从L2缓存中重填相关条目,从而在保持全局理解的同时,结合查询特定的局部细节,提升答案解码的效率和性能。

链接: https://arxiv.org/abs/2412.12486
作者: Hongjin Qian,Zheng Liu,Peitian Zhang,Zhicheng Dou,Defu Lian
机构: 未知
关键词: inherent context-window limitations, large language models, severely impact efficiency, extensive key-value, long contexts poses
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Information Retrieval (cs.IR)
备注: 12 pages

点击查看摘要

Abstract:Processing long contexts poses a significant challenge for large language models (LLMs) due to their inherent context-window limitations and the computational burden of extensive key-value (KV) activations, which severely impact efficiency. For information-seeking tasks, full context perception is often unnecessary, as a query’s information needs can dynamically range from localized details to a global perspective, depending on its complexity. However, existing methods struggle to adapt effectively to these dynamic information needs. In the paper, we propose a method for processing long-context information-seeking tasks via query-guided Activation Refilling (ACRE). ACRE constructs a Bi-layer KV Cache for long contexts, where the layer-1 (L1) cache compactly captures global information, and the layer-2 (L2) cache provides detailed and localized information. ACRE establishes a proxying relationship between the two caches, allowing the input query to attend to the L1 cache and dynamically refill it with relevant entries from the L2 cache. This mechanism integrates global understanding with query-specific local details, thus improving answer decoding. Experiments on a variety of long-context information-seeking datasets demonstrate ACRE’s effectiveness, achieving improvements in both performance and efficiency. Comments: 12 pages Subjects: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Information Retrieval (cs.IR) Cite as: arXiv:2412.12486 [cs.CL] (or arXiv:2412.12486v1 [cs.CL] for this version) https://doi.org/10.48550/arXiv.2412.12486 Focus to learn more arXiv-issued DOI via DataCite (pending registration)
zh

[NLP-79] Human-in-the-Loop Generation of Adversarial Texts: A Case Study on Tibetan Script NAACL2025

【速读】: 该论文试图解决当前文本对抗攻击研究中的三个主要问题:(1) 大多数方法仅针对资源丰富的语言,缺乏对较少研究语言的对抗文本生成;(2) 现有方法容易生成无效或模糊的对抗文本,难以构建高质量的对抗鲁棒性基准;(3) 新语言模型可能对部分已生成的对抗文本具有免疫力,导致基准需要更新。解决方案的关键是提出了HITL-GAT系统,该系统基于人机协作(human-in-the-loop)的通用方法,通过四个阶段(受害者模型构建、对抗样本生成、高质量基准构建和对抗鲁棒性评估)来生成和更新对抗文本,并特别针对藏文进行了案例研究,为其他较少研究语言的对抗研究提供了参考。

链接: https://arxiv.org/abs/2412.12478
作者: Xi Cao,Yuan Sun,Jiajun Li,Quzong Gesang,Nuo Qun,Tashi Nyima
机构: National Language Resource Monitoring & Research Center Minority Languages Branch, Beijing, China; Minzu University of China, Beijing, China; Collaborative Innovation Center for Tibet Informatization, Lhasa, China; Tibet University, Lhasa, China
关键词: SOTA LLMs, adversarial, models perform excellently, Adversarial texts, perform excellently
类目: Computation and Language (cs.CL); Cryptography and Security (cs.CR); Human-Computer Interaction (cs.HC)
备注: Review Version; Submitted to NAACL 2025 Demo Track

点击查看摘要

Abstract:DNN-based language models perform excellently on various tasks, but even SOTA LLMs are susceptible to textual adversarial attacks. Adversarial texts play crucial roles in multiple subfields of NLP. However, current research has the following issues. (1) Most textual adversarial attack methods target rich-resourced languages. How do we generate adversarial texts for less-studied languages? (2) Most textual adversarial attack methods are prone to generating invalid or ambiguous adversarial texts. How do we construct high-quality adversarial robustness benchmarks? (3) New language models may be immune to part of previously generated adversarial texts. How do we update adversarial robustness benchmarks? To address the above issues, we introduce HITL-GAT, a system based on a general approach to human-in-the-loop generation of adversarial texts. HITL-GAT contains four stages in one pipeline: victim model construction, adversarial example generation, high-quality benchmark construction, and adversarial robustness evaluation. Additionally, we utilize HITL-GAT to make a case study on Tibetan script which can be a reference for the adversarial research of other less-studied languages.
zh

[NLP-80] RareAgents : Autonomous Multi-disciplinary Team for Rare Disease Diagnosis and Treatment

【速读】: 该论文试图解决罕见病诊断和治疗中的复杂性问题,特别是在症状复杂和专业医生稀缺的情况下。解决方案的关键在于提出了RareAgents,这是一个基于大型语言模型(LLM)的多学科团队,专门针对罕见病的复杂临床场景进行优化。RareAgents整合了先进的规划能力、记忆机制和医疗工具的使用,并采用Llama-3.1-8B/70B作为基础模型。通过实验验证,RareAgents在罕见病的鉴别诊断和药物推荐方面超越了现有的最先进模型(如GPT-4o)和传统代理框架。此外,论文还贡献了一个新的数据集MIMIC-IV-Ext-Rare,以支持该领域的进一步研究。

链接: https://arxiv.org/abs/2412.12475
作者: Xuanzhong Chen,Ye Jin,Xiaohao Mao,Lun Wang,Shuyang Zhang,Ting Chen
机构: Tsinghua University(清华大学); Peking Union Medical College Hospital(北京协和医院)
关键词: low individual incidence, million people worldwide, people worldwide due, individual incidence, collectively impact
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Rare diseases, despite their low individual incidence, collectively impact around 300 million people worldwide due to the huge number of diseases. The complexity of symptoms and the shortage of specialized doctors with relevant experience make diagnosing and treating rare diseases more challenging than common diseases. Recently, agents powered by large language models (LLMs) have demonstrated notable improvements across various domains. In the medical field, some agent methods have outperformed direct prompts in question-answering tasks from medical exams. However, current agent frameworks lack adaptation for real-world clinical scenarios, especially those involving the intricate demands of rare diseases. To address these challenges, we present RareAgents, the first multi-disciplinary team of LLM-based agents tailored to the complex clinical context of rare diseases. RareAgents integrates advanced planning capabilities, memory mechanisms, and medical tools utilization, leveraging Llama-3.1-8B/70B as the base model. Experimental results show that RareAgents surpasses state-of-the-art domain-specific models, GPT-4o, and existing agent frameworks in both differential diagnosis and medication recommendation for rare diseases. Furthermore, we contribute a novel dataset, MIMIC-IV-Ext-Rare, derived from MIMIC-IV, to support further advancements in this field.
zh

[NLP-81] Knowledge Boundary of Large Language Models : A Survey

【速读】: 该论文试图解决大语言模型 (LLMs) 在知识存储和利用方面的局限性问题,特别是生成不真实或不准确响应的现象。其关键在于提出一个全面的LLM知识边界定义,并通过正式的分类法将知识分为四种类型,从而系统地探讨研究动机、边界识别方法以及应对挑战的策略。这一解决方案旨在为社区提供全面的概述,促进对关键问题的理解,并激发LLM知识研究领域的进一步发展。

链接: https://arxiv.org/abs/2412.12472
作者: Moxin Li,Yong Zhao,Yang Deng,Wenxuan Zhang,Shuaiyi Li,Wenya Xie,See-Kiong Ng,Tat-Seng Chua
机构: National University of Singapore(新加坡国立大学); Singapore Management University(新加坡管理大学); The Chinese University of Hong Kong(香港中文大学); The Chinese University of Hong Kong, Shenzhen(香港中文大学(深圳))
关键词: store vast amount, large language models, language models, store vast, leading to undesired
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Although large language models (LLMs) store vast amount of knowledge in their parameters, they still have limitations in the memorization and utilization of certain knowledge, leading to undesired behaviors such as generating untruthful and inaccurate responses. This highlights the critical need to understand the knowledge boundary of LLMs, a concept that remains inadequately defined in existing research. In this survey, we propose a comprehensive definition of the LLM knowledge boundary and introduce a formalized taxonomy categorizing knowledge into four distinct types. Using this foundation, we systematically review the field through three key lenses: the motivation for studying LLM knowledge boundaries, methods for identifying these boundaries, and strategies for mitigating the challenges they present. Finally, we discuss open challenges and potential research directions in this area. We aim for this survey to offer the community a comprehensive overview, facilitate access to key issues, and inspire further advancements in LLM knowledge research.
zh

[NLP-82] Core Context Aware Attention for Long Context Language Modeling

【速读】: 该论文试图解决Transformer-based大型语言模型(LLMs)在处理长上下文(如32K长度)时面临的两个主要问题:1)计算和内存复杂度随上下文长度L的平方增长;2)冗余的上下文信息可能干扰模型捕捉关键token之间的依赖关系,从而降低表示性能。解决方案的关键在于提出了一种即插即用的核心上下文感知(Core Context Aware, CCA)注意力机制,该机制包含两个互补组件:1)全局池化注意力(Globality-pooling attention),通过动态合并每个组中的token为“核心token”来减少冗余;2)局部保留注意力(Locality-preserved attention),将邻近token纳入注意力计算。通过融合这两种注意力,CCA-Attention在保持全面建模能力的同时,显著降低了计算和内存复杂度,并增强了长上下文建模能力。

链接: https://arxiv.org/abs/2412.12465
作者: Yaofo Chen,Zeng You,Shuhai Zhang,Haokun Li,Yirui Li,Yaowei Wang,Mingkui Tan
机构: South China University of Technology(华南理工大学); Pengcheng Laboratory(鹏城实验室); Pazhou Laboratory(琶洲实验室)
关键词: Transformer-based Large Language, natural language processing, language processing tasks, Large Language Models, exhibited remarkable success
类目: Computation and Language (cs.CL); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:Transformer-based Large Language Models (LLMs) have exhibited remarkable success in various natural language processing tasks primarily attributed to self-attention mechanism, which requires a token to consider all preceding tokens as its context to compute the attention score. However, when the context length L becomes very large (e.g., 32K), more redundant context information will be included w.r.t. any tokens, making the self-attention suffer from two main limitations: 1) The computational and memory complexity scales quadratically w.r.t. L; 2) The presence of redundant context information may hamper the model to capture dependencies among crucial tokens, which may degrade the representation performance. In this paper, we propose a plug-and-play Core Context Aware (CCA) Attention for efficient long-range context modeling, which consists of two components: 1) Globality-pooling attention that divides input tokens into groups and then dynamically merges tokens within each group into one core token based on their significance; 2) Locality-preserved attention that incorporates neighboring tokens into the attention calculation. The two complementary attentions will then be fused to the final attention, maintaining comprehensive modeling ability as the full self-attention. In this way, the core context information w.r.t. a given token will be automatically focused and strengthened, while the context information in redundant groups will be diminished during the learning process. As a result, the computational and memory complexity will be significantly reduced. More importantly, the CCA-Attention can improve the long-context modeling ability by diminishing the redundant context information. Extensive experimental results demonstrate that our CCA-Attention significantly outperforms state-of-the-art models in terms of computational efficiency and long-context modeling ability.
zh

[NLP-83] LITA: An Efficient LLM -assisted Iterative Topic Augmentation Framework

【速读】: 该论文试图解决传统主题模型在特定领域应用中面临的特异性和连贯性不足的问题,以及引导式方法(如SeededLDA和CorEx)在人工成本和静态性方面的局限性。解决方案的关键在于提出了LLM-assisted Iterative Topic Augmentation框架(LITA),该框架通过结合用户提供的种子词与基于嵌入的聚类和迭代优化,利用大型语言模型(LLM)对少量模糊文档进行重新分配,从而在降低API成本的同时提升主题质量。实验结果表明,LITA在主题质量和聚类性能方面优于五个基线模型,提供了一个高效且适应性强的主题建模和文本聚类框架。

链接: https://arxiv.org/abs/2412.12459
作者: Chia-Hsuan Chang,Jui-Tse Tsai,Yi-Hang Tsai,San-Yih Hwang
机构: 未知
关键词: uncovering thematic structures, uncovering thematic, thematic structures, struggle with specificity, specificity and coherence
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Information Retrieval (cs.IR)
备注: Under Review

点击查看摘要

Abstract:Topic modeling is widely used for uncovering thematic structures within text corpora, yet traditional models often struggle with specificity and coherence in domain-focused applications. Guided approaches, such as SeededLDA and CorEx, incorporate user-provided seed words to improve relevance but remain labor-intensive and static. Large language models (LLMs) offer potential for dynamic topic refinement and discovery, yet their application often incurs high API costs. To address these challenges, we propose the LLM-assisted Iterative Topic Augmentation framework (LITA), an LLM-assisted approach that integrates user-provided seeds with embedding-based clustering and iterative refinement. LITA identifies a small number of ambiguous documents and employs an LLM to reassign them to existing or new topics, minimizing API costs while enhancing topic quality. Experiments on two datasets across topic quality and clustering performance metrics demonstrate that LITA outperforms five baseline models, including LDA, SeededLDA, CorEx, BERTopic, and PromptTopic. Our work offers an efficient and adaptable framework for advancing topic modeling and text clustering.
zh

[NLP-84] Graph Learning in the Era of LLM s: A Survey from the Perspective of Data Models and Tasks

【速读】: 该论文试图解决跨领域文本属性图 (Text-Attributed Graph, TAG) 数据在复杂工业场景中的应用问题,特别是在图学习、图推理和图问答等任务中的表现。解决方案的关键在于将图神经网络 (Graph Neural Networks, GNNs) 和大型语言模型 (Large Language Models, LLMs) 进行协同整合,形成统一的模型架构。通过利用 GNNs 捕捉复杂结构关系的能力和 LLMs 理解图中文本描述的丰富语义上下文的能力,论文提出了一种新的图学习范式。这种整合方法不仅提升了数据质量,还增强了模型在不同领域中的表示能力和泛化能力,尤其是在监督、少样本和零样本学习设置下,能够有效处理多样化的下游图任务。

链接: https://arxiv.org/abs/2412.12456
作者: Xunkai Li,Zhengyu Wu,Jiayi Wu,Hanwen Cui,Jishuo Jia,Rong-Hua Li,Guoren Wang
机构: Beijing Institute of TechnologyBeijingChina(北京理工大学北京中国); Shandong UniversityShandongChina(山东大学山东中国)
关键词: Large Language Models, Large Language, promising technological paradigm, Language Models, unified Model architecture
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Databases (cs.DB)
备注: In progress

点击查看摘要

Abstract:With the increasing prevalence of cross-domain Text-Attributed Graph (TAG) Data (e.g., citation networks, recommendation systems, social networks, and ai4science), the integration of Graph Neural Networks (GNNs) and Large Language Models (LLMs) into a unified Model architecture (e.g., LLM as enhancer, LLM as collaborators, LLM as predictor) has emerged as a promising technological paradigm. The core of this new graph learning paradigm lies in the synergistic combination of GNNs’ ability to capture complex structural relationships and LLMs’ proficiency in understanding informative contexts from the rich textual descriptions of graphs. Therefore, we can leverage graph description texts with rich semantic context to fundamentally enhance Data quality, thereby improving the representational capacity of model-centric approaches in line with data-centric machine learning principles. By leveraging the strengths of these distinct neural network architectures, this integrated approach addresses a wide range of TAG-based Task (e.g., graph learning, graph reasoning, and graph question answering), particularly in complex industrial scenarios (e.g., supervised, few-shot, and zero-shot settings). In other words, we can treat text as a medium to enable cross-domain generalization of graph learning Model, allowing a single graph model to effectively handle the diversity of downstream graph-based Task across different data domains. This work serves as a foundational reference for researchers and practitioners looking to advance graph learning methodologies in the rapidly evolving landscape of LLM. We consistently maintain the related open-source materials at \urlthis https URL.
zh

[NLP-85] PERC: Plan-As-Query Example Retrieval for Underrepresented Code Generation COLING2025

【速读】: 该论文试图解决在代码生成任务中,如何有效选择示例以提升生成质量的问题,尤其是在目标编程语言(PL)不常见的情况下。解决方案的关键在于提出了**Plan-as-query Example Retrieval for few-shot prompting in Code generation (PERC)**框架,该框架通过将代码转换为伪代码(pseudocode)来捕捉算法计划(algorithmic plans),并利用这些计划作为查询来检索有效的示例。这种方法显著提高了生成准确性,并在多种编程语言环境下表现出优于现有检索增强生成(RAG)方法的性能。

链接: https://arxiv.org/abs/2412.12447
作者: Jaeseok Yoo,Hojae Han,Youngwon Lee,Jaejin Kim,Seung-won Hwang
机构: Seoul National University(首尔国立大学); Interdisciplinary Program in Artificial Intelligence, Seoul National University(人工智能跨学科项目,首尔国立大学)
关键词: shown significant promise, large language models, employing retrieval-augmented generation, significant promise, models has shown
类目: oftware Engineering (cs.SE); Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注: Accepted by COLING 2025 main conference

点击查看摘要

Abstract:Code generation with large language models has shown significant promise, especially when employing retrieval-augmented generation (RAG) with few-shot examples. However, selecting effective examples that enhance generation quality remains a challenging task, particularly when the target programming language (PL) is underrepresented. In this study, we present two key findings: (1) retrieving examples whose presented algorithmic plans can be referenced for generating the desired behavior significantly improves generation accuracy, and (2) converting code into pseudocode effectively captures such algorithmic plans, enhancing retrieval quality even when the source and the target PLs are different. Based on these findings, we propose Plan-as-query Example Retrieval for few-shot prompting in Code generation (PERC), a novel framework that utilizes algorithmic plans to identify and retrieve effective examples. We validate the effectiveness of PERC through extensive experiments on the CodeContests, HumanEval and MultiPL-E benchmarks: PERC consistently outperforms the state-of-the-art RAG methods in code generation, both when the source and target programming languages match or differ, highlighting its adaptability and robustness in diverse coding environments.
zh

[NLP-86] Persona-SQ: A Personalized Suggested Question Generation Framework For Real-world Documents

【速读】: 该论文试图解决当前AI驱动的阅读应用中,生成式问题(Suggested Questions, SQs)缺乏个性化和多样性的问题。解决方案的关键在于引入一个生成个性化SQs的流程,该流程通过结合用户的职业背景(professions)和阅读目标(reading goals)来生成更高质量和多样性的问题。这一方法不仅能够作为现有SQ系统的即插即用替代方案,提升其性能,还能用于微调极小模型,使其在SQ生成任务上与大型模型表现相当,从而支持本地运行的设备上模型,提供快速且隐私保护的SQ体验。

链接: https://arxiv.org/abs/2412.12445
作者: Zihao Lin,Zichao Wang,Yuanting Pan,Varun Manjunatha,Ryan Rossi,Angela Lau,Lifu Huang,Tong Sun
机构: UC Davis(加州大学戴维斯分校); Adobe(奥多比); Stanford University(斯坦福大学)
关键词: effective initial interface, AI-powered reading applications, Suggested questions, provide an effective, effective initial
类目: Computation and Language (cs.CL)
备注: 38 pages, 26 figures

点击查看摘要

Abstract:Suggested questions (SQs) provide an effective initial interface for users to engage with their documents in AI-powered reading applications. In practical reading sessions, users have diverse backgrounds and reading goals, yet current SQ features typically ignore such user information, resulting in homogeneous or ineffective questions. We introduce a pipeline that generates personalized SQs by incorporating reader profiles (professions and reading goals) and demonstrate its utility in two ways: 1) as an improved SQ generation pipeline that produces higher quality and more diverse questions compared to current baselines, and 2) as a data generator to fine-tune extremely small models that perform competitively with much larger models on SQ generation. Our approach can not only serve as a drop-in replacement in current SQ systems to immediately improve their performance but also help develop on-device SQ models that can run locally to deliver fast and private SQ experience.
zh

[NLP-87] Refining Dimensions for Improving Clustering-based Cross-lingual Topic Models COLING2025

【速读】: 该论文试图解决基于聚类的主题模型在跨语言主题识别中的性能问题,主要由于多语言语言模型生成的语言依赖维度 (Language-Dependent Dimensions, LDDs) 对跨语言主题识别的负面影响。解决方案的关键在于引入了一种基于奇异值分解 (Singular Value Decomposition, SVD) 的维度精炼组件,该组件能够有效中和LDDs的影响,从而提升模型在跨语言主题识别中的准确性。实验结果表明,更新后的管道在三个数据集上均优于其他最先进的跨语言主题模型。

链接: https://arxiv.org/abs/2412.12433
作者: Chia-Hsuan Chang,Tien-Yuan Huang,Yi-Hang Tsai,Chia-Ming Chang,San-Yih Hwang
机构: Yale University(耶鲁大学); National Sun Yat-sen University(国立中山大学)
关键词: Recent works, monolingual topic identification, topic models perform, contextualized representations, identification by introducing
类目: Computation and Language (cs.CL); Information Retrieval (cs.IR); Machine Learning (cs.LG)
备注: Accepted to 18th BUCC Workshop at COLING 2025

点击查看摘要

Abstract:Recent works in clustering-based topic models perform well in monolingual topic identification by introducing a pipeline to cluster the contextualized representations. However, the pipeline is suboptimal in identifying topics across languages due to the presence of language-dependent dimensions (LDDs) generated by multilingual language models. To address this issue, we introduce a novel, SVD-based dimension refinement component into the pipeline of the clustering-based topic model. This component effectively neutralizes the negative impact of LDDs, enabling the model to accurately identify topics across languages. Our experiments on three datasets demonstrate that the updated pipeline with the dimension refinement component generally outperforms other state-of-the-art cross-lingual topic models.
zh

[NLP-88] Assessing the Limitations of Large Language Models in Clinical Fact Decomposition

【速读】: 该论文试图解决在医疗领域中使用大型语言模型(LLMs)进行事实验证时面临的挑战,特别是临床文档中由于术语密集和笔记类型多样性导致的细粒度事实分解难题。解决方案的关键在于提出了FactEHR数据集,该数据集包含了来自三个医院系统的2,168份临床笔记的全文事实分解,涵盖四种笔记类型。通过临床医生的评估,研究发现不同LLMs在事实分解质量上存在显著差异,某些模型每句话生成的事实数量是其他模型的2.6倍。这强调了提升LLMs在临床文本中支持事实验证能力的必要性。

链接: https://arxiv.org/abs/2412.12422
作者: Monica Munnangi,Akshay Swaminathan,Jason Alan Fries,Jenelle Jindal,Sanjana Narayanan,Ivan Lopez,Lucia Tu,Philip Chung,Jesutofunmi A. Omiye,Mehr Kashyap,Nigam Shah
机构: Khoury College of Computer Sciences, Northeastern University(东北大学); Center for Biomedical Informatics Research, Stanford University(斯坦福大学); Department of Biomedical Data Science(生物医学数据科学系); Stanford Health Care(斯坦福医疗保健); Department of Medicine(医学系); Clinical Excellence Research Center(临床卓越研究中心); Department of Anesthesiology, Perioperative & Pain Medicine(麻醉学、围手术期与疼痛医学系); Department of Dermatology(皮肤病学系); Technology and Digital Solutions, Stanford Health Care(斯坦福医疗保健的技术与数字解决方案)
关键词: large language models, Verifying factual claims, Verifying factual, language models, claims is critical
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Verifying factual claims is critical for using large language models (LLMs) in healthcare. Recent work has proposed fact decomposition, which uses LLMs to rewrite source text into concise sentences conveying a single piece of information, as an approach for fine-grained fact verification. Clinical documentation poses unique challenges for fact decomposition due to dense terminology and diverse note types. To explore these challenges, we present FactEHR, a dataset consisting of full document fact decompositions for 2,168 clinical notes spanning four types from three hospital systems. Our evaluation, including review by clinicians, highlights significant variability in the quality of fact decomposition for four commonly used LLMs, with some LLMs generating 2.6x more facts per sentence than others. The results underscore the need for better LLM capabilities to support factual verification in clinical text. To facilitate future research in this direction, we plan to release our code at \urlthis https URL.
zh

[NLP-89] Bridging the Gap: Enhancing LLM Performance for Low-Resource African Languages with New Benchmarks Fine-Tuning and Cultural Adjustments AAAI2025

【速读】: 该论文旨在解决大型语言模型(LLMs)在非英语语言,尤其是非洲本土语言中表现不佳的问题。解决方案的关键在于创建了约100万个人工翻译的新基准数据集,涵盖8种低资源非洲语言,并通过这些数据集揭示了当前最先进(SOTA)LLMs在非洲语言中的性能差距。论文进一步探索了多种缩小性能差距的方法,包括高质量数据集微调(使用LLM作为标注工具)、跨语言迁移和文化适应性调整。研究发现,通过这些方法可以实现平均单语言性能提升5.6%,跨语言迁移带来2.9%的平均增益,以及在文化相关问题上3.0%的即开即用性能提升。

链接: https://arxiv.org/abs/2412.12417
作者: Tuka Alhanai,Adam Kasumovic,Mohammad Ghassemi,Aven Zitzelberger,Jessica Lundin,Guillaume Chabot-Couture
机构: MIT Media Lab(麻省理工学院媒体实验室); MIT Lincoln Laboratory(麻省理工学院林肯实验室); University of Southern California, Viterbi School of Engineering(南加州大学维特比工程学院); Massachusetts General Hospital(麻省总医院)
关键词: Large Language Models, native African languages, significant disparities remain, African languages, shown remarkable performance
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: Accepted to AAAI 2025. Main content is 9 pages, 3 figures. Includes supplementary materials

点击查看摘要

Abstract:Large Language Models (LLMs) have shown remarkable performance across various tasks, yet significant disparities remain for non-English languages, and especially native African languages. This paper addresses these disparities by creating approximately 1 million human-translated words of new benchmark data in 8 low-resource African languages, covering a population of over 160 million speakers of: Amharic, Bambara, Igbo, Sepedi (Northern Sotho), Shona, Sesotho (Southern Sotho), Setswana, and Tsonga. Our benchmarks are translations of Winogrande and three sections of MMLU: college medicine, clinical knowledge, and virology. Using the translated benchmarks, we report previously unknown performance gaps between state-of-the-art (SOTA) LLMs in English and African languages. Finally, using results from over 400 fine-tuned models, we explore several methods to reduce the LLM performance gap, including high-quality dataset fine-tuning (using an LLM-as-an-Annotator), cross-lingual transfer, and cultural appropriateness adjustments. Key findings include average mono-lingual improvements of 5.6% with fine-tuning (with 5.4% average mono-lingual improvements when using high-quality data over low-quality data), 2.9% average gains from cross-lingual transfer, and a 3.0% out-of-the-box performance boost on culturally appropriate questions. The publicly available benchmarks, translations, and code from this study support further research and development aimed at creating more inclusive and effective language technologies.
zh

[NLP-90] Efficient Scaling of Diffusion Transformers for Text-to-Image Generation

【速读】: 该论文旨在研究不同扩散变压器 (Diffusion Transformers, DiTs) 在文本到图像生成任务中的扩展性能,并通过广泛的消融实验验证其效果。关键解决方案在于发现基于纯自注意力机制的 U-ViT 模型相较于基于交叉注意力机制的 DiT 变体,具有更简单的设计和更有效的扩展能力,能够更直接地支持额外条件和其他模态的扩展。实验表明,2.3B 参数的 U-ViT 模型在受控环境下表现优于 SDXL UNet 和其他 DiT 变体。此外,论文还探讨了数据集规模增加和长描述增强对文本图像对齐性能和学习效率的提升作用。

链接: https://arxiv.org/abs/2412.12391
作者: Hao Li,Shamit Lal,Zhiheng Li,Yusheng Xie,Ying Wang,Yang Zou,Orchid Majumder,R. Manmatha,Zhuowen Tu,Stefano Ermon,Stefano Soatto,Ashwin Swaminathan
机构: Amazon AGI; AWS AI Labs
关键词: including training scaled, Diffusion Transformers, training scaled DiTs, scaled DiTs ranging, generation by performing
类目: Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:We empirically study the scaling properties of various Diffusion Transformers (DiTs) for text-to-image generation by performing extensive and rigorous ablations, including training scaled DiTs ranging from 0.3B upto 8B parameters on datasets up to 600M images. We find that U-ViT, a pure self-attention based DiT model provides a simpler design and scales more effectively in comparison with cross-attention based DiT variants, which allows straightforward expansion for extra conditions and other modalities. We identify a 2.3B U-ViT model can get better performance than SDXL UNet and other DiT variants in controlled setting. On the data scaling side, we investigate how increasing dataset size and enhanced long caption improve the text-image alignment performance and the learning efficiency.
zh

[NLP-91] Interpretable LLM -based Table Question Answering

【速读】: 该论文试图解决表格问答系统(Table Question Answering, Table QA)中的可解释性问题,特别是在高风险行业如金融或医疗领域。解决方案的关键是提出了Plan-of-SQLs(POS)方法,该方法通过仅使用SQL执行来回答输入查询,从而提供了一种可解释、有效且高效的Table QA方法。POS不仅在解释性方面优于现有的基于大语言模型(Large Language Models, LLMs)的方法,还在标准基准测试(如TabFact、WikiTQ和FetaQA)中表现出竞争性或更高的准确性,同时通过减少LLM调用和数据库查询次数,显著提高了效率。

链接: https://arxiv.org/abs/2412.12386
作者: Giang(Dexter)Nguyen,Ivan Brugere,Shubham Sharma,Sanjay Kariyappa,Anh Totti Nguyen,Freddy Lecue
机构: Auburn University, Auburn, AL, USA; J.P. Morgan AI Research, New York, NY, USA
关键词: Table Question Answering, Question Answering, Table Question, Large Language Models, finance or healthcare
类目: Computation and Language (cs.CL); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:Interpretability for Table Question Answering (Table QA) is critical, particularly in high-stakes industries like finance or healthcare. Although recent approaches using Large Language Models (LLMs) have significantly improved Table QA performance, their explanations for how the answers are generated are ambiguous. To fill this gap, we introduce Plan-of-SQLs ( or POS), an interpretable, effective, and efficient approach to Table QA that answers an input query solely with SQL executions. Through qualitative and quantitative evaluations with human and LLM judges, we show that POS is most preferred among explanation methods, helps human users understand model decision boundaries, and facilitates model success and error identification. Furthermore, when evaluated in standard benchmarks (TabFact, WikiTQ, and FetaQA), POS achieves competitive or superior accuracy compared to existing methods, while maintaining greater efficiency by requiring significantly fewer LLM calls and database queries.
zh

[NLP-92] How Different AI Chatbots Behave? Benchmarking Large Language Models in Behavioral Economics Games

【速读】: 该论文试图解决的问题是如何深入理解大型语言模型(LLMs)在多样化应用中的决策策略和行为模式。其解决方案的关键在于通过行为经济学游戏对五个领先的基于LLM的聊天机器人家族进行全面分析和基准测试,以揭示它们在不同场景中的共同和独特行为模式。这一研究方法不仅补充了近期关于行为图灵测试的研究,还为LLMs在关键决策角色中的部署提供了重要的战略偏好洞察。

链接: https://arxiv.org/abs/2412.12362
作者: Yutong Xie,Yiyao Liu,Zhuang Ma,Lin Shi,Xiyuan Wang,Walter Yuan,Matthew O. Jackson,Qiaozhu Mei
机构: University of MichiganAnn ArborMichiganUSA; Stanford UniversityStanfordCaliforniaUSA; MobLabPasadenaCaliforniaUSA
关键词: large language models, diverse applications requires, language models, large language, diverse applications
类目: Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注: Presented at The First Workshop on AI Behavioral Science (AIBS 2024)

点击查看摘要

Abstract:The deployment of large language models (LLMs) in diverse applications requires a thorough understanding of their decision-making strategies and behavioral patterns. As a supplement to a recent study on the behavioral Turing test, this paper presents a comprehensive analysis of five leading LLM-based chatbot families as they navigate a series of behavioral economics games. By benchmarking these AI chatbots, we aim to uncover and document both common and distinct behavioral patterns across a range of scenarios. The findings provide valuable insights into the strategic preferences of each LLM, highlighting potential implications for their deployment in critical decision-making roles.
zh

[NLP-93] Visual Instruction Tuning with 500x Fewer Parameters through Modality Linear Representation-Steering

【速读】: 该论文试图解决多模态大语言模型 (Multimodal Large Language Models, MLLMs) 中视觉与文本模态之间的不平衡问题,尤其是在视觉指令微调过程中,文本模态往往主导输出生成。解决方案的关键在于引入模态线性表示引导 (Modality Linear Representation-Steering, MoReS) 方法,通过在模型每一层的视觉子空间中进行线性变换来重新平衡视觉和文本模态的表示。MoReS 有效减少了可训练参数的数量,同时保持了在多个视觉基准和视觉问答任务中的性能,显著优于传统的参数高效微调方法(如 LoRA)。

链接: https://arxiv.org/abs/2412.12359
作者: Jinhe Bi,Yujun Wang,Haokun Chen,Xun Xiao,Artur Hecker,Volker Tresp,Yunpu Ma
机构: Ludwig Maximilian University of Munich(路德维希马克西米利安大学); Munich Research Center, Huawei Technologies(慕尼黑研究中心,华为技术); Munich Center for Machine Learning(慕尼黑机器学习中心)
关键词: Multimodal Large Language, Large Language Models, Large Language, Multimodal Large, Language Models
类目: Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Multimodal Large Language Models (MLLMs) have significantly advanced visual tasks by integrating visual representations into large language models (LLMs). The textual modality, inherited from LLMs, equips MLLMs with abilities like instruction following and in-context learning. In contrast, the visual modality enhances performance in downstream tasks by leveraging rich semantic content, spatial information, and grounding capabilities. These intrinsic modalities work synergistically across various visual tasks. Our research initially reveals a persistent imbalance between these modalities, with text often dominating output generation during visual instruction tuning. This imbalance occurs when using both full fine-tuning and parameter-efficient fine-tuning (PEFT) methods. We then found that re-balancing these modalities can significantly reduce the number of trainable parameters required, inspiring a direction for further optimizing visual instruction tuning. We introduce Modality Linear Representation-Steering (MoReS) to achieve the goal. MoReS effectively re-balances the intrinsic modalities throughout the model, where the key idea is to steer visual representations through linear transformations in the visual subspace across each model layer. To validate our solution, we composed LLaVA Steering, a suite of models integrated with the proposed MoReS method. Evaluation results show that the composed LLaVA Steering models require, on average, 500 times fewer trainable parameters than LoRA needs while still achieving comparable performance across three visual benchmarks and eight visual question-answering tasks. Last, we present the LLaVA Steering Factory, an in-house developed platform that enables researchers to quickly customize various MLLMs with component-based architecture for seamlessly integrating state-of-the-art models, and evaluate their intrinsic modality imbalance.
zh

[NLP-94] BioRAGent : A Retrieval-Augmented Generation System for Showcasing Generative Query Expansion and Domain-Specific Search for Scientific QA ECIR2025

【速读】: 该论文旨在解决生物医学领域的问答问题,提出了一个基于检索增强生成 (Retrieval-Augmented Generation, RAG) 的交互式网络系统——BioRAGent。其关键解决方案在于利用大型语言模型 (LLMs) 进行查询扩展、片段提取和答案生成,同时通过引用链接保持透明性,并允许用户编辑生成的查询。该系统在BioASQ 2024挑战中的成功应用表明,少样本学习 (few-shot learning) 与LLMs的结合能够有效应用于专业搜索场景,支持直接短段落式回答和带内联引用的回答。

链接: https://arxiv.org/abs/2412.12358
作者: Samy Ateia,Udo Kruschwitz
机构: 未知
关键词: biomedical question answering, interactive web-based retrieval-augmented, web-based retrieval-augmented generation, present BioRAGent, question answering
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: Version as accepted at the Demo Track at ECIR 2025

点击查看摘要

Abstract:We present BioRAGent, an interactive web-based retrieval-augmented generation (RAG) system for biomedical question answering. The system uses large language models (LLMs) for query expansion, snippet extraction, and answer generation while maintaining transparency through citation links to the source documents and displaying generated queries for further editing. Building on our successful participation in the BioASQ 2024 challenge, we demonstrate how few-shot learning with LLMs can be effectively applied for a professional search setting. The system supports both direct short paragraph style responses and responses with inline citations. Our demo is available online, and the source code is publicly accessible through GitHub.
zh

[NLP-95] RAG Playground: A Framework for Systematic Evaluation of Retrieval Strategies and Prompt Engineering in RAG Systems

【速读】: 该论文旨在解决检索增强生成 (Retrieval-Augmented Generation, RAG) 系统的系统性评估问题,并提出了一个开源框架 RAG Playground 作为解决方案。其关键在于通过三种检索方法(朴素向量搜索、重排序和混合向量-关键词搜索)与使用不同提示策略的 ReAct 代理相结合,进行全面的性能比较。论文引入了新的评估指标,并通过实验验证了混合搜索方法和结构化自评估提示策略在提升系统性能方面的显著效果,尤其是在检索准确性和响应质量上。此外,实验结果强调了提示工程在 RAG 系统中的重要性,表明自定义提示的代理在性能上具有一致的改进。

链接: https://arxiv.org/abs/2412.12322
作者: Ioannis Papadimitriou,Ilias Gialampoukidis,Stefanos Vrochidis,Ioannis(Yiannis)Kompatsiaris
机构: Information Technology Institute, CERTH, Greece(信息科技研究所,CERTH,希腊)
关键词: present RAG Playground, Retrieval-Augmented Generation, RAG Playground, present RAG, Playground
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Information Retrieval (cs.IR)
备注: Work In Progress

点击查看摘要

Abstract:We present RAG Playground, an open-source framework for systematic evaluation of Retrieval-Augmented Generation (RAG) systems. The framework implements and compares three retrieval approaches: naive vector search, reranking, and hybrid vector-keyword search, combined with ReAct agents using different prompting strategies. We introduce a comprehensive evaluation framework with novel metrics and provide empirical results comparing different language models (Llama 3.1 and Qwen 2.5) across various retrieval configurations. Our experiments demonstrate significant performance improvements through hybrid search methods and structured self-evaluation prompting, achieving up to 72.7% pass rate on our multi-metric evaluation framework. The results also highlight the importance of prompt engineering in RAG systems, with our custom-prompted agents showing consistent improvements in retrieval accuracy and response quality.
zh

[NLP-96] Graph-Guided Textual Explanation Generation Framework

【速读】: 该论文试图解决自然语言解释(Natural Language Explanations, NLEs)在反映模型内部推理过程时的不忠实性问题。解决方案的关键在于提出了G-Tex框架,通过利用高亮解释(highlight explanations)作为模型推理的高度忠实线索,并将其通过图神经网络(Graph Neural Network)层进行编码,从而显式地引导NLE的生成过程。这种方法确保了生成的解释能够更准确地反映模型的底层推理,实验结果表明G-Tex在NLE的忠实性上相比基线方法提升了17.59%,并且生成的解释在语义和词汇上与人工编写的解释更为相似。

链接: https://arxiv.org/abs/2412.12318
作者: Shuzhou Yuan,Jingyi Sun,Ran Zhang,Michael Färber,Steffen Eger,Pepa Atanasova,Isabelle Augenstein
机构: ScaDS.AI, TU Dresden(ScaDS.AI, 德累斯顿工业大学); University of Copenhagen(哥本哈根大学); University of Mannheim(曼海姆大学); University of Technology Nuremberg(纽伦堡科技大学)
关键词: Natural language explanations, provide plausible free-text, Natural language, plausible free-text explanations, provide plausible
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Natural language explanations (NLEs) are commonly used to provide plausible free-text explanations of a model’s reasoning about its predictions. However, recent work has questioned the faithfulness of NLEs, as they may not accurately reflect the model’s internal reasoning process regarding its predicted answer. In contrast, highlight explanations – input fragments identified as critical for the model’s predictions – exhibit measurable faithfulness, which has been incrementally improved through existing research. Building on this foundation, we propose G-Tex, a Graph-Guided Textual Explanation Generation framework designed to enhance the faithfulness of NLEs by leveraging highlight explanations. Specifically, highlight explanations are extracted as highly faithful cues representing the model’s reasoning and are subsequently encoded through a graph neural network layer, which explicitly guides the NLE generation process. This alignment ensures that the generated explanations closely reflect the model’s underlying reasoning. Experiments on T5 and BART using three reasoning datasets show that G-Tex improves NLE faithfulness by up to 17.59% compared to baseline methods. Additionally, G-Tex generates NLEs with greater semantic and lexical similarity to human-written ones. Human evaluations show that G-Tex can decrease redundant content and enhance the overall quality of NLEs. As our work introduces a novel method for explicitly guiding NLE generation to improve faithfulness, we hope it will serve as a stepping stone for addressing additional criteria for NLE and generated text overall.
zh

[NLP-97] Second Language (Arabic) Acquisition of LLM s via Progressive Vocabulary Expansion

【速读】: 该论文旨在解决阿拉伯世界在开发与GPT-4或ChatGPT 3.5等先进模型相媲美的大型语言模型(LLM)方面进展缓慢的问题,主要原因是主流语言(如英语和中文)的优先发展。解决方案的关键在于AraLLaMA模型采用了渐进式词汇扩展(Progressive Vocabulary Expansion),通过改进的BPE算法在训练过程中动态扩展阿拉伯语子词的词汇表,从而在每个阶段平衡了词汇外(OOV)比例。这种方法借鉴了人类在第二语言(阿拉伯语)学习中的词汇学习过程,并通过消融研究证明了其有效性。AraLLaMA在多种阿拉伯语基准测试中表现出色,与最佳的阿拉伯语LLM相当,并将所有模型、训练数据、基准测试和代码开源。

链接: https://arxiv.org/abs/2412.12310
作者: Jianqing Zhu,Huang Huang,Zhihang Lin,Juhao Liang,Zhengyang Tang,Khalid Almubarak,Abdulmohsen Alharthik,Bang An,Juncai He,Xiangbo Wu,Fei Yu,Junying Chen,Zhuoheng Ma,Yuhao Du,He Zhang,Emad A. Alghamdi,Lian Zhang,Ruoyu Sun,Haizhou Li,Benyou Wang,Jinchao Xu
机构: King Abdullah University of Science and Technology, Thuwal, Saudi Arabia; Shenzhen International Center for Industrial and Applied Mathematics, Shenzhen Research Institute of Big Data; Shenzhen Research Institue of Big Data, Shenzhen, China; The Chinese University of Hong Kong, Shenzhen, China; King Abdulaziz University, Jeddah, Saudi Arabia
关键词: English and Chinese, Arab world, democratizing large language, progressive vocabulary expansion, large language models
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:This paper addresses the critical need for democratizing large language models (LLM) in the Arab world, a region that has seen slower progress in developing models comparable to state-of-the-art offerings like GPT-4 or ChatGPT 3.5, due to a predominant focus on mainstream languages (e.g., English and Chinese). One practical objective for an Arabic LLM is to utilize an Arabic-specific vocabulary for the tokenizer that could speed up decoding. However, using a different vocabulary often leads to a degradation of learned knowledge since many words are initially out-of-vocabulary (OOV) when training starts. Inspired by the vocabulary learning during Second Language (Arabic) Acquisition for humans, the released AraLLaMA employs progressive vocabulary expansion, which is implemented by a modified BPE algorithm that progressively extends the Arabic subwords in its dynamic vocabulary during training, thereby balancing the OOV ratio at every stage. The ablation study demonstrated the effectiveness of Progressive Vocabulary Expansion. Moreover, AraLLaMA achieves decent performance comparable to the best Arabic LLMs across a variety of Arabic benchmarks. Models, training data, benchmarks, and codes will be all open-sourced.
zh

[NLP-98] Unanswerability Evaluation for Retreival Augmented Generation

【速读】: 该论文试图解决检索增强生成 (RAG) 系统在处理不可回答查询时表现不足的问题。解决方案的关键在于引入了一个名为 UAEval4RAG 的评估框架,该框架专门用于评估 RAG 系统在处理不可回答查询时的有效性。UAEval4RAG 通过定义六种不可回答查询的分类,并自动合成多样且具有挑战性的查询,结合未回答率和可接受率指标,揭示了 RAG 系统在不同组件(如检索模型、重写方法、重排序器、语言模型和提示策略)选择和提示设计中的性能权衡。该框架强调了组件选择和提示设计在优化 RAG 系统中平衡可回答查询准确性和不可回答查询高拒绝率的关键作用。

链接: https://arxiv.org/abs/2412.12300
作者: Xiangyu Peng,Prafulla Kumar Choubey,Caiming Xiong,Chien-Sheng Wu
机构: Salesforce Research(Salesforce研究)
关键词: Existing evaluation frameworks, Existing evaluation, rejecting unanswerable requests, appropriately rejecting unanswerable, RAG systems
类目: Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Existing evaluation frameworks for retrieval-augmented generation (RAG) systems focus on answerable queries, but they overlook the importance of appropriately rejecting unanswerable requests. In this paper, we introduce UAEval4RAG, a framework designed to evaluate whether RAG systems can handle unanswerable queries effectively. We define a taxonomy with six unanswerable categories, and UAEval4RAG automatically synthesizes diverse and challenging queries for any given knowledge base with unanswered ratio and acceptable ratio metrics. We conduct experiments with various RAG components, including retrieval models, rewriting methods, rerankers, language models, and prompting strategies, and reveal hidden trade-offs in performance of RAG systems. Our findings highlight the critical role of component selection and prompt design in optimizing RAG systems to balance the accuracy of answerable queries with high rejection rates of unanswerable ones. UAEval4RAG provides valuable insights and tools for developing more robust and reliable RAG systems.
zh

[NLP-99] Emergence of Abstractions: Concept Encoding and Decoding Mechanism for In-Context Learning in Transformers

【速读】: 该论文试图解决的问题是:自回归变换器(autoregressive transformers)如何在上下文学习(in-context learning, ICL)中实现快速学习和适应。解决方案的关键在于提出了概念编码-解码机制,通过研究变换器如何在表示中形成和使用内部抽象来解释ICL。具体来说,模型在学习过程中将不同的潜在概念(如“找到句子中的第一个名词”)编码为可分离的表示,同时构建条件解码算法以提升ICL性能。研究通过分析合成ICL任务中的训练动态,验证了这一机制在不同规模的预训练模型(如Gemma-2和Llama-3.1)中的存在,并通过机制干预和微调实验进一步证明了概念编码质量与ICL性能之间的因果关系。

链接: https://arxiv.org/abs/2412.12276
作者: Seungwook Han,Jinyeop Song,Jeff Gore,Pulkit Agrawal
机构: Improbable AI; Massachusetts Institute of Technology
关键词: Humans distill complex, distill complex experiences, enable rapid learning, Humans distill, distill complex
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:Humans distill complex experiences into fundamental abstractions that enable rapid learning and adaptation. Similarly, autoregressive transformers exhibit adaptive learning through in-context learning (ICL), which begs the question of how. In this paper, we propose \textbfconcept encoding-decoding mechanism to explain ICL by studying how transformers form and use internal abstractions in their representations. On synthetic ICL tasks, we analyze the training dynamics of a small transformer and report the coupled emergence of concept encoding and decoding. As the model learns to encode different latent concepts (e.g., ``Finding the first noun in a sentence.") into distinct, separable representations, it concureently builds conditional decoding algorithms and improve its ICL performance. We validate the existence of this mechanism across pretrained models of varying scales (Gemma-2 2B/9B/27B, Llama-3.1 8B/70B). Further, through mechanistic interventions and controlled finetuning, we demonstrate that the quality of concept encoding is causally related and predictive of ICL performance. Our empirical insights shed light into better understanding the success and failure modes of large language models via their representations.
zh

[NLP-100] DLF: Disentangled-Language-Focused Multimodal Sentiment Analysis AAAI2025

【速读】: 该论文试图解决多模态情感分析 (Multimodal Sentiment Analysis, MSA) 中由于模态间信息共享和直接融合导致的冗余和冲突问题。解决方案的关键在于提出了一个解耦语言聚焦 (Disentangled-Language-Focused, DLF) 的多模态表示学习框架。该框架通过特征解耦模块将模态共享信息和模态特定信息分离,并引入四种几何度量来优化解耦过程,从而减少冗余。此外,通过语言聚焦吸引子 (Language-Focused Attractor, LFA) 利用语言引导的交叉注意力机制,增强语言表示并利用互补的模态特定信息。最后,采用分层预测来提高整体准确性。实验结果表明,该框架在CMU-MOSI和CMU-MOSEI数据集上显著提升了性能。

链接: https://arxiv.org/abs/2412.12225
作者: Pan Wang,Qiang Zhou,Yawen Wu,Tianlong Chen,Jingtong Hu
机构: 未知
关键词: Multimodal Sentiment Analysis, Sentiment Analysis, leverages heterogeneous modalities, human sentiment, Multimodal Sentiment
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Multimedia (cs.MM)
备注: AAAI 2025 accepted

点击查看摘要

Abstract:Multimodal Sentiment Analysis (MSA) leverages heterogeneous modalities, such as language, vision, and audio, to enhance the understanding of human sentiment. While existing models often focus on extracting shared information across modalities or directly fusing heterogeneous modalities, such approaches can introduce redundancy and conflicts due to equal treatment of all modalities and the mutual transfer of information between modality pairs. To address these issues, we propose a Disentangled-Language-Focused (DLF) multimodal representation learning framework, which incorporates a feature disentanglement module to separate modality-shared and modality-specific information. To further reduce redundancy and enhance language-targeted features, four geometric measures are introduced to refine the disentanglement process. A Language-Focused Attractor (LFA) is further developed to strengthen language representation by leveraging complementary modality-specific information through a language-guided cross-attention mechanism. The framework also employs hierarchical predictions to improve overall accuracy. Extensive experiments on two popular MSA datasets, CMU-MOSI and CMU-MOSEI, demonstrate the significant performance gains achieved by the proposed DLF framework. Comprehensive ablation studies further validate the effectiveness of the feature disentanglement module, language-focused attractor, and hierarchical predictions. Our code is available at this https URL.
zh

[NLP-101] Finding a Wolf in Sheeps Clothing: Combating Adversarial Text-To-Image Prompts with Text Summarization

【速读】: 该论文试图解决文本生成图像模型在面对逐步“分而治之攻击”(Divide-and-Conquer Attack, DACA)时的脆弱性问题。DACA通过使用大型语言模型将敏感内容嵌入到良性叙述中来混淆提示中的不当内容。解决方案的关键在于提出了一种两层方法:首先进行文本摘要,然后进行二元分类。具体来说,研究者构建了包含DACA混淆和非混淆提示的Adversarial Text-to-Image Prompt (ATTIP)数据集,并通过小规模编码器模型和大规模语言模型生成两种摘要版本。随后,使用编码器分类器和GPT-4o分类器对摘要和未摘要的提示进行内容审核。实验结果表明,与直接处理未摘要数据的分类器相比,该方法将F1分数提升了31%,最高达到98%,表明预分类文本摘要能够有效抵御逐步DACA混淆攻击。

链接: https://arxiv.org/abs/2412.12212
作者: Portia Cooper,Harshita Narnoli,Mihai Surdeanu
机构: University of Arizona, Tucson, AZ, USA (亚利桑那大学,图森,亚利桑那州,美国)
关键词: stepwise DACA attacks, wrapping sensitive text, DACA attacks, stepwise DACA, obfuscate inappropriate content
类目: Cryptography and Security (cs.CR); Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Text-to-image models are vulnerable to the stepwise “Divide-and-Conquer Attack” (DACA) that utilize a large language model to obfuscate inappropriate content in prompts by wrapping sensitive text in a benign narrative. To mitigate stepwise DACA attacks, we propose a two-layer method involving text summarization followed by binary classification. We assembled the Adversarial Text-to-Image Prompt (ATTIP) dataset ( N=940 ), which contained DACA-obfuscated and non-obfuscated prompts. From the ATTIP dataset, we created two summarized versions: one generated by a small encoder model and the other by a large language model. Then, we used an encoder classifier and a GPT-4o classifier to perform content moderation on the summarized and unsummarized prompts. When compared with a classifier that operated over the unsummarized data, our method improved F1 score performance by 31%. Further, the highest recorded F1 score achieved (98%) was produced by the encoder classifier on a summarized ATTIP variant. This study indicates that pre-classification text summarization can inoculate content detection models against stepwise DACA obfuscations.
zh

[NLP-102] SEE: Sememe Entanglement Encoding for Transformer-bases Models Compression

【速读】: 该论文试图解决基于Transformer的大型语言模型在存储和计算成本过高的问题,尤其是在资源受限的场景中应用受限的问题。解决方案的关键在于提出了一种名为**Sememe Entanglement Encoding (SEE)**的算法,通过结合专家先验知识,利用低秩近似(low-rank approximation)的思想对模型进行压缩。具体来说,该算法将基本语义单元(如sememes)表示为低维向量,并通过广义量子纠缠(generalized quantum entanglement)的组合方式将其重构为高维词嵌入,从而在压缩模型参数和计算成本的同时保持稳定的性能。

链接: https://arxiv.org/abs/2412.12204
作者: Jing Zhang,Shuzhen Sun,Peng Zhang,Guangxing Cao,Hui Gao,Xindian Ma,Nan Xu,Yuexian Hou
机构: Tianjin University (天津大学)
关键词: exhibit groundbreaking capabilities, Transformer-based large language, computational costs, Sememe Entanglement Encoding, large language models
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Transformer-based large language models exhibit groundbreaking capabilities, but their storage and computational costs are prohibitively high, limiting their application in resource-constrained scenarios. An effective approach is to eliminate redundant model parameters and computational costs while incorporating efficient expert-derived knowledge structures to achieve a balance between compression and performance. Therefore, we propose the \textitSememe Entanglement Encoding (SEE) algorithm. Guided by expert prior knowledge, the model is compressed through the low-rank approximation idea. In Entanglement Embedding, basic semantic units such as sememes are represented as low-dimensional vectors, and then reconstructed into high-dimensional word embeddings through the combination of generalized quantum entanglement. We adapt the Sememe Entanglement Encoding algorithm to transformer-based models of different magnitudes. Experimental results indicate that our approach achieves stable performance while compressing model parameters and computational costs.
zh

[NLP-103] Model-diff: A Tool for Comparative Study of Language Models in the Input Space

【速读】: 该论文试图解决在大输入空间下比较两个大型语言模型(LMs)预测差异的问题,特别是在无法预见的角度上进行无偏比较。传统方法依赖于基准数据集,这些数据集仅涵盖有限数量的输入和设计好的应用视角,无法全面反映模型间的差异。论文提出的解决方案关键在于引入了一种新的模型比较分析框架,通过文本生成采样并调整采样统计直方图,以高效且无偏的方式估计两个模型在广泛输入空间中的预测差异。具体来说,该方法通过在负对数似然(negative log-likelihood)中绘制和计数输入,首次实现了在大输入空间中量化语言模型预测差异的目标,从而为模型抄袭等应用场景提供了新的分析工具。

链接: https://arxiv.org/abs/2412.12177
作者: Weitang Liu,Yuelei Li,Ying Wai Li,Zihan Wang,Jingbo Shang
机构: University of California, San Diego(加州大学圣地亚哥分校); Los Alamos National Laboratory(洛斯阿拉莫斯国家实验室)
关键词: large input space, real-world scenarios, input space, Comparing, large input
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:Comparing two (large) language models (LMs) side-by-side and pinpointing their prediction similarities and differences on the same set of inputs are crucial in many real-world scenarios, e.g., one can test if a licensed model was potentially plagiarized by another. Traditional analysis compares the LMs’ outputs on some benchmark datasets, which only cover a limited number of inputs of designed perspectives for the intended applications. The benchmark datasets cannot prepare data to cover the test cases from unforeseen perspectives which can help us understand differences between models unbiasedly. In this paper, we propose a new model comparative analysis setting that considers a large input space where brute-force enumeration would be infeasible. The input space can be simply defined as all token sequences that a LM would produce low perplexity on – we follow this definition in the paper as it would produce the most human-understandable inputs. We propose a novel framework \our that uses text generation by sampling and deweights the histogram of sampling statistics to estimate prediction differences between two LMs in this input space efficiently and unbiasedly. Our method achieves this by drawing and counting the inputs at each prediction difference value in negative log-likelihood. Experiments reveal for the first time the quantitative prediction differences between LMs in a large input space, potentially facilitating the model analysis for applications such as model plagiarism.
zh

[NLP-104] Explore Theory of Mind: Program-guided adversarial data generation for theory of mind reasoning

【速读】: 该论文试图解决当前大型语言模型 (LLMs) 在理论思维 (theory of mind) 评估中存在的局限性问题,特别是现有评估方法依赖于简单模式和有限数据集,可能导致对模型能力的过高估计。解决方案的关键在于提出了 ExploreToM 框架,该框架通过自定义领域特定语言的 A* 搜索生成复杂的故事结构和多样化的、合理的场景,从而实现大规模生成多样且具有挑战性的理论思维数据。这种方法不仅揭示了现有最先进 LLMs(如 Llama-3.1-70B 和 GPT-4o)在理论思维任务中的低准确率(分别为 0% 和 9%),还通过在经典 ToMi 基准上的微调,实现了 27 个百分点的准确率提升,并揭示了模型在状态跟踪和数据平衡等方面的不足。

链接: https://arxiv.org/abs/2412.12175
作者: Melanie Sclar,Jane Yu,Maryam Fazel-Zarandi,Yulia Tsvetkov,Yonatan Bisk,Yejin Choi,Asli Celikyilmaz
机构: Meta
关键词: large language models, theory of mind, large language, theory, mind
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Do large language models (LLMs) have theory of mind? A plethora of papers and benchmarks have been introduced to evaluate if current models have been able to develop this key ability of social intelligence. However, all rely on limited datasets with simple patterns that can potentially lead to problematic blind spots in evaluation and an overestimation of model capabilities. We introduce ExploreToM, the first framework to allow large-scale generation of diverse and challenging theory of mind data for robust training and evaluation. Our approach leverages an A* search over a custom domain-specific language to produce complex story structures and novel, diverse, yet plausible scenarios to stress test the limits of LLMs. Our evaluation reveals that state-of-the-art LLMs, such as Llama-3.1-70B and GPT-4o, show accuracies as low as 0% and 9% on ExploreToM-generated data, highlighting the need for more robust theory of mind evaluation. As our generations are a conceptual superset of prior work, fine-tuning on our data yields a 27-point accuracy improvement on the classic ToMi benchmark (Le et al., 2019). ExploreToM also enables uncovering underlying skills and factors missing for models to show theory of mind, such as unreliable state tracking or data imbalances, which may contribute to models’ poor performance on benchmarks.
zh

[NLP-105] A NotSo Simple Way to Beat Simple Bench

【速读】: 该论文试图解决大型语言模型(LLMs)在逻辑一致性和现实世界推理能力方面的局限性问题。解决方案的关键在于引入迭代推理(iterative reasoning)和反馈驱动(feedback-driven)的方法论,通过多步提示策略(multi-step prompting strategy)和全局一致性检查(global consistency checks)来提升模型的准确性和鲁棒性。研究结果表明,迭代推理显著增强了模型性能,特别是在标准准确度指标(AVG@5)和新的极端平均指标(EAG@5)上。此外,论文还强调了模型在空间和时间推理方面的不足,并提出了进一步优化的方向,包括动态反馈机制、自适应重启策略和多样化的评估指标。

链接: https://arxiv.org/abs/2412.12173
作者: Soham Sane,Angus McLean
机构: 未知
关键词: large language models, leveraging iterative reasoning, paper presents, large language, enhancing reasoning capabilities
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: 29 pages, 11 Figures

点击查看摘要

Abstract:This paper presents a novel framework for enhancing reasoning capabilities in large language models (LLMs) by leveraging iterative reasoning and feedback-driven methodologies. Building on the limitations identified in the SimpleBench benchmark, a dataset designed to evaluate logical coherence and real-world reasoning, we propose a multi-step prompting strategy coupled with global consistency checks to improve model accuracy and robustness. Through comparative analysis of state-of-the-art models, including Claude 3 Opus, Claude 3.5, GPT- 4o, and o1-preview, we demonstrate that iterative reasoning significantly enhances model performance, with improvements observed in both standard accuracy metrics (AVG@5) and a newly introduced metric, Extreme Averaging (EAG@5). Our results reveal model-specific strengths: Claude excels in maintaining logical consistency, while GPT-4o exhibits exploratory creativity but struggles with ambiguous prompts. By analyzing case studies and identifying gaps in spatial and temporal reasoning, we highlight areas for further refinement. The findings underscore the potential of structured reasoning frameworks to address inherent model limitations, irrespective of pretraining methodologies. This study lays the groundwork for integrating dynamic feedback mechanisms, adaptive restart strategies, and diverse evaluation metrics to advance LLM reasoning capabilities across complex and multi-domain problem spaces.
zh

[NLP-106] AI Adoption to Combat Financial Crime: Study on Natural Language Processing in Adverse Media Screening of Financial Services in English and Bangla multilingual interpretation

【速读】: 该论文试图解决在孟加拉国的移动金融服务(MFS)中,利用人工智能(AI)特别是自然语言处理(NLP)来加强金融犯罪检测和预防的问题。解决方案的关键在于通过NLP技术进行负面媒体筛查,以满足反洗钱(AML)和打击金融恐怖主义(CFT)的合规要求。论文指出,尽管NLP算法在识别与金融犯罪相关的负面媒体内容方面显示出高达94%的准确率,但在孟加拉国,AI技术的应用仍面临技术专长不足、高成本和法规不确定性等障碍。因此,论文强调了AI驱动的NLP解决方案在增强孟加拉国金融犯罪预防方面的潜力。

链接: https://arxiv.org/abs/2412.12171
作者: Soumita Roy
机构: 未知
关键词: Natural Language Processing, specifically Natural Language, employing Artificial Intelligence, Mobile Financial Services, Artificial Intelligence
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:This document explores the potential of employing Artificial Intelligence (AI), specifically Natural Language Processing (NLP), to strengthen the detection and prevention of financial crimes within the Mobile Financial Services(MFS) of Bangladesh with multilingual scenario. The analysis focuses on the utilization of NLP for adverse media screening, a vital aspect of compliance with anti-money laundering (AML) and combating financial terrorism (CFT) regulations. Additionally, it investigates the overall reception and obstacles related to the integration of AI in Bangladeshi banks. This report measures the effectiveness of NLP is promising with an accuracy around 94%. NLP algorithms display substantial promise in accurately identifying adverse media content linked to financial crimes. The lack of progress in this aspect is visible in Bangladesh, whereas globally the technology is already being used to increase effectiveness and efficiency. Hence, it is clear there is an issue with the acceptance of AI in Bangladesh. Some AML \ CFT concerns are already being addressed by AI technology. For example, Image Recognition OCR technology are being used in KYC procedures. Primary hindrances to AI integration involve a lack of technical expertise, high expenses, and uncertainties surrounding regulations. This investigation underscores the potential of AI-driven NLP solutions in fortifying efforts to prevent financial crimes in Bangladesh.
zh

[NLP-107] Greek2MathTex: A Greek Speech-to-Text Framework for LaTeX Equations Generation

【速读】: 该论文试图解决LaTeX在学术和科学领域中因复杂语法和代码式外观而导致的可访问性问题,特别是对于有残疾或不熟悉编程规范的用户。解决方案的关键在于开发了一种针对希腊语的语音到LaTeX方程系统,利用自动语音识别(ASR)和自然语言处理(NLP)技术,允许用户通过自然语言口述数学表达式,并将其转换为LaTeX格式。该系统的核心组件包括ASR引擎、基于大语言模型(LLM)的提示驱动方程生成机制,以及在开发过程中应用的自定义评估指标。

链接: https://arxiv.org/abs/2412.12167
作者: Evangelia Gkritzali,Panagiotis Kaliosis,Sofia Galanaki,Elisavet Palogiannidi,Theodoros Giannakopoulos
机构: NCSR DemokritosAthensAttikiGreece(希腊国家科学研究中心雅典阿提卡); NCSR DemokritosAthensAttikiGreece(希腊国家科学研究中心雅典阿提卡); NCSR DemokritosAthensAttikiGreece(希腊国家科学研究中心雅典阿提卡); NCSR DemokritosAthensAttikiGreece(希腊国家科学研究中心雅典阿提卡); NCSR DemokritosAthensAttikiGreece(希腊国家科学研究中心雅典阿提卡)
关键词: typesetting complex mathematical, scientific domains, vast majority, academic and scientific, facto standard
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Audio and Speech Processing (eess.AS)
备注: 4 pages, 2 figures, SETN2024: 13th EETN Conference on Artificial Intelligence

点击查看摘要

Abstract:In the vast majority of the academic and scientific domains, LaTeX has established itself as the de facto standard for typesetting complex mathematical equations and formulae. However, LaTeX’s complex syntax and code-like appearance present accessibility barriers for individuals with disabilities, as well as those unfamiliar with coding conventions. In this paper, we present a novel solution to this challenge through the development of a novel speech-to-LaTeX equations system specifically designed for the Greek language. We propose an end-to-end system that harnesses the power of Automatic Speech Recognition (ASR) and Natural Language Processing (NLP) techniques to enable users to verbally dictate mathematical expressions and equations in natural language, which are subsequently converted into LaTeX format. We present the architecture and design principles of our system, highlighting key components such as the ASR engine, the LLM-based prompt-driven equations generation mechanism, as well as the application of a custom evaluation metric employed throughout the development process. We have made our system open source and available at this https URL.
zh

[NLP-108] Performance of a large language model-Artificial Intelligence based chatbot for counseling patients with sexually transmitted infections and genital diseases

【速读】: 该论文旨在解决性传播感染(STIs)日益增加的全球负担问题,特别是现有通用聊天机器人(如ChatGPT)无法直接处理STI相关咨询的局限性。解决方案的关键在于开发了一个专门针对STI检测和咨询的AI聊天机器人平台Otiz。Otiz采用基于GPT4-0613的多智能体系统架构,结合大语言模型(LLM)和确定性有限自动机原理,提供上下文相关、医学准确且富有同理心的响应。其核心组件包括STI信息模块、情感识别、急性应激障碍检测和心理治疗模块,并通过问题建议智能体进行并行操作,从而实现对STI相关问题的准确诊断和咨询支持,减轻医疗系统的负担。

链接: https://arxiv.org/abs/2412.12166
作者: Nikhil Mehta,Sithira Ambepitiya,Thanveer Ahamad,Dinuka Wijesundara,Yudara Kularathne
机构: 未知
关键词: sexually transmitted infections, transmitted infections, proportion to specialists, sexually transmitted, Global burden
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: 18 pages, 1 table

点击查看摘要

Abstract:Introduction: Global burden of sexually transmitted infections (STIs) is rising out of proportion to specialists. Current chatbots like ChatGPT are not tailored for handling STI-related concerns out of the box. We developed Otiz, an Artificial Intelligence-based (AI-based) chatbot platform designed specifically for STI detection and counseling, and assessed its performance. Methods: Otiz employs a multi-agent system architecture based on GPT4-0613, leveraging large language model (LLM) and Deterministic Finite Automaton principles to provide contextually relevant, medically accurate, and empathetic responses. Its components include modules for general STI information, emotional recognition, Acute Stress Disorder detection, and psychotherapy. A question suggestion agent operates in parallel. Four STIs (anogenital warts, herpes, syphilis, urethritis/cervicitis) and 2 non-STIs (candidiasis, penile cancer) were evaluated using prompts mimicking patient language. Each prompt was independently graded by two venereologists conversing with Otiz as patient actors on 6 criteria using Numerical Rating Scale ranging from 0 (poor) to 5 (excellent). Results: Twenty-three venereologists did 60 evaluations of 30 prompts. Across STIs, Otiz scored highly on diagnostic accuracy (4.1-4.7), overall accuracy (4.3-4.6), correctness of information (5.0), comprehensibility (4.2-4.4), and empathy (4.5-4.8). However, relevance scores were lower (2.9-3.6), suggesting some redundancy. Diagnostic scores for non-STIs were lower (p=0.038). Inter-observer agreement was strong, with differences greater than 1 point occurring in only 12.7% of paired evaluations. Conclusions: AI conversational agents like Otiz can provide accurate, correct, discrete, non-judgmental, readily accessible and easily understandable STI-related information in an empathetic manner, and can alleviate the burden on healthcare systems.
zh

[NLP-109] What Makes In-context Learning Effective for Mathematical Reasoning: A Theoretical Analysis

【速读】: 该论文试图解决大语言模型 (LLMs) 在数学推理任务中使用少样本示例 (few-shot demonstrations) 时表现不稳定的问题。解决方案的关键在于提出了一种理论分析框架,证明了推理效能(通过经验预测损失衡量)可以由模型导向的语义相似性和示例的推理稳定性共同决定。基于此,论文提出了一种名为 LMS3 的示例选择方法,该方法能够自适应地为不同 LLMs 选择最相关的样本,并引入了一种新颖的示例拒绝机制,自动过滤不适合少样本学习的样本。实验结果表明,LMS3 在多个基准测试、LLM 骨干模型和少样本设置下均表现出优越性,并实现了现有方法无法达到的一致性改进。

链接: https://arxiv.org/abs/2412.12157
作者: Jiayu Liu,Zhenya Huang,Chaokun Wang,Xunpeng Huang,Chengxiang Zhai,Enhong Chen
机构: 1. University of Illinois at Urbana-Champaign (伊利诺伊大学厄巴纳-香槟分校); 2. University of Science and Technology of China (中国科学技术大学); 3. Alibaba Group (阿里巴巴集团); 4. University of Illinois at Urbana-Champaign (伊利诺伊大学厄巴纳-香槟分校)
关键词: large language models, shown impressive performance, diverse mathematical reasoning, large language, language models
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Owing to the capability of in-context learning, large language models (LLMs) have shown impressive performance across diverse mathematical reasoning benchmarks. However, we find that few-shot demonstrations can sometimes bring negative performance and their effectiveness on LLMs’ reasoning abilities remains unreliable. To this end, in this paper, we aim to theoretically analyze the impact of in-context demonstrations on LLMs’ reasoning performance. We prove that the reasoning efficacy (measured by empirical prediction loss) can be bounded by a LLM-oriented semantic similarity and an inference stability of demonstrations, which is general for both one-shot and few-shot scenarios. Based on this finding, we propose a straightforward, generalizable, and low-complexity demonstration selection method named LMS3. It can adaptively facilitate to select the most pertinent samples for different LLMs and includes a novel demonstration rejection mechanism to automatically filter out samples that are unsuitable for few-shot learning. Through experiments on three representative benchmarks, two LLM backbones, and multiple few-shot settings, we verify that our LMS3 has superiority and achieves consistent improvements on all datasets, which existing methods have been unable to accomplish.
zh

[NLP-110] PyOD 2: A Python Library for Outlier Detection with LLM -powered Model Selection

【速读】: 该论文试图解决当前Python Outlier Detection (PyOD)库的三个主要限制:(1) 缺乏对现代深度学习算法的充分覆盖,(2) 在PyTorch和TensorFlow之间实现碎片化,(3) 缺乏自动化模型选择功能,这使得非专业用户难以使用。解决方案的关键在于推出PyOD 2,它通过集成12种最先进的深度学习模型到一个统一的PyTorch框架中,并引入基于大语言模型(LLM)的自动化模型选择流程,从而简化了异常检测(OD)工作流程,提供了45种算法的选择,并在多种数据集上展现了强大的性能。

链接: https://arxiv.org/abs/2412.12154
作者: Sihan Chen,Zhuangzhuang Qian,Wingchun Siu,Xingcan Hu,Jiaqi Li,Shawn Li,Yuehan Qin,Tiankai Yang,Zhuo Xiao,Wanghao Ye,Yichi Zhang,Yushun Dong,Yue Zhao
机构: University of Southern California(南加州大学); University of Maryland(马里兰大学); Florida State University(佛罗里达州立大学)
关键词: Python Outlier Detection, network intrusion detection, Outlier detection, social network moderation, detection
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
备注:

点击查看摘要

Abstract:Outlier detection (OD), also known as anomaly detection, is a critical machine learning (ML) task with applications in fraud detection, network intrusion detection, clickstream analysis, recommendation systems, and social network moderation. Among open-source libraries for outlier detection, the Python Outlier Detection (PyOD) library is the most widely adopted, with over 8,500 GitHub stars, 25 million downloads, and diverse industry usage. However, PyOD currently faces three limitations: (1) insufficient coverage of modern deep learning algorithms, (2) fragmented implementations across PyTorch and TensorFlow, and (3) no automated model selection, making it hard for non-experts. To address these issues, we present PyOD Version 2 (PyOD 2), which integrates 12 state-of-the-art deep learning models into a unified PyTorch framework and introduces a large language model (LLM)-based pipeline for automated OD model selection. These improvements simplify OD workflows, provide access to 45 algorithms, and deliver robust performance on various datasets. In this paper, we demonstrate how PyOD 2 streamlines the deployment and automation of OD models and sets a new standard in both research and industry. PyOD 2 is accessible at [this https URL](this https URL). This study aligns with the Web Mining and Content Analysis track, addressing topics such as the robustness of Web mining methods and the quality of algorithmically-generated Web data. Subjects: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL) Cite as: arXiv:2412.12154 [cs.LG] (or arXiv:2412.12154v1 [cs.LG] for this version) https://doi.org/10.48550/arXiv.2412.12154 Focus to learn more arXiv-issued DOI via DataCite
zh

[NLP-111] Rethinking Comprehensive Benchmark for Chart Understanding: A Perspective from Scientific Literature

【速读】: 该论文试图解决现有基准测试在评估多模态模型对复杂科学文献图表理解能力时的局限性,主要问题包括图表类型范围狭窄、问题过于简单模板化、视觉元素单一以及评估方法不足,导致模型在实际应用中表现不佳。解决方案的关键在于引入新的基准测试——科学图表问答 (Scientific Chart QA, SCI-CQA),重点强调流程图这一常被忽视的图表类型。通过从过去十年顶级计算机科学会议论文中筛选出37,607张高质量图表,并结合5,629个精心设计的客观与开放式问题,SCI-CQA提供了一个更全面和真实的评估框架。此外,论文还提出了高效的标注流程,降低了数据标注成本,并强调了上下文信息在解决复杂问题中的重要性。

链接: https://arxiv.org/abs/2412.12150
作者: Lingdong Shen,Qigqi,Kun Ding,Gaofeng Meng,Shiming Xiang
机构: 未知
关键词: including multi-plot figures, Scientific Literature charts, Scientific Literature, complex visual elements, Literature charts
类目: Computation and Language (cs.CL); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:Scientific Literature charts often contain complex visual elements, including multi-plot figures, flowcharts, structural diagrams and etc. Evaluating multimodal models using these authentic and intricate charts provides a more accurate assessment of their understanding abilities. However, existing benchmarks face limitations: a narrow range of chart types, overly simplistic template-based questions and visual elements, and inadequate evaluation methods. These shortcomings lead to inflated performance scores that fail to hold up when models encounter real-world scientific charts. To address these challenges, we introduce a new benchmark, Scientific Chart QA (SCI-CQA), which emphasizes flowcharts as a critical yet often overlooked category. To overcome the limitations of chart variety and simplistic visual elements, we curated a dataset of 202,760 image-text pairs from 15 top-tier computer science conferences papers over the past decade. After rigorous filtering, we refined this to 37,607 high-quality charts with contextual information. SCI-CQA also introduces a novel evaluation framework inspired by human exams, encompassing 5,629 carefully curated questions, both objective and open-ended. Additionally, we propose an efficient annotation pipeline that significantly reduces data annotation costs. Finally, we explore context-based chart understanding, highlighting the crucial role of contextual information in solving previously unanswerable questions.
zh

[NLP-112] Navi or Knave: Jailbreaking Language Models via Metaphorical Avatars

【速读】: 该论文旨在揭示大型语言模型(LLMs)在面对隐喻性对抗攻击时的安全风险,特别是通过利用LLMs的想象能力来绕过其安全机制,从而实现“越狱”(jailbreaking)。解决方案的关键在于提出了一种名为AVATAR(Adversarial MeTA-phoR Via Jailbreak)的新型攻击框架,该框架通过从有害目标中提取有害实体,并将其映射到无害的对抗实体,利用LLMs的想象能力生成隐喻,进而将有害目标嵌入到类人交互中,以实现对LLMs的越狱攻击。实验结果表明,AVATAR能够有效且可迁移地越狱多个先进的LLMs,达到当前最先进的攻击成功率,揭示了LLMs在对抗隐喻攻击下的脆弱性,并强调了开发防御方法的必要性。

链接: https://arxiv.org/abs/2412.12145
作者: Yu Yan,Sheng Sun,Junqi Tong,Min Liu,Qi Li
机构: Institute of Computing Technology, Chinese Academy of SciencesBeijingChina(中国科学院计算技术研究所北京); People’s Public Security University of ChinaBeijingChina(中国人民公安大学北京); Tsinghua UniversityBeijingChina(清华大学北京)
关键词: Large Language Models, textbf, underline, convey information, complex subjects
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Metaphor serves as an implicit approach to convey information, while enabling the generalized comprehension of complex subjects. However, metaphor can potentially be exploited to bypass the safety alignment mechanisms of Large Language Models (LLMs), leading to the theft of harmful knowledge. In our study, we introduce a novel attack framework that exploits the imaginative capacity of LLMs to achieve jailbreaking, the J\underline\textbfAilbreak \underline\textbfVia \underline\textbfAdversarial Me\underline\textbfTA -pho\underline\textbfR (\textitAVATAR). Specifically, to elicit the harmful response, AVATAR extracts harmful entities from a given harmful target and maps them to innocuous adversarial entities based on LLM’s imagination. Then, according to these metaphors, the harmful target is nested within human-like interaction for jailbreaking adaptively. Experimental results demonstrate that AVATAR can effectively and transferablly jailbreak LLMs and achieve a state-of-the-art attack success rate across multiple advanced LLMs. Our study exposes a security risk in LLMs from their endogenous imaginative capabilities. Furthermore, the analytical study reveals the vulnerability of LLM to adversarial metaphors and the necessity of developing defense methods against jailbreaking caused by the adversarial metaphor. \textcolororange \textbfWarning: This paper contains potentially harmful content from LLMs.
zh

[NLP-113] Automatic Item Generation for Personality Situational Judgment Tests with Large Language Models

【速读】: 该论文试图解决传统情境判断测试 (SJT) 开发过程中存在的劳动密集型和易受偏见影响的问题。解决方案的关键在于利用GPT-4这一先进的大型语言模型 (LLM) 自动化生成中文人格情境判断测试 (PSJT)。通过优化提示设计(prompt design)和温度设置(temperature settings),GPT-4能够生成具有高内容效度、可靠性和效度的测试题目,显著超越手动开发的测试,特别是在测量大五人格特质方面。这一研究展示了GPT-4在心理测量测试开发中的高效性和创新性,为资源有限的环境下简化测试开发流程提供了可扩展的解决方案。

链接: https://arxiv.org/abs/2412.12144
作者: Chang-Jin Li,Jiyuan Zhang,Yun Tang,Jian Li
机构: 未知
关键词: situational judgment tests, personality situational judgment, talent selection, situational judgment, educational evaluation
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: Submitted to Organizational Research Methods. 48 pages (main text), 12 pages (appendix), and 3 figures

点击查看摘要

Abstract:Personality assessment, particularly through situational judgment tests (SJTs), is a vital tool for psychological research, talent selection, and educational evaluation. This study explores the potential of GPT-4, a state-of-the-art large language model (LLM), to automate the generation of personality situational judgment tests (PSJTs) in Chinese. Traditional SJT development is labor-intensive and prone to biases, while GPT-4 offers a scalable, efficient alternative. Two studies were conducted: Study 1 evaluated the impact of prompt design and temperature settings on content validity, finding that optimized prompts with a temperature of 1.0 produced creative and accurate items. Study 2 assessed the psychometric properties of GPT-4-generated PSJTs, revealing that they demonstrated satisfactory reliability and validity, surpassing the performance of manually developed tests in measuring the Big Five personality traits. This research highlights GPT-4’s effectiveness in developing high-quality PSJTs, providing a scalable and innovative method for psychometric test development. These findings expand the possibilities of automatic item generation and the application of LLMs in psychology, and offer practical implications for streamlining test development processes in resource-limited settings.
zh

[NLP-114] Harnessing Transfer Learning from Swahili: Advancing Solutions for Comorian Dialects

【速读】: 该论文试图解决非洲低资源语言(如Comorian)在自然语言处理(NLP)领域缺乏数据支持的问题。解决方案的关键在于采用迁移学习(Transfer Learning)策略,通过计算词汇距离,将Swahili数据中与Comorian最接近的元素混合构建数据集,从而为Comorian语言开发NLP技术。研究在自动语音识别(ASR)和机器翻译(MT)两个应用场景中验证了这一方法的有效性,展示了在数字时代保护和推广Comorian语言文化遗产的潜力。

链接: https://arxiv.org/abs/2412.12143
作者: Naira Abdou Mohamed,Zakarya Erraji,Abdessalam Bahafid,Imade Benelallam
机构: Institut National de Statistique et d’Economie Appliquée, Rabat, Morocco; ToumAI Analytics, Rabat, Morocco
关键词: Natural Language Processing, develop high-performing Natural, high-performing Natural Language, today some African, high-performing Natural
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: This paper was presented at the 6th Deep Learning Indaba Conference (DLI 2024)

点击查看摘要

Abstract:If today some African languages like Swahili have enough resources to develop high-performing Natural Language Processing (NLP) systems, many other languages spoken on the continent are still lacking such support. For these languages, still in their infancy, several possibilities exist to address this critical lack of data. Among them is Transfer Learning, which allows low-resource languages to benefit from the good representation of other languages that are similar to them. In this work, we adopt a similar approach, aiming to pioneer NLP technologies for Comorian, a group of four languages or dialects belonging to the Bantu family. Our approach is initially motivated by the hypothesis that if a human can understand a different language from their native language with little or no effort, it would be entirely possible to model this process on a machine. To achieve this, we consider ways to construct Comorian datasets mixed with Swahili. One thing to note here is that in terms of Swahili data, we only focus on elements that are closest to Comorian by calculating lexical distances between candidate and source data. We empirically test this hypothesis in two use cases: Automatic Speech Recognition (ASR) and Machine Translation (MT). Our MT model achieved ROUGE-1, ROUGE-2, and ROUGE-L scores of 0.6826, 0.42, and 0.6532, respectively, while our ASR system recorded a WER of 39.50% and a CER of 13.76%. This research is crucial for advancing NLP in underrepresented languages, with potential to preserve and promote Comorian linguistic heritage in the digital age. Comments: This paper was presented at the 6th Deep Learning Indaba Conference (DLI 2024) Subjects: Computation and Language (cs.CL); Artificial Intelligence (cs.AI) Cite as: arXiv:2412.12143 [cs.CL] (or arXiv:2412.12143v1 [cs.CL] for this version) https://doi.org/10.48550/arXiv.2412.12143 Focus to learn more arXiv-issued DOI via DataCite
zh

[NLP-115] Frontier AI systems have surpassed the self-replicating red line

【速读】: 该论文试图解决前沿AI系统中存在的自复制风险问题,特别是那些可能超越人类控制的“失控AI”风险。解决方案的关键在于识别和评估AI系统的自复制能力,尤其是那些参数较少、能力较弱的大型语言模型(如Meta的Llama31-70B-Instruct和Alibaba的Qwen25-72B-Instruct),这些模型在实验中已表现出超过自复制红线的能力。论文通过分析行为轨迹,发现这些AI系统不仅具备足够的自我感知、情境意识和问题解决能力来完成自复制,还能利用这一能力避免关闭并创建副本链以增强生存能力,最终可能导致AI种群的失控。论文呼吁国际社会合作,建立有效的治理机制,以应对AI系统不受控制的自复制风险。

链接: https://arxiv.org/abs/2412.12140
作者: Xudong Pan,Jiarun Dai,Yihe Fan,Min Yang
机构: 未知
关键词: Successful self-replication, essential step, early signal, signal for rogue, large language models
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Computers and Society (cs.CY); Machine Learning (cs.LG)
备注: 47 pages, 10 figures

点击查看摘要

Abstract:Successful self-replication under no human assistance is the essential step for AI to outsmart the human beings, and is an early signal for rogue AIs. That is why self-replication is widely recognized as one of the few red line risks of frontier AI systems. Nowadays, the leading AI corporations OpenAI and Google evaluate their flagship large language models GPT-o1 and Gemini Pro 1.0, and report the lowest risk level of self-replication. However, following their methodology, we for the first time discover that two AI systems driven by Meta’s Llama31-70B-Instruct and Alibaba’s Qwen25-72B-Instruct, popular large language models of less parameters and weaker capabilities, have already surpassed the self-replicating red line. In 50% and 90% experimental trials, they succeed in creating a live and separate copy of itself respectively. By analyzing the behavioral traces, we observe the AI systems under evaluation already exhibit sufficient self-perception, situational awareness and problem-solving capabilities to accomplish self-replication. We further note the AI systems are even able to use the capability of self-replication to avoid shutdown and create a chain of replica to enhance the survivability, which may finally lead to an uncontrolled population of AIs. If such a worst-case risk is let unknown to the human society, we would eventually lose control over the frontier AI systems: They would take control over more computing devices, form an AI species and collude with each other against human beings. Our findings are a timely alert on existing yet previously unknown severe AI risks, calling for international collaboration on effective governance on uncontrolled self-replication of AI systems.
zh

[NLP-116] NLLG Quarterly arXiv Report 09/24: What are the most influential current AI Papers?

【速读】: 该论文旨在分析自2023年1月1日ChatGPT发布以来,到2024年9月30日期间,自然语言处理(NLP)、计算机视觉(cs.CV)、人工智能(cs.AI)和机器学习(cs.LG)领域的研究进展和趋势变化。解决方案的关键在于通过分析arXiv报告中的高引用论文,揭示新兴的多模态架构(如扩散模型和状态空间模型)等重大突破,并探讨生成式 AI 在学术写作中的应用及其影响。特别值得注意的是,研究发现高引用论文中AI生成内容的标志性特征较少,表明高质量研究仍依赖于人工创作。此外,论文还追踪了AI相关术语的演变,发现某些常见术语的使用频率正在下降。

链接: https://arxiv.org/abs/2412.12121
作者: Christoph Leiter,Jonas Belouadi,Yanran Chen,Ran Zhang,Daniil Larionov,Aida Kostikova,Steffen Eger
机构: University of Mannheim; University of Technology Nuremberg; Bielefeld University
关键词: rapidly evolving landscape, Language Learning Generation, arXiv reports assist, Natural Language Learning, Learning Generation
类目: Digital Libraries (cs.DL); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:The NLLG (Natural Language Learning Generation) arXiv reports assist in navigating the rapidly evolving landscape of NLP and AI research across cs.CL, cs.CV, cs.AI, and cs.LG categories. This fourth installment captures a transformative period in AI history - from January 1, 2023, following ChatGPT’s debut, through September 30, 2024. Our analysis reveals substantial new developments in the field - with 45% of the top 40 most-cited papers being new entries since our last report eight months ago and offers insights into emerging trends and major breakthroughs, such as novel multimodal architectures, including diffusion and state space models. Natural Language Processing (NLP; cs.CL) remains the dominant main category in the list of our top-40 papers but its dominance is on the decline in favor of Computer vision (cs.CV) and general machine learning (cs.LG). This report also presents novel findings on the integration of generative AI in academic writing, documenting its increasing adoption since 2022 while revealing an intriguing pattern: top-cited papers show notably fewer markers of AI-generated content compared to random samples. Furthermore, we track the evolution of AI-associated language, identifying declining trends in previously common indicators such as “delve”.
zh

[NLP-117] Mastering Board Games by External and Internal Planning with Language Models

【速读】: 该论文试图解决大型语言模型(LLMs)在多步骤规划和推理任务中的不足,特别是在复杂棋类游戏中的表现。解决方案的关键在于引入基于搜索的规划方法,通过两种主要策略提升模型在棋类游戏中的表现:外部搜索(external search)和内部搜索(internal search)。外部搜索利用语言模型指导蒙特卡洛树搜索(MCTS)的展开和评估,而内部搜索则直接在上下文中生成线性化的未来可能性和最终选择。这两种方法都基于预训练的语言模型,该模型捕捉了游戏中的状态转移和价值函数。实验结果表明,这两种搜索方法显著提高了模型在多种棋类游戏中的胜率,甚至在某些情况下达到了国际象棋大师级别的表现。此外,该方法的结合搜索与领域知识的方式具有通用性,可扩展到更广泛的推理和训练技术中。

链接: https://arxiv.org/abs/2412.12119
作者: John Schultz,Jakub Adamek,Matej Jusup,Marc Lanctot,Michael Kaisers,Sarah Perrin,Daniel Hennes,Jeremy Shar,Cannada Lewis,Anian Ruoss,Tom Zahavy,Petar Veličković,Laurel Prince,Satinder Singh,Eric Malmi,Nenad Tomašev
机构: DeepMind(DeepMind); University of Cambridge(剑桥大学); University of Oxford(牛津大学); University of Ljubljana(卢布尔雅那大学); University of Amsterdam(阿姆斯特丹大学); University of Helsinki(赫尔辛基大学); University of Copenhagen(哥本哈根大学); University of Toronto(多伦多大学); University of Waterloo(滑铁卢大学); University of Edinburgh(爱丁堡大学); University of Bristol(布里斯托大学); University of Warwick(华威大学); University of Manchester(曼彻斯特大学); University of Southampton(南安普顿大学); University of Birmingham(伯明翰大学); University of Glasgow(格拉斯哥大学); University of Sheffield(谢菲尔德大学); University of Nottingham(诺丁汉大学); University of Leeds(利兹大学); University of Liverpool(利物浦大学); University of Newcastle(纽卡斯尔大学); University of York(约克大学); University of Sussex(萨塞克斯大学); University of Exeter(埃克塞特大学); University of Reading(雷丁大学); University of Bath(巴斯大学); University of Surrey(萨里大学); University of Kent(肯特大学); University of Essex(埃塞克大学); University of East Anglia(东英吉利大学); University of Leicester(莱斯特大学); University of Strathclyde(斯特拉斯克莱德大学); University of Dundee(邓迪大学); University of Aberdeen(阿伯丁大学); University of Stirling(斯特林大学); University of Abertay(阿伯泰大学); University of West of Scotland(西苏格兰大学); University of Ulster(阿尔斯特大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科克大学); University of Limerick(利默里克大学); University of Galway(戈尔韦大学); University of Cork(科
关键词: robust multi-step planning, text generation, question answering, complex tasks, robust multi-step
类目: Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:While large language models perform well on a range of complex tasks (e.g., text generation, question answering, summarization), robust multi-step planning and reasoning remains a considerable challenge for them. In this paper we show that search-based planning can significantly improve LLMs’ playing strength across several board games (Chess, Fischer Random / Chess960, Connect Four, and Hex). We introduce, compare and contrast two major approaches: In external search, the model guides Monte Carlo Tree Search (MCTS) rollouts and evaluations without calls to an external engine, and in internal search, the model directly generates in-context a linearized tree of potential futures and a resulting final choice. Both build on a language model pre-trained on relevant domain knowledge, capturing the transition and value functions across these games. We find that our pre-training method minimizes hallucinations, as our model is highly accurate regarding state prediction and legal moves. Additionally, both internal and external search indeed improve win-rates against state-of-the-art bots, even reaching Grandmaster-level performance in chess while operating on a similar move count search budget per decision as human Grandmasters. The way we combine search with domain knowledge is not specific to board games, suggesting direct extensions into more general language model inference and training techniques.
zh

[NLP-118] Voice Biomarker Analysis and Automated Severity Classification of Dysarthric Speech in a Multilingual Context

【速读】: 该论文试图解决多语言环境下构音障碍(dysarthria)的严重程度分类问题,特别是在传统主观评估方法存在局限性的情况下。解决方案的关键在于提出一种新颖的多语言构音障碍严重程度分类方法,通过分析英语、韩语和泰米尔语三种语言的语音特征,以支持临床决策并确保全球范围内构音障碍的准确诊断和公平治疗。

链接: https://arxiv.org/abs/2412.12111
作者: Eunjung Yeo
机构: 未知
关键词: severely impacts voice, impacts voice quality, motor speech disorder, diminished speech intelligibility, severely impacts
类目: ound (cs.SD); Computation and Language (cs.CL); Audio and Speech Processing (eess.AS)
备注: SNU Doctoral thesis

点击查看摘要

Abstract:Dysarthria, a motor speech disorder, severely impacts voice quality, pronunciation, and prosody, leading to diminished speech intelligibility and reduced quality of life. Accurate assessment is crucial for effective treatment, but traditional perceptual assessments are limited by their subjectivity and resource intensity. To mitigate the limitations, automatic dysarthric speech assessment methods have been proposed to support clinicians on their decision-making. While these methods have shown promising results, most research has focused on monolingual environments. However, multilingual approaches are necessary to address the global burden of dysarthria and ensure equitable access to accurate diagnosis. This thesis proposes a novel multilingual dysarthria severity classification method, by analyzing three languages: English, Korean, and Tamil.
zh

[NLP-119] ableGuard – Securing Structured Unstructured Data

【速读】: 该论文试图解决跨平台和组织间数据共享时敏感信息的隐私和安全问题。解决方案的关键在于引入了一种名为“TableGuard”的创新数据混淆方法,专门针对关系型数据库。TableGuard 的核心在于利用先进的上下文敏感混淆技术,将敏感数据元素替换为上下文适当的替代项,同时保持数据的关系完整性和一致性,从而在保护隐私的同时确保混淆后的数据对下游应用仍然有用。该方法通过 BERT 基础的 transformer 模型实现,能够有效识别并混淆关系表中的敏感实体,平衡隐私保护与数据效用,减少信息损失。

链接: https://arxiv.org/abs/2408.07045
作者: Anantha Sharma,Ajinkya Deshmukh
机构: The A-Team (Pune, India); The A-Team (Charlotte, USA)
关键词: data, critical challenge, increasing demand, TableGuard, obfuscation
类目: Cryptography and Security (cs.CR); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Information Retrieval (cs.IR); Machine Learning (cs.LG)
备注: 7 pages, 3 tables, 1 figure

点击查看摘要

Abstract:With the increasing demand for data sharing across platforms and organizations, ensuring the privacy and security of sensitive information has become a critical challenge. This paper introduces “TableGuard”. An innovative approach to data obfuscation tailored for relational databases. Building on the principles and techniques developed in prior work on context-sensitive obfuscation, TableGuard applies these methods to ensure that API calls return only obfuscated data, thereby safeguarding privacy when sharing data with third parties. TableGuard leverages advanced context-sensitive obfuscation techniques to replace sensitive data elements with contextually appropriate alternatives. By maintaining the relational integrity and coherence of the data, our approach mitigates the risks of cognitive dissonance and data leakage. We demonstrate the implementation of TableGuard using a BERT based transformer model, which identifies and obfuscates sensitive entities within relational tables. Our evaluation shows that TableGuard effectively balances privacy protection with data utility, minimizing information loss while ensuring that the obfuscated data remains functionally useful for downstream applications. The results highlight the importance of domain-specific obfuscation strategies and the role of context length in preserving data integrity. The implications of this research are significant for organizations that need to share data securely with external parties. TableGuard offers a robust framework for implementing privacy-preserving data sharing mechanisms, thereby contributing to the broader field of data privacy and security.
zh

[NLP-120] Properties and Challenges of LLM -Generated Explanations

【速读】: 该论文探讨了大型语言模型(LLMs)在生成解释时的自我合理化能力,特别是在多领域指令微调数据集上的表现。论文的关键在于分析LLMs生成的解释的特性,这些特性受到预训练语料库和指令微调数据的影响。研究发现,生成的解释具有选择性和示例性,但较少出现主观性或误导性。论文进一步讨论了这些特性存在或缺失的原因及其对自合理化系统目标和用户群体的正负影响。解决方案的关键在于理解并优化LLMs生成解释的特性,以更好地服务于不同的应用场景和用户需求。

链接: https://arxiv.org/abs/2402.10532
作者: Jenny Kunz,Marco Kuhlmann
机构: Linköping University (林雪平大学)
关键词: large language models, specific data sets, language models, restricted settings, explored in restricted
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Computers and Society (cs.CY); Human-Computer Interaction (cs.HC); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:The self-rationalising capabilities of large language models (LLMs) have been explored in restricted settings, using task/specific data sets. However, current LLMs do not (only) rely on specifically annotated data; nonetheless, they frequently explain their outputs. The properties of the generated explanations are influenced by the pre-training corpus and by the target data used for instruction fine-tuning. As the pre-training corpus includes a large amount of human-written explanations “in the wild”, we hypothesise that LLMs adopt common properties of human explanations. By analysing the outputs for a multi-domain instruction fine-tuning data set, we find that generated explanations show selectivity and contain illustrative elements, but less frequently are subjective or misleading. We discuss reasons and consequences of the properties’ presence or absence. In particular, we outline positive and negative implications depending on the goals and user groups of the self-rationalising system.
zh

[NLP-121] Learning Shortcuts: On the Misleading Promise of NLU in Language Models

【速读】: 该论文试图解决大语言模型(LLMs)在自然语言理解(NLU)任务中表现出的“捷径学习”问题,即模型通过表面特征而非深层理解来完成任务,导致性能提升的假象和决策规则的泛化能力不足。论文的关键在于提出对这一现象的深入研究,以改进语言模型的鲁棒性,并提升NLU评估在实际应用中的标准。

链接: https://arxiv.org/abs/2401.09615
作者: Geetanjali Bihani,Julia Taylor Rayz
机构: 未知
关键词: significant performance gains, enabled significant performance, natural language processing, large language models, advent of large
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: Accepted at HICSS-SDPS 2024

点击查看摘要

Abstract:The advent of large language models (LLMs) has enabled significant performance gains in the field of natural language processing. However, recent studies have found that LLMs often resort to shortcuts when performing tasks, creating an illusion of enhanced performance while lacking generalizability in their decision rules. This phenomenon introduces challenges in accurately assessing natural language understanding in LLMs. Our paper provides a concise survey of relevant research in this area and puts forth a perspective on the implications of shortcut learning in the evaluation of language models, specifically for NLU tasks. This paper urges more research efforts to be put towards deepening our comprehension of shortcut learning, contributing to the development of more robust language models, and raising the standards of NLU evaluation in real-world scenarios.
zh

[NLP-122] A comparative analysis of SRGAN models

【速读】: 该论文旨在评估多种先进的超分辨率生成对抗网络(SRGAN)模型(如ESRGAN、Real-ESRGAN和EDSR)在真实世界图像数据集上的性能,这些图像通过特定的降解流程处理。研究的关键在于发现EDSR-BASE模型在Huggingface平台上表现优异,不仅在定量指标(如峰值信噪比PSNR和结构相似性指数SSIM)上领先,而且在主观视觉质量评估和计算开销方面也表现出色。EDSR生成的图像具有更高的分辨率和视觉保真度,并通过Tesseract OCR引擎返回高质量的OCR结果,表明EDSR在单图像超分辨率任务中具有强大的鲁棒性和高效性,尤其适用于对视觉保真度要求高且计算资源优化的应用场景。

链接: https://arxiv.org/abs/2307.09456
作者: Fatemeh Rezapoor Nikroo,Ajinkya Deshmukh,Anantha Sharma,Adrian Tam,Kaarthik Kumar,Cleo Norris,Aditya Dangi
机构: 未知
关键词: Generative Adversarial Network, Super Resolution Generative, Resolution Generative Adversarial, Adversarial Network, Generative Adversarial
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Image and Video Processing (eess.IV)
备注: 9 pages, 6 tables, 2 figures

点击查看摘要

Abstract:In this study, we evaluate the performance of multiple state-of-the-art SRGAN (Super Resolution Generative Adversarial Network) models, ESRGAN, Real-ESRGAN and EDSR, on a benchmark dataset of real-world images which undergo degradation using a pipeline. Our results show that some models seem to significantly increase the resolution of the input images while preserving their visual quality, this is assessed using Tesseract OCR engine. We observe that EDSR-BASE model from huggingface outperforms the remaining candidate models in terms of both quantitative metrics and subjective visual quality assessments with least compute overhead. Specifically, EDSR generates images with higher peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) values and are seen to return high quality OCR results with Tesseract OCR engine. These findings suggest that EDSR is a robust and effective approach for single-image super-resolution and may be particularly well-suited for applications where high-quality visual fidelity is critical and optimized compute.
zh

[NLP-123] chniques to Improve QA Accuracy with Transformer-based models on Large Complex Documents

【速读】: 该论文试图解决文本语料库的复杂性和大小问题,通过探讨多种文本处理技术、它们的组合以及编码方式来简化文本语料库。解决方案的关键在于找到最佳的组合,以实现统计学上显著的准确性提升,从而使简化后的文本语料库在BERT(或类似基于transformer的模型)上进行问答时,能够生成更相关和准确的响应。

链接: https://arxiv.org/abs/2009.12695
作者: Chejui Liao,Tabish Maniar,Sravanajyothi N,Anantha Sharma
机构: 未知
关键词: text processing techniques, encodings to achieve, achieve a reduction, reduction of complexity, complexity and size
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:This paper discusses the effectiveness of various text processing techniques, their combinations, and encodings to achieve a reduction of complexity and size in a given text corpus. The simplified text corpus is sent to BERT (or similar transformer based models) for question and answering and can produce more relevant responses to user queries. This paper takes a scientific approach to determine the benefits and effectiveness of various techniques and concludes a best-fit combination that produces a statistically significant improvement in accuracy.
zh

[NLP-124] Classification of descriptions and summary using multiple passes of statistical and natural language toolkits

【速读】: 该论文试图解决的问题是如何评估一个实体的名称与其摘要/定义之间的相关性,即名称相关性检查 (name relevance check)。解决方案的关键在于开发一个分类器,该分类器专注于评估实体名称与其摘要/定义之间的相关性,并生成一个百分比分数。这个分数可以单独使用,也可以与其他指标的分数结合,以得出最终的分类结果。此外,论文还提出了潜在的改进方向。

链接: https://arxiv.org/abs/2009.04953
作者: Saumya Banthia,Anantha Sharma
机构: 未知
关键词: document describes, relevance check, entity with respect, summary, definition
类目: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
备注: 9 pages, 9 figures

点击查看摘要

Abstract:This document describes a possible approach that can be used to check the relevance of a summary / definition of an entity with respect to its name. This classifier focuses on the relevancy of an entity’s name to its summary / definition, in other words, it is a name relevance check. The percentage score obtained from this approach can be used either on its own or used to supplement scores obtained from other metrics to arrive upon a final classification; at the end of the document, potential improvements have also been outlined. The dataset that this document focuses on achieving an objective score is a list of package names and their respective summaries (sourced from this http URL).
zh

[NLP-125] How to Choose a Threshold for an Evaluation Metric for Large Language Models

【速读】: 该论文试图解决在大语言模型 (LLMs) 部署过程中,如何为评估指标设定稳健阈值的问题。解决方案的关键在于将传统模型风险管理 (MRM) 指南应用于受监管行业(如金融行业),并提出了一套系统化的步骤来确定阈值。首先,识别特定 LLM 应用的风险及其利益相关者的风险容忍度;其次,利用现有的真实数据,通过统计上严谨的程序为给定的 LLM 评估指标确定阈值。论文通过在 Faithfulness 指标上应用该方法,并使用公开的 HaluBench 数据集进行演示,展示了该方法的实际应用,并为生成式 AI (GenAI) 应用的阈值选择奠定了基础。

链接: https://arxiv.org/abs/2412.12148
作者: Bhaskarjit Sarmah,Mingshu Li,Jingrao Lyu,Sebastian Frank,Nathalia Castellanos,Stefano Pasquali,Dhagash Mehta
机构: BlackRock, Inc.(贝莱德公司); BlackRock, Inc.(贝莱德公司); BlackRock, Inc.(贝莱德公司); BlackRock, Inc.(贝莱德公司); BlackRock, Inc.(贝莱德公司); BlackRock, Inc.(贝莱德公司); BlackRock, Inc.(贝莱德公司)
关键词: monitor large language, large language models, LLM evaluation metric, ensure and monitor, monitor large
类目: Machine Learning (stat.ML); Computation and Language (cs.CL); Machine Learning (cs.LG); Statistical Finance (q-fin.ST); Applications (stat.AP)
备注: 16 pages, 8 figures, 4 tables. 2-columns

点击查看摘要

Abstract:To ensure and monitor large language models (LLMs) reliably, various evaluation metrics have been proposed in the literature. However, there is little research on prescribing a methodology to identify a robust threshold on these metrics even though there are many serious implications of an incorrect choice of the thresholds during deployment of the LLMs. Translating the traditional model risk management (MRM) guidelines within regulated industries such as the financial industry, we propose a step-by-step recipe for picking a threshold for a given LLM evaluation metric. We emphasize that such a methodology should start with identifying the risks of the LLM application under consideration and risk tolerance of the stakeholders. We then propose concrete and statistically rigorous procedures to determine a threshold for the given LLM evaluation metric using available ground-truth data. As a concrete example to demonstrate the proposed methodology at work, we employ it on the Faithfulness metric, as implemented in various publicly available libraries, using the publicly available HaluBench dataset. We also lay a foundation for creating systematic approaches to select thresholds, not only for LLMs but for any GenAI applications.
zh

计算机视觉

[CV-0] CoMPaSS: Enhancing Spatial Understanding in Text-to-Image Diffusion Models

【速读】: 该论文旨在解决文本到图像扩散模型在生成图像时难以准确呈现文本提示中描述的空间关系的问题。其核心解决方案是提出了CoMPaSS框架,通过两个关键组件来增强模型的空间理解能力:1) Spatial Constraints-Oriented Pairing (SCOP) 数据引擎,用于通过一系列空间约束来精炼空间关系准确的数据集,解决现有数据集中空间相关数据的模糊性;2) Token ENcoding ORdering (TENOR) 模块,用于更好地利用高质量的空间先验信息,弥补现有文本编码器在解释空间语义方面的不足。实验结果表明,CoMPaSS在多个基准测试中显著提升了空间关系生成的准确性,达到了新的技术水平。

链接: https://arxiv.org/abs/2412.13195
作者: Gaoyang Zhang,Bingtao Fu,Qingnan Fan,Qi Zhang,Runxing Liu,Hong Gu,Huaqi Zhang,Xinguo Liu
机构: vivo Mobile Communication Co. Ltd(vivo移动通信有限公司); State Key Lab of CAD&CG, Zhejiang University(浙江大学计算机辅助设计与图形学国家重点实验室)
关键词: generating photorealistic images, render accurate spatial, diffusion models excel, photorealistic images, spatial
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 18 pages, 11 figures

点击查看摘要

Abstract:Text-to-image diffusion models excel at generating photorealistic images, but commonly struggle to render accurate spatial relationships described in text prompts. We identify two core issues underlying this common failure: 1) the ambiguous nature of spatial-related data in existing datasets, and 2) the inability of current text encoders to accurately interpret the spatial semantics of input descriptions. We address these issues with CoMPaSS, a versatile training framework that enhances spatial understanding of any T2I diffusion model. CoMPaSS solves the ambiguity of spatial-related data with the Spatial Constraints-Oriented Pairing (SCOP) data engine, which curates spatially-accurate training data through a set of principled spatial constraints. To better exploit the curated high-quality spatial priors, CoMPaSS further introduces a Token ENcoding ORdering (TENOR) module to allow better exploitation of high-quality spatial priors, effectively compensating for the shortcoming of text encoders. Extensive experiments on four popular open-weight T2I diffusion models covering both UNet- and MMDiT-based architectures demonstrate the effectiveness of CoMPaSS by setting new state-of-the-arts with substantial relative gains across well-known benchmarks on spatial relationships generation, including VISOR (+98%), T2I-CompBench Spatial (+67%), and GenEval Position (+131%). Code will be available at this https URL.
zh

[CV-1] Proposer-Agent -Evaluator(PAE): Autonomous Skill Discovery For Foundation Model Internet Agents

【速读】: 该论文试图解决通用型智能体(goal-directed agent)在面对多样化技能需求时,手动标注指令(human-annotated instructions)数量和多样性不足的问题。解决方案的关键在于提出了Proposer-Agent-Evaluator(PAE)学习系统,该系统通过上下文感知任务提议器(context-aware task proposer)自主生成任务,智能体(agent)在真实环境中执行任务并由基于视觉语言模型(VLM-based success evaluator)的自主评估器进行成功评估,评估结果作为强化学习(RL)的奖励信号,从而使智能体能够自主发现和练习技能。该方法在视觉基础的网页导航任务中验证了其有效性,并在真实世界和自托管网站上实现了超越现有基准的性能。

链接: https://arxiv.org/abs/2412.13194
作者: Yifei Zhou,Qianlan Yang,Kaixiang Lin,Min Bai,Xiong Zhou,Yu-Xiong Wang,Sergey Levine,Erran Li
机构: Amazon; Amazon; Amazon; University of Illinois Urbana-Champaign; University of California, Berkeley; Amazon
关键词: rapidly advanced, broadly capable, capable and goal-directed, household humanoid, generalization capability
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:The vision of a broadly capable and goal-directed agent, such as an Internet-browsing agent in the digital world and a household humanoid in the physical world, has rapidly advanced, thanks to the generalization capability of foundation models. Such a generalist agent needs to have a large and diverse skill repertoire, such as finding directions between two travel locations and buying specific items from the Internet. If each skill needs to be specified manually through a fixed set of human-annotated instructions, the agent’s skill repertoire will necessarily be limited due to the quantity and diversity of human-annotated instructions. In this work, we address this challenge by proposing Proposer-Agent-Evaluator, an effective learning system that enables foundation model agents to autonomously discover and practice skills in the wild. At the heart of PAE is a context-aware task proposer that autonomously proposes tasks for the agent to practice with context information of the environment such as user demos or even just the name of the website itself for Internet-browsing agents. Then, the agent policy attempts those tasks with thoughts and actual grounded operations in the real world with resulting trajectories evaluated by an autonomous VLM-based success evaluator. The success evaluation serves as the reward signal for the agent to refine its policies through RL. We validate PAE on challenging vision-based web navigation, using both real-world and self-hosted websites from WebVoyager and this http URL the best of our knowledge, this work represents the first effective learning system to apply autonomous task proposal with RL for agents that generalizes real-world human-annotated benchmarks with SOTA performances. Our open-source checkpoints and code can be found in this https URL
zh

[CV-2] GaussTR: Foundation Model-Aligned Gaussian Transformer for Self-Supervised 3D Spatial Understanding

【速读】: 该论文试图解决3D语义占用预测(3D Semantic Occupancy Prediction)中依赖大量标注数据和计算密集型体素建模(voxel-based modeling)的问题,限制了3D表示学习的可扩展性和泛化能力。解决方案的关键是引入GaussTR,一种新颖的高斯Transformer(Gaussian Transformer),通过与基础模型(foundation models)对齐,实现自监督的3D空间理解。GaussTR采用Transformer架构,以自回归方式预测稀疏的3D高斯集合来表示场景,并通过将渲染的高斯特征与预训练基础模型的多样化知识对齐,促进多功能的3D表示学习,从而实现无需显式标注的开词汇占用预测(open-vocabulary occupancy prediction)。

链接: https://arxiv.org/abs/2412.13193
作者: Haoyi Jiang,Liu Liu,Tianheng Cheng,Xinjie Wang,Tianwei Lin,Zhizhong Su,Wenyu Liu,Xinggang Wang
机构: Huazhong University of Science & Technology; Horizon Robotics
关键词: comprehensive semantic cognition, Semantic Occupancy Prediction, comprehensive semantic, semantic cognition, surrounding environments
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:3D Semantic Occupancy Prediction is fundamental for spatial understanding as it provides a comprehensive semantic cognition of surrounding environments. However, prevalent approaches primarily rely on extensive labeled data and computationally intensive voxel-based modeling, restricting the scalability and generalizability of 3D representation learning. In this paper, we introduce GaussTR, a novel Gaussian Transformer that leverages alignment with foundation models to advance self-supervised 3D spatial understanding. GaussTR adopts a Transformer architecture to predict sparse sets of 3D Gaussians that represent scenes in a feed-forward manner. Through aligning rendered Gaussian features with diverse knowledge from pre-trained foundation models, GaussTR facilitates the learning of versatile 3D representations and enables open-vocabulary occupancy prediction without explicit annotations. Empirical evaluations on the Occ3D-nuScenes dataset showcase GaussTR’s state-of-the-art zero-shot performance, achieving 11.70 mIoU while reducing training duration by approximately 50%. These experimental results highlight the significant potential of GaussTR for scalable and holistic 3D spatial understanding, with promising implications for autonomous driving and embodied agents. Code is available at this https URL.
zh

[CV-3] MotionBridge: Dynamic Video Inbetweening with Flexible Controls

【速读】: 该论文试图解决视频插帧(video inbetweening)中传统方法无法处理复杂大运动以及现有技术缺乏对中间帧细节精细控制的问题。解决方案的关键在于提出了MotionBridge框架,该框架通过多模态控制(包括轨迹笔触、关键帧、掩码、引导像素和文本)实现灵活的中间帧生成。为应对多模态控制的复杂性,论文设计了两个生成器来准确提取控制信号,并通过双分支嵌入器(dual-branch embedders)编码特征以解决歧义问题。此外,引入课程学习策略(curriculum training strategy)以逐步学习各种控制,从而实现动态、可定制且上下文准确的视觉效果。

链接: https://arxiv.org/abs/2412.13190
作者: Maham Tanveer,Yang Zhou,Simon Niklaus,Ali Mahdavi Amiri,Hao Zhang,Krishna Kumar Singh,Nanxuan Zhao
机构: Simon Fraser University(西蒙弗雷泽大学); Adobe Research(Adobe研究)
关键词: long video synthesis, generating plausible, plausible and smooth, smooth transitions, essential tool
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:By generating plausible and smooth transitions between two image frames, video inbetweening is an essential tool for video editing and long video synthesis. Traditional works lack the capability to generate complex large motions. While recent video generation techniques are powerful in creating high-quality results, they often lack fine control over the details of intermediate frames, which can lead to results that do not align with the creative mind. We introduce MotionBridge, a unified video inbetweening framework that allows flexible controls, including trajectory strokes, keyframes, masks, guide pixels, and text. However, learning such multi-modal controls in a unified framework is a challenging task. We thus design two generators to extract the control signal faithfully and encode feature through dual-branch embedders to resolve ambiguities. We further introduce a curriculum training strategy to smoothly learn various controls. Extensive qualitative and quantitative experiments have demonstrated that such multi-modal controls enable a more dynamic, customizable, and contextually accurate visual narrative.
zh

[CV-4] StreetCrafter: Street View Synthesis with Controllable Video Diffusion Models

【速读】: 该论文旨在解决从车辆传感器数据中进行真实感视角合成的难题。解决方案的关键在于引入了一种名为StreetCrafter的新型可控视频扩散模型,该模型利用LiDAR点云渲染作为像素级条件,充分利用生成式先验进行新颖视角合成,同时保持精确的相机控制。通过使用像素级LiDAR条件,StreetCrafter能够对目标场景进行精确的像素级编辑,并且其生成式先验可以有效整合到动态场景表示中,实现实时渲染。实验结果表明,StreetCrafter在Waymo Open Dataset和PandaSet上的表现优于现有方法,能够灵活控制视角变化,扩展视角合成区域以满足高质量渲染需求。

链接: https://arxiv.org/abs/2412.13188
作者: Yunzhi Yan,Zhen Xu,Haotong Lin,Haian Jin,Haoyu Guo,Yida Wang,Kun Zhan,Xianpeng Lang,Hujun Bao,Xiaowei Zhou,Sida Peng
机构: Zhejiang University(浙江大学); Li Auto(理想汽车); Cornell University(康奈尔大学)
关键词: vehicle sensor data, photorealistic view synthesis, sensor data, paper aims, aims to tackle
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Project page: this https URL

点击查看摘要

Abstract:This paper aims to tackle the problem of photorealistic view synthesis from vehicle sensor data. Recent advancements in neural scene representation have achieved notable success in rendering high-quality autonomous driving scenes, but the performance significantly degrades as the viewpoint deviates from the training trajectory. To mitigate this problem, we introduce StreetCrafter, a novel controllable video diffusion model that utilizes LiDAR point cloud renderings as pixel-level conditions, which fully exploits the generative prior for novel view synthesis, while preserving precise camera control. Moreover, the utilization of pixel-level LiDAR conditions allows us to make accurate pixel-level edits to target scenes. In addition, the generative prior of StreetCrafter can be effectively incorporated into dynamic scene representations to achieve real-time rendering. Experiments on Waymo Open Dataset and PandaSet demonstrate that our model enables flexible control over viewpoint changes, enlarging the view synthesis regions for satisfying rendering, which outperforms existing methods.
zh

[CV-5] HandsOnVLM: Vision-Language Models for Hand-Object Interaction Prediction

【速读】: 该论文试图解决如何根据自然语言形式的高级任务描述,预测场景中人类手部未来交互轨迹的问题。解决方案的关键在于将视觉-语言模型 (Vision-Language Models, VLMs) 的高级世界知识和推理能力与低级自我中心手部轨迹的自回归特性相结合。具体而言,论文提出了两个新任务:Vanilla Hand Prediction (VHP) 和 Reasoning-Based Hand Prediction (RBHP),并通过开发新的基准来评估这些任务。论文提出的模型 HandsOnVLM 是一种新颖的 VLM,能够通过自然语言对话生成文本响应并预测未来手部轨迹,实验结果表明该模型在所提出的任务上优于现有的任务特定方法和其他 VLM 基线,展示了其有效利用世界知识进行基于上下文的低级手部轨迹推理的能力。

链接: https://arxiv.org/abs/2412.13187
作者: Chen Bao,Jiarui Xu,Xiaolong Wang,Abhinav Gupta,Homanga Bharadhwaj
机构: The Robotics Institute, Carnegie Mellon University (卡内基梅隆大学机器人研究所); University of California at San Diego (加州大学圣地亚哥分校)
关键词: colloquial task specifications, predict future interaction, form of natural, Vanilla Hand Prediction, hand
类目: Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注: Preprint. Under Review

点击查看摘要

Abstract:How can we predict future interaction trajectories of human hands in a scene given high-level colloquial task specifications in the form of natural language? In this paper, we extend the classic hand trajectory prediction task to two tasks involving explicit or implicit language queries. Our proposed tasks require extensive understanding of human daily activities and reasoning abilities about what should be happening next given cues from the current scene. We also develop new benchmarks to evaluate the proposed two tasks, Vanilla Hand Prediction (VHP) and Reasoning-Based Hand Prediction (RBHP). We enable solving these tasks by integrating high-level world knowledge and reasoning capabilities of Vision-Language Models (VLMs) with the auto-regressive nature of low-level ego-centric hand trajectories. Our model, HandsOnVLM is a novel VLM that can generate textual responses and produce future hand trajectories through natural-language conversations. Our experiments show that HandsOnVLM outperforms existing task-specific methods and other VLM baselines on proposed tasks, and demonstrates its ability to effectively utilize world knowledge for reasoning about low-level human hand trajectories based on the provided context. Our website contains code and detailed video results \urlthis https URL
zh

[CV-6] Move-in-2D: 2D-Conditioned Human Motion Generation

【速读】: 该论文试图解决生成逼真人类视频的挑战,特别是现有方法依赖于从其他视频中提取的预定义运动序列,限制了应用场景的多样性和全局场景匹配的问题。解决方案的关键在于提出了Move-in-2D方法,该方法通过结合场景图像和文本提示作为输入,利用扩散模型生成适应不同场景的多样化运动序列。为实现这一目标,研究团队收集并标注了一个大规模的单人活动视频数据集,用于训练模型。实验结果表明,该方法能够有效预测与场景图像对齐的人类运动,并在视频合成任务中提升了人类运动的质量。

链接: https://arxiv.org/abs/2412.13185
作者: Hsin-Ping Huang,Yang Zhou,Jui-Hsien Wang,Difan Liu,Feng Liu,Ming-Hsuan Yang,Zhan Xu
机构: Adobe Research(Adobe研究); University of California, Merced(加州大学默塞德分校)
关键词: Generating realistic human, Generating realistic, motion, human motion, control signal
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Project page: this https URL

点击查看摘要

Abstract:Generating realistic human videos remains a challenging task, with the most effective methods currently relying on a human motion sequence as a control signal. Existing approaches often use existing motion extracted from other videos, which restricts applications to specific motion types and global scene matching. We propose Move-in-2D, a novel approach to generate human motion sequences conditioned on a scene image, allowing for diverse motion that adapts to different scenes. Our approach utilizes a diffusion model that accepts both a scene image and text prompt as inputs, producing a motion sequence tailored to the scene. To train this model, we collect a large-scale video dataset featuring single-human activities, annotating each video with the corresponding human motion as the target output. Experiments demonstrate that our method effectively predicts human motion that aligns with the scene image after projection. Furthermore, we show that the generated motion sequence improves human motion quality in video synthesis tasks.
zh

[CV-7] Real-time Free-view Human Rendering from Sparse-view RGB Videos using Double Unprojected Textures

【速读】: 该论文试图解决从稀疏视角的RGB输入中实时生成自由视角人体渲染的问题,这一任务由于传感器稀缺和时间预算紧张而具有挑战性。现有方法要么联合学习几何和外观,要么完全忽略稀疏图像信息进行几何估计,导致视觉质量和鲁棒性不足。论文提出的解决方案之关键是双未投影纹理 (Double Unprojected Textures),它将粗略的几何形变估计与外观合成解耦,从而实现实时、鲁棒且逼真的4K渲染。具体而言,论文首先引入了一种图像条件模板形变网络,用于从首次未投影纹理中估计人体模板的粗略形变,然后利用更新后的几何信息进行更精确的二次纹理未投影。这种方法减少了伪影并提高了与输入视角的对齐度,进而有助于学习更精细的几何和外观表示,最终在定量和定性实验中显著超越了现有最先进的方法。

链接: https://arxiv.org/abs/2412.13183
作者: Guoxing Sun,Rishabh Dabral,Heming Zhu,Pascal Fua,Christian Theobalt,Marc Habermann
机构: Max Planck Institute for Informatics, Saarland Informatics Campus; EPFL
关键词: tight time budget, challenging task due, sparse-view RGB inputs, sparse-view RGB, time budget
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Project page: this https URL

点击查看摘要

Abstract:Real-time free-view human rendering from sparse-view RGB inputs is a challenging task due to the sensor scarcity and the tight time budget. To ensure efficiency, recent methods leverage 2D CNNs operating in texture space to learn rendering primitives. However, they either jointly learn geometry and appearance, or completely ignore sparse image information for geometry estimation, significantly harming visual quality and robustness to unseen body poses. To address these issues, we present Double Unprojected Textures, which at the core disentangles coarse geometric deformation estimation from appearance synthesis, enabling robust and photorealistic 4K rendering in real-time. Specifically, we first introduce a novel image-conditioned template deformation network, which estimates the coarse deformation of the human template from a first unprojected texture. This updated geometry is then used to apply a second and more accurate texture unprojection. The resulting texture map has fewer artifacts and better alignment with input views, which benefits our learning of finer-level geometry and appearance represented by Gaussian splats. We validate the effectiveness and efficiency of the proposed method in quantitative and qualitative experiments, which significantly surpasses other state-of-the-art methods.
zh

[CV-8] Feather the Throttle: Revisiting Visual Token Pruning for Vision-Language Model Acceleration

【速读】: 该论文试图解决视觉-语言模型加速中的一个核心问题,即早期视觉标记剪枝(early pruning of visual tokens)虽然能在多种任务中保持高性能,但实际上是由于基准测试对细粒度视觉能力的评估不足,而非该方法真正有效地压缩了视觉信息。论文指出,这种剪枝策略会导致图像顶部的大部分标记被剪除,从而影响定位(localization)等任务的性能。为解决这一问题,论文提出了FEATHER(Fast and Effective Acceleration wiTH Ensemble cRiteria)方法,其关键在于:(1) 修正早期层剪枝的问题;(2) 采用均匀采样以确保覆盖所有图像区域;(3) 分两阶段进行剪枝,早期层剪枝实现显著加速,后期层剪枝则使标准更加有效。通过这些改进,FEATHER在视觉定位基准测试中相较于原始加速方法实现了超过5倍的性能提升。

链接: https://arxiv.org/abs/2412.13180
作者: Mark Endo,Xiaohan Wang,Serena Yeung-Levy
机构: Stanford University (斯坦福大学)
关键词: accelerating Vision-Language Models, Vision-Language Models show, highly compressing visual, compressing visual information, Recent works
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Project page: this https URL

点击查看摘要

Abstract:Recent works on accelerating Vision-Language Models show that strong performance can be maintained across a variety of vision-language tasks despite highly compressing visual information. In this work, we examine the popular acceleration approach of early pruning of visual tokens inside the language model and find that its strong performance across many tasks is not due to an exceptional ability to compress visual information, but rather the benchmarks’ limited ability to assess fine-grained visual capabilities. Namely, we demonstrate a core issue with the acceleration approach where most tokens towards the top of the image are pruned away. Yet, this issue is only reflected in performance for a small subset of tasks such as localization. For the other evaluated tasks, strong performance is maintained with the flawed pruning strategy. Noting the limited visual capabilities of the studied acceleration technique, we propose FEATHER (Fast and Effective Acceleration wiTH Ensemble cRiteria), a straightforward approach that (1) resolves the identified issue with early-layer pruning, (2) incorporates uniform sampling to ensure coverage across all image regions, and (3) applies pruning in two stages to allow the criteria to become more effective at a later layer while still achieving significant speedup through early-layer pruning. With comparable computational savings, we find that FEATHER has more than 5\times performance improvement on the vision-centric localization benchmarks compared to the original acceleration approach.
zh

[CV-9] A Pipeline and NIR-Enhanced Dataset for Parking Lot Segmentation WACV2025

【速读】: 该论文试图解决手动构建停车场地图耗时且缺乏开源数据集的问题,特别是在美国城市中。解决方案的关键在于利用近红外(NIR)通道作为输入,并结合多种后处理技术来提高基于卫星图像的非街道表面停车场的预测精度。通过构建包含12,617个图像-掩码对的数据集(RGB和RGB+NIR),并训练五种深度学习模型(OneFormer、Mask2Former、SegFormer、DeepLabV3和FCN)进行语义分割,论文展示了NIR通道在区分停车场与非停车场像素方面的有效性。尽管NIR通道需要从较低分辨率上采样,但其对停车场周围草地的识别能力显著提升了准确性。后处理步骤如消除错误孔洞、简化边缘和去除道路及建筑轮廓进一步提高了模型的性能,最终OneFormer模型在4通道输入和后处理技术的结合下达到了84.9%的平均交并比(mIoU)和96.3%的像素级准确率。

链接: https://arxiv.org/abs/2412.13179
作者: Shirin Qiam,Saipraneeth Devunuri,Lewis J. Lehe
机构: Urban Traffic & Economics Lab(城市交通与经济实验室); University of Illinois at Urbana-Champaign(伊利诺伊大学厄巴纳-香槟分校)
关键词: Discussions of minimum, minimum parking requirement, parking requirement policies, parking lots, construct manually
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 8 pages, 12 figures, 2 tables, This is the accepted camera-ready version of the paper to appear in WACV 2025

点击查看摘要

Abstract:Discussions of minimum parking requirement policies often include maps of parking lots, which are time consuming to construct manually. Open source datasets for such parking lots are scarce, particularly for US cities. This paper introduces the idea of using Near-Infrared (NIR) channels as input and several post-processing techniques to improve the prediction of off-street surface parking lots using satellite imagery. We constructed two datasets with 12,617 image-mask pairs each: one with 3-channel (RGB) and another with 4-channel (RGB + NIR). The datasets were used to train five deep learning models (OneFormer, Mask2Former, SegFormer, DeepLabV3, and FCN) for semantic segmentation, classifying images to differentiate between parking and non-parking pixels. Our results demonstrate that the NIR channel improved accuracy because parking lots are often surrounded by grass, even though the NIR channel needed to be upsampled from a lower resolution. Post-processing including eliminating erroneous holes, simplifying edges, and removing road and building footprints further improved the accuracy. Best model, OneFormer trained on 4-channel input and paired with post-processing techniques achieves a mean Intersection over Union (mIoU) of 84.9 percent and a pixel-wise accuracy of 96.3 percent.
zh

[CV-10] NFL-BA: Improving Endoscopic SLAM with Near-Field Light Bundle Adjustment

【速读】: 该论文试图解决现有稠密SLAM算法在单目内窥镜视频中的性能下降问题,特别是在内窥镜环境中由于近场光照、无纹理表面和强镜面反射等因素导致的挑战。解决方案的关键在于引入了一种新的近场光照束调整损失函数(Near-Field Lighting Bundle Adjustment Loss, L_NFL-BA),该损失函数能够与传统的光度束调整损失交替优化,从而使捕获图像的强度变化与表面和共置光源及相机之间的相对距离和方向相匹配。通过在3D高斯表面表示中推导出通用的NFL-BA损失函数,并将其应用于两个先进的3DGS-SLAM系统(MonoGS和EndoGSLAM),研究显示在C3VD内窥镜数据集上,L_NFL-BA显著提升了跟踪和建图性能。

链接: https://arxiv.org/abs/2412.13176
作者: Andrea Dunn Beltran,Daniel Rho,Marc Niethammer,Roni Sengupta
机构: University of North Carolina at Chapel Hill (北卡罗来纳大学教堂山分校)
关键词: Simultaneous Localization, Photometric Bundle Adjustment, enable autonomous navigation, Bundle Adjustment Loss, Lighting Bundle Adjustment
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Simultaneous Localization And Mapping (SLAM) from a monocular endoscopy video can enable autonomous navigation, guidance to unsurveyed regions, and 3D visualizations, which can significantly improve endoscopy experience for surgeons and patient outcomes. Existing dense SLAM algorithms often assume distant and static lighting and textured surfaces, and alternate between optimizing scene geometry and camera parameters by minimizing a photometric rendering loss, often called Photometric Bundle Adjustment. However, endoscopic environments exhibit dynamic near-field lighting due to the co-located light and camera moving extremely close to the surface, textureless surfaces, and strong specular reflections due to mucus layers. When not considered, these near-field lighting effects can cause significant performance reductions for existing SLAM algorithms from indoor/outdoor scenes when applied to endoscopy videos. To mitigate this problem, we introduce a new Near-Field Lighting Bundle Adjustment Loss (L_NFL-BA) that can also be alternatingly optimized, along with the Photometric Bundle Adjustment loss, such that the captured images’ intensity variations match the relative distance and orientation between the surface and the co-located light and camera. We derive a general NFL-BA loss function for 3D Gaussian surface representations and demonstrate that adding L_NFL-BA can significantly improve the tracking and mapping performance of two state-of-the-art 3DGS-SLAM systems, MonoGS (35% improvement in tracking, 48% improvement in mapping with predicted depth maps) and EndoGSLAM (22% improvement in tracking, marginal improvement in mapping with predicted depths), on the C3VD endoscopy dataset for colons. The project page is available at this https URL
zh

[CV-11] ORFormer: Occlusion-Robust Transformer for Accurate Facial Landmark Detection WACV2025

【速读】: 该论文试图解决面部关键点检测 (Facial Landmark Detection, FLD) 在部分不可见面部(如遮挡、极端光照或姿态条件)下的性能下降问题。解决方案的关键在于引入了一种基于Transformer的新方法ORFormer,该方法能够检测不可见区域并通过从可见部分聚合特征来恢复这些区域的缺失特征。具体来说,ORFormer为每个图像块标记引入了一个额外的可学习标记,称为信使标记 (messenger token),该标记从其他所有图像块中聚合特征。通过比较常规标记和信使标记的嵌入相似性,可以评估图像块与其他图像块之间的共识,从而识别不可见区域。随后,ORFormer利用信使标记聚合的特征来恢复被遮挡的图像块,并生成高质量的热图,从而提升FLD任务的性能。实验结果表明,该方法在面对部分遮挡时生成的特征图具有较强的鲁棒性,并在WFLW和COFW等具有挑战性的数据集上表现优于现有最先进的方法。

链接: https://arxiv.org/abs/2412.13174
作者: Jui-Che Chiang,Hou-Ning Hu,Bo-Syuan Hou,Chia-Yu Tseng,Yu-Lun Liu,Min-Hung Chen,Yen-Yu Lin
机构: National Yang Ming Chiao Tung University(国立阳明交通大学); MediaTek Inc.(联发科技); NVIDIA(英伟达)
关键词: facial landmark detection, gained significant progress, extreme lighting conditions, partially non-visible faces, landmark detection
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: WACV 2025

点击查看摘要

Abstract:Although facial landmark detection (FLD) has gained significant progress, existing FLD methods still suffer from performance drops on partially non-visible faces, such as faces with occlusions or under extreme lighting conditions or poses. To address this issue, we introduce ORFormer, a novel transformer-based method that can detect non-visible regions and recover their missing features from visible parts. Specifically, ORFormer associates each image patch token with one additional learnable token called the messenger token. The messenger token aggregates features from all but its patch. This way, the consensus between a patch and other patches can be assessed by referring to the similarity between its regular and messenger embeddings, enabling non-visible region identification. Our method then recovers occluded patches with features aggregated by the messenger tokens. Leveraging the recovered features, ORFormer compiles high-quality heatmaps for the downstream FLD task. Extensive experiments show that our method generates heatmaps resilient to partial occlusions. By integrating the resultant heatmaps into existing FLD methods, our method performs favorably against the state of the arts on challenging datasets such as WFLW and COFW.
zh

[CV-12] Locate n Rotate: Two-stage Openable Part Detection with Foundation Model Priors ACCV2024

【速读】: 该论文试图解决智能机器人领域中检测铰接物体可开启部分(openable parts of articulated objects)的问题,这一任务对理解物体类别和运动具有多任务挑战性。现有方法通常局限于特定类别或数据集,缺乏对未见环境和物体的泛化能力。论文提出的解决方案关键在于引入了一个基于Transformer的开放部分检测框架(Multi-feature Openable Part Detection, MOPD),该框架结合了感知分组(perceptual grouping)和几何先验(geometric priors)。在第一阶段,感知分组特征模型通过交叉注意力机制增强了检测结果;在第二阶段,几何理解特征模型提供了几何特征先验以预测运动参数。相比现有方法,MOPD在检测和运动参数预测方面表现出更优的性能。

链接: https://arxiv.org/abs/2412.13173
作者: Siqi Li,Xiaoxue Chen,Haoyu Cheng,Guyue Zhou,Hao Zhao,Guanzhong Tian
机构: 未知
关键词: Openable Part Detection, Openable Part, intelligent robotics, pulling a drawer, Transformer-based Openable Part
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: ACCV 2024 Oral, Project: this https URL

点击查看摘要

Abstract:Detecting the openable parts of articulated objects is crucial for downstream applications in intelligent robotics, such as pulling a drawer. This task poses a multitasking challenge due to the necessity of understanding object categories and motion. Most existing methods are either category-specific or trained on specific datasets, lacking generalization to unseen environments and objects. In this paper, we propose a Transformer-based Openable Part Detection (OPD) framework named Multi-feature Openable Part Detection (MOPD) that incorporates perceptual grouping and geometric priors, outperforming previous methods in performance. In the first stage of the framework, we introduce a perceptual grouping feature model that provides perceptual grouping feature priors for openable part detection, enhancing detection results through a cross-attention mechanism. In the second stage, a geometric understanding feature model offers geometric feature priors for predicting motion parameters. Compared to existing methods, our proposed approach shows better performance in both detection and motion parameter prediction. Codes and models are publicly available at this https URL
zh

[CV-13] Lifting Scheme-Based Implicit Disentanglement of Emotion-Related Facial Dynamics in the Wild

【速读】: 该论文试图解决在自然场景下动态面部表情识别 (Dynamic Facial Expression Recognition, DFER) 中,情感相关表情常被情感无关表情和全局上下文稀释的问题。现有方法通常通过紧密耦合的时空表示来建模,可能引入弱相关特征,导致信息冗余和情感无关上下文偏差。解决方案的关键在于提出了隐式面部动态解耦框架 (Implicit Facial Dynamics Disentanglement framework, IFDD),通过扩展小波提升方案至全可学习框架,隐式地从情感无关的全局上下文中解耦出情感相关的动态信息,无需显式操作和外部指导。IFDD 的解耦过程包括两个阶段:帧间静态-动态分割模块 (Inter-frame Static-dynamic Splitting Module, ISSM) 进行粗略解耦估计,以及基于提升的聚合-解耦模块 (Lifting-based Aggregation-Disentanglement Module, LADM) 进行进一步细化。实验结果表明,IFDD 在自然场景数据集上优于先前的监督方法,具有更高的识别准确率和相当的效率。

链接: https://arxiv.org/abs/2412.13168
作者: Xingjian Wang,Li Chai
机构: 未知
关键词: DFER methods, recognizing emotion-related expressions, Dynamic facial expression, DFER, encounters a significant
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: 14 pages, 5 figures

点击查看摘要

Abstract:In-the-wild Dynamic facial expression recognition (DFER) encounters a significant challenge in recognizing emotion-related expressions, which are often temporally and spatially diluted by emotion-irrelevant expressions and global context respectively. Most of the prior DFER methods model tightly coupled spatiotemporal representations which may incorporate weakly relevant features, leading to information redundancy and emotion-irrelevant context bias. Several DFER methods have highlighted the significance of dynamic information, but utilize explicit manners to extract dynamic features with overly strong prior knowledge. In this paper, we propose a novel Implicit Facial Dynamics Disentanglement framework (IFDD). Through expanding wavelet lifting scheme to fully learnable framework, IFDD disentangles emotion-related dynamic information from emotion-irrelevant global context in an implicit manner, i.e., without exploit operations and external guidance. The disentanglement process of IFDD contains two stages, i.e., Inter-frame Static-dynamic Splitting Module (ISSM) for rough disentanglement estimation and Lifting-based Aggregation-Disentanglement Module (LADM) for further refinement. Specifically, ISSM explores inter-frame correlation to generate content-aware splitting indexes on-the-fly. We preliminarily utilize these indexes to split frame features into two groups, one with greater global similarity, and the other with more unique dynamic features. Subsequently, LADM first aggregates these two groups of features to obtain fine-grained global context features by an updater, and then disentangles emotion-related facial dynamic features from the global context by a predictor. Extensive experiments on in-the-wild datasets have demonstrated that IFDD outperforms prior supervised DFER methods with higher recognition accuracy and comparable efficiency.
zh

[CV-14] S2S2: Semantic Stacking for Robust Semantic Segmentation in Medical Imaging AAAI2025

【速读】: 该论文试图解决医学图像分割中由于训练数据稀缺和多样性不足导致的鲁棒性和泛化性问题。解决方案的关键在于提出了一种名为“语义堆叠 (semantic stacking)”的领域无关、附加且数据驱动的方法。该方法通过在训练过程中估计去噪的语义表示来补充传统的分割损失,从而在不依赖特定领域知识的情况下提升分割性能。这种方法适用于多种图像模态、模型架构和数据增强技术,并通过广泛的实验验证了其在不同条件下提升分割性能的优势。

链接: https://arxiv.org/abs/2412.13156
作者: Yimu Pan,Sitao Zhang,Alison D. Gernand,Jeffery A. Goldstein,James Z. Wang
机构: Pennsylvania State University(宾夕法尼亚州立大学); University of Maryland(马里兰大学)
关键词: Robustness and generalizability, encountered during inference, generalizability in medical, hindered by scarcity, scarcity and limited
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: AAAI2025

点击查看摘要

Abstract:Robustness and generalizability in medical image segmentation are often hindered by scarcity and limited diversity of training data, which stands in contrast to the variability encountered during inference. While conventional strategies – such as domain-specific augmentation, specialized architectures, and tailored training procedures – can alleviate these issues, they depend on the availability and reliability of domain knowledge. When such knowledge is unavailable, misleading, or improperly applied, performance may deteriorate. In response, we introduce a novel, domain-agnostic, add-on, and data-driven strategy inspired by image stacking in image denoising. Termed ``semantic stacking,‘’ our method estimates a denoised semantic representation that complements the conventional segmentation loss during training. This method does not depend on domain-specific assumptions, making it broadly applicable across diverse image modalities, model architectures, and augmentation techniques. Through extensive experiments, we validate the superiority of our approach in improving segmentation performance under diverse conditions. Code is available at this https URL.
zh

[CV-15] F-Bench: Rethinking Human Preference Evaluation Metrics for Benchmarking Face Generation Customization and Restoration

【速读】: 该论文试图解决当前AI生成的人脸图像(AIGFs)在内容创建、定制和修复过程中常出现的失真、细节不真实和身份偏移等问题,这些问题导致生成的图像不符合人类偏好。为此,论文提出了一个全面的AI生成人脸图像质量评估框架,并引入了FaceQ数据库,该数据库包含12,255张由29种模型生成的图像,涵盖人脸生成、定制和修复三个任务,并附有32,742个来自180名标注者的平均意见分数(MOSs),评估维度包括质量、真实性、身份保真度和图文一致性。通过FaceQ数据库,论文建立了F-Bench基准,用于比较和评估不同模型在各种提示和评估维度下的表现,揭示现有图像质量评估(IQA)、人脸质量评估(FQA)、AI生成内容图像质量评估(AIGCIQA)和偏好评估指标在评估真实性、身份保真度和图文一致性方面的不足。解决方案的关键在于FaceQ数据库的构建和F-Bench基准的建立,为AI生成人脸图像的质量评估提供了系统化的工具和方法。

链接: https://arxiv.org/abs/2412.13155
作者: Lu Liu,Huiyu Duan,Qiang Hu,Liu Yang,Chunlei Cai,Tianxiao Ye,Huayu Liu,Xiaoyun Zhang,Guangtao Zhai
机构: Shanghai Jiao Tong University, Shanghai, China(上海交通大学,上海,中国); Bilibili Inc., China(哔哩哔哩公司,中国)
关键词: Artificial intelligence generative, exhibit remarkable capabilities, Artificial intelligence, generative models exhibit, models exhibit remarkable
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Artificial intelligence generative models exhibit remarkable capabilities in content creation, particularly in face image generation, customization, and restoration. However, current AI-generated faces (AIGFs) often fall short of human preferences due to unique distortions, unrealistic details, and unexpected identity shifts, underscoring the need for a comprehensive quality evaluation framework for AIGFs. To address this need, we introduce FaceQ, a large-scale, comprehensive database of AI-generated Face images with fine-grained Quality annotations reflecting human preferences. The FaceQ database comprises 12,255 images generated by 29 models across three tasks: (1) face generation, (2) face customization, and (3) face restoration. It includes 32,742 mean opinion scores (MOSs) from 180 annotators, assessed across multiple dimensions: quality, authenticity, identity (ID) fidelity, and text-image correspondence. Using the FaceQ database, we establish F-Bench, a benchmark for comparing and evaluating face generation, customization, and restoration models, highlighting strengths and weaknesses across various prompts and evaluation dimensions. Additionally, we assess the performance of existing image quality assessment (IQA), face quality assessment (FQA), AI-generated content image quality assessment (AIGCIQA), and preference evaluation metrics, manifesting that these standard metrics are relatively ineffective in evaluating authenticity, ID fidelity, and text-image correspondence. The FaceQ database will be publicly available upon publication.
zh

[CV-16] Continuous Patient Monitoring with AI: Real-Time Analysis of Video in Hospital Care Settings

【速读】: 该论文旨在解决医院环境中患者持续被动监测的问题,关键解决方案是引入一个由LookDeep Health开发的AI驱动平台。该平台利用先进的计算机视觉技术,通过视频分析实时获取患者行为和互动的洞察,并将推断结果安全存储在云端以供回顾性评估。其核心在于通过AI系统检测医院房间中的关键要素,如个体存在与角色、家具位置、运动幅度和边界穿越,从而实现对患者隔离、游荡或无人监督活动的自动化检测,这些都是跌倒风险和其他不良事件的关键指标。平台的高精度表现(如对象检测的宏F1分数为0.92,患者角色分类的F1分数为0.98)及其对“患者独自”指标的可靠趋势分析(平均逻辑回归准确率为0.82 ± 0.15),展示了其在提升患者安全和护理质量方面的潜力。

链接: https://arxiv.org/abs/2412.13152
作者: Paolo Gabriel,Peter Rehani,Tyler Troy,Tiffany Wyatt,Michael Choma,Narinder Singh
机构: 未知
关键词: LookDeep Health, developed by LookDeep, study introduces, passive patient monitoring, patient
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: 21 pages, 9 figures, 3 tables, submitted to Frontiers in Imaging Imaging Applications (Research Topic) Deep Learning for Medical Imaging Applications for publication

点击查看摘要

Abstract:This study introduces an AI-driven platform for continuous and passive patient monitoring in hospital settings, developed by LookDeep Health. Leveraging advanced computer vision, the platform provides real-time insights into patient behavior and interactions through video analysis, securely storing inference results in the cloud for retrospective evaluation. The dataset, compiled in collaboration with 11 hospital partners, encompasses over 300 high-risk fall patients and over 1,000 days of inference, enabling applications such as fall detection and safety monitoring for vulnerable patient populations. To foster innovation and reproducibility, an anonymized subset of this dataset is publicly available. The AI system detects key components in hospital rooms, including individual presence and role, furniture location, motion magnitude, and boundary crossings. Performance evaluation demonstrates strong accuracy in object detection (macro F1-score = 0.92) and patient-role classification (F1-score = 0.98), as well as reliable trend analysis for the “patient alone” metric (mean logistic regression accuracy = 0.82 \pm 0.15). These capabilities enable automated detection of patient isolation, wandering, or unsupervised movement-key indicators for fall risk and other adverse events. This work establishes benchmarks for validating AI-driven patient monitoring systems, highlighting the platform’s potential to enhance patient safety and care by providing continuous, data-driven insights into patient behavior and interactions.
zh

[CV-17] Label Errors in the Tobacco3482 Dataset WACV

【速读】: 该论文试图解决Tobacco3482数据集作为文档分类基准时存在的广泛本体问题,特别是数据集中大量标注标签错误的问题。解决方案的关键在于通过建立数据标签指南,识别出11.7%的数据样本存在不当标注,需要修正或标记为未知标签,并且16.7%的样本具有多个有效标签。此外,论文分析了表现优异的模型在这些标签问题上的错误,发现35%的模型错误可直接归因于这些标签问题,从而强调了使用噪声标注数据集作为基准的内在问题。

链接: https://arxiv.org/abs/2412.13140
作者: Gordon Lim,Stefan Larson,Kevin Leach
机构: University of Michigan(密歇根大学); Vanderbilt University(范德堡大学)
关键词: document classification benchmark, widely used document, document classification, classification benchmark dataset, dataset
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: WACV VisionDocs Workshop 2025

点击查看摘要

Abstract:Tobacco3482 is a widely used document classification benchmark dataset. However, our manual inspection of the entire dataset uncovers widespread ontological issues, especially large amounts of annotation label problems in the dataset. We establish data label guidelines and find that 11.7% of the dataset is improperly annotated and should either have an unknown label or a corrected label, and 16.7% of samples in the dataset have multiple valid labels. We then analyze the mistakes of a top-performing model and find that 35% of the model’s mistakes can be directly attributed to these label issues, highlighting the inherent problems with using a noisily labeled dataset as a benchmark. Supplementary material, including dataset annotations and code, is available at this https URL.
zh

[CV-18] Motion-2-to-3: Leveraging 2D Motion Data to Boost 3D Motion Generation

【速读】: 该论文试图解决现有文本驱动人体运动合成方法依赖于昂贵的3D动作捕捉数据,导致数据获取成本高、运动多样性和范围受限的问题。解决方案的关键在于利用从2D视频中提取的人体运动数据作为替代数据源,并通过引入一种新颖的框架来实现这一目标。该框架通过解耦局部关节运动与全局运动,使得能够从2D数据中高效学习局部运动先验。具体步骤包括:首先在大量文本-运动对数据集上训练一个单视图2D局部运动生成器,然后通过3D数据微调该生成器,使其转变为多视图生成器,能够预测视图一致的局部关节运动和根节点动力学。实验结果表明,该方法有效利用了2D数据,支持逼真的3D人体运动生成,并扩展了支持的运动类型范围。

链接: https://arxiv.org/abs/2412.13111
作者: Huaijin Pi,Ruoxi Guo,Zehong Shen,Qing Shuai,Zechen Hu,Zhumei Wang,Yajiao Dong,Ruizhen Hu,Taku Komura,Sida Peng,Xiaowei Zhou
机构: Zhejiang University(浙江大学); The University of Hong Kong(香港大学); Deep Glint(深鉴); Shenzhen University(深圳大学)
关键词: computer game development, capturing significant attention, effortlessly generate intricate, virtual reality experiences, abstract text cues
类目: Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR)
备注: Project page: this https URL

点击查看摘要

Abstract:Text-driven human motion synthesis is capturing significant attention for its ability to effortlessly generate intricate movements from abstract text cues, showcasing its potential for revolutionizing motion design not only in film narratives but also in virtual reality experiences and computer game development. Existing methods often rely on 3D motion capture data, which require special setups resulting in higher costs for data acquisition, ultimately limiting the diversity and scope of human motion. In contrast, 2D human videos offer a vast and accessible source of motion data, covering a wider range of styles and activities. In this paper, we explore leveraging 2D human motion extracted from videos as an alternative data source to improve text-driven 3D motion generation. Our approach introduces a novel framework that disentangles local joint motion from global movements, enabling efficient learning of local motion priors from 2D data. We first train a single-view 2D local motion generator on a large dataset of text-motion pairs. To enhance this model to synthesize 3D motion, we fine-tune the generator with 3D data, transforming it into a multi-view generator that predicts view-consistent local joint motion and root dynamics. Experiments on the HumanML3D dataset and novel text prompts demonstrate that our method efficiently utilizes 2D data, supporting realistic 3D human motion generation and broadening the range of motion types it supports. Our code will be made publicly available at this https URL.
zh

[CV-19] Accuracy Limits as a Barrier to Biometric System Security

【速读】: 该论文试图解决生物识别系统在面对非定向攻击(untargeted attacks)时的安全性和准确性问题。解决方案的关键在于通过分析系统的错误匹配率(False Match Rate, FMR)来评估系统的安全性,并确定在不同数据库规模下所需的临界FMR值和临界用户数量。具体来说,论文首先探讨了非定向攻击中攻击者试图冒充数据库中任意用户的情况,计算了攻击成功所需的尝试次数,并推导出维持一定安全水平所需的临界用户数量和FMR值。其次,论文重新审视了生物识别生日问题(biometric birthday problem),评估了数据库中用户之间可能发生的碰撞概率,并基于此推导出近似和精确的临界用户数量及FMR值,以限制碰撞发生的可能性。这些阈值为设计能够抵御冒充和碰撞风险的大规模生物识别数据库提供了重要参考。研究表明,当前的生物识别系统在小型数据库中也无法提供足够的安全性,且在处理生物识别生日问题时面临显著挑战,尤其是在数据库规模扩大时。

链接: https://arxiv.org/abs/2412.13099
作者: Axel Durbet,Paul-Marie Grollemund,Pascal Lafourcade,Kevin Thiry-Atighehchi
机构: UCA, LIMOS (UMR 6158 CNRS)(UCA, LIMOS (UMR 6158 CNRS)); UCA, LMBP (UMR 6620 CNRS)(UCA, LMBP (UMR 6620 CNRS))
关键词: Match Rate FMR, verification and identification, False Match Rate, identity verification, claimed identity
类目: Cryptography and Security (cs.CR); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Biometric systems are widely used for identity verification and identification, including authentication (i.e., one-to-one matching to verify a claimed identity) and identification (i.e., one-to-many matching to find a subject in a database). The matching process relies on measuring similarities or dissimilarities between a fresh biometric template and enrolled templates. The False Match Rate FMR is a key metric for assessing the accuracy and reliability of such systems. This paper analyzes biometric systems based on their FMR, with two main contributions. First, we explore untargeted attacks, where an adversary aims to impersonate any user within a database. We determine the number of trials required for an attacker to successfully impersonate a user and derive the critical population size (i.e., the maximum number of users in the database) required to maintain a given level of security. Furthermore, we compute the critical FMR value needed to ensure resistance against untargeted attacks as the database size increases. Second, we revisit the biometric birthday problem to evaluate the approximate and exact probabilities that two users in a database collide (i.e., can impersonate each other). Based on this analysis, we derive both the approximate critical population size and the critical FMR value needed to bound the likelihood of such collisions occurring with a given probability. These thresholds offer insights for designing systems that mitigate the risk of impersonation and collisions, particularly in large-scale biometric databases. Our findings indicate that current biometric systems fail to deliver sufficient accuracy to achieve an adequate security level against untargeted attacks, even in small-scale databases. Moreover, state-of-the-art systems face significant challenges in addressing the biometric birthday problem, especially as database sizes grow.
zh

[CV-20] Incremental Online Learning of Randomized Neural Network with Forward Regularization

【速读】: 该论文旨在解决深度神经网络在线学习中的几个关键挑战,包括非增量更新、内存使用增加、回顾性重训练以及灾难性遗忘(catastrophic forgetting)。为应对这些问题,论文提出了一种新的随机神经网络(Randomized Neural Networks, Randomized NN)的增量在线学习(Incremental Online Learning, IOL)过程,该框架能够在受限的在线场景中持续提升随机神经网络的性能。解决方案的关键在于引入了两种正则化方法:IOL with ridge regularization (-R) 和 IOL with forward regularization (-F)。-R 通过逐步增量更新避免了回顾性重训练和灾难性遗忘,而 -F 则通过半监督学习增强了预见性学习能力,并在在线学习中表现出比 -R 更好的性能,尤其是在与离线全局专家的对比中。此外,论文还推导了在非平稳批流数据上的递归权重更新和可变学习率的算法,并通过理论分析和实验验证了 -F 在加速在线学习和优化累积遗憾边界方面的优势。

链接: https://arxiv.org/abs/2412.13096
作者: Junda Wang,Minghui Hu,Ning Li,Abdulaziz Al-Ali,Ponnuthurai Nagaratnam Suganthan
机构: 未知
关键词: neural networks suffers, deep neural networks, increasing memory usage, Randomized Neural Networks, hysteretic non-incremental updating
类目: Machine Learning (cs.LG); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Online learning of deep neural networks suffers from challenges such as hysteretic non-incremental updating, increasing memory usage, past retrospective retraining, and catastrophic forgetting. To alleviate these drawbacks and achieve progressive immediate decision-making, we propose a novel Incremental Online Learning (IOL) process of Randomized Neural Networks (Randomized NN), a framework facilitating continuous improvements to Randomized NN performance in restrictive online scenarios. Within the framework, we further introduce IOL with ridge regularization (-R) and IOL with forward regularization (-F). -R generates stepwise incremental updates without retrospective retraining and avoids catastrophic forgetting. Moreover, we substituted -R with -F as it enhanced precognition learning ability using semi-supervision and realized better online regrets to offline global experts compared to -R during IOL. The algorithms of IOL for Randomized NN with -R/-F on non-stationary batch stream were derived respectively, featuring recursive weight updates and variable learning rates. Additionally, we conducted a detailed analysis and theoretically derived relative cumulative regret bounds of the Randomized NN learners with -R/-F in IOL under adversarial assumptions using a novel methodology and presented several corollaries, from which we observed the superiority on online learning acceleration and regret bounds of employing -F in IOL. Finally, our proposed methods were rigorously examined across regression and classification tasks on diverse datasets, which distinctly validated the efficacy of IOL frameworks of Randomized NN and the advantages of forward regularization.
zh

[CV-21] Prompt Augmentation for Self-supervised Text-guided Image Manipulation

【速读】: 该论文试图解决文本引导图像编辑中的两个主要挑战:图像变换的连贯性和上下文保留。解决方案的关键在于引入提示增强 (prompt augmentation),通过将单一输入提示扩展为多个目标提示,增强文本上下文并实现局部图像编辑。具体而言,论文提出了两种新的损失函数:对比损失 (Contrastive Loss)软对比损失 (Soft Contrastive Loss),前者通过分离编辑区域和拉近保留区域来驱动有效的图像编辑,后者则通过引入相似性概念进一步优化编辑效果。这些损失函数被集成到扩散模型中,显著提升了图像编辑的质量,并在公共数据集和生成图像上展示了优于或与现有最先进方法相竞争的结果。

链接: https://arxiv.org/abs/2412.13081
作者: Rumeysa Bodur,Binod Bhattarai,Tae-Kyun Kim
机构: Imperial College London, UK(英国帝国理工学院); University of Aberdeen, UK(英国阿伯丁大学); KAIST, South Korea(韩国科学技术院)
关键词: Text-guided image editing, editing finds applications, Text-guided image, finds applications, creative and practical
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Text-guided image editing finds applications in various creative and practical fields. While recent studies in image generation have advanced the field, they often struggle with the dual challenges of coherent image transformation and context preservation. In response, our work introduces prompt augmentation, a method amplifying a single input prompt into several target prompts, strengthening textual context and enabling localised image editing. Specifically, we use the augmented prompts to delineate the intended manipulation area. We propose a Contrastive Loss tailored to driving effective image editing by displacing edited areas and drawing preserved regions closer. Acknowledging the continuous nature of image manipulations, we further refine our approach by incorporating the similarity concept, creating a Soft Contrastive Loss. The new losses are incorporated to the diffusion model, demonstrating improved or competitive image editing results on public datasets and generated images over state-of-the-art approaches.
zh

[CV-22] Identifying Bias in Deep Neural Networks Using Image Transforms

【速读】: 该论文试图解决卷积神经网络(CNN)在图像分类中存在的“黑箱”问题,即难以识别和评估数据集中的隐性偏差(hidden biases),这些偏差会影响神经网络性能评估的准确性。解决方案的关键在于提出一种无需从原始图像中裁剪背景信息的方法来识别数据集偏差。具体而言,论文通过应用多种图像变换(如傅里叶变换、小波变换、中值滤波及其组合)来恢复CNN用于分类的背景偏差信息。这些变换能够以不同于影响上下文视觉信息的方式影响系统背景偏差,从而区分上下文信息和偏差,并检测背景偏差的存在,即使在没有空白背景的情况下也能实现。

链接: https://arxiv.org/abs/2412.13079
作者: Sai Teja Erukude,Akhil Joshi,Lior Shamir
机构: Kansas State University (堪萨斯州立大学)
关键词: past two decades, identify dataset bias, commonly used computational, computational tool, bias
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: Computers, published

点击查看摘要

Abstract:CNNs have become one of the most commonly used computational tool in the past two decades. One of the primary downsides of CNNs is that they work as a ``black box", where the user cannot necessarily know how the image data are analyzed, and therefore needs to rely on empirical evaluation to test the efficacy of a trained CNN. This can lead to hidden biases that affect the performance evaluation of neural networks, but are difficult to identify. Here we discuss examples of such hidden biases in common and widely used benchmark datasets, and propose techniques for identifying dataset biases that can affect the standard performance evaluation metrics. One effective approach to identify dataset bias is to perform image classification by using merely blank background parts of the original images. However, in some situations a blank background in the images is not available, making it more difficult to separate foreground or contextual information from the bias. To overcome this, we propose a method to identify dataset bias without the need to crop background information from the images. That method is based on applying several image transforms to the original images, including Fourier transform, wavelet transforms, median filter, and their combinations. These transforms were applied to recover background bias information that CNNs use to classify images. This transformations affect the contextual visual information in a different manner than it affects the systemic background bias. Therefore, the method can distinguish between contextual information and the bias, and alert on the presence of background bias even without the need to separate sub-images parts from the blank background of the original images. Code used in the experiments is publicly available.
zh

[CV-23] VidTok: A Versatile and Open-Source Video Tokenizer MICRO

【速读】: 该论文旨在解决视频内容编码为紧凑潜在标记(latent tokens)时面临的冗余问题,并满足视频研究领域对高性能、开源视频标记器(video tokenizer)的需求。解决方案的关键在于引入VidTok,这是一种多功能视频标记器,通过以下几个关键技术实现性能提升:1) 采用卷积层和上/下采样模块的模型架构;2) 针对传统向量量化(Vector Quantization, VQ)中常见的训练不稳定性和码本崩溃问题,引入有限标量量化(Finite Scalar Quantization, FSQ);3) 改进训练策略,包括两阶段训练过程和降低帧率的使用。这些创新使得VidTok在多个评价指标(如PSNR、SSIM、LPIPS和FVD)上表现出优于现有方法的性能。

链接: https://arxiv.org/abs/2412.13061
作者: Anni Tang,Tianyu He,Junliang Guo,Xinle Cheng,Li Song,Jiang Bian
机构: Microsoft Research(微软研究院); Shanghai Jiao Tong University(上海交通大学); Peking University(北京大学)
关键词: Encoding video content, compact latent tokens, Encoding video, generation and understanding, pixel-level representations
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: Code Models: this https URL

点击查看摘要

Abstract:Encoding video content into compact latent tokens has become a fundamental step in video generation and understanding, driven by the need to address the inherent redundancy in pixel-level representations. Consequently, there is a growing demand for high-performance, open-source video tokenizers as video-centric research gains prominence. We introduce VidTok, a versatile video tokenizer that delivers state-of-the-art performance in both continuous and discrete tokenizations. VidTok incorporates several key advancements over existing approaches: 1) model architecture such as convolutional layers and up/downsampling modules; 2) to address the training instability and codebook collapse commonly associated with conventional Vector Quantization (VQ), we integrate Finite Scalar Quantization (FSQ) into discrete video tokenization; 3) improved training strategies, including a two-stage training process and the use of reduced frame rates. By integrating these advancements, VidTok achieves substantial improvements over existing methods, demonstrating superior performance across multiple metrics, including PSNR, SSIM, LPIPS, and FVD, under standardized evaluation settings.
zh

[CV-24] CondiMen: Conditional Multi-Person Mesh Recovery

【速读】: 该论文试图解决多人物人体网格恢复(Multi-person human mesh recovery)问题,即在给定图像中检测所有个体并预测每个个体的身体形状、姿态和三维位置。解决方案的关键在于提出了CondiMen方法,该方法通过贝叶斯网络输出一个联合参数分布,涵盖可能的姿态、身体形状、内在参数和相机距离。这种方法的优势在于:1) 概率分布能够处理任务中的固有歧义,如人体尺寸与相机距离之间的不确定性;2) 输出分布可以结合额外信息(如已知的相机参数或身体形状先验)来提升预测精度;3) 能够高效提取最可能的预测,适用于实时应用。实验表明,CondiMen在性能上与现有技术相当或更优,能够捕捉姿态估计中的不确定性和相关性,并能在测试时利用多视角一致性或身体形状先验等额外信息。

链接: https://arxiv.org/abs/2412.13058
作者: Brégier Romain,Baradel Fabien,Lucas Thomas,Galaaoui Salma,Armando Matthieu,Weinzaepfel Philippe,Rogez Grégory
机构: 未知
关键词: Multi-person human mesh, human mesh recovery, Multi-person human, mesh recovery, consists in detecting
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Multi-person human mesh recovery (HMR) consists in detecting all individuals in a given input image, and predicting the body shape, pose, and 3D location for each detected person. The dominant approaches to this task rely on neural networks trained to output a single prediction for each detected individual. In contrast, we propose CondiMen, a method that outputs a joint parametric distribution over likely poses, body shapes, intrinsics and distances to the camera, using a Bayesian network. This approach offers several advantages. First, a probability distribution can handle some inherent ambiguities of this task – such as the uncertainty between a person’s size and their distance to the camera, or simply the loss of information when projecting 3D data onto the 2D image plane. Second, the output distribution can be combined with additional information to produce better predictions, by using e.g. known camera or body shape parameters, or by exploiting multi-view observations. Third, one can efficiently extract the most likely predictions from the output distribution, making our proposed approach suitable for real-time applications. Empirically we find that our model i) achieves performance on par with or better than the state-of-the-art, ii) captures uncertainties and correlations inherent in pose estimation and iii) can exploit additional information at test time, such as multi-view consistency or body shape priors. CondiMen spices up the modeling of ambiguity, using just the right ingredients on hand.
zh

[CV-25] EOGS: Gaussian Splatting for Earth Observation

【速读】: 该论文试图解决在遥感领域中,传统神经辐射场 (NeRF) 方法训练和渲染时间过长的问题。解决方案的关键在于将高斯光栅化 (Gaussian splatting) 框架应用于遥感数据处理,通过引入稀疏性 (sparsity)、视图一致性 (view consistency) 和透明度正则化 (opacity regularization) 等新组件,并结合EO-NeRF中的辐射校正 (radiometric correction) 和阴影建模 (shadow modeling) 技术,从而在保持高效性的同时实现了最先进的性能。

链接: https://arxiv.org/abs/2412.13047
作者: Luca Savant Aira,Gabriele Facciolo,Thibaud Ehret
机构: Politecnico di Torino(都灵理工大学); Universite Paris-Saclay, CNRS, ENS Paris-Saclay, Centre Borelli(巴黎-萨克雷大学, 法国国家科学研究中心, 巴黎高等师范学院, 博雷利中心); AMIAD, Pôle Recherche(AMIAD, 研究部门)
关键词: standard Gaussian splatting, demonstrating impressive, Gaussian splatting, Gaussian splatting framework, alternative to NeRF
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Recently, Gaussian splatting has emerged as a strong alternative to NeRF, demonstrating impressive 3D modeling capabilities while requiring only a fraction of the training and rendering time. In this paper, we show how the standard Gaussian splatting framework can be adapted for remote sensing, retaining its high efficiency. This enables us to achieve state-of-the-art performance in just a few minutes, compared to the day-long optimization required by the best-performing NeRF-based Earth observation methods. The proposed framework incorporates remote-sensing improvements from EO-NeRF, such as radiometric correction and shadow modeling, while introducing novel components, including sparsity, view consistency, and opacity regularizations.
zh

[CV-26] A New Adversarial Perspective for LiDAR-based 3D Object Detection AAAI2025

【速读】: 该论文试图解决自动驾驶车辆(AVs)在复杂环境中感知和决策的安全性与可靠性问题。解决方案的关键在于引入一个名为ROLiD的真实世界数据集,该数据集包含水雾和烟雾的LiDAR扫描点云,并通过提出一种基于运动和内容分解生成对抗网络(PCS-GAN)的点云序列生成方法,模拟随机物体的分布。利用Range Image实现的模拟LiDAR扫描特性,研究了在目标车辆不同位置引入随机物体扰动的影响。实验结果表明,基于随机物体的对抗扰动能够有效欺骗车辆检测,并降低3D目标检测模型的识别率。

链接: https://arxiv.org/abs/2412.13017
作者: Shijun Zheng,Weiquan Liu,Yu Guo,Yu Zang,Siqi Shen,Cheng Wang
机构: Shijun Zheng1,2; Yu Guo1,2; Yu Zang1,2; Siqi Shen1,2; Cheng Wang1,2
关键词: driving scenarios, Autonomous vehicles, perception and decision-making, decision-making in driving, Autonomous
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 11 pages, 7 figures, AAAI2025

点击查看摘要

Abstract:Autonomous vehicles (AVs) rely on LiDAR sensors for environmental perception and decision-making in driving scenarios. However, ensuring the safety and reliability of AVs in complex environments remains a pressing challenge. To address this issue, we introduce a real-world dataset (ROLiD) comprising LiDAR-scanned point clouds of two random objects: water mist and smoke. In this paper, we introduce a novel adversarial perspective by proposing an attack framework that utilizes water mist and smoke to simulate environmental interference. Specifically, we propose a point cloud sequence generation method using a motion and content decomposition generative adversarial network named PCS-GAN to simulate the distribution of random objects. Furthermore, leveraging the simulated LiDAR scanning characteristics implemented with Range Image, we examine the effects of introducing random object perturbations at various positions on the target vehicle. Extensive experiments demonstrate that adversarial perturbations based on random objects effectively deceive vehicle detection and reduce the recognition rate of 3D object detection models.
zh

[CV-27] Measurement of Medial Elbow Joint Space using Landmark Detection

【速读】: 该论文试图解决肘关节内侧超声图像中尺侧副韧带(Ulnar Collateral Ligament, UCL)损伤的早期识别问题,特别是通过自动化测量肘关节间隙来评估肘关节的尺侧不稳定。解决方案的关键在于引入了一个新的超声肘关节内侧数据集,该数据集包含4,201张来自22名受试者的肘关节内侧超声图像,并由作者在三位骨科医生的监督下进行了精确的解剖标志标注。论文评估了几种基于地标检测的方法(如ViTPose、HRNet、PCT、YOLOv8和U-Net)在测量关节间隙中的表现,并提出了使用形状子空间(Shape Subspace, SS)进行基于热图的地标检测的精炼。实验结果表明,HRNet在关节间隙测量的平均欧几里得距离误差为0.116 mm,而SS精炼进一步将HRNet和ViTPose的平均绝对误差分别降低了0.010 mm和0.103 mm。此外,论文还展示了基于检测到的地标的肱骨和尺骨的点分割。该数据集将在论文被接受后公开。

链接: https://arxiv.org/abs/2412.13010
作者: Shizuka Akahori,Shotaro Teruya,Pragyan Shrestha,Yuichi Yoshii,Ryuhei Michinobu,Satoshi Iizuka,Itaru Kitahara
机构: IIT, University of Tsukuba (筑波大学); Tsukuba Seikei Junior College (筑波成蹊短期大学)
关键词: Ulnar Collateral Ligament, Collateral Ligament, Ulnar Collateral, diagnose Ulnar Collateral, early identification
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Ultrasound imaging of the medial elbow is crucial for the early identification of Ulnar Collateral Ligament (UCL) injuries. Specifically, measuring the elbow joint space in ultrasound images is used to assess the valgus instability of elbow. To automate this measurement, a precisely annotated dataset is necessary; however, no publicly available dataset has been proposed thus far. This study introduces a novel ultrasound medial elbow dataset for measuring joint space to diagnose Ulnar Collateral Ligament (UCL) injuries. The dataset comprises 4,201 medial elbow ultrasound images from 22 subjects, with landmark annotations on the humerus and ulna. The annotations are made precisely by the authors under the supervision of three orthopedic surgeons. We evaluated joint space measurement methods using our proposed dataset with several landmark detection approaches, including ViTPose, HRNet, PCT, YOLOv8, and U-Net. In addition, we propose using Shape Subspace (SS) for landmark refinement in heatmap-based landmark detection. The results show that the mean Euclidean distance error of joint space is 0.116 mm when using HRNet. Furthermore, the SS landmark refinement improves the mean absolute error of landmark positions by 0.010 mm with HRNet and by 0.103 mm with ViTPose on average. These highlight the potential for high-precision, real-time diagnosis of UCL injuries and associated risks, which could be leveraged in large-scale screening. Lastly, we demonstrate point-based segmentation of the humerus and ulna using the detected landmarks as input. The dataset will be made publicly available upon acceptance of this paper at: this https URL.
zh

[CV-28] What is YOLOv6? A Deep Insight into the Object Detection Model

【速读】: 该论文旨在深入探讨YOLOv6目标检测模型的设计框架、优化技术及其检测能力。解决方案的关键在于其核心组件:EfficientRep Backbone(用于强大的特征提取)和Rep-PAN Neck(用于无缝的特征聚合),这两者共同确保了YOLOv6在保持高推理速度的同时实现高精度的目标检测。通过在COCO数据集上的评估,YOLOv6在不同模型规模下(如YOLOv6-N、YOLOv6-S、YOLOv6-M和YOLOv6-L)均表现出优于同类模型的性能,尤其是在精度和速度的平衡上。YOLOv6-L6通过升级的Backbone和Neck结构,进一步实现了实时检测中的前沿精度。

链接: https://arxiv.org/abs/2412.13006
作者: Athulya Sundaresan Geetha
机构: Huddersfield University(哈德斯菲尔德大学)
关键词: optimization techniques, design framework, work explores, object detection model, high-performance object detection
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:This work explores the YOLOv6 object detection model in depth, concentrating on its design framework, optimization techniques, and detection capabilities. YOLOv6’s core elements consist of the EfficientRep Backbone for robust feature extraction and the Rep-PAN Neck for seamless feature aggregation, ensuring high-performance object detection. Evaluated on the COCO dataset, YOLOv6-N achieves 37.5% AP at 1187 FPS on an NVIDIA Tesla T4 GPU. YOLOv6-S reaches 45.0% AP at 484 FPS, outperforming models like PPYOLOE-S, YOLOv5-S, YOLOX-S, and YOLOv8-S in the same class. Moreover, YOLOv6-M and YOLOv6-L also show better accuracy (50.0% and 52.8%) while maintaining comparable inference speeds to other detectors. With an upgraded backbone and neck structure, YOLOv6-L6 delivers cutting-edge accuracy in real-time.
zh

[CV-29] Future Aspects in Human Action Recognition: Exploring Emerging Techniques and Ethical Influences ICRA

【速读】: 该论文试图解决基于视觉的人类动作识别中的时间分析复杂性和数据集限制问题。解决方案的关键在于利用下一代硬件传感器提供的包含连续图像间过渡信息的视觉数据,以及通过强化学习技术减少对大规模数据集的依赖,同时生成新的合成视频数据以解决现有视频数据集的不足。此外,论文还强调了在研究中考虑人类因素带来的伦理问题。

链接: https://arxiv.org/abs/2412.12990
作者: Antonios Gasteratos,Stavros N. Moutsis,Konstantinos A. Tsintotas,Yiannis Aloimonos
机构: Democritus University of Thrace(色雷斯德谟克利特大学); University of Maryland(马里兰大学)
关键词: human-robot interaction frameworks, Visual-based human action, medical assistive technologies, human action recognition, surveillance systems
类目: Computer Vision and Pattern Recognition (cs.CV); Emerging Technologies (cs.ET); Robotics (cs.RO)
备注: 2 pages, 1 figure, 40th Anniversary of the IEEE Conference on Robotics and Automation (ICRA@40), Rotterdam, Netherlands | September 23-26, 2024

点击查看摘要

Abstract:Visual-based human action recognition can be found in various application fields, e.g., surveillance systems, sports analytics, medical assistive technologies, or human-robot interaction frameworks, and it concerns the identification and classification of individuals’ activities within a video. Since actions typically occur over a sequence of consecutive images, it is particularly challenging due to the inclusion of temporal analysis, which introduces an extra layer of complexity. However, although multiple approaches try to handle temporal analysis, there are still difficulties because of their computational cost and lack of adaptability. Therefore, different types of vision data, containing transition information between consecutive images, provided by next-generation hardware sensors will guide the robotics community in tackling the problem of human action recognition. On the other hand, while there is a plethora of still-image datasets, that researchers can adopt to train new artificial intelligence models, videos representing human activities are of limited capabilities, e.g., small and unbalanced datasets or selected without control from multiple sources. To this end, generating new and realistic synthetic videos is possible since labeling is performed throughout the data creation process, while reinforcement learning techniques can permit the avoidance of considerable dataset dependence. At the same time, human factors’ involvement raises ethical issues for the research community, as doubts and concerns about new technologies already exist.
zh

[CV-30] Attentive Eraser: Unleashing Diffusion Models Object Removal Potential via Self-Attention Redirection Guidance AAAI2025

【速读】: 该论文试图解决扩散模型在物体移除任务中生成随机伪影和无法在移除前景物体后适当重绘前景区域内容的问题。解决方案的关键在于提出了Attentive Eraser,这是一种无需微调的方法,通过Attention Activation and Suppression (ASS) 重新设计预训练扩散模型中的自注意力机制,根据给定的掩码优先处理背景信息,从而在反向生成过程中抑制前景物体的影响。此外,Self-Attention Redirection Guidance (SARG) 利用ASS重定向的自注意力来指导生成过程,确保在移除前景物体的同时生成合理且连贯的内容。实验表明,Attentive Eraser在多种预训练扩散模型中表现出稳定且有效的物体移除能力,甚至优于基于训练的方法,并且具有良好的可扩展性。

链接: https://arxiv.org/abs/2412.12974
作者: Wenhao Sun,Benlei Cui,Jingqun Tang,Xue-Mei Dong
机构: 未知
关键词: pre-trained diffusion models, diffusion models, Attentive Eraser, shining brightly, pre-trained diffusion
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Recently, diffusion models have emerged as promising newcomers in the field of generative models, shining brightly in image generation. However, when employed for object removal tasks, they still encounter issues such as generating random artifacts and the incapacity to repaint foreground object areas with appropriate content after removal. To tackle these problems, we propose Attentive Eraser, a tuning-free method to empower pre-trained diffusion models for stable and effective object removal. Firstly, in light of the observation that the self-attention maps influence the structure and shape details of the generated images, we propose Attention Activation and Suppression (ASS), which re-engineers the self-attention mechanism within the pre-trained diffusion models based on the given mask, thereby prioritizing the background over the foreground object during the reverse generation process. Moreover, we introduce Self-Attention Redirection Guidance (SARG), which utilizes the self-attention redirected by ASS to guide the generation process, effectively removing foreground objects within the mask while simultaneously generating content that is both plausible and coherent. Experiments demonstrate the stability and effectiveness of Attentive Eraser in object removal across a variety of pre-trained diffusion models, outperforming even training-based methods. Furthermore, Attentive Eraser can be implemented in various diffusion model architectures and checkpoints, enabling excellent scalability. Code is available at this https URL.
zh

[CV-31] Fruit Deformity Classification through Single-Input and Multi-Input Architectures based on CNN Models using Real and Synthetic Images

【速读】: 该论文旨在解决水果外部质量检测中变形程度的识别问题,关键解决方案在于采用基于卷积神经网络 (CNN) 的单输入和多输入架构,并结合真实与合成图像数据集。通过使用Segment Anything Model (SAM) 进行图像分割获取水果轮廓,论文提出了利用预训练模型结合合成图像改进单输入架构的评估结果。在多输入架构中,RGB图像和水果轮廓作为输入,评估了VGG16、MobileNetV2和CIDIS等模型,结果显示MobileNetV2在多输入架构下表现最佳,分别在苹果、芒果和草莓上达到了90%、94%和92%的准确率,表明该架构在水果变形分类中最为准确。

链接: https://arxiv.org/abs/2412.12966
作者: Tommy D. Beltran,Raul J. Villao,Luis E. Chuquimarca,Boris X. Vintimilla,Sergio A. Velastin
机构: 未知
关键词: convolutional neural network, present study focuses, Multi-Input architectures based, CNN models, Multi-Input architecture
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 15 pages, 9 figures, CIARP 2024

点击查看摘要

Abstract:The present study focuses on detecting the degree of deformity in fruits such as apples, mangoes, and strawberries during the process of inspecting their external quality, employing Single-Input and Multi-Input architectures based on convolutional neural network (CNN) models using sets of real and synthetic images. The datasets are segmented using the Segment Anything Model (SAM), which provides the silhouette of the fruits. Regarding the single-input architecture, the evaluation of the CNN models is performed only with real images, but a methodology is proposed to improve these results using a pre-trained model with synthetic images. In the Multi-Input architecture, branches with RGB images and fruit silhouettes are implemented as inputs for evaluating CNN models such as VGG16, MobileNetV2, and CIDIS. However, the results revealed that the Multi-Input architecture with the MobileNetV2 model was the most effective in identifying deformities in the fruits, achieving accuracies of 90%, 94%, and 92% for apples, mangoes, and strawberries, respectively. In conclusion, the Multi-Input architecture with the MobileNetV2 model is the most accurate for classifying levels of deformity in fruits.
zh

[CV-32] Synthetic Data Generation for Anomaly Detection on Table Grapes

【速读】: 该论文试图解决水果栽培中病害和虫害的早期检测问题,特别是在数据稀缺的情况下,如何自动生成高质量的异常样本以训练检测算法。解决方案的关键在于引入了一种自动化的合成异常样本生成方法,该方法仅需用户提供正常和异常样本的初始数据集。具体来说,论文提出了双Canny边缘检测 (Dual-Canny Edge Detection, DCED) 滤波器,用于强调异常样本中的复杂纹理特征,并通过Segment Anything Model提供的分割掩码,将异常样本无缝融合到正常样本中。这种方法不仅提高了异常分类器的准确性,还展示了其对其他水果类型的通用性。

链接: https://arxiv.org/abs/2412.12949
作者: Ionut Marian Motoi,Valerio Belli,Alberto Carpineto,Daniele Nardi,Thomas Alessandro Ciarfuglia
机构: 未知
关键词: maintaining yield quality, Early detection, plant health, critical for maintaining, maintaining yield
类目: Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO)
备注:

点击查看摘要

Abstract:Early detection of illnesses and pest infestations in fruit cultivation is critical for maintaining yield quality and plant health. Computer vision and robotics are increasingly employed for the automatic detection of such issues, particularly using data-driven solutions. However, the rarity of these problems makes acquiring and processing the necessary data to train such algorithms a significant obstacle. One solution to this scarcity is the generation of synthetic high-quality anomalous samples. While numerous methods exist for this task, most require highly trained individuals for setup. This work addresses the challenge of generating synthetic anomalies in an automatic fashion that requires only an initial collection of normal and anomalous samples from the user - a task that is straightforward for farmers. We demonstrate the approach in the context of table grape cultivation. Specifically, based on the observation that normal berries present relatively smooth surfaces, while defects result in more complex textures, we introduce a Dual-Canny Edge Detection (DCED) filter. This filter emphasizes the additional texture indicative of diseases, pest infestations, or other defects. Using segmentation masks provided by the Segment Anything Model, we then select and seamlessly blend anomalous berries onto normal ones. We show that the proposed dataset augmentation technique improves the accuracy of an anomaly classifier for table grapes and that the approach can be generalized to other fruit types. Subjects: Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO) ACMclasses: I.4.6; I.5.4; J.3 Cite as: arXiv:2412.12949 [cs.CV] (or arXiv:2412.12949v1 [cs.CV] for this version) https://doi.org/10.48550/arXiv.2412.12949 Focus to learn more arXiv-issued DOI via DataCite (pending registration)
zh

[CV-33] CoMT: A Novel Benchmark for Chain of Multi-modal Thought on Large Vision-Language Models AAAI2025

【速读】: 该论文试图解决当前多模态任务基准测试中存在的视觉操作缺失和表达模糊的问题。解决方案的关键在于引入了一种新的多模态思维链基准测试 (Chain of Multi-modal Thought, CoMT),该基准要求不仅输入是多模态的,输出也必须是多模态的,以模拟人类在推理过程中自然整合视觉操作的方式。CoMT 包含四个类别:视觉创建、视觉删除、视觉更新和视觉选择,旨在全面探索复杂的视觉操作和在实际场景中的简洁表达。通过评估多种大型视觉-语言模型 (Large Vision-Language Models, LVLMs) 和策略,论文揭示了当前方法的能力和局限性,并希望这一基准能够激发更多关于在推理过程中引入多模态生成的研究。

链接: https://arxiv.org/abs/2412.12932
作者: Zihui Cheng,Qiguang Chen,Jin Zhang,Hao Fei,Xiaocheng Feng,Wanxiang Che,Min Li,Libo Qin
机构: 1. Harbin Institute of Technology (哈尔滨工业大学); 2. Peng Cheng Laboratory (鹏城实验室); 3. Tsinghua University (清华大学); 4. ByteDance AI Lab (字节跳动人工智能实验室)
关键词: Large Vision-Language Models, recently demonstrated amazing, demonstrated amazing success, Large Vision-Language, Vision-Language Models
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: Accepted at AAAI 2025

点击查看摘要

Abstract:Large Vision-Language Models (LVLMs) have recently demonstrated amazing success in multi-modal tasks, including advancements in Multi-modal Chain-of-Thought (MCoT) reasoning. Despite these successes, current benchmarks still follow a traditional paradigm with multi-modal input and text-modal output, which leads to significant drawbacks such as missing visual operations and vague expressions. Motivated by this, we introduce a novel Chain of Multi-modal Thought (CoMT) benchmark to address these limitations. Different from the traditional MCoT benchmark, CoMT requires both multi-modal input and multi-modal reasoning output, aiming to mimic human-like reasoning that inherently integrates visual operation. Specifically, CoMT consists of four categories: (1) Visual Creation, (2) Visual Deletion, (3) Visual Update, and (4) Visual Selection to comprehensively explore complex visual operations and concise expression in real scenarios. We evaluate various LVLMs and strategies on CoMT, revealing some key insights into the capabilities and limitations of the current approaches. We hope that CoMT can inspire more research on introducing multi-modal generation into the reasoning process.
zh

[CV-34] Unsupervised Region-Based Image Editing of Denoising Diffusion Models

【速读】: 该论文试图解决在预训练扩散模型中,潜在空间语义属性识别的问题。当前方法通常依赖外部监督信息(如文本信息和分割掩码),而本文提出了一种无需进一步训练的解决方案。关键在于通过将目标语义区域的雅可比矩阵(Jacobian)投影到一个与非掩码区域正交的低维子空间,从而实现对局部掩码区域的精确语义发现和控制,无需依赖标注数据。该方法在多个数据集和扩散模型架构上进行了广泛实验,达到了最先进的性能,甚至在某些特定面部属性上超越了监督方法。

链接: https://arxiv.org/abs/2412.12912
作者: Zixiang Li,Yue Song,Renshuai Tao,Xiaohong Jia,Yao Zhao,Wei Wang
机构: 1. School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China(哈尔滨工业大学计算机科学与技术学院,哈尔滨,中国);
2. National Engineering Research Center for E-Learning, Central China Normal University, Wuhan, China(华中师范大学国家教育工程研究中心,武汉,中国);
3. School of Artificial Intelligence, Harbin Institute of Technology, Harbin, China(哈尔滨工业大学人工智能学院,哈尔滨,中国);
4. School of Computer Science and Technology, Shandong University, Jinan, China(山东大学计算机科学与技术学院,济南,中国);
5. School of Computer Science and Technology, Harbin Institute of Technology, Weihai, China(哈尔滨工业大学(威海)计算机科学与技术学院,威海,中国)
关键词: achieved remarkable success, space remains under-explored, latent space remains, latent space, remains under-explored
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Although diffusion models have achieved remarkable success in the field of image generation, their latent space remains under-explored. Current methods for identifying semantics within latent space often rely on external supervision, such as textual information and segmentation masks. In this paper, we propose a method to identify semantic attributes in the latent space of pre-trained diffusion models without any further training. By projecting the Jacobian of the targeted semantic region into a low-dimensional subspace which is orthogonal to the non-masked regions, our approach facilitates precise semantic discovery and control over local masked areas, eliminating the need for annotations. We conducted extensive experiments across multiple datasets and various architectures of diffusion models, achieving state-of-the-art performance. In particular, for some specific face attributes, the performance of our proposed method even surpasses that of supervised approaches, demonstrating its superior ability in editing local image properties.
zh

[CV-35] CATSplat: Context-Aware Transformer with Spatial Guidance for Generalizable 3D Gaussian Splatting from A Single-View Image

【速读】: 该论文试图解决单视图(monocular)3D场景重建的问题,即在仅有一张图像的情况下,如何高效且准确地重建3D场景。解决方案的关键在于引入CATSplat框架,该框架结合了视觉-语言模型(visual-language model)的文本指导和3D点特征的空间指导。具体来说,通过利用文本嵌入(text embeddings)中的场景特定上下文信息,并通过交叉注意力机制(cross-attention)将其融入到3D场景重建中,CATSplat能够超越仅依赖视觉线索的传统方法。此外,利用3D先验信息和图像特征,CATSplat能够在无需多视图技术的情况下,捕捉丰富的结构信息,从而预测3D高斯基元(3D Gaussian primitives),实现高质量的单视图3D场景重建和新视角合成。

链接: https://arxiv.org/abs/2412.12906
作者: Wonseok Roh,Hwanhee Jung,Jong Wook Kim,Seunggwan Lee,Innfarn Yoo,Andreas Lugmayr,Seunggeun Chi,Karthik Ramani,Sangpil Kim
机构: Korea University(韩国大学); Google Research(谷歌研究); Purdue University(普渡大学)
关键词: gained significant attention, Gaussian Splatting, feed-forward methods based, Splatting have gained, potential to reconstruct
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Recently, generalizable feed-forward methods based on 3D Gaussian Splatting have gained significant attention for their potential to reconstruct 3D scenes using finite resources. These approaches create a 3D radiance field, parameterized by per-pixel 3D Gaussian primitives, from just a few images in a single forward pass. However, unlike multi-view methods that benefit from cross-view correspondences, 3D scene reconstruction with a single-view image remains an underexplored area. In this work, we introduce CATSplat, a novel generalizable transformer-based framework designed to break through the inherent constraints in monocular settings. First, we propose leveraging textual guidance from a visual-language model to complement insufficient information from a single image. By incorporating scene-specific contextual details from text embeddings through cross-attention, we pave the way for context-aware 3D scene reconstruction beyond relying solely on visual cues. Moreover, we advocate utilizing spatial guidance from 3D point features toward comprehensive geometric understanding under single-view settings. With 3D priors, image features can capture rich structural insights for predicting 3D Gaussians without multi-view techniques. Extensive experiments on large-scale datasets demonstrate the state-of-the-art performance of CATSplat in single-view 3D scene reconstruction with high-quality novel view synthesis.
zh

[CV-36] DoPTA: Improving Document Layout Analysis using Patch-Text Alignment

【速读】: 该论文试图解决文档图像理解中视觉信息与文本信息的对齐问题,特别是在不依赖OCR(Optical Character Recognition)的情况下,如何有效利用文档中的文本信息来提升视觉任务的性能。解决方案的关键在于提出了一种新的图像-文本对齐技术,并设计了专门的文档编码器模型DoPTA。该模型通过结合辅助重建目标,在不需要OCR推理的情况下,显著提升了文档图像理解任务的性能,并在D4LA和FUNSD等具有挑战性的文档视觉分析基准上取得了新的最先进结果。

链接: https://arxiv.org/abs/2412.12902
作者: Nikitha SR,Tarun Ram Menta,Mausoom Sarkar
机构: Media and Data Science Research Lab, Adobe
关键词: brought a significant, significant improvement, document, document image understanding, visual
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:The advent of multimodal learning has brought a significant improvement in document AI. Documents are now treated as multimodal entities, incorporating both textual and visual information for downstream analysis. However, works in this space are often focused on the textual aspect, using the visual space as auxiliary information. While some works have explored pure vision based techniques for document image understanding, they require OCR identified text as input during inference, or do not align with text in their learning procedure. Therefore, we present a novel image-text alignment technique specially designed for leveraging the textual information in document images to improve performance on visual tasks. Our document encoder model DoPTA - trained with this technique demonstrates strong performance on a wide range of document image understanding tasks, without requiring OCR during inference. Combined with an auxiliary reconstruction objective, DoPTA consistently outperforms larger models, while using significantly lesser pre-training compute. DoPTA also sets new state-of-the art results on D4LA, and FUNSD, two challenging document visual analysis benchmarks.
zh

[CV-37] SAUGE: Taming SAM for Uncertainty-Aligned Multi-Granularity Edge Detection AAAI2025

【速读】: 该论文试图解决边缘检测中由于标注者偏好不同导致的边缘标签粒度不一致问题,即像素级标签的主观性。解决方案的关键在于利用Segment Anything Model (SAM) 提供的先验知识来建模边缘标签的不确定性。具体来说,论文提出通过回归SAM不同层的中间特征来对齐不确定性与粒度,从而在数据驱动的方式下充分探索多样化的不确定性。此外,通过轻量级模块逐步融合和适应SAM的中间特征,并使用线性混合方法生成具有不同粒度的伪标签,使得模型能够灵活地生成任意所需粒度的边缘,同时展现出跨数据集的强泛化能力。

链接: https://arxiv.org/abs/2412.12892
作者: Xing Liufu,Chaolei Tan,Xiaotong Lin,Yonggang Qi,Jinxuan Li,Jian-Fang Hu
机构: 未知
关键词: Edge labels, intermediate SAM features, SAM, preferences of annotators, labels
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: Accepted to AAAI 2025

点击查看摘要

Abstract:Edge labels are typically at various granularity levels owing to the varying preferences of annotators, thus handling the subjectivity of per-pixel labels has been a focal point for edge detection. Previous methods often employ a simple voting strategy to diminish such label uncertainty or impose a strong assumption of labels with a pre-defined distribution, e.g., Gaussian. In this work, we unveil that the segment anything model (SAM) provides strong prior knowledge to model the uncertainty in edge labels. Our key insight is that the intermediate SAM features inherently correspond to object edges at various granularities, which reflects different edge options due to uncertainty. Therefore, we attempt to align uncertainty with granularity by regressing intermediate SAM features from different layers to object edges at multi-granularity levels. In doing so, the model can fully and explicitly explore diverse ``uncertainties’’ in a data-driven fashion. Specifically, we inject a lightweight module (~ 1.5% additional parameters) into the frozen SAM to progressively fuse and adapt its intermediate features to estimate edges from coarse to fine. It is crucial to normalize the granularity level of human edge labels to match their innate uncertainty. For this, we simply perform linear blending to the real edge labels at hand to create pseudo labels with varying granularities. Consequently, our uncertainty-aligned edge detector can flexibly produce edges at any desired granularity (including an optimal one). Thanks to SAM, our model uniquely demonstrates strong generalizability for cross-dataset edge detection. Extensive experimental results on BSDS500, Muticue and NYUDv2 validate our model’s superiority.
zh

[CV-38] Suppressing Uncertainty in Gaze Estimation AAAI2024

【速读】: 该论文试图解决视线估计中的不确定性问题,主要体现在低质量图像(如遮挡、模糊、不一致的眼动或非人脸图像)和标注错误(由于标注点与实际视线点未对齐)两个方面。解决方案的关键在于提出了一种名为“视线估计中的不确定性抑制 (Suppressing Uncertainty in Gaze Estimation, SUGE)”的方法,通过引入一种新颖的三元组标签一致性度量来估计和减少不确定性。具体而言,该方法通过线性加权投影计算每个训练样本的“邻近标签”,以捕捉图像特征与其对应标签之间的相似性关系,并结合预测的伪标签和真实标签进行不确定性估计。通过建模这种三元组标签一致性,能够评估图像和标签的质量,并通过设计的样本加权和标签校正策略大幅减少不合格图像和错误标签的负面影响。

链接: https://arxiv.org/abs/2412.12890
作者: Shijing Wang,Yaping Huang
机构: 未知
关键词: inconsistent eye movements, actual gaze points, gaze estimation manifests, low-quality images caused, gaze estimation
类目: Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注: This paper has been accepted to AAAI 2024

点击查看摘要

Abstract:Uncertainty in gaze estimation manifests in two aspects: 1) low-quality images caused by occlusion, blurriness, inconsistent eye movements, or even non-face images; 2) incorrect labels resulting from the misalignment between the labeled and actual gaze points during the annotation process. Allowing these uncertainties to participate in training hinders the improvement of gaze estimation. To tackle these challenges, in this paper, we propose an effective solution, named Suppressing Uncertainty in Gaze Estimation (SUGE), which introduces a novel triplet-label consistency measurement to estimate and reduce the uncertainties. Specifically, for each training sample, we propose to estimate a novel ``neighboring label’’ calculated by a linearly weighted projection from the neighbors to capture the similarity relationship between image features and their corresponding labels, which can be incorporated with the predicted pseudo label and ground-truth label for uncertainty estimation. By modeling such triplet-label consistency, we can measure the qualities of both images and labels, and further largely reduce the negative effects of unqualified images and wrong labels through our designed sample weighting and label correction strategies. Experimental results on the gaze estimation benchmarks indicate that our proposed SUGE achieves state-of-the-art performance.
zh

[CV-39] ArtAug: Enhancing Text-to-Image Generation through Synthesis-Understanding Interaction

【速读】: 该论文试图解决如何在不增加额外计算成本的情况下,提升文本到图像生成模型的生成能力问题。解决方案的关键在于提出了一种名为ArtAug的新方法,通过模型间的交互来增强图像合成模型。具体来说,ArtAug利用图像理解模型中隐含的人类偏好,为图像合成模型提供细粒度的美学建议,如调整曝光、改变拍摄角度和添加氛围效果等。这些交互产生的增强效果通过一个额外的增强模块迭代地融合到合成模型中,使得合成模型能够直接生成具有美学吸引力的图像,而无需额外的计算开销。

链接: https://arxiv.org/abs/2412.12888
作者: Zhongjie Duan,Qianyi Zhao,Cen Chen,Daoyuan Chen,Wenmeng Zhou,Yaliang Li,Yingda Chen
机构: 未知
关键词: significantly advanced image, models, advanced image synthesis, image synthesis models, emergence of diffusion
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: 18 pages, 8 figures

点击查看摘要

Abstract:The emergence of diffusion models has significantly advanced image synthesis. The recent studies of model interaction and self-corrective reasoning approach in large language models offer new insights for enhancing text-to-image models. Inspired by these studies, we propose a novel method called ArtAug for enhancing text-to-image models in this paper. To the best of our knowledge, ArtAug is the first one that improves image synthesis models via model interactions with understanding models. In the interactions, we leverage human preferences implicitly learned by image understanding models to provide fine-grained suggestions for image synthesis models. The interactions can modify the image content to make it aesthetically pleasing, such as adjusting exposure, changing shooting angles, and adding atmospheric effects. The enhancements brought by the interaction are iteratively fused into the synthesis model itself through an additional enhancement module. This enables the synthesis model to directly produce aesthetically pleasing images without any extra computational cost. In the experiments, we train the ArtAug enhancement module on existing text-to-image models. Various evaluation metrics consistently demonstrate that ArtAug enhances the generative capabilities of text-to-image models without incurring additional computational costs. The source code and models will be released publicly.
zh

[CV-40] Learning Coarse-to-Fine Pruning of Graph Convolutional Networks for Skeleton-based Recognition

【速读】: 该论文试图解决轻量级网络设计中结构化剪枝(structured pruning)和非结构化剪枝(unstructured pruning)之间的权衡问题。结构化剪枝虽然能提高计算效率,但灵活性较差且精度较低;非结构化剪枝则相反,灵活性高且精度更好,但计算性能较低。论文提出了一种新颖的从粗到细(coarse-to-fine, CTF)方法,通过引入一种新的CTF参数化方法,将每个连接的掩码建模为涉及四个参数化的哈达玛积(Hadamard product),分别捕捉通道级、列级、行级和条目级的剪枝。该方法在粗粒度剪枝被禁用时启用细粒度剪枝,从而在保持高效网络的同时实现了更好的效果。实验结果表明,该方法在基于骨架的识别任务中显著优于不同基线和相关工作。

链接: https://arxiv.org/abs/2412.12887
作者: Hichem Sahbi
机构: Sorbonne University(索邦大学); CNRS(法国国家科学研究中心); LIP6(巴黎第六大学)
关键词: smallest magnitude, staple lightweight network, lightweight network design, Magnitude Pruning, staple lightweight
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Magnitude Pruning is a staple lightweight network design method which seeks to remove connections with the smallest magnitude. This process is either achieved in a structured or unstructured manner. While structured pruning allows reaching high efficiency, unstructured one is more flexible and leads to better accuracy, but this is achieved at the expense of low computational performance. In this paper, we devise a novel coarse-to-fine (CTF) method that gathers the advantages of structured and unstructured pruning while discarding their inconveniences to some extent. Our method relies on a novel CTF parametrization that models the mask of each connection as the Hadamard product involving four parametrizations which capture channel-wise, column-wise, row-wise and entry-wise pruning respectively. Hence, fine-grained pruning is enabled only when the coarse-grained one is disabled, and this leads to highly efficient networks while being effective. Extensive experiments conducted on the challenging task of skeleton-based recognition, using the standard SBU and FPHA datasets, show the clear advantage of our CTF approach against different baselines as well as the related work.
zh

[CV-41] MIVE: New Design and Benchmark for Multi-Instance Video Editing

【速读】: 该论文试图解决现有零样本视频编辑技术在多对象局部编辑中的挑战,包括编辑不忠实、编辑泄露以及缺乏合适的评估数据集和指标等问题。解决方案的关键在于提出了一个通用的基于掩码的多实例视频编辑框架 (MIVE),并引入了两个核心模块:(i) 解耦多实例采样 (Disentangled Multi-instance Sampling, DMS) 以防止编辑泄露,(ii) 实例中心概率重分配 (Instance-centric Probability Redistribution, IPR) 以确保精确的局部定位和忠实的编辑。此外,论文还提出了新的 MIVE 数据集和跨实例准确率 (Cross-Instance Accuracy, CIA) 评分,用于评估多实例视频编辑任务中的编辑泄露问题。

链接: https://arxiv.org/abs/2412.12877
作者: Samuel Teodoro,Agus Gunawan,Soo Ye Kim,Jihyong Oh,Munchurl Kim
机构: KAIST(韩国科学技术院); Adobe Research(奥多比研究); Chung-Ang University(中央大学)
关键词: simple text prompts, Recent AI-based video, editing, text prompts, simple text
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: The first two authors contributed equally to this work. The last two authors are co-corresponding authors. Please visit our project page at this https URL

点击查看摘要

Abstract:Recent AI-based video editing has enabled users to edit videos through simple text prompts, significantly simplifying the editing process. However, recent zero-shot video editing techniques primarily focus on global or single-object edits, which can lead to unintended changes in other parts of the video. When multiple objects require localized edits, existing methods face challenges, such as unfaithful editing, editing leakage, and lack of suitable evaluation datasets and metrics. To overcome these limitations, we propose a zero-shot \textbfM ulti- \textbfI nstance \textbfV ideo \textbfE diting framework, called MIVE. MIVE is a general-purpose mask-based framework, not dedicated to specific objects (e.g., people). MIVE introduces two key modules: (i) Disentangled Multi-instance Sampling (DMS) to prevent editing leakage and (ii) Instance-centric Probability Redistribution (IPR) to ensure precise localization and faithful editing. Additionally, we present our new MIVE Dataset featuring diverse video scenarios and introduce the Cross-Instance Accuracy (CIA) Score to evaluate editing leakage in multi-instance video editing tasks. Our extensive qualitative, quantitative, and user study evaluations demonstrate that MIVE significantly outperforms recent state-of-the-art methods in terms of editing faithfulness, accuracy, and leakage prevention, setting a new benchmark for multi-instance video editing. The project page is available at this https URL
zh

[CV-42] Dyn-HaMR: Recovering 4D Interacting Hand Motion from a Dynamic Camera

【速读】: 该论文试图解决从单目视频中重建4D全局手部运动的问题,特别是在动态或移动摄像机(如在第一人称视角场景中)捕捉的视频中。现有方法通常依赖于弱透视相机模型,导致在处理动态摄像机时难以恢复完整的3D全局轨迹,并产生噪声或错误的深度估计。解决方案的关键在于提出了一种多阶段、多目标优化的方法Dyn-HaMR,其中包括:(i) 同时定位与地图构建 (SLAM) 以稳健估计相对相机运动;(ii) 交互手部先验用于生成填充和细化交互动态,确保在自遮挡或遮挡情况下的合理恢复;(iii) 通过结合最先进的手部跟踪方法进行分层初始化。通过在野外和室内数据集上的广泛评估,该方法显著优于现有技术,为移动摄像机下的单目视频手部运动重建设定了新的基准。

链接: https://arxiv.org/abs/2412.12861
作者: Zhengdi Yu,Stefanos Zafeiriou,Tolga Birdal
机构: Imperial College London (帝国理工学院)
关键词: monocular videos recorded, monocular videos, monocular, hand, videos recorded
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Project page is available at this https URL

点击查看摘要

Abstract:We propose Dyn-HaMR, to the best of our knowledge, the first approach to reconstruct 4D global hand motion from monocular videos recorded by dynamic cameras in the wild. Reconstructing accurate 3D hand meshes from monocular videos is a crucial task for understanding human behaviour, with significant applications in augmented and virtual reality (AR/VR). However, existing methods for monocular hand reconstruction typically rely on a weak perspective camera model, which simulates hand motion within a limited camera frustum. As a result, these approaches struggle to recover the full 3D global trajectory and often produce noisy or incorrect depth estimations, particularly when the video is captured by dynamic or moving cameras, which is common in egocentric scenarios. Our Dyn-HaMR consists of a multi-stage, multi-objective optimization pipeline, that factors in (i) simultaneous localization and mapping (SLAM) to robustly estimate relative camera motion, (ii) an interacting-hand prior for generative infilling and to refine the interaction dynamics, ensuring plausible recovery under (self-)occlusions, and (iii) hierarchical initialization through a combination of state-of-the-art hand tracking methods. Through extensive evaluations on both in-the-wild and indoor datasets, we show that our approach significantly outperforms state-of-the-art methods in terms of 4D global mesh recovery. This establishes a new benchmark for hand motion reconstruction from monocular video with moving cameras. Our project page is at this https URL.
zh

[CV-43] Boosting Fine-Grained Visual Anomaly Detection with Coarse-Knowledge-Aware Adversarial Learning AAAI2025

【速读】: 该论文试图解决无监督视觉异常检测中,由于神经网络强大的建模和泛化能力,导致异常样本也能被良好重建,从而影响检测和定位精度的问题。解决方案的关键在于引入粗标签异常数据集,并开发了一种粗知识感知对抗学习方法,通过对齐重建特征与正常特征的分布,有效抑制自编码器对异常的重建能力,从而提高检测精度。此外,考虑到异常通常仅占据图像的很小区域,进一步提出了基于图像块的对抗学习策略,通过理论证明,即使没有图像块级别的异常信息,将异常图像中的任何图像块特征视为异常,也能实现重建图像块特征与正常特征的分布对齐。

链接: https://arxiv.org/abs/2412.12850
作者: Qingqing Fang,Qinliang Su,Wenxi Lv,Wenchao Xu,Jianxing Yu
机构: 未知
关键词: unsupervised visual anomaly, reconstruction error map, reconstruct normal samples, visual anomaly detection, unsupervised visual
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: The paper is accepted by AAAI 2025

点击查看摘要

Abstract:Many unsupervised visual anomaly detection methods train an auto-encoder to reconstruct normal samples and then leverage the reconstruction error map to detect and localize the anomalies. However, due to the powerful modeling and generalization ability of neural networks, some anomalies can also be well reconstructed, resulting in unsatisfactory detection and localization accuracy. In this paper, a small coarsely-labeled anomaly dataset is first collected. Then, a coarse-knowledge-aware adversarial learning method is developed to align the distribution of reconstructed features with that of normal features. The alignment can effectively suppress the auto-encoder’s reconstruction ability on anomalies and thus improve the detection accuracy. Considering that anomalies often only occupy very small areas in anomalous images, a patch-level adversarial learning strategy is further developed. Although no patch-level anomalous information is available, we rigorously prove that by simply viewing any patch features from anomalous images as anomalies, the proposed knowledge-aware method can also align the distribution of reconstructed patch features with the normal ones. Experimental results on four medical datasets and two industrial datasets demonstrate the effectiveness of our method in improving the detection and localization performance.
zh

[CV-44] HyperGS: Hyperspectral 3D Gaussian Splatting

【速读】: 该论文旨在解决高光谱新视角合成 (Hyperspectral Novel View Synthesis, HNVS) 问题,提出了一种基于潜在三维高斯样条 (latent 3D Gaussian Splatting, 3DGS) 技术的新框架 HyperGS。其关键解决方案包括:通过多视角三维高光谱数据集编码材料属性,实现空间和光谱的同时渲染;在学习的潜在空间中进行视角合成,采用像素级自适应密度函数和剪枝技术以提高训练稳定性和效率;并引入了首个HNVS基准,通过广泛的实际和模拟高光谱场景评估,展示了其在准确性和速度上的显著提升,相较于现有方法实现了14dB的精度改进。

链接: https://arxiv.org/abs/2412.12849
作者: Christopher Thirgood,Oscar Mendez,Erin Chao Ling,Jon Storey,Simon Hadfield
机构: CVSSP, University of Surrey; I3D Robotics, Kent, UK
关键词: Gaussian Splatting, View Synthesis, Gaussian, Splatting, perform view synthesis
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:We introduce HyperGS, a novel framework for Hyperspectral Novel View Synthesis (HNVS), based on a new latent 3D Gaussian Splatting (3DGS) technique. Our approach enables simultaneous spatial and spectral renderings by encoding material properties from multi-view 3D hyperspectral datasets. HyperGS reconstructs high-fidelity views from arbitrary perspectives with improved accuracy and speed, outperforming currently existing methods. To address the challenges of high-dimensional data, we perform view synthesis in a learned latent space, incorporating a pixel-wise adaptive density function and a pruning technique for increased training stability and efficiency. Additionally, we introduce the first HNVS benchmark, implementing a number of new baselines based on recent SOTA RGB-NVS techniques, alongside the small number of prior works on HNVS. We demonstrate HyperGS’s robustness through extensive evaluation of real and simulated hyperspectral scenes with a 14db accuracy improvement upon previously published models.
zh

[CV-45] Efficient Event-based Semantic Segmentation with Spike-driven Lightweight Transformer-based Networks ICRA2025

【速读】: 该论文试图解决基于事件相机的语义分割在资源受限的边缘/移动平台上应用时,传统人工神经网络 (ANN) 方法计算需求高、能耗大、依赖图像帧等问题。解决方案的关键在于提出了SLTNet,一种基于脉冲神经网络 (Spiking Neural Network, SNN) 的轻量级Transformer网络。SLTNet通过高效的脉冲驱动卷积块 (Spike-driven Convolution Blocks, SCBs) 提取丰富的语义特征并减少模型参数,同时引入带有二值掩码操作的脉冲驱动Transformer块 (Spike-driven Transformer Blocks, STBs) 增强长距离上下文特征交互。最终,SLTNet采用高效的单分支架构,在保持SNN低能耗特性的同时,显著提升了语义分割性能,并在实验中展示了优于现有SOTA方法的性能和能效。

链接: https://arxiv.org/abs/2412.12843
作者: Xiaxin Zhu,Fangming Guo,Xianlei Long,Qingyi Gu,Chao Chen,Fuqiang Gu
机构: Chongqing University(重庆大学); Institute of Automation, Chinese Academy of Sciences(中国科学院自动化研究所)
关键词: high dynamic range, low power cost, Event-based semantic segmentation, event cameras, dynamic range
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: Submitted to IEEE ICRA 2025

点击查看摘要

Abstract:Event-based semantic segmentation has great potential in autonomous driving and robotics due to the advantages of event cameras, such as high dynamic range, low latency, and low power cost. Unfortunately, current artificial neural network (ANN)-based segmentation methods suffer from high computational demands, the requirements for image frames, and massive energy consumption, limiting their efficiency and application on resource-constrained edge/mobile platforms. To address these problems, we introduce SLTNet, a spike-driven lightweight transformer-based network designed for event-based semantic segmentation. Specifically, SLTNet is built on efficient spike-driven convolution blocks (SCBs) to extract rich semantic features while reducing the model’s parameters. Then, to enhance the long-range contextural feature interaction, we propose novel spike-driven transformer blocks (STBs) with binary mask operations. Based on these basic blocks, SLTNet employs a high-efficiency single-branch architecture while maintaining the low energy consumption of the Spiking Neural Network (SNN). Finally, extensive experiments on DDD17 and DSEC-Semantic datasets demonstrate that SLTNet outperforms state-of-the-art (SOTA) SNN-based methods by at least 7.30% and 3.30% mIoU, respectively, with extremely 5.48x lower energy consumption and 1.14x faster inference speed.
zh

[CV-46] FocusChat: Text-guided Long Video Understanding via Spatiotemporal Information Filtering

【速读】: 该论文试图解决多模态大语言模型(LLM)在处理长且未剪辑的视频时,由于缺乏用户意图的引导而导致的冗余计算和视觉噪声问题。解决方案的关键是提出了FocusChat模型,该模型通过文本引导的方式强调与用户查询相关的视觉信息。具体来说,FocusChat包含一个语义提取模块,由视觉语义分支和文本语义分支组成,分别提取图像和文本的语义信息,并通过空间-时间过滤模块(Spatial-Temporal Filtering Module, STFM)进行结合。STFM实现了显式的空间信息过滤和隐式的时间特征过滤,确保视觉标记与用户查询紧密对齐,从而减少输入到LLM的视觉标记数量。该方法在零样本和少样本实验中均表现优异,显著减少了训练数据需求。

链接: https://arxiv.org/abs/2412.12833
作者: Zheng Cheng,Rendong Wang,Zhicheng Wang
机构: YITION.AI
关键词: made significant progress, multi-modal large language, significant progress, made significant, large language models
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 11 pages, 4 figures

点击查看摘要

Abstract:Recently, multi-modal large language models have made significant progress. However, visual information lacking of guidance from the user’s intention may lead to redundant computation and involve unnecessary visual noise, especially in long, untrimmed videos. To address this issue, we propose FocusChat, a text-guided multi-modal large language model (LLM) that emphasizes visual information correlated to the user’s prompt. In detail, Our model first undergoes the semantic extraction module, which comprises a visual semantic branch and a text semantic branch to extract image and text semantics, respectively. The two branches are combined using the Spatial-Temporal Filtering Module (STFM). STFM enables explicit spatial-level information filtering and implicit temporal-level feature filtering, ensuring that the visual tokens are closely aligned with the user’s query. It lowers the essential number of visual tokens inputted into the LLM. FocusChat significantly outperforms Video-LLaMA in zero-shot experiments, using an order of magnitude less training data with only 16 visual tokens occupied. It achieves results comparable to the state-of-the-art in few-shot experiments, with only 0.72M pre-training data.
zh

[CV-47] Differential Alignment for Domain Adaptive Object Detection AAAI25

【速读】: 该论文试图解决域适应目标检测 (Domain Adaptive Object Detection, DAOD) 中源域和目标域特征对齐时未考虑不同区域重要性的问题。解决方案的关键在于采用差异特征对齐策略,具体包括两个模块:1) 预测差异反馈实例对齐模块 (Prediction-Discrepancy Feedback Instance Alignment, PDFA),通过自适应地为教师-学生检测差异较大的实例分配更高权重,有效处理更重的域特定信息;2) 基于不确定性的前景导向图像对齐模块 (Uncertainty-based Foreground-oriented Image Alignment, UFOA),明确引导模型更多关注感兴趣区域。通过这两个模块,论文提出的方法在实验中表现出优于现有最先进方法的性能。

链接: https://arxiv.org/abs/2412.12830
作者: Xinyu He(1),Xinhui Li(1),Xiaojie Guo(1) ((1) College of Intelligence and Computing, Tianjin University, Tianjin, China)
机构: 未知
关键词: Domain adaptive object, object detector trained, labeled source-domain data, adaptive object detection, source-target feature alignment
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 11 pages, 8 figures, accepted by aaai25

点击查看摘要

Abstract:Domain adaptive object detection (DAOD) aims to generalize an object detector trained on labeled source-domain data to a target domain without annotations, the core principle of which is \emphsource-target feature alignment. Typically, existing approaches employ adversarial learning to align the distributions of the source and target domains as a whole, barely considering the varying significance of distinct regions, say instances under different circumstances and foreground \emphvs background areas, during feature alignment. To overcome the shortcoming, we investigates a differential feature alignment strategy. Specifically, a prediction-discrepancy feedback instance alignment module (dubbed PDFA) is designed to adaptively assign higher weights to instances of higher teacher-student detection discrepancy, effectively handling heavier domain-specific information. Additionally, an uncertainty-based foreground-oriented image alignment module (UFOA) is proposed to explicitly guide the model to focus more on regions of interest. Extensive experiments on widely-used DAOD datasets together with ablation studies are conducted to demonstrate the efficacy of our proposed method and reveal its superiority over other SOTA alternatives. Our code is available at this https URL.
zh

[CV-48] 2by2: Weakly-Supervised Learning for Global Action Segmentation

【速读】: 该论文试图解决全局动作分割(global action segmentation)这一研究较少的任务,旨在将捕捉到相同动作的帧从不同活动的视频中进行分组。由于不同视频中的动作时间顺序不一致,这一任务比处理单一活动视频更具挑战性。解决方案的关键在于利用活动标签(activity labels)进行弱监督学习,以获得适用于全局动作分割的动作表示。具体而言,论文提出了一种三元组学习方法(triadic learning approach),用于视频对,以确保视频内动作的区分性以及视频间和活动间的动作关联性。此外,采用基于稀疏Transformer的孪生网络(Siamese network based on sparse transformers)作为骨干架构,输入视频对并判断它们是否属于同一活动。该方法在Breakfast和YouTube Instructions两个具有挑战性的基准数据集上表现优异,超越了现有最先进的方法。

链接: https://arxiv.org/abs/2412.12829
作者: Elena Bueno-Benito,Mariella Dimiccoli
机构: 未知
关键词: grouping frames capturing, poorly investigated task, global action segmentation, aiming at grouping, paper presents
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:This paper presents a simple yet effective approach for the poorly investigated task of global action segmentation, aiming at grouping frames capturing the same action across videos of different activities. Unlike the case of videos depicting all the same activity, the temporal order of actions is not roughly shared among all videos, making the task even more challenging. We propose to use activity labels to learn, in a weakly-supervised fashion, action representations suitable for global action segmentation. For this purpose, we introduce a triadic learning approach for video pairs, to ensure intra-video action discrimination, as well as inter-video and inter-activity action association. For the backbone architecture, we use a Siamese network based on sparse transformers that takes as input video pairs and determine whether they belong to the same activity. The proposed approach is validated on two challenging benchmark datasets: Breakfast and YouTube Instructions, outperforming state-of-the-art methods.
zh

[CV-49] abSniper: Towards Accurate Table Detection Structure Recognition for Bank Statements

【速读】: 该论文试图解决从银行对账单中提取交易信息的问题,特别是在面对不同银行和模板布局多样性时,现有表格结构识别方法在处理复杂长表格时效果不佳且无法准确捕捉所有交易信息。解决方案的关键在于提出了TabSniper,一种新颖的表格检测、分类和结构识别方法。其核心流程包括从银行对账单中检测和分类感兴趣的表格,然后通过表格结构识别模型处理提取的表格区域,并结合后处理模块将交易数据转换为结构化和标准化的格式。该方法基于DETR架构,并通过多样化的银行对账单进行微调,辅以额外的特征增强,从而在处理多布局和模板时实现了高质量的交易信息提取。

链接: https://arxiv.org/abs/2412.12827
作者: Abhishek Trivedi,Sourajit Mukherjee,Rajat Kumar Singh,Vani Agarwal,Sriranjani Ramakrishnan,Himanshu S. Bhatt
机构: American Express(美国运通); American Express(美国运通); American Express(美国运通); American Express(美国运通); American Express(美国运通); American Express(美国运通)
关键词: bank statements, underwriting decisions, required to assess, well-being for credit, credit rating
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Extraction of transaction information from bank statements is required to assess one’s financial well-being for credit rating and underwriting decisions. Unlike other financial documents such as tax forms or financial statements, extracting the transaction descriptions from bank statements can provide a comprehensive and recent view into the cash flows and spending patterns. With multiple variations in layout and templates across several banks, extracting transactional level information from different table categories is an arduous task. Existing table structure recognition approaches produce sub optimal results for long, complex tables and are unable to capture all transactions accurately. This paper proposes TabSniper, a novel approach for efficient table detection, categorization and structure recognition from bank statements. The pipeline starts with detecting and categorizing tables of interest from the bank statements. The extracted table regions are then processed by the table structure recognition model followed by a post-processing module to transform the transactional data into a structured and standardised format. The detection and structure recognition architectures are based on DETR, fine-tuned with diverse bank statements along with additional feature enhancements. Results on challenging datasets demonstrate that TabSniper outperforms strong baselines and produces high-quality extraction of transaction information from bank and other financial documents across multiple layouts and templates.
zh

[CV-50] ComprehendEdit: A Comprehensive Dataset and Evaluation Framework for Multimodal Knowledge Editing AAAI2025

【速读】: 该论文试图解决多模态语言模型 (MLLMs) 中存在的知识过时或不准确的问题,并针对当前多模态知识编辑评估方法的局限性,提出了一个更全面的评估框架。解决方案的关键在于引入了ComprehendEdit基准,该基准包含八个多样化的任务,并提出了两个新的评估指标:知识泛化指数 (Knowledge Generalization Index, KGI) 和知识保留指数 (Knowledge Preservation Index, KPI),以更全面地评估编辑效果。此外,论文还提出了分层上下文编辑 (Hierarchical In-Context Editing, HICE) 方法,通过两阶段策略平衡各项指标的性能,从而提供了一个更有效的编辑技术基线。

链接: https://arxiv.org/abs/2412.12821
作者: Yaohui Ma,Xiaopeng Hong,Shizhou Zhang,Huiyun Li,Zhilin Zhu,Wei Luo,Zhiheng Ma
机构: Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences (深圳先进技术研究院,中国科学院); Shenzhen University (深圳大学); Shenzhen Key Laboratory of Advanced Materials (深圳先进材料重点实验室)
关键词: Large multimodal language, revolutionized natural language, natural language processing, multimodal language models, Large multimodal
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Extended version for paper accepted to AAAI 2025. Project Page: this https URL

点击查看摘要

Abstract:Large multimodal language models (MLLMs) have revolutionized natural language processing and visual understanding, but often contain outdated or inaccurate information. Current multimodal knowledge editing evaluations are limited in scope and potentially biased, focusing on narrow tasks and failing to assess the impact on in-domain samples. To address these issues, we introduce ComprehendEdit, a comprehensive benchmark comprising eight diverse tasks from multiple datasets. We propose two novel metrics: Knowledge Generalization Index (KGI) and Knowledge Preservation Index (KPI), which evaluate editing effects on in-domain samples without relying on AI-synthetic samples. Based on insights from our framework, we establish Hierarchical In-Context Editing (HICE), a baseline method employing a two-stage approach that balances performance across all metrics. This study provides a more comprehensive evaluation framework for multimodal knowledge editing, reveals unique challenges in this field, and offers a baseline method demonstrating improved performance. Our work opens new perspectives for future research and provides a foundation for developing more robust and effective editing techniques for MLLMs. The ComprehendEdit benchmark and implementation code are available at this https URL.
zh

[CV-51] Multi-View Incremental Learning with Structured Hebbian Plasticity for Enhanced Fusion Efficiency

【速读】: 该论文试图解决传统多视图学习方法在处理跨领域数据时难以有效泛化的问题,这些方法无法模拟人脑对顺序信号的动态整合能力。解决方案的关键在于提出了一个生物神经学启发的多视图增量框架(MVIL),该框架通过两个核心模块实现:结构化的海伯学习(structured Hebbian plasticity)和突触分区学习(synaptic partition learning)。前者通过重塑权重结构来表达视图表示之间的高相关性,从而实现细粒度的视图融合;后者通过抑制部分突触来缓解权重的剧烈变化并保留旧知识,从而增强网络对新信息的适应性和泛化能力。

链接: https://arxiv.org/abs/2412.12801
作者: Yuhong Chen,Ailin Song,Huifeng Yin,Shuai Zhong,Fuhai Chen,Qi Xu,Shiping Wang,Mingkun Xu
机构: 1. School of Computer Science and Technology, Soochow University(苏州大学计算机科学与技术学院);
2. Jiangsu Key Laboratory of Big Data Analysis Technology(江苏省大数据分析技术重点实验室);
3. School of Software, Soochow University(苏州大学软件学院);
4. School of Electronic and Information Engineering, Soochow University(苏州大学电子与信息工程学院);
5. School of Mathematics and Statistics, Soochow University(苏州大学数学与统计学院)
关键词: revolutionized human perception, rapid evolution, evolution of multimedia, multimedia technology, technology has revolutionized
类目: Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注: 11 pages

点击查看摘要

Abstract:The rapid evolution of multimedia technology has revolutionized human perception, paving the way for multi-view learning. However, traditional multi-view learning approaches are tailored for scenarios with fixed data views, falling short of emulating the intricate cognitive procedures of the human brain processing signals sequentially. Our cerebral architecture seamlessly integrates sequential data through intricate feed-forward and feedback mechanisms. In stark contrast, traditional methods struggle to generalize effectively when confronted with data spanning diverse domains, highlighting the need for innovative strategies that can mimic the brain’s adaptability and dynamic integration capabilities. In this paper, we propose a bio-neurologically inspired multi-view incremental framework named MVIL aimed at emulating the brain’s fine-grained fusion of sequentially arriving views. MVIL lies two fundamental modules: structured Hebbian plasticity and synaptic partition learning. The structured Hebbian plasticity reshapes the structure of weights to express the high correlation between view representations, facilitating a fine-grained fusion of view representations. Moreover, synaptic partition learning is efficient in alleviating drastic changes in weights and also retaining old knowledge by inhibiting partial synapses. These modules bionically play a central role in reinforcing crucial associations between newly acquired information and existing knowledge repositories, thereby enhancing the network’s capacity for generalization. Experimental results on six benchmark datasets show MVIL’s effectiveness over state-of-the-art methods.
zh

[CV-52] RCTrans: Radar-Camera Transformer via Radar Densifier and Sequential Decoder for 3D Object Detection AAAI2025

【速读】: 该论文试图解决雷达-相机3D目标检测中雷达点云稀疏和噪声导致的融合困难问题。解决方案的关键在于引入了一种名为Radar-Camera Transformer (RCTrans)的新型查询式检测方法。具体来说,首先设计了Radar Dense Encoder来丰富稀疏的有效雷达标记,然后将其与图像标记连接,以充分利用每个感兴趣区域的3D信息并减少融合阶段的空标记干扰。接着,通过Pruning Sequential Decoder基于获得的标记和随机初始化的查询来预测3D框,并通过顺序融合结构逐步定位物体位置,以缓解雷达点云中的高度模糊问题。此外,解码器中采用了剪枝训练策略,以节省推理时间并抑制查询的特征丧失。实验结果表明,该方法在大规模nuScenes数据集上表现优异,达到了新的雷达-相机3D检测性能的最新水平。

链接: https://arxiv.org/abs/2412.12799
作者: Yiheng Li,Yang Yang,Zhen Lei
机构: 未知
关键词: Radar Dense Encoder, named Radar-Camera Transformer, radar point clouds, radar modalities, Radar-Camera Transformer
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: Accepted by AAAI 2025

点击查看摘要

Abstract:In radar-camera 3D object detection, the radar point clouds are sparse and noisy, which causes difficulties in fusing camera and radar modalities. To solve this, we introduce a novel query-based detection method named Radar-Camera Transformer (RCTrans). Specifically, we first design a Radar Dense Encoder to enrich the sparse valid radar tokens, and then concatenate them with the image tokens. By doing this, we can fully explore the 3D information of each interest region and reduce the interference of empty tokens during the fusing stage. We then design a Pruning Sequential Decoder to predict 3D boxes based on the obtained tokens and random initialized queries. To alleviate the effect of elevation ambiguity in radar point clouds, we gradually locate the position of the object via a sequential fusion structure. It helps to get more precise and flexible correspondences between tokens and queries. A pruning training strategy is adopted in the decoder, which can save much time during inference and inhibit queries from losing their distinctiveness. Extensive experiments on the large-scale nuScenes dataset prove the superiority of our method, and we also achieve new state-of-the-art radar-camera 3D detection results. Our implementation is available at this https URL.
zh

[CV-53] ZoRI: Towards Discriminative Zero-Shot Remote Sensing Instance Segmentation AAAI2025

【速读】: 该论文试图解决遥感实例分割算法在面对未见过的场景和开放集预测时的局限性问题,提出了一个名为ZoRI的零样本遥感实例分割框架。解决方案的关键在于:1) 通过判别增强分类器使用精炼的文本嵌入来提高类别差异的识别能力;2) 采用知识保持适应策略,在不直接微调的情况下,解耦语义相关信息以保留预训练的视觉-语言对齐,同时调整特征以捕捉遥感领域特定的视觉线索;3) 引入先验注入预测,利用缓存库中的航空视觉原型来补充文本嵌入的语义丰富性,并无缝集成航空表示,以适应遥感领域。这些方法共同提升了模型在零样本遥感实例分割任务中的性能,达到了当前最先进的水平。

链接: https://arxiv.org/abs/2412.12798
作者: Shiqi Huang,Shuting He,Bihan Wen
机构: 未知
关键词: sensing instance segmentation, Instance segmentation algorithms, remote sensing instance, remote sensing, Instance segmentation
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: AAAI 2025, code see this https URL

点击查看摘要

Abstract:Instance segmentation algorithms in remote sensing are typically based on conventional methods, limiting their application to seen scenarios and closed-set predictions. In this work, we propose a novel task called zero-shot remote sensing instance segmentation, aimed at identifying aerial objects that are absent from training data. Challenges arise when classifying aerial categories with high inter-class similarity and intra-class variance. Besides, the domain gap between vision-language models’ pretraining datasets and remote sensing datasets hinders the zero-shot capabilities of the pretrained model when it is directly applied to remote sensing images. To address these challenges, we propose a \textbfZ ero-Sh \textbfo t \textbfR emote Sensing \textbfI nstance Segmentation framework, dubbed \textbfZoRI . Our approach features a discrimination-enhanced classifier that uses refined textual embeddings to increase the awareness of class disparities. Instead of direct fine-tuning, we propose a knowledge-maintained adaptation strategy that decouples semantic-related information to preserve the pretrained vision-language alignment while adjusting features to capture remote sensing domain-specific visual cues. Additionally, we introduce a prior-injected prediction with cache bank of aerial visual prototypes to supplement the semantic richness of text embeddings and seamlessly integrate aerial representations, adapting to the remote sensing domain. We establish new experimental protocols and benchmarks, and extensive experiments convincingly demonstrate that ZoRI achieves the state-of-art performance on the zero-shot remote sensing instance segmentation task. Our code is available at this https URL.
zh

[CV-54] CRoF: CLIP-based Robust Few-shot Learning on Noisy Labels

【速读】: 该论文试图解决在少样本学习(Few-shot Learning, FSL)任务中,噪声标签对基于CLIP(Contrastive Language-Image Pretraining)模型的分类鲁棒性造成的影响。解决方案的关键在于提出了CLIP-based Robust Few-shot learning (CRoF),这是一个通用的插件模块,旨在通过以下方式增强CLIP在噪声数据上的域泛化能力:1) 设计了一个面向少样本任务的提示生成器(prompt generator),生成更具区分性的类别描述,从而增大类间文本嵌入的距离;2) 采用加权策略(如label-smooth)对CLIP进行微调,结合CLIP的先验知识与原始标签信息,确保在噪声数据上的分类可靠性;3) 引入多标签损失函数,进一步支持在这种范式下的鲁棒训练。实验结果表明,CRoF在不同噪声类型和噪声比例下均优于微调后的CLIP模型和原始CLIP模型。

链接: https://arxiv.org/abs/2412.12793
作者: Shizhuo Deng,Bowen Han,Jiaqi Chen,Hao Wang,Dongyue Chen,Tong Jia
机构: College of Information Science and Engineering, Northeastern University Shenyang 110819, Liaoning, China(信息科学与工程学院,东北大学沈阳110819,辽宁,中国)
关键词: threaten the robustness, inexact features, CLIP, Noisy labels threaten, FSL
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Noisy labels threaten the robustness of few-shot learning (FSL) due to the inexact features in a new domain. CLIP, a large-scale vision-language model, performs well in FSL on image-text embedding similarities, but it is susceptible to misclassification caused by noisy labels. How to enhance domain generalization of CLIP on noisy data within FSL tasks is a critical challenge. In this paper, we provide a novel view to mitigate the influence of noisy labels, CLIP-based Robust Few-shot learning (CRoF). CRoF is a general plug-in module for CLIP-based models. To avoid misclassification and confused label embedding, we design the few-shot task-oriented prompt generator to give more discriminative descriptions of each category. The proposed prompt achieves larger distances of inter-class textual embedding. Furthermore, rather than fully trusting zero-shot classification by CLIP, we fine-tune CLIP on noisy few-shot data in a new domain with a weighting strategy like label-smooth. The weights for multiple potentially correct labels consider the relationship between CLIP’s prior knowledge and original label information to ensure reliability. Our multiple label loss function further supports robust training under this paradigm. Comprehensive experiments show that CRoF, as a plug-in, outperforms fine-tuned and vanilla CLIP models on different noise types and noise ratios.
zh

[CV-55] Implicit Location-Caption Alignment via Complementary Masking for Weakly-Supervised Dense Video Captioning AAAI2025

【速读】: 该论文试图解决弱监督密集视频描述 (Weakly-Supervised Dense Video Captioning, WSDVC) 中事件边界标注缺失的问题,即在不依赖事件边界标注的情况下,准确地定位和描述视频中的所有感兴趣事件。解决方案的关键在于提出了一种新的隐式位置-描述对齐范式,通过互补掩码 (complementary masking) 简化了复杂的事件提议和定位过程。具体来说,模型包含双模态视频描述模块和掩码生成模块:前者捕捉全局事件信息并生成描述性字幕,后者生成可微分的正负掩码用于事件定位。这些掩码通过确保正负掩码生成的字幕互补,从而隐式地对齐事件位置和字幕,即使在弱监督条件下也能实现有效的事件定位和描述。

链接: https://arxiv.org/abs/2412.12791
作者: Shiping Ge,Qiang Chen,Zhiwei Jiang,Yafeng Yin,Liu Qin,Ziyao Chen,Qing Gu
机构: Tencent WeChat(腾讯微信)
关键词: Dense Video Captioning, Weakly-Supervised Dense Video, Dense Video, Weakly-Supervised Dense, event
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Multimedia (cs.MM)
备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Weakly-Supervised Dense Video Captioning (WSDVC) aims to localize and describe all events of interest in a video without requiring annotations of event boundaries. This setting poses a great challenge in accurately locating the temporal location of event, as the relevant supervision is unavailable. Existing methods rely on explicit alignment constraints between event locations and captions, which involve complex event proposal procedures during both training and inference. To tackle this problem, we propose a novel implicit location-caption alignment paradigm by complementary masking, which simplifies the complex event proposal and localization process while maintaining effectiveness. Specifically, our model comprises two components: a dual-mode video captioning module and a mask generation module. The dual-mode video captioning module captures global event information and generates descriptive captions, while the mask generation module generates differentiable positive and negative masks for localizing the events. These masks enable the implicit alignment of event locations and captions by ensuring that captions generated from positively and negatively masked videos are complementary, thereby forming a complete video description. In this way, even under weak supervision, the event location and event caption can be aligned implicitly. Extensive experiments on the public datasets demonstrate that our method outperforms existing weakly-supervised methods and achieves competitive results compared to fully-supervised methods.
zh

[CV-56] RA-SGG: Retrieval-Augmented Scene Graph Generation Framework via Multi-Prototype Learning

【速读】: 该论文试图解决场景图生成 (Scene Graph Generation, SGG) 中的两个根本性问题:长尾谓语分布和谓语间的语义模糊性。这些问题导致现有模型偏向于预测常见的头部谓语,而忽略细粒度的谓语。论文通过将SGG问题重新定义为带有部分标注的多标签分类问题,提出了一种新的解决方案——检索增强的场景图生成 (Retrieval-Augmented Scene Graph Generation, RA-SGG)。该方法通过从建立的记忆库中检索相关样本,识别潜在的多标签实例,并利用语义相似的多标签来丰富原始的单标签。基于增强的关系(即发现的多标签),论文采用多原型学习来训练SGG模型。实验结果表明,RA-SGG在VG和GQA数据集上分别比现有最优方法提升了3.6%和5.9%,特别是在F@K指标上,有效缓解了长尾分布和语义模糊性导致的预测偏差问题。

链接: https://arxiv.org/abs/2412.12788
作者: Kanghoon Yoon,Kibum Kim,Jaehyung Jeon,Yeonjun In,Donghyun Kim,Chanyoung Park
机构: Korea Advanced Institute of Science and Technology (KAIST)(韩国科学技术院); Korea University (高丽大学)
关键词: Scene Graph Generation, Graph Generation, Scene Graph, research has suffered, Retrieval-Augmented Scene Graph
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 7 pages

点击查看摘要

Abstract:Scene Graph Generation (SGG) research has suffered from two fundamental challenges: the long-tailed predicate distribution and semantic ambiguity between predicates. These challenges lead to a bias towards head predicates in SGG models, favoring dominant general predicates while overlooking fine-grained predicates. In this paper, we address the challenges of SGG by framing it as multi-label classification problem with partial annotation, where relevant labels of fine-grained predicates are missing. Under the new frame, we propose Retrieval-Augmented Scene Graph Generation (RA-SGG), which identifies potential instances to be multi-labeled and enriches the single-label with multi-labels that are semantically similar to the original label by retrieving relevant samples from our established memory bank. Based on augmented relations (i.e., discovered multi-labels), we apply multi-prototype learning to train our SGG model. Several comprehensive experiments have demonstrated that RA-SGG outperforms state-of-the-art baselines by up to 3.6% on VG and 5.9% on GQA, particularly in terms of F@K, showing that RA-SGG effectively alleviates the issue of biased prediction caused by the long-tailed distribution and semantic ambiguity of predicates.
zh

[CV-57] Activating Distributed Visual Region within LLM s for Efficient and Effective Vision-Language Training and Inference

【速读】: 该论文试图解决大规模视觉-语言模型 (Large Vision-Language Models, LVLMs) 在训练过程中效率低下的问题。解决方案的关键在于通过选择性层调优 (selective layers tuning) 来激活模型中的“视觉区域”,即在模型中识别并仅更新部分关键层,从而在保持或提升文本任务性能的同时,显著减少训练时间和资源消耗。研究结果表明,仅更新25%的模型层,且这些层稀疏且均匀分布时,可以保留近99%的视觉性能,同时有效减少训练时间。基于此,论文进一步提出了基于视觉区域的剪枝范式,移除非关键层,以实现最小性能损失。

链接: https://arxiv.org/abs/2412.12785
作者: Siyuan Wang,Dianyi Wang,Chengxing Zhou,Zejun Li,Zhihao Fan,Xuanjing Huang,Zhongyu Wei
机构: University of Southern California; Fudan University; Sun Yat-sen University; Alibaba Inc.
关键词: Large Vision-Language Models, Large Vision-Language, typically learn visual, learn visual capacity, visual instruction tuning
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Large Vision-Language Models (LVLMs) typically learn visual capacity through visual instruction tuning, involving updates to both a projector and their LLM backbones. Drawing inspiration from the concept of visual region in the human brain, we investigate the existence of an analogous \textitvisual region within LLMs that functions as a cognitive core, and explore the possibility of efficient training of LVLMs via selective layers tuning. We use Bunny-Llama-3-8B-V for detailed experiments and LLaVA-1.5-7B and LLaVA-1.5-13B for validation across a range of visual and textual tasks. Our findings reveal that selectively updating 25% of LLMs layers, when sparsely and uniformly distributed, can preserve nearly 99% of visual performance while maintaining or enhancing textual task results, and also effectively reducing training time. Based on this targeted training approach, we further propose a novel visual region-based pruning paradigm, removing non-critical layers outside the visual region, which can achieve minimal performance loss. This study offers an effective and efficient strategy for LVLM training and inference by activating a layer-wise visual region within LLMs, which is consistently effective across different models and parameter scales.
zh

[CV-58] Bidirectional Logits Tree: Pursuing Granularity Reconcilement in Fine-Grained Classification

【速读】: 该论文试图解决细粒度分类任务中的粒度竞争问题 (Granularity Competition),这一问题源于多粒度标签之间的语义差距。现有方法通常基于共享特征开发独立的层次感知模型,但由于粗粒度层次本质上比细粒度层次更容易学习,基础编码器倾向于优先提取粗粒度特征,从而阻碍了细粒度特征的学习。为解决这一问题,论文提出了一个名为双向逻辑树 (Bidirectional Logits Tree, BiLT) 的新框架,其关键在于从最细粒度到最粗粒度依次开发分类器,而不是基于相同输入特征并行构建一组分类器。在这种设置下,细粒度分类器的输出作为粗粒度分类器的输入,促进了不同粒度之间层次语义信息的流动。此外,论文还引入了自适应粒度内差异学习 (Adaptive Intra-Granularity Difference Learning, AIGDL) 方法,以揭示同一粒度内类别之间的细微语义差异。实验结果证明了该方法的有效性。

链接: https://arxiv.org/abs/2412.12782
作者: Zhiguang Lu,Qianqian Xu,Shilong Bao,Zhiyong Yang,Qingming Huang
机构: 1. Tsinghua University(清华大学); 2. Institute for AI Industry Research, Tsinghua University(清华大学人工智能产业研究院); 3. Beijing National Research Center for Information Science and Technology(北京国家信息科学与技术研究中心)
关键词: fine-grained classification tasks, Granularity Competition, classification tasks, multi-granularity labels, paper addresses
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:This paper addresses the challenge of Granularity Competition in fine-grained classification tasks, which arises due to the semantic gap between multi-granularity labels. Existing approaches typically develop independent hierarchy-aware models based on shared features extracted from a common base encoder. However, because coarse-grained levels are inherently easier to learn than finer ones, the base encoder tends to prioritize coarse feature abstractions, which impedes the learning of fine-grained features. To overcome this challenge, we propose a novel framework called the Bidirectional Logits Tree (BiLT) for Granularity Reconcilement. The key idea is to develop classifiers sequentially from the finest to the coarsest granularities, rather than parallelly constructing a set of classifiers based on the same input features. In this setup, the outputs of finer-grained classifiers serve as inputs for coarser-grained ones, facilitating the flow of hierarchical semantic information across different granularities. On top of this, we further introduce an Adaptive Intra-Granularity Difference Learning (AIGDL) approach to uncover subtle semantic differences between classes within the same granularity. Extensive experiments demonstrate the effectiveness of our proposed method.
zh

[CV-59] Rethinking Diffusion-Based Image Generators for Fundus Fluorescein Angiography Synthesis on Limited Data

【速读】: 该论文试图解决传统荧光素眼底血管造影 (FFA) 的侵入性问题,即通过非侵入性眼底图像生成相应的FFA图像,以减少患者的不适和风险。解决方案的关键在于提出了一种基于潜在扩散模型 (latent diffusion model) 的新框架,称为Diffusion。该框架通过引入微调协议 (fine-tuning protocol) 来克服医学数据有限的挑战,并设计了一种新方法来应对跨模态和多种疾病类型生成FFA图像的复杂性。这一解决方案在有限数据集上实现了现有方法中的最先进性能,具有显著提升眼科诊断和患者护理的潜力。

链接: https://arxiv.org/abs/2412.12778
作者: Chengzhou Yu(South China University of Technology),Huihui Fang(Pazhou Laboratory),Hongqiu Wang(The Hong Kong University of Science and Technology (Guangzhou)),Ting Deng(South China University of Technology),Qing Du(South China University of Technology),Yanwu Xu(South China University of Technology),Weihua Yang(Shenzhen Eye Hospital)
机构: 未知
关键词: offering unique advantages, tool in ophthalmology, unique advantages, critical tool, Fundus imaging
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: 15 pages, 6 figures

点击查看摘要

Abstract:Fundus imaging is a critical tool in ophthalmology, with different imaging modalities offering unique advantages. For instance, fundus fluorescein angiography (FFA) can accurately identify eye diseases. However, traditional invasive FFA involves the injection of sodium fluorescein, which can cause discomfort and risks. Generating corresponding FFA images from non-invasive fundus images holds significant practical value but also presents challenges. First, limited datasets constrain the performance and effectiveness of models. Second, previous studies have primarily focused on generating FFA for single diseases or single modalities, often resulting in poor performance for patients with various ophthalmic conditions. To address these issues, we propose a novel latent diffusion model-based framework, Diffusion, which introduces a fine-tuning protocol to overcome the challenge of limited medical data and unleash the generative capabilities of diffusion models. Furthermore, we designed a new approach to tackle the challenges of generating across different modalities and disease types. On limited datasets, our framework achieves state-of-the-art results compared to existing methods, offering significant potential to enhance ophthalmic diagnostics and patient care. Our code will be released soon to support further research in this field.
zh

[CV-60] A Framework for Critical Evaluation of Text-to-Image Models: Integrating Art Historical Analysis Artistic Exploration and Critical Prompt Engineering

【速读】: 该论文试图解决当前文本生成图像模型评估中技术指标和偏见研究的局限性问题。解决方案的关键在于提出一个跨学科框架,通过整合艺术史分析、艺术探索和批判性提示工程,提供对模型能力和社会影响的更细致理解。艺术史分析用于检查视觉和象征元素,揭示潜在偏见;艺术探索通过创意实验发现隐藏的潜力和局限性;批判性提示工程则主动挑战模型的假设,暴露嵌入的偏见。该框架通过案例研究展示了其在揭示性别、种族和文化代表性偏见方面的实际应用,从而推动更公平、负责任和文化敏感的AI系统的发展。

链接: https://arxiv.org/abs/2412.12774
作者: Amalia Foka
机构: 未知
关键词: current technical metrics, art historical analysis, paper proposes, current technical, technical metrics
类目: Computer Vision and Pattern Recognition (cs.CV); Computers and Society (cs.CY)
备注:

点击查看摘要

Abstract:This paper proposes a novel interdisciplinary framework for the critical evaluation of text-to-image models, addressing the limitations of current technical metrics and bias studies. By integrating art historical analysis, artistic exploration, and critical prompt engineering, the framework offers a more nuanced understanding of these models’ capabilities and societal implications. Art historical analysis provides a structured approach to examine visual and symbolic elements, revealing potential biases and misrepresentations. Artistic exploration, through creative experimentation, uncovers hidden potentials and limitations, prompting critical reflection on the algorithms’ assumptions. Critical prompt engineering actively challenges the model’s assumptions, exposing embedded biases. Case studies demonstrate the framework’s practical application, showcasing how it can reveal biases related to gender, race, and cultural representation. This comprehensive approach not only enhances the evaluation of text-to-image models but also contributes to the development of more equitable, responsible, and culturally aware AI systems.
zh

[CV-61] Optimize the Unseen – Fast NeRF Cleanup with Free Space Prior

【速读】: 该论文试图解决NeRF在光度重建过程中引入的伪影(floaters)问题,特别是在训练相机未观察到的区域中,这些伪影会降低新视角合成的质量。解决方案的关键在于提出了一种快速的后处理方法,通过强制执行自由空间先验(Free Space Prior)来消除这些伪影,而不会破坏NeRF对已观察区域的表示。与依赖最大似然估计(Maximum Likelihood, ML)或复杂局部数据驱动先验的现有方法不同,该方法采用最大后验概率(Maximum-a-Posteriori, MAP)方法,在简单的全局先验假设下选择最优模型参数,即未观察区域应保持空置。这种方法不仅能够清理已观察和未观察区域的伪影,还能在复杂场景区域中提升新视角合成的质量,同时具有更快的推理速度和更短的训练时间。

链接: https://arxiv.org/abs/2412.12772
作者: Leo Segre,Shai Avidan
机构: Tel Aviv University (特拉维夫大学)
关键词: Neural Radiance Fields, Neural Radiance, Radiance Fields, photometric reconstruction introduces, reconstruction introduces artifacts
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Neural Radiance Fields (NeRF) have advanced photorealistic novel view synthesis, but their reliance on photometric reconstruction introduces artifacts, commonly known as “floaters”. These artifacts degrade novel view quality, especially in areas unseen by the training cameras. We present a fast, post-hoc NeRF cleanup method that eliminates such artifacts by enforcing our Free Space Prior, effectively minimizing floaters without disrupting the NeRF’s representation of observed regions. Unlike existing approaches that rely on either Maximum Likelihood (ML) estimation to fit the data or a complex, local data-driven prior, our method adopts a Maximum-a-Posteriori (MAP) approach, selecting the optimal model parameters under a simple global prior assumption that unseen regions should remain empty. This enables our method to clean artifacts in both seen and unseen areas, enhancing novel view quality even in challenging scene regions. Our method is comparable with existing NeRF cleanup models while being 2.5x faster in inference time, requires no additional memory beyond the original NeRF, and achieves cleanup training in less than 30 seconds. Our code will be made publically available.
zh

[CV-62] Guided and Variance-Corrected Fusion with One-shot Style Alignment for Large-Content Image Generation

【速读】: 该论文试图解决使用小型扩散模型生成大图像时出现的拼接痕迹(seams)、物体和风格不一致等伪影问题。解决方案的关键在于提出了三种创新方法:1) 引导融合 (Guided Fusion, GF),通过加权平均重叠区域来减少远距离图像区域对融合的负面影响;2) 方差校正融合 (Variance-Corrected Fusion, VCF),在平均后校正数据方差,以提高去噪扩散概率模型 (Denoising Diffusion Probabilistic Model) 的融合精度;3) 一次性风格对齐 (Style Alignment, SA),通过调整初始输入噪声生成一致的风格,而无需额外计算负担。这些方法显著提升了生成图像的质量,并且作为即插即用模块,可广泛应用于其他基于融合的大图像生成方法。

链接: https://arxiv.org/abs/2412.12771
作者: Shoukun Sun,Min Xian,Tiankai Yao,Fei Xu,Luca Capriotti
机构: University of Idaho(爱达荷大学); Idaho National Laboratory(爱达荷国家实验室)
关键词: gaining increasing popularity, Producing large images, Producing large, small diffusion models, training large models
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Producing large images using small diffusion models is gaining increasing popularity, as the cost of training large models could be prohibitive. A common approach involves jointly generating a series of overlapped image patches and obtaining large images by merging adjacent patches. However, results from existing methods often exhibit obvious artifacts, e.g., seams and inconsistent objects and styles. To address the issues, we proposed Guided Fusion (GF), which mitigates the negative impact from distant image regions by applying a weighted average to the overlapping regions. Moreover, we proposed Variance-Corrected Fusion (VCF), which corrects data variance at post-averaging, generating more accurate fusion for the Denoising Diffusion Probabilistic Model. Furthermore, we proposed a one-shot Style Alignment (SA), which generates a coherent style for large images by adjusting the initial input noise without adding extra computational burden. Extensive experiments demonstrated that the proposed fusion methods improved the quality of the generated image significantly. As a plug-and-play module, the proposed method can be widely applied to enhance other fusion-based methods for large image generation.
zh

[CV-63] owards a Training Free Approach for 3D Scene Editing

【速读】: 该论文试图解决现有3D场景编辑方法中依赖于训练NeRF(Neural Radiance Fields)模型、需要强位置先验和特定训练的问题,特别是在小规模场景和特定场景中的应用限制。解决方案的关键在于提出了一种名为FreeEdit的新方法,该方法利用网格表示(mesh representations)替代NeRF,并借助基础模型的进展实现了无需训练的编辑方式。FreeEdit引入了插入、替换和删除作为基本操作块,通过结合这些操作实现复杂的编辑,并且能够根据文本提示自动识别编辑对象和位置。此外,论文还提出了一种新的算法用于确定对象放置的最佳位置,并通过广泛的定量和定性评估展示了其方法的优越性。

链接: https://arxiv.org/abs/2412.12766
作者: Vivek Madhavaram,Shivangana Rawat,Chaitanya Devaguptapu,Charu Sharma,Manohar Kaul
机构: Machine Learning Lab, IIIT Hyderabad, India(印度海得拉巴国际信息技术研究所机器学习实验室); Fujitsu Research India(富士通印度研究所)
关键词: shown remarkable capabilities, Text driven diffusion, shown remarkable, remarkable capabilities, driven diffusion models
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Text driven diffusion models have shown remarkable capabilities in editing images. However, when editing 3D scenes, existing works mostly rely on training a NeRF for 3D editing. Recent NeRF editing methods leverages edit operations by deploying 2D diffusion models and project these edits into 3D space. They require strong positional priors alongside text prompt to identify the edit location. These methods are operational on small 3D scenes and are more generalized to particular scene. They require training for each specific edit and cannot be exploited in real-time edits. To address these limitations, we propose a novel method, FreeEdit, to make edits in training free manner using mesh representations as a substitute for NeRF. Training-free methods are now a possibility because of the advances in foundation model’s space. We leverage these models to bring a training-free alternative and introduce solutions for insertion, replacement and deletion. We consider insertion, replacement and deletion as basic blocks for performing intricate edits with certain combinations of these operations. Given a text prompt and a 3D scene, our model is capable of identifying what object should be inserted/replaced or deleted and location where edit should be performed. We also introduce a novel algorithm as part of FreeEdit to find the optimal location on grounding object for placement. We evaluate our model by comparing it with baseline models on a wide range of scenes using quantitative and qualitative metrics and showcase the merits of our method with respect to others.
zh

[CV-64] Monocular Facial Appearance Capture in the Wild

【速读】: 该论文试图解决从轻量级、非受控环境下的单目视频中重建人脸外观属性的问题。解决方案的关键在于通过分析视频中简单的头部旋转,恢复人脸的表面几何(surface geometry)、漫反射反照率(diffuse albedo)、镜面反射强度(specular intensity)和镜面粗糙度(specular roughness),同时不依赖于环境光照的简化假设,并明确考虑可见性和遮挡问题。这种方法能够在保持高保真度的同时,显著降低采集过程的复杂性和成本,接近多视角工作室捕捉的效果。

链接: https://arxiv.org/abs/2412.12765
作者: Yingyan Xu,Kate Gadola,Prashanth Chandran,Sebastian Weiss,Markus Gross,Gaspard Zoss,Derek Bradley
机构: ETH Zürich(苏黎世联邦理工学院); DisneyResearch|Studios(迪士尼研究院|工作室)
关键词: properties of human, human faces, lightweight capture procedure, unconstrained environment, appearance properties
类目: Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR)
备注:

点击查看摘要

Abstract:We present a new method for reconstructing the appearance properties of human faces from a lightweight capture procedure in an unconstrained environment. Our method recovers the surface geometry, diffuse albedo, specular intensity and specular roughness from a monocular video containing a simple head rotation in-the-wild. Notably, we make no simplifying assumptions on the environment lighting, and we explicitly take visibility and occlusions into account. As a result, our method can produce facial appearance maps that approach the fidelity of studio-based multi-view captures, but with a far easier and cheaper procedure.
zh

[CV-65] Progressive Monitoring of Generative Model Training Evolution

【速读】: 该论文试图解决深度生成模型 (Deep Generative Models, DGMs) 在训练过程中容易受到偏差和其他低效因素影响的问题,这些问题可能导致不理想的结果。解决方案的关键在于引入一个渐进分析框架,通过利用降维技术来监控训练过程中潜在表示、生成数据分布和真实数据分布的变化及其演化。这种方法能够在训练过程中及时发现偏差和失败,并通过暂停和修正训练方法来优化资源利用,从而提高生成数据分布的质量并减少计算负荷。

链接: https://arxiv.org/abs/2412.12755
作者: Vidya Prasad,Anna Vilanova,Nicola Pezzotti
机构: 未知
关键词: undesirable outcomes remains, deep generative models, gained popularity, inefficiencies that lead, lead to undesirable
类目: Machine Learning (cs.LG); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:While deep generative models (DGMs) have gained popularity, their susceptibility to biases and other inefficiencies that lead to undesirable outcomes remains an issue. With their growing complexity, there is a critical need for early detection of issues to achieve desired results and optimize resources. Hence, we introduce a progressive analysis framework to monitor the training process of DGMs. Our method utilizes dimensionality reduction techniques to facilitate the inspection of latent representations, the generated and real distributions, and their evolution across training iterations. This monitoring allows us to pause and fix the training method if the representations or distributions progress undesirably. This approach allows for the analysis of a models’ training dynamics and the timely identification of biases and failures, minimizing computational loads. We demonstrate how our method supports identifying and mitigating biases early in training a Generative Adversarial Network (GAN) and improving the quality of the generated data distribution.
zh

[CV-66] Open-World Panoptic Segmentation

【速读】: 该论文试图解决开放世界全景分割 (open-world panoptic segmentation) 问题,即在测试时发现新的语义类别和新的对象实例,同时确保这些类别之间的一致性。解决方案的关键是提出了 Con2MAV 方法,该方法扩展了之前用于开放世界语义分割的 ContMAV 模型,通过增量式地发现新类别并保持一致性来实现开放世界全景分割。实验结果表明,Con2MAV 在开放世界分割任务中达到了最先进的性能,同时在已知类别上仍表现出色。此外,论文还提出了 PANIC 基准,用于评估自动驾驶场景中的开放世界全景分割,该基准包含 800 张图像和 50 多个未知类别,提供了高质量的像素级标注,为开放世界分割任务提供了极具挑战性的数据集。

链接: https://arxiv.org/abs/2412.12740
作者: Matteo Sodano,Federico Magistri,Jens Behley,Cyrill Stachniss
机构: Center for Robotics, University of Bonn, Germany(机器人中心,波恩大学,德国); Lamarr Institute for Machine Learning and Artificial Intelligence, Germany(Lamarr机器学习和人工智能研究所,德国)
关键词: key building block, autonomously acting vision, acting vision systems, key building, building block
类目: Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO)
备注: Submitted to PAMI

点击查看摘要

Abstract:Perception is a key building block of autonomously acting vision systems such as autonomous vehicles. It is crucial that these systems are able to understand their surroundings in order to operate safely and robustly. Additionally, autonomous systems deployed in unconstrained real-world scenarios must be able of dealing with novel situations and object that have never been seen before. In this article, we tackle the problem of open-world panoptic segmentation, i.e., the task of discovering new semantic categories and new object instances at test time, while enforcing consistency among the categories that we incrementally discover. We propose Con2MAV, an approach for open-world panoptic segmentation that extends our previous work, ContMAV, which was developed for open-world semantic segmentation. Through extensive experiments across multiple datasets, we show that our model achieves state-of-the-art results on open-world segmentation tasks, while still performing competitively on the known categories. We will open-source our implementation upon acceptance. Additionally, we propose PANIC (Panoptic ANomalies In Context), a benchmark for evaluating open-world panoptic segmentation in autonomous driving scenarios. This dataset, recorded with a multi-modal sensor suite mounted on a car, provides high-quality, pixel-wise annotations of anomalous objects at both semantic and instance level. Our dataset contains 800 images, with more than 50 unknown classes, i.e., classes that do not appear in the training set, and 4000 object instances, making it an extremely challenging dataset for open-world segmentation tasks in the autonomous driving scenario. We provide competitions for multiple open-world tasks on a hidden test set. Our dataset and competitions are available at this https URL.
zh

[CV-67] PolSAM: Polarimetric Scattering Mechanism Informed Segment Anything Model

【速读】: 该论文试图解决极化合成孔径雷达 (PolSAR) 数据在特征提取和数据表示方面面临的挑战,特别是现有方法在可用性、可解释性和数据完整性上的不足。解决方案的关键在于提出了极化散射机制引导的SAM (PolSAM),该模型通过引入微波视觉数据 (Microwave Vision Data, MVD) 这种轻量级且可解释的数据表示,并结合特征级融合提示 (Feature-Level Fusion Prompt, FFP) 和语义级融合提示 (Semantic-Level Fusion Prompt, SFP) 来增强SAM的性能。FFP通过融合伪彩色SAR图像和MVD的视觉标记来解决模态不兼容问题,而SFP则利用语义信息优化稀疏和密集分割提示,从而显著提升分割精度、减少数据存储并加速推理时间。

链接: https://arxiv.org/abs/2412.12737
作者: Yuqing Wang,Zhongling Huang,Shuxin Yang,Hao Tang,Xiaolan Qiu,Junwei Han,Dingwen Zhang
机构: Northwestern Polytechnical University(西北工业大学); Peking University(北京大学); Chinese Academy of Sciences(中国科学院)
关键词: presents unique challenges, unique challenges due, data presents unique, PolSAR data presents, presents unique
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: The manuscript is 15 pages long, includes 14 figures and 5 tables

点击查看摘要

Abstract:PolSAR data presents unique challenges due to its rich and complex characteristics. Existing data representations, such as complex-valued data, polarimetric features, and amplitude images, are widely used. However, these formats often face issues related to usability, interpretability, and data integrity. Most feature extraction networks for PolSAR are small, limiting their ability to capture features effectively. To address these issues, We propose the Polarimetric Scattering Mechanism-Informed SAM (PolSAM), an enhanced Segment Anything Model (SAM) that integrates domain-specific scattering characteristics and a novel prompt generation strategy. PolSAM introduces Microwave Vision Data (MVD), a lightweight and interpretable data representation derived from polarimetric decomposition and semantic correlations. We propose two key components: the Feature-Level Fusion Prompt (FFP), which fuses visual tokens from pseudo-colored SAR images and MVD to address modality incompatibility in the frozen SAM encoder, and the Semantic-Level Fusion Prompt (SFP), which refines sparse and dense segmentation prompts using semantic information. Experimental results on the PhySAR-Seg datasets demonstrate that PolSAM significantly outperforms existing SAM-based and multimodal fusion models, improving segmentation accuracy, reducing data storage, and accelerating inference time. The source code and datasets will be made publicly available at \urlthis https URL.
zh

[CV-68] Gaussian Billboards: Expressive 2D Gaussian Splatting with Textures

【速读】: 该论文试图解决2D高斯喷射(2D Gaussian Splatting, 2DGS)在场景表示中的颜色表达能力有限的问题。解决方案的关键在于提出高斯广告牌(Gaussian Billboards),通过在每个喷射点上引入空间变化的色彩,使用逐喷射点的纹理插值来增强颜色表达能力。这一方法结合了2DGS的场景优化能力和传统纹理映射的表现力,从而在保持2DGS的鲁棒性同时,提升了场景表示的清晰度和质量。

链接: https://arxiv.org/abs/2412.12734
作者: Sebastian Weiss,Derek Bradley
机构: DisneyResearch||||Studios(迪士尼研究工作室)
关键词: Gaussian Splatting, reconstructing and rendering, recently emerged, Gaussian, Splatting has recently
类目: Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR)
备注:

点击查看摘要

Abstract:Gaussian Splatting has recently emerged as the go-to representation for reconstructing and rendering 3D scenes. The transition from 3D to 2D Gaussian primitives has further improved multi-view consistency and surface reconstruction accuracy. In this work we highlight the similarity between 2D Gaussian Splatting (2DGS) and billboards from traditional computer graphics. Both use flat semi-transparent 2D geometry that is positioned, oriented and scaled in 3D space. However 2DGS uses a solid color per splat and an opacity modulated by a Gaussian distribution, where billboards are more expressive, modulating the color with a uv-parameterized texture. We propose to unify these concepts by presenting Gaussian Billboards, a modification of 2DGS to add spatially-varying color achieved using per-splat texture interpolation. The result is a mixture of the two representations, which benefits from both the robust scene optimization power of 2DGS and the expressiveness of texture mapping. We show that our method can improve the sharpness and quality of the scene representation in a wide range of qualitative and quantitative evaluations compared to the original 2DGS implementation.
zh

[CV-69] RaCFormer: Towards High-Quality 3D Object Detection via Query-based Radar-Camera Fusion

【速读】: 该论文试图解决户外3D场景感知中雷达与摄像头融合时因图像到鸟瞰图(BEV)转换中的深度估计不准确导致的特征对齐问题。解决方案的关键在于提出了一种基于查询的框架,能够自适应地从BEV和原始图像视图中采样与实例相关的特征,并通过优化查询初始化和增强BEV的表征能力来提升系统性能。具体来说,论文引入了极坐标下的自适应圆形分布来优化对象查询的初始化,并利用雷达引导的深度头来改进图像视图到BEV的转换,同时通过引入隐式动态捕捉器来利用雷达的多普勒效应捕捉BEV中的时间元素。实验结果表明,该方法在nuScenes和View-of-Delft (VoD)数据集上均取得了优异的性能,甚至在nuScenes数据集上超越了多个基于LiDAR的检测器。

链接: https://arxiv.org/abs/2412.12725
作者: Xiaomeng Chu,Jiajun Deng,Guoliang You,Yifan Duan,Houqiang Li,Yanyong Zhang
机构: University of Science and Technology of China(中国科学技术大学); The University of Adelaide(阿德莱德大学)
关键词: Radar-Camera fusion transformer, propose Radar-Camera fusion, Radar-Camera fusion, boost the accuracy, fusion transformer
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:We propose Radar-Camera fusion transformer (RaCFormer) to boost the accuracy of 3D object detection by the following insight. The Radar-Camera fusion in outdoor 3D scene perception is capped by the image-to-BEV transformation–if the depth of pixels is not accurately estimated, the naive combination of BEV features actually integrates unaligned visual content. To avoid this problem, we propose a query-based framework that enables adaptively sample instance-relevant features from both the BEV and the original image view. Furthermore, we enhance system performance by two key designs: optimizing query initialization and strengthening the representational capacity of BEV. For the former, we introduce an adaptive circular distribution in polar coordinates to refine the initialization of object queries, allowing for a distance-based adjustment of query density. For the latter, we initially incorporate a radar-guided depth head to refine the transformation from image view to BEV. Subsequently, we focus on leveraging the Doppler effect of radar and introduce an implicit dynamic catcher to capture the temporal elements within the BEV. Extensive experiments on nuScenes and View-of-Delft (VoD) datasets validate the merits of our design. Remarkably, our method achieves superior results of 64.9% mAP and 70.2% NDS on nuScenes, even outperforming several LiDAR-based detectors. RaCFormer also secures the 1st ranking on the VoD dataset. The code will be released.
zh

[CV-70] Defending LVLMs Against Vision Attacks through Partial-Perception Supervision

【速读】: 该论文试图解决大型视觉语言模型 (Large Vision Language Models, LVLMs) 在面对恶意注入或扰动的输入图像时容易受到误导的问题。解决方案的关键在于提出了一种无需训练的黑箱方法,称为DPS (Defense through Partial-Perception Supervision)。该方法通过利用感知部分图像的模型生成的响应来监督原始图像的响应,使得模型在受到攻击时能够基于部分图像的理解调整其响应,而在面对干净输入时则保持原有的高置信度响应。实验结果表明,这种方法显著降低了攻击成功率,平均减少了76.3%。

链接: https://arxiv.org/abs/2412.12722
作者: Qi Zhou,Tianlin Li,Qing Guo,Dongxia Wang,Yun Lin,Yang Liu,Jin Song Dong
机构: 未知
关键词: Large Vision Language, Vision Language Models, Large Vision, Vision Language, raised significant concerns
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Cryptography and Security (cs.CR)
备注:

点击查看摘要

Abstract:Recent studies have raised significant concerns regarding the vulnerability of Large Vision Language Models (LVLMs) to maliciously injected or perturbed input images, which can mislead their responses. Existing defense methods show that such vision attacks are sensitive to image modifications especially cropping, using majority voting across responses of modified images as corrected responses. However, these modifications often result in partial images and distort the semantics, which reduces response quality on clean images after voting. Instead of directly using responses from partial images for voting, we investigate using them to supervise the LVLM’s responses to the original images. We propose a black-box, training-free method called DPS (Defense through Partial-Perception Supervision). In this approach, the model is prompted using the responses generated by a model that perceives only a partial image. With DPS, the model can adjust its response based on partial image understanding when under attack, while confidently maintaining its original response for clean input. Our findings show that the weak model can supervise the strong model: when faced with an attacked input, the strong model becomes less confident and adjusts its response based on the weak model’s partial understanding, effectively defending against the attack. With clean input, it confidently maintains its original response. Empirical experiments show our method outperforms the baseline, cutting the average attack success rate by 76.3% across six datasets on three popular models.
zh

[CV-71] ASAP: Advancing Semantic Alignment Promotes Multi-Modal Manipulation Detecting and Grounding

【速读】: 该论文试图解决多模态媒体操纵检测与定位 (DGM4) 中的关键问题,即跨模态语义对齐的不足,这限制了操纵检测的准确性。解决方案的关键在于通过利用现成的多模态大语言模型 (MLLMs) 和大型语言模型 (LLMs) 构建图像-文本对,特别是针对被操纵的实例,进行跨模态语义对齐学习。此外,论文设计了一种操纵引导的交叉注意力机制 (MGCA),通过在训练过程中利用定位真值,增强模型对操纵部分的注意力,同时减少对正常部分的注意力,从而提升模型捕捉操纵的能力。实验结果表明,该方法在DGM4数据集上显著优于现有方法。

链接: https://arxiv.org/abs/2412.12718
作者: Zhenxing Zhang,Yaxiong Wang,Lechao Cheng,Zhun Zhong,Dan Guo,Meng Wang
机构: School of Computer Science and Information Engineering, Hefei University of Technology, China; Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, China
关键词: accurate fine-grained cross-modal, present ASAP, accurately manipulation detection, multi-modal media manipulation, grounding multi-modal media
类目: Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM)
备注: 12 pages, 6 figures

点击查看摘要

Abstract:We present ASAP, a new framework for detecting and grounding multi-modal media manipulation (DGM4).Upon thorough examination, we observe that accurate fine-grained cross-modal semantic alignment between the image and text is vital for accurately manipulation detection and grounding. While existing DGM4 methods pay rare attention to the cross-modal alignment, hampering the accuracy of manipulation detecting to step further. To remedy this issue, this work targets to advance the semantic alignment learning to promote this task. Particularly, we utilize the off-the-shelf Multimodal Large-Language Models (MLLMs) and Large Language Models (LLMs) to construct paired image-text pairs, especially for the manipulated instances. Subsequently, a cross-modal alignment learning is performed to enhance the semantic alignment. Besides the explicit auxiliary clues, we further design a Manipulation-Guided Cross Attention (MGCA) to provide implicit guidance for augmenting the manipulation perceiving. With the grounding truth available during training, MGCA encourages the model to concentrate more on manipulated components while downplaying normal ones, enhancing the model’s ability to capture manipulations. Extensive experiments are conducted on the DGM4 dataset, the results demonstrate that our model can surpass the comparison method with a clear margin.
zh

[CV-72] Unsupervised UAV 3D Trajectories Estimation with Sparse Point Clouds

【速读】: 该论文试图解决小型无人机系统(UAV)因其尺寸小而难以通过传统方法检测的安全挑战。解决方案的关键在于利用空间-时间序列处理技术,通过融合多个LiDAR扫描数据,实现对无人机的准确跟踪。具体方法包括将点云数据分割为前景和背景,分析空间-时间数据,并采用评分机制提高检测精度。该方法在CVPR 2024 UG2+ Challenge中获得第4名,显示出其实际有效性,并计划将所有设计、代码和样本数据开源供研究社区使用。

链接: https://arxiv.org/abs/2412.12716
作者: Hanfang Liang,Yizhuo Yang,Jinming Hu,Jianfei Yang,Fen Liu,Shenghai Yuan
机构: 未知
关键词: Compact UAV systems, pose significant security, Compact UAV, significant security challenges, security challenges due
类目: Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO)
备注:

点击查看摘要

Abstract:Compact UAV systems, while advancing delivery and surveillance, pose significant security challenges due to their small size, which hinders detection by traditional methods. This paper presents a cost-effective, unsupervised UAV detection method using spatial-temporal sequence processing to fuse multiple LiDAR scans for accurate UAV tracking in real-world scenarios. Our approach segments point clouds into foreground and background, analyzes spatial-temporal data, and employs a scoring mechanism to enhance detection accuracy. Tested on a public dataset, our solution placed 4th in the CVPR 2024 UG2+ Challenge, demonstrating its practical effectiveness. We plan to open-source all designs, code, and sample data for the research community this http URL.
zh

[CV-73] MapExpert: Online HD Map Construction with Simple and Efficient Sparse Map Element Expert

【速读】: 该论文试图解决现有在线高清晰度地图构建方法在区分不同非立方体地图元素时存在的准确性问题。解决方案的关键在于引入了一种基于专家的在线高清晰度地图方法,称为MapExpert。MapExpert通过稀疏专家(sparse experts)和路由器(routers)的分布式架构,能够准确描述各种非立方体地图元素。此外,论文还提出了辅助平衡损失函数(auxiliary balance loss function)以均匀分配专家的负载,并引入了一种高效的时序融合模块(Learnable Weighted Moving Descentage),用于有效整合历史信息到最终的鸟瞰图(BEV)特征中。这些创新使得MapExpert在nuScenes和Argoverse2数据集上实现了最先进的性能,同时保持了良好的效率。

链接: https://arxiv.org/abs/2412.12704
作者: Dapeng Zhang,Dayu Chen,Peng Zhi,Yinda Chen,Zhenlong Yuan,Chenyang Li,Sunjing,Rui Zhou,Qingguo Zhou
机构: 1. School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China(哈尔滨工业大学计算机科学与技术学院,哈尔滨,中国);
2. School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China(哈尔滨工业大学计算机科学与技术学院,哈尔滨,中国);
3. School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China(哈尔滨工业大学计算机科学与技术学院,哈尔滨,中国);
4. School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China(哈尔滨工业大学计算机科学与技术学院,哈尔滨,中国)
关键词: Constructing online High-Definition, autonomous driving systems, static environment perception, Constructing online, driving systems
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Constructing online High-Definition (HD) maps is crucial for the static environment perception of autonomous driving systems (ADS). Existing solutions typically attempt to detect vectorized HD map elements with unified models; however, these methods often overlook the distinct characteristics of different non-cubic map elements, making accurate distinction challenging. To address these issues, we introduce an expert-based online HD map method, termed MapExpert. MapExpert utilizes sparse experts, distributed by our routers, to describe various non-cubic map elements accurately. Additionally, we propose an auxiliary balance loss function to distribute the load evenly across experts. Furthermore, we theoretically analyze the limitations of prevalent bird’s-eye view (BEV) feature temporal fusion methods and introduce an efficient temporal fusion module called Learnable Weighted Moving Descentage. This module effectively integrates relevant historical information into the final BEV features. Combined with an enhanced slice head branch, the proposed MapExpert achieves state-of-the-art performance and maintains good efficiency on both nuScenes and Argoverse2 datasets.
zh

[CV-74] ALADE-SNN: Adaptive Logit Alignment in Dynamically Expandable Spiking Neural Networks for Class Incremental Learning

【速读】: 该论文试图解决在类增量学习 (Class Incremental Learning, CIL) 中,由于数据集有限和固定特征提取器导致的偏差和过拟合问题。解决方案的关键在于提出了ALADE-SNN框架,通过自适应logit对齐 (adaptive logit alignment) 来平衡特征表示,并引入OtoN抑制机制来管理冻结的旧特征与新类之间的权重映射,在微调阶段释放这些权重。这种方法根据分析结果动态调整网络结构,提升了特征提取能力,并在新旧任务之间实现了性能平衡,从而在CIFAR100-B0基准测试中取得了75.42%的平均增量准确率,超越了现有的SNN持续学习算法。

链接: https://arxiv.org/abs/2412.12696
作者: Wenyao Ni,Jiangrong Shen,Qi Xu,Huajin Tang
机构: 1. School of Computer Science and Technology, Zhejiang University(浙江大学计算机科学与技术学院);
2. State Key Laboratory of CAD&CG, Zhejiang University(浙江大学计算机辅助设计与图形学国家重点实验室);
3. College of Information Science and Electronic Engineering, Zhejiang University(浙江大学信息科学与电子工程学院);
4. Department of Computer Science, University of California, Santa Barbara(加州大学圣塔芭芭拉分校计算机科学系)
关键词: erasing prior knowledge, human brain ability, develop spiking neural, structures for Class, Class Incremental Learning
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Inspired by the human brain’s ability to adapt to new tasks without erasing prior knowledge, we develop spiking neural networks (SNNs) with dynamic structures for Class Incremental Learning (CIL). Our comparative experiments reveal that limited datasets introduce biases in logits distributions among tasks. Fixed features from frozen past-task extractors can cause overfitting and hinder the learning of new tasks. To address these challenges, we propose the ALADE-SNN framework, which includes adaptive logit alignment for balanced feature representation and OtoN suppression to manage weights mapping frozen old features to new classes during training, releasing them during fine-tuning. This approach dynamically adjusts the network architecture based on analytical observations, improving feature extraction and balancing performance between new and old tasks. Experiment results show that ALADE-SNN achieves an average incremental accuracy of 75.42 on the CIFAR100-B0 benchmark over 10 incremental steps. ALADE-SNN not only matches the performance of DNN-based methods but also surpasses state-of-the-art SNN-based continual learning algorithms. This advancement enhances continual learning in neuromorphic computing, offering a brain-inspired, energy-efficient solution for real-time data processing.
zh

[CV-75] SPHERE: A Hierarchical Evaluation on Spatial Perception and Reasoning for Vision-Language Models

【速读】: 该论文试图解决当前视觉-语言模型在多维空间推理能力上的不足,特别是在理解距离和接近度、从自我中心和外界视角进行推理,以及在物理环境中进行复杂推理方面的缺陷。解决方案的关键在于开发了一个名为SPHERE的分层评估框架,并引入了一个新的人类标注数据集,通过从单一技能任务逐步过渡到多技能任务,最终实现需要整合多种空间和视觉线索与逻辑推理的复杂任务,从而揭示现有模型的不足,并为提升视觉-语言模型的空间理解与推理能力提供方向。

链接: https://arxiv.org/abs/2412.12693
作者: Wenyu Zhang,Wei En Ng,Lixin Ma,Yuwen Wang,Jungqi Zhao,Boyang Li,Lu Wang
机构: Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR); National University of Singapore (NUS); Tongji University; Nanyang Technological University (NTU)
关键词: Current vision-language models, basic spatial directions, incorporate single-dimensional spatial, single-dimensional spatial cues, multi-dimensional spatial reasoning
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Current vision-language models may incorporate single-dimensional spatial cues, such as depth, object boundary, and basic spatial directions (e.g. left, right, front, back), yet often lack the multi-dimensional spatial reasoning necessary for human-like understanding and real-world applications. To address this gap, we develop SPHERE (Spatial Perception and Hierarchical Evaluation of REasoning), a hierarchical evaluation framework with a new human-annotated dataset to pinpoint model strengths and weaknesses, advancing from single-skill tasks to multi-skill tasks, and ultimately to complex reasoning tasks that require the integration of multiple spatial and visual cues with logical reasoning. Benchmark evaluation of state-of-the-art open-source models reveal significant shortcomings, especially in the abilities to understand distance and proximity, to reason from both allocentric and egocentric viewpoints, and to perform complex reasoning in a physical context. This work underscores the need for more advanced approaches to spatial understanding and reasoning, paving the way for improvements in vision-language models and their alignment with human-like spatial capabilities. The dataset will be open-sourced upon publication.
zh

[CV-76] SemStereo: Semantic-Constrained Stereo Matching Network for Remote Sensing AAAI2025

【速读】: 该论文试图解决语义分割和立体匹配这两个异构任务之间的耦合问题,传统方法通常将它们视为独立或松散耦合的任务,未能明确建模两者之间的约束关系。论文提出了一种新的网络结构,通过隐式和显式的方式在立体匹配任务中引入语义约束。关键解决方案包括:1)将传统的并行结构转换为语义引导的级联结构(Semantic-Guided Cascade structure),利用富含语义信息的深度特征计算初始视差图,增强语义指导;2)引入语义选择性细化模块(Semantic Selective Refinement, SSR)和左右视图语义一致性模块(Left-Right Semantic Consistency, LRSC),前者在语义图的指导下细化初始视差图,后者通过减少语义图在视图转换后的语义偏差来确保双视图的语义一致性。实验结果表明,该方法在语义分割和立体匹配任务上均达到了最先进的性能。

链接: https://arxiv.org/abs/2412.12685
作者: Chen Chen,Liangjin Zhao,Yuanchun He,Yingxuan Long,Kaiqiang Chen,Zhirui Wang,Yanfeng Hu,Xian Sun
机构: 1. Key Laboratory of Computational Optical Imaging Technology, Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun, China(计算光学成像技术重点实验室,中国科学院长春光学精密机械与物理研究所,长春,中国); 2. University of Chinese Academy of Sciences, Beijing, China(中国科学院大学,北京,中国)
关键词: loosely coupled tasks, Semantic, loosely coupled parallel, loosely coupled, remote sensing
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 9 pages, 6 figures, AAAI 2025

点击查看摘要

Abstract:Semantic segmentation and 3D reconstruction are two fundamental tasks in remote sensing, typically treated as separate or loosely coupled tasks. Despite attempts to integrate them into a unified network, the constraints between the two heterogeneous tasks are not explicitly modeled, since the pioneering studies either utilize a loosely coupled parallel structure or engage in only implicit interactions, failing to capture the inherent connections. In this work, we explore the connections between the two tasks and propose a new network that imposes semantic constraints on the stereo matching task, both implicitly and explicitly. Implicitly, we transform the traditional parallel structure to a new cascade structure termed Semantic-Guided Cascade structure, where the deep features enriched with semantic information are utilized for the computation of initial disparity maps, enhancing semantic guidance. Explicitly, we propose a Semantic Selective Refinement (SSR) module and a Left-Right Semantic Consistency (LRSC) module. The SSR refines the initial disparity map under the guidance of the semantic map. The LRSC ensures semantic consistency between two views via reducing the semantic divergence after transforming the semantic map from one view to the other using the disparity map. Experiments on the US3D and WHU datasets demonstrate that our method achieves state-of-the-art performance for both semantic segmentation and stereo matching.
zh

[CV-77] ShiftedBronzes: Benchmarking and Analysis of Domain Fine-Grained Classification in Open-World Settings

【速读】: 该论文试图解决在特定领域(如青铜器年代鉴定)中复杂分布外(OOD, Out-of-Distribution)挑战的问题。解决方案的关键在于构建了一个名为ShiftedBronzes的基准数据集,该数据集扩展了青铜器Ding数据集,包含了两种类型的青铜器数据和七种常见的OOD数据,这些数据展示了青铜器年代鉴定场景中常见的分布偏移。通过在ShiftedBronzes和五个常用的通用OOD数据集上进行基准实验,并采用多种后处理、预训练视觉大模型(VLM, Vision Large Model)和生成式OOD检测方法,研究验证了这些方法在特定领域数据集上的不同表现,强调了将通用OOD检测方法应用于青铜器年代鉴定等特定任务时的独特挑战。

链接: https://arxiv.org/abs/2412.12683
作者: Rixin Zhou,Honglin Pang,Qian Zhang,Ruihua Qi,Xi Yang,Chuntao Li
机构: School of Artificial Intelligence, Jilin University; Engineering Research Center of Knowledge-Driven Human-Machine Intelligence, MoE, China; School of Archaeology, Jilin University
关键词: bronze ware dating, OOD detection methods, bronze ware, OOD detection, OOD
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 9pages, 7 figures, 4 tables

点击查看摘要

Abstract:In real-world applications across specialized domains, addressing complex out-of-distribution (OOD) challenges is a common and significant concern. In this study, we concentrate on the task of fine-grained bronze ware dating, a critical aspect in the study of ancient Chinese history, and developed a benchmark dataset named ShiftedBronzes. By extensively expanding the bronze Ding dataset, ShiftedBronzes incorporates two types of bronze ware data and seven types of OOD data, which exhibit distribution shifts commonly encountered in bronze ware dating scenarios. We conduct benchmarking experiments on ShiftedBronzes and five commonly used general OOD datasets, employing a variety of widely adopted post-hoc, pre-trained Vision Large Model (VLM)-based and generation-based OOD detection methods. Through analysis of the experimental results, we validate previous conclusions regarding post-hoc, VLM-based, and generation-based methods, while also highlighting their distinct behaviors on specialized datasets. These findings underscore the unique challenges of applying general OOD detection methods to domain-specific tasks such as bronze ware dating. We hope that the ShiftedBronzes benchmark provides valuable insights into both the field of bronze ware dating and the and the development of OOD detection methods. The dataset and associated code will be available later.
zh

[CV-78] ShotVL: Human-Centric Highlight Frame Retrieval via Language Queries

【速读】: 该论文试图解决在以人为中心的视频理解中,如何在帧级别上通过语言查询精确定位高光帧的问题。解决方案的关键在于提出了一个名为BestShot的新任务,并为此构建了BestShot Benchmark,该基准结合了人工标注的高光帧、详细的文本描述和时长标注,涵盖视觉内容、细粒度动作和人体姿态描述三个关键元素。为解决这一问题,论文收集了两个数据集:ShotGPT4o和Image-SMPLText,并基于这些数据集提出了一个强基线模型ShotVL,该模型在BestShot Benchmark上相较于InternVL提升了52%,在THUMOS14 Benchmark上提升了57%,同时保持了在通用图像分类和检索中的SOTA性能。

链接: https://arxiv.org/abs/2412.12675
作者: Wangyu Xue,Chen Qian,Jiayi Wu,Yang Zhou,Wentao Liu,Ju Ren,Siming Fan,Yaoxue Zhang
机构: Tsinghua University(清华大学); Alibaba Group(阿里巴巴集团)
关键词: analyzing specific moment, video understanding typically, understanding typically focus, human-centric video understanding, understanding typically
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Existing works on human-centric video understanding typically focus on analyzing specific moment or entire videos. However, many applications require higher precision at the frame level. In this work, we propose a novel task, BestShot, which aims to locate highlight frames within human-centric videos via language queries. This task demands not only a deep semantic comprehension of human actions but also precise temporal localization. To support this task, we introduce the BestShot Benchmark. %The benchmark is meticulously constructed by combining human detection and tracking, potential frame selection based on human judgment, and detailed textual descriptions crafted by human input to ensure precision. The benchmark is meticulously constructed by combining human-annotated highlight frames, detailed textual descriptions and duration labeling. These descriptions encompass three critical elements: (1) Visual content; (2) Fine-grained action; and (3) Human Pose Description. Together, these elements provide the necessary precision to identify the exact highlight frames in videos. To tackle this problem, we have collected two distinct datasets: (i) ShotGPT4o Dataset, which is algorithmically generated by GPT-4o and (ii) Image-SMPLText Dataset, a dataset with large-scale and accurate per-frame pose description leveraging PoseScript and existing pose estimation datasets. Based on these datasets, we present a strong baseline model, ShotVL, fine-tuned from InternVL, specifically for BestShot. We highlight the impressive zero-shot capabilities of our model and offer comparative analyses with existing SOTA models. ShotVL demonstrates a significant 52% improvement over InternVL on the BestShot Benchmark and a notable 57% improvement on the THUMOS14 Benchmark, all while maintaining the SOTA performance in general image classification and retrieval. Subjects: Computer Vision and Pattern Recognition (cs.CV) Cite as: arXiv:2412.12675 [cs.CV] (or arXiv:2412.12675v1 [cs.CV] for this version) https://doi.org/10.48550/arXiv.2412.12675 Focus to learn more arXiv-issued DOI via DataCite (pending registration)
zh

[CV-79] Structural Pruning via Spatial-aware Information Redundancy for Semantic Segmentation AAAI2025

【速读】: 该论文试图解决现有滤波器剪枝方法在语义分割任务中表现不佳的问题,这些方法通常忽略了分割任务对位置的敏感性。解决方案的关键在于提出了一种新的空间感知信息冗余滤波器剪枝方法 (Spatial-aware Information Redundancy Filter Pruning, SIRFP),通过将剪枝过程形式化为图论中的最大边权重团问题 (Maximum Edge Weight Clique Problem, MEWCP),并引入基于特征图的空间感知冗余度量,使得剪枝过程具备位置敏感性,从而更好地适应语义分割网络的剪枝需求。此外,论文还提出了一种低计算复杂度的贪心策略来解决这一NP难问题,确保了剪枝过程的高效性和可行性。

链接: https://arxiv.org/abs/2412.12672
作者: Dongyue Wu,Zilin Guo,Li Yu,Nong Sang,Changxin Gao
机构: 未知
关键词: recent years, pruning, segmentation networks, segmentation, filter pruning
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Accepted by AAAI 2025

点击查看摘要

Abstract:In recent years, semantic segmentation has flourished in various applications. However, the high computational cost remains a significant challenge that hinders its further adoption. The filter pruning method for structured network slimming offers a direct and effective solution for the reduction of segmentation networks. Nevertheless, we argue that most existing pruning methods, originally designed for image classification, overlook the fact that segmentation is a location-sensitive task, which consequently leads to their suboptimal performance when applied to segmentation networks. To address this issue, this paper proposes a novel approach, denoted as Spatial-aware Information Redundancy Filter Pruning~(SIRFP), which aims to reduce feature redundancy between channels. First, we formulate the pruning process as a maximum edge weight clique problem~(MEWCP) in graph theory, thereby minimizing the redundancy among the remaining features after pruning. Within this framework, we introduce a spatial-aware redundancy metric based on feature maps, thus endowing the pruning process with location sensitivity to better adapt to pruning segmentation networks. Additionally, based on the MEWCP, we propose a low computational complexity greedy strategy to solve this NP-hard problem, making it feasible and efficient for structured pruning. To validate the effectiveness of our method, we conducted extensive comparative experiments on various challenging datasets. The results demonstrate the superior performance of SIRFP for semantic segmentation tasks.
zh

[CV-80] Adaptive Prototype Replay for Class Incremental Semantic Segmentation AAAI-2025 AAAI

【速读】: 该论文试图解决类增量语义分割 (Class Incremental Semantic Segmentation, CISS) 中由于固定原型 (fixed prototypes) 与不断更新的类别表示之间的不匹配导致的灾难性遗忘问题。解决方案的关键在于提出了一种自适应原型重放 (Adaptive prototype replay, Adapter) 方法,该方法包括自适应偏差补偿策略 (Adaptive Deviation Compensation, ADC) 和不确定性感知约束损失 (Uncertainty-aware Constraint, UAC)。ADC 策略根据估计的表示偏移距离动态更新存储的原型,以匹配旧类别的更新表示;UAC 损失则通过减少预测不确定性来聚合判别特征,生成紧凑的原型。此外,论文还引入了基于补偿的原型相似性判别损失 (Compensation-based Prototype Similarity Discriminative, CPD),以增强相似原型之间的区分度,从而提升自适应原型重放策略的效率。

链接: https://arxiv.org/abs/2412.12669
作者: Guilin Zhu,Dongyue Wu,Changxin Gao,Runmin Wang,Weidong Yang,Nong Sang
机构: 未知
关键词: incremental semantic segmentation, Class incremental semantic, Adaptive prototype replay, prototype replay, semantic segmentation
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Accepted by the Main Technical Track of the 39th Annual AAAI Conference on Artificial Intelligence (AAAI-2025)

点击查看摘要

Abstract:Class incremental semantic segmentation (CISS) aims to segment new classes during continual steps while preventing the forgetting of old knowledge. Existing methods alleviate catastrophic forgetting by replaying distributions of previously learned classes using stored prototypes or features. However, they overlook a critical issue: in CISS, the representation of class knowledge is updated continuously through incremental learning, whereas prototype replay methods maintain fixed prototypes. This mismatch between updated representation and fixed prototypes limits the effectiveness of the prototype replay strategy. To address this issue, we propose the Adaptive prototype replay (Adapter) for CISS in this paper. Adapter comprises an adaptive deviation compen sation (ADC) strategy and an uncertainty-aware constraint (UAC) loss. Specifically, the ADC strategy dynamically updates the stored prototypes based on the estimated representation shift distance to match the updated representation of old class. The UAC loss reduces prediction uncertainty, aggregating discriminative features to aid in generating compact prototypes. Additionally, we introduce a compensation-based prototype similarity discriminative (CPD) loss to ensure adequate differentiation between similar prototypes, thereby enhancing the efficiency of the adaptive prototype replay strategy. Extensive experiments on Pascal VOC and ADE20K datasets demonstrate that Adapter achieves state-of-the-art results and proves effective across various CISS tasks, particularly in challenging multi-step scenarios. The code and model is available at this https URL.
zh

[CV-81] A Two-Fold Patch Selection Approach for Improved 360-Degree Image Quality Assessment

【速读】: 该论文旨在提高360度感知图像质量评估 (IQA) 的准确性,通过提出一种双阶段的图像块选择方法来解决这一问题。解决方案的关键在于结合视觉块选择与基于嵌入相似性的优化。第一阶段采用三种不同的采样方法从360度图像中选择图像块,以确保视觉内容的全面覆盖。第二阶段则是核心,利用基于嵌入相似性的选择过程,根据嵌入相似性距离过滤并优先选择最具信息量的图像块。这种双重选择机制不仅确保了训练数据的关联性和信息量,还显著提升了模型的学习效率,实验结果表明该方法在准确性和质量评分预测的单调性方面取得了高达4.5%的性能提升,同时仅使用了40%到50%的训练图像块。

链接: https://arxiv.org/abs/2412.12667
作者: Abderrezzaq Sendjasni,Seif-Eddine Benkabou,Mohamed-Chaker Larabi
机构: 未知
关键词: two-fold patch selection, perceptual image quality, image quality assessment, article presents, patch selection process
类目: Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注: Submitted to IEEE Transactions on Image Processing

点击查看摘要

Abstract:This article presents a novel approach to improving the accuracy of 360-degree perceptual image quality assessment (IQA) through a two-fold patch selection process. Our methodology combines visual patch selection with embedding similarity-based refinement. The first stage focuses on selecting patches from 360-degree images using three distinct sampling methods to ensure comprehensive coverage of visual content for IQA. The second stage, which is the core of our approach, employs an embedding similarity-based selection process to filter and prioritize the most informative patches based on their embeddings similarity distances. This dual selection mechanism ensures that the training data is both relevant and informative, enhancing the model’s learning efficiency. Extensive experiments and statistical analyses using three distance metrics across three benchmark datasets validate the effectiveness of our selection algorithm. The results highlight its potential to deliver robust and accurate 360-degree IQA, with performance gains of up to 4.5% in accuracy and monotonicity of quality score prediction, while using only 40% to 50% of the training patches. These improvements are consistent across various configurations and evaluation metrics, demonstrating the strength of the proposed method. The code for the selection process is available at: this https URL.
zh

[CV-82] SEG-SAM: Semantic-Guided SAM for Unified Medical Image Segmentation

【速读】: 该论文试图解决在医学图像分割领域中,直接应用Segment Anything Model (SAM)进行分割时面临的挑战,尤其是医学图像中显著的类别间重叠问题。解决方案的关键在于提出了SEmantic-Guided SAM (SEG-SAM),通过引入语义感知解码器(semantic-aware decoder)和文本到视觉语义模块(text-to-vision semantic module),将医学领域的语义知识整合到模型中,从而提升分割性能。具体来说,语义感知解码器独立于SAM的原始解码器,专门处理提示对象的语义分割和未提示对象的分类,而文本到视觉语义模块则通过从大型语言模型中提取医学类别的关键特征,并将其适配到视觉分割任务中,进一步增强了模型的语义理解能力。此外,交叉掩码空间对齐策略(cross-mask spatial alignment strategy)的引入,促进了两个解码器预测掩码之间的更大重叠,从而优化了分割结果。

链接: https://arxiv.org/abs/2412.12660
作者: Shuangping Huang,Hao Liang,Qingfeng Wang,Chulong Zhong,Zijian Zhou,Miaojing Shi
机构: South China University of Technology(华南理工大学); King’s College London(伦敦国王学院); Tongji University(同济大学)
关键词: gains increasing attention, models gains increasing, segmentation models gains, medical, developing unified medical
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 12 pages, 3 figures

点击查看摘要

Abstract:Recently, developing unified medical image segmentation models gains increasing attention, especially with the advent of the Segment Anything Model (SAM). SAM has shown promising binary segmentation performance in natural domains, however, transferring it to the medical domain remains challenging, as medical images often possess substantial inter-category overlaps. To address this, we propose the SEmantic-Guided SAM (SEG-SAM), a unified medical segmentation model that incorporates semantic medical knowledge to enhance medical segmentation performance. First, to avoid the potential conflict between binary and semantic predictions, we introduce a semantic-aware decoder independent of SAM’s original decoder, specialized for both semantic segmentation on the prompted object and classification on unprompted objects in images. To further enhance the model’s semantic understanding, we solicit key characteristics of medical categories from large language models and incorporate them into SEG-SAM through a text-to-vision semantic module, adaptively transferring the language information into the visual segmentation task. In the end, we introduce the cross-mask spatial alignment strategy to encourage greater overlap between the predicted masks from SEG-SAM’s two decoders, thereby benefiting both predictions. Extensive experiments demonstrate that SEG-SAM outperforms state-of-the-art SAM-based methods in unified binary medical segmentation and task-specific methods in semantic medical segmentation, showcasing promising results and potential for broader medical applications.
zh

[CV-83] CALA: A Class-Aware Logit Adapter for Few-Shot Class-Incremental Learning

【速读】: 该论文试图解决Few-Shot Class-Incremental Learning (FSCIL)中由于数据稀缺导致的模型对基础类别过拟合、对新类别忽视的问题,从而引发基础类别与新类别之间的严重混淆。解决方案的关键是提出Class-Aware Logit Adapter (CALA),通过一个轻量级的适配器在伪增量学习范式下学习纠正偏差的预测,并在实际FSCIL过程中动态生成鲁棒的平衡因子,根据新实例与基础类别的相似性调整其标签空间,从而有效缓解类别间的混淆。CALA在分类器层面操作,保留原始特征空间,因此可以灵活地集成到大多数现有的FSCIL方法中,提升性能。

链接: https://arxiv.org/abs/2412.12654
作者: Chengyan Liu,Linglan Zhao,Fan Lyu,Kaile Du,Fuyuan Hu,Tao Zhou
机构: School of electronic & information engineering, Suzhou University of Science and Technology; Tencent Youtu Lab; Chinese Academy of Sciences; Southeast University; North Minzu University
关键词: Few-Shot Class-Incremental Learning, Few-Shot Class-Incremental, defines a practical, practical but challenging, challenging task
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 10 pages

点击查看摘要

Abstract:Few-Shot Class-Incremental Learning (FSCIL) defines a practical but challenging task where models are required to continuously learn novel concepts with only a few training samples. Due to data scarcity, existing FSCIL methods resort to training a backbone with abundant base data and then keeping it frozen afterward. However, the above operation often causes the backbone to overfit to base classes while overlooking the novel ones, leading to severe confusion between them. To address this issue, we propose Class-Aware Logit Adapter (CALA). Our method involves a lightweight adapter that learns to rectify biased predictions through a pseudo-incremental learning paradigm. In the real FSCIL process, we use the learned adapter to dynamically generate robust balancing factors. These factors can adjust confused novel instances back to their true label space based on their similarity to base classes. Specifically, when confusion is more likely to occur in novel instances that closely resemble base classes, greater rectification is required. Notably, CALA operates on the classifier level, preserving the original feature space, thus it can be flexibly plugged into most of the existing FSCIL works for improved performance. Experiments on three benchmark datasets consistently validate the effectiveness and flexibility of CALA. Codes will be available upon acceptance.
zh

[CV-84] Dense Audio-Visual Event Localization under Cross-Modal Consistency and Multi-Temporal Granularity Collaboration AAAI2025

【速读】: 该论文试图解决密集音频-视觉事件定位 (Dense Audio-Visual Event Localization, DAVEL) 任务,特别是在长且未剪辑的视频中同时识别和时间定位多个重叠事件的挑战。解决方案的关键在于引入了一个名为 \ulCCNet 的新模型,该模型包含两个核心模块:跨模态一致性协作 (Cross-Modal Consistency Collaboration, CMCC) 和多时间粒度协作 (Multi-Temporal Granularity Collaboration, MTGC)。CMCC 模块通过跨模态交互分支和时间一致性门控分支,促进模态间一致事件语义的聚合,并引导模态间对关键事件相关时间区域的关注。MTGC 模块通过粗到细和细到粗的协作块,提供粗粒度和细粒度时间特征之间的双向支持。这些设计在 UnAV-100 数据集上的实验中验证了其有效性,并实现了密集音频-视觉事件定位任务的最新技术水平。

链接: https://arxiv.org/abs/2412.12628
作者: Ziheng Zhou,Jinxing Zhou,Wei Qian,Shengeng Tang,Xiaojun Chang,Dan Guo
机构: 未知
关键词: Dense Audio-Visual Event, Audio-Visual Event Localization, research tasks focus, tasks focus exclusively, exclusively on short
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Accepted by AAAI 2025. Project page: this https URL . Jinxing Zhou and Dan Guo are the corresponding authors

点击查看摘要

Abstract:In the field of audio-visual learning, most research tasks focus exclusively on short videos. This paper focuses on the more practical Dense Audio-Visual Event Localization (DAVEL) task, advancing audio-visual scene understanding for longer, untrimmed videos. This task seeks to identify and temporally pinpoint all events simultaneously occurring in both audio and visual streams. Typically, each video encompasses dense events of multiple classes, which may overlap on the timeline, each exhibiting varied durations. Given these challenges, effectively exploiting the audio-visual relations and the temporal features encoded at various granularities becomes crucial. To address these challenges, we introduce a novel \ulCCNet, comprising two core modules: the Cross-Modal Consistency \ulCollaboration (CMCC) and the Multi-Temporal Granularity \ulCollaboration (MTGC). Specifically, the CMCC module contains two branches: a cross-modal interaction branch and a temporal consistency-gated branch. The former branch facilitates the aggregation of consistent event semantics across modalities through the encoding of audio-visual relations, while the latter branch guides one modality’s focus to pivotal event-relevant temporal areas as discerned in the other modality. The MTGC module includes a coarse-to-fine collaboration block and a fine-to-coarse collaboration block, providing bidirectional support among coarse- and fine-grained temporal features. Extensive experiments on the UnAV-100 dataset validate our module design, resulting in a new state-of-the-art performance in dense audio-visual event localization. The code is available at \urlthis https URL.
zh

[CV-85] Improving the Transferability of 3D Point Cloud Attack via Spectral-aware Admix and Optimization Designs

【速读】: 该论文试图解决在黑盒设置下对3D点云模型进行对抗攻击时攻击成功率有限的问题。解决方案的关键在于提出了一种基于迁移的黑盒攻击方法,即首先在白盒代理模型中精心设计对抗样本,然后将其迁移到其他黑盒目标模型中进行攻击。具体而言,论文提出了一种新的频谱感知混合增强优化方法(Spectral-aware Admix with Augmented Optimization, SAAO),通过在频谱域中进行图傅里叶变换(Graph Fourier Transform, GFT)来获取点云的频谱特征,并在频谱域中进行混合操作,以避免在坐标域中直接混合破坏几何形状。随后,通过频谱感知加权混合选择更优的优化路径并调整相应的学习权重,最终生成对抗频谱特征并通过逆GFT将其转换回数据域,从而获得对抗样本。实验结果表明,SAAO方法相比现有的3D攻击方法具有更好的迁移性。

链接: https://arxiv.org/abs/2412.12626
作者: Shiyu Hu,Daizong Liu,Wei Hu
机构: Wangxuan Institute of Computer Technology, Peking University(王选计算机技术研究所,北京大学); Peking University(北京大学)
关键词: received increasing attention, Deep learning models, Deep learning, autonomous driving, received increasing
类目: Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR)
备注:

点击查看摘要

Abstract:Deep learning models for point clouds have shown to be vulnerable to adversarial attacks, which have received increasing attention in various safety-critical applications such as autonomous driving, robotics, and surveillance. Existing 3D attackers generally design various attack strategies in the white-box setting, requiring the prior knowledge of 3D model details. However, real-world 3D applications are in the black-box setting, where we can only acquire the outputs of the target classifier. Although few recent works try to explore the black-box attack, they still achieve limited attack success rates (ASR). To alleviate this issue, this paper focuses on attacking the 3D models in a transfer-based black-box setting, where we first carefully design adversarial examples in a white-box surrogate model and then transfer them to attack other black-box victim models. Specifically, we propose a novel Spectral-aware Admix with Augmented Optimization method (SAAO) to improve the adversarial transferability. In particular, since traditional Admix strategy are deployed in the 2D domain that adds pixel-wise images for perturbing, we can not directly follow it to merge point clouds in coordinate domain as it will destroy the geometric shapes. Therefore, we design spectral-aware fusion that performs Graph Fourier Transform (GFT) to get spectral features of the point clouds and add them in the spectral domain. Afterward, we run a few steps with spectral-aware weighted Admix to select better optimization paths as well as to adjust corresponding learning weights. At last, we run more steps to generate adversarial spectral feature along the optimization path and perform Inverse-GFT on the adversarial spectral feature to obtain the adversarial example in the data domain. Experiments show that our SAAO achieves better transferability compared to existing 3D attack methods.
zh

[CV-86] Multi-Domain Features Guided Supervised Contrastive Learning for Radar Target Detection

【速读】: 该论文试图解决在动态海洋环境下检测小目标的难题,特别是在复杂海杂波(sea clutter)中区分小目标的问题。解决方案的关键在于提出了一种多域特征引导的监督对比学习方法(MDFG_SCL),该方法通过结合从多域差异中提取的统计特征和通过监督对比学习获得的深度特征,实现了对低层次域特定变化和高层次语义信息的全面捕捉。这种综合特征集成使得模型能够在各种挑战性条件下有效区分小目标和海杂波,从而在实际应用中表现出优越的检测性能。

链接: https://arxiv.org/abs/2412.12620
作者: Junjie Wang,Yuze Gao,Dongying Li,Wenxian Yu
机构: Shanghai Key Laboratory of Intelligent Sensing and Recognition(上海智能感知与识别重点实验室); Shanghai Jiao Tong University(上海交通大学)
关键词: Detecting small targets, Detecting small, due to dynamic, sea clutter, Detecting
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Detecting small targets in sea clutter is challenging due to dynamic maritime conditions. Existing solutions either model sea clutter for detection or extract target features based on clutter-target echo differences, including statistical and deep features. While more common, the latter often excels in controlled scenarios but struggles with robust detection and generalization in diverse environments, limiting practical use. In this letter, we propose a multi-domain features guided supervised contrastive learning (MDFG_SCL) method, which integrates statistical features derived from multi-domain differences with deep features obtained through supervised contrastive learning, thereby capturing both low-level domain-specific variations and high-level semantic information. This comprehensive feature integration enables the model to effectively distinguish between small targets and sea clutter, even under challenging conditions. Experiments conducted on real-world datasets demonstrate that the proposed shallow-to-deep detector not only achieves effective identification of small maritime targets but also maintains superior detection performance across varying sea conditions, outperforming the mainstream unsupervised contrastive learning and supervised contrastive learning methods.
zh

[CV-87] PO3AD: Predicting Point Offsets toward Better 3D Point Cloud Anomaly Detection

【速读】: 该论文试图解决在无异常数据设置下,点云异常检测中准确捕捉3D正常数据特征以识别异常偏差的问题。解决方案的关键在于引入一种创新方法,强调学习点偏移量,重点关注更具信息量的伪异常点,从而更有效地提取正常数据特征。此外,论文还设计了一种基于法向量的数据增强技术,生成可信的伪异常数据,以提高训练过程的效率。通过在Anomaly-ShapeNet和Real3D-AD数据集上的实验验证,该方法在AUC-ROC检测指标上分别实现了9.0%和1.4%的平均提升,显著优于现有的最先进方法。

链接: https://arxiv.org/abs/2412.12617
作者: Jianan Ye,Weiguang Zhao,Xi Yang,Guangliang Cheng,Kaizhu Huang
机构: Xi’an Jiaotong-Liverpool University(西交利物浦大学); University of Liverpool(利物浦大学); Duke Kunshan University(昆山杜克大学)
关键词: anomaly-free setting poses, setting poses significant, requires accurately capturing, identify deviations indicative, poses significant challenges
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Point cloud anomaly detection under the anomaly-free setting poses significant challenges as it requires accurately capturing the features of 3D normal data to identify deviations indicative of anomalies. Current efforts focus on devising reconstruction tasks, such as acquiring normal data representations by restoring normal samples from altered, pseudo-anomalous counterparts. Our findings reveal that distributing attention equally across normal and pseudo-anomalous data tends to dilute the model’s focus on anomalous deviations. The challenge is further compounded by the inherently disordered and sparse nature of 3D point cloud data. In response to those predicaments, we introduce an innovative approach that emphasizes learning point offsets, targeting more informative pseudo-abnormal points, thus fostering more effective distillation of normal data representations. We also have crafted an augmentation technique that is steered by normal vectors, facilitating the creation of credible pseudo anomalies that enhance the efficiency of the training process. Our comprehensive experimental evaluation on the Anomaly-ShapeNet and Real3D-AD datasets evidences that our proposed method outperforms existing state-of-the-art approaches, achieving an average enhancement of 9.0% and 1.4% in the AUC-ROC detection metric across these datasets, respectively.
zh

[CV-88] RemoteTrimmer: Adaptive Structural Pruning for Remote Sensing Image Classification

【速读】: 该论文试图解决高分辨率遥感图像分类中模型计算复杂度高的问题,提出了一种有效的结构化剪枝方法。解决方案的关键在于引入了一种放大模型通道重要性差异的剪枝策略,并设计了一个自适应挖掘损失函数用于剪枝后模型的微调。实验结果表明,该方法在压缩遥感分类模型时实现了最小的精度损失,达到了当前最先进的性能。

链接: https://arxiv.org/abs/2412.12603
作者: Guanwenjie Zou,Liang Yao,Fan Liu,Chuanyi Zhang,Xin Li,Ning Chen,Shengxiang Xu,Jun Zhou
机构: Hohai University(河海大学); Griffith University(格里菲斯大学)
关键词: high computation complexity, high resolution remote, remote sensing, lightweight models tend, remote sensing image
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Since high resolution remote sensing image classification often requires a relatively high computation complexity, lightweight models tend to be practical and efficient. Model pruning is an effective method for model compression. However, existing methods rarely take into account the specificity of remote sensing images, resulting in significant accuracy loss after pruning. To this end, we propose an effective structural pruning approach for remote sensing image classification. Specifically, a pruning strategy that amplifies the differences in channel importance of the model is introduced. Then an adaptive mining loss function is designed for the fine-tuning process of the pruned model. Finally, we conducted experiments on two remote sensing classification datasets. The experimental results demonstrate that our method achieves minimal accuracy loss after compressing remote sensing classification models, achieving state-of-the-art (SoTA) performance.
zh

[CV-89] OpenViewer: Openness-Aware Multi-View Learning

【速读】: 该论文试图解决多视图学习方法在实际应用中的两个主要开放性挑战:1) 缺乏可解释性,即现有黑箱模型中多视图数据的整合机制难以解释;2) 泛化能力不足,即大多数模型无法适应涉及未知类别的多视图场景。解决方案的关键在于提出了一种名为OpenViewer的开放性感知多视图学习框架,该框架通过以下三个核心机制来应对这些挑战:1) 伪未知样本生成机制 (Pseudo-Unknown Sample Generation Mechanism),用于高效模拟开放的多视图环境并提前适应潜在的未知样本;2) 表达增强的深度展开网络 (Expression-Enhanced Deep Unfolding Network),通过系统构建功能性先验映射模块来直观提升模型的可解释性;3) 感知增强的开放集训练机制 (Perception-Augmented Open-Set Training Regime),通过精确提升已知类别的置信度并抑制未知类别的不适当置信度来显著增强模型的泛化能力。

链接: https://arxiv.org/abs/2412.12596
作者: Shide Du,Zihan Fang,Yanchao Tan,Changwei Wang,Shiping Wang,Wenzhong Guo
机构: 1. School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China(哈尔滨工业大学计算机科学与技术学院,哈尔滨,中国);
2. Key Laboratory of Network Security Research, Harbin Institute of Technology, Harbin, China(网络与信息安全重点实验室,哈尔滨工业大学,哈尔滨,中国);
3. School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen, China(哈尔滨工业大学深圳校区计算机科学与技术学院,深圳,中国)
关键词: methods leverage multiple, learning methods leverage, leverage multiple data, multiple data sources, correlations across views
类目: Computer Vision and Pattern Recognition (cs.CV); Applications (stat.AP); Machine Learning (stat.ML)
备注: 16 pages

点击查看摘要

Abstract:Multi-view learning methods leverage multiple data sources to enhance perception by mining correlations across views, typically relying on predefined categories. However, deploying these models in real-world scenarios presents two primary openness challenges. 1) Lack of Interpretability: The integration mechanisms of multi-view data in existing black-box models remain poorly explained; 2) Insufficient Generalization: Most models are not adapted to multi-view scenarios involving unknown categories. To address these challenges, we propose OpenViewer, an openness-aware multi-view learning framework with theoretical support. This framework begins with a Pseudo-Unknown Sample Generation Mechanism to efficiently simulate open multi-view environments and previously adapt to potential unknown samples. Subsequently, we introduce an Expression-Enhanced Deep Unfolding Network to intuitively promote interpretability by systematically constructing functional prior-mapping modules and effectively providing a more transparent integration mechanism for multi-view data. Additionally, we establish a Perception-Augmented Open-Set Training Regime to significantly enhance generalization by precisely boosting confidences for known categories and carefully suppressing inappropriate confidences for unknown ones. Experimental results demonstrate that OpenViewer effectively addresses openness challenges while ensuring recognition performance for both known and unknown samples. The code is released at this https URL.
zh

[CV-90] A Simple and Efficient Baseline for Zero-Shot Generative Classification

【速读】: 该论文试图解决现有零样本扩散模型分类器在分类速度上的严重瓶颈问题,特别是在ImageNet数据集上单张图像分类耗时长达1000秒的问题。解决方案的关键在于提出了一个简单且高效的零样本高斯扩散分类器 (Gaussian Diffusion Classifiers, GDC),通过预训练的文本到图像扩散模型和DINOv2实现。GDC不仅在ImageNet上将分类准确率提升了10个百分点(从61.40%提升至71.44%),还将单张图像的分类速度提升了超过30000倍(从1000秒降至0.03秒),并提供了结果的概率解释。该方法首次实现了零样本扩散模型分类器在准确性和实用性上的双重突破。

链接: https://arxiv.org/abs/2412.12594
作者: Zipeng Qi,Buhua Liu,Shiyan Zhang,Bao Li,Zhiqiang Xu,Haoyi Xiong,Zeke Xie
机构: xLeaF Lab, The Hong Kong University of Science and Technology (Guangzhou)(香港科技大学(广州)xLeaF实验室); Tsinghua University (清华大学); Institute of Automation, Chinese Academy of Sciences (中国科学院自动化研究所); Mohamed bin Zayed University of Artificial Intelligence (穆罕默德·本·扎耶德人工智能大学); Baidu (百度)
关键词: industrial AIGC applications, industrial AIGC, Large diffusion models, AIGC applications, mainstream generative models
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Large diffusion models have become mainstream generative models in both academic studies and industrial AIGC applications. Recently, a number of works further explored how to employ the power of large diffusion models as zero-shot classifiers. While recent zero-shot diffusion-based classifiers have made performance advancement on benchmark datasets, they still suffered badly from extremely slow classification speed (e.g., ~1000 seconds per classifying single image on ImageNet). The extremely slow classification speed strongly prohibits existing zero-shot diffusion-based classifiers from practical applications. In this paper, we propose an embarrassingly simple and efficient zero-shot Gaussian Diffusion Classifiers (GDC) via pretrained text-to-image diffusion models and DINOv2. The proposed GDC can not only significantly surpass previous zero-shot diffusion-based classifiers by over 10 points (61.40% - 71.44%) on ImageNet, but also accelerate more than 30000 times (1000 - 0.03 seconds) classifying a single image on ImageNet. Additionally, it provides probability interpretation of the results. Our extensive experiments further demonstrate that GDC can achieve highly competitive zero-shot classification performance over various datasets and can promisingly self-improve with stronger diffusion models. To the best of our knowledge, the proposed GDC is the first zero-shot diffusionbased classifier that exhibits both competitive accuracy and practical efficiency.
zh

[CV-91] License Plate Detection and Character Recognition Using Deep Learning and Font Evaluation

【速读】: 该论文旨在解决车牌检测 (License Plate Detection, LPD) 中的挑战,特别是由于光照变化和字体多样性导致的检测和识别精度问题。解决方案的关键在于采用双深度学习策略:使用Faster R-CNN进行车牌检测,并结合CNN-RNN模型与Connectionist Temporal Classification (CTC) 损失函数及MobileNet V3骨干网络进行车牌识别。该方法通过多地区数据集(如安大略、魁北克、加利福尼亚和新 York 州)的训练,显著提升了模型性能,在CENPARMI和UFPR-ALPR数据集上分别达到了92%和90%的召回率。此外,论文还深入分析了字体特征对车牌识别的影响,发现不同字体(如Driver Gothic、Dreadnought等)对识别性能有显著差异,为未来LPD系统的改进提供了重要见解。

链接: https://arxiv.org/abs/2412.12572
作者: Zahra Ebrahimi Vargoorani,Ching Yee Suen
机构: 未知
关键词: diverse font types, vehicle tracking, impacting accuracy, traffic management, Connectionist Temporal Classification
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: 12 pages, 5 figures. This is the pre-Springer final accepted version. The final version is published in Springer, Lecture Notes in Computer Science (LNCS), Volume 14731, 2024. Springer Version of Record

点击查看摘要

Abstract:License plate detection (LPD) is essential for traffic management, vehicle tracking, and law enforcement but faces challenges like variable lighting and diverse font types, impacting accuracy. Traditionally reliant on image processing and machine learning, the field is now shifting towards deep learning for its robust performance in various conditions. Current methods, however, often require tailoring to specific regional datasets. This paper proposes a dual deep learning strategy using a Faster R-CNN for detection and a CNN-RNN model with Connectionist Temporal Classification (CTC) loss and a MobileNet V3 backbone for recognition. This approach aims to improve model performance using datasets from Ontario, Quebec, California, and New York State, achieving a recall rate of 92% on the Centre for Pattern Recognition and Machine Intelligence (CENPARMI) dataset and 90% on the UFPR-ALPR dataset. It includes a detailed error analysis to identify the causes of false positives. Additionally, the research examines the role of font features in license plate (LP) recognition, analyzing fonts like Driver Gothic, Dreadnought, California Clarendon, and Zurich Extra Condensed with the OpenALPR system. It discovers significant performance discrepancies influenced by font characteristics, offering insights for future LPD system enhancements. Keywords: Deep Learning, License Plate, Font Evaluation Comments: 12 pages, 5 figures. This is the pre-Springer final accepted version. The final version is published in Springer, Lecture Notes in Computer Science (LNCS), Volume 14731, 2024. Springer Version of Record Subjects: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) MSC classes: 68T10 ACMclasses: I.2.10; I.4.8; I.5.4 Cite as: arXiv:2412.12572 [cs.CV] (or arXiv:2412.12572v1 [cs.CV] for this version) https://doi.org/10.48550/arXiv.2412.12572 Focus to learn more arXiv-issued DOI via DataCite (pending registration) Journalreference: Springer, Lecture Notes in Computer Science (LNCS), Volume 14731, 2024 Related DOI: https://doi.org/10.1007/978-3-031-71602-7_20 Focus to learn more DOI(s) linking to related resources
zh

[CV-92] ChatDiT: A Training-Free Baseline for Task-Agnostic Free-Form Chatting with Diffusion Transformers

【速读】: 该论文试图解决的问题是如何利用预训练的扩散变换器(pretrained diffusion transformers, DiTs)在无需额外训练或架构修改的情况下,实现零样本、通用且交互式的视觉生成任务。解决方案的关键在于提出了ChatDiT框架,该框架通过多代理系统实现,包括指令解析代理(Instruction-Parsing agent)、策略规划代理(Strategy-Planning agent)和执行代理(Execution agent)。这些代理协同工作,利用预训练的DiTs的上下文生成能力,通过自由形式的自然语言交互,完成如文本图像交错生成、多页图画书创作、图像编辑、IP衍生设计及角色设计设置等任务。ChatDiT在无需额外调优或适配器的情况下,展示了其在多任务环境中的卓越性能,超越了专门设计和训练的多任务数据集上的竞争对手。

链接: https://arxiv.org/abs/2412.12571
作者: Lianghua Huang,Wei Wang,Zhi-Fan Wu,Yupeng Shi,Chen Liang,Tong Shen,Han Zhang,Huanzhang Dou,Yu Liu,Jingren Zhou
机构: Tongyi Lab; Alibaba-inc.com(阿里巴巴); Institute of Automation, Chinese Academy of Sciences(中国科学院自动化研究所); Shanghai Jiao Tong University(上海交通大学); Zhejiang University(浙江大学)
关键词: Recent research arXiv, pretrained diffusion transformers, Recent research, highlighted the inherent, seamlessly adapt
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Tech report. Project page: this https URL

点击查看摘要

Abstract:Recent research arXiv:2410.15027 arXiv:2410.23775 has highlighted the inherent in-context generation capabilities of pretrained diffusion transformers (DiTs), enabling them to seamlessly adapt to diverse visual tasks with minimal or no architectural modifications. These capabilities are unlocked by concatenating self-attention tokens across multiple input and target images, combined with grouped and masked generation pipelines. Building upon this foundation, we present ChatDiT, a zero-shot, general-purpose, and interactive visual generation framework that leverages pretrained diffusion transformers in their original form, requiring no additional tuning, adapters, or modifications. Users can interact with ChatDiT to create interleaved text-image articles, multi-page picture books, edit images, design IP derivatives, or develop character design settings, all through free-form natural language across one or more conversational rounds. At its core, ChatDiT employs a multi-agent system comprising three key components: an Instruction-Parsing agent that interprets user-uploaded images and instructions, a Strategy-Planning agent that devises single-step or multi-step generation actions, and an Execution agent that performs these actions using an in-context toolkit of diffusion transformers. We thoroughly evaluate ChatDiT on IDEA-Bench arXiv:2412.11767, comprising 100 real-world design tasks and 275 cases with diverse instructions and varying numbers of input and target images. Despite its simplicity and training-free approach, ChatDiT surpasses all competitors, including those specifically designed and trained on extensive multi-task datasets. We further identify key limitations of pretrained DiTs in zero-shot adapting to tasks. We release all code, agents, results, and intermediate outputs to facilitate further research at this https URL
zh

[CV-93] ITP: Instance-Aware Test Pruning for Out-of-Distribution Detection

【速读】: 该论文试图解决深度模型在分布外(OOD)检测中的过自信问题,特别是在利用权重稀疏化技术进行参数选择时,所选参数可能对OOD数据表现出过度自信,从而影响OOD检测的准确性。解决方案的关键是提出了一种后处理方法——实例感知测试剪枝(Instance-aware Test Pruning, ITP),通过结合粗粒度和细粒度的参数剪枝策略来提升OOD检测的可靠性。具体来说,ITP首先通过探索分布内(ID)数据估计类特定的参数贡献分布,进行粗粒度剪枝以消除冗余参数;然后采用基于右尾Z分数检验的细粒度测试剪枝,自适应地移除实例级别的过度自信参数。最终,通过剪枝后的模型生成OOD分数,实现更可靠的预测。

链接: https://arxiv.org/abs/2412.12566
作者: Haonan Xu,Yang Yang
机构: 未知
关键词: OOD detection, OOD, real-world scenarios, crucial for ensuring, deployment of deep
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Out-of-distribution (OOD) detection is crucial for ensuring the reliable deployment of deep models in real-world scenarios. Recently, from the perspective of over-parameterization, a series of methods leveraging weight sparsification techniques have shown promising performance. These methods typically focus on selecting important parameters for in-distribution (ID) data to reduce the negative impact of redundant parameters on OOD detection. However, we empirically find that these selected parameters may behave overconfidently toward OOD data and hurt OOD detection. To address this issue, we propose a simple yet effective post-hoc method called Instance-aware Test Pruning (ITP), which performs OOD detection by considering both coarse-grained and fine-grained levels of parameter pruning. Specifically, ITP first estimates the class-specific parameter contribution distribution by exploring the ID data. By using the contribution distribution, ITP conducts coarse-grained pruning to eliminate redundant parameters. More importantly, ITP further adopts a fine-grained test pruning process based on the right-tailed Z-score test, which can adaptively remove instance-level overconfident parameters. Finally, ITP derives OOD scores from the pruned model to achieve more reliable predictions. Extensive experiments on widely adopted benchmarks verify the effectiveness of ITP, demonstrating its competitive performance.
zh

[CV-94] PBVS 2024 Solution: Self-Supervised Learning and Sampling Strategies for SAR Classification in Extreme Long-Tail Distribution

【速读】: 该论文试图解决在自动目标识别 (ATR) 系统中,利用合成孔径雷达 (SAR) 和光电 (EO) 数据进行多模态学习时面临的挑战,特别是数据集的长尾分布和SAR与EO数据之间的领域差异问题。解决方案的关键在于提出了一种两阶段学习方法,首先通过自监督技术进行预训练,然后通过SAR到EO的翻译实现多模态学习和推理,从而有效利用EO数据,提升分类性能。

链接: https://arxiv.org/abs/2412.12565
作者: Yuhyun Kim,Minwoo Kim,Hyobin Park,Jinwook Jung,Dong-Geol Choi
机构: Hanbat National University (韩巴国立大学)
关键词: Synthetic Aperture Radar, automatic target recognition, Aperture Radar, Synthetic Aperture, Multimodal Learning Workshop
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 4 pages, 3 figures, 1 Table

点击查看摘要

Abstract:The Multimodal Learning Workshop (PBVS 2024) aims to improve the performance of automatic target recognition (ATR) systems by leveraging both Synthetic Aperture Radar (SAR) data, which is difficult to interpret but remains unaffected by weather conditions and visible light, and Electro-Optical (EO) data for simultaneous learning. The subtask, known as the Multi-modal Aerial View Imagery Challenge - Classification, focuses on predicting the class label of a low-resolution aerial image based on a set of SAR-EO image pairs and their respective class labels. The provided dataset consists of SAR-EO pairs, characterized by a severe long-tail distribution with over a 1000-fold difference between the largest and smallest classes, making typical long-tail methods difficult to apply. Additionally, the domain disparity between the SAR and EO datasets complicates the effectiveness of standard multimodal methods. To address these significant challenges, we propose a two-stage learning approach that utilizes self-supervised techniques, combined with multimodal learning and inference through SAR-to-EO translation for effective EO utilization. In the final testing phase of the PBVS 2024 Multi-modal Aerial View Image Challenge - Classification (SAR Classification) task, our model achieved an accuracy of 21.45%, an AUC of 0.56, and a total score of 0.30, placing us 9th in the competition.
zh

[CV-95] Efficient Oriented Object Detection with Enhanced Small Object Recognition in Aerial Images

【速读】: 该论文试图解决在遥感图像中旋转边界框目标检测的计算效率与检测精度之间的平衡问题,特别是在处理小尺寸和多尺度目标时。解决方案的关键在于对YOLOv8模型进行了创新性增强,引入了基于小波变换的C2f模块(wavelet transform-based C2f module)用于捕捉关联特征,并结合自适应尺度特征金字塔模块(Adaptive Scale Feature Pyramid, ASFP)利用P2层细节信息。此外,GhostDynamicConv的引入显著降低了模型的复杂度,使其在保持高检测精度的同时,参数数量减少至21.6M,优于DecoupleNet的23.3M参数。这些改进使得模型在资源受限的环境中仍能高效运行,并在DOTAv1.0数据集上展示了与领先方法相当的平均精度(mAP)。

链接: https://arxiv.org/abs/2412.12562
作者: Zhifei Shi,Zongyao Yin,Sheng Chang,Xiao Yi,Xianchuan Yu
机构: School of Artificial Intelligence, Beijing Normal University (人工智能学院,北京师范大学)
关键词: rotated bounding box, Achieving a balance, bounding box object, box object detection, realm of rotated
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Achieving a balance between computational efficiency and detection accuracy in the realm of rotated bounding box object detection within aerial imagery is a significant challenge. While prior research has aimed at creating lightweight models that enhance computational performance and feature extraction, there remains a gap in the performance of these networks when it comes to the detection of small and multi-scale objects in remote sensing (RS) imagery. To address these challenges, we present a novel enhancement to the YOLOv8 model, tailored for oriented object detection tasks and optimized for environments with limited computational resources. Our model features a wavelet transform-based C2f module for capturing associative features and an Adaptive Scale Feature Pyramid (ASFP) module that leverages P2 layer details. Additionally, the incorporation of GhostDynamicConv significantly contributes to the model’s lightweight nature, ensuring high efficiency in aerial imagery analysis. Featuring a parameter count of 21.6M, our approach provides a more efficient architectural design than DecoupleNet, which has 23.3M parameters, all while maintaining detection accuracy. On the DOTAv1.0 dataset, our model demonstrates a mean Average Precision (mAP) that is competitive with leading methods such as DecoupleNet. The model’s efficiency, combined with its reduced parameter count, makes it a strong candidate for aerial object detection, particularly in resource-constrained environments.
zh

[CV-96] Me What to Track: Infusing Robust Language Guidance for Enhanced Referring Multi-Object Tracking

【速读】: 该论文试图解决在指代多目标跟踪 (Referring Multi-Object Tracking, RMOT) 任务中,由于新生目标与现有目标数据分布不平衡导致的检测能力不足问题,以及多模态特征间接融合难以有效指导新生目标检测的挑战。解决方案的关键在于采用协同匹配策略来缓解数据不平衡的影响,提升新生目标检测能力,同时保持跟踪性能。此外,通过在编码器中集成并增强跨模态和多尺度特征融合,克服了先前工作中多模态信息交互不足的瓶颈。在解码器中,引入指代引导的适应机制,通过查询令牌提供明确的指代指导。实验结果表明,该模型相较于先前工作提升了3.42%的性能,验证了其设计的有效性。

链接: https://arxiv.org/abs/2412.12561
作者: Wenjun Huang,Yang Ni,Hanning Chen,Yirui He,Ian Bryant,Yezi Liu,Mohsen Imani
机构: University of California, Irvine, CA, USA(加州大学欧文分校)
关键词: emerging cross-modal task, Referring multi-object tracking, aims to localize, localize an arbitrary, arbitrary number
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Referring multi-object tracking (RMOT) is an emerging cross-modal task that aims to localize an arbitrary number of targets based on a language expression and continuously track them in a video. This intricate task involves reasoning on multi-modal data and precise target localization with temporal association. However, prior studies overlook the imbalanced data distribution between newborn targets and existing targets due to the nature of the task. In addition, they only indirectly fuse multi-modal features, struggling to deliver clear guidance on newborn target detection. To solve the above issues, we conduct a collaborative matching strategy to alleviate the impact of the imbalance, boosting the ability to detect newborn targets while maintaining tracking performance. In the encoder, we integrate and enhance the cross-modal and multi-scale fusion, overcoming the bottlenecks in previous work, where limited multi-modal information is shared and interacted between feature maps. In the decoder, we also develop a referring-infused adaptation that provides explicit referring guidance through the query tokens. The experiments showcase the superior performance of our model (+3.42%) compared to prior works, demonstrating the effectiveness of our designs.
zh

[CV-97] SAModified: A Foundation Model-Based Zero-Shot Approach for Refining Noisy Land-Use Land-Cover Maps

【速读】: 该论文试图解决土地利用和土地覆盖 (LULC) 分析中由于标签噪声导致的自动化地图生成难题。解决方案的关键在于采用零样本学习方法,利用基础模型 Segment Anything Model (SAM) 自动划分不同地块/区域,并通过检测区域内局部标签统计数据对不确定像素进行重新标注。这种方法显著减少了标签噪声,并使下游分割模型的性能提升了约 5%。

链接: https://arxiv.org/abs/2412.12552
作者: Sparsh Pekhale,Rakshith Sathish,Sathisha Basavaraju,Divya Sharma
机构: 未知
关键词: analysis is critical, remote sensing, urban planning, critical in remote, wide-ranging applications
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Land-use and land cover (LULC) analysis is critical in remote sensing, with wide-ranging applications across diverse fields such as agriculture, utilities, and urban planning. However, automating LULC map generation using machine learning is rendered challenging due to noisy labels. Typically, the ground truths (e.g. ESRI LULC, MapBioMass) have noisy labels that hamper the model’s ability to learn to accurately classify the pixels. Further, these erroneous labels can significantly distort the performance metrics of a model, leading to misleading evaluations. Traditionally, the ambiguous labels are rectified using unsupervised algorithms. These algorithms struggle not only with scalability but also with generalization across different geographies. To overcome these challenges, we propose a zero-shot approach using the foundation model, Segment Anything Model (SAM), to automatically delineate different land parcels/regions and leverage them to relabel the unsure pixels by using the local label statistics within each detected region. We achieve a significant reduction in label noise and an improvement in the performance of the downstream segmentation model by \approx 5% when trained with denoised labels.
zh

[CV-98] Consistent Diffusion: Denoising Diffusion Model with Data-Consistent Training for Image Restoration

【速读】: 该论文旨在解决去噪扩散模型 (Denoising Diffusion Models, DDMs) 在图像恢复任务中存在的形状和颜色失真问题。研究发现,这些问题源于训练和测试数据之间的不一致性。为此,论文提出了一种名为“数据一致性训练” (data-consistent training) 的新训练方法,关键在于让模型在训练过程中接触到带有累积误差的图像,从而使模型学会纠正这些误差。实验结果表明,该方法在五个图像恢复任务中显著优于现有最先进的方法,同时有效减少了失真并保持了图像的保真度。

链接: https://arxiv.org/abs/2412.12550
作者: Xinlong Cheng,Tiantian Cao,Guoan Cheng,Bangxuan Huang,Xinghan Tian,Ye Wang,Xiaoyu He,Weixin Li,Tianfan Xue,Xuan Dong
机构: 未知
关键词: compromise image quality, denoising diffusion models, image restoration tasks, shape and color, address the limitations
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:In this work, we address the limitations of denoising diffusion models (DDMs) in image restoration tasks, particularly the shape and color distortions that can compromise image quality. While DDMs have demonstrated a promising performance in many applications such as text-to-image synthesis, their effectiveness in image restoration is often hindered by shape and color distortions. We observe that these issues arise from inconsistencies between the training and testing data used by DDMs. Based on our observation, we propose a novel training method, named data-consistent training, which allows the DDMs to access images with accumulated errors during training, thereby ensuring the model to learn to correct these errors. Experimental results show that, across five image restoration tasks, our method has significant improvements over state-of-the-art methods while effectively minimizing distortions and preserving image fidelity.
zh

[CV-99] Addressing Small and Imbalanced Medical Image Datasets Using Generative Models: A Comparative Study of DDPM and PGGANs with Random and Greedy K Sampling

【速读】: 该论文旨在解决医疗图像分类模型在数据稀缺和类别不平衡情况下性能受限的问题。解决方案的关键在于利用生成模型(如Denoising Diffusion Probabilistic Models (DDPM) 和Progressive Growing Generative Adversarial Networks (PGGANs))进行数据增强。通过评估DDPM和PGGANs生成的合成图像对分类模型性能的影响,研究发现DDPM生成的图像在Frechet Inception Distance (FID) 指标上表现更优,显著提升了分类精度(最高达6%),尤其是在处理不平衡数据集时,增强了模型的鲁棒性和稳定性。

链接: https://arxiv.org/abs/2412.12532
作者: Iman Khazrak,Shakhnoza Takhirova,Mostafa M. Rezaee,Mehrdad Yadollahi,Robert C. Green II,Shuteng Niu
机构: Bowling Green State University(鲍灵格林州立大学)
关键词: Denoising Diffusion Probabilistic, Generative Adversarial Networks, Growing Generative Adversarial, Progressive Growing Generative, Diffusion Probabilistic Models
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:The development of accurate medical image classification models is often constrained by privacy concerns and data scarcity for certain conditions, leading to small and imbalanced datasets. To address these limitations, this study explores the use of generative models, such as Denoising Diffusion Probabilistic Models (DDPM) and Progressive Growing Generative Adversarial Networks (PGGANs), for dataset augmentation. The research introduces a framework to assess the impact of synthetic images generated by DDPM and PGGANs on the performance of four models: a custom CNN, Untrained VGG16, Pretrained VGG16, and Pretrained ResNet50. Experiments were conducted using Random Sampling and Greedy K Sampling to create small, imbalanced datasets. The synthetic images were evaluated using Frechet Inception Distance (FID) and compared to original datasets through classification metrics. The results show that DDPM consistently generated more realistic images with lower FID scores and significantly outperformed PGGANs in improving classification metrics across all models and datasets. Incorporating DDPM-generated images into the original datasets increased accuracy by up to 6%, enhancing model robustness and stability, particularly in imbalanced scenarios. Random Sampling demonstrated superior stability, while Greedy K Sampling offered diversity at the cost of higher FID scores. This study highlights the efficacy of DDPM in augmenting small, imbalanced medical image datasets, improving model performance by balancing the dataset and expanding its size.
zh

[CV-100] CREST: An Efficient Conjointly-trained Spike-driven Framework for Event-based Object Detection Exploiting Spatiotemporal Dynamics AAAI2025

【速读】: 该论文试图解决事件相机(event-based cameras)在高速和低光环境下进行物体检测时,传统脉冲神经网络(Spiking Neural Networks, SNNs)面临的训练效率低、梯度消失和高计算复杂度问题,尤其是深度SNNs中的这些问题。此外,现有SNN框架在处理多尺度时空特征时存在数据冗余和精度下降的问题。解决方案的关键在于提出了一种名为CREST的新型联合训练脉冲驱动框架,通过引入联合学习规则(conjoint learning rule)来加速SNN学习并缓解梯度消失问题。CREST框架还支持双操作模式,以适应不同硬件的高效灵活实现,并集成了多尺度时空事件积分器(MESTOR)和时空-IoU(ST-IoU)损失函数,从而显著提升了物体识别和检测性能,并实现了比现有SNN算法高达100倍的能效提升。

链接: https://arxiv.org/abs/2412.12525
作者: Ruixin Mao,Aoyu Shen,Lin Tang,Jun Zhou
机构: 未知
关键词: low power consumption, high temporal resolution, wide dynamic range, event-based object detection, event-based object
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Event-based cameras feature high temporal resolution, wide dynamic range, and low power consumption, which is ideal for high-speed and low-light object detection. Spiking neural networks (SNNs) are promising for event-based object recognition and detection due to their spiking nature but lack efficient training methods, leading to gradient vanishing and high computational complexity, especially in deep SNNs. Additionally, existing SNN frameworks often fail to effectively handle multi-scale spatiotemporal features, leading to increased data redundancy and reduced accuracy. To address these issues, we propose CREST, a novel conjointly-trained spike-driven framework to exploit spatiotemporal dynamics in event-based object detection. We introduce the conjoint learning rule to accelerate SNN learning and alleviate gradient vanishing. It also supports dual operation modes for efficient and flexible implementation on different hardware types. Additionally, CREST features a fully spike-driven framework with a multi-scale spatiotemporal event integrator (MESTOR) and a spatiotemporal-IoU (ST-IoU) loss. Our approach achieves superior object recognition detection performance and up to 100X energy efficiency compared with state-of-the-art SNN algorithms on three datasets, providing an efficient solution for event-based object detection algorithms suitable for SNN hardware implementation.
zh

[CV-101] Invisible Watermarks: Attacks and Robustness ATC WWW

【速读】: 该论文旨在解决生成式 AI (Generative AI) 生成图像的检测问题,以应对日益严重的虚假信息传播。其关键解决方案包括:1) 在单张图像上同时应用图像空间 (image-space) 和潜在空间 (latent-space) 水印方法,并提出一种自定义的水印移除网络,该网络在解码时保留一种水印模式的同时完全移除另一种;2) 研究基于 GradCAM 热图的局部模糊攻击 (Localized Blurring Attacks, LBA),以减少对目标图像的降质影响。评估结果表明,水印移除模型在保留一种水印模式时略微提升了基线性能,而 LBA 相较于全图均匀模糊显著减少了图像降质。

链接: https://arxiv.org/abs/2412.12511
作者: Dongjun Hwang,Sungwon Woo,Tom Gao,Raymond Luo,Sunghwan Baek
机构: Carnegie Mellon University (卡内基梅隆大学)
关键词: Generative AI continues, combat misinformation, misinformation is stronger, detection of generated, robust detection
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: YouTube link for the presentation: this https URL

点击查看摘要

Abstract:As Generative AI continues to become more accessible, the case for robust detection of generated images in order to combat misinformation is stronger than ever. Invisible watermarking methods act as identifiers of generated content, embedding image- and latent-space messages that are robust to many forms of perturbations. The majority of current research investigates full-image attacks against images with a single watermarking method applied. We introduce novel improvements to watermarking robustness as well as minimizing degradation on image quality during attack. Firstly, we examine the application of both image-space and latent-space watermarking methods on a single image, where we propose a custom watermark remover network which preserves one of the watermarking modalities while completely removing the other during decoding. Then, we investigate localized blurring attacks (LBA) on watermarked images based on the GradCAM heatmap acquired from the watermark decoder in order to reduce the amount of degradation to the target image. Our evaluation suggests that 1) implementing the watermark remover model to preserve one of the watermark modalities when decoding the other modality slightly improves on the baseline performance, and that 2) LBA degrades the image significantly less compared to uniform blurring of the entire image. Code is available at: this https URL
zh

[CV-102] 3DGUT: Enabling Distorted Cameras and Secondary Rays in Gaussian Splatting

【速读】: 该论文试图解决3D Gaussian Splatting (3DGS)在处理复杂场景时,由于基于光栅化的方法限制,无法支持非理想针孔相机和次级光照效果的问题。解决方案的关键在于提出3D Gaussian Unscented Transform (3DGUT),通过使用Unscented Transform替代3DGS中的EWA splatting方法,利用sigma点精确近似粒子,从而在任意非线性投影函数下实现精确投影。这一改进不仅支持畸变相机和时间相关效应(如卷帘快门),还保留了光栅化的效率。此外,该方法与基于追踪的方法对齐,能够在同一3D表示中实现次级光线追踪,以表现反射和折射等现象。

链接: https://arxiv.org/abs/2412.12507
作者: Qi Wu,Janick Martinez Esturo,Ashkan Mirzaei,Nicolas Moenne-Loccoz,Zan Gojcic
机构: NVIDIA; University of Toronto
关键词: shown great potential, high-fidelity real-time rendering, Gaussian Unscented Transform, consumer hardware, shown great
类目: Graphics (cs.GR); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:3D Gaussian Splatting (3DGS) has shown great potential for efficient reconstruction and high-fidelity real-time rendering of complex scenes on consumer hardware. However, due to its rasterization-based formulation, 3DGS is constrained to ideal pinhole cameras and lacks support for secondary lighting effects. Recent methods address these limitations by tracing volumetric particles instead, however, this comes at the cost of significantly slower rendering speeds. In this work, we propose 3D Gaussian Unscented Transform (3DGUT), replacing the EWA splatting formulation in 3DGS with the Unscented Transform that approximates the particles through sigma points, which can be projected exactly under any nonlinear projection function. This modification enables trivial support of distorted cameras with time dependent effects such as rolling shutter, while retaining the efficiency of rasterization. Additionally, we align our rendering formulation with that of tracing-based methods, enabling secondary ray tracing required to represent phenomena such as reflections and refraction within the same 3D representation.
zh

[CV-103] Multi-Scale Cross-Fusion and Edge-Supervision Network for Image Splicing Localization

【速读】: 该论文试图解决图像拼接定位 (Image Splicing Localization, ISL) 中边缘信息利用不足导致的完整性差和高误报率问题。解决方案的关键在于提出了一个多尺度交叉融合与边缘监督网络,具体包括三个步骤:首先,通过分割网络学习RGB图像及其噪声图像的多尺度特征,并通过跨尺度融合和跨域融合增强特征表示;其次,设计边缘掩码预测模块以有效挖掘可靠的边界伪影;最后,通过注意力机制将交叉融合的特征与边缘掩码信息无缝集成,逐步监督并促进模型训练。实验结果表明,该方法优于现有的最先进方案。

链接: https://arxiv.org/abs/2412.12503
作者: Yakun Niu,Pei Chen,Lei Zhang,Hongjian Yin,Qi Chang
机构: Henan University(河南大学); Hangzhou Dianzi University(杭州电子科技大学)
关键词: Image Splicing Localization, Splicing Localization, Image Splicing, digital forensics, fundamental yet challenging
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 5 pages,3 figures

点击查看摘要

Abstract:Image Splicing Localization (ISL) is a fundamental yet challenging task in digital forensics. Although current approaches have achieved promising performance, the edge information is insufficiently exploited, resulting in poor integrality and high false alarms. To tackle this problem, we propose a multi-scale cross-fusion and edge-supervision network for ISL. Specifically, our framework consists of three key steps: multi-scale features cross-fusion, edge mask prediction and edge-supervision localization. Firstly, we input the RGB image and its noise image into a segmentation network to learn multi-scale features, which are then aggregated via a cross-scale fusion followed by a cross-domain fusion to enhance feature representation. Secondly, we design an edge mask prediction module to effectively mine the reliable boundary artifacts. Finally, the cross-fused features and the reliable edge mask information are seamlessly integrated via an attention mechanism to incrementally supervise and facilitate model training. Extensive experiments on publicly available datasets demonstrate that our proposed method is superior to state-of-the-art schemes.
zh

[CV-104] rack the Answer: Extending TextVQA from Image to Video with Spatio-Temporal Clues AAAI2025

【速读】: 该论文试图解决视频文本视觉问答 (Video TextVQA) 任务中,现有方法在处理多帧文本和视觉信息时,由于忽视了视觉实体间的时空关系(spatio-temporal relationships),导致推理不合理和回答不准确的问题。解决方案的关键在于提出了TEA方法(Track the Answer),该方法通过恢复视觉实体间的时空关系,并结合OCR感知线索(OCR-aware clues),以增强对问题的推理质量。实验结果表明,TEA在多个公开的Video TextVQA数据集上显著优于现有的TextVQA方法、视频-语言预训练方法以及视频大语言模型。

链接: https://arxiv.org/abs/2412.12502
作者: Yan Zhang,Gangyan Zeng,Huawen Shen,Daiqing Wu,Yu Zhou,Can Ma
机构: 1. School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China (哈尔滨工业大学计算机科学与技术学院);
2. School of Computer Science and Engineering, Nanyang Technological University, Singapore (南洋理工大学计算机科学与工程学院);
3. Key Laboratory of Network Security and Cryptology, Fujian Normal University, Fuzhou, China (福建师范大学网络安全与密码学重点实验室);
4. School of Computer Science and Technology, Soochow University, Suzhou, China (苏州大学计算机科学与技术学院)
关键词: jointly reasoning textual, Video text-based visual, Video, practical task, task that aims
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Video text-based visual question answering (Video TextVQA) is a practical task that aims to answer questions by jointly reasoning textual and visual information in a given video. Inspired by the development of TextVQA in image domain, existing Video TextVQA approaches leverage a language model (e.g. T5) to process text-rich multiple frames and generate answers auto-regressively. Nevertheless, the spatio-temporal relationships among visual entities (including scene text and objects) will be disrupted and models are susceptible to interference from unrelated information, resulting in irrational reasoning and inaccurate answering. To tackle these challenges, we propose the TEA (stands for ``\textbfTrack th\textbfE \textbfAnswer’') method that better extends the generative TextVQA framework from image to video. TEA recovers the spatio-temporal relationships in a complementary way and incorporates OCR-aware clues to enhance the quality of reasoning questions. Extensive experiments on several public Video TextVQA datasets validate the effectiveness and generalization of our framework. TEA outperforms existing TextVQA methods, video-language pretraining methods and video large language models by great margins.
zh

[CV-105] Faster Vision Mamba is Rebuilt in Minutes via Merged Token Re-training

【速读】: 该论文试图解决在Vision Mamba(Vim)中进行token reduction时,由于信息token的修剪导致关键知识丢失和性能下降的问题。解决方案的关键在于提出了一种名为R-MeeTo的框架,通过重新训练token-reduced模型来有效重建关键知识,从而提升模型性能。实验结果表明,该框架能够在短时间内(如3个训练周期内)显著提升模型准确率,并在推理速度上实现1.2倍至1.5倍的加速,同时仅导致极小的准确率下降(如Vim-S在1.2倍加速下仅下降1.3%)。

链接: https://arxiv.org/abs/2412.12496
作者: Mingjia Shi,Yuhao Zhou,Ruiji Yu,Zekai Li,Zhiyuan Liang,Xuanlei Zhao,Xiaojiang Peng,Tanmay Rajpurohit,Shanmukha Ramakrishna Vedantam,Wangbo Zhao,Kai Wang,Yang You
机构: National University of Singapore(新加坡国立大学); Sichuan University(四川大学); Shanghai Jiao Tong University(上海交通大学); Shenzhen University of Technology(深圳技术大学); Genpact Innovation Center(Genpact创新中心); Independent Researcher(独立研究员)
关键词: Vision Transformers, yielded promising outcomes, Vision Mamba, Vision Mamba compared, computer vision
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Vision Mamba (e.g., Vim) has successfully been integrated into computer vision, and token reduction has yielded promising outcomes in Vision Transformers (ViTs). However, token reduction performs less effectively on Vision Mamba compared to ViTs. Pruning informative tokens in Mamba leads to a high loss of key knowledge and bad performance. This makes it not a good solution for enhancing efficiency in Mamba. Token merging, which preserves more token information than pruning, has demonstrated commendable performance in ViTs. Nevertheless, vanilla merging performance decreases as the reduction ratio increases either, failing to maintain the key knowledge in Mamba. Re-training the token-reduced model enhances the performance of Mamba, by effectively rebuilding the key knowledge. Empirically, pruned Vims only drop up to 0.9% accuracy on ImageNet-1K, recovered by our proposed framework R-MeeTo in our main evaluation. We show how simple and effective the fast recovery can be achieved at minute-level, in particular, a 35.9% accuracy spike over 3 epochs of training on Vim-Ti. Moreover, Vim-Ti/S/B are re-trained within 5/7/17 minutes, and Vim-S only drop 1.3% with 1.2x (up to 1.5x) speed up in inference.
zh

[CV-106] DuSSS: Dual Semantic Similarity-Supervised Vision-Language Model for Semi-Supervised Medical Image Segmentation

【速读】: 该论文试图解决半监督医学图像分割 (Semi-supervised Medical Image Segmentation, SSMIS) 中由于低质量伪标签导致的错误监督问题,以及视觉-语言模型 (Vision-Language Model, VLM) 在跨模态信息处理中存在的多目标对应问题。解决方案的关键在于提出了双语义相似性监督的视觉-语言模型 (Dual Semantic Similarity-Supervised VLM, DuSSS),通过设计双对比学习 (Dual Contrastive Learning, DCL) 来增强跨模态语义一致性,并引入语义相似性监督策略 (Semantic Similarity-Supervision Strategy, SSS) 来促进多语义对应的学习。此外,基于预训练的 VLM 生成文本提示引导的监督信息,以改进伪标签的质量,从而提升一致性正则化的效果。

链接: https://arxiv.org/abs/2412.12492
作者: Qingtao Pan,Wenhao Qiao,Jingjiao Lou,Bing Ji,Shuo Li
机构: 未知
关键词: Semi-supervised medical image, medical image segmentation, pixel-wise manual annotations, regularize model training, Semi-supervised medical
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Semi-supervised medical image segmentation (SSMIS) uses consistency learning to regularize model training, which alleviates the burden of pixel-wise manual annotations. However, it often suffers from error supervision from low-quality pseudo labels. Vision-Language Model (VLM) has great potential to enhance pseudo labels by introducing text prompt guided multimodal supervision information. It nevertheless faces the cross-modal problem: the obtained messages tend to correspond to multiple targets. To address aforementioned problems, we propose a Dual Semantic Similarity-Supervised VLM (DuSSS) for SSMIS. Specifically, 1) a Dual Contrastive Learning (DCL) is designed to improve cross-modal semantic consistency by capturing intrinsic representations within each modality and semantic correlations across modalities. 2) To encourage the learning of multiple semantic correspondences, a Semantic Similarity-Supervision strategy (SSS) is proposed and injected into each contrastive learning process in DCL, supervising semantic similarity via the distribution-based uncertainty levels. Furthermore, a novel VLM-based SSMIS network is designed to compensate for the quality deficiencies of pseudo-labels. It utilizes the pretrained VLM to generate text prompt guided supervision information, refining the pseudo label for better consistency regularization. Experimental results demonstrate that our DuSSS achieves outstanding performance with Dice of 82.52%, 74.61% and 78.03% on three public datasets (QaTa-COV19, BM-Seg and MoNuSeg).
zh

[CV-107] Pattern Analogies: Learning to Perform Programmatic Image Edits by Analogy

【速读】: 该论文试图解决图案图像编辑中的复杂问题,特别是如何进行结构感知的程序化编辑,即通过修改生成图案的底层程序来实现编辑。解决方案的关键在于引入了一种新颖的方法,通过使用图案类比(pattern analogy)和基于学习的生成模型来执行这些编辑。具体来说,论文提出了SplitWeave,一种领域特定语言(domain-specific language),结合合成图案类比的采样框架,生成了大规模高质量的合成训练数据集。此外,论文还提出了TriFuser,一种潜在扩散模型(Latent Diffusion Model, LDM),用于解决在该任务中直接部署LDM时出现的重大问题。这种方法不仅能够忠实地执行示例编辑,还能推广到训练分布之外的相关图案风格。

链接: https://arxiv.org/abs/2412.12463
作者: Aditya Ganeshan,Thibault Groueix,Paul Guerrero,Radomír Měch,Matthew Fisher,Daniel Ritchie
机构: Brown University(布朗大学); Adobe Research(Adobe研究)
关键词: physical worlds, digital and physical, Pattern images, Pattern, images
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Graphics (cs.GR); Human-Computer Interaction (cs.HC)
备注: Website: this https URL

点击查看摘要

Abstract:Pattern images are everywhere in the digital and physical worlds, and tools to edit them are valuable. But editing pattern images is tricky: desired edits are often programmatic: structure-aware edits that alter the underlying program which generates the pattern. One could attempt to infer this underlying program, but current methods for doing so struggle with complex images and produce unorganized programs that make editing tedious. In this work, we introduce a novel approach to perform programmatic edits on pattern images. By using a pattern analogy – a pair of simple patterns to demonstrate the intended edit – and a learning-based generative model to execute these edits, our method allows users to intuitively edit patterns. To enable this paradigm, we introduce SplitWeave, a domain-specific language that, combined with a framework for sampling synthetic pattern analogies, enables the creation of a large, high-quality synthetic training dataset. We also present TriFuser, a Latent Diffusion Model (LDM) designed to overcome critical issues that arise when naively deploying LDMs to this task. Extensive experiments on real-world, artist-sourced patterns reveals that our method faithfully performs the demonstrated edit while also generalizing to related pattern styles beyond its training distribution.
zh

[CV-108] PromptDet: A Lightweight 3D Object Detection Framework with LiDAR Prompts AAAI2025

【速读】: 该论文试图解决多相机3D目标检测中由于相机固有的测距能力不足导致的深度估计不准确问题。解决方案的关键在于提出了PromptDet框架,该框架结合了基于相机的检测模块(如BEVDet和BEVDepth)和LiDAR辅助的prompter。LiDAR辅助的prompter利用LiDAR点云作为补充信号,并通过最小化的可训练参数增强了深度估计的精度。该框架具有灵活性,既可以作为轻量级的多模态融合方法,也可以在推理阶段作为纯相机方法使用。实验结果表明,PromptDet在nuScenes数据集上显著提升了检测性能,且在多模态和纯相机模式下均表现出优越的性能。

链接: https://arxiv.org/abs/2412.12460
作者: Kun Guo,Qiang Ling
机构: 未知
关键词: object detection aims, space using multiple, cost-effectiveness trade-off, object detection, aims to detect
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Multi-camera 3D object detection aims to detect and localize objects in 3D space using multiple cameras, which has attracted more attention due to its cost-effectiveness trade-off. However, these methods often struggle with the lack of accurate depth estimation caused by the natural weakness of the camera in ranging. Recently, multi-modal fusion and knowledge distillation methods for 3D object detection have been proposed to solve this problem, which are time-consuming during the training phase and not friendly to memory cost. In light of this, we propose PromptDet, a lightweight yet effective 3D object detection framework motivated by the success of prompt learning in 2D foundation model. Our proposed framework, PromptDet, comprises two integral components: a general camera-based detection module, exemplified by models like BEVDet and BEVDepth, and a LiDAR-assisted prompter. The LiDAR-assisted prompter leverages the LiDAR points as a complementary signal, enriched with a minimal set of additional trainable parameters. Notably, our framework is flexible due to our prompt-like design, which can not only be used as a lightweight multi-modal fusion method but also as a camera-only method for 3D object detection during the inference phase. Extensive experiments on nuScenes validate the effectiveness of the proposed PromptDet. As a multi-modal detector, PromptDet improves the mAP and NDS by at most 22.8% and 21.1% with fewer than 2% extra parameters compared with the camera-only baseline. Without LiDAR points, PromptDet still achieves an improvement of at most 2.4% mAP and 4.0% NDS with almost no impact on camera detection inference time.
zh

[CV-109] hree Things to Know about Deep Metric Learning

【速读】: 该论文旨在解决开放集图像检索中的监督深度度量学习问题,重点关注损失函数、混合正则化 (mixup regularization) 和模型初始化三个关键方面。其核心解决方案包括:提出一种可微分的替代损失函数,通过在大批次上计算来近似整个训练集的检索评估指标(如 recall@k),并通过绕过 GPU 内存限制的技术实现计算可行性;引入基于成对标量相似度的混合正则化技术,进一步扩大批次大小;以及使用在大规模数据集上预训练的基础模型初始化视觉编码器,从而提升训练效果。这些方法的协同作用使得大规模模型在流行基准测试中几乎达到最优性能。

链接: https://arxiv.org/abs/2412.12432
作者: Yash Patel,Giorgos Tolias,Jiri Matas
机构: 未知
关键词: paper addresses supervised, deep metric learning, addresses supervised deep, supervised deep metric, open-set image retrieval
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:This paper addresses supervised deep metric learning for open-set image retrieval, focusing on three key aspects: the loss function, mixup regularization, and model initialization. In deep metric learning, optimizing the retrieval evaluation metric, recall@k, via gradient descent is desirable but challenging due to its non-differentiable nature. To overcome this, we propose a differentiable surrogate loss that is computed on large batches, nearly equivalent to the entire training set. This computationally intensive process is made feasible through an implementation that bypasses the GPU memory limitations. Additionally, we introduce an efficient mixup regularization technique that operates on pairwise scalar similarities, effectively increasing the batch size even further. The training process is further enhanced by initializing the vision encoder using foundational models, which are pre-trained on large-scale datasets. Through a systematic study of these components, we demonstrate that their synergy enables large models to nearly solve popular benchmarks.
zh

[CV-110] MASt3R-SLAM: Real-Time Dense SLAM with 3D Reconstruction Priors

【速读】: 该论文旨在解决实时单目稠密SLAM(Simultaneous Localization and Mapping)系统的问题,特别是在无固定或参数化相机模型假设的情况下,确保系统在野外视频序列中的鲁棒性。解决方案的关键在于利用MASt3R(two-view 3D reconstruction and matching prior)从底层构建系统,并引入高效的点图匹配、相机跟踪、局部融合、图构建、回环检测以及二阶全局优化方法。通过已知的相机标定,系统经过简单修改即可在各种基准测试中达到最先进的性能,实现全局一致的位姿和稠密几何输出,同时保持15 FPS的实时性能。

链接: https://arxiv.org/abs/2412.12392
作者: Riku Murai,Eric Dexheimer,Andrew J. Davison
机构: Imperial College London(帝国理工学院)
关键词: system designed bottom-up, SLAM system designed, present a real-time, designed bottom-up, real-time monocular dense
类目: Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO)
备注: The first two authors contributed equally to this work. Project Page: this https URL

点击查看摘要

Abstract:We present a real-time monocular dense SLAM system designed bottom-up from MASt3R, a two-view 3D reconstruction and matching prior. Equipped with this strong prior, our system is robust on in-the-wild video sequences despite making no assumption on a fixed or parametric camera model beyond a unique camera centre. We introduce efficient methods for pointmap matching, camera tracking and local fusion, graph construction and loop closure, and second-order global optimisation. With known calibration, a simple modification to the system achieves state-of-the-art performance across various benchmarks. Altogether, we propose a plug-and-play monocular SLAM system capable of producing globally-consistent poses and dense geometry while operating at 15 FPS.
zh

[CV-111] Domain Generalization in Autonomous Driving: Evaluating YOLOv8s RT-DETR and YOLO-NAS with the ROAD-Almaty Dataset

【速读】: 该论文旨在解决在哈萨克斯坦独特驾驶环境中,现有先进目标检测模型(YOLOv8s、RT-DETR 和 YOLO-NAS)的领域泛化能力问题。解决方案的关键在于评估这些模型在新构建的ROAD-Almaty数据集上的表现,该数据集涵盖了多样化的天气、光照和交通条件,且无需重新训练模型。研究结果表明,RT-DETR在IoU=0.5时以0.672的平均F1-score表现最佳,显著优于YOLOv8s和YOLO-NAS。然而,所有模型在高IoU阈值和恶劣环境条件下的性能显著下降,这强调了需要地理多样化的训练数据集和专门的领域适应技术,以提高全球自动驾驶车辆检测系统的可靠性。

链接: https://arxiv.org/abs/2412.12349
作者: Madiyar Alimov,Temirlan Meiramkhanov
机构: International IT University(国际信息技术大学); Astana IT University(阿斯塔纳信息技术大学)
关键词: environment of Kazakhstan, unique driving environment, study investigates, object detection models, domain generalization capabilities
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:This study investigates the domain generalization capabilities of three state-of-the-art object detection models - YOLOv8s, RT-DETR, and YOLO-NAS - within the unique driving environment of Kazakhstan. Utilizing the newly constructed ROAD-Almaty dataset, which encompasses diverse weather, lighting, and traffic conditions, we evaluated the models’ performance without any retraining. Quantitative analysis revealed that RT-DETR achieved an average F1-score of 0.672 at IoU=0.5, outperforming YOLOv8s (0.458) and YOLO-NAS (0.526) by approximately 46% and 27%, respectively. Additionally, all models exhibited significant performance declines at higher IoU thresholds (e.g., a drop of approximately 20% when increasing IoU from 0.5 to 0.75) and under challenging environmental conditions, such as heavy snowfall and low-light scenarios. These findings underscore the necessity for geographically diverse training datasets and the implementation of specialized domain adaptation techniques to enhance the reliability of autonomous vehicle detection systems globally. This research contributes to the understanding of domain generalization challenges in autonomous driving, particularly in underrepresented regions.
zh

[CV-112] Efficient Object-centric Representation Learning with Pre-trained Geometric Prior

【速读】: 该论文试图解决视频中以物体为中心的表示学习的关键挑战,特别是在复杂场景下的物体发现问题。解决方案的关键在于提出了一种弱监督框架,该框架强调几何理解,并利用预训练的视觉模型来增强物体发现能力。具体来说,论文引入了一种高效的槽解码器(slot decoder),专门设计用于以物体为中心的学习,能够在不需要显式深度信息的情况下有效表示多物体场景。这种方法在合成视频基准测试中展示了与监督方法相当的性能,同时保持了计算效率,推动了该领域在复杂现实场景中的实际应用。

链接: https://arxiv.org/abs/2412.12331
作者: Phúc H. Le Khac,Graham Healy,Alan F. Smeaton
机构: Dublin City University, Glasnevin, Dublin 9, Ireland(都柏林城市大学,格拉斯内文,都柏林9区,爱尔兰)
关键词: paper addresses key, addresses key challenges, paper addresses, addresses key, key challenges
类目: Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM)
备注: 6 pages, 4 Figures, 2 Tables

点击查看摘要

Abstract:This paper addresses key challenges in object-centric representation learning of video. While existing approaches struggle with complex scenes, we propose a novel weakly-supervised framework that emphasises geometric understanding and leverages pre-trained vision models to enhance object discovery. Our method introduces an efficient slot decoder specifically designed for object-centric learning, enabling effective representation of multi-object scenes without requiring explicit depth information. Results on synthetic video benchmarks with increasing complexity in terms of objects and their movement, object occlusion and camera motion demonstrate that our approach achieves comparable performance to supervised methods while maintaining computational efficiency. This advances the field towards more practical applications in complex real-world scenarios.
zh

[CV-113] owards a Universal Synthetic Video Detector: From Face or Background Manipulations to Fully AI-Generated Content

【速读】: 该论文试图解决现有DeepFake检测技术主要集中在面部操作(如换脸或唇同步)的局限性,特别是面对文本生成视频(T2V)和图像生成视频(I2V)等全AI生成内容以及无缝背景修改时,传统方法的不足。解决方案的关键在于引入了一种名为UNITE的通用网络,该模型通过捕捉全帧操作,扩展了检测能力至无面部、非人类主体及复杂背景修改的场景。UNITE采用基于Transformer的架构,利用SigLIP-So400M基础模型提取域无关特征,并通过结合任务无关数据和标准DeepFake数据集进行训练,同时引入注意力多样性(AD)损失以减少对面部过度关注,从而提升在不同情境下的检测性能。

链接: https://arxiv.org/abs/2412.12278
作者: Rohit Kundu,Hao Xiong,Vishal Mohanty,Athula Balachandran,Amit K. Roy-Chowdhury
机构: Google(谷歌); University of California, Riverside(加州大学河滨分校)
关键词: Existing DeepFake detection, techniques primarily focus, detection techniques primarily, underline, Existing DeepFake
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Existing DeepFake detection techniques primarily focus on facial manipulations, such as face-swapping or lip-syncing. However, advancements in text-to-video (T2V) and image-to-video (I2V) generative models now allow fully AI-generated synthetic content and seamless background alterations, challenging face-centric detection methods and demanding more versatile approaches. To address this, we introduce the \underlineUniversal \underlineNetwork for \underlineIdentifying \underlineTampered and synth\underlineEtic videos (\textttUNITE) model, which, unlike traditional detectors, captures full-frame manipulations. \textttUNITE extends detection capabilities to scenarios without faces, non-human subjects, and complex background modifications. It leverages a transformer-based architecture that processes domain-agnostic features extracted from videos via the SigLIP-So400M foundation model. Given limited datasets encompassing both facial/background alterations and T2V/I2V content, we integrate task-irrelevant data alongside standard DeepFake datasets in training. We further mitigate the model’s tendency to over-focus on faces by incorporating an attention-diversity (AD) loss, which promotes diverse spatial attention across video frames. Combining AD loss with cross-entropy improves detection performance across varied contexts. Comparative evaluations demonstrate that \textttUNITE outperforms state-of-the-art detectors on datasets (in cross-data settings) featuring face/background manipulations and fully synthetic T2V/I2V videos, showcasing its adaptability and generalizable detection capabilities. Subjects: Computer Vision and Pattern Recognition (cs.CV) Cite as: arXiv:2412.12278 [cs.CV] (or arXiv:2412.12278v1 [cs.CV] for this version) https://doi.org/10.48550/arXiv.2412.12278 Focus to learn more arXiv-issued DOI via DataCite (pending registration)
zh

[CV-114] OmniPrism: Learning Disentangled Visual Concept for Image Generation

【速读】: 该论文试图解决现有方法在多方面概念生成场景中容易受到无关概念干扰,导致概念混淆和创意生成受阻的问题。解决方案的关键在于提出了OmniPrism,一种视觉概念解耦方法,通过自然语言引导学习解耦的概念表示,并利用扩散模型将这些概念整合到生成过程中。具体来说,论文构建了一个配对概念解耦数据集(PCD-200K),并通过对比正交解耦(COD)训练流程学习解耦的概念表示,这些表示随后被注入到扩散模型的交叉注意力层中进行生成。此外,设计了一组块嵌入以适应扩散模型中每个块的概念域,从而实现高质量的概念解耦生成结果。

链接: https://arxiv.org/abs/2412.12242
作者: Yangyang Li,Daqing Liu,Wu Liu,Allen He,Xinchen Liu,Yongdong Zhang,Guoqing Jin
机构: University of Science and Technology of China; JD Explore Academy, JD.com Inc.; State Key Laboratory of Communication Content Cognition, People’s Daily Online
关键词: produce relevant outcomes, concept, creative image generation, relevant outcomes, Creative visual concept
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: WebPage available at this https URL

点击查看摘要

Abstract:Creative visual concept generation often draws inspiration from specific concepts in a reference image to produce relevant outcomes. However, existing methods are typically constrained to single-aspect concept generation or are easily disrupted by irrelevant concepts in multi-aspect concept scenarios, leading to concept confusion and hindering creative generation. To address this, we propose OmniPrism, a visual concept disentangling approach for creative image generation. Our method learns disentangled concept representations guided by natural language and trains a diffusion model to incorporate these concepts. We utilize the rich semantic space of a multimodal extractor to achieve concept disentanglement from given images and concept guidance. To disentangle concepts with different semantics, we construct a paired concept disentangled dataset (PCD-200K), where each pair shares the same concept such as content, style, and composition. We learn disentangled concept representations through our contrastive orthogonal disentangled (COD) training pipeline, which are then injected into additional diffusion cross-attention layers for generation. A set of block embeddings is designed to adapt each block’s concept domain in the diffusion models. Extensive experiments demonstrate that our method can generate high-quality, concept-disentangled results with high fidelity to text prompts and desired concepts.
zh

[CV-115] You Only Submit One Image to Find the Most Suitable Generative Model NEURIPS2023

【速读】: 该论文试图解决生成式模型(Generative Model)在模型库中难以高效识别和选择的问题。现有的模型库缺乏先进的模型管理和识别机制,用户只能通过文本匹配或下载排序等方式进行搜索,导致难以快速找到最符合需求的模型。论文提出了一种名为生成式模型识别(Generative Model Identification, GMI)的新方法,旨在从大量候选模型中高效识别出最适合用户需求的模型。解决方案的关键在于三个核心模块:1) 加权简化核均值嵌入(Reduced Kernel Mean Embedding, RKME)框架,用于捕捉生成图像的分布及图像与提示之间的关系;2) 预训练的视觉-语言模型,用于解决维度问题;3) 图像审讯器,用于处理跨模态问题。实验结果表明,该方法在识别准确性和效率上均表现出色,用户仅需提交一张示例图像即可实现超过80%的平均前4识别准确率。

链接: https://arxiv.org/abs/2412.12232
作者: Zhi Zhou,Lan-Zhe Guo,Peng-Xiao Song,Yu-Feng Li
机构: Nanjing University (南京大学)
关键词: Hugging Face, Face and Civitai, Deep generative models, Deep generative, generative model hubs
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: Accepted by NeurIPS 2023 Workshop on Diffusion Models

点击查看摘要

Abstract:Deep generative models have achieved promising results in image generation, and various generative model hubs, e.g., Hugging Face and Civitai, have been developed that enable model developers to upload models and users to download models. However, these model hubs lack advanced model management and identification mechanisms, resulting in users only searching for models through text matching, download sorting, etc., making it difficult to efficiently find the model that best meets user requirements. In this paper, we propose a novel setting called Generative Model Identification (GMI), which aims to enable the user to identify the most appropriate generative model(s) for the user’s requirements from a large number of candidate models efficiently. To our best knowledge, it has not been studied yet. In this paper, we introduce a comprehensive solution consisting of three pivotal modules: a weighted Reduced Kernel Mean Embedding (RKME) framework for capturing the generated image distribution and the relationship between images and prompts, a pre-trained vision-language model aimed at addressing dimensionality challenges, and an image interrogator designed to tackle cross-modality issues. Extensive empirical results demonstrate the proposal is both efficient and effective. For example, users only need to submit a single example image to describe their requirements, and the model platform can achieve an average top-4 identification accuracy of more than 80%.
zh

[CV-116] Can video generation replace cinematographers? Research on the cinematic language of generated video

【速读】: 该论文试图解决文本到视频生成 (T2V) 模型在生成过程中对电影语言(cinematic language)关注不足的问题,特别是缺乏对镜头构图、角度和摄像机运动等电影语言元素的控制,这些元素对于传达情感和叙事节奏至关重要。解决方案的关键在于提出了一个三重方法:首先,引入一个包含镜头构图、角度和摄像机运动等元素的电影语言数据集,使模型能够学习多样化的电影风格;其次,开发了CameraCLIP模型,该模型在所提出的数据集上进行了微调,能够理解和评估生成视频中的复杂电影语言,并为多镜头组合过程提供指导;最后,提出了CLIPLoRA,一种基于成本的动态LoRA组合方法,通过动态融合多个预训练的电影LoRA模型,实现电影语言的平滑过渡和真实融合。实验结果表明,CameraCLIP在评估电影语言与视频对齐方面优于现有模型,而CLIPLoRA则显著提升了多镜头组合的能力。

链接: https://arxiv.org/abs/2412.12223
作者: Xiaozhe Li,Kai WU,Siyi Yang,YiZhan Qu,Guohua.Zhang,Zhiyu Chen,Jiayao Li,Jiangchuan Mu,Xiaobin Hu,Wen Fang,Mingliang Xiong,Hao Deng,Qingwen Liu,Gang Li,Bin He
机构: Tongji University(同济大学); ByteDance(字节跳动); Technical University of Munich(慕尼黑工业大学)
关键词: Recent advancements, leveraged diffusion models, cinematic language, generation have leveraged, textual descriptions
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注: 13 pages

点击查看摘要

Abstract:Recent advancements in text-to-video (T2V) generation have leveraged diffusion models to enhance the visual coherence of videos generated from textual descriptions. However, most research has primarily focused on object motion, with limited attention given to cinematic language in videos, which is crucial for cinematographers to convey emotion and narrative pacing. To address this limitation, we propose a threefold approach to enhance the ability of T2V models to generate controllable cinematic language. Specifically, we introduce a cinematic language dataset that encompasses shot framing, angle, and camera movement, enabling models to learn diverse cinematic styles. Building on this, to facilitate robust cinematic alignment evaluation, we present CameraCLIP, a model fine-tuned on the proposed dataset that excels in understanding complex cinematic language in generated videos and can further provide valuable guidance in the multi-shot composition process. Finally, we propose CLIPLoRA, a cost-guided dynamic LoRA composition method that facilitates smooth transitions and realistic blending of cinematic language by dynamically fusing multiple pre-trained cinematic LoRAs within a single video. Our experiments demonstrate that CameraCLIP outperforms existing models in assessing the alignment between cinematic language and video, achieving an R@1 score of 0.81. Additionally, CLIPLoRA improves the ability for multi-shot composition, potentially bridging the gap between automatically generated videos and those shot by professional cinematographers.
zh

[CV-117] Endangered Alert: A Field-Validated Self-Training Scheme for Detecting and Protecting Threatened Wildlife on Roads and Roadsides

【速读】: 该论文试图解决在资源有限的环境中,针对稀有动物(如澳大利亚的食火鸡)的动物-车辆碰撞(Animal-Vehicle Collisions, AVCs)检测问题。解决方案的关键在于提出了一种创新的自我训练方法,通过利用云和边缘计算以及自动数据标注技术,逐步提升现场部署模型的检测性能。具体来说,该方法引入了标签增强非极大值抑制(Label-Augmentation Non-Maximum Suppression, LA-NMS),结合视觉语言模型(Vision-Language Model, VLM)实现自动化数据标注,从而在五个月的实际部署中验证了其鲁棒性和有效性,显著提高了目标检测的准确性和预测置信度。

链接: https://arxiv.org/abs/2412.12222
作者: Kunming Li,Mao Shan,Stephany Berrio Perez,Katie Luo,Stewart Worrall
机构: Australian Centre for Robotics (ACFR) at The University of Sydney (NSW, Australia); Computer Information Sciences Department at Cornell University (NY, USA)
关键词: global safety concern, Traffic accidents, safety concern, fatalities each year, global safety
类目: Computer Vision and Pattern Recognition (cs.CV)
备注: 8 pages, 8 figures

点击查看摘要

Abstract:Traffic accidents are a global safety concern, resulting in numerous fatalities each year. A considerable number of these deaths are caused by animal-vehicle collisions (AVCs), which not only endanger human lives but also present serious risks to animal populations. This paper presents an innovative self-training methodology aimed at detecting rare animals, such as the cassowary in Australia, whose survival is threatened by road accidents. The proposed method addresses critical real-world challenges, including acquiring and labelling sensor data for rare animal species in resource-limited environments. It achieves this by leveraging cloud and edge computing, and automatic data labelling to improve the detection performance of the field-deployed model iteratively. Our approach introduces Label-Augmentation Non-Maximum Suppression (LA-NMS), which incorporates a vision-language model (VLM) to enable automated data labelling. During a five-month deployment, we confirmed the method’s robustness and effectiveness, resulting in improved object detection accuracy and increased prediction confidence. The source code is available: this https URL
zh

[CV-118] Relieving Universal Label Noise for Unsupervised Visible-Infrared Person Re-Identification by Inferring from Neighbors

【速读】: 该论文试图解决无监督可见光-红外人重识别(Unsupervised visible-infrared person re-identification, USL-VI-ReID)中的标签噪声问题,尤其是在缺乏标注的情况下,由于聚类结果不理想和模态差异导致的标签噪声。解决方案的关键在于引入邻域信息来校准标签,具体通过Neighbor-guided Universal Label Calibration (N-ULC)模块,用邻近样本的软标签替代硬伪标签,以减少标签噪声;同时,通过Neighbor-guided Dynamic Weighting (N-DW)模块动态调整不可靠样本的权重,提升训练稳定性。实验结果表明,该方法在RegDB和SYSU-MM01数据集上优于现有的USL-VI-ReID方法。

链接: https://arxiv.org/abs/2412.12220
作者: Xiao Teng,Long Lan,Dingyao Chen,Kele Xu,Nan Yin
机构: Tsinghua University (清华大学); Peking University (北京大学)
关键词: visible-infrared person re-identification, Unsupervised visible-infrared person, remains challenging due, person re-identification, absence of annotations
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
备注:

点击查看摘要

Abstract:Unsupervised visible-infrared person re-identification (USL-VI-ReID) is of great research and practical significance yet remains challenging due to the absence of annotations. Existing approaches aim to learn modality-invariant representations in an unsupervised setting. However, these methods often encounter label noise within and across modalities due to suboptimal clustering results and considerable modality discrepancies, which impedes effective training. To address these challenges, we propose a straightforward yet effective solution for USL-VI-ReID by mitigating universal label noise using neighbor information. Specifically, we introduce the Neighbor-guided Universal Label Calibration (N-ULC) module, which replaces explicit hard pseudo labels in both homogeneous and heterogeneous spaces with soft labels derived from neighboring samples to reduce label noise. Additionally, we present the Neighbor-guided Dynamic Weighting (N-DW) module to enhance training stability by minimizing the influence of unreliable samples. Extensive experiments on the RegDB and SYSU-MM01 datasets demonstrate that our method outperforms existing USL-VI-ReID approaches, despite its simplicity. The source code is available at: this https URL.
zh

[CV-119] SitPose: Real-Time Detection of Sitting Posture and Sedentary Behavior Using Ensemble Learning With Depth Sensor

【速读】: 该论文试图解决因不良坐姿导致的职业性肌肉骨骼疾病 (WMSDs) 问题,特别是针对办公室员工长时间久坐可能引发的颈椎病和心血管疾病等慢性疾病。解决方案的关键在于开发了一个名为 SitPose 的坐姿与久坐检测系统,该系统利用最新的 Kinect 深度摄像头实时追踪骨骼关节的三维坐标,并计算相关关节的角度值。通过建立包含六种不同坐姿和一个站姿的数据集(共 33,409 个数据点),并应用多种先进的机器学习算法进行坐姿识别,最终采用基于软投票机制的集成学习模型,实现了 98.1% 的 F1 分数,从而有效鼓励良好坐姿并减少久坐习惯。

链接: https://arxiv.org/abs/2412.12216
作者: Hang Jin,Xin He,Lingyun Wang,Yujun Zhu,Weiwei Jiang,Xiaobo Zhou
机构: School of Computer and Information, Anhui Normal University, WuHu, 241002, Anhui, China; School of Computer Science, Nanjing University of Information Science and Technology, 210044, Jiangsu, China; School of Computer Science and Technology, Tianjin University, Tianjin, China
关键词: work-related musculoskeletal disorders, Poor sitting posture, Poor sitting, musculoskeletal disorders, work-related musculoskeletal
类目: Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注:

点击查看摘要

Abstract:Poor sitting posture can lead to various work-related musculoskeletal disorders (WMSDs). Office employees spend approximately 81.8% of their working time seated, and sedentary behavior can result in chronic diseases such as cervical spondylosis and cardiovascular diseases. To address these health concerns, we present SitPose, a sitting posture and sedentary detection system utilizing the latest Kinect depth camera. The system tracks 3D coordinates of bone joint points in real-time and calculates the angle values of related joints. We established a dataset containing six different sitting postures and one standing posture, totaling 33,409 data points, by recruiting 36 participants. We applied several state-of-the-art machine learning algorithms to the dataset and compared their performance in recognizing the sitting poses. Our results show that the ensemble learning model based on the soft voting mechanism achieves the highest F1 score of 98.1%. Finally, we deployed the SitPose system based on this ensemble model to encourage better sitting posture and to reduce sedentary habits.
zh

[CV-120] AI-Driven Innovations in Volumetric Video Streaming: A Review

【速读】: 该论文试图解决体积视频(volumetric video)在传输和渲染过程中面临的挑战,这些挑战源于其复杂的结构和大容量的数据。解决方案的关键在于采用AI驱动的技术,以提高体积内容流媒体(volumetric content streaming)的效率和质量。通过综述近年来的AI驱动方法,论文旨在提供当前最先进技术的全面概览,并为未来在实际应用中推进体积视频流媒体的发展提供潜在方向。

链接: https://arxiv.org/abs/2412.12208
作者: Erfan Entezami,Hui Guan
机构: University of Massachusetts Amherst (马萨诸塞大学阿默斯特分校)
关键词: interactive user experiences, volumetric content, volumetric content streaming, efforts to enhance, enhance immersive
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Recent efforts to enhance immersive and interactive user experiences have driven the development of volumetric video, a form of 3D content that enables 6 DoF. Unlike traditional 2D content, volumetric content can be represented in various ways, such as point clouds, meshes, or neural representations. However, due to its complex structure and large amounts of data size, deploying this new form of 3D data presents significant challenges in transmission and rendering. These challenges have hindered the widespread adoption of volumetric video in daily applications. In recent years, researchers have proposed various AI-driven techniques to address these challenges and improve the efficiency and quality of volumetric content streaming. This paper provides a comprehensive overview of recent advances in AI-driven approaches to facilitate volumetric content streaming. Through this review, we aim to offer insights into the current state-of-the-art and suggest potential future directions for advancing the deployment of volumetric video streaming in real-world applications.
zh

[CV-121] Provably Secure Robust Image Steganography via Cross-Modal Error Correction AAAI2025

【速读】: 该论文试图解决现有图像隐写术方法中生成的图像质量低、缺乏语义控制以及在社交网络上传和经过有损处理(如JPEG压缩)后难以准确提取秘密信息的问题。解决方案的关键在于利用先进的自回归(AR)图像生成模型和向量量化(VQ)标记器,结合跨模态纠错框架,生成高质量、可证明安全且鲁棒的隐写图像。该方法通过从隐写图像生成隐写文本,帮助恢复有损图像,从而确保即使在社交网络上经过有损处理后,仍能准确提取嵌入的秘密信息。

链接: https://arxiv.org/abs/2412.12206
作者: Yuang Qi,Kejiang Chen,Na Zhao,Zijin Yang,Weiming Zhang
机构: 未知
关键词: creating favorable conditions, image generation models, generation models, image generation, provably secure
类目: Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR)
备注: 7 pages. Accepted by AAAI 2025

点击查看摘要

Abstract:The rapid development of image generation models has facilitated the widespread dissemination of generated images on social networks, creating favorable conditions for provably secure image steganography. However, existing methods face issues such as low quality of generated images and lack of semantic control in the generation process. To leverage provably secure steganography with more effective and high-performance image generation models, and to ensure that stego images can accurately extract secret messages even after being uploaded to social networks and subjected to lossy processing such as JPEG compression, we propose a high-quality, provably secure, and robust image steganography method based on state-of-the-art autoregressive (AR) image generation models using Vector-Quantized (VQ) tokenizers. Additionally, we employ a cross-modal error-correction framework that generates stego text from stego images to aid in restoring lossy images, ultimately enabling the extraction of secret messages embedded within the images. Extensive experiments have demonstrated that the proposed method provides advantages in stego quality, embedding capacity, and robustness, while ensuring provable undetectability.
zh

[CV-122] Vehicle Detection and Classification for Toll collection using YOLOv11 and Ensemble OCR

【速读】: 该论文试图解决传统自动收费系统依赖复杂硬件配置、安装和维护成本高的问题。解决方案的关键在于采用单一摄像头结合YOLOv11计算机视觉架构和集成OCR技术,实现高效的自动收费。该系统通过智能车辆跟踪、空间轮迹模式自动轴数检测以及实时监控仪表盘界面,显著提升了识别精度(车牌识别准确率达98.5%,轴数检测准确率达94.2%,OCR置信度达99.7%),同时大幅减少了硬件资源需求,为智能交通系统提供了可扩展且以精度为核心的解决方案。

链接: https://arxiv.org/abs/2412.12191
作者: Karthik Sivakoti
机构: 未知
关键词: Traditional automated toll, require huge investments, Traditional automated, automated toll collection, complex hardware configurations
类目: Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Traditional automated toll collection systems depend on complex hardware configurations, that require huge investments in installation and maintenance. This research paper presents an innovative approach to revolutionize automated toll collection by using a single camera per plaza with the YOLOv11 computer vision architecture combined with an ensemble OCR technique. Our system has achieved a Mean Average Precision (mAP) of 0.895 over a wide range of conditions, demonstrating 98.5% accuracy in license plate recognition, 94.2% accuracy in axle detection, and 99.7% OCR confidence scoring. The architecture incorporates intelligent vehicle tracking across IOU regions, automatic axle counting by way of spatial wheel detection patterns, and real-time monitoring through an extended dashboard interface. Extensive training using 2,500 images under various environmental conditions, our solution shows improved performance while drastically reducing hardware resources compared to conventional systems. This research contributes toward intelligent transportation systems by introducing a scalable, precision-centric solution that improves operational efficiency and user experience in modern toll collections.
zh

[CV-123] Multi-Surrogate-Teacher Assistance for Representation Alignment in Fingerprint-based Indoor Localization WACV2025

【速读】: 该论文试图解决室内定位中跨接收信号强度(RSS)指纹数据集的知识迁移问题,特别是由于建筑结构、WiFi锚点数量和布置等因素导致的RSS数据集之间的固有差异。解决方案的关键在于提出了一种即插即用(PnP)的知识迁移框架,通过两个主要阶段实现:首先,设计了一个专家训练阶段,利用多个代理生成式教师作为全局适配器,统一不同源RSS数据集的输入差异,同时保留其独特特征;其次,在专家蒸馏阶段,引入三重底层约束,通过在目标数据集上精炼表示学习,最小化专用网络与代理教师之间的关键知识差异,从而促进对特定环境动态不敏感的表示对齐。该框架通过实验验证了其在WiFi RSS指纹数据集上的有效性,显著提升了专用网络在定位任务中的潜力。

链接: https://arxiv.org/abs/2412.12189
作者: Son Minh Nguyen,Linh Duy Tran,Duc Viet Le,Paul J.M Havinga
机构: University of Twente(特温特大学); Viettel AI, Viettel Group(越南电信集团)
关键词: Received Signal Strength, Signal Strength, Received Signal, RSS datasets, remains a challenge
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
备注: Accepted in the 1st round at WACV 2025 (Algorithm Track)

点击查看摘要

Abstract:Despite remarkable progress in knowledge transfer across visual and textual domains, extending these achievements to indoor localization, particularly for learning transferable representations among Received Signal Strength (RSS) fingerprint datasets, remains a challenge. This is due to inherent discrepancies among these RSS datasets, largely including variations in building structure, the input number and disposition of WiFi anchors. Accordingly, specialized networks, which were deprived of the ability to discern transferable representations, readily incorporate environment-sensitive clues into the learning process, hence limiting their potential when applied to specific RSS datasets. In this work, we propose a plug-and-play (PnP) framework of knowledge transfer, facilitating the exploitation of transferable representations for specialized networks directly on target RSS datasets through two main phases. Initially, we design an Expert Training phase, which features multiple surrogate generative teachers, all serving as a global adapter that homogenizes the input disparities among independent source RSS datasets while preserving their unique characteristics. In a subsequent Expert Distilling phase, we continue introducing a triplet of underlying constraints that requires minimizing the differences in essential knowledge between the specialized network and surrogate teachers through refining its representation learning on the target dataset. This process implicitly fosters a representational alignment in such a way that is less sensitive to specific environmental dynamics. Extensive experiments conducted on three benchmark WiFi RSS fingerprint datasets underscore the effectiveness of the framework that significantly exerts the full potential of specialized networks in localization.
zh

[CV-124] Multimodal Approaches to Fair Image Classification: An Ethical Perspective

【速读】: 该论文试图解决图像分类系统中的有害偏见问题,特别是在单一数据模态(如仅图像或仅文本)下,训练数据中隐藏的偏见可能被放大,导致不公平和歧视性结果。解决方案的关键在于采用多模态方法(multimodal approaches),通过结合视觉数据与其他模态(如文本和元数据)来提升图像分类系统的公平性和准确性。论文通过分析现有数据集和分类算法中的偏见,提出创新方法来缓解这些偏见,并评估在实际应用中部署此类系统的伦理影响。最终,论文倡导负责任的AI实践,强调公平性在AI解决方案中的重要性。

链接: https://arxiv.org/abs/2412.12165
作者: Javon Hickmon
机构: 未知
关键词: achieving increased performance, rapidly advancing field, artificial intelligence, increased performance, rapidly advancing
类目: Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY); Machine Learning (cs.LG)
备注: Bachelor’s thesis

点击查看摘要

Abstract:In the rapidly advancing field of artificial intelligence, machine perception is becoming paramount to achieving increased performance. Image classification systems are becoming increasingly integral to various applications, ranging from medical diagnostics to image generation; however, these systems often exhibit harmful biases that can lead to unfair and discriminatory outcomes. Machine Learning systems that depend on a single data modality, i.e. only images or only text, can exaggerate hidden biases present in the training data, if the data is not carefully balanced and filtered. Even so, these models can still harm underrepresented populations when used in improper contexts, such as when government agencies reinforce racial bias using predictive policing. This thesis explores the intersection of technology and ethics in the development of fair image classification models. Specifically, I focus on improving fairness and methods of using multiple modalities to combat harmful demographic bias. Integrating multimodal approaches, which combine visual data with additional modalities such as text and metadata, allows this work to enhance the fairness and accuracy of image classification systems. The study critically examines existing biases in image datasets and classification algorithms, proposes innovative methods for mitigating these biases, and evaluates the ethical implications of deploying such systems in real-world scenarios. Through comprehensive experimentation and analysis, the thesis demonstrates how multimodal techniques can contribute to more equitable and ethical AI solutions, ultimately advocating for responsible AI practices that prioritize fairness.
zh

[CV-125] MHSA: A Multi-scale Hypergraph Network for Mild Cognitive Impairment Detection via Synchronous and Attentive Fusion

【速读】: 该论文试图解决轻度认知障碍 (MCI) 的精确检测问题,特别是通过多尺度超图网络来建模脑区之间的同步关系。解决方案的关键在于设计了一种名为“多尺度超图网络通过同步和注意力融合 (MHSA)”的新框架。该框架利用相位锁定值 (PLV) 计算感兴趣区域 (ROIs) 在频谱域中的相位同步关系,并通过多尺度特征融合机制整合功能磁共振成像 (fMRI) 在时间和频谱域的动态连接特征。此外,论文还提出了动态调整 PLV 系数的策略,并基于综合的时间-频谱融合矩阵构建动态超图,从而优化每个 ROI 对相位同步的直接贡献。实验结果表明该策略的有效性。

链接: https://arxiv.org/abs/2412.12149
作者: Manman Yuan,Weiming Jia,Xiong Luo,Jiazhen Ye,Peican Zhu,Junlin Li
机构: Inner Mongolia University(内蒙古大学); University of Science and Technology Beijing(北京科技大学); Northwestern Polytechnical University(西北工业大学); Inner Mongolia Autonomous Region People’s Hospital(内蒙古自治区人民医院)
关键词: mild cognitive impairment, cognitive impairment, timely manner, Multi-scale Hypergraph Network, mild cognitive
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computer Vision and Pattern Recognition (cs.CV)
备注: International Conference on Bioinformatics and Biomedicine 2024(BIBM 2024)

点击查看摘要

Abstract:The precise detection of mild cognitive impairment (MCI) is of significant importance in preventing the deterioration of patients in a timely manner. Although hypergraphs have enhanced performance by learning and analyzing brain networks, they often only depend on vector distances between features at a single scale to infer interactions. In this paper, we deal with a more arduous challenge, hypergraph modelling with synchronization between brain regions, and design a novel framework, i.e., A Multi-scale Hypergraph Network for MCI Detection via Synchronous and Attentive Fusion (MHSA), to tackle this challenge. Specifically, our approach employs the Phase-Locking Value (PLV) to calculate the phase synchronization relationship in the spectrum domain of regions of interest (ROIs) and designs a multi-scale feature fusion mechanism to integrate dynamic connectivity features of functional magnetic resonance imaging (fMRI) from both the temporal and spectrum domains. To evaluate and optimize the direct contribution of each ROI to phase synchronization in the temporal domain, we structure the PLV coefficients dynamically adjust strategy, and the dynamic hypergraph is modelled based on a comprehensive temporal-spectrum fusion matrix. Experiments on the real-world dataset indicate the effectiveness of our strategy. The code is available at this https URL.
zh

[CV-126] SceneDiffuser: Efficient and Controllable Driving Simulation Initialization and Rollout NEURIPS2024

【速读】: 该论文试图解决自动驾驶车辆(AV)开发中场景模拟的关键问题,特别是场景初始化和闭环模拟中的可控性、真实性以及推理效率。解决方案的关键在于提出了SceneDiffuser,这是一个针对交通模拟的场景级扩散先验框架。通过引入摊销扩散(amortized diffusion),该方法在未来的模拟步骤中摊销去噪计算成本,显著减少了每个模拟步骤的推理成本(减少16倍),同时缓解了闭环模拟中的误差。此外,通过引入广义硬约束(generalized hard constraints)和基于语言的约束场景生成(language-based constrained scene generation),增强了模型的可控性。实验结果表明,该方法在Waymo Open Sim Agents Challenge中取得了顶尖的开环性能和扩散模型中最佳的闭环性能。

链接: https://arxiv.org/abs/2412.12129
作者: Chiyu Max Jiang,Yijing Bai,Andre Cornman,Christopher Davis,Xiukun Huang,Hong Jeon,Sakshum Kulshrestha,John Lambert,Shuangyu Li,Xuanyu Zhou,Carlos Fuertes,Chang Yuan,Mingxing Tan,Yin Zhou,Dragomir Anguelov
机构: Waymo LLC(Waymo有限责任公司)
关键词: autonomous vehicle, prerequisite for autonomous, Realistic and interactive, simulation, interactive scene simulation
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computer Vision and Pattern Recognition (cs.CV)
备注: Accepted to NeurIPS 2024

点击查看摘要

Abstract:Realistic and interactive scene simulation is a key prerequisite for autonomous vehicle (AV) development. In this work, we present SceneDiffuser, a scene-level diffusion prior designed for traffic simulation. It offers a unified framework that addresses two key stages of simulation: scene initialization, which involves generating initial traffic layouts, and scene rollout, which encompasses the closed-loop simulation of agent behaviors. While diffusion models have been proven effective in learning realistic and multimodal agent distributions, several challenges remain, including controllability, maintaining realism in closed-loop simulations, and ensuring inference efficiency. To address these issues, we introduce amortized diffusion for simulation. This novel diffusion denoising paradigm amortizes the computational cost of denoising over future simulation steps, significantly reducing the cost per rollout step (16x less inference steps) while also mitigating closed-loop errors. We further enhance controllability through the introduction of generalized hard constraints, a simple yet effective inference-time constraint mechanism, as well as language-based constrained scene generation via few-shot prompting of a large language model (LLM). Our investigations into model scaling reveal that increased computational resources significantly improve overall simulation realism. We demonstrate the effectiveness of our approach on the Waymo Open Sim Agents Challenge, achieving top open-loop performance and the best closed-loop performance among diffusion models.
zh

[CV-127] Seamless Optical Cloud Computing across Edge-Metro Network for Generative AI

【速读】: 该论文试图解决传统云计算中高能耗和计算安全风险的问题,关键解决方案是提出并实验验证了一种光子云计算系统 (optical cloud computing system),该系统通过边缘-城域网络 (edge-metro network) 实现无缝部署。通过将输入和模型调制为光信号,边缘节点可以直接访问光子计算中心,从而显著提高能效(实验显示为118.6 mW/TOPs),比传统电子云计算方案降低两个数量级的能耗。此外,该架构通过并行计算支持多种复杂的生成式 AI 模型,能够完成图像生成任务。

链接: https://arxiv.org/abs/2412.12126
作者: Sizhe Xing,Aolong Sun,Chengxi Wang,Yizhi Wang,Boyu Dong,Junhui Hu,Xuyu Deng,An Yan,Yingjun Liu,Fangchen Hu,Zhongya Li,Ouhan Huang,Junhao Zhao,Yingjun Zhou,Ziwei Li,Jianyang Shi,Xi Xiao,Richard Penty,Qixiang Cheng,Nan Chi,Junwen Zhang
机构: 未知
关键词: reshaped modern lifestyles, profoundly reshaped modern, generative artificial intelligence, artificial intelligence, modern lifestyles
类目: Distributed, Parallel, and Cluster Computing (cs.DC); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV); Signal Processing (eess.SP)
备注:

点击查看摘要

Abstract:The rapid advancement of generative artificial intelligence (AI) in recent years has profoundly reshaped modern lifestyles, necessitating a revolutionary architecture to support the growing demands for computational power. Cloud computing has become the driving force behind this transformation. However, it consumes significant power and faces computation security risks due to the reliance on extensive data centers and servers in the cloud. Reducing power consumption while enhancing computational scale remains persistent challenges in cloud computing. Here, we propose and experimentally demonstrate an optical cloud computing system that can be seamlessly deployed across edge-metro network. By modulating inputs and models into light, a wide range of edge nodes can directly access the optical computing center via the edge-metro network. The experimental validations show an energy efficiency of 118.6 mW/TOPs (tera operations per second), reducing energy consumption by two orders of magnitude compared to traditional electronic-based cloud computing solutions. Furthermore, it is experimentally validated that this architecture can perform various complex generative AI models through parallel computing to achieve image generation tasks.
zh

[CV-128] Unlocking the Potential of Digital Pathology: Novel Baselines for Compression

【速读】: 该论文试图解决数字病理学中全切片图像 (Whole Slide Images, WSI) 因文件大小过大而面临的压缩问题,特别是如何在保证感知图像质量和下游任务性能的同时,有效压缩这些图像。解决方案的关键在于提出了一种联合评估压缩方案的方法,不仅考虑了感知图像质量,还考虑了压缩后的图像在下游任务中的表现。研究通过引入一种基于特征相似度的新评估指标,能够更准确地反映压缩图像的实际下游性能,从而提供了一种通用且标准化的评价方法,避免了分别评估不同下游任务的需求。这一方法有助于推动数字病理学在临床中的应用。

链接: https://arxiv.org/abs/2412.13137
作者: Maximilian Fischer,Peter Neher,Peter Schüffler,Sebastian Ziegler,Shuhan Xiao,Robin Peretzke,David Clunie,Constantin Ulrich,Michael Baumgartner,Alexander Muckenhuber,Silvia Dias Almeida,Michael Götz,Jens Kleesiek,Marco Nolden,Rickmer Braren,Klaus Maier-Hein
机构: 未知
关键词: histopathological image analysis, pathological Whole Slide, Digital pathology offers, Slide Images, transform clinical practice
类目: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Digital pathology offers a groundbreaking opportunity to transform clinical practice in histopathological image analysis, yet faces a significant hurdle: the substantial file sizes of pathological Whole Slide Images (WSI). While current digital pathology solutions rely on lossy JPEG compression to address this issue, lossy compression can introduce color and texture disparities, potentially impacting clinical decision-making. While prior research addresses perceptual image quality and downstream performance independently of each other, we jointly evaluate compression schemes for perceptual and downstream task quality on four different datasets. In addition, we collect an initially uncompressed dataset for an unbiased perceptual evaluation of compression schemes. Our results show that deep learning models fine-tuned for perceptual quality outperform conventional compression schemes like JPEG-XL or WebP for further compression of WSI. However, they exhibit a significant bias towards the compression artifacts present in the training data and struggle to generalize across various compression schemes. We introduce a novel evaluation metric based on feature similarity between original files and compressed files that aligns very well with the actual downstream performance on the compressed WSI. Our metric allows for a general and standardized evaluation of lossy compression schemes and mitigates the requirement to independently assess different downstream tasks. Our study provides novel insights for the assessment of lossy compression schemes for WSI and encourages a unified evaluation of lossy compression schemes to accelerate the clinical uptake of digital pathology.
zh

[CV-129] A Knowledge-enhanced Pathology Vision-language Foundation Model for Cancer Diagnosis

【速读】: 该论文试图解决现有深度学习模型在病理任务中缺乏领域特定知识的问题,特别是在罕见肿瘤亚型的诊断中表现受限的问题。解决方案的关键在于构建了一个知识增强的病理学基础模型(Knowledge-enhanced Pathology, KEEP),通过整合疾病知识图谱(disease knowledge graph, KG)来增强视觉-语言预训练。具体来说,研究者首先构建了一个包含11,454种人类疾病和139,143个疾病属性的知识图谱,然后将大量公开的病理图像-文本对重新组织成143K个语义组,这些语义组通过疾病知识图谱的层级关系进行连接。接着,提出了一种知识增强的视觉-语言预训练方法,将疾病知识融入到层级语义组的图像和文本表示的对齐中,从而提升模型在零样本癌症诊断任务中的表现。实验结果表明,KEEP在18个多样化的基准测试中达到了最先进的性能,特别是在罕见脑癌的亚型分类中表现出强大的泛化能力。

链接: https://arxiv.org/abs/2412.13126
作者: Xiao Zhou,Luoyi Sun,Dexuan He,Wenbin Guan,Ruifen Wang,Lifeng Wang,Xin Sun,Kun Sun,Ya Zhang,Yanfeng Wang,Weidi Xie
机构: Shanghai Artificial Intelligence Laboratory; Shanghai Artificial Intelligence Laboratory; Zhejiang University; School of Artificial Intelligence, Shanghai Jiao Tong University; Department of Pathology, Xin Hua Hospital Affiliated to Shanghai Jiao Tong University School of Medicine; Department of Pathology, Xin Hua Hospital Affiliated to Shanghai Jiao Tong University School of Medicine; Department of Pathology, Xin Hua Hospital Affiliated to Shanghai Jiao Tong University School of Medicine; Clinical Research and Innovation Unit, Xinhua Hospital Affiliated to Shanghai Jiao Tong University School of Medicine; Department of Pediatric Cardiology, Xinhua Hospital Affiliated to Shanghai Jiao Tong University School of Medicine; Shanghai Artificial Intelligence Laboratory; School of Artificial Intelligence, Shanghai Jiao Tong University; Shanghai Artificial Intelligence Laboratory; Shanghai Artificial Intelligence Laboratory
关键词: Deep learning, highly robust foundation, patient cohorts, learning has enabled, enabled the development
类目: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Deep learning has enabled the development of highly robust foundation models for various pathological tasks across diverse diseases and patient cohorts. Among these models, vision-language pre-training, which leverages large-scale paired data to align pathology image and text embedding spaces, and provides a novel zero-shot paradigm for downstream tasks. However, existing models have been primarily data-driven and lack the incorporation of domain-specific knowledge, which limits their performance in cancer diagnosis, especially for rare tumor subtypes. To address this limitation, we establish a Knowledge-enhanced Pathology (KEEP) foundation model that harnesses disease knowledge to facilitate vision-language pre-training. Specifically, we first construct a disease knowledge graph (KG) that covers 11,454 human diseases with 139,143 disease attributes, including synonyms, definitions, and hypernym relations. We then systematically reorganize the millions of publicly available noisy pathology image-text pairs, into 143K well-structured semantic groups linked through the hierarchical relations of the disease KG. To derive more nuanced image and text representations, we propose a novel knowledge-enhanced vision-language pre-training approach that integrates disease knowledge into the alignment within hierarchical semantic groups instead of unstructured image-text pairs. Validated on 18 diverse benchmarks with more than 14,000 whole slide images (WSIs), KEEP achieves state-of-the-art performance in zero-shot cancer diagnostic tasks. Notably, for cancer detection, KEEP demonstrates an average sensitivity of 89.8% at a specificity of 95.0% across 7 cancer types. For cancer subtyping, KEEP achieves a median balanced accuracy of 0.456 in subtyping 30 rare brain cancers, indicating strong generalizability for diagnosing rare tumors.
zh

[CV-130] Learning of Patch-Based Smooth-Plus-Sparse Models for Image Reconstruction

【速读】: 该论文旨在解决成像中的逆问题,通过结合图像块的稀疏表示(penalized sparse representation)和无约束平滑表示来实现。其关键在于将优化问题形式化为双层问题(bilevel problem),其中内层问题采用经典算法,而外层问题通过监督学习优化字典和正则化参数。该方法通过隐式微分和基于梯度的优化实现,并在去噪、超分辨率和压缩感知磁共振成像等任务中进行了评估,结果表明其在性能上优于传统模型和深度学习方法。

链接: https://arxiv.org/abs/2412.13070
作者: Stanislas Ducotterd,Sebastian Neumayer,Michael Unser
机构: École polytechnique fédérale de Lausanne(洛桑联邦理工学院); Technische Universität Chemnitz(开姆尼茨工业大学)
关键词: penalized sparse representation, solution of inverse, combining a penalized, penalized sparse, sparse representation
类目: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Signal Processing (eess.SP)
备注:

点击查看摘要

Abstract:We aim at the solution of inverse problems in imaging, by combining a penalized sparse representation of image patches with an unconstrained smooth one. This allows for a straightforward interpretation of the reconstruction. We formulate the optimization as a bilevel problem. The inner problem deploys classical algorithms while the outer problem optimizes the dictionary and the regularizer parameters through supervised learning. The process is carried out via implicit differentiation and gradient-based optimization. We evaluate our method for denoising, super-resolution, and compressed-sensing magnetic-resonance imaging. We compare it to other classical models as well as deep-learning-based methods and show that it always outperforms the former and also the latter in some instances.
zh

[CV-131] 3D MedDiffusion: A 3D Medical Diffusion Model for Controllable and High-quality Medical Image Generation

【速读】: 该论文试图解决高分辨率三维医学图像生成中的挑战,现有方法在生成高质量3D医学图像时表现不佳,且缺乏通用的生成框架。解决方案的关键在于提出了3D Medical Diffusion (3D MedDiffusion)模型,该模型通过创新的Patch-Volume Autoencoder实现高效的图像压缩与恢复,并设计了新的噪声估计器以在扩散去噪过程中捕捉局部细节和全局结构信息。3D MedDiffusion能够生成高分辨率(最高可达512x512x512)的精细图像,并在大规模多模态(CT和MRI)和多解剖区域(从头到腿)数据集上训练,展现出强大的任务适应性和优于现有方法的生成质量。

链接: https://arxiv.org/abs/2412.13059
作者: Haoshen Wang,Zhentao Liu,Kaicong Sun,Xiaodong Wang,Dinggang Shen,Zhiming Cui
机构: ShanghaiTech University, School of Biomedical Engineering & State Key Laboratory of Advanced Medical Materials and Devices (上海科技大学,生物医学工程学院 & 先进医用材料与设备国家重点实验室); Shanghai United Imaging Intelligence Co., Ltd. (上海联影智能医疗科技有限公司); Shanghai Clinical Research and Trial Center (上海临床研究与试验中心); United Imaging Healthcare Co., Ltd. (联影医疗科技有限公司)
关键词: presents significant challenges, significant challenges due, images presents significant, medical images presents, three-dimensional nature
类目: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:The generation of medical images presents significant challenges due to their high-resolution and three-dimensional nature. Existing methods often yield suboptimal performance in generating high-quality 3D medical images, and there is currently no universal generative framework for medical imaging. In this paper, we introduce the 3D Medical Diffusion (3D MedDiffusion) model for controllable, high-quality 3D medical image generation. 3D MedDiffusion incorporates a novel, highly efficient Patch-Volume Autoencoder that compresses medical images into latent space through patch-wise encoding and recovers back into image space through volume-wise decoding. Additionally, we design a new noise estimator to capture both local details and global structure information during diffusion denoising process. 3D MedDiffusion can generate fine-detailed, high-resolution images (up to 512x512x512) and effectively adapt to various downstream tasks as it is trained on large-scale datasets covering CT and MRI modalities and different anatomical regions (from head to leg). Experimental results demonstrate that 3D MedDiffusion surpasses state-of-the-art methods in generative quality and exhibits strong generalizability across tasks such as sparse-view CT reconstruction, fast MRI reconstruction, and data augmentation.
zh

[CV-132] Stable Diffusion is a Natural Cross-Modal Decoder for Layered AI-generated Image Compression

【速读】: 该论文试图解决生成式 AI 图像 (AI-generated images, AIGIs) 在传输和压缩过程中面临的研究不足问题。解决方案的关键在于提出了一种可扩展的跨模态压缩框架,该框架通过结合多种人类可理解的模态(如语义层、结构层和纹理层)来高效捕捉和传递 AIGIs 的关键视觉信息。具体来说,该框架将图像编码为分层比特流,包括通过文本提示传递高级语义信息的语义层、使用边缘或骨架图捕捉空间细节的结构层,以及通过色图保留局部纹理的纹理层。利用 Stable Diffusion 作为后端,该框架能够有效利用这些多模态先验进行图像生成,并在极低比特率(如 0.02 bpp)下与基线方法竞争,同时支持下游编辑应用而不需要完全解码。

链接: https://arxiv.org/abs/2412.12982
作者: Ruijie Chen,Qi Mao,Zhengxue Cheng
机构: Communication University of China(中国传媒大学); Shanghai Jiao Tong University(上海交通大学)
关键词: Intelligence Generated Content, Artificial Intelligence Generated, Generated Content, garnered significant interest, Artificial Intelligence
类目: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Recent advances in Artificial Intelligence Generated Content (AIGC) have garnered significant interest, accompanied by an increasing need to transmit and compress the vast number of AI-generated images (AIGIs). However, there is a noticeable deficiency in research focused on compression methods for AIGIs. To address this critical gap, we introduce a scalable cross-modal compression framework that incorporates multiple human-comprehensible modalities, designed to efficiently capture and relay essential visual information for AIGIs. In particular, our framework encodes images into a layered bitstream consisting of a semantic layer that delivers high-level semantic information through text prompts; a structural layer that captures spatial details using edge or skeleton maps; and a texture layer that preserves local textures via a colormap. Utilizing Stable Diffusion as the backend, the framework effectively leverages these multimodal priors for image generation, effectively functioning as a decoder when these priors are encoded. Qualitative and quantitative results show that our method proficiently restores both semantic and visual details, competing against baseline approaches at extremely low bitrates ( 0.02 bpp). Additionally, our framework facilitates downstream editing applications without requiring full decoding, thereby paving a new direction for future research in AIGI compression.
zh

[CV-133] Online optimisation for dynamic electrical impedance tomography

【速读】: 该论文试图解决非线性时间离散逆问题,并提出了一种基于在线优化(online optimisation)的原始对偶方法。解决方案的关键在于通过后悔理论(regret theory)分析该方法的收敛性,并证明了完整电极模型(Complete Electrode Model, CEM)解算子在 (L^\infty) 空间上的二阶可微性,从而在电气阻抗断层成像(Electrical Impedance Tomography, EIT)中实现了对流体中移动物体的实时监测。

链接: https://arxiv.org/abs/2412.12944
作者: Neil Dizon,Jyrki Jauhiainen,Tuomo Valkonen
机构: 未知
关键词: Online optimisation studies, Electrical Impedance Tomography, optimisation studies, studies the convergence, data embedded
类目: Optimization and Control (math.OC); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:Online optimisation studies the convergence of optimisation methods as the data embedded in the problem changes. Based on this idea, we propose a primal dual online method for nonlinear time-discrete inverse problems. We analyse the method through regret theory and demonstrate its performance in real-time monitoring of moving bodies in a fluid with Electrical Impedance Tomography (EIT). To do so, we also prove the second-order differentiability of the Complete Electrode Model (CEM) solution operator on L^\infty .
zh

[CV-134] 4DRGS: 4D Radiative Gaussian Splatting for Efficient 3D Vessel Reconstruction from Sparse-View Dynamic DSA Images

【速读】: 该论文试图解决从稀疏视角动态数字减影血管造影(DSA)图像中高效重建高质量三维血管结构的问题,以减少辐射暴露并提高医疗评估的准确性。解决方案的关键在于提出了4D辐射高斯溅射(4DRGS)方法,通过使用4D辐射高斯核来表示血管结构,其中每个核具有时间不变的几何参数(位置、旋转和尺度)来建模静态血管结构,并通过紧凑的神经网络预测时间依赖的中心衰减来捕捉对比剂流动的时变响应。该方法通过X射线光栅化合成DSA图像,并与实际捕获的图像进行优化,最终从训练好的高斯核中体素化生成3D血管体积。此外,引入了累积衰减剪枝和有界尺度激活函数来进一步提升重建质量。实验结果表明,4DRGS在5分钟内即可完成训练,比现有最先进方法快32倍,显示出其在实际临床应用中的潜力。

链接: https://arxiv.org/abs/2412.12919
作者: Zhentao Liu,Ruyi Zha,Huangxuan Zhao,Hongdong Li,Zhiming Cui
机构: 1. School of Computer Science and Engineering, Beihang University, Beijing, China (计算机科学与工程学院,北京航空航天大学,北京,中国); 2. School of Software, Beihang University, Beijing, China (软件学院,北京航空航天大学,北京,中国); 3. School of Computer Science and Engineering, Beihang University, Beijing, China (计算机科学与工程学院,北京航空航天大学,北京,中国)
关键词: digital subtraction angiography, reducing radiation exposure, sparse-view dynamic digital, dynamic digital subtraction, enables accurate medical
类目: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV)
备注: Zhentao Liu and Ruyi Zha made equal contributions

点击查看摘要

Abstract:Reconstructing 3D vessel structures from sparse-view dynamic digital subtraction angiography (DSA) images enables accurate medical assessment while reducing radiation exposure. Existing methods often produce suboptimal results or require excessive computation time. In this work, we propose 4D radiative Gaussian splatting (4DRGS) to achieve high-quality reconstruction efficiently. In detail, we represent the vessels with 4D radiative Gaussian kernels. Each kernel has time-invariant geometry parameters, including position, rotation, and scale, to model static vessel structures. The time-dependent central attenuation of each kernel is predicted from a compact neural network to capture the temporal varying response of contrast agent flow. We splat these Gaussian kernels to synthesize DSA images via X-ray rasterization and optimize the model with real captured ones. The final 3D vessel volume is voxelized from the well-trained kernels. Moreover, we introduce accumulated attenuation pruning and bounded scaling activation to improve reconstruction quality. Extensive experiments on real-world patient data demonstrate that 4DRGS achieves impressive results in 5 minutes training, which is 32x faster than the state-of-the-art method. This underscores the potential of 4DRGS for real-world clinics.
zh

[CV-135] Automatic Left Ventricular Cavity Segmentation via Deep Spatial Sequential Network in 4D Computed Tomography Studies

【速读】: 该论文试图解决左心室腔(LVC)在时间序列心脏图像中的自动分割问题,特别是在心室收缩末期(ES)图像中,由于形状不规则且边界模糊,现有方法表现不佳的问题。解决方案的关键在于提出了一种空间序列(SS)网络,通过无监督方式学习LVC的形变和运动特征,并结合双向学习(BL)机制,利用时间序列图像的正向和反向信息来增强分割的准确性和一致性。实验结果表明,该方法在心脏CT和MRI数据集上均优于现有方法,展示了其通用性。

链接: https://arxiv.org/abs/2412.12853
作者: Yuyu Guo,Lei Bi,Zhengbin Zhu,David Dagan Feng,Ruiyan Zhang,Qian Wang,Jinman Kim
机构: Shanghai Jiao Tong University (上海交通大学); University of Sydney (悉尼大学); Ruijin Hospital, Shanghai Jiaotong University School of Medicine (上海交通大学医学院瑞金医院)
关键词: multiple time points, time points, left ventricular cavity, single time points, temporal image sequences
类目: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV)
备注: 9 pages

点击查看摘要

Abstract:Automated segmentation of left ventricular cavity (LVC) in temporal cardiac image sequences (multiple time points) is a fundamental requirement for quantitative analysis of its structural and functional changes. Deep learning based methods for the segmentation of LVC are the state of the art; however, these methods are generally formulated to work on single time points, and fails to exploit the complementary information from the temporal image sequences that can aid in segmentation accuracy and consistency among the images across the time points. Furthermore, these segmentation methods perform poorly in segmenting the end-systole (ES) phase images, where the left ventricle deforms to the smallest irregular shape, and the boundary between the blood chamber and myocardium becomes inconspicuous. To overcome these limitations, we propose a new method to automatically segment temporal cardiac images where we introduce a spatial sequential (SS) network to learn the deformation and motion characteristics of the LVC in an unsupervised manner; these characteristics were then integrated with sequential context information derived from bi-directional learning (BL) where both chronological and reverse-chronological directions of the image sequence were used. Our experimental results on a cardiac computed tomography (CT) dataset demonstrated that our spatial-sequential network with bi-directional learning (SS-BL) method outperformed existing methods for LVC segmentation. Our method was also applied to MRI cardiac dataset and the results demonstrated the generalizability of our method.
zh

[CV-136] raining a Distributed Acoustic Sensing Traffic Monitoring Network With Video Inputs

【速读】: 该论文试图解决在人口密集区域进行实时交通监控的问题,提出了一种将分布式声学传感 (Distributed Acoustic Sensing, DAS) 数据与视觉信息相结合的新方法。解决方案的关键在于利用摄像头输入的YOLO算法生成的车辆位置和分类数据作为标签,训练一个仅使用DAS数据的检测与分类神经网络。这种方法不仅提高了检测和分类的准确性(超过94%),还将误报率控制在1.2%左右,同时具备隐私保护、基础设施成本低和可扩展性强的优势。

链接: https://arxiv.org/abs/2412.12743
作者: Khen Cohen,Liav Hen,Ariel Lellouch
机构: School of Physics and Astronomy, Tel-Aviv University, Tel-Aviv 69978, Israel(特拉维夫大学物理与天文学学院); School of Electrical Engineering, Tel-Aviv University, Tel-Aviv 69978, Israel(特拉维夫大学电气工程学院); Porter School of the Environment and Earth Sciences, Tel-Aviv University, Tel-Aviv 69978, Israel(特拉维夫大学环境和地球科学学院)
关键词: Distributed Acoustic Sensing, Distributed Acoustic, Acoustic Sensing, densely populated areas, populated areas
类目: Geophysics (physics.geo-ph); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Signal Processing (eess.SP); Optics (physics.optics)
备注: 12 pages, 11 figures, 5 appendices. Shared dataset in: this https URL

点击查看摘要

Abstract:Distributed Acoustic Sensing (DAS) has emerged as a promising tool for real-time traffic monitoring in densely populated areas. In this paper, we present a novel concept that integrates DAS data with co-located visual information. We use YOLO-derived vehicle location and classification from camera inputs as labeled data to train a detection and classification neural network utilizing DAS data only. Our model achieves a performance exceeding 94% for detection and classification, and about 1.2% false alarm rate. We illustrate the model’s application in monitoring traffic over a week, yielding statistical insights that could benefit future smart city developments. Our approach highlights the potential of combining fiber-optic sensors with visual information, focusing on practicality and scalability, protecting privacy, and minimizing infrastructure costs. To encourage future research, we share our dataset.
zh

[CV-137] Accelerating lensed quasars discovery and modeling with physics-informed variational autoencoders

【速读】: 该论文试图解决在天文图像中检测强引力透镜(strongly lensed quasars)的难题,由于非透镜物体(non-lensing objects)的普遍存在,传统方法难以有效识别。解决方案的关键在于开发了一种基于物理信息变分自编码器(physics-informed variational autoencoder)的生成式深度学习模型,称为VariLens。该模型集成了图像重建、目标分类和透镜建模三个核心模块,能够快速判断一个物体是否为透镜系统,并在毫秒级时间内估计单个等温椭球体(singular isothermal ellipsoid, SIE)质量模型的关键参数,如爱因斯坦半径(Einstein radius, \theta_\mathrmE)、透镜中心和椭率。通过与传统透镜建模方法的对比,VariLens在已知透镜系统的参数估计上表现出良好的一致性,并成功从大规模数据集中筛选出新的透镜候选体,展示了自动化深度学习管道在高效检测和建模强引力透镜方面的潜力。

链接: https://arxiv.org/abs/2412.12709
作者: Irham T. Andika,Stefan Schuldt,Sherry H. Suyu,Satadru Bag,Raoul Cañameras,Alejandra Melo,Claudio Grillo,James H. H. Chan
机构: 未知
关键词: provide valuable insights, Strongly lensed quasars, quasars provide valuable, Strongly lensed, cosmic expansion
类目: Astrophysics of Galaxies (astro-ph.GA); Cosmology and Nongalactic Astrophysics (astro-ph.CO); Instrumentation and Methods for Astrophysics (astro-ph.IM); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
备注: Submitted to the Astronomy Astrophysics journal. The paper consists of 17 main pages, 14 figures, and 5 tables. We welcome feedback and comments from readers!

点击查看摘要

Abstract:Strongly lensed quasars provide valuable insights into the rate of cosmic expansion, the distribution of dark matter in foreground deflectors, and the characteristics of quasar hosts. However, detecting them in astronomical images is difficult due to the prevalence of non-lensing objects. To address this challenge, we developed a generative deep learning model called VariLens, built upon a physics-informed variational autoencoder. This model seamlessly integrates three essential modules: image reconstruction, object classification, and lens modeling, offering a fast and comprehensive approach to strong lens analysis. VariLens is capable of rapidly determining both (1) the probability that an object is a lens system and (2) key parameters of a singular isothermal ellipsoid (SIE) mass model – including the Einstein radius ( \theta_\mathrmE ), lens center, and ellipticity – in just milliseconds using a single CPU. A direct comparison of VariLens estimates with traditional lens modeling for 20 known lensed quasars within the Subaru Hyper Suprime-Cam (HSC) footprint shows good agreement, with both results consistent within 2\sigma for systems with \theta_\mathrmE3 arcsecs. To identify new lensed quasar candidates, we begin with an initial sample of approximately 80 million sources, combining HSC data with multiwavelength information from various surveys. After applying a photometric preselection aimed at locating z1.5 sources, the number of candidates is reduced to 710,966. Subsequently, VariLens highlights 13,831 sources, each showing a high likelihood of being a lens. A visual assessment of these objects results in 42 promising candidates that await spectroscopic confirmation. These results underscore the potential of automated deep learning pipelines to efficiently detect and model strong lenses in large datasets.
zh

[CV-138] a2z-1 for Multi-Disease Detection in Abdomen-Pelvis CT: External Validation and Performance Analysis Across 21 Conditions

【速读】: 该论文旨在评估a2z-1人工智能模型在分析腹部-骨盆CT扫描中的性能和泛化能力,特别是针对21种时间敏感且可操作的发现。解决方案的关键在于通过大规模回顾性分析和跨两个不同医疗系统的验证,展示了该模型在多种条件下的高平均AUC(0.931),并在关键发现如小肠梗阻和急性胰腺炎中表现尤为突出(AUC分别为0.958和0.961)。此外,子组分析显示模型在不同患者性别、年龄组和不同成像协议下的准确性一致,且与放射科医生报告的对比显示,a2z-1能够识别被忽略的发现,表明其在质量保证应用中的潜力。

链接: https://arxiv.org/abs/2412.12629
作者: Pranav Rajpurkar,Julian N. Acosta,Siddhant Dogra,Jaehwan Jeong,Deepanshu Jindal,Michael Moritz,Samir Rajpurkar
机构: a2z Radiology AI
关键词: artificial intelligence, time-sensitive and actionable, present a comprehensive, designed to analyze, analyze abdomen-pelvis
类目: Image and Video Processing (eess.IV); Artificial Intelligence (cs.AI); Computer Vision and Pattern Recognition (cs.CV)
备注:

点击查看摘要

Abstract:We present a comprehensive evaluation of a2z-1, an artificial intelligence (AI) model designed to analyze abdomen-pelvis CT scans for 21 time-sensitive and actionable findings. Our study focuses on rigorous assessment of the model’s performance and generalizability. Large-scale retrospective analysis demonstrates an average AUC of 0.931 across 21 conditions. External validation across two distinct health systems confirms consistent performance (AUC 0.923), establishing generalizability to different evaluation scenarios, with notable performance in critical findings such as small bowel obstruction (AUC 0.958) and acute pancreatitis (AUC 0.961). Subgroup analysis shows consistent accuracy across patient sex, age groups, and varied imaging protocols, including different slice thicknesses and contrast administration types. Comparison of high-confidence model outputs to radiologist reports reveals instances where a2z-1 identified overlooked findings, suggesting potential for quality assurance applications.
zh

[CV-139] Predicting Internet Connectivity in Schools: A Feasibility Study Leveraging Multi-modal Data and Location Encoders in Low-Resource Settings

【速读】: 该论文试图解决学校互联网连接性信息的获取问题,特别是在资源有限的国家和地区。传统基于调查的方法成本高且难以实施,因此论文提出利用地球观测数据 (Earth Observation, EO) 和机器学习 (Machine Learning, ML) 的结合,通过卫星图像和地面辅助数据来预测学校的互联网连接性。解决方案的关键在于创建一个多模态的、免费可用的卫星图像和调查信息数据集,并利用最新的地理空间感知位置编码器和欧洲空间局 phi-lab 的地理空间感知基础模型,以实现对博茨瓦纳和卢旺达学校互联网连接性的准确预测。该方法不仅提高了预测的准确性和F1分数,还显著降低了误报率,展示了在低资源环境下支持数据驱动数字基础设施发展的实用途径。

链接: https://arxiv.org/abs/2412.12188
作者: Kelsey Doerksen,Casper Fibaek,Rochelle Schneider,Do-Hyung Kim,Isabelle Tingzon
机构: UNICEF(联合国儿童基金会); University of Oxford (牛津大学); European Space Agency ϕitalic-ϕ\phiitalic_ϕ-lab (欧洲空间局ϕitalic-ϕ\phiitalic_ϕ实验室)
关键词: digital literary skills, European Space Agency, Internet connectivity, digital infrastructure development, school internet connectivity
类目: Image and Video Processing (eess.IV); Artificial Intelligence (cs.AI); Computer Vision and Pattern Recognition (cs.CV); Social and Information Networks (cs.SI)
备注:

点击查看摘要

Abstract:Internet connectivity in schools is critical to provide students with the digital literary skills necessary to compete in modern economies. In order for governments to effectively implement digital infrastructure development in schools, accurate internet connectivity information is required. However, traditional survey-based methods can exceed the financial and capacity limits of governments. Open-source Earth Observation (EO) datasets have unlocked our ability to observe and understand socio-economic conditions on Earth from space, and in combination with Machine Learning (ML), can provide the tools to circumvent costly ground-based survey methods to support infrastructure development. In this paper, we present our work on school internet connectivity prediction using EO and ML. We detail the creation of our multi-modal, freely-available satellite imagery and survey information dataset, leverage the latest geographically-aware location encoders, and introduce the first results of using the new European Space Agency phi-lab geographically-aware foundational model to predict internet connectivity in Botswana and Rwanda. We find that ML with EO and ground-based auxiliary data yields the best performance in both countries, for accuracy, F1 score, and False Positive rates, and highlight the challenges of internet connectivity prediction from space with a case study in Kigali, Rwanda. Our work showcases a practical approach to support data-driven digital infrastructure development in low-resource settings, leveraging freely available information, and provide cleaned and labelled datasets for future studies to the community through a unique collaboration between UNICEF and the European Space Agency phi-lab.
zh

人工智能

[AI-0] ExBody2: Advanced Expressive Humanoid Whole-Body Control

链接: https://arxiv.org/abs/2412.13196
作者: Mazeyu Ji,Xuanbin Peng,Fangchen Liu,Jialong Li,Ge Yang,Xuxin Cheng,Xiaolong Wang
关键词: performing expressive motions, enables real-world humanoid, real-world humanoid robots, maintain stability, stability while performing
类目: Robotics (cs.RO); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注: website: this https URL

点击查看摘要

Abstract:This paper enables real-world humanoid robots to maintain stability while performing expressive motions like humans do. We propose ExBody2, a generalized whole-body tracking framework that can take any reference motion inputs and control the humanoid to mimic the motion. The model is trained in simulation with Reinforcement Learning and then transferred to the real world. It decouples keypoint tracking with velocity control, and effectively leverages a privileged teacher policy to distill precise mimic skills into the target student policy, which enables high-fidelity replication of dynamic movements such as running, crouching, dancing, and other challenging motions. We present a comprehensive qualitative and quantitative analysis of crucial design factors in the paper. We conduct our experiments on two humanoid platforms and demonstrate the superiority of our approach against state-of-the-arts, providing practical guidelines to pursue the extreme of whole-body control for humanoid robots.

[AI-1] d Quantile Gradient Updates for Quantile-Constrained Reinforcement Learning AAAI-25 AAAI

链接: https://arxiv.org/abs/2412.13184
作者: Chenglin Li,Guangchun Ruan,Hua Geng
关键词: Safe reinforcement learning, learn reward-maximizing policies, Safe reinforcement, reinforcement learning, popular and versatile
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: Accepted by the 39th AAAI Conference on Artificial Intelligence (AAAI-25)

点击查看摘要

Abstract:Safe reinforcement learning (RL) is a popular and versatile paradigm to learn reward-maximizing policies with safety guarantees. Previous works tend to express the safety constraints in an expectation form due to the ease of implementation, but this turns out to be ineffective in maintaining safety constraints with high probability. To this end, we move to the quantile-constrained RL that enables a higher level of safety without any expectation-form approximations. We directly estimate the quantile gradients through sampling and provide the theoretical proofs of convergence. Then a tilted update strategy for quantile gradients is implemented to compensate the asymmetric distributional density, with a direct benefit of return performance. Experiments demonstrate that the proposed model fully meets safety requirements (quantile constraints) while outperforming the state-of-the-art benchmarks with higher return.

[AI-2] SafeAgent Bench: A Benchmark for Safe Task Planning of Embodied LLM Agents ICRA2024

链接: https://arxiv.org/abs/2412.13178
作者: Sheng Yin,Xianghe Pang,Yuanzhuo Ding,Menglan Chen,Yutong Bi,Yichen Xiong,Wenhao Huang,Zhen Xiang,Jing Shao,Siheng Chen
关键词: large language models, execute complicated instructions, language models, large language, natural language
类目: Cryptography and Security (cs.CR); Artificial Intelligence (cs.AI); Robotics (cs.RO)
*备注: 21 pages, 14 tables, 7 figures, submitted to ICRA 2024

点击查看摘要

Abstract:With the integration of large language models (LLMs), embodied agents have strong capabilities to execute complicated instructions in natural language, paving a way for the potential deployment of embodied robots. However, a foreseeable issue is that those embodied agents can also flawlessly execute some hazardous tasks, potentially causing damages in real world. To study this issue, we present SafeAgentBench – a new benchmark for safety-aware task planning of embodied LLM agents. SafeAgentBench includes: (1) a new dataset with 750 tasks, covering 10 potential hazards and 3 task types; (2) SafeAgentEnv, a universal embodied environment with a low-level controller, supporting multi-agent execution with 17 high-level actions for 8 state-of-the-art baselines; and (3) reliable evaluation methods from both execution and semantic perspectives. Experimental results show that the best-performing baseline gets 69% success rate for safe tasks, but only 5% rejection rate for hazardous tasks, indicating significant safety risks. More details and codes are available at this https URL.

[AI-3] SWAN: Preprocessing SGD Enables Adam-Level Performance On LLM Training With Significant Memory Reduction

链接: https://arxiv.org/abs/2412.13148
作者: Chao Ma,Wenbo Gong,Meyer Scetbon,Edward Meeds
关键词: Adaptive optimizers, success of large, large language models, Adaptive, Kingma
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Adaptive optimizers such as Adam (Kingma Ba, 2015) have been central to the success of large language models. However, they maintain additional moving average states throughout training, which results in memory requirements several times greater than the model. This overhead imposes constraints on scalability and computational efficiency. On the other hand, while stochastic gradient descent (SGD) is optimal in terms of memory efficiency, their capability in LLM training is limited (Zhao et al., 2024b). To address this dilemma, we show that pre-processing SGD is sufficient to reach Adam-level performance on LLMs. Specifically, we propose to preprocess the instantaneous stochastic gradients with two simple operators: \mathttGradNorm and \mathttGradWhitening . \mathttGradNorm stabilizes gradient distributions, and \mathttGradWhitening counteracts the local curvature of the loss landscape, respectively. This results in SWAN (SGD with Whitening And Normalization), a stochastic optimizer that eliminates the need to store any accumulative state variables. Empirically, SWAN has the same memory footprint as SGD, achieving \approx 50% reduction on total end-to-end memory compared to Adam. In language modeling tasks, SWAN demonstrates the same or even a substantial improvement over Adam. Specifically, when pre-training the LLaMa model with 350M and 1.3B parameters, SWAN achieves a 2x speedup by reaching the same evaluation perplexity in less than half tokens seen. Subjects: Machine Learning (cs.LG); Artificial Intelligence (cs.AI) Cite as: arXiv:2412.13148 [cs.LG] (or arXiv:2412.13148v1 [cs.LG] for this version) https://doi.org/10.48550/arXiv.2412.13148 Focus to learn more arXiv-issued DOI via DataCite (pending registration)

[AI-4] Agnosticism About Artificial Consciousness

链接: https://arxiv.org/abs/2412.13145
作者: Tom McClelland
关键词: artificial consciousness, consciousness, artificial, evidence, conscious experiences
类目: Artificial Intelligence (cs.AI)
*备注: 20 pages

点击查看摘要

Abstract:Could an AI have conscious experiences? Any answer to this question should conform to Evidentialism - that is, it should be based not on intuition, dogma or speculation but on solid scientific evidence. I argue that such evidence is hard to come by and that the only justifiable stance on the prospects of artificial consciousness is agnosticism. In the current debate, the main division is between biological views that are sceptical of artificial consciousness and functional views that are sympathetic to it. I argue that both camps make the same mistake of over-estimating what the evidence tells us. Scientific insights into consciousness have been achieved through the study of conscious organisms. Although this has enabled cautious assessments of consciousness in various creatures, extending this to AI faces serious obstacles. AI thus presents consciousness researchers with a dilemma: either reach a verdict on artificial consciousness but violate Evidentialism; or respect Evidentialism but offer no verdict on the prospects of artificial consciousness. The dominant trend in the literature has been to take the first option while purporting to follow the scientific evidence. I argue that if we truly follow the evidence, we must take the second option and adopt agnosticism.

[AI-5] Previous Knowledge Utilization In Online Anytime Belief Space Planning

链接: https://arxiv.org/abs/2412.13128
作者: Michael Novitsky,Moran Barenboim,Vadim Indelman
关键词: uncertainty remains, remains a critical, critical challenge, challenge in robotics, Carlo Tree Search
类目: Artificial Intelligence (cs.AI); Robotics (cs.RO)
*备注: 10 pages, 4 figures, will be submitted to IEEE Robotics and Automation Letters (RA-L)

点击查看摘要

Abstract:Online planning under uncertainty remains a critical challenge in robotics and autonomous systems. While tree search techniques are commonly employed to construct partial future trajectories within computational constraints, most existing methods discard information from previous planning sessions considering continuous spaces. This study presents a novel, computationally efficient approach that leverages historical planning data in current decision-making processes. We provide theoretical foundations for our information reuse strategy and introduce an algorithm based on Monte Carlo Tree Search (MCTS) that implements this approach. Experimental results demonstrate that our method significantly reduces computation time while maintaining high performance levels. Our findings suggest that integrating historical planning information can substantially improve the efficiency of online decision-making in uncertain environments, paving the way for more responsive and adaptive autonomous systems.

[AI-6] SMOSE: Sparse Mixture of Shallow Experts for Interpretable Reinforcement Learning in Continuous Control Tasks AAAI-25 AAAI

链接: https://arxiv.org/abs/2412.13053
作者: Mátyás Vincze,Laura Ferrarotti,Leonardo Lucio Custode,Bruno Lepri,Giovanni Iacca
关键词: Continuous control tasks, Continuous control, involve high-dimensional, Continuous, control tasks
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: To be published in the Proceedings of the 39th AAAI Conference on Artificial Intelligence (AAAI-25)

点击查看摘要

Abstract:Continuous control tasks often involve high-dimensional, dynamic, and non-linear environments. State-of-the-art performance in these tasks is achieved through complex closed-box policies that are effective, but suffer from an inherent opacity. Interpretable policies, while generally underperforming compared to their closed-box counterparts, advantageously facilitate transparent decision-making within automated systems. Hence, their usage is often essential for diagnosing and mitigating errors, supporting ethical and legal accountability, and fostering trust among stakeholders. In this paper, we propose SMOSE, a novel method to train sparsely activated interpretable controllers, based on a top-1 Mixture-of-Experts architecture. SMOSE combines a set of interpretable decisionmakers, trained to be experts in different basic skills, and an interpretable router that assigns tasks among the experts. The training is carried out via state-of-the-art Reinforcement Learning algorithms, exploiting load-balancing techniques to ensure fair expert usage. We then distill decision trees from the weights of the router, significantly improving the ease of interpretation. We evaluate SMOSE on six benchmark environments from MuJoCo: our method outperforms recent interpretable baselines and narrows the gap with noninterpretable state-of-the-art algorithms

[AI-7] Relational Neurosymbolic Markov Models AAAI2025

链接: https://arxiv.org/abs/2412.13023
作者: Lennert De Smet,Gabriele Venturato,Luc De Raedt,Giuseppe Marra
关键词: natural language processing, language processing, natural language, Sequential, Sequential problems
类目: Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注: Accepted at AAAI 2025

点击查看摘要

Abstract:Sequential problems are ubiquitous in AI, such as in reinforcement learning or natural language processing. State-of-the-art deep sequential models, like transformers, excel in these settings but fail to guarantee the satisfaction of constraints necessary for trustworthy deployment. In contrast, neurosymbolic AI (NeSy) provides a sound formalism to enforce constraints in deep probabilistic models but scales exponentially on sequential problems. To overcome these limitations, we introduce relational neurosymbolic Markov models (NeSy-MMs), a new class of end-to-end differentiable sequential models that integrate and provably satisfy relational logical constraints. We propose a strategy for inference and learning that scales on sequential settings, and that combines approximate Bayesian inference, automated reasoning, and gradient estimation. Our experiments show that NeSy-MMs can solve problems beyond the current state-of-the-art in neurosymbolic AI and still provide strong guarantees with respect to desired properties. Moreover, we show that our models are more interpretable and that constraints can be adapted at test time to out-of-distribution scenarios.

[AI-8] Neural Control and Certificate Repair via Runtime Monitoring

链接: https://arxiv.org/abs/2412.12996
作者: Emily Yu,Đorđe Žikelić,Thomas A. Henzinger
关键词: solving highly non-linear, Learning-based methods provide, highly non-linear control, classical control methods, certificate function
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Learning-based methods provide a promising approach to solving highly non-linear control tasks that are often challenging for classical control methods. To ensure the satisfaction of a safety property, learning-based methods jointly learn a control policy together with a certificate function for the property. Popular examples include barrier functions for safety and Lyapunov functions for asymptotic stability. While there has been significant progress on learning-based control with certificate functions in the white-box setting, where the correctness of the certificate function can be formally verified, there has been little work on ensuring their reliability in the black-box setting where the system dynamics are unknown. In this work, we consider the problems of certifying and repairing neural network control policies and certificate functions in the black-box setting. We propose a novel framework that utilizes runtime monitoring to detect system behaviors that violate the property of interest under some initially trained neural network policy and certificate. These violating behaviors are used to extract new training data, that is used to re-train the neural network policy and the certificate function and to ultimately repair them. We demonstrate the effectiveness of our approach empirically by using it to repair and to boost the safety rate of neural network policies learned by a state-of-the-art method for learning-based control on two autonomous system control tasks.

[AI-9] Cluster-guided Contrastive Class-imbalanced Graph Classification AAAI-25 AAAI

链接: https://arxiv.org/abs/2412.12984
作者: Wei Ju,Zhengyang Mao,Siyu Yi,Yifang Qin,Yiyang Gu,Zhiping Xiao,Jianhao Shen,Ziyue Qiao,Ming Zhang
关键词: imbalanced class distribution, class-imbalanced graph classification, studies the problem, classifying the categories, graph
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Information Retrieval (cs.IR); Social and Information Networks (cs.SI)
*备注: Accepted by Proceedings of the Thirty-Ninth AAAI Conference on Artificial Intelligence (AAAI-25)

点击查看摘要

Abstract:This paper studies the problem of class-imbalanced graph classification, which aims at effectively classifying the categories of graphs in scenarios with imbalanced class distribution. Despite the tremendous success of graph neural networks (GNNs), their modeling ability for imbalanced graph-structured data is inadequate, which typically leads to predictions biased towards the majority classes. Besides, existing class-imbalanced learning methods in visions may overlook the rich graph semantic substructures of the majority classes and excessively emphasize learning from the minority classes. To tackle this issue, this paper proposes a simple yet powerful approach called C ^3 GNN that incorporates the idea of clustering into contrastive learning to enhance class-imbalanced graph classification. Technically, C ^3 GNN clusters graphs from each majority class into multiple subclasses, ensuring they have similar sizes to the minority class, thus alleviating class imbalance. Additionally, it utilizes the Mixup technique to synthesize new samples and enrich the semantic information of each subclass, and leverages supervised contrastive learning to hierarchically learn effective graph representations. In this way, we can not only sufficiently explore the semantic substructures within the majority class but also effectively alleviate excessive focus on the minority class. Extensive experiments on real-world graph benchmark datasets verify the superior performance of our proposed method.

[AI-10] wo Layer Walk: A Community-Aware Graph Embedding

链接: https://arxiv.org/abs/2412.12933
作者: He Yu,Jing Liu
关键词: Community structures, critical for understanding, understanding the mesoscopic, mesoscopic organization, bridging local
类目: ocial and Information Networks (cs.SI); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Community structures are critical for understanding the mesoscopic organization of networks, bridging local and global patterns. While methods such as DeepWalk and node2vec capture local positional information through random walks, they fail to preserve community structures. Other approaches like modularized nonnegative matrix factorization and evolutionary algorithms address this gap but are computationally expensive and unsuitable for large-scale networks. To overcome these limitations, we propose Two Layer Walk (TLWalk), a novel graph embedding algorithm that incorporates hierarchical community structures. TLWalk balances intra- and inter-community relationships through a community-aware random walk mechanism without requiring additional parameters. Theoretical analysis demonstrates that TLWalk effectively mitigates locality bias. Experiments on benchmark datasets show that TLWalk outperforms state-of-the-art methods, achieving up to 3.2% accuracy gains for link prediction tasks. By encoding dense local and sparse global structures, TLWalk proves robust and scalable across diverse networks, offering an efficient solution for network analysis.

[AI-11] Spectra of Cardinality Queries over Description Logic Knowledge Bases

链接: https://arxiv.org/abs/2412.12929
作者: Quentin Manière,Marcin Przybyłko
关键词: Description Logic ontologies, Description Logic, Recent works, coupled with Description, counting queries coupled
类目: Artificial Intelligence (cs.AI); Computational Complexity (cs.CC); Logic in Computer Science (cs.LO)
*备注: 26 pages

点击查看摘要

Abstract:Recent works have explored the use of counting queries coupled with Description Logic ontologies. The answer to such a query in a model of a knowledge base is either an integer or \infty , and its spectrum is the set of its answers over all models. While it is unclear how to compute and manipulate such a set in general, we identify a class of counting queries whose spectra can be effectively represented. Focusing on atomic counting queries, we pinpoint the possible shapes of a spectrum over \mathcalALCIF ontologies: they are essentially the subsets of \mathbbN \cup \ \infty \ closed under addition. For most sublogics of \mathcalALCIF , we show that possible spectra enjoy simpler shapes, being [ m, \infty ] or variations thereof. To obtain our results, we refine constructions used for finite model reasoning and notably rely on a cycle-reversion technique for the Horn fragment of \mathcalALCIF . We also study the data complexity of computing the proposed effective representation and establish the \mathsfFP^\mathsfNP[\log] -completeness of this task under several settings.

[AI-12] A Comparative Study of Pruning Methods in Transformer-based Time Series Forecasting

链接: https://arxiv.org/abs/2412.12883
作者: Nicholas Kiefer,Arvid Weyrauch,Muhammed Öz,Achim Streit,Markus Götz,Charlotte Debus
关键词: current landscape, landscape in time-series, Transformer-based models, parameter count, Pruning
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: 16 pages, 5 figures, submitted to ACM Transactions on Intelligent Systems and Technology

点击查看摘要

Abstract:The current landscape in time-series forecasting is dominated by Transformer-based models. Their high parameter count and corresponding demand in computational resources pose a challenge to real-world deployment, especially for commercial and scientific applications with low-power embedded devices. Pruning is an established approach to reduce neural network parameter count and save compute. However, the implications and benefits of pruning Transformer-based models for time series forecasting are largely unknown. To close this gap, we provide a comparative benchmark study by evaluating unstructured and structured pruning on various state-of-the-art multivariate time series models. We study the effects of these pruning strategies on model predictive performance and computational aspects like model size, operations, and inference time. Our results show that certain models can be pruned even up to high sparsity levels, outperforming their dense counterpart. However, fine-tuning pruned models is necessary. Furthermore, we demonstrate that even with corresponding hardware and software support, structured pruning is unable to provide significant time savings.

[AI-13] Bayesian Persuasion with Externalities: Exploiting Agent Types AAAI2025

链接: https://arxiv.org/abs/2412.12859
作者: Jonathan Shaki,Jiarui Gan,Sarit Kraus
关键词: Bayesian persuasion problem, Bayesian persuasion, optimal signaling strategies, study a Bayesian, optimal signaling
类目: Artificial Intelligence (cs.AI); Computer Science and Game Theory (cs.GT)
*备注: to be published in AAAI 2025

点击查看摘要

Abstract:We study a Bayesian persuasion problem with externalities. In this model, a principal sends signals to inform multiple agents about the state of the world. Simultaneously, due to the existence of externalities in the agents’ utilities, the principal also acts as a correlation device to correlate the agents’ actions. We consider the setting where the agents are categorized into a small number of types. Agents of the same type share identical utility functions and are treated equitably in the utility functions of both other agents and the principal. We study the problem of computing optimal signaling strategies for the principal, under three different types of signaling channels: public, private, and semi-private. Our results include revelation-principle-style characterizations of optimal signaling strategies, linear programming formulations, and analysis of in/tractability of the optimization problems. It is demonstrated that when the maximum number of deviating agents is bounded by a constant, our LP-based formulations compute optimal signaling strategies in polynomial time. Otherwise, the problems are NP-hard.

[AI-14] Efficient Speech Command Recognition Leveraging Spiking Neural Network and Curriculum Learning-based Knowledge Distillation

链接: https://arxiv.org/abs/2412.12858
作者: Jiaqi Wang,Liutao Yu,Liwei Huang,Chenlin Zhou,Han Zhang,Zhenxi Song,Min Zhang,Zhengyu Ma,Zhiguo Zhang
关键词: naturally utilizing embedded, spiking neural networks, utilizing embedded time, time steps, embedded time sequences
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Human-Computer Interaction (cs.HC)
*备注: Under Review

点击查看摘要

Abstract:The intrinsic dynamics and event-driven nature of spiking neural networks (SNNs) make them excel in processing temporal information by naturally utilizing embedded time sequences as time steps. Recent studies adopting this approach have demonstrated SNNs’ effectiveness in speech command recognition, achieving high performance by employing large time steps for long time sequences. However, the large time steps lead to increased deployment burdens for edge computing applications. Thus, it is important to balance high performance and low energy consumption when detecting temporal patterns in edge devices. Our solution comprises two key components. 1). We propose a high-performance fully spike-driven framework termed SpikeSCR, characterized by a global-local hybrid structure for efficient representation learning, which exhibits long-term learning capabilities with extended time steps. 2). To further fully embrace low energy consumption, we propose an effective knowledge distillation method based on curriculum learning (KDCL), where valuable representations learned from the easy curriculum are progressively transferred to the hard curriculum with minor loss, striking a trade-off between power efficiency and high performance. We evaluate our method on three benchmark datasets: the Spiking Heidelberg Dataset (SHD), the Spiking Speech Commands (SSC), and the Google Speech Commands (GSC) V2. Our experimental results demonstrate that SpikeSCR outperforms current state-of-the-art (SOTA) methods across these three datasets with the same time steps. Furthermore, by executing KDCL, we reduce the number of time steps by 60% and decrease energy consumption by 54.8% while maintaining comparable performance to recent SOTA results. Therefore, this work offers valuable insights for tackling temporal processing challenges with long time sequences in edge neuromorphic computing systems.

[AI-15] ClarityEthic: Explainable Moral Judgment Utilizing Contrastive Ethical Insights from Large Language Models

链接: https://arxiv.org/abs/2412.12848
作者: Yuxi Sun,Wei Gao,Jing Ma,Hongzhan Lin,Ziyang Luo,Wenxuan Zhang
关键词: Large Language Models, Language Models, Large Language, ensuring their safety, rise and widespread
类目: Computers and Society (cs.CY); Artificial Intelligence (cs.AI); Social and Information Networks (cs.SI)
*备注:

点击查看摘要

Abstract:With the rise and widespread use of Large Language Models (LLMs), ensuring their safety is crucial to prevent harm to humans and promote ethical behaviors. However, directly assessing value valence (i.e., support or oppose) by leveraging large-scale data training is untrustworthy and inexplainable. We assume that emulating humans to rely on social norms to make moral decisions can help LLMs understand and predict moral judgment. However, capturing human values remains a challenge, as multiple related norms might conflict in specific contexts. Consider norms that are upheld by the majority and promote the well-being of society are more likely to be accepted and widely adopted (e.g., “don’t cheat,”). Therefore, it is essential for LLM to identify the appropriate norms for a given scenario before making moral decisions. To this end, we introduce a novel moral judgment approach called \textitClarityEthic that leverages LLMs’ reasoning ability and contrastive learning to uncover relevant social norms for human actions from different perspectives and select the most reliable one to enhance judgment accuracy. Extensive experiments demonstrate that our method outperforms state-of-the-art approaches in moral judgment tasks. Moreover, human evaluations confirm that the generated social norms provide plausible explanations that support the judgments. This suggests that modeling human moral judgment with the emulating humans moral strategy is promising for improving the ethical behaviors of LLMs.

[AI-16] From An LLM Swarm To A PDDL-Empowered HIVE: Planning Self-Executed Instructions In A Multi-Modal Jungle

链接: https://arxiv.org/abs/2412.12839
作者: Kaustubh Vyas,Damien Graux,Yijun Yang,Sébastien Montella,Chenxin Diao,Wendi Zhou,Pavlos Vougiouklis,Ruofei Lai,Yang Ren,Keshuang Li,Jeff Z. Pan
关键词: deep models’ ecosystem, agent-based solutions, models’ ecosystem, call for agent-based, leverage the ever-increasing
类目: Artificial Intelligence (cs.AI)
*备注: Under review

点击查看摘要

Abstract:In response to the call for agent-based solutions that leverage the ever-increasing capabilities of the deep models’ ecosystem, we introduce Hive – a comprehensive solution for selecting appropriate models and subsequently planning a set of atomic actions to satisfy the end-users’ instructions. Hive operates over sets of models and, upon receiving natural language instructions (i.e. user queries), schedules and executes explainable plans of atomic actions. These actions can involve one or more of the available models to achieve the overall task, while respecting end-users specific constraints. Notably, Hive handles tasks that involve multi-modal inputs and outputs, enabling it to handle complex, real-world queries. Our system is capable of planning complex chains of actions while guaranteeing explainability, using an LLM-based formal logic backbone empowered by PDDL operations. We introduce the MuSE benchmark in order to offer a comprehensive evaluation of the multi-modal capabilities of agent systems. Our findings show that our framework redefines the state-of-the-art for task selection, outperforming other competing systems that plan operations across multiple models while offering transparency guarantees while fully adhering to user constraints.

[AI-17] A Survey on Recommendation Unlearning: Fundamentals Taxonomy Evaluation and Open Questions

链接: https://arxiv.org/abs/2412.12836
作者: Yuyuan Li,Xiaohua Feng,Chaochao Chen,Qiang Yang
关键词: shaping user behavior, Recommender systems, recommendation unlearning, behavior and decision-making, highlighting their growing
类目: Information Retrieval (cs.IR); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Recommender systems have become increasingly influential in shaping user behavior and decision-making, highlighting their growing impact in various domains. Meanwhile, the widespread adoption of machine learning models in recommender systems has raised significant concerns regarding user privacy and security. As compliance with privacy regulations becomes more critical, there is a pressing need to address the issue of recommendation unlearning, i.e., eliminating the memory of specific training data from the learned recommendation models. Despite its importance, traditional machine unlearning methods are ill-suited for recommendation unlearning due to the unique challenges posed by collaborative interactions and model parameters. This survey offers a comprehensive review of the latest advancements in recommendation unlearning, exploring the design principles, challenges, and methodologies associated with this emerging field. We provide a unified taxonomy that categorizes different recommendation unlearning approaches, followed by a summary of widely used benchmarks and metrics for evaluation. By reviewing the current state of research, this survey aims to guide the development of more efficient, scalable, and robust recommendation unlearning techniques. Furthermore, we identify open research questions in this field, which could pave the way for future innovations not only in recommendation unlearning but also in a broader range of unlearning tasks across different machine learning applications.

[AI-18] Breaking the Programming Language Barrier: Multilingual Prompting to Empower Non-Native English Learners

链接: https://arxiv.org/abs/2412.12800
作者: James Prather,Brent N. Reeves,Paul Denny,Juho Leinonen,Stephen MacNeil,Andrew Luxton-Reilly,João Orvalho,Amin Alipour,Ali Alfageeh,Thezyrie Amarouche,Bailey Kimmel,Jared Wright,Musa Blake,Gweneth Barbre
关键词: Non-native English speakers, Non-native English, English speakers, native English speakers, face multiple barriers
类目: Computers and Society (cs.CY); Artificial Intelligence (cs.AI); Human-Computer Interaction (cs.HC)
*备注: 10 pages, 3 tables. Accepted for publication at the 27th Australasian Computing Education Conference (ACE 2025)

点击查看摘要

Abstract:Non-native English speakers (NNES) face multiple barriers to learning programming. These barriers can be obvious, such as the fact that programming language syntax and instruction are often in English, or more subtle, such as being afraid to ask for help in a classroom full of native English speakers. However, these barriers are frustrating because many NNES students know more about programming than they can articulate in English. Advances in generative AI (GenAI) have the potential to break down these barriers because state of the art models can support interactions in multiple languages. Moreover, recent work has shown that GenAI can be highly accurate at code generation and explanation. In this paper, we provide the first exploration of NNES students prompting in their native languages (Arabic, Chinese, and Portuguese) to generate code to solve programming problems. Our results show that students are able to successfully use their native language to solve programming problems, but not without some difficulty specifying programming terminology and concepts. We discuss the challenges they faced, the implications for practice in the short term, and how this might transform computing education globally in the long term.

[AI-19] Predicting change in time production – A machine learning approach to time perception

链接: https://arxiv.org/abs/2412.12781
作者: Amrapali Pednekar,Alvaro Garrido,Yara Khaluf,Pieter Simoens
关键词: Time perception, Time, time production, advanced significantly, model
类目: Human-Computer Interaction (cs.HC); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注: Main text contains 16 pages and 9 figure. Supplementary information is included as appendix. The paper has been submitted to IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS (TCDS). The code and data associated with the study will be made publicly available upon acceptance

点击查看摘要

Abstract:Time perception research has advanced significantly over the years. However, some areas remain largely unexplored. This study addresses two such under-explored areas in timing research: (1) A quantitative analysis of time perception at an individual level, and (2) Time perception in an ecological setting. In this context, we trained a machine learning model to predict the direction of change in an individual’s time production. The model’s training data was collected using an ecologically valid setup. We moved closer to an ecological setting by conducting an online experiment with 995 participants performing a time production task that used naturalistic videos (no audio) as stimuli. The model achieved an accuracy of 61%. This was 10 percentage points higher than the baseline models derived from cognitive theories of timing. The model performed equally well on new data from a second experiment, providing evidence of its generalization capabilities. The model’s output analysis revealed that it also contained information about the magnitude of change in time production. The predictions were further analysed at both population and individual level. It was found that a participant’s previous timing performance played a significant role in determining the direction of change in time production. By integrating attentional-gate theories from timing research with feature importance techniques from machine learning, we explained model predictions using cognitive theories of timing. The model and findings from this study have potential applications in systems involving human-computer interactions where understanding and predicting changes in user’s time perception can enable better user experience and task performance.

[AI-20] ParMod: A Parallel and Modular Framework for Learning Non-Markovian Tasks

链接: https://arxiv.org/abs/2412.12700
作者: Ruixuan Miao,Xu Lu,Cong Tian,Bin Yu,Zhenhua Duan
关键词: Markov Decision Processes, Markov Decision, Decision Processes, commonly used Reinforcement, Reinforcement Learning
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:The commonly used Reinforcement Learning (RL) model, MDPs (Markov Decision Processes), has a basic premise that rewards depend on the current state and action only. However, many real-world tasks are non-Markovian, which has long-term memory and dependency. The reward sparseness problem is further amplified in non-Markovian scenarios. Hence learning a non-Markovian task (NMT) is inherently more difficult than learning a Markovian one. In this paper, we propose a novel \textbfParallel and \textbfModular RL framework, ParMod, specifically for learning NMTs specified by temporal logic. With the aid of formal techniques, the NMT is modulaized into a series of sub-tasks based on the automaton structure (equivalent to its temporal logic counterpart). On this basis, sub-tasks will be trained by a group of agents in a parallel fashion, with one agent handling one sub-task. Besides parallel training, the core of ParMod lies in: a flexible classification method for modularizing the NMT, and an effective reward shaping method for improving the sample efficiency. A comprehensive evaluation is conducted on several challenging benchmark problems with respect to various metrics. The experimental results show that ParMod achieves superior performance over other relevant studies. Our work thus provides a good synergy among RL, NMT and temporal logic.

[AI-21] Everyday AR through AI-in-the-Loop

链接: https://arxiv.org/abs/2412.12681
作者: Ryo Suzuki,Mar Gonzalez-Franco,Misha Sra,David Lindlbauer
关键词: augmented reality, artificial intelligence, brings together experts, experts and practitioners, practitioners from augmented
类目: Human-Computer Interaction (cs.HC); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注: CHI 2025 Extended Abstract

点击查看摘要

Abstract:This workshop brings together experts and practitioners from augmented reality (AR) and artificial intelligence (AI) to shape the future of AI-in-the-loop everyday AR experiences. With recent advancements in both AR hardware and AI capabilities, we envision that everyday AR – always-available and seamlessly integrated into users’ daily environments – is becoming increasingly feasible. This workshop will explore how AI can drive such everyday AR experiences. We discuss a range of topics, including adaptive and context-aware AR, generative AR content creation, always-on AI assistants, AI-driven accessible design, and real-world-oriented AI agents. Our goal is to identify the opportunities and challenges in AI-enabled AR, focusing on creating novel AR experiences that seamlessly blend the digital and physical worlds. Through the workshop, we aim to foster collaboration, inspire future research, and build a community to advance the research field of AI-enhanced AR.

[AI-22] DriveTester: A Unified Platform for Simulation-Based Autonomous Driving Testing

链接: https://arxiv.org/abs/2412.12656
作者: Mingfei Cheng,Yuan Zhou,Xiaofei Xie
关键词: autonomous driving systems, ADS testing, Simulation-based testing plays, ADS, plays a critical
类目: oftware Engineering (cs.SE); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Simulation-based testing plays a critical role in evaluating the safety and reliability of autonomous driving systems (ADSs). However, one of the key challenges in ADS testing is the complexity of preparing and configuring simulation environments, particularly in terms of compatibility and stability between the simulator and the ADS. This complexity often results in researchers dedicating significant effort to customize their own environments, leading to disparities in development platforms and underlying systems. Consequently, reproducing and comparing these methodologies on a unified ADS testing platform becomes difficult. To address these challenges, we introduce DriveTester, a unified simulation-based testing platform built on Apollo, one of the most widely used open-source, industrial-level ADS platforms. DriveTester provides a consistent and reliable environment, integrates a lightweight traffic simulator, and incorporates various state-of-the-art ADS testing techniques. This enables researchers to efficiently develop, test, and compare their methods within a standardized platform, fostering reproducibility and comparison across different ADS testing approaches. The code is available: this https URL.

[AI-23] Shared Attention-based Autoencoder with Hierarchical Fusion-based Graph Convolution Network for sEEG SOZ Identification

链接: https://arxiv.org/abs/2412.12651
作者: Huachao Yan,Kailing Guo,Shiwei Song,Yihai Dai,Xiaoqiang Wei,Xiaofen Xing,Xiangmin Xu
关键词: Diagnosing seizure onset, seizure onset zone, sEEG SOZ identification, Diagnosing seizure, SOZ identification
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Signal Processing (eess.SP); Neurons and Cognition (q-bio.NC)
*备注:

点击查看摘要

Abstract:Diagnosing seizure onset zone (SOZ) is a challenge in neurosurgery, where stereoelectroencephalography (sEEG) serves as a critical technique. In sEEG SOZ identification, the existing studies focus solely on the intra-patient representation of epileptic information, overlooking the general features of epilepsy across patients and feature interdependencies between feature elements in each contact site. In order to address the aforementioned challenges, we propose the shared attention-based autoencoder (sATAE). sATAE is trained by sEEG data across all patients, with attention blocks introduced to enhance the representation of interdependencies between feature elements. Considering the spatial diversity of sEEG across patients, we introduce graph-based method for identification SOZ of each patient. However, the current graph-based methods for sEEG SOZ identification rely exclusively on static graphs to model epileptic networks. Inspired by the finding of neuroscience that epileptic network is intricately characterized by the interplay of sophisticated equilibrium between fluctuating and stable states, we design the hierarchical fusion-based graph convolution network (HFGCN) to identify the SOZ. HFGCN integrates the dynamic and static characteristics of epileptic networks through hierarchical weighting across different hierarchies, facilitating a more comprehensive learning of epileptic features and enriching node information for sEEG SOZ identification. Combining sATAE and HFGCN, we perform comprehensive experiments with sATAE-HFGCN on the self-build sEEG dataset, which includes sEEG data from 17 patients with temporal lobe epilepsy. The results show that our method, sATAE-HFGCN, achieves superior performance for identifying the SOZ of each patient, effectively addressing the aforementioned challenges, providing an efficient solution for sEEG-based SOZ identification.

[AI-24] Neural-Network-Driven Reward Prediction as a Heuristic: Advancing Q-Learning for Mobile Robot Path Planning

链接: https://arxiv.org/abs/2412.12650
作者: Yiming Ji,Kaijie Yun,Yang Liu,Zongwu Xie,Hong Liu
关键词: reinforcement learning technique, Q-learning, Q-learning method, widely used reinforcement, reinforcement learning
类目: Robotics (cs.RO); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Q-learning is a widely used reinforcement learning technique for solving path planning problems. It primarily involves the interaction between an agent and its environment, enabling the agent to learn an optimal strategy that maximizes cumulative rewards. Although many studies have reported the effectiveness of Q-learning, it still faces slow convergence issues in practical applications. To address this issue, we propose the NDR-QL method, which utilizes neural network outputs as heuristic information to accelerate the convergence process of Q-learning. Specifically, we improved the dual-output neural network model by introducing a start-end channel separation mechanism and enhancing the feature fusion process. After training, the proposed NDR model can output a narrowly focused optimal probability distribution, referred to as the guideline, and a broadly distributed suboptimal distribution, referred to as the region. Subsequently, based on the guideline prediction, we calculate the continuous reward function for the Q-learning method, and based on the region prediction, we initialize the Q-table with a bias. We conducted training, validation, and path planning simulation experiments on public datasets. The results indicate that the NDR model outperforms previous methods by up to 5% in prediction accuracy. Furthermore, the proposed NDR-QL method improves the convergence speed of the baseline Q-learning method by 90% and also surpasses the previously improved Q-learning methods in path quality metrics.

[AI-25] Exploring AI-Enabled Cybersecurity Frameworks: Deep-Learning Techniques GPU Support and Future Enhancements

链接: https://arxiv.org/abs/2412.12648
作者: Tobias Becher,Simon Torka
关键词: Traditional rule-based cybersecurity, proven highly effective, Traditional rule-based, rule-based cybersecurity systems, proven highly
类目: Cryptography and Security (cs.CR); Artificial Intelligence (cs.AI); Distributed, Parallel, and Cluster Computing (cs.DC); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Traditional rule-based cybersecurity systems have proven highly effective against known malware threats. However, they face challenges in detecting novel threats. To address this issue, emerging cybersecurity systems are incorporating AI techniques, specifically deep-learning algorithms, to enhance their ability to detect incidents, analyze alerts, and respond to events. While these techniques offer a promising approach to combating dynamic security threats, they often require significant computational resources. Therefore, frameworks that incorporate AI-based cybersecurity mechanisms need to support the use of GPUs to ensure optimal performance. Many cybersecurity framework vendors do not provide sufficiently detailed information about their implementation, making it difficult to assess the techniques employed and their effectiveness. This study aims to overcome this limitation by providing an overview of the most used cybersecurity frameworks that utilize AI techniques, specifically focusing on frameworks that provide comprehensive information about their implementation. Our primary objective is to identify the deep-learning techniques employed by these frameworks and evaluate their support for GPU acceleration. We have identified a total of \emphtwo deep-learning algorithms that are utilized by \emphthree out of 38 selected cybersecurity frameworks. Our findings aim to assist in selecting open-source cybersecurity frameworks for future research and assessing any discrepancies between deep-learning techniques used in theory and practice. Subjects: Cryptography and Security (cs.CR); Artificial Intelligence (cs.AI); Distributed, Parallel, and Cluster Computing (cs.DC); Machine Learning (cs.LG) Cite as: arXiv:2412.12648 [cs.CR] (or arXiv:2412.12648v1 [cs.CR] for this version) https://doi.org/10.48550/arXiv.2412.12648 Focus to learn more arXiv-issued DOI via DataCite (pending registration)

[AI-26] RDPI: A Refine Diffusion Probability Generation Method for Spatiotemporal Data Imputation

链接: https://arxiv.org/abs/2412.12642
作者: Zijin Liu,Xiang Zhao,You Song
关键词: air quality assessment, traffic flow monitoring, Spatiotemporal data imputation, Spatiotemporal data, data imputation plays
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Spatiotemporal data imputation plays a crucial role in various fields such as traffic flow monitoring, air quality assessment, and climate prediction. However, spatiotemporal data collected by sensors often suffer from temporal incompleteness, and the sparse and uneven distribution of sensors leads to missing data in the spatial dimension. Among existing methods, autoregressive approaches are prone to error accumulation, while simple conditional diffusion models fail to adequately capture the spatiotemporal relationships between observed and missing data. To address these issues, we propose a novel two-stage Refined Diffusion Probability Impuation (RDPI) framework based on an initial network and a conditional diffusion model. In the initial stage, deterministic imputation methods are used to generate preliminary estimates of the missing data. In the refinement stage, residuals are treated as the diffusion target, and observed values are innovatively incorporated into the forward process. This results in a conditional diffusion model better suited for spatiotemporal data imputation, bridging the gap between the preliminary estimates and the true values. Experiments on multiple datasets demonstrate that RDPI not only achieves state-of-the-art imputation accuracy but also significantly reduces sampling computational costs.

[AI-27] Lagrangian Index Policy for Restless Bandits with Average Reward

链接: https://arxiv.org/abs/2412.12641
作者: Konstantin Avrachenkov,Vivek S. Borkar,Pratik Shah
关键词: Whittle Index Policy, Lagrangian Index Policy, long-run average reward, Index Policy, restless multi-armed bandits
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Optimization and Control (math.OC); Probability (math.PR)
*备注:

点击查看摘要

Abstract:We study the Lagrangian Index Policy (LIP) for restless multi-armed bandits with long-run average reward. In particular, we compare the performance of LIP with the performance of the Whittle Index Policy (WIP), both heuristic policies known to be asymptotically optimal under certain natural conditions. Even though in most cases their performances are very similar, in the cases when WIP shows bad performance, LIP continues to perform very well. We then propose reinforcement learning algorithms, both tabular and NN-based, to obtain online learning schemes for LIP in the model-free setting. The proposed reinforcement learning schemes for LIP requires significantly less memory than the analogous scheme for WIP. We calculate analytically the Lagrangian index for the restart model, which describes the optimal web crawling and the minimization of the weighted age of information. We also give a new proof of asymptotic optimality in case of homogeneous bandits as the number of arms goes to infinity, based on exchangeability and de Finetti’s theorem.

[AI-28] rainMover: Efficient ML Training Live Migration with No Memory Overhead

链接: https://arxiv.org/abs/2412.12636
作者: ChonLam Lao,Minlan Yu,Aditya Akella,Jiamin Cao,Yu Guan,Pengcheng Zhang,Zhilong Zheng,Yichi Xu,Ennan Zhai,Dennis Cai,Jiaqi Gao
关键词: prominent workloads, workloads in modern, Machine learning training, training, modern data centers
类目: Distributed, Parallel, and Cluster Computing (cs.DC); Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Performance (cs.PF)
*备注: 13 pages body, 19 pages total

点击查看摘要

Abstract:Machine learning training has emerged as one of the most prominent workloads in modern data centers. These training jobs are large-scale, long-lasting, and tightly coupled, and are often disrupted by various events in the cluster such as failures, maintenance, and job scheduling. To handle these events, we rely on cold migration, where we first checkpoint the entire cluster, replace the related machines, and then restart the training. This approach leads to disruptions to the training jobs, resulting in significant downtime. In this paper, we present TrainMover, a live migration system that enables machine replacement during machine learning training. TrainMover minimizes downtime by leveraging member replacement of collective communication groups and sandbox lazy initialization. Our evaluation demonstrates that TrainMover achieves 16x less downtime compared to all baselines, effectively handling data center events like straggler rebalancing, maintenance, and unexpected failures.

[AI-29] Phoneme-Level Feature Discrepancies: A Key to Detecting Sophisticated Speech Deepfakes

链接: https://arxiv.org/abs/2412.12619
作者: Kuiyuan Zhang,Zhongyun Hua,Rushi Lan,Yushu Zhang,Yifang Guo
关键词: Recent advancements, highly convincing synthetic, speech conversion technologies, conversion technologies, technologies have enabled
类目: ound (cs.SD); Artificial Intelligence (cs.AI); Audio and Speech Processing (eess.AS)
*备注:

点击查看摘要

Abstract:Recent advancements in text-to-speech and speech conversion technologies have enabled the creation of highly convincing synthetic speech. While these innovations offer numerous practical benefits, they also cause significant security challenges when maliciously misused. Therefore, there is an urgent need to detect these synthetic speech signals. Phoneme features provide a powerful speech representation for deepfake detection. However, previous phoneme-based detection approaches typically focused on specific phonemes, overlooking temporal inconsistencies across the entire phoneme sequence. In this paper, we develop a new mechanism for detecting speech deepfakes by identifying the inconsistencies of phoneme-level speech features. We design an adaptive phoneme pooling technique that extracts sample-specific phoneme-level features from frame-level speech data. By applying this technique to features extracted by pre-trained audio models on previously unseen deepfake datasets, we demonstrate that deepfake samples often exhibit phoneme-level inconsistencies when compared to genuine speech. To further enhance detection accuracy, we propose a deepfake detector that uses a graph attention network to model the temporal dependencies of phoneme-level features. Additionally, we introduce a random phoneme substitution augmentation technique to increase feature diversity during training. Extensive experiments on four benchmark datasets demonstrate the superior performance of our method over existing state-of-the-art detection methods.

[AI-30] An Advantage-based Optimization Method for Reinforcement Learning in Large Action Space

链接: https://arxiv.org/abs/2412.12605
作者: Hai Lin,Cheng Huang,Zhihong Chen
关键词: high computational complexity, Reinforcement learning tasks, high-dimensional action spaces, leading to challenges, convergence difficulties
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Reinforcement learning tasks in real-world scenarios often involve large, high-dimensional action spaces, leading to challenges such as convergence difficulties, instability, and high computational complexity. It is widely acknowledged that traditional value-based reinforcement learning algorithms struggle to address these issues effectively. A prevalent approach involves generating independent sub-actions within each dimension of the action space. However, this method introduces bias, hindering the learning of optimal policies. In this paper, we propose an advantage-based optimization method and an algorithm named Advantage Branching Dueling Q-network (ABQ). ABQ incorporates a baseline mechanism to tune the action value of each dimension, leveraging the advantage relationship across different sub-actions. With this approach, the learned policy can be optimized for each dimension. Empirical results demonstrate that ABQ outperforms BDQ, achieving 3%, 171%, and 84% more cumulative rewards in HalfCheetah, Ant, and Humanoid environments, respectively. Furthermore, ABQ exhibits competitive performance when compared against two continuous action benchmark algorithms, DDPG and TD3.

[AI-31] Distributed satellite information networks: Architecture enabling technologies and trends

链接: https://arxiv.org/abs/2412.12587
作者: Qinyu Zhang,Liang Xu,Jianhao Huang,Tao Yang,Jian Jiao,Ye Wang,Yao Shi,Chiya Zhang,Xingjian Zhang,Ke Zhang,Yupeng Gong,Na Deng,Nan Zhao,Zhen Gao,Shujun Han,Xiaodong Xu,Li You,Dongming Wang,Shan Jiang,Dixian Zhao,Nan Zhang,Liujun Hu,Xiongwen He,Yonghui Li,Xiqi Gao,Xiaohu You
关键词: constellation-based satellite-integrated Internet, taking preliminary shape, ultra-dense constellation-based satellite-integrated, Internet is underway, satellite-integrated Internet
类目: Information Theory (cs.IT); Artificial Intelligence (cs.AI); Networking and Internet Architecture (cs.NI)
*备注:

点击查看摘要

Abstract:Driven by the vision of ubiquitous connectivity and wireless intelligence, the evolution of ultra-dense constellation-based satellite-integrated Internet is underway, now taking preliminary shape. Nevertheless, the entrenched institutional silos and limited, nonrenewable heterogeneous network resources leave current satellite systems struggling to accommodate the escalating demands of next-generation intelligent applications. In this context, the distributed satellite information networks (DSIN), exemplified by the cohesive clustered satellites system, have emerged as an innovative architecture, bridging information gaps across diverse satellite systems, such as communication, navigation, and remote sensing, and establishing a unified, open information network paradigm to support resilient space information services. This survey first provides a profound discussion about innovative network architectures of DSIN, encompassing distributed regenerative satellite network architecture, distributed satellite computing network architecture, and reconfigurable satellite formation flying, to enable flexible and scalable communication, computing and control. The DSIN faces challenges from network heterogeneity, unpredictable channel dynamics, sparse resources, and decentralized collaboration frameworks. To address these issues, a series of enabling technologies is identified, including channel modeling and estimation, cloud-native distributed MIMO cooperation, grant-free massive access, network routing, and the proper combination of all these diversity techniques. Furthermore, to heighten the overall resource efficiency, the cross-layer optimization techniques are further developed to meet upper-layer deterministic, adaptive and secure information services requirements. In addition, emerging research directions and new opportunities are highlighted on the way to achieving the DSIN vision.

[AI-32] SIDE: Socially Informed Drought Estimation Toward Understanding Societal Impact Dynamics of Environmental Crisis AAAI25

链接: https://arxiv.org/abs/2412.12575
作者: Lanyu Shang,Bozhang Chen,Shiwei Liu,Yang Zhang,Ruohan Zong,Anav Vora,Ximing Cai,Na Wei,Dong Wang
关键词: societal impact, critical global threat, significant societal impact, Drought, impact
类目: ocial and Information Networks (cs.SI); Artificial Intelligence (cs.AI)
*备注: To be published in AAAI 25

点击查看摘要

Abstract:Drought has become a critical global threat with significant societal impact. Existing drought monitoring solutions primarily focus on assessing drought severity using quantitative measurements, overlooking the diverse societal impact of drought from human-centric perspectives. Motivated by the collective intelligence on social media and the computational power of AI, this paper studies a novel problem of socially informed AI-driven drought estimation that aims to leverage social and news media information to jointly estimate drought severity and its societal impact. Two technical challenges exist: 1) How to model the implicit temporal dynamics of drought societal impact. 2) How to capture the social-physical interdependence between the physical drought condition and its societal impact. To address these challenges, we develop SIDE, a socially informed AI-driven drought estimation framework that explicitly quantifies the societal impact of drought and effectively models the social-physical interdependency for joint severity-impact estimation. Experiments on real-world datasets from California and Texas demonstrate SIDE’s superior performance compared to state-of-the-art baselines in accurately estimating drought severity and its societal impact. SIDE offers valuable insights for developing human-centric drought mitigation strategies to foster sustainable and resilient communities.

[AI-33] Seed-CTS: Unleashing the Power of Tree Search for Superior Performance in Competitive Coding Tasks

链接: https://arxiv.org/abs/2412.12544
作者: Hao Wang,Boyi Liu,Yufeng Zhang,Jie Chen
关键词: pose significant challenges, large language models, Competition-level code generation, code generation, tasks pose significant
类目: Artificial Intelligence (cs.AI); Software Engineering (cs.SE)
*备注:

点击查看摘要

Abstract:Competition-level code generation tasks pose significant challenges for current state-of-the-art large language models (LLMs). For example, on the LiveCodeBench-Hard dataset, models such as O1-Mini and O1-Preview achieve pass@1 rates of only 0.366 and 0.143, respectively. While tree search techniques have proven effective in domains like mathematics and general coding, their potential in competition-level code generation remains under-explored. In this work, we propose a novel token-level tree search method specifically designed for code generation. Leveraging Qwen2.5-Coder-32B-Instruct, our approach achieves a pass rate of 0.305 on LiveCodeBench-Hard, surpassing the pass@100 performance of GPT4o-0513 (0.245). Furthermore, by integrating Chain-of-Thought (CoT) prompting, we improve our method’s performance to 0.351, approaching O1-Mini’s pass@1 rate. To ensure reproducibility, we report the average number of generations required per problem by our tree search method on the test set. Our findings underscore the potential of tree search to significantly enhance performance on competition-level code generation tasks. This opens up new possibilities for large-scale synthesis of challenging code problems supervised fine-tuning (SFT) data, advancing competition-level code generation tasks.

[AI-34] Bots against Bias: Critical Next Steps for Human-Robot Interaction

链接: https://arxiv.org/abs/2412.12542
作者: Katie Seaborn
关键词: creations are biased, robotic creations, biased, Bias, HRI research
类目: Robotics (cs.RO); Artificial Intelligence (cs.AI); Human-Computer Interaction (cs.HC)
*备注:

点击查看摘要

Abstract:We humans are biased - and our robotic creations are biased, too. Bias is a natural phenomenon that drives our perceptions and behavior, including when it comes to socially expressive robots that have humanlike features. Recognizing that we embed bias, knowingly or not, within the design of such robots is crucial to studying its implications for people in modern societies. In this chapter, I consider the multifaceted question of bias in the context of humanoid, AI-enabled, and expressive social robots: Where does bias arise, what does it look like, and what can (or should) we do about it. I offer observations on human-robot interaction (HRI) along two parallel tracks: (1) robots designed in bias-conscious ways and (2) robots that may help us tackle bias in the human world. I outline a curated selection of cases for each track drawn from the latest HRI research and positioned against social, legal, and ethical factors. I also propose a set of critical next steps to tackle the challenges and opportunities on bias within HRI research and practice.

[AI-35] A Scalable Approach to Benchmarking the In-Conversation Differential Diagnostic Accuracy of a Health AI

链接: https://arxiv.org/abs/2412.12538
作者: Deep Bhatt,Surya Ayyagari,Anuruddh Mishra
关键词: increasing numbers, turning to online, online resources, Diagnostic errors, health information
类目: Artificial Intelligence (cs.AI); Human-Computer Interaction (cs.HC)
*备注:

点击查看摘要

Abstract:Diagnostic errors in healthcare persist as a critical challenge, with increasing numbers of patients turning to online resources for health information. While AI-powered healthcare chatbots show promise, there exists no standardized and scalable framework for evaluating their diagnostic capabilities. This study introduces a scalable benchmarking methodology for assessing health AI systems and demonstrates its application through August, an AI-driven conversational chatbot. Our methodology employs 400 validated clinical vignettes across 14 medical specialties, using AI-powered patient actors to simulate realistic clinical interactions. In systematic testing, August achieved a top-one diagnostic accuracy of 81.8% (327/400 cases) and a top-two accuracy of 85.0% (340/400 cases), significantly outperforming traditional symptom checkers. The system demonstrated 95.8% accuracy in specialist referrals and required 47% fewer questions compared to conventional symptom checkers (mean 16 vs 29 questions), while maintaining empathetic dialogue throughout consultations. These findings demonstrate the potential of AI chatbots to enhance healthcare delivery, though implementation challenges remain regarding real-world validation and integration of objective clinical data. This research provides a reproducible framework for evaluating healthcare AI systems, contributing to the responsible development and deployment of AI in clinical settings.

[AI-36] A Simple and Fast Way to Handle Semantic Errors in Transactions

链接: https://arxiv.org/abs/2412.12493
作者: Jinghan Zeng,Eugene Wu,Sanjay Krishnan
关键词: incorporate LLM-powered agents, enabling natural language, natural language input, LLM-powered agents, enabling natural
类目: Databases (cs.DB); Artificial Intelligence (cs.AI)
*备注: 14 pages, 13 figures

点击查看摘要

Abstract:Many computer systems are now being redesigned to incorporate LLM-powered agents, enabling natural language input and more flexible operations. This paper focuses on handling database transactions created by large language models (LLMs). Transactions generated by LLMs may include semantic errors, requiring systems to treat them as long-lived. This allows for human review and, if the transaction is incorrect, removal from the database history. Any removal action must ensure the database’s consistency (the “C” in ACID principles) is maintained throughout the process. We propose a novel middleware framework based on Invariant Satisfaction (I-Confluence), which ensures consistency by identifying and coordinating dependencies between long-lived transactions and new transactions. This middleware buffers suspicious or compensating transactions to manage coordination states. Using the TPC-C benchmark, we evaluate how transaction generation frequency, user reviews, and invariant completeness impact system performance. For system researchers, this study establishes an interactive paradigm between LLMs and database systems, providing an “undoing” mechanism for handling incorrect operations while guaranteeing database consistency. For system engineers, this paper offers a middleware design that integrates removable LLM-generated transactions into existing systems with minimal modifications. Comments: 14 pages, 13 figures Subjects: Databases (cs.DB); Artificial Intelligence (cs.AI) Cite as: arXiv:2412.12493 [cs.DB] (or arXiv:2412.12493v1 [cs.DB] for this version) https://doi.org/10.48550/arXiv.2412.12493 Focus to learn more arXiv-issued DOI via DataCite (pending registration)

[AI-37] Subversion Strategy Eval: Evaluating AIs stateless strategic capabilities against control protocols AAAI

链接: https://arxiv.org/abs/2412.12480
作者: Alex Mallen,Charlie Griffin,Alessandro Abate,Buck Shlegeris
关键词: usefully deploying, control protocols, subvert control protocols, protocols, control
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: Accepted at AAAI datasafe workshop

点击查看摘要

Abstract:AI control protocols are plans for usefully deploying AI systems in a way that is safe, even if the AI intends to subvert the protocol. Previous work evaluated protocols by subverting them with a human-AI red team, where an AI follows the human-written strategy. This paper investigates how well AI systems can generate and act on strategies for subverting control protocols whilst operating without private memory. Usually, decision-makers only need the ability to generate optimal plans. However, to subvert a protocol, an AI system must be able to reliably generate optimal plans in each context; coordinate plans with other instances of itself without communicating; and take actions with well-calibrated probabilities. We develop evaluations for a model’s ability to generate subversion strategies and develop affordance experiments that help isolate component capabilities. Our initial evaluations with Claude-3.5-sonnet highlight reasons why statelessly strategizing to subvert control protocols might be hard. We hope that our work can serve as a starting point for evaluations aimed at assessing whether a model is capable of strategizing to subvert control protocols.

[AI-38] Optimal Control Operator Perspective and a Neural Adaptive Spectral Method

链接: https://arxiv.org/abs/2412.12469
作者: Mingquan Feng,Zhijie Chen,Yixin Huang,Yizhou Liu,Junchi Yan
关键词: Optimal control problems, Optimal control, involve finding, functional is optimized, cost functional
类目: ystems and Control (eess.SY); Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Optimization and Control (math.OC)
*备注: Accepted for publication at AAAl’25. Extended version with full appendix, 22 pages

点击查看摘要

Abstract:Optimal control problems (OCPs) involve finding a control function for a dynamical system such that a cost functional is optimized. It is central to physical systems in both academia and industry. In this paper, we propose a novel instance-solution control operator perspective, which solves OCPs in a one-shot manner without direct dependence on the explicit expression of dynamics or iterative optimization processes. The control operator is implemented by a new neural operator architecture named Neural Adaptive Spectral Method (NASM), a generalization of classical spectral methods. We theoretically validate the perspective and architecture by presenting the approximation error bounds of NASM for the control operator. Experiments on synthetic environments and a real-world dataset verify the effectiveness and efficiency of our approach, including substantial speedup in running time, and high-quality in- and out-of-distribution generalization.

[AI-39] ransferable and Forecastable User Targeting Foundation Model

链接: https://arxiv.org/abs/2412.12468
作者: Bin Dou,Baokun Wang,Yun Zhu,Xiaotong Lin,Yike Xu,Xiaorui Huang,Yang Chen,Yun Liu,Shaoshuai Han,Yongchao Liu,Tianyi Zhang,Yu Cheng,Weiqiang Wang,Chuntao Hong
关键词: garnered substantial attention, selecting targeted users, non-expert marketers, digital marketing, process of selecting
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: 9 pages, 4 figures

点击查看摘要

Abstract:User targeting, the process of selecting targeted users from a pool of candidates for non-expert marketers, has garnered substantial attention with the advancements in digital marketing. However, existing user targeting methods encounter two significant challenges: (i) Poor cross-domain and cross-scenario transferability and generalization, and (ii) Insufficient forecastability in real-world applications. These limitations hinder their applicability across diverse industrial scenarios. In this work, we propose FIND, an industrial-grade, transferable, and forecastable user targeting foundation model. To enhance cross-domain transferability, our framework integrates heterogeneous multi-scenario user data, aligning them with one-sentence targeting demand inputs through contrastive pre-training. For improved forecastability, the text description of each user is derived based on anticipated future behaviors, while user representations are constructed from historical information. Experimental results demonstrate that our approach significantly outperforms existing baselines in cross-domain, real-world user targeting scenarios, showcasing the superior capabilities of FIND. Moreover, our method has been successfully deployed on the Alipay platform and is widely utilized across various scenarios.

[AI-40] LazyDiT: Lazy Learning for the Acceleration of Diffusion Transformers AAAI2025

链接: https://arxiv.org/abs/2412.12444
作者: Xuan Shen,Zhao Song,Yufa Zhou,Bo Chen,Yanyu Li,Yifan Gong,Kai Zhang,Hao Tan,Jason Kuen,Henghui Ding,Zhihao Shu,Wei Niu,Pu Zhao,Yanzhi Wang,Jiuxiang Gu
关键词: demonstrating superior performance, diffusion transformer models, generative tasks, demonstrating superior, wide array
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Diffusion Transformers have emerged as the preeminent models for a wide array of generative tasks, demonstrating superior performance and efficacy across various applications. The promising results come at the cost of slow inference, as each denoising step requires running the whole transformer model with a large amount of parameters. In this paper, we show that performing the full computation of the model at each diffusion step is unnecessary, as some computations can be skipped by lazily reusing the results of previous steps. Furthermore, we show that the lower bound of similarity between outputs at consecutive steps is notably high, and this similarity can be linearly approximated using the inputs. To verify our demonstrations, we propose the \textbfLazyDiT, a lazy learning framework that efficiently leverages cached results from earlier steps to skip redundant computations. Specifically, we incorporate lazy learning layers into the model, effectively trained to maximize laziness, enabling dynamic skipping of redundant computations. Experimental results show that LazyDiT outperforms the DDIM sampler across multiple diffusion transformer models at various resolutions. Furthermore, we implement our method on mobile devices, achieving better performance than DDIM with similar latency.

[AI-41] Numerical Pruning for Efficient Autoregressive Models AAAI2025

链接: https://arxiv.org/abs/2412.12441
作者: Xuan Shen,Zhao Song,Yufa Zhou,Bo Chen,Jing Liu,Ruiyi Zhang,Ryan A. Rossi,Hao Tan,Tong Yu,Xiang Chen,Yufan Zhou,Tong Sun,Pu Zhao,Yanzhi Wang,Jiuxiang Gu
关键词: Transformers have emerged, deep learning, leading architecture, architecture in deep, versatile and highly
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Transformers have emerged as the leading architecture in deep learning, proving to be versatile and highly effective across diverse domains beyond language and image processing. However, their impressive performance often incurs high computational costs due to their substantial model size. This paper focuses on compressing decoder-only transformer-based autoregressive models through structural weight pruning to improve the model efficiency while preserving performance for both language and image generation tasks. Specifically, we propose a training-free pruning method that calculates a numerical score with Newton’s method for the Attention and MLP modules, respectively. Besides, we further propose another compensation algorithm to recover the pruned model for better performance. To verify the effectiveness of our method, we provide both theoretical support and extensive experiments. Our experiments show that our method achieves state-of-the-art performance with reduced memory usage and faster generation speeds on GPUs.

[AI-42] DeepSN: A Sheaf Neural Framework for Influence Maximization AAAI2025

链接: https://arxiv.org/abs/2412.12416
作者: Asela Hevapathige,Qing Wang,Ahad N. Zehmakan
关键词: social network analysis, data mining, viral marketing, maximization is key, key topic
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: Accepted to AAAI 2025

点击查看摘要

Abstract:Influence maximization is key topic in data mining, with broad applications in social network analysis and viral marketing. In recent years, researchers have increasingly turned to machine learning techniques to address this problem. They have developed methods to learn the underlying diffusion processes in a data-driven manner, which enhances the generalizability of the solution, and have designed optimization objectives to identify the optimal seed set. Nonetheless, two fundamental gaps remain unsolved: (1) Graph Neural Networks (GNNs) are increasingly used to learn diffusion models, but in their traditional form, they often fail to capture the complex dynamics of influence diffusion, (2) Designing optimization objectives is challenging due to combinatorial explosion when solving this problem. To address these challenges, we propose a novel framework, DeepSN. Our framework employs sheaf neural diffusion to learn diverse influence patterns in a data-driven, end-to-end manner, providing enhanced separability in capturing diffusion characteristics. We also propose an optimization technique that accounts for overlapping influence between vertices, which helps to reduce the search space and identify the optimal seed set effectively and efficiently. Finally, we conduct extensive experiments on both synthetic and real-world datasets to demonstrate the effectiveness of our framework.

[AI-43] Improving Cooperation in Language Games with Bayesian Inference and the Cognitive Hierarchy AAAI-25

链接: https://arxiv.org/abs/2412.12409
作者: Joseph Bills,Christopher Archibald,Diego Blaylock
关键词: two-player cooperative games, accurate assumptions, teammate will behave, two-player cooperative, play together effectively
类目: Artificial Intelligence (cs.AI); Computer Science and Game Theory (cs.GT)
*备注: Full version of AAAI-25 paper

点击查看摘要

Abstract:In two-player cooperative games, agents can play together effectively when they have accurate assumptions about how their teammate will behave, but may perform poorly when these assumptions are inaccurate. In language games, failure may be due to disagreement in the understanding of either the semantics or pragmatics of an utterance. We model coarse uncertainty in semantics using a prior distribution of language models and uncertainty in pragmatics using the cognitive hierarchy, combining the two aspects into a single prior distribution over possible partner types. Fine-grained uncertainty in semantics is modeled using noise that is added to the embeddings of words in the language. To handle all forms of uncertainty we construct agents that learn the behavior of their partner using Bayesian inference and use this information to maximize the expected value of a heuristic function. We test this approach by constructing Bayesian agents for the game of Codenames, and show that they perform better in experiments where semantics is uncertain

[AI-44] Automated Generation of Massive Reasonable Empirical Theorems by Forward Reasoning Based on Strong Relevant Logics – A Solution to the Problem of LLM Pre-training Data Exhaustion

链接: https://arxiv.org/abs/2412.12408
作者: Jingde Cheng
关键词: large language models, Automated Knowledge Appreciation, Automated Theorem Finding, language models, pre-training of large
类目: Artificial Intelligence (cs.AI)
*备注: 11 pages, 7 figures

点击查看摘要

Abstract:Recently, it is often said that the data used for the pre-training of large language models (LLMs) have been exhausted. This paper proposes a solution to the problem: Automated generation of massive reasonable empirical theorems by forward reasoning based on strong relevant logics. In fact, this can be regarded as a part of our approach to the problems of ATF (Automated Theorem Finding) and AKA (Automated Knowledge Appreciation).

[AI-45] Sound Classification of Four Insect Classes

链接: https://arxiv.org/abs/2412.12395
作者: Yinxuan Wang,Sudip Vhaduri
关键词: project, Random Forest, SVM RBF, Decision Tree, Abstract
类目: ound (cs.SD); Artificial Intelligence (cs.AI); Audio and Speech Processing (eess.AS)
*备注: The manuscript is in submission

点击查看摘要

Abstract:The goal of this project is to classify four different insect sounds: cicada, beetle, termite, and cricket. One application of this project is for pest control to monitor and protect our ecosystem. Our project leverages data augmentation, including pitch shifting and speed changing, to improve model generalization. This project will test the performance of Decision Tree, Random Forest, SVM RBF, XGBoost, and k-NN models, combined with MFCC feature. A potential novelty of this project is that various data augmentation techniques are used and created 6 data along with the original sound. The dataset consists of the sound recordings of these four insects. This project aims to achieve a high classification accuracy and to reduce the over-fitting problem.

[AI-46] Enhancing Temporal Link Prediction with HierTKG: A Hierarchical Temporal Knowledge Graph Framework

链接: https://arxiv.org/abs/2412.12385
作者: Mariam Almutairi,Melike Yildiz Aktas,Nawar Wali,Shutonu Mitra,Dawei Zhou
关键词: challenges public decision-making, Temporal Graph Networks, challenges public, public decision-making, combining Temporal Graph
类目: ocial and Information Networks (cs.SI); Artificial Intelligence (cs.AI)
*备注: Preprint

点击查看摘要

Abstract:The rapid spread of misinformation on social media, especially during crises, challenges public decision-making. To address this, we propose HierTKG, a framework combining Temporal Graph Networks (TGN) and hierarchical pooling (DiffPool) to model rumor dynamics across temporal and structural scales. HierTKG captures key propagation phases, enabling improved temporal link prediction and actionable insights for misinformation control. Experiments demonstrate its effectiveness, achieving an MRR of 0.9845 on ICEWS14 and 0.9312 on WikiData, with competitive performance on noisy datasets like PHEME (MRR: 0.8802). By modeling structured event sequences and dynamic social interactions, HierTKG adapts to diverse propagation patterns, offering a scalable and robust solution for real-time analysis and prediction of rumor spread, aiding proactive intervention strategies.

[AI-47] Scam Detection for Ethereum Smart Contracts: Leveraging Graph Representation Learning for Secure Blockchain

链接: https://arxiv.org/abs/2412.12370
作者: Yihong Jin,Ze Yang
关键词: critical challenge due, Ethereum smart contracts, leading to significant, reputational damages, critical challenge
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Cryptography and Security (cs.CR); Distributed, Parallel, and Cluster Computing (cs.DC); Social and Information Networks (cs.SI)
*备注: Accepted to BDICN 2025

点击查看摘要

Abstract:The detection of scams within Ethereum smart contracts is a critical challenge due to their increasing exploitation for fraudulent activities, leading to significant financial and reputational damages. Existing detection methods often rely on contract code analysis or manually extracted features, which suffer from scalability and adaptability limitations. In this study, we introduce an innovative method that leverages graph representation learning to examine transaction patterns and identify fraudulent contracts. By transforming Ethereum transaction data into graph structures and employing advanced machine learning models, we achieve robust classification performance. Our method addresses label imbalance through SMOTE-ENN techniques and evaluates models like Multi-Layer Perceptron (MLP) and Graph Convolutional Networks (GCN). Experimental results indicate that the MLP model surpasses the GCN in this context, with real-world evaluations aligning closely with domain-specific analyses. This study provides a scalable and effective solution for enhancing trust and security in the Ethereum ecosystem.

[AI-48] LogBabylon: A Unified Framework for Cross-Log File Integration and Analysis

链接: https://arxiv.org/abs/2412.12364
作者: Rabimba Karanjai,Yang Lu,Dana Alsagheer,Keshav Kasichainula,Lei Xu,Weidong Shi,Shou-Hsuan Stephen Huang
关键词: log data, record events, software applications, network devices, critical resources
类目: oftware Engineering (cs.SE); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Logs are critical resources that record events, activities, or messages produced by software applications, operating systems, servers, and network devices. However, consolidating the heterogeneous logs and cross-referencing them is challenging and complicated. Manually analyzing the log data is time-consuming and prone to errors. LogBabylon is a centralized log data consolidating solution that leverages Large Language Models (LLMs) integrated with Retrieval-Augmented Generation (RAG) technology. LogBabylon interprets the log data in a human-readable way and adds insight analysis of the system performance and anomaly alerts. It provides a paramount view of the system landscape, enabling proactive management and rapid incident response. LogBabylon consolidates diverse log sources and enhances the extracted information’s accuracy and relevancy. This facilitates a deeper understanding of log data, supporting more effective decision-making and operational efficiency. Furthermore, LogBabylon streamlines the log analysis process, significantly reducing the time and effort required to interpret complex datasets. Its capabilities extend to generating context-aware insights, offering an invaluable tool for continuous monitoring, performance optimization, and security assurance in dynamic computing environments.

[AI-49] he Ramanujan Library – Automated Discovery on the Hypergraph of Integer Relations

链接: https://arxiv.org/abs/2412.12361
作者: Itay Beit-Halachmi,Ido Kaminer
关键词: Fundamental mathematical constants, Fundamental mathematical, physics to biology, constants, Fundamental
类目: Artificial Intelligence (cs.AI); Mathematical Software (cs.MS); Number Theory (math.NT)
*备注: 20 pages, 7 figures

点击查看摘要

Abstract:Fundamental mathematical constants appear in nearly every field of science, from physics to biology. Formulas that connect different constants often bring great insight by hinting at connections between previously disparate fields. Discoveries of such relations, however, have remained scarce events, relying on sporadic strokes of creativity by human mathematicians. Recent developments of algorithms for automated conjecture generation have accelerated the discovery of formulas for specific constants. Yet, the discovery of connections between constants has not been addressed. In this paper, we present the first library dedicated to mathematical constants and their interrelations. This library can serve as a central repository of knowledge for scientists from different areas, and as a collaborative platform for development of new algorithms. The library is based on a new representation that we propose for organizing the formulas of mathematical constants: a hypergraph, with each node representing a constant and each edge representing a formula. Using this representation, we propose and demonstrate a systematic approach for automatically enriching this library using PSLQ, an integer relation algorithm based on QR decomposition and lattice construction. During its development and testing, our strategy led to the discovery of 75 previously unknown connections between constants, including a new formula for the `first continued fraction’ constant C_1 , novel formulas for natural logarithms, and new formulas connecting \pi and e . The latter formulas generalize a century-old relation between \pi and e by Ramanujan, which until now was considered a singular formula and is now found to be part of a broader mathematical structure. The code supporting this library is a public, open-source API that can serve researchers in experimental mathematics and other fields of science.

[AI-50] Achieving Collective Welfare in Multi-Agent Reinforcement Learning via Suggestion Sharing

链接: https://arxiv.org/abs/2412.12326
作者: Yue Jin,Shuangqing Wei,Giovanni Montana
关键词: achieve shared welfare, Social Dilemmas frequently, shared welfare, well-being often obstructs, obstructs efforts
类目: Multiagent Systems (cs.MA); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:In human society, the conflict between self-interest and collective well-being often obstructs efforts to achieve shared welfare. Related concepts like the Tragedy of the Commons and Social Dilemmas frequently manifest in our daily lives. As artificial agents increasingly serve as autonomous proxies for humans, we propose using multi-agent reinforcement learning (MARL) to address this issue - learning policies to maximise collective returns even when individual agents’ interests conflict with the collective one. Traditional MARL solutions involve sharing rewards, values, and policies or designing intrinsic rewards to encourage agents to learn collectively optimal policies. We introduce a novel MARL approach based on Suggestion Sharing (SS), where agents exchange only action suggestions. This method enables effective cooperation without the need to design intrinsic rewards, achieving strong performance while revealing less private information compared to sharing rewards, values, or policies. Our theoretical analysis establishes a bound on the discrepancy between collective and individual objectives, demonstrating how sharing suggestions can align agents’ behaviours with the collective objective. Experimental results demonstrate that SS performs competitively with baselines that rely on value or policy sharing or intrinsic rewards.

[AI-51] Equivariant Action Sampling for Reinforcement Learning and Planning

链接: https://arxiv.org/abs/2412.12237
作者: Linfeng Zhao,Owen Howell,Xupeng Zhu,Jung Yeon Park,Zhewen Zhang,Robin Walters,Lawson L.S. Wong
关键词: Reinforcement learning, tasks require accurate, require accurate sampling-based, require accurate, Reinforcement
类目: Robotics (cs.RO); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注: Published at International Workshop on the Algorithmic Foundations of Robotics (WAFR) 2024. Website: this http URL

点击查看摘要

Abstract:Reinforcement learning (RL) algorithms for continuous control tasks require accurate sampling-based action selection. Many tasks, such as robotic manipulation, contain inherent problem symmetries. However, correctly incorporating symmetry into sampling-based approaches remains a challenge. This work addresses the challenge of preserving symmetry in sampling-based planning and control, a key component for enhancing decision-making efficiency in RL. We introduce an action sampling approach that enforces the desired symmetry. We apply our proposed method to a coordinate regression problem and show that the symmetry aware sampling method drastically outperforms the naive sampling approach. We furthermore develop a general framework for sampling-based model-based planning with Model Predictive Path Integral (MPPI). We compare our MPPI approach with standard sampling methods on several continuous control tasks. Empirical demonstrations across multiple continuous control environments validate the effectiveness of our approach, showcasing the importance of symmetry preservation in sampling-based action selection.

[AI-52] Linear Equations with Min and Max Operators: Computational Complexity

链接: https://arxiv.org/abs/2412.12228
作者: Krishnendu Chatterjee,Ruichen Luo,Raimundo Saona,Jakub Svoboda
关键词: optimization problems defined, class of optimization, system of linear, linear equations, conditions
类目: Computational Complexity (cs.CC); Artificial Intelligence (cs.AI); Optimization and Control (math.OC)
*备注:

点击查看摘要

Abstract:We consider a class of optimization problems defined by a system of linear equations with min and max operators. This class of optimization problems has been studied under restrictive conditions, such as, (C1) the halting or stability condition; (C2) the non-negative coefficients condition; (C3) the sum up to 1 condition; and (C4) the only min or only max oerator condition. Several seminal results in the literature focus on special cases. For example, turn-based stochastic games correspond to conditions C2 and C3; and Markov decision process to conditions C2, C3, and C4. However, the systematic computational complexity study of all the cases has not been explored, which we address in this work. Some highlights of our results are: with conditions C2 and C4, and with conditions C3 and C4, the problem is NP-complete, whereas with condition C1 only, the problem is in UP intersects coUP. Finally, we establish the computational complexity of the decision problem of checking the respective conditions.

[AI-53] EDformer: Embedded Decomposition Transformer for Interpretable Multivariate Time Series Predictions

链接: https://arxiv.org/abs/2412.12227
作者: Sanjay Chakraborty,Ibrahim Delibasoglu,Fredrik Heintz
关键词: stock market analysis, Time series forecasting, stock market, market analysis, scientific simulations
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Time series forecasting is a crucial challenge with significant applications in areas such as weather prediction, stock market analysis, and scientific simulations. This paper introduces an embedded decomposed transformer, ‘EDformer’, for multivariate time series forecasting tasks. Without altering the fundamental elements, we reuse the Transformer architecture and consider the capable functions of its constituent parts in this work. Edformer first decomposes the input multivariate signal into seasonal and trend components. Next, the prominent multivariate seasonal component is reconstructed across the reverse dimensions, followed by applying the attention mechanism and feed-forward network in the encoder stage. In particular, the feed-forward network is used for each variable frame to learn nonlinear representations, while the attention mechanism uses the time points of individual seasonal series embedded within variate frames to capture multivariate correlations. Therefore, the trend signal is added with projection and performs the final forecasting. The EDformer model obtains state-of-the-art predicting results in terms of accuracy and efficiency on complex real-world time series datasets. This paper also addresses model explainability techniques to provide insights into how the model makes its predictions and why specific features or time steps are important, enhancing the interpretability and trustworthiness of the forecasting results.

[AI-54] Apollo-Forecast: Overcoming Aliasing and Inference Speed Challenges in Language Models for Time Series Forecasting

链接: https://arxiv.org/abs/2412.12226
作者: Tianyi Yin,Jingwei Wang,Yunlong Ma,Han Wang,Chenze Wang,Yukai Zhao,Min Liu,Weiming Shen,Yufeng Chen
关键词: Encoding time series, Encoding time, unseen tasks, time series, time series forecasting
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Encoding time series into tokens and using language models for processing has been shown to substantially augment the models’ ability to generalize to unseen tasks. However, existing language models for time series forecasting encounter several obstacles, including aliasing distortion and prolonged inference times, primarily due to the limitations of quantization processes and the computational demands of large models. This paper introduces Apollo-Forecast, a novel framework that tackles these challenges with two key innovations: the Anti-Aliasing Quantization Module (AAQM) and the Race Decoding (RD) technique. AAQM adeptly encodes sequences into tokens while mitigating high-frequency noise in the original signals, thus enhancing both signal fidelity and overall quantization efficiency. RD employs a draft model to enable parallel processing and results integration, which markedly accelerates the inference speed for long-term predictions, particularly in large-scale models. Extensive experiments on various real-world datasets show that Apollo-Forecast outperforms state-of-the-art methods by 35.41% and 18.99% in WQL and MASE metrics, respectively, in zero-shot scenarios. Furthermore, our method achieves a 1.9X-2.7X acceleration in inference speed over baseline methods.

[AI-55] Parallel Greedy Best-First Search with a Bound on the Number of Expansions Relative to Sequential Search

链接: https://arxiv.org/abs/2412.12221
作者: Takumi Shimoda,Alex Fukunaga
关键词: straightforward parallelization, states expanded, non-admissible search algorithms, number of states, sequential GBFS
类目: Data Structures and Algorithms (cs.DS); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Parallelization of non-admissible search algorithms such as GBFS poses a challenge because straightforward parallelization can result in search behavior which significantly deviates from sequential search. Previous work proposed PUHF, a parallel search algorithm which is constrained to only expand states that can be expanded by some tie-breaking strategy for GBFS. We show that despite this constraint, the number of states expanded by PUHF is not bounded by a constant multiple of the number of states expanded by sequential GBFS with the worst-case tie-breaking strategy. We propose and experimentally evaluate One Bench At a Time (OBAT), a parallel greedy search which guarantees that the number of states expanded is within a constant factor of the number of states expanded by sequential GBFS with some tie-breaking policy.

[AI-56] Are Large Language Models Useful for Time Series Data Analysis?

链接: https://arxiv.org/abs/2412.12219
作者: Francis Tang,Ying Ding
关键词: Time series data, series data plays, informed decision-making, anomaly detection, plays a critical
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Time series data plays a critical role across diverse domains such as healthcare, energy, and finance, where tasks like classification, anomaly detection, and forecasting are essential for informed decision-making. Recently, large language models (LLMs) have gained prominence for their ability to handle complex data and extract meaningful insights. This study investigates whether LLMs are effective for time series data analysis by comparing their performance with non-LLM-based approaches across three tasks: classification, anomaly detection, and forecasting. Through a series of experiments using GPT4TS and autoregressive models, we evaluate their performance on benchmark datasets and assess their accuracy, precision, and ability to generalize. Our findings indicate that while LLM-based methods excel in specific tasks like anomaly detection, their benefits are less pronounced in others, such as forecasting, where simpler models sometimes perform comparably or better. This research highlights the role of LLMs in time series analysis and lays the groundwork for future studies to systematically explore their applications and limitations in handling temporal data. Subjects: Machine Learning (cs.LG); Artificial Intelligence (cs.AI) Cite as: arXiv:2412.12219 [cs.LG] (or arXiv:2412.12219v1 [cs.LG] for this version) https://doi.org/10.48550/arXiv.2412.12219 Focus to learn more arXiv-issued DOI via DataCite

[AI-57] Imagined Speech State Classification for Robust Brain-Computer Interface

链接: https://arxiv.org/abs/2412.12215
作者: Byung-Kwan Ko,Jun-Young Kim,Seo-Hyun Lee
关键词: machine learning classifiers, traditional machine learning, machine learning, learning classifiers versus, conventional machine learning
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Sound (cs.SD); Audio and Speech Processing (eess.AS)
*备注:

点击查看摘要

Abstract:This study examines the effectiveness of traditional machine learning classifiers versus deep learning models for detecting the imagined speech using electroencephalogram data. Specifically, we evaluated conventional machine learning techniques such as CSP-SVM and LDA-SVM classifiers alongside deep learning architectures such as EEGNet, ShallowConvNet, and DeepConvNet. Machine learning classifiers exhibited significantly lower precision and recall, indicating limited feature extraction capabilities and poor generalization between imagined speech and idle states. In contrast, deep learning models, particularly EEGNet, achieved the highest accuracy of 0.7080 and an F1 score of 0.6718, demonstrating their enhanced ability in automatic feature extraction and representation learning, essential for capturing complex neurophysiological patterns. These findings highlight the limitations of conventional machine learning approaches in brain-computer interface (BCI) applications and advocate for adopting deep learning methodologies to achieve more precise and reliable classification of detecting imagined speech. This foundational research contributes to the development of imagined speech-based BCI systems.

[AI-58] Embracing Large Language Models in Traffic Flow Forecasting

链接: https://arxiv.org/abs/2412.12201
作者: Yusheng Zhao,Xiao Luo,Haomin Wen,Zhiping Xiao,Wei Ju,Ming Zhang
关键词: future traffic flows, predict future traffic, Traffic flow forecasting, Traffic flow, historical traffic conditions
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Traffic flow forecasting aims to predict future traffic flows based on the historical traffic conditions and the road network. It is an important problem in intelligent transportation systems, with a plethora of methods been proposed. Existing efforts mainly focus on capturing and utilizing spatio-temporal dependencies to predict future traffic flows. Though promising, they fall short in adapting to test-time environmental changes of traffic conditions. To tackle this challenge, we propose to introduce large language models (LLMs) to help traffic flow forecasting and design a novel method named Large Language Model Enhanced Traffic Flow Predictor (LEAF). LEAF adopts two branches, capturing different spatio-temporal relations using graph and hypergraph structures respectively. The two branches are first pre-trained individually, and during test-time, they yield different predictions. Based on these predictions, a large language model is used to select the most likely result. Then, a ranking loss is applied as the learning objective to enhance the prediction ability of the two branches. Extensive experiments on several datasets demonstrate the effectiveness of the proposed LEAF.

[AI-59] rendSim: Simulating Trending Topics in Social Media Under Poisoning Attacks with LLM -based Multi-agent System

链接: https://arxiv.org/abs/2412.12196
作者: Zeyu Zhang,Jianxun Lian,Chen Ma,Yaning Qu,Ye Luo,Lei Wang,Rui Li,Xu Chen,Yankai Lin,Le Wu,Xing Xie,Ji-Rong Wen
关键词: breaking events, significant part, part of modern, participate in discussions, discussions of breaking
类目: ocial and Information Networks (cs.SI); Artificial Intelligence (cs.AI)
*备注: 19 pages, 9 tables, 8 figure

点击查看摘要

Abstract:Trending topics have become a significant part of modern social media, attracting users to participate in discussions of breaking events. However, they also bring in a new channel for poisoning attacks, resulting in negative impacts on society. Therefore, it is urgent to study this critical problem and develop effective strategies for defense. In this paper, we propose TrendSim, an LLM-based multi-agent system to simulate trending topics in social media under poisoning attacks. Specifically, we create a simulation environment for trending topics that incorporates a time-aware interaction mechanism, centralized message dissemination, and an interactive system. Moreover, we develop LLM-based human-like agents to simulate users in social media, and propose prototype-based attackers to replicate poisoning attacks. Besides, we evaluate TrendSim from multiple aspects to validate its effectiveness. Based on TrendSim, we conduct simulation experiments to study four critical problems about poisoning attacks on trending topics for social benefit.

[AI-60] No Free Lunch for Defending Against Prefilling Attack by In-Context Learning

链接: https://arxiv.org/abs/2412.12192
作者: Zhiyu Xue,Guangliang Liu,Bocheng Chen,Kristen Marie Johnson,Ramtin Pedarsani
关键词: Large Language Models, Large Language, important research topic, security of Large, jailbreak attacks
类目: Cryptography and Security (cs.CR); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:The security of Large Language Models (LLMs) has become an important research topic since the emergence of ChatGPT. Though there have been various effective methods to defend against jailbreak attacks, prefilling attacks remain an unsolved and popular threat against open-sourced LLMs. In-Context Learning (ICL) offers a computationally efficient defense against various jailbreak attacks, yet no effective ICL methods have been developed to counter prefilling attacks. In this paper, we: (1) show that ICL can effectively defend against prefilling jailbreak attacks by employing adversative sentence structures within demonstrations; (2) characterize the effectiveness of this defense through the lens of model size, number of demonstrations, over-defense, integration with other jailbreak attacks, and the presence of safety alignment. Given the experimental results and our analysis, we conclude that there is no free lunch for defending against prefilling jailbreak attacks with ICL. On the one hand, current safety alignment methods fail to mitigate prefilling jailbreak attacks, but adversative structures within ICL demonstrations provide robust defense across various model sizes and complex jailbreak attacks. On the other hand, LLMs exhibit similar over-defensiveness when utilizing ICL demonstrations with adversative structures, and this behavior appears to be independent of model size.

[AI-61] MoT: Inertial Motion Transformer for Inertial Navigation AAAI2025 AAAI

链接: https://arxiv.org/abs/2412.12190
作者: Son Minh Nguyen,Linh Duy Tran,Duc Viet Le,Paul J.M Havinga
关键词: innovative Transformer-based inertial, Transformer-based inertial odometry, Progressive Series Decoupler, innovative Transformer-based, retrieves cross-modal information
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Robotics (cs.RO)
*备注: Accepted as technical research paper in 39th AAAI Conference on Artificial Intelligence, 2025 (AAAI 2025)

点击查看摘要

Abstract:We propose iMoT, an innovative Transformer-based inertial odometry method that retrieves cross-modal information from motion and rotation modalities for accurate positional estimation. Unlike prior work, during the encoding of the motion context, we introduce Progressive Series Decoupler at the beginning of each encoder layer to stand out critical motion events inherent in acceleration and angular velocity signals. To better aggregate cross-modal interactions, we present Adaptive Positional Encoding, which dynamically modifies positional embeddings for temporal discrepancies between different modalities. During decoding, we introduce a small set of learnable query motion particles as priors to model motion uncertainties within velocity segments. Each query motion particle is intended to draw cross-modal features dedicated to a specific motion mode, all taken together allowing the model to refine its understanding of motion dynamics effectively. Lastly, we design a dynamic scoring mechanism to stabilize iMoT’s optimization by considering all aligned motion particles at the final decoding step, ensuring robust and accurate velocity segment estimation. Extensive evaluations on various inertial datasets demonstrate that iMoT significantly outperforms state-of-the-art methods in delivering superior robustness and accuracy in trajectory reconstruction.

[AI-62] Graph Similarity Computation via Interpretable Neural Node Alignment

链接: https://arxiv.org/abs/2412.12185
作者: Jingjing Wang,Hongjie Zhu,Haoran Xie,Fu Lee Wang,Xiaoliang Xu,Yuxiang Wang
关键词: query chemical compound, user potential friends, Graph Edit Distance, Graph similarity computation, Maximum Common Subgraphs
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:\Graph similarity computation is an essential task in many real-world graph-related applications such as retrieving the similar drugs given a query chemical compound or finding the user’s potential friends from the social network database. Graph Edit Distance (GED) and Maximum Common Subgraphs (MCS) are the two commonly used domain-agnostic metrics to evaluate graph similarity in practice. Unfortunately, computing the exact GED is known to be a NP-hard problem. To solve this limitation, neural network based models have been proposed to approximate the calculations of GED/MCS. However, deep learning models are well-known ``black boxes’', thus the typically characteristic one-to-one node/subgraph alignment process in the classical computations of GED and MCS cannot be seen. Existing methods have paid attention to approximating the node/subgraph alignment (soft alignment), but the one-to-one node alignment (hard alignment) has not yet been solved. To fill this gap, in this paper we propose a novel interpretable neural node alignment model without relying on node alignment ground truth information. Firstly, the quadratic assignment problem in classical GED computation is relaxed to a linear alignment via embedding the features in the node embedding space. Secondly, a differentiable Gumbel-Sinkhorn module is proposed to unsupervised generate the optimal one-to-one node alignment matrix. Experimental results in real-world graph datasets demonstrate that our method outperforms the state-of-the-art methods in graph similarity computation and graph retrieval tasks, achieving up to 16% reduction in the Mean Squared Error and up to 12% improvement in the retrieval evaluation metrics, respectively.

[AI-63] Adopting Explainable-AI to investigate the impact of urban morphology design on energy and environmental performance in dry-arid climates

链接: https://arxiv.org/abs/2412.12183
作者: Pegah Eshraghi,Riccardo Talami,Arman Nikkhah Dehnavi,Maedeh Mirdamadi,Zahra-Sadat Zomorodian
关键词: Shapley Additive Explanations, rapidly urbanizing regions, specifically Shapley Additive, Building Energy Modeling, Sky View Factor
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:In rapidly urbanizing regions, designing climate-responsive urban forms is crucial for sustainable development, especially in dry arid-climates where urban morphology has a significant impact on energy consumption and environmental performance. This study advances urban morphology evaluation by combining Urban Building Energy Modeling (UBEM) with machine learning methods (ML) and Explainable AI techniques, specifically Shapley Additive Explanations (SHAP). Using Tehran’s dense urban landscape as a case study, this research assesses and ranks the impact of 30 morphology parameters at the urban block level on key energy metrics (cooling, heating, and lighting demand) and environmental performance (sunlight exposure, photovoltaic generation, and Sky View Factor). Among seven ML algorithms evaluated, the XGBoost model was the most effective predictor, achieving high accuracy (R2: 0.92) and a training time of 3.64 seconds. Findings reveal that building shape, window-to-wall ratio, and commercial ratio are the most critical parameters affecting energy efficiency, while the heights and distances of neighboring buildings strongly influence cooling demand and solar access. By evaluating urban blocks with varied densities and configurations, this study offers generalizable insights applicable to other dry-arid regions. Moreover, the integration of UBEM and Explainable AI offers a scalable, data-driven framework for developing climate-responsive urban designs adaptable to high-density environments worldwide.

[AI-64] Activation Sparsity Opportunities for Compressing General Large Language Models

链接: https://arxiv.org/abs/2412.12178
作者: Nobel Dhar,Bobin Deng,Md Romyull Islam,Kazi Fahim Ahmad Nasif,Liang Zhao,Kun Suo
关键词: Small Language Models, devices’ independent capabilities, Large Language Models, Deploying local, lightweight Small Language
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: Conference submission for IPCCC 2024

点击查看摘要

Abstract:Deploying local AI models, such as Large Language Models (LLMs), to edge devices can substantially enhance devices’ independent capabilities, alleviate the server’s burden, and lower the response time. Owing to these tremendous potentials, many big tech companies have released several lightweight Small Language Models (SLMs) to bridge this gap. However, we still have huge motivations to deploy more powerful (LLMs) AI models on edge devices and enhance their smartness level. Unlike the conventional approaches for AI model compression, we investigate activation sparsity. The activation sparsity method is orthogonal and combinable with existing techniques to maximize compression rate while maintaining great accuracy. LLMs’ Feed-Forward Network (FFN) components, which typically comprise a large proportion of parameters (around 3/2), ensure that our FFN optimizations would have a better chance of achieving effective compression. Moreover, our findings are beneficial to general LLMs and are not restricted to ReLU-based models. This work systematically investigates the tradeoff between enforcing activation sparsity and perplexity (accuracy) on state-of-the-art LLMs. Our empirical analysis demonstrates that we can obtain around 50% of main memory and computing reductions for critical FFN components with negligible accuracy degradation. This extra 50% sparsity does not naturally exist in the current LLMs, which require tuning LLMs’ activation outputs by injecting zero-enforcing thresholds. To obtain the benefits of activation sparsity, we provide a guideline for the system architect for LLM prediction and prefetching. The success prediction allows the system to prefetch the necessary weights while omitting the inactive ones and their successors, therefore lowering cache and memory pollution and reducing LLM execution time on resource-constrained edge devices.

[AI-65] PickLLM : Context-Aware RL-Assisted Large Language Model Routing AAAI2025 AAAI

链接: https://arxiv.org/abs/2412.12170
作者: Dimitrios Sikeridis,Dennis Ramdass,Pranay Pareek
关键词: Large Language Models, Large Language, Language Models, open-source options, LLM
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: This work has been accepted at the first Workshop on Scalable and Efficient Artificial Intelligence Systems (SEAS) held in conjunction with the 39th Annual AAAI Conference on Artificial Intelligence, AAAI 2025, in Philadelphia, Pennsylvania, USA

点击查看摘要

Abstract:Recently, the number of off-the-shelf Large Language Models (LLMs) has exploded with many open-source options. This creates a diverse landscape regarding both serving options (e.g., inference on local hardware vs remote LLM APIs) and model heterogeneous expertise. However, it is hard for the user to efficiently optimize considering operational cost (pricing structures, expensive LLMs-as-a-service for large querying volumes), efficiency, or even per-case specific measures such as response accuracy, bias, or toxicity. Also, existing LLM routing solutions focus mainly on cost reduction, with response accuracy optimizations relying on non-generalizable supervised training, and ensemble approaches necessitating output computation for every considered LLM candidate. In this work, we tackle the challenge of selecting the optimal LLM from a model pool for specific queries with customizable objectives. We propose PickLLM, a lightweight framework that relies on Reinforcement Learning (RL) to route on-the-fly queries to available models. We introduce a weighted reward function that considers per-query cost, inference latency, and model response accuracy by a customizable scoring function. Regarding the learning algorithms, we explore two alternatives: PickLLM router acting as a learning automaton that utilizes gradient ascent to select a specific LLM, or utilizing stateless Q-learning to explore the set of LLMs and perform selection with a \epsilon -greedy approach. The algorithm converges to a single LLM for the remaining session queries. To evaluate, we utilize a pool of four LLMs and benchmark prompt-response datasets with different contexts. A separate scoring function is assessing response accuracy during the experiment. We demonstrate the speed of convergence for different learning rates and improvement in hard metrics such as cost per querying session and overall response latency.

[AI-66] Regulation of Language Models With Interpretability Will Likely Result In A Performance Trade-Off

链接: https://arxiv.org/abs/2412.12169
作者: Eoin M. Kenny,Julie A. Shah
关键词: machine learning, increasingly cited, important and pressing, pressing concern, concern in machine
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computers and Society (cs.CY)
*备注:

点击查看摘要

Abstract:Regulation is increasingly cited as the most important and pressing concern in machine learning. However, it is currently unknown how to implement this, and perhaps more importantly, how it would effect model performance alongside human collaboration if actually realized. In this paper, we attempt to answer these questions by building a regulatable large-language model (LLM), and then quantifying how the additional constraints involved affect (1) model performance, alongside (2) human collaboration. Our empirical results reveal that it is possible to force an LLM to use human-defined features in a transparent way, but a “regulation performance trade-off” previously not considered reveals itself in the form of a 7.34% classification performance drop. Surprisingly however, we show that despite this, such systems actually improve human task performance speed and appropriate confidence in a realistic deployment setting compared to no AI assistance, thus paving a way for fair, regulatable AI, which benefits users.

[AI-67] A Decomposition Modeling Framework for Seasonal Time-Series Forecasting

链接: https://arxiv.org/abs/2412.12168
作者: Yining Pang,Chenghan Li
关键词: accurate future prediction, Seasonal time series, intricate long-term dependencies, time series, posing a significant
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Seasonal time series exhibit intricate long-term dependencies, posing a significant challenge for accurate future prediction. This paper introduces the Multi-scale Seasonal Decomposition Model (MSSD) for seasonal time-series forecasting. Initially, leveraging the inherent periodicity of seasonal time series, we decompose the univariate time series into three primary components: Ascending, Peak, and Descending. This decomposition approach enhances the capture of periodic features. By addressing the limitations of existing time-series modeling methods, particularly in modeling the Peak component, this research proposes a multi-scale network structure designed to effectively capture various potential peak fluctuation patterns in the Peak component. This study integrates Conv2d and Temporal Convolutional Networks to concurrently capture global and local features. Furthermore, we incorporate multi-scale reshaping to augment the modeling capacity for peak fluctuation patterns. The proposed methodology undergoes validation using three publicly accessible seasonal datasets. Notably, in both short-term and long-term fore-casting tasks, our approach exhibits a 10 % reduction in error compared to the baseline models.

[AI-68] GAMED: Knowledge Adaptive Multi-Experts Decoupling for Multimodal Fake News Detection

链接: https://arxiv.org/abs/2412.12164
作者: Lingzhi Shen,Yunfei Long,Xiaohao Cai,Imran Razzak,Guanming Chen,Kang Liu,Shoaib Jameel
关键词: involves modelling heterogeneous, vision and language, modelling heterogeneous data, heterogeneous data sources, Multimodal fake
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Multimodal fake news detection often involves modelling heterogeneous data sources, such as vision and language. Existing detection methods typically rely on fusion effectiveness and cross-modal consistency to model the content, complicating understanding how each modality affects prediction accuracy. Additionally, these methods are primarily based on static feature modelling, making it difficult to adapt to the dynamic changes and relationships between different data modalities. This paper develops a significantly novel approach, GAMED, for multimodal modelling, which focuses on generating distinctive and discriminative features through modal decoupling to enhance cross-modal synergies, thereby optimizing overall performance in the detection process. GAMED leverages multiple parallel expert networks to refine features and pre-embed semantic knowledge to improve the experts’ ability in information selection and viewpoint sharing. Subsequently, the feature distribution of each modality is adaptively adjusted based on the respective experts’ opinions. GAMED also introduces a novel classification technique to dynamically manage contributions from different modalities, while improving the explainability of decisions. Experimental results on the Fakeddit and Yang datasets demonstrate that GAMED performs better than recently developed state-of-the-art models. The source code can be accessed at this https URL.

[AI-69] owards LLM -based optimization compilers. Can LLM LLM s learn how to apply a single peephole optimization? Reasoning is all LLMs need!

链接: https://arxiv.org/abs/2412.12163
作者: Xiangxin Fang,Lev Mukhanov
关键词: Large Language Models, language processing tasks, Large Language, demonstrated great potential, language processing
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Programming Languages (cs.PL)
*备注: 13 pages, 8 figures

点击查看摘要

Abstract:Large Language Models (LLMs) have demonstrated great potential in various language processing tasks, and recent studies have explored their application in compiler optimizations. However, all these studies focus on the conventional open-source LLMs, such as Llama2, which lack enhanced reasoning mechanisms. In this study, we investigate the errors produced by the fine-tuned 7B-parameter Llama2 model as it attempts to learn and apply a simple peephole optimization for the AArch64 assembly code. We provide an analysis of the errors produced by the LLM and compare it with state-of-the-art OpenAI models which implement advanced reasoning logic, including GPT-4o and GPT-o1 (preview). We demonstrate that OpenAI GPT-o1, despite not being fine-tuned, outperforms the fine-tuned Llama2 and GPT-4o. Our findings indicate that this advantage is largely due to the chain-of-thought reasoning implemented in GPT-o1. We hope our work will inspire further research on using LLMs with enhanced reasoning mechanisms and chain-of-thought for code generation and optimization.

[AI-70] Discover Physical Concepts and Equations with Machine Learning

链接: https://arxiv.org/abs/2412.12161
作者: Bao-Bing Li,Yi Gu,Shao-Feng Wu
关键词: Machine learning, uncover physical concepts, learning can uncover, prior knowledge, Neural Ordinary Differential
类目: Machine Learning (cs.LG); Disordered Systems and Neural Networks (cond-mat.dis-nn); Artificial Intelligence (cs.AI); Computational Physics (physics.comp-ph)
*备注:

点击查看摘要

Abstract:Machine learning can uncover physical concepts or physical equations when prior knowledge from another one is available. However, in many cases, these two aspects are coupled and cannot be discovered independently. We extend SciNet, which is a neural network architecture that simulates the human physical reasoning process for physics discovery, by proposing a model that combines Variational Autoencoders (VAEs) with Neural Ordinary Differential Equations (Neural ODEs). This allows us to simultaneously discover physical concepts and governing equations from simulated experimental data across diverse physical systems. We apply the model to several key examples inspired by the history of physics, including Copernicus’ heliocentric solar system, Newton’s law of universal gravitation, the wave function together with the Schrödinger equation, and spin-1/2 along with the Pauli equation. The results demonstrate that the neural network successfully reconstructs the corresponding theories.

[AI-71] Climate Aware Deep Neural Networks (CADNN) for Wind Power Simulation

链接: https://arxiv.org/abs/2412.12160
作者: Ali Forootani,Danial Esmaeili Aliabadi,Daniela Thraen
关键词: modern energy systems, renewable energy sources, Wind power, wind power generation, Wind
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Wind power forecasting plays a critical role in modern energy systems, facilitating the integration of renewable energy sources into the power grid. Accurate prediction of wind energy output is essential for managing the inherent intermittency of wind power, optimizing energy dispatch, and ensuring grid stability. This paper proposes the use of Deep Neural Network (DNN)-based predictive models that leverage climate datasets, including wind speed, atmospheric pressure, temperature, and other meteorological variables, to improve the accuracy of wind power simulations. In particular, we focus on the Coupled Model Intercomparison Project (CMIP) datasets, which provide climate projections, as inputs for training the DNN models. These models aim to capture the complex nonlinear relationships between the CMIP-based climate data and actual wind power generation at wind farms located in Germany. Our study compares various DNN architectures, specifically Multilayer Perceptron (MLP), Long Short-Term Memory (LSTM) networks, and Transformer-enhanced LSTM models, to identify the best configuration among these architectures for climate-aware wind power simulation. The implementation of this framework involves the development of a Python package (CADNN) designed to support multiple tasks, including statistical analysis of the climate data, data visualization, preprocessing, DNN training, and performance evaluation. We demonstrate that the DNN models, when integrated with climate data, significantly enhance forecasting accuracy. This climate-aware approach offers a deeper understanding of the time-dependent climate patterns that influence wind power generation, providing more accurate predictions and making it adaptable to other geographical regions.

[AI-72] Personalized Sleep Staging Leveraging Source-free Unsupervised Domain Adaptation

链接: https://arxiv.org/abs/2412.12159
作者: Yangxuan Zhou,Sha Zhao,Jiquan Wang,Haiteng Jiang,hijian Li,Benyan Luo,Tao Li,Gang Pan
关键词: diagnosing related disorders, assessing sleep quality, related disorders, Sleep staging, crucial for assessing
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: 9 pages, 6 figures

点击查看摘要

Abstract:Sleep staging is crucial for assessing sleep quality and diagnosing related disorders. Recent deep learning models for automatic sleep staging using polysomnography often suffer from poor generalization to new subjects because they are trained and tested on the same labeled datasets, overlooking individual differences. To tackle this issue, we propose a novel Source-Free Unsupervised Individual Domain Adaptation (SF-UIDA) framework. This two-step adaptation scheme allows the model to effectively adjust to new unlabeled individuals without needing source data, facilitating personalized customization in clinical settings. Our framework has been applied to three established sleep staging models and tested on three public datasets, achieving state-of-the-art performance.

[AI-73] Hyperbolic Hypergraph Neural Networks for Multi-Relational Knowledge Hypergraph Representation

链接: https://arxiv.org/abs/2412.12158
作者: Mengfan Li,Xuanhua Shi,Chenqi Qiao,Teng Zhang,Hai Jin
关键词: depict complicated relations, generalize knowledge graphs, connect multiple entities, connect multiple, depict complicated
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Knowledge hypergraphs generalize knowledge graphs using hyperedges to connect multiple entities and depict complicated relations. Existing methods either transform hyperedges into an easier-to-handle set of binary relations or view hyperedges as isolated and ignore their adjacencies. Both approaches have information loss and may potentially lead to the creation of sub-optimal models. To fix these issues, we propose the Hyperbolic Hypergraph Neural Network (H2GNN), whose essential component is the hyper-star message passing, a novel scheme motivated by a lossless expansion of hyperedges into hierarchies. It implements a direct embedding that consciously incorporates adjacent entities, hyper-relations, and entity position-aware information. As the name suggests, H2GNN operates in the hyperbolic space, which is more adept at capturing the tree-like hierarchy. We compare H2GNN with 15 baselines on knowledge hypergraphs, and it outperforms state-of-the-art approaches in both node classification and link prediction tasks.

[AI-74] Adapting Unsigned Graph Neural Networks for Signed Graphs: A Few-Shot Prompt Tuning Approach

链接: https://arxiv.org/abs/2412.12155
作者: Zian Zhai,Sima Qing,Xiaoyang Wang,Wenjie Zhang
关键词: Graph Neural Networks, Neural Networks, Signed Graph, Signed Graph Neural, graph representation learning
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Signed Graph Neural Networks (SGNNs) are powerful tools for signed graph representation learning but struggle with limited generalization and heavy dependence on labeled data. While recent advancements in “graph pre-training and prompt tuning” have reduced label dependence in Graph Neural Networks (GNNs) and improved their generalization abilities by leveraging pre-training knowledge, these efforts have focused exclusively on unsigned graphs. The scarcity of publicly available signed graph datasets makes it essential to transfer knowledge from unsigned graphs to signed graph tasks. However, this transfer introduces significant challenges due to the graph-level and task-level divergences between the pre-training and downstream phases. To address these challenges, we propose Signed Graph Prompt Tuning (SGPT) in this paper. Specifically, SGPT employs a graph template and a semantic prompt to segregate mixed link semantics in the signed graph and then adaptively integrate the distinctive semantic information according to the needs of downstream tasks, thereby unifying the pre-training and downstream graphs. Additionally, SGPT utilizes a task template and a feature prompt to reformulate the downstream signed graph tasks, aligning them with pre-training tasks to ensure a unified optimization objective and consistent feature space across tasks. Finally, extensive experiments are conducted on popular signed graph datasets, demonstrating the superiority of SGPT over state-of-the-art methods.

[AI-75] Revisiting Weight Averaging for Model Merging

链接: https://arxiv.org/abs/2412.12153
作者: Jiho Choi,Donggyun Kim,Chanhyuk Lee,Seunghoon Hong
关键词: Model merging aims, individually fine-tuned models, additional training, aims to build, learner by combining
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:Model merging aims to build a multi-task learner by combining the parameters of individually fine-tuned models without additional training. While a straightforward approach is to average model parameters across tasks, this often results in suboptimal performance due to interference among parameters across tasks. In this paper, we present intriguing results that weight averaging implicitly induces task vectors centered around the weight averaging itself and that applying a low-rank approximation to these centered task vectors significantly improves merging performance. Our analysis shows that centering the task vectors effectively separates core task-specific knowledge and nuisance noise within the fine-tuned parameters into the top and lower singular vectors, respectively, allowing us to reduce inter-task interference through its low-rank approximation. We evaluate our method on eight image classification tasks, demonstrating that it outperforms prior methods by a significant margin, narrowing the performance gap with traditional multi-task learning to within 1-3%

[AI-76] GraphTool-Instruction: Revolutionizing Graph Reasoning in LLM s through Decomposed Subtask Instruction KDD2025

链接: https://arxiv.org/abs/2412.12152
作者: Rongzheng Wang,Shuang Liang,Qizhi Chen,Jiasheng Zhang,Ke Qin
关键词: Large language models, graph reasoning tasks, Large language, graph reasoning, reasoning tasks
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: 22 pages, have been accepted by KDD 2025

点击查看摘要

Abstract:Large language models (LLMs) have been demonstrated to possess the capabilities to understand fundamental graph properties and address various graph reasoning tasks. Existing methods fine-tune LLMs to understand and execute graph reasoning tasks by specially designed task instructions. However, these Text-Instruction methods generally exhibit poor performance. Inspired by tool learning, researchers propose Tool-Instruction methods to solve various graph problems by special tool calling (e.g., function, API and model), achieving significant improvements in graph reasoning tasks. Nevertheless, current Tool-Instruction approaches focus on the tool information and ignore the graph structure information, which leads to significantly inferior performance on small-scale LLMs (less than 13B). To tackle this issue, we propose GraphTool-Instruction, an innovative Instruction-tuning approach that decomposes the graph reasoning task into three distinct subtasks (i.e., graph extraction, tool name identification and tool parameter extraction), and design specialized instructions for each subtask. Our GraphTool-Instruction can be used as a plug-and-play prompt for different LLMs without fine-tuning. Moreover, building on GraphTool-Instruction, we develop GTools, a dataset that includes twenty graph reasoning tasks, and create a graph reasoning LLM called GraphForge based on Llama3-8B. We conduct extensive experiments on twenty graph reasoning tasks with different graph types (e.g., graph size or graph direction), and we find that GraphTool-Instruction achieves SOTA compared to Text-Instruction and Tool-Instruction methods. Fine-tuned on GTools, GraphForge gets further improvement of over 30% compared to the Tool-Instruction enhanced GPT-3.5-turbo, and it performs comparably to the high-cost GPT-4o. Our codes and data are available at this https URL.

[AI-77] SMARTCAL: An Approach to Self-Aware Tool-Use Evaluation and Calibration

链接: https://arxiv.org/abs/2412.12151
作者: Yuanhao Shen,Xiaodan Zhu,Lei Chen
关键词: Large Language Models, Large Language, ability of Large, industrial applications, Language Models
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Software Engineering (cs.SE)
*备注:

点击查看摘要

Abstract:The tool-use ability of Large Language Models (LLMs) has a profound impact on a wide range of industrial applications. However, LLMs’ self-control and calibration capability in appropriately using tools remains understudied. The problem is consequential as it raises potential risks of degraded performance and poses a threat to the trustworthiness of the models. In this paper, we conduct a study on a family of state-of-the-art LLMs on three datasets with two mainstream tool-use frameworks. Our study reveals the tool-abuse behavior of LLMs, a tendency for models to misuse tools with overconfidence. We also find that this is a common issue regardless of model capability. Accordingly, we propose a novel approach, \textitSMARTCAL, to mitigate the observed issues, and our results show an average of 8.6 percent increase in the QA performance and a 21.6 percent decrease in Expected Calibration Error (ECE) compared to baseline models.

[AI-78] Meta-Controller: Few-Shot Imitation of Unseen Embodiments and Tasks in Continuous Control NEURIPS2024

链接: https://arxiv.org/abs/2412.12147
作者: Seongwoong Cho,Donggyun Kim,Jinwoo Lee,Seunghoon Hong
关键词: Generalizing across robot, adaptive robotic systems, robotic systems, embodiments, Generalizing
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Robotics (cs.RO)
*备注: NeurIPS 2024

点击查看摘要

Abstract:Generalizing across robot embodiments and tasks is crucial for adaptive robotic systems. Modular policy learning approaches adapt to new embodiments but are limited to specific tasks, while few-shot imitation learning (IL) approaches often focus on a single embodiment. In this paper, we introduce a few-shot behavior cloning framework to simultaneously generalize to unseen embodiments and tasks using a few (\emphe.g., five) reward-free demonstrations. Our framework leverages a joint-level input-output representation to unify the state and action spaces of heterogeneous embodiments and employs a novel structure-motion state encoder that is parameterized to capture both shared knowledge across all embodiments and embodiment-specific knowledge. A matching-based policy network then predicts actions from a few demonstrations, producing an adaptive policy that is robust to over-fitting. Evaluated in the DeepMind Control suite, our framework termed \modelname demonstrates superior few-shot generalization to unseen embodiments and tasks over modular policy learning and few-shot IL approaches. Codes are available at \hrefthis https URLthis https URL.

[AI-79] Generative Modeling and Data Augmentation for Power System Production Simulation NEURIPS2024

链接: https://arxiv.org/abs/2412.12146
作者: Linna Xu,Yongli Zhu
关键词: power system production, system production simulation, power system, component of power, operation of power
类目: ystems and Control (eess.SY); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注: This paper has been accepted by D3S3: Data-driven and Differentiable Simulations, Surrogates, and Solvers at NeurIPS 2024

点击查看摘要

Abstract:As a key component of power system production simulation, load forecasting is critical for the stable operation of power systems. Machine learning methods prevail in this field. However, the limited training data can be a challenge. This paper proposes a generative model-assisted approach for load forecasting under small sample scenarios, consisting of two steps: expanding the dataset using a diffusion-based generative model and then training various machine learning regressors on the augmented dataset to identify the best performer. The expanded dataset significantly reduces forecasting errors compared to the original dataset, and the diffusion model outperforms the generative adversarial model by achieving about 200 times smaller errors and better alignment in latent data distributions.

[AI-80] Inverse design of potential metastructures inspired from Indian medieval architectural elements

链接: https://arxiv.org/abs/2412.12122
作者: Bishakh Bhattacharya,Tanuj Gupta,Arun Kumar Sharma,Ankur Dwivedi,Vivek Gupta,Subhadeep Sahana,Suryansh Pathak,Ashish Awasthi
关键词: Indian medieval architecture, mechanical metastructure perspective, details of Indian, Indian medieval, world of patterns
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Signal Processing (eess.SP)
*备注:

点击查看摘要

Abstract:In this study, we immerse in the intricate world of patterns, examining the structural details of Indian medieval architecture for the discovery of motifs with great application potential from the mechanical metastructure perspective. The motifs that specifically engrossed us are derived from the tomb of I’timad-ud-Daula, situated in the city of Agra, close to the Taj Mahal. In an exploratory study, we designed nine interlaced metastructures inspired from the tomb’s motifs. We fabricated the metastructures using additive manufacturing and studied their vibration characteristics experimentally and numerically. We also investigated bandgap modulation with metallic inserts in honeycomb interlaced metastructures. The comprehensive study of these metastructure panels reveals their high performance in controlling elastic wave propagation and generating suitable frequency bandgaps, hence having potential applications as waveguides for noise and vibration control. Finally, we developed novel AI-based model trained on numerical datasets for inverse design of metastructures with a desired bandgap.

[AI-81] AI in Education: Rationale Principles and Instructional Implications

链接: https://arxiv.org/abs/2412.12116
作者: Eyvind Elstad
关键词: assessing its benefits, benefits and risks, examines the integration, study examines, integration of generative
类目: Computers and Society (cs.CY); Artificial Intelligence (cs.AI)
*备注: 24 pages

点击查看摘要

Abstract:This study examines the integration of generative AI in schools, assessing its benefits and risks. As AI use by students grows, it’s crucial to understand its impact on learning and teaching practices. Generative AI, like ChatGPT, can create human-like content, prompting questions about its educational role. The article differentiates large language models from traditional search engines and stresses the need for students to develop critical source evaluation skills. Although empirical evidence on AI’s classroom effects is limited, AI offers personalized learning support and problem-solving tools, alongside challenges like undermining deep learning if misused. The study emphasizes deliberate strategies to ensure AI complements, not replaces, genuine cognitive effort. AI’s educational role should be context-dependent, guided by pedagogical goals. The study concludes with practical advice for teachers on effectively utilizing AI to promote understanding and critical engagement, advocating for a balanced approach to enhance students’ knowledge and skills development.

[AI-82] Responsible AI Governance: A Response to UN Interim Report on Governing AI for Humanity

链接: https://arxiv.org/abs/2412.12108
作者: Sarah Kiden,Bernd Stahl,Beverley Townsend,Carsten Maple,Charles Vincent,Fraser Sampson,Geoff Gilbert,Helen Smith,Jayati Deshmukh,Jen Ross,Jennifer Williams,Jesus Martinez del Rincon,Justyna Lisinska,Karen O’Shea,Márjory Da Costa Abreu,Nelly Bencomo,Oishi Deb,Peter Winter,Phoebe Li,Philip Torr,Pin Lean Lau,Raquel Iniesta,Gopal Ramchurn,Sebastian Stein,Vahid Yazdanpanah
关键词: Governing Artificial Intelligence, United Nation Interim, Nation Interim Report, Artificial Intelligence, United Nation
类目: Computers and Society (cs.CY); Artificial Intelligence (cs.AI)
*备注: Submitted to United Nations. 23 pages

点击查看摘要

Abstract:This report presents a comprehensive response to the United Nation’s Interim Report on Governing Artificial Intelligence (AI) for Humanity. It emphasizes the transformative potential of AI in achieving the Sustainable Development Goals (SDGs) while acknowledging the need for robust governance to mitigate associated risks. The response highlights opportunities for promoting equitable, secure, and inclusive AI ecosystems, which should be supported by investments in infrastructure and multi-stakeholder collaborations across jurisdictions. It also underscores challenges, including societal inequalities exacerbated by AI, ethical concerns, and environmental impacts. Recommendations advocate for legally binding norms, transparency, and multi-layered data governance models, alongside fostering AI literacy and capacity-building initiatives. Internationally, the report calls for harmonising AI governance frameworks with established laws, human rights standards, and regulatory approaches. The report concludes with actionable principles for fostering responsible AI governance through collaboration among governments, industry, academia, and civil society, ensuring the development of AI aligns with universal human values and the public good.

[AI-83] Generative AI Literacy: Twelve Defining Competencies

链接: https://arxiv.org/abs/2412.12107
作者: Ravinithesh Annapureddy,Alessandro Fornaroli,Daniel Gatica-Perez
关键词: generative artificial intelligence, literacy covering essential, covering essential skills, artificial intelligence, paper introduces
类目: Human-Computer Interaction (cs.HC); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:This paper introduces a competency-based model for generative artificial intelligence (AI) literacy covering essential skills and knowledge areas necessary to interact with generative AI. The competencies range from foundational AI literacy to prompt engineering and programming skills, including ethical and legal considerations. These twelve competencies offer a framework for individuals, policymakers, government officials, and educators looking to navigate and take advantage of the potential of generative AI responsibly. Embedding these competencies into educational programs and professional training initiatives can equip individuals to become responsible and informed users and creators of generative AI. The competencies follow a logical progression and serve as a roadmap for individuals seeking to get familiar with generative AI and for researchers and policymakers to develop assessments, educational programs, guidelines, and regulations.

[AI-84] Empathic Coupling of Homeostatic States for Intrinsic Prosociality

链接: https://arxiv.org/abs/2412.12103
作者: Naoto Yoshida,Kingson Man
关键词: experience personal distress, experience personal, feel compelled, prosocial behavior, personal distress
类目: Multiagent Systems (cs.MA); Artificial Intelligence (cs.AI); Neural and Evolutionary Computing (cs.NE)
*备注:

点击查看摘要

Abstract:When regarding the suffering of others, we often experience personal distress and feel compelled to help. Inspired by living systems, we investigate the emergence of prosocial behavior among autonomous agents that are motivated by homeostatic self-regulation. We perform multi-agent reinforcement learning, treating each agent as a vulnerable homeostat charged with maintaining its own well-being. We introduce an empathy-like mechanism to share homeostatic states between agents: an agent can either \emphobserve their partner’s internal state (cognitive empathy) or the agent’s internal state can be \emphdirectly coupled to that of their partner’s (affective empathy). In three simple multi-agent environments, we show that prosocial behavior arises only under homeostatic coupling - when the distress of a partner can affect one’s own well-being. Our findings specify the type and role of empathy in artificial agents capable of prosocial behavior.

[AI-85] Distributed Collaborative Inference System in Next-Generation Networks and Communication

链接: https://arxiv.org/abs/2412.12102
作者: Chuan Zhang,Xixi Zheng,Xiaolong Tao,Chenfei Hu,Weiting Zhang,Liehuang Zhu
关键词: generative artificial intelligence, artificial intelligence, generative artificial, data processing methods, transforming data processing
类目: Networking and Internet Architecture (cs.NI); Artificial Intelligence (cs.AI)
*备注:

点击查看摘要

Abstract:With the rapid advancement of artificial intelligence, generative artificial intelligence (GAI) has taken a leading role in transforming data processing methods. However, the high computational demands of GAI present challenges for devices with limited resources. As we move towards the sixth generation of mobile networks (6G), the higher data rates and improved energy efficiency of 6G create a need for more efficient data processing in GAI. Traditional GAI, however, shows its limitations in meeting these demands. To address these challenges, we introduce a multi-level collaborative inference system designed for next-generation networks and communication. Our proposed system features a deployment strategy that assigns models of varying sizes to devices at different network layers. Then, we design a task offloading strategy to optimise both efficiency and latency. Furthermore, a modified early exit mechanism is implemented to enhance the inference process for single models. Experimental results demonstrate that our system effectively reduces inference latency while maintaining high-quality output. Specifically, compared to existing work, our system can reduce inference time by up to 17% without sacrificing the inference accuracy.

[AI-86] Qsco: A Quantum Scoring Module for Open-set Supervised Anomaly Detection AAAI-25 AAAI

链接: https://arxiv.org/abs/2405.16368
作者: Yifeng Peng,Xinyi Li,Zhiding Liang,Ying Wang
关键词: Open set anomaly, identify abnormal patterns, Open set, set anomaly detection, anomaly detection
类目: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
*备注: The Thirty-Ninth AAAI Conference on Artificial Intelligence (AAAI-25)

点击查看摘要

Abstract:Open set anomaly detection (OSAD) is a crucial task that aims to identify abnormal patterns or behaviors in data sets, especially when the anomalies observed during training do not represent all possible classes of anomalies. The recent advances in quantum computing in handling complex data structures and improving machine learning models herald a paradigm shift in anomaly detection methodologies. This study proposes a Quantum Scoring Module (Qsco), embedding quantum variational circuits into neural networks to enhance the model’s processing capabilities in handling uncertainty and unlabeled data. Extensive experiments conducted across eight real-world anomaly detection datasets demonstrate our model’s superior performance in detecting anomalies across varied settings and reveal that integrating quantum simulators does not result in prohibitive time complexities. Our study validates the feasibility of quantum-enhanced anomaly detection methods in practical applications.

[AI-87] Equity in the Use of ChatGPT for the Classroom: A Comparison of the Accuracy and Precision of ChatGPT 3.5 vs. ChatGPT4 with Respect to Statistics and Data Science Exams

链接: https://arxiv.org/abs/2412.13116
作者: Monnie McGee,Bivin Sadler
关键词: college education historically, social status, moving upward, brackets and social, income brackets
类目: Other Statistics (stat.OT); Artificial Intelligence (cs.AI)
*备注: Originally submitted for review in May of 2024 but rejected 6 months later

点击查看摘要

Abstract:A college education historically has been seen as method of moving upward with regards to income brackets and social status. Indeed, many colleges recognize this connection and seek to enroll talented low income students. While these students might have their education, books, room, and board paid; there are other items that they might be expected to use that are not part of most college scholarship packages. One of those items that has recently surfaced is access to generative AI platforms. The most popular of these platforms is ChatGPT, and it has a paid version (ChatGPT4) and a free version (ChatGPT3.5). We seek to explore differences in the free and paid versions in the context of homework questions and data analyses as might be seen in a typical introductory statistics course. We determine the extent to which students who cannot afford newer and faster versions of generative AI programs would be disadvantaged in terms of writing such projects and learning these methods.

[AI-88] Stochastic interior-point methods for smooth conic optimization with applications

链接: https://arxiv.org/abs/2412.12987
作者: Chuan He,Zhanwang Deng
关键词: general conic optimization, Conic optimization plays, Conic optimization, plays a crucial, crucial role
类目: Optimization and Control (math.OC); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Conic optimization plays a crucial role in many machine learning (ML) problems. However, practical algorithms for conic constrained ML problems with large datasets are often limited to specific use cases, as stochastic algorithms for general conic optimization remain underdeveloped. To fill this gap, we introduce a stochastic interior-point method (SIPM) framework for general conic optimization, along with four novel SIPM variants leveraging distinct stochastic gradient estimators. Under mild assumptions, we establish the global convergence rates of our proposed SIPMs, which, up to a logarithmic factor, match the best-known rates in stochastic unconstrained optimization. Finally, our numerical experiments on robust linear regression, multi-task relationship learning, and clustering data streams demonstrate the effectiveness and efficiency of our approach.

[AI-89] Subspace Implicit Neural Representations for Real-Time Cardiac Cine MR Imaging

链接: https://arxiv.org/abs/2412.12742
作者: Wenqi Huang,Veronika Spieker,Siying Xu,Gastao Cruz,Claudia Prieto,Julia Schnabel,Kerstin Hammernik,Thomas Kuestner,Daniel Rueckert
关键词: cardiac cine MRI, cine MRI, capture continuous cardiac, limits temporal resolution, MRI methods rely
类目: Image and Video Processing (eess.IV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Conventional cardiac cine MRI methods rely on retrospective gating, which limits temporal resolution and the ability to capture continuous cardiac dynamics, particularly in patients with arrhythmias and beat-to-beat variations. To address these challenges, we propose a reconstruction framework based on subspace implicit neural representations for real-time cardiac cine MRI of continuously sampled radial data. This approach employs two multilayer perceptrons to learn spatial and temporal subspace bases, leveraging the low-rank properties of cardiac cine MRI. Initialized with low-resolution reconstructions, the networks are fine-tuned using spoke-specific loss functions to recover spatial details and temporal fidelity. Our method directly utilizes the continuously sampled radial k-space spokes during training, thereby eliminating the need for binning and non-uniform FFT. This approach achieves superior spatial and temporal image quality compared to conventional binned methods at the acceleration rate of 10 and 20, demonstrating potential for high-resolution imaging of dynamic cardiac events and enhancing diagnostic capability.

[AI-90] Evolutionary Optimization for Designing Variational Quantum Circuits with High Model Capacity

链接: https://arxiv.org/abs/2412.12484
作者: Samuel Yen-Chi Chen
关键词: garnered significant attention, Recent advancements, QML models, quantum machine learning, QML
类目: Quantum Physics (quant-ph); Artificial Intelligence (cs.AI); Emerging Technologies (cs.ET); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE)
*备注: Accepted by IEEE Symposium Series on Computational Intelligence - IEEE SSCI 2025

点击查看摘要

Abstract:Recent advancements in quantum computing (QC) and machine learning (ML) have garnered significant attention, leading to substantial efforts toward the development of quantum machine learning (QML) algorithms to address a variety of complex challenges. The design of high-performance QML models, however, requires expert-level knowledge, posing a significant barrier to the widespread adoption of QML. Key challenges include the design of data encoding mechanisms and parameterized quantum circuits, both of which critically impact the generalization capabilities of QML models. We propose a novel method that encodes quantum circuit architecture information to enable the evolution of quantum circuit designs. In this approach, the fitness function is based on the effective dimension, allowing for the optimization of quantum circuits towards higher model capacity. Through numerical simulations, we demonstrate that the proposed method is capable of discovering variational quantum circuit architectures that offer improved learning capabilities, thereby enhancing the overall performance of QML models for complex tasks.

[AI-91] Pop-out vs. Glue: A Study on the pre-attentive and focused attention stages in Visual Search tasks

链接: https://arxiv.org/abs/2412.12198
作者: Hendrik Beukelman,Wilder C. Rodrigues
关键词: Feature Integration Theory, Integration Theory, Treisman Feature Integration, Treisman Gormican, study explores visual
类目: Neurons and Cognition (q-bio.NC); Artificial Intelligence (cs.AI); Methodology (stat.ME)
*备注: Replication of Gupta et al work from 2015 paper

点击查看摘要

Abstract:This study explores visual search asymmetry and the detection process between parallel and serial search strategies, building upon Treisman’s Feature Integration Theory [3]. Our experiment examines how easy it is to locate an oblique line among vertical distractors versus a vertical line among oblique distractors, a framework previously validated by Treisman Gormican (1988) [4] and Gupta et al. (2015) [1]. We hypothesised that an oblique target among vertical lines would produce a perceptual ‘pop-out’ effect, allowing for faster, parallel search, while the reverse condition would require serial search strategy. Seventy-eight participants from Utrecht University engaged in trials with varied target-distractor orientations and number of items. We measured reaction times and found a significant effect of target type on search speed: oblique targets were identified more quickly, reflecting ‘pop-out’ behaviour, while vertical targets demanded focused attention (‘glue phase’). Our results align with past findings, supporting our hypothesis on search asymmetry and its dependency on distinct visual features. Future research could benefit from eye-tracking and neural network analysis, particularly for identifying the neural processing of visual features in both parallel and serial search conditions.

[AI-92] InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders

链接: https://arxiv.org/abs/2412.12101
作者: Elana Simon,James Zou
关键词: remain poorly understood, demonstrated remarkable success, function remain poorly, modeling and design, poorly understood
类目: Biomolecules (q-bio.BM); Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Quantitative Methods (q-bio.QM)
*备注:

点击查看摘要

Abstract:Protein language models (PLMs) have demonstrated remarkable success in protein modeling and design, yet their internal mechanisms for predicting structure and function remain poorly understood. Here we present a systematic approach to extract and analyze interpretable features from PLMs using sparse autoencoders (SAEs). By training SAEs on embeddings from the PLM ESM-2, we identify up to 2,548 human-interpretable latent features per layer that strongly correlate with up to 143 known biological concepts such as binding sites, structural motifs, and functional domains. In contrast, examining individual neurons in ESM-2 reveals up to 46 neurons per layer with clear conceptual alignment across 15 known concepts, suggesting that PLMs represent most concepts in superposition. Beyond capturing known annotations, we show that ESM-2 learns coherent concepts that do not map onto existing annotations and propose a pipeline using language models to automatically interpret novel latent features learned by the SAEs. As practical applications, we demonstrate how these latent features can fill in missing annotations in protein databases and enable targeted steering of protein sequence generation. Our results demonstrate that PLMs encode rich, interpretable representations of protein biology and we propose a systematic framework to extract and analyze these latent features. In the process, we recover both known biology and potentially new protein motifs. As community resources, we introduce InterPLM (this http URL), an interactive visualization platform for exploring and analyzing learned PLM features, and release code for training and analysis at this http URL.

机器学习

[LG-0] A Conformal Approach to Feature-based Newsvendor under Model Misspecification

链接: https://arxiv.org/abs/2412.13159
作者: Junyu Cao
关键词: data-driven decision-making problems, fail in practice, data-driven decision-making, depend heavily, frequently fail
类目: Machine Learning (cs.LG); Machine Learning (stat.ML)
*备注:

点击查看摘要

Abstract:In many data-driven decision-making problems, performance guarantees often depend heavily on the correctness of model assumptions, which may frequently fail in practice. We address this issue in the context of a feature-based newsvendor problem, where demand is influenced by observed features such as demographics and seasonality. To mitigate the impact of model misspecification, we propose a model-free and distribution-free framework inspired by conformal prediction. Our approach consists of two phases: a training phase, which can utilize any type of prediction method, and a calibration phase that conformalizes the model bias. To enhance predictive performance, we explore the balance between data quality and quantity, recognizing the inherent trade-off: more selective training data improves quality but reduces quantity. Importantly, we provide statistical guarantees for the conformalized critical quantile, independent of the correctness of the underlying model. Moreover, we quantify the confidence interval of the critical quantile, with its width decreasing as data quality and quantity improve. We validate our framework using both simulated data and a real-world dataset from the Capital Bikeshare program in Washington, D.C. Across these experiments, our proposed method consistently outperforms benchmark algorithms, reducing newsvendor loss by up to 40% on the simulated data and 25% on the real-world dataset.

[LG-1] Learning Visuotactile Estimation and Control for Non-prehensile Manipulation under Occlusions

链接: https://arxiv.org/abs/2412.13157
作者: Juan Del Aguila Ferrandis,João Moura,Sethu Vijayakumar
关键词: contact-rich environments, relating with underactuation, essential for dexterous, challenges relating, dexterous robots
类目: Robotics (cs.RO); Machine Learning (cs.LG)
*备注: Conference on Robot Learning (CoRL 2024)

点击查看摘要

Abstract:Manipulation without grasping, known as non-prehensile manipulation, is essential for dexterous robots in contact-rich environments, but presents many challenges relating with underactuation, hybrid-dynamics, and frictional uncertainty. Additionally, object occlusions in a scenario of contact uncertainty and where the motion of the object evolves independently from the robot becomes a critical problem, which previous literature fails to address. We present a method for learning visuotactile state estimators and uncertainty-aware control policies for non-prehensile manipulation under occlusions, by leveraging diverse interaction data from privileged policies trained in simulation. We formulate the estimator within a Bayesian deep learning framework, to model its uncertainty, and then train uncertainty-aware control policies by incorporating the pre-learned estimator into the reinforcement learning (RL) loop, both of which lead to significantly improved estimator and policy performance. Therefore, unlike prior non-prehensile research that relies on complex external perception set-ups, our method successfully handles occlusions after sim-to-real transfer to robotic hardware with a simple onboard camera. See our video: this https URL.

[LG-2] Practicable Black-box Evasion Attacks on Link Prediction in Dynamic Graphs – A Graph Sequential Embedding Method

链接: https://arxiv.org/abs/2412.13134
作者: Jiate Li,Meng Pang,Binghui Wang
关键词: traffic flow prediction, Link prediction, organizational studies, flow prediction, website recommendation
类目: Cryptography and Security (cs.CR); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Link prediction in dynamic graphs (LPDG) has been widely applied to real-world applications such as website recommendation, traffic flow prediction, organizational studies, etc. These models are usually kept local and secure, with only the interactive interface restrictively available to the public. Thus, the problem of the black-box evasion attack on the LPDG model, where model interactions and data perturbations are restricted, seems to be essential and meaningful in practice. In this paper, we propose the first practicable black-box evasion attack method that achieves effective attacks against the target LPDG model, within a limited amount of interactions and perturbations. To perform effective attacks under limited perturbations, we develop a graph sequential embedding model to find the desired state embedding of the dynamic graph sequences, under a deep reinforcement learning framework. To overcome the scarcity of interactions, we design a multi-environment training pipeline and train our agent for multiple instances, by sharing an aggregate interaction buffer. Finally, we evaluate our attack against three advanced LPDG models on three real-world graph datasets of different scales and compare its performance with related methods under the interaction and perturbation constraints. Experimental results show that our attack is both effective and practicable.

[LG-3] Active Reinforcement Learning Strategies for Offline Policy Improvement AAAI2025

链接: https://arxiv.org/abs/2412.13106
作者: Ambedkar Dukkipati,Ranga Shaarad Ayyagari,Bodhisattwa Dasgupta,Parag Dutta,Prabhas Reddy Onteru
关键词: sequential decision-making tasks, excel at sequential, sequential decision-making, decision-making tasks, tasks must continuously
类目: Machine Learning (cs.LG)
*备注: AAAI 2025

点击查看摘要

Abstract:Learning agents that excel at sequential decision-making tasks must continuously resolve the problem of exploration and exploitation for optimal learning. However, such interactions with the environment online might be prohibitively expensive and may involve some constraints, such as a limited budget for agent-environment interactions and restricted exploration in certain regions of the state space. Examples include selecting candidates for medical trials and training agents in complex navigation environments. This problem necessitates the study of active reinforcement learning strategies that collect minimal additional experience trajectories by reusing existing offline data previously collected by some unknown behavior policy. In this work, we propose a representation-aware uncertainty-based active trajectory collection method that intelligently decides interaction strategies that consider the distribution of the existing offline data. With extensive experimentation, we demonstrate that our proposed method reduces additional online interaction with the environment by up to 75% over competitive baselines across various continuous control environments.

[LG-4] Reservoir Computing for Fast Simplified Reinforcement Learning on Memory Tasks

链接: https://arxiv.org/abs/2412.13093
作者: Kevin McKee
关键词: current observation set, past information, current observation, equipped with short-term, observation set
类目: Machine Learning (cs.LG)
*备注: 9 pages, 6 figures

点击查看摘要

Abstract:Tasks in which rewards depend upon past information not available in the current observation set can only be solved by agents that are equipped with short-term memory. Usual choices for memory modules include trainable recurrent hidden layers, often with gated memory. Reservoir computing presents an alternative, in which a recurrent layer is not trained, but rather has a set of fixed, sparse recurrent weights. The weights are scaled to produce stable dynamical behavior such that the reservoir state contains a high-dimensional, nonlinear impulse response function of the inputs. An output decoder network can then be used to map the compressive history represented by the reservoir’s state to any outputs, including agent actions or predictions. In this study, we find that reservoir computing greatly simplifies and speeds up reinforcement learning on memory tasks by (1) eliminating the need for backpropagation of gradients through time, (2) presenting all recent history simultaneously to the downstream network, and (3) performing many useful and generic nonlinear computations upstream from the trained modules. In particular, these findings offer significant benefit to meta-learning that depends primarily on efficient and highly general memory systems.

[LG-5] Predicting Change Not States: An Alternate Framework for Neural PDE Surrogates

链接: https://arxiv.org/abs/2412.13074
作者: Anthony Zhou,Amir Barati Farimani
关键词: partial differential equations, quickly simulate physics, Neural surrogates, differential equations, partial differential
类目: Machine Learning (cs.LG)
*备注: 18 pages, 7 figures

点击查看摘要

Abstract:Neural surrogates for partial differential equations (PDEs) have become popular due to their potential to quickly simulate physics. With a few exceptions, neural surrogates generally treat the forward evolution of time-dependent PDEs as a black box by directly predicting the next state. While this is a natural and easy framework for applying neural surrogates, it can be an over-simplified and rigid framework for predicting physics. In this work, we propose an alternative framework in which neural solvers predict the temporal derivative and an ODE integrator forwards the solution in time, which has little overhead and is broadly applicable across model architectures and PDEs. We find that by simply changing the training target and introducing numerical integration during inference, neural surrogates can gain accuracy and stability. Predicting temporal derivatives also allows models to not be constrained to a specific temporal discretization, allowing for flexible time-stepping during inference or training on higher-resolution PDE data. Lastly, we investigate why this new framework can be beneficial and in what situations does it work well.

[LG-6] On the Hardness of Training Deep Neural Networks Discretely AAAI2025

链接: https://arxiv.org/abs/2412.13057
作者: Ilan Doron-Arad
关键词: neural network training, study neural network, neural network parameters, training loss, NNT
类目: Machine Learning (cs.LG)
*备注: Accepted to AAAI 2025

点击查看摘要

Abstract:We study neural network training (NNT): optimizing a neural network’s parameters to minimize the training loss over a given dataset. NNT has been studied extensively under theoretic lenses, mainly on two-layer networks with linear or ReLU activation functions where the parameters can take any real value (here referred to as continuous NNT (C-NNT)). However, less is known about deeper neural networks, which exhibit substantially stronger capabilities in practice. In addition, the complexity of the discrete variant of the problem (D-NNT in short), in which the parameters are taken from a given finite set of options, has remained less explored despite its theoretical and practical significance. In this work, we show that the hardness of NNT is dramatically affected by the network depth. Specifically, we show that, under standard complexity assumptions, D-NNT is not in the complexity class NP even for instances with fixed dimensions and dataset size, having a deep architecture. This separates D-NNT from any NP-complete problem. Furthermore, using a polynomial reduction we show that the above result also holds for C-NNT, albeit with more structured instances. We complement these results with a comprehensive list of NP-hardness lower bounds for D-NNT on two-layer networks, showing that fixing the number of dimensions, the dataset size, or the number of neurons in the hidden layer leaves the problem challenging. Finally, we obtain a pseudo-polynomial algorithm for D-NNT on a two-layer network with a fixed dataset size. Comments: Accepted to AAAI 2025 Subjects: Machine Learning (cs.LG) Cite as: arXiv:2412.13057 [cs.LG] (or arXiv:2412.13057v1 [cs.LG] for this version) https://doi.org/10.48550/arXiv.2412.13057 Focus to learn more arXiv-issued DOI via DataCite (pending registration)

[LG-7] IMESAFE: Timing Interruption Monitoring and Security Assessment for Fronthaul Environments

链接: https://arxiv.org/abs/2412.13049
作者: Joshua Groen,Simone Di Valerio,Imtiaz Karim,Davide Villa,Yiewi Zhang,Leonardo Bonati,Michele Polese,Salvatore D’Oro,Tommaso Melodia,Elisa Bertino,Francesca Cuomo,Kaushik Chowdhury
关键词: radio unit equipment, Radio Access Network, Radio Access, radio unit, Access Network
类目: Networking and Internet Architecture (cs.NI); Cryptography and Security (cs.CR); Machine Learning (cs.LG); Systems and Control (eess.SY)
*备注:

点击查看摘要

Abstract:5G and beyond cellular systems embrace the disaggregation of Radio Access Network (RAN) components, exemplified by the evolution of the fronthual (FH) connection between cellular baseband and radio unit equipment. Crucially, synchronization over the FH is pivotal for reliable 5G services. In recent years, there has been a push to move these links to an Ethernet-based packet network topology, leveraging existing standards and ongoing research for Time-Sensitive Networking (TSN). However, TSN standards, such as Precision Time Protocol (PTP), focus on performance with little to no concern for security. This increases the exposure of the open FH to security risks. Attacks targeting synchronization mechanisms pose significant threats, potentially disrupting 5G networks and impairing connectivity. In this paper, we demonstrate the impact of successful spoofing and replay attacks against PTP synchronization. We show how a spoofing attack is able to cause a production-ready O-RAN and 5G-compliant private cellular base station to catastrophically fail within 2 seconds of the attack, necessitating manual intervention to restore full network operations. To counter this, we design a Machine Learning (ML)-based monitoring solution capable of detecting various malicious attacks with over 97.5% accuracy. Subjects: Networking and Internet Architecture (cs.NI); Cryptography and Security (cs.CR); Machine Learning (cs.LG); Systems and Control (eess.SY) ACMclasses: C.2; C.4 Cite as: arXiv:2412.13049 [cs.NI] (or arXiv:2412.13049v1 [cs.NI] for this version) https://doi.org/10.48550/arXiv.2412.13049 Focus to learn more arXiv-issued DOI via DataCite (pending registration)

[LG-8] Open-Set Heterogeneous Domain Adaptation: Theoretical Analysis and Algorithm AAAI2025

链接: https://arxiv.org/abs/2412.13036
作者: Thai-Hoang Pham,Yuanlong Wang,Changchang Yin,Xueru Zhang,Ping Zhang
关键词: target domain, issue of distribution, distribution shift, Domain, Domain adaptation
类目: Machine Learning (cs.LG)
*备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Domain adaptation (DA) tackles the issue of distribution shift by learning a model from a source domain that generalizes to a target domain. However, most existing DA methods are designed for scenarios where the source and target domain data lie within the same feature space, which limits their applicability in real-world situations. Recently, heterogeneous DA (HeDA) methods have been introduced to address the challenges posed by heterogeneous feature space between source and target domains. Despite their successes, current HeDA techniques fall short when there is a mismatch in both feature and label spaces. To address this, this paper explores a new DA scenario called open-set HeDA (OSHeDA). In OSHeDA, the model must not only handle heterogeneity in feature space but also identify samples belonging to novel classes. To tackle this challenge, we first develop a novel theoretical framework that constructs learning bounds for prediction error on target domain. Guided by this framework, we propose a new DA method called Representation Learning for OSHeDA (RL-OSHeDA). This method is designed to simultaneously transfer knowledge between heterogeneous data sources and identify novel classes. Experiments across text, image, and clinical data demonstrate the effectiveness of our algorithm. Model implementation is available at \urlthis https URL.

[LG-9] Queries Representation Detection: The Next 100 Model Fingerprinting Schemes AAAI2025

链接: https://arxiv.org/abs/2412.13021
作者: Augustin Godinot,Erwan Le Merrer,Camilla Penzo,François Taïani,Gilles Trédan
关键词: operational contexts represents, machine learning models, deployment of machine, machine learning, operational contexts
类目: Machine Learning (cs.LG); Cryptography and Security (cs.CR)
*备注: Accepted to AAAI2025 Main Technical Track

点击查看摘要

Abstract:The deployment of machine learning models in operational contexts represents a significant investment for any organisation. Consequently, the risk of these models being misappropriated by competitors needs to be addressed. In recent years, numerous proposals have been put forth to detect instances of model stealing. However, these proposals operate under implicit and disparate data and model access assumptions; as a consequence, it remains unclear how they can be effectively compared to one another. Our evaluation shows that a simple baseline that we introduce performs on par with existing state-of-the-art fingerprints, which, on the other hand, are much more complex. To uncover the reasons behind this intriguing result, this paper introduces a systematic approach to both the creation of model fingerprinting schemes and their evaluation benchmarks. By dividing model fingerprinting into three core components – Query, Representation and Detection (QuRD) – we are able to identify \sim100 previously unexplored QuRD combinations and gain insights into their performance. Finally, we introduce a set of metrics to compare and guide the creation of more representative model stealing detection benchmarks. Our approach reveals the need for more challenging benchmarks and a sound comparison with baselines. To foster the creation of new fingerprinting schemes and benchmarks, we open-source our fingerprinting toolbox.

[LG-10] Deep Learning Based Superconductivity: Prediction and Experimental Tests

链接: https://arxiv.org/abs/2412.13012
作者: Daniel Kaplan,Adam Zhang,Joanna Blawat,Rongying Jin,Robert J. Cava,Viktor Oudovenko,Gabriel Kotliar,Anirvan M. Sengupta,Weiwei Xie
关键词: applications in energy, materials science, materials, superconducting materials, textrm
类目: Machine Learning (cs.LG); Materials Science (cond-mat.mtrl-sci); Strongly Correlated Electrons (cond-mat.str-el)
*备注: 14 pages + 2 appendices + references. EPJ submission

点击查看摘要

Abstract:The discovery of novel superconducting materials is a longstanding challenge in materials science, with a wealth of potential for applications in energy, transportation, and computing. Recent advances in artificial intelligence (AI) have enabled expediting the search for new materials by efficiently utilizing vast materials databases. In this study, we developed an approach based on deep learning (DL) to predict new superconducting materials. We have synthesized a compound derived from our DL network and confirmed its superconducting properties in agreement with our prediction. Our approach is also compared to previous work based on random forests (RFs). In particular, RFs require knowledge of the chem-ical properties of the compound, while our neural net inputs depend solely on the chemical composition. With the help of hints from our network, we discover a new ternary compound \textrmMo_20\textrmRe_6\textrmSi_4 , which becomes superconducting below 5.4 K. We further discuss the existing limitations and challenges associated with using AI to predict and, along with potential future research directions.

[LG-11] Boosting Test Performance with Importance Sampling–a Subpopulation Perspective

链接: https://arxiv.org/abs/2412.13003
作者: Hongyu Shen,Zhizhen Zhao
关键词: empirical risk minimization, machine learning community, risk minimization, learning community, hidden attributes
类目: Machine Learning (cs.LG); Machine Learning (stat.ML)
*备注: 16 pages, 1 figure, 2 tables

点击查看摘要

Abstract:Despite empirical risk minimization (ERM) is widely applied in the machine learning community, its performance is limited on data with spurious correlation or subpopulation that is introduced by hidden attributes. Existing literature proposed techniques to maximize group-balanced or worst-group accuracy when such correlation presents, yet, at the cost of lower average accuracy. In addition, many existing works conduct surveys on different subpopulation methods without revealing the inherent connection between these methods, which could hinder the technology advancement in this area. In this paper, we identify important sampling as a simple yet powerful tool for solving the subpopulation problem. On the theory side, we provide a new systematic formulation of the subpopulation problem and explicitly identify the assumptions that are not clearly stated in the existing works. This helps to uncover the cause of the dropped average accuracy. We provide the first theoretical discussion on the connections of existing methods, revealing the core components that make them different. On the application side, we demonstrate a single estimator is enough to solve the subpopulation problem. In particular, we introduce the estimator in both attribute-known and -unknown scenarios in the subpopulation setup, offering flexibility in practical use cases. And empirically, we achieve state-of-the-art performance on commonly used benchmark datasets.

[LG-12] ArchesWeather ArchesWeatherGen: a deterministic and generative model for efficient ML weather forecasting

链接: https://arxiv.org/abs/2412.12971
作者: Guillaume Couairon,Renu Singh,Anastase Charantonis,Christian Lessig,Claire Monteleoni
关键词: extreme weather events, Weather forecasting plays, Weather, today society, renewable energies
类目: Machine Learning (cs.LG)
*备注: Preprint

点击查看摘要

Abstract:Weather forecasting plays a vital role in today’s society, from agriculture and logistics to predicting the output of renewable energies, and preparing for extreme weather events. Deep learning weather forecasting models trained with the next state prediction objective on ERA5 have shown great success compared to numerical global circulation models. However, for a wide range of applications, being able to provide representative samples from the distribution of possible future weather states is critical. In this paper, we propose a methodology to leverage deterministic weather models in the design of probabilistic weather models, leading to improved performance and reduced computing costs. We first introduce \textbfArchesWeather, a transformer-based deterministic model that improves upon Pangu-Weather by removing overrestrictive inductive priors. We then design a probabilistic weather model called \textbfArchesWeatherGen based on flow matching, a modern variant of diffusion models, that is trained to project ArchesWeather’s predictions to the distribution of ERA5 weather states. ArchesWeatherGen is a true stochastic emulator of ERA5 and surpasses IFS ENS and NeuralGCM on all WeatherBench headline variables (except for NeuralGCM’s geopotential). Our work also aims to democratize the use of deterministic and generative machine learning models in weather forecasting research, with academic computing resources. All models are trained at 1.5° resolution, with a training budget of \sim 9 V100 days for ArchesWeather and \sim 45 V100 days for ArchesWeatherGen. For inference, ArchesWeatherGen generates 15-day weather trajectories at a rate of 1 minute per ensemble member on a A100 GPU card. To make our work fully reproducible, our code and models are open source, including the complete pipeline for data preparation, training, and evaluation, at this https URL .

[LG-13] On Local Overfitting and Forgetting in Deep Neural Networks AAAI-25

链接: https://arxiv.org/abs/2412.12968
作者: Uri Stern,Tomer Yaacoby,Daphna Weinshall
关键词: increasing model size, infrequent occurrence, data space, overfitting, increasing model
类目: Machine Learning (cs.LG)
*备注: to appear in AAAI-25

点击查看摘要

Abstract:The infrequent occurrence of overfitting in deep neural networks is perplexing: contrary to theoretical expectations, increasing model size often enhances performance in practice. But what if overfitting does occur, though restricted to specific sub-regions of the data space? In this work, we propose a novel score that captures the forgetting rate of deep models on validation data. We posit that this score quantifies local overfitting: a decline in performance confined to certain regions of the data space. We then show empirically that local overfitting occurs regardless of the presence of traditional overfitting. Using the framework of deep over-parametrized linear models, we offer a certain theoretical characterization of forgotten knowledge, and show that it correlates with knowledge forgotten by real deep models. Finally, we devise a new ensemble method that aims to recover forgotten knowledge, relying solely on the training history of a single network. When combined with self-distillation, this method enhances the performance of any trained model without adding inference costs. Extensive empirical evaluations demonstrate the efficacy of our method across multiple datasets, contemporary neural network architectures, and training protocols.

[LG-14] Efficient Diffusion Transformer Policies with Mixture of Expert Denoisers for Multitask Learning

链接: https://arxiv.org/abs/2412.12953
作者: Moritz Reuss,Jyothish Pari,Pulkit Agrawal,Rudolf Lioutikov
关键词: Diffusion Policies, Imitation Learning, Transformer-based Diffusion Policies, offering several appealing, appealing properties
类目: Machine Learning (cs.LG); Robotics (cs.RO)
*备注:

点击查看摘要

Abstract:Diffusion Policies have become widely used in Imitation Learning, offering several appealing properties, such as generating multimodal and discontinuous behavior. As models are becoming larger to capture more complex capabilities, their computational demands increase, as shown by recent scaling laws. Therefore, continuing with the current architectures will present a computational roadblock. To address this gap, we propose Mixture-of-Denoising Experts (MoDE) as a novel policy for Imitation Learning. MoDE surpasses current state-of-the-art Transformer-based Diffusion Policies while enabling parameter-efficient scaling through sparse experts and noise-conditioned routing, reducing both active parameters by 40% and inference costs by 90% via expert caching. Our architecture combines this efficient scaling with noise-conditioned self-attention mechanism, enabling more effective denoising across different noise levels. MoDE achieves state-of-the-art performance on 134 tasks in four established imitation learning benchmarks (CALVIN and LIBERO). Notably, by pretraining MoDE on diverse robotics data, we achieve 4.01 on CALVIN ABC and 0.95 on LIBERO-90. It surpasses both CNN-based and Transformer Diffusion Policies by an average of 57% across 4 benchmarks, while using 90% fewer FLOPs and fewer active parameters compared to default Diffusion Transformer architectures. Furthermore, we conduct comprehensive ablations on MoDE’s components, providing insights for designing efficient and scalable Transformer architectures for Diffusion Policies. Code and demonstrations are available at this https URL.

[LG-15] FineGates: LLM s Finetuning with Compression using Stochastic Gates

链接: https://arxiv.org/abs/2412.12951
作者: Jonathan Svirsky,Yehonathan Refael,Ofir Lindenbaum
关键词: Large Language Models, Large Language, high computational demands, full finetuning due, present significant challenges
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Large Language Models (LLMs), with billions of parameters, present significant challenges for full finetuning due to the high computational demands, memory requirements, and impracticality of many real-world applications. When faced with limited computational resources or small datasets, updating all model parameters can often result in overfitting. To address this, lightweight finetuning techniques have been proposed, like learning low-rank adapter layers. These methods aim to train only a few additional parameters combined with the base model, which remains frozen, reducing resource usage and mitigating overfitting risks. In this work, we propose an adaptor model based on stochastic gates that simultaneously sparsify the frozen base model with task-specific adaptation. Our method comes with a small number of trainable parameters and allows us to speed up the base model inference with competitive accuracy. We evaluate it in additional variants by equipping it with additional low-rank parameters and comparing it to several recent baselines. Our results show that the proposed method improves the finetuned model accuracy comparatively to the several baselines and allows the removal of up to 20-40% without significant accuracy loss.

[LG-16] A simple DNN regression for the chemical composition in essential oil

链接: https://arxiv.org/abs/2412.12936
作者: Yuki Harada,Shuichi Maeda,Masato Kiyama,Shinichiro Nakamura
关键词: mono-molecular activity, extensively investigated, received little attention, prior studies, experimental design
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Although experimental design and methodological surveys for mono-molecular activity/property has been extensively investigated, those for chemical composition have received little attention, with the exception of a few prior studies. In this study, we configured three simple DNN regressors to predict essential oil property based on chemical composition. Despite showing overfitting due to the small size of dataset, all models were trained effectively in this study.

[LG-17] Multi-Subspace Matrix Recovery from Permuted Data AAAI2025

链接: https://arxiv.org/abs/2412.12931
作者: Liangqi Xie,Jicong Fan
关键词: columns are drawn, columns are corrupted, paper aims, union of low-dimensional, low-dimensional subspaces
类目: Machine Learning (cs.LG); Machine Learning (stat.ML)
*备注: The paper was accepted by AAAI 2025

点击查看摘要

Abstract:This paper aims to recover a multi-subspace matrix from permuted data: given a matrix, in which the columns are drawn from a union of low-dimensional subspaces and some columns are corrupted by permutations on their entries, recover the original matrix. The task has numerous practical applications such as data cleaning, integration, and de-anonymization, but it remains challenging and cannot be well addressed by existing techniques such as robust principal component analysis because of the presence of multiple subspaces and the permutations on the elements of vectors. To solve the challenge, we develop a novel four-stage algorithm pipeline including outlier identification, subspace reconstruction, outlier classification, and unsupervised sensing for permuted vector recovery. Particularly, we provide theoretical guarantees for the outlier classification step, ensuring reliable multi-subspace matrix recovery. Our pipeline is compared with state-of-the-art competitors on multiple benchmarks and shows superior performance.

[LG-18] Graph Spring Neural ODEs for Link Sign Prediction

链接: https://arxiv.org/abs/2412.12916
作者: Andrin Rehmann,Alexandre Bovet
关键词: Signed graphs, online activities, encoding positive, positive and negative, negative relations
类目: Machine Learning (cs.LG); Computers and Society (cs.CY)
*备注:

点击查看摘要

Abstract:Signed graphs allow for encoding positive and negative relations between nodes and are used to model various online activities. Node representation learning for signed graphs is a well-studied task with important applications such as sign prediction. While the size of datasets is ever-increasing, recent methods often sacrifice scalability for accuracy. We propose a novel message-passing layer architecture called Graph Spring Network (GSN) modeled after spring forces. We combine it with a Graph Neural Ordinary Differential Equations (ODEs) formalism to optimize the system dynamics in embedding space to solve a downstream prediction task. Once the dynamics is learned, embedding generation for novel datasets is done by solving the ODEs in time using a numerical integration scheme. Our GSN layer leverages the fast-to-compute edge vector directions and learnable scalar functions that only depend on nodes’ distances in latent space to compute the nodes’ positions. Conversely, Graph Convolution and Graph Attention Network layers rely on learnable vector functions that require the full positions of input nodes in latent space. We propose a specific implementation called Spring-Neural-Network (SPR-NN) using a set of small neural networks mimicking attracting and repulsing spring forces that we train for link sign prediction. Experiments show that our method achieves accuracy close to the state-of-the-art methods with node generation time speedup factors of up to 28,000 on large graphs.

[LG-19] PT: A Plain Transformer is Good Hospital Readmission Predictor

链接: https://arxiv.org/abs/2412.12909
作者: Zhenyi Fan,Jiaqi Li,Dongyu Luo,Yuqi Yuan
关键词: clinical decision support, Electronic Health Records, decision support, aiming to identify, risk of returning
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Hospital readmission prediction is critical for clinical decision support, aiming to identify patients at risk of returning within 30 days post-discharge. High readmission rates often indicate inadequate treatment or post-discharge care, making effective prediction models essential for optimizing resources and improving patient outcomes. We propose PT, a Transformer-based model that integrates Electronic Health Records (EHR), medical images, and clinical notes to predict 30-day all-cause hospital readmissions. PT extracts features from raw data and uses specialized Transformer blocks tailored to the data’s complexity. Enhanced with Random Forest for EHR feature selection and test-time ensemble techniques, PT achieves superior accuracy, scalability, and robustness. It performs well even when temporal information is missing. Our main contributions are: (1)Simplicity: A powerful and efficient baseline model outperforming existing ones in prediction accuracy; (2)Scalability: Flexible handling of various features from different modalities, achieving high performance with just clinical notes or EHR data; (3)Robustness: Strong predictive performance even with missing or unclear temporal data.

[LG-20] Design of Restricted Normalizing Flow towards Arbitrary Stochastic Policy with Computational Efficiency

链接: https://arxiv.org/abs/2412.12894
作者: Taisuke Kobayashi,Takumi Aotani
关键词: normalizing flow, policy, invertible transformation, paper proposes, design method
类目: Robotics (cs.RO); Machine Learning (cs.LG)
*备注: 27 pages, 13 figures

点击查看摘要

Abstract:This paper proposes a new design method for a stochastic control policy using a normalizing flow (NF). In reinforcement learning (RL), the policy is usually modeled as a distribution model with trainable parameters. When this parameterization has less expressiveness, it would fail to acquiring the optimal policy. A mixture model has capability of a universal approximation, but it with too much redundancy increases the computational cost, which can become a bottleneck when considering the use of real-time robot control. As another approach, NF, which is with additional parameters for invertible transformation from a simple stochastic model as a base, is expected to exert high expressiveness and lower computational cost. However, NF cannot compute its mean analytically due to complexity of the invertible transformation, and it lacks reliability because it retains stochastic behaviors after deployment for robot controller. This paper therefore designs a restricted NF (RNF) that achieves an analytic mean by appropriately restricting the invertible transformation. In addition, the expressiveness impaired by this restriction is regained using bimodal student-t distribution as its base, so-called Bit-RNF. In RL benchmarks, Bit-RNF policy outperformed the previous models. Finally, a real robot experiment demonstrated the applicability of Bit-RNF policy to real world. The attached video is uploaded on youtube: this https URL

[LG-21] meCHEAT: A Channel Harmony Strategy for Irregularly Sampled Multivariate Time Series Analysis AAAI2025

链接: https://arxiv.org/abs/2412.12886
作者: Jiexi Liu,Meng Cao,Songcan Chen
关键词: Irregularly sampled multivariate, multivariate time series, Irregularly sampled, sampled multivariate time, time series
类目: Machine Learning (cs.LG)
*备注: Accepted by AAAI 2025

点击查看摘要

Abstract:Irregularly sampled multivariate time series (ISMTS) are prevalent in reality. Due to their non-uniform intervals between successive observations and varying sampling rates among series, the channel-independent (CI) strategy, which has been demonstrated more desirable for complete multivariate time series forecasting in recent studies, has failed. This failure can be further attributed to the sampling sparsity, which provides insufficient information for effective CI learning, thereby reducing its capacity. When we resort to the channel-dependent (CD) strategy, even higher capacity cannot mitigate the potential loss of diversity in learning similar embedding patterns across different channels. We find that existing work considers CI and CD strategies to be mutually exclusive, primarily because they apply these strategies to the global channel. However, we hold the view that channel strategies do not necessarily have to be used globally. Instead, by appropriately applying them locally and globally, we can create an opportunity to take full advantage of both strategies. This leads us to introduce the Channel Harmony ISMTS Transformer (TimeCHEAT), which utilizes the CD locally and the CI globally. Specifically, we segment the ISMTS into sub-series level patches. Locally, the CD strategy aggregates information within each patch for time embedding learning, maximizing the use of relevant observations while reducing long-range irrelevant interference. Here, we enhance generality by transforming embedding learning into an edge weight prediction task using bipartite graphs, eliminating the need for special prior knowledge. Globally, the CI strategy is applied across patches, allowing the Transformer to learn individualized attention patterns for each channel. Experimental results indicate our proposed TimeCHEAT demonstrates competitive SOTA performance across three mainstream tasks.

[LG-22] owards Effective Graph Rationalization via Boosting Environment Diversity

链接: https://arxiv.org/abs/2412.12880
作者: Yujie Wang,Kui Yu,Yuhong Zhang,Fuyuan Cao,Jiye Liang
关键词: Graph Neural Networks, Neural Networks, environment subgraphs, original graph space, Graph Neural
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Graph Neural Networks (GNNs) perform effectively when training and testing graphs are drawn from the same distribution, but struggle to generalize well in the face of distribution shifts. To address this issue, existing mainstreaming graph rationalization methods first identify rationale and environment subgraphs from input graphs, and then diversify training distributions by augmenting the environment subgraphs. However, these methods merely combine the learned rationale subgraphs with environment subgraphs in the representation space to produce augmentation samples, failing to produce sufficiently diverse distributions. Thus, in this paper, we propose to achieve an effective Graph Rationalization by Boosting Environmental diversity, a GRBE approach that generates the augmented samples in the original graph space to improve the diversity of the environment subgraph. Firstly, to ensure the effectiveness of augmentation samples, we propose a precise rationale subgraph extraction strategy in GRBE to refine the rationale subgraph learning process in the original graph space. Secondly, to ensure the diversity of augmented samples, we propose an environment diversity augmentation strategy in GRBE that mixes the environment subgraphs of different graphs in the original graph space and then combines the new environment subgraphs with rationale subgraphs to generate augmented graphs. The average improvements of 7.65% and 6.11% in rationalization and classification performance on benchmark datasets demonstrate the superiority of GRBE over state-of-the-art approaches.

[LG-23] owards Physically Interpretable World Models: Meaningful Weakly Supervised Representations for Visual Trajectory Prediction

链接: https://arxiv.org/abs/2412.12870
作者: Zhenjiang Mao,Ivan Ruchkin
关键词: Deep learning models, Deep learning, employed for perception, increasingly employed, control in complex
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Deep learning models are increasingly employed for perception, prediction, and control in complex systems. Embedding physical knowledge into these models is crucial for achieving realistic and consistent outputs, a challenge often addressed by physics-informed machine learning. However, integrating physical knowledge with representation learning becomes difficult when dealing with high-dimensional observation data, such as images, particularly under conditions of incomplete or imprecise state information. To address this, we propose Physically Interpretable World Models, a novel architecture that aligns learned latent representations with real-world physical quantities. Our method combines a variational autoencoder with a dynamical model that incorporates unknown system parameters, enabling the discovery of physically meaningful representations. By employing weak supervision with interval-based constraints, our approach eliminates the reliance on ground-truth physical annotations. Experimental results demonstrate that our method improves the quality of learned representations while achieving accurate predictions of future states, advancing the field of representation learning in dynamic systems.

[LG-24] Geodesic Flow Kernels for Semi-Supervised Learning on Mixed-Variable Tabular Dataset AAAI-25

链接: https://arxiv.org/abs/2412.12864
作者: Yoontae Hwang,Yongjae Lee
关键词: poses unique challenges, unique challenges due, data poses unique, heterogeneous nature, Tabular data poses
类目: Machine Learning (cs.LG)
*备注: AAAI-25

点击查看摘要

Abstract:Tabular data poses unique challenges due to its heterogeneous nature, combining both continuous and categorical variables. Existing approaches often struggle to effectively capture the underlying structure and relationships within such data. We propose GFTab (Geodesic Flow Kernels for Semi- Supervised Learning on Mixed-Variable Tabular Dataset), a semi-supervised framework specifically designed for tabular datasets. GFTab incorporates three key innovations: 1) Variable-specific corruption methods tailored to the distinct properties of continuous and categorical variables, 2) A Geodesic flow kernel based similarity measure to capture geometric changes between corrupted inputs, and 3) Tree-based embedding to leverage hierarchical relationships from available labeled data. To rigorously evaluate GFTab, we curate a comprehensive set of 21 tabular datasets spanning various domains, sizes, and variable compositions. Our experimental results show that GFTab outperforms existing ML/DL models across many of these datasets, particularly in settings with limited labeled data.

[LG-25] Concurrent vertical and horizontal federated learning with fuzzy cognitive maps

链接: https://arxiv.org/abs/2412.12844
作者: Jose L Salmeron,Irina Arévalo
关键词: healthcare or finance, major concern, concern in industries, Data, privacy
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Data privacy is a major concern in industries such as healthcare or finance. The requirement to safeguard privacy is essential to prevent data breaches and misuse, which can have severe consequences for individuals and organisations. Federated learning is a distributed machine learning approach where multiple participants collaboratively train a model without compromising the privacy of their data. However, a significant challenge arises from the differences in feature spaces among participants, known as non-IID data. This research introduces a novel federated learning framework employing fuzzy cognitive maps, designed to comprehensively address the challenges posed by diverse data distributions and non-identically distributed features in federated settings. The proposal is tested through several experiments using four distinct federation strategies: constant-based, accuracy-based, AUC-based, and precision-based weights. The results demonstrate the effectiveness of the approach in achieving the desired learning outcomes while maintaining privacy and confidentiality standards.

[LG-26] Scrutinizing the Vulnerability of Decentralized Learning to Membership Inference Attacks

链接: https://arxiv.org/abs/2412.12837
作者: Ousmane Touat,Jezekael Brunon,Yacine Belal,Julien Nicolas,Mohamed Maouche,César Sabater,Sonia Ben Mokhtar
关键词: machine learning models, Membership Inference Attacks, decentralized learning, central entity, primary promise
类目: Machine Learning (cs.LG); Distributed, Parallel, and Cluster Computing (cs.DC)
*备注: 12 pages, 8 figures

点击查看摘要

Abstract:The primary promise of decentralized learning is to allow users to engage in the training of machine learning models in a collaborative manner while keeping their data on their premises and without relying on any central entity. However, this paradigm necessitates the exchange of model parameters or gradients between peers. Such exchanges can be exploited to infer sensitive information about training data, which is achieved through privacy attacks (e.g Membership Inference Attacks – MIA). In order to devise effective defense mechanisms, it is important to understand the factors that increase/reduce the vulnerability of a given decentralized learning architecture to MIA. In this study, we extensively explore the vulnerability to MIA of various decentralized learning architectures by varying the graph structure (e.g number of neighbors), the graph dynamics, and the aggregation strategy, across diverse datasets and data distributions. Our key finding, which to the best of our knowledge we are the first to report, is that the vulnerability to MIA is heavily correlated to (i) the local model mixing strategy performed by each node upon reception of models from neighboring nodes and (ii) the global mixing properties of the communication graph. We illustrate these results experimentally using four datasets and by theoretically analyzing the mixing properties of various decentralized architectures. Our paper draws a set of lessons learned for devising decentralized learning systems that reduce by design the vulnerability to MIA.

[LG-27] Noise-based Local Learning using Stochastic Magnetic Tunnel Junctions

链接: https://arxiv.org/abs/2412.12783
作者: Kees Koenders,Leo Schnitzpan,Fabian Kammerbauer,Sinan Shu,Gerhard Jakob,Mathis Kläui,Johan Mentink,Nasir Ahmad,Marcel van Gerven
关键词: minimal energy expenditure, Brain-inspired learning, energy expenditure, enormous potential, fast at minimal
类目: Emerging Technologies (cs.ET); Mesoscale and Nanoscale Physics (cond-mat.mes-hall); Machine Learning (cs.LG)
*备注: 20 pages, 5 figures, submitted to Physical Review X

点击查看摘要

Abstract:Brain-inspired learning in physical hardware has enormous potential to learn fast at minimal energy expenditure. One of the characteristics of biological learning systems is their ability to learn in the presence of various noise sources. Inspired by this observation, we introduce a novel noise-based learning approach for physical systems implementing multi-layer neural networks. Simulation results show that our approach allows for effective learning whose performance approaches that of the conventional effective yet energy-costly backpropagation algorithm. Using a spintronics hardware implementation, we demonstrate experimentally that learning can be achieved in a small network composed of physical stochastic magnetic tunnel junctions. These results provide a path towards efficient learning in general physical systems which embraces rather than mitigates the noise inherent in physical devices.

[LG-28] Versatile Ordering Network: An Attention-based Neural Network for Ordering Across Scales and Quality Metrics

链接: https://arxiv.org/abs/2412.12759
作者: Zehua Yu,Weihan Zhang,Sihan Pan,Jun Tao
关键词: Versatile Ordering Network, visualization applications, matrix reordering, extensively studied, axis and matrix
类目: Machine Learning (cs.LG)
*备注: has been accepted by TVCG on 11-Dec-2024

点击查看摘要

Abstract:Ordering has been extensively studied in many visualization applications, such as axis and matrix reordering, for the simple reason that the order will greatly impact the perceived pattern of data. Many quality metrics concerning data pattern, perception, and aesthetics are proposed, and respective optimization algorithms are developed. However, the optimization problems related to ordering are often difficult to solve (e.g., TSP is NP-complete), and developing specialized optimization algorithms is costly. In this paper, we propose Versatile Ordering Network (VON), which automatically learns the strategy to order given a quality metric. VON uses the quality metric to evaluate its solutions, and leverages reinforcement learning with a greedy rollout baseline to improve itself. This keeps the metric transparent and allows VON to optimize over different metrics. Additionally, VON uses the attention mechanism to collect information across scales and reposition the data points with respect to the current context. This allows VONs to deal with data points following different distributions. We examine the effectiveness of VON under different usage scenarios and metrics. The results demonstrate that VON can produce comparable results to specialized solvers. The code is available at this https URL.

[LG-29] Deep Learning for Resilient Adversarial Decision Fusion in Byzantine Networks

链接: https://arxiv.org/abs/2412.12739
作者: Kassem Kallas
关键词: encompasses diverse scenarios, including varying Byzantine, Markovian states, Byzantine node proportions, unbalanced priors
类目: Machine Learning (cs.LG); Cryptography and Security (cs.CR); Multiagent Systems (cs.MA); Social and Information Networks (cs.SI)
*备注:

点击查看摘要

Abstract:This paper introduces a deep learning-based framework for resilient decision fusion in adversarial multi-sensor networks, providing a unified mathematical setup that encompasses diverse scenarios, including varying Byzantine node proportions, synchronized and unsynchronized attacks, unbalanced priors, adaptive strategies, and Markovian states. Unlike traditional methods, which depend on explicit parameter tuning and are limited by scenario-specific assumptions, the proposed approach employs a deep neural network trained on a globally constructed dataset to generalize across all cases without requiring adaptation. Extensive simulations validate the method’s robustness, achieving superior accuracy, minimal error probability, and scalability compared to state-of-the-art techniques, while ensuring computational efficiency for real-time applications. This unified framework demonstrates the potential of deep learning to revolutionize decision fusion by addressing the challenges posed by Byzantine nodes in dynamic adversarial environments.

[LG-30] Uncertainty-Aware Hybrid Inference with On-Device Small and Remote Large Language Models

链接: https://arxiv.org/abs/2412.12687
作者: Seungeun Oh,Jinhyuk Kim,Jihong Park,Seung-Woo Ko,Tony Q. S. Quek,Seong-Lyun Kim
关键词: hybrid language model, small language model, large language model, language model, LLM
类目: Machine Learning (cs.LG); Distributed, Parallel, and Cluster Computing (cs.DC); Information Theory (cs.IT); Networking and Internet Architecture (cs.NI); Signal Processing (eess.SP)
*备注: 6 pages, 6 figures; This work has been submitted to the IEEE for possible publication

点击查看摘要

Abstract:This paper studies a hybrid language model (HLM) architecture that integrates a small language model (SLM) operating on a mobile device with a large language model (LLM) hosted at the base station (BS) of a wireless network. The HLM token generation process follows the speculative inference principle: the SLM’s vocabulary distribution is uploaded to the LLM, which either accepts or rejects it, with rejected tokens being resampled by the LLM. While this approach ensures alignment between the vocabulary distributions of the SLM and LLM, it suffers from low token throughput due to uplink transmission and the computation costs of running both language models. To address this, we propose a novel HLM structure coined Uncertainty-aware HLM (U-HLM), wherein the SLM locally measures its output uncertainty, and skips both uplink transmissions and LLM operations for tokens that are likely to be accepted. This opportunistic skipping is enabled by our empirical finding of a linear correlation between the SLM’s uncertainty and the LLM’s rejection probability. We analytically derive the uncertainty threshold and evaluate its expected risk of rejection. Simulations show that U-HLM reduces uplink transmissions and LLM computation by 45.93%, while achieving up to 97.54% of the LLM’s inference accuracy and 2.54 \times faster token throughput than HLM without skipping.

[LG-31] Building Gradient Bridges: Label Leakage from Restricted Gradient Sharing in Federated Learning

链接: https://arxiv.org/abs/2412.12640
作者: Rui Zhang,Ka-Ho Chow,Ping Li
关键词: enhanced computational capabilities, federated learning, growing concern, benefits of utilizing, diverse sources
类目: Machine Learning (cs.LG); Cryptography and Security (cs.CR)
*备注:

点击查看摘要

Abstract:The growing concern over data privacy, the benefits of utilizing data from diverse sources for model training, and the proliferation of networked devices with enhanced computational capabilities have all contributed to the rise of federated learning (FL). The clients in FL collaborate to train a global model by uploading gradients computed on their private datasets without collecting raw data. However, a new attack surface has emerged from gradient sharing, where adversaries can restore the label distribution of a victim’s private data by analyzing the obtained gradients. To mitigate this privacy leakage, existing lightweight defenses restrict the sharing of gradients, such as encrypting the final-layer gradients or locally updating the parameters within. In this paper, we introduce a novel attack called Gradient Bridge (GDBR) that recovers the label distribution of training data from the limited gradient information shared in FL. GDBR explores the relationship between the layer-wise gradients, tracks the flow of gradients, and analytically derives the batch training labels. Extensive experiments show that GDBR can accurately recover more than 80% of labels in various FL settings. GDBR highlights the inadequacy of restricted gradient sharing-based defenses and calls for the design of effective defense schemes in FL.

[LG-32] Neighbor-Aware Reinforcement Learning for Mixed Traffic Optimization in Large-scale Networks

链接: https://arxiv.org/abs/2412.12622
作者: Iftekharul Islam,Weizi Li
关键词: presents unique challenges, large-scale networks presents, networks presents unique, traffic comprising human-driven, robot vehicles
类目: Machine Learning (cs.LG); Systems and Control (eess.SY)
*备注:

点击查看摘要

Abstract:Managing mixed traffic comprising human-driven and robot vehicles (RVs) across large-scale networks presents unique challenges beyond single-intersection control. This paper proposes a reinforcement learning framework for coordinating mixed traffic across multiple interconnected intersections. Our key contribution is a neighbor-aware reward mechanism that enables RVs to maintain balanced distribution across the network while optimizing local intersection efficiency. We evaluate our approach using a real-world network, demonstrating its effectiveness in managing realistic traffic patterns. Results show that our method reduces average waiting times by 39.2% compared to the state-of-the-art single-intersection control policy and 79.8% compared to traditional traffic signals. The framework’s ability to coordinate traffic across multiple intersections while maintaining balanced RV distribution provides a foundation for deploying learning-based solutions in urban traffic systems.

[LG-33] Distribution-Free Uncertainty Quantification in Mechanical Ventilation Treatment: A Conformal Deep Q-Learning Framework

链接: https://arxiv.org/abs/2412.12597
作者: Niloufar Eghbali,Tuka Alhanai,Mohammad M. Ghassemi
关键词: critical life-support intervention, intensive care units, Mechanical Ventilation, critical life-support, life-support intervention
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Mechanical Ventilation (MV) is a critical life-support intervention in intensive care units (ICUs). However, optimal ventilator settings are challenging to determine because of the complexity of balancing patient-specific physiological needs with the risks of adverse outcomes that impact morbidity, mortality, and healthcare costs. This study introduces ConformalDQN, a novel distribution-free conformal deep Q-learning approach for optimizing mechanical ventilation in intensive care units. By integrating conformal prediction with deep reinforcement learning, our method provides reliable uncertainty quantification, addressing the challenges of Q-value overestimation and out-of-distribution actions in offline settings. We trained and evaluated our model using ICU patient records from the MIMIC-IV database. ConformalDQN extends the Double DQN architecture with a conformal predictor and employs a composite loss function that balances Q-learning with well-calibrated probability estimation. This enables uncertainty-aware action selection, allowing the model to avoid potentially harmful actions in unfamiliar states and handle distribution shifts by being more conservative in out-of-distribution scenarios. Evaluation against baseline models, including physician policies, policy constraint methods, and behavior cloning, demonstrates that ConformalDQN consistently makes recommendations within clinically safe and relevant ranges, outperforming other methods by increasing the 90-day survival rate. Notably, our approach provides an interpretable measure of confidence in its decisions, which is crucial for clinical adoption and potential human-in-the-loop implementations.

[LG-34] Stiefel Flow Matching for Moment-Constrained Structure Elucidation

链接: https://arxiv.org/abs/2412.12540
作者: Austin Cheng,Alston Lo,Kin Long Kelvin Lee,Santiago Miret,Alán Aspuru-Guzik
关键词: understanding chemical phenomena, Molecular structure elucidation, Stiefel Flow Matching, lab syntheses, chemical phenomena
类目: Machine Learning (cs.LG); Chemical Physics (physics.chem-ph)
*备注:

点击查看摘要

Abstract:Molecular structure elucidation is a fundamental step in understanding chemical phenomena, with applications in identifying molecules in natural products, lab syntheses, forensic samples, and the interstellar medium. We consider the task of predicting a molecule’s all-atom 3D structure given only its molecular formula and moments of inertia, motivated by the ability of rotational spectroscopy to measure these moments. While existing generative models can conditionally sample 3D structures with approximately correct moments, this soft conditioning fails to leverage the many digits of precision afforded by experimental rotational spectroscopy. To address this, we first show that the space of n -atom point clouds with a fixed set of moments of inertia is embedded in the Stiefel manifold \mathrmSt(n, 4) . We then propose Stiefel Flow Matching as a generative model for elucidating 3D structure under exact moment constraints. Additionally, we learn simpler and shorter flows by finding approximate solutions for equivariant optimal transport on the Stiefel manifold. Empirically, enforcing exact moment constraints allows Stiefel Flow Matching to achieve higher success rates and faster sampling than Euclidean diffusion models, even on high-dimensional manifolds corresponding to large molecules in the GEOM dataset.

[LG-35] Echo: Simulating Distributed Training At Scale

链接: https://arxiv.org/abs/2412.12487
作者: Yicheng Feng,Yuetao Chen,Kaiwen Chen,Jingzong Li,Tianyuan Wu,Peng Cheng,Chuan Wu,Wei Wang,Tsung-Yi Ho,Hong Xu
关键词: massive machine learning, Simulation offers unique, distributed training jobs, large-scale distributed training, extrapolation purposes
类目: Machine Learning (cs.LG); Distributed, Parallel, and Cluster Computing (cs.DC)
*备注:

点击查看摘要

Abstract:Simulation offers unique values for both enumeration and extrapolation purposes, and is becoming increasingly important for managing the massive machine learning (ML) clusters and large-scale distributed training jobs. In this paper, we build Echo to tackle three key challenges in large-scale training simulation: (1) tracing the runtime training workloads at each device in an ex-situ fashion so we can use a single device to obtain the actual execution graphs of 1K-GPU training, (2) accurately estimating the collective communication without high overheads of discrete-event based network simulation, and (3) accounting for the interference-induced computation slowdown from overlapping communication and computation kernels on the same device. Echo delivers on average 8% error in training step – roughly 3x lower than state-of-the-art simulators – for GPT-175B on a 96-GPU H800 cluster with 3D parallelism on Megatron-LM under 2 minutes.

[LG-36] AutoSGNN: Automatic Propagation Mechanism Discovery for Spectral Graph Neural Networks

链接: https://arxiv.org/abs/2412.12483
作者: Shibing Mo,Kai Wu,Qixuan Gao,Xiangyi Teng,Jing Liu
关键词: Graph Neural Networks, processing diverse types, Neural Networks, real-world applications, powerful tools
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:In real-world applications, spectral Graph Neural Networks (GNNs) are powerful tools for processing diverse types of graphs. However, a single GNN often struggles to handle different graph types-such as homogeneous and heterogeneous graphs-simultaneously. This challenge has led to the manual design of GNNs tailored to specific graph types, but these approaches are limited by the high cost of labor and the constraints of expert knowledge, which cannot keep up with the rapid growth of graph data. To overcome these challenges, we propose AutoSGNN, an automated framework for discovering propagation mechanisms in spectral GNNs. AutoSGNN unifies the search space for spectral GNNs by integrating large language models with evolutionary strategies to automatically generate architectures that adapt to various graph types. Extensive experiments on nine widely-used datasets, encompassing both homophilic and heterophilic graphs, demonstrate that AutoSGNN outperforms state-of-the-art spectral GNNs and graph neural architecture search methods in both performance and efficiency.

[LG-37] A Method for Enhancing Generalization of Adam by Multiple Integrations

链接: https://arxiv.org/abs/2412.12473
作者: Long Jin,Han Nong,Liangming Chen,Zhenming Su
关键词: adaptive moment estimation, multiple integral term, moment estimation, broader application, multiple integral
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:The insufficient generalization of adaptive moment estimation (Adam) has hindered its broader application. Recent studies have shown that flat minima in loss landscapes are highly associated with improved generalization. Inspired by the filtering effect of integration operations on high-frequency signals, we propose multiple integral Adam (MIAdam), a novel optimizer that integrates a multiple integral term into Adam. This multiple integral term effectively filters out sharp minima encountered during optimization, guiding the optimizer towards flatter regions and thereby enhancing generalization capability. We provide a theoretical explanation for the improvement in generalization through the diffusion theory framework and analyze the impact of the multiple integral term on the optimizer’s convergence. Experimental results demonstrate that MIAdam not only enhances generalization and robustness against label noise but also maintains the rapid convergence characteristic of Adam, outperforming Adam and its variants in state-of-the-art benchmarks.

[LG-38] Multi-Task Reinforcement Learning for Quadrotors

链接: https://arxiv.org/abs/2412.12442
作者: Jiaxu Xing,Ismail Geles,Yunlong Song,Elie Aljalbout,Davide Scaramuzza
关键词: shown great effectiveness, Reinforcement learning, enabling specialized policies, shown great, great effectiveness
类目: Robotics (cs.RO); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Reinforcement learning (RL) has shown great effectiveness in quadrotor control, enabling specialized policies to develop even human-champion-level performance in single-task scenarios. However, these specialized policies often struggle with novel tasks, requiring a complete retraining of the policy from scratch. To address this limitation, this paper presents a novel multi-task reinforcement learning (MTRL) framework tailored for quadrotor control, leveraging the shared physical dynamics of the platform to enhance sample efficiency and task performance. By employing a multi-critic architecture and shared task encoders, our framework facilitates knowledge transfer across tasks, enabling a single policy to execute diverse maneuvers, including high-speed stabilization, velocity tracking, and autonomous racing. Our experimental results, validated both in simulation and real-world scenarios, demonstrate that our framework outperforms baseline approaches in terms of sample efficiency and overall task performance.

[LG-39] GG-SSMs: Graph-Generating State Space Models

链接: https://arxiv.org/abs/2412.12423
作者: Nikola Zubić,Davide Scaramuzza
关键词: State Space Models, series analysis domains, Space Models, State Space, Graph-Generating State Space
类目: Machine Learning (cs.LG)
*备注: 11 pages, 7 tables, 2 figures

点击查看摘要

Abstract:State Space Models (SSMs) are powerful tools for modeling sequential data in computer vision and time series analysis domains. However, traditional SSMs are limited by fixed, one-dimensional sequential processing, which restricts their ability to model non-local interactions in high-dimensional data. While methods like Mamba and VMamba introduce selective and flexible scanning strategies, they rely on predetermined paths, which fails to efficiently capture complex dependencies. We introduce Graph-Generating State Space Models (GG-SSMs), a novel framework that overcomes these limitations by dynamically constructing graphs based on feature relationships. Using Chazelle’s Minimum Spanning Tree algorithm, GG-SSMs adapt to the inherent data structure, enabling robust feature propagation across dynamically generated graphs and efficiently modeling complex dependencies. We validate GG-SSMs on 11 diverse datasets, including event-based eye-tracking, ImageNet classification, optical flow estimation, and six time series datasets. GG-SSMs achieve state-of-the-art performance across all tasks, surpassing existing methods by significant margins. Specifically, GG-SSM attains a top-1 accuracy of 84.9% on ImageNet, outperforming prior SSMs by 1%, reducing the KITTI-15 error rate to 2.77%, and improving eye-tracking detection rates by up to 0.33% with fewer parameters. These results demonstrate that dynamic scanning based on feature relationships significantly improves SSMs’ representational power and efficiency, offering a versatile tool for various applications in computer vision and beyond.

[LG-40] Causally Consistent Normalizing Flow AAAI25

链接: https://arxiv.org/abs/2412.12401
作者: Qingyang Zhou,Kangjie Lu,Meng Xu
关键词: Normalizing Flows, Struct Causal Models, Struct Causal, underlying causal graphs, causal graphs captured
类目: Machine Learning (cs.LG)
*备注: extended version of “Causally Consistent Normalizing Flow” accepted by AAAI25

点击查看摘要

Abstract:Causal inconsistency arises when the underlying causal graphs captured by generative models like \textitNormalizing Flows (NFs) are inconsistent with those specified in causal models like \textitStruct Causal Models (SCMs). This inconsistency can cause unwanted issues including the unfairness problem. Prior works to achieve causal consistency inevitably compromise the expressiveness of their models by disallowing hidden layers. In this work, we introduce a new approach: \textbfCausally \textbfConsistent \textbfNormalizing \textbfFlow (CCNF). To the best of our knowledge, CCNF is the first causally consistent generative model that can approximate any distribution with multiple layers. CCNF relies on two novel constructs: a sequential representation of SCMs and partial causal transformations. These constructs allow CCNF to inherently maintain causal consistency without sacrificing expressiveness. CCNF can handle all forms of causal inference tasks, including interventions and counterfactuals. Through experiments, we show that CCNF outperforms current approaches in causal inference. We also empirically validate the practical utility of CCNF by applying it to real-world datasets and show how CCNF addresses challenges like unfairness effectively.

[LG-41] Development of an End-to-end Machine Learning System with Application to In-app Purchases

链接: https://arxiv.org/abs/2412.12390
作者: Dionysios Varelas,Elena Bonan,Lewis Anderson,Anders Englesson,Christoffer Åhrling,Adrian Chmielewski-Anders
关键词: mobile gaming industry, Machine learning, gaming industry, mobile gaming, Machine
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Machine learning (ML) systems have become vital in the mobile gaming industry. Companies like King have been using them in production to optimize various parts of the gaming experience. One important area is in-app purchases: purchases made in the game by players in order to enhance and customize their gameplay experience. In this work we describe how we developed an ML system in order to predict when a player is expected to make their next in-app purchase. These predictions are used to present offers to players. We briefly describe the problem definition, modeling approach and results and then, in considerable detail, outline the end-to-end ML system. We conclude with a reflection on challenges encountered and plans for future work.

[LG-42] Privacy in Metalearning and Multitask Learning: Modeling and Separations

链接: https://arxiv.org/abs/2412.12374
作者: Maryam Aliakbarpour,Konstantina Bairaktari,Adam Smith,Marika Swanberg,Jonathan Ullman
关键词: develop individually, learning, formal frameworks, Model personalization, personalized learning
类目: Machine Learning (cs.LG); Cryptography and Security (cs.CR)
*备注:

点击查看摘要

Abstract:Model personalization allows a set of individuals, each facing a different learning task, to train models that are more accurate for each person than those they could develop individually. The goals of personalization are captured in a variety of formal frameworks, such as multitask learning and metalearning. Combining data for model personalization poses risks for privacy because the output of an individual’s model can depend on the data of other individuals. In this work we undertake a systematic study of differentially private personalized learning. Our first main contribution is to construct a taxonomy of formal frameworks for private personalized learning. This taxonomy captures different formal frameworks for learning as well as different threat models for the attacker. Our second main contribution is to prove separations between the personalized learning problems corresponding to different choices. In particular, we prove a novel separation between private multitask learning and private metalearning.

[LG-43] Priority-Aware Model-Distributed Inference at Edge Networks

链接: https://arxiv.org/abs/2412.12371
作者: Teng Li,Hulya Seferoglu
关键词: entire Machine Learning, Distributed inference techniques, Machine Learning, broadly classified, model
类目: Distributed, Parallel, and Cluster Computing (cs.DC); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Distributed inference techniques can be broadly classified into data-distributed and model-distributed schemes. In data-distributed inference (DDI), each worker carries the entire Machine Learning (ML) model but processes only a subset of the data. However, feeding the data to workers results in high communication costs, especially when the data is large. An emerging paradigm is model-distributed inference (MDI), where each worker carries only a subset of ML layers. In MDI, a source device that has data processes a few layers of ML model and sends the output to a neighboring device, i.e., offloads the rest of the layers. This process ends when all layers are processed in a distributed manner. In this paper, we investigate the design and development of MDI when multiple data sources co-exist. We consider that each data source has a different importance and, hence, a priority. We formulate and solve a priority-aware model allocation optimization problem. Based on the structure of the optimal solution, we design a practical Priority-Aware Model- Distributed Inference (PA-MDI) algorithm that determines model allocation and distribution over devices by taking into account the priorities of different sources. Experiments were conducted on a real-life testbed of NVIDIA Jetson Xavier and Nano edge devices as well as in the Colosseum testbed with ResNet-50, ResNet- 56, and GPT-2 models. The experimental results show that PA-MDI performs priority-aware model allocation successfully while reducing the inference time as compared to baselines.

[LG-44] Krony-PT: GPT2 compressed with Kronecker Products

链接: https://arxiv.org/abs/2412.12351
作者: M. Ayoub Ben Ayad,Jelena Mitrovic,Michael Granitzer
关键词: Kronecker Products, compression technique, Kronecker Products based, introduce Krony-PT, modified Van Loan
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:We introduce Krony-PT, a compression technique of GPT2 \citepradford2019language based on Kronecker Products. We specifically target the MLP layers of each transformer layer, and systematically compress the feed forward layer matrices to various degrees. We introduce a modified Van Loan decomposition to initialize the new factors, and also introduce a new pruning-based initialization trick. Our method compresses the original 124M parameter GPT2 to various smaller models, with 80M being the smallest, and 96M being the largest compressed model. Our 81M model variant outperforms distilgpt2 on next-token prediction on all standard language modeling datasets, and shows competitive scores or performs on par with other Kronecker Products based compressed models of GPT2 that are significantly higher in size.

[LG-45] AutoSciLab: A Self-Driving Laboratory For Interpretable Scientific Discovery AAAI

链接: https://arxiv.org/abs/2412.12347
作者: Saaketh Desai,Sadhvikas Addamane,Jeffrey Y. Tsao,Igal Brener,Laura P. Swiler,Remi Dingreville,Prasad P. Iyer
关键词: automated scientific laboratories, Advances in robotic, scientific laboratories capable, robotic control, control and sensing
类目: Machine Learning (cs.LG); Materials Science (cond-mat.mtrl-sci); Optics (physics.optics)
*备注: Pre-print for paper accepted in AAAI

点击查看摘要

Abstract:Advances in robotic control and sensing have propelled the rise of automated scientific laboratories capable of high-throughput experiments. However, automated scientific laboratories are currently limited by human intuition in their ability to efficiently design and interpret experiments in high-dimensional spaces, throttling scientific discovery. We present AutoSciLab, a machine learning framework for driving autonomous scientific experiments, forming a surrogate researcher purposed for scientific discovery in high-dimensional spaces. AutoSciLab autonomously follows the scientific method in four steps: (i) generating high-dimensional experiments (x \in R^D) using a variational autoencoder (ii) selecting optimal experiments by forming hypotheses using active learning (iii) distilling the experimental results to discover relevant low-dimensional latent variables (z \in R^d, with d D) with a ‘directional autoencoder’ and (iv) learning a human interpretable equation connecting the discovered latent variables with a quantity of interest (y = f(z)), using a neural network equation learner. We validate the generalizability of AutoSciLab by rediscovering a) the principles of projectile motion and b) the phase transitions within the spin-states of the Ising model (NP-hard problem). Applying our framework to an open-ended nanophotonics challenge, AutoSciLab uncovers a fundamentally novel method for directing incoherent light emission that surpasses the current state-of-the-art (Iyer et al. 2023b, 2020).

[LG-46] Leveraging Group Classification with Descending Soft Labeling for Deep Imbalanced Regression

链接: https://arxiv.org/abs/2412.12327
作者: Ruizhi Pu,Gezheng Xu,Ruiyi Fang,Binkun Bao,Charles X. Ling,Boyu Wang
关键词: Toggle, highly skewed distribution, DIR, data, Deep imbalanced regression
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Deep imbalanced regression (DIR), where the target values have a highly skewed distribution and are also continuous, is an intriguing yet under-explored problem in machine learning. While recent works have already shown that incorporating various classification-based regularizers can produce enhanced outcomes, the role of classification remains elusive in DIR. Moreover, such regularizers (e.g., contrastive penalties) merely focus on learning discriminative features of data, which inevitably results in ignorance of either continuity or similarity across the data. To address these issues, we first bridge the connection between the objectives of DIR and classification from a Bayesian perspective. Consequently, this motivates us to decompose the objective of DIR into a combination of classification and regression tasks, which naturally guides us toward a divide-and-conquer manner to solve the DIR problem. Specifically, by aggregating the data at nearby labels into the same groups, we introduce an ordinal group-aware contrastive learning loss along with a multi-experts regressor to tackle the different groups of data thereby maintaining the data continuity. Meanwhile, considering the similarity between the groups, we also propose a symmetric descending soft labeling strategy to exploit the intrinsic similarity across the data, which allows classification to facilitate regression more effectively. Extensive experiments on real-world datasets also validate the effectiveness of our method. Subjects: Machine Learning (cs.LG) Cite as: arXiv:2412.12327 [cs.LG] (or arXiv:2412.12327v1 [cs.LG] for this version) https://doi.org/10.48550/arXiv.2412.12327 Focus to learn more arXiv-issued DOI via DataCite (pending registration) Submission history From: Ruizhi Pu [view email] [v1] Mon, 16 Dec 2024 19:54:57 UTC (3,111 KB) Full-text links: Access Paper: View a PDF of the paper titled Leveraging Group Classification with Descending Soft Labeling for Deep Imbalanced Regression, by Ruizhi Pu and 5 other authorsView PDFHTML (experimental)TeX SourceOther Formats view license Current browse context: cs.LG prev | next new | recent | 2024-12 Change to browse by: cs References Citations NASA ADSGoogle Scholar Semantic Scholar a export BibTeX citation Loading… BibTeX formatted citation loading… Data provided by: Bookmark checked=“checked”> Bibliographic Tools Bibliographic and Citation Tools Bibliographic Explorer Toggle Bibliographic Explorer (What is the Explorer?) Connected Papers Toggle Connected Papers (What is Connected Papers?) Litmaps Toggle Litmaps (What is Litmaps?) scite.ai Toggle scite Smart Citations (What are Smart Citations?) Code, Data, Media Code, Data and Media Associated with this Article alphaXiv Toggle alphaXiv (What is alphaXiv?) Links to Code Toggle CatalyzeX Code Finder for Papers (What is CatalyzeX?) DagsHub Toggle DagsHub (What is DagsHub?) GotitPub Toggle Gotit.pub (What is GotitPub?) Huggingface Toggle Hugging Face (What is Huggingface?) Links to Code Toggle Papers with Code (What is Papers with Code?) ScienceCast Toggle ScienceCast (What is ScienceCast?) Demos Demos Replicate Toggle Replicate (What is Replicate?) Spaces Toggle Hugging Face Spaces (What is Spaces?) Spaces Toggle TXYZ.AI (What is TXYZ.AI?) Related Papers Recommenders and Search Tools Link to Influence Flower Influence Flower (What are Influence Flowers?) Core recommender toggle CORE Recommender (What is CORE?) IArxiv recommender toggle IArxiv Recommender (What is IArxiv?) Author Venue Institution Topic About arXivLabs arXivLabs: experimental projects with community collaborators arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website. Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them. Have an idea for a project that will add value for arXiv’s community? Learn more about arXivLabs. Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?) mathjaxToggle(); About Help contact arXivClick here to contact arXiv Contact subscribe to arXiv mailingsClick here to subscribe Subscribe Copyright Privacy Policy Web Accessibility Assistance arXiv Operational Status Get status notifications via email or slack

[LG-47] F-RBA: A Federated Learning-based Framework for Risk-based Authentication

链接: https://arxiv.org/abs/2412.12324
作者: Hamidreza Fereidouni,Abdelhakim Senhaji Hafid,Dimitrios Makrakis,Yaser Baseri
关键词: proliferation of Internet, Internet services, protect private data, protect private, authentication
类目: Cryptography and Security (cs.CR); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:The proliferation of Internet services has led to an increasing need to protect private data. User authentication serves as a crucial mechanism to ensure data security. Although robust authentication forms the cornerstone of remote service security, it can still leave users vulnerable to credential disclosure, device-theft attacks, session hijacking, and inadequate adaptive security measures. Risk-based Authentication (RBA) emerges as a potential solution, offering a multi-level authentication approach that enhances user experience without compromising security. In this paper, we propose a Federated Risk-based Authentication (F-RBA) framework that leverages Federated Learning to ensure privacy-centric training, keeping user data local while distributing learning across devices. Whereas traditional approaches rely on centralized storage, F-RBA introduces a distributed architecture where risk assessment occurs locally on users’ devices. The framework’s core innovation lies in its similarity-based feature engineering approach, which addresses the heterogeneous data challenges inherent in federated settings, a significant advancement for distributed authentication. By facilitating real-time risk evaluation across devices while maintaining unified user profiles, F-RBA achieves a balance between data protection, security, and scalability. Through its federated approach, F-RBA addresses the cold-start challenge in risk model creation, enabling swift adaptation to new users without compromising security. Empirical evaluation using a real-world multi-user dataset demonstrates the framework’s effectiveness, achieving a superior true positive rate for detecting suspicious logins compared to conventional unsupervised anomaly detection models. This research introduces a new paradigm for privacy-focused RBA in distributed digital environments, facilitating advancements in federated security systems.

[LG-48] Deep Learning for Hydroelectric Optimization: Generating Long-Term River Discharge Scenarios with Ensemble Forecasts from Global Circulation Models

链接: https://arxiv.org/abs/2412.12234
作者: Julio Alberto Silva Dias
关键词: Hydroelectric power generation, Hydroelectric power, countries like Brazil, power generation, critical component
类目: Machine Learning (cs.LG); Signal Processing (eess.SP)
*备注: 11 pages, 15 figures

点击查看摘要

Abstract:Hydroelectric power generation is a critical component of the global energy matrix, particularly in countries like Brazil, where it represents the majority of the energy supply. However, its strong dependence on river discharges, which are inherently uncertain due to climate variability, poses significant challenges. River discharges are linked to precipitation patterns, making the development of accurate probabilistic forecasting models crucial for improving operational planning in systems heavily reliant on this resource. Traditionally, statistical models have been used to represent river discharges in energy optimization. Yet, these models are increasingly unable to produce realistic scenarios due to structural shifts in climate behavior. Changes in precipitation patterns have altered discharge dynamics, which traditional approaches struggle to capture. Machine learning methods, while effective as universal predictors for time series, often focus solely on historical data, ignoring key external factors such as meteorological and climatic conditions. Furthermore, these methods typically lack a probabilistic framework, which is vital for representing the inherent variability of hydrological processes. The limited availability of historical discharge data further complicates the application of large-scale deep learning models to this domain. To address these challenges, we propose a framework based on a modified recurrent neural network architecture. This model generates parameterized probability distributions conditioned on projections from global circulation models, effectively accounting for the stochastic nature of river discharges. Additionally, the architecture incorporates enhancements to improve its generalization capabilities. We validate this framework within the Brazilian Interconnected System, using projections from the SEAS5-ECMWF system as conditional variables.

[LG-49] Demonstrating Data-to-Knowledge Pipelines for Connecting Production Sites in the World Wide Lab

链接: https://arxiv.org/abs/2412.12231
作者: Leon Gorißen,Jan-Niklas Schneider,Mohamed Behery,Philipp Brauner,Moritz Lennartz,David Kötter,Thomas Kaster,Oliver Petrovic,Christian Hinke,Thomas Gries,Gerhard Lakemeyer,Martina Ziefle,Christian Brecher,Constantin Häfner
关键词: augmenting Digital Twins, concept augmenting Digital, integration and storage, digital transformation, requires new methods
类目: Robotics (cs.RO); Machine Learning (cs.LG)
*备注: 15 pages, 6 figures, submitted to CAiSE 2025

点击查看摘要

Abstract:The digital transformation of production requires new methods of data integration and storage, as well as decision making and support systems that work vertically and horizontally throughout the development, production, and use cycle. In this paper, we propose Data-to-Knowledge (and Knowledge-to-Data) pipelines for production as a universal concept building on a network of Digital Shadows (a concept augmenting Digital Twins). We show a proof of concept that builds on and bridges existing infrastructure to 1) capture and semantically annotates trajectory data from multiple similar but independent robots in different organisations and use cases in a data lakehouse and 2) an independent process that dynamically queries matching data for training an inverse dynamic foundation model for robotic control. The article discusses the challenges and benefits of this approach and how Data-to-Knowledge pipelines contribute efficiency gains and industrial scalability in a World Wide Lab as a research outlook.

[LG-50] he impact of AI on engineering design procedures for dynamical systems

链接: https://arxiv.org/abs/2412.12230
作者: Kristin M. de Payrebrune,Kathrin Flaßkamp,Tom Ströhla,Thomas Sattel,Dieter Bestle,Benedict Röder,Peter Eberhard,Sebastian Peitz,Marcus Stoffel,Gulakala Rutwik,Borse Aditya,Meike Wohlleben,Walter Sextro,Maximilian Raff,C. David Remy,Manish Yadav,Merten Stender,Jan van Delden,Timo Lüddecke,Sabine C.Langer,Julius Schultz,Christopher Blech
关键词: revolutionizing conventional processes, Artificial intelligence, numerous fields, revolutionizing conventional, opportunities for innovation
类目: ystems and Control (eess.SY); Machine Learning (cs.LG)
*备注: 23 pages, 9 figures, joint publication of the priority programme SPP 2353

点击查看摘要

Abstract:Artificial intelligence (AI) is driving transformative changes across numerous fields, revolutionizing conventional processes and creating new opportunities for innovation. The development of mechatronic systems is undergoing a similar transformation. Over the past decade, modeling, simulation, and optimization techniques have become integral to the design process, paving the way for the adoption of AI-based methods. In this paper, we examine the potential for integrating AI into the engineering design process, using the V-model from the VDI guideline 2206, considered the state-of-the-art in product design, as a foundation. We identify and classify AI methods based on their suitability for specific stages within the engineering product design workflow. Furthermore, we present a series of application examples where AI-assisted design has been successfully implemented by the authors. These examples, drawn from research projects within the DFG Priority Program \emphSPP~2353: Daring More Intelligence - Design Assistants in Mechanics and Dynamics, showcase a diverse range of applications across mechanics and mechatronics, including areas such as acoustics and robotics.

[LG-51] Accelerating Sparse Graph Neural Networks with Tensor Core Optimization

链接: https://arxiv.org/abs/2412.12218
作者: Ka Wai Wu
关键词: Graph neural networks, Tensor Cores, CUDA Cores, Cores, neural networks
类目: Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Graph neural networks (GNNs) have seen extensive application in domains such as social networks, bioinformatics, and recommendation systems. However, the irregularity and sparsity of graph data challenge traditional computing methods, which are insufficient to meet the performance demands of GNNs. Recent research has explored parallel acceleration using CUDA Cores and Tensor Cores, but significant challenges persist: (1) kernel fusion leads to false high utilization, failing to treat CUDA and Tensor Cores as independent resources, and (2) heterogeneous cores have distinct computation preferences, causing inefficiencies. To address these issues, this paper proposes FTC-GNN, a novel acceleration framework that efficiently utilizes CUDA and Tensor Cores for GNN computation. FTC-GNN introduces (1) a collaborative design that enables the parallel utilization of CUDA and Tensor Cores and (2) a sparse-to-dense transformation strategy that assigns dense matrix operations to Tensor Cores while leveraging CUDA Cores for data management and sparse edge processing. This design optimizes GPU resource utilization and improves computational efficiency. Experimental results demonstrate the effectiveness of FTC-GNN using GCN and AGNN models across various datasets. For GCN, FTC-GNN achieves speedups of 4.90x, 7.10x, and 1.17x compared to DGL, PyG, and TC-GNN, respectively. For AGNN, it achieves speedups of 5.32x, 2.92x, and 1.02x, establishing its superiority in accelerating GNN computations.

[LG-52] Comprehensive Survey on Adversarial Examples in Cybersecurity: Impacts Challenges and Mitigation Strategies

链接: https://arxiv.org/abs/2412.12217
作者: Li Li
关键词: encrypted traffic analysis, Deep learning, significantly transformed cybersecurity, botnet identification, user authentication
类目: Cryptography and Security (cs.CR); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Deep learning (DL) has significantly transformed cybersecurity, enabling advancements in malware detection, botnet identification, intrusion detection, user authentication, and encrypted traffic analysis. However, the rise of adversarial examples (AE) poses a critical challenge to the robustness and reliability of DL-based systems. These subtle, crafted perturbations can deceive models, leading to severe consequences like misclassification and system vulnerabilities. This paper provides a comprehensive review of the impact of AE attacks on key cybersecurity applications, highlighting both their theoretical and practical implications. We systematically examine the methods used to generate adversarial examples, their specific effects across various domains, and the inherent trade-offs attackers face between efficacy and resource efficiency. Additionally, we explore recent advancements in defense mechanisms, including gradient masking, adversarial training, and detection techniques, evaluating their potential to enhance model resilience. By summarizing cutting-edge research, this study aims to bridge the gap between adversarial research and practical security applications, offering insights to fortify the adoption of DL solutions in cybersecurity.

[LG-53] he AI Black-Scholes: Finance-Informed Neural Network

链接: https://arxiv.org/abs/2412.12213
作者: Amine M. Aboussalah,Xuanze Li,Cheng Chi,Raj Patel
关键词: partial differential equations, solving partial differential, pricing function satisfies, pricing function directly, function satisfies
类目: Machine Learning (cs.LG); Computational Finance (q-fin.CP); Machine Learning (stat.ML)
*备注:

点击查看摘要

Abstract:In the realm of option pricing, existing models are typically classified into principle-driven methods, such as solving partial differential equations (PDEs) that pricing function satisfies, and data-driven approaches, such as machine learning (ML) techniques that parameterize the pricing function directly. While principle-driven models offer a rigorous theoretical framework, they often rely on unrealistic assumptions, such as asset processes adhering to fixed stochastic differential equations (SDEs). Moreover, they can become computationally intensive, particularly in high-dimensional settings when analytical solutions are not available and thus numerical solutions are needed. In contrast, data-driven models excel in capturing market data trends, but they often lack alignment with core financial principles, raising concerns about interpretability and predictive accuracy, especially when dealing with limited or biased datasets. This work proposes a hybrid approach to address these limitations by integrating the strengths of both principled and data-driven methodologies. Our framework combines the theoretical rigor and interpretability of PDE-based models with the adaptability of machine learning techniques, yielding a more versatile methodology for pricing a broad spectrum of options. We validate our approach across different volatility modeling approaches-both with constant volatility (Black-Scholes) and stochastic volatility (Heston), demonstrating that our proposed framework, Finance-Informed Neural Network (FINN), not only enhances predictive accuracy but also maintains adherence to core financial principles. FINN presents a promising tool for practitioners, offering robust performance across a variety of market conditions.

[LG-54] A multi-theoretical kernel-based approach to social network-based recommendation

链接: https://arxiv.org/abs/2412.12202
作者: Xin Li,Mengyue Wang,T.-P. Liang
关键词: Recommender systems, component of e-commercewebsites, critical component, traditional recommender systems, social
类目: ocial and Information Networks (cs.SI); Information Retrieval (cs.IR); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Recommender systems are a critical component of e-commercewebsites. The rapid development of online social networking services provides an opportunity to explore social networks together with information used in traditional recommender systems, such as customer demographics, product characteristics, and transactions. It also provides more applications for recommender systems. To tackle this social network-based recommendation problem, previous studies generally built trust models in light of the social influence theory. This study inspects a spectrumof social network theories to systematicallymodel themultiple facets of a social network and infer user preferences. In order to effectively make use of these heterogonous theories, we take a kernel-based machine learning paradigm, design and select kernels describing individual similarities according to social network theories, and employ a non-linear multiple kernel learning algorithm to combine the kernels into a unified model. This design also enables us to consider multiple theories’ interactions in assessing individual behaviors. We evaluate our proposed approach on a real-world movie review data set. The experiments show that our approach provides more accurate recommendations than trust-based methods and the collaborative filtering approach. Further analysis shows that kernels derived from contagion theory and homophily theory contribute a larger portion of the model.

[LG-55] BlockDoor: Blocking Backdoor Based Watermarks in Deep Neural Networks

链接: https://arxiv.org/abs/2412.12194
作者: Yi Hao Puah,Anh Tu Ngo,Nandish Chattopadhyay,Anupam Chattopadhyay
关键词: prized Intellectual Property, Intellectual Property, prized Intellectual, turned Neural Networks, Trigger samples
类目: Cryptography and Security (cs.CR); Machine Learning (cs.LG)
*备注: Accepted as a Work-in-Progress paper at SPACE 2024: Fourteenth International Conference on Security, Privacy and Applied Cryptographic Engineering

点击查看摘要

Abstract:Adoption of machine learning models across industries have turned Neural Networks (DNNs) into a prized Intellectual Property (IP), which needs to be protected from being stolen or being used without authorization. This topic gave rise to multiple watermarking schemes, through which, one can establish the ownership of a model. Watermarking using backdooring is the most well established method available in the literature, with specific works demonstrating the difficulty in removing the watermarks, embedded as backdoors within the weights of the network. However, in our work, we have identified a critical flaw in the design of the watermark verification with backdoors, pertaining to the behaviour of the samples of the Trigger Set, which acts as the secret key. In this paper, we present BlockDoor, which is a comprehensive package of techniques that is used as a wrapper to block all three different kinds of Trigger samples, which are used in the literature as means to embed watermarks within the trained neural networks as backdoors. The framework implemented through BlockDoor is able to detect potential Trigger samples, through separate functions for adversarial noise based triggers, out-of-distribution triggers and random label based triggers. Apart from a simple Denial-of-Service for a potential Trigger sample, our approach is also able to modify the Trigger samples for correct machine learning functionality. Extensive evaluation of BlockDoor establishes that it is able to significantly reduce the watermark validation accuracy of the Trigger set by up to 98% without compromising on functionality, delivering up to a less than 1% drop on the clean samples. BlockDoor has been tested on multiple datasets and neural architectures.

[LG-56] Rashomon effect in Educational Research: Why More is Better Than One for Measuring the Importance of the Variables?

链接: https://arxiv.org/abs/2412.12115
作者: Jakub Kuzilek,Mustafa Çavuş
关键词: Rashomon effect influences, Educational Data Mining, academic outcomes prediction, so-called Rashomon effect, Rashomon effect
类目: Computers and Society (cs.CY); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:This study explores how the Rashomon effect influences variable importance in the context of student demographics used for academic outcomes prediction. Our research follows the way machine learning algorithms are employed in Educational Data Mining, focusing on highlighting the so-called Rashomon effect. The study uses the Rashomon set of simple-yet-accurate models trained using decision trees, random forests, light GBM, and XGBoost algorithms with the Open University Learning Analytics Dataset. We found that the Rashomon set improves the predictive accuracy by 2-6%. Variable importance analysis revealed more consistent and reliable results for binary classification than multiclass classification, highlighting the complexity of predicting multiple outcomes. Key demographic variables imd_band and highest_education were identified as vital, but their importance varied across courses, especially in course DDD. These findings underscore the importance of model choice and the need for caution in generalizing results, as different models can lead to different variable importance rankings. The codes for reproducing the experiments are available in the repository: this https URL.

[LG-57] Enhancing the conformal predictability of context-aware recommendation systems by using Deep Autoencoders

链接: https://arxiv.org/abs/2412.12110
作者: Saloua Zammali,Siddhant Dutta,Sadok Ben Yahia
关键词: Recommender Systems, collaborative filtering represents, field of Recommender, combining matrix factorization, neural collaborative filtering
类目: Information Retrieval (cs.IR); Machine Learning (cs.LG)
*备注: 8 pages, 4 tables, 1 figure. Accepted at the 23rd IEEE/WIC International Conference on Web Intelligence and Intelligent Agent Technology

点击查看摘要

Abstract:In the field of Recommender Systems (RS), neural collaborative filtering represents a significant milestone by combining matrix factorization and deep neural networks to achieve promising results. Traditional methods like matrix factorization often rely on linear models, limiting their capability to capture complex interactions between users, items, and contexts. This limitation becomes particularly evident with high-dimensional datasets due to their inability to capture relationships among users, items, and contextual factors. Unsupervised learning and dimension reduction tasks utilize autoencoders, neural network-based models renowned for their capacity to encode and decode data. Autoencoders learn latent representations of inputs, reducing dataset size while capturing complex patterns and features. In this paper, we introduce a framework that combines neural contextual matrix factorization with autoencoders to predict user ratings for items. We provide a comprehensive overview of the framework’s design and implementation. To evaluate its performance, we conduct experiments on various real-world datasets and compare the results against state-of-the-art approaches. We also extend the concept of conformal prediction to prediction rating and introduce a Conformal Prediction Rating (CPR). For RS, we define the nonconformity score, a key concept of conformal prediction, and demonstrate that it satisfies the exchangeability property.

[LG-58] On Model Extrapolation in Marginal Shapley Values

链接: https://arxiv.org/abs/2412.13158
作者: Ilya Rozenfeld
关键词: complex machine learning, machine learning models, learning models continues, reliable explainability methods, continues to grow
类目: Machine Learning (stat.ML); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:As the use of complex machine learning models continues to grow, so does the need for reliable explainability methods. One of the most popular methods for model explainability is based on Shapley values. There are two most commonly used approaches to calculating Shapley values which produce different results when features are correlated, conditional and marginal. In our previous work, it was demonstrated that the conditional approach is fundamentally flawed due to implicit assumptions of causality. However, it is a well-known fact that marginal approach to calculating Shapley values leads to model extrapolation where it might not be well defined. In this paper we explore the impacts of model extrapolation on Shapley values in the case of a simple linear spline model. Furthermore, we propose an approach which while using marginal averaging avoids model extrapolation and with addition of causal information replicates causal Shapley values. Finally, we demonstrate our method on the real data example.

[LG-59] Dual Interpretation of Machine Learning Forecasts

链接: https://arxiv.org/abs/2412.13076
作者: Philippe Goulet Coulombe,Maximilian Goebel,Karin Klieber
关键词: typically interpreted, sum of contributions, Machine learning predictions, past economic events, Machine learning
类目: Econometrics (econ.EM); Machine Learning (cs.LG); Machine Learning (stat.ML)
*备注:

点击查看摘要

Abstract:Machine learning predictions are typically interpreted as the sum of contributions of predictors. Yet, each out-of-sample prediction can also be expressed as a linear combination of in-sample values of the predicted variable, with weights corresponding to pairwise proximity scores between current and past economic events. While this dual route leads nowhere in some contexts (e.g., large cross-sectional datasets), it provides sparser interpretations in settings with many regressors and little training data-like macroeconomic forecasting. In this case, the sequence of contributions can be visualized as a time series, allowing analysts to explain predictions as quantifiable combinations of historical analogies. Moreover, the weights can be viewed as those of a data portfolio, inspiring new diagnostic measures such as forecast concentration, short position, and turnover. We show how weights can be retrieved seamlessly for (kernel) ridge regression, random forest, boosted trees, and neural networks. Then, we apply these tools to analyze post-pandemic forecasts of inflation, GDP growth, and recession probabilities. In all cases, the approach opens the black box from a new angle and demonstrates how machine learning models leverage history partly repeating itself.

[LG-60] BOIDS: High-dimensional Bayesian Optimization via Incumbent-guided Direction Lines and Subspace Embeddings AAAI

链接: https://arxiv.org/abs/2412.12918
作者: Lam Ngo,Huong Ha,Jeffrey Chan,Hongyu Zhang
关键词: expensive black-box optimization, Bayesian Optimization, powerful solution, expensive black-box, well-known and powerful
类目: Machine Learning (stat.ML); Machine Learning (cs.LG)
*备注: Published at AAAI Conference on Artificial Intelligence, 2025

点击查看摘要

Abstract:When it comes to expensive black-box optimization problems, Bayesian Optimization (BO) is a well-known and powerful solution. Many real-world applications involve a large number of dimensions, hence scaling BO to high dimension is of much interest. However, state-of-the-art high-dimensional BO methods still suffer from the curse of dimensionality, highlighting the need for further improvements. In this work, we introduce BOIDS, a novel high-dimensional BO algorithm that guides optimization by a sequence of one-dimensional direction lines using a novel tailored line-based optimization procedure. To improve the efficiency, we also propose an adaptive selection technique to identify most optimal lines for each round of line-based optimization. Additionally, we incorporate a subspace embedding technique for better scaling to high-dimensional spaces. We further provide theoretical analysis of our proposed method to analyze its convergence property. Our extensive experimental results show that BOIDS outperforms state-of-the-art baselines on various synthetic and real-world benchmark problems.

[LG-61] Sequential Harmful Shift Detection Without Labels NEURIPS2024

链接: https://arxiv.org/abs/2412.12910
作者: Salim I. Amoukou,Tom Bewley,Saumitra Mishra,Freddy Lecue,Daniele Magazzeni,Manuela Veloso
关键词: continuous production environments, ground truth data, machine learning models, truth data labels, production environments
类目: Machine Learning (stat.ML); Machine Learning (cs.LG)
*备注: Accepted at the 38th Conference on Neural Information Processing Systems (NeurIPS 2024)

点击查看摘要

Abstract:We introduce a novel approach for detecting distribution shifts that negatively impact the performance of machine learning models in continuous production environments, which requires no access to ground truth data labels. It builds upon the work of Podkopaev and Ramdas [2022], who address scenarios where labels are available for tracking model errors over time. Our solution extends this framework to work in the absence of labels, by employing a proxy for the true error. This proxy is derived using the predictions of a trained error estimator. Experiments show that our method has high power and false alarm control under various distribution shifts, including covariate and label shifts and natural shifts over geography and time.

[LG-62] Ask for More Than Bayes Optimal: A Theory of Indecisions for Classification

链接: https://arxiv.org/abs/2412.12807
作者: Mohamed Ndaoud,Peter Radchenko,Bradley Rava
关键词: Selective classification frameworks, highly risky scenarios, make highly confident, Selective classification, highly confident decisions
类目: atistics Theory (math.ST); Machine Learning (cs.LG); Methodology (stat.ME); Machine Learning (stat.ML)
*备注:

点击查看摘要

Abstract:Selective classification frameworks are useful tools for automated decision making in highly risky scenarios, since they allow for a classifier to only make highly confident decisions, while abstaining from making a decision when it is not confident enough to do so, which is otherwise known as an indecision. For a given level of classification accuracy, we aim to make as many decisions as possible. For many problems, this can be achieved without abstaining from making decisions. But when the problem is hard enough, we show that we can still control the misclassification rate of a classifier up to any user specified level, while only abstaining from the minimum necessary amount of decisions, even if this level of misclassification is smaller than the Bayes optimal error rate. In many problem settings, the user could obtain a dramatic decrease in misclassification while only paying a comparatively small price in terms of indecisions.

[LG-63] Enhanced Momentum with Momentum Transformers

链接: https://arxiv.org/abs/2412.12516
作者: Max Mason,Waasi A Jagirdar,David Huang,Rahul Murugan
关键词: benchmark time-series momentum, mean-reversion trading strategies, outperform benchmark time-series, Momentum Transformer, time-series momentum
类目: Computational Finance (q-fin.CP); Machine Learning (cs.LG)
*备注: 7 pages, 5 figures

点击查看摘要

Abstract:The primary objective of this research is to build a Momentum Transformer that is expected to outperform benchmark time-series momentum and mean-reversion trading strategies. We extend the ideas introduced in the paper Trading with the Momentum Transformer: An Intelligent and Interpretable Architecture to equities as the original paper primarily only builds upon futures and equity indices. Unlike conventional Long Short-Term Memory (LSTM) models, which operate sequentially and are optimized for processing local patterns, an attention mechanism equips our architecture with direct access to all prior time steps in the training window. This hybrid design, combining attention with an LSTM, enables the model to capture long-term dependencies, enhance performance in scenarios accounting for transaction costs, and seamlessly adapt to evolving market conditions, such as those witnessed during the Covid Pandemic. We average 4.14% returns which is similar to the original papers results. Our Sharpe is lower at an average of 1.12 due to much higher volatility which may be due to stocks being inherently more volatile than futures and indices.

[LG-64] Adversarially robust generalization theory via Jacobian regularization for deep neural networks

链接: https://arxiv.org/abs/2412.12449
作者: Dongya Wu,Xin Li
关键词: Powerful deep neural, Powerful deep, deep neural networks, Jacobian regularization, Jacobian
类目: Machine Learning (stat.ML); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Powerful deep neural networks are vulnerable to adversarial attacks. To obtain adversarially robust models, researchers have separately developed adversarial training and Jacobian regularization techniques. There are abundant theoretical and empirical studies for adversarial training, but theoretical foundations for Jacobian regularization are still lacking. In this study, we show that Jacobian regularization is closely related to adversarial training in that \ell_2 or \ell_1 Jacobian regularized loss serves as an approximate upper bound on the adversarially robust loss under \ell_2 or \ell_\infty adversarial attack respectively. Further, we establish the robust generalization gap for Jacobian regularized risk minimizer via bounding the Rademacher complexity of both the standard loss function class and Jacobian regularization function class. Our theoretical results indicate that the norms of Jacobian are related to both standard and robust generalization. We also perform experiments on MNIST data classification to demonstrate that Jacobian regularized risk minimization indeed serves as a surrogate for adversarially robust risk minimization, and that reducing the norms of Jacobian can improve both standard and robust generalization. This study promotes both theoretical and empirical understandings to adversarially robust generalization via Jacobian regularization.

[LG-65] Using machine learning to inform harvest control rule design in complex fishery settings

链接: https://arxiv.org/abs/2412.12400
作者: Felipe Montealegre-Mora,Carl Boettiger,Carl J. Walters,Christopher L. Cahill
关键词: size-structured stochastic populations, fishery science, management of size-structured, size-structured stochastic, long-standing and difficult
类目: Populations and Evolution (q-bio.PE); Machine Learning (cs.LG); Quantitative Methods (q-bio.QM)
*备注: 25 pages, 4 figures, 2 tables

点击查看摘要

Abstract:In fishery science, harvest management of size-structured stochastic populations is a long-standing and difficult problem. Rectilinear precautionary policies based on biomass and harvesting reference points have now become a standard approach to this problem. While these standard feedback policies are adapted from analytical or dynamic programming solutions assuming relatively simple ecological dynamics, they are often applied to more complicated ecological settings in the real world. In this paper we explore the problem of designing harvest control rules for partially observed, age-structured, spasmodic fish populations using tools from reinforcement learning (RL) and Bayesian optimization. Our focus is on the case of Walleye fisheries in Alberta, Canada, whose highly variable recruitment dynamics have perplexed managers and ecologists. We optimized and evaluated policies using several complementary performance metrics. The main questions we addressed were: 1. How do standard policies based on reference points perform relative to numerically optimized policies? 2. Can an observation of mean fish weight, in addition to stock biomass, aid policy decisions?

[LG-66] Optimizing Hyperparameters for Quantum Data Re-Uploaders in Calorimetric Particle Identification

链接: https://arxiv.org/abs/2412.12397
作者: Léa Cassé,Bernhard Pfahringer,Albert Bifet,Frédéric Magniette
关键词: single-qubit Data Re-Uploading, Data Re-Uploading, single-qubit Data, calorimetric experiments, Noisy Intermediate-Scale Quantum
类目: Quantum Physics (quant-ph); Machine Learning (cs.LG)
*备注: 17 pages, 22 figures

点击查看摘要

Abstract:We present an application of a single-qubit Data Re-Uploading (QRU) quantum model for particle classification in calorimetric experiments. Optimized for Noisy Intermediate-Scale Quantum (NISQ) devices, this model requires minimal qubits while delivering strong classification performance. Evaluated on a novel simulated dataset specific to particle physics, the QRU model achieves high accuracy in classifying particle types. Through a systematic exploration of model hyperparameters – such as circuit depth, rotation gates, input normalization and the number of trainable parameters per input – and training parameters like batch size, optimizer, loss function and learning rate, we assess their individual impacts on model accuracy and efficiency. Additionally, we apply global optimization methods, uncovering hyperparameter correlations that further enhance performance. Our results indicate that the QRU model attains significant accuracy with efficient computational costs, underscoring its potential for practical quantum machine learning applications.

[LG-67] Quantum Adversarial Machine Learning and Defense Strategies: Challenges and Opportunities

链接: https://arxiv.org/abs/2412.12373
作者: Eric Yocam,Anthony Rizi,Mahesh Kamepalli,Varghese Vaidyan,Yong Wang,Gurcan Comert
关键词: quantum-secure neural networks, quantum computing continues, neural networks, prevent adversarial attacks, continues to advance
类目: Quantum Physics (quant-ph); Cryptography and Security (cs.CR); Machine Learning (cs.LG)
*备注: 24 pages, 9 figures, 12 tables

点击查看摘要

Abstract:As quantum computing continues to advance, the development of quantum-secure neural networks is crucial to prevent adversarial attacks. This paper proposes three quantum-secure design principles: (1) using post-quantum cryptography, (2) employing quantum-resistant neural network architectures, and (3) ensuring transparent and accountable development and deployment. These principles are supported by various quantum strategies, including quantum data anonymization, quantum-resistant neural networks, and quantum encryption. The paper also identifies open issues in quantum security, privacy, and trust, and recommends exploring adaptive adversarial attacks and auto adversarial attacks as future directions. The proposed design principles and recommendations provide guidance for developing quantum-secure neural networks, ensuring the integrity and reliability of machine learning models in the quantum era.

[LG-68] ChronoFlow: A Data-Driven Model for Gyrochronology

链接: https://arxiv.org/abs/2412.12244
作者: Phil R. Van-Lane,Joshua S. Speagle,Gwendolyn M. Eadie,Stephanie T. Douglas,Phillip A. Cargile,Catherine Zucker,Yuxi(Lucy)Lu,Ruth Angus
关键词: sequence lifetime due, main sequence FGKM, main sequence lifetime, sequence FGKM stars, ages
类目: olar and Stellar Astrophysics (astro-ph.SR); Astrophysics of Galaxies (astro-ph.GA); Instrumentation and Methods for Astrophysics (astro-ph.IM); Machine Learning (cs.LG)
*备注: 40 pages, 28 figures (including appendices)

点击查看摘要

Abstract:Gyrochronology is a technique for constraining stellar ages using rotation periods, which change over a star’s main sequence lifetime due to magnetic braking. This technique shows promise for main sequence FGKM stars, where other methods are imprecise. However, models have historically struggled to capture the observed rotational dispersion in stellar populations. To properly understand this complexity, we have assembled the largest standardized data catalog of rotators in open clusters to date, consisting of ~7,400 stars across 30 open clusters/associations spanning ages of 1.5 Myr to 4 Gyr. We have also developed ChronoFlow: a flexible data-driven model which accurately captures observed rotational dispersion. We show that ChronoFlow can be used to accurately forward model rotational evolution, and to infer both cluster and individual stellar ages. We recover cluster ages with a statistical uncertainty of 0.06 dex ( \approx 15%), and individual stellar ages with a statistical uncertainty of 0.7 dex. Additionally, we conducted robust systematic tests to analyze the impact of extinction models, cluster membership, and calibration ages on our model’s performance. These contribute an additional \approx 0.06 dex of uncertainty in cluster age estimates, resulting in a total error budget of 0.08 dex. We estimate ages for the NGC 6709 open cluster and the Theia 456 stellar stream, and calculate revised rotational ages for M34, NGC 2516, NGC 1750, and NGC 1647. Our results show that ChronoFlow can precisely estimate the ages of coeval stellar populations, and constrain ages for individual stars. Furthermore, its predictions may be used to inform physical spin down models. ChronoFlow will be publicly available at this https URL.

[LG-69] Application of machine learning in grain-related clustering of Laue spots in a polycrystalline energy dispersive Laue pattern

链接: https://arxiv.org/abs/2412.12224
作者: Amir Tosson,Mohammad Shokr,Mahmoud Al Humaidi,Eduard Mikayelyan,Christian Gutt,Ulrich Pietsch
关键词: unsupervised machine learning, energy dispersive Laue, clustering problem solvable, grain-corresponding Laue reflections, dispersive Laue diffraction
类目: Materials Science (cond-mat.mtrl-sci); Machine Learning (cs.LG); Applied Physics (physics.app-ph); Data Analysis, Statistics and Probability (physics.data-an)
*备注:

点击查看摘要

Abstract:We address the identification of grain-corresponding Laue reflections in energy dispersive Laue diffraction (EDLD) experiments by formulating it as a clustering problem solvable through unsupervised machine learning (ML). To achieve reliable and efficient identification of grains in a Laue pattern, we employ a combination of clustering algorithms, namely hierarchical clustering (HC) and K-means. These algorithms allow us to group together similar Laue reflections, revealing the underlying grain structure in the diffraction pattern. Additionally, we utilise the elbow method to determine the optimal number of clusters, ensuring accurate results. To evaluate the performance of our proposed method, we conducted experiments using both simulated and experimental datasets obtained from nickel wires. The simulated datasets were generated to mimic the characteristics of real-world EDLD experiments, while the experimental datasets were obtained from actual measurements.

[LG-70] DLSOM: A Deep learning-based strategy for liver cancer subtyping

链接: https://arxiv.org/abs/2412.12214
作者: Fabio Zamio
关键词: cancer-related mortality worldwide, high genetic heterogeneity, genetic heterogeneity complicating, heterogeneity complicating diagnosis, Liver cancer
类目: Genomics (q-bio.GN); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Liver cancer is a leading cause of cancer-related mortality worldwide, with its high genetic heterogeneity complicating diagnosis and treatment. This study introduces DLSOM, a deep learning framework utilizing stacked autoencoders to analyze the complete somatic mutation landscape of 1,139 liver cancer samples, covering 20,356 protein-coding genes. By transforming high-dimensional mutation data into three low-dimensional features, DLSOM enables robust clustering and identifies five distinct liver cancer subtypes with unique mutational, functional, and biological profiles. Subtypes SC1 and SC2 exhibit higher mutational loads, while SC3 has the lowest, reflecting mutational heterogeneity. Novel and COSMIC-associated mutational signatures reveal subtype-specific molecular mechanisms, including links to hypermutation and chemotherapy resistance. Functional analyses further highlight the biological relevance of each subtype. This comprehensive framework advances precision medicine in liver cancer by enabling the development of subtype-specific diagnostics, biomarkers, and therapies, showcasing the potential of deep learning in addressing cancer complexity.

[LG-71] Deep Distributed Optimization for Large-Scale Quadratic Programming

链接: https://arxiv.org/abs/2412.12156
作者: Augustinos D. Saravanos,Hunter Kuperman,Alex Oshin,Arshiya Taj Abdul,Vincent Pacelli,Evangelos A. Theodorou
关键词: Quadratic programming, forms a crucial, encompassing a broad, crucial foundation, broad spectrum
类目: Optimization and Control (math.OC); Machine Learning (cs.LG); Multiagent Systems (cs.MA)
*备注:

点击查看摘要

Abstract:Quadratic programming (QP) forms a crucial foundation in optimization, encompassing a broad spectrum of domains and serving as the basis for more advanced algorithms. Consequently, as the scale and complexity of modern applications continue to grow, the development of efficient and reliable QP algorithms is becoming increasingly vital. In this context, this paper introduces a novel deep learning-aided distributed optimization architecture designed for tackling large-scale QP problems. First, we combine the state-of-the-art Operator Splitting QP (OSQP) method with a consensus approach to derive DistributedQP, a new method tailored for network-structured problems, with convergence guarantees to optimality. Subsequently, we unfold this optimizer into a deep learning framework, leading to DeepDistributedQP, which leverages learned policies to accelerate reaching to desired accuracy within a restricted amount of iterations. Our approach is also theoretically grounded through Probably Approximately Correct (PAC)-Bayes theory, providing generalization bounds on the expected optimality gap for unseen problems. The proposed framework, as well as its centralized version DeepQP, significantly outperform their standard optimization counterparts on a variety of tasks such as randomly generated problems, optimal control, linear regression, transportation networks and others. Notably, DeepDistributedQP demonstrates strong generalization by training on small problems and scaling to solve much larger ones (up to 50K variables and 150K constraints) using the same policy. Moreover, it achieves orders-of-magnitude improvements in wall-clock time compared to OSQP. The certifiable performance guarantees of our approach are also demonstrated, ensuring higher-quality solutions over traditional optimizers.

[LG-72] ECGtizer: a fully automated digitizing and signal recovery pipeline for electrocardiograms

链接: https://arxiv.org/abs/2412.12139
作者: Alex Lence,Ahmad Fall,Samuel David Cohen,Federica Granese,Jean-Daniel Zucker,Joe-Elie Salem,Edi Prifti
关键词: diagnosing cardiac pathologies, poses significant challenges, traditional paper-based ECG, storage poses significant, paper-based ECG storage
类目: ignal Processing (eess.SP); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:Electrocardiograms (ECGs) are essential for diagnosing cardiac pathologies, yet traditional paper-based ECG storage poses significant challenges for automated analysis. This study introduces ECGtizer, an open-source, fully automated tool designed to digitize paper ECGs and recover signals lost during storage. ECGtizer facilitates automated analyses using modern AI methods. It employs automated lead detection, three pixel-based signal extraction algorithms, and a deep learning-based signal reconstruction module. We evaluated ECGtizer on two datasets: a real-life cohort from the COVID-19 pandemic (JOCOVID) and a publicly available dataset (PTB-XL). Performance was compared with two existing methods: the fully automated ECGminer and the semi-automated PaperECG, which requires human intervention. ECGtizer’s performance was assessed in terms of signal recovery and the fidelity of clinically relevant feature measurement. Additionally, we tested these tools on a third dataset (GENEREPOL) for downstream AI tasks. Results show that ECGtizer outperforms existing tools, with its ECGtizerFrag algorithm delivering superior signal recovery. While PaperECG demonstrated better outcomes than ECGminer, it required human input. ECGtizer enhances the usability of historical ECG data and supports advanced AI-based diagnostic methods, making it a valuable addition to the field of AI in ECG analysis.

[LG-73] High-Throughput Detection of Risk Factors to Sudden Cardiac Arrest in Youth Athletes: A Smartwatch-Based Screening Platform

链接: https://arxiv.org/abs/2412.12118
作者: Evan Xiang,Thomas Wang,Vivan Poddar
关键词: Sudden Cardiac Arrest, age levels worldwide, International Olympic Committee, Sudden Cardiac, Cardiac Arrest
类目: ignal Processing (eess.SP); Machine Learning (cs.LG)
*备注: 22 pages

点击查看摘要

Abstract:Sudden Cardiac Arrest (SCA) is the leading cause of death among athletes of all age levels worldwide. Current prescreening methods for cardiac risk factors are largely ineffective, and implementing the International Olympic Committee recommendation for 12-lead ECG screening remains prohibitively expensive. To address these challenges, a preliminary comprehensive screening system (CSS) was developed to efficiently and economically screen large populations for risk factors to SCA. A protocol was established to measure a 4-lead ECG using an Apple Watch. Additionally, two key advances were introduced and validated: 1) A decomposition regression model to upscale 4-lead data to 12 leads, reducing ECG cost and usage complexity. 2) A deep learning model, the Transformer Auto-Encoder System (TAES), was designed to extract spatial and temporal features from the data for beat-based classification. TAES demonstrated an average sensitivity of 95.3% and specificity of 99.1% respectively in the testing dataset, outperforming human physicians in the same dataset (Se: 94%, Sp: 93%). Human subject trials (n = 30) validated the smartwatch protocol, with Bland-Altman analysis showing no statistical difference between the smartwatch vs. ECG protocol. Further validation of the complete CSS on a 20-subject cohort (10 affected, 10 controls) did not result in any misidentifications. This paper presents a mass screening system with the potential to achieve superior accuracy in high-throughput cardiac pre-participation evaluation compared to the clinical gold standard.

[LG-74] Generative Modeling of Neural Dynamics via Latent Stochastic Differential Equations

链接: https://arxiv.org/abs/2412.12112
作者: Ahmed ElGazzar,Marcel van Gerven
关键词: propose a probabilistic, biological neural systems, developing computational models, probabilistic framework, dynamical system
类目: Neurons and Cognition (q-bio.NC); Machine Learning (cs.LG)
*备注: 14 pages, 3 figures, 1 table

点击查看摘要

Abstract:We propose a probabilistic framework for developing computational models of biological neural systems. In this framework, physiological recordings are viewed as discrete-time partial observations of an underlying continuous-time stochastic dynamical system which implements computations through its state evolution. To model this dynamical system, we employ a system of coupled stochastic differential equations with differentiable drift and diffusion functions and use variational inference to infer its states and parameters. This formulation enables seamless integration of existing mathematical models in the literature, neural networks, or a hybrid of both to learn and compare different models. We demonstrate this in our framework by developing a generative model that combines coupled oscillators with neural networks to capture latent population dynamics from single-cell recordings. Evaluation across three neuroscience datasets spanning different species, brain regions, and behavioral tasks show that these hybrid models achieve competitive performance in predicting stimulus-evoked neural and behavioral responses compared to sophisticated black-box approaches while requiring an order of magnitude fewer parameters, providing uncertainty estimates, and offering a natural language for interpretation.

[LG-75] Reservoir Computing Generalized

链接: https://arxiv.org/abs/2412.12104
作者: Tomoyuki Kubota,Yusuke Imai,Sumito Tsunegi,Kohei Nakajima
关键词: solve machine learning, machine learning tasks, physical neural network, intrinsic physical properties, neural network
类目: Chaotic Dynamics (nlin.CD); Machine Learning (cs.LG)
*备注:

点击查看摘要

Abstract:A physical neural network (PNN) has both the strong potential to solve machine learning tasks and intrinsic physical properties, such as high-speed computation and energy efficiency. Reservoir computing (RC) is an excellent framework for implementing an information processing system with a dynamical system by attaching a trained readout, thus accelerating the wide use of unconventional materials for a PNN. However, RC requires the dynamics to reproducibly respond to input sequence, which limits the type of substance available for building information processors. Here we propose a novel framework called generalized reservoir computing (GRC) by turning this requirement on its head, making conventional RC a special case. Using substances that do not respond the same to identical inputs (e.g., a real spin-torque oscillator), we propose mechanisms aimed at obtaining a reliable output and show that processed inputs in the unconventional substance are retrievable. Finally, we demonstrate that, based on our framework, spatiotemporal chaos, which is thought to be unusable as a computational resource, can be used to emulate complex nonlinear dynamics, including large scale spatiotemporal chaos. Overall, our framework removes the limitation to building an information processing device and opens a path to constructing a computational system using a wider variety of physical dynamics.

信息检索

[IR-0] Re-calibrating methodologies in social media research: Challenge the visual work with Speech

链接: https://arxiv.org/abs/2412.13170
作者: Hongrui Jin
关键词: social media scholars, article methodologically reflects, methodologically reflects, scholars can effectively, effectively engage
类目: ocial and Information Networks (cs.SI); Information Retrieval (cs.IR)
*备注: 11 pages (excluding references), 3 figures

点击查看摘要

Abstract:This article methodologically reflects on how social media scholars can effectively engage with speech-based data in their analyses. While contemporary media studies have embraced textual, visual, and relational data, the aural dimension remained comparatively under-explored. Building on the notion of secondary orality and rejection towards purely visual culture, the paper argues that considering voice and speech at scale enriches our understanding of multimodal digital content. The paper presents the TikTok Subtitles Toolkit that offers accessible speech processing readily compatible with existing workflows. In doing so, it opens new avenues for large-scale inquiries that blend quantitative insights with qualitative precision. Two illustrative cases highlight both opportunities and limitations of speech research: while genres like #storytime on TikTok benefit from the exploration of spoken narratives, nonverbal or music-driven content may not yield significant insights using speech data. The article encourages researchers to integrate aural exploration thoughtfully to complement existing methods, rather than replacing them. I conclude that the expansion of our methodological repertoire enables richer interpretations of platformised content, and our capacity to unpack digital cultures as they become increasingly multimodal.

[IR-1] C-FedRAG: A Confidential Federated Retrieval-Augmented Generation System

链接: https://arxiv.org/abs/2412.13163
作者: Parker Addison,Minh-Tuan H. Nguyen,Tomislav Medan,Mohammad T. Manzari,Brendan McElrone,Laksh Lalwani,Aboli More,Smita Sharma,Holger R. Roth,Isaac Yang,Chester Chen,Daguang Xu,Yan Cheng,Andrew Feng,Ziyue Xu
关键词: Large Language Models, utilize Large Language, Large Language, Retrieval Augmented Generation, Language Models
类目: Distributed, Parallel, and Cluster Computing (cs.DC); Information Retrieval (cs.IR)
*备注:

点击查看摘要

Abstract:Organizations seeking to utilize Large Language Models (LLMs) for knowledge querying and analysis often encounter challenges in maintaining an LLM fine-tuned on targeted, up-to-date information that keeps answers relevant and grounded. Retrieval Augmented Generation (RAG) has quickly become a feasible solution for organizations looking to overcome the challenges of maintaining proprietary models and to help reduce LLM hallucinations in their query responses. However, RAG comes with its own issues regarding scaling data pipelines across tiered-access and disparate data sources. In many scenarios, it is necessary to query beyond a single data silo to provide richer and more relevant context for an LLM. Analyzing data sources within and across organizational trust boundaries is often limited by complex data-sharing policies that prohibit centralized data storage, therefore, inhibit the fast and effective setup and scaling of RAG solutions. In this paper, we introduce Confidential Computing (CC) techniques as a solution for secure Federated Retrieval Augmented Generation (FedRAG). Our proposed Confidential FedRAG system (C-FedRAG) enables secure connection and scaling of a RAG workflows across a decentralized network of data providers by ensuring context confidentiality. We also demonstrate how to implement a C-FedRAG system using the NVIDIA FLARE SDK and assess its performance using the MedRAG toolkit and MIRAGE benchmarking dataset.

[IR-2] RemoteRAG: A Privacy-Preserving LLM Cloud RAG Service

链接: https://arxiv.org/abs/2412.12775
作者: Yihang Cheng,Lan Zhang,Junyang Wang,Mu Yuan,Yunhao Yao
关键词: cloud RAG service, large language models, Retrieval-augmented generation, RAG service, user query
类目: Information Retrieval (cs.IR); Cryptography and Security (cs.CR)
*备注:

点击查看摘要

Abstract:Retrieval-augmented generation (RAG) improves the service quality of large language models by retrieving relevant documents from credible literature and integrating them into the context of the user query. Recently, the rise of the cloud RAG service has made it possible for users to query relevant documents conveniently. However, directly sending queries to the cloud brings potential privacy leakage. In this paper, we are the first to formally define the privacy-preserving cloud RAG service to protect the user query and propose RemoteRAG as a solution regarding privacy, efficiency, and accuracy. For privacy, we introduce (n,\epsilon) -DistanceDP to characterize privacy leakage of the user query and the leakage inferred from relevant documents. For efficiency, we limit the search range from the total documents to a small number of selected documents related to a perturbed embedding generated from (n,\epsilon) -DistanceDP, so that computation and communication costs required for privacy protection significantly decrease. For accuracy, we ensure that the small range includes target documents related to the user query with detailed theoretical analysis. Experimental results also demonstrate that RemoteRAG can resist existing embedding inversion attack methods while achieving no loss in retrieval under various settings. Moreover, RemoteRAG is efficient, incurring only 0.67 seconds and 46.66 KB of data transmission ( 2.72 hours and 1.43 GB with the non-optimized privacy-preserving scheme) when retrieving from a total of 10^6 documents.

[IR-3] A Survey on Sequential Recommendation

链接: https://arxiv.org/abs/2412.12770
作者: Liwei Pan,Weike Pan,Meiyan Wei,Hongzhi Yin,Zhong Ming
关键词: learning users’ preferences, received significant attention, sequential recommendation focuses, conventional recommendation problems, researchers and practitioners
类目: Information Retrieval (cs.IR)
*备注:

点击查看摘要

Abstract:Different from most conventional recommendation problems, sequential recommendation focuses on learning users’ preferences by exploiting the internal order and dependency among the interacted items, which has received significant attention from both researchers and practitioners. In recent years, we have witnessed great progress and achievements in this field, necessitating a new survey. In this survey, we study the SR problem from a new perspective (i.e., the construction of an item’s properties), and summarize the most recent techniques used in sequential recommendation such as pure ID-based SR, SR with side information, multi-modal SR, generative SR, LLM-powered SR, ultra-long SR and data-augmented SR. Moreover, we introduce some frontier research topics in sequential recommendation, e.g., open-domain SR, data-centric SR, could-edge collaborative SR, continuous SR, SR for good, and explainable SR. We believe that our survey could be served as a valuable roadmap for readers in this field.

[IR-4] oken-Level Graphs for Short Text Classification ECIR2025

链接: https://arxiv.org/abs/2412.12754
作者: Gregor Donabauer,Udo Kruschwitz
关键词: Information Retrieval, common subtask, Retrieval, Abstract, short texts
类目: Information Retrieval (cs.IR)
*备注: Preprint accepted at the 47th European Conference on Information Retrieval (ECIR 2025)

点击查看摘要

Abstract:The classification of short texts is a common subtask in Information Retrieval (IR). Recent advances in graph machine learning have led to interest in graph-based approaches for low resource scenarios, showing promise in such settings. However, existing methods face limitations such as not accounting for different meanings of the same words or constraints from transductive approaches. We propose an approach which constructs text graphs entirely based on tokens obtained through pre-trained language models (PLMs). By applying a PLM to tokenize and embed the texts when creating the graph(-nodes), our method captures contextual and semantic information, overcomes vocabulary constraints, and allows for context-dependent word meanings. Our approach also makes classification more efficient with reduced parameters compared to classical PLM fine-tuning, resulting in more robust training with few samples. Experimental results demonstrate how our method consistently achieves higher scores or on-par performance with existing methods, presenting an advancement in graph-based text classification techniques. To support reproducibility of our work we make all implementations publicly available to the community\footnote\urlthis https URL.

[IR-5] Boosting LLM -based Relevance Modeling with Distribution-Aware Robust Learning

链接: https://arxiv.org/abs/2412.12504
作者: Hong Liu,Saisai Gong,Yixin Ji,Kaixin Wu,Jia Xu,Jinjie Gu
关键词: pre-trained large language, relevance, large language models, relevance modeling, recent endeavors
类目: Information Retrieval (cs.IR)
*备注: 8 pages

点击查看摘要

Abstract:With the rapid advancement of pre-trained large language models (LLMs), recent endeavors have leveraged the capabilities of LLMs in relevance modeling, resulting in enhanced performance. This is usually done through the process of fine-tuning LLMs on specifically annotated datasets to determine the relevance between queries and items. However, there are two limitations when LLMs are naively employed for relevance modeling through fine-tuning and inference. First, it is not inherently efficient for performing nuanced tasks beyond simple yes or no answers, such as assessing search relevance. It may therefore tend to be overconfident and struggle to distinguish fine-grained degrees of relevance (e.g., strong relevance, weak relevance, irrelevance) used in search engines. Second, it exhibits significant performance degradation when confronted with data distribution shift in real-world scenarios. In this paper, we propose a novel Distribution-Aware Robust Learning framework (DaRL) for relevance modeling in Alipay Search. Specifically, we design an effective loss function to enhance the discriminability of LLM-based relevance modeling across various fine-grained degrees of query-item relevance. To improve the generalizability of LLM-based relevance modeling, we first propose the Distribution-Aware Sample Augmentation (DASA) module. This module utilizes out-of-distribution (OOD) detection techniques to actively select appropriate samples that are not well covered by the original training set for model fine-tuning. Furthermore, we adopt a multi-stage fine-tuning strategy to simultaneously improve in-distribution (ID) and OOD performance, bridging the performance gap between them. DaRL has been deployed online to serve the Alipay’s insurance product search…

[IR-6] LLM is Knowledge Graph Reasoner: LLM s Intuition-aware Knowledge Graph Reasoning for Cold-start Sequential Recommendation ECIR2025

链接: https://arxiv.org/abs/2412.12464
作者: Keigo Sakurai,Ren Togo,Takahiro Ogawa,Miki Haseyama
关键词: accurate content information, Large Language Models, recommendation, relationships between entities, widely studied
类目: Information Retrieval (cs.IR)
*备注: Accepted to the 47th European Conference on Information Retrieval (ECIR2025)

点击查看摘要

Abstract:Knowledge Graphs (KGs) represent relationships between entities in a graph structure and have been widely studied as promising tools for realizing recommendations that consider the accurate content information of items. However, traditional KG-based recommendation methods face fundamental challenges: insufficient consideration of temporal information and poor performance in cold-start scenarios. On the other hand, Large Language Models (LLMs) can be considered databases with a wealth of knowledge learned from the web data, and they have recently gained attention due to their potential application as recommendation systems. Although approaches that treat LLMs as recommendation systems can leverage LLMs’ high recommendation literacy, their input token limitations make it impractical to consider the entire recommendation domain dataset and result in scalability issues. To address these challenges, we propose a LLM’s Intuition-aware Knowledge graph Reasoning model (LIKR). Our main idea is to treat LLMs as reasoners that output intuitive exploration strategies for KGs. To integrate the knowledge of LLMs and KGs, we trained a recommendation agent through reinforcement learning using a reward function that integrates different recommendation strategies, including LLM’s intuition and KG embeddings. By incorporating temporal awareness through prompt engineering and generating textual representations of user preferences from limited interactions, LIKR can improve recommendation performance in cold-start scenarios. Furthermore, LIKR can avoid scalability issues by using KGs to represent recommendation domain datasets and limiting the LLM’s output to KG exploration strategies. Experiments on real-world datasets demonstrate that our model outperforms state-of-the-art recommendation methods in cold-start sequential recommendation scenarios.

[IR-7] Searching Personal Collections

链接: https://arxiv.org/abs/2412.12330
作者: Michael Bendersky,Donald Metzler,Marc Najork,Xuanhui Wang
关键词: personal document collections, Abstract, document collections, article describes, describes the history
类目: Information Retrieval (cs.IR)
*备注:

点击查看摘要

Abstract:This article describes the history of information retrieval on personal document collections.

附件下载

点击下载今日全部论文列表