LLM Advances: Oct. 16, 2025 - Latest Research

by ADMIN 46 views

Dive into the forefront of Large Language Model (LLM) research with this curated collection of the latest papers, published on October 16, 2025. This digest covers critical areas like Knowledge Editing, Model Editing, GUI Agents, Steering Vectors, and Efficient LLMs. For a more immersive experience and access to the complete papers, including a better reading experience, please visit the Github page.

Knowledge Editing

Knowledge Editing in LLMs is rapidly evolving, with researchers continually refining how models learn and update information. The ability to modify a model's knowledge base without retraining the entire system is crucial for keeping LLMs current and accurate. The papers in this section explore diverse methods, from model editing and unlearning to more nuanced techniques that enhance factual recall and multi-hop question answering. This field is essential for addressing issues such as model alignment and the incorporation of real-time knowledge.

Paper Highlights:

  • KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning: This technical report dives deep into knowledge updating techniques, offering valuable insights into the mechanisms behind model editing.
  • STEAM: A Semantic-Level Knowledge Editing Framework for Large Language Models: STEAM presents a semantic-level framework, aiming to improve how LLMs process and apply knowledge.
  • AnyEdit: Edit Any Knowledge Encoded in Language Models: This paper introduces AnyEdit, which could revolutionize how we correct and add information to LLMs.
  • ACE: Attribution-Controlled Knowledge Editing for Multi-hop Factual Recall: This research focuses on improving the accuracy of LLMs in multi-hop question answering, with a focus on attribution.
  • From Injection to Defense: Constructing Edit-Based Fingerprints for Large Language Models: Explores the creation of edit-based fingerprints to protect against malicious injections into LLMs.
Title Date Comment
KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning 2025-10-14 Technical report
STEAM: A Semantic-Level Knowledge Editing Framework for Large Language Models 2025-10-12
Accep...

Accepted to EMNLP 2025 (Findings)

AnyEdit: Edit Any Knowledge Encoded in Language Models 2025-10-10
ACE: Attribution-Controlled Knowledge Editing for Multi-hop Factual Recall 2025-10-09
From Injection to Defense: Constructing Edit-Based Fingerprints for Large Language Models 2025-10-08 preprint
Aligning Language Models with Real-time Knowledge Editing 2025-10-07 Pre-print
When Disagreements Elicit Robustness: Investigating Self-Repair Capabilities under LLM Multi-Agent Disagreements 2025-10-02 Working in progress
Energy-Regularized Sequential Model Editing on Hyperspheres 2025-10-01
The c...

The code is available at https://github.com/PlusLabNLP/SPHERE. arXiv admin note: text overlap with arXiv:2410.02355 by other authors

Knowledge Editing with Subspace-Aware Key-Value Mappings 2025-09-29
25 pa...

25 pages, 12 figures, 10 tables

CaKE: Circuit-aware Editing Enables Generalizable Knowledge Learners 2025-09-23 EMNLP 2025
Consistency-Aware Parameter-Preserving Knowledge Editing Framework for Multi-Hop Question Answering 2025-09-23
Submi...

Submitted to ICASSP 2026

WikiBigEdit: Understanding the Limits of Lifelong Knowledge Editing in LLMs 2025-09-21
publi...

published at ICML 2025

Avoiding Knowledge Edit Skipping in Multi-hop Question Answering with Guided Decomposition 2025-09-09
Accep...

Accepted in EMNLP Findings 2025

MEMIT-Merge: Addressing MEMIT's Key-Value Conflicts in Same-Subject Batch Editing for LLMs 2025-09-09
Accep...

Accepted by ACL2025 findings

Knowledge Editing through Chain-of-Thought 2025-09-07

Model Editing

Model Editing is closely related to Knowledge Editing, but focuses more on how to directly alter the parameters of a pre-trained LLM. This includes modifying specific weights, adapting the model's behavior, and correcting factual inaccuracies. Research in model editing often explores the trade-offs between efficiency, performance, and the preservation of the model's overall capabilities. This area is crucial for making LLMs more reliable and adaptable to new information.

Paper Highlights:

  • KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning: Also appears in knowledge editing, highlighting the overlap between the fields.
  • ECMSim: A high-performance web simulation of cardiac ECM remodeling through integrated ODE-based signaling and diffusion: Although specialized, the research could have broader implications for how models are designed.
  • Diff-XYZ: A Benchmark for Evaluating Diff Understanding: Introduces a benchmark designed to assess the understanding of differences in models.
  • SPEED: Scalable, Precise, and Efficient Concept Erasure for Diffusion Models: Focuses on concept erasure in diffusion models, a technique relevant to model editing.
Title Date Comment
KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning 2025-10-14 Technical report
ECMSim: A high-performance web simulation of cardiac ECM remodeling through integrated ODE-based signaling and diffusion 2025-10-14
Diff-XYZ: A Benchmark for Evaluating Diff Understanding 2025-10-14
SPEED: Scalable, Precise, and Efficient Concept Erasure for Diffusion Models 2025-10-14
This ...

This version has been temporarily withdrawn for procedural review purposes. The withdrawal is unrelated to the technical content of the paper

EMSEdit: Efficient Multi-Step Meta-Learning-based Model Editing 2025-10-14
Exploring and Leveraging Class Vectors for Classifier Editing 2025-10-13
Accep...

Accepted in NeurIPS 2025

CoSPED: Consistent Soft Prompt Targeted Data Extraction and Defense 2025-10-13
Rethinking the Residual Distribution of Locate-then-Editing Methods in Model Editing 2025-10-13 NeurIPS 2025
AnyEdit: Edit Any Knowledge Encoded in Language Models 2025-10-10
Transmuting prompts into weights 2025-10-09
POME: Post Optimization Model Edit via Muon-style Projection 2025-10-08
Energy-Regularized Sequential Model Editing on Hyperspheres 2025-10-01
The c...

The code is available at https://github.com/PlusLabNLP/SPHERE. arXiv admin note: text overlap with arXiv:2410.02355 by other authors

Is Model Editing Built on Sand? Revealing Its Illusory Success and Fragile Foundation 2025-10-01
This ...

This is a work in progress. Comments and suggestions are welcome

Resolving UnderEdit & OverEdit with Iterative & Neighbor-Assisted Model Editing 2025-10-01
Accep...

Accepted at EMNLP 2025 as Findings

Fine-tuning Done Right in Model Editing 2025-09-29

GUI Agent

GUI Agents are LLMs designed to interact with graphical user interfaces (GUIs). These agents can automate tasks, navigate applications, and perform actions within a visual environment. Research in this area is focused on improving the agent's ability to understand visual inputs, plan actions, and execute tasks effectively. This is crucial for enabling more accessible and versatile AI systems.

Paper Highlights:

  • GUI-Shift: Enhancing VLM-Based GUI Agents through Self-supervised Reinforcement Learning: Focuses on the use of VLM-based GUI agents and self-supervised reinforcement learning for enhanced performance.
  • Auto-scaling Continuous Memory for GUI Agent: This research investigates auto-scaling continuous memory to increase the operational effectiveness of GUI agents.
  • ReInAgent: A Context-Aware GUI Agent Enabling Human-in-the-Loop Mobile Task Navigation: Provides a more advanced approach to GUI interaction by integrating human feedback for mobile task navigation.
  • PAL-UI: Planning with Active Look-back for Vision-Based GUI Agents: Enhances the planning capabilities of vision-based GUI agents.
Title Date Comment
GUI-Shift: Enhancing VLM-Based GUI Agents through Self-supervised Reinforcement Learning 2025-10-10
Auto-scaling Continuous Memory for GUI Agent 2025-10-10
ReInAgent: A Context-Aware GUI Agent Enabling Human-in-the-Loop Mobile Task Navigation 2025-10-09
PAL-UI: Planning with Active Look-back for Vision-Based GUI Agents 2025-10-04 Under Review
GTA1: GUI Test-time Scaling Agent 2025-10-03
GUI-PRA: Process Reward Agent for GUI Tasks 2025-10-03
GUI-KV: Efficient GUI Agents via KV Cache with Spatio-Temporal Awareness 2025-10-01
GUI-R1 : A Generalist R1-Style Vision-Language Action Model For GUI Agents 2025-10-01
Ferret-UI Lite: Lessons from Building Small On-Device GUI Agents 2025-09-30
Adaptive and Resource-efficient Agentic AI Systems for Mobile and Embedded Devices: A Survey 2025-09-30
MobileIPL: Enhancing Mobile Agents Thinking Process via Iterative Preference Learning 2025-09-29
9 pag...

9 pages, 8 figures, 7 tables

Retrieval-augmented GUI Agents with Generative Guidelines 2025-09-29
Accep...

Accepted to EMNLP 2025 (Main Conference)

Efficient Multi-turn RL for GUI Agents via Decoupled Training and Adaptive Data Curation 2025-09-28
GUI-Shepherd: Reliable Process Reward and Verification for Long-Sequence GUI Tasks 2025-09-28
GUI Agents: A Survey 2025-09-26
Accep...

Accepted to Findings of ACL 2025

Steering Vector

Steering Vectors are used to control LLMs by modifying the model's behavior. This technique allows researchers to fine-tune the outputs of LLMs without retraining the model. This is used for tasks like controlling the tone, style, and content of generated text. This section includes research on vector refinement and bias-only adaptation to steer LLM responses.

Paper Highlights:

  • FlexAC: Towards Flexible Control of Associative Reasoning in Multimodal Large Language Models: Focuses on enhancing the control of associative reasoning within multimodal LLMs.
  • Distribution-Aligned Decoding for Efficient LLM Task Adaptation: This research explores methods to optimize LLM performance through distribution-aligned decoding.
  • Transmuting prompts into weights: Shows the potential to transform prompts into weights, which changes how LLMs are controlled.
  • How Language Models Conflate Logical Validity with Plausibility: A Representational Analysis of Content Effects: Investigates how LLMs treat logical validity versus plausibility.
  • Prototype-Based Dynamic Steering for Large Language Models: Proposes the use of prototypes for dynamic steering.
Title Date Comment
FlexAC: Towards Flexible Control of Associative Reasoning in Multimodal Large Language Models 2025-10-14
19 pa...

19 pages, 11 figures. Accepted by the 39th Conference on Neural Information Processing Systems (NeurIPS 2025)

Distribution-Aligned Decoding for Efficient LLM Task Adaptation 2025-10-12
Accep...

Accepted by NeurIPS'25

Transmuting prompts into weights 2025-10-09
How Language Models Conflate Logical Validity with Plausibility: A Representational Analysis of Content Effects 2025-10-08
Prototype-Based Dynamic Steering for Large Language Models 2025-10-07
Activation Steering with a Feedback Controller 2025-10-05
9 pag...

9 pages in the main text. Under Review

Enhancing LLM Steering through Sparse Autoencoder-Based Vector Refinement 2025-10-03
19 pa...

19 pages, 11 figures, 7 tables

Steering LLM Reasoning Through Bias-Only Adaptation 2025-10-01 EMNLP 2025
Small Vectors, Big Effects: A Mechanistic Study of RL-Induced Reasoning via Steering Vectors 2025-10-01 Preprint
VISOR++: Universal Visual Inputs based Steering for Large Vision Language Models 2025-09-29
EasySteer: A Unified Framework for High-Performance and Extensible LLM Steering 2025-09-29
proje...

project: https://github.com/ZJU-REAL/EasySteer

Pixels Versus Priors: Controlling Knowledge Priors in Vision-Language Models through Visual Counterfacts 2025-09-29
Learning to Ponder: Adaptive Reasoning in Latent Space 2025-09-29
Toward Preference-aligned Large Language Models via Residual-based Model Steering 2025-09-28
Improving LLM Reasoning through Interpretable Role-Playing Steering 2025-09-28
Accep...

Accepted at EMNLP 2025 Findings

Efficient LLM

Efficient LLMs focuses on improving the speed, memory usage, and computational cost of LLMs. This involves methods such as model compression, quantization, and other techniques to reduce the resources needed for running and deploying these models. These advancements are crucial for making LLMs more accessible and practical across various applications.

Paper Highlights:

  • Efficient LLM Inference over Heterogeneous Edge Networks with Speculative Decoding: Explores speculative decoding to enhance inference speed in edge networks.
  • Distribution-Aligned Decoding for Efficient LLM Task Adaptation: Examines distribution-aligned decoding to optimize LLM performance.
  • LLMSynthor: Macro-Aligned Micro-Records Synthesis with Large Language Models: Focuses on synthesis using LLMs for macro-aligned micro-records.
  • Evaluating LLM-Based Process Explanations under Progressive Behavioral-Input Reduction: Process explanations are evaluated under reduced behavioral inputs.
Title Date Comment
Efficient LLM Inference over Heterogeneous Edge Networks with Speculative Decoding 2025-10-13
Distribution-Aligned Decoding for Efficient LLM Task Adaptation 2025-10-12
Accep...

Accepted by NeurIPS'25

LLMSynthor: Macro-Aligned Micro-Records Synthesis with Large Language Models 2025-10-11
Evaluating LLM-Based Process Explanations under Progressive Behavioral-Input Reduction 2025-10-10
12 pa...

12 pages, 2 figures, 3 tables; to appear in Enterprise Design, Operations, and Computing. EDOC 2025 Workshops, Lecture Notes in Business Information Processing (LNBIP), Springer, 2025. Part of 29th International Conference on Enterprise Design, Operations, and Computing (EDOC)

FLRC: Fine-grained Low-Rank Compressor for Efficient LLM Inference 2025-10-10
Accep...

Accepted by EMNLP 2025

Layout-Aware Parsing Meets Efficient LLMs: A Unified, Scalable Framework for Resume Information Extraction and Evaluation 2025-10-10
Spotlight Attention: Towards Efficient LLM Generation via Non-linear Hashing-based KV Cache Retrieval 2025-10-09
LLMs on a Budget? Say HOLA 2025-10-09
Accep...

Accepted at EMNLP 2025 (Industry Track)

PTEB: Towards Robust Text Embedding Evaluation via Stochastic Paraphrasing at Evaluation Time with LLMs 2025-10-08
VecInfer: Efficient LLM Inference with Low-Bit KV Cache via Outlier-Suppressed Vector Quantization 2025-10-07
MetaLLMix : An XAI Aided LLM-Meta-learning Based Approach for Hyper-parameters Optimization 2025-10-07
Activation-Informed Pareto-Guided Low-Rank Compression for Efficient LLM/VLM 2025-10-07
COSPADI: Compressing LLMs via Calibration-Guided Sparse Dictionary Learning 2025-10-06
Beyond Manuals and Tasks: Instance-Level Context Learning for LLM Agents 2025-10-06
SpikingMamba: Towards Energy-Efficient Large Language Models via Knowledge Distillation from Mamba 2025-10-06