LLM Advances: Oct. 16, 2025 - Latest Research
Dive into the forefront of Large Language Model (LLM) research with this curated collection of the latest papers, published on October 16, 2025. This digest covers critical areas like Knowledge Editing, Model Editing, GUI Agents, Steering Vectors, and Efficient LLMs. For a more immersive experience and access to the complete papers, including a better reading experience, please visit the Github page.
Knowledge Editing
Knowledge Editing in LLMs is rapidly evolving, with researchers continually refining how models learn and update information. The ability to modify a model's knowledge base without retraining the entire system is crucial for keeping LLMs current and accurate. The papers in this section explore diverse methods, from model editing and unlearning to more nuanced techniques that enhance factual recall and multi-hop question answering. This field is essential for addressing issues such as model alignment and the incorporation of real-time knowledge.
Paper Highlights:
- KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning: This technical report dives deep into knowledge updating techniques, offering valuable insights into the mechanisms behind model editing.
- STEAM: A Semantic-Level Knowledge Editing Framework for Large Language Models: STEAM presents a semantic-level framework, aiming to improve how LLMs process and apply knowledge.
- AnyEdit: Edit Any Knowledge Encoded in Language Models: This paper introduces AnyEdit, which could revolutionize how we correct and add information to LLMs.
- ACE: Attribution-Controlled Knowledge Editing for Multi-hop Factual Recall: This research focuses on improving the accuracy of LLMs in multi-hop question answering, with a focus on attribution.
- From Injection to Defense: Constructing Edit-Based Fingerprints for Large Language Models: Explores the creation of edit-based fingerprints to protect against malicious injections into LLMs.
Title | Date | Comment |
---|---|---|
KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning | 2025-10-14 | Technical report |
STEAM: A Semantic-Level Knowledge Editing Framework for Large Language Models | 2025-10-12 | Accep...Accepted to EMNLP 2025 (Findings) |
AnyEdit: Edit Any Knowledge Encoded in Language Models | 2025-10-10 | |
ACE: Attribution-Controlled Knowledge Editing for Multi-hop Factual Recall | 2025-10-09 | |
From Injection to Defense: Constructing Edit-Based Fingerprints for Large Language Models | 2025-10-08 | preprint |
Aligning Language Models with Real-time Knowledge Editing | 2025-10-07 | Pre-print |
When Disagreements Elicit Robustness: Investigating Self-Repair Capabilities under LLM Multi-Agent Disagreements | 2025-10-02 | Working in progress |
Energy-Regularized Sequential Model Editing on Hyperspheres | 2025-10-01 | The c...The code is available at https://github.com/PlusLabNLP/SPHERE. arXiv admin note: text overlap with arXiv:2410.02355 by other authors |
Knowledge Editing with Subspace-Aware Key-Value Mappings | 2025-09-29 | 25 pa...25 pages, 12 figures, 10 tables |
CaKE: Circuit-aware Editing Enables Generalizable Knowledge Learners | 2025-09-23 | EMNLP 2025 |
Consistency-Aware Parameter-Preserving Knowledge Editing Framework for Multi-Hop Question Answering | 2025-09-23 | Submi...Submitted to ICASSP 2026 |
WikiBigEdit: Understanding the Limits of Lifelong Knowledge Editing in LLMs | 2025-09-21 | publi...published at ICML 2025 |
Avoiding Knowledge Edit Skipping in Multi-hop Question Answering with Guided Decomposition | 2025-09-09 | Accep...Accepted in EMNLP Findings 2025 |
MEMIT-Merge: Addressing MEMIT's Key-Value Conflicts in Same-Subject Batch Editing for LLMs | 2025-09-09 | Accep...Accepted by ACL2025 findings |
Knowledge Editing through Chain-of-Thought | 2025-09-07 |
Model Editing
Model Editing is closely related to Knowledge Editing, but focuses more on how to directly alter the parameters of a pre-trained LLM. This includes modifying specific weights, adapting the model's behavior, and correcting factual inaccuracies. Research in model editing often explores the trade-offs between efficiency, performance, and the preservation of the model's overall capabilities. This area is crucial for making LLMs more reliable and adaptable to new information.
Paper Highlights:
- KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning: Also appears in knowledge editing, highlighting the overlap between the fields.
- ECMSim: A high-performance web simulation of cardiac ECM remodeling through integrated ODE-based signaling and diffusion: Although specialized, the research could have broader implications for how models are designed.
- Diff-XYZ: A Benchmark for Evaluating Diff Understanding: Introduces a benchmark designed to assess the understanding of differences in models.
- SPEED: Scalable, Precise, and Efficient Concept Erasure for Diffusion Models: Focuses on concept erasure in diffusion models, a technique relevant to model editing.
Title | Date | Comment | |
---|---|---|---|
KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning | 2025-10-14 | Technical report | |
ECMSim: A high-performance web simulation of cardiac ECM remodeling through integrated ODE-based signaling and diffusion | 2025-10-14 | ||
Diff-XYZ: A Benchmark for Evaluating Diff Understanding | 2025-10-14 | ||
SPEED: Scalable, Precise, and Efficient Concept Erasure for Diffusion Models | 2025-10-14 | This ...This version has been temporarily withdrawn for procedural review purposes. The withdrawal is unrelated to the technical content of the paper |
|
EMSEdit: Efficient Multi-Step Meta-Learning-based Model Editing | 2025-10-14 | ||
Exploring and Leveraging Class Vectors for Classifier Editing | 2025-10-13 | Accep...Accepted in NeurIPS 2025 |
|
CoSPED: Consistent Soft Prompt Targeted Data Extraction and Defense | 2025-10-13 | ||
Rethinking the Residual Distribution of Locate-then-Editing Methods in Model Editing | 2025-10-13 | NeurIPS 2025 | |
AnyEdit: Edit Any Knowledge Encoded in Language Models | 2025-10-10 | ||
Transmuting prompts into weights | 2025-10-09 | ||
POME: Post Optimization Model Edit via Muon-style Projection | 2025-10-08 | ||
Energy-Regularized Sequential Model Editing on Hyperspheres | 2025-10-01 | The c...The code is available at https://github.com/PlusLabNLP/SPHERE. arXiv admin note: text overlap with arXiv:2410.02355 by other authors |
|
Is Model Editing Built on Sand? Revealing Its Illusory Success and Fragile Foundation | 2025-10-01 | This ...This is a work in progress. Comments and suggestions are welcome |
|
Resolving UnderEdit & OverEdit with Iterative & Neighbor-Assisted Model Editing | 2025-10-01 | Accep...Accepted at EMNLP 2025 as Findings |
|
Fine-tuning Done Right in Model Editing | 2025-09-29 |
GUI Agent
GUI Agents are LLMs designed to interact with graphical user interfaces (GUIs). These agents can automate tasks, navigate applications, and perform actions within a visual environment. Research in this area is focused on improving the agent's ability to understand visual inputs, plan actions, and execute tasks effectively. This is crucial for enabling more accessible and versatile AI systems.
Paper Highlights:
- GUI-Shift: Enhancing VLM-Based GUI Agents through Self-supervised Reinforcement Learning: Focuses on the use of VLM-based GUI agents and self-supervised reinforcement learning for enhanced performance.
- Auto-scaling Continuous Memory for GUI Agent: This research investigates auto-scaling continuous memory to increase the operational effectiveness of GUI agents.
- ReInAgent: A Context-Aware GUI Agent Enabling Human-in-the-Loop Mobile Task Navigation: Provides a more advanced approach to GUI interaction by integrating human feedback for mobile task navigation.
- PAL-UI: Planning with Active Look-back for Vision-Based GUI Agents: Enhances the planning capabilities of vision-based GUI agents.
Title | Date | Comment | |
---|---|---|---|
GUI-Shift: Enhancing VLM-Based GUI Agents through Self-supervised Reinforcement Learning | 2025-10-10 | ||
Auto-scaling Continuous Memory for GUI Agent | 2025-10-10 | ||
ReInAgent: A Context-Aware GUI Agent Enabling Human-in-the-Loop Mobile Task Navigation | 2025-10-09 | ||
PAL-UI: Planning with Active Look-back for Vision-Based GUI Agents | 2025-10-04 | Under Review | |
GTA1: GUI Test-time Scaling Agent | 2025-10-03 | ||
GUI-PRA: Process Reward Agent for GUI Tasks | 2025-10-03 | ||
GUI-KV: Efficient GUI Agents via KV Cache with Spatio-Temporal Awareness | 2025-10-01 | ||
GUI-R1 : A Generalist R1-Style Vision-Language Action Model For GUI Agents | 2025-10-01 | ||
Ferret-UI Lite: Lessons from Building Small On-Device GUI Agents | 2025-09-30 | ||
Adaptive and Resource-efficient Agentic AI Systems for Mobile and Embedded Devices: A Survey | 2025-09-30 | ||
MobileIPL: Enhancing Mobile Agents Thinking Process via Iterative Preference Learning | 2025-09-29 | 9 pag...9 pages, 8 figures, 7 tables |
|
Retrieval-augmented GUI Agents with Generative Guidelines | 2025-09-29 | Accep...Accepted to EMNLP 2025 (Main Conference) |
|
Efficient Multi-turn RL for GUI Agents via Decoupled Training and Adaptive Data Curation | 2025-09-28 | ||
GUI-Shepherd: Reliable Process Reward and Verification for Long-Sequence GUI Tasks | 2025-09-28 | ||
GUI Agents: A Survey | 2025-09-26 | Accep...Accepted to Findings of ACL 2025 |
Steering Vector
Steering Vectors are used to control LLMs by modifying the model's behavior. This technique allows researchers to fine-tune the outputs of LLMs without retraining the model. This is used for tasks like controlling the tone, style, and content of generated text. This section includes research on vector refinement and bias-only adaptation to steer LLM responses.
Paper Highlights:
- FlexAC: Towards Flexible Control of Associative Reasoning in Multimodal Large Language Models: Focuses on enhancing the control of associative reasoning within multimodal LLMs.
- Distribution-Aligned Decoding for Efficient LLM Task Adaptation: This research explores methods to optimize LLM performance through distribution-aligned decoding.
- Transmuting prompts into weights: Shows the potential to transform prompts into weights, which changes how LLMs are controlled.
- How Language Models Conflate Logical Validity with Plausibility: A Representational Analysis of Content Effects: Investigates how LLMs treat logical validity versus plausibility.
- Prototype-Based Dynamic Steering for Large Language Models: Proposes the use of prototypes for dynamic steering.
Title | Date | Comment | |
---|---|---|---|
FlexAC: Towards Flexible Control of Associative Reasoning in Multimodal Large Language Models | 2025-10-14 | 19 pa...19 pages, 11 figures. Accepted by the 39th Conference on Neural Information Processing Systems (NeurIPS 2025) |
|
Distribution-Aligned Decoding for Efficient LLM Task Adaptation | 2025-10-12 | Accep...Accepted by NeurIPS'25 |
|
Transmuting prompts into weights | 2025-10-09 | ||
How Language Models Conflate Logical Validity with Plausibility: A Representational Analysis of Content Effects | 2025-10-08 | ||
Prototype-Based Dynamic Steering for Large Language Models | 2025-10-07 | ||
Activation Steering with a Feedback Controller | 2025-10-05 | 9 pag...9 pages in the main text. Under Review |
|
Enhancing LLM Steering through Sparse Autoencoder-Based Vector Refinement | 2025-10-03 | 19 pa...19 pages, 11 figures, 7 tables |
|
Steering LLM Reasoning Through Bias-Only Adaptation | 2025-10-01 | EMNLP 2025 | |
Small Vectors, Big Effects: A Mechanistic Study of RL-Induced Reasoning via Steering Vectors | 2025-10-01 | Preprint | |
VISOR++: Universal Visual Inputs based Steering for Large Vision Language Models | 2025-09-29 | ||
EasySteer: A Unified Framework for High-Performance and Extensible LLM Steering | 2025-09-29 | proje... |
|
Pixels Versus Priors: Controlling Knowledge Priors in Vision-Language Models through Visual Counterfacts | 2025-09-29 | ||
Learning to Ponder: Adaptive Reasoning in Latent Space | 2025-09-29 | ||
Toward Preference-aligned Large Language Models via Residual-based Model Steering | 2025-09-28 | ||
Improving LLM Reasoning through Interpretable Role-Playing Steering | 2025-09-28 | Accep...Accepted at EMNLP 2025 Findings |
Efficient LLM
Efficient LLMs focuses on improving the speed, memory usage, and computational cost of LLMs. This involves methods such as model compression, quantization, and other techniques to reduce the resources needed for running and deploying these models. These advancements are crucial for making LLMs more accessible and practical across various applications.
Paper Highlights:
- Efficient LLM Inference over Heterogeneous Edge Networks with Speculative Decoding: Explores speculative decoding to enhance inference speed in edge networks.
- Distribution-Aligned Decoding for Efficient LLM Task Adaptation: Examines distribution-aligned decoding to optimize LLM performance.
- LLMSynthor: Macro-Aligned Micro-Records Synthesis with Large Language Models: Focuses on synthesis using LLMs for macro-aligned micro-records.
- Evaluating LLM-Based Process Explanations under Progressive Behavioral-Input Reduction: Process explanations are evaluated under reduced behavioral inputs.
Title | Date | Comment | |
---|---|---|---|
Efficient LLM Inference over Heterogeneous Edge Networks with Speculative Decoding | 2025-10-13 | ||
Distribution-Aligned Decoding for Efficient LLM Task Adaptation | 2025-10-12 | Accep...Accepted by NeurIPS'25 |
|
LLMSynthor: Macro-Aligned Micro-Records Synthesis with Large Language Models | 2025-10-11 | ||
Evaluating LLM-Based Process Explanations under Progressive Behavioral-Input Reduction | 2025-10-10 | 12 pa...12 pages, 2 figures, 3 tables; to appear in Enterprise Design, Operations, and Computing. EDOC 2025 Workshops, Lecture Notes in Business Information Processing (LNBIP), Springer, 2025. Part of 29th International Conference on Enterprise Design, Operations, and Computing (EDOC) |
|
FLRC: Fine-grained Low-Rank Compressor for Efficient LLM Inference | 2025-10-10 | Accep...Accepted by EMNLP 2025 |
|
Layout-Aware Parsing Meets Efficient LLMs: A Unified, Scalable Framework for Resume Information Extraction and Evaluation | 2025-10-10 | ||
Spotlight Attention: Towards Efficient LLM Generation via Non-linear Hashing-based KV Cache Retrieval | 2025-10-09 | ||
LLMs on a Budget? Say HOLA | 2025-10-09 | Accep...Accepted at EMNLP 2025 (Industry Track) |
|
PTEB: Towards Robust Text Embedding Evaluation via Stochastic Paraphrasing at Evaluation Time with LLMs | 2025-10-08 | ||
VecInfer: Efficient LLM Inference with Low-Bit KV Cache via Outlier-Suppressed Vector Quantization | 2025-10-07 | ||
MetaLLMix : An XAI Aided LLM-Meta-learning Based Approach for Hyper-parameters Optimization | 2025-10-07 | ||
Activation-Informed Pareto-Guided Low-Rank Compression for Efficient LLM/VLM | 2025-10-07 | ||
COSPADI: Compressing LLMs via Calibration-Guided Sparse Dictionary Learning | 2025-10-06 | ||
Beyond Manuals and Tasks: Instance-Level Context Learning for LLM Agents | 2025-10-06 | ||
SpikingMamba: Towards Energy-Efficient Large Language Models via Knowledge Distillation from Mamba | 2025-10-06 |