← All Topics
⚙️

Fine-tuning & PEFT

Parameter-efficient fine-tuning, LoRA, instruction tuning, and supervised fine-tuning.

30 papers in the last 30 daysRSS feed
Heterogeneous Scientific Foundation Model Collaboration

Zihao Li, Jiaru Zou, Feihao Fang et al.

cs.AIcs.CLcs.LGApr 30, 2026

Agentic large language model systems have demonstrated strong capabilities. However, their reliance on language as the universal interface fundamentally limits their applicability to many real-world p

Post-Optimization Adaptive Rank Allocation for LoRA

Vishnuprasadh Kumaravelu, Sunil Gupta, P. K. Srijith

cs.AIApr 30, 2026

Exponential growth in the scale of modern foundation models has led to the widespread adoption of Low-Rank Adaptation (LoRA) as a parameter-efficient fine-tuning technique. However, standard LoRA impl

Exploration Hacking: Can LLMs Learn to Resist RL Training?

Eyon Jang, Damon Falck, Joschka Braun et al.

cs.LGcs.CLApr 30, 2026

Reinforcement learning (RL) has become essential to the post-training of large language models (LLMs) for reasoning, agentic capabilities and alignment. Successful RL relies on sufficient exploration

Why Self-Supervised Encoders Want to Be Normal

Yuval Domb

cs.ITcs.AIcs.LGApr 30, 2026

We develop a geometric and information-theoretic framework for encoder-decoder learning built on the Information Bottleneck (IB) principle. Recasting IB as a rate-distortion problem with Kullback-Leib

Sampling two-dimensional spin systems with transformers

Piotr Białas, Piotr Korcyl, Tomasz Stebel et al.

cond-mat.dis-nncond-mat.stat-mechcs.LGhep-latApr 30, 2026

Autoregressive Neural Networks based on dense or convolutional layers have recently been shown to be a viable strategy for generating classical spin systems. Unlike these methods, sampling with transf

Computing Equilibrium beyond Unilateral Deviation

Mingyang Liu, Gabriele Farina, Asuman Ozdaglar

cs.GTcs.AIcs.CCcs.LGApr 30, 2026

Most familiar equilibrium concepts, such as Nash and correlated equilibrium, guarantee only that no single player can improve their utility by deviating unilaterally. They offer no guarantees against

Track Fine-tuning & PEFT — Get notified when new papers are scored

Sign up free and get daily digests tailored to your research interests.

Sign up free