Skip to content
星际流动

CROP: Token-Efficient Reasoning in Large Language Models via Regularized Prompt Optimization

发布
采集
学术前沿 5.5 分 — Token-efficient reasoning via prompt optimization, addresses cost-efficiency in LLM inference
原文: cs.CL updates on arXiv.org

评分 5.5 · 来源:cs.CL updates on arXiv.org · 发布于 2026-04-17

评分依据:Token-efficient reasoning via prompt optimization, addresses cost-efficiency in LLM inference

arXiv:2604.14214v1 Announce Type: new Abstract: Large Language Models utilizing reasoning techniques improve task performance but incur significant latency and token costs due to verbose generation. Existing automatic prompt optimization(APO) frameworks target task accuracy exclusively at the expense of generating long reasoning traces. We propose Cost-Regularized Optimization of Prompts (CROP), an APO method that introduces regularization on response length by generating textual feedback in addition to standard accuracy feedback.