评分 5.5 · 来源:cs.CL updates on arXiv.org · 发布于 2026-04-17
评分依据:Token-efficient reasoning via prompt optimization, addresses cost-efficiency in LLM inference
arXiv:2604.14214v1 Announce Type: new Abstract: Large Language Models utilizing reasoning techniques improve task performance but incur significant latency and token costs due to verbose generation. Existing automatic prompt optimization(APO) frameworks target task accuracy exclusively at the expense of generating long reasoning traces. We propose Cost-Regularized Optimization of Prompts (CROP), an APO method that introduces regularization on response length by generating textual feedback in addition to standard accuracy feedback.