|国家预印本平台
| 注册

国家预印本平台

中国首发,全球知晓

热点论文更多 >
MedXIAOHE: A Comprehensive Recipe for Building Medical MLLMs

We present MedXIAOHE, a medical vision-language foundation model designed to advance general-purpose medical understanding and reasoning in real-world clinical applications. MedXIAOHE achieves state-of-the-art performance across diverse medical benchmarks and surpasses leading closed-source multimodal systems on multiple capabilities. To achieve this, we propose an entity-aware continual pretraining framework that organizes heterogeneous medical corpora to broaden knowledge coverage and reduce long-tail gaps (e.g., rare diseases). For medical expert-level reasoning and interaction, MedXIAOHE incorporates diverse medical reasoning patterns via reinforcement learning and tool-augmented agentic training, enabling multi-step diagnostic reasoning with verifiable decision traces. To improve reliability in real-world use, MedXIAOHE integrates user-preference rubrics, evidence-grounded reasoning, and low-hallucination long-form report generation, with improved adherence to medical instructions. We release this report to document our practical design choices, scaling insights, and evaluation framework, hoping to inspire further research.

Baorong Shi;Bo Cui;Boyuan Jiang;Deli Yu;Fang Qian;Haihua Yang;Huichao Wang;Jiale Chen;Jianfei Pan;Jieqiong Cao;Jinghao Lin;Kai Wu;Lin Yang;Shengsheng Yao;Tao Chen;Xiaojun Xiao;Xiaozhong Ji;Xu Wang;Yijun He;Zhixiong Yang发表时间:2026-02-16
R-Diverse: Mitigating Diversity Illusion in Self-Play LLM Training

Self-play bootstraps LLM reasoning through an iterative Challenger-Solver loop: the Challenger is trained to generate questions that target the Solver's capabilities, and the Solver is optimized on the generated data to expand its reasoning skills. However, existing frameworks like R-Zero often exhibit non-sustained improvement, where early gains degrade as self-play continues. We identify a key failure mode, Diversity Illusion, where the Solver's training signals appear diverse yet collapse into recurring underlying patterns. It manifests as (1) Local Diversity Illusion, where diversity is enforced only within-batch, inducing cross-iteration mode cycling; and (2) Surface Diversity Illusion, where questions vary superficially but require near-identical reasoning skills. To mitigate them, we propose R-Diverse with two aligned innovations: Memory-Augmented Penalty (MAP), which uses a persistent memory bank to discourage recycling across iterations, and Skill-Aware Measurement (SAM), which evaluates diversity by the reasoning skills exercised rather than surface variation of questions. Across 10 math and general reasoning benchmarks, R-Diverse sustains gains over more iterations and consistently outperforms prior self-play methods. Code is available at https://github.com/Gengsheng-Li/R-Diverse.

Gengsheng Li;Jinghan He;Shijie Wang;Dan Zhang;Ruiqi Liu;Renrui Zhang;Zijun Yao;Junfeng Fang;Haiyun Guo;Jinqiao Wang发表时间:2026-02-16
Consistency of Large Reasoning Models Under Multi-Turn Attacks

Large reasoning models with reasoning capabilities achieve state-of-the-art performance on complex tasks, but their robustness under multi-turn adversarial pressure remains underexplored. We evaluate nine frontier reasoning models under adversarial attacks. Our findings reveal that reasoning confers meaningful but incomplete robustness: most reasoning models studied significantly outperform instruction-tuned baselines, yet all exhibit distinct vulnerability profiles, with misleading suggestions universally effective and social pressure showing model-specific efficacy. Through trajectory analysis, we identify five failure modes (Self-Doubt, Social Conformity, Suggestion Hijacking, Emotional Susceptibility, and Reasoning Fatigue) with the first two accounting for 50% of failures. We further demonstrate that Confidence-Aware Response Generation (CARG), effective for standard LLMs, fails for reasoning models due to overconfidence induced by extended reasoning traces; counterintuitively, random confidence embedding outperforms targeted extraction. Our results highlight that reasoning capabilities do not automatically confer adversarial robustness and that confidence-based defenses require fundamental redesign for reasoning models.

Yubo Li;Ramayya Krishnan;Rema Padman发表时间:2026-02-16
Prior-Guided Symbolic Regression: Towards Scientific Consistency in Equation Discovery

Symbolic Regression (SR) aims to discover interpretable equations from observational data, with the potential to reveal underlying principles behind natural phenomena. However, existing approaches often fall into the Pseudo-Equation Trap: producing equations that fit observations well but remain inconsistent with fundamental scientific principles. A key reason is that these approaches are dominated by empirical risk minimization, lacking explicit constraints to ensure scientific consistency. To bridge this gap, we propose PG-SR, a prior-guided SR framework built upon a three-stage pipeline consisting of warm-up, evolution, and refinement. Throughout the pipeline, PG-SR introduces a prior constraint checker that explicitly encodes domain priors as executable constraint programs, and employs a Prior Annealing Constrained Evaluation (PACE) mechanism during the evolution stage to progressively steer discovery toward scientifically consistent regions. Theoretically, we prove that PG-SR reduces the Rademacher complexity of the hypothesis space, yielding tighter generalization bounds and establishing a guarantee against pseudo-equations. Experimentally, PG-SR outperforms state-of-the-art baselines across diverse domains, maintaining robustness to varying prior quality, noisy data, and data scarcity.

Jing Xiao;Xinhai Chen;Jiaming Peng;Qinglin Wang;Menghan Jia;Zhiquan Lai;Guangping Yu;Dongsheng Li;Tiejun Li;Jie Liu发表时间:2026-02-16
A sparse overview on sparse resultants

In this survey, we give an overview of advances in the theory and computation of sparse resultants. First, we examine the construction and proof of the Canny-Emiris formula, which gives a rational determinantal formula. Second, we discuss and compare the latter with the computation of the sparse resultant as the determinant of the Koszul complex given by $n + 1$ nef divisors in a toric variety. Finally, we cover techniques for computing the Newton polytope of sparse resultants.

Carles Checa;Ioannis Z. Emiris;Christos Konaxis发表时间:2026-02-16
中国预印本平台发展联盟
合作期刊
国际预印本仓储