Showing 1–20 of 37 results
/ Date/ Name
Jun 18, 2024Adaptive Token Biaser: Knowledge Editing via Biasing Key EntitiesSep 16, 2024StruEdit: Structured Outputs Enable the Fast and Accurate Knowledge Editing for Large Language ModelsMar 30, 2024Is Factuality Enhancement a Free Lunch For LLMs? Better Factuality Can Lead to Worse Context-FaithfulnessJul 2, 2025Representation Entanglement for Generation: Training Diffusion Transformers Is Much Easier Than You ThinkApr 10, 2025Kimi-VL Technical ReportMar 29, 2025Efficient Inference for Large Reasoning Models: A SurveyJun 7, 2023Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs EvaluationsJul 15, 2024Spider2-V: How Far Are Multimodal Agents From Automating Data Science and Engineering Workflows?May 26, 2023Efficient Detection of LLM-generated Texts with a Bayesian Surrogate ModelOct 16, 2024Meta-Unlearning on Diffusion Models: Preventing Relearning Unlearned ConceptsNov 12, 2024Spider 2.0: Evaluating Language Models on Real-World Enterprise Text-to-SQL WorkflowsJan 31, 2026How Far Are LLMs from Professional Poker Players? Revisiting Game-Theoretic Reasoning with Agentic Tool UseAug 2, 2025Personalized Safety Alignment for Text-to-Image Diffusion ModelsJan 30, 2025GuardReasoner: Towards Reasoning-based LLM SafeguardsMay 29, 2023From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation FrameworkOct 15, 2021Textual Backdoor Attacks Can Be More Harmful via Two Simple TricksJun 16, 2025ExtendAttack: Attacking Servers of LRMs via Extending ReasoningFeb 17, 2025LanP: Rethinking the Impact of Language Priors in Large Vision-Language ModelsFeb 16, 2024Universal Prompt Optimizer for Safe Text-to-Image GenerationMay 16, 2025GuardReasoner-VL: Safeguarding VLMs via Reinforced Reasoning