arxiv:2510.21285
Yingz
KigYzi
AI & ML interests
None yet
Recent Activity
authored
a paper
about 1 month ago
When Models Outthink Their Safety: Mitigating Self-Jailbreak in Large
Reasoning Models with Chain-of-Guardrails
upvoted
a
paper
about 1 month ago
When Models Outthink Their Safety: Mitigating Self-Jailbreak in Large
Reasoning Models with Chain-of-Guardrails
liked
a model
about 2 months ago
jiawei-ucas/Qwen-2.5-7B-ConsistentChat
Organizations
None yet