Search

Jiaqi Huang's Homepage
Jiaqi Huang's Homepage
  • Home
  • Publications
  • Contact
  • Light Dark Automatic

Featured Publications

Tianyu Zhang, Zixuan Zhao, Jiaqi Huang, Jingyu Hua, Sheng Zhong
May, 2024 In DLSP 2024
Subtoxic Questions: Dive Into Attitude Change of LLM’s Response in Jailbreak Attempts

As Large Language Models (LLMs) of Prompt Jailbreaking are getting more and more attention, it is of great significance to raise a generalized research paradigm to evaluate attack strengths and a basic model to conduct subtler experiments. In this paper, we propose a novel approach by focusing on a set of target questions that are inherently more sensitive to jailbreak prompts, aiming to circumvent the limitations posed by enhanced LLM security. Through designing and analyzing these sensitive questions, this paper reveals a more effective method of identifying vulnerabilities in LLMs, thereby contributing to the advancement of LLM security. This research not only challenges existing jailbreaking methodologies but also fortifies LLMs against potential exploits.

PDF Cite

Recent Publications

Quickly discover relevant content by filtering publications.
Haonan Chen, Junxiao Li, Ruihai Wu, Yiwei Liu, Yiwen Hou, Zhixuan Xu, Jingxiang Guo, Chongkai Gao, Zhenyu Wei, Shensi Xu, Jiaqi Huang, Lin Shao (2025). MetaFold: Language-Guided Multi-Category Garment Folding Framework via Trajectory Generation and Foundation Model. In IROS 2025.

PDF Cite Code Dataset Project

Tianyu Zhang, Zixuan Zhao, Jiaqi Huang, Jingyu Hua, Sheng Zhong (2024). Subtoxic Questions: Dive Into Attitude Change of LLM’s Response in Jailbreak Attempts. In DLSP 2024.

PDF Cite

© 2025 Me. This work is licensed under CC BY NC ND 4.0

Published with Wowchemy — the free, open source website builder that empowers creators.

Cite
Copy Download