no code implementations • 29 Feb 2024 • He Zhu, Wenjia Zhang, Nuoxian Huang, Boyang Li, Luyao Niu, Zipei Fan, Tianle Lun, Yicheng Tao, Junyou Su, Zhaoya Gong, Chenyu Fang, Xing Liu
In the field of urban planning, general-purpose large language models often struggle to meet the specific needs of planners.
1 code implementation • 28 Feb 2024 • Hongchao Zhang, Luyao Niu, Andrew Clark, Radha Poovendran
Control barrier function (CBF)-based approaches have been proposed to guarantee the safety of robotic systems.
1 code implementation • 19 Feb 2024 • Fengqing Jiang, Zhangchen Xu, Luyao Niu, Zhen Xiang, Bhaskar Ramasubramanian, Bo Li, Radha Poovendran
In this paper, we propose a novel ASCII art-based jailbreak attack and introduce a comprehensive benchmark Vision-in-Text Challenge (ViTC) to evaluate the capabilities of LLMs in recognizing prompts that cannot be solely interpreted by semantics.
1 code implementation • 14 Feb 2024 • Zhangchen Xu, Fengqing Jiang, Luyao Niu, Jinyuan Jia, Bill Yuchen Lin, Radha Poovendran
Our results show that SafeDecoding significantly reduces the attack success rate and harmfulness of jailbreak attacks without compromising the helpfulness of responses to benign user queries.
no code implementations • 12 Feb 2024 • Dinuka Sahabandu, Xiaojun Xu, Arezoo Rajabi, Luyao Niu, Bhaskar Ramasubramanian, Bo Li, Radha Poovendran
We propose and analyze an adaptive adversary that can retrain a Trojaned DNN and is also aware of SOTA output-based Trojaned model detectors.
no code implementations • 10 Jan 2024 • Zhangchen Xu, Fengqing Jiang, Luyao Niu, Jinyuan Jia, Radha Poovendran
Our results show that the global model learned with Brave in the presence of adversaries achieves comparable classification accuracy to a global model trained in the absence of any adversary.
no code implementations • 7 Nov 2023 • Fengqing Jiang, Zhangchen Xu, Luyao Niu, Boxin Wang, Jinyuan Jia, Bo Li, Radha Poovendran
Successful exploits of the identified vulnerabilities result in the users receiving responses tailored to the intent of a threat initiator.
1 code implementation • 30 Aug 2023 • Arezoo Rajabi, Surudhi Asokraj, Fengqing Jiang, Luyao Niu, Bhaskar Ramasubramanian, Jim Ritcey, Radha Poovendran
An adversary carrying out a backdoor attack embeds a predefined perturbation called a trigger into a small subset of input samples and trains the DNN such that the presence of the trigger in the input results in an adversary-desired output class.
no code implementations • 4 Apr 2023 • Abdullah Al Maruf, Luyao Niu, Bhaskar Ramasubramanian, Andrew Clark, Radha Poovendran
We then propose a distributed MARL algorithm called the CVaR QD-Learning algorithm, and establish that value functions of individual agents reaches consensus.
no code implementations • 3 Dec 2022 • Arezoo Rajabi, Dinuka Sahabandu, Luyao Niu, Bhaskar Ramasubramanian, Radha Poovendran
Overfitted models have been shown to be susceptible to query-based attacks such as membership inference attacks (MIAs).
no code implementations • 22 Aug 2022 • Luyao Niu, Zhouchi Li, Andrew Clark
We develop a class of fault-tolerant finite time convergence control barrier functions (CBFs) to guarantee that a dynamical system reaches a set within finite time almost surely in the presence of malicious attacks.
no code implementations • 11 Aug 2022 • Hongchao Zhang, Shiyu Cheng, Luyao Niu, Andrew Clark
We prove that the synthesized control input guarantees system safety using control barrier certificates.
no code implementations • 13 Jul 2022 • Dinuka Sahabandu, Arezoo Rajabi, Luyao Niu, Bo Li, Bhaskar Ramasubramanian, Radha Poovendran
The results show that (i) with Submodular Trojan algorithm, the adversary needs to embed a Trojan trigger into a very small fraction of samples to achieve high accuracy on both Trojan and clean samples, and (ii) the MM Trojan algorithm yields a trained Trojan model that evades detection with probability 1.
no code implementations • 28 Sep 2021 • Luyao Niu, Hongchao Zhang, Andrew Clark
By satisfying the constructed CBF constraint at each sampling time, we guarantee the unknown sampled-data system is safe for all time.
no code implementations • 3 Aug 2021 • Luyao Niu, Dinuka Sahabandu, Andrew Clark, Radha Poovendran
In this paper, we study the controlled islanding problem of a power system under disturbances introduced by a malicious adversary.
no code implementations • 29 Mar 2021 • Bhaskar Ramasubramanian, Luyao Niu, Andrew Clark, Radha Poovendran
In this paper, we consider a setting where an autonomous agent has to learn behaviors in an unknown environment.
no code implementations • 28 Feb 2021 • Zhouchi Li, Luyao Niu, Andrew Clark
For each possible set of compromised sensors, we maintain a state estimator disregarding the sensors in that set, and calculate the optimal LQG control input at each time based on this estimate.