no code implementations • 15 Mar 2024 • Eugene Jang, Jian Cui, Dayeon Yim, Youngjin Jin, Jin-Woo Chung, Seungwon Shin, YongJae lee
We use our domain-customized methodology to train CyBERTuned, a cybersecurity domain language model that outperforms other cybersecurity PLMs on most tasks.
1 code implementation • 22 Jan 2024 • Yu Sun, Gaojian Xiong, Xianxun Yao, Kailang Ma, Jian Cui
Deep gradient inversion attacks expose a serious threat to Federated Learning (FL) by accurately recovering private data from shared gradients.
no code implementations • 6 Jan 2024 • Zilong Lin, Jian Cui, Xiaojing Liao, XiaoFeng Wang
The underground exploitation of large language models (LLMs) for malicious services (i. e., Malla) is witnessing an uptick, amplifying the cyber threat landscape and posing questions about the trustworthiness of LLM technologies.
no code implementations • 15 May 2023 • Youngjin Jin, Eugene Jang, Jian Cui, Jin-Woo Chung, YongJae lee, Seungwon Shin
Recent research has suggested that there are clear differences in the language used in the Dark Web compared to that of the Surface Web.
no code implementations • 4 Apr 2023 • Joe Yue-Hei Ng, Kevin McCloskey, Jian Cui, Vincent R. Meijer, Erica Brand, Aaron Sarna, Nita Goyal, Christopher Van Arsdale, Scott Geraedts
Contrails (condensation trails) are line-shaped ice clouds caused by aircraft and are likely the largest contributor of aviation-induced climate change.
1 code implementation • International Conference on Learning Representations 2023 • Kailang Ma, Yu Sun, Jian Cui, Dawei Li, Zhenyu Guan and Jianwei Liu
Furthermore, we demonstrate that our method facilitates the existing gradient inversion attacks by exploiting the recovered labels, with an increase of 6-7 in PSNR on both MNIST and CIFAR100.
1 code implementation • 12 Feb 2022 • Jian Cui, Liqiang Yuan, Zhaoxiang Wang, Ruilin Li, Tianzi Jiang
In addition, we also find that the quality of the interpretation results is inconsistent for individual samples despite when a method with an overall good performance is used.
1 code implementation • 21 Nov 2021 • Jian Cui, Zirui Lan, Tianhu Zheng, Yisi Liu, Olga Sourina, Lipo Wang, Wolfgang Müller-Wittig
For EEG-based drowsiness recognition, it is desirable to use subject-independent recognition since conducting calibration on each subject is time-consuming.
no code implementations • 13 Sep 2021 • Jian Cui, Kwanwoo Kim, Seung Ho Na, Seungwon Shin
We then propose Meta-Path instance encoding and aggregation methods to capture the temporal information of user engagement and produce news representation end-to-end.
1 code implementation • 30 May 2021 • Jian Cui, Zirui Lan, Olga Sourina, Wolfgang Müller-Wittig
Results show that the model achieves an average accuracy of 78. 35% on 11 subjects for leave-one-out cross-subject drowsiness recognition, which is higher than the conventional baseline methods of 53. 40%-72. 68% and state-of-the-art deep learning methods of 71. 75%-75. 19%.
1 code implementation • 30 May 2021 • Jian Cui, Zirui Lan, Yisi Liu, Ruilin Li, Fan Li, Olga Sourina, Wolfgang Mueller-Wittig
Driver drowsiness is one of main factors leading to road fatalities and hazards in the transportation industry.
2 code implementations • IEEE Transactions on Knowledge and Data Engineering 2020 • Zheng Wang, Xiaojun Ye, Chaokun Wang, Jian Cui, Philip S. Yu
Network embedding, aiming to project a network into a low-dimensional space, is increasingly becoming a focus of network research.