Search Results for author: Siyuan Ma

Found 12 papers, 5 papers with code

JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks

no code implementations3 Apr 2024 Weidi Luo, Siyuan Ma, Xiaogeng Liu, XIAOYU GUO, Chaowei Xiao

With the rapid advancements in Multimodal Large Language Models (MLLMs), securing these models against malicious inputs while aligning them with human values has emerged as a critical challenge.

Understanding News Creation Intents: Frame, Dataset, and Method

1 code implementation27 Dec 2023 Zhengjia Wang, Danding Wang, Qiang Sheng, Juan Cao, Silong Su, Yifan Sun, Beizhe Hu, Siyuan Ma

As the disruptive changes in the media economy and the proliferation of alternative news media outlets, news intent has progressively deviated from ethical standards that serve the public interest.

Philosophy

Summary of ChatGPT-Related Research and Perspective Towards the Future of Large Language Models

no code implementations4 Apr 2023 Yiheng Liu, Tianle Han, Siyuan Ma, Jiayue Zhang, Yuanyuan Yang, Jiaming Tian, Hao He, Antong Li, Mengshen He, Zhengliang Liu, Zihao Wu, Lin Zhao, Dajiang Zhu, Xiang Li, Ning Qiang, Dingang Shen, Tianming Liu, Bao Ge

This paper presents a comprehensive survey of ChatGPT-related (GPT-3. 5 and GPT-4) research, state-of-the-art large language models (LLM) from the GPT series, and their prospective applications across diverse domains.

Learning the joint distribution of two sequences using little or no paired data

no code implementations6 Dec 2022 Soroosh Mariooryad, Matt Shannon, Siyuan Ma, Tom Bagby, David Kao, Daisy Stanton, Eric Battenberg, RJ Skerry-Ryan

We present a noisy channel generative model of two sequences, for example text and speech, which enables uncovering the association between the two modalities when limited paired data is available.

Variational Inference

Gated Transformer Networks for Multivariate Time Series Classification

2 code implementations26 Mar 2021 Minghao Liu, Shengqi Ren, Siyuan Ma, Jiahui Jiao, Yizhou Chen, Zhiguang Wang, Wei Song

In this work, we explored a simple extension of the current Transformer Networks with gating, named Gated Transformer Networks (GTN) for the multivariate time series classification problem.

Classification General Classification +3

Reconciling modern machine learning practice and the bias-variance trade-off

2 code implementations28 Dec 2018 Mikhail Belkin, Daniel Hsu, Siyuan Ma, Soumik Mandal

This connection between the performance and the structure of machine learning models delineates the limits of classical analyses, and has implications for both the theory and practice of machine learning.

BIG-bench Machine Learning

On exponential convergence of SGD in non-convex over-parametrized learning

no code implementations6 Nov 2018 Raef Bassily, Mikhail Belkin, Siyuan Ma

Large over-parametrized models learned via stochastic gradient descent (SGD) methods have become a key element in modern machine learning.

BIG-bench Machine Learning

Kernel machines that adapt to GPUs for effective large batch training

2 code implementations15 Jun 2018 Siyuan Ma, Mikhail Belkin

In this paper we develop the first analytical framework that extends linear scaling to match the parallel computing capacity of a resource.

To understand deep learning we need to understand kernel learning

no code implementations ICML 2018 Mikhail Belkin, Siyuan Ma, Soumik Mandal

Certain key phenomena of deep learning are manifested similarly in kernel methods in the modern "overfitted" regime.

Generalization Bounds

The Power of Interpolation: Understanding the Effectiveness of SGD in Modern Over-parametrized Learning

no code implementations ICML 2018 Siyuan Ma, Raef Bassily, Mikhail Belkin

We show that there is a critical batch size $m^*$ such that: (a) SGD iteration with mini-batch size $m\leq m^*$ is nearly equivalent to $m$ iterations of mini-batch size $1$ (\emph{linear scaling regime}).

Diving into the shallows: a computational perspective on large-scale shallow learning

1 code implementation NeurIPS 2017 Siyuan Ma, Mikhail Belkin

An analysis based on the spectral properties of the kernel demonstrates that only a vanishingly small portion of the function space is reachable after a polynomial number of gradient descent iterations.

Cannot find the paper you are looking for? You can Submit a new open access paper.