Search Results for author: Wissam Antoun

Found 9 papers, 5 papers with code

From Text to Source: Results in Detecting Large Language Model-Generated Content

no code implementations23 Sep 2023 Wissam Antoun, Benoît Sagot, Djamé Seddah

The research also explores Model Attribution, encompassing source model identification, model family, and model size classification, in addition to quantization and watermarking detection.

Attribute Language Modelling +3

Towards a Robust Detection of Language Model Generated Text: Is ChatGPT that Easy to Detect?

no code implementations9 Jun 2023 Wissam Antoun, Virginie Mouilleron, Benoît Sagot, Djamé Seddah

This paper proposes a methodology for developing and evaluating ChatGPT detectors for French text, with a focus on investigating their robustness on out-of-domain data and against common attack schemes.

Adversarial Text Language Modelling

Data-Efficient French Language Modeling with CamemBERTa

no code implementations2 Jun 2023 Wissam Antoun, Benoît Sagot, Djamé Seddah

In this paper, we introduce CamemBERTa, a French DeBERTa model that builds upon the DeBERTaV3 architecture and training objective.

Dependency Parsing FLUE +5

AraGPT2: Pre-Trained Transformer for Arabic Language Generation

1 code implementation EACL (WANLP) 2021 Wissam Antoun, Fady Baly, Hazem Hajj

In this paper, we develop the first advanced Arabic language generation model, AraGPT2, trained from scratch on a large Arabic corpus of internet text and news articles.

Language Modelling News Generation +1

AraELECTRA: Pre-Training Text Discriminators for Arabic Language Understanding

1 code implementation EACL (WANLP) 2021 Wissam Antoun, Fady Baly, Hazem Hajj

Advances in English language representation enabled a more sample-efficient pre-training task by Efficiently Learning an Encoder that Classifies Token Replacements Accurately (ELECTRA).

Language Modelling Masked Language Modeling +5

Multi-Task Learning using AraBert for Offensive Language Detection

no code implementations LREC 2020 Dj, Marc ji, Fady Baly, Wissam Antoun, Hazem Hajj

The shared task on Offensive Language Detection at the OSACT4 has aimed at achieving state of art profane language detection methods for Arabic social media.

Language Modelling Multi-Task Learning

AraBERT: Transformer-based Model for Arabic Language Understanding

3 code implementations LREC 2020 Wissam Antoun, Fady Baly, Hazem Hajj

Recently, with the surge of transformers based models, language-specific BERT based models have proven to be very efficient at language understanding, provided they are pre-trained on a very large corpus.

named-entity-recognition Named Entity Recognition +3

hULMonA: The Universal Language Model in Arabic

1 code implementation WS 2019 Obeida ElJundi, Wissam Antoun, Nour El Droubi, Hazem Hajj, Wassim El-Hajj, Khaled Shaban

Experiment results show that the developed hULMonA and multi-lingual ULM are able to generalize well to multiple Arabic data sets and achieve new state of the art results in Arabic Sentiment Analysis for some of the tested sets.

Arabic Sentiment Analysis General Classification +4

Cannot find the paper you are looking for? You can Submit a new open access paper.