Search Results for author: Leonard Lausen

Found 9 papers, 4 papers with code

Testing the Limits of Unified Sequence to Sequence LLM Pretraining on Diverse Table Data Tasks

no code implementations1 Oct 2023 Soumajyoti Sarkar, Leonard Lausen

Tables stored in databases and tables which are present in web pages and articles account for a large part of semi-structured data that is available on the internet.

Question Answering Semantic Parsing

Large Language Models of Code Fail at Completing Code with Potential Bugs

1 code implementation NeurIPS 2023 Tuan Dinh, Jinman Zhao, Samson Tan, Renato Negrinho, Leonard Lausen, Sheng Zha, George Karypis

We find that the presence of potential bugs significantly degrades the generation performance of the high-performing Code-LLMs.

Code Completion

Better Context Makes Better Code Language Models: A Case Study on Function Call Argument Completion

no code implementations1 Jun 2023 Hengzhi Pei, Jinman Zhao, Leonard Lausen, Sheng Zha, George Karypis

To better solve this task, we query a program analyzer for information relevant to a given function call, and consider ways to provide the analyzer results to different code completion models during inference and training.

Code Completion Program Synthesis

Parameter and Data Efficient Continual Pre-training for Robustness to Dialectal Variance in Arabic

no code implementations8 Nov 2022 Soumajyoti Sarkar, Kaixiang Lin, Sailik Sengupta, Leonard Lausen, Sheng Zha, Saab Mansour

While prior research studies have tried to adapt these multilingual models for dialectal variants of Arabic, it still remains a challenging problem owing to the lack of sufficient monolingual dialectal data and parallel translation data of such dialectal variants.

Avg Language Modelling +1

Exploring the Role of Task Transferability in Large-Scale Multi-Task Learning

no code implementations NAACL 2022 Vishakh Padmakumar, Leonard Lausen, Miguel Ballesteros, Sheng Zha, He He, George Karypis

Recent work has found that multi-task training with a large number of diverse tasks can uniformly improve downstream performance on unseen target tasks.

Multi-Task Learning Representation Learning

GluonCV and GluonNLP: Deep Learning in Computer Vision and Natural Language Processing

4 code implementations9 Jul 2019 Jian Guo, He He, Tong He, Leonard Lausen, Mu Li, Haibin Lin, Xingjian Shi, Chenguang Wang, Junyuan Xie, Sheng Zha, Aston Zhang, Hang Zhang, Zhi Zhang, Zhongyue Zhang, Shuai Zheng, Yi Zhu

We present GluonCV and GluonNLP, the deep learning toolkits for computer vision and natural language processing based on Apache MXNet (incubating).

NSML: A Machine Learning Platform That Enables You to Focus on Your Models

no code implementations16 Dec 2017 Nako Sung, Minkyu Kim, Hyunwoo Jo, Youngil Yang, Jingwoong Kim, Leonard Lausen, Youngkwan Kim, Gayoung Lee, Dong-Hyun Kwak, Jung-Woo Ha, Sunghun Kim

However, researchers are still required to perform a non-trivial amount of manual tasks such as GPU allocation, training status tracking, and comparison of models with different hyperparameter settings.

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.