no code implementations • 18 Sep 2023 • Yang Zhang, YuFei Wang, Kai Wang, Quan Z. Sheng, Lina Yao, Adnan Mahmood, Wei Emma Zhang, Rongying Zhao
Such information could be incorporated into LLMs pre-training and improve the text representation in LLMs.
no code implementations • 4 Aug 2023 • Munazza Zaib, Wei Emma Zhang, Quan Z. Sheng, Subhash Sagar, Adnan Mahmood, Yang Zhang
In this paper, we propose a framework, DHS-ConvQA (Dynamic History Selection in Conversational Question Answering), that first generates the context and question entities for all the history turns, which are then pruned on the basis of similarity they share in common with the question at hand.
no code implementations • 14 Apr 2023 • Munazza Zaib, Quan Z. Sheng, Wei Emma Zhang, Adnan Mahmood
However, these sequential questions are sometimes left implicit and thus require the resolution of some natural language phenomena such as anaphora and ellipsis.
no code implementations • 8 Jul 2022 • Venus Haghighi, Behnaz Soltani, Adnan Mahmood, Quan Z. Sheng, Jian Yang
Anomaly detection in attributed networks has received a considerable attention in recent years due to its applications in a wide range of domains such as finance, network security, and medicine.
no code implementations • 8 Jul 2022 • Behnaz Soltani, Venus Haghighi, Adnan Mahmood, Quan Z. Sheng, Lina Yao
The main challenges of FL is that end devices usually possess various computation and communication capabilities and their training data are not independent and identically distributed (non-IID).
no code implementations • 2 Jun 2021 • Munazza Zaib, Wei Emma Zhang, Quan Z. Sheng, Adnan Mahmood, Yang Zhang
Question answering (QA) systems provide a way of querying the information available in various formats including, but not limited to, unstructured and structured data in natural languages.
no code implementations • 23 Apr 2021 • Munazza Zaib, Dai Hoang Tran, Subhash Sagar, Adnan Mahmood, Wei E. Zhang, Quan Z. Sheng
On one hand, we introduce a framework based on a publically available pre-trained language model called BERT for incorporating history turns into the system.