no code implementations • 3 Oct 2022 • Lauren J. Wong, Sean McPherson, Alan J. Michaels
Transfer learning (TL) techniques, which leverage prior knowledge gained from data with different distributions to achieve higher performance and reduced training time, are often used in computer vision (CV) and natural language processing (NLP), but have yet to be fully utilized in the field of radio frequency machine learning (RFML).
no code implementations • 16 Jun 2022 • Lauren J. Wong, Sean McPherson, Alan J. Michaels
The use of transfer learning (TL) techniques has become common practice in fields such as computer vision (CV) and natural language processing (NLP).
no code implementations • 7 May 2022 • William H. Clark IV, Alan J. Michaels
While the model's deployed performance is dependent on numerous variables within the scope of machine learning, beyond that of the training data itself, the effect of the dataset is isolated in this work to better understand the role training data plays in the problem.
no code implementations • 1 Oct 2020 • Lauren J. Wong, William H. Clark IV, Bryse Flowers, R. Michael Buehrer, Alan J. Michaels, William C. Headley
While deep machine learning technologies are now pervasive in state-of-the-art image recognition and natural language processing applications, only in recent years have these technologies started to sufficiently mature in applications related to wireless communications.
no code implementations • 1 Oct 2020 • William H. Clark IV, Steven Hauser, William C. Headley, Alan J. Michaels
Looking into the Radio Frequency Machine Learning (RFML) field of Automatic Modulation Classification (AMC) as an example of a tool used for situational awareness, the use of synthetic, captured, and augmented data are examined and compared to provide insights about the quantity and quality of the available data necessary to achieve desired performance levels.