Transfer Learning
2850 papers with code • 7 benchmarks • 15 datasets
Transfer Learning is a machine learning technique where a model trained on one task is re-purposed and fine-tuned for a related, but different task. The idea behind transfer learning is to leverage the knowledge learned from a pre-trained model to solve a new, but related problem. This can be useful in situations where there is limited data available to train a new model from scratch, or when the new task is similar enough to the original task that the pre-trained model can be adapted to the new problem with only minor modifications.
( Image credit: Subodh Malgonde )
Libraries
Use these libraries to find Transfer Learning models and implementationsDatasets
Subtasks
Latest papers with no code
Generation of Uncorrelated Residual Variables for Chemical Process Fault Diagnosis via Transfer Learning-based Input-Output Decoupled Network
However, traditional methods exhibit limited effectiveness in modeling high-dimensional nonlinearity and big data, and the decoupling idea has not been well-valued in data-driven frameworks.
Enhancing Action Recognition from Low-Quality Skeleton Data via Part-Level Knowledge Distillation
Skeleton-based action recognition is vital for comprehending human-centric videos and has applications in diverse domains.
EkoHate: Abusive Language and Hate Speech Detection for Code-switched Political Discussions on Nigerian Twitter
However, little or none has been done in the detection of abusive language and hate speech in Nigeria.
Transfer Learning Enhanced Single-choice Decision for Multi-choice Question Answering
Multi-choice Machine Reading Comprehension (MMRC) aims to select the correct answer from a set of options based on a given passage and question.
Toxicity Classification in Ukrainian
The task of toxicity detection is still a relevant task, especially in the context of safe and fair LMs development.
Remote Sensing Image Enhancement through Spatiotemporal Filtering
The newly developed filter proved that it can enhance the accuracy of classification using transfer learning by about 5%, 15%, and 2% for the three experiments respectively.
FTL: Transfer Learning Nonlinear Plasma Dynamic Transitions in Low Dimensional Embeddings via Deep Neural Networks
Deep learning algorithms provide a new paradigm to study high-dimensional dynamical behaviors, such as those in fusion plasma systems.
Causally Abstracted Multi-armed Bandits
Multi-armed bandits (MAB) and causal MABs (CMAB) are established frameworks for decision-making problems.
A Novel Spike Transformer Network for Depth Estimation from Event Cameras via Cross-modality Knowledge Distillation
This necessitates the development of innovative, spike-aware algorithms tailored for event cameras, a task compounded by the irregularity, continuity, noise, and spatial and temporal characteristics inherent in spiking data. Harnessing the strong generalization capabilities of transformer neural networks for spatiotemporal data, we propose a purely spike-driven spike transformer network for depth estimation from spiking camera data.
2M-NER: Contrastive Learning for Multilingual and Multimodal NER with Language and Modal Fusion
To tackle this challenging MMNER task on the dataset, we introduce a new model called 2M-NER, which aligns the text and image representations using contrastive learning and integrates a multimodal collaboration module to effectively depict the interactions between the two modalities.