Natural Language Processing • Language Models • 93 methods
Transformers are a type of neural network architecture that have several properties that make them effective for modeling data with long-range dependencies. They generally feature a combination of multi-headed attention mechanisms, residual connections, layer normalization, feedforward connections, and positional embeddings.
Subcategories
Method | Year | Papers |
---|---|---|
2017 | 9476 | |
2023 | 5925 | |
2018 | 5075 | |
2020 | 1353 | |
2019 | 750 | |
2018 | 673 | |
2019 | 555 | |
2019 | 544 | |
2019 | 456 | |
2020 | 158 | |
2019 | 156 | |
2019 | 155 | |
2019 | 126 | |
2020 | 119 | |
2022 | 112 | |
2020 | 108 | |
2020 | 83 | |
2020 | 78 | |
2020 | 72 | |
2020 | 66 | |
2019 | 61 | |
2020 | 59 | |
2019 | 55 | |
2020 | 49 | |
2019 | 48 | |
2019 | 43 | |
2019 | 35 | |
2019 | 30 | |
2020 | 29 | |
2000 | 28 | |
2019 | 21 | |
2021 | 18 | |
2018 | 17 | |
2021 | 16 | |
2020 | 15 | |
2020 | 14 | |
2020 | 13 | |
2019 | 12 | |
2021 | 10 | |
2020 | 9 | |
2021 | 8 | |
2022 | 8 | |
2020 | 6 | |
2021 | 6 | |
2020 | 5 | |
2021 | 5 | |
2021 | 5 | |
2020 | 5 | |
2021 | 3 | |
2019 | 3 | |
2021 | 3 | |
2019 | 3 | |
2021 | 3 | |
2019 | 3 | |
2020 | 3 | |
2020 | 3 | |
2020 | 3 | |
2020 | 2 | |
2019 | 2 | |
2021 | 2 | |
2021 | 2 | |
2020 | 2 | |
2021 | 2 | |
2020 | 2 | |
2021 | 2 | |
2022 | 2 | |
2021 | 2 | |
2020 | 1 | |
2020 | 1 | |
2021 | 1 | |
2021 | 1 | |
2019 | 1 | |
2020 | 1 | |
2020 | 1 | |
2021 | 1 | |
2020 | 1 | |
2020 | 1 | |
2020 | 1 | |
2020 | 1 | |
2018 | 1 | |
2020 | 1 | |
2020 | 1 | |
2021 | 1 | |
2021 | 1 | |
2020 | 1 | |
2019 | 1 | |
2020 | 1 | |
2019 | 1 |