Search Results

Operationalizing a National Digital Library: The Case for a Norwegian Transformer Model

2 code implementations NoDaLiDa 2021

In this work, we show the process of building a large-scale training set from digital and digitized collections at a national library.

Language Modelling Optical Character Recognition +1

Robust Digital-Twin Localization via An RGBD-based Transformer Network and A Comprehensive Evaluation on a Mobile Dataset

3 code implementations24 Sep 2023

The potential of digital-twin technology, involving the creation of precise digital replicas of physical objects, to reshape AR experiences in 3D object tracking and localization scenarios is significant.

3D Object Tracking Object +1

Fast Autofocusing using Tiny Transformer Networks for Digital Holographic Microscopy

1 code implementation15 Mar 2022

Tiny DL models are proposed and compared such as a tiny Vision Transformer (TViT), tiny VGG16 (TVGG) and a tiny Swin-Transfomer (TSwinT).

Deep Learning Framework for Measuring the Digital Strategy of Companies from Earnings Calls

1 code implementation COLING 2020

Companies today are racing to leverage the latest digital technologies, such as artificial intelligence, blockchain, and cloud computing.

Cloud Computing Clustering +2

Nougat: Neural Optical Understanding for Academic Documents

2 code implementations25 Aug 2023

Scientific knowledge is predominantly stored in books and scientific journals, often in the form of PDFs.

Optical Character Recognition Optical Character Recognition (OCR)

LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding

5 code implementations ACL 2021

Pre-training of text and layout has proved effective in a variety of visually-rich document understanding tasks due to its effective model architecture and the advantage of large-scale unlabeled scanned/digital-born documents.

Document Image Classification Document Layout Analysis +6

Supervised Multimodal Bitransformers for Classifying Images and Text

6 code implementations6 Sep 2019

Self-supervised bidirectional transformer models such as BERT have led to dramatic improvements in a wide variety of textual classification tasks.

 Ranked #1 on Natural Language Inference on V-SNLI (using extra training data)

General Classification Natural Language Inference

Dynamic Routing Between Capsules

78 code implementations NeurIPS 2017

We use the length of the activity vector to represent the probability that the entity exists and its orientation to represent the instantiation parameters.

Image Classification