WebLINX: Real-World Website Navigation with Multi-Turn Dialogue

8 Feb 2024  ยท  Xing Han Lรน, Zdenฤ›k Kasner, Siva Reddy ยท

We propose the problem of conversational web navigation, where a digital agent controls a web browser and follows user instructions to solve real-world tasks in a multi-turn dialogue fashion. To support this problem, we introduce WEBLINX - a large-scale benchmark of 100K interactions across 2300 expert demonstrations of conversational web navigation. Our benchmark covers a broad range of patterns on over 150 real-world websites and can be used to train and evaluate agents in diverse scenarios. Due to the magnitude of information present, Large Language Models (LLMs) cannot process entire web pages in real-time. To solve this bottleneck, we design a retrieval-inspired model that efficiently prunes HTML pages by ranking relevant elements. We use the selected elements, along with screenshots and action history, to assess a variety of models for their ability to replicate human behavior when navigating the web. Our experiments span from small text-only to proprietary multimodal LLMs. We find that smaller finetuned decoders surpass the best zero-shot LLMs (including GPT-4V), but also larger finetuned multimodal models which were explicitly pretrained on screenshots. However, all finetuned models struggle to generalize to unseen websites. Our findings highlight the need for large multimodal models that can generalize to novel settings. Our code, data and models are available for research: https://mcgill-nlp.github.io/weblinx

PDF Abstract

Datasets


Introduced in the Paper:

WebLINX

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Conversational Web Navigation WebLINX GPT-3.5T (Zero-Shot) Overall score 8.51 # 17
Intent Match 42.77 # 15
Element (IoU) 8.62 # 15
Text (F1) 3.45 # 17
Conversational Web Navigation WebLINX GPT-4V (Zero-Shot) Overall score 10.45 # 16
Intent Match 42.36 # 16
Element (IoU) 10.91 # 13
Text (F1) 6.21 # 16
Conversational Web Navigation WebLINX GPT-4T (Zero-Shot) Overall score 10.72 # 15
Intent Match 41.66 # 17
Element (IoU) 10.85 # 14
Text (F1) 6.75 # 15
Conversational Web Navigation WebLINX Llama-2-13B Overall score 25.21 # 1
Intent Match 81.91 # 4
Element (IoU) 22.82 # 1
Text (F1) 26.60 # 2
Conversational Web Navigation WebLINX Pix2Act-282M Overall score 12.51 # 14
Intent Match 79.71 # 10
Element (IoU) 6.20 # 17
Text (F1) 16.40 # 10
Conversational Web Navigation WebLINX MindAct-250M Overall score 12.63 # 13
Intent Match 74.25 # 14
Element (IoU) 12.05 # 12
Text (F1) 7.67 # 14
Conversational Web Navigation WebLINX MindAct-780M Overall score 15.13 # 11
Intent Match 75.87 # 13
Element (IoU) 13.39 # 11
Text (F1) 13.58 # 12
Conversational Web Navigation WebLINX Flan-T5-250M Overall score 14.99 # 12
Intent Match 79.69 # 11
Element (IoU) 14.86 # 10
Text (F1) 9.21 # 13
Conversational Web Navigation WebLINX Pix2Act-1.3B Overall score 16.88 # 10
Intent Match 81.80 # 5
Element (IoU) 8.28 # 16
Text (F1) 25.21 # 6
Conversational Web Navigation WebLINX Flan-T5-780M Overall score 17.27 # 9
Intent Match 80.02 # 8
Element (IoU) 15.36 # 9
Text (F1) 14.05 # 11
Conversational Web Navigation WebLINX MindAct-3B Overall score 20.94 # 7
Intent Match 79.89 # 9
Element (IoU) 16.50 # 7
Text (F1) 23.16 # 7
Conversational Web Navigation WebLINX Fuyu-8B Overall score 19.97 # 8
Intent Match 80.07 # 7
Element (IoU) 15.70 # 8
Text (F1) 22.30 # 9
Conversational Web Navigation WebLINX GPT-3.5F Overall score 21.22 # 6
Intent Match 77.56 # 12
Element (IoU) 18.64 # 6
Text (F1) 22.39 # 8
Conversational Web Navigation WebLINX Flan-T5-3B Overall score 23.77 # 4
Intent Match 81.14 # 6
Element (IoU) 20.31 # 5
Text (F1) 25.75 # 5
Conversational Web Navigation WebLINX S-LLaMA-1.3B Overall score 23.73 # 5
Intent Match 83.32 # 2
Element (IoU) 20.54 # 4
Text (F1) 25.85 # 4
Conversational Web Navigation WebLINX Llama-2-7B Overall score 24.57 # 3
Intent Match 82.64 # 3
Element (IoU) 22.26 # 3
Text (F1) 26.50 # 3
Conversational Web Navigation WebLINX S-LLaMA-2.7B Overall score 25.02 # 2
Intent Match 84.00 # 1
Element (IoU) 22.60 # 2
Text (F1) 27.17 # 1

Methods


No methods listed for this paper. Add relevant methods here