Large Language Models for Scientific Question Answering: an Extensive Analysis of the SciQA Benchmark thumbnail
slide-image
Pause
Mute
Subtitles not available
Playback speed
0.25
0.5
0.75
1
1.25
1.5
1.75
2
Full screen

Large Language Models for Scientific Question Answering: an Extensive Analysis of the SciQA Benchmark

Published on Jun 18, 202439 Views

The SciQA benchmark for scientific question answering aims to represent a challenging task for next-generation question-answering systems on which vanilla large language models fail. In this article,

Related categories

Chapter list

Large Language Models for Scientific Question Answering00:00
The Task00:31
The Used Benchmark02:11
The Used Benchmark – Example I - 103:38
The Used Benchmark – Example I - 203:53
The Benchmark Composition - 104:15
The Benchmark Composition - 204:33
Approaches - 105:06
Approaches - 205:40
The Models05:49
The Adopted Metrics06:52
T507:15
GPT-207:58
Dolly08:24
GPT-3.509:34
Summary (F1)10:05
Category10:37
Conclusions12:32
Future Work13:49
Thank You!15:27