Skip to content

Code for experiments on Domain Adaptation for the Claim Detection task

Notifications You must be signed in to change notification settings

SamiNenno/Domain-Adaptation-of-Claim-Detection

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Domain-Adaptation-of-Claim-Detection

Claim Detection is the task of spotting checkworthy sentences in a given text corpus, i.e. sentences that are potentially mis- or disinformation, interesting to fact-checkers, or interesting to the general public to know if they are true. In the recent years there have been multiple data sets and models published and there are many models that show strong results on the test sets. But how do models that are trained on claim detection data sets adapt to new domains such as new text types or topics? This question is important because in real-world applications, domain adaptation is crucial. Fact-checkers focus on political debates, Social Media, or blog posts and deal with topics ranging from immigration over climate change to COVID-19. For a successful application, claim detection models have to function across domains. I tried to simulate a real-world application that requires strong domain adaptation, by training several models on one of the data sets and test it the remaining data sets individually. The code for these experiments can be found here.

Data sets for the Claim Detection task

alt text

Performance of various ML models on all data set combinations (Recall/Precision)

alt text

About

Code for experiments on Domain Adaptation for the Claim Detection task

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages