On the impact of AI on social science data quality and reproducibility

stefandietze 0 views 43 slides Oct 15, 2025
Slide 1
Slide 1 of 43
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20
Slide 21
21
Slide 22
22
Slide 23
23
Slide 24
24
Slide 25
25
Slide 26
26
Slide 27
27
Slide 28
28
Slide 29
29
Slide 30
30
Slide 31
31
Slide 32
32
Slide 33
33
Slide 34
34
Slide 35
35
Slide 36
36
Slide 37
37
Slide 38
38
Slide 39
39
Slide 40
40
Slide 41
41
Slide 42
42
Slide 43
43

About This Presentation

Research talk at LIRMM, Montpellier, France on 14 October 2025.
Abstract: Throughout the last decades, the social sciences have increasingly adopted novel forms of research data, e.g. data mined from the web and social media platforms. This together with the recent advances in artificial intelligen...


Slide Content

Social scientific data quality and reproducibility in the AI
era: challenges and pathways
LIRMM, 14 October2025
Stefan Dietze

Social scienceresearchischanging
▪Emergence of large volumes of
behavioral data (e.g. from social
media) has introduced new
research field (CSS), methods and
data
2

Behavioral web dataforthesocial sciences
▪Online discourse (e.g. in social media, online news)
▪Social web activity streams (posts, shares, likes, follows etc)
▪Web search behaviour, e.g. browsing, navigation or search engine interactions
▪Low-level behavioral traces (scrolling, mouse movements, gaze behavior etc)
▪General characteristics
oClose to users& theirpersonal (potentially sensitive) information
oLarge and heterogeneous
3

New kindsofdatarequirenewkindsofmethods
Methods widely used (e.g. for social media analysis) :
▪Time series analysis
(auto-regressive models, ARIMA etc)
▪Network/graph analysis
▪Dictionary-based methods
(e.g. for sentiment analysis)
▪Tailored machine learning models
(trained from scratch)
▪Pretrained open source language models (e.g. BERT)
▪Pretrained proprietary LLMs (like GPT/ChatGPT)
Substantial differences with
respect to:
▪Scalability (ability to handle
larger volumes of data)
▪Robustness (ability to handle
noisy or biased data)
▪Efficiency (compute/resource
requirements)
▪Transparency & interpretability
▪Reproducibility
„AI“
5

BeyondbasicuseofAI fordataanalysis: LLMs forsimulatinghuman behavior
Santurkar, S., et al., Whose Opinions Do Language Models Reflect?, International Conference on Machine Learning (ICML2023)
6

Different modelshavedifferent biases
(e.g. income, politicalleaning)
Steering themodelwithpersonas does
not leadtogrouprepresentativeness
LLMs arebiasedand intransparent
Santurkar, S., et al., Whose Opinions Do Language Models Reflect?, International Conference on Machine Learning (ICML2023)
7

Different modelshavedifferent biases
(e.g. income, politicalleaning)
Steering themodelwithpersonas does
not leadtogrouprepresentativeness
LLMs arebiasedand intransparent
Santurkar, S., et al., Whose Opinions Do Language Models Reflect?, International Conference on Machine Learning (ICML2023)
Key challenges:
▪LLMs arebiased(and wedo not fully understandbiases)
▪Provenanceofresponsesintransparent (modeland data)
▪LLMs not a goodchoicewhenrepresentativityand provenancematters
▪Access todataiscrucialto(a) understandpretrainedLLMs, (b) trainourown models/methods, (c)
mineopinionsfromactualdataratherthanopaqueblackboxes(LLMs)
8

Can AI actually„conduct“ research(„replaceresearchers“)?
The claim / hype:
▪AI startup claimed their ACL2025 paper (leading
A* NLP/AI conference) was “autonomously
created by AI” (research, experiments, writing)
▪Paper retracted/withdrawn later
But: real wave of research investigating AI
capabilities to conduct research, e.g.:
▪Identify SotA& research gaps (Si et al., 2025)
▪Reproduce research code (Boginet al., 2024)
▪Replicate research (Staraceet al., 2025)
9

Paradigmshift towardslesstransparent / reproduciblemodelsin NLP & CSS
Sristava, A., et al., Beyond the imitation game: quantifying & extrapolating the capabilities of language models (2022)
Model performanceincreaseswithsize
(and intransparency)
Large (and proprietary/ lessreproducible) models
areprevalentin CSS: modeladoptionat AAAI ICWSM
11

Reproducibilitycrisis: whatisthesituationin CS & AI?
▪Reproducibility crisis across disciplines: 90% agree (Baker, 2012)
▪In CS: experimental apparatus = “compute environment” => better controllable variables =>
reproducibility should be easier (compared to fields like sociology, physics, biology)
▪But: only 63.5% of CS papers successfully replicated (Raff, 2019),andonly 4% from papers
alone(Pineau et al., 2019)
▪Underspecificationof methods/experiments not seen in other disciplines
▪Negative impact of AI & deep learning (Dacremaet al., 2019)
Baker, M. 1,500 scientists lift the lid on reproducibility, Nature 533,2016
Raff, E., A step toward quantifying independently reproducible machine learning research. In Advances in Neural Information Processing Systems 2019.
Pineau,J., et. al., Improving Reproducibility in Machine Learning Research, Journal of Machine Learning Research 22 (2021) 1-20
Dacrema, M. F., et al., 2019. Are wereallymakingmuchprogress? A worryinganalysisofrecentneuralrecommendationapproaches. ACM RecSys2019.
13

Reproducibility: „A worringanalysis of neural recommender approaches”
MajorityofDL-basedmethodsisNOT reproducible
(„reproducibilitycrisis“)
Even thereproducibleonesdo NOT beatsimple baselines
(„benchmarking/ state-of-the-art crisis“)
Dacrema, M. F., et al., 2019. Are wereallymakingmuchprogress? A worryinganalysisofrecentneuralrecommendationapproaches. ACM RecSys2019.

Beyondreproducibility: do benchmarksassessgeneralisabilelearnings?
Example: Twitter bot detection
Chris Hays, Zachary Schutzman, Manish Raghavan, Erin Walk, and Philipp Zimmer. 2023. SimplisticCollection and LabelingPractices Limit theUtility ofBenchmark
Datasets forTwitter Bot Detection. ACM WebConf2023
18
„Shortcuts“ in thedata

Beyondreproducibility: do benchmarksassessgeneralisabilelearnings?
Example: Twitter bot detection
Chris Hays, Zachary Schutzman, Manish Raghavan, Erin Walk, and Philipp Zimmer. 2023. SimplisticCollection and LabelingPractices Limit theUtility ofBenchmark
Datasets forTwitter Bot Detection. ACM WebConf2023
Take-aways
▪Benchmark datasets don’t represent real-world data/problems but contain shortcuts
(“shortcut learning” => poor generalisability)
▪Benchmarking, i.e. understanding what is state-of-the-art in AI/NLP is hard
19
„Shortcuts“ in thedata

Addressingreproducibility& generalisabilityin CSS/AI research?
1.Empowering researchers to find state-of-the-art methods
(“benchmarking / state-of-the-art crisis”)
2.Improving the interpretability of scholarly reporting
(“reporting problem”)
3.Ensuring data availability & access
(“access problem”)
Reproducibility
Replicability
Robustness
Generalisability
22

Overview
1.Empowering researchers to find state-of-the-art methods
(“benchmarking / state-of-the-art crisis”)
2.Improving the interpretability of scholarly reporting
(“reporting problem”)
3.Ensuring data availability & access
(“access problem”)
23

Key challenge: howtoidentifyhigh qualitymethods?
Howtofind SotAmethodsforgiventask(e.g. stancedetectionon specifictweet sample)?
•Review literature: labor-intensive, methods oftenpoorlycited/ not traceable
•Code/modelrepositories(e.g. HuggingFace, GitHub): lack context(e.g. relatedresearch,
comparisonswithothermethods etc)
•Ad-hoc choices(„I usewhatI know“)
Benchmarking ofAI/CS methods
•Use ofstandardevaluationcorpora& metricstocomparemethodperformance/ quality
•In theory: benchmarksassesswhethera publishedmethodisgood/bad/state-of-the-art
•In practice: benchmarksand benchmarkingpractices(egbaselinechoices) areflawed, e.g.
do not evaluategeneralisability
24

EvaluatinggeneralisabilityofNLP models
Feger, M., Boland, K., Dietze, S., Limited Generalizability in Argument Mining: State-Of-The-Art Models Learn Datasets, Not Arguments, In ACL2025.
26
Example case: argument mining in tweets/social media posts as established NLP task

Feger, M., Boland, K., Dietze, S., Limited Generalizability in Argument Mining: State-Of-The-Art Models Learn Datasets, Not Arguments, In ACL2025.
Do modelsactuallygeneralise?
•Train-on-one-test-on-another
(dataset) experimentson 17 AM
datasets
•Usingstate-of-the-art
Transformer-basedlanguage
models(BERT, RoBERTa, WRAP)
•Models do not generalise(„do not
learntodetectarguments“):
performancedegradeswhen
modelsaretestedon OOD data
27
Realistic benchmarking: evaluating generalisabilityof NLP models

Feger, M., Boland, K., Dietze, S., Limited Generalizability in Argument Mining: State-Of-The-Art Models Learn Datasets, Not Arguments, In ACL2025.
•Leave-one-out crossvalidation:
modelstrainedon all datasetsbut
thetargetdataset(rows)
•Performance degradation
significant(despitemorediverse
trainingdata)
•Performance dropparticularlyfor
datasetsthatseemed„easy“ to
learn
28
Realistic benchmarking: evaluating generalisabilityof NLP models

Promotingmorerealisticbenchmarkingpracticesand corporaat CLEF2026
https://clef2026.clef-initiative.eu/
29

FindingAI methods forthesocial sciences: GESIS Methods Hub
Release in Q3 2025
Integrated intoGESIS Search
•Platform forfinding, sharing& usingdata
science& AI methods
•Empoweringsocial scientistswith&
withouttechnicalexpertisetouse
complexstate-of-the-art methods & LLMs
•GESIS-curatedand community-based
methods and tutorials
•Focus on reproducibility, quality, citability
(DOIs), benchmarking, provenance
https://methodshub.gesis.org
30

Overview
1.Empowering researchers to find state-of-the-art methods
(“benchmarking / state-of-the-art crisis”)
2.Improving the interpretability of scholarly reporting
(“reporting problem”)
3.Ensuring data availability & access
(“access problem”)
31

Primary documentationofscientificoutput: unstructuredpublications
•Unsupported claims: e.g. over-generalization of claims or claimed
significance w/o statistical testing
•Informal citations of datasets & computational methods/code
(e.g. insufficient adoption of DOIs/PIDs)
•Broken citations (e.g. URLs are not accessible anymore or
code/data was changed)
•Ambiguous description of dataset/method adoption (e.g. sampling
methods from a large dataset)
•Mis-or underspecificationof ML models or training procedure (e.g.
training/test splits)
32

Reproducibilitychecklists
toenforcereproducibility
Momeni, F. et al., Checklists for Computational Reproducibility in the Social Sciences: Insights from Literature & Survey Evaluation. ACM Rep2025
34
•Checklists as common tool
(see also ACL, NeuRIPSetc)

Mining scholarlypapersforinformationaboutML models& data
Goal
▪Automatically mining papers (NLP) to
understand dataset, software and machine
learning method adoption
▪Creating a large knowledge base of ML
methods, tasks, datasets and how they are
used (cited) => e.g. GESIS Methods Hub
35

Otto, W., Zloch, M., Gan, L., Karmakar, S., Dietze, S. (2023). GSAP-NER: A Novel Task, Corpus, and Baseline for Scholarly EntityExtraction Focused on
Machine Learning Models and Datasets. In Findings of the Association for Computational Linguistics: EMNLP 2023
Approach
1.Manual annotationof > 54K mentions of
models, datasets etcin 100 publications
2.Finetuning PLMs for automatically detecting ML
model and dataset mentions
3.Applying trained modelson large publication
corpora (e.g. from ICWSM)
Mining scholarly papers for information about ML models & data
36

Detectingmodel, taskand datasetmentions: modelperformance
Otto, W., Zloch, M., Gan, L., Karmakar, S., Dietze, S. (2023). GSAP-NER: A Novel Task, Corpus, and Baseline for Scholarly EntityExtraction Focused on
Machine Learning Models and Datasets. In Findings of the Association for Computational Linguistics: EMNLP 2023
37

Understanding methods and datain CSS (AAAI ICWSM publications)
Tasks Methods
38

Understanding methods and datain CSS (AAAI ICWSM publications)
CitationsofML modelsovertime Citationsofdatasourcesovertime
39

MethodMiner: a toolforminingtask, dataset& modelmentions
40
Otto, W., Upadhyaya, S., Gan, L., Silva, K. (2025), Track MachineLearning in YourResearch Domain. In2nd Conference on Research Data Infrastructure (CoRDI)

SharedAI task@ ACL2025: miningdata, model, softwarementions
https://sdproc.org/2025/somd25.html
45

Overview
1.Empowering researchers to find state-of-the-art methods
(“benchmarking / state-of-the-art crisis”)
2.Improving the interpretability of scholarly reporting
(“reporting problem”)
3.Ensuring data availability & access
(“access problem”)
46

47
Challenge: dependencieson 3rd partygatekeepers
Behavioral data is not distributed as the web but tied
to platforms/gatekeepers

Challenge: volatility& decayofweb data
•Data isnot persistent
•Example: deletionratiooftweets
between25-29 %
•Differsbetweendifferent samples
Khan, M.T., Dimitrov, D., Dietze, S., Characterization of Tweet Deletion Patterns in the Context of COVID-19 Discourse and Polarization, ACM Hypertext 2025
48

Challenge: data evolution impacts methods (quality/reproducibility)
▪Vocabulary evolves: e.g. vocabulary shift, over-
/underrepresentation of topics/vocabulary in
particular time periods (e.g. Twitter COVID19-
discourse 2020 vs prior periods)
▪PLMs/LLMs require frequent training and
updates (and continuous access to data)
Source: Hombaiahet al., “Dynamic Language Models for continuously evolving Content”, SIGKDD2021
49

Responsible social media archiving @ GESIS: examples
X/Twitter (https://data.gesis.org/tweetskb)
▪Sampling: 1% -randomsample
▪Dataset size: > 14 billiontweets
▪Time period: Feb 2013 -June 2023
Telegram(https://data.gesis.org/telescope)
▪Sampling: seedlists+ snowballsampling
▪Dataset: ~120M messages from ~71K public channels andmetadata for
~500K channels
▪Time period: Feb 2024 and running
Fact-checkedclaims(https://data.gesis.org/claimskg)
▪Sampling method: 13 factchecking websites
▪Dataset: 74066 claims and 72128 claim reviews
▪Time period: claims published between 1996 –2023
4Chan
▪Sampling method: all boards
▪Dataset size: 4,676,378 threads, 264,898,231 posts
▪Time period: Nov 2023 and running
▪In preparation: BlueSky, YouTube, …
https://www.gesis.org/gesis-web-data
51

Case study: harvesting 1% of Twitter/X
53
▪Complete 1% sample of all tweets
(14 billion tweets between 04/2013 –
05/2023)
▪Legal, ethical and licensing constraints:
social media data is sensitive (!)
▪Data sharing via:
▪Secure data access (online/offline
secure data access)
▪Public, non-sensitive dataoffers
Distributed redundant crawlersovertime

NLP methods for generating non-sensitive data offers
http://dbpedia.org/page/COVID-19
negative emotion
hasEmotionIntensity"0.25“
positive emotion
hasEmotionIntensity"0.73“
http://dbpedia.org/page/COVID-19_vaccine
Dimitrov, D., Fafalios, P., Yu, R., Zhu, X., Zloch, M., Dietze, S., TweetsCOV19 –A KB of Semantically Annotated Tweets about the COVID-19 Pandemic, CIKM2020
Motivation
Providing derived, non-sensitive dataproductsfrom
rawarchives
Approach
•Offeringtweet metadataand derivedfeaturesthat
capturetweet semantics, e.g.:
•Entities(e.g. “China Virus” => dbp:COVID-19)
•Sentiments
•Georeferences
•Arguments/stances
•Large, non-sensitive dataproductssuch asTweetsKB
(https://data.gesis.org/tweetskb/), TweetsCOV19
(https://data.gesis.org/tweetscov19/),
> 3 bnannotatedtweets
54

Germany suspends
vaccinations with Astra
Zeneca
Twitter discourse on
“Impfbereitschaft” /
„Vaccination hesitancy“
TweetsKB as social science research corpus
Investigating vaccine hesitancy in DACH countries
https://dd4p.gesis.org/
Boland, K. et al., Data for policy-making in times of crisis -a computational analysis of German online discourses about COVID-19 vaccinations, JMIR2025
Germany suspends
vaccinations with Astra
Zeneca
55

TeleScope: a longitudinal corpusofTelegramdiscourse
▪Telegram channels: public, only admin can post
▪Decentralised: no registry of channels available
▪Continuous data collection of currently 1.2 M channels
through snowball sampling (300 seed channels)
▪Full message history collected for > 70 K public channels;
approx. 120 M messages so far
▪Message interaction data computed for whole dataset
(forwards, views) to facilitate Twitter-like analysis
Gangopadhyay, S., Dessi, D., Dimitrov, D., Dietze, S., TeleScope: A Longitudinal Dataset for Investigating Online Discourse and Information Interaction on
Telegram, AAAI ICWSM2025
56

57
Responsible social media archiving @ GESIS
https://www.gesis.org/gesis-web-data
57

Take-aways: towardsbettermethodquality& reproducibility
•Web dataarchivingforresearchcommunity
•Non-sensitive datacorpora(e.g. TweetsKB) & secureaccess
•Legal conditionsforsafe useofweb data& methods
Findingmethods&
understandingSotA
Reporting quality
Data access
•Incentivisingbetterreportinghabits(e.g. DOIs, citations) through
reproducibilitychecklists
•Automatedminingofmethod/datacitations
•Method curation& documentation(Methods Hub)
•Betterbenchmarkingpractices(evaluatinggeneralisability)
•Community engagementin benchmarkingand sharedtasks
Culture change& interdisciplinarycollaboration
61

https://stefandietze.net
https://gesis.org/en/kts
Thank you!