On the impact of AI on social science data quality and reproducibility
stefandietze
0 views
43 slides
Oct 15, 2025
Slide 1 of 43
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
About This Presentation
Research talk at LIRMM, Montpellier, France on 14 October 2025.
Abstract: Throughout the last decades, the social sciences have increasingly adopted novel forms of research data, e.g. data mined from the web and social media platforms. This together with the recent advances in artificial intelligen...
Research talk at LIRMM, Montpellier, France on 14 October 2025.
Abstract: Throughout the last decades, the social sciences have increasingly adopted novel forms of research data, e.g. data mined from the web and social media platforms. This together with the recent advances in artificial intelligence (AI) and related areas, e.g. natural language processing (NLP), led to a much more widespread adoption of diverse computational methods, including techniques from machine learning and, most prominently, large language models. However, increasingly complex computational methods lead to new challenges with respect to transparency, reproducibility and overall quality of social science research and data, further elevating an already widely recognised reproducibility crisis. This talk will, one the one hand, introduce challenges posed by the use of AI-based methods in social science research. On the other hand, it will show pathways to address such problems. Examples are works geared towards sharing computational (AI) methods in the social sciences in a reproducible and citable way, for understanding and tracing adoption of and relations between methods and datasets at large scale, e.g. in social science research in general (e.g. by mining scientific publications) or novel ways for providing access to sensitive research data in the social sciences (e.g. social media data) to facilitate reproducible research without violating ethical or legal constraints or principles.
Size: 3.74 MB
Language: en
Added: Oct 15, 2025
Slides: 43 pages
Slide Content
Social scientific data quality and reproducibility in the AI
era: challenges and pathways
LIRMM, 14 October2025
Stefan Dietze
Social scienceresearchischanging
▪Emergence of large volumes of
behavioral data (e.g. from social
media) has introduced new
research field (CSS), methods and
data
2
Behavioral web dataforthesocial sciences
▪Online discourse (e.g. in social media, online news)
▪Social web activity streams (posts, shares, likes, follows etc)
▪Web search behaviour, e.g. browsing, navigation or search engine interactions
▪Low-level behavioral traces (scrolling, mouse movements, gaze behavior etc)
▪General characteristics
oClose to users& theirpersonal (potentially sensitive) information
oLarge and heterogeneous
3
New kindsofdatarequirenewkindsofmethods
Methods widely used (e.g. for social media analysis) :
▪Time series analysis
(auto-regressive models, ARIMA etc)
▪Network/graph analysis
▪Dictionary-based methods
(e.g. for sentiment analysis)
▪Tailored machine learning models
(trained from scratch)
▪Pretrained open source language models (e.g. BERT)
▪Pretrained proprietary LLMs (like GPT/ChatGPT)
Substantial differences with
respect to:
▪Scalability (ability to handle
larger volumes of data)
▪Robustness (ability to handle
noisy or biased data)
▪Efficiency (compute/resource
requirements)
▪Transparency & interpretability
▪Reproducibility
„AI“
5
BeyondbasicuseofAI fordataanalysis: LLMs forsimulatinghuman behavior
Santurkar, S., et al., Whose Opinions Do Language Models Reflect?, International Conference on Machine Learning (ICML2023)
6
Different modelshavedifferent biases
(e.g. income, politicalleaning)
Steering themodelwithpersonas does
not leadtogrouprepresentativeness
LLMs arebiasedand intransparent
Santurkar, S., et al., Whose Opinions Do Language Models Reflect?, International Conference on Machine Learning (ICML2023)
7
Different modelshavedifferent biases
(e.g. income, politicalleaning)
Steering themodelwithpersonas does
not leadtogrouprepresentativeness
LLMs arebiasedand intransparent
Santurkar, S., et al., Whose Opinions Do Language Models Reflect?, International Conference on Machine Learning (ICML2023)
Key challenges:
▪LLMs arebiased(and wedo not fully understandbiases)
▪Provenanceofresponsesintransparent (modeland data)
▪LLMs not a goodchoicewhenrepresentativityand provenancematters
▪Access todataiscrucialto(a) understandpretrainedLLMs, (b) trainourown models/methods, (c)
mineopinionsfromactualdataratherthanopaqueblackboxes(LLMs)
8
Can AI actually„conduct“ research(„replaceresearchers“)?
The claim / hype:
▪AI startup claimed their ACL2025 paper (leading
A* NLP/AI conference) was “autonomously
created by AI” (research, experiments, writing)
▪Paper retracted/withdrawn later
But: real wave of research investigating AI
capabilities to conduct research, e.g.:
▪Identify SotA& research gaps (Si et al., 2025)
▪Reproduce research code (Boginet al., 2024)
▪Replicate research (Staraceet al., 2025)
9
Paradigmshift towardslesstransparent / reproduciblemodelsin NLP & CSS
Sristava, A., et al., Beyond the imitation game: quantifying & extrapolating the capabilities of language models (2022)
Model performanceincreaseswithsize
(and intransparency)
Large (and proprietary/ lessreproducible) models
areprevalentin CSS: modeladoptionat AAAI ICWSM
11
Reproducibilitycrisis: whatisthesituationin CS & AI?
▪Reproducibility crisis across disciplines: 90% agree (Baker, 2012)
▪In CS: experimental apparatus = “compute environment” => better controllable variables =>
reproducibility should be easier (compared to fields like sociology, physics, biology)
▪But: only 63.5% of CS papers successfully replicated (Raff, 2019),andonly 4% from papers
alone(Pineau et al., 2019)
▪Underspecificationof methods/experiments not seen in other disciplines
▪Negative impact of AI & deep learning (Dacremaet al., 2019)
Baker, M. 1,500 scientists lift the lid on reproducibility, Nature 533,2016
Raff, E., A step toward quantifying independently reproducible machine learning research. In Advances in Neural Information Processing Systems 2019.
Pineau,J., et. al., Improving Reproducibility in Machine Learning Research, Journal of Machine Learning Research 22 (2021) 1-20
Dacrema, M. F., et al., 2019. Are wereallymakingmuchprogress? A worryinganalysisofrecentneuralrecommendationapproaches. ACM RecSys2019.
13
Reproducibility: „A worringanalysis of neural recommender approaches”
MajorityofDL-basedmethodsisNOT reproducible
(„reproducibilitycrisis“)
Even thereproducibleonesdo NOT beatsimple baselines
(„benchmarking/ state-of-the-art crisis“)
Dacrema, M. F., et al., 2019. Are wereallymakingmuchprogress? A worryinganalysisofrecentneuralrecommendationapproaches. ACM RecSys2019.
Beyondreproducibility: do benchmarksassessgeneralisabilelearnings?
Example: Twitter bot detection
Chris Hays, Zachary Schutzman, Manish Raghavan, Erin Walk, and Philipp Zimmer. 2023. SimplisticCollection and LabelingPractices Limit theUtility ofBenchmark
Datasets forTwitter Bot Detection. ACM WebConf2023
18
„Shortcuts“ in thedata
Beyondreproducibility: do benchmarksassessgeneralisabilelearnings?
Example: Twitter bot detection
Chris Hays, Zachary Schutzman, Manish Raghavan, Erin Walk, and Philipp Zimmer. 2023. SimplisticCollection and LabelingPractices Limit theUtility ofBenchmark
Datasets forTwitter Bot Detection. ACM WebConf2023
Take-aways
▪Benchmark datasets don’t represent real-world data/problems but contain shortcuts
(“shortcut learning” => poor generalisability)
▪Benchmarking, i.e. understanding what is state-of-the-art in AI/NLP is hard
19
„Shortcuts“ in thedata
Addressingreproducibility& generalisabilityin CSS/AI research?
1.Empowering researchers to find state-of-the-art methods
(“benchmarking / state-of-the-art crisis”)
2.Improving the interpretability of scholarly reporting
(“reporting problem”)
3.Ensuring data availability & access
(“access problem”)
Reproducibility
Replicability
Robustness
Generalisability
22
Overview
1.Empowering researchers to find state-of-the-art methods
(“benchmarking / state-of-the-art crisis”)
2.Improving the interpretability of scholarly reporting
(“reporting problem”)
3.Ensuring data availability & access
(“access problem”)
23
EvaluatinggeneralisabilityofNLP models
Feger, M., Boland, K., Dietze, S., Limited Generalizability in Argument Mining: State-Of-The-Art Models Learn Datasets, Not Arguments, In ACL2025.
26
Example case: argument mining in tweets/social media posts as established NLP task
Feger, M., Boland, K., Dietze, S., Limited Generalizability in Argument Mining: State-Of-The-Art Models Learn Datasets, Not Arguments, In ACL2025.
Do modelsactuallygeneralise?
•Train-on-one-test-on-another
(dataset) experimentson 17 AM
datasets
•Usingstate-of-the-art
Transformer-basedlanguage
models(BERT, RoBERTa, WRAP)
•Models do not generalise(„do not
learntodetectarguments“):
performancedegradeswhen
modelsaretestedon OOD data
27
Realistic benchmarking: evaluating generalisabilityof NLP models
Feger, M., Boland, K., Dietze, S., Limited Generalizability in Argument Mining: State-Of-The-Art Models Learn Datasets, Not Arguments, In ACL2025.
•Leave-one-out crossvalidation:
modelstrainedon all datasetsbut
thetargetdataset(rows)
•Performance degradation
significant(despitemorediverse
trainingdata)
•Performance dropparticularlyfor
datasetsthatseemed„easy“ to
learn
28
Realistic benchmarking: evaluating generalisabilityof NLP models
Overview
1.Empowering researchers to find state-of-the-art methods
(“benchmarking / state-of-the-art crisis”)
2.Improving the interpretability of scholarly reporting
(“reporting problem”)
3.Ensuring data availability & access
(“access problem”)
31
Primary documentationofscientificoutput: unstructuredpublications
•Unsupported claims: e.g. over-generalization of claims or claimed
significance w/o statistical testing
•Informal citations of datasets & computational methods/code
(e.g. insufficient adoption of DOIs/PIDs)
•Broken citations (e.g. URLs are not accessible anymore or
code/data was changed)
•Ambiguous description of dataset/method adoption (e.g. sampling
methods from a large dataset)
•Mis-or underspecificationof ML models or training procedure (e.g.
training/test splits)
32
Reproducibilitychecklists
toenforcereproducibility
Momeni, F. et al., Checklists for Computational Reproducibility in the Social Sciences: Insights from Literature & Survey Evaluation. ACM Rep2025
34
•Checklists as common tool
(see also ACL, NeuRIPSetc)
Mining scholarlypapersforinformationaboutML models& data
Goal
▪Automatically mining papers (NLP) to
understand dataset, software and machine
learning method adoption
▪Creating a large knowledge base of ML
methods, tasks, datasets and how they are
used (cited) => e.g. GESIS Methods Hub
35
Otto, W., Zloch, M., Gan, L., Karmakar, S., Dietze, S. (2023). GSAP-NER: A Novel Task, Corpus, and Baseline for Scholarly EntityExtraction Focused on
Machine Learning Models and Datasets. In Findings of the Association for Computational Linguistics: EMNLP 2023
Approach
1.Manual annotationof > 54K mentions of
models, datasets etcin 100 publications
2.Finetuning PLMs for automatically detecting ML
model and dataset mentions
3.Applying trained modelson large publication
corpora (e.g. from ICWSM)
Mining scholarly papers for information about ML models & data
36
Detectingmodel, taskand datasetmentions: modelperformance
Otto, W., Zloch, M., Gan, L., Karmakar, S., Dietze, S. (2023). GSAP-NER: A Novel Task, Corpus, and Baseline for Scholarly EntityExtraction Focused on
Machine Learning Models and Datasets. In Findings of the Association for Computational Linguistics: EMNLP 2023
37
MethodMiner: a toolforminingtask, dataset& modelmentions
40
Otto, W., Upadhyaya, S., Gan, L., Silva, K. (2025), Track MachineLearning in YourResearch Domain. In2nd Conference on Research Data Infrastructure (CoRDI)
Overview
1.Empowering researchers to find state-of-the-art methods
(“benchmarking / state-of-the-art crisis”)
2.Improving the interpretability of scholarly reporting
(“reporting problem”)
3.Ensuring data availability & access
(“access problem”)
46
47
Challenge: dependencieson 3rd partygatekeepers
Behavioral data is not distributed as the web but tied
to platforms/gatekeepers
Challenge: volatility& decayofweb data
•Data isnot persistent
•Example: deletionratiooftweets
between25-29 %
•Differsbetweendifferent samples
Khan, M.T., Dimitrov, D., Dietze, S., Characterization of Tweet Deletion Patterns in the Context of COVID-19 Discourse and Polarization, ACM Hypertext 2025
48
Challenge: data evolution impacts methods (quality/reproducibility)
▪Vocabulary evolves: e.g. vocabulary shift, over-
/underrepresentation of topics/vocabulary in
particular time periods (e.g. Twitter COVID19-
discourse 2020 vs prior periods)
▪PLMs/LLMs require frequent training and
updates (and continuous access to data)
Source: Hombaiahet al., “Dynamic Language Models for continuously evolving Content”, SIGKDD2021
49
Responsible social media archiving @ GESIS: examples
X/Twitter (https://data.gesis.org/tweetskb)
▪Sampling: 1% -randomsample
▪Dataset size: > 14 billiontweets
▪Time period: Feb 2013 -June 2023
Telegram(https://data.gesis.org/telescope)
▪Sampling: seedlists+ snowballsampling
▪Dataset: ~120M messages from ~71K public channels andmetadata for
~500K channels
▪Time period: Feb 2024 and running
Fact-checkedclaims(https://data.gesis.org/claimskg)
▪Sampling method: 13 factchecking websites
▪Dataset: 74066 claims and 72128 claim reviews
▪Time period: claims published between 1996 –2023
4Chan
▪Sampling method: all boards
▪Dataset size: 4,676,378 threads, 264,898,231 posts
▪Time period: Nov 2023 and running
▪In preparation: BlueSky, YouTube, …
https://www.gesis.org/gesis-web-data
51
Case study: harvesting 1% of Twitter/X
53
▪Complete 1% sample of all tweets
(14 billion tweets between 04/2013 –
05/2023)
▪Legal, ethical and licensing constraints:
social media data is sensitive (!)
▪Data sharing via:
▪Secure data access (online/offline
secure data access)
▪Public, non-sensitive dataoffers
Distributed redundant crawlersovertime
NLP methods for generating non-sensitive data offers
http://dbpedia.org/page/COVID-19
negative emotion
hasEmotionIntensity"0.25“
positive emotion
hasEmotionIntensity"0.73“
http://dbpedia.org/page/COVID-19_vaccine
Dimitrov, D., Fafalios, P., Yu, R., Zhu, X., Zloch, M., Dietze, S., TweetsCOV19 –A KB of Semantically Annotated Tweets about the COVID-19 Pandemic, CIKM2020
Motivation
Providing derived, non-sensitive dataproductsfrom
rawarchives
Approach
•Offeringtweet metadataand derivedfeaturesthat
capturetweet semantics, e.g.:
•Entities(e.g. “China Virus” => dbp:COVID-19)
•Sentiments
•Georeferences
•Arguments/stances
•Large, non-sensitive dataproductssuch asTweetsKB
(https://data.gesis.org/tweetskb/), TweetsCOV19
(https://data.gesis.org/tweetscov19/),
> 3 bnannotatedtweets
54
Germany suspends
vaccinations with Astra
Zeneca
Twitter discourse on
“Impfbereitschaft” /
„Vaccination hesitancy“
TweetsKB as social science research corpus
Investigating vaccine hesitancy in DACH countries
https://dd4p.gesis.org/
Boland, K. et al., Data for policy-making in times of crisis -a computational analysis of German online discourses about COVID-19 vaccinations, JMIR2025
Germany suspends
vaccinations with Astra
Zeneca
55
TeleScope: a longitudinal corpusofTelegramdiscourse
▪Telegram channels: public, only admin can post
▪Decentralised: no registry of channels available
▪Continuous data collection of currently 1.2 M channels
through snowball sampling (300 seed channels)
▪Full message history collected for > 70 K public channels;
approx. 120 M messages so far
▪Message interaction data computed for whole dataset
(forwards, views) to facilitate Twitter-like analysis
Gangopadhyay, S., Dessi, D., Dimitrov, D., Dietze, S., TeleScope: A Longitudinal Dataset for Investigating Online Discourse and Information Interaction on
Telegram, AAAI ICWSM2025
56
57
Responsible social media archiving @ GESIS
https://www.gesis.org/gesis-web-data
57