250405_HW_LabSeminar[CoLAKE: Contextualized Language and Knowledge Embedding].pptx

thanhdowork 39 views 12 slides Apr 07, 2025
Slide 1
Slide 1 of 12
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12

About This Presentation

CoLAKE: Contextualized Language and Knowledge Embedding


Slide Content

K im Hyun W oo Network Science Lab The Catholic University of Korea E-mail : [email protected] CoLAKE : Contextualized Language and Knowledge Embedding

LLM with KG Deep contextualized language models pre-trained on large-scale unlabeled corpora have achived significant improvement on a wide range of NLP tasks However, they are shown to have difficulty capturing factual knowledge Sentence is not a “FACT”, it is a “Indicator of FACT”. PLM learn relations between the sentences. And by the learning process, the model seems like it know real world knowledge. But Knowledge graph(KG) is literally FACT. It shows relations between entities, direct knowledge. By put KG in PLM, we can make model can capture factual knowledge. Introduction

KEPLER: aims to benefit both sides jointly learn language model and knowledge embedding. However, KEPLER does not directly learn embeddings for each entity but learns to generate entity embedding with PLMs from descriptions. K-BERT: it is similar with this model, but K-BERT injecting triplet during fine-tuning Previous work LLM with KG

LLM with KG Model Architecture

LLM with KG Model Architecture By calculating attention socre only if tokens are connected, they can preserve structural information . (It works like GAT) On training stage, the model predicts masked words and entities, model learn contextualized representation for entities

LLM with KG Experiment They do three type of tasks. Kno wledge-driven tasks Entity typing Relation Extraction Knowledge-Probing Language understanding tasks Word-Kno wledge Graph Completion

Knowledge-Driven tasks LLM with KG

Knowledge Probing LLM with KG

Language understanding tasks LLM with KG

Word-Knowledge Graph Completion LLM with KG

They propose the model to jointly learn contextualized representation for language and knowledge. they show CoLAKE is essentially a powerful GNN that is structure aware. And demonstrate performance in various task. LLM with KG Conclusion

QA