250310_JH_Seminar[Translating Embeddings for Modeling Multi-relational Data].pptx

thanhdowork 95 views 14 slides Mar 10, 2025
Slide 1
Slide 1 of 14
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14

About This Presentation

Translating Embeddings for Modeling Multi-relational Data


Slide Content

Cho junhee Network Science Lab The Catholic University of Korea E-mail : [email protected] Translating Embeddings for Modeling Multi-relational Data

Abstract Objective: Embed entities and relationships into a low-dimensional vector space. Key Features: Simple, easy to train, and requires fewer parameters. Scales efficiently to large datasets. Methodology: Relationships are modeled as translations between entity embeddings. Current Status

Introduction Directed graph where nodes are entities and edges represent relationships in the form (head, label, tail). Objection of this paper is to develop an efficient tool to automatically complete KBs without extra knowledge Multi-Relational Data & Its Importance

Introduction Modeling Multi-Relational Data = Extract local/global connectivity patterns to predict relationships. Since multiple types of relationships and entities are involved simultaneously, a more generalized approach is required. To handle multiple data, they tried to increasing the expressivity and universality of the model imitations of Complex Models Difficult to interpret due to high expressivity. High computational cost and risk of overfitting/underfitting. Research Insight: Simpler linear models can perform almost as well as complex ones, balancing accuracy & scalability. Modeling Multi-Relational Data: Challenges & Approaches

Introduction Hierarchical Relationships in KBs Common in knowledge bases → Naturally represented by translations Tree Structure Representation: Siblings → Close in embedding space Parent-Child → Translation along the y-axis Zero translation → Represents equivalence (e.g., sibling relationships) 1-to-1 Relationships (e.g., Country-Capital) Found in word embeddings from free text Suggests that translation naturally models structured relationships Motivation Behind TransE

Introduction Algorithm

Translation-based model Goal Learn entity (h,t) and relation (ℓ) embeddings in vector space Model relations as translations in embedding space If (h,ℓ,t) is true: Otherwise:   TransE Model Overview

Translation-based model Loss Function & Negative Sampling Objective: Lower energy for positive triplets Higher energy for corrupted triplets

Translation-based model Corrupted Triplets (Negative Sampling) Randomly replace head or tail (not both) Encourages model to learn triplet validity Loss Function & Negative Sampling

Translation-based model Stochastic Gradient Descent (SGD) : Optimization over h,ℓ,t embeddings Assitional Constraints : L2-normalization of entity embeddings No constraints for relations (ℓ) Why? To p revents trivial solutions (infinite norm growth) Optimization & Training Process

Translation-based model Stochastic Gradient Descent (SGD) : Optimization over h,ℓ,t embeddings Assitional Constraints : L2-normalization of entity embeddings No constraints for relations (ℓ) Why? To p revents trivial solutions (infinite norm growth) Optimization & Training Process

Experiment wordnet, Freebase Data sets

Experiment Link prediction result

Conclusion Advantages: Highly efficient & fast (SGD-based training) Scalable to large-scale datasets Simple yet powerful model Limitations: Struggles with 1:N, N:1, N:N relations Cannot model symmetric relationships effectively Does not consider relation-specific transformations (→ Improved models: TransH, TransR) Motivation Behind TransE
Tags