0%

Representing Text for Joint Embedding of Text and Knowledge Bases 笔记

Perface

This is the notes when I was reading Representing Text for Joint Embedding of Text and Knowledge Bases paper.

Introduction Notes

This paper was based on jointly learning continuous representations for knowledge base and textual relations by Riedel et al. [1]. The textual relations of two entities were extracted by the lexicalized dependency path. The previous work made textual relation learn its continuous latent representation from the pattern of its co-occurrences in the knowledge graph.

The authors think that there are large common sub-structure (similar words and dependency paths) for synonymous textual relations. Therefore, they raised sharing parameters among related dependency paths. Their contribution is using a CNN to derive continuous representations for textual relations and this boosts the performance of link prediction for entity pairs have textual information.

Related Work Notes

The author introduce three types of related work for link prediction, using KB, using text and using both of them. Their work is built on the approach of Riedel et al 1 and the extension of the DISTMULT model 2 (a simple variant of a bilinear model) and jointly train the representations of KB and text relations.

Models for knowledge base completion Notes

$\textbf{Terminologies}$:

  • RDF triples: $(e_s, r, e_o)$, where $e_s$ is subject entity, $e_o$ is object entity and $r$ is relation.
  • Task: Predict relations between entities. Rank candidate entities for given queries like $(e_s, r, ?)$, $(?, r, e_o)$

They followed previous work 2 to represent both textual and knowledge base relations.
$\textbf{Textual relations}$: Use lexicalized dependency paths, like for entities $\text{BARACK OBAMA}$ and $\text{UNITED STATES}$, the textual relation is

The model uses a score between $(0, 1)$ to rank the probabilty of the triples.

Basic Models

$\textbf{Model E and F}$: they were used for a combined KB+text graph 1.

$\textbf{Model F}$:

  • K-dimention vector for both relation $r$ and entity pair $(e_s, e_o)$
  • Scoring function is $f(e_s, r, e_o) = v(r)^Tv(e_s, e_o)$
  • Do not share parameters for different entity pairs having same entity.

$\textbf{Model E}$:

  • Two K dimention vector $r_s, r_o$ for relations $r$
  • Each entity only has its own K-dimention vector (no vector for entity pair)
  • Scoring function is $f(e_s, r, e_o) = v(r_s)^Tv(e_s) + v(r_o)^Tv(e_o)$
  • For query $(e_s, r, ?)$, the ranking depends on $r$, not $e_s$

$\textbf{Model DISTMULT}$:

  • A special form of a bilinear model like RESCAL 3
  • K dimention vector for each entity $e_i$ and each relation $r$
  • Scoring function is $f(e_s, r, e_o) = v(r)^T(v(e_s) \circ v(e_o))$, where $\circ$ denotes the element-wise vector product
  • The entity pairs sharing an entiry will also share parameters, and ranking result for $(e_s ,r, ?)$ will depend on the $e_s$

CONV: Compositional Representations of Textual Relations

Because for the different dependency paths, the structure may be similar, the authors use CNN to encode the dependency arc and words as K-dimention vector and feed into CNN.

  • Assume each dependency arc or word is $e$
  • Firstly, map to embeddings $v = Ve$
  • The hidden layer is that $h = tanh(W^{-1}v{-1} + W^0v + W^1v{+1} + b)$
  • Use max-pooling to represent the final dependency paths $r = max{h}$

Loss function

For predicting entity, the authors use the function like negative loglikelihood

  • $p(e_s|e_o, r; \theta)$ is analogously
  • $\theta$ is the parameters
  • $Neg()$ is the negative sample for $(e_s, r, ?)$ (Because the whole samples are too large)

Because the textual relations task only takes part of the task (main is using KB relation), we use $\tau$ as weight for loss fucntion

So Final loss fucntion is

Reference

-------------本文结束感谢您的阅读-------------