besskge.bess.EmbeddingMovingBessKGE

class besskge.bess.EmbeddingMovingBessKGE(negative_sampler, score_fn, loss_fn=None, evaluation=None, return_scores=False, augment_negative=False)[source]

Compute negative scores on the shard where the positive triples are scored (namely the head shard). This requires moving the embedding of negative entities between shards, which can be done with a single AllToAll collective.

Each triple is scored against a total number of entities equal to n_negative * n_shard if negative sample sharing is disabled, or to n_negative * n_shard * B otherwise (see BessKGE.forward()) for “h”, “t” corruption scheme, n_negative * n_shard * (B > 2 ? B // 2 : 1) for “ht”.

Initialize BESS-KGE module.

Parameters:
  • negative_sampler (ShardedNegativeSampler) – Sampler of negative entities.

  • score_fn (BaseScoreFunction) – Scoring function.

  • loss_fn (Optional[BaseLossFunction]) – Loss function, required when training. Default: None.

  • evaluation (Optional[Evaluation]) – Evaluation module, for computing metrics on device. Default: None.

  • return_scores (bool) – If True, return positive and negative scores of batches to the host. Default: False.

  • augment_negative (bool) – If True, augment sampled negative entities with the head/tails (according to the corruption scheme) of other positive triples in the micro-batch. Default: False.

forward(head, relation, tail, negative, triple_mask=None, triple_weight=None, negative_mask=None)

The forward step.

Comprises of four phases:

  1. Gather relevant embeddings from local memory;

  2. Share embeddings with other devices through collective operators;

  3. Score positive and negative triples;

  4. Compute loss/metrics.

Each device scores n_shard * positive_per_partition positive triples.

Parameters:
  • head (Tensor) – shape: (1, n_shard, positive_per_partition) Head indices.

  • relation (Tensor) – shape: (1, n_shard, positive_per_partition) Relation indices.

  • tail (Tensor) – shape: (1, n_shard, positive_per_partition) Tail indices.

  • triple_mask (Optional[Tensor]) – shape: (1, n_shard, positive_per_partition) Mask to filter the triples in the micro-batch before computing metrics.

  • negative (Tensor) – shape: (1, n_shard, B, padded_negative) Indices of negative entities, with B = 1, 2 or n_shard * positive_per_partition.

  • triple_weight (Optional[Tensor]) – shape: (1, n_shard * positive_per_partition,) or (1,) Weights of positive triples.

  • negative_mask (Optional[Tensor]) – shape: (1, B, n_shard, padded_negative) Mask to identify padding negatives, to discard when computing metrics.

Return type:

Dict[str, Any]

Returns:

Micro-batch loss, scores and metrics.

property n_embedding_parameters: int

Returns the number of trainable parameters in the embedding tables

score_batch(head, relation, tail, negative)[source]

Compute positive and negative scores for the micro-batch.

Parameters:
Return type:

Tuple[Tensor, Tensor]

Returns:

Positive (shape: (n_shard * positive_per_partition,)) and negative (shape: (n_shard * positive_per_partition, n_negative)) scores of the micro-batch.