besskge.bess.ScoreMovingBessKGE

class besskge.bess.ScoreMovingBessKGE(negative_sampler, score_fn, loss_fn=None, evaluation=None, return_scores=False, augment_negative=False)[source]

Compute negative scores on the shard where the negative entities are stored. This avoids moving embeddings between shards (convenient when the number of negative entities is very large, for example when scoring queries against all entities in the knowledge graph, or when using a large embedding size).

AllGather collectives are required to replicate queries on all devices, so that they can be scored against the local negative entities. An AllToAll collective is then used to send the scores back to the correct device.

For the number of negative samples scored for each triple, see the corresponding value documented in EmbeddingMovingBessKGE and, if using negative sample sharing, multiply that by n_shard.

Does not support local sampling or negative augmentation.

Initialize BESS-KGE module.

Parameters:
  • negative_sampler (ShardedNegativeSampler) – Sampler of negative entities.

  • score_fn (BaseScoreFunction) – Scoring function.

  • loss_fn (Optional[BaseLossFunction]) – Loss function, required when training. Default: None.

  • evaluation (Optional[Evaluation]) – Evaluation module, for computing metrics on device. Default: None.

  • return_scores (bool) – If True, return positive and negative scores of batches to the host. Default: False.

  • augment_negative (bool) – If True, augment sampled negative entities with the head/tails (according to the corruption scheme) of other positive triples in the micro-batch. Default: False.

forward(head, relation, tail, negative, triple_mask=None, triple_weight=None, negative_mask=None)

The forward step.

Comprises of four phases:

  1. Gather relevant embeddings from local memory;

  2. Share embeddings with other devices through collective operators;

  3. Score positive and negative triples;

  4. Compute loss/metrics.

Each device scores n_shard * positive_per_partition positive triples.

Parameters:
  • head (Tensor) – shape: (1, n_shard, positive_per_partition) Head indices.

  • relation (Tensor) – shape: (1, n_shard, positive_per_partition) Relation indices.

  • tail (Tensor) – shape: (1, n_shard, positive_per_partition) Tail indices.

  • triple_mask (Optional[Tensor]) – shape: (1, n_shard, positive_per_partition) Mask to filter the triples in the micro-batch before computing metrics.

  • negative (Tensor) – shape: (1, n_shard, B, padded_negative) Indices of negative entities, with B = 1, 2 or n_shard * positive_per_partition.

  • triple_weight (Optional[Tensor]) – shape: (1, n_shard * positive_per_partition,) or (1,) Weights of positive triples.

  • negative_mask (Optional[Tensor]) – shape: (1, B, n_shard, padded_negative) Mask to identify padding negatives, to discard when computing metrics.

Return type:

Dict[str, Any]

Returns:

Micro-batch loss, scores and metrics.

property n_embedding_parameters: int

Returns the number of trainable parameters in the embedding tables

score_batch(head, relation, tail, negative)[source]

Compute positive and negative scores for the micro-batch.

Parameters:
Return type:

Tuple[Tensor, Tensor]

Returns:

Positive (shape: (n_shard * positive_per_partition,)) and negative (shape: (n_shard * positive_per_partition, n_negative)) scores of the micro-batch.