Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add omnibus_embedding_pairwise to pipeline module #823

Merged
merged 6 commits into from
Aug 19, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
1 change: 1 addition & 0 deletions graspologic/pipeline/embed/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,3 +13,4 @@
from .adjacency_spectral_embedding import adjacency_spectral_embedding
from .embeddings import Embeddings, EmbeddingsView
from .laplacian_spectral_embedding import laplacian_spectral_embedding
from .omnibus_embedding import omnibus_embedding_pairwise
285 changes: 285 additions & 0 deletions graspologic/pipeline/embed/omnibus_embedding.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,285 @@
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.

import warnings
from typing import List, Optional, Tuple, Union

import networkx as nx
import numpy as np
from beartype import beartype

from graspologic.embed import OmnibusEmbed
from graspologic.preconditions import check_argument, is_real_weighted
from graspologic.utils import (
augment_diagonal,
largest_connected_component,
pass_to_ranks,
remove_loops,
)

from . import __SVD_SOLVER_TYPES
from ._elbow import _index_of_elbow
from .embeddings import Embeddings


@beartype
def omnibus_embedding_pairwise(
graphs: List[Union[nx.Graph, nx.OrderedGraph, nx.DiGraph, nx.OrderedDiGraph]],
dimensions: int = 100,
elbow_cut: Optional[int] = None,
svd_solver_algorithm: str = "randomized",
svd_solver_iterations: int = 5,
svd_seed: Optional[int] = None,
weight_attribute: str = "weight",
) -> List[Tuple[Embeddings, Embeddings]]:
"""
Generates a pairwise omnibus embedding for each pair of graphs in a list of graphs using the adjacency matrix.
If given graphs A, B, and C, the embeddings will be computed for A, B and B, C.

If the node labels differ between each pair of graphs, then those nodes will only be found in the resulting embedding
if they exist in the largest connected component of the union of all edges across all graphs in the time series.

Graphs will always have their diagonal augmented. In other words, a self-loop
will be created for each node with a weight corresponding to the weighted degree.

Lastly, all weights will be rescaled based on their relative rank in the graph,
which is beneficial in minimizing anomalous results if some edge weights are
extremely atypical of the rest of the graph.

Parameters
----------
graphs : List[Union[nx.Graph, nx.OrderedGraph, nx.DiGraph, nx.OrderedDiGraph]]
A list of undirected or directed graphs. The graphs **must**:

- be fully numerically weighted (every edge must have a real, numeric weight
or else it will be treated as an unweighted graph)
- be a basic graph (meaning it should not be a multigraph; if you have a
multigraph you must first decide how you want to handle the weights of the
edges between two nodes, whether summed, averaged, last-wins,
maximum-weight-only, etc)
dimensions : int (default=100)
Dimensions to use for the svd solver.
For undirected graphs, if ``elbow_cut==None``, you will receive an embedding
that has ``nodes`` rows and ``dimensions`` columns.
For directed graphs, if ``elbow_cut==None``, you will receive an embedding that
has ``nodes`` rows and ``2*dimensions`` columns.
If ``elbow_cut`` is specified to be not ``None``, we will cut the embedding at
``elbow_cut`` elbow, but the provided ``dimensions`` will be used in the
creation of the SVD.
elbow_cut : Optional[int] (default=None)
Using a process described by Zhu & Ghodsi in their paper "Automatic
dimensionality selection from the scree plot via the use of profile likelihood",
truncate the dimensionality of the return on the ``elbow_cut``-th elbow.
By default this value is ``None`` but can be used to reduce the dimensionality
of the returned tensors.
svd_solver_algorithm : str (default="randomized")
allowed values: {'randomized', 'full', 'truncated'}

SVD solver to use:

- 'randomized'
Computes randomized svd using
:func:`sklearn.utils.extmath.randomized_svd`
- 'full'
Computes full svd using :func:`scipy.linalg.svd`
Does not support ``graph`` input of type scipy.sparse.csr_matrix
- 'truncated'
Computes truncated svd using :func:`scipy.sparse.linalg.svds`
svd_solver_iterations : int (default=5)
Number of iterations for randomized SVD solver. Not used by 'full' or
'truncated'. The default is larger than the default in randomized_svd
to handle sparse matrices that may have large slowly decaying spectrum.
svd_seed : Optional[int] (default=None)
Used to seed the PRNG used in the ``randomized`` svd solver algorithm.
weight_attribute : str (default="weight")
The edge dictionary key that contains the weight of the edge.

Returns
-------
List[Tuple[Embeddings, Embeddings]]

Raises
------
beartype.roar.BeartypeCallHintPepParamException if parameters do not match type hints
ValueError if values are not within appropriate ranges or allowed values

See Also
--------
graspologic.pipeline.embed.Embeddings
graspologic.embed.OmnibusEmbed
graspologic.embed.AdjacencySpectralEmbed
graspologic.embed.select_svd

References
----------
.. [1] Levin, K., Athreya, A., Tang, M., Lyzinski, V., & Priebe, C. E. (2017,
November). A central limit theorem for an omnibus embedding of multiple random
dot product graphs. In Data Mining Workshops (ICDMW), 2017 IEEE International
Conference on (pp. 964-967). IEEE.

.. [2] Sussman, D.L., Tang, M., Fishkind, D.E., Priebe, C.E. "A
Consistent Adjacency Spectral Embedding for Stochastic Blockmodel Graphs,"
Journal of the American Statistical Association, Vol. 107(499), 2012

.. [3] Levin, K., Roosta-Khorasani, F., Mahoney, M. W., & Priebe, C. E. (2018).
Out-of-sample extension of graph adjacency spectral embedding. PMLR: Proceedings
of Machine Learning Research, 80, 2975-2984.

.. [4] Zhu, M. and Ghodsi, A. (2006). Automatic dimensionality selection from the
scree plot via the use of profile likelihood. Computational Statistics & Data
Analysis, 51(2), pp.918-930.
"""
check_argument(len(graphs) > 1, "more than one graph is required")

check_argument(dimensions >= 1, "dimensions must be positive")

check_argument(elbow_cut is None or elbow_cut >= 1, "elbow_cut must be positive")

check_argument(
svd_solver_algorithm in __SVD_SOLVER_TYPES,
f"svd_solver_algorithm must be one of the values in {','.join(__SVD_SOLVER_TYPES)}",
)

check_argument(svd_solver_iterations >= 1, "svd_solver_iterations must be positive")

check_argument(
svd_seed is None or 0 <= svd_seed <= 2 ** 32 - 1,
"svd_seed must be a nonnegative, 32-bit integer",
)

weight_attribute = _graphs_precondition_checks(graphs, weight_attribute)

graph_embeddings = []

# create a graph that contains all nodes and edges across the entire corpus
union_graph = graphs[0].copy()
for graph in graphs[1:]:
union_graph.add_edges_from(graph.edges())

union_graph_lcc = largest_connected_component(union_graph)
union_graph_lcc_nodes = union_graph_lcc.nodes()
union_node_ids = np.array(list(union_graph_lcc_nodes))
Comment on lines +154 to +161
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@nyecarr maybe you can use this in #824 too?


previous_graph = graphs[0].copy()

for graph in graphs[1:]:
current_graph = graph.copy()

# assure both graphs contain the exact same node set
# by removing nodes or adding isolates as needed
_sync_nodes(previous_graph, union_graph_lcc_nodes)
_sync_nodes(current_graph, union_graph_lcc_nodes)

# remove self loops, run pass to ranks and diagonal augmentation
previous_graph_augmented = _augment_graph(
previous_graph, union_graph_lcc_nodes, weight_attribute
)
current_graph_augmented = _augment_graph(
current_graph, union_graph_lcc_nodes, weight_attribute
)

model = OmnibusEmbed(
n_components=dimensions,
n_elbows=None, # we will do elbow cuts
algorithm=svd_solver_algorithm,
n_iter=svd_solver_iterations,
check_lcc=False,
diag_aug=False,
concat=False,
svd_seed=svd_seed,
)

previous_embedding, current_embedding = model.fit_transform(
graphs=[previous_graph_augmented, current_graph_augmented]
)

previous_embedding_cut = _elbow_cut_if_needed(
elbow_cut, graph.is_directed(), model.singular_values_, previous_embedding
)

current_embedding_cut = _elbow_cut_if_needed(
elbow_cut, graph.is_directed(), model.singular_values_, current_embedding
)

graph_embeddings.append(
(
Embeddings(union_node_ids, previous_embedding_cut),
Embeddings(union_node_ids, current_embedding_cut),
)
)

return graph_embeddings


def _graphs_precondition_checks(graphs, weight_attribute):
is_directed = graphs[0].is_directed()

for graph in graphs:
check_argument(
is_directed == graph.is_directed(),
"graphs must either be all directed or all undirected",
)

check_argument(
not graph.is_multigraph(),
"Multigraphs are not supported; you must determine how to represent at most "
"one edge between any two nodes, and handle the corresponding weights "
"accordingly",
)

if not is_real_weighted(graph, weight_attribute=weight_attribute):
warnings.warn(
f"Graphs with edges that do not have a real numeric weight set for every "
f"{weight_attribute} attribute on every edge are treated as an unweighted "
f"graph - which presumes all weights are `1.0`. If this is incorrect, "
f"please add a '{weight_attribute}' attribute to every edge with a real, "
f"numeric value (e.g. an integer or a float) and call this function again."
)
weight_attribute = None # this supercedes what the user said, because
# not all of the weights are real numbers, if they exist at all
# this weight=1.0 treatment actually happens in nx.to_scipy_sparse_matrix()

return weight_attribute


def _elbow_cut_if_needed(elbow_cut, is_directed, singular_values, embedding):
if elbow_cut is None:
if is_directed:
embedding = np.concatenate(embedding, axis=1)
else:
column_index = _index_of_elbow(singular_values, elbow_cut)

if is_directed:
left, right = embedding
left = left[:, :column_index]
right = right[:, :column_index]
embedding = np.concatenate((left, right), axis=1)
else:
embedding = embedding[:, :column_index]

return embedding


def _augment_graph(graph, node_ids, weight_attribute):
graph_as_array = nx.to_numpy_array(
daxpryce marked this conversation as resolved.
Show resolved Hide resolved
graph, weight=weight_attribute, nodelist=node_ids
)

graphs_loops_removed = remove_loops(graph_as_array)
graphs_ranked = pass_to_ranks(graphs_loops_removed)
graphs_diag_augmented = augment_diagonal(graphs_ranked)

return graphs_diag_augmented


def _sync_nodes(graph_to_reduce, set_of_valid_nodes):
to_remove = []
for n in graph_to_reduce.nodes():
if n not in set_of_valid_nodes:
to_remove.append(n)

graph_to_reduce.remove_nodes_from(to_remove)

for node in set_of_valid_nodes:
if not graph_to_reduce.has_node(node):
graph_to_reduce.add_node(node)
4 changes: 4 additions & 0 deletions tests/pipeline/embed/test_adjacency_spectral_embedding.py
Original file line number Diff line number Diff line change
Expand Up @@ -139,3 +139,7 @@ def test_elbow_cuts(self):
digraph, dimensions=dimensions, elbow_cut=2, svd_seed=1234
)
self.assertTrue(results.embeddings().shape[1] < dimensions * 2)
results = adjacency_spectral_embedding(
digraph, dimensions=dimensions, elbow_cut=None, svd_seed=1234
)
self.assertTrue(results.embeddings().shape[1] == dimensions * 2)
4 changes: 4 additions & 0 deletions tests/pipeline/embed/test_laplacian_spectral_embedding.py
Original file line number Diff line number Diff line change
Expand Up @@ -157,3 +157,7 @@ def test_elbow_cuts(self):
digraph, dimensions=dimensions, elbow_cut=2, svd_seed=1234
)
self.assertTrue(results.embeddings().shape[1] < dimensions * 2)
results = laplacian_spectral_embedding(
digraph, dimensions=dimensions, elbow_cut=None, svd_seed=1234
)
self.assertTrue(results.embeddings().shape[1] == dimensions * 2)