| import os |
| import copy |
| import math |
| import pickle |
| import random |
| import collections |
| import numpy as np |
| import networkx as nx |
| import torch |
| import torch.nn.functional as F |
| from torch_geometric.data import Data |
|
|
| TRAIN_MODE = "train" |
| VAL_MODE = "val" |
| TEST_MODE = "test" |
| |
|
|
| GMN_DATA_TYPE = "gmn" |
| PYG_DATA_TYPE = "pyg" |
|
|
| GraphCollection = collections.namedtuple( |
| 'GraphCollection', |
| ['from_idx', 'to_idx', 'node_features', 'edge_features', 'graph_idx', 'num_graphs'] |
| ) |
|
|
| class SubgraphIsomorphismDataset: |
| def __init__(self, mode, dataset_name, dataset_size, batch_size, data_type, dataset_base_path, experiment, dataset_path_override=None, device=None): |
| |
| assert mode in [TRAIN_MODE, VAL_MODE, TEST_MODE] |
| self.mode = mode |
| self.dataset_name = dataset_name |
| self.dataset_size = dataset_size |
| self.max_node_set_size = {"small": 15, "large": 20}[dataset_size] |
| self.batch_size = batch_size |
| self.data_type = data_type |
| self.dataset_base_path = dataset_base_path |
| self.device = experiment.device if experiment else (device if device else 'cuda:0') |
| self.batch_setting = None |
| self.dataset_path_override = dataset_path_override |
|
|
| self.load_graphs(experiment=experiment) |
| self.preprocess_subgraphs_to_pyG_data() |
| self.build_adjacency_info() |
|
|
| self.max_edge_set_size = max( |
| max([graph.number_of_edges() for graph in self.query_graphs]), |
| max([graph.number_of_edges() for graph in self.corpus_graphs]) |
| ) |
|
|
| def load_graphs(self, experiment): |
| dataset_accessor = lambda file_name: os.path.join( |
| self.dataset_base_path, self.dataset_path_override or f"{self.dataset_size}_dataset", |
| "splits", self.mode, file_name |
| ) |
| |
| |
| pair_count = f"{80 if self.dataset_size == 'small' else 240}k" |
| mode_prefix = "test" if "test" in self.mode else self.mode |
| query_graph_file = dataset_accessor(f"{mode_prefix}_{self.dataset_name}{pair_count}_query_subgraphs.pkl") |
| self.query_graphs = pickle.load(open(query_graph_file, 'rb')) |
| num_query_graphs = len(self.query_graphs) |
| if experiment: |
| experiment.log("loaded %s query graphs from %s", self.mode, query_graph_file) |
|
|
| |
| relationships_file = query_graph_file.replace("query_subgraphs", "rel_nx_is_subgraph_iso") |
| self.relationships = pickle.load(open(relationships_file, 'rb')) |
| if experiment: |
| experiment.log("loaded %s relationships from %s", self.mode, relationships_file) |
|
|
| assert list(self.relationships.keys()) == list(range(num_query_graphs)) |
|
|
| |
| corpus_graph_file = os.path.join( |
| os.path.dirname(os.path.dirname(os.path.dirname(query_graph_file))), "corpus", |
| f"{self.dataset_name}{pair_count}_corpus_subgraphs.pkl" |
| ) |
| self.corpus_graphs = pickle.load(open(corpus_graph_file, 'rb')) |
| if experiment: |
| experiment.log("loaded corpus graphs from %s", corpus_graph_file) |
|
|
| self.pos_pairs, self.neg_pairs = [], [] |
| for query_idx in range(num_query_graphs): |
| for corpus_idx in self.relationships[query_idx]['pos']: |
| self.pos_pairs.append((query_idx, corpus_idx)) |
| for corpus_idx in self.relationships[query_idx]['neg']: |
| self.neg_pairs.append((query_idx, corpus_idx)) |
|
|
| def create_pyG_object(self, graph): |
| num_nodes = graph.number_of_nodes() |
| features = torch.ones(num_nodes, 1, dtype=torch.float, device=self.device) |
|
|
| edges = list(graph.edges) |
| doubled_edges = [[x, y] for (x, y) in edges] + [[y, x] for (x, y) in edges] |
| edge_index = torch.tensor(np.array(doubled_edges).T, dtype=torch.int64, device=self.device) |
| return Data(x = features, edge_index = edge_index), num_nodes |
|
|
| def preprocess_subgraphs_to_pyG_data(self): |
| self.query_graph_data, self.query_graph_sizes = zip( |
| *[self.create_pyG_object(query_graph) for query_graph in self.query_graphs] |
| ) |
| self.corpus_graph_data, self.corpus_graph_sizes = zip( |
| *[self.create_pyG_object(corpus_graph) for corpus_graph in self.corpus_graphs] |
| ) |
| |
| def build_adjacency_info(self): |
| def adj_list_from_graph_list(graphs): |
| adj_list = [] |
| for graph in graphs: |
| unpadded_adj = torch.tensor(nx.adjacency_matrix(graph).todense(), dtype=torch.float, device=self.device) |
| assert unpadded_adj.shape[0] == unpadded_adj.shape[1] |
| num_nodes = len(unpadded_adj) |
| padded_adj = F.pad(unpadded_adj, pad = (0, self.max_node_set_size - num_nodes, 0, self.max_node_set_size - num_nodes)) |
| adj_list.append(padded_adj) |
| return adj_list |
|
|
| self.query_adj_list = adj_list_from_graph_list(self.query_graphs) |
| self.corpus_adj_list = adj_list_from_graph_list(self.corpus_graphs) |
| |
| def _pack_batch(self, graphs): |
| from_idx = [] |
| to_idx = [] |
| graph_idx = [] |
| all_graphs = [individual_graph for graph_tuple in graphs for individual_graph in graph_tuple] |
|
|
| total_nodes, total_edges = 0, 0 |
| for idx, graph in enumerate(all_graphs): |
| num_nodes = graph.number_of_nodes() |
| num_edges = graph.number_of_edges() |
| edges = np.array(graph.edges(), dtype=np.int32) |
| |
| from_idx.append(edges[:, 0] + total_nodes) |
| to_idx.append(edges[:, 1] + total_nodes) |
| graph_idx.append(np.ones(num_nodes, dtype=np.int32) * idx) |
|
|
| total_nodes += num_nodes |
| total_edges += num_edges |
| |
| return GraphCollection( |
| from_idx = torch.tensor(np.concatenate(from_idx, axis=0), dtype=torch.int64, device=self.device), |
| to_idx = torch.tensor(np.concatenate(to_idx, axis=0), dtype=torch.int64, device=self.device), |
| graph_idx = torch.tensor(np.concatenate(graph_idx, axis=0), dtype=torch.int64, device=self.device), |
| num_graphs = len(all_graphs), |
| node_features = torch.ones(total_nodes, 1, dtype=torch.float, device=self.device), |
| edge_features = torch.ones(total_edges, 1, dtype=torch.float, device=self.device) |
| ) |
|
|
| def create_stratified_batches(self): |
| self.batch_setting = 'stratified' |
| random.shuffle(self.pos_pairs), random.shuffle(self.neg_pairs) |
| pos_to_neg_ratio = len(self.pos_pairs) / len(self.neg_pairs) |
|
|
| num_pos_per_batch = math.ceil(pos_to_neg_ratio/(1 + pos_to_neg_ratio) * self.batch_size) |
| num_neg_per_batch = self.batch_size - num_pos_per_batch |
|
|
| batches_pos, batches_neg = [], [] |
| labels_pos, labels_neg = [], [] |
| for idx in range(0, len(self.pos_pairs), num_pos_per_batch): |
| elements_remaining = len(self.pos_pairs) - idx |
| elements_chosen = min(num_pos_per_batch, elements_remaining) |
| batches_pos.append(self.pos_pairs[idx : idx + elements_chosen]) |
| labels_pos.append([1.0] * elements_chosen) |
| for idx in range(0, len(self.neg_pairs), num_neg_per_batch): |
| elements_remaining = len(self.neg_pairs) - idx |
| elements_chosen = min(num_neg_per_batch, elements_remaining) |
| batches_neg.append(self.neg_pairs[idx : idx + elements_chosen]) |
| labels_neg.append([0.0] * elements_chosen) |
|
|
| self.num_batches = min(len(batches_pos), len(batches_neg)) |
| self.batches = [pos + neg for (pos, neg) in zip(batches_pos[:self.num_batches], batches_neg[:self.num_batches])] |
| self.labels = [pos + neg for (pos, neg) in zip(labels_pos[:self.num_batches], labels_neg[:self.num_batches])] |
|
|
| return self.num_batches |
|
|
| def create_custom_batches(self, pair_list): |
| self.batch_setting = 'custom' |
| self.batches = [] |
| for idx in range(0, len(pair_list), self.batch_size): |
| self.batches.append(pair_list[idx : idx + self.batch_size]) |
| |
| self.num_batches = len(self.batches) |
| return self.num_batches |
|
|
| def fetch_batch_by_id(self, idx): |
| assert idx < self.num_batches |
| batch = self.batches[idx] |
|
|
| query_graph_idxs, corpus_graph_idxs = zip(*batch) |
| |
| if self.data_type == GMN_DATA_TYPE: |
| query_graphs = [self.query_graphs[idx] for idx in query_graph_idxs] |
| corpus_graphs = [self.corpus_graphs[idx] for idx in corpus_graph_idxs] |
| all_graphs = self._pack_batch(zip(query_graphs, corpus_graphs)) |
| elif self.data_type == PYG_DATA_TYPE: |
| query_graphs = [self.query_graph_data[idx] for idx in query_graph_idxs] |
| corpus_graphs = [self.corpus_graph_data[idx] for idx in corpus_graph_idxs] |
| all_graphs = list(zip(query_graphs, corpus_graphs)) |
|
|
| query_graph_sizes = [self.query_graph_sizes[idx] for idx in query_graph_idxs] |
| corpus_graph_sizes = [self.corpus_graph_sizes[idx] for idx in corpus_graph_idxs] |
| all_graph_sizes = list(zip(query_graph_sizes, corpus_graph_sizes)) |
|
|
| query_graph_adjs = [self.query_adj_list[idx] for idx in query_graph_idxs] |
| corpus_graph_adjs = [self.corpus_adj_list[idx] for idx in corpus_graph_idxs] |
| all_graph_adjs = list(zip(query_graph_adjs, corpus_graph_adjs)) |
|
|
| if self.batch_setting == 'stratified': |
| target = torch.tensor(np.array(self.labels[idx]), dtype=torch.float, device=self.device) |
| return all_graphs, all_graph_sizes, target, all_graph_adjs |
| elif self.batch_setting == 'custom': |
| return all_graphs, all_graph_sizes, None, all_graph_adjs |
| else: |
| raise NotImplementedError |
|
|
| def get_datasets(dataset_config, experiment, data_type, modes=['train', 'val']): |
| return { |
| mode: SubgraphIsomorphismDataset( |
| mode = mode, experiment = experiment, |
| data_type = data_type, **dataset_config |
| ) for mode in modes |
| } |