| import numpy as np |
| import itertools as itools |
| from ridge_utils.interpdata import sincinterp2D, gabor_xfm2D, lanczosinterp2D |
|
|
| class DataSequence(object): |
| """DataSequence class provides a nice interface for handling data that is both continuous |
| and discretely chunked. For example, semantic projections of speech stimuli must be |
| considered both at the level of single words (which are continuous throughout the stimulus) |
| and at the level of TRs (which contain discrete chunks of words). |
| """ |
| def __init__(self, data, split_inds, data_times=None, tr_times=None): |
| """Initializes the DataSequence with the given [data] object (which can be any iterable) |
| and a collection of [split_inds], which should be the indices where the data is split into |
| separate TR chunks. |
| """ |
| self.data = data |
| self.split_inds = split_inds |
| self.data_times = data_times |
| self.tr_times = tr_times |
|
|
| def mapdata(self, fun): |
| """Creates a new DataSequence where each element of [data] is produced by mapping the |
| function [fun] onto this DataSequence's [data]. |
| |
| The [split_inds] are preserved exactly. |
| """ |
| return DataSequence(self, list(map(fun, self.data)), self.split_inds) |
|
|
| def chunks(self): |
| """Splits the stored [data] into the discrete chunks and returns them. |
| """ |
| return np.split(self.data, self.split_inds) |
|
|
| def data_to_chunk_ind(self, dataind): |
| """Returns the index of the chunk containing the data with the given index. |
| """ |
| zc = np.zeros((len(self.data),)) |
| zc[dataind] = 1.0 |
| ch = np.array([ch.sum() for ch in np.split(zc, self.split_inds)]) |
| return np.nonzero(ch)[0][0] |
|
|
| def chunk_to_data_ind(self, chunkind): |
| """Returns the indexes of the data contained in the chunk with the given index. |
| """ |
| return list(np.split(np.arange(len(self.data)), self.split_inds)[chunkind]) |
|
|
| def chunkmeans(self): |
| """Splits the stored [data] into the discrete chunks, then takes the mean of each chunk |
| (this is assuming that [data] is a numpy array) and returns the resulting matrix with |
| one row per chunk. |
| """ |
| dsize = self.data.shape[1] |
| outmat = np.zeros((len(self.split_inds)+1, dsize)) |
| for ci, c in enumerate(self.chunks()): |
| if len(c): |
| outmat[ci] = np.vstack(c).mean(0) |
|
|
| return outmat |
|
|
| def chunksums(self, interp="rect", **kwargs): |
| """Splits the stored [data] into the discrete chunks, then takes the sum of each chunk |
| (this is assuming that [data] is a numpy array) and returns the resulting matrix with |
| one row per chunk. |
| |
| If [interp] is "sinc", the signal will be downsampled using a truncated sinc filter |
| instead of a rectangular filter. |
| |
| if [interp] is "lanczos", the signal will be downsampled using a Lanczos filter. |
| |
| [kwargs] are passed to the interpolation function. |
| """ |
| if interp=="sinc": |
| |
| return sincinterp2D(self.data, self.data_times, self.tr_times, **kwargs) |
| elif interp=="lanczos": |
| |
| return lanczosinterp2D(self.data, self.data_times, self.tr_times, **kwargs) |
| elif interp=="gabor": |
| |
| return np.abs(gabor_xfm2D(self.data.T, self.data_times, self.tr_times, **kwargs)).T |
| else: |
| dsize = self.data.shape[1] |
| outmat = np.zeros((len(self.split_inds)+1, dsize)) |
| for ci, c in enumerate(self.chunks()): |
| if len(c): |
| outmat[ci] = np.vstack(c).sum(0) |
| |
| return outmat |
|
|
| def copy(self): |
| """Returns a copy of this DataSequence. |
| """ |
| return DataSequence(list(self.data), self.split_inds.copy(), self.data_times, self.tr_times) |
| |
| @classmethod |
| def from_grid(cls, grid_transcript, trfile): |
| """Creates a new DataSequence from a [grid_transript] and a [trfile]. |
| grid_transcript should be the product of the 'make_simple_transcript' method of TextGrid. |
| """ |
| data_entries = list(zip(*grid_transcript))[2] |
| if isinstance(data_entries[0], str): |
| data = list(map(str.lower, list(zip(*grid_transcript))[2])) |
| else: |
| data = data_entries |
| word_starts = np.array(list(map(float, list(zip(*grid_transcript))[0]))) |
| word_ends = np.array(list(map(float, list(zip(*grid_transcript))[1]))) |
| word_avgtimes = (word_starts + word_ends)/2.0 |
| |
| tr = trfile.avgtr |
| trtimes = trfile.get_reltriggertimes() |
| |
| split_inds = [(word_starts<(t+tr)).sum() for t in trtimes][:-1] |
| return cls(data, split_inds, word_avgtimes, trtimes+tr/2.0) |
|
|
| @classmethod |
| def from_chunks(cls, chunks): |
| """The inverse operation of DataSequence.chunks(), this function concatenates |
| the [chunks] and infers split_inds. |
| """ |
| lens = list(map(len, chunks)) |
| split_inds = np.cumsum(lens)[:-1] |
| |
| data = list(itools.chain(*map(list, chunks))) |
| return cls(data, split_inds) |
|
|