Skip to content

Commit ff91f9a

Browse files
authored
remove to_dict in postprocessing (#1455)
1 parent c3cefcf commit ff91f9a

File tree

3 files changed

+4
-13
lines changed

3 files changed

+4
-13
lines changed

spikeinterface/postprocessing/alignsorting.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ def __init__(self, sorting, unit_peak_shifts):
3939
self.register_recording(sorting._recording)
4040

4141
self._kwargs = {
42-
'sorting': sorting.to_dict(),
42+
'sorting': sorting,
4343
'unit_peak_shifts': unit_peak_shifts
4444
}
4545

spikeinterface/postprocessing/principal_component.py

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -333,13 +333,7 @@ def run_for_all_spikes(self, file_path=None, **job_kwargs):
333333
# and run
334334
func = _all_pc_extractor_chunk
335335
init_func = _init_work_all_pc_extractor
336-
n_jobs = ensure_n_jobs(recording, job_kwargs.get('n_jobs', None))
337-
if n_jobs == 1:
338-
init_args = (recording,)
339-
else:
340-
init_args = (recording.to_dict(),)
341-
init_args = init_args + (all_pcs_args, spike_times, spike_labels, we.nbefore, we.nafter,
342-
unit_channels, pca_model)
336+
init_args = (recording, all_pcs_args, spike_times, spike_labels, we.nbefore, we.nafter, unit_channels, pca_model)
343337
processor = ChunkRecordingExecutor(recording, func, init_func, init_args, job_name='extract PCs', **job_kwargs)
344338
processor.run()
345339

spikeinterface/postprocessing/spike_amplitudes.py

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -68,16 +68,13 @@ def _run(self, **job_kwargs):
6868
func = _spike_amplitudes_chunk
6969
init_func = _init_worker_spike_amplitudes
7070
n_jobs = ensure_n_jobs(recording, job_kwargs.get('n_jobs', None))
71-
if n_jobs == 1:
72-
init_args = (recording, sorting)
73-
else:
71+
if n_jobs != 1:
7472
# TODO: avoid dumping sorting and use spike vector and peak pipeline instead
7573
assert sorting.check_if_dumpable(), (
7674
"The soring object is not dumpable and cannot be processed in parallel. You can use the "
7775
"`sorting.save()` function to make it dumpable"
7876
)
79-
init_args = (recording.to_dict(), sorting.to_dict())
80-
init_args = init_args + (extremum_channels_index, peak_shifts, return_scaled)
77+
init_args = (recording, sorting, extremum_channels_index, peak_shifts, return_scaled)
8178
processor = ChunkRecordingExecutor(recording, func, init_func, init_args,
8279
handle_returns=True, job_name='extract amplitudes', **job_kwargs)
8380
out = processor.run()

0 commit comments

Comments
 (0)