""" test_distance.py Tests the isi- and spike-distance computation Copyright 2014, Mario Mulansky Distributed under the BSD License """ from __future__ import print_function import numpy as np from copy import copy from numpy.testing import assert_equal, assert_almost_equal, \ assert_array_almost_equal import pyspike as spk from pyspike import SpikeTrain import os TEST_PATH = os.path.dirname(os.path.realpath(__file__)) def test_isi(): # generate two spike trains: t1 = SpikeTrain([0.2, 0.4, 0.6, 0.7], 1.0) t2 = SpikeTrain([0.3, 0.45, 0.8, 0.9, 0.95], 1.0) # pen&paper calculation of the isi distance expected_times = [0.0, 0.2, 0.3, 0.4, 0.45, 0.6, 0.7, 0.8, 0.9, 0.95, 1.0] expected_isi = [0.1/0.3, 0.1/0.3, 0.05/0.2, 0.05/0.2, 0.15/0.35, 0.25/0.35, 0.05/0.35, 0.2/0.3, 0.25/0.3, 0.25/0.3] expected_times = np.array(expected_times) expected_isi = np.array(expected_isi) expected_isi_val = sum((expected_times[1:] - expected_times[:-1]) * expected_isi)/(expected_times[-1]-expected_times[0]) f = spk.isi_profile(t1, t2) # print("ISI: ", f.y) print("ISI value:", expected_isi_val) assert_equal(f.x, expected_times) assert_array_almost_equal(f.y, expected_isi, decimal=15) assert_equal(f.avrg(), expected_isi_val) assert_equal(spk.isi_distance(t1, t2), expected_isi_val) # check with some equal spike times t1 = SpikeTrain([0.2, 0.4, 0.6], [0.0, 1.0]) t2 = SpikeTrain([0.1, 0.4, 0.5, 0.6], [0.0, 1.0]) expected_times = [0.0, 0.1, 0.2, 0.4, 0.5, 0.6, 1.0] expected_isi = [0.1/0.3, 0.1/0.3, 0.1/0.3, 0.1/0.2, 0.1/0.2, 0.0/0.5] expected_times = np.array(expected_times) expected_isi = np.array(expected_isi) expected_isi_val = sum((expected_times[1:] - expected_times[:-1]) * expected_isi)/(expected_times[-1]-expected_times[0]) f = spk.isi_profile(t1, t2) assert_equal(f.x, expected_times) assert_array_almost_equal(f.y, expected_isi, decimal=15) assert_equal(f.avrg(), expected_isi_val) assert_equal(spk.isi_distance(t1, t2), expected_isi_val) def test_spike(): # generate two spike trains: t1 = SpikeTrain([0.0, 2.0, 5.0, 8.0], 10.0) t2 = SpikeTrain([0.0, 1.0, 5.0, 9.0], 10.0) expected_times = np.array([0.0, 1.0, 2.0, 5.0, 8.0, 9.0, 10.0]) f = spk.spike_profile(t1, t2) assert_equal(f.x, expected_times) # from SPIKY: y_all = np.array([0.000000000000000000, 0.555555555555555580, 0.222222222222222210, 0.305555555555555580, 0.255102040816326536, 0.000000000000000000, 0.000000000000000000, 0.255102040816326536, 0.255102040816326536, 0.285714285714285698, 0.285714285714285698, 0.285714285714285698]) #assert_array_almost_equal(f.y1, y_all[::2]) assert_array_almost_equal(f.y2, y_all[1::2]) assert_almost_equal(f.avrg(), 0.186309523809523814, decimal=15) assert_equal(spk.spike_distance(t1, t2), f.avrg()) t1 = SpikeTrain([0.2, 0.4, 0.6, 0.7], 1.0) t2 = SpikeTrain([0.3, 0.45, 0.8, 0.9, 0.95], 1.0) # pen&paper calculation of the spike distance expected_times = [0.0, 0.2, 0.3, 0.4, 0.45, 0.6, 0.7, 0.8, 0.9, 0.95, 1.0] s1 = np.array([0.1, 0.1, (0.1*0.1+0.05*0.1)/0.2, 0.05, (0.05*0.15 * 2)/0.2, 0.15, 0.1, (0.1*0.1+0.1*0.2)/0.3, (0.1*0.2+0.1*0.1)/0.3, (0.1*0.05+0.1*0.25)/0.3, 0.1]) s2 = np.array([0.1, (0.1*0.2+0.1*0.1)/0.3, 0.1, (0.1*0.05 * 2)/.15, 0.05, (0.05*0.2+0.1*0.15)/0.35, (0.05*0.1+0.1*0.25)/0.35, 0.1, 0.1, 0.05, 0.05]) isi1 = np.array([0.2, 0.2, 0.2, 0.2, 0.2, 0.1, 0.3, 0.3, 0.3, 0.3]) isi2 = np.array([0.3, 0.3, 0.15, 0.15, 0.35, 0.35, 0.35, 0.1, 0.05, 0.05]) expected_y1 = (s1[:-1]*isi2+s2[:-1]*isi1) / (0.5*(isi1+isi2)**2) expected_y2 = (s1[1:]*isi2+s2[1:]*isi1) / (0.5*(isi1+isi2)**2) expected_times = np.array(expected_times) expected_y1 = np.array(expected_y1) expected_y2 = np.array(expected_y2) expected_spike_val = sum((expected_times[1:] - expected_times[:-1]) * (expected_y1+expected_y2)/2) expected_spike_val /= (expected_times[-1]-expected_times[0]) print("SPIKE value:", expected_spike_val) f = spk.spike_profile(t1, t2) assert_equal(f.x, expected_times) assert_array_almost_equal(f.y1, expected_y1, decimal=15) assert_array_almost_equal(f.y2, expected_y2, decimal=15) assert_almost_equal(f.avrg(), expected_spike_val, decimal=15) assert_almost_equal(spk.spike_distance(t1, t2), expected_spike_val, decimal=15) # check with some equal spike times t1 = SpikeTrain([0.2, 0.4, 0.6], [0.0, 1.0]) t2 = SpikeTrain([0.1, 0.4, 0.5, 0.6], [0.0, 1.0]) expected_times = [0.0, 0.1, 0.2, 0.4, 0.5, 0.6, 1.0] # due to the edge correction in the beginning, s1 and s2 are different # for left and right values s1_r = np.array([0.1, (0.1*0.1+0.1*0.1)/0.2, 0.1, 0.0, 0.0, 0.0, 0.0]) s1_l = np.array([0.1, (0.1*0.1+0.1*0.1)/0.2, 0.1, 0.0, 0.0, 0.0, 0.0]) # s2_r = np.array([0.1*0.1/0.3, 0.1*0.3/0.3, 0.1*0.2/0.3, # 0.0, 0.1, 0.0, 0.0]) # s2_l = np.array([0.1*0.1/0.3, 0.1*0.1/0.3, 0.1*0.2/0.3, 0.0, # 0.1, 0.0, 0.0]) # eero's edge correction: s2_r = np.array([0.1, 0.1*0.3/0.3, 0.1*0.2/0.3, 0.0, 0.1, 0.0, 0.0]) s2_l = np.array([0.1, 0.1*0.3/0.3, 0.1*0.2/0.3, 0.0, 0.1, 0.0, 0.0]) isi1 = np.array([0.2, 0.2, 0.2, 0.2, 0.2, 0.4]) isi2 = np.array([0.3, 0.3, 0.3, 0.1, 0.1, 0.4]) expected_y1 = (s1_r[:-1]*isi2+s2_r[:-1]*isi1) / (0.5*(isi1+isi2)**2) expected_y2 = (s1_l[1:]*isi2+s2_l[1:]*isi1) / (0.5*(isi1+isi2)**2) expected_times = np.array(expected_times) expected_y1 = np.array(expected_y1) expected_y2 = np.array(expected_y2) expected_spike_val = sum((expected_times[1:] - expected_times[:-1]) * (expected_y1+expected_y2)/2) expected_spike_val /= (expected_times[-1]-expected_times[0]) f = spk.spike_profile(t1, t2) assert_equal(f.x, expected_times) assert_array_almost_equal(f.y1, expected_y1, decimal=14) assert_array_almost_equal(f.y2, expected_y2, decimal=14) assert_almost_equal(f.avrg(), expected_spike_val, decimal=16) assert_almost_equal(spk.spike_distance(t1, t2), expected_spike_val, decimal=16) def test_spike_sync(): spikes1 = SpikeTrain([1.0, 2.0, 3.0], 4.0) spikes2 = SpikeTrain([2.1], 4.0) expected_x = np.array([0.0, 1.0, 2.0, 2.1, 3.0, 4.0]) expected_y = np.array([0.0, 0.0, 1.0, 1.0, 0.0, 0.0]) f = spk.spike_sync_profile(spikes1, spikes2) assert_array_almost_equal(f.x, expected_x, decimal=16) assert_array_almost_equal(f.y, expected_y, decimal=16) assert_almost_equal(spk.spike_sync(spikes1, spikes2), 0.5, decimal=16) # test with some small max_tau, spike_sync should be 0 assert_almost_equal(spk.spike_sync(spikes1, spikes2, max_tau=0.05), 0.0, decimal=16) spikes2 = SpikeTrain([3.1], 4.0) assert_almost_equal(spk.spike_sync(spikes1, spikes2), 0.5, decimal=16) spikes2 = SpikeTrain([1.1], 4.0) expected_x = np.array([0.0, 1.0, 1.1, 2.0, 3.0, 4.0]) expected_y = np.array([1.0, 1.0, 1.0, 0.0, 0.0, 0.0]) f = spk.spike_sync_profile(spikes1, spikes2) assert_array_almost_equal(f.x, expected_x, decimal=16) assert_array_almost_equal(f.y, expected_y, decimal=16) assert_almost_equal(spk.spike_sync(spikes1, spikes2), 0.5, decimal=16) spikes2 = SpikeTrain([0.9], 4.0) assert_almost_equal(spk.spike_sync(spikes1, spikes2), 0.5, decimal=16) spikes2 = SpikeTrain([3.0], 4.0) assert_almost_equal(spk.spike_sync(spikes1, spikes2), 0.5, decimal=16) spikes2 = SpikeTrain([1.0], 4.0) assert_almost_equal(spk.spike_sync(spikes1, spikes2), 0.5, decimal=16) spikes2 = SpikeTrain([1.5, 3.0], 4.0) assert_almost_equal(spk.spike_sync(spikes1, spikes2), 0.4, decimal=16) def check_multi_profile(profile_func, profile_func_multi, dist_func_multi): # generate spike trains: t1 = SpikeTrain([0.2, 0.4, 0.6, 0.7], 1.0) t2 = SpikeTrain([0.3, 0.45, 0.8, 0.9, 0.95], 1.0) t3 = SpikeTrain([0.2, 0.4, 0.6], 1.0) t4 = SpikeTrain([0.1, 0.4, 0.5, 0.6], 1.0) spike_trains = [t1, t2, t3, t4] f12 = profile_func(t1, t2) f13 = profile_func(t1, t3) f14 = profile_func(t1, t4) f23 = profile_func(t2, t3) f24 = profile_func(t2, t4) f34 = profile_func(t3, t4) f_multi = profile_func_multi(spike_trains, [0, 1]) assert f_multi.almost_equal(f12, decimal=14) d = dist_func_multi(spike_trains, [0, 1]) assert_equal(f_multi.avrg(), d) f_multi1 = profile_func_multi(spike_trains, [1, 2, 3]) f_multi2 = profile_func_multi(spike_trains[1:]) assert f_multi1.almost_equal(f_multi2, decimal=14) d = dist_func_multi(spike_trains, [1, 2, 3]) assert_almost_equal(f_multi1.avrg(), d, decimal=14) f = copy(f12) f.add(f13) f.add(f23) f.mul_scalar(1.0/3) f_multi = profile_func_multi(spike_trains, [0, 1, 2]) assert f_multi.almost_equal(f, decimal=14) d = dist_func_multi(spike_trains, [0, 1, 2]) assert_almost_equal(f_multi.avrg(), d, decimal=14) f.mul_scalar(3) # revert above normalization f.add(f14) f.add(f24) f.add(f34) f.mul_scalar(1.0/6) f_multi = profile_func_multi(spike_trains) assert f_multi.almost_equal(f, decimal=14) def test_multi_isi(): check_multi_profile(spk.isi_profile, spk.isi_profile_multi, spk.isi_distance_multi) def test_multi_spike(): check_multi_profile(spk.spike_profile, spk.spike_profile_multi, spk.spike_distance_multi) def test_multi_spike_sync(): # some basic multivariate check spikes1 = SpikeTrain([100, 300, 400, 405, 410, 500, 700, 800, 805, 810, 815, 900], 1000) spikes2 = SpikeTrain([100, 200, 205, 210, 295, 350, 400, 510, 600, 605, 700, 910], 1000) spikes3 = SpikeTrain([100, 180, 198, 295, 412, 420, 510, 640, 695, 795, 820, 920], 1000) assert_almost_equal(spk.spike_sync(spikes1, spikes2), 0.5, decimal=15) assert_almost_equal(spk.spike_sync(spikes1, spikes3), 0.5, decimal=15) assert_almost_equal(spk.spike_sync(spikes2, spikes3), 0.5, decimal=15) f = spk.spike_sync_profile_multi([spikes1, spikes2, spikes3]) # hands on definition of the average multivariate spike synchronization # expected = (f1.integral() + f2.integral() + f3.integral()) / \ # (np.sum(f1.mp[1:-1])+np.sum(f2.mp[1:-1])+np.sum(f3.mp[1:-1])) expected = 0.5 assert_almost_equal(f.avrg(), expected, decimal=15) assert_almost_equal(spk.spike_sync_multi([spikes1, spikes2, spikes3]), expected, decimal=15) # multivariate regression test spike_trains = spk.load_spike_trains_from_txt( os.path.join(TEST_PATH, "SPIKE_Sync_Test.txt"), edges=[0, 4000]) # extract all spike times spike_times = np.array([]) for st in spike_trains: spike_times = np.append(spike_times, st.spikes) spike_times = np.unique(np.sort(spike_times)) f = spk.spike_sync_profile_multi(spike_trains) assert_equal(spike_times, f.x[1:-1]) assert_equal(len(f.x), len(f.y)) assert_equal(np.sum(f.y[1:-1]), 39932) assert_equal(np.sum(f.mp[1:-1]), 85554) def check_dist_matrix(dist_func, dist_matrix_func): # generate spike trains: t1 = SpikeTrain([0.2, 0.4, 0.6, 0.7], 1.0) t2 = SpikeTrain([0.3, 0.45, 0.8, 0.9, 0.95], 1.0) t3 = SpikeTrain([0.2, 0.4, 0.6], 1.0) t4 = SpikeTrain([0.1, 0.4, 0.5, 0.6], 1.0) spike_trains = [t1, t2, t3, t4] f12 = dist_func(t1, t2) f13 = dist_func(t1, t3) f14 = dist_func(t1, t4) f23 = dist_func(t2, t3) f24 = dist_func(t2, t4) f34 = dist_func(t3, t4) f_matrix = dist_matrix_func(spike_trains) # check zero diagonal for i in range(4): assert_equal(0.0, f_matrix[i, i]) for i in range(4): for j in range(i+1, 4): assert_equal(f_matrix[i, j], f_matrix[j, i]) assert_equal(f12, f_matrix[1, 0]) assert_equal(f13, f_matrix[2, 0]) assert_equal(f14, f_matrix[3, 0]) assert_equal(f23, f_matrix[2, 1]) assert_equal(f24, f_matrix[3, 1]) assert_equal(f34, f_matrix[3, 2]) def test_isi_matrix(): check_dist_matrix(spk.isi_distance, spk.isi_distance_matrix) def test_spike_matrix(): check_dist_matrix(spk.spike_distance, spk.spike_distance_matrix) def test_spike_sync_matrix(): check_dist_matrix(spk.spike_sync, spk.spike_sync_matrix) def test_regression_spiky(): # standard example st1 = SpikeTrain(np.arange(100, 1201, 100), 1300) st2 = SpikeTrain(np.arange(100, 1201, 110), 1300) isi_dist = spk.isi_distance(st1, st2) assert_almost_equal(isi_dist, 9.0909090909090939e-02, decimal=15) isi_profile = spk.isi_profile(st1, st2) assert_equal(isi_profile.y, 0.1/1.1 * np.ones_like(isi_profile.y)) spike_dist = spk.spike_distance(st1, st2) assert_equal(spike_dist, 0.211058782487353908) spike_sync = spk.spike_sync(st1, st2) assert_equal(spike_sync, 8.6956521739130432e-01) # multivariate check spike_trains = spk.load_spike_trains_from_txt( os.path.join(TEST_PATH, "PySpike_testdata.txt"), (0.0, 4000.0)) isi_dist = spk.isi_distance_multi(spike_trains) # get the full precision from SPIKY assert_almost_equal(isi_dist, 0.17051816816999129656, decimal=15) spike_profile = spk.spike_profile_multi(spike_trains) assert_equal(len(spike_profile.y1)+len(spike_profile.y2), 1252) spike_dist = spk.spike_distance_multi(spike_trains) # get the full precision from SPIKY assert_almost_equal(spike_dist, 0.25188056475463755, decimal=15) spike_sync = spk.spike_sync_multi(spike_trains) # get the full precision from SPIKY assert_equal(spike_sync, 0.7183531505298066) # Eero's edge correction example st1 = SpikeTrain([0.5, 1.5, 2.5], 6.0) st2 = SpikeTrain([3.5, 4.5, 5.5], 6.0) f = spk.spike_profile(st1, st2) expected_times = np.array([0.0, 0.5, 1.5, 2.5, 3.5, 4.5, 5.5, 6.0]) y_all = np.array([0.271604938271605, 0.271604938271605, 0.271604938271605, 0.617283950617284, 0.617283950617284, 0.444444444444444, 0.285714285714286, 0.285714285714286, 0.444444444444444, 0.617283950617284, 0.617283950617284, 0.271604938271605, 0.271604938271605, 0.271604938271605]) expected_y1 = y_all[::2] expected_y2 = y_all[1::2] assert_equal(f.x, expected_times) assert_array_almost_equal(f.y1, expected_y1, decimal=14) assert_array_almost_equal(f.y2, expected_y2, decimal=14) def test_multi_variate_subsets(): spike_trains = spk.load_spike_trains_from_txt( os.path.join(TEST_PATH, "PySpike_testdata.txt"), (0.0, 4000.0)) sub_set = [1, 3, 5, 7] spike_trains_sub_set = [spike_trains[i] for i in sub_set] v1 = spk.isi_distance_multi(spike_trains_sub_set) v2 = spk.isi_distance_multi(spike_trains, sub_set) assert_equal(v1, v2) v1 = spk.spike_distance_multi(spike_trains_sub_set) v2 = spk.spike_distance_multi(spike_trains, sub_set) assert_equal(v1, v2) v1 = spk.spike_sync_multi(spike_trains_sub_set) v2 = spk.spike_sync_multi(spike_trains, sub_set) assert_equal(v1, v2) if __name__ == "__main__": test_isi() test_spike() test_spike_sync() test_multi_isi() test_multi_spike() test_multi_spike_sync() test_isi_matrix() test_spike_matrix() test_spike_sync_matrix() test_regression_spiky() test_multi_variate_subsets()