Scheduled service maintenance on November 22


On Friday, November 22, 2024, between 06:00 CET and 18:00 CET, GIN services will undergo planned maintenance. Extended service interruptions should be expected. We will try to keep downtimes to a minimum, but recommend that users avoid critical tasks, large data uploads, or DOI requests during this time.

We apologize for any inconvenience.

1
1

example_blackrock.py 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290
  1. # -*- coding: utf-8 -*-
  2. """
  3. Example code for loading and processing of a recording of the reach-
  4. to-grasp experiments conducted at the Institute de Neurosciences de la Timone
  5. by Thomas Brochier and Alexa Riehle from the original Blackrock files and
  6. odML files, using the custom loading routine `ReachGraspIO`.
  7. Authors: Julia Sprenger, Lyuba Zehl, Michael Denker
  8. Copyright (c) 2017, Institute of Neuroscience and Medicine (INM-6),
  9. Forschungszentrum Juelich, Germany
  10. All rights reserved.
  11. Redistribution and use in source and binary forms, with or without
  12. modification, are permitted provided that the following conditions are met:
  13. * Redistributions of source code must retain the above copyright notice, this
  14. list of conditions and the following disclaimer.
  15. * Redistributions in binary form must reproduce the above copyright notice,
  16. this list of conditions and the following disclaimer in the documentation
  17. and/or other materials provided with the distribution.
  18. * Neither the names of the copyright holders nor the names of the contributors
  19. may be used to endorse or promote products derived from this software without
  20. specific prior written permission.
  21. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
  22. ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
  23. WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
  24. DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
  25. FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
  26. DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
  27. SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
  28. CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
  29. OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
  30. OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
  31. """
  32. import os
  33. import numpy as np
  34. import matplotlib.pyplot as plt
  35. import quantities as pq
  36. from neo import Block, Segment
  37. from elephant.signal_processing import butter
  38. from reachgraspio import reachgraspio
  39. from neo.utils import cut_segment_by_epoch, add_epoch, get_events
  40. from neo_utils import load_segment
  41. # =============================================================================
  42. # Load data
  43. #
  44. # As a first step, we partially load the data file into memory as a Neo object.
  45. # =============================================================================
  46. # Specify the path to the recording session to load, eg,
  47. # '/home/user/l101210-001'
  48. session_name = "i140703-001"
  49. # session_name = "l101210-001"
  50. session_path = os.path.join('..', 'datasets_blackrock', session_name)
  51. odml_dir = os.path.join('..', 'datasets_blackrock')
  52. # Open the session for reading
  53. session = reachgraspio.ReachGraspIO(session_path, odml_directory=odml_dir)
  54. # Read a the complete dataset in lazy mode generating all neo objects,
  55. # but not loading data into memory. The lazy neo structure will contain objects
  56. # to capture all recorded data types (time series at 1000Hz (ns2) and 30kHz (ns6)
  57. # scaled to units of voltage, sorted spike trains, spike waveforms and events)
  58. # of the recording session and return it as a Neo Block. The
  59. # time shift of the ns2 signal (LFP) induced by the online filter is
  60. # automatically corrected for by a heuristic factor stored in the metadata
  61. # (correct_filter_shifts=True).
  62. block = session.read_block(lazy=True, correct_filter_shifts=True)
  63. # Validate there is only a single Segment present in the block
  64. assert len(block.segments) == 1
  65. # loading data content of all data objects during the first 300 seconds
  66. data_segment = load_segment(block.segments[0], time_range=(None, 300*pq.s))
  67. # =============================================================================
  68. # Create offline filtered LFP
  69. #
  70. # Here, we construct one offline filtered LFP from each ns5 (monkey L) or ns6
  71. # (monkey N) raw recording trace. For monkey N, this filtered LFP can be
  72. # compared to the LFPs in the ns2 file (note that monkey L contains only
  73. # behavioral signals in the ns2 file). Also, we assign telling names to each
  74. # Neo AnalogSignal, which is used for plotting later on in this script.
  75. # =============================================================================
  76. target_channel_id = 62
  77. nsx_to_anasig_name = {2: 'LFP signal (online filtered)',
  78. 5: 'raw signal',
  79. 6: 'raw signal'}
  80. filtered_anasig = None
  81. raw_anasig = None
  82. # identify neuronal signals and provide labels for plotting
  83. for anasig in data_segment.analogsignals:
  84. # skip non-neuronal signals
  85. if not anasig.annotations['neural_signal']:
  86. continue
  87. # identify nsx source of signals in this AnalogSignal object
  88. if 'nsx' in anasig.annotations:
  89. nsx = anasig.annotations['nsx']
  90. else:
  91. nsx = np.unique(anasig.array_annotations['nsx'])
  92. assert len(nsx) == 1, 'Different nsx sources in AnalogSignal'
  93. nsx = nsx[0]
  94. if nsx == 2:
  95. # AnalogSignal is LFP from ns2
  96. anasig.name = f'LFP (online filter, ns2)'
  97. filtered_anasig = anasig
  98. elif nsx in [5, 6]:
  99. # AnalogSignal is raw signal from ns5 or ns6
  100. anasig.name = f'raw (ns{nsx})'
  101. raw_anasig = anasig
  102. # Create LFP signal by filtering raw signal if not present already
  103. if filtered_anasig is None:
  104. # Use the Elephant library to filter the signal, filter only target channel
  105. channel_ids = np.asarray(raw_anasig.array_annotations['channel_ids'], dtype=int)
  106. target_channel_index = np.where(target_channel_id == channel_ids)[0]
  107. raw_signal = raw_anasig[:, target_channel_index]
  108. f_anasig = butter(
  109. raw_signal,
  110. highpass_freq=None,
  111. lowpass_freq=250 * pq.Hz,
  112. order=4)
  113. if 'nsx' in anasig.annotations:
  114. nsx = anasig.annotations['nsx']
  115. else:
  116. nsx = anasig.array_annotations["nsx"][0]
  117. f_anasig.name = f'LFP (offline filtered ns{nsx})'
  118. # Attach all offline filtered LFPs to the segment of data
  119. data_segment.analogsignals.append(f_anasig)
  120. # =============================================================================
  121. # Construct analysis epochs
  122. #
  123. # In this step we extract and cut the data into time segments (termed analysis
  124. # epochs) that we wish to analyze. We contrast these analysis epochs to the
  125. # behavioral trials that are defined by the experiment as occurrence of a Trial
  126. # Start (TS-ON) event in the experiment. Concretely, here our analysis epochs
  127. # are constructed as a cutout of 25ms of data around the TS-ON event of all
  128. # successful behavioral trials.
  129. # =============================================================================
  130. # Get Trial Start (TS-ON) events of all successful behavioral trials
  131. # (corresponds to performance code 255, which is accessed for convenience and
  132. # better legibility in the dictionary attribute performance_codes of the
  133. # ReachGraspIO class).
  134. #
  135. # To this end, we filter all event objects of the loaded data to match the name
  136. # "TrialEvents", which is the Event object containing all Events available (see
  137. # documentation of ReachGraspIO). From this Event object we extract only events
  138. # matching "TS-ON" and the desired trial performance code (which are
  139. # annotations of the Event object).
  140. start_events = get_events(
  141. data_segment,
  142. name='TrialEvents',
  143. trial_event_labels='TS-ON',
  144. performance_in_trial=session.performance_codes['correct_trial'])
  145. print('got start events.')
  146. # Extract single Neo Event object containing all TS-ON triggers
  147. assert len(start_events) == 1
  148. start_event = start_events[0]
  149. # Construct analysis epochs from 10ms before the TS-ON of a successful
  150. # behavioral trial to 15ms after TS-ON. The name "analysis_epochs" is given to
  151. # the resulting Neo Epoch object. The object is not attached to the Neo
  152. # Segment. The parameter event2 of add_epoch() is left empty, since we are
  153. # cutting around a single event, as opposed to cutting between two events.
  154. pre = -10 * pq.ms
  155. post = 15 * pq.ms
  156. epoch = add_epoch(
  157. data_segment,
  158. event1=start_event, event2=None,
  159. pre=pre, post=post,
  160. attach_result=False,
  161. name='analysis_epochs',
  162. array_annotations=start_event.array_annotations)
  163. print('added epoch.')
  164. # Create new segments of data cut according to the analysis epochs of the
  165. # 'analysis_epochs' Neo Epoch object. The time axes of all segments are aligned
  166. # such that each segment starts at time 0 (parameter reset_times); annotations
  167. # describing the analysis epoch are carried over to the segments. A new Neo
  168. # Block named "data_cut_to_analysis_epochs" is created to capture all cut
  169. # analysis epochs. For execution time reason, we are only considering the
  170. # first 10 epochs here.
  171. cut_trial_block = Block(name="data_cut_to_analysis_epochs")
  172. cut_trial_block.segments = cut_segment_by_epoch(
  173. data_segment, epoch[:10], reset_time=True)
  174. # =============================================================================
  175. # Plot data
  176. # =============================================================================
  177. # Determine the first existing trial ID i from the Event object containing all
  178. # start events. Then, by calling the filter() function of the Neo Block
  179. # "data_cut_to_analysis_epochs" containing the data cut into the analysis
  180. # epochs, we ask to return all Segments annotated by the behavioral trial ID i.
  181. # In this case this call should return one matching analysis epoch around TS-ON
  182. # belonging to behavioral trial ID i. For monkey N, this is trial ID 1, for
  183. # monkey L this is trial ID 2 since trial ID 1 is not a correct trial.
  184. trial_id = int(np.min(start_event.array_annotations['trial_id']))
  185. trial_segments = cut_trial_block.filter(
  186. targdict={"trial_id": trial_id}, objects=Segment)
  187. assert len(trial_segments) == 1
  188. trial_segment = trial_segments[0]
  189. # Create figure
  190. fig = plt.figure(facecolor='w')
  191. time_unit = pq.CompoundUnit('1./30000*s')
  192. amplitude_unit = pq.microvolt
  193. nsx_colors = {2: 'k', 5: 'r', 6: 'b'}
  194. # Loop through all AnalogSignal objects and plot the signal of the target channel
  195. # in a color corresponding to its sampling frequency (i.e., originating from the ns2/ns5 or ns2/ns6).
  196. for i, anasig in enumerate(trial_segment.analogsignals):
  197. # only visualize neural data
  198. if anasig.annotations['neural_signal']:
  199. if 'nsx' in anasig.annotations:
  200. nsx = anasig.annotations['nsx']
  201. else:
  202. nsx = anasig.array_annotations['nsx'][0]
  203. channel_ids = np.asarray(anasig.array_annotations['channel_ids'], dtype=int)
  204. target_channel_index = np.where(channel_ids == target_channel_id)[0]
  205. target_signal = anasig[:, target_channel_index]
  206. plt.plot(
  207. target_signal.times.rescale(time_unit),
  208. target_signal.squeeze().rescale(amplitude_unit),
  209. label=target_signal.name,
  210. color=nsx_colors[nsx])
  211. # Loop through all spike trains and plot the spike time, and overlapping the
  212. # wave form of the spike used for spike sorting stored separately in the nev
  213. # file.
  214. for st in trial_segment.spiketrains:
  215. color = np.random.rand(3,)
  216. if st.annotations['channel_id'] == target_channel_id:
  217. for spike_id, spike in enumerate(st):
  218. # Plot spike times
  219. plt.axvline(
  220. spike.rescale(time_unit).magnitude,
  221. color=color,
  222. label='Unit ID %i' % st.annotations['unit_id'])
  223. # Plot waveforms
  224. waveform = st.waveforms[spike_id, 0, :]
  225. waveform_times = np.arange(len(waveform))*time_unit + spike
  226. plt.plot(
  227. waveform_times.rescale(time_unit).magnitude,
  228. waveform.rescale(amplitude_unit),
  229. '--',
  230. linewidth=2,
  231. color=color,
  232. zorder=0)
  233. # Loop through all events
  234. for event in trial_segment.events:
  235. if event.name == 'TrialEvents':
  236. for ev_id, ev in enumerate(event):
  237. plt.axvline(
  238. ev.rescale(time_unit),
  239. alpha=0.2,
  240. linewidth=3,
  241. linestyle='dashed',
  242. label=f'event {event.array_annotations["trial_event_labels"][ev_id]}')
  243. # Finishing touches on the plot
  244. plt.autoscale(enable=True, axis='x', tight=True)
  245. plt.xlabel(time_unit.name)
  246. plt.ylabel(amplitude_unit.name)
  247. plt.legend(loc=4, fontsize=10)
  248. # Save plot
  249. file_name = 'example_plot_from_blackrock_%s' % session_name
  250. for file_format in ['eps', 'png', 'pdf']:
  251. fig.savefig(file_name + '.%s' % file_format, dpi=400, format=file_format)