1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
use anyhow::{anyhow, Result};
use chrono::{DateTime, Utc};
use flate2::read::GzDecoder;
use ndarray::{s, Array2};
use std::fs::File;
use std::io::{BufRead, BufReader, Read};
use std::path::Path;
const DELIMITER: u8 = b'\t';
pub type Float = f64;
const UV_SCALE: Float = 1000.0;
/// Struct representing a reader for EEG data stored in `.easy` files.
///
/// This struct is responsible for parsing and storing the data from a `.easy` file,
/// which may include EEG signals, accelerometer data, and associated markers. The struct
/// loads the data from `.easy` and `.easy.gz` and optional `.info` files, provides methods for
/// accessing the data, and tracks relevant metadata about the file, including the
/// start date and number of channels.
#[derive(Debug)]
#[allow(dead_code)]
pub struct EasyReader {
verbose: bool,
/// Path to the `.easy` file being read.
///
/// This is the full path to the `.easy` file that contains the EEG and accelerometer data.
/// The file is parsed to extract the signals and metadata.
filepath: String,
/// Base name of the file without the extension.
///
/// This is derived from the `filepath` and excludes the extension (e.g., `.easy` or `.easy.gz`).
/// It is used for naming related files like the `.info` file.
basename: String,
/// The extension of the file (either "easy" or "easy.gz").
///
/// This is used to identify the file type and determine how to process it.
extension: String,
/// Root of the file name (file path without extension).
///
/// Used to construct the path for the associated `.info` file.
filenameroot: String,
/// Path to the associated `.info` file.
///
/// If available, this file provides information about the electrode names and other metadata.
infofilepath: String,
/// Flag indicating whether accelerometer data is present.
///
/// This flag is set to `true` if accelerometer data is found in the `.easy` file or the `.info` file.
acc_data: bool,
/// List of electrode names.
///
/// If the `.info` file is available, this field will contain the names of the EEG channels (electrodes).
/// If the `.info` file is not present, this will be populated with default channel names.
electrodes: Vec<String>,
/// Number of EEG channels.
///
/// This represents the number of electrodes in the dataset (excluding accelerometer data).
/// It is determined from the `.info` file or the `.easy` file.
num_channels: Option<usize>,
/// Start date of the EEG recording.
///
/// This date is extracted from the first timestamp in the `.easy` file. It represents the
/// time when the EEG recording began.
eegstartdate: Option<String>,
/// Array representing the time vector of the dataset in seconds.
///
/// This array contains the time of each sample relative to the start of the recording.
np_time: Option<Array2<Float>>,
/// 2D array of EEG data.
///
/// This is a 2D array where each row represents an EEG sample, and each column represents
/// an individual channel (electrode). The data is in microvolts (uV).
np_eeg: Option<Array2<Float>>,
/// 2D array of stimulus data (optional).
///
/// If present, this array contains stimulus information related to the EEG recording. It is typically used
/// for event-marking or stimulus presentation data, but it may not always be available.
np_stim: Option<Array2<Float>>,
/// 2D array of accelerometer data.
///
/// If accelerometer data is available, this array will contain the 3-axis accelerometer readings for each sample.
/// The data represents the X, Y, and Z axes of the accelerometer. The array has shape `(num_samples, 3)`.
np_acc: Option<Array2<Float>>,
/// Array of markers associated with the EEG data.
///
/// This array holds marker values that can represent events, triggers, or annotations
/// in the EEG signal. Markers are typically used to mark specific moments in time during the recording.
np_markers: Option<Array2<Float>>,
/// Log of the events related to the processing of the `.easy` file.
///
/// This is a collection of strings that logs important events, like the creation of the `EasyReader` instance
/// and when key steps in the file processing were completed. This can be useful for debugging and tracking processing.
log: Vec<String>,
}
impl EasyReader {
/// Initializes a new `EasyReader` instance from the given file path.
pub fn new(filepath: &str, verbose: bool) -> Result<Self> {
if verbose {
println!("Initializing in file path: {}", filepath);
}
let extension;
let (filenameroot, basename) = if filepath.ends_with(".easy.gz") {
extension = "easy.gz".to_string();
let filenameroot = filepath.trim_end_matches(".gz");
let basename = Path::new(filepath)
.file_name()
.unwrap()
.to_str()
.unwrap()
.trim_end_matches(".gz")
.to_string();
(filenameroot.to_string(), basename)
} else if filepath.ends_with(".easy") {
extension = "easy".to_string();
let filenameroot = filepath.trim_end_matches(".easy");
let basename = Path::new(filepath)
.file_name()
.unwrap()
.to_str()
.unwrap()
.trim_end_matches(".easy")
.to_string();
(filenameroot.to_string(), basename)
} else {
return Err(anyhow!("ERROR: Proposed file has wrong extension."));
};
let infofilepath = format!("{}.info", filenameroot);
let mut reader = EasyReader {
verbose,
filepath: filepath.to_string(),
basename,
extension,
filenameroot,
infofilepath,
acc_data: false,
electrodes: Vec::new(),
num_channels: None,
eegstartdate: None,
np_time: None,
np_eeg: None,
np_stim: None,
np_acc: None,
np_markers: None,
log: vec![format!("capsule created: {}", Utc::now())],
};
// Try to read the info file
reader.get_info()?;
Ok(reader)
}
/// Reads and processes the `.info` file for metadata about channels and accelerometer data.
fn get_info(&mut self) -> Result<()> {
let file = File::open(&self.infofilepath);
match file {
Ok(file) => {
let reader = BufReader::new(file);
let mut electrodes = Vec::new();
let mut acc_data = false;
for line in reader.lines() {
let line = line.unwrap();
if line.contains("Channel ") {
let electrode = line.split_whitespace().last().unwrap().to_string();
electrodes.push(electrode);
}
if line.contains("Accelerometer data: ") {
acc_data = true;
}
}
self.electrodes = electrodes;
self.acc_data = acc_data;
self.num_channels = Some(self.electrodes.len());
Ok(())
}
Err(_) => {
// If no info file is found, read the .easy file to determine the number of channels
self.read_easy_file_for_channels()
}
}
}
/// Reads the `.easy` file to determine the number of channels based on the file structure.
fn read_easy_file_for_channels(&mut self) -> Result<()> {
let reader = self.get_file_reader(&self.filepath)?;
let mut rdr = csv::ReaderBuilder::new()
.delimiter(DELIMITER)
.has_headers(false)
.from_reader(reader);
// Read the first 5 lines to determine number of columns
let mut header = rdr.records().take(5);
let first_record = header.next().unwrap().unwrap();
let num_columns = first_record.len();
let num_channels = if [13, 25, 37].contains(&num_columns) {
num_columns - 5
} else if [10, 22, 34].contains(&num_columns) {
num_columns - 2
} else {
return Err(anyhow!("Number of columns mismatch with expected values."));
};
self.num_channels = Some(num_channels);
self.electrodes = (1..=num_channels).map(|x| format!("Ch{}", x)).collect();
Ok(())
}
/// Reads and processes raw EEG and accelerometer data from the `.easy` file.
///
/// This method reads the `.easy` file (or the data section of it), converts the EEG data
/// into microvolts (uV), and extracts time, accelerometer, and marker data. It stores the
/// resulting data in the struct's fields (e.g., `np_eeg`, `np_time`, `np_acc`, `np_markers`).
/// It also logs key processing steps and ensures that the number of channels is consistent
/// with the data found in the file.
///
/// # Returns
///
/// - `Ok(())` if the data was successfully read and processed.
/// - `Err(String)` if there was an error reading or processing the file data. The error
/// string provides details about the failure (e.g., column mismatches or data format issues).
///
/// # Details
///
/// - The function expects the `.easy` file to have the following general format:
/// EEG data followed by accelerometer data (if available), markers, and timestamps.
/// - The EEG data is divided by channels, and the accelerometer data (if present) consists
/// of three columns representing X, Y, and Z axes.
pub fn parse_data(&mut self) -> Result<()> {
let reader = self.get_file_reader(&self.filepath)?;
let mut rdr = csv::ReaderBuilder::new()
.delimiter(DELIMITER)
.has_headers(false)
.from_reader(reader);
let mut records = rdr.records();
let first_record = records.next().unwrap().unwrap();
if self.verbose {
println!("first_record - {first_record:?}");
}
let num_columns = first_record.len();
let num_channels = if [13, 25, 37].contains(&num_columns) {
num_columns - 5
} else if [10, 22, 34].contains(&num_columns) {
num_columns - 2
} else {
return Err(anyhow!("Number of columns mismatch with expected values."));
};
// Handle timestamp
let timestamp = first_record[first_record.len() - 1].parse::<u64>().unwrap();
if let Some(start_date) = DateTime::from_timestamp((timestamp / 1000) as i64, 0) {
self.eegstartdate = Some(start_date.format("%Y-%m-%d %H:%M:%S").to_string());
}
if self.verbose {
println!("Number of channels detected: {}", num_channels);
println!(
"First sample recorded: {}",
self.eegstartdate.clone().unwrap()
);
}
// Read the rest of the file into numpy-like data
let mut eeg_data = Vec::new();
let mut acc_data = Vec::new();
let mut markers = Vec::new();
for record in records {
let record = record.unwrap();
let eeg_values: Vec<Float> = record
.iter()
.take(num_channels)
.map(|x| x.parse::<Float>().unwrap())
.map(|f| f / UV_SCALE)
.collect();
let acc_values: Vec<Float> = record
.iter()
.skip(num_channels)
.take(3)
.map(|x| x.parse::<Float>().unwrap())
.collect();
let marker_value: Float = record[num_channels + 3].parse().unwrap();
eeg_data.push(eeg_values);
acc_data.push(acc_values);
markers.push(marker_value);
}
self.np_eeg = Some(
Array2::from_shape_vec(
(eeg_data.len(), num_channels),
eeg_data.into_iter().flatten().collect(),
)
.unwrap(),
);
self.np_acc = Some(
Array2::from_shape_vec(
(acc_data.len(), 3),
acc_data.into_iter().flatten().collect(),
)
.unwrap(),
);
self.np_markers = Some(Array2::from_shape_vec((markers.len(), 1), markers).unwrap());
Ok(())
}
/// Reads and processes raw EEG and accelerometer data from the `.easy` file in a streaming manner.
///
/// This function reads the `.easy` file in chunks and processes each chunk as it is read. This approach
/// helps to minimize memory usage when dealing with large files by avoiding the need to load the entire
/// file into memory at once.
///
/// The function uses a callback (`process_chunk`) to handle each chunk of data. The callback is invoked
/// after processing each chunk, and it receives the following data:
/// - `eeg_chunk`: A `Vec<Vec<f32>>` representing a chunk of EEG data (one row per sample, one column per channel).
/// - `acc_chunk`: A `Vec<Vec<f32>>` representing a chunk of accelerometer data (three values per sample: X, Y, Z).
/// - `markers_chunk`: A `Vec<f32>` representing the marker data for each sample in the chunk.
///
/// The chunk size can be customized by passing a `chunk_size` value (in number of rows). If no chunk size
/// is provided, the default chunk size will be `1000` rows.
///
/// # Parameters:
/// - `chunk_size`: An optional parameter specifying the number of rows to process per chunk. If `None`
/// is provided, the default chunk size will be `1000`.
/// - `process_chunk`: A callback function that takes three arguments: `eeg_chunk`, `acc_chunk`, and
/// `markers_chunk`. This function will be called once a chunk is read and parsed.
///
/// # Returns:
/// - `Ok(())` if the data was successfully read and processed.
/// - `Err(String)` if there was an error
pub fn stream<F>(&mut self, chunk_size: Option<usize>, mut process_chunk: F) -> Result<()>
where
F: FnMut(Vec<Vec<Float>>, Vec<Vec<Float>>, Vec<Float>), // Callback to process each chunk of data
{
let chunk_size = match chunk_size {
Some(chunk_size) => chunk_size,
None => 1000,
};
let reader = self.get_file_reader(&self.filepath)?;
let mut rdr = csv::ReaderBuilder::new()
.delimiter(DELIMITER)
.has_headers(false)
.from_reader(reader);
let mut records = rdr.records();
let first_record = records.next().unwrap().unwrap();
let num_columns = first_record.len();
let num_channels = if [13, 25, 37].contains(&num_columns) {
num_columns - 5
} else if [10, 22, 34].contains(&num_columns) {
num_columns - 2
} else {
return Err(anyhow!("Number of columns mismatch with expected values."));
};
// Handle timestamp
let timestamp = first_record[first_record.len() - 1].parse::<u64>().unwrap();
if let Some(start_date) = DateTime::from_timestamp((timestamp / 1000) as i64, 0) {
self.eegstartdate = Some(start_date.format("%Y-%m-%d %H:%M:%S").to_string());
}
if self.verbose {
println!(
"First sample recorded: {}",
self.eegstartdate.clone().unwrap()
);
}
// Process the records in chunks
let mut eeg_chunk = Vec::new();
let mut acc_chunk = Vec::new();
let mut markers_chunk = Vec::new();
for record in records {
let record = record.unwrap();
// Process EEG data (channels)
let eeg_values: Vec<Float> = record
.iter()
.take(num_channels)
.map(|x| x.parse::<Float>().unwrap())
.map(|f| f / UV_SCALE)
.collect();
eeg_chunk.push(eeg_values);
// Process accelerometer data (3 axes)
let acc_values: Vec<Float> = record
.iter()
.skip(num_channels)
.take(3)
.map(|x| x.parse::<Float>().unwrap())
.collect();
acc_chunk.push(acc_values);
// Process marker data
let marker_value: Float = record[num_channels + 3].parse().unwrap();
markers_chunk.push(marker_value);
// Once a chunk is ready, call the callback to process the chunk
if eeg_chunk.len() >= chunk_size {
// Process every 1000 rows as a chunk
process_chunk(eeg_chunk.clone(), acc_chunk.clone(), markers_chunk.clone());
// Clear the chunk data after processing
eeg_chunk.clear();
acc_chunk.clear();
markers_chunk.clear();
}
}
// Process any remaining data in the chunk
if !eeg_chunk.is_empty() {
process_chunk(eeg_chunk, acc_chunk, markers_chunk);
}
Ok(())
}
/// Helper function to get a reader for the file, whether it's gzipped or not.
fn get_file_reader(&self, filepath: &str) -> Result<Box<dyn Read>> {
if filepath.ends_with(".gz") {
let file = File::open(filepath).map_err(|e| anyhow!(e.to_string()))?;
let decoder = GzDecoder::new(file);
Ok(Box::new(decoder))
} else {
let file = File::open(filepath).map_err(|e| anyhow!(e.to_string()))?;
Ok(Box::new(file))
}
}
/// Prints a summary of the `EasyReader` instance, displaying important metadata and previews of data.
///
/// This function outputs the file path, base name, extension, number of channels, EEG start date,
/// and any log entries related to the processing steps. It also prints the first few rows of the EEG,
/// accelerometer, and markers data, if available. This method avoids printing the entire datasets.
pub fn print_summary(&self) {
// Print metadata
println!("File Path: {}", self.filepath);
println!("Base Name: {}", self.basename);
println!("Extension: {}", self.extension);
match &self.num_channels {
Some(channels) => println!("Number of Channels: {}", channels),
None => println!("Number of Channels: Not available"),
}
match &self.eegstartdate {
Some(start_date) => println!("EEG Start Date: {}", start_date),
None => println!("EEG Start Date: Not available"),
}
// Print a preview of EEG data (first 5 samples)
match &self.np_eeg {
Some(eeg) => {
let total_samples = eeg.shape()[0];
println!("\nEEG Data (First 5 of {total_samples} Samples):");
let preview_count = total_samples.min(5); // Preview the first 5 samples or total samples if less than 5
let preview: Vec<Vec<Float>> = eeg
.slice(s![..preview_count, ..]) // Get the first `preview_count` rows and all columns
.axis_iter(ndarray::Axis(0)) // Iterate over rows
.map(|row| row.to_owned().to_vec()) // Convert each row into a Vec<Float>
.collect(); // Collect all rows into a Vec<Vec<Float>>
for (i, row) in preview.iter().enumerate() {
println!("Sample {}: {:?}", i + 1, row);
}
println!(
"Showing {} out of {} EEG samples.",
preview_count, total_samples
);
}
None => println!("EEG Data: Not available"),
}
// Print a preview of accelerometer data (first 5 samples if available)
match &self.np_acc {
Some(acc) => {
let total_samples = acc.shape()[0];
println!("\nAccelerometer Data (First 5 of {total_samples} Samples):");
let preview_count = total_samples.min(5); // Preview the first 5 samples or total samples if less than 5
let preview: Vec<Vec<Float>> = acc
.slice(s![..preview_count, ..]) // Get the first `preview_count` rows and all columns
.axis_iter(ndarray::Axis(0)) // Iterate over rows
.map(|row| row.to_owned().to_vec()) // Convert each row into a Vec<f32>
.collect(); // Collect all rows into a Vec<Vec<Float>>
for (i, row) in preview.iter().enumerate() {
println!("Sample {}: {:?}", i + 1, row);
}
}
None => println!("Accelerometer Data: Not available"),
}
// Print a preview of markers (first 5 samples if available)
match &self.np_markers {
Some(markers) => {
let total_samples = markers.shape()[0];
println!("\nMarkers Data (First 5 of {total_samples} Samples):");
let preview_count = total_samples.min(5); // Preview the first 5 samples or total samples if less than 5
let (preview, _) = markers
.slice(s![..preview_count, ..]) // Get the first `preview_count` elements
.to_owned() // Copy the values from the slice
.into_raw_vec_and_offset(); // Convert it into a Vec<Float>
for (i, marker) in preview.iter().enumerate() {
println!("Marker {}: {}", i + 1, marker);
}
}
None => println!("Markers Data: Not available"),
}
// Print log entries
println!("\nLog Entries:");
for entry in &self.log {
println!("- {}", entry);
}
}
}