nodedb 0.0.0-beta.1

Local-first, real-time, edge-to-cloud hybrid database for multi-modal workloads
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
//! Bridge between the timeseries-specific columnar types and the shared
//! `nodedb-columnar` crate types.
//!
//! The timeseries engine has `Symbol` columns (dictionary-encoded tags) that
//! don't exist in the shared columnar format. This bridge converts timeseries
//! drain results into the shared format for segment writing, handling Symbol
//! columns by writing them as Int64 (their u32 symbol IDs cast to i64).
//!
//! Symbol dictionaries are persisted separately alongside the segment.

use nodedb_columnar::format::BlockStats;
use nodedb_columnar::memtable::ColumnData as SharedColumnData;
use nodedb_columnar::predicate::ScanPredicate;
use nodedb_columnar::reader::{DecodedColumn, SegmentReader};
use nodedb_columnar::writer::SegmentWriter;
use nodedb_types::columnar::{ColumnDef, ColumnType as SharedColumnType, ColumnarSchema};

use super::columnar_memtable::{
    ColumnData as TsColumnData, ColumnType as TsColumnType, ColumnarDrainResult,
};

/// Convert a timeseries ColumnarSchema to a shared ColumnarSchema.
///
/// `Symbol` columns are mapped to `Int64` in the shared schema (the u32
/// symbol IDs are stored as i64). The symbol dictionary is persisted
/// separately.
pub fn ts_schema_to_shared(ts_schema: &super::columnar_memtable::ColumnarSchema) -> ColumnarSchema {
    let columns: Vec<ColumnDef> = ts_schema
        .columns
        .iter()
        .map(|(name, ts_type)| {
            let shared_type = ts_column_type_to_shared(*ts_type);
            // All timeseries columns are non-nullable (metrics always present).
            ColumnDef::required(name.clone(), shared_type)
        })
        .collect();

    // Schema validation can't fail here since we're converting known-good types.
    ColumnarSchema {
        columns,
        version: 1,
    }
}

/// Map a timeseries ColumnType to a shared ColumnType.
fn ts_column_type_to_shared(ts_type: TsColumnType) -> SharedColumnType {
    match ts_type {
        TsColumnType::Timestamp => SharedColumnType::Timestamp,
        TsColumnType::Float64 => SharedColumnType::Float64,
        TsColumnType::Int64 => SharedColumnType::Int64,
        // Symbol columns store u32 IDs — represented as Int64 in shared format.
        TsColumnType::Symbol => SharedColumnType::Int64,
    }
}

/// Convert timeseries column data to shared ColumnData for the SegmentWriter.
///
/// Returns a Vec of shared ColumnData plus the row count.
pub fn ts_drain_to_shared_columns(drain: &ColumnarDrainResult) -> (Vec<SharedColumnData>, usize) {
    let row_count = drain.row_count as usize;
    let mut shared_columns = Vec::with_capacity(drain.columns.len());

    for (i, ts_col) in drain.columns.iter().enumerate() {
        let ts_type = drain.schema.columns[i].1;
        let shared = ts_column_to_shared(ts_col, ts_type, row_count);
        shared_columns.push(shared);
    }

    (shared_columns, row_count)
}

/// Convert a single timeseries ColumnData to a shared ColumnData.
fn ts_column_to_shared(
    ts_col: &TsColumnData,
    ts_type: TsColumnType,
    row_count: usize,
) -> SharedColumnData {
    match (ts_col, ts_type) {
        (TsColumnData::Timestamp(values), TsColumnType::Timestamp) => SharedColumnData::Timestamp {
            values: values.clone(),
            valid: vec![true; row_count],
        },
        (TsColumnData::Float64(values), TsColumnType::Float64) => SharedColumnData::Float64 {
            values: values.clone(),
            valid: vec![true; row_count],
        },
        (TsColumnData::Int64(values), TsColumnType::Int64) => SharedColumnData::Int64 {
            values: values.clone(),
            valid: vec![true; row_count],
        },
        (TsColumnData::Symbol(sym_ids), TsColumnType::Symbol) => {
            // Convert u32 symbol IDs to i64 for the shared Int64 column.
            SharedColumnData::Int64 {
                values: sym_ids.iter().map(|&id| id as i64).collect(),
                valid: vec![true; row_count],
            }
        }
        // Fallback: shouldn't happen if schema is consistent.
        _ => SharedColumnData::Int64 {
            values: vec![0; row_count],
            valid: vec![false; row_count],
        },
    }
}

/// Write a timeseries drain result as a shared columnar segment.
///
/// Returns the segment bytes. The caller is responsible for persisting the
/// symbol dictionaries separately.
pub fn write_ts_drain_as_segment(
    drain: &ColumnarDrainResult,
) -> Result<Vec<u8>, nodedb_columnar::ColumnarError> {
    let shared_schema = ts_schema_to_shared(&drain.schema);
    let (shared_columns, row_count) = ts_drain_to_shared_columns(drain);

    let writer = SegmentWriter::new(nodedb_columnar::writer::PROFILE_TIMESERIES);
    writer.write_segment(&shared_schema, &shared_columns, row_count)
}

/// Read block statistics from a shared segment footer for timeseries queries.
///
/// Returns the timestamp column's block stats (for time-range predicate pushdown).
pub fn extract_timestamp_block_stats(
    segment_data: &[u8],
) -> Result<Vec<BlockStats>, nodedb_columnar::ColumnarError> {
    let reader = SegmentReader::open(segment_data)?;
    let footer = reader.footer();

    // Timestamp is always column 0 in timeseries schemas.
    if footer.columns.is_empty() {
        return Ok(Vec::new());
    }
    Ok(footer.columns[0].block_stats.clone())
}

// ---------------------------------------------------------------------------
// Read path: shared segment → timeseries query data
// ---------------------------------------------------------------------------

/// Result of scanning a shared-format segment for timeseries data.
pub struct TsScanResult {
    /// Timestamp values (microseconds or milliseconds depending on schema).
    pub timestamps: Vec<i64>,
    /// Value column data (f64 for metrics).
    pub values: Vec<f64>,
    /// Validity masks — false for rows filtered out by predicates or block skip.
    pub ts_valid: Vec<bool>,
    pub val_valid: Vec<bool>,
}

/// Scan a shared-format segment for timeseries data with time-range filtering.
///
/// Uses block-level predicate pushdown: blocks whose timestamp range doesn't
/// overlap `[start_ms, end_ms]` are skipped entirely (not decompressed).
///
/// `ts_col_idx` is the timestamp column index (typically 0).
/// `val_col_idx` is the value column index to read.
pub fn scan_shared_segment(
    segment_data: &[u8],
    ts_col_idx: usize,
    val_col_idx: usize,
    start_ms: i64,
    end_ms: i64,
) -> Result<TsScanResult, nodedb_columnar::ColumnarError> {
    let reader = SegmentReader::open(segment_data)?;

    // Build predicates for block-level time-range skip.
    // Timestamp blocks with max < start or min > end can be skipped.
    let predicates = vec![
        ScanPredicate::gte(ts_col_idx, start_ms as f64),
        ScanPredicate::lte(ts_col_idx, end_ms as f64),
    ];

    // Read timestamp column with predicate pushdown.
    let ts_decoded = reader.read_column_filtered(ts_col_idx, &predicates)?;
    let (timestamps, ts_valid) = decoded_to_i64(ts_decoded);

    // Read value column with same predicate pushdown (ensures row alignment).
    let val_decoded = reader.read_column_filtered(val_col_idx, &predicates)?;
    let (values, val_valid) = decoded_to_f64(val_decoded);

    Ok(TsScanResult {
        timestamps,
        values,
        ts_valid,
        val_valid,
    })
}

/// Scan a shared-format segment, returning only rows within a time range.
///
/// Unlike `scan_shared_segment` which does block-level filtering, this
/// additionally applies row-level filtering, returning only exact matches.
pub fn scan_shared_segment_filtered(
    segment_data: &[u8],
    ts_col_idx: usize,
    val_col_idx: usize,
    start_ms: i64,
    end_ms: i64,
) -> Result<(Vec<i64>, Vec<f64>), nodedb_columnar::ColumnarError> {
    let scan = scan_shared_segment(segment_data, ts_col_idx, val_col_idx, start_ms, end_ms)?;

    // Row-level filter: keep only rows where ts is within range AND valid.
    let mut filtered_ts = Vec::new();
    let mut filtered_vals = Vec::new();

    for i in 0..scan.timestamps.len() {
        if scan.ts_valid[i]
            && scan.val_valid[i]
            && scan.timestamps[i] >= start_ms
            && scan.timestamps[i] <= end_ms
        {
            filtered_ts.push(scan.timestamps[i]);
            filtered_vals.push(scan.values[i]);
        }
    }

    Ok((filtered_ts, filtered_vals))
}

/// Read a single column from a shared segment by index.
///
/// For timeseries queries that need an arbitrary column (e.g., tag columns
/// for GROUP BY).
pub fn read_column_from_shared(
    segment_data: &[u8],
    col_idx: usize,
) -> Result<DecodedColumn, nodedb_columnar::ColumnarError> {
    let reader = SegmentReader::open(segment_data)?;
    reader.read_column(col_idx)
}

/// Extract i64 values and validity from a DecodedColumn.
fn decoded_to_i64(col: DecodedColumn) -> (Vec<i64>, Vec<bool>) {
    match col {
        DecodedColumn::Int64 { values, valid } => (values, valid),
        DecodedColumn::Timestamp { values, valid } => (values, valid),
        DecodedColumn::Float64 { values, valid } => {
            (values.into_iter().map(|f| f as i64).collect(), valid)
        }
        _ => (Vec::new(), Vec::new()),
    }
}

/// Extract f64 values and validity from a DecodedColumn.
fn decoded_to_f64(col: DecodedColumn) -> (Vec<f64>, Vec<bool>) {
    match col {
        DecodedColumn::Float64 { values, valid } => (values, valid),
        DecodedColumn::Int64 { values, valid } => {
            (values.into_iter().map(|i| i as f64).collect(), valid)
        }
        DecodedColumn::Timestamp { values, valid } => {
            (values.into_iter().map(|i| i as f64).collect(), valid)
        }
        _ => (Vec::new(), Vec::new()),
    }
}

#[cfg(test)]
mod tests {
    use nodedb_types::timeseries::{IngestResult, MetricSample};

    use super::super::columnar_memtable::{ColumnarMemtable, ColumnarMemtableConfig};
    use super::*;

    fn default_config() -> ColumnarMemtableConfig {
        ColumnarMemtableConfig {
            max_memory_bytes: crate::engine::timeseries::memtable::DEFAULT_MEMTABLE_BUDGET_BYTES,
            hard_memory_limit: 80 * 1024 * 1024,
            max_tag_cardinality: 100_000,
        }
    }

    #[test]
    fn schema_conversion() {
        let ts_schema = super::super::columnar_memtable::ColumnarSchema {
            columns: vec![
                ("timestamp".into(), TsColumnType::Timestamp),
                ("value".into(), TsColumnType::Float64),
                ("host".into(), TsColumnType::Symbol),
            ],
            timestamp_idx: 0,
            codecs: vec![nodedb_codec::ColumnCodec::Auto; 3],
        };

        let shared = ts_schema_to_shared(&ts_schema);
        assert_eq!(shared.columns.len(), 3);
        assert_eq!(shared.columns[0].column_type, SharedColumnType::Timestamp);
        assert_eq!(shared.columns[1].column_type, SharedColumnType::Float64);
        // Symbol → Int64 in shared format.
        assert_eq!(shared.columns[2].column_type, SharedColumnType::Int64);
    }

    #[test]
    fn drain_to_shared_segment_roundtrip() {
        let mut mt = ColumnarMemtable::new_metric(default_config());

        for i in 0..100 {
            let result = mt.ingest_metric(
                i % 10,
                MetricSample {
                    timestamp_ms: 1000 + i as i64,
                    value: i as f64 * 0.5,
                },
            );
            assert_ne!(result, IngestResult::Rejected);
        }

        let drain = mt.drain();
        assert_eq!(drain.row_count, 100);

        // Write as shared segment.
        let segment = write_ts_drain_as_segment(&drain).expect("write segment");

        // Read back and verify.
        let reader = nodedb_columnar::reader::SegmentReader::open(&segment).expect("open");
        assert_eq!(reader.row_count(), 100);
        assert_eq!(reader.column_count(), 2); // timestamp + value

        // Verify timestamp column.
        let ts_col = reader.read_column(0).expect("read ts");
        match ts_col {
            nodedb_columnar::reader::DecodedColumn::Int64 { values, valid } => {
                assert_eq!(values.len(), 100);
                assert_eq!(values[0], 1000);
                assert_eq!(values[99], 1099);
                assert!(valid.iter().all(|&v| v));
            }
            _ => panic!("expected Int64 for timestamps"),
        }

        // Verify footer has timeseries profile tag.
        let footer = reader.footer();
        assert_eq!(
            footer.profile_tag,
            nodedb_columnar::writer::PROFILE_TIMESERIES
        );
    }

    #[test]
    fn timestamp_block_stats_extraction() {
        let mut mt = ColumnarMemtable::new_metric(default_config());
        for i in 0..50 {
            mt.ingest_metric(
                1,
                MetricSample {
                    timestamp_ms: 2000 + i as i64,
                    value: 1.0,
                },
            );
        }

        let drain = mt.drain();
        let segment = write_ts_drain_as_segment(&drain).expect("write");
        let stats = extract_timestamp_block_stats(&segment).expect("stats");

        assert_eq!(stats.len(), 1); // 50 rows < 1024 block size = 1 block.
        assert_eq!(stats[0].min, 2000.0);
        assert_eq!(stats[0].max, 2049.0);
        assert_eq!(stats[0].row_count, 50);
    }

    // -- Read path tests --

    fn write_test_segment(count: usize, start_ts: i64) -> Vec<u8> {
        let mut mt = ColumnarMemtable::new_metric(default_config());
        for i in 0..count {
            mt.ingest_metric(
                (i % 5) as u64,
                MetricSample {
                    timestamp_ms: start_ts + i as i64,
                    value: i as f64 * 0.25,
                },
            );
        }
        let drain = mt.drain();
        write_ts_drain_as_segment(&drain).expect("write")
    }

    #[test]
    fn scan_full_range() {
        let segment = write_test_segment(100, 1000);

        let result = scan_shared_segment(&segment, 0, 1, 1000, 1099).expect("scan");

        assert_eq!(result.timestamps.len(), 100);
        assert_eq!(result.values.len(), 100);
        assert!(result.ts_valid.iter().all(|&v| v));
        assert_eq!(result.timestamps[0], 1000);
        assert_eq!(result.timestamps[99], 1099);
        assert!((result.values[0] - 0.0).abs() < 0.001);
        assert!((result.values[4] - 1.0).abs() < 0.001);
    }

    #[test]
    fn scan_filtered_range() {
        let segment = write_test_segment(100, 1000);

        // Only rows with ts in [1050, 1060].
        let (ts, vals) = scan_shared_segment_filtered(&segment, 0, 1, 1050, 1060).expect("scan");

        assert_eq!(ts.len(), 11); // 1050..=1060 = 11 rows.
        assert_eq!(ts[0], 1050);
        assert_eq!(ts[10], 1060);
        assert_eq!(vals.len(), 11);
    }

    #[test]
    fn scan_empty_range() {
        let segment = write_test_segment(100, 1000);

        // Range outside segment data.
        let (ts, vals) = scan_shared_segment_filtered(&segment, 0, 1, 5000, 6000).expect("scan");

        assert!(ts.is_empty());
        assert!(vals.is_empty());
    }

    #[test]
    fn scan_block_level_skip() {
        // Create a large segment with multiple blocks (>1024 rows).
        let segment = write_test_segment(3000, 0);

        // Query only the last block: ts in [2500, 2999].
        let result = scan_shared_segment(&segment, 0, 1, 2500, 2999).expect("scan");

        // Block 0 [0..1023] and block 1 [1024..2047] should be skipped.
        // Block 2 [2048..2999] should be read.
        // Skipped blocks produce null-filled rows (valid=false).
        let valid_count = result.ts_valid.iter().filter(|&&v| v).count();
        // Block 2 has 952 rows (3000-2048), all valid.
        assert_eq!(valid_count, 952);

        // Use filtered scan to get only the matching rows.
        let (ts, _) = scan_shared_segment_filtered(&segment, 0, 1, 2500, 2999).expect("filtered");
        assert_eq!(ts.len(), 500); // 2500..=2999 = 500 rows.
    }

    #[test]
    fn read_arbitrary_column() {
        let segment = write_test_segment(50, 1000);

        let col = read_column_from_shared(&segment, 1).expect("read value col");
        match col {
            DecodedColumn::Float64 { values, valid } => {
                assert_eq!(values.len(), 50);
                assert!(valid.iter().all(|&v| v));
                assert!((values[0] - 0.0).abs() < 0.001);
                assert!((values[4] - 1.0).abs() < 0.001);
            }
            _ => panic!("expected Float64 for value column"),
        }
    }
}