1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
use crate::table::{record_batch_to_rows, schema_to_field_desc};
use async_trait::async_trait;
use convergence::engine::{Engine, Portal};
use convergence::protocol::{ErrorResponse, FieldDescription, SqlState};
use convergence::protocol_ext::DataRowBatch;
use datafusion::error::DataFusionError;
use datafusion::prelude::*;
use sqlparser::ast::Statement;
use std::sync::Arc;
fn df_err_to_sql(err: DataFusionError) -> ErrorResponse {
ErrorResponse::error(SqlState::DATA_EXCEPTION, err.to_string())
}
pub struct DataFusionPortal {
df: Arc<dyn DataFrame>,
}
#[async_trait]
impl Portal for DataFusionPortal {
async fn fetch(&mut self, batch: &mut DataRowBatch) -> Result<(), ErrorResponse> {
for arrow_batch in self.df.collect().await.map_err(df_err_to_sql)? {
record_batch_to_rows(&arrow_batch, batch)?;
}
Ok(())
}
}
pub struct DataFusionEngine {
ctx: ExecutionContext,
}
impl DataFusionEngine {
pub fn new(ctx: ExecutionContext) -> Self {
Self { ctx }
}
}
#[async_trait]
impl Engine for DataFusionEngine {
type PortalType = DataFusionPortal;
async fn prepare(&mut self, statement: &Statement) -> Result<Vec<FieldDescription>, ErrorResponse> {
let plan = self.ctx.sql(&statement.to_string()).map_err(df_err_to_sql)?;
schema_to_field_desc(&plan.schema().clone().into())
}
async fn create_portal(&mut self, statement: &Statement) -> Result<Self::PortalType, ErrorResponse> {
let df = self.ctx.sql(&statement.to_string()).map_err(df_err_to_sql)?;
Ok(DataFusionPortal { df })
}
}