use csv;
use config::{Config, Delimiter};
use select::SelectColumns;
use util::{self, ImmutableRecordHelpers};
use CliResult;
use moonblade::AggregationProgram;
use moonblade::GroupAggregationProgram;
use cmd::moonblade::{
get_moonblade_aggregations_function_help, get_moonblade_cheatsheet,
get_moonblade_functions_help, MoonbladeErrorPolicy,
};
static USAGE: &str = "
Group a CSV file by values contained in a given column then aggregate data per
group using a custom aggregation expression.
The result of running the command will be a CSV file containing a \"group\"
column containing the value representing each group and additional columns for
each computed aggregation.
You can, for instance, compute the sum of a column per group:
$ xan groupby user_name 'sum(retweet_count)' file.csv
You can use dynamic expressions to mangle the data before aggregating it:
$ xan groupby user_name 'sum(retweet_count + replies_count)' file.csv
You can perform multiple aggregations at once:
$ xan groupby user_name 'sum(retweet_count), mean(retweet_count), max(replies_count)' file.csv
You can rename the output columns using the 'as' syntax:
$ xan groupby user_name 'sum(n) as sum, max(replies_count) as \"Max Replies\"' file.csv
For a quick review of the capabilities of the script language, use
the --cheatsheet flag.
For a list of available aggregation functions, use the --aggs flag.
If you want to list available functions, use the --functions flag.
Usage:
xan groupby [options] <column> <expression> [<input>]
xan groupby --help
xan groupby --cheatsheet
xan groupby --aggs
xan groupby --functions
groupby options:
--group-column <name> Name of the column containing group values.
[default: group].
-S, --sorted Use this flag to indicate that the file is already sorted on the
group column, in which case the command will be able to considerably
optimize memory usage.
-e, --errors <policy> What to do with evaluation errors. One of:
- \"panic\": exit on first error
- \"ignore\": ignore row altogether
- \"log\": print error to stderr
[default: panic].
-p, --parallel Whether to use parallelization to speed up computations.
Will automatically select a suitable number of threads to use
based on your number of cores.
Common options:
-h, --help Display this message
-o, --output <file> Write output to <file> instead of stdout.
-n, --no-headers When set, the first row will not be evaled
as headers.
-d, --delimiter <arg> The field delimiter for reading CSV data.
Must be a single character. [default: ,]
";
#[derive(Deserialize)]
struct Args {
arg_column: SelectColumns,
arg_expression: String,
arg_input: Option<String>,
flag_no_headers: bool,
flag_output: Option<String>,
flag_delimiter: Option<Delimiter>,
flag_aggs: bool,
flag_cheatsheet: bool,
flag_functions: bool,
flag_group_column: String,
flag_sorted: bool,
flag_errors: String,
flag_parallel: bool,
}
pub fn run(argv: &[&str]) -> CliResult<()> {
let args: Args = util::get_args(USAGE, argv)?;
if args.flag_aggs {
println!("{}", get_moonblade_aggregations_function_help());
return Ok(());
}
if args.flag_cheatsheet {
println!("{}", get_moonblade_cheatsheet());
return Ok(());
}
if args.flag_functions {
println!("{}", get_moonblade_functions_help());
return Ok(());
}
let error_policy = MoonbladeErrorPolicy::from_restricted(&args.flag_errors)?;
let rconf = Config::new(&args.arg_input)
.delimiter(args.flag_delimiter)
.no_headers(args.flag_no_headers)
.select(args.arg_column);
let mut rdr = rconf.reader()?;
let mut wtr = Config::new(&args.flag_output).writer()?;
let headers = rdr.byte_headers()?;
let column_index = rconf.single_selection(headers)?;
let mut record = csv::ByteRecord::new();
if args.flag_sorted {
let mut program = AggregationProgram::parse(&args.arg_expression, headers)?;
let mut current_group: Option<Vec<u8>> = None;
record.push_field(args.flag_group_column.as_bytes());
record.extend(program.headers());
wtr.write_byte_record(&record)?;
let mut index: usize = 0;
while rdr.read_byte_record(&mut record)? {
index += 1;
let group = record[column_index].to_vec();
match current_group.as_ref() {
None => {
current_group = Some(group);
}
Some(group_name) => {
if group_name != &group {
wtr.write_byte_record(
&program.finalize(args.flag_parallel).prepend(group_name),
)?;
program.clear();
current_group = Some(group);
}
}
};
program
.run_with_record(index, &record)
.or_else(|error| error_policy.handle_error(index, error))?;
}
if let Some(group_name) = current_group {
wtr.write_byte_record(&program.finalize(args.flag_parallel).prepend(&group_name))?;
}
} else {
let mut program = GroupAggregationProgram::parse(&args.arg_expression, headers)?;
record.push_field(args.flag_group_column.as_bytes());
record.extend(program.headers());
wtr.write_byte_record(&record)?;
let mut index: usize = 0;
while rdr.read_byte_record(&mut record)? {
index += 1;
let group = record[column_index].to_vec();
program
.run_with_record(group, index, &record)
.or_else(|error| error_policy.handle_error(index, error))?;
}
for (group, group_record) in program.into_byte_records(args.flag_parallel) {
wtr.write_byte_record(&group_record.prepend(&group))?;
}
}
Ok(wtr.flush()?)
}