iroha-internal 0.1.1

Internal functions, struct for Iroha
mod helper;
pub mod ty;
pub use helper::TokenizableError;

use syn::{Type, Error};
use proc_macro2::TokenStream;
use ty::*;
use syn::export::ToTokens;
use helper::assert_angle_args;

pub fn get_wrapped_value(ty: &Type, value_path: TokenStream, as_ref: bool, clone: bool) -> Result<TokenStream, Error> {
    let ref_token = if as_ref {
        quote::quote! {&}
    } else {
        TokenStream::new()
    };

    let clone_token = if clone {
        quote::quote! {.clone()}
    } else {
        TokenStream::new()
    };
    if let Type::Path(type_path) = ty {
        let last_segment = type_path.path.segments.iter().rev().next().unwrap();
        let args = assert_angle_args(&last_segment.arguments)?;
        match last_segment.ident.to_string().as_str() {
            "Vec" => TokenizableVec::<String>::convert_token_stream(args, &value_path),
            "HashMap" => TokenizableHashMap::<String, String>::convert_token_stream(args, &value_path),
            "HashSet" => TokenizableHashSet::<String>::convert_token_stream(args, &value_path),
            "String" => TokenizableString::convert_token_stream(args, &value_path),
            "Option" => TokenizableOption::<String>::convert_token_stream(args, &value_path),
            "Result" => TokenizableResult::<String, TokenizableError>::convert_token_stream(args, &value_path),
            _ => Ok(quote::quote! {
                #ref_token#value_path#clone_token
            })
        }
    } else {
        Ok(quote::quote! {
            #ref_token#value_path#clone_token
        })
    }
}

pub fn get_wrapper(ty: &Type) -> TokenStream {
    if let Type::Path(type_path) = ty {
        let last_segment = type_path.path.segments.iter().rev().next().unwrap();
        let arguments = &last_segment.arguments;
        match last_segment.ident.to_string().as_str() {
            "Vec" => Some(TokenizableVec::<String>::type_name(arguments)),
            "HashMap" => Some(TokenizableHashMap::<String, String>::type_name(arguments)),
            "HashSet" => Some(TokenizableHashSet::<String>::type_name(arguments)),
            "String" => Some(TokenizableString::type_name(arguments)),
            "Result" =>Some(TokenizableResult::<String, TokenizableError>::type_name(arguments)),
            "Option" => Some(TokenizableOption::<String>::type_name(arguments)),
            _ => None
        }
    } else {
        None
    }.unwrap_or_else(|| ty.to_token_stream())
}