extern crate base64;
extern crate md5;
use bytes::Bytes;
use futures::TryStreamExt;
use hyper::{Body, Client};
use hyper_tls::HttpsConnector;
use std::collections::HashMap;
use time::OffsetDateTime;
use super::request_trait::{Request, ResponseData};
use crate::bucket::Bucket;
use crate::command::Command;
use crate::command::HttpMethod;
use crate::error::S3Error;
use tokio_stream::StreamExt;
pub use crate::request::tokio_backend::HyperRequest as RequestImpl;
pub use tokio::io::AsyncRead;
pub use tokio::io::{AsyncWrite, AsyncWriteExt};
pub use tokio_stream::Stream;
use tracing::{event, span, Level};
use crate::request::request_trait::ResponseDataStream;
pub struct HyperRequest<'a> {
pub bucket: &'a Bucket,
pub path: &'a str,
pub command: Command<'a>,
pub datetime: OffsetDateTime,
}
#[async_trait::async_trait]
impl<'a> Request for HyperRequest<'a> {
type Response = http::Response<Body>;
type HeaderMap = http::header::HeaderMap;
async fn response(&self) -> Result<http::Response<Body>, S3Error> {
let headers = match self.headers() {
Ok(headers) => headers,
Err(e) => return Err(e),
};
let https_connector = HttpsConnector::new();
let client = Client::builder().build::<_, hyper::Body>(https_connector);
let method = match self.command.http_verb() {
HttpMethod::Delete => http::Method::DELETE,
HttpMethod::Get => http::Method::GET,
HttpMethod::Post => http::Method::POST,
HttpMethod::Put => http::Method::PUT,
HttpMethod::Head => http::Method::HEAD,
};
let request = {
let mut request = http::Request::builder()
.method(method)
.uri(self.url()?.as_str());
for (header, value) in headers.iter() {
request = request.header(header, value);
}
request.body(Body::from(self.request_body()))?
};
let span = span!(
Level::DEBUG,
"rust-s3-async",
bucket = self.bucket.name(),
command = self.command.to_string(),
path = self.path,
second = self.datetime.second(),
minute = self.datetime.minute(),
hour = self.datetime.hour(),
day = self.datetime.day(),
month = self.datetime.month() as u8,
year = self.datetime.year()
);
let _enter = span.enter();
let response = client.request(request).await?;
event!(Level::DEBUG, status_code = response.status().as_u16(),);
if cfg!(feature = "fail-on-err") && !response.status().is_success() {
let status = response.status().as_u16();
let text =
String::from_utf8(hyper::body::to_bytes(response.into_body()).await?.into())?;
return Err(S3Error::HttpFailWithBody(status, text));
}
Ok(response)
}
async fn response_data(&self, etag: bool) -> Result<ResponseData, S3Error> {
let response = self.response().await?;
let status_code = response.status().as_u16();
let mut headers = response.headers().clone();
let response_headers = headers
.clone()
.iter()
.map(|(k, v)| {
(
k.to_string(),
v.to_str()
.unwrap_or("could-not-decode-header-value")
.to_string(),
)
})
.collect::<HashMap<String, String>>();
let body_vec = if etag {
if let Some(etag) = headers.remove("ETag") {
Bytes::from(etag.to_str()?.to_string())
} else {
Bytes::from("")
}
} else {
hyper::body::to_bytes(response.into_body()).await?
};
Ok(ResponseData::new(body_vec, status_code, response_headers))
}
async fn response_data_to_writer<T: tokio::io::AsyncWrite + Send + Unpin>(
&self,
writer: &mut T,
) -> Result<u16, S3Error> {
let response = self.response().await?;
let status_code = response.status();
let mut stream = response.into_body().into_stream();
while let Some(item) = stream.next().await {
writer.write_all(&item?).await?;
}
Ok(status_code.as_u16())
}
async fn response_data_to_stream(&self) -> Result<ResponseDataStream, S3Error> {
let response = self.response().await?;
let status_code = response.status();
let stream = response.into_body().into_stream().map_err(S3Error::Hyper);
Ok(ResponseDataStream {
bytes: Box::pin(stream),
status_code: status_code.as_u16(),
})
}
async fn response_header(&self) -> Result<(Self::HeaderMap, u16), S3Error> {
let response = self.response().await?;
let status_code = response.status().as_u16();
let headers = response.headers().clone();
Ok((headers, status_code))
}
fn datetime(&self) -> OffsetDateTime {
self.datetime
}
fn bucket(&self) -> Bucket {
self.bucket.clone()
}
fn command(&self) -> Command {
self.command.clone()
}
fn path(&self) -> String {
self.path.to_string()
}
}
impl<'a> HyperRequest<'a> {
pub fn new(
bucket: &'a Bucket,
path: &'a str,
command: Command<'a>,
) -> Result<HyperRequest<'a>, S3Error> {
bucket.credentials_refresh()?;
Ok(Self {
bucket,
path,
command,
datetime: OffsetDateTime::now_utc(),
})
}
}
#[cfg(test)]
mod tests {
use crate::bucket::Bucket;
use crate::command::Command;
use crate::request::tokio_backend::HyperRequest;
use crate::request::Request;
use awscreds::Credentials;
use http::header::{HOST, RANGE};
fn fake_credentials() -> Credentials {
let access_key = "AKIAIOSFODNN7EXAMPLE";
let secert_key = "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY";
Credentials::new(Some(access_key), Some(secert_key), None, None, None).unwrap()
}
#[test]
fn url_uses_https_by_default() {
let region = "custom-region".parse().unwrap();
let bucket = Bucket::new("my-first-bucket", region, fake_credentials()).unwrap();
let path = "/my-first/path";
let request = HyperRequest::new(&bucket, path, Command::GetObject).unwrap();
assert_eq!(request.url().unwrap().scheme(), "https");
let headers = request.headers().unwrap();
let host = headers.get(HOST).unwrap();
assert_eq!(*host, "my-first-bucket.custom-region".to_string());
}
#[test]
fn url_uses_https_by_default_path_style() {
let region = "custom-region".parse().unwrap();
let bucket = Bucket::new("my-first-bucket", region, fake_credentials())
.unwrap()
.with_path_style();
let path = "/my-first/path";
let request = HyperRequest::new(&bucket, path, Command::GetObject).unwrap();
assert_eq!(request.url().unwrap().scheme(), "https");
let headers = request.headers().unwrap();
let host = headers.get(HOST).unwrap();
assert_eq!(*host, "custom-region".to_string());
}
#[test]
fn url_uses_scheme_from_custom_region_if_defined() {
let region = "http://custom-region".parse().unwrap();
let bucket = Bucket::new("my-second-bucket", region, fake_credentials()).unwrap();
let path = "/my-second/path";
let request = HyperRequest::new(&bucket, path, Command::GetObject).unwrap();
assert_eq!(request.url().unwrap().scheme(), "http");
let headers = request.headers().unwrap();
let host = headers.get(HOST).unwrap();
assert_eq!(*host, "my-second-bucket.custom-region".to_string());
}
#[test]
fn url_uses_scheme_from_custom_region_if_defined_with_path_style() {
let region = "http://custom-region".parse().unwrap();
let bucket = Bucket::new("my-second-bucket", region, fake_credentials())
.unwrap()
.with_path_style();
let path = "/my-second/path";
let request = HyperRequest::new(&bucket, path, Command::GetObject).unwrap();
assert_eq!(request.url().unwrap().scheme(), "http");
let headers = request.headers().unwrap();
let host = headers.get(HOST).unwrap();
assert_eq!(*host, "custom-region".to_string());
}
#[test]
fn test_get_object_range_header() {
let region = "http://custom-region".parse().unwrap();
let bucket = Bucket::new("my-second-bucket", region, fake_credentials())
.unwrap()
.with_path_style();
let path = "/my-second/path";
let request = HyperRequest::new(
&bucket,
path,
Command::GetObjectRange {
start: 0,
end: None,
},
)
.unwrap();
let headers = request.headers().unwrap();
let range = headers.get(RANGE).unwrap();
assert_eq!(range, "bytes=0-");
let request = HyperRequest::new(
&bucket,
path,
Command::GetObjectRange {
start: 0,
end: Some(1),
},
)
.unwrap();
let headers = request.headers().unwrap();
let range = headers.get(RANGE).unwrap();
assert_eq!(range, "bytes=0-1");
}
}