use std::ffi::{OsStr, OsString};
use std::future::Future;
use std::num::NonZeroUsize;
use std::ops::Deref;
use std::ops::Range;
use std::os::unix::prelude::OsStrExt;
use std::pin::Pin;
use std::sync::atomic::{AtomicBool, AtomicU64, Ordering};
use std::sync::{Arc, Mutex};
use std::task::{Context, Poll};
use std::time::{Duration, Instant};
use mountpoint_s3_crt::auth::credentials::{
CredentialsProvider, CredentialsProviderChainDefaultOptions, CredentialsProviderProfileOptions,
};
use mountpoint_s3_crt::auth::signing_config::SigningConfig;
use mountpoint_s3_crt::common::allocator::Allocator;
use mountpoint_s3_crt::common::string::AwsString;
use mountpoint_s3_crt::common::uri::Uri;
use mountpoint_s3_crt::http::request_response::{Header, Headers, Message};
use mountpoint_s3_crt::io::channel_bootstrap::{ClientBootstrap, ClientBootstrapOptions};
use mountpoint_s3_crt::io::event_loop::EventLoopGroup;
use mountpoint_s3_crt::io::host_resolver::{AddressKinds, HostResolver, HostResolverDefaultOptions};
use mountpoint_s3_crt::io::retry_strategy::{ExponentialBackoffJitterMode, RetryStrategy, StandardRetryOptions};
use mountpoint_s3_crt::s3::client::{
init_signing_config, ChecksumConfig, Client, ClientConfig, MetaRequest, MetaRequestOptions, MetaRequestResult,
MetaRequestType, RequestMetrics, RequestType,
};
use async_trait::async_trait;
use futures::channel::oneshot;
use percent_encoding::{percent_encode, AsciiSet, NON_ALPHANUMERIC};
use pin_project::{pin_project, pinned_drop};
use thiserror::Error;
use tracing::{debug, error, trace, Span};
use self::get_object::S3GetObjectRequest;
use self::put_object::S3PutObjectRequest;
use crate::endpoint_config::EndpointConfig;
use crate::endpoint_config::EndpointError;
use crate::object_client::*;
use crate::user_agent::UserAgent;
macro_rules! request_span {
($self:expr, $method:expr, $($field:tt)*) => {{
let counter = $self.next_request_counter();
let span = tracing::warn_span!(target: "mountpoint_s3_client::s3_crt_client::request", $method, id = counter, $($field)*);
span.in_scope(|| tracing::debug!("new request"));
span
}};
($self:expr, $method:expr) => { request_span!($self, $method,) };
}
pub(crate) mod delete_object;
pub(crate) mod get_object;
pub(crate) mod get_object_attributes;
pub(crate) mod head_object;
pub(crate) mod list_objects;
pub(crate) mod put_object;
pub(crate) mod head_bucket;
pub use head_bucket::HeadBucketError;
macro_rules! event {
($level:expr, $($args:tt)*) => {
match $level {
::tracing::Level::ERROR => ::tracing::event!(::tracing::Level::ERROR, $($args)*),
::tracing::Level::WARN => ::tracing::event!(::tracing::Level::WARN, $($args)*),
::tracing::Level::INFO => ::tracing::event!(::tracing::Level::INFO, $($args)*),
::tracing::Level::DEBUG => ::tracing::event!(::tracing::Level::DEBUG, $($args)*),
::tracing::Level::TRACE => ::tracing::event!(::tracing::Level::TRACE, $($args)*),
}
}
}
#[derive(Debug, Clone)]
pub struct S3ClientConfig {
auth_config: S3ClientAuthConfig,
throughput_target_gbps: f64,
part_size: usize,
endpoint_config: EndpointConfig,
user_agent: Option<UserAgent>,
request_payer: Option<String>,
bucket_owner: Option<String>,
max_attempts: Option<NonZeroUsize>,
}
impl Default for S3ClientConfig {
fn default() -> Self {
Self {
auth_config: Default::default(),
throughput_target_gbps: 10.0,
part_size: 8 * 1024 * 1024,
endpoint_config: EndpointConfig::new("us-east-1"),
user_agent: None,
request_payer: None,
bucket_owner: None,
max_attempts: None,
}
}
}
impl S3ClientConfig {
pub fn new() -> Self {
Self::default()
}
#[must_use = "S3ClientConfig follows a builder pattern"]
pub fn auth_config(mut self, auth_config: S3ClientAuthConfig) -> Self {
self.auth_config = auth_config;
self
}
#[must_use = "S3ClientConfig follows a builder pattern"]
pub fn part_size(mut self, part_size: usize) -> Self {
self.part_size = part_size;
self
}
#[must_use = "S3ClientConfig follows a builder pattern"]
pub fn throughput_target_gbps(mut self, throughput_target_gbps: f64) -> Self {
self.throughput_target_gbps = throughput_target_gbps;
self
}
#[must_use = "S3ClientConfig follows a builder pattern"]
pub fn endpoint_config(mut self, endpoint_config: EndpointConfig) -> Self {
self.endpoint_config = endpoint_config;
self
}
#[must_use = "S3ClientConfig follows a builder pattern"]
pub fn user_agent(mut self, user_agent: UserAgent) -> Self {
self.user_agent = Some(user_agent);
self
}
#[must_use = "S3ClientConfig follows a builder pattern"]
pub fn request_payer(mut self, request_payer: &str) -> Self {
self.request_payer = Some(request_payer.to_owned());
self
}
#[must_use = "S3ClientConfig follows a builder pattern"]
pub fn bucket_owner(mut self, bucket_owner: &str) -> Self {
self.bucket_owner = Some(bucket_owner.to_owned());
self
}
#[must_use = "S3ClientConfig follows a builder pattern"]
pub fn max_attempts(mut self, max_attempts: NonZeroUsize) -> Self {
self.max_attempts = Some(max_attempts);
self
}
}
#[derive(Debug, Clone, Default)]
pub enum S3ClientAuthConfig {
#[default]
Default,
NoSigning,
Profile(String),
Provider(CredentialsProvider),
}
#[derive(Debug, Clone)]
pub struct S3CrtClient {
inner: Arc<S3CrtClientInner>,
}
impl S3CrtClient {
pub fn new(config: S3ClientConfig) -> Result<Self, NewClientError> {
Ok(Self {
inner: Arc::new(S3CrtClientInner::new(config)?),
})
}
pub fn endpoint_config(&self) -> EndpointConfig {
self.inner.endpoint_config.clone()
}
#[doc(hidden)]
pub fn event_loop_group(&self) -> EventLoopGroup {
self.inner.event_loop_group.clone()
}
}
#[derive(Debug)]
struct S3CrtClientInner {
s3_client: Client,
event_loop_group: EventLoopGroup,
endpoint_config: EndpointConfig,
allocator: Allocator,
next_request_counter: AtomicU64,
user_agent_header: String,
request_payer: Option<String>,
part_size: usize,
bucket_owner: Option<String>,
credentials_provider: Option<CredentialsProvider>,
host_resolver: HostResolver,
}
impl S3CrtClientInner {
fn new(config: S3ClientConfig) -> Result<Self, NewClientError> {
let allocator = Allocator::default();
let mut event_loop_group = EventLoopGroup::new_default(&allocator, None, || {}).unwrap();
let resolver_options = HostResolverDefaultOptions {
max_entries: 8,
event_loop_group: &mut event_loop_group,
};
let mut host_resolver = HostResolver::new_default(&allocator, &resolver_options).unwrap();
let bootstrap_options = ClientBootstrapOptions {
event_loop_group: &mut event_loop_group,
host_resolver: &mut host_resolver,
};
let mut client_bootstrap = ClientBootstrap::new(&allocator, &bootstrap_options).unwrap();
let mut client_config = ClientConfig::new();
let retry_strategy = {
let mut retry_strategy_options = StandardRetryOptions::default(&mut event_loop_group);
let max_attempts = std::env::var("AWS_MAX_ATTEMPTS")
.ok()
.and_then(|s| s.parse::<usize>().ok())
.or_else(|| config.max_attempts.map(|m| m.get()))
.unwrap_or(3);
retry_strategy_options.backoff_retry_options.max_retries = max_attempts.saturating_sub(1);
retry_strategy_options.backoff_retry_options.backoff_scale_factor = Duration::from_millis(500);
retry_strategy_options.backoff_retry_options.jitter_mode = ExponentialBackoffJitterMode::Full;
RetryStrategy::standard(&allocator, &retry_strategy_options).unwrap()
};
trace!("constructing client with auth config {:?}", config.auth_config);
let credentials_provider = match config.auth_config {
S3ClientAuthConfig::Default => {
let credentials_chain_default_options = CredentialsProviderChainDefaultOptions {
bootstrap: &mut client_bootstrap,
};
CredentialsProvider::new_chain_default(&allocator, credentials_chain_default_options)
.map_err(NewClientError::ProviderFailure)?
}
S3ClientAuthConfig::NoSigning => {
CredentialsProvider::new_anonymous(&allocator).map_err(NewClientError::ProviderFailure)?
}
S3ClientAuthConfig::Profile(profile_name) => {
let credentials_profile_options = CredentialsProviderProfileOptions {
bootstrap: &mut client_bootstrap,
profile_name_override: &profile_name,
};
CredentialsProvider::new_profile(&allocator, credentials_profile_options)
.map_err(NewClientError::ProviderFailure)?
}
S3ClientAuthConfig::Provider(provider) => provider,
};
let endpoint_config = config.endpoint_config;
client_config.region(endpoint_config.get_region());
let signing_config = init_signing_config(
endpoint_config.get_region(),
credentials_provider.clone(),
None,
None,
None,
);
client_config.express_support(true);
client_config.signing_config(signing_config);
client_config
.client_bootstrap(client_bootstrap)
.retry_strategy(retry_strategy);
client_config.throughput_target_gbps(config.throughput_target_gbps);
if !(5 * 1024 * 1024..=5 * 1024 * 1024 * 1024).contains(&config.part_size) {
return Err(NewClientError::InvalidConfiguration(
"part size must be at between 5MiB and 5GiB".into(),
));
}
client_config.part_size(config.part_size);
let user_agent = config.user_agent.unwrap_or_else(|| UserAgent::new(None));
let user_agent_header = user_agent.build();
let s3_client = Client::new(&allocator, client_config).map_err(NewClientError::CrtError)?;
Ok(Self {
allocator,
s3_client,
event_loop_group,
endpoint_config,
next_request_counter: AtomicU64::new(0),
user_agent_header,
request_payer: config.request_payer,
part_size: config.part_size,
bucket_owner: config.bucket_owner,
credentials_provider: Some(credentials_provider),
host_resolver,
})
}
fn new_request_template(&self, method: &str, bucket: &str) -> Result<S3Message, ConstructionError> {
let endpoint = self.endpoint_config.resolve_for_bucket(bucket)?;
let uri = endpoint.uri()?;
trace!(?uri, "resolved endpoint");
let signing_config = if let Some(credentials_provider) = &self.credentials_provider {
let auth_scheme = match endpoint.auth_scheme() {
Ok(auth_scheme) => auth_scheme,
Err(e) => {
error!(error=?e, "no auth scheme for endpoint");
return Err(e.into());
}
};
trace!(?auth_scheme, "resolved auth scheme");
let algorithm = Some(auth_scheme.scheme_name());
let service = Some(auth_scheme.signing_name());
let use_double_uri_encode = Some(!auth_scheme.disable_double_encoding());
Some(init_signing_config(
auth_scheme.signing_region(),
credentials_provider.clone(),
algorithm,
service,
use_double_uri_encode,
))
} else {
None
};
let hostname = uri.host_name().to_str().unwrap();
let path_prefix = uri.path().to_os_string().into_string().unwrap();
let port = uri.host_port();
let hostname_header = if port > 0 {
format!("{}:{}", hostname, port)
} else {
hostname.to_string()
};
let mut message = Message::new_request(&self.allocator)?;
message.set_request_method(method)?;
message.add_header(&Header::new("Host", hostname_header))?;
message.add_header(&Header::new("accept", "application/xml"))?;
message.add_header(&Header::new("User-Agent", &self.user_agent_header))?;
if let Some(ref payer) = self.request_payer {
message.add_header(&Header::new("x-amz-request-payer", payer))?;
}
if let Some(ref owner) = self.bucket_owner {
message.add_header(&Header::new("x-amz-expected-bucket-owner", owner))?;
}
Ok(S3Message {
inner: message,
uri,
path_prefix,
checksum_config: None,
signing_config,
})
}
fn new_meta_request_options(message: S3Message, meta_request_type: MetaRequestType) -> MetaRequestOptions {
let mut options = MetaRequestOptions::new();
if let Some(checksum_config) = message.checksum_config {
options.checksum_config(checksum_config);
}
if let Some(signing_config) = message.signing_config {
options.signing_config(signing_config);
}
options
.message(message.inner)
.endpoint(message.uri)
.request_type(meta_request_type);
options
}
fn make_meta_request<T: Send + 'static, E: std::error::Error + Send + 'static>(
&self,
message: S3Message,
meta_request_type: MetaRequestType,
request_span: Span,
on_headers: impl FnMut(&Headers, i32) + Send + 'static,
on_body: impl FnMut(u64, &[u8]) + Send + 'static,
on_finish: impl FnOnce(&MetaRequestResult) -> Result<T, Option<ObjectClientError<E, S3RequestError>>>
+ Send
+ 'static,
) -> Result<S3HttpRequest<T, E>, S3RequestError> {
let options = Self::new_meta_request_options(message, meta_request_type);
self.make_meta_request_from_options(options, request_span, |_| {}, on_headers, on_body, on_finish)
}
fn make_meta_request_from_options<T: Send + 'static, E: std::error::Error + Send + 'static>(
&self,
mut options: MetaRequestOptions,
request_span: Span,
on_request_finish: impl Fn(&RequestMetrics) + Send + 'static,
mut on_headers: impl FnMut(&Headers, i32) + Send + 'static,
mut on_body: impl FnMut(u64, &[u8]) + Send + 'static,
on_meta_request_finish: impl FnOnce(&MetaRequestResult) -> Result<T, Option<ObjectClientError<E, S3RequestError>>>
+ Send
+ 'static,
) -> Result<S3HttpRequest<T, E>, S3RequestError> {
let (tx, rx) = oneshot::channel::<ObjectClientResult<T, E, S3RequestError>>();
let span_telemetry = request_span.clone();
let span_body = request_span.clone();
let span_finish = request_span;
let endpoint = options.get_endpoint().expect("S3Message always has an endpoint");
let hostname = endpoint.host_name().to_str().unwrap().to_owned();
let host_resolver = self.host_resolver.clone();
let start_time = Instant::now();
let first_body_part = Arc::new(AtomicBool::new(true));
let first_body_part_clone = Arc::clone(&first_body_part);
let total_bytes = Arc::new(AtomicU64::new(0));
let total_bytes_clone = Arc::clone(&total_bytes);
options
.on_telemetry(move |metrics| {
let _guard = span_telemetry.enter();
on_request_finish(metrics);
let http_status = metrics.status_code();
let request_canceled = metrics.is_canceled();
let request_failure = http_status.map(|status| !(200..299).contains(&status)).unwrap_or(!request_canceled);
let crt_error = Some(metrics.error()).filter(|e| e.is_err());
let request_type = request_type_to_metrics_string(metrics.request_type());
let request_id = metrics.request_id().unwrap_or_else(|| "<unknown>".into());
let duration = metrics.total_duration();
let ttfb = metrics.time_to_first_byte();
let range = metrics.response_headers().and_then(|headers| extract_range_header(&headers));
let message = if request_failure {
"S3 request failed"
} else if request_canceled {
"S3 request canceled"
} else {
"S3 request finished"
};
debug!(%request_type, ?crt_error, http_status, ?range, ?duration, ?ttfb, %request_id, "{}", message);
trace!(detailed_metrics=?metrics, "S3 request completed");
let op = span_telemetry.metadata().map(|m| m.name()).unwrap_or("unknown");
if let Some(ttfb) = ttfb {
metrics::histogram!("s3.requests.first_byte_latency_us", "op" => op, "type" => request_type).record(ttfb.as_micros() as f64);
}
metrics::histogram!("s3.requests.total_latency_us", "op" => op, "type" => request_type).record(duration.as_micros() as f64);
metrics::counter!("s3.requests", "op" => op, "type" => request_type).increment(1);
if request_failure {
metrics::counter!("s3.requests.failures", "op" => op, "type" => request_type, "status" => http_status.unwrap_or(-1).to_string()).increment(1);
} else if request_canceled {
metrics::counter!("s3.requests.canceled", "op" => op, "type" => request_type).increment(1);
}
})
.on_headers(move |headers, response_status| {
(on_headers)(headers, response_status);
})
.on_body(move |range_start, data| {
let _guard = span_body.enter();
if first_body_part.compare_exchange(true, false, Ordering::SeqCst, Ordering::SeqCst).ok() == Some(true) {
let latency = start_time.elapsed().as_micros() as f64;
let op = span_body.metadata().map(|m| m.name()).unwrap_or("unknown");
metrics::histogram!("s3.meta_requests.first_byte_latency_us", "op" => op).record(latency);
}
total_bytes.fetch_add(data.len() as u64, Ordering::SeqCst);
trace!(start = range_start, length = data.len(), "body part received");
(on_body)(range_start, data);
})
.on_finish(move |request_result| {
let _guard = span_finish.enter();
let op = span_finish.metadata().map(|m| m.name()).unwrap_or("unknown");
let duration = start_time.elapsed();
metrics::counter!("s3.meta_requests", "op" => op).increment(1);
metrics::histogram!("s3.meta_requests.total_latency_us", "op" => op).record(duration.as_micros() as f64);
if first_body_part_clone.compare_exchange(true, false, Ordering::SeqCst, Ordering::SeqCst).ok() == Some(true) {
let latency = duration.as_micros() as f64;
metrics::histogram!("s3.meta_requests.first_byte_latency_us", "op" => op).record(latency);
}
let total_bytes = total_bytes_clone.load(Ordering::SeqCst);
if op == "get_object" {
emit_throughput_metric(total_bytes, duration, op);
}
let hostname_awsstring = AwsString::from_str(&hostname, &Allocator::default());
if let Ok(host_count) = host_resolver.get_host_address_count(&hostname_awsstring, AddressKinds::a()) {
metrics::gauge!("s3.client.host_count", "host" => hostname).set(host_count as f64);
}
let status_code = request_result.response_status;
let log_level = if (200..=399).contains(&status_code) || status_code == 404 || request_result.is_canceled() {
tracing::Level::DEBUG
} else {
tracing::Level::WARN
};
let result = on_meta_request_finish(&request_result);
let result = result.map_err(|e| e.or_else(|| try_parse_generic_error(&request_result).map(ObjectClientError::ClientError)));
let result = match result {
Ok(t) => {
event!(log_level, ?duration, "meta request finished");
Ok(t)
}
Err(maybe_err) => {
let request_id = match &request_result.error_response_headers {
Some(headers) => headers.get("x-amz-request-id").map(|s| s.value().to_string_lossy().to_string()).ok(),
None => None,
};
let request_id = request_id.unwrap_or_else(|| "<unknown>".into());
let message = if request_result.is_canceled() {
"meta request canceled"
} else {
"meta request failed"
};
if let Some(error) = &maybe_err {
event!(log_level, ?duration, %request_id, ?error, message);
debug!("meta request result: {:?}", request_result);
} else {
event!(log_level, ?duration, %request_id, ?request_result, message);
}
if request_result.is_canceled() {
metrics::counter!("s3.meta_requests.canceled", "op" => op).increment(1);
} else {
let error_status = if request_result.response_status >= 100 {
request_result.response_status
} else {
-request_result.crt_error.raw_error()
};
metrics::counter!("s3.meta_requests.failures", "op" => op, "status" => format!("{error_status}")).increment(1);
}
Err(maybe_err.unwrap_or_else(|| ObjectClientError::ClientError(S3RequestError::ResponseError(request_result))))
}
};
let _ = tx.send(result);
});
let meta_request = self.s3_client.make_meta_request(options)?;
Self::poll_client_metrics(&self.s3_client);
Ok(S3HttpRequest {
receiver: rx,
meta_request,
})
}
fn make_simple_http_request<E: std::error::Error + Send + 'static>(
&self,
message: S3Message,
request_type: MetaRequestType,
request_span: Span,
on_error: impl FnOnce(&MetaRequestResult) -> Option<E> + Send + 'static,
) -> Result<S3HttpRequest<Vec<u8>, E>, S3RequestError> {
let options = Self::new_meta_request_options(message, request_type);
self.make_simple_http_request_from_options(options, request_span, |_| {}, on_error, |_, _| ())
}
fn make_simple_http_request_from_options<E: std::error::Error + Send + 'static>(
&self,
options: MetaRequestOptions,
request_span: Span,
on_request_finish: impl Fn(&RequestMetrics) + Send + 'static,
on_error: impl FnOnce(&MetaRequestResult) -> Option<E> + Send + 'static,
on_headers: impl FnMut(&Headers, i32) + Send + 'static,
) -> Result<S3HttpRequest<Vec<u8>, E>, S3RequestError> {
let body: Arc<Mutex<Vec<u8>>> = Default::default();
let body_clone = Arc::clone(&body);
self.make_meta_request_from_options(
options,
request_span,
on_request_finish,
on_headers,
move |offset, data| {
let mut body = body_clone.lock().unwrap();
assert_eq!(offset as usize, body.len());
body.extend_from_slice(data);
},
move |result| {
if result.is_err() {
Err(on_error(result).map(ObjectClientError::ServiceError))
} else {
Ok(std::mem::take(&mut *body.lock().unwrap()))
}
},
)
}
fn poll_client_metrics(s3_client: &Client) {
let metrics = s3_client.poll_client_metrics();
metrics::gauge!("s3.client.num_requests_being_processed").set(metrics.num_requests_tracked_requests as f64);
metrics::gauge!("s3.client.num_requests_being_prepared").set(metrics.num_requests_being_prepared as f64);
metrics::gauge!("s3.client.request_queue_size").set(metrics.request_queue_size as f64);
metrics::gauge!("s3.client.num_auto_default_network_io").set(metrics.num_auto_default_network_io as f64);
metrics::gauge!("s3.client.num_auto_ranged_get_network_io").set(metrics.num_auto_ranged_get_network_io as f64);
metrics::gauge!("s3.client.num_auto_ranged_put_network_io").set(metrics.num_auto_ranged_put_network_io as f64);
metrics::gauge!("s3.client.num_auto_ranged_copy_network_io")
.set(metrics.num_auto_ranged_copy_network_io as f64);
metrics::gauge!("s3.client.num_total_network_io").set(metrics.num_total_network_io() as f64);
metrics::gauge!("s3.client.num_requests_stream_queued_waiting")
.set(metrics.num_requests_stream_queued_waiting as f64);
metrics::gauge!("s3.client.num_requests_streaming_response")
.set(metrics.num_requests_streaming_response as f64);
let buffer_pool_stats = s3_client.poll_buffer_pool_usage_stats();
metrics::gauge!("s3.client.buffer_pool.mem_limit").set(buffer_pool_stats.mem_limit as f64);
metrics::gauge!("s3.client.buffer_pool.primary_cutoff").set(buffer_pool_stats.primary_cutoff as f64);
metrics::gauge!("s3.client.buffer_pool.primary_used").set(buffer_pool_stats.primary_used as f64);
metrics::gauge!("s3.client.buffer_pool.primary_allocated").set(buffer_pool_stats.primary_allocated as f64);
metrics::gauge!("s3.client.buffer_pool.primary_reserved").set(buffer_pool_stats.primary_reserved as f64);
metrics::gauge!("s3.client.buffer_pool.primary_num_blocks").set(buffer_pool_stats.primary_num_blocks as f64);
metrics::gauge!("s3.client.buffer_pool.secondary_reserved").set(buffer_pool_stats.secondary_reserved as f64);
metrics::gauge!("s3.client.buffer_pool.secondary_used").set(buffer_pool_stats.secondary_used as f64);
}
fn next_request_counter(&self) -> u64 {
self.next_request_counter.fetch_add(1, Ordering::SeqCst)
}
}
#[derive(Debug)]
struct S3Message {
inner: Message,
uri: Uri,
path_prefix: String,
checksum_config: Option<ChecksumConfig>,
signing_config: Option<SigningConfig>,
}
impl S3Message {
fn set_header(
&mut self,
header: &Header<impl AsRef<OsStr>, impl AsRef<OsStr>>,
) -> Result<(), mountpoint_s3_crt::common::error::Error> {
self.inner.set_header(header)
}
fn set_request_path_and_query<P: AsRef<OsStr>>(
&mut self,
path: impl AsRef<OsStr>,
query: impl AsRef<[(P, P)]>,
) -> Result<(), mountpoint_s3_crt::common::error::Error> {
const URLENCODE_QUERY_FRAGMENT: &AsciiSet =
&NON_ALPHANUMERIC.remove(b'-').remove(b'.').remove(b'_').remove(b'~');
const URLENCODE_PATH_FRAGMENT: &AsciiSet = &URLENCODE_QUERY_FRAGMENT.remove(b'/');
fn write_encoded_fragment(s: &mut OsString, piece: impl AsRef<OsStr>, encoding: &'static AsciiSet) {
let iter = percent_encode(piece.as_ref().as_bytes(), encoding);
s.extend(iter.map(|s| OsStr::from_bytes(s.as_bytes())));
}
let space_needed = self.path_prefix.len()
+ path.as_ref().len()
+ query
.as_ref()
.iter()
.map(|(key, value)| key.as_ref().len() + value.as_ref().len() + 2) .sum::<usize>();
let mut full_path = OsString::with_capacity(space_needed);
write_encoded_fragment(&mut full_path, &self.path_prefix, URLENCODE_PATH_FRAGMENT);
write_encoded_fragment(&mut full_path, &path, URLENCODE_PATH_FRAGMENT);
if !query.as_ref().is_empty() {
full_path.push("?");
for (i, (key, value)) in query.as_ref().iter().enumerate() {
if i != 0 {
full_path.push("&");
}
write_encoded_fragment(&mut full_path, key, URLENCODE_QUERY_FRAGMENT);
full_path.push("=");
write_encoded_fragment(&mut full_path, value, URLENCODE_QUERY_FRAGMENT);
}
}
self.inner.set_request_path(full_path)
}
fn set_request_path(&mut self, path: impl AsRef<OsStr>) -> Result<(), mountpoint_s3_crt::common::error::Error> {
self.set_request_path_and_query::<&str>(path, &[])
}
fn set_checksum_config(&mut self, checksum_config: Option<ChecksumConfig>) {
self.checksum_config = checksum_config;
}
}
#[derive(Debug)]
#[pin_project(PinnedDrop)]
struct S3HttpRequest<T, E> {
#[pin]
receiver: oneshot::Receiver<ObjectClientResult<T, E, S3RequestError>>,
meta_request: MetaRequest,
}
impl<T: Send, E: Send> Future for S3HttpRequest<T, E> {
type Output = ObjectClientResult<T, E, S3RequestError>;
fn poll(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
let this = self.project();
this.receiver.poll(cx).map(|result| {
result.unwrap_or_else(|err| {
Err(ObjectClientError::ClientError(S3RequestError::InternalError(Box::new(
err,
))))
})
})
}
}
#[pinned_drop]
impl<T, E> PinnedDrop for S3HttpRequest<T, E> {
fn drop(self: Pin<&mut Self>) {
self.meta_request.cancel();
}
}
#[derive(Error, Debug)]
#[non_exhaustive]
pub enum NewClientError {
#[error("invalid S3 endpoint")]
InvalidEndpoint(#[from] EndpointError),
#[error("invalid AWS credentials")]
ProviderFailure(#[source] mountpoint_s3_crt::common::error::Error),
#[error("invalid configuration: {0}")]
InvalidConfiguration(String),
#[error("Unknown CRT error")]
CrtError(#[source] mountpoint_s3_crt::common::error::Error),
}
#[derive(Error, Debug)]
pub enum S3RequestError {
#[error("Internal S3 client error")]
InternalError(#[source] Box<dyn std::error::Error + Send + Sync>),
#[error("Unknown CRT error")]
CrtError(#[from] mountpoint_s3_crt::common::error::Error),
#[error("Failed to construct request")]
ConstructionFailure(#[from] ConstructionError),
#[error("Unknown response error: {0:?}")]
ResponseError(MetaRequestResult),
#[error("Wrong region (expecting {0})")]
IncorrectRegion(String),
#[error("Forbidden: {0}")]
Forbidden(String),
#[error("No signing credentials found")]
NoSigningCredentials,
#[error("Request canceled")]
RequestCanceled,
}
impl S3RequestError {
fn construction_failure(inner: impl Into<ConstructionError>) -> Self {
S3RequestError::ConstructionFailure(inner.into())
}
}
#[derive(Error, Debug)]
pub enum ConstructionError {
#[error("Unknown CRT error")]
CrtError(#[from] mountpoint_s3_crt::common::error::Error),
#[error("Invalid S3 endpoint")]
InvalidEndpoint(#[from] EndpointError),
}
fn request_type_to_metrics_string(request_type: RequestType) -> &'static str {
match request_type {
RequestType::Unknown => "Default",
RequestType::HeadObject => "HeadObject",
RequestType::GetObject => "GetObject",
RequestType::ListParts => "ListParts",
RequestType::CreateMultipartUpload => "CreateMultipartUpload",
RequestType::UploadPart => "UploadPart",
RequestType::AbortMultipartUpload => "AbortMultipartUpload",
RequestType::CompleteMultipartUpload => "CompleteMultipartUpload",
RequestType::UploadPartCopy => "UploadPartCopy",
}
}
fn extract_range_header(headers: &Headers) -> Option<Range<u64>> {
let header = headers.get("Content-Range").ok()?;
let value = header.value().to_str()?;
if !value.starts_with("bytes ") {
return None;
}
let (_, value) = value.split_at("bytes ".len());
let (range, _) = value.split_once('/')?;
let (start, end) = range.split_once('-')?;
let start = start.parse::<u64>().ok()?;
let end = end.parse::<u64>().ok()?;
Some(start..end + 1)
}
fn try_parse_generic_error(request_result: &MetaRequestResult) -> Option<S3RequestError> {
fn try_parse_redirect(request_result: &MetaRequestResult) -> Option<S3RequestError> {
let headers = request_result.error_response_headers.as_ref()?;
let region_header = headers.get("x-amz-bucket-region").ok()?;
let region = region_header.value().to_owned().into_string().ok()?;
Some(S3RequestError::IncorrectRegion(region))
}
fn try_parse_forbidden(request_result: &MetaRequestResult) -> Option<S3RequestError> {
let Some(body) = request_result.error_response_body.as_ref() else {
return Some(S3RequestError::Forbidden("<no message>".to_owned()));
};
let error_elem = xmltree::Element::parse(body.as_bytes()).ok()?;
let error_code = error_elem.get_child("Code")?;
let error_code_str = error_code.get_text()?;
if request_result.response_status == 403
|| matches!(
error_code_str.deref(),
"AccessDenied" | "InvalidToken" | "ExpiredToken" | "SignatureDoesNotMatch"
)
{
let message = error_elem
.get_child("Message")
.and_then(|e| e.get_text())
.unwrap_or(error_code_str);
Some(S3RequestError::Forbidden(message.into_owned()))
} else {
None
}
}
fn try_parse_no_credentials(request_result: &MetaRequestResult) -> Option<S3RequestError> {
let crt_error_code = request_result.crt_error.raw_error();
if crt_error_code == mountpoint_s3_crt_sys::aws_auth_errors::AWS_AUTH_SIGNING_NO_CREDENTIALS as i32 {
Some(S3RequestError::NoSigningCredentials)
} else {
Some(S3RequestError::CrtError(crt_error_code.into()))
}
}
fn try_parse_canceled_request(request_result: &MetaRequestResult) -> Option<S3RequestError> {
request_result.is_canceled().then_some(S3RequestError::RequestCanceled)
}
match request_result.response_status {
301 => try_parse_redirect(request_result),
400 => try_parse_forbidden(request_result).or_else(|| try_parse_redirect(request_result)),
403 => try_parse_forbidden(request_result),
0 => try_parse_no_credentials(request_result).or_else(|| try_parse_canceled_request(request_result)),
_ => None,
}
}
fn emit_throughput_metric(bytes: u64, duration: Duration, op: &'static str) {
let throughput_mbps = bytes as f64 / 1024.0 / 1024.0 / duration.as_secs_f64();
const MEGABYTE: u64 = 1024 * 1024;
let bucket = if bytes < MEGABYTE {
"<1MiB"
} else if bytes <= 16 * MEGABYTE {
"1-16MiB"
} else {
">16MiB"
};
metrics::histogram!("s3.meta_requests.throughput_mibs", "op" => op, "size" => bucket).record(throughput_mbps);
}
#[cfg_attr(not(docs_rs), async_trait)]
impl ObjectClient for S3CrtClient {
type GetObjectResult = S3GetObjectRequest;
type PutObjectRequest = S3PutObjectRequest;
type ClientError = S3RequestError;
fn part_size(&self) -> Option<usize> {
Some(self.inner.part_size)
}
async fn delete_object(
&self,
bucket: &str,
key: &str,
) -> ObjectClientResult<DeleteObjectResult, DeleteObjectError, Self::ClientError> {
self.delete_object(bucket, key).await
}
async fn get_object(
&self,
bucket: &str,
key: &str,
range: Option<Range<u64>>,
if_match: Option<ETag>,
) -> ObjectClientResult<Self::GetObjectResult, GetObjectError, Self::ClientError> {
self.get_object(bucket, key, range, if_match)
}
async fn list_objects(
&self,
bucket: &str,
continuation_token: Option<&str>,
delimiter: &str,
max_keys: usize,
prefix: &str,
) -> ObjectClientResult<ListObjectsResult, ListObjectsError, Self::ClientError> {
self.list_objects(bucket, continuation_token, delimiter, max_keys, prefix)
.await
}
async fn head_object(
&self,
bucket: &str,
key: &str,
) -> ObjectClientResult<HeadObjectResult, HeadObjectError, Self::ClientError> {
self.head_object(bucket, key).await
}
async fn put_object(
&self,
bucket: &str,
key: &str,
params: &PutObjectParams,
) -> ObjectClientResult<Self::PutObjectRequest, PutObjectError, Self::ClientError> {
self.put_object(bucket, key, params).await
}
async fn get_object_attributes(
&self,
bucket: &str,
key: &str,
max_parts: Option<usize>,
part_number_marker: Option<usize>,
object_attributes: &[ObjectAttribute],
) -> ObjectClientResult<GetObjectAttributesResult, GetObjectAttributesError, Self::ClientError> {
self.get_object_attributes(bucket, key, max_parts, part_number_marker, object_attributes)
.await
}
}
#[cfg(test)]
mod tests {
use mountpoint_s3_crt::common::error::Error;
use std::assert_eq;
use super::*;
use test_case::test_case;
#[test_case(4 * 1024 * 1024; "less than 5MiB")]
#[test_case(6 * 1024 * 1024 * 1024; "greater than 5GiB")]
fn client_new_fails_with_invalid_part_size(part_size: usize) {
let config = S3ClientConfig {
part_size,
..Default::default()
};
S3CrtClient::new(config).expect_err("creating a new client should fail");
}
#[test]
fn test_user_agent_with_prefix() {
let user_agent_prefix = String::from("someprefix");
let expected_user_agent = "someprefix mountpoint-s3-client/";
let config = S3ClientConfig {
user_agent: Some(UserAgent::new(Some(user_agent_prefix))),
..Default::default()
};
let client = S3CrtClient::new(config).expect("Create test client");
let mut message = client
.inner
.new_request_template("GET", "doc-example-bucket")
.expect("new request template expected");
let headers = message.inner.get_headers().expect("Expected a block of HTTP headers");
let user_agent_header = headers
.get("User-Agent")
.expect("User Agent Header expected with given prefix");
let user_agent_header_value = user_agent_header.value();
assert!(user_agent_header_value
.to_string_lossy()
.starts_with(expected_user_agent));
}
#[test]
fn test_user_agent_without_prefix() {
let expected_user_agent = "mountpoint-s3-client/";
let config: S3ClientConfig = Default::default();
let client = S3CrtClient::new(config).expect("Create test client");
let mut message = client
.inner
.new_request_template("GET", "doc-example-bucket")
.expect("new request template expected");
let headers = message.inner.get_headers().expect("Expected a block of HTTP headers");
let user_agent_header = headers
.get("User-Agent")
.expect("User Agent Header expected with given prefix");
let user_agent_header_value = user_agent_header.value();
assert!(user_agent_header_value
.to_string_lossy()
.starts_with(expected_user_agent));
}
#[test_case("bytes 200-1000/67589" => Some(200..1001))]
#[test_case("bytes 200-1000/*" => Some(200..1001))]
#[test_case("bytes 200-1000" => None)]
#[test_case("bytes */67589" => None)]
#[test_case("octets 200-1000]" => None)]
fn parse_content_range(range: &str) -> Option<Range<u64>> {
let mut headers = Headers::new(&Allocator::default()).unwrap();
let header = Header::new("Content-Range", range);
headers.add_header(&header).unwrap();
extract_range_header(&headers)
}
#[test]
fn test_expected_bucket_owner() {
let expected_bucket_owner = "111122223333";
let config: S3ClientConfig = S3ClientConfig::new().bucket_owner("111122223333");
let client = S3CrtClient::new(config).expect("Create test client");
let mut message = client
.inner
.new_request_template("GET", "doc-example-bucket")
.expect("new request template expected");
let headers = message.inner.get_headers().expect("Expected a block of HTTP headers");
let expected_bucket_owner_header = headers
.get("x-amz-expected-bucket-owner")
.expect("the headers should contain x-amz-expected-bucket-owner");
let expected_bucket_owner_value = expected_bucket_owner_header.value();
assert!(expected_bucket_owner_value
.to_string_lossy()
.starts_with(expected_bucket_owner));
}
fn make_result(
response_status: i32,
body: impl Into<OsString>,
bucket_region_header: Option<&str>,
) -> MetaRequestResult {
let error_response_headers = bucket_region_header.map(|h| {
let mut headers = Headers::new(&Allocator::default()).unwrap();
headers.add_header(&Header::new("x-amz-bucket-region", h)).unwrap();
headers
});
MetaRequestResult {
response_status,
crt_error: 1i32.into(),
error_response_headers,
error_response_body: Some(body.into()),
}
}
#[test]
fn parse_301_redirect() {
let body = br#"<?xml version="1.0" encoding="UTF-8"?><Error><Code>PermanentRedirect</Code><Message>The bucket you are attempting to access must be addressed using the specified endpoint. Please send all future requests to this endpoint.</Message><Endpoint>DOC-EXAMPLE-BUCKET.s3-us-west-2.amazonaws.com</Endpoint><Bucket>DOC-EXAMPLE-BUCKET</Bucket><RequestId>CM0Z9YFABRVSWXDJ</RequestId><HostId>HHmbUixasrJ02DlkOSCvJId897Jm0ERHuE2XMkSn2Oax1J/ad2+AU9nFrODN1ay13cWFgIAYBnI=</HostId></Error>"#;
let result = make_result(301, OsStr::from_bytes(&body[..]), Some("us-west-2"));
let result = try_parse_generic_error(&result);
let Some(S3RequestError::IncorrectRegion(region)) = result else {
panic!("wrong result, got: {:?}", result);
};
assert_eq!(region, "us-west-2");
}
#[test]
fn parse_403_access_denied() {
let body = br#"<?xml version="1.0" encoding="UTF-8"?><Error><Code>AccessDenied</Code><Message>Access Denied</Message><RequestId>CM0R497NB0WAQ977</RequestId><HostId>w1TqUKGaIuNAIgzqm/L2azuzgEBINxTngWPbV1iH2IvpLsVCCTKHJTh4HsGp4JnggHqVkA+KN1MGqHDw1+WEuA==</HostId></Error>"#;
let result = make_result(403, OsStr::from_bytes(&body[..]), None);
let result = try_parse_generic_error(&result);
let Some(S3RequestError::Forbidden(message)) = result else {
panic!("wrong result, got: {:?}", result);
};
assert_eq!(message, "Access Denied");
}
#[test]
fn parse_400_invalid_token() {
let body = br#"<?xml version="1.0" encoding="UTF-8"?><Error><Code>InvalidToken</Code><Message>The provided token is malformed or otherwise invalid.</Message><Token-0>THEREALTOKENGOESHERE</Token-0><RequestId>CBFNVADDAZ8661HK</RequestId><HostId>rb5dpgYeIFxi8p5BzVK8s8wG/nQ4a7C5kMBp/KWIT4bvOUihugpssMTy7xS0mispbz6IIaX8W1g=</HostId></Error>"#;
let result = make_result(400, OsStr::from_bytes(&body[..]), None);
let result = try_parse_generic_error(&result);
let Some(S3RequestError::Forbidden(message)) = result else {
panic!("wrong result, got: {:?}", result);
};
assert_eq!(message, "The provided token is malformed or otherwise invalid.");
}
#[test]
fn parse_400_expired_token() {
let body = br#"<?xml version="1.0" encoding="UTF-8"?><Error><Code>ExpiredToken</Code><Message>The provided token has expired.</Message><Token-0>THEREALTOKENGOESHERE</Token-0><RequestId>RFXW0E15XSRPJYSW</RequestId><HostId>djitP7S+g43JSzR4pMOJpOO3RYpQUOUsmD4AqhRe3v24+JB/c+vwOEZgI8A35KDUe1cqQ5yKHwg=</HostId></Error>"#;
let result = make_result(400, OsStr::from_bytes(&body[..]), None);
let result = try_parse_generic_error(&result);
let Some(S3RequestError::Forbidden(message)) = result else {
panic!("wrong result, got: {:?}", result);
};
assert_eq!(message, "The provided token has expired.");
}
#[test]
fn parse_400_redirect() {
let body = br#"<?xml version="1.0" encoding="UTF-8"?><Error><Code>AuthorizationHeaderMalformed</Code><Message>The authorization header is malformed; the region \'us-east-1\' is wrong; expecting \'us-west-2\'</Message><Region>us-west-2</Region><RequestId>VR3NH4JF5F39GB66</RequestId><HostId>ZDzYFC1w0E5K34+ZCAnvh9ZiGaAhvx5COyZVYTUnKvSP/694xCiXmJ2AEGZd5T1Epy9vB4EOOjk=</HostId></Error>"#;
let result = make_result(400, OsStr::from_bytes(&body[..]), Some("us-west-2"));
let result = try_parse_generic_error(&result);
let Some(S3RequestError::IncorrectRegion(region)) = result else {
panic!("wrong result, got: {:?}", result);
};
assert_eq!(region, "us-west-2");
}
#[test]
fn parse_403_signature_does_not_match() {
let body = br#"<?xml version="1.0" encoding="UTF-8"?><Error><Code>SignatureDoesNotMatch</Code><Message>The request signature we calculated does not match the signature you provided. Check your key and signing method.</Message><AWSAccessKeyId>ASIASMEXAMPLE0000000</AWSAccessKeyId><StringToSign>EXAMPLE</StringToSign><SignatureProvided>EXAMPLE</SignatureProvided><StringToSignBytes>EXAMPLE</StringToSignBytes><CanonicalRequest>EXAMPLE</CanonicalRequest><CanonicalRequestBytes>EXAMPLE</CanonicalRequestBytes><RequestId>A1F516XX5M8AATSQ</RequestId><HostId>qs9dULIp5ABM7U+H8nGfzKtMYTxvqxIVvOYZ8lEFBDyTF4Fe+876Y4bLptG4mb+PTZFyG4yaUjg=</HostId></Error>"#;
let result = make_result(403, OsStr::from_bytes(&body[..]), None);
let result = try_parse_generic_error(&result);
let Some(S3RequestError::Forbidden(message)) = result else {
panic!("wrong result, got: {:?}", result);
};
assert_eq!(message, "The request signature we calculated does not match the signature you provided. Check your key and signing method.");
}
#[test]
fn parse_403_made_up_error() {
let body = br#"<?xml version="1.0" encoding="UTF-8"?><Error><Code>NotARealError</Code><Message>This error is made up.</Message><RequestId>CM0R497NB0WAQ977</RequestId><HostId>w1TqUKGaIuNAIgzqm/L2azuzgEBINxTngWPbV1iH2IvpLsVCCTKHJTh4HsGp4JnggHqVkA+KN1MGqHDw1+WEuA==</HostId></Error>"#;
let result = make_result(403, OsStr::from_bytes(&body[..]), None);
let result = try_parse_generic_error(&result);
let Some(S3RequestError::Forbidden(message)) = result else {
panic!("wrong result, got: {:?}", result);
};
assert_eq!(message, "This error is made up.");
}
fn make_crt_error_result(response_status: i32, crt_error: Error) -> MetaRequestResult {
MetaRequestResult {
response_status,
crt_error,
error_response_headers: None,
error_response_body: None,
}
}
#[test]
fn parse_no_signing_credential_error() {
let error_code = mountpoint_s3_crt_sys::aws_auth_errors::AWS_AUTH_SIGNING_NO_CREDENTIALS as i32;
let result = make_crt_error_result(0, error_code.into());
let result = try_parse_generic_error(&result);
let Some(S3RequestError::NoSigningCredentials) = result else {
panic!("wrong result, got: {:?}", result);
};
}
#[test]
fn parse_test_other_crt_error() {
let error_code = mountpoint_s3_crt_sys::aws_auth_errors::AWS_AUTH_SIGNING_UNSUPPORTED_ALGORITHM as i32;
let result = make_crt_error_result(0, error_code.into());
let result = try_parse_generic_error(&result);
let Some(S3RequestError::CrtError(error)) = result else {
panic!("wrong result, got: {:?}", result);
};
assert_eq!(error, error_code.into());
}
}