use std::collections::BTreeSet;
use std::fs::File;
use std::io::{self, BufReader};
use std::ops::Range;
use std::sync::Arc;
use async_trait::async_trait;
use bytes::{Buf, Bytes};
use chrono::{DateTime, Utc};
use futures::future::BoxFuture;
use futures::{stream::BoxStream, StreamExt, TryStreamExt};
use percent_encoding::{percent_encode, NON_ALPHANUMERIC};
use reqwest::header::RANGE;
use reqwest::{header, Client, Method, Response, StatusCode};
use snafu::{ResultExt, Snafu};
use tokio::io::AsyncWrite;
use crate::client::retry::RetryExt;
use crate::{
client::{oauth::OAuthProvider, token::TokenCache},
multipart::{CloudMultiPartUpload, CloudMultiPartUploadImpl, UploadPart},
path::{Path, DELIMITER},
util::{format_http_range, format_prefix},
GetResult, ListResult, MultipartId, ObjectMeta, ObjectStore, Result, RetryConfig,
};
#[derive(Debug, Snafu)]
enum Error {
#[snafu(display("Unable to open service account file: {}", source))]
OpenCredentials { source: std::io::Error },
#[snafu(display("Unable to decode service account file: {}", source))]
DecodeCredentials { source: serde_json::Error },
#[snafu(display("Got invalid XML response for {} {}: {}", method, url, source))]
InvalidXMLResponse {
source: quick_xml::de::DeError,
method: String,
url: String,
data: Bytes,
},
#[snafu(display("Error performing list request: {}", source))]
ListRequest { source: reqwest::Error },
#[snafu(display("Error performing get request {}: {}", path, source))]
GetRequest {
source: reqwest::Error,
path: String,
},
#[snafu(display("Error performing delete request {}: {}", path, source))]
DeleteRequest {
source: reqwest::Error,
path: String,
},
#[snafu(display("Error performing copy request {}: {}", path, source))]
CopyRequest {
source: reqwest::Error,
path: String,
},
#[snafu(display("Error performing put request: {}", source))]
PutRequest { source: reqwest::Error },
#[snafu(display("Error decoding object size: {}", source))]
InvalidSize { source: std::num::ParseIntError },
#[snafu(display("Missing bucket name"))]
MissingBucketName {},
#[snafu(display("Missing service account path"))]
MissingServiceAccountPath,
}
impl From<Error> for super::Error {
fn from(err: Error) -> Self {
match err {
Error::GetRequest { source, path }
| Error::DeleteRequest { source, path }
| Error::CopyRequest { source, path }
if matches!(source.status(), Some(StatusCode::NOT_FOUND)) =>
{
Self::NotFound {
path,
source: Box::new(source),
}
}
_ => Self::Generic {
store: "GCS",
source: Box::new(err),
},
}
}
}
#[derive(serde::Deserialize, Debug)]
struct ServiceAccountCredentials {
pub private_key: String,
pub client_email: String,
#[serde(default = "default_gcs_base_url")]
pub gcs_base_url: String,
#[serde(default = "default_disable_oauth")]
pub disable_oauth: bool,
}
fn default_gcs_base_url() -> String {
"https://storage.googleapis.com".to_owned()
}
fn default_disable_oauth() -> bool {
false
}
#[derive(serde::Deserialize, Debug)]
#[serde(rename_all = "camelCase")]
struct ListResponse {
next_page_token: Option<String>,
#[serde(default)]
prefixes: Vec<String>,
#[serde(default)]
items: Vec<Object>,
}
#[derive(serde::Deserialize, Debug)]
struct Object {
name: String,
size: String,
updated: DateTime<Utc>,
}
#[derive(serde::Deserialize, Debug)]
#[serde(rename_all = "PascalCase")]
struct InitiateMultipartUploadResult {
upload_id: String,
}
#[derive(serde::Serialize, Debug)]
#[serde(rename_all = "PascalCase", rename(serialize = "Part"))]
struct MultipartPart {
#[serde(rename = "$unflatten=PartNumber")]
part_number: usize,
#[serde(rename = "$unflatten=ETag")]
e_tag: String,
}
#[derive(serde::Serialize, Debug)]
#[serde(rename_all = "PascalCase")]
struct CompleteMultipartUpload {
#[serde(rename = "Part", default)]
parts: Vec<MultipartPart>,
}
#[derive(Debug)]
pub struct GoogleCloudStorage {
client: Arc<GoogleCloudStorageClient>,
}
impl std::fmt::Display for GoogleCloudStorage {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
write!(f, "GoogleCloudStorage({})", self.client.bucket_name)
}
}
#[derive(Debug)]
struct GoogleCloudStorageClient {
client: Client,
base_url: String,
oauth_provider: Option<OAuthProvider>,
token_cache: TokenCache<String>,
bucket_name: String,
bucket_name_encoded: String,
retry_config: RetryConfig,
max_list_results: Option<String>,
}
impl GoogleCloudStorageClient {
async fn get_token(&self) -> Result<String> {
if let Some(oauth_provider) = &self.oauth_provider {
Ok(self
.token_cache
.get_or_insert_with(|| {
oauth_provider.fetch_token(&self.client, &self.retry_config)
})
.await?)
} else {
Ok("".to_owned())
}
}
fn object_url(&self, path: &Path) -> String {
let encoded =
percent_encoding::utf8_percent_encode(path.as_ref(), NON_ALPHANUMERIC);
format!(
"{}/storage/v1/b/{}/o/{}",
self.base_url, self.bucket_name_encoded, encoded
)
}
async fn get_request(
&self,
path: &Path,
range: Option<Range<usize>>,
head: bool,
) -> Result<Response> {
let token = self.get_token().await?;
let url = self.object_url(path);
let mut builder = self.client.request(Method::GET, url);
if let Some(range) = range {
builder = builder.header(RANGE, format_http_range(range));
}
let alt = match head {
true => "json",
false => "media",
};
let response = builder
.bearer_auth(token)
.query(&[("alt", alt)])
.send_retry(&self.retry_config)
.await
.context(GetRequestSnafu {
path: path.as_ref(),
})?
.error_for_status()
.context(GetRequestSnafu {
path: path.as_ref(),
})?;
Ok(response)
}
async fn put_request(&self, path: &Path, payload: Bytes) -> Result<()> {
let token = self.get_token().await?;
let url = format!(
"{}/upload/storage/v1/b/{}/o",
self.base_url, self.bucket_name_encoded
);
self.client
.request(Method::POST, url)
.bearer_auth(token)
.header(header::CONTENT_TYPE, "application/octet-stream")
.header(header::CONTENT_LENGTH, payload.len())
.query(&[("uploadType", "media"), ("name", path.as_ref())])
.body(payload)
.send_retry(&self.retry_config)
.await
.context(PutRequestSnafu)?
.error_for_status()
.context(PutRequestSnafu)?;
Ok(())
}
async fn multipart_initiate(&self, path: &Path) -> Result<MultipartId> {
let token = self.get_token().await?;
let url = format!("{}/{}/{}", self.base_url, self.bucket_name_encoded, path);
let response = self
.client
.request(Method::POST, &url)
.bearer_auth(token)
.header(header::CONTENT_TYPE, "application/octet-stream")
.header(header::CONTENT_LENGTH, "0")
.query(&[("uploads", "")])
.send_retry(&self.retry_config)
.await
.context(PutRequestSnafu)?
.error_for_status()
.context(PutRequestSnafu)?;
let data = response.bytes().await.context(PutRequestSnafu)?;
let result: InitiateMultipartUploadResult = quick_xml::de::from_reader(
data.as_ref().reader(),
)
.context(InvalidXMLResponseSnafu {
method: "POST".to_string(),
url,
data,
})?;
Ok(result.upload_id)
}
async fn multipart_cleanup(
&self,
path: &str,
multipart_id: &MultipartId,
) -> Result<()> {
let token = self.get_token().await?;
let url = format!("{}/{}/{}", self.base_url, self.bucket_name_encoded, path);
self.client
.request(Method::DELETE, &url)
.bearer_auth(token)
.header(header::CONTENT_TYPE, "application/octet-stream")
.header(header::CONTENT_LENGTH, "0")
.query(&[("uploadId", multipart_id)])
.send_retry(&self.retry_config)
.await
.context(PutRequestSnafu)?
.error_for_status()
.context(PutRequestSnafu)?;
Ok(())
}
async fn delete_request(&self, path: &Path) -> Result<()> {
let token = self.get_token().await?;
let url = self.object_url(path);
let builder = self.client.request(Method::DELETE, url);
builder
.bearer_auth(token)
.send_retry(&self.retry_config)
.await
.context(DeleteRequestSnafu {
path: path.as_ref(),
})?
.error_for_status()
.context(DeleteRequestSnafu {
path: path.as_ref(),
})?;
Ok(())
}
async fn copy_request(
&self,
from: &Path,
to: &Path,
if_not_exists: bool,
) -> Result<()> {
let token = self.get_token().await?;
let source =
percent_encoding::utf8_percent_encode(from.as_ref(), NON_ALPHANUMERIC);
let destination =
percent_encoding::utf8_percent_encode(to.as_ref(), NON_ALPHANUMERIC);
let url = format!(
"{}/storage/v1/b/{}/o/{}/copyTo/b/{}/o/{}",
self.base_url,
self.bucket_name_encoded,
source,
self.bucket_name_encoded,
destination
);
let mut builder = self.client.request(Method::POST, url);
if if_not_exists {
builder = builder.query(&[("ifGenerationMatch", "0")]);
}
builder
.bearer_auth(token)
.send_retry(&self.retry_config)
.await
.context(CopyRequestSnafu {
path: from.as_ref(),
})?
.error_for_status()
.context(CopyRequestSnafu {
path: from.as_ref(),
})?;
Ok(())
}
async fn list_request(
&self,
prefix: Option<&str>,
delimiter: bool,
page_token: Option<&str>,
) -> Result<ListResponse> {
let token = self.get_token().await?;
let url = format!(
"{}/storage/v1/b/{}/o",
self.base_url, self.bucket_name_encoded
);
let mut query = Vec::with_capacity(4);
if delimiter {
query.push(("delimiter", DELIMITER))
}
if let Some(prefix) = &prefix {
query.push(("prefix", prefix))
}
if let Some(page_token) = page_token {
query.push(("pageToken", page_token))
}
if let Some(max_results) = &self.max_list_results {
query.push(("maxResults", max_results))
}
let response: ListResponse = self
.client
.request(Method::GET, url)
.query(&query)
.bearer_auth(token)
.send_retry(&self.retry_config)
.await
.context(ListRequestSnafu)?
.error_for_status()
.context(ListRequestSnafu)?
.json()
.await
.context(ListRequestSnafu)?;
Ok(response)
}
fn list_paginated(
&self,
prefix: Option<&Path>,
delimiter: bool,
) -> Result<BoxStream<'_, Result<ListResponse>>> {
let prefix = format_prefix(prefix);
enum ListState {
Start,
HasMore(String),
Done,
}
Ok(futures::stream::unfold(ListState::Start, move |state| {
let prefix = prefix.clone();
async move {
let page_token = match &state {
ListState::Start => None,
ListState::HasMore(page_token) => Some(page_token.as_str()),
ListState::Done => {
return None;
}
};
let resp = match self
.list_request(prefix.as_deref(), delimiter, page_token)
.await
{
Ok(resp) => resp,
Err(e) => return Some((Err(e), state)),
};
let next_state = match &resp.next_page_token {
Some(token) => ListState::HasMore(token.clone()),
None => ListState::Done,
};
Some((Ok(resp), next_state))
}
})
.boxed())
}
}
fn reqwest_error_as_io(err: reqwest::Error) -> io::Error {
if err.is_builder() || err.is_request() {
io::Error::new(io::ErrorKind::InvalidInput, err)
} else if err.is_status() {
match err.status() {
Some(StatusCode::NOT_FOUND) => io::Error::new(io::ErrorKind::NotFound, err),
Some(StatusCode::BAD_REQUEST) => {
io::Error::new(io::ErrorKind::InvalidInput, err)
}
Some(_) => io::Error::new(io::ErrorKind::Other, err),
None => io::Error::new(io::ErrorKind::Other, err),
}
} else if err.is_timeout() {
io::Error::new(io::ErrorKind::TimedOut, err)
} else if err.is_connect() {
io::Error::new(io::ErrorKind::NotConnected, err)
} else {
io::Error::new(io::ErrorKind::Other, err)
}
}
struct GCSMultipartUpload {
client: Arc<GoogleCloudStorageClient>,
encoded_path: String,
multipart_id: MultipartId,
}
impl CloudMultiPartUploadImpl for GCSMultipartUpload {
fn put_multipart_part(
&self,
buf: Vec<u8>,
part_idx: usize,
) -> BoxFuture<'static, Result<(usize, UploadPart), io::Error>> {
let upload_id = self.multipart_id.clone();
let url = format!(
"{}/{}/{}",
self.client.base_url, self.client.bucket_name_encoded, self.encoded_path
);
let client = Arc::clone(&self.client);
Box::pin(async move {
let token = client
.get_token()
.await
.map_err(|err| io::Error::new(io::ErrorKind::Other, err))?;
let response = client
.client
.request(Method::PUT, &url)
.bearer_auth(token)
.query(&[
("partNumber", format!("{}", part_idx + 1)),
("uploadId", upload_id),
])
.header(header::CONTENT_TYPE, "application/octet-stream")
.header(header::CONTENT_LENGTH, format!("{}", buf.len()))
.body(buf)
.send_retry(&client.retry_config)
.await
.map_err(reqwest_error_as_io)?
.error_for_status()
.map_err(reqwest_error_as_io)?;
let content_id = response
.headers()
.get("ETag")
.ok_or_else(|| {
io::Error::new(
io::ErrorKind::InvalidData,
"response headers missing ETag",
)
})?
.to_str()
.map_err(|err| io::Error::new(io::ErrorKind::InvalidData, err))?
.to_string();
Ok((part_idx, UploadPart { content_id }))
})
}
fn complete(
&self,
completed_parts: Vec<Option<UploadPart>>,
) -> BoxFuture<'static, Result<(), io::Error>> {
let client = Arc::clone(&self.client);
let upload_id = self.multipart_id.clone();
let url = format!(
"{}/{}/{}",
self.client.base_url, self.client.bucket_name_encoded, self.encoded_path
);
Box::pin(async move {
let parts: Vec<MultipartPart> = completed_parts
.into_iter()
.enumerate()
.map(|(part_number, maybe_part)| match maybe_part {
Some(part) => Ok(MultipartPart {
e_tag: part.content_id,
part_number: part_number + 1,
}),
None => Err(io::Error::new(
io::ErrorKind::Other,
format!("Missing information for upload part {:?}", part_number),
)),
})
.collect::<Result<Vec<MultipartPart>, io::Error>>()?;
let token = client
.get_token()
.await
.map_err(|err| io::Error::new(io::ErrorKind::Other, err))?;
let upload_info = CompleteMultipartUpload { parts };
let data = quick_xml::se::to_string(&upload_info)
.map_err(|err| io::Error::new(io::ErrorKind::Other, err))?
.replace(""", "\"");
client
.client
.request(Method::POST, &url)
.bearer_auth(token)
.query(&[("uploadId", upload_id)])
.body(data)
.send_retry(&client.retry_config)
.await
.map_err(reqwest_error_as_io)?
.error_for_status()
.map_err(reqwest_error_as_io)?;
Ok(())
})
}
}
#[async_trait]
impl ObjectStore for GoogleCloudStorage {
async fn put(&self, location: &Path, bytes: Bytes) -> Result<()> {
self.client.put_request(location, bytes).await
}
async fn put_multipart(
&self,
location: &Path,
) -> Result<(MultipartId, Box<dyn AsyncWrite + Unpin + Send>)> {
let upload_id = self.client.multipart_initiate(location).await?;
let encoded_path =
percent_encode(location.to_string().as_bytes(), NON_ALPHANUMERIC).to_string();
let inner = GCSMultipartUpload {
client: Arc::clone(&self.client),
encoded_path,
multipart_id: upload_id.clone(),
};
Ok((upload_id, Box::new(CloudMultiPartUpload::new(inner, 8))))
}
async fn abort_multipart(
&self,
location: &Path,
multipart_id: &MultipartId,
) -> Result<()> {
self.client
.multipart_cleanup(location.as_ref(), multipart_id)
.await?;
Ok(())
}
async fn get(&self, location: &Path) -> Result<GetResult> {
let response = self.client.get_request(location, None, false).await?;
let stream = response
.bytes_stream()
.map_err(|source| crate::Error::Generic {
store: "GCS",
source: Box::new(source),
})
.boxed();
Ok(GetResult::Stream(stream))
}
async fn get_range(&self, location: &Path, range: Range<usize>) -> Result<Bytes> {
let response = self
.client
.get_request(location, Some(range), false)
.await?;
Ok(response.bytes().await.context(GetRequestSnafu {
path: location.as_ref(),
})?)
}
async fn head(&self, location: &Path) -> Result<ObjectMeta> {
let response = self.client.get_request(location, None, true).await?;
let object = response.json().await.context(GetRequestSnafu {
path: location.as_ref(),
})?;
convert_object_meta(&object)
}
async fn delete(&self, location: &Path) -> Result<()> {
self.client.delete_request(location).await
}
async fn list(
&self,
prefix: Option<&Path>,
) -> Result<BoxStream<'_, Result<ObjectMeta>>> {
let stream = self
.client
.list_paginated(prefix, false)?
.map_ok(|r| {
futures::stream::iter(
r.items.into_iter().map(|x| convert_object_meta(&x)),
)
})
.try_flatten()
.boxed();
Ok(stream)
}
async fn list_with_delimiter(&self, prefix: Option<&Path>) -> Result<ListResult> {
let mut stream = self.client.list_paginated(prefix, true)?;
let mut common_prefixes = BTreeSet::new();
let mut objects = Vec::new();
while let Some(result) = stream.next().await {
let response = result?;
for p in response.prefixes {
common_prefixes.insert(Path::parse(p)?);
}
objects.reserve(response.items.len());
for object in &response.items {
objects.push(convert_object_meta(object)?);
}
}
Ok(ListResult {
common_prefixes: common_prefixes.into_iter().collect(),
objects,
})
}
async fn copy(&self, from: &Path, to: &Path) -> Result<()> {
self.client.copy_request(from, to, false).await
}
async fn copy_if_not_exists(&self, from: &Path, to: &Path) -> Result<()> {
self.client.copy_request(from, to, true).await
}
}
fn reader_credentials_file(
service_account_path: impl AsRef<std::path::Path>,
) -> Result<ServiceAccountCredentials> {
let file = File::open(service_account_path).context(OpenCredentialsSnafu)?;
let reader = BufReader::new(file);
Ok(serde_json::from_reader(reader).context(DecodeCredentialsSnafu)?)
}
#[derive(Debug, Default)]
pub struct GoogleCloudStorageBuilder {
bucket_name: Option<String>,
service_account_path: Option<String>,
client: Option<Client>,
retry_config: RetryConfig,
}
impl GoogleCloudStorageBuilder {
pub fn new() -> Self {
Default::default()
}
pub fn with_bucket_name(mut self, bucket_name: impl Into<String>) -> Self {
self.bucket_name = Some(bucket_name.into());
self
}
pub fn with_service_account_path(
mut self,
service_account_path: impl Into<String>,
) -> Self {
self.service_account_path = Some(service_account_path.into());
self
}
pub fn with_retry(mut self, retry_config: RetryConfig) -> Self {
self.retry_config = retry_config;
self
}
#[cfg(test)]
pub fn with_client(mut self, client: Client) -> Self {
self.client = Some(client);
self
}
pub fn build(self) -> Result<GoogleCloudStorage> {
let Self {
bucket_name,
service_account_path,
client,
retry_config,
} = self;
let bucket_name = bucket_name.ok_or(Error::MissingBucketName {})?;
let service_account_path =
service_account_path.ok_or(Error::MissingServiceAccountPath)?;
let client = client.unwrap_or_else(Client::new);
let credentials = reader_credentials_file(service_account_path)?;
let scope = "https://www.googleapis.com/auth/devstorage.full_control";
let audience = "https://www.googleapis.com/oauth2/v4/token".to_string();
let oauth_provider = (!credentials.disable_oauth)
.then(|| {
OAuthProvider::new(
credentials.client_email,
credentials.private_key,
scope.to_string(),
audience,
)
})
.transpose()?;
let encoded_bucket_name =
percent_encode(bucket_name.as_bytes(), NON_ALPHANUMERIC).to_string();
Ok(GoogleCloudStorage {
client: Arc::new(GoogleCloudStorageClient {
client,
base_url: credentials.gcs_base_url,
oauth_provider,
token_cache: Default::default(),
bucket_name,
bucket_name_encoded: encoded_bucket_name,
retry_config,
max_list_results: None,
}),
})
}
}
fn convert_object_meta(object: &Object) -> Result<ObjectMeta> {
let location = Path::parse(&object.name)?;
let last_modified = object.updated;
let size = object.size.parse().context(InvalidSizeSnafu)?;
Ok(ObjectMeta {
location,
last_modified,
size,
})
}
#[cfg(test)]
mod test {
use std::env;
use bytes::Bytes;
use crate::{
tests::{
get_nonexistent_object, list_uses_directories_correctly, list_with_delimiter,
put_get_delete_list, rename_and_copy, stream_get,
},
Error as ObjectStoreError, ObjectStore,
};
use super::*;
const NON_EXISTENT_NAME: &str = "nonexistentname";
macro_rules! maybe_skip_integration {
() => {{
dotenv::dotenv().ok();
let required_vars = ["OBJECT_STORE_BUCKET", "GOOGLE_SERVICE_ACCOUNT"];
let unset_vars: Vec<_> = required_vars
.iter()
.filter_map(|&name| match env::var(name) {
Ok(_) => None,
Err(_) => Some(name),
})
.collect();
let unset_var_names = unset_vars.join(", ");
let force = std::env::var("TEST_INTEGRATION");
if force.is_ok() && !unset_var_names.is_empty() {
panic!(
"TEST_INTEGRATION is set, \
but variable(s) {} need to be set",
unset_var_names
)
} else if force.is_err() {
eprintln!(
"skipping Google Cloud integration test - set {}TEST_INTEGRATION to run",
if unset_var_names.is_empty() {
String::new()
} else {
format!("{} and ", unset_var_names)
}
);
return;
} else {
GoogleCloudStorageBuilder::new()
.with_bucket_name(
env::var("OBJECT_STORE_BUCKET")
.expect("already checked OBJECT_STORE_BUCKET")
)
.with_service_account_path(
env::var("GOOGLE_SERVICE_ACCOUNT")
.expect("already checked GOOGLE_SERVICE_ACCOUNT")
)
.with_client(
Client::builder()
.danger_accept_invalid_certs(true)
.build()
.expect("Error creating http client for testing")
)
}
}};
}
#[tokio::test]
async fn gcs_test() {
let integration = maybe_skip_integration!().build().unwrap();
put_get_delete_list(&integration).await;
list_uses_directories_correctly(&integration).await;
list_with_delimiter(&integration).await;
rename_and_copy(&integration).await;
if integration.client.base_url == default_gcs_base_url() {
stream_get(&integration).await;
}
}
#[tokio::test]
async fn gcs_test_get_nonexistent_location() {
let integration = maybe_skip_integration!().build().unwrap();
let location = Path::from_iter([NON_EXISTENT_NAME]);
let err = integration.get(&location).await.unwrap_err();
assert!(
matches!(err, ObjectStoreError::NotFound { .. }),
"unexpected error type: {}",
err
);
}
#[tokio::test]
async fn gcs_test_get_nonexistent_bucket() {
let integration = maybe_skip_integration!()
.with_bucket_name(NON_EXISTENT_NAME)
.build()
.unwrap();
let location = Path::from_iter([NON_EXISTENT_NAME]);
let err = get_nonexistent_object(&integration, Some(location))
.await
.unwrap_err();
assert!(
matches!(err, ObjectStoreError::NotFound { .. }),
"unexpected error type: {}",
err
);
}
#[tokio::test]
async fn gcs_test_delete_nonexistent_location() {
let integration = maybe_skip_integration!().build().unwrap();
let location = Path::from_iter([NON_EXISTENT_NAME]);
let err = integration.delete(&location).await.unwrap_err();
assert!(
matches!(err, ObjectStoreError::NotFound { .. }),
"unexpected error type: {}",
err
);
}
#[tokio::test]
async fn gcs_test_delete_nonexistent_bucket() {
let integration = maybe_skip_integration!()
.with_bucket_name(NON_EXISTENT_NAME)
.build()
.unwrap();
let location = Path::from_iter([NON_EXISTENT_NAME]);
let err = integration.delete(&location).await.unwrap_err();
assert!(
matches!(err, ObjectStoreError::NotFound { .. }),
"unexpected error type: {}",
err
);
}
#[tokio::test]
async fn gcs_test_put_nonexistent_bucket() {
let integration = maybe_skip_integration!()
.with_bucket_name(NON_EXISTENT_NAME)
.build()
.unwrap();
let location = Path::from_iter([NON_EXISTENT_NAME]);
let data = Bytes::from("arbitrary data");
let err = integration
.put(&location, data)
.await
.unwrap_err()
.to_string();
assert!(
err.contains(
"Error performing put request: HTTP status client error (404 Not Found)"
),
"{}",
err
)
}
}