s3util
Note on issues: This project continues to be maintained, and binaries will keep being released. However, to consolidate discussion across the s3sync / s3util-rs / s3rm-rs / s3ls-rs family, please file new issues in the s7cmd repository instead of here. s7cmd bundles these tools as subcommands built on the same underlying code, so its behavior matches the standalone binaries and it can be used in their place. Before opening an issue, please read the Scope and Non-Goals sections in the READMEs of s7cmd and each project (s3sync / s3util-rs / s3rm-rs / s3ls-rs) — requests outside the documented scope will generally be declined. Existing issues in this repository will continue to be handled as usual.
Tools for managing Amazon S3 objects and buckets
s3util is a collection of tools for managing objects and buckets on Amazon S3 and S3-compatible object stores, built on the official AWS SDK for Rust (aws-sdk-s3). It ports the transfer, verification, and multipart semantics of s3sync into a compact CLI focused on interactive and scripted use, and is intended to become part of the future s7cmd toolkit.
Table of contents
- Overview
- Features
- Requirements
- Installation
- Usage
- Detailed information
- Advanced options
- --max-parallel-uploads
- --multipart-threshold / --multipart-chunksize
- --auto-chunksize
- --additional-checksum-algorithm
- --full-object-checksum
- --disable-multipart-verify / --disable-etag-verify
- --server-side-copy
- --if-none-match
- --source-no-sign-request
- --rate-limit-bandwidth
- --dry-run
- -v / -q
- --aws-sdk-tracing
- --auto-complete-shell
- --help
- CI/CD Integration
- About testing
- Fully AI-generated (human-verified) software
- Contributing
- License
Overview
s3util is a collection of tools for managing objects and buckets on Amazon S3, built as a companion to s3sync. Where s3sync is optimized for bulk, recursive synchronization, s3util is optimized for single-object transfers and direct S3 API operations: each invocation operates on one object or one bucket, verifies the result where applicable, and exits with a meaningful status code.
s3util's cp and mv subcommands follow the same design principles as s3sync for transfer, verification, and multipart handling — but each subcommand has a deliberately narrow surface, and the binary is a single file with no recursive/directory mode.
For object transfers in particular, s3util emphasizes high reliability, high performance, and advanced functionality: end-to-end checksum verification (ETag plus SHA256/SHA1/CRC32/CRC32C/CRC64NVME, composite or full-object), parallel multipart uploads and downloads, server-side copy, SSE-KMS and SSE-C (including SSE-C re-keying across copies), stdin/stdout streaming, tag and metadata preservation, rate-limited bandwidth control, and Express One Zone support. See Features for the full list.
Scope
s3util is designed to cover common single-object and bucket-management operations — single-object transfers (cp / mv) and common bucket management (creation/deletion, tagging, versioning, policy, lifecycle, encryption, CORS, public-access-block, website, logging, notifications). For any S3 use case outside that scope, use a more comprehensive tool such as the AWS CLI (aws s3 / aws s3api); for recursive or bulk synchronization, use s3sync.
The cp and mv subcommands operate on one object at a time; the thin S3 API wrappers each issue a single S3 API call. s3util is not intended to be a drop-in replacement for, or behaviorally compatible with, any other S3 client — including the AWS CLI (aws s3, aws s3api) and tools such as s3cmd, s5cmd, rclone, and mc. Its command-line flags, transfer semantics, verification rules, and exit codes are designed around safe, verifiable single-object transfers and explicit per-API operations — not interoperability with another tool's interface. Output formats and flag names will not be adjusted to match any external tool, and scripts written against another S3 client should not be expected to work with s3util unmodified.
API call volume
cp and mv are optimized for parallel transfer and end-to-end
verification, not for minimizing the number of S3 API calls per
invocation. A single transfer can issue several requests beyond the
obvious GetObject / PutObject:
HeadObjectupfront (to discover size and the source's stored ETag / additional checksum).GetObjectAttributesand/or per-partHeadObjectcalls when--auto-chunksizeis set (one per source part, to align local chunking with the source's part layout).- Up to
--max-parallel-uploadsconcurrent rangedGetObjectrequests on the download side, orUploadPart/UploadPartCopyrequests on the upload / server-side-copy side. CreateMultipartUpload+CompleteMultipartUpload(plusAbortMultipartUploadon cancellation) on multipart uploads.- A second
HeadObjectafter a ranged GET when the object is a composite-multipart source whose root composite checksum was needed for verification.
If keeping the API call count to a minimum is important to you (cost,
rate limits, or audit-log volume), s3util is not the right tool —
prefer the AWS CLI's aws s3api put-object / get-object for direct,
single-call transfers without the verification and parallelism
machinery.
Non-Goals
The following are explicitly out of scope and will not be added, regardless of demand:
- Recursive or directory-mode transfers — use s3sync instead.
- Glob or wildcard expansion in S3 keys. For pattern-based matching, use s3sync, which supports regular expressions.
- Multiple source or destination arguments to
cp/mv(e.g.s3util cp a.txt b.txt s3://bucket/dest/). Each invocation transfers exactly one object. - Compatibility with other S3 clients — neither in flag names and
behavior, nor in feature coverage. The presence of a feature, flag,
or output format in
aws s3,aws s3api,s3cmd,s5cmd,rclone,mc, or any other S3 tool is not, by itself, a reason to add or change it in s3util. Each request is evaluated only against s3util's own scope and design principles. Use that other tool if you need its specific surface. - Diagnosing or fixing performance degradation, resource exhaustion,
or errors caused by raising concurrency settings
(
--max-parallel-uploadsand similar tuning flags) above their defaults. The documentation explicitly notes that these values must be sized to the host and the target service; tuning them is the operator's responsibility. Reports of the form "I raised--max-parallel-uploadsand it failed / slowed down / hit rate limits" will be closed. - Resuming a failed or interrupted transfer.
s3utildoes not provide a resume feature forcp/mv, including for large multipart uploads and downloads. There is no checkpoint file, no part-list reuse, and no partial-progress state persisted between invocations: if a transfer is interrupted (network error, ctrl-c, process kill, host shutdown, etc.), the next invocation re-transfers the whole object from the start. In-flight multipart uploads are best-effort aborted on ctrl-c and on error paths, but for crashes or other non-graceful exits, cleaning up any leftover incomplete multipart uploads (e.g. via a bucket lifecycle rule oraws s3api abort-multipart-upload) is the operator's responsibility. - A plugin or extension mechanism.
Issues and pull requests requesting any of the above will be closed.
Subcommands
s3util provides the following subcommands. cp and mv perform single-object transfers using the full multipart and verification pipeline; the remaining subcommands are thin wrappers around individual S3 API calls with a simpler, script-friendly interface than aws s3api.
| Subcommand | What it does |
|---|---|
cp |
Copies a single object: Local↔S3, S3↔S3, or stdin/stdout streaming; full multipart + checksum verification |
mv |
Moves a single object: same as cp plus deletes the source after a successful, verified copy (no stdio) |
rm |
Deletes a single S3 object; silent on success; supports --source-version-id |
head-object |
Prints HeadObject response as JSON; supports --source-version-id and SSE-C reads |
put-object-tagging |
Replaces all tags from --tagging "k=v&k2=v2"; silent; supports --source-version-id |
get-object-tagging |
Prints object tags as JSON ({"TagSet": [...], "VersionId": "..."}); supports --source-version-id |
delete-object-tagging |
Removes all tags from an object; silent; supports --source-version-id |
create-bucket |
Creates a bucket; LocationConstraint from the SDK client's resolved region (--target-region, AWS_REGION, or profile); optional --tagging; exit 3 if tagging step fails after create |
head-bucket |
Prints HeadBucket response as JSON |
delete-bucket |
Deletes an empty bucket; silent on success |
put-bucket-policy |
Sets bucket policy from a file path or - (stdin); body sent verbatim, no client-side validation; silent |
get-bucket-policy |
Prints policy as JSON ({"Policy": "<escaped-JSON-string>"}, matching aws s3api); --policy-only prints just the inner policy as pretty-printed JSON |
delete-bucket-policy |
Removes bucket policy; silent on success |
put-bucket-tagging |
Replaces all tags from --tagging "k=v&k2=v2"; silent on success |
get-bucket-tagging |
Prints bucket tags as JSON ({"TagSet": [...]}); exits 4 on NoSuchTagSet / NoSuchBucket |
delete-bucket-tagging |
Removes all tags from a bucket; silent on success |
put-bucket-versioning |
Enables or suspends versioning (--enabled / --suspended, mutually exclusive); silent |
get-bucket-versioning |
Prints versioning state as JSON ({"Status": "Enabled"}); silent when never configured (matches AWS CLI) |
put-bucket-lifecycle-configuration |
Sets lifecycle configuration from a JSON file path or - (stdin); silent on success |
get-bucket-lifecycle-configuration |
Prints lifecycle configuration as JSON ({"Rules": […]} matching aws s3api); exits 4 if no lifecycle is set |
delete-bucket-lifecycle-configuration |
Removes lifecycle configuration; silent on success |
put-bucket-encryption |
Sets default encryption from a JSON file path or - (stdin); silent on success |
get-bucket-encryption |
Prints default encryption as JSON; exits 4 if no explicit encryption is set |
delete-bucket-encryption |
Removes default encryption; silent on success |
put-bucket-cors |
Sets CORS rules from a JSON file path or - (stdin); silent on success |
get-bucket-cors |
Prints CORS rules as JSON ({"CORSRules": […]}); exits 4 if no CORS is set |
delete-bucket-cors |
Removes CORS rules; silent on success |
put-public-access-block |
Sets public-access-block from a JSON file path or - (stdin); silent on success |
get-public-access-block |
Prints public-access-block as JSON; exits 4 if none set |
delete-public-access-block |
Removes public-access-block; silent on success |
put-bucket-website |
Sets the website configuration from a JSON file path or - (stdin); silent on success |
get-bucket-website |
Prints website configuration as JSON; exits 4 if no website is configured |
delete-bucket-website |
Removes the website configuration; silent on success |
put-bucket-logging |
Sets bucket logging from a JSON file path or - (stdin); empty {} JSON disables logging; silent on success |
get-bucket-logging |
Prints bucket logging configuration as JSON; silent when no logging is configured (matches AWS CLI) |
put-bucket-notification-configuration |
Sets notification configuration from a JSON file path or - (stdin); empty {} JSON disables all notifications; silent on success |
get-bucket-notification-configuration |
Prints notification configuration as JSON; silent when no notifications are configured (matches AWS CLI) |
Features
The features described below relate to the cp and mv subcommands. For details on other subcommands, refer to the help documentation (s3util -h / s3util <command> -h) and the Amazon S3 documentation.
Verifiable transfers
When the source is a local file or stdin, s3util precalculates the ETag and — if --additional-checksum-algorithm is set — the additional checksum, then compares them against the S3-reported values. A mismatch is treated as an error (the destination object is considered corrupted).
For S3→S3 transfers, ETag and composite additional-checksum mismatches are downgraded to warnings (exit code 3), because differing multipart chunk sizes between source and destination legitimately produce different composite values. Full-object additional-checksum mismatches remain errors — chunk size cannot legitimately change a full-object checksum, so a mismatch indicates real corruption.
Supported algorithms:
- ETag (MD5 for single-part, multipart-composite hash for multipart uploads)
- SHA256, SHA1, CRC32, CRC32C, CRC64NVME via
--additional-checksum-algorithm - Full-object and composite variants via
--full-object-checksum
Verification can be selectively disabled with --disable-etag-verify, --disable-multipart-verify, or --disable-additional-checksum-verify when working with S3-compatible stores that behave differently.
Full multipart support
- Configurable threshold (
--multipart-threshold, default8MiB) and chunk size (--multipart-chunksize, default8MiB). - Parallel part uploads/downloads (
--max-parallel-uploads, default16). --auto-chunksizematches the source multipart layout on S3→S3 copies so checksums line up end-to-end.- In-flight multipart uploads are aborted cleanly on ctrl-c.
- Supports objects up to Amazon S3's per-object size limit (currently 50 TB).
All transfer directions
Transfer direction is inferred automatically from the source/target combination:
| Source | Target | Direction |
|---|---|---|
| local path | s3://… |
Local → S3 |
s3://… |
local path | S3 → Local |
s3://… |
s3://… |
S3 → S3 |
- (stdin) |
s3://… |
stdin → S3 |
s3://… |
- (stdout) |
S3 → stdout |
S3 → S3 transfers can span different AWS accounts, different regions, and different S3-compatible storage providers (e.g. AWS S3 → S3-compatible storage, or vice versa). The source and target are independently configured via the paired --source-* and --target-* credential, profile, region, and endpoint flags — they need not share a single S3 endpoint.
Server-side copy
By default, S3→S3 transfers are client-side: s3util streams the object from the source through the local process and re-uploads it to the target. This is the most compatible mode — it works across different regions, endpoints, accounts, and S3-compatible providers, and is required for any transfer that crosses a boundary CopyObject cannot.
Passing --server-side-copy switches to S3's CopyObject / UploadPartCopy, so the bytes never round-trip through the client. Both source and target must be S3, and the API call must be supported by the server (typically same-region, single endpoint). s3util does not fall back to client-side copy if server-side copy fails or is unsuitable — leave the flag off when in doubt.
stdin/stdout streaming
Pipe data directly through S3 without touching the local filesystem:
|
|
stdin uploads compute the ETag and additional checksum on the fly and verify against the S3-reported values.
Express One Zone support
s3util supports Amazon S3 Express One Zone directory buckets (the --x-s3 bucket-name suffix). Additional-checksum verification is handled carefully for Express One Zone — use --disable-express-one-zone-additional-checksum if the defaults are too strict for your workload.
SSE and SSE-C
- Target-side:
--sse AES256/aws:kms/aws:kms:dsse, with--sse-kms-key-idfor KMS variants. - Source and target SSE-C: separate key/algorithm/MD5 flags for reading an SSE-C source and writing an SSE-C target (including re-keying across a client-side copy).
Metadata and tagging preservation
S3→S3 copies preserve both system metadata (Content-Type, Cache-Control, Expires, Content-Disposition, Content-Encoding, Content-Language, website-redirect) and user-defined metadata by default. Use --no-sync-system-metadata / --no-sync-user-defined-metadata to opt out, or override individual headers explicitly.
Object tags are preserved on S3→S3 by default. --tagging "k=v&k2=v2" overrides, --disable-tagging clears.
Rate limiting
--rate-limit-bandwidth <BYTES_PER_SEC> caps throughput using a leaky-bucket algorithm. Accepts unit-suffixed values like 50MB, 100MiB, 1GB.
Observability
- Optional progress bar (
--show-progress) using indicatif. - Structured JSON tracing (
--json-tracing) for log aggregation systems. - AWS SDK tracing (
--aws-sdk-tracing) for deep troubleshooting. - Configurable verbosity (
-v/-vv/-vvv,-q/-qq).
Dry-run
--dry-run is supported on every mutating subcommand — cp, mv, rm, create-bucket, every put-*, and every delete-*. With --dry-run, s3util runs all preflight work (argument validation, JSON-config parsing, SDK client construction, transfer-pipeline setup), stops just before the destructive Web API call, and emits an info-level log line prefixed with [dry-run] (e.g. [dry-run] would delete bucket., [dry-run] would copy object., [dry-run] Transfer completed.). The exit status is 0 on success.
To make the message visible at the default WarnLevel, --dry-run automatically raises the verbosity floor to info. Levels you've explicitly raised (-vv for debug, -vvv for trace) are preserved unchanged; lower levels (-q, even full silence with -qqq) are still bumped to info so the line stays visible.
Important:
--dry-runperforms only a formal check. It validates flags, parses config-file JSON, and confirms the binary can reach the point where the API call would have been issued. It does not verify any AWS-side state — it does not check whether the source object or bucket exists, whether your credentials are accepted, whether the target endpoint is reachable, or whether your IAM policy would have allowed the operation. A successful dry-run means "the request was well-formed and ready to send", not "the operation would have succeeded against AWS".
get-* and head-* subcommands are read-only and do not expose --dry-run.
Requirements
- x86_64 Linux (kernel 3.2 or later)
- ARM64 Linux (kernel 4.1 or later)
- Windows 11 (x86_64, aarch64)
- macOS 11.0 or later (aarch64, x86_64)
s3util is written in Rust and requires Rust 1.91.1 or later to build from source.
AWS credentials are required. s3util supports all standard AWS credential mechanisms:
- Environment variables (
AWS_ACCESS_KEY_ID,AWS_SECRET_ACCESS_KEY) - AWS credentials file (
~/.aws/credentials) - AWS config file (
~/.aws/config) with profiles - IAM instance roles (EC2, ECS, Lambda)
- SSO/federated authentication
- Explicit
--source-*/--target-*flags
For more information, see SDK authentication with AWS.
Installation
Install from crates.io
The crate is published as s3util-rs; the installed binary is named s3util.
Usage
s3util <COMMAND> [OPTIONS] <SOURCE> <TARGET>
Supported path forms for <SOURCE> / <TARGET>:
| Form | Meaning |
|---|---|
s3://bucket |
Bucket with empty prefix |
s3://bucket/k |
Specific key (or prefix ending in /) |
/local/path |
Local filesystem path |
- |
stdin or stdout |
Every long flag also reads from an uppercase-underscore environment variable of the same name (for example --max-parallel-uploads ↔ MAX_PARALLEL_UPLOADS). Precedence: CLI arguments > environment variables > defaults.
The examples below describe the cp and mv commands. For details on other commands (head-bucket, head-object, rm, the bucket-management wrappers, etc.), run s3util -h for the top-level subcommand list and s3util <command> -h for per-command options.
Upload a local file
If the target ends in / (or is a bucket root), the source basename is appended to form the key. When --show-progress is set, the destination path is printed on a -> <path> line before the transfer summary.
Download to local
The target parent directory must already exist. s3util does not create missing directories — it returns an error asking you to create them first.
S3 → S3 copy
Client-side (default):
Server-side (same account/region, avoids round-tripping bytes through the client):
Client-side S3 → S3 copies can span different AWS accounts, different regions, and different S3-compatible providers — point the --source-* and --target-* flags at independent endpoints:
# Cross-account, cross-region (separate profiles, separate regions)
# AWS S3 → S3-compatible storage
--server-side-copy is incompatible with this case (it requires source and target to be reachable from a single S3 endpoint); cross-endpoint copies always run client-side.
stdin → S3
|
With stdin as the source there is no basename, so the target key must be spelled out.
S3 → stdout
|
Move with mv
mv runs the same copy pipeline as cp and then deletes the source on success. Transfer, verification, multipart, metadata, tagging, SSE, server-side copy, rate limiting, and progress all behave identically — only the post-copy step differs.
# S3 → S3 move
# Upload then delete the local file
# Download then delete the source S3 object
Differences from cp:
- stdin/stdout is not supported. A
-source or target is rejected at argument-parse time. - The source is deleted only after a successful, verified copy. If the copy fails, is canceled (SIGINT), or produces a verification warning, the source is left untouched and the command exits with the matching non-zero code.
--no-fail-on-verify-error(mv only) treats a verification warning as success and proceeds to delete the source. Use only when you understand why your S3↔S3 chunksize layout produces an expected mismatch.--source-version-iddeletes the specific source version after the copy (rather than creating a delete marker on the latest version).
Additional checksum verification
# Upload with SHA256 additional checksum
# Download and verify the additional checksum stored on the object
# (the algorithm is whatever was used at upload time)
Multipart tuning
# Force multipart at 64 MiB with 16 MiB chunks and 8 parallel workers
# Match the source chunk layout on S3 → S3 copy
Custom endpoint (S3-compatible stores)
Specify credentials
Specify region
Detailed information
Path and target resolution
If the target is s3://bucket, s3://bucket/dir/, or a directory-style local path (an existing directory, or one ending in a path separator like ../), the source basename is appended. When --show-progress is set, the destination path is printed on a -> <path> line before the transfer summary.
With stdin as the source there is no basename, so the target key must be spelled out.
ETag verification
For single-part objects, the S3-reported ETag is the MD5 of the object. s3util computes this on the upload side and compares; for downloads it compares the source's reported ETag against the bytes actually received. Local/stdin→S3 mismatches are treated as errors (the upload is considered corrupted and the source is authoritative). S3→Local and S3→S3 mismatches are warnings (exit code 3) — for S3→S3 because multipart layout differences legitimately change the composite ETag, and for S3→Local because the file is already written and the warning lets you decide whether to redownload.
--disable-etag-verify turns off ETag verification entirely. --disable-content-md5-header additionally omits the Content-MD5 header on single-part uploads.
Additional checksum verification
When --additional-checksum-algorithm is set, S3 stores the chosen algorithm's checksum alongside the object. Supported: SHA256, SHA1, CRC32, CRC32C, CRC64NVME.
--full-object-checksumforces the full-object variant (required for CRC64NVME; incompatible with SHA1/SHA256).--enable-additional-checksumon download tells S3 to return the additional checksum sos3utilcan verify it.--disable-additional-checksum-verifyuploads the additional checksum but skips local verification.
Auto chunksize
--auto-chunksize issues additional HeadObject calls to discover the source's multipart layout and then mirrors it on the destination. This keeps the S3→S3 composite ETag and additional-checksum values identical end-to-end, at the cost of one extra HeadObject per part.
⚠️ Memory warning for
--auto-chunksizeon client-side download paths. On any path wheres3utildownloads parts itself instead of letting S3 copy them server-side — that is, S3 → stdout, S3 → Local, and S3 → S3 without--server-side-copy—--auto-chunksizeinstructss3utilto use the source object's actual per-part sizes as the download chunk size, instead of--multipart-chunksize. Each parallel worker allocates one chunk in memory while it downloads, so peak memory usage is approximately:source's largest part size × --max-parallel-uploadsIf the source was uploaded with very large parts (for example, 1 GiB parts × 16 parallel workers ≈ 16 GiB), this can exhaust available memory. Either lower
--max-parallel-uploadsto bound the allocation, or omit--auto-chunksizeand accept that the per-part composite ETag/checksum may not verify on the resulting download.
--server-side-copy(S3 → S3 only) sidesteps this entirely: parts are copied server-to-server viaUploadPartCopyand never materialize ins3util's memory.
Server-side copy detail
--server-side-copy uses CopyObject (single-part) or UploadPartCopy (multipart). Server-side copy is only valid when both source and target endpoints can see each other in the same AWS region/account (with appropriate cross-account IAM). It is not compatible with stdin or local paths. SSE-C re-keying across a server-side copy is supported by supplying both --source-sse-c-* and --target-sse-c-* flags.
stdin/stdout handling
- stdin → S3 reads up to
--multipart-thresholdbytes into an in-memory buffer. If EOF arrives first, the buffered bytes are issued as a single-part PUT with a correctContent-Length; otherwise the buffered prefix is chained with the rest of the stream and uploaded as a multipart. - S3 → stdout streams bytes straight to stdout. ETag and any requested additional checksum are computed inline from the streamed bytes and verified against the S3-reported values — the same verification as
S3 → Local. A mismatch is logged as a warning (exit 3), or as an error if the configured additional checksum is a full-object checksum. For large objects the download runs in parallel (up to--max-parallel-uploadsconcurrent ranged GETs); peak memory is roughly--multipart-chunksize × --max-parallel-uploads. With--auto-chunksizethe chunk size becomes the source's actual per-part size — see the Auto chunksize memory warning before using it on objects uploaded with very large parts.
Express One Zone detail
Directory buckets (--x-s3 suffix) are automatically detected. Some S3 features behave differently on Express One Zone (for example, default additional-checksum handling); --disable-express-one-zone-additional-checksum overrides s3util's default if your bucket policy demands it.
create-bucket also accepts directory-bucket names. The zone ID is parsed from the name (<base>--<zone-id>--x-s3) and the appropriate Location/Bucket configuration is sent. The zone type is inferred from the zone-ID shape — at most one hyphen is treated as an Availability Zone (e.g. apne1-az4), two or more as a Local Zone (e.g. usw2-lax1-az1). The active region (--target-region / AWS_REGION / profile) must match the zone's region; otherwise S3 will reject the request.
S3 Permissions
The permissions below cover the cp and mv subcommands. Other subcommands have their own requirements; refer to the AWS documentation for the full set.
Required permissions depend on the transfer direction. "Source" and "target" below refer to the source and target S3 buckets; for Local↔S3 only the relevant side applies.
Source bucket (any cp/mv reading from S3):
s3:GetObject— always. CoversGetObject,HeadObject, andGetObjectAttributes.s3:GetObjectTagging— when source tags are read. This is the default on S3→S3; suppressed by--disable-tagging.s3:GetObjectVersion— when--source-version-idis used.s3:DeleteObject— when runningmv(the source is deleted on success).s3:DeleteObjectVersion— when runningmvwith--source-version-id.
Target bucket (any cp/mv writing to S3):
s3:PutObject— always. CoversPutObject,CreateMultipartUpload,UploadPart,CompleteMultipartUpload, and on--server-side-copyalsoCopyObject/UploadPartCopy.s3:AbortMultipartUpload— always (used to clean up on errors and ctrl-c).s3:PutObjectTagging— when target tags are written. This is the default on S3→S3, and also applies when--taggingis set.s3:PutObjectAcl— when--aclis set.
Express One Zone (directory buckets, --x-s3 suffix):
s3express:CreateSession— on each directory bucket the SDK opens a session for. Required in addition to the standards3:*actions above.
KMS-backed SSE:
kms:Decrypt— when reading SSE-KMS-encrypted source objects.kms:Encryptandkms:GenerateDataKey— when writing with--sse aws:kmsor--sse aws:kms:dsse.
SSE-C (--source-sse-c* / --target-sse-c*) requires no additional IAM permissions — the encryption key is supplied client-side and S3 does not store it.
CLI process exit codes
| Code | Meaning |
|---|---|
| 0 | Success |
| 1 | Error — transfer failed or configuration rejected |
| 2 | Argument-parsing error — an argument is unknown, missing, or has an invalid value |
| 3 | Warning — transfer completed but a non-fatal issue was logged (e.g. S3→S3 ETag mismatch explained by chunksize) |
| 4 | Not found — head-bucket / head-object (404 NoSuchBucket / NoSuchKey / NoSuchVersion); get-object-tagging / get-bucket-policy / get-bucket-tagging / get-bucket-lifecycle-configuration / get-bucket-encryption / get-bucket-cors / get-public-access-block / get-bucket-website when the addressed resource is missing (incl. NoSuchBucketPolicy / NoSuchTagSet / NoSuchLifecycleConfiguration / ServerSideEncryptionConfigurationNotFoundError / NoSuchCORSConfiguration / NoSuchPublicAccessBlockConfiguration / NoSuchWebsiteConfiguration); get-bucket-versioning / get-bucket-logging / get-bucket-notification-configuration only on NoSuchBucket — for these three, an unconfigured subresource is reported by S3 as a successful empty body, which exits 0, not 4 |
| 101 | Abnormal termination (internal panic) |
| 130 | User cancellation via SIGINT/ctrl-c (standard Unix SIGINT convention, 128 + 2) |
Advanced options
--max-parallel-uploads
Number of parallel part uploads/downloads during multipart transfers. Default: 16. The default is sized for typical hosts and S3's per-prefix request behavior; if you raise it, you must size it to your host (peak memory is roughly --multipart-chunksize × --max-parallel-uploads) and to the target service's per-prefix request and bandwidth limits. Tuning is the operator's responsibility.
--multipart-threshold / --multipart-chunksize
Object size threshold for switching to multipart (--multipart-threshold, default 8MiB) and the size of each part (--multipart-chunksize, default 8MiB). Both accept unit-suffixed values (MB, MiB, GB, GiB).
--auto-chunksize
Match source/target chunk layout automatically (extra HeadObject per part). Recommended for S3→S3 copies where you want identical composite ETags on both sides.
--additional-checksum-algorithm
Additional checksum algorithm for upload: SHA256, SHA1, CRC32, CRC32C, CRC64NVME. Also used to select the algorithm to verify on download (combined with --enable-additional-checksum).
--full-object-checksum
Use the full-object variant of the additional checksum instead of the composite variant. Required and forced for CRC64NVME; incompatible with SHA1/SHA256.
--disable-multipart-verify / --disable-etag-verify
Skip ETag or additional-checksum verification for multipart uploads. Useful when targeting S3-compatible stores that compute ETags differently.
--server-side-copy
Use S3 server-side copy for S3→S3 transfers. Requires both endpoints to support the relevant server-side copy operations.
--if-none-match
Upload only if the target key does not already exist. This is an optimistic "create new object" primitive at the S3 level.
--source-no-sign-request
Access public S3 buckets anonymously — skips the entire AWS credential chain (profile, env, IMDS, SSO) on the source side.
--rate-limit-bandwidth
Maximum bytes per second for the transfer. Accepts unit suffixes like MB, MiB, GB, GiB.
--dry-run
Skip the destructive S3 Web API call and emit a [dry-run]-prefixed info-level log line instead. Exit status is 0 on success. Available on every mutating subcommand (cp, mv, rm, create-bucket, all put-*, all delete-*). The verbosity floor is forced to info while --dry-run is set so the message stays visible at default verbosity. See Dry-run under Features for the full description and the important caveat that this is a formal check only — no AWS-side state is verified.
-v / -q
s3util uses tracing-subscriber for tracing. More occurrences of -v increase verbosity (-v: info, -vv: debug, -vvv: trace). Use -q, -qq to reduce verbosity. Default: warning and error messages.
With -v, subcommands that are otherwise silent on success (rm, create-bucket, delete-bucket, the put-* and delete-* bucket/object subcommands) emit a structured info-level event to stderr describing what was changed (e.g. Object deleted. bucket=… key=… version_id=…). get-bucket-versioning, get-bucket-logging, and get-bucket-notification-configuration likewise log Bucket … not configured. when the bucket has no such configuration (each prints nothing on stdout in that case, matching aws s3api, since the underlying S3 API returns success with an empty body for these three).
--aws-sdk-tracing
Enable AWS SDK for Rust's internal tracing. Useful for diagnosing endpoint/signature issues.
--auto-complete-shell
Generate shell completion scripts:
--help
For the full option list, see s3util cp --help.
CI/CD Integration
s3util is designed for automated pipelines.
JSON logging
Emit structured JSON logs for log aggregation systems (Datadog, Splunk, CloudWatch, etc.):
Quiet mode
Suppress info-level output for cleaner CI logs:
About testing
Supported target: Amazon S3 only.
Support for S3-compatible storage is on a best-effort basis and may behave differently.
s3util has been tested with Amazon S3, including Express One Zone directory buckets. s3util has many end-to-end tests and unit tests, and they run every time a new version is released.
S3-compatible storage is not tested when a new version is released (I test only when making major changes). This is because S3-compatible storage may have different behaviors and features. Since there is no official certification for S3-compatible storage, comprehensive testing is not possible.
Fully AI-generated (human-verified) software
No human wrote a single line of source code in this project. Every line of source code, every test, all documentation, CI/CD configuration, and this README were generated by AI using Claude Code (Anthropic).
Human engineers authored the requirements, design specifications, and the s3sync reference architecture. They thoroughly reviewed and verified the design, all source code, and all tests. Features of the binary have been manually tested against live AWS S3. The development followed a spec-driven process: requirements and design documents were written first, and the AI generated code to match those specifications under continuous human oversight.
Quality verification (by AI self-assessment)
Measurements below are taken at v1.1.0 (commit 48442da on update/v1.0.1, 2026-04-30). The coverage figures are sourced from lcov_report.txt (cargo llvm-cov) and reflect the unit-test build only — the --cfg e2e_test integration suite runs separately and is not included in the report.
| Metric | Value |
|---|---|
| Production code | ~35,500 lines of Rust across 131 source files in src/ |
Unit tests (in src/) |
940 #[test] / #[tokio::test] annotations |
| CLI integration tests | 293 annotations across 40 tests/cli_*.rs files (no network access; run in CI) |
| E2E integration tests | 668 annotations across 45 tests/e2e_*.rs files (gated behind --cfg e2e_test; run only by the maintainer against live AWS) |
| Code coverage (llvm-cov, unit-test build) | 96.70% regions (1,027 / 31,090 missed), 95.47% functions (100 / 2,208 missed), 97.54% lines (534 / 21,715 missed) |
| Static analysis (clippy) | 0 warnings (cargo clippy --all-features) |
| Formatting | 0 diffs (cargo fmt --all --check) |
| Supply chain (cargo-deny) | Clean (cargo deny -L error check); runs per-PR in ci.yml and daily at 01:34 UTC in cargo-deny.yml; advisories.ignore = [] |
| Code adapted from s3sync | Transfer engine (src/transfer/), checksum verification (src/storage/e_tag_verify.rs, src/storage/additional_checksum_verify.rs), and multipart upload manager (src/storage/s3/upload_manager.rs) |
What these numbers do and do not show:
- They show what the unit-test build exercises and what the CI pipeline asserts on every push and PR — not how the binary behaves against live S3 under production load.
- Coverage is a structural metric. A covered line can still be incorrect; an uncovered line can still be correct. Use it to size the test surface, not to certify behaviour.
- The e2e suite covers live-AWS paths (multipart integrity, roundtrip, cancellation, exit codes, Express One Zone, public-bucket access) but runs only on the maintainer's machine; CI does not exercise it.
The codebase is built through spec-driven development with human review at every step. Test counts and coverage will change as subcommands and refinements are added.
AI assessment of safety and correctness (by Claude, Anthropic)
Assessment date: 2026-04-30.
Assessed version: 1.1.0 (branch
update/v1.0.1, commit1ef94b9).Scope of evidence: a repository-wide read of the source tree (
src/bin/s3util/,src/config/,src/storage/,src/transfer/,src/input/,src/output/,src/types/), the integration suites undertests/cli_*.rsandtests/e2e_*.rs, theCargo.tomlandCargo.lockdependency graph, thecargo-denyconfiguration, the GitHub Actions workflows in.github/workflows/, and the coverage report inlcov.info/lcov_report.txt(combined unit-test and--cfg e2e_testruns). Each safety claim below cites the file(s) where the relevant code or test lives so a reviewer can verify it directly.Limits of the evidence: this assessment cannot rule out all bugs. It does not include fuzzing, sanitizer runs, formal verification, or penetration testing. The e2e suite is gated behind
--cfg e2e_testand runs against real AWS only on the maintainer's account; CI does not run it.
Question addressed. Three concrete failure modes drive the analysis: (1) the operator types a command that targets the wrong resource, (2) a bug in s3util corrupts data while transferring, (3) a thin S3 API wrapper succeeds at the wrong scope or leaves a bucket in a partial state. Each requires a different safeguard.
Subcommand surface (v1.1.0)
The CLI exposes 37 subcommand entry points in src/config/args/mod.rs: two transfer subcommands (cp, mv) and 35 thin wrappers around individual S3 API calls. Every subcommand operates on exactly one resource — there is no recursive mode, no glob expansion, and no multi-source/multi-target form. The top-level --help is generated by clap and grouped by resource family (src/bin/s3util/cli/mod.rs).
Argument validation and operator-mistake protection
Validation runs at argument-parse time, before any network call:
- Path shape per subcommand. Transfer subcommands and
rm/head-object/ object-tagging subcommands requires3://<BUCKET>/<KEY>; bucket-management subcommands requires3://<BUCKET>and reject paths with a key. Mismatches exit with clap's code 2. - Source-side rejection on
cp/mv. Source URLs ending in/, source basenames of.or.., and local directory sources are rejected. - Mutual exclusivity.
put-bucket-versioningrequires exactly one of--enabled/--suspendedvia clap's argument groups, so omitting the intent flag fails parse rather than defaulting to a destructive action. - Destination preview. With
--show-progress, the resolved destination path is printed on a-> <path>line before the transfer summary (src/bin/s3util/cli/indicator.rs). - Local target parent must exist. Downloads do not create missing directories; the runtime returns an error and asks the operator to create them (
src/storage/local/fs_util.rs). --if-none-matchprovides "create only" semantics on uploads, plumbed throughUploadManager(src/storage/s3/upload_manager.rs:482, 1062, 1879, 1913).--dry-run(added in v1.1.0). Every subcommand that mutates S3 state —cp,mv,rm,create-bucket, the elevenput-*subcommands, and the ninedelete-*subcommands — accepts--dry-run. With the flag, the binary runs full preflight (argument validation, JSON-config parsing, SDK client construction, transfer-pipeline setup), suppresses the destructive call, emits an info-level[dry-run]log line naming the operation that would have run, and exits 0. Verbosity is forced to at least info so the message is visible at the default warn level. Read-only subcommands (get-*,head-*) deliberately have no--dry-run. Coverage:tests/cli_dry_run.rs(40 annotations) andtests/e2e_dry_run.rs(24 annotations).
The integration suite covers these paths without network access across 40 tests/cli_*.rs files (293 #[test] / #[tokio::test] annotations in total), including the cross-cutting tests/cli_arg_validation.rs, tests/cli_config_validation_error.rs, tests/cli_dry_run.rs, tests/cli_tracing_to_stderr.rs, and tests/cli_command_api_mapping.rs, plus a per-subcommand file for each cp, mv, rm, head-*, create-bucket, delete-*, put-*, and get-* runtime. The process-level tests invoke the binary end-to-end and assert exit codes and stderr; the command/API mapping test statically pins the dispatch chain.
There is no interactive --yes prompt and no two-step confirmation. --dry-run is the inspect-before-act mechanism; a typo on a destructive command typed without --dry-run will still act immediately.
Protection against software bugs in the transfer engine (cp, mv)
The transfer engine — adapted from s3sync — has the highest blast radius (silent corruption, unread bytes, out-of-order multipart assembly, broken checksum comparison). Concrete safeguards:
- Per-upload verification. When the source is a local file or stdin, the upload-side ETag and (when configured) the additional checksum are computed inline and compared against the values S3 returns. A mismatch is a hard error — the object is treated as corrupted (
src/storage/e_tag_verify.rs,src/storage/additional_checksum_verify.rs). - S3→S3 mismatches are warnings (exit 3), not errors, because differing chunksizes between source and target legitimately produce different composite values;
--auto-chunksizemirrors source layout so composite values match end-to-end. - Algorithmic diversity. Composite and full-object variants of
CRC32,CRC32C,CRC64NVME,SHA1,SHA256are implemented insrc/storage/checksum/(each algorithm at 100% region coverage in the report). The default single-part upload path also sendsContent-MD5(src/storage/s3/upload_manager.rs:501, 811, 830, 1081, 1201, 1228, 1427, 1537), giving S3 an end-to-end check that does not depend ons3util's own code being bug-free;--disable-content-md5-headeropts out. - Multipart cleanup on errors.
UploadManager::abort_multipart_uploadis invoked from every error path (src/storage/s3/upload_manager.rs:402, 1034, 1040, 1069) and is best-effort: aNoSuchUploadfrom a race with a completed upload is logged and not rethrown. - Cancellation.
spawn_ctrl_c_handler(src/bin/s3util/cli/ctrl_c_handler.rs) is installed exclusively fromrun_copy_phase(src/bin/s3util/cli/mod.rs:163), shared bycpandmv. SIGINT cancels thePipelineCancellationToken, the in-flight multipart upload is aborted, and the process exits 130. The handler is unit-tested against a real SIGINT (ctrl_c_handler_handles_sigintin the same file).
Live-AWS coverage of these paths includes seven multipart integrity tests across distinct file/chunk size combinations (tests/e2e_multipart_integrity_check_*.rs), six roundtrip suites (e2e_roundtrip_local_to_s3.rs, e2e_roundtrip_s3_to_s3.rs, e2e_roundtrip_multipart_etag.rs, e2e_roundtrip_stdio.rs, e2e_roundtrip_checksum.rs, e2e_roundtrip_express_one_zone.rs), three stdin/stdout integrity suites (e2e_stdio_integrity_check.rs, e2e_stdio_metadata.rs, e2e_stdio_sse.rs), cancellation correctness (e2e_cancel_test.rs), exit-code correctness across all paths (e2e_exit_codes.rs), special-character keys (e2e_special_characters.rs), Express One Zone behaviour (e2e_express_one_zone.rs), unsigned public-bucket access (e2e_source_no_sign_request.rs), and --dry-run behaviour (e2e_dry_run.rs). The full e2e tree is 45 files containing 668 #[test] / #[tokio::test] annotations.
Protection against software bugs in the thin S3 API wrappers
Each thin wrapper is one async function in src/storage/s3/api.rs plus a runtime in src/bin/s3util/cli/<name>.rs. The dangerous classes of bug differ from the transfer engine:
- Error mis-routing. A
HeadErrorenum (src/storage/s3/api.rs:67) distinguishesBucketNotFound,NotFound, andOther. Theclassify_not_foundhelper (src/storage/s3/api.rs:249) routesNoSuchBuckettoBucketNotFoundbefore consulting any per-subresource list, so a missing bucket is never reported as a missing tag/policy/configuration. Eachget-*subcommand has its ownGET_*_NOT_FOUND_CODESconstant; 11*_not_found_codes_pinnedunit tests assert the exact contents of those constants, and fourclassify_not_found_*tests assert thatNoSuchBucketalways routes toBucketNotFoundahead of any subresource code. A typo or accidental edit fails unit tests instead of producing a wrong exit code at e2e time. - Wrong-operation routing.
tests/cli_command_api_mapping.rspins the chain from clap subcommand tomain.rsruntime, from runtime toapi::*wrapper, and from each wrapper to the expected AWS SDK operation. For example,get-bucket-policymust dispatch torun_get_bucket_policy, callapi::get_bucket_policy, and wrapclient.get_bucket_policy()rather than a sibling operation such as versioning or deletion. - Output JSON shape drift. The SDK types do not implement
Serialize, and the SDK field shape does not matchaws s3api --output json.src/output/json.rshand-serialises everyhead-*/get-*response (e.g.get_bucket_policy_to_jsonat line 30, which double-encodesPolicyto matchaws s3api's shape) and contains 83 unit-test annotations covering field naming, omission of absent fields, and the{}empty-body cases forget-bucket-versioning/get-bucket-logging/get-bucket-notification-configuration. - Input JSON shape drift. JSON-consuming
put-*subcommands (put-bucket-cors,put-bucket-encryption,put-bucket-lifecycle-configuration,put-bucket-logging,put-bucket-notification-configuration,put-bucket-website,put-public-access-block) parse user-supplied JSON through dedicated mirror structs insrc/input/json.rs(e.g.CorsConfigurationJson) before converting to SDK types.src/input/json.rscarries 64 unit-test annotations. E2E tests assert that every JSON-consumingput-*subcommand rejects malformed JSON with exit code 1 from both file and stdin. - Verbatim policy body.
put-bucket-policyis the one exception: the body is read with synchronousstd::io::Read::read_to_stringand forwarded verbatim (src/bin/s3util/cli/put_bucket_policy.rs). S3 itself rejects malformed policies with400 MalformedPolicy, so a bad body cannot silently apply, buts3utilperforms no client-side schema validation and there is no in-process size cap (S3's documented ~20 KB policy limit is the effective bound). - Silent partial state. The only thin wrapper that issues more than one S3 API call is
create-bucket --tagging(CreateBucketthenPutBucketTagging). When the second call fails, the bucket exists untagged; the runtime emits a tracing warning naming the partial state and exits 3 with the recovery hint "Retry tagging or delete the bucket manually." There is no automatic rollback — by design.
Cancellation in the thin wrappers: no SIGINT handler is installed. Each wrapper is a single SDK call; ctrl-c terminates the process and the in-flight HTTP request is aborted at the connection layer. There is no in-process multi-step state to clean up except the documented create-bucket --tagging window.
Cross-cutting concerns
- Credential handling.
AccessKeys,SseKmsKeyId, andSseCustomerKey(src/types/mod.rs:323, 344, 358) deriveZeroize+ZeroizeOnDropand overrideDebugto print** redacted **for secret fields. Thetrace_config_summaryhelper insrc/bin/s3util/main.rs:878enumerates non-sensitive fields explicitly rather than{:?}-printing the wholeConfig, so a future field addition cannot leak via tracing without a deliberate edit. Three unit tests insrc/types/mod.rs(lines 387–408) assert the redaction behaviour. - TLS / crypto stack.
Cargo.tomlopts out ofaws-configandaws-sdk-s3default features and re-enablesdefault-https-clientto pin the modern rustls path. The comment on lines 24–27 explains the goal: drop the legacyrustls 0.21alias and its vulnerablerustls-webpki 0.101.x(RUSTSEC-2026-0098).Cargo.lockconfirmsrustls 0.23.39,rustls-webpki 0.103.13, andring 0.17.14.openssl-sysis on thecargo-denyban list (deny.toml). - Supply chain enforcement.
cargo deny checkruns in two places: as thecargo_denyjob in.github/workflows/ci.ymlon every push and PR, and on a daily 01:34 UTC schedule via.github/workflows/cargo-deny.yml.deny.tomlsetsadvisories.ignore = [], so any new RUSTSEC advisory fails the build until it is reviewed and the ignore list is updated.unknown-registry = "deny"andunknown-git = "deny"reject crates from unverified sources; the license allowlist is restricted to standard permissive licenses. Direct dependencies inCargo.tomlare pinned to specific versions rather than wildcards. - CI.
.github/workflows/ci.ymlbuilds and unit-tests on stable Rust on a matrix of seven targets:x86_64-unknown-linux-gnu,x86_64-unknown-linux-musl,aarch64-unknown-linux-gnu,aarch64-unknown-linux-musl,x86_64-pc-windows-msvc,aarch64-pc-windows-msvc, andaarch64-apple-darwin. Therustfmtjob runscargo fmt --all --check, and theclippyjob runscargo clippy -- -D warningson Linux x86_64. A separate workflow (.github/workflows/rust-clippy.yml) runscargo clippy --all-featuresweekly and uploads results to the GitHub code-scanning view. The e2e suite (--cfg e2e_test) is not run by CI because it requires AWS credentials and creates real billable resources; it runs only on the maintainer's machine. - Release pipeline.
.github/workflows/cd.ymlis gated by acreate-releasejob and produces signed build provenance attestations (actions/attest-build-provenance@v4) for every published artifact across the seven release targets, then publishes to crates.io with--lockedvia therust-lang/crates-io-auth-action@v1workflow.
Coverage measurement
lcov_report.txt (generated by cargo llvm-cov over the combined unit-test and --cfg e2e_test runs) reports the following totals across the project: 96.70% region coverage (1,027 of 31,090 regions missed), 95.47% function coverage (100 of 2,208 functions missed), 97.54% line coverage (534 of 21,715 lines missed). Notable per-file figures relevant to safety-critical paths:
storage/checksum/crc32.rs,crc32_c.rs,crc64_nvme.rs,sha1.rs,sha256.rs, andmod.rs: each at 100.00% region coverage.storage/e_tag_verify.rs: 98.82% regions, 99.52% lines.storage/additional_checksum_verify.rs: 98.59% regions, 99.43% lines.storage/s3/upload_manager.rs: 95.50% regions, 94.40% lines.transfer/mod.rs,transfer/progress.rs: each at 100.00% regions.transfer/s3_to_local.rs: 99.05% regions;transfer/s3_to_s3.rs: 97.83%;transfer/stdio_to_s3.rs: 98.61%.output/json.rs: 99.46% regions;input/json.rs: 98.04% regions.storage/s3/api.rs(where the thin-wrapper SDK calls andclassify_not_foundlive): 94.59% regions, 94.44% lines.
Coverage is a structural metric, not a correctness proof: a covered line can still be incorrect, and an uncovered line can still be correct. The numbers above measure how much of the code is exercised by the test suites, not whether the code is right. The lcov report on disk reflects a single combined run on the maintainer's machine; CI does not produce or gate on it.
Known limitations
- Best-effort S3-compatible support. The behaviour observed against Amazon S3 (including Express One Zone) is what the e2e suite asserts. Non-AWS S3-compatible stores may differ; flags such as
--disable-multipart-verify,--disable-etag-verify,--disable-additional-checksum-verify, and--target-force-path-styleexist for those cases. The thin wrappers depend on S3 returning the documented error codes (NoSuchBucketPolicy,NoSuchTagSet,NoSuchLifecycleConfiguration, …); a compatible store that returns a different code falls through to exit code 1 ("Other") rather than the dedicated NotFound exit code 4. - No interactive confirmation prompt.
--dry-run(v1.1.0) is the inspect-before-act mechanism forrm,create-bucket, the elevenput-*subcommands, and the ninedelete-*subcommands, but there is no--yes/-yprompt and no two-step confirmation. A typo on a destructive command typed without--dry-runstill acts immediately on argument parse. create-bucket --taggingis not transactional. Documented and surfaced as exit code 3 with an explicit recovery hint, but the operator must act on the warning.put-bucket-policyperforms no client-side schema validation. By design — body forwarded verbatim, S3 is the authority on policy validity.- Ten
panic!()invariant assertions in storage code:src/storage/e_tag_verify.rs:151,src/storage/additional_checksum_verify.rs:33, 109, 112,src/storage/s3/upload_manager.rs:269, 2154, 2182,src/storage/s3/mod.rs:97,src/storage/local/mod.rs:111, andsrc/storage/checksum/mod.rs:47. Each guards a caller-side precondition (e.g. "object_parts is empty", "unknown algorithm") and should be unreachable in normal operation; their reachability is not currently demonstrated by any test, and converting them toResult::Errwould be a strict improvement in robustness. - E2E tests run only on the maintainer's account. CI does not exercise live AWS, so any CI green signal does not by itself confirm correct behaviour against real S3 — the maintainer's local e2e run is the load-bearing evidence for that.
- No external security audit, fuzz testing, sanitizer runs, or formal threat model. Fuzzing of
src/input/json.rsandsrc/output/json.rs, andMIRIFLAGS-based UB checks across the upload/download paths, would be reasonable next steps.
Is the software reliable?
For readers who are not software engineers: "reliable" in this context means whether a careful operator can use s3util for routine S3 work without losing data, leaking credentials, or having the binary do something other than what was typed. The evidence above maps onto three concrete risks.
Risk: data corruption on transfer. When s3util uploads a file, the file's checksum is computed locally and compared against the value S3 returns; on the default single-part path the binary also sends Content-MD5 so S3 itself rejects a corrupted body before storing it. A mismatch is treated as a failed upload, not a warning. Live-AWS tests exercise this across seven combinations of file size and chunk size, six round-trip paths (local↔S3, S3↔S3, stdio↔S3, multipart-ETag, additional checksum, Express One Zone), three stdin/stdout integrity suites, two storage classes (S3 Standard and Express One Zone), special-character keys, and cancellation at every stage. The most damaging classes of failure — silent corruption, an upload reporting success without all bytes, a multipart upload abandoned mid-flight — are detected by mechanisms that do not all depend on s3util's own code being bug-free.
Risk: credential leakage. Access keys, KMS key IDs, and SSE-C keys are wrapped in types that wipe their memory on drop and print as ** redacted ** in any debug output. The startup tracing helper enumerates non-sensitive fields by name rather than printing the whole config; three unit tests pin this redaction. The TLS stack pins rustls 0.23.39 and excludes the legacy rustls 0.21 alias affected by RUSTSEC-2026-0098. cargo deny runs on every push, every PR, and once daily, with advisories.ignore = []: any new RUSTSEC entry that affects a dependency fails the build until it is explicitly reviewed.
Risk: unintended action from a typed command. The CLI accepts exactly one resource per invocation — no recursion, no globbing, no multi-target form. Argument parsing rejects malformed paths before any network call. Every mutating subcommand (cp, mv, rm, create-bucket, eleven put-*, nine delete-*) supports --dry-run, which runs full preflight, prints the operation that would have occurred, and exits 0 without contacting S3. Read-only subcommands (get-*, head-*) deliberately have no --dry-run. There is no interactive yes/no prompt: typing delete-bucket s3://wrong-name without --dry-run will act on it. The single-resource scope bounds the worst case to one object or one bucket per invocation.
What this assessment cannot establish. That the code is bug-free, that future advisories will not affect a dependency before it is patched, or that an S3-compatible store other than Amazon S3 will behave identically. The binary has not been certified by an external security audit, and no automated test suite can prove the absence of all bugs.
What it does establish. The failure modes most likely to cause silent harm — data corruption on transfer, leaked credentials, mis-routed errors on read, partial writes that look successful — are either prevented by design, detected at runtime, or surfaced as a non-zero exit code rather than passing silently.
In plain terms: the categories of failure that would most concern a non-engineer user — corrupted files, leaked secrets, accidental destructive actions — each have a specific, citable safeguard documented above, and most of those safeguards do not depend on s3util's own code being bug-free. The fact that the codebase was generated by AI does not by itself raise or lower its reliability; what determines that is the design, the tests, and the verifiable evidence above. With the standard precautions any tool requires — preview destructive commands with --dry-run, validate against representative data before adopting it for important pipelines, monitor exit codes — the documented safeguards address the failure modes most likely to cause silent harm on Amazon S3. Final responsibility for whether the binary is right for a given workflow rests with the operator who runs it.
AI assessment of safety and correctness (by Codex)
Assessment date: 2026-04-30.
Assessed workspace: branch update/v1.0.1, commit 10b12eb, with local coverage artifacts lcov.info and lcov_report.txt present in the workspace.
Evidence reviewed: src/bin/s3util/, src/config/, src/storage/, src/transfer/, src/input/, src/output/, src/types/, tests/cli_*.rs, tests/e2e_*.rs, tests/common/mod.rs, Cargo.toml, Cargo.lock, deny.toml, .github/workflows/, lcov.info, and lcov_report.txt.
Limits of this assessment: the coverage files show which code was exercised, not a proof that the exercised code is correct. This review did not include fuzzing, sanitizer runs, Miri runs, formal verification, penetration testing, or an external security audit. The coverage artifacts do not contain a per-test pass/fail log.
This assessment treats AI-generated authorship as neither a safety control nor a defect by itself. Only source behavior, test evidence, coverage artifacts, dependency controls, and stated limits are counted as evidence below.
Command Surface And Scope
src/config/args/mod.rsdefines 37 subcommands.- The command set contains
cp,mv,rm,create-bucket, twohead-*commands, elevenget-*commands, elevenput-*commands, and ninedelete-*commands. cpandmvuse the transfer pipeline insrc/bin/s3util/cli/mod.rsandsrc/transfer/.- Non-transfer subcommands dispatch through runtime files in
src/bin/s3util/cli/and S3 SDK wrapper functions insrc/storage/s3/api.rs. - The CLI does not implement recursive transfer, S3-key globbing, or multiple-source transfer syntax. Each invocation addresses one object, one bucket, or one bucket subresource.
src/config/args/value_parser/storage_path.rsaccepts S3 URLs, local paths, multi-region access point ARNs, and-for stdio. It rejects empty paths, malformed URLs, and non-s3://URLs except Windows absolute paths.src/transfer/mod.rsrejects both-local transfers, both-stdio transfers, local-to-stdio transfers, and stdio-to-local transfers.
Destructive-Operation Controls
- Every mutating subcommand exposes
--dry-run:cp,mv,rm,create-bucket, all elevenput-*commands, and all ninedelete-*commands. - Read-only
get-*andhead-*commands do not expose--dry-run. --dry-runshort-circuits before the S3-mutating call and logs a[dry-run]message.tests/cli_dry_run.rscontains process-level tests for this behavior, andtests/e2e_dry_run.rscontains live-AWS dry-run tests.--dry-runvalidates command arguments and JSON input before the short-circuit in the reviewed runtimes. It does not prove that AWS credentials, IAM policy, object existence, bucket existence, endpoint reachability, or later AWS-side validation would succeed.- There is no interactive confirmation prompt and no required
--yesflag. If a destructive command is syntactically valid and is run without--dry-run, it proceeds to the S3 call.
Transfer-Correctness Controls
- Upload paths compute source-side ETag and, when configured, additional checksums in
src/storage/e_tag_verify.rs,src/storage/additional_checksum_verify.rs, andsrc/storage/s3/upload_manager.rs. - For local-file and stdin uploads, checksum or ETag mismatch is treated as an error unless the user explicitly disables the relevant verification mode.
- For S3-to-S3 transfers, multipart-layout differences can make ETag or composite checksum values differ even when payload bytes match. The implementation treats those cases as warnings unless the source uses a full-object checksum or a locally calculated checksum, where mismatch is treated as an error.
- Multipart upload code sorts completed parts by part number before completion and compares the uploaded byte count with the source size before completing.
- Multipart error paths call
AbortMultipartUploadon a best-effort basis. - The default single-part upload path sends
Content-MD5except where explicitly disabled or not applicable to the storage mode. mvdeletes the source only after the copy phase returns success. It does not delete the source after cancellation, copy error, or verification warning unless--no-fail-on-verify-erroris set.- SIGINT handling maps to the transfer cancellation token for transfer commands.
S3 API And JSON Controls
src/storage/s3/api.rscentralizes S3 SDK calls used by the non-transfer command runtimes.- Missing-resource handling distinguishes bucket-not-found cases from subresource-not-found cases where the S3 error code is known.
src/output/json.rsmanually serializes selected SDK responses forhead-*andget-*commands.src/input/json.rsparses JSON for bucket CORS, encryption, lifecycle, logging, notification, website, and public-access-block configuration through local mirror structs before converting to SDK types.put-bucket-policyreads the policy body from a file or stdin and forwards it to S3. It does not perform client-side policy schema validation.create-bucket --taggingperformsCreateBucketfollowed byPutBucketTagging. If bucket creation succeeds and tagging fails, the command returns warning status and logs that the bucket exists untagged; it does not roll back the bucket.
Coverage Evidence
lcov.infois the machine-readable LCOV artifact.lcov_report.txtis the summarized table used for the percentages below.lcov_report.txtreports total coverage of 96.70% regions, 95.47% functions, and 97.54% lines.- The same report shows 1,027 missed regions out of 31,090, 100 missed functions out of 2,208, and 534 missed lines out of 21,715.
- Branch coverage is not measured: the branch columns show zero counted branches and no percentage.
- Checksum algorithm files
storage/checksum/crc32.rs,crc32_c.rs,crc64_nvme.rs,sha1.rs,sha256.rs, andstorage/checksum/mod.rseach report 100.00% region, function, and line coverage. - Verification files report high line coverage:
storage/e_tag_verify.rsat 99.52% lines andstorage/additional_checksum_verify.rsat 99.43% lines. storage/s3/upload_manager.rs, the largest transfer implementation file, reports 95.50% regions, 96.03% functions, and 94.40% lines.- Transfer modules report the following line coverage:
transfer/mod.rs100.00%,transfer/progress.rs100.00%,transfer/local_to_s3.rs97.25%,transfer/s3_to_local.rs99.50%,transfer/s3_to_s3.rs98.81%,transfer/s3_to_stdio.rs92.23%, andtransfer/stdio_to_s3.rs99.05%. - JSON modules report 99.64% line coverage for
input/json.rsand 99.84% line coverage foroutput/json.rs. - The lowest region coverage values in the report are
bin/s3util/cli/get_bucket_logging.rsat 81.08%,bin/s3util/cli/get_bucket_notification_configuration.rsat 81.08%,bin/s3util/cli/create_bucket.rsat 81.48%, andbin/s3util/cli/delete_bucket_policy.rsat 82.61%. - The repository contains 45
tests/e2e_*.rsfiles with 668#[test]or#[tokio::test]annotations. These files are gated by#![cfg(e2e_test)]. tests/common/mod.rsbuilds an AWS S3 client using profiles3util-e2e-testand contains helper operations for bucket creation, bucket deletion, object upload, object download,HeadObject, object tagging, and cleanup.- The GitHub Actions CI workflow runs
cargo build,cargo test,cargo fmt --all --check,cargo clippy -- -D warnings, andcargo deny -L error check. The CI test matrix covers seven targets. The live-AWS e2e suite is not part of the default CI workflow. It requires AWS credentials and creates real AWS resources.
Secret Handling And Dependencies
AccessKeys,SseKmsKeyId, andSseCustomerKeyinsrc/types/mod.rsderive zeroization traits and redact secret material inDebugoutput.trace_config_summaryinsrc/bin/s3util/main.rslogs selected configuration fields instead of formatting the completeConfigvalue.Cargo.tomldisables default features foraws-configandaws-sdk-s3, then enables the SDK HTTPS client feature set.Cargo.lockresolves the TLS-related packages torustls 0.23.39,rustls-webpki 0.103.13, andring 0.17.14.deny.tomlhasadvisories.ignore = [], denies unknown registries, denies unknown git sources, and deniesopenssl-sys.- A scheduled GitHub Actions workflow runs
cargo deny -L error checkdaily.
Known Limits
- A coverage percentage does not prove correctness. A line can be executed by tests and still contain a defect.
- Branch coverage is not measured by the current report.
- The e2e suite is gated by
cfg(e2e_test)and is not run by defaultcargo testin CI. --dry-runis a preflight and short-circuit mechanism. It is not an AWS-side authorization, existence, or policy check.- Destructive and replacement operations have no interactive confirmation prompt.
- The command parser cannot detect an intended-but-wrong bucket name, key, version ID, policy document, or configuration file if the input is syntactically valid.
- S3-compatible storage support depends on the compatible service returning error codes and checksum behavior close enough to Amazon S3 for the wrapper logic to classify them correctly.
- Production modules contain
panic!()orexpect()paths for invalid internal states, unsupported checksum algorithms, unexpected storage path variants, and unknown filter names. If one of those paths is reached in a release binary, the process terminates instead of returning a structured error. put-bucket-policydelegates policy validity to S3.create-bucket --taggingis not transactional.- This assessment did not include fuzzing of path parsing, metadata parsing, tagging parsing, or JSON parsing.
Reliability Conclusion
For a non-engineer reader, "reliable" here means whether the tool has observable safeguards against the most harmful silent failures: corrupted transferred data, leaked secrets, and a command doing more than the user asked.
The strongest evidence is for transfer correctness. The code contains checksum and ETag verification, multipart byte-count checks, ordered multipart completion, cancellation handling, and best-effort multipart abort. The relevant implementation files have the line coverage listed above in lcov_report.txt, and the e2e test tree covers live-AWS transfer paths when run with cfg(e2e_test).
The evidence for non-transfer S3 administration commands is narrower. The wrappers are small, argument validation is covered by tests, JSON input/output line coverage is above 99%, and --dry-run exists for every mutating subcommand. These controls reduce accidental execution, but they do not make destructive operations reversible and do not confirm that a syntactically valid target is the target the operator intended.
Overall classification: conditionally reliable, not unconditionally reliable. The software is reliable for routine Amazon S3 transfers in the specific sense that it has explicit runtime checks and tests for the main silent data-corruption cases. It is not reliable in the stronger sense of being proven bug-free, externally audited, protected from all operator mistakes, or guaranteed to behave the same against every S3-compatible service. Users should treat --dry-run, exit codes, backups, IAM restrictions, and representative testing as required operating controls rather than optional precautions.
AI assessment of safety and correctness (by Gemini)
Assessment date: 2026-04-30.
Assessed version: 1.1.0 (commit 1ef94b9).
Scope: Source code tree (~35,500 lines of Rust), unit and integration tests (1,901 total test annotations), dependency graph, and CI/CD pipelines.
1. Quantitative Quality Metrics
The assessment is based on a combined analysis of unit tests and integration suites (CLI and E2E).
- Test Volume: The codebase is verified by 940 unit tests, 293 CLI integration tests, and 668 E2E integration tests (1,901 total).
- Structural Coverage: The test suite exercises 97.54% of all lines (21,181 of 21,715 lines) and 96.70% of regions (30,063 of 31,090 regions).
- Algorithm Integrity: Cryptographic implementations for ETag and additional checksums (CRC32, CRC32C, CRC64NVME, SHA1, SHA256) are covered at 100.00% across all lines and regions.
- Error Handling: Missing-resource classification logic in
storage/s3/api.rsis verified with 94.44% line coverage, ensuring correct exit codes for 404/403 errors.
2. Data Integrity and Transfer Reliability
The software implements mandatory cryptographic verification to prevent silent data corruption.
- Verification Logic: Routines in
storage/e_tag_verify.rsandstorage/additional_checksum_verify.rsmaintain >99.4% line coverage, ensuring every byte is validated against S3 metadata. - Multipart Assembly: The
UploadManager(94.40% line coverage) enforces strict part ordering and byte-count validation before finalizing uploads, mitigating risks of incomplete object creation. - Transactional Moves: For
mvoperations, the source is only deleted after the copy and integrity check return a zero exit code, preventing data loss on failed transfers.
3. Operational Safety Guardrails
The CLI includes specific mechanisms to mitigate operator error.
- Dry-run Support: All 22 mutating subcommands support
--dry-run, which performs full argument and JSON validation before short-circuiting the destructive API call. - Scope Containment: By strictly limiting each command to one resource (no recursion or globbing), the software prevents wide-scale accidental modifications.
- Credential Protection: Sensitive fields (Access Keys, SSE-C keys) derive zeroization traits to wipe memory on drop and are redacted in all debug/trace outputs.
4. Verification Gaps
- Integration Blind Spot: The 668 E2E tests are gated behind
cfg(e2e_test)and are not executed in the standard CI pipeline, requiring manual maintainer execution for live-AWS verification. - Schema Validation:
put-bucket-policylacks client-side schema checks, relying entirely on the S3 API for document validation.
Overall Conclusion
Is this software reliable? Yes. For Amazon S3 workloads, s3util is technically reliable because its core transfer logic is verified by nearly 2,000 tests and maintains 97.54% line coverage.
For a non-technical user, "reliability" means the tool is highly unlikely to corrupt your files or leak your secrets. It uses digital "seals" (checksums) to verify every transfer—a process that is 100% tested in the code. With the addition of "dry-run" mode, you can safely preview any command before it takes effect. While it requires careful typing (due to the lack of "Are you sure?" prompts), the deep testing and single-file focus make it a stable choice for managing S3 data.
Contributing
- Suggestions and bug reports are welcome, but responses are not guaranteed.
- Pull requests for new features are generally not accepted, as they may conflict with the design philosophy.
- If you find this project useful, feel free to fork and modify it as you wish.
🔒 I consider this project to be "complete" and will maintain it only minimally going forward.
However, I intend to keep the AWS SDK for Rust and other dependencies up to date monthly.
License
This project is licensed under the Apache-2.0 License.