pub struct DatasetWriteGuard<'a> { /* private fields */ }Methods from Deref<Target = Dataset>§
Sourcepub async fn checkout_version(
&self,
version: impl Into<Ref>,
) -> Result<Dataset, Error>
pub async fn checkout_version( &self, version: impl Into<Ref>, ) -> Result<Dataset, Error>
Check out a dataset version with a ref
pub fn branches(&self) -> Branches<'_>
Sourcepub async fn checkout_latest(&mut self) -> Result<(), Error>
pub async fn checkout_latest(&mut self) -> Result<(), Error>
Check out the latest version of the dataset
Sourcepub async fn checkout_branch(&self, branch: &str) -> Result<Dataset, Error>
pub async fn checkout_branch(&self, branch: &str) -> Result<Dataset, Error>
Check out the latest version of the branch
Sourcepub async fn create_branch(
&mut self,
branch: &str,
version: impl Into<Ref>,
store_params: Option<ObjectStoreParams>,
) -> Result<Dataset, Error>
pub async fn create_branch( &mut self, branch: &str, version: impl Into<Ref>, store_params: Option<ObjectStoreParams>, ) -> Result<Dataset, Error>
This is a two-phase operation:
- Create the branch dataset by shallow cloning.
- Create the branch metadata (a.k.a.
BranchContents).
These two phases are not atomic. We consider BranchContents as the source of truth
for the branch.
The cleanup procedure should:
- Clean up zombie branch datasets that have no related
BranchContents. - Delete broken
BranchContentsentries that have no related branch dataset.
If create_branch stops at phase 1, it may leave a zombie branch dataset,
which can be cleaned up later. Such a zombie dataset may cause a branch creation
failure if we use the same name to create_branch. In that case, you need to call
force_delete_branch to interactively clean up the zombie dataset.
pub async fn delete_branch(&mut self, branch: &str) -> Result<(), Error>
Sourcepub async fn force_delete_branch(&mut self, branch: &str) -> Result<(), Error>
pub async fn force_delete_branch(&mut self, branch: &str) -> Result<(), Error>
Delete the branch even if the BranchContents is not found. This could be useful when we have zombie branches and want to clean them up immediately.
pub async fn list_branches( &self, ) -> Result<HashMap<String, BranchContents>, Error>
Sourcepub async fn append(
&mut self,
batches: impl RecordBatchReader + Send + 'static,
params: Option<WriteParams>,
) -> Result<(), Error>
pub async fn append( &mut self, batches: impl RecordBatchReader + Send + 'static, params: Option<WriteParams>, ) -> Result<(), Error>
Append to existing Dataset with a stream of RecordBatchs
Returns void result or Returns Error
pub fn branch_location(&self) -> BranchLocation
pub fn find_branch_location( &self, branch_name: &str, ) -> Result<BranchLocation, Error>
pub fn manifest_location(&self) -> &ManifestLocation
Sourcepub fn delta(&self) -> DatasetDeltaBuilder
pub fn delta(&self) -> DatasetDeltaBuilder
Create a delta::DatasetDeltaBuilder to explore changes between dataset versions.
§Example
let delta = dataset.delta()
.compared_against_version(5)
.build()?;
let inserted = delta.get_inserted_rows().await?;pub async fn latest_manifest( &self, ) -> Result<(Arc<Manifest>, ManifestLocation), Error>
Sourcepub async fn read_transaction(&self) -> Result<Option<Transaction>, Error>
pub async fn read_transaction(&self) -> Result<Option<Transaction>, Error>
Read the transaction file for this version of the dataset.
If there was no transaction file written for this version of the dataset then this will return None.
Sourcepub async fn read_transaction_by_version(
&self,
version: u64,
) -> Result<Option<Transaction>, Error>
pub async fn read_transaction_by_version( &self, version: u64, ) -> Result<Option<Transaction>, Error>
Read the transaction file for this version of the dataset.
If there was no transaction file written for this version of the dataset then this will return None.
Sourcepub async fn get_transactions(
&self,
recent_transactions: usize,
) -> Result<Vec<Option<Transaction>>, Error>
pub async fn get_transactions( &self, recent_transactions: usize, ) -> Result<Vec<Option<Transaction>>, Error>
List transactions for the dataset, up to a maximum number.
This method iterates through dataset versions, starting from the current version,
and collects the transaction for each version. It stops when either recent_transactions
is reached or there are no more versions.
§Arguments
recent_transactions- Maximum number of transactions to return
§Returns
A vector of optional transactions. Each element corresponds to a version, and may be None if no transaction file exists for that version.
Sourcepub async fn restore(&mut self) -> Result<(), Error>
pub async fn restore(&mut self) -> Result<(), Error>
Restore the currently checked out version of the dataset as the latest version.
Sourcepub fn cleanup_old_versions(
&self,
older_than: TimeDelta,
delete_unverified: Option<bool>,
error_if_tagged_old_versions: Option<bool>,
) -> Pin<Box<dyn Future<Output = Result<RemovalStats, Error>> + Send + '_>>
pub fn cleanup_old_versions( &self, older_than: TimeDelta, delete_unverified: Option<bool>, error_if_tagged_old_versions: Option<bool>, ) -> Pin<Box<dyn Future<Output = Result<RemovalStats, Error>> + Send + '_>>
Removes old versions of the dataset from disk
This function will remove all versions of the dataset that are older than the provided timestamp. This function will not remove the current version of the dataset.
Once a version is removed it can no longer be checked out or restored. Any data unique to that version will be lost.
§Arguments
older_than- Versions older than this will be deleted.delete_unverified- If false (the default) then files will only be deleted if they are listed in at least one manifest. Otherwise these files will be kept since they cannot be distinguished from an in-progress transaction. Set to true to delete these files if you are sure there are no other in-progress dataset operations.
§Returns
RemovalStats- Statistics about the removal operation
Sourcepub fn cleanup_with_policy(
&self,
policy: CleanupPolicy,
) -> Pin<Box<dyn Future<Output = Result<RemovalStats, Error>> + Send + '_>>
pub fn cleanup_with_policy( &self, policy: CleanupPolicy, ) -> Pin<Box<dyn Future<Output = Result<RemovalStats, Error>> + Send + '_>>
Removes old versions of the dataset from storage
This function will remove all versions of the dataset that satisfies the given policy. This function will not remove the current version of the dataset.
Once a version is removed it can no longer be checked out or restored. Any data unique to that version will be lost.
§Arguments
policy-CleanupPolicydetermines the behaviour of cleanup.
§Returns
RemovalStats- Statistics about the removal operation
Sourcepub async fn count_rows(&self, filter: Option<String>) -> Result<usize, Error>
pub async fn count_rows(&self, filter: Option<String>) -> Result<usize, Error>
Count the number of rows in the dataset.
It offers a fast path of counting rows by just computing via metadata.
Sourcepub async fn take(
&self,
row_indices: &[u64],
projection: impl Into<ProjectionRequest>,
) -> Result<RecordBatch, Error>
pub async fn take( &self, row_indices: &[u64], projection: impl Into<ProjectionRequest>, ) -> Result<RecordBatch, Error>
Take rows by indices.
Sourcepub async fn take_rows(
&self,
row_ids: &[u64],
projection: impl Into<ProjectionRequest>,
) -> Result<RecordBatch, Error>
pub async fn take_rows( &self, row_ids: &[u64], projection: impl Into<ProjectionRequest>, ) -> Result<RecordBatch, Error>
Take Rows by the internal ROW ids.
In Lance format, each row has a unique u64 id, which is used to identify the row globally.
let schema = dataset.schema().clone();
let row_ids = vec![0, 4, 7];
let rows = dataset.take_rows(&row_ids, schema).await.unwrap();
// We can have more fine-grained control over the projection, i.e., SQL projection.
let projection = ProjectionRequest::from_sql([("identity", "id * 2")]);
let rows = dataset.take_rows(&row_ids, projection).await.unwrap();pub fn take_builder( self: &Arc<Dataset>, row_ids: &[u64], projection: impl Into<ProjectionRequest>, ) -> Result<TakeBuilder, Error>
Sourcepub async fn take_blobs(
self: &Arc<Dataset>,
row_ids: &[u64],
column: impl AsRef<str>,
) -> Result<Vec<BlobFile>, Error>
pub async fn take_blobs( self: &Arc<Dataset>, row_ids: &[u64], column: impl AsRef<str>, ) -> Result<Vec<BlobFile>, Error>
Take BlobFile by row ids (row address).
Sourcepub async fn take_blobs_by_addresses(
self: &Arc<Dataset>,
row_addrs: &[u64],
column: impl AsRef<str>,
) -> Result<Vec<BlobFile>, Error>
pub async fn take_blobs_by_addresses( self: &Arc<Dataset>, row_addrs: &[u64], column: impl AsRef<str>, ) -> Result<Vec<BlobFile>, Error>
Take BlobFile by row addresses.
Row addresses are u64 values encoding (fragment_id << 32) | row_offset.
Use this method when you already have row addresses, for example from
a scan with with_row_address(). For row IDs (stable identifiers), use
Self::take_blobs. For row indices (offsets), use
Self::take_blobs_by_indices.
Sourcepub async fn take_blobs_by_indices(
self: &Arc<Dataset>,
row_indices: &[u64],
column: impl AsRef<str>,
) -> Result<Vec<BlobFile>, Error>
pub async fn take_blobs_by_indices( self: &Arc<Dataset>, row_indices: &[u64], column: impl AsRef<str>, ) -> Result<Vec<BlobFile>, Error>
Take BlobFile by row indices (offsets in the dataset).
Sourcepub fn take_scan(
&self,
row_ranges: Pin<Box<dyn Stream<Item = Result<Range<u64>, Error>> + Send>>,
projection: Arc<Schema>,
batch_readahead: usize,
) -> DatasetRecordBatchStream
pub fn take_scan( &self, row_ranges: Pin<Box<dyn Stream<Item = Result<Range<u64>, Error>> + Send>>, projection: Arc<Schema>, batch_readahead: usize, ) -> DatasetRecordBatchStream
Get a stream of batches based on iterator of ranges of row numbers.
This is an experimental API. It may change at any time.
Sourcepub async fn delete(&mut self, predicate: &str) -> Result<(), Error>
pub async fn delete(&mut self, predicate: &str) -> Result<(), Error>
Delete rows based on a predicate.
Sourcepub async fn add_bases(
self: &Arc<Dataset>,
new_bases: Vec<BasePath>,
transaction_properties: Option<HashMap<String, String>>,
) -> Result<Dataset, Error>
pub async fn add_bases( self: &Arc<Dataset>, new_bases: Vec<BasePath>, transaction_properties: Option<HashMap<String, String>>, ) -> Result<Dataset, Error>
Add new base paths to the dataset.
This method allows you to register additional storage locations (buckets) that can be used for future data writes. The base paths are added to the dataset’s manifest and can be referenced by name in subsequent write operations.
§Arguments
new_bases- A vector oflance_table::format::BasePathobjects representing the new storage locations to add. Each base path should have a unique name and path.
§Returns
Returns a new Dataset instance with the updated manifest containing the
new base paths.
pub async fn count_deleted_rows(&self) -> Result<usize, Error>
pub fn object_store(&self) -> &ObjectStore
Sourcepub fn storage_options(&self) -> Option<&HashMap<String, String>>
pub fn storage_options(&self) -> Option<&HashMap<String, String>>
Returns the storage options used when opening this dataset, if any.
Sourcepub fn storage_options_provider(
&self,
) -> Option<Arc<dyn StorageOptionsProvider>>
pub fn storage_options_provider( &self, ) -> Option<Arc<dyn StorageOptionsProvider>>
Returns the storage options provider used when opening this dataset, if any.
pub fn data_dir(&self) -> Path
pub fn indices_dir(&self) -> Path
pub fn session(&self) -> Arc<Session>
pub fn version(&self) -> Version
Sourcepub async fn index_cache_entry_count(&self) -> usize
pub async fn index_cache_entry_count(&self) -> usize
Get the number of entries currently in the index cache.
Sourcepub async fn index_cache_hit_rate(&self) -> f32
pub async fn index_cache_hit_rate(&self) -> f32
Get cache hit ratio.
pub fn cache_size_bytes(&self) -> u64
Sourcepub async fn latest_version_id(&self) -> Result<u64, Error>
pub async fn latest_version_id(&self) -> Result<u64, Error>
Get the latest version of the dataset This is meant to be a fast path for checking if a dataset has changed. This is why we don’t return the full version struct.
pub fn count_fragments(&self) -> usize
Sourcepub fn empty_projection(self: &Arc<Dataset>) -> Projection
pub fn empty_projection(self: &Arc<Dataset>) -> Projection
Similar to Self::schema, but only returns fields that are not marked as blob columns Creates a new empty projection into the dataset schema
Sourcepub fn full_projection(self: &Arc<Dataset>) -> Projection
pub fn full_projection(self: &Arc<Dataset>) -> Projection
Creates a projection that includes all columns in the dataset
Sourcepub fn get_fragments(&self) -> Vec<FileFragment>
pub fn get_fragments(&self) -> Vec<FileFragment>
Get fragments.
If filter is provided, only fragments with the given name will be returned.
pub fn get_fragment(&self, fragment_id: usize) -> Option<FileFragment>
pub fn fragments(&self) -> &Arc<Vec<Fragment>>
pub fn get_frags_from_ordered_ids( &self, ordered_ids: &[u32], ) -> Vec<Option<FileFragment>>
Sourcepub async fn num_small_files(&self, max_rows_per_group: usize) -> usize
pub async fn num_small_files(&self, max_rows_per_group: usize) -> usize
Gets the number of files that are so small they don’t even have a full group. These are considered too small because reading many of them is much less efficient than reading a single file because the separate files split up what would otherwise be single IO requests into multiple.
pub async fn validate(&self) -> Result<(), Error>
Sourcepub async fn migrate_manifest_paths_v2(&mut self) -> Result<(), Error>
pub async fn migrate_manifest_paths_v2(&mut self) -> Result<(), Error>
Migrate the dataset to use the new manifest path scheme.
This function will rename all V1 manifests to ManifestNamingScheme::V2. These paths provide more efficient opening of datasets with many versions on object stores.
This function is idempotent, and can be run multiple times without changing the state of the object store.
However, it should not be run while other concurrent operations are happening. And it should also run until completion before resuming other operations.
let mut dataset = Dataset::write(data, "memory://test", None).await.unwrap();
assert_eq!(dataset.manifest_location().naming_scheme, ManifestNamingScheme::V1);
dataset.migrate_manifest_paths_v2().await.unwrap();
assert_eq!(dataset.manifest_location().naming_scheme, ManifestNamingScheme::V2);Sourcepub async fn shallow_clone(
&mut self,
target_path: &str,
version: impl Into<Ref>,
store_params: Option<ObjectStoreParams>,
) -> Result<Dataset, Error>
pub async fn shallow_clone( &mut self, target_path: &str, version: impl Into<Ref>, store_params: Option<ObjectStoreParams>, ) -> Result<Dataset, Error>
Shallow clone the target version into a new dataset at target_path. ‘target_path’: the uri string to clone the dataset into. ‘version’: the version cloned from, could be a version number or tag. ‘store_params’: the object store params to use for the new dataset.
Sourcepub fn sql(&self, sql: &str) -> SqlQueryBuilder
pub fn sql(&self, sql: &str) -> SqlQueryBuilder
Run a SQL query against the dataset. The underlying SQL engine is DataFusion. Please refer to the DataFusion documentation for supported SQL syntax.
Sourcepub async fn add_columns(
&mut self,
transforms: NewColumnTransform,
read_columns: Option<Vec<String>>,
batch_size: Option<u32>,
) -> Result<(), Error>
pub async fn add_columns( &mut self, transforms: NewColumnTransform, read_columns: Option<Vec<String>>, batch_size: Option<u32>, ) -> Result<(), Error>
Append new columns to the dataset.
Sourcepub async fn alter_columns(
&mut self,
alterations: &[ColumnAlteration],
) -> Result<(), Error>
pub async fn alter_columns( &mut self, alterations: &[ColumnAlteration], ) -> Result<(), Error>
Modify columns in the dataset, changing their name, type, or nullability.
If only changing the name or nullability of a column, this is a zero-copy operation and any indices will be preserved. If changing the type of a column, the data for that column will be rewritten and any indices will be dropped. The old column data will not be immediately deleted. To remove it, call optimize::compact_files() and then cleanup::cleanup_old_versions() on the dataset.
Sourcepub async fn drop_columns(&mut self, columns: &[&str]) -> Result<(), Error>
pub async fn drop_columns(&mut self, columns: &[&str]) -> Result<(), Error>
Remove columns from the dataset.
This is a metadata-only operation and does not remove the data from the underlying storage. In order to remove the data, you must subsequently call optimize::compact_files() to rewrite the data without the removed columns and then call cleanup::cleanup_old_versions() to remove the old files.
Sourcepub async fn drop(&mut self, columns: &[&str]) -> Result<(), Error>
👎Deprecated since 0.9.12: Please use drop_columns instead.
pub async fn drop(&mut self, columns: &[&str]) -> Result<(), Error>
drop_columns instead.Drop columns from the dataset and return updated dataset. Note that this is a zero-copy operation and column is not physically removed from the dataset. Parameters:
columns: the list of column names to drop.
Sourcepub async fn merge(
&mut self,
stream: impl RecordBatchReader + Send + 'static,
left_on: &str,
right_on: &str,
) -> Result<(), Error>
pub async fn merge( &mut self, stream: impl RecordBatchReader + Send + 'static, left_on: &str, right_on: &str, ) -> Result<(), Error>
Merge this dataset with another arrow Table / Dataset, and returns a new version of dataset.
Parameters:
stream: the stream ofRecordBatchto merge.left_on: the column name to join on the left side (self).right_on: the column name to join on the right side (stream).
Returns: a new version of dataset.
It performs a left-join on the two datasets.
Sourcepub async fn delete_config_keys(
&mut self,
delete_keys: &[&str],
) -> Result<(), Error>
👎Deprecated: Use the new update_config(values, replace) method - pass None values to delete keys
pub async fn delete_config_keys( &mut self, delete_keys: &[&str], ) -> Result<(), Error>
Delete keys from the config.
Sourcepub fn update_metadata(
&mut self,
values: impl IntoIterator<Item = impl Into<UpdateMapEntry>>,
) -> UpdateMetadataBuilder<'_>
pub fn update_metadata( &mut self, values: impl IntoIterator<Item = impl Into<UpdateMapEntry>>, ) -> UpdateMetadataBuilder<'_>
Update table metadata.
Pass None for a value to remove that key.
Use .replace() to replace the entire metadata map instead of merging.
Returns the updated metadata map after the operation.
// Update single key
dataset.update_metadata([("key", "value")]).await?;
// Remove a key
dataset.update_metadata([("to_delete", None)]).await?;
// Clear all metadata
dataset.update_metadata([] as [UpdateMapEntry; 0]).replace().await?;
// Replace full metadata
dataset.update_metadata([("k1", "v1"), ("k2", "v2")]).replace().await?;Sourcepub fn update_config(
&mut self,
values: impl IntoIterator<Item = impl Into<UpdateMapEntry>>,
) -> UpdateMetadataBuilder<'_>
pub fn update_config( &mut self, values: impl IntoIterator<Item = impl Into<UpdateMapEntry>>, ) -> UpdateMetadataBuilder<'_>
Update config.
Pass None for a value to remove that key.
Use .replace() to replace the entire config map instead of merging.
Returns the updated config map after the operation.
// Update single key
dataset.update_config([("key", "value")]).await?;
// Remove a key
dataset.update_config([("to_delete", None)]).await?;
// Clear all config
dataset.update_config([] as [UpdateMapEntry; 0]).replace().await?;
// Replace full config
dataset.update_config([("k1", "v1"), ("k2", "v2")]).replace().await?;Sourcepub fn update_schema_metadata(
&mut self,
values: impl IntoIterator<Item = impl Into<UpdateMapEntry>>,
) -> UpdateMetadataBuilder<'_>
pub fn update_schema_metadata( &mut self, values: impl IntoIterator<Item = impl Into<UpdateMapEntry>>, ) -> UpdateMetadataBuilder<'_>
Update schema metadata.
Pass None for a value to remove that key.
Use .replace() to replace the entire schema metadata map instead of merging.
Returns the updated schema metadata map after the operation.
// Update single key
dataset.update_schema_metadata([("key", "value")]).await?;
// Remove a key
dataset.update_schema_metadata([("to_delete", None)]).await?;
// Clear all schema metadata
dataset.update_schema_metadata([] as [UpdateMapEntry; 0]).replace().await?;
// Replace full schema metadata
dataset.update_schema_metadata([("k1", "v1"), ("k2", "v2")]).replace().await?;Sourcepub async fn replace_schema_metadata(
&mut self,
new_values: impl IntoIterator<Item = (String, String)>,
) -> Result<(), Error>
👎Deprecated: Use the new update_schema_metadata(values).replace() instead
pub async fn replace_schema_metadata( &mut self, new_values: impl IntoIterator<Item = (String, String)>, ) -> Result<(), Error>
Update schema metadata
Sourcepub fn update_field_metadata(&mut self) -> UpdateFieldMetadataBuilder<'_>
pub fn update_field_metadata(&mut self) -> UpdateFieldMetadataBuilder<'_>
Update field metadata
// Update metadata by field path
dataset.update_field_metadata()
.update("path.to_field", [("key", "value")])?
.await?;
// Update metadata by field id
dataset.update_field_metadata()
.update(12, [("key", "value")])?
.await?;
// Clear field metadata
dataset.update_field_metadata()
.replace("path.to_field", [] as [UpdateMapEntry; 0])?
.replace(12, [] as [UpdateMapEntry; 0])?
.await?;
// Replace field metadata
dataset.update_field_metadata()
.replace("field_name", [("k1", "v1"), ("k2", "v2")])?
.await?;Sourcepub async fn replace_field_metadata(
&mut self,
new_values: impl IntoIterator<Item = (u32, HashMap<String, String>)>,
) -> Result<(), Error>
pub async fn replace_field_metadata( &mut self, new_values: impl IntoIterator<Item = (u32, HashMap<String, String>)>, ) -> Result<(), Error>
Update field metadata
Trait Implementations§
Source§impl Deref for DatasetWriteGuard<'_>
impl Deref for DatasetWriteGuard<'_>
Auto Trait Implementations§
impl<'a> Freeze for DatasetWriteGuard<'a>
impl<'a> !RefUnwindSafe for DatasetWriteGuard<'a>
impl<'a> Send for DatasetWriteGuard<'a>
impl<'a> Sync for DatasetWriteGuard<'a>
impl<'a> Unpin for DatasetWriteGuard<'a>
impl<'a> !UnwindSafe for DatasetWriteGuard<'a>
Blanket Implementations§
§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
Source§impl<T> Downcast for Twhere
T: Any,
impl<T> Downcast for Twhere
T: Any,
Source§fn into_any(self: Box<T>) -> Box<dyn Any>
fn into_any(self: Box<T>) -> Box<dyn Any>
Box<dyn Trait> (where Trait: Downcast) to Box<dyn Any>, which can then be
downcast into Box<dyn ConcreteType> where ConcreteType implements Trait.Source§fn into_any_rc(self: Rc<T>) -> Rc<dyn Any>
fn into_any_rc(self: Rc<T>) -> Rc<dyn Any>
Rc<Trait> (where Trait: Downcast) to Rc<Any>, which can then be further
downcast into Rc<ConcreteType> where ConcreteType implements Trait.Source§fn as_any(&self) -> &(dyn Any + 'static)
fn as_any(&self) -> &(dyn Any + 'static)
&Trait (where Trait: Downcast) to &Any. This is needed since Rust cannot
generate &Any’s vtable from &Trait’s.Source§fn as_any_mut(&mut self) -> &mut (dyn Any + 'static)
fn as_any_mut(&mut self) -> &mut (dyn Any + 'static)
&mut Trait (where Trait: Downcast) to &Any. This is needed since Rust cannot
generate &mut Any’s vtable from &mut Trait’s.Source§impl<T> DowncastSend for T
impl<T> DowncastSend for T
Source§impl<T> DowncastSync for T
impl<T> DowncastSync for T
Source§impl<T> FmtForward for T
impl<T> FmtForward for T
Source§fn fmt_binary(self) -> FmtBinary<Self>where
Self: Binary,
fn fmt_binary(self) -> FmtBinary<Self>where
Self: Binary,
self to use its Binary implementation when Debug-formatted.Source§fn fmt_display(self) -> FmtDisplay<Self>where
Self: Display,
fn fmt_display(self) -> FmtDisplay<Self>where
Self: Display,
self to use its Display implementation when
Debug-formatted.Source§fn fmt_lower_exp(self) -> FmtLowerExp<Self>where
Self: LowerExp,
fn fmt_lower_exp(self) -> FmtLowerExp<Self>where
Self: LowerExp,
self to use its LowerExp implementation when
Debug-formatted.Source§fn fmt_lower_hex(self) -> FmtLowerHex<Self>where
Self: LowerHex,
fn fmt_lower_hex(self) -> FmtLowerHex<Self>where
Self: LowerHex,
self to use its LowerHex implementation when
Debug-formatted.Source§fn fmt_octal(self) -> FmtOctal<Self>where
Self: Octal,
fn fmt_octal(self) -> FmtOctal<Self>where
Self: Octal,
self to use its Octal implementation when Debug-formatted.Source§fn fmt_pointer(self) -> FmtPointer<Self>where
Self: Pointer,
fn fmt_pointer(self) -> FmtPointer<Self>where
Self: Pointer,
self to use its Pointer implementation when
Debug-formatted.Source§fn fmt_upper_exp(self) -> FmtUpperExp<Self>where
Self: UpperExp,
fn fmt_upper_exp(self) -> FmtUpperExp<Self>where
Self: UpperExp,
self to use its UpperExp implementation when
Debug-formatted.Source§fn fmt_upper_hex(self) -> FmtUpperHex<Self>where
Self: UpperHex,
fn fmt_upper_hex(self) -> FmtUpperHex<Self>where
Self: UpperHex,
self to use its UpperHex implementation when
Debug-formatted.Source§impl<T> Instrument for T
impl<T> Instrument for T
Source§fn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
Source§fn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
Source§impl<T> IntoEither for T
impl<T> IntoEither for T
Source§fn into_either(self, into_left: bool) -> Either<Self, Self>
fn into_either(self, into_left: bool) -> Either<Self, Self>
self into a Left variant of Either<Self, Self>
if into_left is true.
Converts self into a Right variant of Either<Self, Self>
otherwise. Read moreSource§fn into_either_with<F>(self, into_left: F) -> Either<Self, Self>
fn into_either_with<F>(self, into_left: F) -> Either<Self, Self>
self into a Left variant of Either<Self, Self>
if into_left(&self) returns true.
Converts self into a Right variant of Either<Self, Self>
otherwise. Read moreSource§impl<T> Pipe for Twhere
T: ?Sized,
impl<T> Pipe for Twhere
T: ?Sized,
Source§fn pipe<R>(self, func: impl FnOnce(Self) -> R) -> Rwhere
Self: Sized,
fn pipe<R>(self, func: impl FnOnce(Self) -> R) -> Rwhere
Self: Sized,
Source§fn pipe_ref<'a, R>(&'a self, func: impl FnOnce(&'a Self) -> R) -> Rwhere
R: 'a,
fn pipe_ref<'a, R>(&'a self, func: impl FnOnce(&'a Self) -> R) -> Rwhere
R: 'a,
self and passes that borrow into the pipe function. Read moreSource§fn pipe_ref_mut<'a, R>(&'a mut self, func: impl FnOnce(&'a mut Self) -> R) -> Rwhere
R: 'a,
fn pipe_ref_mut<'a, R>(&'a mut self, func: impl FnOnce(&'a mut Self) -> R) -> Rwhere
R: 'a,
self and passes that borrow into the pipe function. Read moreSource§fn pipe_borrow<'a, B, R>(&'a self, func: impl FnOnce(&'a B) -> R) -> R
fn pipe_borrow<'a, B, R>(&'a self, func: impl FnOnce(&'a B) -> R) -> R
Source§fn pipe_borrow_mut<'a, B, R>(
&'a mut self,
func: impl FnOnce(&'a mut B) -> R,
) -> R
fn pipe_borrow_mut<'a, B, R>( &'a mut self, func: impl FnOnce(&'a mut B) -> R, ) -> R
Source§fn pipe_as_ref<'a, U, R>(&'a self, func: impl FnOnce(&'a U) -> R) -> R
fn pipe_as_ref<'a, U, R>(&'a self, func: impl FnOnce(&'a U) -> R) -> R
self, then passes self.as_ref() into the pipe function.Source§fn pipe_as_mut<'a, U, R>(&'a mut self, func: impl FnOnce(&'a mut U) -> R) -> R
fn pipe_as_mut<'a, U, R>(&'a mut self, func: impl FnOnce(&'a mut U) -> R) -> R
self, then passes self.as_mut() into the pipe
function.Source§fn pipe_deref<'a, T, R>(&'a self, func: impl FnOnce(&'a T) -> R) -> R
fn pipe_deref<'a, T, R>(&'a self, func: impl FnOnce(&'a T) -> R) -> R
self, then passes self.deref() into the pipe function.Source§impl<T> Pointable for T
impl<T> Pointable for T
Source§impl<T> PolicyExt for Twhere
T: ?Sized,
impl<T> PolicyExt for Twhere
T: ?Sized,
Source§impl<T> Tap for T
impl<T> Tap for T
Source§fn tap_borrow<B>(self, func: impl FnOnce(&B)) -> Self
fn tap_borrow<B>(self, func: impl FnOnce(&B)) -> Self
Borrow<B> of a value. Read moreSource§fn tap_borrow_mut<B>(self, func: impl FnOnce(&mut B)) -> Self
fn tap_borrow_mut<B>(self, func: impl FnOnce(&mut B)) -> Self
BorrowMut<B> of a value. Read moreSource§fn tap_ref<R>(self, func: impl FnOnce(&R)) -> Self
fn tap_ref<R>(self, func: impl FnOnce(&R)) -> Self
AsRef<R> view of a value. Read moreSource§fn tap_ref_mut<R>(self, func: impl FnOnce(&mut R)) -> Self
fn tap_ref_mut<R>(self, func: impl FnOnce(&mut R)) -> Self
AsMut<R> view of a value. Read moreSource§fn tap_deref<T>(self, func: impl FnOnce(&T)) -> Self
fn tap_deref<T>(self, func: impl FnOnce(&T)) -> Self
Deref::Target of a value. Read moreSource§fn tap_deref_mut<T>(self, func: impl FnOnce(&mut T)) -> Self
fn tap_deref_mut<T>(self, func: impl FnOnce(&mut T)) -> Self
Deref::Target of a value. Read moreSource§fn tap_dbg(self, func: impl FnOnce(&Self)) -> Self
fn tap_dbg(self, func: impl FnOnce(&Self)) -> Self
.tap() only in debug builds, and is erased in release builds.Source§fn tap_mut_dbg(self, func: impl FnOnce(&mut Self)) -> Self
fn tap_mut_dbg(self, func: impl FnOnce(&mut Self)) -> Self
.tap_mut() only in debug builds, and is erased in release
builds.Source§fn tap_borrow_dbg<B>(self, func: impl FnOnce(&B)) -> Self
fn tap_borrow_dbg<B>(self, func: impl FnOnce(&B)) -> Self
.tap_borrow() only in debug builds, and is erased in release
builds.Source§fn tap_borrow_mut_dbg<B>(self, func: impl FnOnce(&mut B)) -> Self
fn tap_borrow_mut_dbg<B>(self, func: impl FnOnce(&mut B)) -> Self
.tap_borrow_mut() only in debug builds, and is erased in release
builds.Source§fn tap_ref_dbg<R>(self, func: impl FnOnce(&R)) -> Self
fn tap_ref_dbg<R>(self, func: impl FnOnce(&R)) -> Self
.tap_ref() only in debug builds, and is erased in release
builds.Source§fn tap_ref_mut_dbg<R>(self, func: impl FnOnce(&mut R)) -> Self
fn tap_ref_mut_dbg<R>(self, func: impl FnOnce(&mut R)) -> Self
.tap_ref_mut() only in debug builds, and is erased in release
builds.Source§fn tap_deref_dbg<T>(self, func: impl FnOnce(&T)) -> Self
fn tap_deref_dbg<T>(self, func: impl FnOnce(&T)) -> Self
.tap_deref() only in debug builds, and is erased in release
builds.