pub struct BigQueryReadClient<T> { /* private fields */ }
Expand description
BigQuery Read API.
The Read API can be used to read data from BigQuery.
Implementations§
Source§impl<T> BigQueryReadClient<T>where
T: GrpcService<BoxBody>,
T::Error: Into<StdError>,
T::ResponseBody: Body<Data = Bytes> + Send + 'static,
<T::ResponseBody as Body>::Error: Into<StdError> + Send,
impl<T> BigQueryReadClient<T>where
T: GrpcService<BoxBody>,
T::Error: Into<StdError>,
T::ResponseBody: Body<Data = Bytes> + Send + 'static,
<T::ResponseBody as Body>::Error: Into<StdError> + Send,
pub fn new(inner: T) -> Self
pub fn with_origin(inner: T, origin: Uri) -> Self
pub fn with_interceptor<F>(
inner: T,
interceptor: F,
) -> BigQueryReadClient<InterceptedService<T, F>>where
F: Interceptor,
T::ResponseBody: Default,
T: Service<Request<BoxBody>, Response = Response<<T as GrpcService<BoxBody>>::ResponseBody>>,
<T as Service<Request<BoxBody>>>::Error: Into<StdError> + Send + Sync,
Sourcepub fn send_compressed(self, encoding: CompressionEncoding) -> Self
pub fn send_compressed(self, encoding: CompressionEncoding) -> Self
Compress requests with the given encoding.
This requires the server to support it otherwise it might respond with an error.
Sourcepub fn accept_compressed(self, encoding: CompressionEncoding) -> Self
pub fn accept_compressed(self, encoding: CompressionEncoding) -> Self
Enable decompressing responses.
Sourcepub fn max_decoding_message_size(self, limit: usize) -> Self
pub fn max_decoding_message_size(self, limit: usize) -> Self
Limits the maximum size of a decoded message.
Default: 4MB
Sourcepub fn max_encoding_message_size(self, limit: usize) -> Self
pub fn max_encoding_message_size(self, limit: usize) -> Self
Limits the maximum size of an encoded message.
Default: usize::MAX
Sourcepub async fn create_read_session(
&mut self,
request: impl IntoRequest<CreateReadSessionRequest>,
) -> Result<Response<ReadSession>, Status>
pub async fn create_read_session( &mut self, request: impl IntoRequest<CreateReadSessionRequest>, ) -> Result<Response<ReadSession>, Status>
Creates a new read session. A read session divides the contents of a BigQuery table into one or more streams, which can then be used to read data from the table. The read session also specifies properties of the data to be read, such as a list of columns or a push-down filter describing the rows to be returned.
A particular row can be read by at most one stream. When the caller has reached the end of each stream in the session, then all the data in the table has been read.
Data is assigned to each stream such that roughly the same number of rows can be read from each stream. Because the server-side unit for assigning data is collections of rows, the API does not guarantee that each stream will return the same number or rows. Additionally, the limits are enforced based on the number of pre-filtered rows, so some filters can lead to lopsided assignments.
Read sessions automatically expire 6 hours after they are created and do not require manual clean-up by the caller.
Sourcepub async fn read_rows(
&mut self,
request: impl IntoRequest<ReadRowsRequest>,
) -> Result<Response<Streaming<ReadRowsResponse>>, Status>
pub async fn read_rows( &mut self, request: impl IntoRequest<ReadRowsRequest>, ) -> Result<Response<Streaming<ReadRowsResponse>>, Status>
Reads rows from the stream in the format prescribed by the ReadSession. Each response contains one or more table rows, up to a maximum of 100 MiB per response; read requests which attempt to read individual rows larger than 100 MiB will fail.
Each request also returns a set of stream statistics reflecting the current state of the stream.
Sourcepub async fn split_read_stream(
&mut self,
request: impl IntoRequest<SplitReadStreamRequest>,
) -> Result<Response<SplitReadStreamResponse>, Status>
pub async fn split_read_stream( &mut self, request: impl IntoRequest<SplitReadStreamRequest>, ) -> Result<Response<SplitReadStreamResponse>, Status>
Splits a given ReadStream
into two ReadStream
objects. These
ReadStream
objects are referred to as the primary and the residual
streams of the split. The original ReadStream
can still be read from in
the same manner as before. Both of the returned ReadStream
objects can
also be read from, and the rows returned by both child streams will be
the same as the rows read from the original stream.
Moreover, the two child streams will be allocated back-to-back in the
original ReadStream
. Concretely, it is guaranteed that for streams
original, primary, and residual, that original[0-j] = primary[0-j] and
original[j-n] = residual[0-m] once the streams have been read to
completion.
Trait Implementations§
Source§impl<T: Clone> Clone for BigQueryReadClient<T>
impl<T: Clone> Clone for BigQueryReadClient<T>
Source§fn clone(&self) -> BigQueryReadClient<T>
fn clone(&self) -> BigQueryReadClient<T>
1.0.0 · Source§fn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source
. Read moreAuto Trait Implementations§
impl<T> !Freeze for BigQueryReadClient<T>
impl<T> RefUnwindSafe for BigQueryReadClient<T>where
T: RefUnwindSafe,
impl<T> Send for BigQueryReadClient<T>where
T: Send,
impl<T> Sync for BigQueryReadClient<T>where
T: Sync,
impl<T> Unpin for BigQueryReadClient<T>where
T: Unpin,
impl<T> UnwindSafe for BigQueryReadClient<T>where
T: UnwindSafe,
Blanket Implementations§
Source§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
Source§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
Source§impl<T> CloneToUninit for Twhere
T: Clone,
impl<T> CloneToUninit for Twhere
T: Clone,
Source§impl<T> Instrument for T
impl<T> Instrument for T
Source§fn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
Source§fn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
Source§impl<T> IntoRequest<T> for T
impl<T> IntoRequest<T> for T
Source§fn into_request(self) -> Request<T>
fn into_request(self) -> Request<T>
T
in a tonic::Request