[][src]Struct csv_async::AsyncDeserializer

pub struct AsyncDeserializer<R>(_);

A already configured CSV serde deserializer.

A CSV deserializer takes as input CSV data and transforms that into standard Rust values. The reader reads CSV data is as a sequence of records, where a record is either a sequence of string fields or structure with derived serde::Deserialize interface.

Configuration

A CSV deserializer has convenient constructor method from_reader. and from_reader. However, if you want to configure the CSV deserializer to use a different delimiter or quote character (among many other things), then you should use a AsyncReaderBuilder to construct a AsyncDeserializer. For example, to change the field delimiter:

use std::error::Error;
use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncReaderBuilder;
 
#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    pop: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = indoc::indoc! {"
        city;country;pop
        Boston;United States;4628910
    "};
    let mut rdr = AsyncReaderBuilder::new()
        .delimiter(b';')
        .create_deserializer(data.as_bytes());

    let mut records = rdr.deserialize::<Row>();
    assert_eq!(records.next().await.unwrap()?,
               Row {city: "Boston".to_string(), 
                    country: "United States".to_string(), 
                    pop: 4628910 });
    Ok(())
}

Error handling

In general, CSV parsing does not ever return an error. That is, there is no such thing as malformed CSV data. Instead, this reader will prioritize finding a parse over rejecting CSV data that it does not understand. This choice was inspired by other popular CSV parsers, but also because it is pragmatic. CSV data varies wildly, so even if the CSV data is malformed, it might still be possible to work with the data. In the land of CSV, there is no "right" or "wrong," only "right" and "less right."

With that said, a number of errors can occur while reading CSV data:

  • By default, all records in CSV data must have the same number of fields. If a record is found with a different number of fields than a prior record, then an error is returned. This behavior can be disabled by enabling flexible parsing via the flexible method on AsyncReaderBuilder.
  • When reading CSV data from a resource (like a file), it is possible for reading from the underlying resource to fail. This will return an error. For subsequent calls to the reader after encountering a such error (unless seek is used), it will behave as if end of file had been reached, in order to avoid running into infinite loops when still attempting to read the next record when one has errored.
  • When reading CSV data into String or &str fields (e.g., via a StringRecord), UTF-8 is strictly enforced. If CSV data is invalid UTF-8, then an error is returned. If you want to read invalid UTF-8, then you should use the byte oriented APIs such as ByteRecord. If you need explicit support for another encoding entirely, then you'll need to use another crate to transcode your CSV data to UTF-8 before parsing it.
  • When using Serde to deserialize CSV data into Rust types, it is possible for a number of additional errors to occur. For example, deserializing a field xyz into an i32 field will result in an error.

For more details on the precise semantics of errors, see the Error type.

Implementations

impl<'r, R> AsyncDeserializer<R> where
    R: AsyncRead + Unpin + 'r, 
[src]

pub fn from_reader(rdr: R) -> AsyncDeserializer<R>[src]

Create a new CSV parser with a default configuration for the given reader.

To customize CSV parsing, use a ReaderBuilder.

Example

use std::error::Error;
use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncDeserializer;
 
#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    pop: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,pop
Boston,United States,4628910
Concord,United States,42695
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());
    let mut records = rdr.into_deserialize::<Row>();
    while let Some(record) = records.next().await {
        println!("{:?}", record?);
    }
    Ok(())
}

pub fn deserialize<D: 'r>(&'r mut self) -> DeserializeRecordsStream<'r, R, D> where
    D: DeserializeOwned
[src]

Returns a borrowed stream over deserialized records.

Each item yielded by this stream is a Result<D, Error>. Therefore, in order to access the record, callers must handle the possibility of error (typically with ?).

If has_headers was enabled via a ReaderBuilder (which is the default), then this does not include the first record. Additionally, if has_headers is enabled, then deserializing into a struct will automatically align the values in each row to the fields of a struct based on the header row.

Frequently turbo-fish notation is needed while calling this function: rdr.deserialize::<RecordTyme>();

Example

This shows how to deserialize CSV data into normal Rust structures. The fields of the header row are used to match up the values in each row to the fields of the struct.

use std::error::Error;

use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncDeserializer;

#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    #[serde(rename = "popcount")]
    population: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,popcount
Boston,United States,4628910
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());
    let mut iter = rdr.deserialize();

    if let Some(result) = iter.next().await {
        let record: Row = result?;
        assert_eq!(record, Row {
            city: "Boston".to_string(),
            country: "United States".to_string(),
            population: 4628910,
        });
        Ok(())
    } else {
        Err(From::from("expected at least one record but got none"))
    }
}

Rules

For the most part, any Rust type that maps straight-forwardly to a CSV record is supported. This includes maps, structs, tuples and tuple structs. Other Rust types, such as Vecs, arrays, and enums have a more complicated story. In general, when working with CSV data, one should avoid nested sequences as much as possible.

Maps, structs, tuples and tuple structs map to CSV records in a simple way. Tuples and tuple structs decode their fields in the order that they are defined. Structs will do the same only if has_headers has been disabled using ReaderBuilder, otherwise, structs and maps are deserialized based on the fields defined in the header row. (If there is no header row, then deserializing into a map will result in an error.)

Nested sequences are supported in a limited capacity. Namely, they are flattened. As a result, it's often useful to use a Vec to capture a "tail" of fields in a record:

use std::error::Error;

use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncReaderBuilder;

#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    label: String,
    values: Vec<i32>,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "foo,1,2,3";
    let mut rdr = AsyncReaderBuilder::new()
        .has_headers(false)
        .create_deserializer(data.as_bytes());
    let mut iter = rdr.deserialize();

    if let Some(result) = iter.next().await {
        let record: Row = result?;
        assert_eq!(record, Row {
            label: "foo".to_string(),
            values: vec![1, 2, 3],
        });
        Ok(())
    } else {
        Err(From::from("expected at least one record but got none"))
    }
}

In the above example, adding another field to the Row struct after the values field will result in a deserialization error. This is because the deserializer doesn't know when to stop reading fields into the values vector, so it will consume the rest of the fields in the record leaving none left over for the additional field.

Finally, simple enums in Rust can be deserialized as well. Namely, enums must either be variants with no arguments or variants with a single argument. Variants with no arguments are deserialized based on which variant name the field matches. Variants with a single argument are deserialized based on which variant can store the data. The latter is only supported when using "untagged" enum deserialization. The following example shows both forms in action:

use std::error::Error;

use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncDeserializer;

#[derive(Debug, Deserialize, PartialEq)]
struct Row {
    label: Label,
    value: Number,
}

#[derive(Debug, Deserialize, PartialEq)]
#[serde(rename_all = "lowercase")]
enum Label {
    Celsius,
    Fahrenheit,
}

#[derive(Debug, Deserialize, PartialEq)]
#[serde(untagged)]
enum Number {
    Integer(i64),
    Float(f64),
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
label,value
celsius,22.2222
fahrenheit,72
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());
    let mut iter = rdr.deserialize();

    // Read the first record.
    if let Some(result) = iter.next().await {
        let record: Row = result?;
        assert_eq!(record, Row {
            label: Label::Celsius,
            value: Number::Float(22.2222),
        });
    } else {
        return Err(From::from(
            "expected at least two records but got none"));
    }

    // Read the second record.
    if let Some(result) = iter.next().await {
        let record: Row = result?;
        assert_eq!(record, Row {
            label: Label::Fahrenheit,
            value: Number::Integer(72),
        });
        Ok(())
    } else {
        Err(From::from(
            "expected at least two records but got only one"))
    }
}

pub fn deserialize_with_pos<D: 'r>(
    &'r mut self
) -> DeserializeRecordsStreamPos<'r, R, D> where
    D: DeserializeOwned
[src]

Returns a borrowed stream over pairs of deserialized record and position in reader stream before record read.

Each item yielded by this stream is a (Result<D, Error>, Position). Therefore, in order to access the record, callers must handle the possibility of error (typically with ?).

Frequently turbo-fish notation is needed while calling this function: rdr.deserialize_with_pos::<RecordTyme>();

Example

This shows how to deserialize CSV data into normal Rust structures. The fields of the header row are used to match up the values in each row to the fields of the struct.

use std::error::Error;

use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncDeserializer;

#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    population: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,population
Boston,United States,4628910
Concord,United States,42695
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());
    let mut iter = rdr.deserialize_with_pos();

    if let Some((result, pos)) = iter.next().await {
        let record: Row = result?;
        assert_eq!(record, Row {
            city: "Boston".to_string(),
            country: "United States".to_string(),
            population: 4628910,
        });
        assert_eq!(pos.byte(), 24);
        assert_eq!(pos.line(), 2);
        assert_eq!(pos.record(), 1);
    } else {
        return Err(From::from("expected at least one record but got none"));
    }
    if let Some((result, pos)) = iter.next().await {
        let record: Row = result?;
        assert_eq!(record, Row {
            city: "Concord".to_string(),
            country: "United States".to_string(),
            population: 42695,
        });
        assert_eq!(pos.byte(), 53);
        assert_eq!(pos.line(), 3);
        assert_eq!(pos.record(), 2);
    } else {
        return Err(From::from("expected at least two records but got one only"));
    }
    assert!(iter.next().await.is_none());
    Ok(())
}

pub fn into_deserialize<D: 'r>(self) -> DeserializeRecordsIntoStream<'r, R, D> where
    D: DeserializeOwned
[src]

Returns a owned stream over deserialized records.

Each item yielded by this stream is a Result<D, Error>. Therefore, in order to access the record, callers must handle the possibility of error (typically with ?).

If has_headers was enabled via a ReaderBuilder (which is the default), then this does not include the first record. Additionally, if has_headers is enabled, then deserializing into a struct will automatically align the values in each row to the fields of a struct based on the header row.

Frequently turbo-fish notation is needed while calling this function: rdr.into_deserialize::<RecordTyme>();

Example

This shows how to deserialize CSV data into normal Rust structs. The fields of the header row are used to match up the values in each row to the fields of the struct.

use std::error::Error;

use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncDeserializer;

#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    #[serde(rename = "popcount")]
    population: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,popcount
Boston,United States,4628910
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());
    let mut iter = rdr.into_deserialize();

    if let Some(result) = iter.next().await {
        let record: Row = result?;
        assert_eq!(record, Row {
            city: "Boston".to_string(),
            country: "United States".to_string(),
            population: 4628910,
        });
        Ok(())
    } else {
        Err(From::from("expected at least one record but got none"))
    }
}

pub fn into_deserialize_with_pos<D: 'r>(
    self
) -> DeserializeRecordsIntoStreamPos<'r, R, D> where
    D: DeserializeOwned
[src]

Returns a owned stream over pairs of deserialized record and position in reader stream before record read.

pub async fn headers<'_, '_>(&'_ mut self) -> Result<&'_ StringRecord>[src]

Returns a reference to the first row read by this parser.

If no row has been read yet, then this will force parsing of the first row.

If there was a problem parsing the row or if it wasn't valid UTF-8, then this returns an error.

If the underlying reader emits EOF before any data, then this returns an empty record.

Note that this method may be used regardless of whether has_headers was enabled (but it is enabled by default).

Example

This example shows how to get the header row of CSV data. Notice that the header row does not appear as a record in the iterator!

use std::error::Error;
use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncDeserializer;
 
#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    pop: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,pop
Boston,United States,4628910
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());

    // We can read the headers before iterating.
    {
    // `headers` borrows from the reader, so we put this in its
    // own scope. That way, the borrow ends before we try iterating
    // below. Alternatively, we could clone the headers.
    let headers = rdr.headers().await?;
    assert_eq!(headers, vec!["city", "country", "pop"]);
    }

    {
    let mut records = rdr.deserialize::<Row>();
    assert_eq!(records.next().await.unwrap()?, 
               Row {city: "Boston".to_string(), 
                    country: "United States".to_string(), 
                    pop: 4628910 });
    assert!(records.next().await.is_none());
    }

    // We can also read the headers after iterating.
    let headers = rdr.headers().await?;
    assert_eq!(headers, vec!["city", "country", "pop"]);
    Ok(())
}

pub async fn byte_headers<'_, '_>(&'_ mut self) -> Result<&'_ ByteRecord>[src]

Returns a reference to the first row read by this parser as raw bytes.

If no row has been read yet, then this will force parsing of the first row.

If there was a problem parsing the row then this returns an error.

If the underlying reader emits EOF before any data, then this returns an empty record.

Note that this method may be used regardless of whether has_headers was enabled (but it is enabled by default).

Example

This example shows how to get the header row of CSV data. Notice that the header row does not appear as a record in the iterator!

use std::error::Error;
use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::AsyncDeserializer;

#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    #[serde(rename = "pop")]
    population: u64,
}
 
async fn example() -> Result<(), Box<dyn Error>> {
    let data = indoc::indoc! {"
        city,country,pop
        Boston,United States,4628910
    "};
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());

    // We can read the headers before iterating.
    {
    // `headers` borrows from the reader, so we put this in its
    // own scope. That way, the borrow ends before we try iterating
    // below. Alternatively, we could clone the headers.
    let headers = rdr.byte_headers().await?;
    assert_eq!(headers, vec!["city", "country", "pop"]);
    }

    {
    let mut records = rdr.deserialize::<Row>();
    assert_eq!(records.next().await.unwrap()?, 
               Row {city: "Boston".to_string(), 
                    country: "United States".to_string(), 
                    population: 4628910 });
    assert!(records.next().await.is_none());
    }

    // We can also read the headers after iterating.
    let headers = rdr.byte_headers().await?;
    assert_eq!(headers, vec!["city", "country", "pop"]);
    Ok(())
}

pub fn set_headers(&mut self, headers: StringRecord)[src]

Set the headers of this CSV parser manually.

This overrides any other setting (including set_byte_headers). Any automatic detection of headers is disabled. This may be called at any time.

Example

use std::error::Error;
use csv_async::{AsyncDeserializer, StringRecord};

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,pop
Boston,United States,4628910
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());

    assert_eq!(rdr.headers().await?, vec!["city", "country", "pop"]);
    rdr.set_headers(StringRecord::from(vec!["a", "b", "c"]));
    assert_eq!(rdr.headers().await?, vec!["a", "b", "c"]);

    Ok(())
}

pub fn set_byte_headers(&mut self, headers: ByteRecord)[src]

Set the headers of this CSV parser manually as raw bytes.

This overrides any other setting (including set_headers). Any automatic detection of headers is disabled. This may be called at any time.

Example

use std::error::Error;
use csv_async::{AsyncDeserializer, ByteRecord};

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,pop
Boston,United States,4628910
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());

    assert_eq!(rdr.byte_headers().await?, vec!["city", "country", "pop"]);
    rdr.set_byte_headers(ByteRecord::from(vec!["a", "b", "c"]));
    assert_eq!(rdr.byte_headers().await?, vec!["a", "b", "c"]);

    Ok(())
}

pub async fn read_record<'_, '_>(
    &'_ mut self,
    record: &'_ mut StringRecord
) -> Result<bool>
[src]

Read a single row into the given record. Returns false when no more records could be read.

If has_headers was enabled via a ReaderBuilder (which is the default), then this will never read the first record.

This method is useful when you want to read records as fast as as possible. It's less ergonomic than an iterator, but it permits the caller to reuse the StringRecord allocation, which usually results in higher throughput.

Records read via this method are guaranteed to have a position set on them, even if the reader is at EOF or if an error is returned.

Example

use std::error::Error;
use csv_async::{AsyncDeserializer, StringRecord};

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,pop
Boston,United States,4628910
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());
    let mut record = StringRecord::new();

    if rdr.read_record(&mut record).await? {
        assert_eq!(record, vec!["Boston", "United States", "4628910"]);
        Ok(())
    } else {
        Err(From::from("expected at least one record but got none"))
    }
}

pub async fn read_byte_record<'_, '_>(
    &'_ mut self,
    record: &'_ mut ByteRecord
) -> Result<bool>
[src]

Read a single row into the given byte record. Returns false when no more records could be read.

If has_headers was enabled via a ReaderBuilder (which is the default), then this will never read the first record.

This method is useful when you want to read records as fast as as possible. It's less ergonomic than an iterator, but it permits the caller to reuse the ByteRecord allocation, which usually results in higher throughput.

Records read via this method are guaranteed to have a position set on them, even if the reader is at EOF or if an error is returned.

Example

use std::error::Error;
use csv_async::{ByteRecord, AsyncDeserializer};

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,pop
Boston,United States,4628910
";
    let mut rdr = AsyncDeserializer::from_reader(data.as_bytes());
    let mut record = ByteRecord::new();

    if rdr.read_byte_record(&mut record).await? {
        assert_eq!(record, vec!["Boston", "United States", "4628910"]);
        Ok(())
    } else {
        Err(From::from("expected at least one record but got none"))
    }
}

pub fn position(&self) -> &Position[src]

Return the current position of this CSV deserializer.

Because of borrowing rules this function can only be used when there is no alive deserializer (which borrows mutable reader). To know position during deserialization, deserialize_with_pos should be used as shown in below example.

The byte offset in the position returned can be used to seek this deserializer. In particular, seeking to a position returned here on the same data will result in parsing the same subsequent record.

Example: reading the position

use std::error::Error;
use futures::io;
use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::{AsyncDeserializer, Position};

#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    popcount: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,popcount
Boston,United States,4628910
Concord,United States,42695
";
    let mut rdr = AsyncDeserializer::from_reader(io::Cursor::new(data));
    let mut iter = rdr.deserialize_with_pos::<Row>();
    let mut pos_at_boston = Position::new();
    while let Some((rec, pos)) = iter.next().await {
        if rec?.city == "Boston" {
            pos_at_boston = pos;
        }
    }
    drop(iter); // releases rdr borrow by iter
    let pos_at_end = rdr.position();
 
    assert_eq!(pos_at_boston.byte(),  22);
    assert_eq!(pos_at_boston.line(),   2);
    assert_eq!(pos_at_boston.record(), 1);
    assert_eq!(pos_at_end.byte(),  79);
    assert_eq!(pos_at_end.line(),   4);
    assert_eq!(pos_at_end.record(), 3);
    Ok(())
}

pub fn is_done(&self) -> bool[src]

Returns true if and only if this reader has been exhausted.

When this returns true, no more records can be read from this reader (unless it has been seek to another position).

Example

use std::error::Error;
use futures::io;
use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::{AsyncDeserializer, Position};

#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    popcount: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,popcount
Boston,United States,4628910
Concord,United States,42695
";
    let mut rdr = AsyncDeserializer::from_reader(io::Cursor::new(data));
    assert!(!rdr.is_done());
    {
        let mut records = rdr.deserialize::<Row>();
        while let Some(record) = records.next().await {
            let _ = record?;
        }
    }
    assert!(rdr.is_done());
    Ok(())
}

pub fn has_headers(&self) -> bool[src]

Returns true if and only if this reader has been configured to interpret the first record as a header record.

pub fn get_ref(&self) -> &R[src]

Returns a reference to the underlying reader.

pub fn get_mut(&mut self) -> &mut R[src]

Returns a mutable reference to the underlying reader.

pub fn into_inner(self) -> R[src]

Unwraps this CSV reader, returning the underlying reader.

Note that any leftover data inside this reader's internal buffer is lost.

impl<R: AsyncRead + AsyncSeek + Unpin> AsyncDeserializer<R>[src]

pub async fn seek<'_>(&'_ mut self, pos: Position) -> Result<()>[src]

Seeks the underlying reader to the position given.

This comes with a few caveats:

  • Any internal buffer associated with this reader is cleared.
  • If the given position does not correspond to a position immediately before the start of a record, then the behavior of this reader is unspecified.
  • Any special logic that skips the first record in the CSV reader when reading or iterating over records is disabled.

If the given position has a byte offset equivalent to the current position, then no seeking is performed.

If the header row has not already been read, then this will attempt to read the header row before seeking. Therefore, it is possible that this returns an error associated with reading CSV data.

Note that seeking is performed based only on the byte offset in the given position. Namely, the record or line numbers in the position may be incorrect, but this will cause any future position generated by this CSV reader to be similarly incorrect.

Example: seek to parse a record twice

use std::error::Error;
use futures::io;
use futures::stream::StreamExt;
use serde::Deserialize;
use csv_async::{AsyncDeserializer, Position};

#[derive(Debug, Deserialize, Eq, PartialEq)]
struct Row {
    city: String,
    country: String,
    popcount: u64,
}

async fn example() -> Result<(), Box<dyn Error>> {
    let data = "\
city,country,popcount
Boston,United States,4628910
Concord,United States,42695
";
    let mut rdr = AsyncDeserializer::from_reader(io::Cursor::new(data));
    let mut records = rdr.deserialize_with_pos::<Row>();
    let mut pos_at_boston = Position::new();
    while let Some((rec, pos)) = records.next().await {
        if rec?.city == "Boston" {
            pos_at_boston = pos;
            // no break here - we are reading data to end
        }
    }
    drop(records); // releases rdr borrowed by records

    // Now seek the reader back to `pos_at_boston`. This will let us read the
    // Boston's record again.
    rdr.seek(pos_at_boston).await?;
    let mut records = rdr.into_deserialize::<Row>();
    if let Some(rec) = records.next().await {
        assert_eq!(rec?.city, "Boston");
        Ok(())
    } else {
        Err(From::from("After seek we should be before last record, but was at end of stream"))
    }
}

pub async fn seek_raw<'_>(
    &'_ mut self,
    seek_from: SeekFrom,
    pos: Position
) -> Result<()>
[src]

This is like seek, but provides direct control over how the seeking operation is performed via io::SeekFrom.

The pos position given should correspond the position indicated by seek_from, but there is no requirement. If the pos position given is incorrect, then the position information returned by this reader will be similarly incorrect.

If the header row has not already been read, then this will attempt to read the header row before seeking. Therefore, it is possible that this returns an error associated with reading CSV data.

Unlike seek, this will always cause an actual seek to be performed.

Trait Implementations

impl<R: Debug> Debug for AsyncDeserializer<R>[src]

Auto Trait Implementations

impl<R> RefUnwindSafe for AsyncDeserializer<R> where
    R: RefUnwindSafe

impl<R> Send for AsyncDeserializer<R> where
    R: Send

impl<R> Sync for AsyncDeserializer<R> where
    R: Sync

impl<R> Unpin for AsyncDeserializer<R> where
    R: Unpin

impl<R> UnwindSafe for AsyncDeserializer<R> where
    R: UnwindSafe

Blanket Implementations

impl<T> Any for T where
    T: 'static + ?Sized
[src]

impl<T> Borrow<T> for T where
    T: ?Sized
[src]

impl<T> BorrowMut<T> for T where
    T: ?Sized
[src]

impl<T> From<T> for T[src]

impl<T, U> Into<U> for T where
    U: From<T>, 
[src]

impl<T, U> TryFrom<U> for T where
    U: Into<T>, 
[src]

type Error = Infallible

The type returned in the event of a conversion error.

impl<T, U> TryInto<U> for T where
    U: TryFrom<T>, 
[src]

type Error = <U as TryFrom<T>>::Error

The type returned in the event of a conversion error.