use std::collections::HashSet;
use anyhow::Context;
use librqbit_core::lengths::{ChunkInfo, Lengths, ValidPieceIndex};
use peer_binary_protocol::Piece;
use tracing::{debug, trace};
use crate::{
bitv::{BitV, BoxBitV},
file_info::FileInfo,
type_aliases::{FileInfos, FilePriorities, BF, BS},
};
pub struct ChunkTracker {
queue_pieces: BF,
chunk_status: BF,
have: BoxBitV,
selected: BF,
per_file_bytes: Vec<u64>,
lengths: Lengths,
hns: HaveNeededSelected,
}
#[derive(Default, Debug, PartialEq, Eq, Clone, Copy)]
pub struct HaveNeededSelected {
pub have_bytes: u64,
pub needed_bytes: u64,
pub selected_bytes: u64,
}
impl HaveNeededSelected {
pub const fn progress(&self) -> u64 {
self.selected_bytes - self.needed_bytes
}
pub const fn total(&self) -> u64 {
self.selected_bytes
}
pub const fn finished(&self) -> bool {
self.needed_bytes == 0
}
}
fn compute_chunk_have_status(lengths: &Lengths, have_pieces: &BS) -> anyhow::Result<BF> {
if have_pieces.len() < lengths.total_pieces() as usize {
anyhow::bail!(
"bug: have_pieces.len() < lengths.total_pieces(); {} < {}",
have_pieces.len(),
lengths.total_pieces()
);
}
let required_size = lengths.chunk_bitfield_bytes();
let vec = vec![0u8; required_size];
let mut chunk_bf = BF::from_boxed_slice(vec.into_boxed_slice());
for piece in lengths.iter_piece_infos() {
let chunks = lengths.chunks_per_piece(piece.piece_index) as usize;
let offset = (lengths.default_chunks_per_piece() * piece.piece_index.get()) as usize;
let range = offset..(offset + chunks);
if have_pieces[piece.piece_index.get() as usize] {
chunk_bf
.get_mut(range.clone())
.with_context(|| {
format!("bug in bitvec: error getting range {range:?} from chunk_bf")
})?
.fill(true);
}
}
Ok(chunk_bf)
}
fn compute_queued_pieces_unchecked(have_pieces: &BS, selected_pieces: &BS) -> BF {
use core::ops::BitAnd;
use core::ops::Not;
have_pieces
.to_bitvec()
.not()
.bitand(selected_pieces)
.into_boxed_bitslice()
}
fn compute_queued_pieces(have_pieces: &BS, selected_pieces: &BS) -> anyhow::Result<BF> {
if have_pieces.len() != selected_pieces.len() {
anyhow::bail!(
"have_pieces.len() != selected_pieces.len(), {} != {}",
have_pieces.len(),
selected_pieces.len()
);
}
Ok(compute_queued_pieces_unchecked(
have_pieces,
selected_pieces,
))
}
#[derive(Debug)]
pub enum ChunkMarkingResult {
PreviouslyCompleted,
NotCompleted,
Completed,
}
impl ChunkTracker {
pub fn new(
have_pieces: BoxBitV,
selected_pieces: BF,
lengths: Lengths,
file_infos: &FileInfos,
) -> anyhow::Result<Self> {
let needed_pieces = compute_queued_pieces(have_pieces.as_slice(), &selected_pieces)
.context("error computing needed pieces")?;
let mut ct = Self {
chunk_status: compute_chunk_have_status(&lengths, have_pieces.as_slice())
.context("error computing chunk status")?,
queue_pieces: needed_pieces,
selected: selected_pieces,
lengths,
have: have_pieces,
hns: HaveNeededSelected::default(),
per_file_bytes: vec![0; file_infos.len()],
};
ct.recalculate_per_file_bytes(file_infos);
ct.hns = ct.calc_hns();
Ok(ct)
}
fn recalculate_per_file_bytes(&mut self, file_infos: &FileInfos) {
for (slot, fi) in self.per_file_bytes.iter_mut().zip(file_infos.iter()) {
*slot = fi
.piece_range
.clone()
.filter(|p| self.have.as_slice()[*p as usize])
.map(|id| {
self.lengths
.size_of_piece_in_file(id, fi.offset_in_torrent, fi.len)
})
.sum();
}
}
pub fn get_lengths(&self) -> &Lengths {
&self.lengths
}
pub fn get_have_pieces(&self) -> &dyn BitV {
&*self.have
}
pub fn get_have_pieces_mut(&mut self) -> &mut dyn BitV {
&mut *self.have
}
pub fn reserve_needed_piece(&mut self, index: ValidPieceIndex) {
self.queue_pieces.set(index.get() as usize, false)
}
pub fn get_hns(&self) -> &HaveNeededSelected {
&self.hns
}
fn calc_hns(&self) -> HaveNeededSelected {
let mut hns = HaveNeededSelected::default();
for piece in self.lengths.iter_piece_infos() {
let id = piece.piece_index.get() as usize;
let len = piece.len as u64;
let is_have = self.have.as_slice()[id];
let is_selected = self.selected[id];
let is_needed = is_selected && !is_have;
hns.have_bytes += len * (is_have as u64);
hns.selected_bytes += len * (is_selected as u64);
hns.needed_bytes += len * (is_needed as u64);
}
hns
}
pub(crate) fn iter_queued_pieces<'a>(
&'a self,
file_priorities: &'a FilePriorities,
file_infos: &'a FileInfos,
) -> impl Iterator<Item = ValidPieceIndex> + 'a {
file_priorities
.iter()
.filter_map(|p| Some((*p, file_infos.get(*p)?)))
.filter(|(id, f)| self.per_file_bytes[*id] != f.len)
.flat_map(|(_id, f)| f.iter_piece_priorities())
.filter(|id| self.queue_pieces[*id])
.filter_map(|id| id.try_into().ok())
.filter_map(|id| self.lengths.validate_piece_index(id))
}
pub(crate) fn is_piece_have(&self, id: ValidPieceIndex) -> bool {
self.have.as_slice()[id.get() as usize]
}
pub fn mark_piece_broken_if_not_have(&mut self, index: ValidPieceIndex) {
if self
.have
.as_slice()
.get(index.get() as usize)
.map(|r| *r)
.unwrap_or_default()
{
return;
}
debug!("marking piece={} as broken", index);
self.queue_pieces.set(index.get() as usize, true);
if let Some(s) = self.chunk_status.get_mut(self.lengths.chunk_range(index)) {
s.fill(false);
}
}
pub fn mark_piece_downloaded(&mut self, idx: ValidPieceIndex) {
let id = idx.get() as usize;
if !self.have.as_slice()[id] {
self.have.as_slice_mut().set(id, true);
let len = self.lengths.piece_length(idx) as u64;
self.hns.have_bytes += len;
if self.selected[id] {
self.hns.needed_bytes -= len;
}
}
}
pub fn is_chunk_ready_to_upload(&self, chunk: &ChunkInfo) -> bool {
self.have
.as_slice()
.get(chunk.piece_index.get() as usize)
.map(|b| *b)
.unwrap_or(false)
}
pub fn get_remaining_bytes(&self) -> u64 {
self.hns.needed_bytes
}
pub fn mark_chunk_downloaded<ByteBuf>(
&mut self,
piece: &Piece<ByteBuf>,
) -> Option<ChunkMarkingResult>
where
ByteBuf: AsRef<[u8]>,
{
let chunk_info = self.lengths.chunk_info_from_received_data(
self.lengths.validate_piece_index(piece.index)?,
piece.begin,
piece.block.as_ref().len().try_into().unwrap(),
)?;
let chunk_range = self.lengths.chunk_range(chunk_info.piece_index);
let chunk_range = self.chunk_status.get_mut(chunk_range).unwrap();
if chunk_range.all() {
return Some(ChunkMarkingResult::PreviouslyCompleted);
}
chunk_range.set(chunk_info.chunk_index as usize, true);
trace!(
"piece={}, chunk_info={:?}, bits={:?}",
piece.index,
chunk_info,
chunk_range,
);
if chunk_range.all() {
return Some(ChunkMarkingResult::Completed);
}
Some(ChunkMarkingResult::NotCompleted)
}
pub fn update_only_files(
&mut self,
file_lengths_iterator: impl IntoIterator<Item = u64>,
new_only_files: &HashSet<usize>,
) -> anyhow::Result<HaveNeededSelected> {
let mut piece_it = self.lengths.iter_piece_infos();
let mut current_piece = piece_it
.next()
.context("bug: iter_piece_infos() returned empty iterator")?;
let mut current_piece_selected = false;
let mut current_piece_remaining = current_piece.len;
let mut have_bytes = 0u64;
let mut selected_bytes = 0u64;
let mut needed_bytes = 0u64;
for (idx, len) in file_lengths_iterator.into_iter().enumerate() {
let file_required = new_only_files.contains(&idx);
let mut remaining_file_len = len;
while remaining_file_len > 0 {
current_piece_selected |= len > 0 && file_required;
let shift = std::cmp::min(current_piece_remaining as u64, remaining_file_len);
if shift == 0 {
anyhow::bail!("bug: shift = 0, this shouldn't have happened")
}
remaining_file_len -= shift;
current_piece_remaining -= TryInto::<u32>::try_into(shift)?;
if current_piece_remaining == 0 {
let current_piece_have =
self.have.as_slice()[current_piece.piece_index.get() as usize];
if current_piece_have {
have_bytes += current_piece.len as u64;
}
if current_piece_selected {
selected_bytes += current_piece.len as u64;
}
if current_piece_selected && !current_piece_have {
needed_bytes += current_piece.len as u64;
}
self.selected.set(
current_piece.piece_index.get() as usize,
current_piece_selected,
);
match (current_piece_selected, current_piece_have) {
(true, true) => {}
(true, false) => {
self.mark_piece_broken_if_not_have(current_piece.piece_index)
}
(false, true) => {}
(false, false) => {
self.queue_pieces
.set(current_piece.piece_index.get() as usize, false);
}
}
if current_piece.piece_index != self.lengths.last_piece_id() {
current_piece = piece_it.next().context(
"bug: iter_piece_infos() pieces ended earlier than expected",
)?;
current_piece_selected = false;
current_piece_remaining = current_piece.len;
}
}
}
}
let res = HaveNeededSelected {
have_bytes,
needed_bytes,
selected_bytes,
};
self.hns = res;
Ok(res)
}
pub(crate) fn get_selected_pieces(&self) -> &BF {
&self.selected
}
pub fn is_file_finished(&self, file_info: &FileInfo) -> bool {
self.have
.as_slice()
.get(file_info.piece_range_usize())
.map(|r| r.all())
.unwrap_or(true)
}
pub(crate) fn is_finished(&self) -> bool {
self.get_hns().finished()
}
pub fn per_file_have_bytes(&self) -> &[u64] {
&self.per_file_bytes
}
pub fn update_file_have_on_piece_completed(
&mut self,
piece_id: ValidPieceIndex,
file_id: usize,
file_info: &FileInfo,
) -> u64 {
let diff_have = self.lengths.size_of_piece_in_file(
piece_id.get(),
file_info.offset_in_torrent,
file_info.len,
);
self.per_file_bytes[file_id] += diff_have;
file_info.len.saturating_sub(self.per_file_bytes[file_id])
}
}
#[cfg(test)]
mod tests {
use librqbit_core::{constants::CHUNK_SIZE, lengths::Lengths};
use std::collections::HashSet;
use crate::{bitv::BitV, chunk_tracker::HaveNeededSelected, type_aliases::BF};
use super::{compute_chunk_have_status, ChunkTracker};
#[test]
fn test_compute_chunk_status() {
let piece_length = CHUNK_SIZE * 2 + 1;
let l = Lengths::new(piece_length as u64 * 2 + 1, piece_length).unwrap();
assert_eq!(l.total_pieces(), 3);
assert_eq!(l.default_chunks_per_piece(), 3);
assert_eq!(l.total_chunks(), 7);
{
let mut have_pieces =
BF::from_boxed_slice(vec![u8::MAX; l.piece_bitfield_bytes()].into_boxed_slice());
have_pieces.set(0, false);
let chunks = compute_chunk_have_status(&l, &have_pieces).unwrap();
assert!(!chunks[0]);
assert!(!chunks[1]);
assert!(!chunks[2]);
assert!(chunks[3]);
assert!(chunks[4]);
assert!(chunks[5]);
assert!(chunks[6]);
}
{
let mut have_pieces =
BF::from_boxed_slice(vec![u8::MAX; l.piece_bitfield_bytes()].into_boxed_slice());
have_pieces.set(1, false);
let chunks = compute_chunk_have_status(&l, &have_pieces).unwrap();
dbg!(&chunks);
assert!(chunks[0]);
assert!(chunks[1]);
assert!(chunks[2]);
assert!(!chunks[3]);
assert!(!chunks[4]);
assert!(!chunks[5]);
assert!(chunks[6]);
}
{
let mut have_pieces =
BF::from_boxed_slice(vec![u8::MAX; l.piece_bitfield_bytes()].into_boxed_slice());
have_pieces.set(2, false);
let chunks = compute_chunk_have_status(&l, &have_pieces).unwrap();
dbg!(&chunks);
assert!(chunks[0]);
assert!(chunks[1]);
assert!(chunks[2]);
assert!(chunks[3]);
assert!(chunks[4]);
assert!(chunks[5]);
assert!(!chunks[6]);
}
{
let piece_length = CHUNK_SIZE * 2;
let l = Lengths::new(piece_length as u64 * 2 + 1, piece_length).unwrap();
assert_eq!(l.total_pieces(), 3);
assert_eq!(l.default_chunks_per_piece(), 2);
assert_eq!(l.total_chunks(), 5);
{
let mut have_pieces = BF::from_boxed_slice(
vec![u8::MAX; l.piece_bitfield_bytes()].into_boxed_slice(),
);
have_pieces.set(1, false);
let chunks = compute_chunk_have_status(&l, &have_pieces).unwrap();
dbg!(&chunks);
assert!(chunks[0]);
assert!(chunks[1]);
assert!(!chunks[2]);
assert!(!chunks[3]);
assert!(chunks[4]);
}
{
let mut have_pieces = BF::from_boxed_slice(
vec![u8::MAX; l.piece_bitfield_bytes()].into_boxed_slice(),
);
have_pieces.set(2, false);
let chunks = compute_chunk_have_status(&l, &have_pieces).unwrap();
dbg!(&chunks);
assert!(chunks[0]);
assert!(chunks[1]);
assert!(chunks[2]);
assert!(chunks[3]);
assert!(!chunks[4]);
}
}
}
#[test]
fn test_update_only_files() {
let piece_len = CHUNK_SIZE * 2 + 1;
let total_len = piece_len as u64 * 2 + 1;
let l = Lengths::new(total_len, piece_len).unwrap();
assert_eq!(l.total_pieces(), 3);
assert_eq!(l.total_chunks(), 7);
let all_files = [
piece_len as u64, 1, 0, piece_len as u64, ];
let bf_len = l.piece_bitfield_bytes();
let initial_have = BF::from_boxed_slice(vec![0u8; bf_len].into_boxed_slice());
let initial_selected = BF::from_boxed_slice(vec![u8::MAX; bf_len].into_boxed_slice());
let mut ct = ChunkTracker::new(
initial_have.clone().into_dyn(),
initial_selected.clone(),
l,
&Default::default(),
)
.unwrap();
assert_eq!(
ct.update_only_files(all_files.into_iter(), &HashSet::from_iter([0, 1, 2, 3]))
.unwrap(),
HaveNeededSelected {
have_bytes: 0,
selected_bytes: total_len,
needed_bytes: total_len,
}
);
assert_eq!(ct.have.as_slice(), initial_have.as_bitslice());
assert_eq!(ct.queue_pieces, initial_selected);
println!("Select only the first file.");
assert_eq!(
ct.update_only_files(all_files, &HashSet::from_iter([0]))
.unwrap(),
HaveNeededSelected {
have_bytes: 0,
selected_bytes: all_files[0],
needed_bytes: all_files[0],
}
);
assert!(ct.queue_pieces[0]);
assert!(!ct.queue_pieces[1]);
assert!(!ct.queue_pieces[2]);
assert_eq!(
ct.update_only_files(all_files, &HashSet::from_iter([1]))
.unwrap(),
HaveNeededSelected {
have_bytes: 0,
selected_bytes: piece_len as u64,
needed_bytes: piece_len as u64,
}
);
assert!(!ct.queue_pieces[0]);
assert!(ct.queue_pieces[1]);
assert!(!ct.queue_pieces[2]);
assert_eq!(
ct.update_only_files(all_files, &HashSet::from_iter([2]))
.unwrap(),
HaveNeededSelected {
have_bytes: 0,
selected_bytes: 0,
needed_bytes: 0,
}
);
assert!(!ct.queue_pieces[0]);
assert!(!ct.queue_pieces[1]);
assert!(!ct.queue_pieces[2]);
assert_eq!(
ct.update_only_files(all_files, &HashSet::from_iter([3]))
.unwrap(),
HaveNeededSelected {
have_bytes: 0,
selected_bytes: (piece_len + 1) as u64,
needed_bytes: (piece_len + 1) as u64,
}
);
assert!(!ct.queue_pieces[0]);
assert!(ct.queue_pieces[1]);
assert!(ct.queue_pieces[2]);
assert_eq!(
ct.update_only_files(all_files, &HashSet::from_iter([0, 3]))
.unwrap(),
HaveNeededSelected {
have_bytes: 0,
selected_bytes: all_files[0] + all_files[3] + 1,
needed_bytes: all_files[0] + all_files[3] + 1,
}
);
assert!(ct.queue_pieces[0]);
assert!(ct.queue_pieces[1]);
assert!(ct.queue_pieces[2]);
assert_eq!(
ct.update_only_files(all_files, &HashSet::from_iter([0, 1, 2, 3]))
.unwrap(),
HaveNeededSelected {
have_bytes: 0,
selected_bytes: total_len,
needed_bytes: total_len
}
);
assert!(ct.queue_pieces[0]);
assert!(ct.queue_pieces[1]);
assert!(ct.queue_pieces[2]);
}
}