#![deny(missing_docs)]
use crate::activity::ActivityMask;
use crate::api::lookahead::*;
use crate::api::{EncoderConfig, EncoderStatus, FrameType, Packet};
use crate::color::ChromaSampling::Cs400;
use crate::cpu_features::CpuFeatureLevel;
use crate::dist::get_satd;
use crate::encoder::*;
use crate::frame::*;
use crate::partition::*;
use crate::rate::{
RCState, FRAME_NSUBTYPES, FRAME_SUBTYPE_I, FRAME_SUBTYPE_P,
FRAME_SUBTYPE_SEF,
};
use crate::scenechange::SceneChangeDetector;
use crate::stats::EncoderStats;
use crate::tiling::Area;
use crate::util::Pixel;
use arrayvec::ArrayVec;
use log::Level::Info;
use rust_hawktracer::*;
use std::cmp;
use std::collections::{BTreeMap, BTreeSet};
use std::env;
use std::fs;
use std::path::PathBuf;
use std::sync::Arc;
#[derive(Debug, Clone, Copy)]
pub struct InterConfig {
reorder: bool,
pub(crate) multiref: bool,
pub(crate) pyramid_depth: u64,
pub(crate) group_input_len: u64,
group_output_len: u64,
pub(crate) switch_frame_interval: u64,
}
impl InterConfig {
pub(crate) fn new(enc_config: &EncoderConfig) -> InterConfig {
let reorder = !enc_config.low_latency;
let pyramid_depth = if reorder { 2 } else { 0 };
let group_input_len = 1 << pyramid_depth;
let group_output_len = group_input_len + pyramid_depth;
let switch_frame_interval = enc_config.switch_frame_interval;
assert!(switch_frame_interval % group_input_len == 0);
InterConfig {
reorder,
multiref: reorder || enc_config.speed_settings.multiref,
pyramid_depth,
group_input_len,
group_output_len,
switch_frame_interval,
}
}
pub(crate) fn get_idx_in_group_output(
&self, output_frameno_in_gop: u64,
) -> u64 {
debug_assert!(output_frameno_in_gop > 0);
(output_frameno_in_gop - 1) % self.group_output_len
}
pub(crate) fn get_order_hint(
&self, output_frameno_in_gop: u64, idx_in_group_output: u64,
) -> u32 {
debug_assert!(output_frameno_in_gop > 0);
let group_idx = (output_frameno_in_gop - 1) / self.group_output_len;
let offset = if idx_in_group_output < self.pyramid_depth {
self.group_input_len >> idx_in_group_output
} else {
idx_in_group_output - self.pyramid_depth + 1
};
(self.group_input_len * group_idx + offset) as u32
}
pub(crate) fn get_level(&self, idx_in_group_output: u64) -> u64 {
if !self.reorder {
0
} else if idx_in_group_output < self.pyramid_depth {
idx_in_group_output
} else {
pos_to_lvl(
idx_in_group_output - self.pyramid_depth + 1,
self.pyramid_depth,
)
}
}
pub(crate) fn get_slot_idx(&self, level: u64, order_hint: u32) -> u32 {
if level == 0 {
(order_hint >> self.pyramid_depth) & 3
} else {
3 + level as u32
}
}
pub(crate) const fn get_show_frame(&self, idx_in_group_output: u64) -> bool {
idx_in_group_output >= self.pyramid_depth
}
pub(crate) fn get_show_existing_frame(
&self, idx_in_group_output: u64,
) -> bool {
self.reorder
&& self.get_show_frame(idx_in_group_output)
&& (idx_in_group_output - self.pyramid_depth + 1).count_ones() == 1
&& idx_in_group_output != self.pyramid_depth
}
pub(crate) fn get_input_frameno(
&self, output_frameno_in_gop: u64, gop_input_frameno_start: u64,
) -> u64 {
if output_frameno_in_gop == 0 {
gop_input_frameno_start
} else {
let idx_in_group_output =
self.get_idx_in_group_output(output_frameno_in_gop);
let order_hint =
self.get_order_hint(output_frameno_in_gop, idx_in_group_output);
gop_input_frameno_start + order_hint as u64
}
}
const fn max_reordering_latency(&self) -> u64 {
self.group_input_len
}
pub(crate) fn keyframe_lookahead_distance(&self) -> u64 {
self.max_reordering_latency() + 1
}
pub(crate) fn allowed_ref_frames(&self) -> &[RefType] {
use crate::partition::RefType::*;
if self.reorder {
&ALL_INTER_REFS
} else if self.multiref {
&[LAST_FRAME, LAST2_FRAME, LAST3_FRAME, GOLDEN_FRAME]
} else {
&[LAST_FRAME]
}
}
}
#[derive(Clone)]
pub(crate) struct FrameData<T: Pixel> {
pub(crate) fi: FrameInvariants<T>,
pub(crate) fs: FrameState<T>,
}
impl<T: Pixel> FrameData<T> {
pub(crate) fn new(fi: FrameInvariants<T>, frame: Arc<Frame<T>>) -> Self {
let fs = FrameState::new_with_frame(&fi, frame);
FrameData { fi, fs }
}
}
type FrameQueue<T> = BTreeMap<u64, Option<Arc<Frame<T>>>>;
type FrameDataQueue<T> = BTreeMap<u64, FrameData<T>>;
pub(crate) struct ContextInner<T: Pixel> {
pub(crate) frame_count: u64,
pub(crate) limit: Option<u64>,
pub(crate) output_frameno: u64,
pub(super) inter_cfg: InterConfig,
pub(super) frames_processed: u64,
pub(super) frame_q: FrameQueue<T>,
pub(super) frame_data: FrameDataQueue<T>,
keyframes: BTreeSet<u64>,
keyframes_forced: BTreeSet<u64>,
packet_data: Vec<u8>,
gop_output_frameno_start: BTreeMap<u64, u64>,
pub(crate) gop_input_frameno_start: BTreeMap<u64, u64>,
keyframe_detector: SceneChangeDetector,
pub(crate) config: Arc<EncoderConfig>,
seq: Arc<Sequence>,
pub(crate) rc_state: RCState,
maybe_prev_log_base_q: Option<i64>,
next_lookahead_frame: u64,
next_lookahead_output_frameno: u64,
opaque_q: BTreeMap<u64, Box<dyn std::any::Any + Send>>,
}
impl<T: Pixel> ContextInner<T> {
pub fn new(enc: &EncoderConfig) -> Self {
let packet_data = TEMPORAL_DELIMITER.to_vec();
let mut keyframes = BTreeSet::new();
keyframes.insert(0);
let maybe_ac_qi_max =
if enc.quantizer < 255 { Some(enc.quantizer as u8) } else { None };
let seq = Arc::new(Sequence::new(enc));
let inter_cfg = InterConfig::new(enc);
let lookahead_distance = inter_cfg.keyframe_lookahead_distance() as usize;
if log_enabled!(Info) {
let tiling = seq.tiling;
if tiling.tile_count() == 1 {
info!("Using 1 tile");
} else {
info!(
"Using {} tiles ({}x{})",
tiling.tile_count(),
tiling.cols,
tiling.rows
);
}
}
ContextInner {
frame_count: 0,
limit: None,
inter_cfg,
output_frameno: 0,
frames_processed: 0,
frame_q: BTreeMap::new(),
frame_data: BTreeMap::new(),
keyframes,
keyframes_forced: BTreeSet::new(),
packet_data,
gop_output_frameno_start: BTreeMap::new(),
gop_input_frameno_start: BTreeMap::new(),
keyframe_detector: SceneChangeDetector::new(
*enc,
CpuFeatureLevel::default(),
lookahead_distance,
seq.clone(),
true,
),
config: Arc::new(*enc),
seq,
rc_state: RCState::new(
enc.width as i32,
enc.height as i32,
enc.time_base.den as i64,
enc.time_base.num as i64,
enc.bitrate,
maybe_ac_qi_max,
enc.min_quantizer,
enc.max_key_frame_interval as i32,
enc.reservoir_frame_delay,
),
maybe_prev_log_base_q: None,
next_lookahead_frame: 1,
next_lookahead_output_frameno: 0,
opaque_q: BTreeMap::new(),
}
}
#[hawktracer(send_frame)]
pub fn send_frame(
&mut self, frame: Option<Arc<Frame<T>>>, params: Option<FrameParameters>,
) -> Result<(), EncoderStatus> {
let input_frameno = self.frame_count;
let is_flushing = frame.is_none();
if !is_flushing {
self.frame_count += 1;
}
self.frame_q.insert(input_frameno, frame);
if let Some(params) = params {
if params.frame_type_override == FrameTypeOverride::Key {
self.keyframes_forced.insert(input_frameno);
}
if let Some(op) = params.opaque {
self.opaque_q.insert(input_frameno, op);
}
}
if !self.needs_more_frame_q_lookahead(self.next_lookahead_frame) {
let lookahead_frames = self
.frame_q
.range(self.next_lookahead_frame - 1..)
.filter_map(|(&_input_frameno, frame)| frame.clone())
.collect::<Vec<_>>();
if is_flushing {
for cur_lookahead_frames in
std::iter::successors(Some(&lookahead_frames[..]), |s| s.get(1..))
{
if cur_lookahead_frames.len() < 2 {
break;
}
self.compute_keyframe_placement(cur_lookahead_frames);
}
} else {
self.compute_keyframe_placement(&lookahead_frames);
}
}
self.compute_frame_invariants();
Ok(())
}
fn needs_more_frame_q_lookahead(&self, input_frameno: u64) -> bool {
let lookahead_end = self.frame_q.keys().last().cloned().unwrap_or(0);
let frames_needed =
input_frameno + self.inter_cfg.keyframe_lookahead_distance() + 1;
lookahead_end < frames_needed && self.needs_more_frames(lookahead_end)
}
pub fn needs_more_fi_lookahead(&self) -> bool {
let ready_frames = self.get_rdo_lookahead_frames().count();
ready_frames < self.config.rdo_lookahead_frames + 1
&& self.needs_more_frames(self.next_lookahead_frame)
}
pub fn needs_more_frames(&self, frame_count: u64) -> bool {
self.limit.map(|limit| frame_count < limit).unwrap_or(true)
}
fn get_rdo_lookahead_frames(
&self,
) -> impl Iterator<Item = (&u64, &FrameData<T>)> {
self
.frame_data
.iter()
.skip_while(move |(&output_frameno, _)| {
output_frameno < self.output_frameno
})
.filter(|(_, data)| !data.fi.invalid && !data.fi.show_existing_frame)
.take(self.config.rdo_lookahead_frames + 1)
}
fn next_keyframe_input_frameno(
&self, gop_input_frameno_start: u64, ignore_limit: bool,
) -> u64 {
let next_detected = self
.keyframes
.iter()
.find(|&&input_frameno| input_frameno > gop_input_frameno_start)
.cloned();
let mut next_limit =
gop_input_frameno_start + self.config.max_key_frame_interval;
if !ignore_limit && self.limit.is_some() {
next_limit = next_limit.min(self.limit.unwrap());
}
if next_detected.is_none() {
return next_limit;
}
cmp::min(next_detected.unwrap(), next_limit)
}
fn set_frame_properties(
&mut self, output_frameno: u64,
) -> Result<(), EncoderStatus> {
let fi = self.build_frame_properties(output_frameno)?;
let frame =
self.frame_q.get(&fi.input_frameno).as_ref().unwrap().as_ref().unwrap();
self.frame_data.insert(output_frameno, FrameData::new(fi, frame.clone()));
Ok(())
}
#[allow(unused)]
pub fn build_dump_properties() -> PathBuf {
let mut data_location = PathBuf::new();
if env::var_os("RAV1E_DATA_PATH").is_some() {
data_location.push(&env::var_os("RAV1E_DATA_PATH").unwrap());
fs::create_dir_all(data_location.clone()).unwrap();
data_location
} else {
data_location.push(&env::current_dir().unwrap());
data_location.push(".lookahead_data");
fs::create_dir_all(data_location.clone()).unwrap();
data_location
}
}
fn build_frame_properties(
&mut self, output_frameno: u64,
) -> Result<FrameInvariants<T>, EncoderStatus> {
let (prev_gop_output_frameno_start, prev_gop_input_frameno_start) =
if output_frameno == 0 {
(0, 0)
} else {
(
self.gop_output_frameno_start[&(output_frameno - 1)],
self.gop_input_frameno_start[&(output_frameno - 1)],
)
};
self
.gop_output_frameno_start
.insert(output_frameno, prev_gop_output_frameno_start);
self
.gop_input_frameno_start
.insert(output_frameno, prev_gop_input_frameno_start);
let output_frameno_in_gop =
output_frameno - self.gop_output_frameno_start[&output_frameno];
let mut input_frameno = self.inter_cfg.get_input_frameno(
output_frameno_in_gop,
self.gop_input_frameno_start[&output_frameno],
);
if self.needs_more_frame_q_lookahead(input_frameno) {
return Err(EncoderStatus::NeedMoreData);
}
if output_frameno_in_gop > 0 {
let next_keyframe_input_frameno = self.next_keyframe_input_frameno(
self.gop_input_frameno_start[&output_frameno],
false,
);
let prev_input_frameno =
self.frame_data[&(output_frameno - 1)].fi.input_frameno;
if input_frameno >= next_keyframe_input_frameno {
if !self.inter_cfg.reorder
|| ((output_frameno_in_gop - 1) % self.inter_cfg.group_output_len
== 0
&& prev_input_frameno == (next_keyframe_input_frameno - 1))
{
input_frameno = next_keyframe_input_frameno;
match self.frame_q.get(&input_frameno) {
Some(Some(_)) => {}
_ => {
return Err(EncoderStatus::NeedMoreData);
}
}
*self.gop_output_frameno_start.get_mut(&output_frameno).unwrap() =
output_frameno;
*self.gop_input_frameno_start.get_mut(&output_frameno).unwrap() =
next_keyframe_input_frameno;
} else {
let fi = FrameInvariants::new_inter_frame(
&self.frame_data[&(output_frameno - 1)].fi,
&self.inter_cfg,
self.gop_input_frameno_start[&output_frameno],
output_frameno_in_gop,
next_keyframe_input_frameno,
self.config.error_resilient,
);
assert!(fi.invalid);
return Ok(fi);
}
}
}
match self.frame_q.get(&input_frameno) {
Some(Some(_)) => {}
_ => {
return Err(EncoderStatus::NeedMoreData);
}
}
let frame_type = if self.keyframes.contains(&input_frameno) {
FrameType::KEY
} else {
FrameType::INTER
};
if frame_type == FrameType::KEY {
*self.gop_output_frameno_start.get_mut(&output_frameno).unwrap() =
output_frameno;
*self.gop_input_frameno_start.get_mut(&output_frameno).unwrap() =
input_frameno;
}
let output_frameno_in_gop =
output_frameno - self.gop_output_frameno_start[&output_frameno];
if output_frameno_in_gop == 0 {
let fi = FrameInvariants::new_key_frame(
self.config.clone(),
self.seq.clone(),
self.gop_input_frameno_start[&output_frameno],
);
assert!(!fi.invalid);
Ok(fi)
} else {
let next_keyframe_input_frameno = self.next_keyframe_input_frameno(
self.gop_input_frameno_start[&output_frameno],
false,
);
let fi = FrameInvariants::new_inter_frame(
&self.frame_data[&(output_frameno - 1)].fi,
&self.inter_cfg,
self.gop_input_frameno_start[&output_frameno],
output_frameno_in_gop,
next_keyframe_input_frameno,
self.config.error_resilient,
);
assert!(!fi.invalid);
Ok(fi)
}
}
pub(crate) fn done_processing(&self) -> bool {
self.limit.map(|limit| self.frames_processed == limit).unwrap_or(false)
}
#[hawktracer(compute_lookahead_motion_vectors)]
fn compute_lookahead_motion_vectors(&mut self, output_frameno: u64) {
let qps = {
let frame_data = self.frame_data.get(&output_frameno).unwrap();
let fti = frame_data.fi.get_frame_subtype();
self.rc_state.select_qi(
self,
output_frameno,
fti,
self.maybe_prev_log_base_q,
)
};
let frame_data = self.frame_data.get_mut(&output_frameno).unwrap();
let fs = &mut frame_data.fs;
let fi = &mut frame_data.fi;
if fi.invalid || fi.show_existing_frame {
return;
}
#[cfg(feature = "dump_lookahead_data")]
{
let data_location = Self::build_dump_properties();
let plane = &fs.input_qres;
let mut file_name = format!("{:010}-qres", fi.input_frameno);
let buf: Vec<_> = plane.iter().map(|p| p.as_()).collect();
image::GrayImage::from_vec(
plane.cfg.width as u32,
plane.cfg.height as u32,
buf,
)
.unwrap()
.save(data_location.join(file_name).with_extension("png"))
.unwrap();
let plane = &fs.input_hres;
file_name = format!("{:010}-hres", fi.input_frameno);
let buf: Vec<_> = plane.iter().map(|p| p.as_()).collect();
image::GrayImage::from_vec(
plane.cfg.width as u32,
plane.cfg.height as u32,
buf,
)
.unwrap()
.save(data_location.join(file_name).with_extension("png"))
.unwrap();
}
if self.output_frameno == output_frameno {
let rfs = Arc::new(ReferenceFrame {
order_hint: fi.order_hint,
width: fi.width as u32,
height: fi.height as u32,
render_width: fi.render_width,
render_height: fi.render_height,
frame: fs.input.clone(),
input_hres: fs.input_hres.clone(),
input_qres: fs.input_qres.clone(),
cdfs: fs.cdfs,
frame_me_stats: fs.frame_me_stats.clone(),
output_frameno,
segmentation: fs.segmentation,
});
for i in 0..(REF_FRAMES as usize) {
if (fi.refresh_frame_flags & (1 << i)) != 0 {
fi.lookahead_rec_buffer.frames[i] = Some(Arc::clone(&rfs));
fi.lookahead_rec_buffer.deblock[i] = fs.deblock;
}
}
return;
}
fi.rec_buffer = fi.lookahead_rec_buffer.clone();
fi.set_quantizers(&qps);
compute_motion_vectors(fi, fs, &self.inter_cfg);
fi.lookahead_me_stats = fs.frame_me_stats.clone();
#[cfg(feature = "dump_lookahead_data")]
{
use crate::partition::RefType::*;
let data_location = Self::build_dump_properties();
let file_name = format!("{:010}-mvs", fi.input_frameno);
let second_ref_frame = if !self.inter_cfg.multiref {
LAST_FRAME
} else if fi.idx_in_group_output == 0 {
LAST2_FRAME
} else {
ALTREF_FRAME
};
let index = if second_ref_frame.to_index() != 0 { 0 } else { 1 };
let me_stats = &fs.frame_me_stats[index];
use byteorder::{NativeEndian, WriteBytesExt};
let mut buf = vec![];
buf.write_u64::<NativeEndian>(me_stats.rows as u64).unwrap();
buf.write_u64::<NativeEndian>(me_stats.cols as u64).unwrap();
for y in 0..me_stats.rows {
for x in 0..me_stats.cols {
let mv = me_stats[y][x].mv;
buf.write_i16::<NativeEndian>(mv.row).unwrap();
buf.write_i16::<NativeEndian>(mv.col).unwrap();
}
}
::std::fs::write(
data_location.join(file_name).with_extension("bin"),
buf,
)
.unwrap();
}
let rfs = Arc::new(ReferenceFrame {
order_hint: fi.order_hint,
width: fi.width as u32,
height: fi.height as u32,
render_width: fi.render_width,
render_height: fi.render_height,
frame: fs.input.clone(),
input_hres: fs.input_hres.clone(),
input_qres: fs.input_qres.clone(),
cdfs: fs.cdfs,
frame_me_stats: fs.frame_me_stats.clone(),
output_frameno,
segmentation: fs.segmentation,
});
for i in 0..(REF_FRAMES as usize) {
if (fi.refresh_frame_flags & (1 << i)) != 0 {
fi.lookahead_rec_buffer.frames[i] = Some(Arc::clone(&rfs));
fi.lookahead_rec_buffer.deblock[i] = fs.deblock;
}
}
}
#[hawktracer(compute_lookahead_intra_costs)]
fn compute_lookahead_intra_costs(&mut self, output_frameno: u64) {
let frame_data = self.frame_data.get(&output_frameno).unwrap();
let fi = &frame_data.fi;
if fi.invalid || fi.show_existing_frame {
return;
}
self
.frame_data
.get_mut(&output_frameno)
.unwrap()
.fi
.lookahead_intra_costs = estimate_intra_costs(
&*self.frame_q[&fi.input_frameno].as_ref().unwrap(),
fi.sequence.bit_depth,
fi.cpu_feature_level,
);
}
#[hawktracer(compute_keyframe_placement)]
pub fn compute_keyframe_placement(
&mut self, lookahead_frames: &[Arc<Frame<T>>],
) {
if self.keyframes_forced.contains(&self.next_lookahead_frame)
|| self.keyframe_detector.analyze_next_frame(
lookahead_frames,
self.next_lookahead_frame,
*self.keyframes.iter().last().unwrap(),
)
{
self.keyframes.insert(self.next_lookahead_frame);
}
self.next_lookahead_frame += 1;
}
#[hawktracer(compute_frame_invariants)]
pub fn compute_frame_invariants(&mut self) {
while self.set_frame_properties(self.next_lookahead_output_frameno).is_ok()
{
self
.compute_lookahead_motion_vectors(self.next_lookahead_output_frameno);
if self.config.temporal_rdo() {
self.compute_lookahead_intra_costs(self.next_lookahead_output_frameno);
}
self.next_lookahead_output_frameno += 1;
}
}
#[hawktracer(compute_block_importances)]
fn compute_block_importances(&mut self) {
if self.frame_data[&self.output_frameno].fi.show_existing_frame {
return;
}
let output_framenos = self
.get_rdo_lookahead_frames()
.map(|(&output_frameno, _)| output_frameno)
.collect::<Vec<_>>();
assert_eq!(output_framenos[0], self.output_frameno);
for output_frameno in output_framenos.iter() {
let fi = &mut self.frame_data.get_mut(output_frameno).unwrap().fi;
for x in fi.block_importances.iter_mut() {
*x = 0.;
}
}
let bsize = BlockSize::from_width_and_height(
IMPORTANCE_BLOCK_SIZE,
IMPORTANCE_BLOCK_SIZE,
);
for &output_frameno in output_framenos.iter().skip(1).rev() {
if self.frame_data.get(&output_frameno).unwrap().fi.frame_type
== FrameType::KEY
{
continue;
}
let output_frame_data = self.frame_data.remove(&output_frameno).unwrap();
let fi = &output_frame_data.fi;
let frame = self.frame_q[&fi.input_frameno].as_ref().unwrap();
let mut unique_indices = ArrayVec::<[_; 3]>::new();
for (mv_index, &rec_index) in fi.ref_frames.iter().enumerate() {
if unique_indices.iter().find(|&&(_, r)| r == rec_index).is_none() {
unique_indices.push((mv_index, rec_index));
}
}
let bit_depth = self.config.bit_depth;
let frame_data = &mut self.frame_data;
let len = unique_indices.len();
unique_indices.iter().for_each(|&(mv_index, rec_index)| {
let reference =
fi.rec_buffer.frames[rec_index as usize].as_ref().unwrap();
let reference_frame = &reference.frame;
let reference_output_frameno = reference.output_frameno;
let me_stats = &fi.lookahead_me_stats[mv_index];
assert_ne!(reference_output_frameno, output_frameno);
if let Some(reference_frame_block_importances) = frame_data
.get_mut(&reference_output_frameno)
.map(|data| &mut data.fi.block_importances)
{
update_block_importances(
fi,
me_stats,
frame,
reference_frame,
bit_depth,
bsize,
len,
reference_frame_block_importances,
);
#[hawktracer(update_block_importances)]
fn update_block_importances<T: Pixel>(
fi: &FrameInvariants<T>, me_stats: &crate::me::FrameMEStats,
frame: &Frame<T>, reference_frame: &Frame<T>, bit_depth: usize,
bsize: BlockSize, len: usize,
reference_frame_block_importances: &mut [f32],
) {
let plane_org = &frame.planes[0];
let plane_ref = &reference_frame.planes[0];
(0..fi.h_in_imp_b)
.zip(fi.lookahead_intra_costs.chunks_exact(fi.w_in_imp_b))
.zip(fi.block_importances.chunks_exact(fi.w_in_imp_b))
.for_each(|((y, lookahead_intra_costs), block_importances)| {
(0..fi.w_in_imp_b).for_each(|x| {
let mv = me_stats[y * 2][x * 2].mv;
let reference_x =
x as i64 * IMP_BLOCK_SIZE_IN_MV_UNITS + mv.col as i64;
let reference_y =
y as i64 * IMP_BLOCK_SIZE_IN_MV_UNITS + mv.row as i64;
let region_org = plane_org.region(Area::Rect {
x: (x * IMPORTANCE_BLOCK_SIZE) as isize,
y: (y * IMPORTANCE_BLOCK_SIZE) as isize,
width: IMPORTANCE_BLOCK_SIZE,
height: IMPORTANCE_BLOCK_SIZE,
});
let region_ref = plane_ref.region(Area::Rect {
x: reference_x as isize
/ IMP_BLOCK_MV_UNITS_PER_PIXEL as isize,
y: reference_y as isize
/ IMP_BLOCK_MV_UNITS_PER_PIXEL as isize,
width: IMPORTANCE_BLOCK_SIZE,
height: IMPORTANCE_BLOCK_SIZE,
});
let inter_cost = get_satd(
®ion_org,
®ion_ref,
bsize,
bit_depth,
fi.cpu_feature_level,
) as f32;
let intra_cost = lookahead_intra_costs[x] as f32;
let future_importance = block_importances[x];
let propagate_fraction = if intra_cost <= inter_cost {
0.
} else {
1. - inter_cost / intra_cost
};
let propagate_amount = (intra_cost + future_importance)
* propagate_fraction
/ len as f32;
let mut propagate =
|block_x_in_mv_units, block_y_in_mv_units, fraction| {
let x = block_x_in_mv_units / IMP_BLOCK_SIZE_IN_MV_UNITS;
let y = block_y_in_mv_units / IMP_BLOCK_SIZE_IN_MV_UNITS;
if x >= 0
&& y >= 0
&& (x as usize) < fi.w_in_imp_b
&& (y as usize) < fi.h_in_imp_b
{
reference_frame_block_importances
[y as usize * fi.w_in_imp_b + x as usize] +=
propagate_amount * fraction;
}
};
let top_left_block_x = (reference_x
- if reference_x < 0 {
IMP_BLOCK_SIZE_IN_MV_UNITS - 1
} else {
0
})
/ IMP_BLOCK_SIZE_IN_MV_UNITS
* IMP_BLOCK_SIZE_IN_MV_UNITS;
let top_left_block_y = (reference_y
- if reference_y < 0 {
IMP_BLOCK_SIZE_IN_MV_UNITS - 1
} else {
0
})
/ IMP_BLOCK_SIZE_IN_MV_UNITS
* IMP_BLOCK_SIZE_IN_MV_UNITS;
debug_assert!(reference_x >= top_left_block_x);
debug_assert!(reference_y >= top_left_block_y);
let top_right_block_x =
top_left_block_x + IMP_BLOCK_SIZE_IN_MV_UNITS;
let top_right_block_y = top_left_block_y;
let bottom_left_block_x = top_left_block_x;
let bottom_left_block_y =
top_left_block_y + IMP_BLOCK_SIZE_IN_MV_UNITS;
let bottom_right_block_x = top_right_block_x;
let bottom_right_block_y = bottom_left_block_y;
let top_left_block_fraction = ((top_right_block_x
- reference_x)
* (bottom_left_block_y - reference_y))
as f32
/ IMP_BLOCK_AREA_IN_MV_UNITS as f32;
propagate(
top_left_block_x,
top_left_block_y,
top_left_block_fraction,
);
let top_right_block_fraction = ((reference_x
+ IMP_BLOCK_SIZE_IN_MV_UNITS
- top_right_block_x)
* (bottom_left_block_y - reference_y))
as f32
/ IMP_BLOCK_AREA_IN_MV_UNITS as f32;
propagate(
top_right_block_x,
top_right_block_y,
top_right_block_fraction,
);
let bottom_left_block_fraction =
((top_right_block_x - reference_x)
* (reference_y + IMP_BLOCK_SIZE_IN_MV_UNITS
- bottom_left_block_y)) as f32
/ IMP_BLOCK_AREA_IN_MV_UNITS as f32;
propagate(
bottom_left_block_x,
bottom_left_block_y,
bottom_left_block_fraction,
);
let bottom_right_block_fraction =
((reference_x + IMP_BLOCK_SIZE_IN_MV_UNITS
- top_right_block_x)
* (reference_y + IMP_BLOCK_SIZE_IN_MV_UNITS
- bottom_left_block_y)) as f32
/ IMP_BLOCK_AREA_IN_MV_UNITS as f32;
propagate(
bottom_right_block_x,
bottom_right_block_y,
bottom_right_block_fraction,
);
});
});
}
}
});
self.frame_data.insert(output_frameno, output_frame_data);
}
if !output_framenos.is_empty() {
let fi = &mut self.frame_data.get_mut(&output_framenos[0]).unwrap().fi;
let block_importances = fi.block_importances.iter();
let lookahead_intra_costs = fi.lookahead_intra_costs.iter();
let distortion_scales = fi.distortion_scales.iter_mut();
for ((&propagate_cost, &intra_cost), distortion_scale) in
block_importances.zip(lookahead_intra_costs).zip(distortion_scales)
{
*distortion_scale = crate::rdo::distortion_scale_for(
propagate_cost as f64,
intra_cost as f64,
);
}
#[cfg(feature = "dump_lookahead_data")]
{
use byteorder::{NativeEndian, WriteBytesExt};
let mut buf = vec![];
let data_location = Self::build_dump_properties();
let file_name = format!("{:010}-imps", fi.input_frameno);
buf.write_u64::<NativeEndian>(fi.h_in_imp_b as u64).unwrap();
buf.write_u64::<NativeEndian>(fi.w_in_imp_b as u64).unwrap();
buf.write_u64::<NativeEndian>(fi.get_frame_subtype() as u64).unwrap();
for y in 0..fi.h_in_imp_b {
for x in 0..fi.w_in_imp_b {
buf
.write_f32::<NativeEndian>(f64::from(
fi.distortion_scales[y * fi.w_in_imp_b + x],
) as f32)
.unwrap();
}
}
::std::fs::write(
data_location.join(file_name).with_extension("bin"),
buf,
)
.unwrap();
}
}
}
pub(crate) fn encode_packet(
&mut self, cur_output_frameno: u64,
) -> Result<Packet<T>, EncoderStatus> {
if self.frame_data.get(&cur_output_frameno).unwrap().fi.show_existing_frame
{
if !self.rc_state.ready() {
return Err(EncoderStatus::NotReady);
}
let frame_data = self.frame_data.get_mut(&cur_output_frameno).unwrap();
let sef_data = encode_show_existing_frame(
&frame_data.fi,
&mut frame_data.fs,
&self.inter_cfg,
);
let bits = (sef_data.len() * 8) as i64;
self.packet_data.extend(sef_data);
self.rc_state.update_state(
bits,
FRAME_SUBTYPE_SEF,
frame_data.fi.show_frame,
0,
false,
false,
);
let (rec, source) = if frame_data.fi.show_frame {
(Some(frame_data.fs.rec.clone()), Some(frame_data.fs.input.clone()))
} else {
(None, None)
};
self.output_frameno += 1;
let input_frameno = frame_data.fi.input_frameno;
let frame_type = frame_data.fi.frame_type;
let qp = frame_data.fi.base_q_idx;
let enc_stats = frame_data.fs.enc_stats.clone();
self.finalize_packet(
rec,
source,
input_frameno,
frame_type,
qp,
enc_stats,
)
} else if let Some(Some(_)) = self
.frame_q
.get(&self.frame_data.get(&cur_output_frameno).unwrap().fi.input_frameno)
{
if !self.rc_state.ready() {
return Err(EncoderStatus::NotReady);
}
let mut frame_data =
self.frame_data.get(&cur_output_frameno).cloned().unwrap();
let fti = frame_data.fi.get_frame_subtype();
let qps = self.rc_state.select_qi(
self,
cur_output_frameno,
fti,
self.maybe_prev_log_base_q,
);
frame_data.fi.set_quantizers(&qps);
if self.rc_state.needs_trial_encode(fti) {
let mut trial_fs = frame_data.fs.clone();
let data =
encode_frame(&frame_data.fi, &mut trial_fs, &self.inter_cfg);
self.rc_state.update_state(
(data.len() * 8) as i64,
fti,
frame_data.fi.show_frame,
qps.log_target_q,
true,
false,
);
let qps = self.rc_state.select_qi(
self,
cur_output_frameno,
fti,
self.maybe_prev_log_base_q,
);
frame_data.fi.set_quantizers(&qps);
}
frame_data.fi.activity_mask = ActivityMask::default();
let data =
encode_frame(&frame_data.fi, &mut frame_data.fs, &self.inter_cfg);
let enc_stats = frame_data.fs.enc_stats.clone();
self.maybe_prev_log_base_q = Some(qps.log_base_q);
self.rc_state.update_state(
(data.len() * 8) as i64,
fti,
frame_data.fi.show_frame,
qps.log_target_q,
false,
false,
);
self.packet_data.extend(data);
let planes =
if frame_data.fi.sequence.chroma_sampling == Cs400 { 1 } else { 3 };
Arc::make_mut(&mut frame_data.fs.rec).pad(
frame_data.fi.width,
frame_data.fi.height,
planes,
);
let (rec, source) = if frame_data.fi.show_frame {
(Some(frame_data.fs.rec.clone()), Some(frame_data.fs.input.clone()))
} else {
(None, None)
};
update_rec_buffer(
cur_output_frameno,
&mut frame_data.fi,
&frame_data.fs,
);
let rec_buffer = frame_data.fi.rec_buffer.clone();
for subsequent_fi in self
.frame_data
.iter_mut()
.skip_while(|(&output_frameno, _)| {
output_frameno <= cur_output_frameno
})
.map(|(_, frame_data)| &mut frame_data.fi)
.filter(|fi| !fi.invalid)
.take_while(|fi| fi.frame_type != FrameType::KEY)
{
subsequent_fi.rec_buffer = rec_buffer.clone();
subsequent_fi.set_ref_frame_sign_bias();
if !subsequent_fi.show_existing_frame {
break;
}
}
self.frame_data.insert(cur_output_frameno, frame_data);
let frame_data = &self.frame_data.get(&cur_output_frameno).unwrap();
let fi = &frame_data.fi;
self.output_frameno += 1;
if fi.show_frame {
let input_frameno = fi.input_frameno;
let frame_type = fi.frame_type;
let qp = fi.base_q_idx;
self.finalize_packet(
rec,
source,
input_frameno,
frame_type,
qp,
enc_stats,
)
} else {
Err(EncoderStatus::Encoded)
}
} else {
Err(EncoderStatus::NeedMoreData)
}
}
#[hawktracer(receive_packet)]
pub fn receive_packet(&mut self) -> Result<Packet<T>, EncoderStatus> {
if self.done_processing() {
return Err(EncoderStatus::LimitReached);
}
if self.needs_more_fi_lookahead() {
return Err(EncoderStatus::NeedMoreData);
}
self.output_frameno = self
.frame_data
.iter()
.skip_while(|(&output_frameno, _)| output_frameno < self.output_frameno)
.find(|(_, data)| !data.fi.invalid)
.map(|(&output_frameno, _)| output_frameno)
.ok_or(EncoderStatus::NeedMoreData)?;
let input_frameno = self.frame_data[&self.output_frameno].fi.input_frameno;
if !self.needs_more_frames(input_frameno) {
return Err(EncoderStatus::LimitReached);
}
if self.config.temporal_rdo() {
self.compute_block_importances();
}
let cur_output_frameno = self.output_frameno;
let mut ret = self.encode_packet(cur_output_frameno);
if let Ok(ref mut pkt) = ret {
self.garbage_collect(pkt.input_frameno);
pkt.opaque = self.opaque_q.remove(&pkt.input_frameno);
}
ret
}
fn finalize_packet(
&mut self, rec: Option<Arc<Frame<T>>>, source: Option<Arc<Frame<T>>>,
input_frameno: u64, frame_type: FrameType, qp: u8,
enc_stats: EncoderStats,
) -> Result<Packet<T>, EncoderStatus> {
let data = self.packet_data.clone();
self.packet_data.clear();
if write_temporal_delimiter(&mut self.packet_data).is_err() {
return Err(EncoderStatus::Failure);
}
self.frames_processed += 1;
Ok(Packet {
data,
rec,
source,
input_frameno,
frame_type,
qp,
enc_stats,
opaque: None,
})
}
fn garbage_collect(&mut self, cur_input_frameno: u64) {
if cur_input_frameno == 0 {
return;
}
let frame_q_start = self.frame_q.keys().next().cloned().unwrap_or(0);
for i in frame_q_start..cur_input_frameno {
self.frame_q.remove(&i);
}
if self.output_frameno < 2 {
return;
}
let fi_start = self.frame_data.keys().next().cloned().unwrap_or(0);
for i in fi_start..(self.output_frameno - 1) {
self.frame_data.remove(&i);
self.gop_output_frameno_start.remove(&i);
self.gop_input_frameno_start.remove(&i);
}
}
pub(crate) fn guess_frame_subtypes(
&self, nframes: &mut [i32; FRAME_NSUBTYPES + 1],
reservoir_frame_delay: i32,
) -> (i32, i32) {
for fti in 0..=FRAME_NSUBTYPES {
nframes[fti] = 0;
}
let mut prev_keyframe_input_frameno = *self
.gop_input_frameno_start
.get(&self.output_frameno)
.unwrap_or_else(|| {
assert!(self.output_frameno == 0);
&0
});
let mut prev_keyframe_output_frameno = *self
.gop_output_frameno_start
.get(&self.output_frameno)
.unwrap_or_else(|| {
assert!(self.output_frameno == 0);
&0
});
let mut prev_keyframe_ntus = 0;
let mut prev_keyframe_nframes = 0;
let mut acc: [i32; FRAME_NSUBTYPES + 1] = [0; FRAME_NSUBTYPES + 1];
fn collect_counts(
nframes: &mut [i32; FRAME_NSUBTYPES + 1],
acc: &mut [i32; FRAME_NSUBTYPES + 1],
) {
for fti in 0..=FRAME_NSUBTYPES {
nframes[fti] += acc[fti];
acc[fti] = 0;
}
acc[FRAME_SUBTYPE_I] += 1;
}
let mut output_frameno = self.output_frameno;
let mut ntus = 0;
let mut nframes_total = 0;
while ntus < reservoir_frame_delay {
let output_frameno_in_gop =
output_frameno - prev_keyframe_output_frameno;
let is_kf =
if let Some(frame_data) = self.frame_data.get(&output_frameno) {
if frame_data.fi.frame_type == FrameType::KEY {
prev_keyframe_input_frameno = frame_data.fi.input_frameno;
debug_assert!(frame_data.fi.show_frame);
true
} else {
false
}
} else {
output_frameno_in_gop == 0
};
if is_kf {
collect_counts(nframes, &mut acc);
prev_keyframe_output_frameno = output_frameno;
prev_keyframe_ntus = ntus;
prev_keyframe_nframes = nframes_total;
output_frameno += 1;
ntus += 1;
nframes_total += 1;
continue;
}
let idx_in_group_output =
self.inter_cfg.get_idx_in_group_output(output_frameno_in_gop);
let input_frameno = prev_keyframe_input_frameno
+ self
.inter_cfg
.get_order_hint(output_frameno_in_gop, idx_in_group_output)
as u64;
let next_keyframe_input_frameno =
self.next_keyframe_input_frameno(prev_keyframe_input_frameno, true);
if input_frameno >= next_keyframe_input_frameno {
if 1
+ (output_frameno - prev_keyframe_output_frameno)
/ self.inter_cfg.group_output_len
* self.inter_cfg.group_input_len
>= next_keyframe_input_frameno - prev_keyframe_input_frameno
{
collect_counts(nframes, &mut acc);
prev_keyframe_input_frameno = input_frameno;
prev_keyframe_output_frameno = output_frameno;
prev_keyframe_ntus = ntus;
prev_keyframe_nframes = nframes_total;
output_frameno += 1;
ntus += 1;
}
output_frameno += 1;
continue;
}
if self.inter_cfg.get_show_existing_frame(idx_in_group_output) {
acc[FRAME_SUBTYPE_SEF] += 1;
} else {
let fti = FRAME_SUBTYPE_P
+ (self.inter_cfg.get_level(idx_in_group_output) as usize);
acc[fti] += 1;
nframes_total += 1;
}
if self.inter_cfg.get_show_frame(idx_in_group_output) {
ntus += 1;
}
output_frameno += 1;
}
if prev_keyframe_output_frameno <= self.output_frameno {
collect_counts(nframes, &mut acc);
(nframes_total, ntus)
} else {
(prev_keyframe_nframes, prev_keyframe_ntus)
}
}
}