mod import_collector;
use std::{
collections::{HashMap, HashSet},
fs,
path::{Path, PathBuf},
process::ExitCode,
str::FromStr,
sync::atomic::{AtomicUsize, Ordering},
};
use anyhow::Result;
use bpaf::Bpaf;
use cargo_metadata::{Dependency, Metadata, MetadataCommand, Package};
use rayon::iter::{IntoParallelRefIterator, ParallelIterator};
use walkdir::{DirEntry, WalkDir};
use crate::import_collector::collect_imports;
#[derive(Debug)]
pub struct Error;
#[derive(Debug, Clone, Bpaf)]
#[bpaf(options("shear"))]
pub struct CargoShearOptions {
#[bpaf(long)]
fix: bool,
#[bpaf(positional("PATH"), fallback(PathBuf::from(".")))]
path: PathBuf,
}
pub struct CargoShear {
options: CargoShearOptions,
unused_dependencies: AtomicUsize,
}
type Deps = HashSet<String>;
impl CargoShear {
#[must_use]
pub fn new(options: CargoShearOptions) -> Self {
Self { options, unused_dependencies: AtomicUsize::default() }
}
#[must_use]
pub fn run(self) -> ExitCode {
match self.shear() {
Ok(()) => {
let has_deps = self.unused_dependencies.load(Ordering::SeqCst) > 0;
ExitCode::from(u8::from(has_deps))
}
Err(err) => {
println!("{err}");
ExitCode::from(2)
}
}
}
fn shear(&self) -> Result<()> {
let metadata = MetadataCommand::new().current_dir(&self.options.path).exec()?;
let workspace_root = metadata.workspace_root.as_std_path();
let package_dependencies = metadata
.workspace_packages()
.par_iter()
.map(|package| self.shear_package(workspace_root, package))
.collect::<Result<Vec<Deps>>>()?
.into_iter()
.fold(HashSet::new(), |a, b| a.union(&b).cloned().collect());
self.shear_workspace(&metadata, &package_dependencies)
}
fn shear_workspace(&self, metadata: &Metadata, all_pkg_deps: &Deps) -> Result<()> {
if metadata.workspace_packages().len() <= 1 {
return Ok(());
}
let metadata_path = metadata.workspace_root.as_std_path();
let cargo_toml_path = metadata_path.join("Cargo.toml");
let metadata = cargo_toml::Manifest::from_path(&cargo_toml_path)?;
let Some(workspace) = &metadata.workspace else { return Ok(()) };
let workspace_deps = workspace.dependencies.keys().cloned().collect::<HashSet<String>>();
let unused_deps = workspace_deps.difference(all_pkg_deps).cloned().collect::<Vec<_>>();
if unused_deps.is_empty() {
return Ok(());
}
println!("root: {unused_deps:?}");
self.try_fix_package(&cargo_toml_path, &unused_deps)?;
self.unused_dependencies.fetch_add(unused_deps.len(), Ordering::SeqCst);
Ok(())
}
fn shear_package(&self, workspace_root: &Path, package: &Package) -> Result<Deps> {
let dir = package
.manifest_path
.parent()
.unwrap_or_else(|| panic!("failed to get parent path {}", &package.manifest_path))
.as_std_path();
let dependency_names =
package.dependencies.iter().map(Self::dependency_name).collect::<Deps>();
let package_deps_map = dependency_names
.iter()
.map(|dep_name| {
let mod_name = dep_name.clone().replace('-', "_").to_lowercase();
(mod_name, dep_name.clone())
})
.collect::<HashMap<String, String>>();
let ignored_names = package
.metadata
.as_object()
.and_then(|object| object.get("cargo-shear").and_then(|object| object.get("ignored")))
.and_then(|ignored| {
ignored.as_array().map(|ignored| {
ignored.iter().filter_map(|item| item.as_str()).collect::<HashSet<_>>()
})
})
.unwrap_or_default();
let mod_names = package_deps_map
.keys()
.filter(|name| !ignored_names.contains(name.as_str()))
.cloned()
.collect::<HashSet<_>>();
let rust_file_deps = Self::get_package_dependencies_from_rust_files(package)?;
let unused_deps = mod_names.difference(&rust_file_deps).collect::<Vec<_>>();
if unused_deps.is_empty() {
return Ok(dependency_names);
}
let unused_dep_names =
unused_deps.into_iter().map(|name| package_deps_map[name].clone()).collect::<Vec<_>>();
self.try_fix_package(package.manifest_path.as_std_path(), &unused_dep_names)?;
let path = dir.strip_prefix(workspace_root).unwrap_or(dir);
println!("{path:?}: {unused_dep_names:?}");
self.unused_dependencies.fetch_add(unused_dep_names.len(), Ordering::SeqCst);
let dependency_names = dependency_names
.difference(&HashSet::from_iter(unused_dep_names))
.cloned()
.collect::<Deps>();
Ok(dependency_names)
}
fn get_package_dependencies_from_rust_files(package: &Package) -> Result<Deps> {
Ok(Self::get_package_rust_files(package)
.par_iter()
.map(|path| Self::process_rust_source(path))
.collect::<Result<Vec<Deps>>>()?
.into_iter()
.fold(HashSet::new(), |a, b| a.union(&b).cloned().collect()))
}
fn get_package_rust_files(package: &Package) -> Vec<PathBuf> {
package
.targets
.iter()
.flat_map(|target| {
if target.kind.iter().any(|s| s == "custom-build") {
vec![target.src_path.clone().into_std_path_buf()]
} else {
let target_dir = target.src_path.parent().unwrap_or_else(|| {
panic!("failed to get parentp path {}", &target.src_path)
});
WalkDir::new(target_dir)
.into_iter()
.filter_map(Result::ok)
.filter(|e| e.path().extension().is_some_and(|ext| ext == "rs"))
.map(DirEntry::into_path)
.collect::<Vec<_>>()
}
})
.collect()
}
fn dependency_name(dependency: &Dependency) -> String {
dependency.rename.as_ref().unwrap_or(&dependency.name).clone()
}
fn process_rust_source(path: &Path) -> Result<Deps> {
let source_text = fs::read_to_string(path)?;
let imports = collect_imports(&source_text)?;
Ok(imports)
}
fn try_fix_package(&self, cargo_toml_path: &Path, unused_dep_names: &[String]) -> Result<()> {
if !self.options.fix {
return Ok(());
}
let manifest = fs::read_to_string(cargo_toml_path)?;
let mut manifest = toml_edit::DocumentMut::from_str(&manifest)?;
if let Some(dependencies) = manifest
.get_mut("workspace")
.and_then(|item| item.as_table_mut())
.and_then(|table| table.get_mut("dependencies"))
.and_then(|item| item.as_table_mut())
{
for k in unused_dep_names {
dependencies.remove(k);
}
}
for table_name in ["dependencies", "dev-dependencies", "build-dependencies"] {
if let Some(dependencies) =
manifest.get_mut(table_name).and_then(|item| item.as_table_mut())
{
for k in unused_dep_names {
dependencies.remove(k);
}
}
}
let serialized = manifest.to_string();
fs::write(cargo_toml_path, serialized)?;
Ok(())
}
}