1
0
Fork 0
mirror of synced 2024-05-04 12:33:11 +12:00
czkawka/czkawka_core/src/big_file.rs
Rafał Mikrut 378fa1fd6e
Excluded extensions and krokiet new features (#1184)
* AVC

* Import split

* Default thread size

* Hen

* Allowed extensions

* Perf

* Connect

* Excluded

* Zmiany

* Optimization

* 4.10

* At once

* Included

* Chang

* VD

* VD

* Hashes

* Wersja

* SD

* Up

* Up

* 2024

* Dup

* Slint files

* Added  select

* Selections

* Fix

* LTO

* Actions

* Added popup delete

* AB

* V4

* Release

* LTO

* Basic moving

* Commonsy

* Moving probably works

* Popup move
2024-02-14 17:45:25 +01:00

186 lines
5.9 KiB
Rust

use std::fs;
use std::io::Write;
use crossbeam_channel::{Receiver, Sender};
use fun_time::fun_time;
use humansize::{format_size, BINARY};
use log::debug;
use rayon::prelude::*;
use crate::common_dir_traversal::{DirTraversalBuilder, DirTraversalResult, FileEntry, ProgressData, ToolType};
use crate::common_tool::{CommonData, CommonToolData, DeleteMethod};
use crate::common_traits::{DebugPrint, PrintResults};
#[derive(Copy, Clone, Eq, PartialEq)]
pub enum SearchMode {
BiggestFiles,
SmallestFiles,
}
#[derive(Default)]
pub struct Info {
pub number_of_real_files: usize,
}
pub struct BigFile {
common_data: CommonToolData,
information: Info,
big_files: Vec<FileEntry>,
number_of_files_to_check: usize,
search_mode: SearchMode,
}
impl BigFile {
pub fn new() -> Self {
Self {
common_data: CommonToolData::new(ToolType::BigFile),
information: Info::default(),
big_files: Default::default(),
number_of_files_to_check: 50,
search_mode: SearchMode::BiggestFiles,
}
}
#[fun_time(message = "find_big_files", level = "info")]
pub fn find_big_files(&mut self, stop_receiver: Option<&Receiver<()>>, progress_sender: Option<&Sender<ProgressData>>) {
self.prepare_items();
if !self.look_for_big_files(stop_receiver, progress_sender) {
self.common_data.stopped_search = true;
return;
}
self.delete_files();
self.debug_print();
}
// #[fun_time(message = "look_for_big_files", level = "debug")]
fn look_for_big_files(&mut self, stop_receiver: Option<&Receiver<()>>, progress_sender: Option<&Sender<ProgressData>>) -> bool {
let result = DirTraversalBuilder::new()
.group_by(|_fe| ())
.stop_receiver(stop_receiver)
.progress_sender(progress_sender)
.common_data(&self.common_data)
.minimal_file_size(1)
.maximal_file_size(u64::MAX)
.max_stage(0)
.build()
.run();
match result {
DirTraversalResult::SuccessFiles { grouped_file_entries, warnings } => {
let mut all_files = grouped_file_entries.into_values().flatten().collect::<Vec<_>>();
all_files.par_sort_unstable_by_key(|fe| fe.size);
if self.search_mode == SearchMode::BiggestFiles {
all_files.reverse();
}
if all_files.len() > self.number_of_files_to_check {
all_files.truncate(self.number_of_files_to_check);
}
self.big_files = all_files;
self.common_data.text_messages.warnings.extend(warnings);
debug!("check_files - Found {} biggest/smallest files.", self.big_files.len());
true
}
DirTraversalResult::Stopped => false,
}
}
fn delete_files(&mut self) {
match self.common_data.delete_method {
DeleteMethod::Delete => {
for file_entry in &self.big_files {
if fs::remove_file(&file_entry.path).is_err() {
self.common_data.text_messages.warnings.push(file_entry.path.to_string_lossy().to_string());
}
}
}
DeleteMethod::None => {
//Just do nothing
}
_ => unreachable!(),
}
}
}
impl Default for BigFile {
fn default() -> Self {
Self::new()
}
}
impl DebugPrint for BigFile {
fn debug_print(&self) {
if !cfg!(debug_assertions) {
return;
}
println!("### INDIVIDUAL DEBUG PRINT ###");
println!("Big files size {} in {} groups", self.information.number_of_real_files, self.big_files.len());
println!("Number of files to check - {:?}", self.number_of_files_to_check);
self.debug_print_common();
println!("-----------------------------------------");
}
}
impl PrintResults for BigFile {
fn write_results<T: Write>(&self, writer: &mut T) -> std::io::Result<()> {
writeln!(
writer,
"Results of searching {:?} with excluded directories {:?} and excluded items {:?}",
self.common_data.directories.included_directories,
self.common_data.directories.excluded_directories,
self.common_data.excluded_items.get_excluded_items()
)?;
if self.information.number_of_real_files != 0 {
if self.search_mode == SearchMode::BiggestFiles {
writeln!(writer, "{} the biggest files.\n\n", self.information.number_of_real_files)?;
} else {
writeln!(writer, "{} the smallest files.\n\n", self.information.number_of_real_files)?;
}
for file_entry in &self.big_files {
writeln!(writer, "{} ({}) - {:?}", format_size(file_entry.size, BINARY), file_entry.size, file_entry.path)?;
}
} else {
write!(writer, "Not found any files.").unwrap();
}
Ok(())
}
fn save_results_to_file_as_json(&self, file_name: &str, pretty_print: bool) -> std::io::Result<()> {
self.save_results_to_file_as_json_internal(file_name, &self.big_files, pretty_print)
}
}
impl CommonData for BigFile {
fn get_cd(&self) -> &CommonToolData {
&self.common_data
}
fn get_cd_mut(&mut self) -> &mut CommonToolData {
&mut self.common_data
}
}
impl BigFile {
pub fn set_search_mode(&mut self, search_mode: SearchMode) {
self.search_mode = search_mode;
}
pub const fn get_big_files(&self) -> &Vec<FileEntry> {
&self.big_files
}
pub const fn get_information(&self) -> &Info {
&self.information
}
pub fn set_number_of_files_to_check(&mut self, number_of_files_to_check: usize) {
self.number_of_files_to_check = number_of_files_to_check;
}
}