summaryrefslogtreecommitdiffstats
path: root/src
diff options
context:
space:
mode:
authorClementTsang <clementjhtsang@gmail.com>2020-01-28 21:24:52 -0500
committerClementTsang <clementjhtsang@gmail.com>2020-01-28 21:24:52 -0500
commitabe8a4bb861988daf67e37803b05d94507eddbda (patch)
treebf95c88c1bb7255e5156836fa4b5eac6ffcc78d8 /src
parent57aa15978ce2678c196f8ee93d9d45b9ff6f568d (diff)
Moved temp and drives over...
Diffstat (limited to 'src')
-rw-r--r--src/app.rs36
-rw-r--r--src/app/data_farmer.rs (renamed from src/app/data_janitor.rs)94
-rw-r--r--src/app/data_harvester.rs65
-rw-r--r--src/app/data_harvester/cpu.rs7
-rw-r--r--src/app/data_harvester/disks.rs53
-rw-r--r--src/app/data_harvester/processes.rs16
-rw-r--r--src/app/data_harvester/temperature.rs18
-rw-r--r--src/canvas.rs10
-rw-r--r--src/constants.rs2
-rw-r--r--src/data_conversion.rs98
-rw-r--r--src/main.rs15
11 files changed, 217 insertions, 197 deletions
diff --git a/src/app.rs b/src/app.rs
index 3e672f6c..5b398c93 100644
--- a/src/app.rs
+++ b/src/app.rs
@@ -2,10 +2,10 @@ pub mod data_harvester;
use data_harvester::{processes, temperature};
use std::time::Instant;
-pub mod data_janitor;
-use data_janitor::*;
+pub mod data_farmer;
+use data_farmer::*;
-use crate::{canvas, constants, data_conversion::ConvertedProcessData, utils::error::Result};
+use crate::{canvas, constants, data_conversion::ConvertedProcessHarvest, utils::error::Result};
mod process_killer;
@@ -76,7 +76,7 @@ pub struct App {
pub show_help: bool,
pub show_dd: bool,
pub dd_err: Option<String>,
- to_delete_process_list: Option<Vec<ConvertedProcessData>>,
+ to_delete_process_list: Option<Vec<ConvertedProcessHarvest>>,
pub is_frozen: bool,
pub left_legend: bool,
pub use_current_cpu_total: bool,
@@ -196,19 +196,13 @@ impl App {
self.enable_grouping
}
- pub fn toggle_searching(&mut self) {
+ pub fn enable_searching(&mut self) {
if !self.is_in_dialog() {
match self.current_application_position {
ApplicationPosition::Process | ApplicationPosition::ProcessSearch => {
- if self.enable_searching {
- // Toggle off
- self.enable_searching = false;
- self.current_application_position = ApplicationPosition::Process;
- } else {
- // Toggle on
- self.enable_searching = true;
- self.current_application_position = ApplicationPosition::ProcessSearch;
- }
+ // Toggle on
+ self.enable_searching = true;
+ self.current_application_position = ApplicationPosition::ProcessSearch;
}
_ => {}
}
@@ -397,7 +391,7 @@ impl App {
} else {
match caught_char {
'/' => {
- self.toggle_searching();
+ self.enable_searching();
}
'd' => {
if let ApplicationPosition::Process = self.current_application_position {
@@ -405,7 +399,7 @@ impl App {
self.awaiting_second_char = false;
self.second_char = ' ';
let current_process = if self.is_grouped() {
- let mut res: Vec<ConvertedProcessData> = Vec::new();
+ let mut res: Vec<ConvertedProcessHarvest> = Vec::new();
for pid in &self.canvas_data.grouped_process_data
[self.currently_selected_process_position as usize]
.group
@@ -530,7 +524,7 @@ impl App {
Ok(())
}
- pub fn get_current_highlighted_process_list(&self) -> Option<Vec<ConvertedProcessData>> {
+ pub fn get_current_highlighted_process_list(&self) -> Option<Vec<ConvertedProcessHarvest>> {
self.to_delete_process_list.clone()
}
@@ -633,10 +627,10 @@ impl App {
}
ApplicationPosition::Temp => {
self.currently_selected_temperature_position =
- self.data.list_of_temperature_sensor.len() as i64 - 1
+ self.data.temperature_sensors.len() as i64 - 1
}
ApplicationPosition::Disk => {
- self.currently_selected_disk_position = self.data.list_of_disks.len() as i64 - 1
+ self.currently_selected_disk_position = self.data.disks.len() as i64 - 1
}
ApplicationPosition::Cpu => {
self.currently_selected_cpu_table_position =
@@ -698,7 +692,7 @@ impl App {
fn change_temp_position(&mut self, num_to_change_by: i64) {
if self.currently_selected_temperature_position + num_to_change_by >= 0
&& self.currently_selected_temperature_position + num_to_change_by
- < self.data.list_of_temperature_sensor.len() as i64
+ < self.data.temperature_sensors.len() as i64
{
self.currently_selected_temperature_position += num_to_change_by;
}
@@ -707,7 +701,7 @@ impl App {
fn change_disk_position(&mut self, num_to_change_by: i64) {
if self.currently_selected_disk_position + num_to_change_by >= 0
&& self.currently_selected_disk_position + num_to_change_by
- < self.data.list_of_disks.len() as i64
+ < self.data.disks.len() as i64
{
self.currently_selected_disk_position += num_to_change_by;
}
diff --git a/src/app/data_janitor.rs b/src/app/data_farmer.rs
index e6e5f336..0acb7343 100644
--- a/src/app/data_janitor.rs
+++ b/src/app/data_farmer.rs
@@ -1,6 +1,18 @@
-use crate::data_harvester::{cpu, mem, network, Data};
-/// In charge of cleaning and managing data. I couldn't think of a better
-/// name for the file.
+use crate::data_harvester::{cpu, disks, mem, network, processes, temperature, Data};
+/// In charge of cleaning, processing, and managing data. I couldn't think of
+/// a better name for the file. Since I called data collection "harvesting",
+/// then this is the farmer I guess.
+///
+/// Essentially the main goal is to shift the initial calculation and distribution
+/// of joiner points and data to one central location that will only do it
+/// *once* upon receiving the data --- as opposed to doing it on canvas draw,
+/// which will be a costly process.
+///
+/// This will also handle the *cleaning* of stale data. That should be done
+/// in some manner (timer on another thread, some loop) that will occasionally
+/// call the purging function. Failure to do so *will* result in a growing
+/// memory usage and higher CPU usage - you will be trying to process more and
+/// more points as this is used!
use std::time::Instant;
use std::vec::Vec;
@@ -15,8 +27,9 @@ pub struct TimedData {
pub cpu_data: Vec<JoinedDataPoints>,
pub mem_data: JoinedDataPoints,
pub swap_data: JoinedDataPoints,
- pub temp_data: JoinedDataPoints,
- pub io_data: JoinedDataPoints,
+ // Unused for now
+ // pub io_data : JoinedDataPoints
+ // pub temp_data: JoinedDataPoints,
}
/// AppCollection represents the pooled data stored within the main app
@@ -36,6 +49,12 @@ pub struct DataCollection {
pub memory_harvest: mem::MemHarvest,
pub swap_harvest: mem::MemHarvest,
pub cpu_harvest: cpu::CPUHarvest,
+ pub process_harvest: processes::ProcessHarvest,
+ pub disk_harvest: Vec<disks::DiskHarvest>,
+ pub io_harvest: disks::IOHarvest,
+ pub io_labels: Vec<(u64, u64)>,
+ io_prev: Vec<(u64, u64)>,
+ pub temp_harvest: Vec<temperature::TempHarvest>,
}
impl Default for DataCollection {
@@ -47,12 +66,20 @@ impl Default for DataCollection {
memory_harvest: mem::MemHarvest::default(),
swap_harvest: mem::MemHarvest::default(),
cpu_harvest: cpu::CPUHarvest::default(),
+ process_harvest: processes::ProcessHarvest::default(),
+ disk_harvest: Vec::default(),
+ io_harvest: disks::IOHarvest::default(),
+ io_labels: Vec::default(),
+ io_prev: Vec::default(),
+ temp_harvest: Vec::default(),
}
}
}
impl DataCollection {
- pub fn clean_data(&mut self) {}
+ pub fn clean_data(&mut self) {
+ // TODO: [OPT] To implement to clean
+ }
pub fn eat_data(&mut self, harvested_data: &Data) {
let harvested_time = harvested_data.last_collection_time;
@@ -67,6 +94,14 @@ impl DataCollection {
// CPU
self.eat_cpu(&harvested_data, &harvested_time, &mut new_entry);
+ // Temp
+ self.eat_temp(&harvested_data, &harvested_time, &mut new_entry);
+
+ // Disks
+ self.eat_disks(&harvested_data, &harvested_time, &mut new_entry);
+
+ // Processes
+
// And we're done eating.
self.current_instant = harvested_time;
self.timed_data_vec.push((harvested_time, new_entry));
@@ -147,7 +182,7 @@ impl DataCollection {
// Note this only pre-calculates the data points - the names will be
// within the local copy of cpu_harvest. Since it's all sequential
// it probably doesn't matter anyways.
- for (itx, cpu) in harvested_data.cpu.cpu_vec.iter().enumerate() {
+ for (itx, cpu) in harvested_data.cpu.iter().enumerate() {
let cpu_joining_pts = if let Some((time, last_pt)) = self.timed_data_vec.last() {
generate_joining_points(
&time,
@@ -165,6 +200,51 @@ impl DataCollection {
self.cpu_harvest = harvested_data.cpu.clone();
}
+
+ fn eat_temp(
+ &mut self, harvested_data: &Data, _harvested_time: &Instant, _new_entry: &mut TimedData,
+ ) {
+ // TODO: [PO] To implement
+ self.temp_harvest = harvested_data.temperature_sensors.clone();
+ }
+
+ fn eat_disks(
+ &mut self, harvested_data: &Data, harvested_time: &Instant, _new_entry: &mut TimedData,
+ ) {
+ // TODO: [PO] To implement
+
+ let time_since_last_harvest = harvested_time
+ .duration_since(self.current_instant)
+ .as_secs_f64();
+
+ for (itx, device) in harvested_data.disks.iter().enumerate() {
+ if let Some(trim) = device.name.split('/').last() {
+ let io_device = harvested_data.io.get(trim);
+ if let Some(io) = io_device {
+ let io_r_pt = io.read_bytes;
+ let io_w_pt = io.write_bytes;
+
+ if self.io_labels.len() <= itx {
+ self.io_prev.push((io_r_pt, io_w_pt));
+ self.io_labels.push((0, 0));
+ } else {
+ let r_rate = ((io_r_pt - self.io_prev[itx].0) as f64
+ / time_since_last_harvest)
+ .round() as u64;
+ let w_rate = ((io_w_pt - self.io_prev[itx].1) as f64
+ / time_since_last_harvest)
+ .round() as u64;
+
+ self.io_labels[itx] = (r_rate, w_rate);
+ self.io_prev[itx] = (io_r_pt, io_w_pt);
+ }
+ }
+ }
+ }
+
+ self.disk_harvest = harvested_data.disks.clone();
+ self.io_harvest = harvested_data.io.clone();
+ }
}
pub fn generate_joining_points(
diff --git a/src/app/data_harvester.rs b/src/app/data_harvester.rs
index 4127f09f..fc468f7e 100644
--- a/src/app/data_harvester.rs
+++ b/src/app/data_harvester.rs
@@ -17,23 +17,17 @@ fn set_if_valid<T: std::clone::Clone>(result: &Result<T>, value_to_set: &mut T)
}
}
-fn push_if_valid<T: std::clone::Clone>(result: &Result<T>, vector_to_push: &mut Vec<T>) {
- if let Ok(result) = result {
- vector_to_push.push(result.clone());
- }
-}
-
#[derive(Clone, Debug)]
pub struct Data {
pub cpu: cpu::CPUHarvest,
- pub list_of_io: Vec<disks::IOPackage>,
pub memory: mem::MemHarvest,
pub swap: mem::MemHarvest,
- pub list_of_temperature_sensor: Vec<temperature::TempData>,
+ pub temperature_sensors: Vec<temperature::TempHarvest>,
pub network: network::NetworkHarvest,
- pub list_of_processes: Vec<processes::ProcessData>,
- pub grouped_list_of_processes: Option<Vec<processes::ProcessData>>,
- pub list_of_disks: Vec<disks::DiskData>,
+ pub list_of_processes: Vec<processes::ProcessHarvest>,
+ pub grouped_list_of_processes: Option<Vec<processes::ProcessHarvest>>,
+ pub disks: Vec<disks::DiskHarvest>,
+ pub io: disks::IOHarvest,
pub last_collection_time: Instant,
}
@@ -41,13 +35,13 @@ impl Default for Data {
fn default() -> Self {
Data {
cpu: cpu::CPUHarvest::default(),
- list_of_io: Vec::default(),
memory: mem::MemHarvest::default(),
swap: mem::MemHarvest::default(),
- list_of_temperature_sensor: Vec::default(),
+ temperature_sensors: Vec::default(),
list_of_processes: Vec::default(),
grouped_list_of_processes: None,
- list_of_disks: Vec::default(),
+ disks: Vec::default(),
+ io: disks::IOHarvest::default(),
network: network::NetworkHarvest::default(),
last_collection_time: Instant::now(),
}
@@ -56,11 +50,11 @@ impl Default for Data {
impl Data {
pub fn first_run_cleanup(&mut self) {
- self.list_of_io = Vec::new();
- self.list_of_temperature_sensor = Vec::new();
+ self.io = disks::IOHarvest::default();
+ self.temperature_sensors = Vec::new();
self.list_of_processes = Vec::new();
self.grouped_list_of_processes = None;
- self.list_of_disks = Vec::new();
+ self.disks = Vec::new();
self.network.first_run_cleanup();
self.memory = mem::MemHarvest::default();
@@ -149,6 +143,20 @@ impl DataState {
// CPU
self.data.cpu = cpu::get_cpu_data_list(&self.sys);
+ // Disks
+ if let Ok(disks) = disks::get_disk_usage_list().await {
+ self.data.disks = disks;
+ }
+ if let Ok(io) = disks::get_io_usage_list(false).await {
+ self.data.io = io;
+ }
+
+ // Temp
+ if let Ok(temp) = temperature::get_temperature_data(&self.sys, &self.temperature_type).await
+ {
+ self.data.temperature_sensors = temp;
+ }
+
// What we want to do: For timed data, if there is an error, just do not add. For other data, just don't update!
set_if_valid(
&processes::get_sorted_processes_list(
@@ -163,19 +171,7 @@ impl DataState {
&mut self.data.list_of_processes,
);
- set_if_valid(
- &disks::get_disk_usage_list().await,
- &mut self.data.list_of_disks,
- );
- push_if_valid(
- &disks::get_io_usage_list(false).await,
- &mut self.data.list_of_io,
- );
- set_if_valid(
- &temperature::get_temperature_data(&self.sys, &self.temperature_type).await,
- &mut self.data.list_of_temperature_sensor,
- );
-
+ // Update time
self.data.last_collection_time = current_instant;
// Filter out stale timed entries
@@ -192,15 +188,6 @@ impl DataState {
for stale in stale_list {
self.prev_pid_stats.remove(&stale);
}
- self.data.list_of_io = self
- .data
- .list_of_io
- .iter()
- .cloned()
- .filter(|entry| {
- clean_instant.duration_since(entry.instant).as_secs() <= self.stale_max_seconds
- })
- .collect::<Vec<_>>();
self.last_clean = clean_instant;
}
diff --git a/src/app/data_harvester/cpu.rs b/src/app/data_harvester/cpu.rs
index c3276827..136a36dd 100644
--- a/src/app/data_harvester/cpu.rs
+++ b/src/app/data_harvester/cpu.rs
@@ -6,10 +6,7 @@ pub struct CPUData {
pub cpu_usage: f64,
}
-#[derive(Default, Debug, Clone)]
-pub struct CPUHarvest {
- pub cpu_vec: Vec<CPUData>,
-}
+pub type CPUHarvest = Vec<CPUData>;
pub fn get_cpu_data_list(sys: &System) -> CPUHarvest {
let cpu_data = sys.get_processor_list();
@@ -22,5 +19,5 @@ pub fn get_cpu_data_list(sys: &System) -> CPUHarvest {
});
}
- CPUHarvest { cpu_vec }
+ cpu_vec
}
diff --git a/src/app/data_harvester/disks.rs b/src/app/data_harvester/disks.rs
index bf564ab2..7a7f52e2 100644
--- a/src/app/data_harvester/disks.rs
+++ b/src/app/data_harvester/disks.rs
@@ -1,11 +1,10 @@
use futures::stream::StreamExt;
use heim::units::information;
-use std::time::Instant;
#[derive(Debug, Clone, Default)]
-pub struct DiskData {
- pub name: Box<str>,
- pub mount_point: Box<str>,
+pub struct DiskHarvest {
+ pub name: String,
+ pub mount_point: String,
pub free_space: u64,
pub used_space: u64,
pub total_space: u64,
@@ -13,18 +12,13 @@ pub struct DiskData {
#[derive(Clone, Debug)]
pub struct IOData {
- pub mount_point: Box<str>,
pub read_bytes: u64,
pub write_bytes: u64,
}
-#[derive(Debug, Clone)]
-pub struct IOPackage {
- pub io_hash: std::collections::HashMap<String, IOData>,
- pub instant: Instant,
-}
+pub type IOHarvest = std::collections::HashMap<String, IOData>;
-pub async fn get_io_usage_list(get_physical: bool) -> crate::utils::error::Result<IOPackage> {
+pub async fn get_io_usage_list(get_physical: bool) -> crate::utils::error::Result<IOHarvest> {
let mut io_hash: std::collections::HashMap<String, IOData> = std::collections::HashMap::new();
if get_physical {
let mut physical_counter_stream = heim::disk::io_counters_physical();
@@ -34,7 +28,6 @@ pub async fn get_io_usage_list(get_physical: bool) -> crate::utils::error::Resul
io_hash.insert(
mount_point.to_string(),
IOData {
- mount_point: Box::from(mount_point),
read_bytes: io.read_bytes().get::<information::megabyte>(),
write_bytes: io.write_bytes().get::<information::megabyte>(),
},
@@ -48,7 +41,6 @@ pub async fn get_io_usage_list(get_physical: bool) -> crate::utils::error::Resul
io_hash.insert(
mount_point.to_string(),
IOData {
- mount_point: Box::from(mount_point),
read_bytes: io.read_bytes().get::<information::byte>(),
write_bytes: io.write_bytes().get::<information::byte>(),
},
@@ -56,14 +48,11 @@ pub async fn get_io_usage_list(get_physical: bool) -> crate::utils::error::Resul
}
}
- Ok(IOPackage {
- io_hash,
- instant: Instant::now(),
- })
+ Ok(io_hash)
}
-pub async fn get_disk_usage_list() -> crate::utils::error::Result<Vec<DiskData>> {
- let mut vec_disks: Vec<DiskData> = Vec::new();
+pub async fn get_disk_usage_list() -> crate::utils::error::Result<Vec<DiskHarvest>> {
+ let mut vec_disks: Vec<DiskHarvest> = Vec::new();
let mut partitions_stream = heim::disk::partitions_physical();
while let Some(part) = partitions_stream.next().await {
@@ -71,23 +60,21 @@ pub async fn get_disk_usage_list() -> crate::utils::error::Result<Vec<DiskData>>
let partition = part;
let usage = heim::disk::usage(partition.mount_point().to_path_buf()).await?;
- vec_disks.push(DiskData {
+ vec_disks.push(DiskHarvest {
free_space: usage.free().get::<information::byte>(),
used_space: usage.used().get::<information::byte>(),
total_space: usage.total().get::<information::byte>(),
- mount_point: Box::from(
- partition
- .mount_point()
- .to_str()
- .unwrap_or("Name Unavailable"),
- ),
- name: Box::from(
- partition
- .device()
- .unwrap_or_else(|| std::ffi::OsStr::new("Name Unavailable"))
- .to_str()
- .unwrap_or("Name Unavailable"),
- ),
+ mount_point: (partition
+ .mount_point()
+ .to_str()
+ .unwrap_or("Name Unavailable"))
+ .to_string(),
+ name: (partition
+ .device()
+ .unwrap_or_else(|| std::ffi::OsStr::new("Name Unavailable"))
+ .to_str()
+ .unwrap_or("Name Unavailable"))
+ .to_string(),
});
}
}
diff --git a/src/app/data_harvester/processes.rs b/src/app/data_harvester/processes.rs
index 25d3a6d4..c72c6d01 100644
--- a/src/app/data_harvester/processes.rs
+++ b/src/app/data_harvester/processes.rs
@@ -18,7 +18,7 @@ impl Default for ProcessSorting {
}
#[derive(Debug, Clone, Default)]
-pub struct ProcessData {
+pub struct ProcessHarvest {
pub pid: u32,
pub cpu_usage_percent: f64,
pub mem_usage_percent: f64,
@@ -182,9 +182,9 @@ fn convert_ps(
process: &str, cpu_usage: f64, cpu_percentage: f64,
prev_pid_stats: &mut HashMap<String, (f64, Instant)>, use_current_cpu_total: bool,
curr_time: &Instant,
-) -> std::io::Result<ProcessData> {
+) -> std::io::Result<ProcessHarvest> {
if process.trim().to_string().is_empty() {
- return Ok(ProcessData {
+ return Ok(ProcessHarvest {
pid: 0,
name: "".to_string(),
mem_usage_percent: 0.0,
@@ -205,7 +205,7 @@ fn convert_ps(
.parse::<f64>()
.unwrap_or(0_f64);
- Ok(ProcessData {
+ Ok(ProcessHarvest {
pid,
name,
mem_usage_percent,
@@ -225,8 +225,8 @@ pub fn get_sorted_processes_list(
sys: &System, prev_idle: &mut f64, prev_non_idle: &mut f64,
prev_pid_stats: &mut std::collections::HashMap<String, (f64, Instant)>,
use_current_cpu_total: bool, mem_total_kb: u64, curr_time: &Instant,
-) -> crate::utils::error::Result<Vec<ProcessData>> {
- let mut process_vector: Vec<ProcessData> = Vec::new();
+) -> crate::utils::error::Result<Vec<ProcessHarvest>> {
+ let mut process_vector: Vec<ProcessHarvest> = Vec::new();
if cfg!(target_os = "linux") {
// Linux specific - this is a massive pain... ugh.
@@ -282,7 +282,7 @@ pub fn get_sorted_processes_list(
process_val.name().to_string()
};
- process_vector.push(ProcessData {
+ process_vector.push(ProcessHarvest {
pid: process_val.pid() as u32,
name,
mem_usage_percent: process_val.memory() as f64 * 100.0 / mem_total_kb as f64,
@@ -296,7 +296,7 @@ pub fn get_sorted_processes_list(
}
pub fn sort_processes(
- process_vector: &mut Vec<ProcessData>, sorting_method: &ProcessSorting, reverse_order: bool,
+ process_vector: &mut Vec<ProcessHarvest>, sorting_method: &ProcessSorting, reverse_order: bool,
) {
// Always sort alphabetically first!
process_vector.sort_by(|a, b| get_ordering(&a.name, &b.name, false));
diff --git a/src/app/data_harvester/temperature.rs b/src/app/data_harvester/temperature.rs
index 1bee113e..77ffa4d6 100644
--- a/src/app/data_harvester/temperature.rs
+++ b/src/app/data_harvester/temperature.rs
@@ -3,9 +3,9 @@ use heim::units::thermodynamic_temperature;
use std::cmp::Ordering;
use sysinfo::{ComponentExt, System, SystemExt};
-#[derive(Debug, Clone)]
-pub struct TempData {
- pub component_name: Box<str>,
+#[derive(Default, Debug, Clone)]
+pub struct TempHarvest {
+ pub component_name: String,
pub temperature: f32,
}
@@ -24,15 +24,15 @@ impl Default for TemperatureType {
pub async fn get_temperature_data(
sys: &System, temp_type: &TemperatureType,
-) -> crate::utils::error::Result<Vec<TempData>> {
- let mut temperature_vec: Vec<TempData> = Vec::new();
+) -> crate::utils::error::Result<Vec<TempHarvest>> {
+ let mut temperature_vec: Vec<TempHarvest> = Vec::new();
if cfg!(target_os = "linux") {
let mut sensor_data = heim::sensors::temperatures();
while let Some(sensor) = sensor_data.next().await {
if let Ok(sensor) = sensor {
- temperature_vec.push(TempData {
- component_name: Box::from(sensor.unit()),
+ temperature_vec.push(TempHarvest {
+ component_name: sensor.unit().to_string(),
temperature: match temp_type {
TemperatureType::Celsius => sensor
.current()
@@ -52,8 +52,8 @@ pub async fn get_temperature_data(
} else {
let sensor_data = sys.get_components_list();
for component in sensor_data {
- temperature_vec.push(TempData {
- component_name: Box::from(component.get_label()),
+ temperature_vec.push(TempHarvest {
+ component_name: component.get_label().to_string(),
temperature: match temp_type {
TemperatureType::Celsius => component.get_temperature(),
TemperatureType::Kelvin => component.get_temperature() + 273.15,
diff --git a/src/canvas.rs b/src/canvas.rs
index 1cb7e4b6..2c3ce977 100644
--- a/src/canvas.rs
+++ b/src/canvas.rs
@@ -1,6 +1,6 @@
use crate::{
app, constants,
- data_conversion::{ConvertedCpuData, ConvertedProcessData},
+ data_conversion::{ConvertedCpuData, ConvertedProcessHarvest},
utils::{error, gen_util::*},
};
use std::cmp::max;
@@ -94,8 +94,8 @@ pub struct CanvasData {
pub network_data_tx: Vec<(f64, f64)>,
pub disk_data: Vec<Vec<String>>,
pub temp_sensor_data: Vec<Vec<String>>,
- pub process_data: Vec<ConvertedProcessData>,
- pub grouped_process_data: Vec<ConvertedProcessData>,
+ pub process_data: Vec<ConvertedProcessHarvest>,
+ pub grouped_process_data: Vec<ConvertedProcessHarvest>,
pub mem_label: String,
pub swap_label: String,
pub mem_data: Vec<(f64, f64)>,
@@ -951,7 +951,7 @@ fn draw_search_field<B: backend::Backend>(
fn draw_processes_table<B: backend::Backend>(
f: &mut Frame<B>, app_state: &mut app::App, draw_loc: Rect,
) {
- let process_data: &[ConvertedProcessData] = if app_state.is_grouped() {
+ let process_data: &[ConvertedProcessHarvest] = if app_state.is_grouped() {
&app_state.canvas_data.grouped_process_data
} else {
&app_state.canvas_data.process_data
@@ -971,7 +971,7 @@ fn draw_processes_table<B: backend::Backend>(
app_state.currently_selected_process_position,
);
- let sliced_vec: Vec<ConvertedProcessData> = (&process_data[start_position as usize..]).to_vec();
+ let sliced_vec: Vec<ConvertedProcessHarvest> = (&process_data[start_position as usize..]).to_vec();
let mut process_counter = 0;
// Draw!
diff --git a/src/constants.rs b/src/constants.rs
index 1dad85a9..1b93ea1b 100644
--- a/src/constants.rs
+++ b/src/constants.rs
@@ -1,7 +1,7 @@
// TODO: Store like three minutes of data, then change how much is shown based on scaling!
pub const STALE_MAX_MILLISECONDS: u64 = 180 * 1000; // We wish to store at most 60 seconds worth of data. This may change in the future, or be configurable.
pub const TIME_STARTS_FROM: u64 = 60 * 1000;
-pub const TICK_RATE_IN_MILLISECONDS: u64 = 200; // We use this as it's a good value to work with.
+pub const TICK_RATE_IN_MILLISECONDS: u64 = 200; // How fast the screen refreshes
pub const DEFAULT_REFRESH_RATE_IN_MILLISECONDS: u128 = 1000;
pub const MAX_KEY_TIMEOUT_IN_MILLISECONDS: u128 = 1000;
pub const NUM_COLOURS: i32 = 256;
diff --git a/src/data_conversion.rs b/src/data_conversion.rs
index 1b5dd0ee..a01842f2 100644
--- a/src/data_conversion.rs
+++ b/src/data_conversion.rs
@@ -2,8 +2,9 @@
//! can actually handle.
use crate::{
+ app::data_farmer,
app::data_harvester,
- app::data_janitor,
+ app::App,
constants,
utils::gen_util::{get_exact_byte_values, get_simple_byte_values},
};
@@ -21,7 +22,7 @@ pub struct ConvertedNetworkData {
}
#[derive(Clone, Default, Debug)]
-pub struct ConvertedProcessData {
+pub struct ConvertedProcessHarvest {
pub pid: u32,
pub name: String,
pub cpu_usage: String,
@@ -55,15 +56,16 @@ impl From<&CpuPoint> for (f64, f64) {
}
}
-pub fn update_temp_row(
- app_data: &data_harvester::Data, temp_type: &data_harvester::temperature::TemperatureType,
-) -> Vec<Vec<String>> {
+pub fn update_temp_row(app: &App) -> Vec<Vec<String>> {
let mut sensor_vector: Vec<Vec<String>> = Vec::new();
- if (&app_data.list_of_temperature_sensor).is_empty() {
+ let current_data = &app.data_collection;
+ let temp_type = &app.temperature_type;
+
+ if current_data.temp_harvest.is_empty() {
sensor_vector.push(vec!["No Sensors Found".to_string(), "".to_string()])
} else {
- for sensor in &app_data.list_of_temperature_sensor {
+ for sensor in &current_data.temp_harvest {
sensor_vector.push(vec![
sensor.component_name.to_string(),
(sensor.temperature.ceil() as u64).to_string()
@@ -79,44 +81,18 @@ pub fn update_temp_row(
sensor_vector
}
-pub fn update_disk_row(app_data: &data_harvester::Data) -> Vec<Vec<String>> {
+pub fn update_disk_row(current_data: &data_farmer::DataCollection) -> Vec<Vec<String>> {
let mut disk_vector: Vec<Vec<String>> = Vec::new();
- for disk in &app_data.list_of_disks {
- let io_activity = {
- let mut final_result = ("0B/s".to_string(), "0B/s".to_string());
- if app_data.list_of_io.len() > 2 {
- if let Some(io_package) = &app_data.list_of_io.last() {
- if let Some(trimmed_mount) = disk.name.to_string().split('/').last() {
- let prev_io_package = &app_data.list_of_io[app_data.list_of_io.len() - 2];
-
- let io_hashmap = &io_package.io_hash;
- let prev_io_hashmap = &prev_io_package.io_hash;
- let time_difference = io_package
- .instant
- .duration_since(prev_io_package.instant)
- .as_secs_f64();
- if io_hashmap.contains_key(trimmed_mount)
- && prev_io_hashmap.contains_key(trimmed_mount)
- {
- // Ideally change this...
- let ele = &io_hashmap[trimmed_mount];
- let prev = &prev_io_hashmap[trimmed_mount];
- let read_bytes_per_sec = ((ele.read_bytes - prev.read_bytes) as f64
- / time_difference) as u64;
- let write_bytes_per_sec = ((ele.write_bytes - prev.write_bytes) as f64
- / time_difference) as u64;
- let converted_read = get_simple_byte_values(read_bytes_per_sec, false);
- let converted_write =
- get_simple_byte_values(write_bytes_per_sec, false);
- final_result = (
- format!("{:.*}{}/s", 0, converted_read.0, converted_read.1),
- format!("{:.*}{}/s", 0, converted_write.0, converted_write.1),
- );
- }
- }
- }
- }
- final_result
+ for (itx, disk) in current_data.disk_harvest.iter().enumerate() {
+ let io_activity = if current_data.io_labels.len() > itx {
+ let converted_read = get_simple_byte_values(current_data.io_labels[itx].0, false);
+ let converted_write = get_simple_byte_values(current_data.io_labels[itx].1, false);
+ (
+ format!("{:.*}{}/s", 0, converted_read.0, converted_read.1),
+ format!("{:.*}{}/s", 0, converted_write.0, converted_write.1),
+ )
+ } else {
+ ("0B/s".to_string(), "0B/s".to_string())
};
let converted_free_space = get_simple_byte_values(disk.free_space, false);
@@ -143,8 +119,8 @@ pub fn update_disk_row(app_data: &data_harvester::Data) -> Vec<Vec<String>> {
pub fn simple_update_process_row(
app_data: &data_harvester::Data, matching_string: &str, use_pid: bool,
-) -> (Vec<ConvertedProcessData>, Vec<ConvertedProcessData>) {
- let process_vector: Vec<ConvertedProcessData> = app_data
+) -> (Vec<ConvertedProcessHarvest>, Vec<ConvertedProcessHarvest>) {
+ let process_vector: Vec<ConvertedProcessHarvest> = app_data
.list_of_processes
.iter()
.filter(|process| {
@@ -161,7 +137,7 @@ pub fn simple_update_process_row(
.map(|process| return_mapped_process(process))
.collect::<Vec<_>>();
- let mut grouped_process_vector: Vec<ConvertedProcessData> = Vec::new();
+ let mut grouped_process_vector: Vec<ConvertedProcessHarvest> = Vec::new();
if let Some(grouped_list_of_processes) = &app_data.grouped_list_of_processes {
grouped_process_vector = grouped_list_of_processes
.iter()
@@ -186,8 +162,8 @@ pub fn simple_update_process_row(
pub fn regex_update_process_row(
app_data: &data_harvester::Data, regex_matcher: &std::result::Result<Regex, regex::Error>,
use_pid: bool,
-) -> (Vec<ConvertedProcessData>, Vec<ConvertedProcessData>) {
- let process_vector: Vec<ConvertedProcessData> = app_data
+) -> (Vec<ConvertedProcessHarvest>, Vec<ConvertedProcessHarvest>) {
+ let process_vector: Vec<ConvertedProcessHarvest> = app_data
.list_of_processes
.iter()
.filter(|process| {
@@ -204,7 +180,7 @@ pub fn regex_update_process_row(
.map(|process| return_mapped_process(process))
.collect::<Vec<_>>();
- let mut grouped_process_vector: Vec<ConvertedProcessData> = Vec::new();
+ let mut grouped_process_vector: Vec<ConvertedProcessHarvest> = Vec::new();
if let Some(grouped_list_of_processes) = &app_data.grouped_list_of_processes {
grouped_process_vector = grouped_list_of_processes
.iter()
@@ -226,8 +202,10 @@ pub fn regex_update_process_row(
(process_vector, grouped_process_vector)
}
-fn return_mapped_process(process: &data_harvester::processes::ProcessData) -> ConvertedProcessData {
- ConvertedProcessData {
+fn return_mapped_process(
+ process: &data_harvester::processes::ProcessHarvest,
+) -> ConvertedProcessHarvest {
+ ConvertedProcessHarvest {
pid: process.pid,
name: process.name.to_string(),
cpu_usage: format!("{:.1}%", process.cpu_usage_percent),
@@ -237,7 +215,7 @@ fn return_mapped_process(process: &data_harvester::processes::ProcessData) -> Co
}
pub fn update_cpu_data_points(
- show_avg_cpu: bool, current_data: &data_janitor::DataCollection,
+ show_avg_cpu: bool, current_data: &data_farmer::DataCollection,
) -> Vec<ConvertedCpuData> {
let mut cpu_data_vector: Vec<ConvertedCpuData> = Vec::new();
let current_time = current_data.current_instant;
@@ -260,9 +238,7 @@ pub fn update_cpu_data_points(
cpu_data_vector[itx_offset].cpu_name = if show_avg_cpu && itx_offset == 0 {
"AVG".to_string()
} else {
- current_data.cpu_harvest.cpu_vec[itx]
- .cpu_name
- .to_uppercase()
+ current_data.cpu_harvest[itx].cpu_name.to_uppercase()
};
}
@@ -285,7 +261,7 @@ pub fn update_cpu_data_points(
cpu_data_vector
}
-pub fn update_mem_data_points(current_data: &data_janitor::DataCollection) -> Vec<(f64, f64)> {
+pub fn update_mem_data_points(current_data: &data_farmer::DataCollection) -> Vec<(f64, f64)> {
let mut result: Vec<(f64, f64)> = Vec::new();
let current_time = current_data.current_instant;