mirror of
https://github.com/LibreQoE/LibreQoS.git
synced 2025-02-25 18:55:32 -06:00
RAM and CPU charts
This commit is contained in:
parent
2e2c290b1a
commit
3caa007fe9
@ -0,0 +1,33 @@
|
||||
import {BaseDashlet} from "./base_dashlet";
|
||||
import {CpuHistogram} from "../graphs/cpu_graph";
|
||||
|
||||
export class CpuDash extends BaseDashlet{
|
||||
constructor(slot) {
|
||||
super(slot);
|
||||
}
|
||||
|
||||
title() {
|
||||
return "CPU Utilization";
|
||||
}
|
||||
|
||||
subscribeTo() {
|
||||
return [ "Cpu" ];
|
||||
}
|
||||
|
||||
buildContainer() {
|
||||
let base = super.buildContainer();
|
||||
base.appendChild(this.graphDiv());
|
||||
return base;
|
||||
}
|
||||
|
||||
setup() {
|
||||
super.setup();
|
||||
this.graph = new CpuHistogram(this.graphDivId());
|
||||
}
|
||||
|
||||
onMessage(msg) {
|
||||
if (msg.event === "Cpu") {
|
||||
this.graph.update(msg.data);
|
||||
}
|
||||
}
|
||||
}
|
@ -12,6 +12,8 @@ import {Top10EndpointsByCountry} from "./endpoints_by_country";
|
||||
import {EtherProtocols} from "./ether_protocols";
|
||||
import {IpProtocols} from "./ip_protocols";
|
||||
import {Worst10Retransmits} from "./worst10_retransmits";
|
||||
import {CpuDash} from "./cpu_dash";
|
||||
import {RamDash} from "./ram_dash";
|
||||
|
||||
export const DashletMenu = [
|
||||
{ name: "Throughput Bits/Second", tag: "throughputBps", size: 3 },
|
||||
@ -28,6 +30,8 @@ export const DashletMenu = [
|
||||
{ name: "Top 10 Endpoints by Country", tag: "top10endpointsCountry", size: 6 },
|
||||
{ name: "Ether Protocols", tag: "etherProtocols", size: 6 },
|
||||
{ name: "IP Protocols", tag: "ipProtocols", size: 6 },
|
||||
{ name: "CPU Utilization", tag: "cpu", size: 6 },
|
||||
{ name: "RAM Utilization", tag: "ram", size: 6 },
|
||||
];
|
||||
|
||||
export function widgetFactory(widgetName, count) {
|
||||
@ -47,6 +51,8 @@ export function widgetFactory(widgetName, count) {
|
||||
case "top10endpointsCountry" : widget = new Top10EndpointsByCountry(count); break;
|
||||
case "etherProtocols" : widget = new EtherProtocols(count); break;
|
||||
case "ipProtocols" : widget = new IpProtocols(count); break;
|
||||
case "cpu" : widget = new CpuDash(count); break;
|
||||
case "ram" : widget = new RamDash(count); break;
|
||||
default: {
|
||||
console.log("I don't know how to construct a widget of type [" + widgetName + "]");
|
||||
return null;
|
||||
|
@ -0,0 +1,37 @@
|
||||
import {BaseDashlet} from "./base_dashlet";
|
||||
import {RttHistogram} from "../graphs/rtt_histo";
|
||||
import {RamPie} from "../graphs/ram_pie";
|
||||
|
||||
export class RamDash extends BaseDashlet{
|
||||
constructor(slot) {
|
||||
super(slot);
|
||||
}
|
||||
|
||||
title() {
|
||||
return "RAM Utilization";
|
||||
}
|
||||
|
||||
subscribeTo() {
|
||||
return [ "Ram" ];
|
||||
}
|
||||
|
||||
buildContainer() {
|
||||
let base = super.buildContainer();
|
||||
base.appendChild(this.graphDiv());
|
||||
return base;
|
||||
}
|
||||
|
||||
setup() {
|
||||
super.setup();
|
||||
this.graph = new RamPie(this.graphDivId());
|
||||
}
|
||||
|
||||
onMessage(msg) {
|
||||
if (msg.event === "Ram") {
|
||||
let total = msg.data.total;
|
||||
let used = msg.data.used;
|
||||
let free = total - used;
|
||||
this.graph.update(free, used);
|
||||
}
|
||||
}
|
||||
}
|
@ -0,0 +1,46 @@
|
||||
import {DashboardGraph} from "./dashboard_graph";
|
||||
import {lerpGreenToRedViaOrange} from "../helpers/scaling";
|
||||
|
||||
export class CpuHistogram extends DashboardGraph {
|
||||
constructor(id) {
|
||||
super(id);
|
||||
let d = [];
|
||||
let axis = [];
|
||||
for (let i=0; i<20; i++) {
|
||||
d.push({
|
||||
value: i,
|
||||
itemStyle: {color: lerpGreenToRedViaOrange(20-i, 20)},
|
||||
});
|
||||
axis.push(i.toString());
|
||||
}
|
||||
this.option = {
|
||||
xAxis: {
|
||||
type: 'category',
|
||||
data: axis,
|
||||
},
|
||||
yAxis: {
|
||||
type: 'value',
|
||||
min: 0,
|
||||
max: 100,
|
||||
},
|
||||
series: {
|
||||
data: d,
|
||||
type: 'bar',
|
||||
}
|
||||
};
|
||||
this.option && this.chart.setOption(this.option);
|
||||
}
|
||||
|
||||
update(cpu) {
|
||||
this.chart.hideLoading();
|
||||
this.option.series.data = [];
|
||||
for (let i=0; i<cpu.length; i++) {
|
||||
this.option.series.data.push({
|
||||
value: cpu[i],
|
||||
itemStyle: {color: lerpGreenToRedViaOrange(100-cpu[i], 100)},
|
||||
});
|
||||
}
|
||||
this.chart.setOption(this.option);
|
||||
}
|
||||
}
|
||||
|
@ -0,0 +1,37 @@
|
||||
import {DashboardGraph} from "./dashboard_graph";
|
||||
import {scaleNumber} from "../helpers/scaling";
|
||||
|
||||
export class RamPie extends DashboardGraph {
|
||||
constructor(id) {
|
||||
super(id);
|
||||
this.option = {
|
||||
series: [
|
||||
{
|
||||
type: 'pie',
|
||||
radius: '50%',
|
||||
data: [
|
||||
{ name: 'Free', value: 0 },
|
||||
{ name: 'Used', value: 0 }
|
||||
],
|
||||
color: [
|
||||
'green', 'orange'
|
||||
],
|
||||
label: {
|
||||
color: '#aaa'
|
||||
}
|
||||
}
|
||||
],
|
||||
tooltip: {
|
||||
trigger: 'item',
|
||||
},
|
||||
}
|
||||
this.option && this.chart.setOption(this.option);
|
||||
}
|
||||
|
||||
update(free, used) {
|
||||
this.chart.hideLoading();
|
||||
this.option.series[0].data[0].value = free;
|
||||
this.option.series[0].data[1].value = used;
|
||||
this.chart.setOption(this.option);
|
||||
}
|
||||
}
|
@ -1,4 +1,5 @@
|
||||
import {DashboardGraph} from "./dashboard_graph";
|
||||
import {lerpGreenToRedViaOrange} from "../helpers/scaling";
|
||||
|
||||
export class RttHistogram extends DashboardGraph {
|
||||
constructor(id) {
|
||||
@ -37,16 +38,3 @@ export class RttHistogram extends DashboardGraph {
|
||||
}
|
||||
}
|
||||
|
||||
function lerpGreenToRedViaOrange(value, max) {
|
||||
let r = 0;
|
||||
let g = 0;
|
||||
let b = 0;
|
||||
if (value < max / 2) {
|
||||
r = 255;
|
||||
g = Math.floor(255 * value / (max / 2));
|
||||
} else {
|
||||
r = Math.floor(255 * (max - value) / (max / 2));
|
||||
g = 255;
|
||||
}
|
||||
return `rgb(${r}, ${g}, ${b})`;
|
||||
}
|
@ -38,4 +38,18 @@ export function rttCircleSpan(rtt) {
|
||||
span.style.color = colorRamp(rtt);
|
||||
span.innerText = "⬤";
|
||||
return span;
|
||||
}
|
||||
|
||||
export function lerpGreenToRedViaOrange(value, max) {
|
||||
let r = 0;
|
||||
let g = 0;
|
||||
let b = 0;
|
||||
if (value < max / 2) {
|
||||
r = 255;
|
||||
g = Math.floor(255 * value / (max / 2));
|
||||
} else {
|
||||
r = Math.floor(255 * (max - value) / (max / 2));
|
||||
g = 255;
|
||||
}
|
||||
return `rgb(${r}, ${g}, ${b})`;
|
||||
}
|
@ -13,6 +13,7 @@ use crate::node_manager::ws::ticker::channel_ticker;
|
||||
pub fn websocket_router() -> Router {
|
||||
let channels = PubSub::new();
|
||||
tokio::spawn(channel_ticker(channels.clone()));
|
||||
tokio::spawn(ticker::system_info::cache::update_cache());
|
||||
Router::new()
|
||||
.route("/ws", get(ws_handler))
|
||||
.layer(Extension(channels))
|
||||
|
@ -15,4 +15,6 @@ pub enum PublishedChannels {
|
||||
EndpointsByCountry,
|
||||
EtherProtocols,
|
||||
IpProtocols,
|
||||
Cpu,
|
||||
Ram,
|
||||
}
|
||||
|
@ -6,6 +6,7 @@ mod top_10;
|
||||
mod ipstats_conversion;
|
||||
mod top_flows;
|
||||
mod flow_endpoints;
|
||||
pub mod system_info;
|
||||
|
||||
use std::sync::Arc;
|
||||
use crate::node_manager::ws::publish_subscribe::PubSub;
|
||||
@ -29,6 +30,8 @@ pub(super) async fn channel_ticker(channels: Arc<PubSub>) {
|
||||
flow_endpoints::endpoints_by_country(channels.clone()),
|
||||
flow_endpoints::ether_protocols(channels.clone()),
|
||||
flow_endpoints::ip_protocols(channels.clone()),
|
||||
system_info::cpu_info(channels.clone()),
|
||||
system_info::ram_info(channels.clone()),
|
||||
);
|
||||
}
|
||||
}
|
47
src/rust/lqosd/src/node_manager/ws/ticker/system_info.rs
Normal file
47
src/rust/lqosd/src/node_manager/ws/ticker/system_info.rs
Normal file
@ -0,0 +1,47 @@
|
||||
use std::sync::Arc;
|
||||
use serde_json::json;
|
||||
use crate::node_manager::ws::publish_subscribe::PubSub;
|
||||
use crate::node_manager::ws::published_channels::PublishedChannels;
|
||||
use crate::node_manager::ws::ticker::system_info::cache::{CPU_USAGE, NUM_CPUS, RAM_USED, TOTAL_RAM};
|
||||
|
||||
pub mod cache;
|
||||
|
||||
pub async fn cpu_info(channels: Arc<PubSub>) {
|
||||
if !channels.is_channel_alive(PublishedChannels::Cpu).await {
|
||||
return;
|
||||
}
|
||||
|
||||
let usage: Vec<u32> = CPU_USAGE
|
||||
.iter()
|
||||
.take(NUM_CPUS.load(std::sync::atomic::Ordering::Relaxed))
|
||||
.map(|cpu| cpu.load(std::sync::atomic::Ordering::Relaxed))
|
||||
.collect();
|
||||
|
||||
let message = json!(
|
||||
{
|
||||
"event": PublishedChannels::Cpu.to_string(),
|
||||
"data": usage,
|
||||
}
|
||||
).to_string();
|
||||
channels.send(PublishedChannels::Cpu, message).await;
|
||||
}
|
||||
|
||||
pub async fn ram_info(channels: Arc<PubSub>) {
|
||||
if !channels.is_channel_alive(PublishedChannels::Ram).await {
|
||||
return;
|
||||
}
|
||||
|
||||
let ram_usage = RAM_USED.load(std::sync::atomic::Ordering::Relaxed);
|
||||
let total_ram = TOTAL_RAM.load(std::sync::atomic::Ordering::Relaxed);
|
||||
|
||||
let message = json!(
|
||||
{
|
||||
"event": PublishedChannels::Ram.to_string(),
|
||||
"data": {
|
||||
"total" : total_ram,
|
||||
"used" : ram_usage,
|
||||
},
|
||||
}
|
||||
).to_string();
|
||||
channels.send(PublishedChannels::Ram, message).await;
|
||||
}
|
@ -0,0 +1,57 @@
|
||||
use std::sync::atomic::{AtomicU32, AtomicU64, AtomicUsize};
|
||||
use std::time::Duration;
|
||||
|
||||
use once_cell::sync::Lazy;
|
||||
|
||||
const MAX_CPUS_COUNTED: usize = 128;
|
||||
|
||||
/// Stores overall CPU usage
|
||||
pub static CPU_USAGE: Lazy<[AtomicU32; MAX_CPUS_COUNTED]> =
|
||||
Lazy::new(build_empty_cpu_list);
|
||||
|
||||
/// Total number of CPUs detected
|
||||
pub static NUM_CPUS: AtomicUsize = AtomicUsize::new(0);
|
||||
|
||||
/// Total RAM used (bytes)
|
||||
pub static RAM_USED: AtomicU64 = AtomicU64::new(0);
|
||||
|
||||
/// Total RAM installed (bytes)
|
||||
pub static TOTAL_RAM: AtomicU64 = AtomicU64::new(0);
|
||||
|
||||
fn build_empty_cpu_list() -> [AtomicU32; MAX_CPUS_COUNTED] {
|
||||
let mut temp = Vec::with_capacity(MAX_CPUS_COUNTED);
|
||||
for _ in 0..MAX_CPUS_COUNTED {
|
||||
temp.push(AtomicU32::new(0));
|
||||
}
|
||||
temp.try_into().expect("This should never happen, sizes are constant.")
|
||||
}
|
||||
|
||||
pub async fn update_cache() {
|
||||
use sysinfo::System;
|
||||
let mut sys = System::new_all();
|
||||
tokio::time::sleep(Duration::from_secs(10)).await;
|
||||
|
||||
let mut interval = tokio::time::interval(tokio::time::Duration::from_secs(1));
|
||||
loop {
|
||||
interval.tick().await; // Once per second
|
||||
|
||||
sys.refresh_cpu();
|
||||
sys.refresh_memory();
|
||||
|
||||
sys
|
||||
.cpus()
|
||||
.iter()
|
||||
.enumerate()
|
||||
.map(|(i, cpu)| (i, cpu.cpu_usage() as u32)) // Always rounds down
|
||||
.for_each(|(i, cpu)| {
|
||||
CPU_USAGE[i].store(cpu, std::sync::atomic::Ordering::Relaxed)
|
||||
});
|
||||
|
||||
NUM_CPUS
|
||||
.store(sys.cpus().len(), std::sync::atomic::Ordering::Relaxed);
|
||||
RAM_USED
|
||||
.store(sys.used_memory(), std::sync::atomic::Ordering::Relaxed);
|
||||
TOTAL_RAM
|
||||
.store(sys.total_memory(), std::sync::atomic::Ordering::Relaxed);
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue
Block a user