tabbed frametype filtering
This commit is contained in:
621
analyzer/tui/textual/app_v2.py.backup
Normal file
621
analyzer/tui/textual/app_v2.py.backup
Normal file
@@ -0,0 +1,621 @@
|
||||
"""
|
||||
StreamLens Textual Application V2 - TipTop-Inspired Design
|
||||
Modern TUI with real-time metrics, sparklines, and professional monitoring aesthetic
|
||||
"""
|
||||
|
||||
from textual.app import App, ComposeResult
|
||||
from textual.containers import Container, Horizontal, Vertical, ScrollableContainer
|
||||
from textual.widgets import Header, Footer, Static, DataTable, Label, TabPane
|
||||
from textual.reactive import reactive
|
||||
from textual.timer import Timer
|
||||
from textual.events import MouseDown, MouseMove
|
||||
from typing import TYPE_CHECKING
|
||||
from rich.text import Text
|
||||
from rich.console import Group
|
||||
from rich.panel import Panel
|
||||
from rich.table import Table
|
||||
import time
|
||||
import signal
|
||||
import sys
|
||||
import datetime
|
||||
from pathlib import Path
|
||||
import subprocess
|
||||
import platform
|
||||
|
||||
from .widgets.sparkline import SparklineWidget
|
||||
from .widgets.metric_card import MetricCard
|
||||
from .widgets.flow_table_v2 import EnhancedFlowTable
|
||||
from .widgets.filtered_flow_view import FilteredFlowView
|
||||
from ...reporting import FlowReportGenerator
|
||||
from .widgets.split_flow_details import FlowMainDetailsPanel, SubFlowDetailsPanel
|
||||
from .widgets.debug_panel import DebugPanel
|
||||
from .widgets.progress_bar import ParsingProgressBar
|
||||
from ...analysis.background_analyzer import BackgroundAnalyzer
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from ...analysis.core import EthernetAnalyzer
|
||||
|
||||
|
||||
class StreamLensAppV2(App):
|
||||
"""
|
||||
StreamLens TipTop-Inspired Interface
|
||||
|
||||
Features:
|
||||
- Real-time metrics with sparklines
|
||||
- Color-coded quality indicators
|
||||
- Compact information display
|
||||
- Multi-column layout
|
||||
- Smooth live updates
|
||||
"""
|
||||
|
||||
CSS_PATH = "styles/streamlens_v2.tcss"
|
||||
ENABLE_COMMAND_PALETTE = False
|
||||
AUTO_FOCUS = None
|
||||
|
||||
BINDINGS = [
|
||||
("q", "quit", "Quit"),
|
||||
("1", "select_filter('1')", "Overview"),
|
||||
("2", "select_filter('2')", "Frame Type 2"),
|
||||
("3", "select_filter('3')", "Frame Type 3"),
|
||||
("4", "select_filter('4')", "Frame Type 4"),
|
||||
("5", "select_filter('5')", "Frame Type 5"),
|
||||
("6", "select_filter('6')", "Frame Type 6"),
|
||||
("7", "select_filter('7')", "Frame Type 7"),
|
||||
("8", "select_filter('8')", "Frame Type 8"),
|
||||
("9", "select_filter('9')", "Frame Type 9"),
|
||||
("0", "select_filter('0')", "Frame Type 10"),
|
||||
("alt+1", "sort_table_column(0)", "Sort by column 1"),
|
||||
("alt+2", "sort_table_column(1)", "Sort by column 2"),
|
||||
("alt+3", "sort_table_column(2)", "Sort by column 3"),
|
||||
("alt+4", "sort_table_column(3)", "Sort by column 4"),
|
||||
("alt+5", "sort_table_column(4)", "Sort by column 5"),
|
||||
("alt+6", "sort_table_column(5)", "Sort by column 6"),
|
||||
("alt+7", "sort_table_column(6)", "Sort by column 7"),
|
||||
("alt+8", "sort_table_column(7)", "Sort by column 8"),
|
||||
("alt+9", "sort_table_column(8)", "Sort by column 9"),
|
||||
("alt+0", "sort_table_column(9)", "Sort by column 10"),
|
||||
("p", "toggle_pause", "Pause"),
|
||||
("d", "show_details", "Details"),
|
||||
("v", "toggle_view_mode", "Toggle View"),
|
||||
("r", "generate_report", "Generate Report"),
|
||||
("o", "copy_outliers", "Copy Outliers"),
|
||||
("?", "toggle_help", "Help"),
|
||||
]
|
||||
|
||||
# Reactive attributes
|
||||
total_flows = reactive(0)
|
||||
total_packets = reactive(0)
|
||||
packets_per_sec = reactive(0.0)
|
||||
bytes_per_sec = reactive(0.0)
|
||||
enhanced_flows = reactive(0)
|
||||
outlier_count = reactive(0)
|
||||
debug_visible = reactive(False) # Hide debug panel for now
|
||||
|
||||
# Update timers
|
||||
metric_timer: Timer = None
|
||||
flow_timer: Timer = None
|
||||
|
||||
def __init__(self, analyzer: 'EthernetAnalyzer'):
|
||||
super().__init__()
|
||||
self.analyzer = analyzer
|
||||
self.title = "StreamLens"
|
||||
self.sub_title = "Network Flow Analysis"
|
||||
self.paused = False
|
||||
|
||||
# Background parsing support - Use single thread to avoid race conditions in frame reference tracking
|
||||
self.background_analyzer = BackgroundAnalyzer(
|
||||
analyzer=analyzer,
|
||||
num_threads=1, # Single-threaded to prevent race conditions in outlier frame references
|
||||
batch_size=1000,
|
||||
progress_callback=self._on_progress_update,
|
||||
flow_update_callback=self._on_flow_update
|
||||
)
|
||||
self.pcap_file = None
|
||||
|
||||
|
||||
# Metrics history for sparklines
|
||||
self.packets_history = []
|
||||
self.bytes_history = []
|
||||
self.flows_history = []
|
||||
self.max_history = 60 # 60 seconds of history
|
||||
|
||||
def compose(self) -> ComposeResult:
|
||||
"""Create TipTop-inspired layout"""
|
||||
yield Header()
|
||||
|
||||
with Container(id="main-container"):
|
||||
# Progress bar for PCAP loading (initially hidden)
|
||||
yield ParsingProgressBar(id="progress-bar")
|
||||
|
||||
# Ultra-compact metrics bar
|
||||
with Horizontal(id="metrics-bar"):
|
||||
yield MetricCard("Flows", f"{self.total_flows}", id="flows-metric")
|
||||
yield MetricCard("Pkts/s", f"{self.packets_per_sec:.0f}", id="packets-metric")
|
||||
yield MetricCard("Vol/s", self._format_bytes_per_sec(self.bytes_per_sec), id="volume-metric")
|
||||
yield MetricCard("Enhanced", f"{self.enhanced_flows}", color="success", id="enhanced-metric")
|
||||
yield MetricCard("Outliers", f"{self.outlier_count}", color="warning" if self.outlier_count > 0 else "normal", id="outliers-metric")
|
||||
|
||||
# Main content area with conditional debug panel
|
||||
with Horizontal(id="content-area"):
|
||||
# Left - Filtered flow view with frame type buttons
|
||||
yield FilteredFlowView(
|
||||
self.analyzer,
|
||||
id="filtered-flow-view",
|
||||
classes="panel-wide"
|
||||
)
|
||||
|
||||
# Middle - Flow details
|
||||
with Vertical(id="flow-panels"):
|
||||
yield FlowMainDetailsPanel(id="main-flow-details")
|
||||
yield SubFlowDetailsPanel(id="sub-flow-details")
|
||||
|
||||
# Right - Debug panel (conditionally visible)
|
||||
if self.debug_visible:
|
||||
yield DebugPanel(id="debug-panel")
|
||||
|
||||
yield Footer()
|
||||
|
||||
def on_mount(self) -> None:
|
||||
"""Initialize the application with TipTop-style updates"""
|
||||
try:
|
||||
debug_panel = self.query_one("#debug-panel", DebugPanel)
|
||||
debug_panel.add_debug_message("APP: Application mounted, checking panels...")
|
||||
|
||||
try:
|
||||
main_panel = self.query_one("#main-flow-details", FlowMainDetailsPanel)
|
||||
sub_panel = self.query_one("#sub-flow-details", SubFlowDetailsPanel)
|
||||
debug_panel.add_debug_message("APP: Both panels found successfully")
|
||||
except Exception as e:
|
||||
debug_panel.add_debug_message(f"APP: Panel query failed: {e}")
|
||||
except:
|
||||
pass # Debug panel not visible
|
||||
|
||||
# Set initial subtitle with view mode
|
||||
try:
|
||||
flow_table = self.query_one("#flow-table", EnhancedFlowTable)
|
||||
view_mode = flow_table.get_current_view_mode()
|
||||
status = "PAUSED" if self.paused else "LIVE"
|
||||
self.sub_title = f"Network Flow Analysis - {status} - {view_mode} VIEW"
|
||||
except:
|
||||
pass
|
||||
|
||||
self.update_metrics()
|
||||
|
||||
# Set up update intervals (slower during parsing to reduce CPU usage)
|
||||
self.metric_timer = self.set_interval(5.0, self.update_metrics) # 0.2Hz for slower background updates
|
||||
self.flow_timer = self.set_interval(10.0, self.update_flows) # 0.1Hz for slower fallback flow updates
|
||||
|
||||
# Initialize sparkline history
|
||||
self._initialize_history()
|
||||
|
||||
# Set initial focus to the flow table for immediate keyboard navigation
|
||||
self.call_after_refresh(self._set_initial_focus)
|
||||
|
||||
def _set_initial_focus(self):
|
||||
"""Set initial focus to the filtered flow view after widgets are ready"""
|
||||
try:
|
||||
flow_view = self.query_one("#filtered-flow-view", FilteredFlowView)
|
||||
flow_view.flow_table.focus()
|
||||
except Exception:
|
||||
# If flow view isn't ready yet, try again after a short delay
|
||||
self.set_timer(0.1, self._set_initial_focus)
|
||||
|
||||
def _initialize_history(self):
|
||||
"""Initialize metrics history arrays"""
|
||||
current_time = time.time()
|
||||
for _ in range(self.max_history):
|
||||
self.packets_history.append(0)
|
||||
self.bytes_history.append(0)
|
||||
self.flows_history.append(0)
|
||||
|
||||
def update_metrics(self) -> None:
|
||||
"""Update real-time metrics and sparklines"""
|
||||
if self.paused:
|
||||
return
|
||||
|
||||
# Get current metrics
|
||||
summary = self.analyzer.get_summary()
|
||||
self.total_flows = summary.get('unique_flows', 0)
|
||||
self.total_packets = summary.get('total_packets', 0)
|
||||
|
||||
# Calculate rates (simplified for now)
|
||||
# In real implementation, track deltas over time
|
||||
current_time = time.time()
|
||||
if not hasattr(self, '_start_time'):
|
||||
self._start_time = current_time
|
||||
|
||||
elapsed = max(1, current_time - self._start_time)
|
||||
self.packets_per_sec = self.total_packets / elapsed
|
||||
self.bytes_per_sec = summary.get('total_bytes', 0) / elapsed
|
||||
|
||||
# Count enhanced and outliers (thread-safe access)
|
||||
enhanced = 0
|
||||
outliers = 0
|
||||
try:
|
||||
# Use background analyzer's thread-safe flow access
|
||||
flows = self.background_analyzer.get_current_flows()
|
||||
for flow in flows.values():
|
||||
if flow.enhanced_analysis.decoder_type != "Standard":
|
||||
enhanced += 1
|
||||
# Use frame-type-specific outliers instead of flow-level outliers
|
||||
outliers += sum(len(ft_stats.outlier_frames) for ft_stats in flow.frame_types.values())
|
||||
except Exception:
|
||||
# Fallback to direct access if background analyzer not available
|
||||
for flow in self.analyzer.flows.values():
|
||||
if flow.enhanced_analysis.decoder_type != "Standard":
|
||||
enhanced += 1
|
||||
# Use frame-type-specific outliers instead of flow-level outliers
|
||||
outliers += sum(len(ft_stats.outlier_frames) for ft_stats in flow.frame_types.values())
|
||||
|
||||
self.enhanced_flows = enhanced
|
||||
self.outlier_count = outliers
|
||||
|
||||
# Update metric cards
|
||||
self._update_metric_cards()
|
||||
|
||||
# Update sparklines (removed - no longer in left panel)
|
||||
# self._update_sparklines()
|
||||
|
||||
def _update_metric_cards(self):
|
||||
"""Update the metric card displays"""
|
||||
# Update flows metric
|
||||
flows_card = self.query_one("#flows-metric", MetricCard)
|
||||
flows_card.update_value(f"{self.total_flows}")
|
||||
|
||||
# Update packets/s with color coding
|
||||
packets_card = self.query_one("#packets-metric", MetricCard)
|
||||
packets_card.update_value(f"{self.packets_per_sec:.1f}")
|
||||
if self.packets_per_sec > 10000:
|
||||
packets_card.color = "warning"
|
||||
elif self.packets_per_sec > 50000:
|
||||
packets_card.color = "error"
|
||||
else:
|
||||
packets_card.color = "success"
|
||||
|
||||
# Update volume/s
|
||||
volume_card = self.query_one("#volume-metric", MetricCard)
|
||||
volume_card.update_value(self._format_bytes_per_sec(self.bytes_per_sec))
|
||||
|
||||
# Update enhanced flows
|
||||
enhanced_card = self.query_one("#enhanced-metric", MetricCard)
|
||||
enhanced_card.update_value(f"{self.enhanced_flows}")
|
||||
|
||||
# Update outliers with color
|
||||
outliers_card = self.query_one("#outliers-metric", MetricCard)
|
||||
outliers_card.update_value(f"{self.outlier_count}")
|
||||
if self.outlier_count > 100:
|
||||
outliers_card.color = "error"
|
||||
elif self.outlier_count > 10:
|
||||
outliers_card.color = "warning"
|
||||
else:
|
||||
outliers_card.color = "normal"
|
||||
|
||||
def _update_sparklines(self):
|
||||
"""Update sparkline charts with latest data"""
|
||||
# Add new data points
|
||||
self.packets_history.append(self.packets_per_sec)
|
||||
self.bytes_history.append(self.bytes_per_sec)
|
||||
self.flows_history.append(self.total_flows)
|
||||
|
||||
# Keep only recent history
|
||||
if len(self.packets_history) > self.max_history:
|
||||
self.packets_history.pop(0)
|
||||
self.bytes_history.pop(0)
|
||||
self.flows_history.pop(0)
|
||||
|
||||
# Update sparkline widgets
|
||||
flow_spark = self.query_one("#flow-rate-spark", SparklineWidget)
|
||||
flow_spark.update_data(self.flows_history)
|
||||
|
||||
packet_spark = self.query_one("#packet-rate-spark", SparklineWidget)
|
||||
packet_spark.update_data(self.packets_history)
|
||||
|
||||
def update_flows(self) -> None:
|
||||
"""Update flow table data"""
|
||||
if self.paused:
|
||||
return
|
||||
|
||||
# Update filtered flow view
|
||||
flow_view = self.query_one("#filtered-flow-view", FilteredFlowView)
|
||||
flow_view.refresh_frame_types()
|
||||
flow_view.refresh_flow_data()
|
||||
def _on_progress_update(self, progress):
|
||||
"""Handle progress updates from background parser"""
|
||||
try:
|
||||
# Use call_from_thread to safely update UI from background thread
|
||||
self.call_from_thread(self._update_progress_ui, progress)
|
||||
except Exception:
|
||||
# Ignore errors during shutdown
|
||||
pass
|
||||
|
||||
def _update_progress_ui(self, progress):
|
||||
"""Update progress UI (called from main thread)"""
|
||||
try:
|
||||
progress_bar = self.query_one("#progress-bar", ParsingProgressBar)
|
||||
|
||||
if progress.error:
|
||||
progress_bar.show_error(progress.error)
|
||||
elif progress.is_complete:
|
||||
progress_bar.complete_parsing()
|
||||
# Trigger frame type button creation now that parsing is complete
|
||||
self._create_frame_type_buttons()
|
||||
else:
|
||||
# Start progress if this is the first update
|
||||
if not progress_bar.is_visible and progress.total_packets > 0:
|
||||
progress_bar.start_parsing(progress.total_packets)
|
||||
|
||||
# Update progress
|
||||
progress_bar.update_progress(
|
||||
progress.processed_packets,
|
||||
progress.total_packets,
|
||||
progress.packets_per_second,
|
||||
progress.estimated_time_remaining
|
||||
)
|
||||
except Exception as e:
|
||||
# Progress bar widget may not be available yet
|
||||
pass
|
||||
|
||||
def _on_flow_update(self):
|
||||
"""Handle flow data updates from background parser"""
|
||||
try:
|
||||
# Use call_from_thread to safely update UI from background thread
|
||||
self.call_from_thread(self._update_flow_ui)
|
||||
except Exception:
|
||||
# Ignore errors during shutdown
|
||||
pass
|
||||
|
||||
def _update_flow_ui(self):
|
||||
"""Update flow UI (called from main thread)"""
|
||||
try:
|
||||
# Update filtered flow view - frame types first for dynamic button creation
|
||||
flow_view = self.query_one("#filtered-flow-view", FilteredFlowView)
|
||||
flow_view.refresh_frame_types() # This will create buttons as frame types are detected
|
||||
flow_view.refresh_flow_data()
|
||||
|
||||
# Also trigger button creation if parsing is complete but buttons haven't been created yet
|
||||
if not self.analyzer.is_parsing and not getattr(flow_view, '_buttons_created', False):
|
||||
self._create_frame_type_buttons()
|
||||
|
||||
# Also update metrics in real-time
|
||||
self.update_metrics()
|
||||
except Exception:
|
||||
# Flow view widget may not be available yet
|
||||
pass
|
||||
|
||||
def _create_frame_type_buttons(self):
|
||||
"""Create frame type buttons now that parsing is complete"""
|
||||
try:
|
||||
flow_view = self.query_one("#filtered-flow-view", FilteredFlowView)
|
||||
# Force refresh of frame types now that parsing is complete
|
||||
flow_view.refresh_frame_types()
|
||||
flow_view.refresh_flow_data()
|
||||
except Exception as e:
|
||||
# Flow view widget may not be available yet
|
||||
pass
|
||||
|
||||
def start_background_parsing(self, pcap_file: str):
|
||||
"""Start parsing PCAP file in background"""
|
||||
self.pcap_file = pcap_file
|
||||
|
||||
# Start background parsing
|
||||
self.background_analyzer.start_parsing(pcap_file)
|
||||
|
||||
def stop_background_parsing(self):
|
||||
"""Stop background parsing"""
|
||||
self.background_analyzer.stop_parsing()
|
||||
|
||||
def cleanup(self):
|
||||
"""Cleanup resources when app shuts down"""
|
||||
try:
|
||||
self.background_analyzer.cleanup()
|
||||
# Cancel any pending timers
|
||||
if self.metric_timer:
|
||||
self.metric_timer.stop()
|
||||
if self.flow_timer:
|
||||
self.flow_timer.stop()
|
||||
except Exception as e:
|
||||
# Don't let cleanup errors prevent shutdown
|
||||
pass
|
||||
|
||||
def on_enhanced_flow_table_flow_selected(self, event: EnhancedFlowTable.FlowSelected) -> None:
|
||||
"""Handle flow selection events"""
|
||||
try:
|
||||
debug_panel = self.query_one("#debug-panel", DebugPanel)
|
||||
flow_info = f"{event.flow.src_ip}:{event.flow.src_port}" if event.flow else "None"
|
||||
debug_panel.add_debug_message(f"APP: Flow selected - {flow_info}, subflow={event.subflow_type}")
|
||||
except:
|
||||
pass # Debug panel not visible
|
||||
|
||||
if event.flow:
|
||||
# Update main flow details panel
|
||||
main_panel = self.query_one("#main-flow-details", FlowMainDetailsPanel)
|
||||
main_panel.update_flow(event.flow)
|
||||
|
||||
# Update sub-flow details panel
|
||||
sub_panel = self.query_one("#sub-flow-details", SubFlowDetailsPanel)
|
||||
sub_panel.update_flow(event.flow, event.subflow_type)
|
||||
|
||||
|
||||
def _format_bytes_per_sec(self, bps: float) -> str:
|
||||
"""Format bytes per second with appropriate units"""
|
||||
if bps >= 1_000_000_000:
|
||||
return f"{bps / 1_000_000_000:.1f} GB/s"
|
||||
elif bps >= 1_000_000:
|
||||
return f"{bps / 1_000_000:.1f} MB/s"
|
||||
elif bps >= 1_000:
|
||||
return f"{bps / 1_000:.1f} KB/s"
|
||||
else:
|
||||
return f"{bps:.0f} B/s"
|
||||
|
||||
def action_toggle_pause(self) -> None:
|
||||
"""Toggle pause state"""
|
||||
self.paused = not self.paused
|
||||
status = "PAUSED" if self.paused else "LIVE"
|
||||
|
||||
# Update subtitle
|
||||
self.sub_title = f"Network Flow Analysis - {status}"
|
||||
|
||||
def action_select_filter(self, number: str) -> None:
|
||||
"""Select frame type filter by number key"""
|
||||
try:
|
||||
flow_view = self.query_one("#filtered-flow-view", FilteredFlowView)
|
||||
flow_view.action_select_filter(number)
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
def action_sort_table_column(self, column_index: int) -> None:
|
||||
"""Sort table by column index"""
|
||||
try:
|
||||
flow_view = self.query_one("#filtered-flow-view", FilteredFlowView)
|
||||
flow_view.action_sort_column(column_index)
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
def action_show_details(self) -> None:
|
||||
"""Show detailed view for selected flow"""
|
||||
# TODO: Implement detailed flow modal
|
||||
pass
|
||||
|
||||
def action_toggle_view_mode(self) -> None:
|
||||
"""Toggle between different display modes"""
|
||||
# For now, this could cycle through different column layouts
|
||||
# or show more/less detail in the frame type views
|
||||
pass
|
||||
|
||||
|
||||
def on_mouse_down(self, event: MouseDown) -> None:
|
||||
"""Prevent default mouse down behavior to disable mouse interaction."""
|
||||
event.prevent_default()
|
||||
|
||||
def on_mouse_move(self, event: MouseMove) -> None:
|
||||
"""Prevent default mouse move behavior to disable mouse interaction."""
|
||||
event.prevent_default()
|
||||
|
||||
def action_generate_report(self) -> None:
|
||||
"""Generate comprehensive flow analysis report"""
|
||||
try:
|
||||
# Generate timestamp-based filename
|
||||
timestamp = datetime.datetime.now().strftime("%Y%m%d_%H%M%S")
|
||||
output_file = f"streamlens_flow_report_{timestamp}.md"
|
||||
|
||||
# Create report generator
|
||||
report_generator = FlowReportGenerator(self.analyzer)
|
||||
|
||||
# Generate report (markdown format)
|
||||
report_content = report_generator.generate_report(output_file, "markdown")
|
||||
|
||||
# Show success notification in the footer
|
||||
self.sub_title = f"✅ Report generated: {output_file}"
|
||||
|
||||
# Set a timer to restore the original subtitle
|
||||
self.set_timer(3.0, self._restore_subtitle)
|
||||
|
||||
except Exception as e:
|
||||
# Show error notification
|
||||
self.sub_title = f"❌ Report generation failed: {str(e)}"
|
||||
self.set_timer(3.0, self._restore_subtitle)
|
||||
|
||||
def _restore_subtitle(self) -> None:
|
||||
"""Restore the original subtitle"""
|
||||
status = "PAUSED" if self.paused else "LIVE"
|
||||
self.sub_title = f"Network Flow Analysis - {status}"
|
||||
|
||||
def action_copy_outliers(self) -> None:
|
||||
"""Copy outlier frame information to clipboard"""
|
||||
try:
|
||||
# Get selected flow from the filtered view
|
||||
flow_view = self.query_one("#filtered-flow-view", FilteredFlowView)
|
||||
# For now, get the first flow (could be improved to use actual selection)
|
||||
flows = list(self.analyzer.flows.values())
|
||||
selected_flow = flows[0] if flows else None
|
||||
|
||||
if not selected_flow:
|
||||
self.sub_title = "⚠️ No flow selected"
|
||||
self.set_timer(2.0, self._restore_subtitle)
|
||||
return
|
||||
|
||||
# Build frame-type-specific outlier information
|
||||
outlier_info = []
|
||||
outlier_info.append(f"Flow: {selected_flow.src_ip}:{selected_flow.src_port} → {selected_flow.dst_ip}:{selected_flow.dst_port}")
|
||||
outlier_info.append(f"Protocol: {selected_flow.transport_protocol}")
|
||||
outlier_info.append(f"Total Packets: {selected_flow.frame_count}")
|
||||
|
||||
# Calculate total frame-type-specific outliers
|
||||
total_frame_type_outliers = sum(len(ft_stats.outlier_frames) for ft_stats in selected_flow.frame_types.values())
|
||||
outlier_info.append(f"Total Frame-Type Outliers: {total_frame_type_outliers}")
|
||||
|
||||
if total_frame_type_outliers > 0:
|
||||
outlier_info.append(f"\n=== Frame Type Outlier Analysis ===")
|
||||
|
||||
# Show outliers per frame type
|
||||
for frame_type, ft_stats in sorted(selected_flow.frame_types.items(), key=lambda x: len(x[1].outlier_frames), reverse=True):
|
||||
if ft_stats.outlier_frames:
|
||||
outlier_info.append(f"\n{frame_type}: {len(ft_stats.outlier_frames)} outliers")
|
||||
outlier_info.append(f" Frames: {', '.join(map(str, sorted(ft_stats.outlier_frames)))}")
|
||||
outlier_info.append(f" Avg ΔT: {ft_stats.avg_inter_arrival * 1000:.3f} ms")
|
||||
outlier_info.append(f" Std σ: {ft_stats.std_inter_arrival * 1000:.3f} ms")
|
||||
outlier_info.append(f" 3σ Threshold: {(ft_stats.avg_inter_arrival + 3 * ft_stats.std_inter_arrival) * 1000:.3f} ms")
|
||||
|
||||
# Show enhanced outlier information for this frame type
|
||||
if hasattr(ft_stats, 'enhanced_outlier_details') and ft_stats.enhanced_outlier_details:
|
||||
outlier_info.append(f" Enhanced Outlier Details:")
|
||||
for frame_num, prev_frame_num, inter_time in sorted(ft_stats.enhanced_outlier_details[:5]):
|
||||
deviation = (inter_time - ft_stats.avg_inter_arrival) / ft_stats.std_inter_arrival if ft_stats.std_inter_arrival > 0 else 0
|
||||
outlier_info.append(f" Frame {frame_num} (from {prev_frame_num}): {inter_time * 1000:.3f} ms ({deviation:.1f}σ)")
|
||||
if len(ft_stats.enhanced_outlier_details) > 5:
|
||||
outlier_info.append(f" ... and {len(ft_stats.enhanced_outlier_details) - 5} more")
|
||||
elif ft_stats.outlier_details:
|
||||
outlier_info.append(f" Outlier Details:")
|
||||
for frame_num, inter_time in sorted(ft_stats.outlier_details[:5]):
|
||||
deviation = (inter_time - ft_stats.avg_inter_arrival) / ft_stats.std_inter_arrival if ft_stats.std_inter_arrival > 0 else 0
|
||||
outlier_info.append(f" Frame {frame_num}: {inter_time * 1000:.3f} ms ({deviation:.1f}σ)")
|
||||
if len(ft_stats.outlier_details) > 5:
|
||||
outlier_info.append(f" ... and {len(ft_stats.outlier_details) - 5} more")
|
||||
else:
|
||||
outlier_info.append("\nNo frame-type-specific timing outliers detected.")
|
||||
|
||||
# Copy to clipboard
|
||||
clipboard_text = "\n".join(outlier_info)
|
||||
self._copy_to_clipboard(clipboard_text)
|
||||
|
||||
# Show success notification
|
||||
total_frame_type_outliers = sum(len(ft_stats.outlier_frames) for ft_stats in selected_flow.frame_types.values())
|
||||
self.sub_title = f"✅ Copied {total_frame_type_outliers} frame-type outliers to clipboard"
|
||||
self.set_timer(2.0, self._restore_subtitle)
|
||||
|
||||
except Exception as e:
|
||||
self.sub_title = f"❌ Failed to copy: {str(e)}"
|
||||
self.set_timer(2.0, self._restore_subtitle)
|
||||
|
||||
def _copy_to_clipboard(self, text: str) -> None:
|
||||
"""Copy text to system clipboard"""
|
||||
system = platform.system()
|
||||
|
||||
if system == "Darwin": # macOS
|
||||
process = subprocess.Popen(['pbcopy'], stdin=subprocess.PIPE)
|
||||
process.communicate(text.encode('utf-8'))
|
||||
elif system == "Linux":
|
||||
# Try xclip first, then xsel
|
||||
try:
|
||||
process = subprocess.Popen(['xclip', '-selection', 'clipboard'], stdin=subprocess.PIPE)
|
||||
process.communicate(text.encode('utf-8'))
|
||||
except FileNotFoundError:
|
||||
try:
|
||||
process = subprocess.Popen(['xsel', '--clipboard', '--input'], stdin=subprocess.PIPE)
|
||||
process.communicate(text.encode('utf-8'))
|
||||
except FileNotFoundError:
|
||||
raise Exception("Neither xclip nor xsel found. Please install one.")
|
||||
elif system == "Windows":
|
||||
process = subprocess.Popen(['clip'], stdin=subprocess.PIPE, shell=True)
|
||||
process.communicate(text.encode('utf-8'))
|
||||
else:
|
||||
raise Exception(f"Unsupported platform: {system}")
|
||||
|
||||
def action_quit(self) -> None:
|
||||
"""Quit the application with proper cleanup"""
|
||||
self.cleanup()
|
||||
self.exit()
|
||||
|
||||
def on_unmount(self) -> None:
|
||||
"""Called when app is being unmounted - ensure cleanup"""
|
||||
self.cleanup()
|
||||
Reference in New Issue
Block a user