wlan_telemetry/
lib.rs

1// Copyright 2024 The Fuchsia Authors. All rights reserved.
2// Use of this source code is governed by a BSD-style license that can be
3// found in the LICENSE file.
4use anyhow::{Context as _, Error, format_err};
5use fuchsia_inspect::Node as InspectNode;
6use futures::channel::mpsc;
7use futures::{Future, StreamExt, select};
8use log::error;
9use std::boxed::Box;
10use windowed_stats::experimental::inspect::TimeMatrixClient;
11use wlan_common::bss::BssDescription;
12use {
13    fidl_fuchsia_power_battery as fidl_battery, fidl_fuchsia_wlan_ieee80211 as fidl_ieee80211,
14    fidl_fuchsia_wlan_internal as fidl_internal, fuchsia_async as fasync,
15    fuchsia_inspect_auto_persist as auto_persist, wlan_legacy_metrics_registry as metrics,
16};
17
18mod processors;
19pub(crate) mod util;
20pub use crate::processors::connect_disconnect::DisconnectInfo;
21pub use crate::processors::power::{IfacePowerLevel, UnclearPowerDemand};
22pub use crate::processors::scan::ScanResult;
23pub use crate::processors::toggle_events::ClientConnectionsToggleEvent;
24pub use util::sender::TelemetrySender;
25#[cfg(test)]
26mod testing;
27
28#[derive(Debug)]
29pub enum TelemetryEvent {
30    ConnectResult {
31        result: fidl_ieee80211::StatusCode,
32        bss: Box<BssDescription>,
33    },
34    Disconnect {
35        info: DisconnectInfo,
36    },
37    // We should maintain docstrings if we can see any possibility of ambiguity for an enum
38    /// Client connections enabled or disabled
39    ClientConnectionsToggle {
40        event: ClientConnectionsToggleEvent,
41    },
42    ClientIfaceCreated {
43        iface_id: u16,
44    },
45    ClientIfaceDestroyed {
46        iface_id: u16,
47    },
48    IfaceCreationFailure,
49    IfaceDestructionFailure,
50    ScanStart,
51    ScanResult {
52        result: ScanResult,
53    },
54    IfacePowerLevelChanged {
55        iface_power_level: IfacePowerLevel,
56        iface_id: u16,
57    },
58    /// System suspension imminent
59    SuspendImminent,
60    /// Unclear power level requested by policy layer
61    UnclearPowerDemand(UnclearPowerDemand),
62    BatteryChargeStatus(fidl_battery::ChargeStatus),
63    RecoveryEvent,
64    SmeTimeout,
65    ChipPowerUpFailure,
66    ResetTxPowerScenario,
67    SetTxPowerScenario {
68        scenario: fidl_internal::TxPowerScenario,
69    },
70}
71
72/// Attempts to connect to the Cobalt service.
73pub async fn setup_cobalt_proxy()
74-> Result<fidl_fuchsia_metrics::MetricEventLoggerProxy, anyhow::Error> {
75    let cobalt_svc = fuchsia_component::client::connect_to_protocol::<
76        fidl_fuchsia_metrics::MetricEventLoggerFactoryMarker,
77    >()
78    .context("failed to connect to metrics service")?;
79
80    let (cobalt_proxy, cobalt_server) =
81        fidl::endpoints::create_proxy::<fidl_fuchsia_metrics::MetricEventLoggerMarker>();
82
83    let project_spec = fidl_fuchsia_metrics::ProjectSpec {
84        customer_id: Some(metrics::CUSTOMER_ID),
85        project_id: Some(metrics::PROJECT_ID),
86        ..Default::default()
87    };
88
89    match cobalt_svc.create_metric_event_logger(&project_spec, cobalt_server).await {
90        Ok(_) => Ok(cobalt_proxy),
91        Err(err) => Err(format_err!("failed to create metrics event logger: {:?}", err)),
92    }
93}
94
95/// Attempts to create a disconnected FIDL channel with types matching the Cobalt service. This
96/// allows for a fallback with a uniform code path in case of a failure to connect to Cobalt.
97pub fn setup_disconnected_cobalt_proxy()
98-> Result<fidl_fuchsia_metrics::MetricEventLoggerProxy, anyhow::Error> {
99    // Create a disconnected proxy
100    Ok(fidl::endpoints::create_proxy::<fidl_fuchsia_metrics::MetricEventLoggerMarker>().0)
101}
102
103pub fn setup_persistence_req_sender()
104-> Result<(auto_persist::PersistenceReqSender, impl Future<Output = ()>), anyhow::Error> {
105    fuchsia_component::client::connect_to_protocol::<
106        fidl_fuchsia_diagnostics_persist::DataPersistenceMarker,
107    >()
108    .map(auto_persist::create_persistence_req_sender)
109}
110
111/// Creates a disconnected channel with the same types as the persistence service. This allows for
112/// a fallback with a uniform code path in case of a failure to connect to the persistence service.
113pub fn setup_disconnected_persistence_req_sender() -> auto_persist::PersistenceReqSender {
114    let (sender, _receiver) = mpsc::channel::<String>(1);
115    // Note: because we drop the receiver here, be careful about log spam when sending
116    // tags through the `sender` below. This is automatically handled by `auto_persist::AutoPersist`
117    // because it only logs the first time sending fails, so just use that wrapper type instead of
118    // writing directly to this channel.
119    sender
120}
121
122/// How often to refresh time series stats. Also how often to request packet counters.
123const TELEMETRY_QUERY_INTERVAL: zx::MonotonicDuration = zx::MonotonicDuration::from_seconds(10);
124
125pub fn serve_telemetry(
126    cobalt_proxy: fidl_fuchsia_metrics::MetricEventLoggerProxy,
127    monitor_svc_proxy: fidl_fuchsia_wlan_device_service::DeviceMonitorProxy,
128    inspect_node: InspectNode,
129    inspect_path: &str,
130    persistence_req_sender: auto_persist::PersistenceReqSender,
131) -> (TelemetrySender, impl Future<Output = Result<(), Error>>) {
132    let (sender, mut receiver) =
133        mpsc::channel::<TelemetryEvent>(util::sender::TELEMETRY_EVENT_BUFFER_SIZE);
134    let sender = TelemetrySender::new(sender);
135
136    // Inspect nodes to hold time series and metadata for other nodes
137    const METADATA_NODE_NAME: &str = "metadata";
138    let inspect_metadata_node = inspect_node.create_child(METADATA_NODE_NAME);
139    let inspect_metadata_path = format!("{inspect_path}/{METADATA_NODE_NAME}");
140    let inspect_time_series_node = inspect_node.create_child("time_series");
141    let driver_specific_time_series_node = inspect_time_series_node.create_child("driver_specific");
142    let driver_counters_time_series_node =
143        driver_specific_time_series_node.create_child("counters");
144    let driver_gauges_time_series_node = driver_specific_time_series_node.create_child("gauges");
145
146    let time_matrix_client = TimeMatrixClient::new(inspect_time_series_node.clone_weak());
147    let driver_counters_time_series_client =
148        TimeMatrixClient::new(driver_counters_time_series_node.clone_weak());
149    let driver_gauges_time_series_client =
150        TimeMatrixClient::new(driver_gauges_time_series_node.clone_weak());
151
152    // Create and initialize modules
153    let connect_disconnect = processors::connect_disconnect::ConnectDisconnectLogger::new(
154        cobalt_proxy.clone(),
155        &inspect_node,
156        &inspect_metadata_node,
157        &inspect_metadata_path,
158        persistence_req_sender,
159        &time_matrix_client,
160    );
161    let iface_logger = processors::iface::IfaceLogger::new(cobalt_proxy.clone());
162    let power_logger = processors::power::PowerLogger::new(cobalt_proxy.clone(), &inspect_node);
163    let recovery_logger = processors::recovery::RecoveryLogger::new(cobalt_proxy.clone());
164    let mut scan_logger =
165        processors::scan::ScanLogger::new(cobalt_proxy.clone(), &time_matrix_client);
166    let sme_timeout_logger = processors::sme_timeout::SmeTimeoutLogger::new(cobalt_proxy.clone());
167    let mut toggle_logger =
168        processors::toggle_events::ToggleLogger::new(cobalt_proxy.clone(), &inspect_node);
169    let tx_power_scenario_logger =
170        processors::tx_power_scenario::TxPowerScenarioLogger::new(cobalt_proxy.clone());
171
172    let client_iface_counters_logger =
173        processors::client_iface_counters::ClientIfaceCountersLogger::new(
174            cobalt_proxy,
175            monitor_svc_proxy,
176            &inspect_metadata_node,
177            &inspect_metadata_path,
178            &time_matrix_client,
179            driver_counters_time_series_client,
180            driver_gauges_time_series_client,
181        );
182
183    let fut = async move {
184        // Prevent the inspect nodes from being dropped while the loop is running.
185        let _inspect_node = inspect_node;
186        let _inspect_metadata_node = inspect_metadata_node;
187        let _inspect_time_series_node = inspect_time_series_node;
188        let _driver_specific_time_series_node = driver_specific_time_series_node;
189        let _driver_counters_time_series_node = driver_counters_time_series_node;
190        let _driver_gauges_time_series_node = driver_gauges_time_series_node;
191
192        let mut telemetry_interval = fasync::Interval::new(TELEMETRY_QUERY_INTERVAL);
193        loop {
194            select! {
195                event = receiver.next() => {
196                    let Some(event) = event else {
197                        error!("Telemetry event stream unexpectedly terminated.");
198                        return Err(format_err!("Telemetry event stream unexpectedly terminated."));
199                    };
200                    use TelemetryEvent::*;
201                    match event {
202                        ConnectResult { result, bss } => {
203                            connect_disconnect.handle_connect_attempt(result, &bss).await;
204                        }
205                        Disconnect { info } => {
206                            connect_disconnect.log_disconnect(&info).await;
207                            power_logger.handle_iface_disconnect(info.iface_id).await;
208                        }
209                        ClientConnectionsToggle { event } => {
210                            toggle_logger.handle_toggle_event(event).await;
211                        }
212                        ClientIfaceCreated { iface_id } => {
213                            client_iface_counters_logger.handle_iface_created(iface_id).await;
214                        }
215                        ClientIfaceDestroyed { iface_id } => {
216                            client_iface_counters_logger.handle_iface_destroyed(iface_id).await;
217                            power_logger.handle_iface_destroyed(iface_id).await;
218                        }
219                        IfaceCreationFailure => {
220                            iface_logger.handle_iface_creation_failure().await;
221                        }
222                        IfaceDestructionFailure => {
223                            iface_logger.handle_iface_destruction_failure().await;
224                        }
225                        ScanStart => {
226                            scan_logger.handle_scan_start().await;
227                        }
228                        ScanResult { result } => {
229                            scan_logger.handle_scan_result(result).await;
230                        }
231                        IfacePowerLevelChanged { iface_power_level, iface_id } => {
232                            power_logger.log_iface_power_event(iface_power_level, iface_id).await;
233                        }
234                        // TODO(b/340921554): either watch for suspension directly in the library,
235                        // or plumb this from callers once suspend mechanisms are integrated
236                        SuspendImminent => {
237                            power_logger.handle_suspend_imminent().await;
238                            connect_disconnect.handle_suspend_imminent().await;
239                        }
240                        UnclearPowerDemand(demand) => {
241                            power_logger.handle_unclear_power_demand(demand).await;
242                        }
243                        ChipPowerUpFailure => {
244                            power_logger.handle_chip_power_up_failure().await;
245                        }
246                        BatteryChargeStatus(charge_status) => {
247                            scan_logger.handle_battery_charge_status(charge_status).await;
248                            toggle_logger.handle_battery_charge_status(charge_status).await;
249                        }
250                        RecoveryEvent => {
251                            recovery_logger.handle_recovery_event().await;
252                        }
253                        SmeTimeout => {
254                            sme_timeout_logger.handle_sme_timeout_event().await;
255                        }
256                        ResetTxPowerScenario => {
257                            tx_power_scenario_logger.handle_sar_reset().await;
258                        }
259                        SetTxPowerScenario {scenario} => {
260                            tx_power_scenario_logger.handle_set_sar(scenario).await;
261                        }
262                    }
263                }
264                _ = telemetry_interval.next() => {
265                    connect_disconnect.handle_periodic_telemetry().await;
266                    client_iface_counters_logger.handle_periodic_telemetry().await;
267                }
268            }
269        }
270    };
271    (sender, fut)
272}