diff --git a/Cargo.lock b/Cargo.lock index 38b0b4fa..b9c4d46b 100644 --- a/Cargo.lock +++ b/Cargo.lock @@ -637,6 +637,7 @@ dependencies = [ "databroker-proto", "futures", "glob-match", + "indexmap 2.8.0", "jemallocator", "jsonwebtoken", "kuksa", @@ -1594,7 +1595,7 @@ dependencies = [ "futures-sink", "futures-util", "http", - "indexmap 2.7.1", + "indexmap 2.8.0", "slab", "tokio", "tokio-util", @@ -1910,9 +1911,9 @@ dependencies = [ [[package]] name = "indexmap" -version = "2.7.1" +version = "2.8.0" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "8c9c992b02b5b4c94ea26e32fe5bccb7aa7d9f390ab5c1221ff895bc7ea8b652" +checksum = "3954d50fe15b02142bf25d3b8bdadb634ec3948f103d04ffe3031bc8fe9d7058" dependencies = [ "equivalent", "hashbrown 0.15.2", @@ -2513,7 +2514,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index" checksum = "b4c5cc86750666a3ed20bdaf5ca2a0344f9c67674cae0515bec2da16fbaa47db" dependencies = [ "fixedbitset", - "indexmap 2.7.1", + "indexmap 2.8.0", ] [[package]] diff --git a/databroker/Cargo.toml b/databroker/Cargo.toml index 266daf58..27cd7279 100644 --- a/databroker/Cargo.toml +++ b/databroker/Cargo.toml @@ -57,6 +57,7 @@ serde_json = "1.0" jsonwebtoken = "9.1.0" regex = "1.7.1" glob-match = "0.2.1" +indexmap = "2.8.0" jemallocator = { version = "0.5.0", optional = true } lazy_static = "1.4.0" @@ -67,7 +68,7 @@ async-trait = "0.1.82" # VISS axum = { version = "0.6.20", optional = true, features = ["ws"] } chrono = { version = "0.4.31", optional = true, features = ["std"] } -uuid = { version = "1.4.1", optional = true, features = ["v4"] } +uuid = { version = "1.4.1", features = ["v4"] } # OTEL opentelemetry = { version = "0.19.0", optional = true, features = ["rt-tokio", "trace"] } @@ -83,7 +84,7 @@ sd-notify = "0.4.1" default = ["tls"] tls = ["tonic/tls", "kuksa-common/tls", "kuksa/tls"] jemalloc = ["dep:jemallocator"] -viss = ["dep:axum", "dep:chrono", "dep:uuid"] +viss = ["dep:axum", "dep:chrono"] libtest = [] otel = ["dep:chrono", "dep:opentelemetry", "dep:opentelemetry-otlp", "dep:opentelemetry-semantic-conventions", "dep:tracing-opentelemetry"] diff --git a/databroker/src/broker.rs b/databroker/src/broker.rs index 78d691ae..2ff474f5 100644 --- a/databroker/src/broker.rs +++ b/databroker/src/broker.rs @@ -11,22 +11,25 @@ * SPDX-License-Identifier: Apache-2.0 ********************************************************************************/ +use crate::filter::filter_manager::FilterManager; use crate::permissions::{PermissionError, Permissions}; pub use crate::types; use crate::query; -pub use crate::types::{ChangeType, DataType, DataValue, EntryType}; +pub use crate::types::{ChangeType, DataType, DataValue, EntryType, SignalId, TimeInterval}; +use indexmap::IndexMap; use tokio::sync::{broadcast, mpsc, RwLock}; -use tokio_stream::wrappers::BroadcastStream; -use tokio_stream::wrappers::ReceiverStream; +use tokio::time::Instant; +use tokio_stream::wrappers::{BroadcastStream, ReceiverStream}; use tokio_stream::{Stream, StreamExt}; use std::collections::{HashMap, HashSet}; use std::convert::TryFrom; use std::sync::atomic::{AtomicI32, Ordering}; use std::sync::Arc; -use std::time::SystemTime; +use std::time::{Duration, SystemTime}; +use uuid::Uuid; use crate::query::{CompiledQuery, ExecutionInput}; use crate::types::ExecutionInputImplData; @@ -49,6 +52,15 @@ pub enum ActuationError { TransmissionFailure, } +#[derive(Debug)] +pub enum RegisterSignalError { + NotFound, + PermissionDenied, + PermissionExpired, + SignalAlreadyRegistered, + TransmissionFailure, +} + #[derive(Debug, PartialEq)] pub enum UpdateError { NotFound, @@ -124,7 +136,8 @@ pub struct Database { pub struct Subscriptions { actuation_subscriptions: Vec, query_subscriptions: Vec, - change_subscriptions: Vec, + change_subscriptions: HashMap, + signal_provider_subscriptions: HashMap, } #[derive(Debug, Clone)] @@ -171,6 +184,7 @@ pub struct DataBroker { version: String, commit_sha: String, shutdown_trigger: broadcast::Sender<()>, + filter_manager: Arc>, } #[async_trait::async_trait] @@ -182,6 +196,19 @@ pub trait ActuationProvider { fn is_available(&self) -> bool; } +#[async_trait::async_trait] +pub trait SignalProvider: Send + Sync + 'static { + async fn update_filter( + &self, + update_fiters: HashMap>, + ) -> Result<(), (RegisterSignalError, String)>; + fn is_available(&self) -> bool; + async fn get_signals_values_from_provider( + &mut self, + signals_ids: Vec, + ) -> Result; +} + #[derive(Clone)] pub struct ActuationChange { pub id: i32, @@ -194,6 +221,16 @@ pub struct ActuationSubscription { permissions: Permissions, } +pub struct GetValuesProviderResponse { + pub entries: IndexMap, +} + +pub struct SignalProviderSubscription { + vss_ids: HashSet, + signal_provider: Box, + permissions: Permissions, +} + pub struct QuerySubscription { query: query::CompiledQuery, sender: mpsc::Sender, @@ -202,8 +239,10 @@ pub struct QuerySubscription { pub struct ChangeSubscription { entries: HashMap>, - sender: broadcast::Sender, + sender: broadcast::Sender>, permissions: Permissions, + interval_duration: Option, + last_emitted: Arc>, } #[derive(Debug)] @@ -757,8 +796,20 @@ impl Subscriptions { } #[cfg_attr(feature="otel", tracing::instrument(name="subscriptions_add_change_subscription",skip(self, subscription), fields(timestamp=chrono::Utc::now().to_string())))] - pub fn add_change_subscription(&mut self, subscription: ChangeSubscription) { - self.change_subscriptions.push(subscription) + pub fn add_change_subscription(&mut self, subscription: ChangeSubscription) -> Uuid { + let uuid = Uuid::new_v4(); + self.change_subscriptions.insert(uuid, subscription); + uuid + } + + pub fn add_signal_provider_subscription( + &mut self, + subscription: SignalProviderSubscription, + ) -> Uuid { + let uuid = Uuid::new_v4(); + self.signal_provider_subscriptions + .insert(uuid, subscription); + uuid } #[cfg_attr( @@ -786,7 +837,7 @@ impl Subscriptions { } } - for sub in &self.change_subscriptions { + for sub in self.change_subscriptions.values() { match sub.notify(changed, db).await { Ok(_) => {} Err(err) => error = Some(err), @@ -821,29 +872,57 @@ impl Subscriptions { true } }); - self.change_subscriptions.retain(|sub| { - if sub.sender.receiver_count() == 0 { - info!("Subscriber gone: removing subscription"); + + self.actuation_subscriptions.retain(|sub| { + if !sub.actuation_provider.is_available() { + info!("Actuation Provider gone: removing provided actuation"); false } else if sub.permissions.is_expired() { - info!("Permissions of Subscriber expired: removing subscription"); + info!("Permissions of Provider expired: removing provided actuation"); false } else { true } }); + } - self.actuation_subscriptions.retain(|sub| { - if !sub.actuation_provider.is_available() { - info!("Provider gone: removing provided actuation"); + #[cfg_attr(feature="otel", tracing::instrument(name="signal_provider_subscriptions_cleanup", skip(self), fields(timestamp=chrono::Utc::now().to_string())))] + pub fn cleanup_signal_providers_subscriptions(&mut self) -> HashMap> { + let mut closed_signal_providers: HashMap> = HashMap::new(); + self.signal_provider_subscriptions + .retain(|provider_uuid, signal_provider| { + if !signal_provider.signal_provider.is_available() { + closed_signal_providers.insert(*provider_uuid, signal_provider.vss_ids.clone()); + info!("Singal Provider gone: removing provided"); + false + } else if signal_provider.permissions.is_expired() { + closed_signal_providers.insert(*provider_uuid, signal_provider.vss_ids.clone()); + info!("Permissions of Provider expired: removing provided"); + false + } else { + true + } + }); + closed_signal_providers + } + + #[cfg_attr(feature="otel", tracing::instrument(name="change_subscriptions_cleanup", skip(self), fields(timestamp=chrono::Utc::now().to_string())))] + pub fn cleanup_change_subscriptions(&mut self) -> Vec { + let mut closed_subscriptions_uuids: Vec = Vec::new(); + self.change_subscriptions.retain(|uuid, sub| { + if sub.sender.receiver_count() == 0 { + closed_subscriptions_uuids.push(*uuid); + info!("Subscriber gone: removing subscription"); false } else if sub.permissions.is_expired() { - info!("Permissions of Provider expired: removing provided actuation"); + closed_subscriptions_uuids.push(*uuid); + info!("Permissions of Subscriber expired: removing subscription"); false } else { true } }); + closed_subscriptions_uuids } } @@ -857,6 +936,13 @@ impl ChangeSubscription { changed: Option<&HashMap>>, db: &Database, ) -> Result<(), NotificationError> { + if self.interval_duration.is_some() + && self.last_emitted.read().await.elapsed().as_millis() + < self.interval_duration.unwrap().as_millis() + { + return Ok(()); + } + let db_read = db.authorized_read_access(&self.permissions); match changed { Some(changed) => { @@ -919,8 +1005,12 @@ impl ChangeSubscription { if notifications.updates.is_empty() { Ok(()) } else { - match self.sender.send(notifications) { - Ok(_number_of_receivers) => Ok(()), + match self.sender.send(Some(notifications)) { + Ok(_number_of_receivers) => { + let mut last_emitted = self.last_emitted.write().await; + *last_emitted = Instant::now(); + Ok(()) + } Err(err) => { debug!("Send error for entry{}: ", err); Err(NotificationError {}) @@ -963,8 +1053,12 @@ impl ChangeSubscription { } notifications }; - match self.sender.send(notifications) { - Ok(_number_of_receivers) => Ok(()), + match self.sender.send(Some(notifications)) { + Ok(_number_of_receivers) => { + let mut last_emitted = self.last_emitted.write().await; + *last_emitted = Instant::now(); + Ok(()) + } Err(err) => { debug!("Send error for entry{}: ", err); Err(NotificationError {}) @@ -973,6 +1067,9 @@ impl ChangeSubscription { } } } + async fn send_none_to_receiver(&self) { + let _ = self.sender.send(None); + } } impl QuerySubscription { @@ -1670,7 +1767,8 @@ impl AuthorizedAccess<'_, '_> { &self, valid_entries: HashMap>, buffer_size: Option, - ) -> Result, SubscriptionError> { + interval_ms: Option, + ) -> Result>, SubscriptionError> { if valid_entries.is_empty() { return Err(SubscriptionError::InvalidInput); } @@ -1686,11 +1784,21 @@ impl AuthorizedAccess<'_, '_> { 1 }; + let (time_interval, interval_duration) = match interval_ms { + Some(milliseconds) => ( + Some(TimeInterval::new(milliseconds)), + Some(Duration::from_millis(milliseconds.into())), + ), + None => (None, None), + }; + let (sender, receiver) = broadcast::channel(channel_capacity); let subscription = ChangeSubscription { - entries: valid_entries, + entries: valid_entries.clone(), sender, permissions: self.permissions.clone(), + interval_duration, + last_emitted: Arc::new(RwLock::new(Instant::now())), }; { @@ -1701,14 +1809,28 @@ impl AuthorizedAccess<'_, '_> { } } - self.broker + let uuid_subscription = self + .broker .subscriptions .write() .await .add_change_subscription(subscription); - let stream = BroadcastStream::new(receiver).filter_map(move |result| match result { - Ok(message) => Some(message), + if time_interval.is_some() { + let _ = self + .propagate_new_filter_to_provider( + time_interval.unwrap(), + valid_entries + .keys() + .map(|signal_id| SignalId::new(*signal_id)) + .collect(), + uuid_subscription, + ) + .await; + } + + let stream = BroadcastStream::new(receiver).map(move |result| match result { + Ok(message) => message, Err(err) => { warn!( "Slow subscriber with capacity {} lagged and missed signal updates: {}", @@ -2048,6 +2170,198 @@ impl AuthorizedAccess<'_, '_> { )), } } + + async fn propagate_new_filter_to_provider( + &self, + time_interval: TimeInterval, + signal_ids: Vec, + uuid_subscription: Uuid, + ) { + let new_update_filter = self + .broker + .filter_manager + .write() + .await + .add_new_update_filter(signal_ids, time_interval, uuid_subscription) + .into_iter() + .map(|(key, value)| (key, Some(value))) + .collect(); + + DataBroker::update_filter_to_providers( + new_update_filter, + &self + .broker + .subscriptions + .read() + .await + .signal_provider_subscriptions, + ) + .await; + } + + pub async fn register_signals( + &self, + vss_ids: HashSet, + signal_provider: Box, + ) -> Result { + let registered_vss_ids: HashSet = self + .broker + .subscriptions + .read() + .await + .signal_provider_subscriptions + .iter() + .flat_map(|provider_entry| provider_entry.1.vss_ids.clone()) + .collect(); + + let intersection: HashSet<&SignalId> = vss_ids + .iter() + .filter(|&x| registered_vss_ids.contains(x)) + .collect(); + if !intersection.is_empty() { + let message = format!( + "Providers for the following vss_ids already registered: {:?}", + intersection + ); + return Err((RegisterSignalError::SignalAlreadyRegistered, message)); + } + + let signal_subscription = SignalProviderSubscription { + vss_ids, + signal_provider, + permissions: self.permissions.clone(), + }; + let provider_uuid = self + .broker + .subscriptions + .write() + .await + .add_signal_provider_subscription(signal_subscription); + + Ok(provider_uuid) + } + + pub async fn publish_provider_error( + &self, + provider_uuid: Uuid, + ) -> Result<(), Vec<(i32, UpdateError)>> { + let subscriptions = self.broker.subscriptions.read().await; + let key_set = &subscriptions + .signal_provider_subscriptions + .get(&provider_uuid) + .as_ref() + .unwrap() + .vss_ids; + + let entries_updates: Vec<(i32, EntryUpdate)> = key_set + .iter() + .map(|signal_id| { + ( + signal_id.id(), + EntryUpdate { + path: None, + datapoint: None, + actuator_target: None, + entry_type: None, + data_type: None, + description: None, + allowed: None, + min: None, + max: None, + unit: None, + }, + ) + }) + .collect(); + self.update_entries(entries_updates).await + } + + pub async fn valid_provider_publish_signals( + &self, + provider_uuid: Uuid, + signal_request: &HashSet, + ) -> bool { + let subscriptions = self.broker.subscriptions.read().await; + let key_set = &subscriptions + .signal_provider_subscriptions + .get(&provider_uuid) + .as_ref() + .unwrap() + .vss_ids; + key_set.is_subset(signal_request) + } + + pub async fn get_values_broker( + &self, + vss_signals: Vec, + ) -> Result { + let mut subscriptions = self.broker.subscriptions.write().await; + let mut entries_response: IndexMap = IndexMap::new(); + + // if there are providers connected then forward the get value request to them + if !subscriptions.signal_provider_subscriptions.is_empty() { + // Step 1: find the interseccion of the requested signals and each provider's signals + for (_, provider) in subscriptions.signal_provider_subscriptions.iter_mut() { + let intersection_signals_request: Vec = provider + .vss_ids + .intersection(&vss_signals.clone().into_iter().collect()) + .copied() + .collect(); + + //Step 2: check if there is any possible ReadError while reading any of the signals + for signal_id in &intersection_signals_request { + match self + .broker + .database + .read() + .await + .authorized_read_access(self.permissions) + .get_entry_by_id(signal_id.id()) + { + Ok(_) => {} + Err(err) => return Err((err, signal_id.id())), + } + } + + //Step 3: send to each provider the intersection signals requested + let response = provider + .signal_provider + .get_signals_values_from_provider(intersection_signals_request.clone()) + .await; + + //Step 4: collect responses from providers + match response { + Ok(value) => { + entries_response.extend(value.entries); + } + Err(_) => { + //Step 5: if provider did not return any item, then return the datapoint in database + for signal_id in &intersection_signals_request { + match self.get_datapoint(signal_id.id()).await { + Ok(datapoint) => { + entries_response.insert(*signal_id, datapoint.clone()); + } + Err(err) => return Err((err, signal_id.id())), + } + } + } + } + } + } else { + // if not just send to the client database last database values + for signal_id in &vss_signals { + match self.get_datapoint(signal_id.id()).await { + Ok(datapoint) => { + entries_response.insert(*signal_id, datapoint.clone()); + } + Err(err) => return Err((err, signal_id.id())), + } + } + } + Ok(GetValuesProviderResponse { + entries: entries_response, + }) + } } impl DataBroker { @@ -2060,6 +2374,7 @@ impl DataBroker { version: version.into(), commit_sha: commit_sha.into(), shutdown_trigger, + filter_manager: Default::default(), } } @@ -2077,18 +2392,115 @@ impl DataBroker { pub fn start_housekeeping_task(&self) { info!("Starting housekeeping task"); let subscriptions = self.subscriptions.clone(); + let filter_manager = self.filter_manager.clone(); tokio::spawn(async move { let mut interval = tokio::time::interval(std::time::Duration::from_secs(1)); + let mut connected_providers_count = 0; + loop { interval.tick().await; subscriptions.write().await.cleanup(); // Cleanup dropped subscriptions + + // clean up disconnected providers + let closed_signal_providers = subscriptions + .write() + .await + .cleanup_signal_providers_subscriptions(); + + if !closed_signal_providers.is_empty() { + // Inform remaining subscriptions about not available providers + { + let remaining_subscriptions = + &mut subscriptions.write().await.change_subscriptions; + + // check which subscription contains a signal of the closed providers: + for (_, provider_signals_set) in closed_signal_providers { + for (_, subscription) in remaining_subscriptions.iter_mut() { + let key_set: HashSet<_> = + subscription.entries.keys().cloned().collect(); + if key_set.is_subset( + &provider_signals_set + .iter() + .map(|signal| signal.id()) + .collect(), + ) { + subscription.send_none_to_receiver().await; + } + } + + // TODO + //set disconnected provider's signals to none and update subscription with a notify + } + } + } + + // clean up disconnected subscriptions + let closed_change_subscriptions = + subscriptions.write().await.cleanup_change_subscriptions(); + + let new_connected_providers_count = subscriptions + .read() + .await + .signal_provider_subscriptions + .len(); + + // If closed subscription update new filters to providers or new provider connected + if !closed_change_subscriptions.is_empty() + || new_connected_providers_count > connected_providers_count + { + if new_connected_providers_count > connected_providers_count { + connected_providers_count += 1; + } + let new_update_filter = filter_manager + .write() + .await + .remove_filter_by_subscription_uuid(closed_change_subscriptions); + + Self::update_filter_to_providers( + new_update_filter, + &subscriptions.read().await.signal_provider_subscriptions, + ) + .await; + } + + if new_connected_providers_count < connected_providers_count { + connected_providers_count -= 1; + } } }); } + async fn update_filter_to_providers( + update_signal_intervals: HashMap>, + providers: &HashMap, + ) { + for provider in providers.values() { + // Find which provider contains any of the signals to be updated with new interval. + let update_signal_map: HashMap> = provider + .vss_ids + .iter() + .filter_map(|&signal_id| update_signal_intervals.get_key_value(&signal_id)) + .map(|(&k, v)| { + let signal_id = k; + let interval_ms = *v; + (signal_id, interval_ms) + }) + .collect(); + + // Send to each provider the new (signal_id, new_interval) map. + if !update_signal_map.is_empty() { + provider + .signal_provider + .update_filter(update_signal_map) + .await + .unwrap(); + } + } + } + pub async fn shutdown(&self) { // Drain subscriptions let mut subscriptions = self.subscriptions.write().await; @@ -4305,6 +4717,7 @@ pub mod tests { .subscribe( HashMap::from([(id1, HashSet::from([Field::Datapoint]))]), buffer_size, + None, ) .await .expect("subscription should succeed"); @@ -4312,15 +4725,17 @@ pub mod tests { // Stream should yield initial notification with current values i.e. NotAvailable match stream.next().await { Some(entry) => { - assert_eq!(entry.updates.len(), 1); - assert_eq!( - entry.updates[0].update.path, - Some("test.datapoint1".to_string()) - ); - assert_eq!( - entry.updates[0].update.datapoint.as_ref().unwrap().value, - DataValue::NotAvailable - ); + if let Some(value) = entry { + assert_eq!(value.updates.len(), 1); + assert_eq!( + value.updates[0].update.path, + Some("test.datapoint1".to_string()) + ); + assert_eq!( + value.updates[0].update.datapoint.as_ref().unwrap().value, + DataValue::NotAvailable + ); + } } None => { panic!("did not expect stream end") @@ -4353,15 +4768,17 @@ pub mod tests { // Value has been set, expect the next item in stream to match. match stream.next().await { Some(entry) => { - assert_eq!(entry.updates.len(), 1); - assert_eq!( - entry.updates[0].update.path, - Some("test.datapoint1".to_string()) - ); - assert_eq!( - entry.updates[0].update.datapoint.as_ref().unwrap().value, - DataValue::Int32(101) - ); + if let Some(value) = entry { + assert_eq!(value.updates.len(), 1); + assert_eq!( + value.updates[0].update.path, + Some("test.datapoint1".to_string()) + ); + assert_eq!( + value.updates[0].update.datapoint.as_ref().unwrap().value, + DataValue::Int32(101) + ); + } } None => { panic!("did not expect stream end") @@ -4415,6 +4832,7 @@ pub mod tests { HashMap::from([(id1, HashSet::from([Field::Datapoint]))]), // 1001 is just outside valid range 0-1000 Some(1001), + None, ) .await { diff --git a/databroker/src/filter/filter_manager.rs b/databroker/src/filter/filter_manager.rs new file mode 100644 index 00000000..6d47d05c --- /dev/null +++ b/databroker/src/filter/filter_manager.rs @@ -0,0 +1,256 @@ +/******************************************************************************** +* Copyright (c) 2025 Contributors to the Eclipse Foundation +* +* See the NOTICE file(s) distributed with this work for additional +* information regarding copyright ownership. +* +* This program and the accompanying materials are made available under the +* terms of the Apache License 2.0 which is available at +* http://www.apache.org/licenses/LICENSE-2.0 +* +* SPDX-License-Identifier: Apache-2.0 +********************************************************************************/ + +use std::collections::{BTreeSet, HashMap}; +use uuid::Uuid; + +use crate::types::{SignalId, TimeInterval}; + +type SubscriptionUuid = Uuid; +/// +/// FilterManager: +/// Contains a small database with the signal_id, interval_ms and subscription_uuid +/// +#[derive(Default)] +pub struct FilterManager { + database: HashMap>, +} + +impl FilterManager { + /// + /// Returns a HashMap containing each signal_id present in the map, + /// along with its corresponding lowest associated filter. + /// + fn get_lowest_filter_interval_per_signal(&self) -> HashMap { + self.database + .clone() + .into_iter() + .filter_map(|(signal_id, set)| { + set.first() + .map(|&(lowest_interval, _)| (signal_id, lowest_interval)) + }) + .collect() + } + + /// + /// Add a new filter to each signal_id in the database and returns + /// the new update filter map containing ONLY signals_ids along with + /// their new lowest interval_ms. + /// + pub fn add_new_update_filter( + &mut self, + signal_ids: Vec, + sample_interval: TimeInterval, + subscription_uuid: Uuid, + ) -> HashMap { + // Get the signals ids and their intervals before updating anything. + let current_lowest_interval_per_signal = self.get_lowest_filter_interval_per_signal(); + + // Insert new pair of (sample_interval, subscription_uuid) for each signal_id + for signal_id in signal_ids { + self.database + .entry(signal_id) + .or_default() + .insert((sample_interval, subscription_uuid)); + } + + // Get the signals ids and their intervals after updating anything. + let updated_lowest_interval_per_signal = self.get_lowest_filter_interval_per_signal(); + + // Return only the signals whose lowest interval_ms as changed. + let update_only_new_interval_filter = if !current_lowest_interval_per_signal.is_empty() { + updated_lowest_interval_per_signal + .into_iter() + .filter_map(|(signal_id, lowest_interval)| { + let value = current_lowest_interval_per_signal.get(&signal_id); + match value { + Some(sample_interval) => { + if *sample_interval != lowest_interval { + Some((signal_id, lowest_interval)) + } else { + None + } + } + None => Some((signal_id, lowest_interval)), + } + }) + .collect() + } else { + updated_lowest_interval_per_signal + }; + update_only_new_interval_filter + } + + /// + /// Remove all intervals associated to an subscription_uuid and return + /// the new update filter map containing ONLY signals_ids along with + /// their new lowest interval_ms. + /// + pub fn remove_filter_by_subscription_uuid( + &mut self, + target: Vec, + ) -> HashMap> { + // Get the signals ids and their intervals before removing anything. + let current_lowest_interval_per_signal = self.get_lowest_filter_interval_per_signal(); + + // Remove intervals and/or signals by uuid + for target_uuid in &target { + // Collect keys whose BTreeSet becomes empty after removal. + let mut empty_keys = Vec::new(); + + // Iterate over each key and its associated BTreeSet. + for (key, set) in self.database.iter_mut() { + // Find the tuple in the BTreeSet that has the matching Uuid. + if let Some(item) = set.iter().find(|(_, uuid)| *uuid == *target_uuid).cloned() { + // Remove the found tuple. + set.remove(&item); + // If that BTreeSet is now empty, mark this key for removal. + if set.is_empty() { + empty_keys.push(*key); + } + } + } + + // Remove keys that have empty BTreeSets. + for key in empty_keys { + self.database.remove(&key); + } + } + + // Get the signals ids and their intervals after removing intervals. + let updated_lowest_interval_per_signal = self.get_lowest_filter_interval_per_signal(); + + // Return only the signals whose lowest interval_ms as changed. + let update_only_new_interval_filter = if !target.is_empty() { + current_lowest_interval_per_signal + .into_iter() + .filter_map(|(signal_id, lowest_inteval)| { + let value = updated_lowest_interval_per_signal.get(&signal_id); + match value { + Some(sample_interval) => { + if *sample_interval != lowest_inteval { + Some((signal_id, Some(*sample_interval))) + } else { + None + } + } + None => Some((signal_id, None)), + } + }) + .collect() + } else { + updated_lowest_interval_per_signal + .into_iter() + .map(|(key, value)| (key, Some(value))) + .collect() + }; + update_only_new_interval_filter + } +} + +#[cfg(test)] +pub mod tests { + use super::*; + + #[tokio::test] + async fn add_filter() { + let mut filter_manager = FilterManager::default(); + + let (singal_id_1, sample_interval_1) = (SignalId::new(1), TimeInterval::new(10)); + let (singal_id_2, sample_interval_2) = (SignalId::new(2), TimeInterval::new(100)); + let (singal_id_3, sample_interval_3) = (SignalId::new(3), TimeInterval::new(1000)); + let (all_singals_ids, min_sample_interval) = ( + vec![singal_id_1, singal_id_2, singal_id_3], + TimeInterval::new(1), + ); + + filter_manager.add_new_update_filter(vec![singal_id_1], sample_interval_1, Uuid::default()); + filter_manager.add_new_update_filter(vec![singal_id_2], sample_interval_2, Uuid::default()); + filter_manager.add_new_update_filter(vec![singal_id_3], sample_interval_3, Uuid::default()); + + let updated_filter_with_lowest_interval = filter_manager.add_new_update_filter( + all_singals_ids, + min_sample_interval, + Uuid::default(), + ); + + assert_eq!( + updated_filter_with_lowest_interval, + HashMap::from([ + (singal_id_1, min_sample_interval), + (singal_id_2, min_sample_interval), + (singal_id_3, min_sample_interval) + ]) + ); + } + + #[tokio::test] + async fn remove_filter() { + let mut filter_manager = FilterManager::default(); + + let (singal_id_1, sample_interval_1, sub_1) = + (SignalId::new(1), TimeInterval::new(10), Uuid::new_v4()); + let (singal_id_2, sample_interval_2, sub_2) = + (SignalId::new(2), TimeInterval::new(100), Uuid::new_v4()); + let (singal_id_3, sample_interval_3, sub_3) = + (SignalId::new(3), TimeInterval::new(1000), Uuid::new_v4()); + let (all_singals_ids, min_sample_interval, all_uuid) = ( + vec![singal_id_1, singal_id_2, singal_id_3], + TimeInterval::new(1), + Uuid::new_v4(), + ); + + filter_manager.add_new_update_filter(vec![singal_id_1], sample_interval_1, sub_1); + filter_manager.add_new_update_filter(vec![singal_id_2], sample_interval_2, sub_2); + filter_manager.add_new_update_filter(vec![singal_id_3], sample_interval_3, sub_3); + + let updated_filter_with_lowest_interval = + filter_manager.add_new_update_filter(all_singals_ids, min_sample_interval, all_uuid); + + assert_eq!( + updated_filter_with_lowest_interval, + HashMap::from([ + (singal_id_1, min_sample_interval), + (singal_id_2, min_sample_interval), + (singal_id_3, min_sample_interval) + ]) + ); + let update_filter_after_remove = + filter_manager.remove_filter_by_subscription_uuid(vec![all_uuid]); + + assert_eq!( + update_filter_after_remove, + HashMap::from([ + (singal_id_1, Some(sample_interval_1)), + (singal_id_2, Some(sample_interval_2)), + (singal_id_3, Some(sample_interval_3)) + ]) + ); + + let update_filter_after_remove_sub1_and_sub2 = + filter_manager.remove_filter_by_subscription_uuid(vec![sub_1, sub_2]); + + assert_eq!( + update_filter_after_remove_sub1_and_sub2, + HashMap::from([(singal_id_1, None), (singal_id_2, None)]) + ); + + let update_filter_after_remove_sub3 = + filter_manager.remove_filter_by_subscription_uuid(vec![sub_3]); + + assert_eq!( + update_filter_after_remove_sub3, + HashMap::from([(singal_id_3, None)]) + ); + } +} diff --git a/databroker/src/filter/mod.rs b/databroker/src/filter/mod.rs new file mode 100644 index 00000000..9ace61e8 --- /dev/null +++ b/databroker/src/filter/mod.rs @@ -0,0 +1,14 @@ +/******************************************************************************** +* Copyright (c) 2025 Contributors to the Eclipse Foundation +* +* See the NOTICE file(s) distributed with this work for additional +* information regarding copyright ownership. +* +* This program and the accompanying materials are made available under the +* terms of the Apache License 2.0 which is available at +* http://www.apache.org/licenses/LICENSE-2.0 +* +* SPDX-License-Identifier: Apache-2.0 +********************************************************************************/ + +pub mod filter_manager; diff --git a/databroker/src/grpc/kuksa_val_v1/val.rs b/databroker/src/grpc/kuksa_val_v1/val.rs index e35c8a24..2cf3dc0f 100644 --- a/databroker/src/grpc/kuksa_val_v1/val.rs +++ b/databroker/src/grpc/kuksa_val_v1/val.rs @@ -610,7 +610,7 @@ impl proto::val_server::Val for broker::DataBroker { } } - match broker.subscribe(entries, None).await { + match broker.subscribe(entries, None, None).await { Ok(stream) => { let stream = convert_to_proto_stream(stream); Ok(tonic::Response::new(Box::pin(stream))) @@ -748,22 +748,27 @@ fn convert_to_data_entry_error(path: &String, error: &broker::UpdateError) -> Da #[cfg_attr(feature="otel", tracing::instrument(name="kuksa_val_v1_convert_to_proto_stream", skip(input), fields(timestamp=chrono::Utc::now().to_string())))] fn convert_to_proto_stream( - input: impl Stream, + input: impl Stream>, ) -> impl Stream> { - input.map(move |item| { - let mut updates = Vec::new(); - for update in item.updates { - updates.push(proto::EntryUpdate { - entry: Some(proto::DataEntry::from(update.update)), - fields: update - .fields - .iter() - .map(|field| proto::Field::from(field) as i32) - .collect(), - }); + input.map(move |item| match item { + Some(entry) => { + let mut updates = Vec::new(); + for update in entry.updates { + updates.push(proto::EntryUpdate { + entry: Some(proto::DataEntry::from(update.update)), + fields: update + .fields + .iter() + .map(|field| proto::Field::from(field) as i32) + .collect(), + }); + } + let response = proto::SubscribeResponse { updates }; + Ok(response) + } + None => { + todo!() } - let response = proto::SubscribeResponse { updates }; - Ok(response) }) } diff --git a/databroker/src/grpc/kuksa_val_v2/conversions.rs b/databroker/src/grpc/kuksa_val_v2/conversions.rs index 1affebb0..14bcef3c 100644 --- a/databroker/src/grpc/kuksa_val_v2/conversions.rs +++ b/databroker/src/grpc/kuksa_val_v2/conversions.rs @@ -538,3 +538,21 @@ impl broker::ActuationError { } } } + +impl broker::RegisterSignalError { + pub fn to_tonic_status(&self, message: String) -> tonic::Status { + match self { + broker::RegisterSignalError::NotFound => tonic::Status::not_found(message), + broker::RegisterSignalError::PermissionDenied => { + tonic::Status::permission_denied(message) + } + broker::RegisterSignalError::PermissionExpired => { + tonic::Status::unauthenticated(message) + } + broker::RegisterSignalError::SignalAlreadyRegistered => { + tonic::Status::already_exists(message) + } + broker::RegisterSignalError::TransmissionFailure => tonic::Status::data_loss(message), + } + } +} diff --git a/databroker/src/grpc/kuksa_val_v2/val.rs b/databroker/src/grpc/kuksa_val_v2/val.rs index f6f0de3c..0931c1bd 100644 --- a/databroker/src/grpc/kuksa_val_v2/val.rs +++ b/databroker/src/grpc/kuksa_val_v2/val.rs @@ -11,15 +11,18 @@ * SPDX-License-Identifier: Apache-2.0 ********************************************************************************/ +use indexmap::IndexMap; use std::{collections::HashMap, pin::Pin}; +use uuid::Uuid; use crate::{ broker::{ - self, ActuationChange, ActuationProvider, AuthorizedAccess, ReadError, SubscriptionError, + self, ActuationChange, ActuationProvider, AuthorizedAccess, GetValuesProviderResponse, + ReadError, RegisterSignalError, SignalProvider, SubscriptionError, }, glob::Matcher, permissions::Permissions, - types::DataValue, + types::{DataValue, SignalId, TimeInterval}, }; use databroker_proto::kuksa::val::v2::{ @@ -33,17 +36,25 @@ use databroker_proto::kuksa::val::v2::{ use kuksa::proto::v2::{ signal_id, ActuateRequest, ActuateResponse, BatchActuateStreamRequest, ErrorCode, - ListMetadataResponse, ProvideActuationResponse, + ListMetadataResponse, ProvideActuationResponse, ProvideSignalResponse, }; use std::collections::HashSet; -use tokio::{select, sync::mpsc}; -use tokio_stream::{wrappers::ReceiverStream, Stream, StreamExt}; +use tokio::{ + select, + sync::{broadcast, mpsc}, + time::timeout, +}; +use tokio_stream::{ + wrappers::{BroadcastStream, ReceiverStream}, + Stream, StreamExt, +}; use tracing::debug; const MAX_REQUEST_PATH_LENGTH: usize = 1000; pub struct Provider { sender: mpsc::Sender>, + receiver: Option>, } #[async_trait::async_trait] @@ -89,6 +100,98 @@ impl ActuationProvider for Provider { } } +#[async_trait::async_trait] +impl SignalProvider for Provider { + async fn update_filter( + &self, + update_filter: HashMap>, + ) -> Result<(), (RegisterSignalError, String)> { + let mut filters_update: HashMap = HashMap::new(); + for (signal_id, time_interval) in update_filter { + let min_sample_interval = time_interval.map(|time_interval| proto::SampleInterval { + interval_ms: time_interval.interval_ms(), + }); + + let filter = proto::Filter { + duration_ms: 0, + min_sample_interval, + }; + filters_update.insert(signal_id.id(), filter); + } + + let filter_request = proto::UpdateFilterRequest { + request_id: 1, + filters_update, + }; + + let filter_stream_request = + open_provider_stream_response::Action::UpdateFilterRequest(filter_request); + + let response = OpenProviderStreamResponse { + action: Some(filter_stream_request), + }; + + let result = self.sender.send(Ok(response)).await; + if result.is_err() { + return Err(( + RegisterSignalError::TransmissionFailure, + "An error occured while sending the data".to_string(), + )); + } + return Ok(()); + } + + fn is_available(&self) -> bool { + !self.sender.is_closed() + } + + async fn get_signals_values_from_provider( + &mut self, + signals_ids: Vec, + ) -> Result { + let request = OpenProviderStreamResponse { + action: Some( + open_provider_stream_response::Action::GetProviderValueRequest( + proto::GetProviderValueRequest { + request_id: 0, + signal_ids: signals_ids.iter().map(|signal_id| signal_id.id()).collect(), + }, + ), + ), + }; + + let result = self.sender.send(Ok(request)).await; + match result { + Ok(_) => {} + Err(err) => { + debug!("{}", err.to_string()); + } + } + match self.receiver.as_mut() { + // Here is assumed that provider will return a response immediately, todo -> timeout configuration + Some(receiver) => { + match timeout(tokio::time::Duration::from_secs(1), receiver.next()).await { + Ok(Some(value)) => match value { + Ok(value) => { + let mut entries_map = IndexMap::new(); + for (id, datapoint) in value.entries { + entries_map.insert(SignalId::new(id), (&datapoint).into()); + } + return Ok(GetValuesProviderResponse { + entries: entries_map, + }); + } + Err(_) => Err(()), + }, + // Either the stream ended (None) or the timeout elapsed (Err). + Ok(None) | Err(_) => Err(()), + } + } + None => Err(()), + } + } +} + #[tonic::async_trait] impl proto::val_server::Val for broker::DataBroker { // Returns (GRPC error code): @@ -114,23 +217,25 @@ impl proto::val_server::Val for broker::DataBroker { let request = request.into_inner(); let signal_id = match get_signal(request.signal_id, &broker).await { - Ok(signal_id) => signal_id, + Ok(signal_id) => SignalId::new(signal_id), Err(err) => return Err(err), }; - let datapoint = match broker.get_datapoint(signal_id).await { + let datapoint = match broker.get_values_broker(Vec::from([signal_id])).await { Ok(datapoint) => datapoint, - Err(ReadError::NotFound) => return Err(tonic::Status::not_found("Path not found")), - Err(ReadError::PermissionDenied) => { + Err((ReadError::NotFound, _)) => { + return Err(tonic::Status::not_found("Path not found")) + } + Err((ReadError::PermissionDenied, _)) => { return Err(tonic::Status::permission_denied("Permission denied")) } - Err(ReadError::PermissionExpired) => { + Err((ReadError::PermissionExpired, _)) => { return Err(tonic::Status::unauthenticated("Permission expired")) } }; Ok(tonic::Response::new(proto::GetValueResponse { - data_point: datapoint.into(), + data_point: datapoint.entries.values().next().unwrap().clone().into(), })) } @@ -157,38 +262,43 @@ impl proto::val_server::Val for broker::DataBroker { let requested = request.into_inner().signal_ids; let mut response_datapoints = Vec::new(); + let mut signals_requested = Vec::new(); + for request in requested { - let signal_id = match get_signal(Some(request), &broker).await { - Ok(signal_id) => signal_id, + match get_signal(Some(request), &broker).await { + Ok(signal_id) => { + signals_requested.push(SignalId::new(signal_id)); + } Err(err) => return Err(err), }; + } - match broker.get_datapoint(signal_id).await { - Ok(datapoint) => { + match broker.get_values_broker(signals_requested).await { + Ok(response) => { + for (_, datapoint) in response.entries { let proto_datapoint_opt: Option = datapoint.into(); - //let proto_datapoint: proto::Datapoint = proto_datapoint_opt.into(); response_datapoints.push(proto_datapoint_opt.unwrap()); } - Err(ReadError::NotFound) => { - return Err(tonic::Status::not_found(format!( - "Path not found (id: {})", - signal_id - ))); - } - Err(ReadError::PermissionDenied) => { - return Err(tonic::Status::permission_denied(format!( - "Permission denied(id: {})", - signal_id - ))) - } - Err(ReadError::PermissionExpired) => { - return Err(tonic::Status::unauthenticated(format!( - "Permission expired (id: {})", - signal_id - ))) - } - }; - } + } + Err((ReadError::NotFound, signal_id)) => { + return Err(tonic::Status::not_found(format!( + "Path not found (id: {})", + signal_id + ))); + } + Err((ReadError::PermissionDenied, signal_id)) => { + return Err(tonic::Status::permission_denied(format!( + "Permission denied(id: {})", + signal_id + ))) + } + Err((ReadError::PermissionExpired, signal_id)) => { + return Err(tonic::Status::unauthenticated(format!( + "Permission expired (id: {})", + signal_id + ))) + } + }; Ok(tonic::Response::new(proto::GetValuesResponse { data_points: response_datapoints, @@ -248,8 +358,20 @@ impl proto::val_server::Val for broker::DataBroker { ); } + let interval_ms = if let Some(filter) = request.filter { + filter + .min_sample_interval + .map(|min_sample_interval| min_sample_interval.interval_ms) + } else { + None + }; + match broker - .subscribe(valid_requests, Some(request.buffer_size as usize)) + .subscribe( + valid_requests, + Some(request.buffer_size as usize), + interval_ms, + ) .await { Ok(stream) => { @@ -321,8 +443,20 @@ impl proto::val_server::Val for broker::DataBroker { ); } + let interval_ms = if let Some(filter) = request.filter { + filter + .min_sample_interval + .map(|min_sample_interval| min_sample_interval.interval_ms) + } else { + None + }; + match broker - .subscribe(valid_requests, Some(request.buffer_size as usize)) + .subscribe( + valid_requests, + Some(request.buffer_size as usize), + interval_ms, + ) .await { Ok(stream) => { @@ -730,11 +864,13 @@ impl proto::val_server::Val for broker::DataBroker { let broker = self.clone(); // Create stream (to be returned) let (response_stream_sender, response_stream_receiver) = mpsc::channel(10); + let (get_value_sender, _) = broadcast::channel(10); // Listening on stream tokio::spawn(async move { let permissions = permissions; let broker = broker.authorized_access(&permissions); + let mut local_provider_uuid: Option = None; loop { select! { message = stream.message() => { @@ -751,15 +887,18 @@ impl proto::val_server::Val for broker::DataBroker { } }, Some(PublishValuesRequest(publish_values_request)) => { - let response = publish_values(&broker, &publish_values_request).await; - if let Some(value) = response { - if let Err(err) = response_stream_sender.send(Ok(value)).await { - debug!("Failed to send error response: {}", err); + if local_provider_uuid.is_some() { + let response = publish_values(&broker, local_provider_uuid.unwrap(), &publish_values_request).await; + if let Some(value) = response { + if let Err(err) = response_stream_sender.send(value).await { + debug!("Failed to send error response: {}", err); + } } + } else if let Err(err) = response_stream_sender.send(Err(tonic::Status::aborted("Provider has not claimed yet the signals, please call ProvideSignalRequest first"))).await { + debug!("Failed to send error response: {}", err); } }, Some(BatchActuateStreamResponse(batch_actuate_stream_response)) => { - if let Some(error) = batch_actuate_stream_response.error { match error.code() { ErrorCode::Ok => {}, @@ -783,17 +922,39 @@ impl proto::val_server::Val for broker::DataBroker { } }, - Some(ProvideSignalRequest(_provide_signal_request)) => { - todo!(); + Some(ProvideSignalRequest(provide_signal_request)) => { + let response = register_provided_signals(&broker, &provide_signal_request, response_stream_sender.clone(), get_value_sender.subscribe()).await; + match response { + Ok(value) => { + local_provider_uuid = Some(value.0); + if let Err(err) = response_stream_sender.send(Ok(value.1)).await + { + debug!("Failed to send response: {}", err) + } + }, + Err(tonic_error) => { + if let Err(err) = response_stream_sender.send(Err(tonic_error)).await + { + debug!("Failed to send tonic error: {}", err) + } + } + } } Some(UpdateFilterResponse(_update_filter_response)) => { - todo!(); + debug!("Filter response received from provider {}", local_provider_uuid.unwrap()); } - Some(GetProviderValueResponse(_get_provider_value_response)) => { - todo!(); + Some(GetProviderValueResponse(get_provider_value_response)) => { + if let Err(err) = get_value_sender.send(get_provider_value_response) + { + debug!("Failed to send tonic error: {}", err) + } } - Some(ProviderErrorIndication(_provide_error_indication)) => { - todo!(); + Some(ProviderErrorIndication(_provider_error_indication)) => { + if local_provider_uuid.is_some() { + publish_provider_error(&broker, local_provider_uuid.unwrap()).await; + } else if let Err(err) = response_stream_sender.send(Err(tonic::Status::aborted("Provider has not claimed yet any signals, please call ProvideSignalRequest first"))).await { + debug!("Failed to send error response: {}", err); + } } None => { @@ -882,7 +1043,10 @@ async fn provide_actuation( all_vss_ids.extend(vss_ids); all_vss_ids.extend(resolved_vss_ids); - let provider = Provider { sender }; + let provider = Provider { + sender, + receiver: None, + }; match broker .provide_actuation(all_vss_ids, Box::new(provider)) @@ -906,14 +1070,67 @@ async fn provide_actuation( } } +async fn register_provided_signals( + broker: &AuthorizedAccess<'_, '_>, + request: &databroker_proto::kuksa::val::v2::ProvideSignalRequest, + sender: mpsc::Sender>, + receiver: broadcast::Receiver, +) -> Result<(Uuid, OpenProviderStreamResponse), tonic::Status> { + let provider = Provider { + sender, + receiver: Some(BroadcastStream::new(receiver)), + }; + + let all_vss_ids = request + .signals_sample_intervals + .keys() + .map(|signal| SignalId::new(*signal)) + .collect(); + + match broker + .register_signals(all_vss_ids, Box::new(provider)) + .await + { + Ok(provider_uuid) => { + let provide_signal_response = ProvideSignalResponse {}; + let response = OpenProviderStreamResponse { + action: Some( + open_provider_stream_response::Action::ProvideSignalResponse( + provide_signal_response, + ), + ), + }; + Ok((provider_uuid, response)) + } + Err(error) => Err(error.0.to_tonic_status(error.1)), + } +} + +async fn publish_provider_error(broker: &AuthorizedAccess<'_, '_>, provide_uuid: Uuid) { + match broker.publish_provider_error(provide_uuid).await { + Ok(_) => {} + Err(entries) => { + let keys: HashSet = entries.iter().map(|(key, _)| key.to_string()).collect(); + let unique_string = keys.into_iter().collect::>().join(", "); + debug!( + "The provider {} error could not invalidate the entries {}", + provide_uuid, unique_string + ); + } + } +} + async fn publish_values( broker: &AuthorizedAccess<'_, '_>, + provider_uuid: Uuid, request: &databroker_proto::kuksa::val::v2::PublishValuesRequest, -) -> Option { +) -> Option> { + let mut request_signal_set: HashSet = HashSet::new(); let ids: Vec<(i32, broker::EntryUpdate)> = request .data_points .iter() .map(|(id, datapoint)| { + request_signal_set.insert(SignalId::new(*id)); ( *id, broker::EntryUpdate { @@ -932,22 +1149,30 @@ async fn publish_values( }) .collect(); - // TODO check if provider is allowed to update the entries for the provided signals? - match broker.update_entries(ids).await { - Ok(_) => None, - Err(err) => Some(OpenProviderStreamResponse { - action: Some( - open_provider_stream_response::Action::PublishValuesResponse( - PublishValuesResponse { - request_id: request.request_id, - status: err - .iter() - .map(|(id, error)| (*id, proto::Error::from(error))) - .collect(), - }, + if broker + .valid_provider_publish_signals(provider_uuid, &request_signal_set) + .await + { + match broker.update_entries(ids).await { + Ok(_) => None, + Err(err) => Some(Ok(OpenProviderStreamResponse { + action: Some( + open_provider_stream_response::Action::PublishValuesResponse( + PublishValuesResponse { + request_id: request.request_id, + status: err + .iter() + .map(|(id, error)| (*id, proto::Error::from(error))) + .collect(), + }, + ), ), - ), - }), + })), + } + } else { + Some(Err(tonic::Status::already_exists( + "Signals already registered by another provider", + ))) } } @@ -983,48 +1208,54 @@ async fn get_signal( } fn convert_to_proto_stream( - input: impl Stream, + input: impl Stream>, size: usize, ) -> impl Stream> { - input.map(move |item| { - let mut entries: HashMap = HashMap::with_capacity(size); - for update in item.updates { - let update_datapoint: Option = match update.update.datapoint { - Some(datapoint) => datapoint.into(), - None => None, - }; - if let Some(dp) = update_datapoint { - entries.insert( - update - .update - .path - .expect("Something wrong with update path of subscriptions!"), - dp, - ); + input.filter_map(move |item| match item { + Some(entry) => { + let mut entries: HashMap = HashMap::with_capacity(size); + for update in entry.updates { + let update_datapoint: Option = match update.update.datapoint { + Some(datapoint) => datapoint.into(), + None => None, + }; + if let Some(dp) = update_datapoint { + entries.insert( + update + .update + .path + .expect("Something wrong with update path of subscriptions!"), + dp, + ); + } } + let response = proto::SubscribeResponse { entries }; + Some(Ok(response)) } - let response = proto::SubscribeResponse { entries }; - Ok(response) + None => None, }) } fn convert_to_proto_stream_id( - input: impl Stream, + input: impl Stream>, size: usize, ) -> impl Stream> { - input.map(move |item| { - let mut entries: HashMap = HashMap::with_capacity(size); - for update in item.updates { - let update_datapoint: Option = match update.update.datapoint { - Some(datapoint) => datapoint.into(), - None => None, - }; - if let Some(dp) = update_datapoint { - entries.insert(update.id, dp); + input.filter_map(move |item| match item { + Some(entry) => { + let mut entries: HashMap = HashMap::with_capacity(size); + for update in entry.updates { + let update_datapoint: Option = match update.update.datapoint { + Some(datapoint) => datapoint.into(), + None => None, + }; + if let Some(dp) = update_datapoint { + entries.insert(update.id, dp); + } } + let response = proto::SubscribeByIdResponse { entries }; + Some(Ok(response)) } - let response = proto::SubscribeByIdResponse { entries }; - Ok(response) + None => None, }) } @@ -1413,7 +1644,7 @@ mod tests { static SIGNAL1: &str = "test.datapoint1"; static SIGNAL2: &str = "test.datapoint2"; - let broker = DataBroker::default(); + let broker: DataBroker = DataBroker::default(); let timestamp = std::time::SystemTime::now(); @@ -2239,7 +2470,7 @@ mod tests { Test open_provider_stream service method */ #[tokio::test(flavor = "multi_thread", worker_threads = 2)] - async fn test_open_provider_stream() { + async fn test_publish_value_request_without_first_claiming_it() { let broker = DataBroker::default(); let authorized_access = broker.authorized_access(&permissions::ALLOW_ALL); let request_id = 1; @@ -2304,18 +2535,8 @@ mod tests { Some(ProvideActuationResponse(_)) => { panic!("Should not happen") } - Some(PublishValuesResponse(publish_values_response)) => { - assert_eq!(publish_values_response.request_id, request_id); - assert_eq!(publish_values_response.status.len(), 1); - match publish_values_response.status.get(&entry_id) { - Some(value) => { - assert_eq!(value.code, 1); - assert_eq!(value.message, "Wrong Type"); - } - None => { - panic!("Should not happen") - } - } + Some(PublishValuesResponse(_)) => { + panic!("Should not happen") } Some(BatchActuateStreamRequest(_)) => { panic!("Should not happen") @@ -2333,8 +2554,8 @@ mod tests { panic!("Should not happen") } }, - Err(_) => { - panic!("Should not happen") + Err(err) => { + assert_eq!(err.code(), tonic::Code::Aborted); } } } @@ -2647,7 +2868,10 @@ mod tests { let vss_ids = vec![vss_id]; let (sender, _) = mpsc::channel(10); - let actuation_provider = Provider { sender }; + let actuation_provider = Provider { + sender, + receiver: None, + }; authorized_access .provide_actuation(vss_ids, Box::new(actuation_provider)) .await @@ -2770,7 +2994,10 @@ mod tests { let vss_ids = vec![vss_id]; let (sender, mut receiver) = mpsc::channel(10); - let actuation_provider = Provider { sender }; + let actuation_provider = Provider { + sender, + receiver: None, + }; authorized_access .provide_actuation(vss_ids, Box::new(actuation_provider)) .await @@ -2868,7 +3095,10 @@ mod tests { let vss_ids = vec![vss_id_abs, vss_id_cruise_control, vss_id_navigation_volume]; let (sender, _receiver) = mpsc::channel(10); - let actuation_provider = Provider { sender }; + let actuation_provider = Provider { + sender, + receiver: None, + }; authorized_access .provide_actuation(vss_ids, Box::new(actuation_provider)) .await @@ -3018,7 +3248,10 @@ mod tests { let vss_ids = vec![vss_id_abs]; let (sender, _receiver) = mpsc::channel(10); - let actuation_provider = Provider { sender }; + let actuation_provider = Provider { + sender, + receiver: None, + }; authorized_access .provide_actuation(vss_ids, Box::new(actuation_provider)) .await @@ -3108,7 +3341,10 @@ mod tests { let vss_ids = vec![vss_id_abs, vss_id_cruise_control]; let (sender, mut receiver) = mpsc::channel(10); - let actuation_provider = Provider { sender }; + let actuation_provider = Provider { + sender, + receiver: None, + }; authorized_access .provide_actuation(vss_ids, Box::new(actuation_provider)) .await @@ -3275,7 +3511,10 @@ mod tests { let vss_ids = vec![vss_id]; let (sender, _) = mpsc::channel(10); - let actuation_provider = Provider { sender }; + let actuation_provider = Provider { + sender, + receiver: None, + }; authorized_access .provide_actuation(vss_ids, Box::new(actuation_provider)) .await @@ -3410,7 +3649,10 @@ mod tests { let vss_ids = vec![vss_id]; let (sender, mut _receiver) = mpsc::channel(10); - let actuation_provider = Provider { sender }; + let actuation_provider = Provider { + sender, + receiver: None, + }; authorized_access .provide_actuation(vss_ids, Box::new(actuation_provider)) .await diff --git a/databroker/src/lib.rs b/databroker/src/lib.rs index b872be65..874fb138 100644 --- a/databroker/src/lib.rs +++ b/databroker/src/lib.rs @@ -13,6 +13,7 @@ pub mod authorization; pub mod broker; +pub mod filter; pub mod glob; pub mod grpc; pub mod open_telemetry; diff --git a/databroker/src/types.rs b/databroker/src/types.rs index c176d410..d24141ef 100644 --- a/databroker/src/types.rs +++ b/databroker/src/types.rs @@ -487,6 +487,34 @@ impl DataValue { } } +#[derive(Debug, Clone, Copy, Eq, PartialEq, Hash)] +pub struct SignalId { + id: i32, +} + +impl SignalId { + pub fn new(id: i32) -> Self { + Self { id } + } + pub fn id(&self) -> i32 { + self.id + } +} + +#[derive(Debug, Clone, Copy, Eq, PartialEq, PartialOrd, Ord)] +pub struct TimeInterval { + ms: u32, +} + +impl TimeInterval { + pub fn new(ms: u32) -> Self { + Self { ms } + } + pub fn interval_ms(&self) -> u32 { + self.ms + } +} + #[derive(Debug)] pub struct ExecutionInputImplData { pub value: DataValue, diff --git a/databroker/src/viss/v2/server.rs b/databroker/src/viss/v2/server.rs index 1af851f3..98f10c83 100644 --- a/databroker/src/viss/v2/server.rs +++ b/databroker/src/viss/v2/server.rs @@ -377,7 +377,13 @@ impl Viss for Server { }); }; - match broker.subscribe(entries, None).await { + let interval_ms = if let Some(Filter::Timebased(timebased)) = &request.filter { + Some(timebased.period) + } else { + None + }; + + match broker.subscribe(entries, None, interval_ms).await { Ok(stream) => { let subscription_id = SubscriptionId::new(); @@ -443,30 +449,38 @@ impl Viss for Server { fn convert_to_viss_stream( subscription_id: SubscriptionId, - stream: impl Stream, + stream: impl Stream>, ) -> impl Stream> { - stream.map(move |mut item| { + stream.map(move |item| { let ts = SystemTime::now().into(); let subscription_id = subscription_id.clone(); - match item.updates.pop() { - Some(item) => match (item.update.path, item.update.datapoint) { - (Some(path), Some(datapoint)) => Ok(SubscriptionEvent { - subscription_id, - data: Data::Object(DataObject { - path: path.into(), - dp: datapoint.into(), + match item { + Some(mut value) => match value.updates.pop() { + Some(item) => match (item.update.path, item.update.datapoint) { + (Some(path), Some(datapoint)) => Ok(SubscriptionEvent { + subscription_id, + data: Data::Object(DataObject { + path: path.into(), + dp: datapoint.into(), + }), + ts, }), - ts, - }), - (_, _) => Err(SubscriptionErrorEvent { + (_, _) => Err(SubscriptionErrorEvent { + subscription_id, + error: Error::InternalServerError, + ts, + }), + }, + None => Err(SubscriptionErrorEvent { subscription_id, error: Error::InternalServerError, ts, }), }, + // if None, it means the provider(is not available), meaning we should return the VISS error service_unavailable None => Err(SubscriptionErrorEvent { subscription_id, - error: Error::InternalServerError, + error: Error::ServiceUnavailable, ts, }), } diff --git a/databroker/src/viss/v2/types.rs b/databroker/src/viss/v2/types.rs index 3f926d4e..a6184413 100644 --- a/databroker/src/viss/v2/types.rs +++ b/databroker/src/viss/v2/types.rs @@ -117,14 +117,14 @@ pub struct SubscribeRequest { pub path: Path, pub request_id: RequestId, pub authorization: Option, - // filter: Option, + pub filter: Option, } #[derive(Serialize)] #[serde(tag = "action", rename = "subscribe", rename_all = "camelCase")] pub struct SubscribeSuccessResponse { - pub request_id: RequestId, pub subscription_id: SubscriptionId, + pub request_id: RequestId, pub ts: Timestamp, } @@ -198,6 +198,8 @@ pub enum Filter { StaticMetadata(StaticMetadataFilter), #[serde(rename = "paths")] Paths(PathsFilter), + #[serde(rename = "timebased")] + Timebased(TimebasedFilter), } #[derive(Deserialize)] @@ -212,6 +214,12 @@ pub struct StaticMetadataFilter { // pub parameters: Option, } +#[derive(Deserialize)] +#[serde(rename_all = "camelCase")] +pub struct TimebasedFilter { + pub period: u32, +} + // Unique id value specified by the client. Returned by the server in the // response and used by the client to link the request and response messages. // The value MAY be an integer or a Universally Unique Identifier (UUID). @@ -403,6 +411,7 @@ pub enum Error { NotFoundInvalidSubscriptionId, InternalServerError, NotImplemented, + ServiceUnavailable, } impl From for ErrorSpec { @@ -490,6 +499,11 @@ impl From for ErrorSpec { // BadGateway 502 bad_gateway The server was acting as a gateway or proxy and received an invalid response from an upstream server. // ServiceUnavailable 503 service_unavailable The server is currently unable to handle the request due to a temporary overload or scheduled maintenance (which may be alleviated after some delay). // GatewayTimeout 504 gateway_timeout The server did not receive a timely response from an upstream server it needed to access in order to complete the request. + Error::ServiceUnavailable => ErrorSpec { + number: 503, + reason: "service_unavailable".into(), + message: "The server is temporarily unable to handle the request.".into(), + }, } } } diff --git a/integration_test/gen_proto/kuksa/val/v1/types_pb2.py b/integration_test/gen_proto/kuksa/val/v1/types_pb2.py new file mode 100644 index 00000000..00462ab3 --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v1/types_pb2.py @@ -0,0 +1,99 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +# -*- coding: utf-8 -*- +# Generated by the protocol buffer compiler. DO NOT EDIT! +# NO CHECKED-IN PROTOBUF GENCODE +# source: kuksa/val/v1/types.proto +# Protobuf Python Version: 5.29.0 +"""Generated protocol buffer code.""" +from google.protobuf import descriptor as _descriptor +from google.protobuf import descriptor_pool as _descriptor_pool +from google.protobuf import runtime_version as _runtime_version +from google.protobuf import symbol_database as _symbol_database +from google.protobuf.internal import builder as _builder +_runtime_version.ValidateProtobufRuntimeVersion( + _runtime_version.Domain.PUBLIC, + 5, + 29, + 0, + '', + 'kuksa/val/v1/types.proto' +) +# @@protoc_insertion_point(imports) + +_sym_db = _symbol_database.Default() + + +from google.protobuf import timestamp_pb2 as google_dot_protobuf_dot_timestamp__pb2 + + +DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\x18kuksa/val/v1/types.proto\x12\x0ckuksa.val.v1\x1a\x1fgoogle/protobuf/timestamp.proto\"\x9d\x01\n\tDataEntry\x12\x0c\n\x04path\x18\x01 \x01(\t\x12&\n\x05value\x18\x02 \x01(\x0b\x32\x17.kuksa.val.v1.Datapoint\x12\x30\n\x0f\x61\x63tuator_target\x18\x03 \x01(\x0b\x32\x17.kuksa.val.v1.Datapoint\x12(\n\x08metadata\x18\n \x01(\x0b\x32\x16.kuksa.val.v1.Metadata\"\xdc\x04\n\tDatapoint\x12-\n\ttimestamp\x18\x01 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12\x10\n\x06string\x18\x0b \x01(\tH\x00\x12\x0e\n\x04\x62ool\x18\x0c \x01(\x08H\x00\x12\x0f\n\x05int32\x18\r \x01(\x11H\x00\x12\x0f\n\x05int64\x18\x0e \x01(\x12H\x00\x12\x10\n\x06uint32\x18\x0f \x01(\rH\x00\x12\x10\n\x06uint64\x18\x10 \x01(\x04H\x00\x12\x0f\n\x05\x66loat\x18\x11 \x01(\x02H\x00\x12\x10\n\x06\x64ouble\x18\x12 \x01(\x01H\x00\x12\x31\n\x0cstring_array\x18\x15 \x01(\x0b\x32\x19.kuksa.val.v1.StringArrayH\x00\x12-\n\nbool_array\x18\x16 \x01(\x0b\x32\x17.kuksa.val.v1.BoolArrayH\x00\x12/\n\x0bint32_array\x18\x17 \x01(\x0b\x32\x18.kuksa.val.v1.Int32ArrayH\x00\x12/\n\x0bint64_array\x18\x18 \x01(\x0b\x32\x18.kuksa.val.v1.Int64ArrayH\x00\x12\x31\n\x0cuint32_array\x18\x19 \x01(\x0b\x32\x19.kuksa.val.v1.Uint32ArrayH\x00\x12\x31\n\x0cuint64_array\x18\x1a \x01(\x0b\x32\x19.kuksa.val.v1.Uint64ArrayH\x00\x12/\n\x0b\x66loat_array\x18\x1b \x01(\x0b\x32\x18.kuksa.val.v1.FloatArrayH\x00\x12\x31\n\x0c\x64ouble_array\x18\x1c \x01(\x0b\x32\x19.kuksa.val.v1.DoubleArrayH\x00\x42\x07\n\x05value\"\xc3\x03\n\x08Metadata\x12)\n\tdata_type\x18\x0b \x01(\x0e\x32\x16.kuksa.val.v1.DataType\x12+\n\nentry_type\x18\x0c \x01(\x0e\x32\x17.kuksa.val.v1.EntryType\x12\x18\n\x0b\x64\x65scription\x18\r \x01(\tH\x01\x88\x01\x01\x12\x14\n\x07\x63omment\x18\x0e \x01(\tH\x02\x88\x01\x01\x12\x18\n\x0b\x64\x65precation\x18\x0f \x01(\tH\x03\x88\x01\x01\x12\x11\n\x04unit\x18\x10 \x01(\tH\x04\x88\x01\x01\x12\x39\n\x11value_restriction\x18\x11 \x01(\x0b\x32\x1e.kuksa.val.v1.ValueRestriction\x12*\n\x08\x61\x63tuator\x18\x14 \x01(\x0b\x32\x16.kuksa.val.v1.ActuatorH\x00\x12&\n\x06sensor\x18\x1e \x01(\x0b\x32\x14.kuksa.val.v1.SensorH\x00\x12,\n\tattribute\x18( \x01(\x0b\x32\x17.kuksa.val.v1.AttributeH\x00\x42\x10\n\x0e\x65ntry_specificB\x0e\n\x0c_descriptionB\n\n\x08_commentB\x0e\n\x0c_deprecationB\x07\n\x05_unit\"\n\n\x08\x41\x63tuator\"\x08\n\x06Sensor\"\x0b\n\tAttribute\"\xfe\x01\n\x10ValueRestriction\x12\x36\n\x06string\x18\x15 \x01(\x0b\x32$.kuksa.val.v1.ValueRestrictionStringH\x00\x12\x33\n\x06signed\x18\x16 \x01(\x0b\x32!.kuksa.val.v1.ValueRestrictionIntH\x00\x12\x36\n\x08unsigned\x18\x17 \x01(\x0b\x32\".kuksa.val.v1.ValueRestrictionUintH\x00\x12=\n\x0e\x66loating_point\x18\x18 \x01(\x0b\x32#.kuksa.val.v1.ValueRestrictionFloatH\x00\x42\x06\n\x04type\"a\n\x13ValueRestrictionInt\x12\x10\n\x03min\x18\x01 \x01(\x12H\x00\x88\x01\x01\x12\x10\n\x03max\x18\x02 \x01(\x12H\x01\x88\x01\x01\x12\x16\n\x0e\x61llowed_values\x18\x03 \x03(\x12\x42\x06\n\x04_minB\x06\n\x04_max\"b\n\x14ValueRestrictionUint\x12\x10\n\x03min\x18\x01 \x01(\x04H\x00\x88\x01\x01\x12\x10\n\x03max\x18\x02 \x01(\x04H\x01\x88\x01\x01\x12\x16\n\x0e\x61llowed_values\x18\x03 \x03(\x04\x42\x06\n\x04_minB\x06\n\x04_max\"c\n\x15ValueRestrictionFloat\x12\x10\n\x03min\x18\x01 \x01(\x01H\x00\x88\x01\x01\x12\x10\n\x03max\x18\x02 \x01(\x01H\x01\x88\x01\x01\x12\x16\n\x0e\x61llowed_values\x18\x03 \x03(\x01\x42\x06\n\x04_minB\x06\n\x04_max\"0\n\x16ValueRestrictionString\x12\x16\n\x0e\x61llowed_values\x18\x03 \x03(\t\"6\n\x05\x45rror\x12\x0c\n\x04\x63ode\x18\x01 \x01(\r\x12\x0e\n\x06reason\x18\x02 \x01(\t\x12\x0f\n\x07message\x18\x03 \x01(\t\"B\n\x0e\x44\x61taEntryError\x12\x0c\n\x04path\x18\x01 \x01(\t\x12\"\n\x05\x65rror\x18\x02 \x01(\x0b\x32\x13.kuksa.val.v1.Error\"\x1d\n\x0bStringArray\x12\x0e\n\x06values\x18\x01 \x03(\t\"\x1b\n\tBoolArray\x12\x0e\n\x06values\x18\x01 \x03(\x08\"\x1c\n\nInt32Array\x12\x0e\n\x06values\x18\x01 \x03(\x11\"\x1c\n\nInt64Array\x12\x0e\n\x06values\x18\x01 \x03(\x12\"\x1d\n\x0bUint32Array\x12\x0e\n\x06values\x18\x01 \x03(\r\"\x1d\n\x0bUint64Array\x12\x0e\n\x06values\x18\x01 \x03(\x04\"\x1c\n\nFloatArray\x12\x0e\n\x06values\x18\x01 \x03(\x02\"\x1d\n\x0b\x44oubleArray\x12\x0e\n\x06values\x18\x01 \x03(\x01*\xa9\x05\n\x08\x44\x61taType\x12\x19\n\x15\x44\x41TA_TYPE_UNSPECIFIED\x10\x00\x12\x14\n\x10\x44\x41TA_TYPE_STRING\x10\x01\x12\x15\n\x11\x44\x41TA_TYPE_BOOLEAN\x10\x02\x12\x12\n\x0e\x44\x41TA_TYPE_INT8\x10\x03\x12\x13\n\x0f\x44\x41TA_TYPE_INT16\x10\x04\x12\x13\n\x0f\x44\x41TA_TYPE_INT32\x10\x05\x12\x13\n\x0f\x44\x41TA_TYPE_INT64\x10\x06\x12\x13\n\x0f\x44\x41TA_TYPE_UINT8\x10\x07\x12\x14\n\x10\x44\x41TA_TYPE_UINT16\x10\x08\x12\x14\n\x10\x44\x41TA_TYPE_UINT32\x10\t\x12\x14\n\x10\x44\x41TA_TYPE_UINT64\x10\n\x12\x13\n\x0f\x44\x41TA_TYPE_FLOAT\x10\x0b\x12\x14\n\x10\x44\x41TA_TYPE_DOUBLE\x10\x0c\x12\x17\n\x13\x44\x41TA_TYPE_TIMESTAMP\x10\r\x12\x1a\n\x16\x44\x41TA_TYPE_STRING_ARRAY\x10\x14\x12\x1b\n\x17\x44\x41TA_TYPE_BOOLEAN_ARRAY\x10\x15\x12\x18\n\x14\x44\x41TA_TYPE_INT8_ARRAY\x10\x16\x12\x19\n\x15\x44\x41TA_TYPE_INT16_ARRAY\x10\x17\x12\x19\n\x15\x44\x41TA_TYPE_INT32_ARRAY\x10\x18\x12\x19\n\x15\x44\x41TA_TYPE_INT64_ARRAY\x10\x19\x12\x19\n\x15\x44\x41TA_TYPE_UINT8_ARRAY\x10\x1a\x12\x1a\n\x16\x44\x41TA_TYPE_UINT16_ARRAY\x10\x1b\x12\x1a\n\x16\x44\x41TA_TYPE_UINT32_ARRAY\x10\x1c\x12\x1a\n\x16\x44\x41TA_TYPE_UINT64_ARRAY\x10\x1d\x12\x19\n\x15\x44\x41TA_TYPE_FLOAT_ARRAY\x10\x1e\x12\x1a\n\x16\x44\x41TA_TYPE_DOUBLE_ARRAY\x10\x1f\x12\x1d\n\x19\x44\x41TA_TYPE_TIMESTAMP_ARRAY\x10 *q\n\tEntryType\x12\x1a\n\x16\x45NTRY_TYPE_UNSPECIFIED\x10\x00\x12\x18\n\x14\x45NTRY_TYPE_ATTRIBUTE\x10\x01\x12\x15\n\x11\x45NTRY_TYPE_SENSOR\x10\x02\x12\x17\n\x13\x45NTRY_TYPE_ACTUATOR\x10\x03*}\n\x04View\x12\x14\n\x10VIEW_UNSPECIFIED\x10\x00\x12\x16\n\x12VIEW_CURRENT_VALUE\x10\x01\x12\x15\n\x11VIEW_TARGET_VALUE\x10\x02\x12\x11\n\rVIEW_METADATA\x10\x03\x12\x0f\n\x0bVIEW_FIELDS\x10\n\x12\x0c\n\x08VIEW_ALL\x10\x14*\x9c\x03\n\x05\x46ield\x12\x15\n\x11\x46IELD_UNSPECIFIED\x10\x00\x12\x0e\n\nFIELD_PATH\x10\x01\x12\x0f\n\x0b\x46IELD_VALUE\x10\x02\x12\x19\n\x15\x46IELD_ACTUATOR_TARGET\x10\x03\x12\x12\n\x0e\x46IELD_METADATA\x10\n\x12\x1c\n\x18\x46IELD_METADATA_DATA_TYPE\x10\x0b\x12\x1e\n\x1a\x46IELD_METADATA_DESCRIPTION\x10\x0c\x12\x1d\n\x19\x46IELD_METADATA_ENTRY_TYPE\x10\r\x12\x1a\n\x16\x46IELD_METADATA_COMMENT\x10\x0e\x12\x1e\n\x1a\x46IELD_METADATA_DEPRECATION\x10\x0f\x12\x17\n\x13\x46IELD_METADATA_UNIT\x10\x10\x12$\n FIELD_METADATA_VALUE_RESTRICTION\x10\x11\x12\x1b\n\x17\x46IELD_METADATA_ACTUATOR\x10\x14\x12\x19\n\x15\x46IELD_METADATA_SENSOR\x10\x1e\x12\x1c\n\x18\x46IELD_METADATA_ATTRIBUTE\x10(B\x0eZ\x0ckuksa/val/v1b\x06proto3') + +_globals = globals() +_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) +_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'kuksa.val.v1.types_pb2', _globals) +if not _descriptor._USE_C_DESCRIPTORS: + _globals['DESCRIPTOR']._loaded_options = None + _globals['DESCRIPTOR']._serialized_options = b'Z\014kuksa/val/v1' + _globals['_DATATYPE']._serialized_start=2306 + _globals['_DATATYPE']._serialized_end=2987 + _globals['_ENTRYTYPE']._serialized_start=2989 + _globals['_ENTRYTYPE']._serialized_end=3102 + _globals['_VIEW']._serialized_start=3104 + _globals['_VIEW']._serialized_end=3229 + _globals['_FIELD']._serialized_start=3232 + _globals['_FIELD']._serialized_end=3644 + _globals['_DATAENTRY']._serialized_start=76 + _globals['_DATAENTRY']._serialized_end=233 + _globals['_DATAPOINT']._serialized_start=236 + _globals['_DATAPOINT']._serialized_end=840 + _globals['_METADATA']._serialized_start=843 + _globals['_METADATA']._serialized_end=1294 + _globals['_ACTUATOR']._serialized_start=1296 + _globals['_ACTUATOR']._serialized_end=1306 + _globals['_SENSOR']._serialized_start=1308 + _globals['_SENSOR']._serialized_end=1316 + _globals['_ATTRIBUTE']._serialized_start=1318 + _globals['_ATTRIBUTE']._serialized_end=1329 + _globals['_VALUERESTRICTION']._serialized_start=1332 + _globals['_VALUERESTRICTION']._serialized_end=1586 + _globals['_VALUERESTRICTIONINT']._serialized_start=1588 + _globals['_VALUERESTRICTIONINT']._serialized_end=1685 + _globals['_VALUERESTRICTIONUINT']._serialized_start=1687 + _globals['_VALUERESTRICTIONUINT']._serialized_end=1785 + _globals['_VALUERESTRICTIONFLOAT']._serialized_start=1787 + _globals['_VALUERESTRICTIONFLOAT']._serialized_end=1886 + _globals['_VALUERESTRICTIONSTRING']._serialized_start=1888 + _globals['_VALUERESTRICTIONSTRING']._serialized_end=1936 + _globals['_ERROR']._serialized_start=1938 + _globals['_ERROR']._serialized_end=1992 + _globals['_DATAENTRYERROR']._serialized_start=1994 + _globals['_DATAENTRYERROR']._serialized_end=2060 + _globals['_STRINGARRAY']._serialized_start=2062 + _globals['_STRINGARRAY']._serialized_end=2091 + _globals['_BOOLARRAY']._serialized_start=2093 + _globals['_BOOLARRAY']._serialized_end=2120 + _globals['_INT32ARRAY']._serialized_start=2122 + _globals['_INT32ARRAY']._serialized_end=2150 + _globals['_INT64ARRAY']._serialized_start=2152 + _globals['_INT64ARRAY']._serialized_end=2180 + _globals['_UINT32ARRAY']._serialized_start=2182 + _globals['_UINT32ARRAY']._serialized_end=2211 + _globals['_UINT64ARRAY']._serialized_start=2213 + _globals['_UINT64ARRAY']._serialized_end=2242 + _globals['_FLOATARRAY']._serialized_start=2244 + _globals['_FLOATARRAY']._serialized_end=2272 + _globals['_DOUBLEARRAY']._serialized_start=2274 + _globals['_DOUBLEARRAY']._serialized_end=2303 +# @@protoc_insertion_point(module_scope) diff --git a/integration_test/gen_proto/kuksa/val/v1/types_pb2.pyi b/integration_test/gen_proto/kuksa/val/v1/types_pb2.pyi new file mode 100644 index 00000000..5432eb04 --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v1/types_pb2.pyi @@ -0,0 +1,825 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +""" +@generated by mypy-protobuf. Do not edit manually! +isort:skip_file +I added V1 as in databroker. Is this good practice?""" + +import builtins +import collections.abc +import google.protobuf.descriptor +import google.protobuf.internal.containers +import google.protobuf.internal.enum_type_wrapper +import google.protobuf.message +import google.protobuf.timestamp_pb2 +import sys +import typing + +if sys.version_info >= (3, 10): + import typing as typing_extensions +else: + import typing_extensions + +DESCRIPTOR: google.protobuf.descriptor.FileDescriptor + +class _DataType: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _DataTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_DataType.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + DATA_TYPE_UNSPECIFIED: _DataType.ValueType # 0 + DATA_TYPE_STRING: _DataType.ValueType # 1 + DATA_TYPE_BOOLEAN: _DataType.ValueType # 2 + DATA_TYPE_INT8: _DataType.ValueType # 3 + DATA_TYPE_INT16: _DataType.ValueType # 4 + DATA_TYPE_INT32: _DataType.ValueType # 5 + DATA_TYPE_INT64: _DataType.ValueType # 6 + DATA_TYPE_UINT8: _DataType.ValueType # 7 + DATA_TYPE_UINT16: _DataType.ValueType # 8 + DATA_TYPE_UINT32: _DataType.ValueType # 9 + DATA_TYPE_UINT64: _DataType.ValueType # 10 + DATA_TYPE_FLOAT: _DataType.ValueType # 11 + DATA_TYPE_DOUBLE: _DataType.ValueType # 12 + DATA_TYPE_TIMESTAMP: _DataType.ValueType # 13 + DATA_TYPE_STRING_ARRAY: _DataType.ValueType # 20 + DATA_TYPE_BOOLEAN_ARRAY: _DataType.ValueType # 21 + DATA_TYPE_INT8_ARRAY: _DataType.ValueType # 22 + DATA_TYPE_INT16_ARRAY: _DataType.ValueType # 23 + DATA_TYPE_INT32_ARRAY: _DataType.ValueType # 24 + DATA_TYPE_INT64_ARRAY: _DataType.ValueType # 25 + DATA_TYPE_UINT8_ARRAY: _DataType.ValueType # 26 + DATA_TYPE_UINT16_ARRAY: _DataType.ValueType # 27 + DATA_TYPE_UINT32_ARRAY: _DataType.ValueType # 28 + DATA_TYPE_UINT64_ARRAY: _DataType.ValueType # 29 + DATA_TYPE_FLOAT_ARRAY: _DataType.ValueType # 30 + DATA_TYPE_DOUBLE_ARRAY: _DataType.ValueType # 31 + DATA_TYPE_TIMESTAMP_ARRAY: _DataType.ValueType # 32 + +class DataType(_DataType, metaclass=_DataTypeEnumTypeWrapper): + """VSS Data type of a signal + + Protobuf doesn't support int8, int16, uint8 or uint16. + These are mapped to int32 and uint32 respectively. + """ + +DATA_TYPE_UNSPECIFIED: DataType.ValueType # 0 +DATA_TYPE_STRING: DataType.ValueType # 1 +DATA_TYPE_BOOLEAN: DataType.ValueType # 2 +DATA_TYPE_INT8: DataType.ValueType # 3 +DATA_TYPE_INT16: DataType.ValueType # 4 +DATA_TYPE_INT32: DataType.ValueType # 5 +DATA_TYPE_INT64: DataType.ValueType # 6 +DATA_TYPE_UINT8: DataType.ValueType # 7 +DATA_TYPE_UINT16: DataType.ValueType # 8 +DATA_TYPE_UINT32: DataType.ValueType # 9 +DATA_TYPE_UINT64: DataType.ValueType # 10 +DATA_TYPE_FLOAT: DataType.ValueType # 11 +DATA_TYPE_DOUBLE: DataType.ValueType # 12 +DATA_TYPE_TIMESTAMP: DataType.ValueType # 13 +DATA_TYPE_STRING_ARRAY: DataType.ValueType # 20 +DATA_TYPE_BOOLEAN_ARRAY: DataType.ValueType # 21 +DATA_TYPE_INT8_ARRAY: DataType.ValueType # 22 +DATA_TYPE_INT16_ARRAY: DataType.ValueType # 23 +DATA_TYPE_INT32_ARRAY: DataType.ValueType # 24 +DATA_TYPE_INT64_ARRAY: DataType.ValueType # 25 +DATA_TYPE_UINT8_ARRAY: DataType.ValueType # 26 +DATA_TYPE_UINT16_ARRAY: DataType.ValueType # 27 +DATA_TYPE_UINT32_ARRAY: DataType.ValueType # 28 +DATA_TYPE_UINT64_ARRAY: DataType.ValueType # 29 +DATA_TYPE_FLOAT_ARRAY: DataType.ValueType # 30 +DATA_TYPE_DOUBLE_ARRAY: DataType.ValueType # 31 +DATA_TYPE_TIMESTAMP_ARRAY: DataType.ValueType # 32 +global___DataType = DataType + +class _EntryType: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _EntryTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_EntryType.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + ENTRY_TYPE_UNSPECIFIED: _EntryType.ValueType # 0 + ENTRY_TYPE_ATTRIBUTE: _EntryType.ValueType # 1 + ENTRY_TYPE_SENSOR: _EntryType.ValueType # 2 + ENTRY_TYPE_ACTUATOR: _EntryType.ValueType # 3 + +class EntryType(_EntryType, metaclass=_EntryTypeEnumTypeWrapper): + """Entry type""" + +ENTRY_TYPE_UNSPECIFIED: EntryType.ValueType # 0 +ENTRY_TYPE_ATTRIBUTE: EntryType.ValueType # 1 +ENTRY_TYPE_SENSOR: EntryType.ValueType # 2 +ENTRY_TYPE_ACTUATOR: EntryType.ValueType # 3 +global___EntryType = EntryType + +class _View: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _ViewEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_View.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + VIEW_UNSPECIFIED: _View.ValueType # 0 + """Unspecified. Equivalent to VIEW_CURRENT_VALUE unless `fields` are explicitly set.""" + VIEW_CURRENT_VALUE: _View.ValueType # 1 + """Populate DataEntry with value.""" + VIEW_TARGET_VALUE: _View.ValueType # 2 + """Populate DataEntry with actuator target.""" + VIEW_METADATA: _View.ValueType # 3 + """Populate DataEntry with metadata.""" + VIEW_FIELDS: _View.ValueType # 10 + """Populate DataEntry only with requested fields.""" + VIEW_ALL: _View.ValueType # 20 + """Populate DataEntry with everything.""" + +class View(_View, metaclass=_ViewEnumTypeWrapper): + """A `View` specifies a set of fields which should + be populated in a `DataEntry` (in a response message) + """ + +VIEW_UNSPECIFIED: View.ValueType # 0 +"""Unspecified. Equivalent to VIEW_CURRENT_VALUE unless `fields` are explicitly set.""" +VIEW_CURRENT_VALUE: View.ValueType # 1 +"""Populate DataEntry with value.""" +VIEW_TARGET_VALUE: View.ValueType # 2 +"""Populate DataEntry with actuator target.""" +VIEW_METADATA: View.ValueType # 3 +"""Populate DataEntry with metadata.""" +VIEW_FIELDS: View.ValueType # 10 +"""Populate DataEntry only with requested fields.""" +VIEW_ALL: View.ValueType # 20 +"""Populate DataEntry with everything.""" +global___View = View + +class _Field: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _FieldEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_Field.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + FIELD_UNSPECIFIED: _Field.ValueType # 0 + """"*" i.e. everything""" + FIELD_PATH: _Field.ValueType # 1 + """path""" + FIELD_VALUE: _Field.ValueType # 2 + """value""" + FIELD_ACTUATOR_TARGET: _Field.ValueType # 3 + """actuator_target""" + FIELD_METADATA: _Field.ValueType # 10 + """metadata.*""" + FIELD_METADATA_DATA_TYPE: _Field.ValueType # 11 + """metadata.data_type""" + FIELD_METADATA_DESCRIPTION: _Field.ValueType # 12 + """metadata.description""" + FIELD_METADATA_ENTRY_TYPE: _Field.ValueType # 13 + """metadata.entry_type""" + FIELD_METADATA_COMMENT: _Field.ValueType # 14 + """metadata.comment""" + FIELD_METADATA_DEPRECATION: _Field.ValueType # 15 + """metadata.deprecation""" + FIELD_METADATA_UNIT: _Field.ValueType # 16 + """metadata.unit""" + FIELD_METADATA_VALUE_RESTRICTION: _Field.ValueType # 17 + """metadata.value_restriction.*""" + FIELD_METADATA_ACTUATOR: _Field.ValueType # 20 + """metadata.actuator.*""" + FIELD_METADATA_SENSOR: _Field.ValueType # 30 + """metadata.sensor.*""" + FIELD_METADATA_ATTRIBUTE: _Field.ValueType # 40 + """metadata.attribute.*""" + +class Field(_Field, metaclass=_FieldEnumTypeWrapper): + """A `Field` corresponds to a specific field of a `DataEntry`. + + It can be used to: + * populate only specific fields of a `DataEntry` response. + * specify which fields of a `DataEntry` should be set as + part of a `Set` request. + * subscribe to only specific fields of a data entry. + * convey which fields of an updated `DataEntry` have changed. + """ + +FIELD_UNSPECIFIED: Field.ValueType # 0 +""""*" i.e. everything""" +FIELD_PATH: Field.ValueType # 1 +"""path""" +FIELD_VALUE: Field.ValueType # 2 +"""value""" +FIELD_ACTUATOR_TARGET: Field.ValueType # 3 +"""actuator_target""" +FIELD_METADATA: Field.ValueType # 10 +"""metadata.*""" +FIELD_METADATA_DATA_TYPE: Field.ValueType # 11 +"""metadata.data_type""" +FIELD_METADATA_DESCRIPTION: Field.ValueType # 12 +"""metadata.description""" +FIELD_METADATA_ENTRY_TYPE: Field.ValueType # 13 +"""metadata.entry_type""" +FIELD_METADATA_COMMENT: Field.ValueType # 14 +"""metadata.comment""" +FIELD_METADATA_DEPRECATION: Field.ValueType # 15 +"""metadata.deprecation""" +FIELD_METADATA_UNIT: Field.ValueType # 16 +"""metadata.unit""" +FIELD_METADATA_VALUE_RESTRICTION: Field.ValueType # 17 +"""metadata.value_restriction.*""" +FIELD_METADATA_ACTUATOR: Field.ValueType # 20 +"""metadata.actuator.*""" +FIELD_METADATA_SENSOR: Field.ValueType # 30 +"""metadata.sensor.*""" +FIELD_METADATA_ATTRIBUTE: Field.ValueType # 40 +"""metadata.attribute.*""" +global___Field = Field + +@typing.final +class DataEntry(google.protobuf.message.Message): + """Describes a VSS entry + When requesting an entry, the amount of information returned can + be controlled by specifying either a `View` or a set of `Field`s. + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + PATH_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + ACTUATOR_TARGET_FIELD_NUMBER: builtins.int + METADATA_FIELD_NUMBER: builtins.int + path: builtins.str + """Defines the full VSS path of the entry. + [field: FIELD_PATH] + """ + @property + def value(self) -> global___Datapoint: + """The value (datapoint) + [field: FIELD_VALUE] + """ + + @property + def actuator_target(self) -> global___Datapoint: + """Actuator target (only used if the entry is an actuator) + [field: FIELD_ACTUATOR_TARGET] + """ + + @property + def metadata(self) -> global___Metadata: + """Metadata for this entry + [field: FIELD_METADATA] + """ + + def __init__( + self, + *, + path: builtins.str = ..., + value: global___Datapoint | None = ..., + actuator_target: global___Datapoint | None = ..., + metadata: global___Metadata | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["actuator_target", b"actuator_target", "metadata", b"metadata", "value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["actuator_target", b"actuator_target", "metadata", b"metadata", "path", b"path", "value", b"value"]) -> None: ... + +global___DataEntry = DataEntry + +@typing.final +class Datapoint(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + TIMESTAMP_FIELD_NUMBER: builtins.int + STRING_FIELD_NUMBER: builtins.int + BOOL_FIELD_NUMBER: builtins.int + INT32_FIELD_NUMBER: builtins.int + INT64_FIELD_NUMBER: builtins.int + UINT32_FIELD_NUMBER: builtins.int + UINT64_FIELD_NUMBER: builtins.int + FLOAT_FIELD_NUMBER: builtins.int + DOUBLE_FIELD_NUMBER: builtins.int + STRING_ARRAY_FIELD_NUMBER: builtins.int + BOOL_ARRAY_FIELD_NUMBER: builtins.int + INT32_ARRAY_FIELD_NUMBER: builtins.int + INT64_ARRAY_FIELD_NUMBER: builtins.int + UINT32_ARRAY_FIELD_NUMBER: builtins.int + UINT64_ARRAY_FIELD_NUMBER: builtins.int + FLOAT_ARRAY_FIELD_NUMBER: builtins.int + DOUBLE_ARRAY_FIELD_NUMBER: builtins.int + string: builtins.str + bool: builtins.bool + int32: builtins.int + int64: builtins.int + uint32: builtins.int + uint64: builtins.int + float: builtins.float + double: builtins.float + @property + def timestamp(self) -> google.protobuf.timestamp_pb2.Timestamp: ... + @property + def string_array(self) -> global___StringArray: ... + @property + def bool_array(self) -> global___BoolArray: ... + @property + def int32_array(self) -> global___Int32Array: ... + @property + def int64_array(self) -> global___Int64Array: ... + @property + def uint32_array(self) -> global___Uint32Array: ... + @property + def uint64_array(self) -> global___Uint64Array: ... + @property + def float_array(self) -> global___FloatArray: ... + @property + def double_array(self) -> global___DoubleArray: ... + def __init__( + self, + *, + timestamp: google.protobuf.timestamp_pb2.Timestamp | None = ..., + string: builtins.str = ..., + bool: builtins.bool = ..., + int32: builtins.int = ..., + int64: builtins.int = ..., + uint32: builtins.int = ..., + uint64: builtins.int = ..., + float: builtins.float = ..., + double: builtins.float = ..., + string_array: global___StringArray | None = ..., + bool_array: global___BoolArray | None = ..., + int32_array: global___Int32Array | None = ..., + int64_array: global___Int64Array | None = ..., + uint32_array: global___Uint32Array | None = ..., + uint64_array: global___Uint64Array | None = ..., + float_array: global___FloatArray | None = ..., + double_array: global___DoubleArray | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["bool", b"bool", "bool_array", b"bool_array", "double", b"double", "double_array", b"double_array", "float", b"float", "float_array", b"float_array", "int32", b"int32", "int32_array", b"int32_array", "int64", b"int64", "int64_array", b"int64_array", "string", b"string", "string_array", b"string_array", "timestamp", b"timestamp", "uint32", b"uint32", "uint32_array", b"uint32_array", "uint64", b"uint64", "uint64_array", b"uint64_array", "value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["bool", b"bool", "bool_array", b"bool_array", "double", b"double", "double_array", b"double_array", "float", b"float", "float_array", b"float_array", "int32", b"int32", "int32_array", b"int32_array", "int64", b"int64", "int64_array", b"int64_array", "string", b"string", "string_array", b"string_array", "timestamp", b"timestamp", "uint32", b"uint32", "uint32_array", b"uint32_array", "uint64", b"uint64", "uint64_array", b"uint64_array", "value", b"value"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["value", b"value"]) -> typing.Literal["string", "bool", "int32", "int64", "uint32", "uint64", "float", "double", "string_array", "bool_array", "int32_array", "int64_array", "uint32_array", "uint64_array", "float_array", "double_array"] | None: ... + +global___Datapoint = Datapoint + +@typing.final +class Metadata(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + DATA_TYPE_FIELD_NUMBER: builtins.int + ENTRY_TYPE_FIELD_NUMBER: builtins.int + DESCRIPTION_FIELD_NUMBER: builtins.int + COMMENT_FIELD_NUMBER: builtins.int + DEPRECATION_FIELD_NUMBER: builtins.int + UNIT_FIELD_NUMBER: builtins.int + VALUE_RESTRICTION_FIELD_NUMBER: builtins.int + ACTUATOR_FIELD_NUMBER: builtins.int + SENSOR_FIELD_NUMBER: builtins.int + ATTRIBUTE_FIELD_NUMBER: builtins.int + data_type: global___DataType.ValueType + """Data type + The VSS data type of the entry (i.e. the value, min, max etc). + + NOTE: protobuf doesn't have int8, int16, uint8 or uint16 which means + that these values must be serialized as int32 and uint32 respectively. + [field: FIELD_METADATA_DATA_TYPE] + """ + entry_type: global___EntryType.ValueType + """Entry type + [field: FIELD_METADATA_ENTRY_TYPE] + """ + description: builtins.str + """Description + Describes the meaning and content of the entry. + [field: FIELD_METADATA_DESCRIPTION] + """ + comment: builtins.str + """Comment [optional] + A comment can be used to provide additional informal information + on a entry. + [field: FIELD_METADATA_COMMENT] + """ + deprecation: builtins.str + """Deprecation [optional] + Whether this entry is deprecated. Can contain recommendations of what + to use instead. + [field: FIELD_METADATA_DEPRECATION] + """ + unit: builtins.str + """Unit [optional] + The unit of measurement + [field: FIELD_METADATA_UNIT] + """ + @property + def value_restriction(self) -> global___ValueRestriction: + """Value restrictions [optional] + Restrict which values are allowed. + Only restrictions matching the DataType {datatype} above are valid. + [field: FIELD_METADATA_VALUE_RESTRICTION] + """ + + @property + def actuator(self) -> global___Actuator: + """[field: FIELD_METADATA_ACTUATOR]""" + + @property + def sensor(self) -> global___Sensor: + """[field: FIELD_METADATA_SENSOR]""" + + @property + def attribute(self) -> global___Attribute: + """[field: FIELD_METADATA_ATTRIBUTE]""" + + def __init__( + self, + *, + data_type: global___DataType.ValueType = ..., + entry_type: global___EntryType.ValueType = ..., + description: builtins.str | None = ..., + comment: builtins.str | None = ..., + deprecation: builtins.str | None = ..., + unit: builtins.str | None = ..., + value_restriction: global___ValueRestriction | None = ..., + actuator: global___Actuator | None = ..., + sensor: global___Sensor | None = ..., + attribute: global___Attribute | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["_comment", b"_comment", "_deprecation", b"_deprecation", "_description", b"_description", "_unit", b"_unit", "actuator", b"actuator", "attribute", b"attribute", "comment", b"comment", "deprecation", b"deprecation", "description", b"description", "entry_specific", b"entry_specific", "sensor", b"sensor", "unit", b"unit", "value_restriction", b"value_restriction"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["_comment", b"_comment", "_deprecation", b"_deprecation", "_description", b"_description", "_unit", b"_unit", "actuator", b"actuator", "attribute", b"attribute", "comment", b"comment", "data_type", b"data_type", "deprecation", b"deprecation", "description", b"description", "entry_specific", b"entry_specific", "entry_type", b"entry_type", "sensor", b"sensor", "unit", b"unit", "value_restriction", b"value_restriction"]) -> None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_comment", b"_comment"]) -> typing.Literal["comment"] | None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_deprecation", b"_deprecation"]) -> typing.Literal["deprecation"] | None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_description", b"_description"]) -> typing.Literal["description"] | None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_unit", b"_unit"]) -> typing.Literal["unit"] | None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["entry_specific", b"entry_specific"]) -> typing.Literal["actuator", "sensor", "attribute"] | None: ... + +global___Metadata = Metadata + +@typing.final +class Actuator(google.protobuf.message.Message): + """///////////////////// + Actuator specific fields + Nothing for now + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___Actuator = Actuator + +@typing.final +class Sensor(google.protobuf.message.Message): + """////////////////////// + Sensor specific + Nothing for now + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___Sensor = Sensor + +@typing.final +class Attribute(google.protobuf.message.Message): + """////////////////////// + Attribute specific + Nothing for now. + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___Attribute = Attribute + +@typing.final +class ValueRestriction(google.protobuf.message.Message): + """Value restriction + + One ValueRestriction{type} for each type, since + they don't make sense unless the types match + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + STRING_FIELD_NUMBER: builtins.int + SIGNED_FIELD_NUMBER: builtins.int + UNSIGNED_FIELD_NUMBER: builtins.int + FLOATING_POINT_FIELD_NUMBER: builtins.int + @property + def string(self) -> global___ValueRestrictionString: ... + @property + def signed(self) -> global___ValueRestrictionInt: + """For signed VSS integers""" + + @property + def unsigned(self) -> global___ValueRestrictionUint: + """For unsigned VSS integers""" + + @property + def floating_point(self) -> global___ValueRestrictionFloat: + """For floating point VSS values (float and double)""" + + def __init__( + self, + *, + string: global___ValueRestrictionString | None = ..., + signed: global___ValueRestrictionInt | None = ..., + unsigned: global___ValueRestrictionUint | None = ..., + floating_point: global___ValueRestrictionFloat | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["floating_point", b"floating_point", "signed", b"signed", "string", b"string", "type", b"type", "unsigned", b"unsigned"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["floating_point", b"floating_point", "signed", b"signed", "string", b"string", "type", b"type", "unsigned", b"unsigned"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["type", b"type"]) -> typing.Literal["string", "signed", "unsigned", "floating_point"] | None: ... + +global___ValueRestriction = ValueRestriction + +@typing.final +class ValueRestrictionInt(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + MIN_FIELD_NUMBER: builtins.int + MAX_FIELD_NUMBER: builtins.int + ALLOWED_VALUES_FIELD_NUMBER: builtins.int + min: builtins.int + max: builtins.int + @property + def allowed_values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + min: builtins.int | None = ..., + max: builtins.int | None = ..., + allowed_values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["_max", b"_max", "_min", b"_min", "max", b"max", "min", b"min"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["_max", b"_max", "_min", b"_min", "allowed_values", b"allowed_values", "max", b"max", "min", b"min"]) -> None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_max", b"_max"]) -> typing.Literal["max"] | None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_min", b"_min"]) -> typing.Literal["min"] | None: ... + +global___ValueRestrictionInt = ValueRestrictionInt + +@typing.final +class ValueRestrictionUint(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + MIN_FIELD_NUMBER: builtins.int + MAX_FIELD_NUMBER: builtins.int + ALLOWED_VALUES_FIELD_NUMBER: builtins.int + min: builtins.int + max: builtins.int + @property + def allowed_values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + min: builtins.int | None = ..., + max: builtins.int | None = ..., + allowed_values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["_max", b"_max", "_min", b"_min", "max", b"max", "min", b"min"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["_max", b"_max", "_min", b"_min", "allowed_values", b"allowed_values", "max", b"max", "min", b"min"]) -> None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_max", b"_max"]) -> typing.Literal["max"] | None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_min", b"_min"]) -> typing.Literal["min"] | None: ... + +global___ValueRestrictionUint = ValueRestrictionUint + +@typing.final +class ValueRestrictionFloat(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + MIN_FIELD_NUMBER: builtins.int + MAX_FIELD_NUMBER: builtins.int + ALLOWED_VALUES_FIELD_NUMBER: builtins.int + min: builtins.float + max: builtins.float + @property + def allowed_values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.float]: + """allowed for doubles/floats not recommended""" + + def __init__( + self, + *, + min: builtins.float | None = ..., + max: builtins.float | None = ..., + allowed_values: collections.abc.Iterable[builtins.float] | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["_max", b"_max", "_min", b"_min", "max", b"max", "min", b"min"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["_max", b"_max", "_min", b"_min", "allowed_values", b"allowed_values", "max", b"max", "min", b"min"]) -> None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_max", b"_max"]) -> typing.Literal["max"] | None: ... + @typing.overload + def WhichOneof(self, oneof_group: typing.Literal["_min", b"_min"]) -> typing.Literal["min"] | None: ... + +global___ValueRestrictionFloat = ValueRestrictionFloat + +@typing.final +class ValueRestrictionString(google.protobuf.message.Message): + """min, max doesn't make much sense for a string""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ALLOWED_VALUES_FIELD_NUMBER: builtins.int + @property + def allowed_values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: ... + def __init__( + self, + *, + allowed_values: collections.abc.Iterable[builtins.str] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["allowed_values", b"allowed_values"]) -> None: ... + +global___ValueRestrictionString = ValueRestrictionString + +@typing.final +class Error(google.protobuf.message.Message): + """Error response shall be an HTTP-like code. + Should follow https://www.w3.org/TR/viss2-transport/#status-codes. + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + CODE_FIELD_NUMBER: builtins.int + REASON_FIELD_NUMBER: builtins.int + MESSAGE_FIELD_NUMBER: builtins.int + code: builtins.int + reason: builtins.str + message: builtins.str + def __init__( + self, + *, + code: builtins.int = ..., + reason: builtins.str = ..., + message: builtins.str = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["code", b"code", "message", b"message", "reason", b"reason"]) -> None: ... + +global___Error = Error + +@typing.final +class DataEntryError(google.protobuf.message.Message): + """Used in get/set requests to report errors for specific entries""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + PATH_FIELD_NUMBER: builtins.int + ERROR_FIELD_NUMBER: builtins.int + path: builtins.str + """vss path""" + @property + def error(self) -> global___Error: ... + def __init__( + self, + *, + path: builtins.str = ..., + error: global___Error | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["error", b"error"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["error", b"error", "path", b"path"]) -> None: ... + +global___DataEntryError = DataEntryError + +@typing.final +class StringArray(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.str] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___StringArray = StringArray + +@typing.final +class BoolArray(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.bool] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___BoolArray = BoolArray + +@typing.final +class Int32Array(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___Int32Array = Int32Array + +@typing.final +class Int64Array(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___Int64Array = Int64Array + +@typing.final +class Uint32Array(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___Uint32Array = Uint32Array + +@typing.final +class Uint64Array(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___Uint64Array = Uint64Array + +@typing.final +class FloatArray(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.float]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.float] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___FloatArray = FloatArray + +@typing.final +class DoubleArray(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.float]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.float] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___DoubleArray = DoubleArray diff --git a/integration_test/gen_proto/kuksa/val/v1/types_pb2_grpc.py b/integration_test/gen_proto/kuksa/val/v1/types_pb2_grpc.py new file mode 100644 index 00000000..a1701be9 --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v1/types_pb2_grpc.py @@ -0,0 +1,37 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +# Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT! +"""Client and server classes corresponding to protobuf-defined services.""" +import grpc +import warnings + + +GRPC_GENERATED_VERSION = '1.71.0' +GRPC_VERSION = grpc.__version__ +_version_not_supported = False + +try: + from grpc._utilities import first_version_is_lower + _version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION) +except ImportError: + _version_not_supported = True + +if _version_not_supported: + raise RuntimeError( + f'The grpc package installed is at version {GRPC_VERSION},' + + f' but the generated code in kuksa/val/v1/types_pb2_grpc.py depends on' + + f' grpcio>={GRPC_GENERATED_VERSION}.' + + f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}' + + f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.' + ) diff --git a/integration_test/gen_proto/kuksa/val/v1/val_pb2.py b/integration_test/gen_proto/kuksa/val/v1/val_pb2.py new file mode 100644 index 00000000..76826c55 --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v1/val_pb2.py @@ -0,0 +1,77 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +# -*- coding: utf-8 -*- +# Generated by the protocol buffer compiler. DO NOT EDIT! +# NO CHECKED-IN PROTOBUF GENCODE +# source: kuksa/val/v1/val.proto +# Protobuf Python Version: 5.29.0 +"""Generated protocol buffer code.""" +from google.protobuf import descriptor as _descriptor +from google.protobuf import descriptor_pool as _descriptor_pool +from google.protobuf import runtime_version as _runtime_version +from google.protobuf import symbol_database as _symbol_database +from google.protobuf.internal import builder as _builder +_runtime_version.ValidateProtobufRuntimeVersion( + _runtime_version.Domain.PUBLIC, + 5, + 29, + 0, + '', + 'kuksa/val/v1/val.proto' +) +# @@protoc_insertion_point(imports) + +_sym_db = _symbol_database.Default() + + +from kuksa.val.v1 import types_pb2 as kuksa_dot_val_dot_v1_dot_types__pb2 + + +DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\x16kuksa/val/v1/val.proto\x12\x0ckuksa.val.v1\x1a\x18kuksa/val/v1/types.proto\"c\n\x0c\x45ntryRequest\x12\x0c\n\x04path\x18\x01 \x01(\t\x12 \n\x04view\x18\x02 \x01(\x0e\x32\x12.kuksa.val.v1.View\x12#\n\x06\x66ields\x18\x03 \x03(\x0e\x32\x13.kuksa.val.v1.Field\"9\n\nGetRequest\x12+\n\x07\x65ntries\x18\x01 \x03(\x0b\x32\x1a.kuksa.val.v1.EntryRequest\"\x89\x01\n\x0bGetResponse\x12(\n\x07\x65ntries\x18\x01 \x03(\x0b\x32\x17.kuksa.val.v1.DataEntry\x12,\n\x06\x65rrors\x18\x02 \x03(\x0b\x32\x1c.kuksa.val.v1.DataEntryError\x12\"\n\x05\x65rror\x18\x03 \x01(\x0b\x32\x13.kuksa.val.v1.Error\"Z\n\x0b\x45ntryUpdate\x12&\n\x05\x65ntry\x18\x01 \x01(\x0b\x32\x17.kuksa.val.v1.DataEntry\x12#\n\x06\x66ields\x18\x02 \x03(\x0e\x32\x13.kuksa.val.v1.Field\"8\n\nSetRequest\x12*\n\x07updates\x18\x01 \x03(\x0b\x32\x19.kuksa.val.v1.EntryUpdate\"_\n\x0bSetResponse\x12\"\n\x05\x65rror\x18\x01 \x01(\x0b\x32\x13.kuksa.val.v1.Error\x12,\n\x06\x65rrors\x18\x02 \x03(\x0b\x32\x1c.kuksa.val.v1.DataEntryError\"C\n\x15StreamedUpdateRequest\x12*\n\x07updates\x18\x01 \x03(\x0b\x32\x19.kuksa.val.v1.EntryUpdate\"j\n\x16StreamedUpdateResponse\x12\"\n\x05\x65rror\x18\x01 \x01(\x0b\x32\x13.kuksa.val.v1.Error\x12,\n\x06\x65rrors\x18\x02 \x03(\x0b\x32\x1c.kuksa.val.v1.DataEntryError\"e\n\x0eSubscribeEntry\x12\x0c\n\x04path\x18\x01 \x01(\t\x12 \n\x04view\x18\x02 \x01(\x0e\x32\x12.kuksa.val.v1.View\x12#\n\x06\x66ields\x18\x03 \x03(\x0e\x32\x13.kuksa.val.v1.Field\"A\n\x10SubscribeRequest\x12-\n\x07\x65ntries\x18\x01 \x03(\x0b\x32\x1c.kuksa.val.v1.SubscribeEntry\"?\n\x11SubscribeResponse\x12*\n\x07updates\x18\x01 \x03(\x0b\x32\x19.kuksa.val.v1.EntryUpdate\"\x16\n\x14GetServerInfoRequest\"6\n\x15GetServerInfoResponse\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x0f\n\x07version\x18\x02 \x01(\t2\x88\x03\n\x03VAL\x12:\n\x03Get\x12\x18.kuksa.val.v1.GetRequest\x1a\x19.kuksa.val.v1.GetResponse\x12:\n\x03Set\x12\x18.kuksa.val.v1.SetRequest\x1a\x19.kuksa.val.v1.SetResponse\x12_\n\x0eStreamedUpdate\x12#.kuksa.val.v1.StreamedUpdateRequest\x1a$.kuksa.val.v1.StreamedUpdateResponse(\x01\x30\x01\x12N\n\tSubscribe\x12\x1e.kuksa.val.v1.SubscribeRequest\x1a\x1f.kuksa.val.v1.SubscribeResponse0\x01\x12X\n\rGetServerInfo\x12\".kuksa.val.v1.GetServerInfoRequest\x1a#.kuksa.val.v1.GetServerInfoResponseB\x0eZ\x0ckuksa/val/v1b\x06proto3') + +_globals = globals() +_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) +_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'kuksa.val.v1.val_pb2', _globals) +if not _descriptor._USE_C_DESCRIPTORS: + _globals['DESCRIPTOR']._loaded_options = None + _globals['DESCRIPTOR']._serialized_options = b'Z\014kuksa/val/v1' + _globals['_ENTRYREQUEST']._serialized_start=66 + _globals['_ENTRYREQUEST']._serialized_end=165 + _globals['_GETREQUEST']._serialized_start=167 + _globals['_GETREQUEST']._serialized_end=224 + _globals['_GETRESPONSE']._serialized_start=227 + _globals['_GETRESPONSE']._serialized_end=364 + _globals['_ENTRYUPDATE']._serialized_start=366 + _globals['_ENTRYUPDATE']._serialized_end=456 + _globals['_SETREQUEST']._serialized_start=458 + _globals['_SETREQUEST']._serialized_end=514 + _globals['_SETRESPONSE']._serialized_start=516 + _globals['_SETRESPONSE']._serialized_end=611 + _globals['_STREAMEDUPDATEREQUEST']._serialized_start=613 + _globals['_STREAMEDUPDATEREQUEST']._serialized_end=680 + _globals['_STREAMEDUPDATERESPONSE']._serialized_start=682 + _globals['_STREAMEDUPDATERESPONSE']._serialized_end=788 + _globals['_SUBSCRIBEENTRY']._serialized_start=790 + _globals['_SUBSCRIBEENTRY']._serialized_end=891 + _globals['_SUBSCRIBEREQUEST']._serialized_start=893 + _globals['_SUBSCRIBEREQUEST']._serialized_end=958 + _globals['_SUBSCRIBERESPONSE']._serialized_start=960 + _globals['_SUBSCRIBERESPONSE']._serialized_end=1023 + _globals['_GETSERVERINFOREQUEST']._serialized_start=1025 + _globals['_GETSERVERINFOREQUEST']._serialized_end=1047 + _globals['_GETSERVERINFORESPONSE']._serialized_start=1049 + _globals['_GETSERVERINFORESPONSE']._serialized_end=1103 + _globals['_VAL']._serialized_start=1106 + _globals['_VAL']._serialized_end=1498 +# @@protoc_insertion_point(module_scope) diff --git a/integration_test/gen_proto/kuksa/val/v1/val_pb2.pyi b/integration_test/gen_proto/kuksa/val/v1/val_pb2.pyi new file mode 100644 index 00000000..7ba79f02 --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v1/val_pb2.pyi @@ -0,0 +1,303 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +""" +@generated by mypy-protobuf. Do not edit manually! +isort:skip_file +******************************************************************************* +Copyright (c) 2022 Contributors to the Eclipse Foundation + +See the NOTICE file(s) distributed with this work for additional +information regarding copyright ownership. + +This program and the accompanying materials are made available under the +terms of the Apache License 2.0 which is available at +http://www.apache.org/licenses/LICENSE-2.0 + +SPDX-License-Identifier: Apache-2.0 +****************************************************************************** +""" + +import builtins +import collections.abc +import google.protobuf.descriptor +import google.protobuf.internal.containers +import google.protobuf.message +import kuksa.val.v1.types_pb2 +import typing + +DESCRIPTOR: google.protobuf.descriptor.FileDescriptor + +@typing.final +class EntryRequest(google.protobuf.message.Message): + """Define which data we want""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + PATH_FIELD_NUMBER: builtins.int + VIEW_FIELD_NUMBER: builtins.int + FIELDS_FIELD_NUMBER: builtins.int + path: builtins.str + view: kuksa.val.v1.types_pb2.View.ValueType + @property + def fields(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[kuksa.val.v1.types_pb2.Field.ValueType]: ... + def __init__( + self, + *, + path: builtins.str = ..., + view: kuksa.val.v1.types_pb2.View.ValueType = ..., + fields: collections.abc.Iterable[kuksa.val.v1.types_pb2.Field.ValueType] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["fields", b"fields", "path", b"path", "view", b"view"]) -> None: ... + +global___EntryRequest = EntryRequest + +@typing.final +class GetRequest(google.protobuf.message.Message): + """Request a set of entries.""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ENTRIES_FIELD_NUMBER: builtins.int + @property + def entries(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___EntryRequest]: ... + def __init__( + self, + *, + entries: collections.abc.Iterable[global___EntryRequest] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["entries", b"entries"]) -> None: ... + +global___GetRequest = GetRequest + +@typing.final +class GetResponse(google.protobuf.message.Message): + """Global errors are specified in `error`. + Errors for individual entries are specified in `errors`. + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ENTRIES_FIELD_NUMBER: builtins.int + ERRORS_FIELD_NUMBER: builtins.int + ERROR_FIELD_NUMBER: builtins.int + @property + def entries(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[kuksa.val.v1.types_pb2.DataEntry]: ... + @property + def errors(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[kuksa.val.v1.types_pb2.DataEntryError]: ... + @property + def error(self) -> kuksa.val.v1.types_pb2.Error: ... + def __init__( + self, + *, + entries: collections.abc.Iterable[kuksa.val.v1.types_pb2.DataEntry] | None = ..., + errors: collections.abc.Iterable[kuksa.val.v1.types_pb2.DataEntryError] | None = ..., + error: kuksa.val.v1.types_pb2.Error | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["error", b"error"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["entries", b"entries", "error", b"error", "errors", b"errors"]) -> None: ... + +global___GetResponse = GetResponse + +@typing.final +class EntryUpdate(google.protobuf.message.Message): + """Define the data we want to set""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ENTRY_FIELD_NUMBER: builtins.int + FIELDS_FIELD_NUMBER: builtins.int + @property + def entry(self) -> kuksa.val.v1.types_pb2.DataEntry: ... + @property + def fields(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[kuksa.val.v1.types_pb2.Field.ValueType]: ... + def __init__( + self, + *, + entry: kuksa.val.v1.types_pb2.DataEntry | None = ..., + fields: collections.abc.Iterable[kuksa.val.v1.types_pb2.Field.ValueType] | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["entry", b"entry"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["entry", b"entry", "fields", b"fields"]) -> None: ... + +global___EntryUpdate = EntryUpdate + +@typing.final +class SetRequest(google.protobuf.message.Message): + """A list of entries to be updated""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + UPDATES_FIELD_NUMBER: builtins.int + @property + def updates(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___EntryUpdate]: ... + def __init__( + self, + *, + updates: collections.abc.Iterable[global___EntryUpdate] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["updates", b"updates"]) -> None: ... + +global___SetRequest = SetRequest + +@typing.final +class SetResponse(google.protobuf.message.Message): + """Global errors are specified in `error`. + Errors for individual entries are specified in `errors`. + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ERROR_FIELD_NUMBER: builtins.int + ERRORS_FIELD_NUMBER: builtins.int + @property + def error(self) -> kuksa.val.v1.types_pb2.Error: ... + @property + def errors(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[kuksa.val.v1.types_pb2.DataEntryError]: ... + def __init__( + self, + *, + error: kuksa.val.v1.types_pb2.Error | None = ..., + errors: collections.abc.Iterable[kuksa.val.v1.types_pb2.DataEntryError] | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["error", b"error"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["error", b"error", "errors", b"errors"]) -> None: ... + +global___SetResponse = SetResponse + +@typing.final +class StreamedUpdateRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + UPDATES_FIELD_NUMBER: builtins.int + @property + def updates(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___EntryUpdate]: ... + def __init__( + self, + *, + updates: collections.abc.Iterable[global___EntryUpdate] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["updates", b"updates"]) -> None: ... + +global___StreamedUpdateRequest = StreamedUpdateRequest + +@typing.final +class StreamedUpdateResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ERROR_FIELD_NUMBER: builtins.int + ERRORS_FIELD_NUMBER: builtins.int + @property + def error(self) -> kuksa.val.v1.types_pb2.Error: ... + @property + def errors(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[kuksa.val.v1.types_pb2.DataEntryError]: ... + def __init__( + self, + *, + error: kuksa.val.v1.types_pb2.Error | None = ..., + errors: collections.abc.Iterable[kuksa.val.v1.types_pb2.DataEntryError] | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["error", b"error"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["error", b"error", "errors", b"errors"]) -> None: ... + +global___StreamedUpdateResponse = StreamedUpdateResponse + +@typing.final +class SubscribeEntry(google.protobuf.message.Message): + """Define what to subscribe to""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + PATH_FIELD_NUMBER: builtins.int + VIEW_FIELD_NUMBER: builtins.int + FIELDS_FIELD_NUMBER: builtins.int + path: builtins.str + view: kuksa.val.v1.types_pb2.View.ValueType + @property + def fields(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[kuksa.val.v1.types_pb2.Field.ValueType]: ... + def __init__( + self, + *, + path: builtins.str = ..., + view: kuksa.val.v1.types_pb2.View.ValueType = ..., + fields: collections.abc.Iterable[kuksa.val.v1.types_pb2.Field.ValueType] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["fields", b"fields", "path", b"path", "view", b"view"]) -> None: ... + +global___SubscribeEntry = SubscribeEntry + +@typing.final +class SubscribeRequest(google.protobuf.message.Message): + """Subscribe to changes in datapoints.""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ENTRIES_FIELD_NUMBER: builtins.int + @property + def entries(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___SubscribeEntry]: ... + def __init__( + self, + *, + entries: collections.abc.Iterable[global___SubscribeEntry] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["entries", b"entries"]) -> None: ... + +global___SubscribeRequest = SubscribeRequest + +@typing.final +class SubscribeResponse(google.protobuf.message.Message): + """A subscription response""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + UPDATES_FIELD_NUMBER: builtins.int + @property + def updates(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___EntryUpdate]: ... + def __init__( + self, + *, + updates: collections.abc.Iterable[global___EntryUpdate] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["updates", b"updates"]) -> None: ... + +global___SubscribeResponse = SubscribeResponse + +@typing.final +class GetServerInfoRequest(google.protobuf.message.Message): + """Nothing yet""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___GetServerInfoRequest = GetServerInfoRequest + +@typing.final +class GetServerInfoResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + NAME_FIELD_NUMBER: builtins.int + VERSION_FIELD_NUMBER: builtins.int + name: builtins.str + version: builtins.str + def __init__( + self, + *, + name: builtins.str = ..., + version: builtins.str = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["name", b"name", "version", b"version"]) -> None: ... + +global___GetServerInfoResponse = GetServerInfoResponse diff --git a/integration_test/gen_proto/kuksa/val/v1/val_pb2_grpc.py b/integration_test/gen_proto/kuksa/val/v1/val_pb2_grpc.py new file mode 100644 index 00000000..08d56b9b --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v1/val_pb2_grpc.py @@ -0,0 +1,325 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +# Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT! +"""Client and server classes corresponding to protobuf-defined services.""" +import grpc +import warnings + +from kuksa.val.v1 import val_pb2 as kuksa_dot_val_dot_v1_dot_val__pb2 + +GRPC_GENERATED_VERSION = '1.71.0' +GRPC_VERSION = grpc.__version__ +_version_not_supported = False + +try: + from grpc._utilities import first_version_is_lower + _version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION) +except ImportError: + _version_not_supported = True + +if _version_not_supported: + raise RuntimeError( + f'The grpc package installed is at version {GRPC_VERSION},' + + f' but the generated code in kuksa/val/v1/val_pb2_grpc.py depends on' + + f' grpcio>={GRPC_GENERATED_VERSION}.' + + f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}' + + f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.' + ) + + +class VALStub(object): + """Note on authorization: + Tokens (auth-token or auth-uuid) are sent as (GRPC / http2) metadata. + + The auth-token is a JWT compliant token as the examples found here: + https://github.com/eclipse-kuksa/kuksa-databroker/tree/main/certificates/jwt + + See also https://github.com/eclipse-kuksa/kuksa-databroker/blob/main/doc/authorization.md#jwt-access-token + + Upon reception of auth-token, server shall generate an auth-uuid in metadata + that the client can use instead of auth-token in subsequent calls. + + """ + + def __init__(self, channel): + """Constructor. + + Args: + channel: A grpc.Channel. + """ + self.Get = channel.unary_unary( + '/kuksa.val.v1.VAL/Get', + request_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.GetRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.GetResponse.FromString, + _registered_method=True) + self.Set = channel.unary_unary( + '/kuksa.val.v1.VAL/Set', + request_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.SetRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.SetResponse.FromString, + _registered_method=True) + self.StreamedUpdate = channel.stream_stream( + '/kuksa.val.v1.VAL/StreamedUpdate', + request_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.StreamedUpdateRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.StreamedUpdateResponse.FromString, + _registered_method=True) + self.Subscribe = channel.unary_stream( + '/kuksa.val.v1.VAL/Subscribe', + request_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.SubscribeRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.SubscribeResponse.FromString, + _registered_method=True) + self.GetServerInfo = channel.unary_unary( + '/kuksa.val.v1.VAL/GetServerInfo', + request_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.GetServerInfoRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.GetServerInfoResponse.FromString, + _registered_method=True) + + +class VALServicer(object): + """Note on authorization: + Tokens (auth-token or auth-uuid) are sent as (GRPC / http2) metadata. + + The auth-token is a JWT compliant token as the examples found here: + https://github.com/eclipse-kuksa/kuksa-databroker/tree/main/certificates/jwt + + See also https://github.com/eclipse-kuksa/kuksa-databroker/blob/main/doc/authorization.md#jwt-access-token + + Upon reception of auth-token, server shall generate an auth-uuid in metadata + that the client can use instead of auth-token in subsequent calls. + + """ + + def Get(self, request, context): + """Get entries + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def Set(self, request, context): + """Set entries + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def StreamedUpdate(self, request_iterator, context): + """Missing associated documentation comment in .proto file.""" + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def Subscribe(self, request, context): + """Subscribe to a set of entries + + Returns a stream of notifications. + + InvalidArgument is returned if the request is malformed. + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def GetServerInfo(self, request, context): + """Shall return information that allows the client to determine + what server/server implementation/version it is talking to + eg. kuksa-databroker 0.5.1 + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + +def add_VALServicer_to_server(servicer, server): + rpc_method_handlers = { + 'Get': grpc.unary_unary_rpc_method_handler( + servicer.Get, + request_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.GetRequest.FromString, + response_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.GetResponse.SerializeToString, + ), + 'Set': grpc.unary_unary_rpc_method_handler( + servicer.Set, + request_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.SetRequest.FromString, + response_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.SetResponse.SerializeToString, + ), + 'StreamedUpdate': grpc.stream_stream_rpc_method_handler( + servicer.StreamedUpdate, + request_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.StreamedUpdateRequest.FromString, + response_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.StreamedUpdateResponse.SerializeToString, + ), + 'Subscribe': grpc.unary_stream_rpc_method_handler( + servicer.Subscribe, + request_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.SubscribeRequest.FromString, + response_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.SubscribeResponse.SerializeToString, + ), + 'GetServerInfo': grpc.unary_unary_rpc_method_handler( + servicer.GetServerInfo, + request_deserializer=kuksa_dot_val_dot_v1_dot_val__pb2.GetServerInfoRequest.FromString, + response_serializer=kuksa_dot_val_dot_v1_dot_val__pb2.GetServerInfoResponse.SerializeToString, + ), + } + generic_handler = grpc.method_handlers_generic_handler( + 'kuksa.val.v1.VAL', rpc_method_handlers) + server.add_generic_rpc_handlers((generic_handler,)) + server.add_registered_method_handlers('kuksa.val.v1.VAL', rpc_method_handlers) + + + # This class is part of an EXPERIMENTAL API. +class VAL(object): + """Note on authorization: + Tokens (auth-token or auth-uuid) are sent as (GRPC / http2) metadata. + + The auth-token is a JWT compliant token as the examples found here: + https://github.com/eclipse-kuksa/kuksa-databroker/tree/main/certificates/jwt + + See also https://github.com/eclipse-kuksa/kuksa-databroker/blob/main/doc/authorization.md#jwt-access-token + + Upon reception of auth-token, server shall generate an auth-uuid in metadata + that the client can use instead of auth-token in subsequent calls. + + """ + + @staticmethod + def Get(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v1.VAL/Get', + kuksa_dot_val_dot_v1_dot_val__pb2.GetRequest.SerializeToString, + kuksa_dot_val_dot_v1_dot_val__pb2.GetResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def Set(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v1.VAL/Set', + kuksa_dot_val_dot_v1_dot_val__pb2.SetRequest.SerializeToString, + kuksa_dot_val_dot_v1_dot_val__pb2.SetResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def StreamedUpdate(request_iterator, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.stream_stream( + request_iterator, + target, + '/kuksa.val.v1.VAL/StreamedUpdate', + kuksa_dot_val_dot_v1_dot_val__pb2.StreamedUpdateRequest.SerializeToString, + kuksa_dot_val_dot_v1_dot_val__pb2.StreamedUpdateResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def Subscribe(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_stream( + request, + target, + '/kuksa.val.v1.VAL/Subscribe', + kuksa_dot_val_dot_v1_dot_val__pb2.SubscribeRequest.SerializeToString, + kuksa_dot_val_dot_v1_dot_val__pb2.SubscribeResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def GetServerInfo(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v1.VAL/GetServerInfo', + kuksa_dot_val_dot_v1_dot_val__pb2.GetServerInfoRequest.SerializeToString, + kuksa_dot_val_dot_v1_dot_val__pb2.GetServerInfoResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) diff --git a/integration_test/gen_proto/kuksa/val/v2/types_pb2.py b/integration_test/gen_proto/kuksa/val/v2/types_pb2.py new file mode 100644 index 00000000..28997035 --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v2/types_pb2.py @@ -0,0 +1,89 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +# -*- coding: utf-8 -*- +# Generated by the protocol buffer compiler. DO NOT EDIT! +# NO CHECKED-IN PROTOBUF GENCODE +# source: kuksa/val/v2/types.proto +# Protobuf Python Version: 5.29.0 +"""Generated protocol buffer code.""" +from google.protobuf import descriptor as _descriptor +from google.protobuf import descriptor_pool as _descriptor_pool +from google.protobuf import runtime_version as _runtime_version +from google.protobuf import symbol_database as _symbol_database +from google.protobuf.internal import builder as _builder +_runtime_version.ValidateProtobufRuntimeVersion( + _runtime_version.Domain.PUBLIC, + 5, + 29, + 0, + '', + 'kuksa/val/v2/types.proto' +) +# @@protoc_insertion_point(imports) + +_sym_db = _symbol_database.Default() + + +from google.protobuf import timestamp_pb2 as google_dot_protobuf_dot_timestamp__pb2 + + +DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\x18kuksa/val/v2/types.proto\x12\x0ckuksa.val.v2\x1a\x1fgoogle/protobuf/timestamp.proto\"^\n\tDatapoint\x12-\n\ttimestamp\x18\x01 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12\"\n\x05value\x18\x02 \x01(\x0b\x32\x13.kuksa.val.v2.Value\"\xaf\x04\n\x05Value\x12\x10\n\x06string\x18\x0b \x01(\tH\x00\x12\x0e\n\x04\x62ool\x18\x0c \x01(\x08H\x00\x12\x0f\n\x05int32\x18\r \x01(\x11H\x00\x12\x0f\n\x05int64\x18\x0e \x01(\x12H\x00\x12\x10\n\x06uint32\x18\x0f \x01(\rH\x00\x12\x10\n\x06uint64\x18\x10 \x01(\x04H\x00\x12\x0f\n\x05\x66loat\x18\x11 \x01(\x02H\x00\x12\x10\n\x06\x64ouble\x18\x12 \x01(\x01H\x00\x12\x31\n\x0cstring_array\x18\x15 \x01(\x0b\x32\x19.kuksa.val.v2.StringArrayH\x00\x12-\n\nbool_array\x18\x16 \x01(\x0b\x32\x17.kuksa.val.v2.BoolArrayH\x00\x12/\n\x0bint32_array\x18\x17 \x01(\x0b\x32\x18.kuksa.val.v2.Int32ArrayH\x00\x12/\n\x0bint64_array\x18\x18 \x01(\x0b\x32\x18.kuksa.val.v2.Int64ArrayH\x00\x12\x31\n\x0cuint32_array\x18\x19 \x01(\x0b\x32\x19.kuksa.val.v2.Uint32ArrayH\x00\x12\x31\n\x0cuint64_array\x18\x1a \x01(\x0b\x32\x19.kuksa.val.v2.Uint64ArrayH\x00\x12/\n\x0b\x66loat_array\x18\x1b \x01(\x0b\x32\x18.kuksa.val.v2.FloatArrayH\x00\x12\x31\n\x0c\x64ouble_array\x18\x1c \x01(\x0b\x32\x19.kuksa.val.v2.DoubleArrayH\x00\x42\r\n\x0btyped_value\"%\n\x0eSampleInterval\x12\x13\n\x0binterval_ms\x18\x01 \x01(\r\"X\n\x06\x46ilter\x12\x13\n\x0b\x64uration_ms\x18\x01 \x01(\r\x12\x39\n\x13min_sample_interval\x18\x02 \x01(\x0b\x32\x1c.kuksa.val.v2.SampleInterval\"2\n\x08SignalID\x12\x0c\n\x02id\x18\x01 \x01(\x05H\x00\x12\x0e\n\x04path\x18\x02 \x01(\tH\x00\x42\x08\n\x06signal\"?\n\x05\x45rror\x12%\n\x04\x63ode\x18\x01 \x01(\x0e\x32\x17.kuksa.val.v2.ErrorCode\x12\x0f\n\x07message\x18\x02 \x01(\t\"\xf1\x02\n\x08Metadata\x12\x0c\n\x04path\x18\t \x01(\t\x12\n\n\x02id\x18\n \x01(\x05\x12)\n\tdata_type\x18\x0b \x01(\x0e\x32\x16.kuksa.val.v2.DataType\x12+\n\nentry_type\x18\x0c \x01(\x0e\x32\x17.kuksa.val.v2.EntryType\x12\x13\n\x0b\x64\x65scription\x18\r \x01(\t\x12\x0f\n\x07\x63omment\x18\x0e \x01(\t\x12\x13\n\x0b\x64\x65precation\x18\x0f \x01(\t\x12\x0c\n\x04unit\x18\x10 \x01(\t\x12+\n\x0e\x61llowed_values\x18\x11 \x01(\x0b\x32\x13.kuksa.val.v2.Value\x12 \n\x03min\x18\x12 \x01(\x0b\x32\x13.kuksa.val.v2.Value\x12 \n\x03max\x18\x13 \x01(\x0b\x32\x13.kuksa.val.v2.Value\x12\x39\n\x13min_sample_interval\x18\x14 \x01(\x0b\x32\x1c.kuksa.val.v2.SampleInterval\"\x1d\n\x0bStringArray\x12\x0e\n\x06values\x18\x01 \x03(\t\"\x1b\n\tBoolArray\x12\x0e\n\x06values\x18\x01 \x03(\x08\"\x1c\n\nInt32Array\x12\x0e\n\x06values\x18\x01 \x03(\x11\"\x1c\n\nInt64Array\x12\x0e\n\x06values\x18\x01 \x03(\x12\"\x1d\n\x0bUint32Array\x12\x0e\n\x06values\x18\x01 \x03(\r\"\x1d\n\x0bUint64Array\x12\x0e\n\x06values\x18\x01 \x03(\x04\"\x1c\n\nFloatArray\x12\x0e\n\x06values\x18\x01 \x03(\x02\"\x1d\n\x0b\x44oubleArray\x12\x0e\n\x06values\x18\x01 \x03(\x01*Y\n\x0b\x46ilterError\x12!\n\x1d\x46ILTER_ERROR_CODE_UNSPECIFIED\x10\x00\x12\'\n#FILTER_ERROR_CODE_UNKNOWN_SINGAL_ID\x10\x01*P\n\rProviderError\x12\x14\n\x10\x43ODE_UNSPECIFIED\x10\x00\x12\x16\n\x12\x43ODE_NETWORK_ERROR\x10\x01\x12\x11\n\rCODE_OVERLOAD\x10\x02*\x97\x01\n\tErrorCode\x12\x1a\n\x16\x45RROR_CODE_UNSPECIFIED\x10\x00\x12\x11\n\rERROR_CODE_OK\x10\x01\x12\x1f\n\x1b\x45RROR_CODE_INVALID_ARGUMENT\x10\x02\x12\x18\n\x14\x45RROR_CODE_NOT_FOUND\x10\x03\x12 \n\x1c\x45RROR_CODE_PERMISSION_DENIED\x10\x04*\xa9\x05\n\x08\x44\x61taType\x12\x19\n\x15\x44\x41TA_TYPE_UNSPECIFIED\x10\x00\x12\x14\n\x10\x44\x41TA_TYPE_STRING\x10\x01\x12\x15\n\x11\x44\x41TA_TYPE_BOOLEAN\x10\x02\x12\x12\n\x0e\x44\x41TA_TYPE_INT8\x10\x03\x12\x13\n\x0f\x44\x41TA_TYPE_INT16\x10\x04\x12\x13\n\x0f\x44\x41TA_TYPE_INT32\x10\x05\x12\x13\n\x0f\x44\x41TA_TYPE_INT64\x10\x06\x12\x13\n\x0f\x44\x41TA_TYPE_UINT8\x10\x07\x12\x14\n\x10\x44\x41TA_TYPE_UINT16\x10\x08\x12\x14\n\x10\x44\x41TA_TYPE_UINT32\x10\t\x12\x14\n\x10\x44\x41TA_TYPE_UINT64\x10\n\x12\x13\n\x0f\x44\x41TA_TYPE_FLOAT\x10\x0b\x12\x14\n\x10\x44\x41TA_TYPE_DOUBLE\x10\x0c\x12\x17\n\x13\x44\x41TA_TYPE_TIMESTAMP\x10\r\x12\x1a\n\x16\x44\x41TA_TYPE_STRING_ARRAY\x10\x14\x12\x1b\n\x17\x44\x41TA_TYPE_BOOLEAN_ARRAY\x10\x15\x12\x18\n\x14\x44\x41TA_TYPE_INT8_ARRAY\x10\x16\x12\x19\n\x15\x44\x41TA_TYPE_INT16_ARRAY\x10\x17\x12\x19\n\x15\x44\x41TA_TYPE_INT32_ARRAY\x10\x18\x12\x19\n\x15\x44\x41TA_TYPE_INT64_ARRAY\x10\x19\x12\x19\n\x15\x44\x41TA_TYPE_UINT8_ARRAY\x10\x1a\x12\x1a\n\x16\x44\x41TA_TYPE_UINT16_ARRAY\x10\x1b\x12\x1a\n\x16\x44\x41TA_TYPE_UINT32_ARRAY\x10\x1c\x12\x1a\n\x16\x44\x41TA_TYPE_UINT64_ARRAY\x10\x1d\x12\x19\n\x15\x44\x41TA_TYPE_FLOAT_ARRAY\x10\x1e\x12\x1a\n\x16\x44\x41TA_TYPE_DOUBLE_ARRAY\x10\x1f\x12\x1d\n\x19\x44\x41TA_TYPE_TIMESTAMP_ARRAY\x10 *q\n\tEntryType\x12\x1a\n\x16\x45NTRY_TYPE_UNSPECIFIED\x10\x00\x12\x18\n\x14\x45NTRY_TYPE_ATTRIBUTE\x10\x01\x12\x15\n\x11\x45NTRY_TYPE_SENSOR\x10\x02\x12\x17\n\x13\x45NTRY_TYPE_ACTUATOR\x10\x03\x42\x0eZ\x0ckuksa/val/v2b\x06proto3') + +_globals = globals() +_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) +_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'kuksa.val.v2.types_pb2', _globals) +if not _descriptor._USE_C_DESCRIPTORS: + _globals['DESCRIPTOR']._loaded_options = None + _globals['DESCRIPTOR']._serialized_options = b'Z\014kuksa/val/v2' + _globals['_FILTERERROR']._serialized_start=1594 + _globals['_FILTERERROR']._serialized_end=1683 + _globals['_PROVIDERERROR']._serialized_start=1685 + _globals['_PROVIDERERROR']._serialized_end=1765 + _globals['_ERRORCODE']._serialized_start=1768 + _globals['_ERRORCODE']._serialized_end=1919 + _globals['_DATATYPE']._serialized_start=1922 + _globals['_DATATYPE']._serialized_end=2603 + _globals['_ENTRYTYPE']._serialized_start=2605 + _globals['_ENTRYTYPE']._serialized_end=2718 + _globals['_DATAPOINT']._serialized_start=75 + _globals['_DATAPOINT']._serialized_end=169 + _globals['_VALUE']._serialized_start=172 + _globals['_VALUE']._serialized_end=731 + _globals['_SAMPLEINTERVAL']._serialized_start=733 + _globals['_SAMPLEINTERVAL']._serialized_end=770 + _globals['_FILTER']._serialized_start=772 + _globals['_FILTER']._serialized_end=860 + _globals['_SIGNALID']._serialized_start=862 + _globals['_SIGNALID']._serialized_end=912 + _globals['_ERROR']._serialized_start=914 + _globals['_ERROR']._serialized_end=977 + _globals['_METADATA']._serialized_start=980 + _globals['_METADATA']._serialized_end=1349 + _globals['_STRINGARRAY']._serialized_start=1351 + _globals['_STRINGARRAY']._serialized_end=1380 + _globals['_BOOLARRAY']._serialized_start=1382 + _globals['_BOOLARRAY']._serialized_end=1409 + _globals['_INT32ARRAY']._serialized_start=1411 + _globals['_INT32ARRAY']._serialized_end=1439 + _globals['_INT64ARRAY']._serialized_start=1441 + _globals['_INT64ARRAY']._serialized_end=1469 + _globals['_UINT32ARRAY']._serialized_start=1471 + _globals['_UINT32ARRAY']._serialized_end=1500 + _globals['_UINT64ARRAY']._serialized_start=1502 + _globals['_UINT64ARRAY']._serialized_end=1531 + _globals['_FLOATARRAY']._serialized_start=1533 + _globals['_FLOATARRAY']._serialized_end=1561 + _globals['_DOUBLEARRAY']._serialized_start=1563 + _globals['_DOUBLEARRAY']._serialized_end=1592 +# @@protoc_insertion_point(module_scope) diff --git a/integration_test/gen_proto/kuksa/val/v2/types_pb2.pyi b/integration_test/gen_proto/kuksa/val/v2/types_pb2.pyi new file mode 100644 index 00000000..8488d14d --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v2/types_pb2.pyi @@ -0,0 +1,588 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +""" +@generated by mypy-protobuf. Do not edit manually! +isort:skip_file +******************************************************************************* +Copyright (c) 2024 Contributors to the Eclipse Foundation + +See the NOTICE file(s) distributed with this work for additional +information regarding copyright ownership. + +This program and the accompanying materials are made available under the +terms of the Apache License 2.0 which is available at +http://www.apache.org/licenses/LICENSE-2.0 + +SPDX-License-Identifier: Apache-2.0 +****************************************************************************** + +Please do not add optional fields due to older proto3 versions limitations +""" + +import builtins +import collections.abc +import google.protobuf.descriptor +import google.protobuf.internal.containers +import google.protobuf.internal.enum_type_wrapper +import google.protobuf.message +import google.protobuf.timestamp_pb2 +import sys +import typing + +if sys.version_info >= (3, 10): + import typing as typing_extensions +else: + import typing_extensions + +DESCRIPTOR: google.protobuf.descriptor.FileDescriptor + +class _FilterError: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _FilterErrorEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_FilterError.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + FILTER_ERROR_CODE_UNSPECIFIED: _FilterError.ValueType # 0 + FILTER_ERROR_CODE_UNKNOWN_SINGAL_ID: _FilterError.ValueType # 1 + +class FilterError(_FilterError, metaclass=_FilterErrorEnumTypeWrapper): ... + +FILTER_ERROR_CODE_UNSPECIFIED: FilterError.ValueType # 0 +FILTER_ERROR_CODE_UNKNOWN_SINGAL_ID: FilterError.ValueType # 1 +global___FilterError = FilterError + +class _ProviderError: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _ProviderErrorEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_ProviderError.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + CODE_UNSPECIFIED: _ProviderError.ValueType # 0 + CODE_NETWORK_ERROR: _ProviderError.ValueType # 1 + CODE_OVERLOAD: _ProviderError.ValueType # 2 + +class ProviderError(_ProviderError, metaclass=_ProviderErrorEnumTypeWrapper): + """Could be extended in the future with more errors""" + +CODE_UNSPECIFIED: ProviderError.ValueType # 0 +CODE_NETWORK_ERROR: ProviderError.ValueType # 1 +CODE_OVERLOAD: ProviderError.ValueType # 2 +global___ProviderError = ProviderError + +class _ErrorCode: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _ErrorCodeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_ErrorCode.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + ERROR_CODE_UNSPECIFIED: _ErrorCode.ValueType # 0 + """Default value, never to be explicitly set,""" + ERROR_CODE_OK: _ErrorCode.ValueType # 1 + ERROR_CODE_INVALID_ARGUMENT: _ErrorCode.ValueType # 2 + ERROR_CODE_NOT_FOUND: _ErrorCode.ValueType # 3 + ERROR_CODE_PERMISSION_DENIED: _ErrorCode.ValueType # 4 + +class ErrorCode(_ErrorCode, metaclass=_ErrorCodeEnumTypeWrapper): ... + +ERROR_CODE_UNSPECIFIED: ErrorCode.ValueType # 0 +"""Default value, never to be explicitly set,""" +ERROR_CODE_OK: ErrorCode.ValueType # 1 +ERROR_CODE_INVALID_ARGUMENT: ErrorCode.ValueType # 2 +ERROR_CODE_NOT_FOUND: ErrorCode.ValueType # 3 +ERROR_CODE_PERMISSION_DENIED: ErrorCode.ValueType # 4 +global___ErrorCode = ErrorCode + +class _DataType: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _DataTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_DataType.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + DATA_TYPE_UNSPECIFIED: _DataType.ValueType # 0 + DATA_TYPE_STRING: _DataType.ValueType # 1 + DATA_TYPE_BOOLEAN: _DataType.ValueType # 2 + DATA_TYPE_INT8: _DataType.ValueType # 3 + DATA_TYPE_INT16: _DataType.ValueType # 4 + DATA_TYPE_INT32: _DataType.ValueType # 5 + DATA_TYPE_INT64: _DataType.ValueType # 6 + DATA_TYPE_UINT8: _DataType.ValueType # 7 + DATA_TYPE_UINT16: _DataType.ValueType # 8 + DATA_TYPE_UINT32: _DataType.ValueType # 9 + DATA_TYPE_UINT64: _DataType.ValueType # 10 + DATA_TYPE_FLOAT: _DataType.ValueType # 11 + DATA_TYPE_DOUBLE: _DataType.ValueType # 12 + DATA_TYPE_TIMESTAMP: _DataType.ValueType # 13 + DATA_TYPE_STRING_ARRAY: _DataType.ValueType # 20 + DATA_TYPE_BOOLEAN_ARRAY: _DataType.ValueType # 21 + DATA_TYPE_INT8_ARRAY: _DataType.ValueType # 22 + DATA_TYPE_INT16_ARRAY: _DataType.ValueType # 23 + DATA_TYPE_INT32_ARRAY: _DataType.ValueType # 24 + DATA_TYPE_INT64_ARRAY: _DataType.ValueType # 25 + DATA_TYPE_UINT8_ARRAY: _DataType.ValueType # 26 + DATA_TYPE_UINT16_ARRAY: _DataType.ValueType # 27 + DATA_TYPE_UINT32_ARRAY: _DataType.ValueType # 28 + DATA_TYPE_UINT64_ARRAY: _DataType.ValueType # 29 + DATA_TYPE_FLOAT_ARRAY: _DataType.ValueType # 30 + DATA_TYPE_DOUBLE_ARRAY: _DataType.ValueType # 31 + DATA_TYPE_TIMESTAMP_ARRAY: _DataType.ValueType # 32 + +class DataType(_DataType, metaclass=_DataTypeEnumTypeWrapper): + """VSS Data type of a signal + + Protobuf doesn't support int8, int16, uint8 or uint16. + These are mapped to int32 and uint32 respectively. + """ + +DATA_TYPE_UNSPECIFIED: DataType.ValueType # 0 +DATA_TYPE_STRING: DataType.ValueType # 1 +DATA_TYPE_BOOLEAN: DataType.ValueType # 2 +DATA_TYPE_INT8: DataType.ValueType # 3 +DATA_TYPE_INT16: DataType.ValueType # 4 +DATA_TYPE_INT32: DataType.ValueType # 5 +DATA_TYPE_INT64: DataType.ValueType # 6 +DATA_TYPE_UINT8: DataType.ValueType # 7 +DATA_TYPE_UINT16: DataType.ValueType # 8 +DATA_TYPE_UINT32: DataType.ValueType # 9 +DATA_TYPE_UINT64: DataType.ValueType # 10 +DATA_TYPE_FLOAT: DataType.ValueType # 11 +DATA_TYPE_DOUBLE: DataType.ValueType # 12 +DATA_TYPE_TIMESTAMP: DataType.ValueType # 13 +DATA_TYPE_STRING_ARRAY: DataType.ValueType # 20 +DATA_TYPE_BOOLEAN_ARRAY: DataType.ValueType # 21 +DATA_TYPE_INT8_ARRAY: DataType.ValueType # 22 +DATA_TYPE_INT16_ARRAY: DataType.ValueType # 23 +DATA_TYPE_INT32_ARRAY: DataType.ValueType # 24 +DATA_TYPE_INT64_ARRAY: DataType.ValueType # 25 +DATA_TYPE_UINT8_ARRAY: DataType.ValueType # 26 +DATA_TYPE_UINT16_ARRAY: DataType.ValueType # 27 +DATA_TYPE_UINT32_ARRAY: DataType.ValueType # 28 +DATA_TYPE_UINT64_ARRAY: DataType.ValueType # 29 +DATA_TYPE_FLOAT_ARRAY: DataType.ValueType # 30 +DATA_TYPE_DOUBLE_ARRAY: DataType.ValueType # 31 +DATA_TYPE_TIMESTAMP_ARRAY: DataType.ValueType # 32 +global___DataType = DataType + +class _EntryType: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _EntryTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_EntryType.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + ENTRY_TYPE_UNSPECIFIED: _EntryType.ValueType # 0 + ENTRY_TYPE_ATTRIBUTE: _EntryType.ValueType # 1 + ENTRY_TYPE_SENSOR: _EntryType.ValueType # 2 + ENTRY_TYPE_ACTUATOR: _EntryType.ValueType # 3 + +class EntryType(_EntryType, metaclass=_EntryTypeEnumTypeWrapper): + """Entry type""" + +ENTRY_TYPE_UNSPECIFIED: EntryType.ValueType # 0 +ENTRY_TYPE_ATTRIBUTE: EntryType.ValueType # 1 +ENTRY_TYPE_SENSOR: EntryType.ValueType # 2 +ENTRY_TYPE_ACTUATOR: EntryType.ValueType # 3 +global___EntryType = EntryType + +@typing.final +class Datapoint(google.protobuf.message.Message): + """A Datapoint represents a timestamped value. + The 'value' field can be explicitly 'None', meaning the Datapoint exists but no value is present. + """ + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + TIMESTAMP_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + @property + def timestamp(self) -> google.protobuf.timestamp_pb2.Timestamp: + """The timestamp of the datapoint.""" + + @property + def value(self) -> global___Value: + """The value associated with the timestamp. If no value is present, this field can be 'None'.""" + + def __init__( + self, + *, + timestamp: google.protobuf.timestamp_pb2.Timestamp | None = ..., + value: global___Value | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["timestamp", b"timestamp", "value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["timestamp", b"timestamp", "value", b"value"]) -> None: ... + +global___Datapoint = Datapoint + +@typing.final +class Value(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + STRING_FIELD_NUMBER: builtins.int + BOOL_FIELD_NUMBER: builtins.int + INT32_FIELD_NUMBER: builtins.int + INT64_FIELD_NUMBER: builtins.int + UINT32_FIELD_NUMBER: builtins.int + UINT64_FIELD_NUMBER: builtins.int + FLOAT_FIELD_NUMBER: builtins.int + DOUBLE_FIELD_NUMBER: builtins.int + STRING_ARRAY_FIELD_NUMBER: builtins.int + BOOL_ARRAY_FIELD_NUMBER: builtins.int + INT32_ARRAY_FIELD_NUMBER: builtins.int + INT64_ARRAY_FIELD_NUMBER: builtins.int + UINT32_ARRAY_FIELD_NUMBER: builtins.int + UINT64_ARRAY_FIELD_NUMBER: builtins.int + FLOAT_ARRAY_FIELD_NUMBER: builtins.int + DOUBLE_ARRAY_FIELD_NUMBER: builtins.int + string: builtins.str + bool: builtins.bool + int32: builtins.int + int64: builtins.int + uint32: builtins.int + uint64: builtins.int + float: builtins.float + double: builtins.float + @property + def string_array(self) -> global___StringArray: ... + @property + def bool_array(self) -> global___BoolArray: ... + @property + def int32_array(self) -> global___Int32Array: ... + @property + def int64_array(self) -> global___Int64Array: ... + @property + def uint32_array(self) -> global___Uint32Array: ... + @property + def uint64_array(self) -> global___Uint64Array: ... + @property + def float_array(self) -> global___FloatArray: ... + @property + def double_array(self) -> global___DoubleArray: ... + def __init__( + self, + *, + string: builtins.str = ..., + bool: builtins.bool = ..., + int32: builtins.int = ..., + int64: builtins.int = ..., + uint32: builtins.int = ..., + uint64: builtins.int = ..., + float: builtins.float = ..., + double: builtins.float = ..., + string_array: global___StringArray | None = ..., + bool_array: global___BoolArray | None = ..., + int32_array: global___Int32Array | None = ..., + int64_array: global___Int64Array | None = ..., + uint32_array: global___Uint32Array | None = ..., + uint64_array: global___Uint64Array | None = ..., + float_array: global___FloatArray | None = ..., + double_array: global___DoubleArray | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["bool", b"bool", "bool_array", b"bool_array", "double", b"double", "double_array", b"double_array", "float", b"float", "float_array", b"float_array", "int32", b"int32", "int32_array", b"int32_array", "int64", b"int64", "int64_array", b"int64_array", "string", b"string", "string_array", b"string_array", "typed_value", b"typed_value", "uint32", b"uint32", "uint32_array", b"uint32_array", "uint64", b"uint64", "uint64_array", b"uint64_array"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["bool", b"bool", "bool_array", b"bool_array", "double", b"double", "double_array", b"double_array", "float", b"float", "float_array", b"float_array", "int32", b"int32", "int32_array", b"int32_array", "int64", b"int64", "int64_array", b"int64_array", "string", b"string", "string_array", b"string_array", "typed_value", b"typed_value", "uint32", b"uint32", "uint32_array", b"uint32_array", "uint64", b"uint64", "uint64_array", b"uint64_array"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["typed_value", b"typed_value"]) -> typing.Literal["string", "bool", "int32", "int64", "uint32", "uint64", "float", "double", "string_array", "bool_array", "int32_array", "int64_array", "uint32_array", "uint64_array", "float_array", "double_array"] | None: ... + +global___Value = Value + +@typing.final +class SampleInterval(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + INTERVAL_MS_FIELD_NUMBER: builtins.int + interval_ms: builtins.int + def __init__( + self, + *, + interval_ms: builtins.int = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["interval_ms", b"interval_ms"]) -> None: ... + +global___SampleInterval = SampleInterval + +@typing.final +class Filter(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + DURATION_MS_FIELD_NUMBER: builtins.int + MIN_SAMPLE_INTERVAL_FIELD_NUMBER: builtins.int + duration_ms: builtins.int + """Duration of the active call. If it is not set, call will last for ever.""" + @property + def min_sample_interval(self) -> global___SampleInterval: + """Min desired sample update interval.""" + + def __init__( + self, + *, + duration_ms: builtins.int = ..., + min_sample_interval: global___SampleInterval | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["min_sample_interval", b"min_sample_interval"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["duration_ms", b"duration_ms", "min_sample_interval", b"min_sample_interval"]) -> None: ... + +global___Filter = Filter + +@typing.final +class SignalID(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ID_FIELD_NUMBER: builtins.int + PATH_FIELD_NUMBER: builtins.int + id: builtins.int + """Numeric identifier to the signal + As of today Databroker assigns arbitrary unique numbers to each registered signal + at startup, meaning that identifiers may change after restarting Databroker. + A mechanism for static identifiers may be introduced in the future. + """ + path: builtins.str + """Full VSS-style path to a specific signal, like "Vehicle.Speed" + Wildcards and paths to branches are not supported. + The given path must be known by the Databroker. + """ + def __init__( + self, + *, + id: builtins.int = ..., + path: builtins.str = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["id", b"id", "path", b"path", "signal", b"signal"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["id", b"id", "path", b"path", "signal", b"signal"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["signal", b"signal"]) -> typing.Literal["id", "path"] | None: ... + +global___SignalID = SignalID + +@typing.final +class Error(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + CODE_FIELD_NUMBER: builtins.int + MESSAGE_FIELD_NUMBER: builtins.int + code: global___ErrorCode.ValueType + message: builtins.str + def __init__( + self, + *, + code: global___ErrorCode.ValueType = ..., + message: builtins.str = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["code", b"code", "message", b"message"]) -> None: ... + +global___Error = Error + +@typing.final +class Metadata(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + PATH_FIELD_NUMBER: builtins.int + ID_FIELD_NUMBER: builtins.int + DATA_TYPE_FIELD_NUMBER: builtins.int + ENTRY_TYPE_FIELD_NUMBER: builtins.int + DESCRIPTION_FIELD_NUMBER: builtins.int + COMMENT_FIELD_NUMBER: builtins.int + DEPRECATION_FIELD_NUMBER: builtins.int + UNIT_FIELD_NUMBER: builtins.int + ALLOWED_VALUES_FIELD_NUMBER: builtins.int + MIN_FIELD_NUMBER: builtins.int + MAX_FIELD_NUMBER: builtins.int + MIN_SAMPLE_INTERVAL_FIELD_NUMBER: builtins.int + path: builtins.str + """Full dot notated path for the signal""" + id: builtins.int + """ID field""" + data_type: global___DataType.ValueType + """Data type + The VSS data type of the entry (i.e. the value, min, max etc). + + NOTE: protobuf doesn't have int8, int16, uint8 or uint16 which means + that these values must be serialized as int32 and uint32 respectively. + """ + entry_type: global___EntryType.ValueType + """Entry type""" + description: builtins.str + """Description + Describes the meaning and content of the entry. + """ + comment: builtins.str + """Comment + A comment can be used to provide additional informal information + on a entry. + """ + deprecation: builtins.str + """Deprecation + Whether this entry is deprecated. Can contain recommendations of what + to use instead. + """ + unit: builtins.str + """Unit + The unit of measurement + """ + @property + def allowed_values(self) -> global___Value: + """Value restrictions checked/enforced by Databroker + Must be of array type + """ + + @property + def min(self) -> global___Value: ... + @property + def max(self) -> global___Value: ... + @property + def min_sample_interval(self) -> global___SampleInterval: + """Minimum sample interval at which its provider can publish the signal value""" + + def __init__( + self, + *, + path: builtins.str = ..., + id: builtins.int = ..., + data_type: global___DataType.ValueType = ..., + entry_type: global___EntryType.ValueType = ..., + description: builtins.str = ..., + comment: builtins.str = ..., + deprecation: builtins.str = ..., + unit: builtins.str = ..., + allowed_values: global___Value | None = ..., + min: global___Value | None = ..., + max: global___Value | None = ..., + min_sample_interval: global___SampleInterval | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["allowed_values", b"allowed_values", "max", b"max", "min", b"min", "min_sample_interval", b"min_sample_interval"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["allowed_values", b"allowed_values", "comment", b"comment", "data_type", b"data_type", "deprecation", b"deprecation", "description", b"description", "entry_type", b"entry_type", "id", b"id", "max", b"max", "min", b"min", "min_sample_interval", b"min_sample_interval", "path", b"path", "unit", b"unit"]) -> None: ... + +global___Metadata = Metadata + +@typing.final +class StringArray(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.str] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___StringArray = StringArray + +@typing.final +class BoolArray(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.bool] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___BoolArray = BoolArray + +@typing.final +class Int32Array(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___Int32Array = Int32Array + +@typing.final +class Int64Array(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___Int64Array = Int64Array + +@typing.final +class Uint32Array(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___Uint32Array = Uint32Array + +@typing.final +class Uint64Array(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___Uint64Array = Uint64Array + +@typing.final +class FloatArray(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.float]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.float] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___FloatArray = FloatArray + +@typing.final +class DoubleArray(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + VALUES_FIELD_NUMBER: builtins.int + @property + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.float]: ... + def __init__( + self, + *, + values: collections.abc.Iterable[builtins.float] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + +global___DoubleArray = DoubleArray diff --git a/integration_test/gen_proto/kuksa/val/v2/types_pb2_grpc.py b/integration_test/gen_proto/kuksa/val/v2/types_pb2_grpc.py new file mode 100644 index 00000000..cfc4a461 --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v2/types_pb2_grpc.py @@ -0,0 +1,37 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +# Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT! +"""Client and server classes corresponding to protobuf-defined services.""" +import grpc +import warnings + + +GRPC_GENERATED_VERSION = '1.71.0' +GRPC_VERSION = grpc.__version__ +_version_not_supported = False + +try: + from grpc._utilities import first_version_is_lower + _version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION) +except ImportError: + _version_not_supported = True + +if _version_not_supported: + raise RuntimeError( + f'The grpc package installed is at version {GRPC_VERSION},' + + f' but the generated code in kuksa/val/v2/types_pb2_grpc.py depends on' + + f' grpcio>={GRPC_GENERATED_VERSION}.' + + f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}' + + f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.' + ) diff --git a/integration_test/gen_proto/kuksa/val/v2/val_pb2.py b/integration_test/gen_proto/kuksa/val/v2/val_pb2.py new file mode 100644 index 00000000..7ed3839e --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v2/val_pb2.py @@ -0,0 +1,145 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +# -*- coding: utf-8 -*- +# Generated by the protocol buffer compiler. DO NOT EDIT! +# NO CHECKED-IN PROTOBUF GENCODE +# source: kuksa/val/v2/val.proto +# Protobuf Python Version: 5.29.0 +"""Generated protocol buffer code.""" +from google.protobuf import descriptor as _descriptor +from google.protobuf import descriptor_pool as _descriptor_pool +from google.protobuf import runtime_version as _runtime_version +from google.protobuf import symbol_database as _symbol_database +from google.protobuf.internal import builder as _builder +_runtime_version.ValidateProtobufRuntimeVersion( + _runtime_version.Domain.PUBLIC, + 5, + 29, + 0, + '', + 'kuksa/val/v2/val.proto' +) +# @@protoc_insertion_point(imports) + +_sym_db = _symbol_database.Default() + + +from gen_proto.kuksa.val.v2 import types_pb2 as kuksa_dot_val_dot_v2_dot_types__pb2 + + +DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\x16kuksa/val/v2/val.proto\x12\x0ckuksa.val.v2\x1a\x18kuksa/val/v2/types.proto\"<\n\x0fGetValueRequest\x12)\n\tsignal_id\x18\x01 \x01(\x0b\x32\x16.kuksa.val.v2.SignalID\"?\n\x10GetValueResponse\x12+\n\ndata_point\x18\x01 \x01(\x0b\x32\x17.kuksa.val.v2.Datapoint\">\n\x10GetValuesRequest\x12*\n\nsignal_ids\x18\x01 \x03(\x0b\x32\x16.kuksa.val.v2.SignalID\"A\n\x11GetValuesResponse\x12,\n\x0b\x64\x61ta_points\x18\x01 \x03(\x0b\x32\x17.kuksa.val.v2.Datapoint\"c\n\x10SubscribeRequest\x12\x14\n\x0csignal_paths\x18\x01 \x03(\t\x12\x13\n\x0b\x62uffer_size\x18\x02 \x01(\r\x12$\n\x06\x66ilter\x18\x03 \x01(\x0b\x32\x14.kuksa.val.v2.Filter\"\x9b\x01\n\x11SubscribeResponse\x12=\n\x07\x65ntries\x18\x01 \x03(\x0b\x32,.kuksa.val.v2.SubscribeResponse.EntriesEntry\x1aG\n\x0c\x45ntriesEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12&\n\x05value\x18\x02 \x01(\x0b\x32\x17.kuksa.val.v2.Datapoint:\x02\x38\x01\"e\n\x14SubscribeByIdRequest\x12\x12\n\nsignal_ids\x18\x01 \x03(\x05\x12\x13\n\x0b\x62uffer_size\x18\x02 \x01(\r\x12$\n\x06\x66ilter\x18\x03 \x01(\x0b\x32\x14.kuksa.val.v2.Filter\"\xa3\x01\n\x15SubscribeByIdResponse\x12\x41\n\x07\x65ntries\x18\x01 \x03(\x0b\x32\x30.kuksa.val.v2.SubscribeByIdResponse.EntriesEntry\x1aG\n\x0c\x45ntriesEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12&\n\x05value\x18\x02 \x01(\x0b\x32\x17.kuksa.val.v2.Datapoint:\x02\x38\x01\"_\n\x0e\x41\x63tuateRequest\x12)\n\tsignal_id\x18\x01 \x01(\x0b\x32\x16.kuksa.val.v2.SignalID\x12\"\n\x05value\x18\x02 \x01(\x0b\x32\x13.kuksa.val.v2.Value\"\x11\n\x0f\x41\x63tuateResponse\"M\n\x13\x42\x61tchActuateRequest\x12\x36\n\x10\x61\x63tuate_requests\x18\x01 \x03(\x0b\x32\x1c.kuksa.val.v2.ActuateRequest\"\x16\n\x14\x42\x61tchActuateResponse\"3\n\x13ListMetadataRequest\x12\x0c\n\x04root\x18\x01 \x01(\t\x12\x0e\n\x06\x66ilter\x18\x02 \x01(\t\"@\n\x14ListMetadataResponse\x12(\n\x08metadata\x18\x01 \x03(\x0b\x32\x16.kuksa.val.v2.Metadata\"m\n\x13PublishValueRequest\x12)\n\tsignal_id\x18\x01 \x01(\x0b\x32\x16.kuksa.val.v2.SignalID\x12+\n\ndata_point\x18\x02 \x01(\x0b\x32\x17.kuksa.val.v2.Datapoint\"\x16\n\x14PublishValueResponse\"\xbf\x01\n\x14PublishValuesRequest\x12\x12\n\nrequest_id\x18\x01 \x01(\r\x12G\n\x0b\x64\x61ta_points\x18\x02 \x03(\x0b\x32\x32.kuksa.val.v2.PublishValuesRequest.DataPointsEntry\x1aJ\n\x0f\x44\x61taPointsEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12&\n\x05value\x18\x02 \x01(\x0b\x32\x17.kuksa.val.v2.Datapoint:\x02\x38\x01\"\xb0\x01\n\x15PublishValuesResponse\x12\x12\n\nrequest_id\x18\x01 \x01(\r\x12?\n\x06status\x18\x02 \x03(\x0b\x32/.kuksa.val.v2.PublishValuesResponse.StatusEntry\x1a\x42\n\x0bStatusEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12\"\n\x05value\x18\x02 \x01(\x0b\x32\x13.kuksa.val.v2.Error:\x02\x38\x01\"O\n\x17ProvideActuationRequest\x12\x34\n\x14\x61\x63tuator_identifiers\x18\x01 \x03(\x0b\x32\x16.kuksa.val.v2.SignalID\"\x1a\n\x18ProvideActuationResponse\"\xd5\x01\n\x14ProvideSignalRequest\x12`\n\x18signals_sample_intervals\x18\x01 \x03(\x0b\x32>.kuksa.val.v2.ProvideSignalRequest.SignalsSampleIntervalsEntry\x1a[\n\x1bSignalsSampleIntervalsEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12+\n\x05value\x18\x02 \x01(\x0b\x32\x1c.kuksa.val.v2.SampleInterval:\x02\x38\x01\"\x17\n\x15ProvideSignalResponse\"S\n\x19\x42\x61tchActuateStreamRequest\x12\x36\n\x10\x61\x63tuate_requests\x18\x01 \x03(\x0b\x32\x1c.kuksa.val.v2.ActuateRequest\"k\n\x1a\x42\x61tchActuateStreamResponse\x12)\n\tsignal_id\x18\x01 \x01(\x0b\x32\x16.kuksa.val.v2.SignalID\x12\"\n\x05\x65rror\x18\x02 \x01(\x0b\x32\x13.kuksa.val.v2.Error\"\xc3\x01\n\x13UpdateFilterRequest\x12\x12\n\nrequest_id\x18\x01 \x01(\r\x12L\n\x0e\x66ilters_update\x18\x02 \x03(\x0b\x32\x34.kuksa.val.v2.UpdateFilterRequest.FiltersUpdateEntry\x1aJ\n\x12\x46iltersUpdateEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12#\n\x05value\x18\x02 \x01(\x0b\x32\x14.kuksa.val.v2.Filter:\x02\x38\x01\"[\n\x14UpdateFilterResponse\x12\x12\n\nrequest_id\x18\x01 \x01(\r\x12/\n\x0c\x66ilter_error\x18\x02 \x01(\x0e\x32\x19.kuksa.val.v2.FilterError\"N\n\x17ProviderErrorIndication\x12\x33\n\x0eprovider_error\x18\x01 \x01(\x0e\x32\x1b.kuksa.val.v2.ProviderError\"A\n\x17GetProviderValueRequest\x12\x12\n\nrequest_id\x18\x01 \x01(\r\x12\x12\n\nsignal_ids\x18\x02 \x03(\x05\"\xbd\x01\n\x18GetProviderValueResponse\x12\x12\n\nrequest_id\x18\x01 \x01(\r\x12\x44\n\x07\x65ntries\x18\x02 \x03(\x0b\x32\x33.kuksa.val.v2.GetProviderValueResponse.EntriesEntry\x1aG\n\x0c\x45ntriesEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12&\n\x05value\x18\x02 \x01(\x0b\x32\x17.kuksa.val.v2.Datapoint:\x02\x38\x01\"\xb1\x04\n\x19OpenProviderStreamRequest\x12J\n\x19provide_actuation_request\x18\x01 \x01(\x0b\x32%.kuksa.val.v2.ProvideActuationRequestH\x00\x12\x44\n\x16publish_values_request\x18\x02 \x01(\x0b\x32\".kuksa.val.v2.PublishValuesRequestH\x00\x12Q\n\x1d\x62\x61tch_actuate_stream_response\x18\x03 \x01(\x0b\x32(.kuksa.val.v2.BatchActuateStreamResponseH\x00\x12\x44\n\x16provide_signal_request\x18\x04 \x01(\x0b\x32\".kuksa.val.v2.ProvideSignalRequestH\x00\x12\x44\n\x16update_filter_response\x18\x05 \x01(\x0b\x32\".kuksa.val.v2.UpdateFilterResponseH\x00\x12M\n\x1bget_provider_value_response\x18\x06 \x01(\x0b\x32&.kuksa.val.v2.GetProviderValueResponseH\x00\x12J\n\x19provider_error_indication\x18\x07 \x01(\x0b\x32%.kuksa.val.v2.ProviderErrorIndicationH\x00\x42\x08\n\x06\x61\x63tion\"\xe6\x03\n\x1aOpenProviderStreamResponse\x12L\n\x1aprovide_actuation_response\x18\x01 \x01(\x0b\x32&.kuksa.val.v2.ProvideActuationResponseH\x00\x12\x46\n\x17publish_values_response\x18\x02 \x01(\x0b\x32#.kuksa.val.v2.PublishValuesResponseH\x00\x12O\n\x1c\x62\x61tch_actuate_stream_request\x18\x03 \x01(\x0b\x32\'.kuksa.val.v2.BatchActuateStreamRequestH\x00\x12\x46\n\x17provide_signal_response\x18\x04 \x01(\x0b\x32#.kuksa.val.v2.ProvideSignalResponseH\x00\x12\x42\n\x15update_filter_request\x18\x05 \x01(\x0b\x32!.kuksa.val.v2.UpdateFilterRequestH\x00\x12K\n\x1aget_provider_value_request\x18\x06 \x01(\x0b\x32%.kuksa.val.v2.GetProviderValueRequestH\x00\x42\x08\n\x06\x61\x63tion\"\x16\n\x14GetServerInfoRequest\"K\n\x15GetServerInfoResponse\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x0f\n\x07version\x18\x02 \x01(\t\x12\x13\n\x0b\x63ommit_hash\x18\x03 \x01(\t2\xae\x07\n\x03VAL\x12I\n\x08GetValue\x12\x1d.kuksa.val.v2.GetValueRequest\x1a\x1e.kuksa.val.v2.GetValueResponse\x12L\n\tGetValues\x12\x1e.kuksa.val.v2.GetValuesRequest\x1a\x1f.kuksa.val.v2.GetValuesResponse\x12N\n\tSubscribe\x12\x1e.kuksa.val.v2.SubscribeRequest\x1a\x1f.kuksa.val.v2.SubscribeResponse0\x01\x12Z\n\rSubscribeById\x12\".kuksa.val.v2.SubscribeByIdRequest\x1a#.kuksa.val.v2.SubscribeByIdResponse0\x01\x12\x46\n\x07\x41\x63tuate\x12\x1c.kuksa.val.v2.ActuateRequest\x1a\x1d.kuksa.val.v2.ActuateResponse\x12N\n\rActuateStream\x12\x1c.kuksa.val.v2.ActuateRequest\x1a\x1d.kuksa.val.v2.ActuateResponse(\x01\x12U\n\x0c\x42\x61tchActuate\x12!.kuksa.val.v2.BatchActuateRequest\x1a\".kuksa.val.v2.BatchActuateResponse\x12U\n\x0cListMetadata\x12!.kuksa.val.v2.ListMetadataRequest\x1a\".kuksa.val.v2.ListMetadataResponse\x12U\n\x0cPublishValue\x12!.kuksa.val.v2.PublishValueRequest\x1a\".kuksa.val.v2.PublishValueResponse\x12k\n\x12OpenProviderStream\x12\'.kuksa.val.v2.OpenProviderStreamRequest\x1a(.kuksa.val.v2.OpenProviderStreamResponse(\x01\x30\x01\x12X\n\rGetServerInfo\x12\".kuksa.val.v2.GetServerInfoRequest\x1a#.kuksa.val.v2.GetServerInfoResponseB\x0eZ\x0ckuksa/val/v2b\x06proto3') + +_globals = globals() +_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) +_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'kuksa.val.v2.val_pb2', _globals) +if not _descriptor._USE_C_DESCRIPTORS: + _globals['DESCRIPTOR']._loaded_options = None + _globals['DESCRIPTOR']._serialized_options = b'Z\014kuksa/val/v2' + _globals['_SUBSCRIBERESPONSE_ENTRIESENTRY']._loaded_options = None + _globals['_SUBSCRIBERESPONSE_ENTRIESENTRY']._serialized_options = b'8\001' + _globals['_SUBSCRIBEBYIDRESPONSE_ENTRIESENTRY']._loaded_options = None + _globals['_SUBSCRIBEBYIDRESPONSE_ENTRIESENTRY']._serialized_options = b'8\001' + _globals['_PUBLISHVALUESREQUEST_DATAPOINTSENTRY']._loaded_options = None + _globals['_PUBLISHVALUESREQUEST_DATAPOINTSENTRY']._serialized_options = b'8\001' + _globals['_PUBLISHVALUESRESPONSE_STATUSENTRY']._loaded_options = None + _globals['_PUBLISHVALUESRESPONSE_STATUSENTRY']._serialized_options = b'8\001' + _globals['_PROVIDESIGNALREQUEST_SIGNALSSAMPLEINTERVALSENTRY']._loaded_options = None + _globals['_PROVIDESIGNALREQUEST_SIGNALSSAMPLEINTERVALSENTRY']._serialized_options = b'8\001' + _globals['_UPDATEFILTERREQUEST_FILTERSUPDATEENTRY']._loaded_options = None + _globals['_UPDATEFILTERREQUEST_FILTERSUPDATEENTRY']._serialized_options = b'8\001' + _globals['_GETPROVIDERVALUERESPONSE_ENTRIESENTRY']._loaded_options = None + _globals['_GETPROVIDERVALUERESPONSE_ENTRIESENTRY']._serialized_options = b'8\001' + _globals['_GETVALUEREQUEST']._serialized_start=66 + _globals['_GETVALUEREQUEST']._serialized_end=126 + _globals['_GETVALUERESPONSE']._serialized_start=128 + _globals['_GETVALUERESPONSE']._serialized_end=191 + _globals['_GETVALUESREQUEST']._serialized_start=193 + _globals['_GETVALUESREQUEST']._serialized_end=255 + _globals['_GETVALUESRESPONSE']._serialized_start=257 + _globals['_GETVALUESRESPONSE']._serialized_end=322 + _globals['_SUBSCRIBEREQUEST']._serialized_start=324 + _globals['_SUBSCRIBEREQUEST']._serialized_end=423 + _globals['_SUBSCRIBERESPONSE']._serialized_start=426 + _globals['_SUBSCRIBERESPONSE']._serialized_end=581 + _globals['_SUBSCRIBERESPONSE_ENTRIESENTRY']._serialized_start=510 + _globals['_SUBSCRIBERESPONSE_ENTRIESENTRY']._serialized_end=581 + _globals['_SUBSCRIBEBYIDREQUEST']._serialized_start=583 + _globals['_SUBSCRIBEBYIDREQUEST']._serialized_end=684 + _globals['_SUBSCRIBEBYIDRESPONSE']._serialized_start=687 + _globals['_SUBSCRIBEBYIDRESPONSE']._serialized_end=850 + _globals['_SUBSCRIBEBYIDRESPONSE_ENTRIESENTRY']._serialized_start=779 + _globals['_SUBSCRIBEBYIDRESPONSE_ENTRIESENTRY']._serialized_end=850 + _globals['_ACTUATEREQUEST']._serialized_start=852 + _globals['_ACTUATEREQUEST']._serialized_end=947 + _globals['_ACTUATERESPONSE']._serialized_start=949 + _globals['_ACTUATERESPONSE']._serialized_end=966 + _globals['_BATCHACTUATEREQUEST']._serialized_start=968 + _globals['_BATCHACTUATEREQUEST']._serialized_end=1045 + _globals['_BATCHACTUATERESPONSE']._serialized_start=1047 + _globals['_BATCHACTUATERESPONSE']._serialized_end=1069 + _globals['_LISTMETADATAREQUEST']._serialized_start=1071 + _globals['_LISTMETADATAREQUEST']._serialized_end=1122 + _globals['_LISTMETADATARESPONSE']._serialized_start=1124 + _globals['_LISTMETADATARESPONSE']._serialized_end=1188 + _globals['_PUBLISHVALUEREQUEST']._serialized_start=1190 + _globals['_PUBLISHVALUEREQUEST']._serialized_end=1299 + _globals['_PUBLISHVALUERESPONSE']._serialized_start=1301 + _globals['_PUBLISHVALUERESPONSE']._serialized_end=1323 + _globals['_PUBLISHVALUESREQUEST']._serialized_start=1326 + _globals['_PUBLISHVALUESREQUEST']._serialized_end=1517 + _globals['_PUBLISHVALUESREQUEST_DATAPOINTSENTRY']._serialized_start=1443 + _globals['_PUBLISHVALUESREQUEST_DATAPOINTSENTRY']._serialized_end=1517 + _globals['_PUBLISHVALUESRESPONSE']._serialized_start=1520 + _globals['_PUBLISHVALUESRESPONSE']._serialized_end=1696 + _globals['_PUBLISHVALUESRESPONSE_STATUSENTRY']._serialized_start=1630 + _globals['_PUBLISHVALUESRESPONSE_STATUSENTRY']._serialized_end=1696 + _globals['_PROVIDEACTUATIONREQUEST']._serialized_start=1698 + _globals['_PROVIDEACTUATIONREQUEST']._serialized_end=1777 + _globals['_PROVIDEACTUATIONRESPONSE']._serialized_start=1779 + _globals['_PROVIDEACTUATIONRESPONSE']._serialized_end=1805 + _globals['_PROVIDESIGNALREQUEST']._serialized_start=1808 + _globals['_PROVIDESIGNALREQUEST']._serialized_end=2021 + _globals['_PROVIDESIGNALREQUEST_SIGNALSSAMPLEINTERVALSENTRY']._serialized_start=1930 + _globals['_PROVIDESIGNALREQUEST_SIGNALSSAMPLEINTERVALSENTRY']._serialized_end=2021 + _globals['_PROVIDESIGNALRESPONSE']._serialized_start=2023 + _globals['_PROVIDESIGNALRESPONSE']._serialized_end=2046 + _globals['_BATCHACTUATESTREAMREQUEST']._serialized_start=2048 + _globals['_BATCHACTUATESTREAMREQUEST']._serialized_end=2131 + _globals['_BATCHACTUATESTREAMRESPONSE']._serialized_start=2133 + _globals['_BATCHACTUATESTREAMRESPONSE']._serialized_end=2240 + _globals['_UPDATEFILTERREQUEST']._serialized_start=2243 + _globals['_UPDATEFILTERREQUEST']._serialized_end=2438 + _globals['_UPDATEFILTERREQUEST_FILTERSUPDATEENTRY']._serialized_start=2364 + _globals['_UPDATEFILTERREQUEST_FILTERSUPDATEENTRY']._serialized_end=2438 + _globals['_UPDATEFILTERRESPONSE']._serialized_start=2440 + _globals['_UPDATEFILTERRESPONSE']._serialized_end=2531 + _globals['_PROVIDERERRORINDICATION']._serialized_start=2533 + _globals['_PROVIDERERRORINDICATION']._serialized_end=2611 + _globals['_GETPROVIDERVALUEREQUEST']._serialized_start=2613 + _globals['_GETPROVIDERVALUEREQUEST']._serialized_end=2678 + _globals['_GETPROVIDERVALUERESPONSE']._serialized_start=2681 + _globals['_GETPROVIDERVALUERESPONSE']._serialized_end=2870 + _globals['_GETPROVIDERVALUERESPONSE_ENTRIESENTRY']._serialized_start=779 + _globals['_GETPROVIDERVALUERESPONSE_ENTRIESENTRY']._serialized_end=850 + _globals['_OPENPROVIDERSTREAMREQUEST']._serialized_start=2873 + _globals['_OPENPROVIDERSTREAMREQUEST']._serialized_end=3434 + _globals['_OPENPROVIDERSTREAMRESPONSE']._serialized_start=3437 + _globals['_OPENPROVIDERSTREAMRESPONSE']._serialized_end=3923 + _globals['_GETSERVERINFOREQUEST']._serialized_start=3925 + _globals['_GETSERVERINFOREQUEST']._serialized_end=3947 + _globals['_GETSERVERINFORESPONSE']._serialized_start=3949 + _globals['_GETSERVERINFORESPONSE']._serialized_end=4024 + _globals['_VAL']._serialized_start=4027 + _globals['_VAL']._serialized_end=4969 +# @@protoc_insertion_point(module_scope) diff --git a/integration_test/gen_proto/kuksa/val/v2/val_pb2.pyi b/integration_test/gen_proto/kuksa/val/v2/val_pb2.pyi new file mode 100644 index 00000000..488fa20f --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v2/val_pb2.pyi @@ -0,0 +1,827 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +""" +@generated by mypy-protobuf. Do not edit manually! +isort:skip_file +******************************************************************************* +Copyright (c) 2024 Contributors to the Eclipse Foundation + +See the NOTICE file(s) distributed with this work for additional +information regarding copyright ownership. + +This program and the accompanying materials are made available under the +terms of the Apache License 2.0 which is available at +http://www.apache.org/licenses/LICENSE-2.0 + +SPDX-License-Identifier: Apache-2.0 +****************************************************************************** + +Please do not add optional fields due to older proto3 versions limitations +""" + +import builtins +import collections.abc +import google.protobuf.descriptor +import google.protobuf.internal.containers +import google.protobuf.message +import kuksa.val.v2.types_pb2 +import typing + +DESCRIPTOR: google.protobuf.descriptor.FileDescriptor + +@typing.final +class GetValueRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + SIGNAL_ID_FIELD_NUMBER: builtins.int + @property + def signal_id(self) -> kuksa.val.v2.types_pb2.SignalID: ... + def __init__( + self, + *, + signal_id: kuksa.val.v2.types_pb2.SignalID | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["signal_id", b"signal_id"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["signal_id", b"signal_id"]) -> None: ... + +global___GetValueRequest = GetValueRequest + +@typing.final +class GetValueResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + DATA_POINT_FIELD_NUMBER: builtins.int + @property + def data_point(self) -> kuksa.val.v2.types_pb2.Datapoint: ... + def __init__( + self, + *, + data_point: kuksa.val.v2.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["data_point", b"data_point"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["data_point", b"data_point"]) -> None: ... + +global___GetValueResponse = GetValueResponse + +@typing.final +class GetValuesRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + SIGNAL_IDS_FIELD_NUMBER: builtins.int + @property + def signal_ids(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[kuksa.val.v2.types_pb2.SignalID]: ... + def __init__( + self, + *, + signal_ids: collections.abc.Iterable[kuksa.val.v2.types_pb2.SignalID] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["signal_ids", b"signal_ids"]) -> None: ... + +global___GetValuesRequest = GetValuesRequest + +@typing.final +class GetValuesResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + DATA_POINTS_FIELD_NUMBER: builtins.int + @property + def data_points(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[kuksa.val.v2.types_pb2.Datapoint]: ... + def __init__( + self, + *, + data_points: collections.abc.Iterable[kuksa.val.v2.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["data_points", b"data_points"]) -> None: ... + +global___GetValuesResponse = GetValuesResponse + +@typing.final +class SubscribeRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + SIGNAL_PATHS_FIELD_NUMBER: builtins.int + BUFFER_SIZE_FIELD_NUMBER: builtins.int + FILTER_FIELD_NUMBER: builtins.int + buffer_size: builtins.int + """Specifies the number of messages that can be buffered for + slow subscribers before the oldest messages are dropped. + Default (0) results in that only latest message is kept. + Maximum value supported is implementation dependent. + """ + @property + def signal_paths(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: ... + @property + def filter(self) -> kuksa.val.v2.types_pb2.Filter: ... + def __init__( + self, + *, + signal_paths: collections.abc.Iterable[builtins.str] | None = ..., + buffer_size: builtins.int = ..., + filter: kuksa.val.v2.types_pb2.Filter | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["filter", b"filter"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["buffer_size", b"buffer_size", "filter", b"filter", "signal_paths", b"signal_paths"]) -> None: ... + +global___SubscribeRequest = SubscribeRequest + +@typing.final +class SubscribeResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class EntriesEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.str + @property + def value(self) -> kuksa.val.v2.types_pb2.Datapoint: ... + def __init__( + self, + *, + key: builtins.str = ..., + value: kuksa.val.v2.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + ENTRIES_FIELD_NUMBER: builtins.int + @property + def entries(self) -> google.protobuf.internal.containers.MessageMap[builtins.str, kuksa.val.v2.types_pb2.Datapoint]: ... + def __init__( + self, + *, + entries: collections.abc.Mapping[builtins.str, kuksa.val.v2.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["entries", b"entries"]) -> None: ... + +global___SubscribeResponse = SubscribeResponse + +@typing.final +class SubscribeByIdRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + SIGNAL_IDS_FIELD_NUMBER: builtins.int + BUFFER_SIZE_FIELD_NUMBER: builtins.int + FILTER_FIELD_NUMBER: builtins.int + buffer_size: builtins.int + """Specifies the number of messages that can be buffered for + slow subscribers before the oldest messages are dropped. + Default (0) results in that only latest message is kept. + Maximum value supported is implementation dependent. + """ + @property + def signal_ids(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + @property + def filter(self) -> kuksa.val.v2.types_pb2.Filter: ... + def __init__( + self, + *, + signal_ids: collections.abc.Iterable[builtins.int] | None = ..., + buffer_size: builtins.int = ..., + filter: kuksa.val.v2.types_pb2.Filter | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["filter", b"filter"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["buffer_size", b"buffer_size", "filter", b"filter", "signal_ids", b"signal_ids"]) -> None: ... + +global___SubscribeByIdRequest = SubscribeByIdRequest + +@typing.final +class SubscribeByIdResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class EntriesEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.int + @property + def value(self) -> kuksa.val.v2.types_pb2.Datapoint: ... + def __init__( + self, + *, + key: builtins.int = ..., + value: kuksa.val.v2.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + ENTRIES_FIELD_NUMBER: builtins.int + @property + def entries(self) -> google.protobuf.internal.containers.MessageMap[builtins.int, kuksa.val.v2.types_pb2.Datapoint]: ... + def __init__( + self, + *, + entries: collections.abc.Mapping[builtins.int, kuksa.val.v2.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["entries", b"entries"]) -> None: ... + +global___SubscribeByIdResponse = SubscribeByIdResponse + +@typing.final +class ActuateRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + SIGNAL_ID_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + @property + def signal_id(self) -> kuksa.val.v2.types_pb2.SignalID: ... + @property + def value(self) -> kuksa.val.v2.types_pb2.Value: ... + def __init__( + self, + *, + signal_id: kuksa.val.v2.types_pb2.SignalID | None = ..., + value: kuksa.val.v2.types_pb2.Value | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["signal_id", b"signal_id", "value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["signal_id", b"signal_id", "value", b"value"]) -> None: ... + +global___ActuateRequest = ActuateRequest + +@typing.final +class ActuateResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___ActuateResponse = ActuateResponse + +@typing.final +class BatchActuateRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ACTUATE_REQUESTS_FIELD_NUMBER: builtins.int + @property + def actuate_requests(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ActuateRequest]: ... + def __init__( + self, + *, + actuate_requests: collections.abc.Iterable[global___ActuateRequest] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["actuate_requests", b"actuate_requests"]) -> None: ... + +global___BatchActuateRequest = BatchActuateRequest + +@typing.final +class BatchActuateResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___BatchActuateResponse = BatchActuateResponse + +@typing.final +class ListMetadataRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ROOT_FIELD_NUMBER: builtins.int + FILTER_FIELD_NUMBER: builtins.int + root: builtins.str + """Root path to be used when listing metadata + Shall correspond to a VSS branch, e.g. "Vehicle", "Vehicle.Cabin" + Metadata for all signals under that branch will be returned unless filtered by filter. + NOTE: Currently Databroker supports also signals and wildcards in root but that may + be removed in a future release! + """ + filter: builtins.str + """NOTE : Currently not considered by Databroker, all signals matching root are returned""" + def __init__( + self, + *, + root: builtins.str = ..., + filter: builtins.str = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["filter", b"filter", "root", b"root"]) -> None: ... + +global___ListMetadataRequest = ListMetadataRequest + +@typing.final +class ListMetadataResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + METADATA_FIELD_NUMBER: builtins.int + @property + def metadata(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[kuksa.val.v2.types_pb2.Metadata]: ... + def __init__( + self, + *, + metadata: collections.abc.Iterable[kuksa.val.v2.types_pb2.Metadata] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["metadata", b"metadata"]) -> None: ... + +global___ListMetadataResponse = ListMetadataResponse + +@typing.final +class PublishValueRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + SIGNAL_ID_FIELD_NUMBER: builtins.int + DATA_POINT_FIELD_NUMBER: builtins.int + @property + def signal_id(self) -> kuksa.val.v2.types_pb2.SignalID: ... + @property + def data_point(self) -> kuksa.val.v2.types_pb2.Datapoint: ... + def __init__( + self, + *, + signal_id: kuksa.val.v2.types_pb2.SignalID | None = ..., + data_point: kuksa.val.v2.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["data_point", b"data_point", "signal_id", b"signal_id"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["data_point", b"data_point", "signal_id", b"signal_id"]) -> None: ... + +global___PublishValueRequest = PublishValueRequest + +@typing.final +class PublishValueResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___PublishValueResponse = PublishValueResponse + +@typing.final +class PublishValuesRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class DataPointsEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.int + @property + def value(self) -> kuksa.val.v2.types_pb2.Datapoint: ... + def __init__( + self, + *, + key: builtins.int = ..., + value: kuksa.val.v2.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + REQUEST_ID_FIELD_NUMBER: builtins.int + DATA_POINTS_FIELD_NUMBER: builtins.int + request_id: builtins.int + """/ Unique request id for the stream that can be used to match the corresponding response.""" + @property + def data_points(self) -> google.protobuf.internal.containers.MessageMap[builtins.int, kuksa.val.v2.types_pb2.Datapoint]: ... + def __init__( + self, + *, + request_id: builtins.int = ..., + data_points: collections.abc.Mapping[builtins.int, kuksa.val.v2.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["data_points", b"data_points", "request_id", b"request_id"]) -> None: ... + +global___PublishValuesRequest = PublishValuesRequest + +@typing.final +class PublishValuesResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class StatusEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.int + @property + def value(self) -> kuksa.val.v2.types_pb2.Error: ... + def __init__( + self, + *, + key: builtins.int = ..., + value: kuksa.val.v2.types_pb2.Error | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + REQUEST_ID_FIELD_NUMBER: builtins.int + STATUS_FIELD_NUMBER: builtins.int + request_id: builtins.int + """/ Unique request id for the stream that can be used to match the corresponding request.""" + @property + def status(self) -> google.protobuf.internal.containers.MessageMap[builtins.int, kuksa.val.v2.types_pb2.Error]: ... + def __init__( + self, + *, + request_id: builtins.int = ..., + status: collections.abc.Mapping[builtins.int, kuksa.val.v2.types_pb2.Error] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["request_id", b"request_id", "status", b"status"]) -> None: ... + +global___PublishValuesResponse = PublishValuesResponse + +@typing.final +class ProvideActuationRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ACTUATOR_IDENTIFIERS_FIELD_NUMBER: builtins.int + @property + def actuator_identifiers(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[kuksa.val.v2.types_pb2.SignalID]: ... + def __init__( + self, + *, + actuator_identifiers: collections.abc.Iterable[kuksa.val.v2.types_pb2.SignalID] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["actuator_identifiers", b"actuator_identifiers"]) -> None: ... + +global___ProvideActuationRequest = ProvideActuationRequest + +@typing.final +class ProvideActuationResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___ProvideActuationResponse = ProvideActuationResponse + +@typing.final +class ProvideSignalRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class SignalsSampleIntervalsEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.int + @property + def value(self) -> kuksa.val.v2.types_pb2.SampleInterval: ... + def __init__( + self, + *, + key: builtins.int = ..., + value: kuksa.val.v2.types_pb2.SampleInterval | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + SIGNALS_SAMPLE_INTERVALS_FIELD_NUMBER: builtins.int + @property + def signals_sample_intervals(self) -> google.protobuf.internal.containers.MessageMap[builtins.int, kuksa.val.v2.types_pb2.SampleInterval]: ... + def __init__( + self, + *, + signals_sample_intervals: collections.abc.Mapping[builtins.int, kuksa.val.v2.types_pb2.SampleInterval] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["signals_sample_intervals", b"signals_sample_intervals"]) -> None: ... + +global___ProvideSignalRequest = ProvideSignalRequest + +@typing.final +class ProvideSignalResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___ProvideSignalResponse = ProvideSignalResponse + +@typing.final +class BatchActuateStreamRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + ACTUATE_REQUESTS_FIELD_NUMBER: builtins.int + @property + def actuate_requests(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ActuateRequest]: ... + def __init__( + self, + *, + actuate_requests: collections.abc.Iterable[global___ActuateRequest] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["actuate_requests", b"actuate_requests"]) -> None: ... + +global___BatchActuateStreamRequest = BatchActuateStreamRequest + +@typing.final +class BatchActuateStreamResponse(google.protobuf.message.Message): + """Message that shall be used by provider to indicate if an actuation request was accepted.""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + SIGNAL_ID_FIELD_NUMBER: builtins.int + ERROR_FIELD_NUMBER: builtins.int + @property + def signal_id(self) -> kuksa.val.v2.types_pb2.SignalID: ... + @property + def error(self) -> kuksa.val.v2.types_pb2.Error: ... + def __init__( + self, + *, + signal_id: kuksa.val.v2.types_pb2.SignalID | None = ..., + error: kuksa.val.v2.types_pb2.Error | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["error", b"error", "signal_id", b"signal_id"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["error", b"error", "signal_id", b"signal_id"]) -> None: ... + +global___BatchActuateStreamResponse = BatchActuateStreamResponse + +@typing.final +class UpdateFilterRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class FiltersUpdateEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.int + @property + def value(self) -> kuksa.val.v2.types_pb2.Filter: ... + def __init__( + self, + *, + key: builtins.int = ..., + value: kuksa.val.v2.types_pb2.Filter | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + REQUEST_ID_FIELD_NUMBER: builtins.int + FILTERS_UPDATE_FIELD_NUMBER: builtins.int + request_id: builtins.int + """/ Unique request id for the stream that can be used to match the corresponding response.""" + @property + def filters_update(self) -> google.protobuf.internal.containers.MessageMap[builtins.int, kuksa.val.v2.types_pb2.Filter]: + """Databroker sends filters to provider. + In case provider restarts, databroker will send local filters stored + to continue the provider sending same signals with same filter. + """ + + def __init__( + self, + *, + request_id: builtins.int = ..., + filters_update: collections.abc.Mapping[builtins.int, kuksa.val.v2.types_pb2.Filter] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["filters_update", b"filters_update", "request_id", b"request_id"]) -> None: ... + +global___UpdateFilterRequest = UpdateFilterRequest + +@typing.final +class UpdateFilterResponse(google.protobuf.message.Message): + """Only returned if there is a filter error on provider side""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + REQUEST_ID_FIELD_NUMBER: builtins.int + FILTER_ERROR_FIELD_NUMBER: builtins.int + request_id: builtins.int + """/ Unique request id for the stream that can be used to match the corresponding request.""" + filter_error: kuksa.val.v2.types_pb2.FilterError.ValueType + def __init__( + self, + *, + request_id: builtins.int = ..., + filter_error: kuksa.val.v2.types_pb2.FilterError.ValueType = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["filter_error", b"filter_error", "request_id", b"request_id"]) -> None: ... + +global___UpdateFilterResponse = UpdateFilterResponse + +@typing.final +class ProviderErrorIndication(google.protobuf.message.Message): + """Send to indicate an error on provider side""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + PROVIDER_ERROR_FIELD_NUMBER: builtins.int + provider_error: kuksa.val.v2.types_pb2.ProviderError.ValueType + def __init__( + self, + *, + provider_error: kuksa.val.v2.types_pb2.ProviderError.ValueType = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["provider_error", b"provider_error"]) -> None: ... + +global___ProviderErrorIndication = ProviderErrorIndication + +@typing.final +class GetProviderValueRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + REQUEST_ID_FIELD_NUMBER: builtins.int + SIGNAL_IDS_FIELD_NUMBER: builtins.int + request_id: builtins.int + """/ Unique request id for the stream that can be used to match the corresponding response.""" + @property + def signal_ids(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... + def __init__( + self, + *, + request_id: builtins.int = ..., + signal_ids: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["request_id", b"request_id", "signal_ids", b"signal_ids"]) -> None: ... + +global___GetProviderValueRequest = GetProviderValueRequest + +@typing.final +class GetProviderValueResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class EntriesEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.int + @property + def value(self) -> kuksa.val.v2.types_pb2.Datapoint: ... + def __init__( + self, + *, + key: builtins.int = ..., + value: kuksa.val.v2.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + REQUEST_ID_FIELD_NUMBER: builtins.int + ENTRIES_FIELD_NUMBER: builtins.int + request_id: builtins.int + """/ Unique request id for the stream that can be used to match the corresponding request.""" + @property + def entries(self) -> google.protobuf.internal.containers.MessageMap[builtins.int, kuksa.val.v2.types_pb2.Datapoint]: ... + def __init__( + self, + *, + request_id: builtins.int = ..., + entries: collections.abc.Mapping[builtins.int, kuksa.val.v2.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["entries", b"entries", "request_id", b"request_id"]) -> None: ... + +global___GetProviderValueResponse = GetProviderValueResponse + +@typing.final +class OpenProviderStreamRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + PROVIDE_ACTUATION_REQUEST_FIELD_NUMBER: builtins.int + PUBLISH_VALUES_REQUEST_FIELD_NUMBER: builtins.int + BATCH_ACTUATE_STREAM_RESPONSE_FIELD_NUMBER: builtins.int + PROVIDE_SIGNAL_REQUEST_FIELD_NUMBER: builtins.int + UPDATE_FILTER_RESPONSE_FIELD_NUMBER: builtins.int + GET_PROVIDER_VALUE_RESPONSE_FIELD_NUMBER: builtins.int + PROVIDER_ERROR_INDICATION_FIELD_NUMBER: builtins.int + @property + def provide_actuation_request(self) -> global___ProvideActuationRequest: + """Inform server of an actuator this provider provides.""" + + @property + def publish_values_request(self) -> global___PublishValuesRequest: + """Publish a value.""" + + @property + def batch_actuate_stream_response(self) -> global___BatchActuateStreamResponse: + """Sent to acknowledge the acceptance of a batch actuate + request. + """ + + @property + def provide_signal_request(self) -> global___ProvideSignalRequest: + """Inform server of a signal this provider provides.""" + + @property + def update_filter_response(self) -> global___UpdateFilterResponse: + """Update filter response""" + + @property + def get_provider_value_response(self) -> global___GetProviderValueResponse: + """GetValue response""" + + @property + def provider_error_indication(self) -> global___ProviderErrorIndication: + """Indication of error on provider side""" + + def __init__( + self, + *, + provide_actuation_request: global___ProvideActuationRequest | None = ..., + publish_values_request: global___PublishValuesRequest | None = ..., + batch_actuate_stream_response: global___BatchActuateStreamResponse | None = ..., + provide_signal_request: global___ProvideSignalRequest | None = ..., + update_filter_response: global___UpdateFilterResponse | None = ..., + get_provider_value_response: global___GetProviderValueResponse | None = ..., + provider_error_indication: global___ProviderErrorIndication | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["action", b"action", "batch_actuate_stream_response", b"batch_actuate_stream_response", "get_provider_value_response", b"get_provider_value_response", "provide_actuation_request", b"provide_actuation_request", "provide_signal_request", b"provide_signal_request", "provider_error_indication", b"provider_error_indication", "publish_values_request", b"publish_values_request", "update_filter_response", b"update_filter_response"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["action", b"action", "batch_actuate_stream_response", b"batch_actuate_stream_response", "get_provider_value_response", b"get_provider_value_response", "provide_actuation_request", b"provide_actuation_request", "provide_signal_request", b"provide_signal_request", "provider_error_indication", b"provider_error_indication", "publish_values_request", b"publish_values_request", "update_filter_response", b"update_filter_response"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["action", b"action"]) -> typing.Literal["provide_actuation_request", "publish_values_request", "batch_actuate_stream_response", "provide_signal_request", "update_filter_response", "get_provider_value_response", "provider_error_indication"] | None: ... + +global___OpenProviderStreamRequest = OpenProviderStreamRequest + +@typing.final +class OpenProviderStreamResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + PROVIDE_ACTUATION_RESPONSE_FIELD_NUMBER: builtins.int + PUBLISH_VALUES_RESPONSE_FIELD_NUMBER: builtins.int + BATCH_ACTUATE_STREAM_REQUEST_FIELD_NUMBER: builtins.int + PROVIDE_SIGNAL_RESPONSE_FIELD_NUMBER: builtins.int + UPDATE_FILTER_REQUEST_FIELD_NUMBER: builtins.int + GET_PROVIDER_VALUE_REQUEST_FIELD_NUMBER: builtins.int + @property + def provide_actuation_response(self) -> global___ProvideActuationResponse: + """Response to a provide actuator request.""" + + @property + def publish_values_response(self) -> global___PublishValuesResponse: + """Acknowledgement that a published value was received.""" + + @property + def batch_actuate_stream_request(self) -> global___BatchActuateStreamRequest: + """Send a batch actuate request to a provider.""" + + @property + def provide_signal_response(self) -> global___ProvideSignalResponse: + """Response to a provide sensor request.""" + + @property + def update_filter_request(self) -> global___UpdateFilterRequest: + """Filter request""" + + @property + def get_provider_value_request(self) -> global___GetProviderValueRequest: + """GetValue request from client forwarded to provider""" + + def __init__( + self, + *, + provide_actuation_response: global___ProvideActuationResponse | None = ..., + publish_values_response: global___PublishValuesResponse | None = ..., + batch_actuate_stream_request: global___BatchActuateStreamRequest | None = ..., + provide_signal_response: global___ProvideSignalResponse | None = ..., + update_filter_request: global___UpdateFilterRequest | None = ..., + get_provider_value_request: global___GetProviderValueRequest | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["action", b"action", "batch_actuate_stream_request", b"batch_actuate_stream_request", "get_provider_value_request", b"get_provider_value_request", "provide_actuation_response", b"provide_actuation_response", "provide_signal_response", b"provide_signal_response", "publish_values_response", b"publish_values_response", "update_filter_request", b"update_filter_request"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["action", b"action", "batch_actuate_stream_request", b"batch_actuate_stream_request", "get_provider_value_request", b"get_provider_value_request", "provide_actuation_response", b"provide_actuation_response", "provide_signal_response", b"provide_signal_response", "publish_values_response", b"publish_values_response", "update_filter_request", b"update_filter_request"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["action", b"action"]) -> typing.Literal["provide_actuation_response", "publish_values_response", "batch_actuate_stream_request", "provide_signal_response", "update_filter_request", "get_provider_value_request"] | None: ... + +global___OpenProviderStreamResponse = OpenProviderStreamResponse + +@typing.final +class GetServerInfoRequest(google.protobuf.message.Message): + """Nothing yet""" + + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + def __init__( + self, + ) -> None: ... + +global___GetServerInfoRequest = GetServerInfoRequest + +@typing.final +class GetServerInfoResponse(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + NAME_FIELD_NUMBER: builtins.int + VERSION_FIELD_NUMBER: builtins.int + COMMIT_HASH_FIELD_NUMBER: builtins.int + name: builtins.str + version: builtins.str + commit_hash: builtins.str + def __init__( + self, + *, + name: builtins.str = ..., + version: builtins.str = ..., + commit_hash: builtins.str = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["commit_hash", b"commit_hash", "name", b"name", "version", b"version"]) -> None: ... + +global___GetServerInfoResponse = GetServerInfoResponse diff --git a/integration_test/gen_proto/kuksa/val/v2/val_pb2_grpc.py b/integration_test/gen_proto/kuksa/val/v2/val_pb2_grpc.py new file mode 100644 index 00000000..21c77f82 --- /dev/null +++ b/integration_test/gen_proto/kuksa/val/v2/val_pb2_grpc.py @@ -0,0 +1,710 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +# Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT! +"""Client and server classes corresponding to protobuf-defined services.""" +import grpc +import warnings + +from gen_proto.kuksa.val.v2 import val_pb2 as kuksa_dot_val_dot_v2_dot_val__pb2 + +GRPC_GENERATED_VERSION = '1.71.0' +GRPC_VERSION = grpc.__version__ +_version_not_supported = False + +try: + from grpc._utilities import first_version_is_lower + _version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION) +except ImportError: + _version_not_supported = True + +if _version_not_supported: + raise RuntimeError( + f'The grpc package installed is at version {GRPC_VERSION},' + + f' but the generated code in kuksa/val/v2/val_pb2_grpc.py depends on' + + f' grpcio>={GRPC_GENERATED_VERSION}.' + + f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}' + + f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.' + ) + + +class VALStub(object): + """Missing associated documentation comment in .proto file.""" + + def __init__(self, channel): + """Constructor. + + Args: + channel: A grpc.Channel. + """ + self.GetValue = channel.unary_unary( + '/kuksa.val.v2.VAL/GetValue', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetValueRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetValueResponse.FromString, + _registered_method=True) + self.GetValues = channel.unary_unary( + '/kuksa.val.v2.VAL/GetValues', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetValuesRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetValuesResponse.FromString, + _registered_method=True) + self.Subscribe = channel.unary_stream( + '/kuksa.val.v2.VAL/Subscribe', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeResponse.FromString, + _registered_method=True) + self.SubscribeById = channel.unary_stream( + '/kuksa.val.v2.VAL/SubscribeById', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeByIdRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeByIdResponse.FromString, + _registered_method=True) + self.Actuate = channel.unary_unary( + '/kuksa.val.v2.VAL/Actuate', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.ActuateRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.ActuateResponse.FromString, + _registered_method=True) + self.ActuateStream = channel.stream_unary( + '/kuksa.val.v2.VAL/ActuateStream', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.ActuateRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.ActuateResponse.FromString, + _registered_method=True) + self.BatchActuate = channel.unary_unary( + '/kuksa.val.v2.VAL/BatchActuate', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.BatchActuateRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.BatchActuateResponse.FromString, + _registered_method=True) + self.ListMetadata = channel.unary_unary( + '/kuksa.val.v2.VAL/ListMetadata', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.ListMetadataRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.ListMetadataResponse.FromString, + _registered_method=True) + self.PublishValue = channel.unary_unary( + '/kuksa.val.v2.VAL/PublishValue', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.PublishValueRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.PublishValueResponse.FromString, + _registered_method=True) + self.OpenProviderStream = channel.stream_stream( + '/kuksa.val.v2.VAL/OpenProviderStream', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.OpenProviderStreamRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.OpenProviderStreamResponse.FromString, + _registered_method=True) + self.GetServerInfo = channel.unary_unary( + '/kuksa.val.v2.VAL/GetServerInfo', + request_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetServerInfoRequest.SerializeToString, + response_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetServerInfoResponse.FromString, + _registered_method=True) + + +class VALServicer(object): + """Missing associated documentation comment in .proto file.""" + + def GetValue(self, request, context): + """Get the latest value of a signal + If the signal exist but does not have a valid value + a DataPoint where value is None shall be returned. + + Returns (GRPC error code): + NOT_FOUND if the requested signal doesn't exist + UNAUTHENTICATED if no credentials provided or credentials has expired + PERMISSION_DENIED if access is denied + INVALID_ARGUMENT if the request is empty or provided path is too long + - MAX_REQUEST_PATH_LENGTH: usize = 1000; + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def GetValues(self, request, context): + """Get the latest values of a set of signals. + The returned list of data points has the same order as the list of the request. + If a requested signal has no value a DataPoint where value is None will be returned. + + Returns (GRPC error code): + NOT_FOUND if any of the requested signals doesn't exist. + UNAUTHENTICATED if no credentials provided or credentials has expired + PERMISSION_DENIED if access is denied for any of the requested signals. + INVALID_ARGUMENT if the request is empty or provided path is too long + - MAX_REQUEST_PATH_LENGTH: usize = 1000; + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def Subscribe(self, request, context): + """Subscribe to a set of signals using string path parameters + Returns (GRPC error code): + NOT_FOUND if any of the signals are non-existant. + UNAUTHENTICATED if no credentials provided or credentials has expired + PERMISSION_DENIED if access is denied for any of the signals. + INVALID_ARGUMENT + - if the request is empty or provided path is too long + MAX_REQUEST_PATH_LENGTH: usize = 1000; + - if buffer_size exceeds the maximum permitted + MAX_BUFFER_SIZE: usize = 1000; + + When subscribing, Databroker shall immediately return the value for all + subscribed entries. + If a value isn't available when subscribing to a it, it should return None + + If a subscriber is slow to consume signals, messages will be buffered up + to the specified buffer_size before the oldest messages are dropped. + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def SubscribeById(self, request, context): + """Subscribe to a set of signals using i32 id parameters + Returns (GRPC error code): + NOT_FOUND if any of the signals are non-existant. + UNAUTHENTICATED if no credentials provided or credentials has expired + PERMISSION_DENIED if access is denied for any of the signals. + INVALID_ARGUMENT + - if the request is empty or provided path is too long + MAX_REQUEST_PATH_LENGTH: usize = 1000; + - if buffer_size exceeds the maximum permitted + MAX_BUFFER_SIZE: usize = 1000; + + When subscribing, Databroker shall immediately return the value for all + subscribed entries. + If a value isn't available when subscribing to a it, it should return None + + If a subscriber is slow to consume signals, messages will be buffered up + to the specified buffer_size before the oldest messages are dropped. + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def Actuate(self, request, context): + """Actuate a single actuator + + Returns (GRPC error code): + NOT_FOUND if the actuator does not exist. + PERMISSION_DENIED if access is denied for the actuator. + UNAUTHENTICATED if no credentials provided or credentials has expired + UNAVAILABLE if there is no provider currently providing the actuator + DATA_LOSS is there is a internal TransmissionFailure + INVALID_ARGUMENT + - if the provided path is not an actuator. + - if the data type used in the request does not match + the data type of the addressed signal + - if the requested value is not accepted, + e.g. if sending an unsupported enum value + - if the provided value is out of the min/max range specified + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def ActuateStream(self, request_iterator, context): + """Actuate a single actuator in a gRPC stream -> Use for low latency and high throughput + + Returns (GRPC error code): + NOT_FOUND if the actuator does not exist. + PERMISSION_DENIED if access is denied for the actuator. + UNAUTHENTICATED if no credentials provided or credentials has expired + UNAVAILABLE if there is no provider currently providing the actuator + DATA_LOSS is there is a internal TransmissionFailure + INVALID_ARGUMENT + - if the provided path is not an actuator. + - if the data type used in the request does not match + the data type of the addressed signal + - if the requested value is not accepted, + e.g. if sending an unsupported enum value + - if the provided value is out of the min/max range specified + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def BatchActuate(self, request, context): + """Actuate simultaneously multiple actuators. + If any error occurs, the entire operation will be aborted + and no single actuator value will be forwarded to the provider. + + Returns (GRPC error code): + NOT_FOUND if any of the actuators are non-existant. + PERMISSION_DENIED if access is denied for any of the actuators. + UNAUTHENTICATED if no credentials provided or credentials has expired + UNAVAILABLE if there is no provider currently providing an actuator + DATA_LOSS is there is a internal TransmissionFailure + INVALID_ARGUMENT + - if any of the provided path is not an actuator. + - if the data type used in the request does not match + the data type of the addressed signal + - if the requested value is not accepted, + e.g. if sending an unsupported enum value + - if any of the provided actuators values are out of the min/max range specified + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def ListMetadata(self, request, context): + """List metadata of signals matching the request. + + Returns (GRPC error code): + NOT_FOUND if the specified root branch does not exist. + UNAUTHENTICATED if no credentials provided or credentials has expired + INVALID_ARGUMENT if the provided path or wildcard is wrong. + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def PublishValue(self, request, context): + """Publish a signal value. Used for low frequency signals (e.g. attributes). + + Returns (GRPC error code): + NOT_FOUND if any of the signals are non-existant. + PERMISSION_DENIED + - if access is denied for any of the signals. + UNAUTHENTICATED if no credentials provided or credentials has expired + INVALID_ARGUMENT + - if the data type used in the request does not match + the data type of the addressed signal + - if the published value is not accepted, + e.g. if sending an unsupported enum value + - if the published value is out of the min/max range specified + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def OpenProviderStream(self, request_iterator, context): + """Open a stream used to provide actuation and/or publishing values using + a streaming interface. Used to provide actuators and to enable high frequency + updates of values. + + The open stream is used for request / response type communication between the + provider and server (where the initiator of a request can vary). + + Errors: + - Provider sends ProvideActuationRequest -> Databroker returns ProvideActuationResponse + Returns (GRPC error code) and closes the stream call (strict case). + NOT_FOUND if any of the signals are non-existant. + PERMISSION_DENIED if access is denied for any of the signals. + UNAUTHENTICATED if no credentials provided or credentials has expired + ALREADY_EXISTS if a provider already claimed the ownership of an actuator + + - Provider sends PublishValuesRequest -> Databroker returns PublishValuesResponse upon error, and nothing upon success + GRPC errors are returned as messages in the stream + response with the signal id `map status = 2;` (permissive case) + NOT_FOUND if a signal is non-existant. + PERMISSION_DENIED + - if access is denied for a signal. + INVALID_ARGUMENT + - if the data type used in the request does not match + the data type of the addressed signal + - if the published value is not accepted, + e.g. if sending an unsupported enum value + - if the published value is out of the min/max range specified + + - Databroker sends BatchActuateStreamRequest -> Provider shall return a BatchActuateStreamResponse, + for every signal requested to indicate if the request was accepted or not. + It is up to the provider to decide if the stream shall be closed, + as of today Databroker will not react on the received error message. + + - Provider sends ProvideSignalRequest -> Databroker returns ProvideSignalResponse + Returns (GRPC error code) and closes the stream call (strict case). + NOT_FOUND if any of the signals are non-existant. + PERMISSION_DENIED if access is denied for any of the signals. + UNAUTHENTICATED if no credentials provided or credentials has expired + ALREADY_EXISTS if a provider already claimed the ownership of any signal. + + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + def GetServerInfo(self, request, context): + """Get server information + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + + +def add_VALServicer_to_server(servicer, server): + rpc_method_handlers = { + 'GetValue': grpc.unary_unary_rpc_method_handler( + servicer.GetValue, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetValueRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetValueResponse.SerializeToString, + ), + 'GetValues': grpc.unary_unary_rpc_method_handler( + servicer.GetValues, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetValuesRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetValuesResponse.SerializeToString, + ), + 'Subscribe': grpc.unary_stream_rpc_method_handler( + servicer.Subscribe, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeResponse.SerializeToString, + ), + 'SubscribeById': grpc.unary_stream_rpc_method_handler( + servicer.SubscribeById, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeByIdRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeByIdResponse.SerializeToString, + ), + 'Actuate': grpc.unary_unary_rpc_method_handler( + servicer.Actuate, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.ActuateRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.ActuateResponse.SerializeToString, + ), + 'ActuateStream': grpc.stream_unary_rpc_method_handler( + servicer.ActuateStream, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.ActuateRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.ActuateResponse.SerializeToString, + ), + 'BatchActuate': grpc.unary_unary_rpc_method_handler( + servicer.BatchActuate, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.BatchActuateRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.BatchActuateResponse.SerializeToString, + ), + 'ListMetadata': grpc.unary_unary_rpc_method_handler( + servicer.ListMetadata, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.ListMetadataRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.ListMetadataResponse.SerializeToString, + ), + 'PublishValue': grpc.unary_unary_rpc_method_handler( + servicer.PublishValue, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.PublishValueRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.PublishValueResponse.SerializeToString, + ), + 'OpenProviderStream': grpc.stream_stream_rpc_method_handler( + servicer.OpenProviderStream, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.OpenProviderStreamRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.OpenProviderStreamResponse.SerializeToString, + ), + 'GetServerInfo': grpc.unary_unary_rpc_method_handler( + servicer.GetServerInfo, + request_deserializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetServerInfoRequest.FromString, + response_serializer=kuksa_dot_val_dot_v2_dot_val__pb2.GetServerInfoResponse.SerializeToString, + ), + } + generic_handler = grpc.method_handlers_generic_handler( + 'kuksa.val.v2.VAL', rpc_method_handlers) + server.add_generic_rpc_handlers((generic_handler,)) + server.add_registered_method_handlers('kuksa.val.v2.VAL', rpc_method_handlers) + + + # This class is part of an EXPERIMENTAL API. +class VAL(object): + """Missing associated documentation comment in .proto file.""" + + @staticmethod + def GetValue(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v2.VAL/GetValue', + kuksa_dot_val_dot_v2_dot_val__pb2.GetValueRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.GetValueResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def GetValues(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v2.VAL/GetValues', + kuksa_dot_val_dot_v2_dot_val__pb2.GetValuesRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.GetValuesResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def Subscribe(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_stream( + request, + target, + '/kuksa.val.v2.VAL/Subscribe', + kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def SubscribeById(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_stream( + request, + target, + '/kuksa.val.v2.VAL/SubscribeById', + kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeByIdRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.SubscribeByIdResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def Actuate(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v2.VAL/Actuate', + kuksa_dot_val_dot_v2_dot_val__pb2.ActuateRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.ActuateResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def ActuateStream(request_iterator, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.stream_unary( + request_iterator, + target, + '/kuksa.val.v2.VAL/ActuateStream', + kuksa_dot_val_dot_v2_dot_val__pb2.ActuateRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.ActuateResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def BatchActuate(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v2.VAL/BatchActuate', + kuksa_dot_val_dot_v2_dot_val__pb2.BatchActuateRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.BatchActuateResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def ListMetadata(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v2.VAL/ListMetadata', + kuksa_dot_val_dot_v2_dot_val__pb2.ListMetadataRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.ListMetadataResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def PublishValue(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v2.VAL/PublishValue', + kuksa_dot_val_dot_v2_dot_val__pb2.PublishValueRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.PublishValueResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def OpenProviderStream(request_iterator, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.stream_stream( + request_iterator, + target, + '/kuksa.val.v2.VAL/OpenProviderStream', + kuksa_dot_val_dot_v2_dot_val__pb2.OpenProviderStreamRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.OpenProviderStreamResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def GetServerInfo(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/kuksa.val.v2.VAL/GetServerInfo', + kuksa_dot_val_dot_v2_dot_val__pb2.GetServerInfoRequest.SerializeToString, + kuksa_dot_val_dot_v2_dot_val__pb2.GetServerInfoResponse.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) diff --git a/integration_test/gen_proto/sdv/databroker/v1/broker_pb2.py b/integration_test/gen_proto/sdv/databroker/v1/broker_pb2.py index 98dc94a8..efcba6aa 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/broker_pb2.py +++ b/integration_test/gen_proto/sdv/databroker/v1/broker_pb2.py @@ -1,12 +1,35 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ # -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! +# NO CHECKED-IN PROTOBUF GENCODE # source: sdv/databroker/v1/broker.proto +# Protobuf Python Version: 5.29.0 """Generated protocol buffer code.""" from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool -from google.protobuf import message as _message -from google.protobuf import reflection as _reflection +from google.protobuf import runtime_version as _runtime_version from google.protobuf import symbol_database as _symbol_database +from google.protobuf.internal import builder as _builder +_runtime_version.ValidateProtobufRuntimeVersion( + _runtime_version.Domain.PUBLIC, + 5, + 29, + 0, + '', + 'sdv/databroker/v1/broker.proto' +) # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() @@ -15,100 +38,45 @@ from gen_proto.sdv.databroker.v1 import types_pb2 as sdv_dot_databroker_dot_v1_dot_types__pb2 -DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\x1esdv/databroker/v1/broker.proto\x12\x11sdv.databroker.v1\x1a\x1dsdv/databroker/v1/types.proto\"*\n\x14GetDatapointsRequest\x12\x12\n\ndatapoints\x18\x01 \x03(\t\"\xb0\x01\n\x12GetDatapointsReply\x12I\n\ndatapoints\x18\x01 \x03(\x0b\x32\x35.sdv.databroker.v1.GetDatapointsReply.DatapointsEntry\x1aO\n\x0f\x44\x61tapointsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12+\n\x05value\x18\x02 \x01(\x0b\x32\x1c.sdv.databroker.v1.Datapoint:\x02\x38\x01\"!\n\x10SubscribeRequest\x12\r\n\x05query\x18\x02 \x01(\t\"\x9c\x01\n\x0eSubscribeReply\x12=\n\x06\x66ields\x18\x01 \x03(\x0b\x32-.sdv.databroker.v1.SubscribeReply.FieldsEntry\x1aK\n\x0b\x46ieldsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12+\n\x05value\x18\x02 \x01(\x0b\x32\x1c.sdv.databroker.v1.Datapoint:\x02\x38\x01\"#\n\x12GetMetadataRequest\x12\r\n\x05names\x18\x01 \x03(\t\"=\n\x10GetMetadataReply\x12)\n\x04list\x18\x01 \x03(\x0b\x32\x1b.sdv.databroker.v1.Metadata2\x9b\x02\n\x06\x42roker\x12_\n\rGetDatapoints\x12\'.sdv.databroker.v1.GetDatapointsRequest\x1a%.sdv.databroker.v1.GetDatapointsReply\x12U\n\tSubscribe\x12#.sdv.databroker.v1.SubscribeRequest\x1a!.sdv.databroker.v1.SubscribeReply0\x01\x12Y\n\x0bGetMetadata\x12%.sdv.databroker.v1.GetMetadataRequest\x1a#.sdv.databroker.v1.GetMetadataReplyb\x06proto3') - - - -_GETDATAPOINTSREQUEST = DESCRIPTOR.message_types_by_name['GetDatapointsRequest'] -_GETDATAPOINTSREPLY = DESCRIPTOR.message_types_by_name['GetDatapointsReply'] -_GETDATAPOINTSREPLY_DATAPOINTSENTRY = _GETDATAPOINTSREPLY.nested_types_by_name['DatapointsEntry'] -_SUBSCRIBEREQUEST = DESCRIPTOR.message_types_by_name['SubscribeRequest'] -_SUBSCRIBEREPLY = DESCRIPTOR.message_types_by_name['SubscribeReply'] -_SUBSCRIBEREPLY_FIELDSENTRY = _SUBSCRIBEREPLY.nested_types_by_name['FieldsEntry'] -_GETMETADATAREQUEST = DESCRIPTOR.message_types_by_name['GetMetadataRequest'] -_GETMETADATAREPLY = DESCRIPTOR.message_types_by_name['GetMetadataReply'] -GetDatapointsRequest = _reflection.GeneratedProtocolMessageType('GetDatapointsRequest', (_message.Message,), { - 'DESCRIPTOR' : _GETDATAPOINTSREQUEST, - '__module__' : 'sdv.databroker.v1.broker_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.GetDatapointsRequest) - }) -_sym_db.RegisterMessage(GetDatapointsRequest) - -GetDatapointsReply = _reflection.GeneratedProtocolMessageType('GetDatapointsReply', (_message.Message,), { - - 'DatapointsEntry' : _reflection.GeneratedProtocolMessageType('DatapointsEntry', (_message.Message,), { - 'DESCRIPTOR' : _GETDATAPOINTSREPLY_DATAPOINTSENTRY, - '__module__' : 'sdv.databroker.v1.broker_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.GetDatapointsReply.DatapointsEntry) - }) - , - 'DESCRIPTOR' : _GETDATAPOINTSREPLY, - '__module__' : 'sdv.databroker.v1.broker_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.GetDatapointsReply) - }) -_sym_db.RegisterMessage(GetDatapointsReply) -_sym_db.RegisterMessage(GetDatapointsReply.DatapointsEntry) - -SubscribeRequest = _reflection.GeneratedProtocolMessageType('SubscribeRequest', (_message.Message,), { - 'DESCRIPTOR' : _SUBSCRIBEREQUEST, - '__module__' : 'sdv.databroker.v1.broker_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.SubscribeRequest) - }) -_sym_db.RegisterMessage(SubscribeRequest) - -SubscribeReply = _reflection.GeneratedProtocolMessageType('SubscribeReply', (_message.Message,), { - - 'FieldsEntry' : _reflection.GeneratedProtocolMessageType('FieldsEntry', (_message.Message,), { - 'DESCRIPTOR' : _SUBSCRIBEREPLY_FIELDSENTRY, - '__module__' : 'sdv.databroker.v1.broker_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.SubscribeReply.FieldsEntry) - }) - , - 'DESCRIPTOR' : _SUBSCRIBEREPLY, - '__module__' : 'sdv.databroker.v1.broker_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.SubscribeReply) - }) -_sym_db.RegisterMessage(SubscribeReply) -_sym_db.RegisterMessage(SubscribeReply.FieldsEntry) - -GetMetadataRequest = _reflection.GeneratedProtocolMessageType('GetMetadataRequest', (_message.Message,), { - 'DESCRIPTOR' : _GETMETADATAREQUEST, - '__module__' : 'sdv.databroker.v1.broker_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.GetMetadataRequest) - }) -_sym_db.RegisterMessage(GetMetadataRequest) - -GetMetadataReply = _reflection.GeneratedProtocolMessageType('GetMetadataReply', (_message.Message,), { - 'DESCRIPTOR' : _GETMETADATAREPLY, - '__module__' : 'sdv.databroker.v1.broker_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.GetMetadataReply) - }) -_sym_db.RegisterMessage(GetMetadataReply) - -_BROKER = DESCRIPTOR.services_by_name['Broker'] -if _descriptor._USE_C_DESCRIPTORS == False: - - DESCRIPTOR._options = None - _GETDATAPOINTSREPLY_DATAPOINTSENTRY._options = None - _GETDATAPOINTSREPLY_DATAPOINTSENTRY._serialized_options = b'8\001' - _SUBSCRIBEREPLY_FIELDSENTRY._options = None - _SUBSCRIBEREPLY_FIELDSENTRY._serialized_options = b'8\001' - _GETDATAPOINTSREQUEST._serialized_start=84 - _GETDATAPOINTSREQUEST._serialized_end=126 - _GETDATAPOINTSREPLY._serialized_start=129 - _GETDATAPOINTSREPLY._serialized_end=305 - _GETDATAPOINTSREPLY_DATAPOINTSENTRY._serialized_start=226 - _GETDATAPOINTSREPLY_DATAPOINTSENTRY._serialized_end=305 - _SUBSCRIBEREQUEST._serialized_start=307 - _SUBSCRIBEREQUEST._serialized_end=340 - _SUBSCRIBEREPLY._serialized_start=343 - _SUBSCRIBEREPLY._serialized_end=499 - _SUBSCRIBEREPLY_FIELDSENTRY._serialized_start=424 - _SUBSCRIBEREPLY_FIELDSENTRY._serialized_end=499 - _GETMETADATAREQUEST._serialized_start=501 - _GETMETADATAREQUEST._serialized_end=536 - _GETMETADATAREPLY._serialized_start=538 - _GETMETADATAREPLY._serialized_end=599 - _BROKER._serialized_start=602 - _BROKER._serialized_end=885 +DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\x1esdv/databroker/v1/broker.proto\x12\x11sdv.databroker.v1\x1a\x1dsdv/databroker/v1/types.proto\"*\n\x14GetDatapointsRequest\x12\x12\n\ndatapoints\x18\x01 \x03(\t\"\xb0\x01\n\x12GetDatapointsReply\x12I\n\ndatapoints\x18\x01 \x03(\x0b\x32\x35.sdv.databroker.v1.GetDatapointsReply.DatapointsEntry\x1aO\n\x0f\x44\x61tapointsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12+\n\x05value\x18\x02 \x01(\x0b\x32\x1c.sdv.databroker.v1.Datapoint:\x02\x38\x01\"\xb4\x01\n\x14SetDatapointsRequest\x12K\n\ndatapoints\x18\x01 \x03(\x0b\x32\x37.sdv.databroker.v1.SetDatapointsRequest.DatapointsEntry\x1aO\n\x0f\x44\x61tapointsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12+\n\x05value\x18\x02 \x01(\x0b\x32\x1c.sdv.databroker.v1.Datapoint:\x02\x38\x01\"\xa9\x01\n\x12SetDatapointsReply\x12\x41\n\x06\x65rrors\x18\x01 \x03(\x0b\x32\x31.sdv.databroker.v1.SetDatapointsReply.ErrorsEntry\x1aP\n\x0b\x45rrorsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\x30\n\x05value\x18\x02 \x01(\x0e\x32!.sdv.databroker.v1.DatapointError:\x02\x38\x01\"!\n\x10SubscribeRequest\x12\r\n\x05query\x18\x02 \x01(\t\"\x9c\x01\n\x0eSubscribeReply\x12=\n\x06\x66ields\x18\x01 \x03(\x0b\x32-.sdv.databroker.v1.SubscribeReply.FieldsEntry\x1aK\n\x0b\x46ieldsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12+\n\x05value\x18\x02 \x01(\x0b\x32\x1c.sdv.databroker.v1.Datapoint:\x02\x38\x01\"#\n\x12GetMetadataRequest\x12\r\n\x05names\x18\x01 \x03(\t\"=\n\x10GetMetadataReply\x12)\n\x04list\x18\x01 \x03(\x0b\x32\x1b.sdv.databroker.v1.Metadata2\xfc\x02\n\x06\x42roker\x12_\n\rGetDatapoints\x12\'.sdv.databroker.v1.GetDatapointsRequest\x1a%.sdv.databroker.v1.GetDatapointsReply\x12_\n\rSetDatapoints\x12\'.sdv.databroker.v1.SetDatapointsRequest\x1a%.sdv.databroker.v1.SetDatapointsReply\x12U\n\tSubscribe\x12#.sdv.databroker.v1.SubscribeRequest\x1a!.sdv.databroker.v1.SubscribeReply0\x01\x12Y\n\x0bGetMetadata\x12%.sdv.databroker.v1.GetMetadataRequest\x1a#.sdv.databroker.v1.GetMetadataReplyb\x06proto3') + +_globals = globals() +_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) +_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'sdv.databroker.v1.broker_pb2', _globals) +if not _descriptor._USE_C_DESCRIPTORS: + DESCRIPTOR._loaded_options = None + _globals['_GETDATAPOINTSREPLY_DATAPOINTSENTRY']._loaded_options = None + _globals['_GETDATAPOINTSREPLY_DATAPOINTSENTRY']._serialized_options = b'8\001' + _globals['_SETDATAPOINTSREQUEST_DATAPOINTSENTRY']._loaded_options = None + _globals['_SETDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_options = b'8\001' + _globals['_SETDATAPOINTSREPLY_ERRORSENTRY']._loaded_options = None + _globals['_SETDATAPOINTSREPLY_ERRORSENTRY']._serialized_options = b'8\001' + _globals['_SUBSCRIBEREPLY_FIELDSENTRY']._loaded_options = None + _globals['_SUBSCRIBEREPLY_FIELDSENTRY']._serialized_options = b'8\001' + _globals['_GETDATAPOINTSREQUEST']._serialized_start=84 + _globals['_GETDATAPOINTSREQUEST']._serialized_end=126 + _globals['_GETDATAPOINTSREPLY']._serialized_start=129 + _globals['_GETDATAPOINTSREPLY']._serialized_end=305 + _globals['_GETDATAPOINTSREPLY_DATAPOINTSENTRY']._serialized_start=226 + _globals['_GETDATAPOINTSREPLY_DATAPOINTSENTRY']._serialized_end=305 + _globals['_SETDATAPOINTSREQUEST']._serialized_start=308 + _globals['_SETDATAPOINTSREQUEST']._serialized_end=488 + _globals['_SETDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_start=226 + _globals['_SETDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_end=305 + _globals['_SETDATAPOINTSREPLY']._serialized_start=491 + _globals['_SETDATAPOINTSREPLY']._serialized_end=660 + _globals['_SETDATAPOINTSREPLY_ERRORSENTRY']._serialized_start=580 + _globals['_SETDATAPOINTSREPLY_ERRORSENTRY']._serialized_end=660 + _globals['_SUBSCRIBEREQUEST']._serialized_start=662 + _globals['_SUBSCRIBEREQUEST']._serialized_end=695 + _globals['_SUBSCRIBEREPLY']._serialized_start=698 + _globals['_SUBSCRIBEREPLY']._serialized_end=854 + _globals['_SUBSCRIBEREPLY_FIELDSENTRY']._serialized_start=779 + _globals['_SUBSCRIBEREPLY_FIELDSENTRY']._serialized_end=854 + _globals['_GETMETADATAREQUEST']._serialized_start=856 + _globals['_GETMETADATAREQUEST']._serialized_end=891 + _globals['_GETMETADATAREPLY']._serialized_start=893 + _globals['_GETMETADATAREPLY']._serialized_end=954 + _globals['_BROKER']._serialized_start=957 + _globals['_BROKER']._serialized_end=1337 # @@protoc_insertion_point(module_scope) diff --git a/integration_test/gen_proto/sdv/databroker/v1/broker_pb2.pyi b/integration_test/gen_proto/sdv/databroker/v1/broker_pb2.pyi index 0d0c4ee3..00ad8a24 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/broker_pb2.pyi +++ b/integration_test/gen_proto/sdv/databroker/v1/broker_pb2.pyi @@ -1,133 +1,256 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ """ @generated by mypy-protobuf. Do not edit manually! isort:skip_file +******************************************************************************* +Copyright (c) 2022 Contributors to the Eclipse Foundation + +See the NOTICE file(s) distributed with this work for additional +information regarding copyright ownership. + +This program and the accompanying materials are made available under the +terms of the Apache License 2.0 which is available at +http://www.apache.org/licenses/LICENSE-2.0 + +SPDX-License-Identifier: Apache-2.0 +****************************************************************************** """ + import builtins +import collections.abc import google.protobuf.descriptor import google.protobuf.internal.containers import google.protobuf.message import gen_proto.sdv.databroker.v1.types_pb2 import typing -import typing_extensions DESCRIPTOR: google.protobuf.descriptor.FileDescriptor +@typing.final class GetDatapointsRequest(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + DATAPOINTS_FIELD_NUMBER: builtins.int @property - def datapoints(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[typing.Text]: + def datapoints(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: """A list of requested data points.""" - pass - def __init__(self, + + def __init__( + self, *, - datapoints: typing.Optional[typing.Iterable[typing.Text]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["datapoints",b"datapoints"]) -> None: ... + datapoints: collections.abc.Iterable[builtins.str] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["datapoints", b"datapoints"]) -> None: ... + global___GetDatapointsRequest = GetDatapointsRequest +@typing.final class GetDatapointsReply(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final class DatapointsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int - key: typing.Text + key: builtins.str @property def value(self) -> sdv.databroker.v1.types_pb2.Datapoint: ... - def __init__(self, + def __init__( + self, *, - key: typing.Text = ..., - value: typing.Optional[sdv.databroker.v1.types_pb2.Datapoint] = ..., - ) -> None: ... - def HasField(self, field_name: typing_extensions.Literal["value",b"value"]) -> builtins.bool: ... - def ClearField(self, field_name: typing_extensions.Literal["key",b"key","value",b"value"]) -> None: ... + key: builtins.str = ..., + value: sdv.databroker.v1.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... DATAPOINTS_FIELD_NUMBER: builtins.int @property - def datapoints(self) -> google.protobuf.internal.containers.MessageMap[typing.Text, sdv.databroker.v1.types_pb2.Datapoint]: + def datapoints(self) -> google.protobuf.internal.containers.MessageMap[builtins.str, sdv.databroker.v1.types_pb2.Datapoint]: """Contains the values of the requested data points. If a requested data point is not available, the corresponding Datapoint will have the respective failure value set. """ - pass - def __init__(self, + + def __init__( + self, *, - datapoints: typing.Optional[typing.Mapping[typing.Text, sdv.databroker.v1.types_pb2.Datapoint]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["datapoints",b"datapoints"]) -> None: ... + datapoints: collections.abc.Mapping[builtins.str, sdv.databroker.v1.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["datapoints", b"datapoints"]) -> None: ... + global___GetDatapointsReply = GetDatapointsReply +@typing.final +class SetDatapointsRequest(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class DatapointsEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.str + @property + def value(self) -> sdv.databroker.v1.types_pb2.Datapoint: ... + def __init__( + self, + *, + key: builtins.str = ..., + value: sdv.databroker.v1.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + DATAPOINTS_FIELD_NUMBER: builtins.int + @property + def datapoints(self) -> google.protobuf.internal.containers.MessageMap[builtins.str, sdv.databroker.v1.types_pb2.Datapoint]: + """A map of data points to set""" + + def __init__( + self, + *, + datapoints: collections.abc.Mapping[builtins.str, sdv.databroker.v1.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["datapoints", b"datapoints"]) -> None: ... + +global___SetDatapointsRequest = SetDatapointsRequest + +@typing.final +class SetDatapointsReply(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final + class ErrorsEntry(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + KEY_FIELD_NUMBER: builtins.int + VALUE_FIELD_NUMBER: builtins.int + key: builtins.str + value: sdv.databroker.v1.types_pb2.DatapointError.ValueType + def __init__( + self, + *, + key: builtins.str = ..., + value: sdv.databroker.v1.types_pb2.DatapointError.ValueType = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... + + ERRORS_FIELD_NUMBER: builtins.int + @property + def errors(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, sdv.databroker.v1.types_pb2.DatapointError.ValueType]: + """A map of errors (if any)""" + + def __init__( + self, + *, + errors: collections.abc.Mapping[builtins.str, sdv.databroker.v1.types_pb2.DatapointError.ValueType] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["errors", b"errors"]) -> None: ... + +global___SetDatapointsReply = SetDatapointsReply + +@typing.final class SubscribeRequest(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + QUERY_FIELD_NUMBER: builtins.int - query: typing.Text + query: builtins.str """Subscribe to a set of data points (or expressions) described by the provided query. The query syntax is a subset of SQL and is described in more detail in the QUERY.md file. """ - - def __init__(self, + def __init__( + self, *, - query: typing.Text = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["query",b"query"]) -> None: ... + query: builtins.str = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["query", b"query"]) -> None: ... + global___SubscribeRequest = SubscribeRequest +@typing.final class SubscribeReply(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final class FieldsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int - key: typing.Text + key: builtins.str @property def value(self) -> sdv.databroker.v1.types_pb2.Datapoint: ... - def __init__(self, + def __init__( + self, *, - key: typing.Text = ..., - value: typing.Optional[sdv.databroker.v1.types_pb2.Datapoint] = ..., - ) -> None: ... - def HasField(self, field_name: typing_extensions.Literal["value",b"value"]) -> builtins.bool: ... - def ClearField(self, field_name: typing_extensions.Literal["key",b"key","value",b"value"]) -> None: ... + key: builtins.str = ..., + value: sdv.databroker.v1.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... FIELDS_FIELD_NUMBER: builtins.int @property - def fields(self) -> google.protobuf.internal.containers.MessageMap[typing.Text, sdv.databroker.v1.types_pb2.Datapoint]: + def fields(self) -> google.protobuf.internal.containers.MessageMap[builtins.str, sdv.databroker.v1.types_pb2.Datapoint]: """Contains the fields specified by the query. If a requested data point value is not available, the corresponding Datapoint will have it's respective failure value set. """ - pass - def __init__(self, + + def __init__( + self, *, - fields: typing.Optional[typing.Mapping[typing.Text, sdv.databroker.v1.types_pb2.Datapoint]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["fields",b"fields"]) -> None: ... + fields: collections.abc.Mapping[builtins.str, sdv.databroker.v1.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["fields", b"fields"]) -> None: ... + global___SubscribeReply = SubscribeReply +@typing.final class GetMetadataRequest(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + NAMES_FIELD_NUMBER: builtins.int @property - def names(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[typing.Text]: + def names(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: """Request metadata for a list of data points referenced by their names. e.g. "Vehicle.Cabin.Seat.Row1.Pos1.Position" or "Vehicle.Speed". If no names are provided, metadata for all known data points will be returned. """ - pass - def __init__(self, + + def __init__( + self, *, - names: typing.Optional[typing.Iterable[typing.Text]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["names",b"names"]) -> None: ... + names: collections.abc.Iterable[builtins.str] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["names", b"names"]) -> None: ... + global___GetMetadataRequest = GetMetadataRequest +@typing.final class GetMetadataReply(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + LIST_FIELD_NUMBER: builtins.int @property def list(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[sdv.databroker.v1.types_pb2.Metadata]: @@ -135,10 +258,12 @@ class GetMetadataReply(google.protobuf.message.Message): doesn't exist (i.e. not known to the Data Broker) the corresponding Metadata isn't part of the returned list. """ - pass - def __init__(self, + + def __init__( + self, *, - list: typing.Optional[typing.Iterable[sdv.databroker.v1.types_pb2.Metadata]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["list",b"list"]) -> None: ... + list: collections.abc.Iterable[sdv.databroker.v1.types_pb2.Metadata] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["list", b"list"]) -> None: ... + global___GetMetadataReply = GetMetadataReply diff --git a/integration_test/gen_proto/sdv/databroker/v1/broker_pb2_grpc.py b/integration_test/gen_proto/sdv/databroker/v1/broker_pb2_grpc.py index 5c87db18..de520847 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/broker_pb2_grpc.py +++ b/integration_test/gen_proto/sdv/databroker/v1/broker_pb2_grpc.py @@ -1,9 +1,42 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ # Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT! """Client and server classes corresponding to protobuf-defined services.""" import grpc +import warnings from gen_proto.sdv.databroker.v1 import broker_pb2 as sdv_dot_databroker_dot_v1_dot_broker__pb2 +GRPC_GENERATED_VERSION = '1.71.0' +GRPC_VERSION = grpc.__version__ +_version_not_supported = False + +try: + from grpc._utilities import first_version_is_lower + _version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION) +except ImportError: + _version_not_supported = True + +if _version_not_supported: + raise RuntimeError( + f'The grpc package installed is at version {GRPC_VERSION},' + + f' but the generated code in sdv/databroker/v1/broker_pb2_grpc.py depends on' + + f' grpcio>={GRPC_GENERATED_VERSION}.' + + f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}' + + f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.' + ) + class BrokerStub(object): """Missing associated documentation comment in .proto file.""" @@ -18,17 +51,22 @@ def __init__(self, channel): '/sdv.databroker.v1.Broker/GetDatapoints', request_serializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.GetDatapointsRequest.SerializeToString, response_deserializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.GetDatapointsReply.FromString, - ) + _registered_method=True) + self.SetDatapoints = channel.unary_unary( + '/sdv.databroker.v1.Broker/SetDatapoints', + request_serializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.SetDatapointsRequest.SerializeToString, + response_deserializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.SetDatapointsReply.FromString, + _registered_method=True) self.Subscribe = channel.unary_stream( '/sdv.databroker.v1.Broker/Subscribe', request_serializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.SubscribeRequest.SerializeToString, response_deserializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.SubscribeReply.FromString, - ) + _registered_method=True) self.GetMetadata = channel.unary_unary( '/sdv.databroker.v1.Broker/GetMetadata', request_serializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.GetMetadataRequest.SerializeToString, response_deserializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.GetMetadataReply.FromString, - ) + _registered_method=True) class BrokerServicer(object): @@ -45,6 +83,13 @@ def GetDatapoints(self, request, context): context.set_details('Method not implemented!') raise NotImplementedError('Method not implemented!') + def SetDatapoints(self, request, context): + """Set a datapoint (values) + """ + context.set_code(grpc.StatusCode.UNIMPLEMENTED) + context.set_details('Method not implemented!') + raise NotImplementedError('Method not implemented!') + def Subscribe(self, request, context): """Subscribe to a set of data points or conditional expressions using the Data Broker Query Syntax (described in QUERY.md) @@ -74,6 +119,11 @@ def add_BrokerServicer_to_server(servicer, server): request_deserializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.GetDatapointsRequest.FromString, response_serializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.GetDatapointsReply.SerializeToString, ), + 'SetDatapoints': grpc.unary_unary_rpc_method_handler( + servicer.SetDatapoints, + request_deserializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.SetDatapointsRequest.FromString, + response_serializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.SetDatapointsReply.SerializeToString, + ), 'Subscribe': grpc.unary_stream_rpc_method_handler( servicer.Subscribe, request_deserializer=sdv_dot_databroker_dot_v1_dot_broker__pb2.SubscribeRequest.FromString, @@ -88,6 +138,7 @@ def add_BrokerServicer_to_server(servicer, server): generic_handler = grpc.method_handlers_generic_handler( 'sdv.databroker.v1.Broker', rpc_method_handlers) server.add_generic_rpc_handlers((generic_handler,)) + server.add_registered_method_handlers('sdv.databroker.v1.Broker', rpc_method_handlers) # This class is part of an EXPERIMENTAL API. @@ -105,11 +156,48 @@ def GetDatapoints(request, wait_for_ready=None, timeout=None, metadata=None): - return grpc.experimental.unary_unary(request, target, '/sdv.databroker.v1.Broker/GetDatapoints', + return grpc.experimental.unary_unary( + request, + target, + '/sdv.databroker.v1.Broker/GetDatapoints', sdv_dot_databroker_dot_v1_dot_broker__pb2.GetDatapointsRequest.SerializeToString, sdv_dot_databroker_dot_v1_dot_broker__pb2.GetDatapointsReply.FromString, - options, channel_credentials, - insecure, call_credentials, compression, wait_for_ready, timeout, metadata) + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) + + @staticmethod + def SetDatapoints(request, + target, + options=(), + channel_credentials=None, + call_credentials=None, + insecure=False, + compression=None, + wait_for_ready=None, + timeout=None, + metadata=None): + return grpc.experimental.unary_unary( + request, + target, + '/sdv.databroker.v1.Broker/SetDatapoints', + sdv_dot_databroker_dot_v1_dot_broker__pb2.SetDatapointsRequest.SerializeToString, + sdv_dot_databroker_dot_v1_dot_broker__pb2.SetDatapointsReply.FromString, + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) @staticmethod def Subscribe(request, @@ -122,11 +210,21 @@ def Subscribe(request, wait_for_ready=None, timeout=None, metadata=None): - return grpc.experimental.unary_stream(request, target, '/sdv.databroker.v1.Broker/Subscribe', + return grpc.experimental.unary_stream( + request, + target, + '/sdv.databroker.v1.Broker/Subscribe', sdv_dot_databroker_dot_v1_dot_broker__pb2.SubscribeRequest.SerializeToString, sdv_dot_databroker_dot_v1_dot_broker__pb2.SubscribeReply.FromString, - options, channel_credentials, - insecure, call_credentials, compression, wait_for_ready, timeout, metadata) + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) @staticmethod def GetMetadata(request, @@ -139,8 +237,18 @@ def GetMetadata(request, wait_for_ready=None, timeout=None, metadata=None): - return grpc.experimental.unary_unary(request, target, '/sdv.databroker.v1.Broker/GetMetadata', + return grpc.experimental.unary_unary( + request, + target, + '/sdv.databroker.v1.Broker/GetMetadata', sdv_dot_databroker_dot_v1_dot_broker__pb2.GetMetadataRequest.SerializeToString, sdv_dot_databroker_dot_v1_dot_broker__pb2.GetMetadataReply.FromString, - options, channel_credentials, - insecure, call_credentials, compression, wait_for_ready, timeout, metadata) + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) diff --git a/integration_test/gen_proto/sdv/databroker/v1/collector_pb2.py b/integration_test/gen_proto/sdv/databroker/v1/collector_pb2.py index 25d69dd4..c5d0b34d 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/collector_pb2.py +++ b/integration_test/gen_proto/sdv/databroker/v1/collector_pb2.py @@ -1,12 +1,35 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ # -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! +# NO CHECKED-IN PROTOBUF GENCODE # source: sdv/databroker/v1/collector.proto +# Protobuf Python Version: 5.29.0 """Generated protocol buffer code.""" from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool -from google.protobuf import message as _message -from google.protobuf import reflection as _reflection +from google.protobuf import runtime_version as _runtime_version from google.protobuf import symbol_database as _symbol_database +from google.protobuf.internal import builder as _builder +_runtime_version.ValidateProtobufRuntimeVersion( + _runtime_version.Domain.PUBLIC, + 5, + 29, + 0, + '', + 'sdv/databroker/v1/collector.proto' +) # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() @@ -17,147 +40,45 @@ DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n!sdv/databroker/v1/collector.proto\x12\x11sdv.databroker.v1\x1a\x1dsdv/databroker/v1/types.proto\"\xba\x01\n\x17UpdateDatapointsRequest\x12N\n\ndatapoints\x18\x01 \x03(\x0b\x32:.sdv.databroker.v1.UpdateDatapointsRequest.DatapointsEntry\x1aO\n\x0f\x44\x61tapointsEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12+\n\x05value\x18\x02 \x01(\x0b\x32\x1c.sdv.databroker.v1.Datapoint:\x02\x38\x01\"\xaf\x01\n\x15UpdateDatapointsReply\x12\x44\n\x06\x65rrors\x18\x01 \x03(\x0b\x32\x34.sdv.databroker.v1.UpdateDatapointsReply.ErrorsEntry\x1aP\n\x0b\x45rrorsEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12\x30\n\x05value\x18\x02 \x01(\x0e\x32!.sdv.databroker.v1.DatapointError:\x02\x38\x01\"\xba\x01\n\x17StreamDatapointsRequest\x12N\n\ndatapoints\x18\x01 \x03(\x0b\x32:.sdv.databroker.v1.StreamDatapointsRequest.DatapointsEntry\x1aO\n\x0f\x44\x61tapointsEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12+\n\x05value\x18\x02 \x01(\x0b\x32\x1c.sdv.databroker.v1.Datapoint:\x02\x38\x01\"\xaf\x01\n\x15StreamDatapointsReply\x12\x44\n\x06\x65rrors\x18\x01 \x03(\x0b\x32\x34.sdv.databroker.v1.StreamDatapointsReply.ErrorsEntry\x1aP\n\x0b\x45rrorsEntry\x12\x0b\n\x03key\x18\x01 \x01(\x05\x12\x30\n\x05value\x18\x02 \x01(\x0e\x32!.sdv.databroker.v1.DatapointError:\x02\x38\x01\"R\n\x19RegisterDatapointsRequest\x12\x35\n\x04list\x18\x01 \x03(\x0b\x32\'.sdv.databroker.v1.RegistrationMetadata\"\x9d\x01\n\x14RegistrationMetadata\x12\x0c\n\x04name\x18\x01 \x01(\t\x12.\n\tdata_type\x18\x02 \x01(\x0e\x32\x1b.sdv.databroker.v1.DataType\x12\x13\n\x0b\x64\x65scription\x18\x03 \x01(\t\x12\x32\n\x0b\x63hange_type\x18\x04 \x01(\x0e\x32\x1d.sdv.databroker.v1.ChangeType\"\x93\x01\n\x17RegisterDatapointsReply\x12H\n\x07results\x18\x01 \x03(\x0b\x32\x37.sdv.databroker.v1.RegisterDatapointsReply.ResultsEntry\x1a.\n\x0cResultsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\x05:\x02\x38\x01\x32\xd3\x02\n\tCollector\x12n\n\x12RegisterDatapoints\x12,.sdv.databroker.v1.RegisterDatapointsRequest\x1a*.sdv.databroker.v1.RegisterDatapointsReply\x12h\n\x10UpdateDatapoints\x12*.sdv.databroker.v1.UpdateDatapointsRequest\x1a(.sdv.databroker.v1.UpdateDatapointsReply\x12l\n\x10StreamDatapoints\x12*.sdv.databroker.v1.StreamDatapointsRequest\x1a(.sdv.databroker.v1.StreamDatapointsReply(\x01\x30\x01\x62\x06proto3') - - -_UPDATEDATAPOINTSREQUEST = DESCRIPTOR.message_types_by_name['UpdateDatapointsRequest'] -_UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY = _UPDATEDATAPOINTSREQUEST.nested_types_by_name['DatapointsEntry'] -_UPDATEDATAPOINTSREPLY = DESCRIPTOR.message_types_by_name['UpdateDatapointsReply'] -_UPDATEDATAPOINTSREPLY_ERRORSENTRY = _UPDATEDATAPOINTSREPLY.nested_types_by_name['ErrorsEntry'] -_STREAMDATAPOINTSREQUEST = DESCRIPTOR.message_types_by_name['StreamDatapointsRequest'] -_STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY = _STREAMDATAPOINTSREQUEST.nested_types_by_name['DatapointsEntry'] -_STREAMDATAPOINTSREPLY = DESCRIPTOR.message_types_by_name['StreamDatapointsReply'] -_STREAMDATAPOINTSREPLY_ERRORSENTRY = _STREAMDATAPOINTSREPLY.nested_types_by_name['ErrorsEntry'] -_REGISTERDATAPOINTSREQUEST = DESCRIPTOR.message_types_by_name['RegisterDatapointsRequest'] -_REGISTRATIONMETADATA = DESCRIPTOR.message_types_by_name['RegistrationMetadata'] -_REGISTERDATAPOINTSREPLY = DESCRIPTOR.message_types_by_name['RegisterDatapointsReply'] -_REGISTERDATAPOINTSREPLY_RESULTSENTRY = _REGISTERDATAPOINTSREPLY.nested_types_by_name['ResultsEntry'] -UpdateDatapointsRequest = _reflection.GeneratedProtocolMessageType('UpdateDatapointsRequest', (_message.Message,), { - - 'DatapointsEntry' : _reflection.GeneratedProtocolMessageType('DatapointsEntry', (_message.Message,), { - 'DESCRIPTOR' : _UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.UpdateDatapointsRequest.DatapointsEntry) - }) - , - 'DESCRIPTOR' : _UPDATEDATAPOINTSREQUEST, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.UpdateDatapointsRequest) - }) -_sym_db.RegisterMessage(UpdateDatapointsRequest) -_sym_db.RegisterMessage(UpdateDatapointsRequest.DatapointsEntry) - -UpdateDatapointsReply = _reflection.GeneratedProtocolMessageType('UpdateDatapointsReply', (_message.Message,), { - - 'ErrorsEntry' : _reflection.GeneratedProtocolMessageType('ErrorsEntry', (_message.Message,), { - 'DESCRIPTOR' : _UPDATEDATAPOINTSREPLY_ERRORSENTRY, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.UpdateDatapointsReply.ErrorsEntry) - }) - , - 'DESCRIPTOR' : _UPDATEDATAPOINTSREPLY, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.UpdateDatapointsReply) - }) -_sym_db.RegisterMessage(UpdateDatapointsReply) -_sym_db.RegisterMessage(UpdateDatapointsReply.ErrorsEntry) - -StreamDatapointsRequest = _reflection.GeneratedProtocolMessageType('StreamDatapointsRequest', (_message.Message,), { - - 'DatapointsEntry' : _reflection.GeneratedProtocolMessageType('DatapointsEntry', (_message.Message,), { - 'DESCRIPTOR' : _STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.StreamDatapointsRequest.DatapointsEntry) - }) - , - 'DESCRIPTOR' : _STREAMDATAPOINTSREQUEST, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.StreamDatapointsRequest) - }) -_sym_db.RegisterMessage(StreamDatapointsRequest) -_sym_db.RegisterMessage(StreamDatapointsRequest.DatapointsEntry) - -StreamDatapointsReply = _reflection.GeneratedProtocolMessageType('StreamDatapointsReply', (_message.Message,), { - - 'ErrorsEntry' : _reflection.GeneratedProtocolMessageType('ErrorsEntry', (_message.Message,), { - 'DESCRIPTOR' : _STREAMDATAPOINTSREPLY_ERRORSENTRY, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.StreamDatapointsReply.ErrorsEntry) - }) - , - 'DESCRIPTOR' : _STREAMDATAPOINTSREPLY, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.StreamDatapointsReply) - }) -_sym_db.RegisterMessage(StreamDatapointsReply) -_sym_db.RegisterMessage(StreamDatapointsReply.ErrorsEntry) - -RegisterDatapointsRequest = _reflection.GeneratedProtocolMessageType('RegisterDatapointsRequest', (_message.Message,), { - 'DESCRIPTOR' : _REGISTERDATAPOINTSREQUEST, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.RegisterDatapointsRequest) - }) -_sym_db.RegisterMessage(RegisterDatapointsRequest) - -RegistrationMetadata = _reflection.GeneratedProtocolMessageType('RegistrationMetadata', (_message.Message,), { - 'DESCRIPTOR' : _REGISTRATIONMETADATA, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.RegistrationMetadata) - }) -_sym_db.RegisterMessage(RegistrationMetadata) - -RegisterDatapointsReply = _reflection.GeneratedProtocolMessageType('RegisterDatapointsReply', (_message.Message,), { - - 'ResultsEntry' : _reflection.GeneratedProtocolMessageType('ResultsEntry', (_message.Message,), { - 'DESCRIPTOR' : _REGISTERDATAPOINTSREPLY_RESULTSENTRY, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.RegisterDatapointsReply.ResultsEntry) - }) - , - 'DESCRIPTOR' : _REGISTERDATAPOINTSREPLY, - '__module__' : 'sdv.databroker.v1.collector_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.RegisterDatapointsReply) - }) -_sym_db.RegisterMessage(RegisterDatapointsReply) -_sym_db.RegisterMessage(RegisterDatapointsReply.ResultsEntry) - -_COLLECTOR = DESCRIPTOR.services_by_name['Collector'] -if _descriptor._USE_C_DESCRIPTORS == False: - - DESCRIPTOR._options = None - _UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY._options = None - _UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY._serialized_options = b'8\001' - _UPDATEDATAPOINTSREPLY_ERRORSENTRY._options = None - _UPDATEDATAPOINTSREPLY_ERRORSENTRY._serialized_options = b'8\001' - _STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY._options = None - _STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY._serialized_options = b'8\001' - _STREAMDATAPOINTSREPLY_ERRORSENTRY._options = None - _STREAMDATAPOINTSREPLY_ERRORSENTRY._serialized_options = b'8\001' - _REGISTERDATAPOINTSREPLY_RESULTSENTRY._options = None - _REGISTERDATAPOINTSREPLY_RESULTSENTRY._serialized_options = b'8\001' - _UPDATEDATAPOINTSREQUEST._serialized_start=88 - _UPDATEDATAPOINTSREQUEST._serialized_end=274 - _UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY._serialized_start=195 - _UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY._serialized_end=274 - _UPDATEDATAPOINTSREPLY._serialized_start=277 - _UPDATEDATAPOINTSREPLY._serialized_end=452 - _UPDATEDATAPOINTSREPLY_ERRORSENTRY._serialized_start=372 - _UPDATEDATAPOINTSREPLY_ERRORSENTRY._serialized_end=452 - _STREAMDATAPOINTSREQUEST._serialized_start=455 - _STREAMDATAPOINTSREQUEST._serialized_end=641 - _STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY._serialized_start=195 - _STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY._serialized_end=274 - _STREAMDATAPOINTSREPLY._serialized_start=644 - _STREAMDATAPOINTSREPLY._serialized_end=819 - _STREAMDATAPOINTSREPLY_ERRORSENTRY._serialized_start=372 - _STREAMDATAPOINTSREPLY_ERRORSENTRY._serialized_end=452 - _REGISTERDATAPOINTSREQUEST._serialized_start=821 - _REGISTERDATAPOINTSREQUEST._serialized_end=903 - _REGISTRATIONMETADATA._serialized_start=906 - _REGISTRATIONMETADATA._serialized_end=1063 - _REGISTERDATAPOINTSREPLY._serialized_start=1066 - _REGISTERDATAPOINTSREPLY._serialized_end=1213 - _REGISTERDATAPOINTSREPLY_RESULTSENTRY._serialized_start=1167 - _REGISTERDATAPOINTSREPLY_RESULTSENTRY._serialized_end=1213 - _COLLECTOR._serialized_start=1216 - _COLLECTOR._serialized_end=1555 +_globals = globals() +_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) +_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'sdv.databroker.v1.collector_pb2', _globals) +if not _descriptor._USE_C_DESCRIPTORS: + DESCRIPTOR._loaded_options = None + _globals['_UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY']._loaded_options = None + _globals['_UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_options = b'8\001' + _globals['_UPDATEDATAPOINTSREPLY_ERRORSENTRY']._loaded_options = None + _globals['_UPDATEDATAPOINTSREPLY_ERRORSENTRY']._serialized_options = b'8\001' + _globals['_STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY']._loaded_options = None + _globals['_STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_options = b'8\001' + _globals['_STREAMDATAPOINTSREPLY_ERRORSENTRY']._loaded_options = None + _globals['_STREAMDATAPOINTSREPLY_ERRORSENTRY']._serialized_options = b'8\001' + _globals['_REGISTERDATAPOINTSREPLY_RESULTSENTRY']._loaded_options = None + _globals['_REGISTERDATAPOINTSREPLY_RESULTSENTRY']._serialized_options = b'8\001' + _globals['_UPDATEDATAPOINTSREQUEST']._serialized_start=88 + _globals['_UPDATEDATAPOINTSREQUEST']._serialized_end=274 + _globals['_UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_start=195 + _globals['_UPDATEDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_end=274 + _globals['_UPDATEDATAPOINTSREPLY']._serialized_start=277 + _globals['_UPDATEDATAPOINTSREPLY']._serialized_end=452 + _globals['_UPDATEDATAPOINTSREPLY_ERRORSENTRY']._serialized_start=372 + _globals['_UPDATEDATAPOINTSREPLY_ERRORSENTRY']._serialized_end=452 + _globals['_STREAMDATAPOINTSREQUEST']._serialized_start=455 + _globals['_STREAMDATAPOINTSREQUEST']._serialized_end=641 + _globals['_STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_start=195 + _globals['_STREAMDATAPOINTSREQUEST_DATAPOINTSENTRY']._serialized_end=274 + _globals['_STREAMDATAPOINTSREPLY']._serialized_start=644 + _globals['_STREAMDATAPOINTSREPLY']._serialized_end=819 + _globals['_STREAMDATAPOINTSREPLY_ERRORSENTRY']._serialized_start=372 + _globals['_STREAMDATAPOINTSREPLY_ERRORSENTRY']._serialized_end=452 + _globals['_REGISTERDATAPOINTSREQUEST']._serialized_start=821 + _globals['_REGISTERDATAPOINTSREQUEST']._serialized_end=903 + _globals['_REGISTRATIONMETADATA']._serialized_start=906 + _globals['_REGISTRATIONMETADATA']._serialized_end=1063 + _globals['_REGISTERDATAPOINTSREPLY']._serialized_start=1066 + _globals['_REGISTERDATAPOINTSREPLY']._serialized_end=1213 + _globals['_REGISTERDATAPOINTSREPLY_RESULTSENTRY']._serialized_start=1167 + _globals['_REGISTERDATAPOINTSREPLY_RESULTSENTRY']._serialized_end=1213 + _globals['_COLLECTOR']._serialized_start=1216 + _globals['_COLLECTOR']._serialized_end=1555 # @@protoc_insertion_point(module_scope) diff --git a/integration_test/gen_proto/sdv/databroker/v1/collector_pb2.pyi b/integration_test/gen_proto/sdv/databroker/v1/collector_pb2.pyi index 87265140..7f0dc30d 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/collector_pb2.pyi +++ b/integration_test/gen_proto/sdv/databroker/v1/collector_pb2.pyi @@ -1,184 +1,252 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ """ @generated by mypy-protobuf. Do not edit manually! isort:skip_file +******************************************************************************* +Copyright (c) 2022 Contributors to the Eclipse Foundation + +See the NOTICE file(s) distributed with this work for additional +information regarding copyright ownership. + +This program and the accompanying materials are made available under the +terms of the Apache License 2.0 which is available at +http://www.apache.org/licenses/LICENSE-2.0 + +SPDX-License-Identifier: Apache-2.0 +****************************************************************************** """ + import builtins +import collections.abc import google.protobuf.descriptor import google.protobuf.internal.containers import google.protobuf.message import gen_proto.sdv.databroker.v1.types_pb2 import typing -import typing_extensions DESCRIPTOR: google.protobuf.descriptor.FileDescriptor +@typing.final class UpdateDatapointsRequest(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final class DatapointsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int key: builtins.int @property def value(self) -> sdv.databroker.v1.types_pb2.Datapoint: ... - def __init__(self, + def __init__( + self, *, key: builtins.int = ..., - value: typing.Optional[sdv.databroker.v1.types_pb2.Datapoint] = ..., - ) -> None: ... - def HasField(self, field_name: typing_extensions.Literal["value",b"value"]) -> builtins.bool: ... - def ClearField(self, field_name: typing_extensions.Literal["key",b"key","value",b"value"]) -> None: ... + value: sdv.databroker.v1.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... DATAPOINTS_FIELD_NUMBER: builtins.int @property def datapoints(self) -> google.protobuf.internal.containers.MessageMap[builtins.int, sdv.databroker.v1.types_pb2.Datapoint]: ... - def __init__(self, + def __init__( + self, *, - datapoints: typing.Optional[typing.Mapping[builtins.int, sdv.databroker.v1.types_pb2.Datapoint]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["datapoints",b"datapoints"]) -> None: ... + datapoints: collections.abc.Mapping[builtins.int, sdv.databroker.v1.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["datapoints", b"datapoints"]) -> None: ... + global___UpdateDatapointsRequest = UpdateDatapointsRequest +@typing.final class UpdateDatapointsReply(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final class ErrorsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int key: builtins.int value: sdv.databroker.v1.types_pb2.DatapointError.ValueType - def __init__(self, + def __init__( + self, *, key: builtins.int = ..., value: sdv.databroker.v1.types_pb2.DatapointError.ValueType = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["key",b"key","value",b"value"]) -> None: ... + ) -> None: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... ERRORS_FIELD_NUMBER: builtins.int @property def errors(self) -> google.protobuf.internal.containers.ScalarMap[builtins.int, sdv.databroker.v1.types_pb2.DatapointError.ValueType]: """If empty, everything went well""" - pass - def __init__(self, + + def __init__( + self, *, - errors: typing.Optional[typing.Mapping[builtins.int, sdv.databroker.v1.types_pb2.DatapointError.ValueType]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["errors",b"errors"]) -> None: ... + errors: collections.abc.Mapping[builtins.int, sdv.databroker.v1.types_pb2.DatapointError.ValueType] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["errors", b"errors"]) -> None: ... + global___UpdateDatapointsReply = UpdateDatapointsReply +@typing.final class StreamDatapointsRequest(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final class DatapointsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int key: builtins.int @property def value(self) -> sdv.databroker.v1.types_pb2.Datapoint: ... - def __init__(self, + def __init__( + self, *, key: builtins.int = ..., - value: typing.Optional[sdv.databroker.v1.types_pb2.Datapoint] = ..., - ) -> None: ... - def HasField(self, field_name: typing_extensions.Literal["value",b"value"]) -> builtins.bool: ... - def ClearField(self, field_name: typing_extensions.Literal["key",b"key","value",b"value"]) -> None: ... + value: sdv.databroker.v1.types_pb2.Datapoint | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... DATAPOINTS_FIELD_NUMBER: builtins.int @property def datapoints(self) -> google.protobuf.internal.containers.MessageMap[builtins.int, sdv.databroker.v1.types_pb2.Datapoint]: ... - def __init__(self, + def __init__( + self, *, - datapoints: typing.Optional[typing.Mapping[builtins.int, sdv.databroker.v1.types_pb2.Datapoint]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["datapoints",b"datapoints"]) -> None: ... + datapoints: collections.abc.Mapping[builtins.int, sdv.databroker.v1.types_pb2.Datapoint] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["datapoints", b"datapoints"]) -> None: ... + global___StreamDatapointsRequest = StreamDatapointsRequest +@typing.final class StreamDatapointsReply(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final class ErrorsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int key: builtins.int value: sdv.databroker.v1.types_pb2.DatapointError.ValueType - def __init__(self, + def __init__( + self, *, key: builtins.int = ..., value: sdv.databroker.v1.types_pb2.DatapointError.ValueType = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["key",b"key","value",b"value"]) -> None: ... + ) -> None: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... ERRORS_FIELD_NUMBER: builtins.int @property def errors(self) -> google.protobuf.internal.containers.ScalarMap[builtins.int, sdv.databroker.v1.types_pb2.DatapointError.ValueType]: """If empty, everything went well""" - pass - def __init__(self, + + def __init__( + self, *, - errors: typing.Optional[typing.Mapping[builtins.int, sdv.databroker.v1.types_pb2.DatapointError.ValueType]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["errors",b"errors"]) -> None: ... + errors: collections.abc.Mapping[builtins.int, sdv.databroker.v1.types_pb2.DatapointError.ValueType] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["errors", b"errors"]) -> None: ... + global___StreamDatapointsReply = StreamDatapointsReply +@typing.final class RegisterDatapointsRequest(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + LIST_FIELD_NUMBER: builtins.int @property def list(self) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___RegistrationMetadata]: ... - def __init__(self, + def __init__( + self, *, - list: typing.Optional[typing.Iterable[global___RegistrationMetadata]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["list",b"list"]) -> None: ... + list: collections.abc.Iterable[global___RegistrationMetadata] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["list", b"list"]) -> None: ... + global___RegisterDatapointsRequest = RegisterDatapointsRequest +@typing.final class RegistrationMetadata(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + NAME_FIELD_NUMBER: builtins.int DATA_TYPE_FIELD_NUMBER: builtins.int DESCRIPTION_FIELD_NUMBER: builtins.int CHANGE_TYPE_FIELD_NUMBER: builtins.int - name: typing.Text + name: builtins.str """Name of the data point (e.g. "Vehicle.Cabin.Seat.Row1.Pos1.Position" or "Vehicle.Speed") """ - data_type: sdv.databroker.v1.types_pb2.DataType.ValueType - description: typing.Text + description: builtins.str change_type: sdv.databroker.v1.types_pb2.ChangeType.ValueType - def __init__(self, + def __init__( + self, *, - name: typing.Text = ..., + name: builtins.str = ..., data_type: sdv.databroker.v1.types_pb2.DataType.ValueType = ..., - description: typing.Text = ..., + description: builtins.str = ..., change_type: sdv.databroker.v1.types_pb2.ChangeType.ValueType = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["change_type",b"change_type","data_type",b"data_type","description",b"description","name",b"name"]) -> None: ... + ) -> None: ... + def ClearField(self, field_name: typing.Literal["change_type", b"change_type", "data_type", b"data_type", "description", b"description", "name", b"name"]) -> None: ... + global___RegistrationMetadata = RegistrationMetadata +@typing.final class RegisterDatapointsReply(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + + @typing.final class ResultsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int - key: typing.Text + key: builtins.str value: builtins.int - def __init__(self, + def __init__( + self, *, - key: typing.Text = ..., + key: builtins.str = ..., value: builtins.int = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["key",b"key","value",b"value"]) -> None: ... + ) -> None: ... + def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ... RESULTS_FIELD_NUMBER: builtins.int @property - def results(self) -> google.protobuf.internal.containers.ScalarMap[typing.Text, builtins.int]: + def results(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.int]: """Maps each data point name passed in RegisterDatapointsRequest to a data point id""" - pass - def __init__(self, + + def __init__( + self, *, - results: typing.Optional[typing.Mapping[typing.Text, builtins.int]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["results",b"results"]) -> None: ... + results: collections.abc.Mapping[builtins.str, builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["results", b"results"]) -> None: ... + global___RegisterDatapointsReply = RegisterDatapointsReply diff --git a/integration_test/gen_proto/sdv/databroker/v1/collector_pb2_grpc.py b/integration_test/gen_proto/sdv/databroker/v1/collector_pb2_grpc.py index cb73b05a..6c5d28d0 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/collector_pb2_grpc.py +++ b/integration_test/gen_proto/sdv/databroker/v1/collector_pb2_grpc.py @@ -1,9 +1,42 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ # Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT! """Client and server classes corresponding to protobuf-defined services.""" import grpc +import warnings from gen_proto.sdv.databroker.v1 import collector_pb2 as sdv_dot_databroker_dot_v1_dot_collector__pb2 +GRPC_GENERATED_VERSION = '1.71.0' +GRPC_VERSION = grpc.__version__ +_version_not_supported = False + +try: + from grpc._utilities import first_version_is_lower + _version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION) +except ImportError: + _version_not_supported = True + +if _version_not_supported: + raise RuntimeError( + f'The grpc package installed is at version {GRPC_VERSION},' + + f' but the generated code in sdv/databroker/v1/collector_pb2_grpc.py depends on' + + f' grpcio>={GRPC_GENERATED_VERSION}.' + + f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}' + + f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.' + ) + class CollectorStub(object): """Missing associated documentation comment in .proto file.""" @@ -18,17 +51,17 @@ def __init__(self, channel): '/sdv.databroker.v1.Collector/RegisterDatapoints', request_serializer=sdv_dot_databroker_dot_v1_dot_collector__pb2.RegisterDatapointsRequest.SerializeToString, response_deserializer=sdv_dot_databroker_dot_v1_dot_collector__pb2.RegisterDatapointsReply.FromString, - ) + _registered_method=True) self.UpdateDatapoints = channel.unary_unary( '/sdv.databroker.v1.Collector/UpdateDatapoints', request_serializer=sdv_dot_databroker_dot_v1_dot_collector__pb2.UpdateDatapointsRequest.SerializeToString, response_deserializer=sdv_dot_databroker_dot_v1_dot_collector__pb2.UpdateDatapointsReply.FromString, - ) + _registered_method=True) self.StreamDatapoints = channel.stream_stream( '/sdv.databroker.v1.Collector/StreamDatapoints', request_serializer=sdv_dot_databroker_dot_v1_dot_collector__pb2.StreamDatapointsRequest.SerializeToString, response_deserializer=sdv_dot_databroker_dot_v1_dot_collector__pb2.StreamDatapointsReply.FromString, - ) + _registered_method=True) class CollectorServicer(object): @@ -108,6 +141,7 @@ def add_CollectorServicer_to_server(servicer, server): generic_handler = grpc.method_handlers_generic_handler( 'sdv.databroker.v1.Collector', rpc_method_handlers) server.add_generic_rpc_handlers((generic_handler,)) + server.add_registered_method_handlers('sdv.databroker.v1.Collector', rpc_method_handlers) # This class is part of an EXPERIMENTAL API. @@ -125,11 +159,21 @@ def RegisterDatapoints(request, wait_for_ready=None, timeout=None, metadata=None): - return grpc.experimental.unary_unary(request, target, '/sdv.databroker.v1.Collector/RegisterDatapoints', + return grpc.experimental.unary_unary( + request, + target, + '/sdv.databroker.v1.Collector/RegisterDatapoints', sdv_dot_databroker_dot_v1_dot_collector__pb2.RegisterDatapointsRequest.SerializeToString, sdv_dot_databroker_dot_v1_dot_collector__pb2.RegisterDatapointsReply.FromString, - options, channel_credentials, - insecure, call_credentials, compression, wait_for_ready, timeout, metadata) + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) @staticmethod def UpdateDatapoints(request, @@ -142,11 +186,21 @@ def UpdateDatapoints(request, wait_for_ready=None, timeout=None, metadata=None): - return grpc.experimental.unary_unary(request, target, '/sdv.databroker.v1.Collector/UpdateDatapoints', + return grpc.experimental.unary_unary( + request, + target, + '/sdv.databroker.v1.Collector/UpdateDatapoints', sdv_dot_databroker_dot_v1_dot_collector__pb2.UpdateDatapointsRequest.SerializeToString, sdv_dot_databroker_dot_v1_dot_collector__pb2.UpdateDatapointsReply.FromString, - options, channel_credentials, - insecure, call_credentials, compression, wait_for_ready, timeout, metadata) + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) @staticmethod def StreamDatapoints(request_iterator, @@ -159,8 +213,18 @@ def StreamDatapoints(request_iterator, wait_for_ready=None, timeout=None, metadata=None): - return grpc.experimental.stream_stream(request_iterator, target, '/sdv.databroker.v1.Collector/StreamDatapoints', + return grpc.experimental.stream_stream( + request_iterator, + target, + '/sdv.databroker.v1.Collector/StreamDatapoints', sdv_dot_databroker_dot_v1_dot_collector__pb2.StreamDatapointsRequest.SerializeToString, sdv_dot_databroker_dot_v1_dot_collector__pb2.StreamDatapointsReply.FromString, - options, channel_credentials, - insecure, call_credentials, compression, wait_for_ready, timeout, metadata) + options, + channel_credentials, + insecure, + call_credentials, + compression, + wait_for_ready, + timeout, + metadata, + _registered_method=True) diff --git a/integration_test/gen_proto/sdv/databroker/v1/types_pb2.py b/integration_test/gen_proto/sdv/databroker/v1/types_pb2.py index 7cf9bc4f..7c13ab22 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/types_pb2.py +++ b/integration_test/gen_proto/sdv/databroker/v1/types_pb2.py @@ -1,13 +1,35 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ # -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! +# NO CHECKED-IN PROTOBUF GENCODE # source: sdv/databroker/v1/types.proto +# Protobuf Python Version: 5.29.0 """Generated protocol buffer code.""" -from google.protobuf.internal import enum_type_wrapper from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool -from google.protobuf import message as _message -from google.protobuf import reflection as _reflection +from google.protobuf import runtime_version as _runtime_version from google.protobuf import symbol_database as _symbol_database +from google.protobuf.internal import builder as _builder +_runtime_version.ValidateProtobufRuntimeVersion( + _runtime_version.Domain.PUBLIC, + 5, + 29, + 0, + '', + 'sdv/databroker/v1/types.proto' +) # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() @@ -16,160 +38,45 @@ from google.protobuf import timestamp_pb2 as google_dot_protobuf_dot_timestamp__pb2 -DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\x1dsdv/databroker/v1/types.proto\x12\x11sdv.databroker.v1\x1a\x1fgoogle/protobuf/timestamp.proto\"\x1d\n\x0bStringArray\x12\x0e\n\x06values\x18\x01 \x03(\t\"\x1b\n\tBoolArray\x12\x0e\n\x06values\x18\x01 \x03(\x08\"\x1c\n\nInt32Array\x12\x0e\n\x06values\x18\x01 \x03(\x11\"\x1c\n\nInt64Array\x12\x0e\n\x06values\x18\x01 \x03(\x12\"\x1d\n\x0bUint32Array\x12\x0e\n\x06values\x18\x01 \x03(\r\"\x1d\n\x0bUint64Array\x12\x0e\n\x06values\x18\x01 \x03(\x04\"\x1c\n\nFloatArray\x12\x0e\n\x06values\x18\x01 \x03(\x02\"\x1d\n\x0b\x44oubleArray\x12\x0e\n\x06values\x18\x01 \x03(\x01\"\xe2\x06\n\tDatapoint\x12-\n\ttimestamp\x18\x01 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12=\n\rfailure_value\x18\n \x01(\x0e\x32$.sdv.databroker.v1.Datapoint.FailureH\x00\x12\x16\n\x0cstring_value\x18\x0b \x01(\tH\x00\x12\x14\n\nbool_value\x18\x0c \x01(\x08H\x00\x12\x15\n\x0bint32_value\x18\r \x01(\x11H\x00\x12\x15\n\x0bint64_value\x18\x0e \x01(\x12H\x00\x12\x16\n\x0cuint32_value\x18\x0f \x01(\rH\x00\x12\x16\n\x0cuint64_value\x18\x10 \x01(\x04H\x00\x12\x15\n\x0b\x66loat_value\x18\x11 \x01(\x02H\x00\x12\x16\n\x0c\x64ouble_value\x18\x12 \x01(\x01H\x00\x12\x36\n\x0cstring_array\x18\x15 \x01(\x0b\x32\x1e.sdv.databroker.v1.StringArrayH\x00\x12\x32\n\nbool_array\x18\x16 \x01(\x0b\x32\x1c.sdv.databroker.v1.BoolArrayH\x00\x12\x34\n\x0bint32_array\x18\x17 \x01(\x0b\x32\x1d.sdv.databroker.v1.Int32ArrayH\x00\x12\x34\n\x0bint64_array\x18\x18 \x01(\x0b\x32\x1d.sdv.databroker.v1.Int64ArrayH\x00\x12\x36\n\x0cuint32_array\x18\x19 \x01(\x0b\x32\x1e.sdv.databroker.v1.Uint32ArrayH\x00\x12\x36\n\x0cuint64_array\x18\x1a \x01(\x0b\x32\x1e.sdv.databroker.v1.Uint64ArrayH\x00\x12\x34\n\x0b\x66loat_array\x18\x1b \x01(\x0b\x32\x1d.sdv.databroker.v1.FloatArrayH\x00\x12\x36\n\x0c\x64ouble_array\x18\x1c \x01(\x0b\x32\x1e.sdv.databroker.v1.DoubleArrayH\x00\"m\n\x07\x46\x61ilure\x12\x11\n\rINVALID_VALUE\x10\x00\x12\x11\n\rNOT_AVAILABLE\x10\x01\x12\x15\n\x11UNKNOWN_DATAPOINT\x10\x02\x12\x11\n\rACCESS_DENIED\x10\x03\x12\x12\n\x0eINTERNAL_ERROR\x10\x04\x42\x07\n\x05value\"\x9d\x01\n\x08Metadata\x12\n\n\x02id\x18\x01 \x01(\x05\x12\x0c\n\x04name\x18\x04 \x01(\t\x12.\n\tdata_type\x18\x05 \x01(\x0e\x32\x1b.sdv.databroker.v1.DataType\x12\x32\n\x0b\x63hange_type\x18\x06 \x01(\x0e\x32\x1d.sdv.databroker.v1.ChangeType\x12\x13\n\x0b\x64\x65scription\x18\x07 \x01(\t*\x84\x03\n\x08\x44\x61taType\x12\n\n\x06STRING\x10\x00\x12\x08\n\x04\x42OOL\x10\x01\x12\x08\n\x04INT8\x10\x02\x12\t\n\x05INT16\x10\x03\x12\t\n\x05INT32\x10\x04\x12\t\n\x05INT64\x10\x05\x12\t\n\x05UINT8\x10\x06\x12\n\n\x06UINT16\x10\x07\x12\n\n\x06UINT32\x10\x08\x12\n\n\x06UINT64\x10\t\x12\t\n\x05\x46LOAT\x10\n\x12\n\n\x06\x44OUBLE\x10\x0b\x12\r\n\tTIMESTAMP\x10\x0c\x12\x10\n\x0cSTRING_ARRAY\x10\x14\x12\x0e\n\nBOOL_ARRAY\x10\x15\x12\x0e\n\nINT8_ARRAY\x10\x16\x12\x0f\n\x0bINT16_ARRAY\x10\x17\x12\x0f\n\x0bINT32_ARRAY\x10\x18\x12\x0f\n\x0bINT64_ARRAY\x10\x19\x12\x0f\n\x0bUINT8_ARRAY\x10\x1a\x12\x10\n\x0cUINT16_ARRAY\x10\x1b\x12\x10\n\x0cUINT32_ARRAY\x10\x1c\x12\x10\n\x0cUINT64_ARRAY\x10\x1d\x12\x0f\n\x0b\x46LOAT_ARRAY\x10\x1e\x12\x10\n\x0c\x44OUBLE_ARRAY\x10\x1f\x12\x13\n\x0fTIMESTAMP_ARRAY\x10 *s\n\x0e\x44\x61tapointError\x12\x15\n\x11UNKNOWN_DATAPOINT\x10\x00\x12\x10\n\x0cINVALID_TYPE\x10\x01\x12\x11\n\rACCESS_DENIED\x10\x02\x12\x12\n\x0eINTERNAL_ERROR\x10\x03\x12\x11\n\rOUT_OF_BOUNDS\x10\x04*7\n\nChangeType\x12\n\n\x06STATIC\x10\x00\x12\r\n\tON_CHANGE\x10\x01\x12\x0e\n\nCONTINUOUS\x10\x02\x62\x06proto3') - -_DATATYPE = DESCRIPTOR.enum_types_by_name['DataType'] -DataType = enum_type_wrapper.EnumTypeWrapper(_DATATYPE) -_DATAPOINTERROR = DESCRIPTOR.enum_types_by_name['DatapointError'] -DatapointError = enum_type_wrapper.EnumTypeWrapper(_DATAPOINTERROR) -_CHANGETYPE = DESCRIPTOR.enum_types_by_name['ChangeType'] -ChangeType = enum_type_wrapper.EnumTypeWrapper(_CHANGETYPE) -STRING = 0 -BOOL = 1 -INT8 = 2 -INT16 = 3 -INT32 = 4 -INT64 = 5 -UINT8 = 6 -UINT16 = 7 -UINT32 = 8 -UINT64 = 9 -FLOAT = 10 -DOUBLE = 11 -TIMESTAMP = 12 -STRING_ARRAY = 20 -BOOL_ARRAY = 21 -INT8_ARRAY = 22 -INT16_ARRAY = 23 -INT32_ARRAY = 24 -INT64_ARRAY = 25 -UINT8_ARRAY = 26 -UINT16_ARRAY = 27 -UINT32_ARRAY = 28 -UINT64_ARRAY = 29 -FLOAT_ARRAY = 30 -DOUBLE_ARRAY = 31 -TIMESTAMP_ARRAY = 32 -UNKNOWN_DATAPOINT = 0 -INVALID_TYPE = 1 -ACCESS_DENIED = 2 -INTERNAL_ERROR = 3 -OUT_OF_BOUNDS = 4 -STATIC = 0 -ON_CHANGE = 1 -CONTINUOUS = 2 - - -_STRINGARRAY = DESCRIPTOR.message_types_by_name['StringArray'] -_BOOLARRAY = DESCRIPTOR.message_types_by_name['BoolArray'] -_INT32ARRAY = DESCRIPTOR.message_types_by_name['Int32Array'] -_INT64ARRAY = DESCRIPTOR.message_types_by_name['Int64Array'] -_UINT32ARRAY = DESCRIPTOR.message_types_by_name['Uint32Array'] -_UINT64ARRAY = DESCRIPTOR.message_types_by_name['Uint64Array'] -_FLOATARRAY = DESCRIPTOR.message_types_by_name['FloatArray'] -_DOUBLEARRAY = DESCRIPTOR.message_types_by_name['DoubleArray'] -_DATAPOINT = DESCRIPTOR.message_types_by_name['Datapoint'] -_METADATA = DESCRIPTOR.message_types_by_name['Metadata'] -_DATAPOINT_FAILURE = _DATAPOINT.enum_types_by_name['Failure'] -StringArray = _reflection.GeneratedProtocolMessageType('StringArray', (_message.Message,), { - 'DESCRIPTOR' : _STRINGARRAY, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.StringArray) - }) -_sym_db.RegisterMessage(StringArray) - -BoolArray = _reflection.GeneratedProtocolMessageType('BoolArray', (_message.Message,), { - 'DESCRIPTOR' : _BOOLARRAY, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.BoolArray) - }) -_sym_db.RegisterMessage(BoolArray) - -Int32Array = _reflection.GeneratedProtocolMessageType('Int32Array', (_message.Message,), { - 'DESCRIPTOR' : _INT32ARRAY, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.Int32Array) - }) -_sym_db.RegisterMessage(Int32Array) - -Int64Array = _reflection.GeneratedProtocolMessageType('Int64Array', (_message.Message,), { - 'DESCRIPTOR' : _INT64ARRAY, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.Int64Array) - }) -_sym_db.RegisterMessage(Int64Array) - -Uint32Array = _reflection.GeneratedProtocolMessageType('Uint32Array', (_message.Message,), { - 'DESCRIPTOR' : _UINT32ARRAY, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.Uint32Array) - }) -_sym_db.RegisterMessage(Uint32Array) - -Uint64Array = _reflection.GeneratedProtocolMessageType('Uint64Array', (_message.Message,), { - 'DESCRIPTOR' : _UINT64ARRAY, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.Uint64Array) - }) -_sym_db.RegisterMessage(Uint64Array) - -FloatArray = _reflection.GeneratedProtocolMessageType('FloatArray', (_message.Message,), { - 'DESCRIPTOR' : _FLOATARRAY, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.FloatArray) - }) -_sym_db.RegisterMessage(FloatArray) - -DoubleArray = _reflection.GeneratedProtocolMessageType('DoubleArray', (_message.Message,), { - 'DESCRIPTOR' : _DOUBLEARRAY, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.DoubleArray) - }) -_sym_db.RegisterMessage(DoubleArray) - -Datapoint = _reflection.GeneratedProtocolMessageType('Datapoint', (_message.Message,), { - 'DESCRIPTOR' : _DATAPOINT, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.Datapoint) - }) -_sym_db.RegisterMessage(Datapoint) - -Metadata = _reflection.GeneratedProtocolMessageType('Metadata', (_message.Message,), { - 'DESCRIPTOR' : _METADATA, - '__module__' : 'sdv.databroker.v1.types_pb2' - # @@protoc_insertion_point(class_scope:sdv.databroker.v1.Metadata) - }) -_sym_db.RegisterMessage(Metadata) - -if _descriptor._USE_C_DESCRIPTORS == False: - - DESCRIPTOR._options = None - _DATATYPE._serialized_start=1358 - _DATATYPE._serialized_end=1746 - _DATAPOINTERROR._serialized_start=1748 - _DATAPOINTERROR._serialized_end=1863 - _CHANGETYPE._serialized_start=1865 - _CHANGETYPE._serialized_end=1920 - _STRINGARRAY._serialized_start=85 - _STRINGARRAY._serialized_end=114 - _BOOLARRAY._serialized_start=116 - _BOOLARRAY._serialized_end=143 - _INT32ARRAY._serialized_start=145 - _INT32ARRAY._serialized_end=173 - _INT64ARRAY._serialized_start=175 - _INT64ARRAY._serialized_end=203 - _UINT32ARRAY._serialized_start=205 - _UINT32ARRAY._serialized_end=234 - _UINT64ARRAY._serialized_start=236 - _UINT64ARRAY._serialized_end=265 - _FLOATARRAY._serialized_start=267 - _FLOATARRAY._serialized_end=295 - _DOUBLEARRAY._serialized_start=297 - _DOUBLEARRAY._serialized_end=326 - _DATAPOINT._serialized_start=329 - _DATAPOINT._serialized_end=1195 - _DATAPOINT_FAILURE._serialized_start=1077 - _DATAPOINT_FAILURE._serialized_end=1186 - _METADATA._serialized_start=1198 - _METADATA._serialized_end=1355 +DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n\x1dsdv/databroker/v1/types.proto\x12\x11sdv.databroker.v1\x1a\x1fgoogle/protobuf/timestamp.proto\"\x1d\n\x0bStringArray\x12\x0e\n\x06values\x18\x01 \x03(\t\"\x1b\n\tBoolArray\x12\x0e\n\x06values\x18\x01 \x03(\x08\"\x1c\n\nInt32Array\x12\x0e\n\x06values\x18\x01 \x03(\x11\"\x1c\n\nInt64Array\x12\x0e\n\x06values\x18\x01 \x03(\x12\"\x1d\n\x0bUint32Array\x12\x0e\n\x06values\x18\x01 \x03(\r\"\x1d\n\x0bUint64Array\x12\x0e\n\x06values\x18\x01 \x03(\x04\"\x1c\n\nFloatArray\x12\x0e\n\x06values\x18\x01 \x03(\x02\"\x1d\n\x0b\x44oubleArray\x12\x0e\n\x06values\x18\x01 \x03(\x01\"\xe2\x06\n\tDatapoint\x12-\n\ttimestamp\x18\x01 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12=\n\rfailure_value\x18\n \x01(\x0e\x32$.sdv.databroker.v1.Datapoint.FailureH\x00\x12\x16\n\x0cstring_value\x18\x0b \x01(\tH\x00\x12\x14\n\nbool_value\x18\x0c \x01(\x08H\x00\x12\x15\n\x0bint32_value\x18\r \x01(\x11H\x00\x12\x15\n\x0bint64_value\x18\x0e \x01(\x12H\x00\x12\x16\n\x0cuint32_value\x18\x0f \x01(\rH\x00\x12\x16\n\x0cuint64_value\x18\x10 \x01(\x04H\x00\x12\x15\n\x0b\x66loat_value\x18\x11 \x01(\x02H\x00\x12\x16\n\x0c\x64ouble_value\x18\x12 \x01(\x01H\x00\x12\x36\n\x0cstring_array\x18\x15 \x01(\x0b\x32\x1e.sdv.databroker.v1.StringArrayH\x00\x12\x32\n\nbool_array\x18\x16 \x01(\x0b\x32\x1c.sdv.databroker.v1.BoolArrayH\x00\x12\x34\n\x0bint32_array\x18\x17 \x01(\x0b\x32\x1d.sdv.databroker.v1.Int32ArrayH\x00\x12\x34\n\x0bint64_array\x18\x18 \x01(\x0b\x32\x1d.sdv.databroker.v1.Int64ArrayH\x00\x12\x36\n\x0cuint32_array\x18\x19 \x01(\x0b\x32\x1e.sdv.databroker.v1.Uint32ArrayH\x00\x12\x36\n\x0cuint64_array\x18\x1a \x01(\x0b\x32\x1e.sdv.databroker.v1.Uint64ArrayH\x00\x12\x34\n\x0b\x66loat_array\x18\x1b \x01(\x0b\x32\x1d.sdv.databroker.v1.FloatArrayH\x00\x12\x36\n\x0c\x64ouble_array\x18\x1c \x01(\x0b\x32\x1e.sdv.databroker.v1.DoubleArrayH\x00\"m\n\x07\x46\x61ilure\x12\x11\n\rINVALID_VALUE\x10\x00\x12\x11\n\rNOT_AVAILABLE\x10\x01\x12\x15\n\x11UNKNOWN_DATAPOINT\x10\x02\x12\x11\n\rACCESS_DENIED\x10\x03\x12\x12\n\x0eINTERNAL_ERROR\x10\x04\x42\x07\n\x05value\"\xe0\x02\n\x08Metadata\x12\n\n\x02id\x18\x01 \x01(\x05\x12\x30\n\nentry_type\x18\x02 \x01(\x0e\x32\x1c.sdv.databroker.v1.EntryType\x12\x0c\n\x04name\x18\x04 \x01(\t\x12.\n\tdata_type\x18\x05 \x01(\x0e\x32\x1b.sdv.databroker.v1.DataType\x12\x32\n\x0b\x63hange_type\x18\x06 \x01(\x0e\x32\x1d.sdv.databroker.v1.ChangeType\x12\x13\n\x0b\x64\x65scription\x18\x07 \x01(\t\x12+\n\x07\x61llowed\x18\n \x01(\x0b\x32\x1a.sdv.databroker.v1.Allowed\x12\x30\n\x03min\x18\x0b \x01(\x0b\x32#.sdv.databroker.v1.ValueRestriction\x12\x30\n\x03max\x18\x0c \x01(\x0b\x32#.sdv.databroker.v1.ValueRestriction\"\x9c\x03\n\x07\x41llowed\x12\x37\n\rstring_values\x18\x01 \x01(\x0b\x32\x1e.sdv.databroker.v1.StringArrayH\x00\x12\x35\n\x0cint32_values\x18\x03 \x01(\x0b\x32\x1d.sdv.databroker.v1.Int32ArrayH\x00\x12\x35\n\x0cint64_values\x18\x04 \x01(\x0b\x32\x1d.sdv.databroker.v1.Int64ArrayH\x00\x12\x37\n\ruint32_values\x18\x05 \x01(\x0b\x32\x1e.sdv.databroker.v1.Uint32ArrayH\x00\x12\x37\n\ruint64_values\x18\x06 \x01(\x0b\x32\x1e.sdv.databroker.v1.Uint64ArrayH\x00\x12\x35\n\x0c\x66loat_values\x18\x07 \x01(\x0b\x32\x1d.sdv.databroker.v1.FloatArrayH\x00\x12\x37\n\rdouble_values\x18\x08 \x01(\x0b\x32\x1e.sdv.databroker.v1.DoubleArrayH\x00\x42\x08\n\x06values\"\xac\x01\n\x10ValueRestriction\x12\x10\n\x06string\x18\x01 \x01(\tH\x00\x12\x0e\n\x04\x62ool\x18\x02 \x01(\x08H\x00\x12\x0f\n\x05int32\x18\x03 \x01(\x11H\x00\x12\x0f\n\x05int64\x18\x04 \x01(\x12H\x00\x12\x10\n\x06uint32\x18\x05 \x01(\rH\x00\x12\x10\n\x06uint64\x18\x06 \x01(\x04H\x00\x12\x0f\n\x05\x66loat\x18\x07 \x01(\x02H\x00\x12\x10\n\x06\x64ouble\x18\x08 \x01(\x01H\x00\x42\r\n\x0btyped_value*\xe0\x02\n\x08\x44\x61taType\x12\n\n\x06STRING\x10\x00\x12\x08\n\x04\x42OOL\x10\x01\x12\x08\n\x04INT8\x10\x02\x12\t\n\x05INT16\x10\x03\x12\t\n\x05INT32\x10\x04\x12\t\n\x05INT64\x10\x05\x12\t\n\x05UINT8\x10\x06\x12\n\n\x06UINT16\x10\x07\x12\n\n\x06UINT32\x10\x08\x12\n\n\x06UINT64\x10\t\x12\t\n\x05\x46LOAT\x10\n\x12\n\n\x06\x44OUBLE\x10\x0b\x12\x10\n\x0cSTRING_ARRAY\x10\x14\x12\x0e\n\nBOOL_ARRAY\x10\x15\x12\x0e\n\nINT8_ARRAY\x10\x16\x12\x0f\n\x0bINT16_ARRAY\x10\x17\x12\x0f\n\x0bINT32_ARRAY\x10\x18\x12\x0f\n\x0bINT64_ARRAY\x10\x19\x12\x0f\n\x0bUINT8_ARRAY\x10\x1a\x12\x10\n\x0cUINT16_ARRAY\x10\x1b\x12\x10\n\x0cUINT32_ARRAY\x10\x1c\x12\x10\n\x0cUINT64_ARRAY\x10\x1d\x12\x0f\n\x0b\x46LOAT_ARRAY\x10\x1e\x12\x10\n\x0c\x44OUBLE_ARRAY\x10\x1f*s\n\x0e\x44\x61tapointError\x12\x15\n\x11UNKNOWN_DATAPOINT\x10\x00\x12\x10\n\x0cINVALID_TYPE\x10\x01\x12\x11\n\rACCESS_DENIED\x10\x02\x12\x12\n\x0eINTERNAL_ERROR\x10\x03\x12\x11\n\rOUT_OF_BOUNDS\x10\x04*q\n\tEntryType\x12\x1a\n\x16\x45NTRY_TYPE_UNSPECIFIED\x10\x00\x12\x15\n\x11\x45NTRY_TYPE_SENSOR\x10\x01\x12\x17\n\x13\x45NTRY_TYPE_ACTUATOR\x10\x02\x12\x18\n\x14\x45NTRY_TYPE_ATTRIBUTE\x10\x03*7\n\nChangeType\x12\n\n\x06STATIC\x10\x00\x12\r\n\tON_CHANGE\x10\x01\x12\x0e\n\nCONTINUOUS\x10\x02\x62\x06proto3') + +_globals = globals() +_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) +_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'sdv.databroker.v1.types_pb2', _globals) +if not _descriptor._USE_C_DESCRIPTORS: + DESCRIPTOR._loaded_options = None + _globals['_DATATYPE']._serialized_start=2143 + _globals['_DATATYPE']._serialized_end=2495 + _globals['_DATAPOINTERROR']._serialized_start=2497 + _globals['_DATAPOINTERROR']._serialized_end=2612 + _globals['_ENTRYTYPE']._serialized_start=2614 + _globals['_ENTRYTYPE']._serialized_end=2727 + _globals['_CHANGETYPE']._serialized_start=2729 + _globals['_CHANGETYPE']._serialized_end=2784 + _globals['_STRINGARRAY']._serialized_start=85 + _globals['_STRINGARRAY']._serialized_end=114 + _globals['_BOOLARRAY']._serialized_start=116 + _globals['_BOOLARRAY']._serialized_end=143 + _globals['_INT32ARRAY']._serialized_start=145 + _globals['_INT32ARRAY']._serialized_end=173 + _globals['_INT64ARRAY']._serialized_start=175 + _globals['_INT64ARRAY']._serialized_end=203 + _globals['_UINT32ARRAY']._serialized_start=205 + _globals['_UINT32ARRAY']._serialized_end=234 + _globals['_UINT64ARRAY']._serialized_start=236 + _globals['_UINT64ARRAY']._serialized_end=265 + _globals['_FLOATARRAY']._serialized_start=267 + _globals['_FLOATARRAY']._serialized_end=295 + _globals['_DOUBLEARRAY']._serialized_start=297 + _globals['_DOUBLEARRAY']._serialized_end=326 + _globals['_DATAPOINT']._serialized_start=329 + _globals['_DATAPOINT']._serialized_end=1195 + _globals['_DATAPOINT_FAILURE']._serialized_start=1077 + _globals['_DATAPOINT_FAILURE']._serialized_end=1186 + _globals['_METADATA']._serialized_start=1198 + _globals['_METADATA']._serialized_end=1550 + _globals['_ALLOWED']._serialized_start=1553 + _globals['_ALLOWED']._serialized_end=1965 + _globals['_VALUERESTRICTION']._serialized_start=1968 + _globals['_VALUERESTRICTION']._serialized_end=2140 # @@protoc_insertion_point(module_scope) diff --git a/integration_test/gen_proto/sdv/databroker/v1/types_pb2.pyi b/integration_test/gen_proto/sdv/databroker/v1/types_pb2.pyi index 735a9207..8b4220cf 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/types_pb2.pyi +++ b/integration_test/gen_proto/sdv/databroker/v1/types_pb2.pyi @@ -1,21 +1,56 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ """ @generated by mypy-protobuf. Do not edit manually! isort:skip_file +******************************************************************************* +Copyright (c) 2022 Contributors to the Eclipse Foundation + +See the NOTICE file(s) distributed with this work for additional +information regarding copyright ownership. + +This program and the accompanying materials are made available under the +terms of the Apache License 2.0 which is available at +http://www.apache.org/licenses/LICENSE-2.0 + +SPDX-License-Identifier: Apache-2.0 +****************************************************************************** + +Please do not add optional fields due to older proto3 versions limitations """ + import builtins +import collections.abc import google.protobuf.descriptor import google.protobuf.internal.containers import google.protobuf.internal.enum_type_wrapper import google.protobuf.message import google.protobuf.timestamp_pb2 +import sys import typing -import typing_extensions + +if sys.version_info >= (3, 10): + import typing as typing_extensions +else: + import typing_extensions DESCRIPTOR: google.protobuf.descriptor.FileDescriptor class _DataType: - ValueType = typing.NewType('ValueType', builtins.int) + ValueType = typing.NewType("ValueType", builtins.int) V: typing_extensions.TypeAlias = ValueType + class _DataTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_DataType.ValueType], builtins.type): DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor STRING: _DataType.ValueType # 0 @@ -30,7 +65,6 @@ class _DataTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumT UINT64: _DataType.ValueType # 9 FLOAT: _DataType.ValueType # 10 DOUBLE: _DataType.ValueType # 11 - TIMESTAMP: _DataType.ValueType # 12 STRING_ARRAY: _DataType.ValueType # 20 BOOL_ARRAY: _DataType.ValueType # 21 INT8_ARRAY: _DataType.ValueType # 22 @@ -43,14 +77,13 @@ class _DataTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumT UINT64_ARRAY: _DataType.ValueType # 29 FLOAT_ARRAY: _DataType.ValueType # 30 DOUBLE_ARRAY: _DataType.ValueType # 31 - TIMESTAMP_ARRAY: _DataType.ValueType # 32 + class DataType(_DataType, metaclass=_DataTypeEnumTypeWrapper): """Data type of a signal Protobuf doesn't support int8, int16, uint8 or uint16. These are mapped to sint32 and uint32 respectively. """ - pass STRING: DataType.ValueType # 0 BOOL: DataType.ValueType # 1 @@ -64,7 +97,6 @@ UINT32: DataType.ValueType # 8 UINT64: DataType.ValueType # 9 FLOAT: DataType.ValueType # 10 DOUBLE: DataType.ValueType # 11 -TIMESTAMP: DataType.ValueType # 12 STRING_ARRAY: DataType.ValueType # 20 BOOL_ARRAY: DataType.ValueType # 21 INT8_ARRAY: DataType.ValueType # 22 @@ -77,13 +109,12 @@ UINT32_ARRAY: DataType.ValueType # 28 UINT64_ARRAY: DataType.ValueType # 29 FLOAT_ARRAY: DataType.ValueType # 30 DOUBLE_ARRAY: DataType.ValueType # 31 -TIMESTAMP_ARRAY: DataType.ValueType # 32 global___DataType = DataType - class _DatapointError: - ValueType = typing.NewType('ValueType', builtins.int) + ValueType = typing.NewType("ValueType", builtins.int) V: typing_extensions.TypeAlias = ValueType + class _DatapointErrorEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_DatapointError.ValueType], builtins.type): DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor UNKNOWN_DATAPOINT: _DatapointError.ValueType # 0 @@ -91,8 +122,8 @@ class _DatapointErrorEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper. ACCESS_DENIED: _DatapointError.ValueType # 2 INTERNAL_ERROR: _DatapointError.ValueType # 3 OUT_OF_BOUNDS: _DatapointError.ValueType # 4 -class DatapointError(_DatapointError, metaclass=_DatapointErrorEnumTypeWrapper): - pass + +class DatapointError(_DatapointError, metaclass=_DatapointErrorEnumTypeWrapper): ... UNKNOWN_DATAPOINT: DatapointError.ValueType # 0 INVALID_TYPE: DatapointError.ValueType # 1 @@ -101,177 +132,213 @@ INTERNAL_ERROR: DatapointError.ValueType # 3 OUT_OF_BOUNDS: DatapointError.ValueType # 4 global___DatapointError = DatapointError +class _EntryType: + ValueType = typing.NewType("ValueType", builtins.int) + V: typing_extensions.TypeAlias = ValueType + +class _EntryTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_EntryType.ValueType], builtins.type): + DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor + ENTRY_TYPE_UNSPECIFIED: _EntryType.ValueType # 0 + ENTRY_TYPE_SENSOR: _EntryType.ValueType # 1 + ENTRY_TYPE_ACTUATOR: _EntryType.ValueType # 2 + ENTRY_TYPE_ATTRIBUTE: _EntryType.ValueType # 3 + +class EntryType(_EntryType, metaclass=_EntryTypeEnumTypeWrapper): ... + +ENTRY_TYPE_UNSPECIFIED: EntryType.ValueType # 0 +ENTRY_TYPE_SENSOR: EntryType.ValueType # 1 +ENTRY_TYPE_ACTUATOR: EntryType.ValueType # 2 +ENTRY_TYPE_ATTRIBUTE: EntryType.ValueType # 3 +global___EntryType = EntryType class _ChangeType: - ValueType = typing.NewType('ValueType', builtins.int) + ValueType = typing.NewType("ValueType", builtins.int) V: typing_extensions.TypeAlias = ValueType + class _ChangeTypeEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[_ChangeType.ValueType], builtins.type): DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor STATIC: _ChangeType.ValueType # 0 """Value never changes""" - ON_CHANGE: _ChangeType.ValueType # 1 """Updates are provided every time the value changes (i.e.""" - CONTINUOUS: _ChangeType.ValueType # 2 """window is open / closed) Value is updated continuously. Broker needs to tell """ -class ChangeType(_ChangeType, metaclass=_ChangeTypeEnumTypeWrapper): - pass +class ChangeType(_ChangeType, metaclass=_ChangeTypeEnumTypeWrapper): ... STATIC: ChangeType.ValueType # 0 """Value never changes""" - ON_CHANGE: ChangeType.ValueType # 1 """Updates are provided every time the value changes (i.e.""" - CONTINUOUS: ChangeType.ValueType # 2 """window is open / closed) Value is updated continuously. Broker needs to tell """ - global___ChangeType = ChangeType - +@typing.final class StringArray(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + VALUES_FIELD_NUMBER: builtins.int @property - def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[typing.Text]: ... - def __init__(self, + def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: ... + def __init__( + self, *, - values: typing.Optional[typing.Iterable[typing.Text]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["values",b"values"]) -> None: ... + values: collections.abc.Iterable[builtins.str] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + global___StringArray = StringArray +@typing.final class BoolArray(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + VALUES_FIELD_NUMBER: builtins.int @property def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ... - def __init__(self, + def __init__( + self, *, - values: typing.Optional[typing.Iterable[builtins.bool]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["values",b"values"]) -> None: ... + values: collections.abc.Iterable[builtins.bool] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + global___BoolArray = BoolArray +@typing.final class Int32Array(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + VALUES_FIELD_NUMBER: builtins.int @property def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... - def __init__(self, + def __init__( + self, *, - values: typing.Optional[typing.Iterable[builtins.int]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["values",b"values"]) -> None: ... + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + global___Int32Array = Int32Array +@typing.final class Int64Array(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + VALUES_FIELD_NUMBER: builtins.int @property def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... - def __init__(self, + def __init__( + self, *, - values: typing.Optional[typing.Iterable[builtins.int]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["values",b"values"]) -> None: ... + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + global___Int64Array = Int64Array +@typing.final class Uint32Array(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + VALUES_FIELD_NUMBER: builtins.int @property def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... - def __init__(self, + def __init__( + self, *, - values: typing.Optional[typing.Iterable[builtins.int]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["values",b"values"]) -> None: ... + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + global___Uint32Array = Uint32Array +@typing.final class Uint64Array(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + VALUES_FIELD_NUMBER: builtins.int @property def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ... - def __init__(self, + def __init__( + self, *, - values: typing.Optional[typing.Iterable[builtins.int]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["values",b"values"]) -> None: ... + values: collections.abc.Iterable[builtins.int] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + global___Uint64Array = Uint64Array +@typing.final class FloatArray(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + VALUES_FIELD_NUMBER: builtins.int @property def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.float]: ... - def __init__(self, + def __init__( + self, *, - values: typing.Optional[typing.Iterable[builtins.float]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["values",b"values"]) -> None: ... + values: collections.abc.Iterable[builtins.float] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + global___FloatArray = FloatArray +@typing.final class DoubleArray(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + VALUES_FIELD_NUMBER: builtins.int @property def values(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.float]: ... - def __init__(self, + def __init__( + self, *, - values: typing.Optional[typing.Iterable[builtins.float]] = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["values",b"values"]) -> None: ... + values: collections.abc.Iterable[builtins.float] | None = ..., + ) -> None: ... + def ClearField(self, field_name: typing.Literal["values", b"values"]) -> None: ... + global___DoubleArray = DoubleArray +@typing.final class Datapoint(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + class _Failure: - ValueType = typing.NewType('ValueType', builtins.int) + ValueType = typing.NewType("ValueType", builtins.int) V: typing_extensions.TypeAlias = ValueType + class _FailureEnumTypeWrapper(google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[Datapoint._Failure.ValueType], builtins.type): DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor INVALID_VALUE: Datapoint._Failure.ValueType # 0 """The data point is known, but doesn't have a valid value""" - NOT_AVAILABLE: Datapoint._Failure.ValueType # 1 """The data point is known, but no value is available""" - UNKNOWN_DATAPOINT: Datapoint._Failure.ValueType # 2 """Unknown datapoint""" - ACCESS_DENIED: Datapoint._Failure.ValueType # 3 """Access denied""" - INTERNAL_ERROR: Datapoint._Failure.ValueType # 4 """Unexpected internal error""" - class Failure(_Failure, metaclass=_FailureEnumTypeWrapper): - pass - + class Failure(_Failure, metaclass=_FailureEnumTypeWrapper): ... INVALID_VALUE: Datapoint.Failure.ValueType # 0 """The data point is known, but doesn't have a valid value""" - NOT_AVAILABLE: Datapoint.Failure.ValueType # 1 """The data point is known, but no value is available""" - UNKNOWN_DATAPOINT: Datapoint.Failure.ValueType # 2 """Unknown datapoint""" - ACCESS_DENIED: Datapoint.Failure.ValueType # 3 """Access denied""" - INTERNAL_ERROR: Datapoint.Failure.ValueType # 4 """Unexpected internal error""" - TIMESTAMP_FIELD_NUMBER: builtins.int FAILURE_VALUE_FIELD_NUMBER: builtins.int STRING_VALUE_FIELD_NUMBER: builtins.int @@ -290,12 +357,8 @@ class Datapoint(google.protobuf.message.Message): UINT64_ARRAY_FIELD_NUMBER: builtins.int FLOAT_ARRAY_FIELD_NUMBER: builtins.int DOUBLE_ARRAY_FIELD_NUMBER: builtins.int - @property - def timestamp(self) -> google.protobuf.timestamp_pb2.Timestamp: - """Timestamp of the value""" - pass failure_value: global___Datapoint.Failure.ValueType - string_value: typing.Text + string_value: builtins.str bool_value: builtins.bool int32_value: builtins.int int64_value: builtins.int @@ -303,6 +366,10 @@ class Datapoint(google.protobuf.message.Message): uint64_value: builtins.int float_value: builtins.float double_value: builtins.float + @property + def timestamp(self) -> google.protobuf.timestamp_pb2.Timestamp: + """Timestamp of the value""" + @property def string_array(self) -> global___StringArray: ... @property @@ -319,11 +386,12 @@ class Datapoint(google.protobuf.message.Message): def float_array(self) -> global___FloatArray: ... @property def double_array(self) -> global___DoubleArray: ... - def __init__(self, + def __init__( + self, *, - timestamp: typing.Optional[google.protobuf.timestamp_pb2.Timestamp] = ..., + timestamp: google.protobuf.timestamp_pb2.Timestamp | None = ..., failure_value: global___Datapoint.Failure.ValueType = ..., - string_value: typing.Text = ..., + string_value: builtins.str = ..., bool_value: builtins.bool = ..., int32_value: builtins.int = ..., int64_value: builtins.int = ..., @@ -331,45 +399,143 @@ class Datapoint(google.protobuf.message.Message): uint64_value: builtins.int = ..., float_value: builtins.float = ..., double_value: builtins.float = ..., - string_array: typing.Optional[global___StringArray] = ..., - bool_array: typing.Optional[global___BoolArray] = ..., - int32_array: typing.Optional[global___Int32Array] = ..., - int64_array: typing.Optional[global___Int64Array] = ..., - uint32_array: typing.Optional[global___Uint32Array] = ..., - uint64_array: typing.Optional[global___Uint64Array] = ..., - float_array: typing.Optional[global___FloatArray] = ..., - double_array: typing.Optional[global___DoubleArray] = ..., - ) -> None: ... - def HasField(self, field_name: typing_extensions.Literal["bool_array",b"bool_array","bool_value",b"bool_value","double_array",b"double_array","double_value",b"double_value","failure_value",b"failure_value","float_array",b"float_array","float_value",b"float_value","int32_array",b"int32_array","int32_value",b"int32_value","int64_array",b"int64_array","int64_value",b"int64_value","string_array",b"string_array","string_value",b"string_value","timestamp",b"timestamp","uint32_array",b"uint32_array","uint32_value",b"uint32_value","uint64_array",b"uint64_array","uint64_value",b"uint64_value","value",b"value"]) -> builtins.bool: ... - def ClearField(self, field_name: typing_extensions.Literal["bool_array",b"bool_array","bool_value",b"bool_value","double_array",b"double_array","double_value",b"double_value","failure_value",b"failure_value","float_array",b"float_array","float_value",b"float_value","int32_array",b"int32_array","int32_value",b"int32_value","int64_array",b"int64_array","int64_value",b"int64_value","string_array",b"string_array","string_value",b"string_value","timestamp",b"timestamp","uint32_array",b"uint32_array","uint32_value",b"uint32_value","uint64_array",b"uint64_array","uint64_value",b"uint64_value","value",b"value"]) -> None: ... - def WhichOneof(self, oneof_group: typing_extensions.Literal["value",b"value"]) -> typing.Optional[typing_extensions.Literal["failure_value","string_value","bool_value","int32_value","int64_value","uint32_value","uint64_value","float_value","double_value","string_array","bool_array","int32_array","int64_array","uint32_array","uint64_array","float_array","double_array"]]: ... + string_array: global___StringArray | None = ..., + bool_array: global___BoolArray | None = ..., + int32_array: global___Int32Array | None = ..., + int64_array: global___Int64Array | None = ..., + uint32_array: global___Uint32Array | None = ..., + uint64_array: global___Uint64Array | None = ..., + float_array: global___FloatArray | None = ..., + double_array: global___DoubleArray | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["bool_array", b"bool_array", "bool_value", b"bool_value", "double_array", b"double_array", "double_value", b"double_value", "failure_value", b"failure_value", "float_array", b"float_array", "float_value", b"float_value", "int32_array", b"int32_array", "int32_value", b"int32_value", "int64_array", b"int64_array", "int64_value", b"int64_value", "string_array", b"string_array", "string_value", b"string_value", "timestamp", b"timestamp", "uint32_array", b"uint32_array", "uint32_value", b"uint32_value", "uint64_array", b"uint64_array", "uint64_value", b"uint64_value", "value", b"value"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["bool_array", b"bool_array", "bool_value", b"bool_value", "double_array", b"double_array", "double_value", b"double_value", "failure_value", b"failure_value", "float_array", b"float_array", "float_value", b"float_value", "int32_array", b"int32_array", "int32_value", b"int32_value", "int64_array", b"int64_array", "int64_value", b"int64_value", "string_array", b"string_array", "string_value", b"string_value", "timestamp", b"timestamp", "uint32_array", b"uint32_array", "uint32_value", b"uint32_value", "uint64_array", b"uint64_array", "uint64_value", b"uint64_value", "value", b"value"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["value", b"value"]) -> typing.Literal["failure_value", "string_value", "bool_value", "int32_value", "int64_value", "uint32_value", "uint64_value", "float_value", "double_value", "string_array", "bool_array", "int32_array", "int64_array", "uint32_array", "uint64_array", "float_array", "double_array"] | None: ... + global___Datapoint = Datapoint +@typing.final class Metadata(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor + ID_FIELD_NUMBER: builtins.int + ENTRY_TYPE_FIELD_NUMBER: builtins.int NAME_FIELD_NUMBER: builtins.int DATA_TYPE_FIELD_NUMBER: builtins.int CHANGE_TYPE_FIELD_NUMBER: builtins.int DESCRIPTION_FIELD_NUMBER: builtins.int + ALLOWED_FIELD_NUMBER: builtins.int + MIN_FIELD_NUMBER: builtins.int + MAX_FIELD_NUMBER: builtins.int id: builtins.int - """Id to be used in "get" and "subscribe" requests. Ids stay valid during - one power cycle, only. - """ - - name: typing.Text + entry_type: global___EntryType.ValueType + name: builtins.str data_type: global___DataType.ValueType change_type: global___ChangeType.ValueType """CONTINUOUS or STATIC or ON_CHANGE""" + description: builtins.str + @property + def allowed(self) -> global___Allowed: + """Value restrictions checked/enforced by Databroker.""" - description: typing.Text - def __init__(self, + @property + def min(self) -> global___ValueRestriction: ... + @property + def max(self) -> global___ValueRestriction: ... + def __init__( + self, *, id: builtins.int = ..., - name: typing.Text = ..., + entry_type: global___EntryType.ValueType = ..., + name: builtins.str = ..., data_type: global___DataType.ValueType = ..., change_type: global___ChangeType.ValueType = ..., - description: typing.Text = ..., - ) -> None: ... - def ClearField(self, field_name: typing_extensions.Literal["change_type",b"change_type","data_type",b"data_type","description",b"description","id",b"id","name",b"name"]) -> None: ... + description: builtins.str = ..., + allowed: global___Allowed | None = ..., + min: global___ValueRestriction | None = ..., + max: global___ValueRestriction | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["allowed", b"allowed", "max", b"max", "min", b"min"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["allowed", b"allowed", "change_type", b"change_type", "data_type", b"data_type", "description", b"description", "entry_type", b"entry_type", "id", b"id", "max", b"max", "min", b"min", "name", b"name"]) -> None: ... + global___Metadata = Metadata + +@typing.final +class Allowed(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + STRING_VALUES_FIELD_NUMBER: builtins.int + INT32_VALUES_FIELD_NUMBER: builtins.int + INT64_VALUES_FIELD_NUMBER: builtins.int + UINT32_VALUES_FIELD_NUMBER: builtins.int + UINT64_VALUES_FIELD_NUMBER: builtins.int + FLOAT_VALUES_FIELD_NUMBER: builtins.int + DOUBLE_VALUES_FIELD_NUMBER: builtins.int + @property + def string_values(self) -> global___StringArray: ... + @property + def int32_values(self) -> global___Int32Array: ... + @property + def int64_values(self) -> global___Int64Array: ... + @property + def uint32_values(self) -> global___Uint32Array: ... + @property + def uint64_values(self) -> global___Uint64Array: ... + @property + def float_values(self) -> global___FloatArray: ... + @property + def double_values(self) -> global___DoubleArray: ... + def __init__( + self, + *, + string_values: global___StringArray | None = ..., + int32_values: global___Int32Array | None = ..., + int64_values: global___Int64Array | None = ..., + uint32_values: global___Uint32Array | None = ..., + uint64_values: global___Uint64Array | None = ..., + float_values: global___FloatArray | None = ..., + double_values: global___DoubleArray | None = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["double_values", b"double_values", "float_values", b"float_values", "int32_values", b"int32_values", "int64_values", b"int64_values", "string_values", b"string_values", "uint32_values", b"uint32_values", "uint64_values", b"uint64_values", "values", b"values"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["double_values", b"double_values", "float_values", b"float_values", "int32_values", b"int32_values", "int64_values", b"int64_values", "string_values", b"string_values", "uint32_values", b"uint32_values", "uint64_values", b"uint64_values", "values", b"values"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["values", b"values"]) -> typing.Literal["string_values", "int32_values", "int64_values", "uint32_values", "uint64_values", "float_values", "double_values"] | None: ... + +global___Allowed = Allowed + +@typing.final +class ValueRestriction(google.protobuf.message.Message): + DESCRIPTOR: google.protobuf.descriptor.Descriptor + + STRING_FIELD_NUMBER: builtins.int + BOOL_FIELD_NUMBER: builtins.int + INT32_FIELD_NUMBER: builtins.int + INT64_FIELD_NUMBER: builtins.int + UINT32_FIELD_NUMBER: builtins.int + UINT64_FIELD_NUMBER: builtins.int + FLOAT_FIELD_NUMBER: builtins.int + DOUBLE_FIELD_NUMBER: builtins.int + string: builtins.str + bool: builtins.bool + int32: builtins.int + int64: builtins.int + uint32: builtins.int + uint64: builtins.int + float: builtins.float + double: builtins.float + def __init__( + self, + *, + string: builtins.str = ..., + bool: builtins.bool = ..., + int32: builtins.int = ..., + int64: builtins.int = ..., + uint32: builtins.int = ..., + uint64: builtins.int = ..., + float: builtins.float = ..., + double: builtins.float = ..., + ) -> None: ... + def HasField(self, field_name: typing.Literal["bool", b"bool", "double", b"double", "float", b"float", "int32", b"int32", "int64", b"int64", "string", b"string", "typed_value", b"typed_value", "uint32", b"uint32", "uint64", b"uint64"]) -> builtins.bool: ... + def ClearField(self, field_name: typing.Literal["bool", b"bool", "double", b"double", "float", b"float", "int32", b"int32", "int64", b"int64", "string", b"string", "typed_value", b"typed_value", "uint32", b"uint32", "uint64", b"uint64"]) -> None: ... + def WhichOneof(self, oneof_group: typing.Literal["typed_value", b"typed_value"]) -> typing.Literal["string", "bool", "int32", "int64", "uint32", "uint64", "float", "double"] | None: ... + +global___ValueRestriction = ValueRestriction diff --git a/integration_test/gen_proto/sdv/databroker/v1/types_pb2_grpc.py b/integration_test/gen_proto/sdv/databroker/v1/types_pb2_grpc.py index 2daafffe..90e2b445 100644 --- a/integration_test/gen_proto/sdv/databroker/v1/types_pb2_grpc.py +++ b/integration_test/gen_proto/sdv/databroker/v1/types_pb2_grpc.py @@ -1,4 +1,37 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ # Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT! """Client and server classes corresponding to protobuf-defined services.""" import grpc +import warnings + +GRPC_GENERATED_VERSION = '1.71.0' +GRPC_VERSION = grpc.__version__ +_version_not_supported = False + +try: + from grpc._utilities import first_version_is_lower + _version_not_supported = first_version_is_lower(GRPC_VERSION, GRPC_GENERATED_VERSION) +except ImportError: + _version_not_supported = True + +if _version_not_supported: + raise RuntimeError( + f'The grpc package installed is at version {GRPC_VERSION},' + + f' but the generated code in sdv/databroker/v1/types_pb2_grpc.py depends on' + + f' grpcio>={GRPC_GENERATED_VERSION}.' + + f' Please upgrade your grpc module to grpcio>={GRPC_GENERATED_VERSION}' + + f' or downgrade your generated code using grpcio-tools<={GRPC_VERSION}.' + ) diff --git a/integration_test/provider.py b/integration_test/provider.py new file mode 100644 index 00000000..4f18816e --- /dev/null +++ b/integration_test/provider.py @@ -0,0 +1,303 @@ +#!/usr/bin/env python3 +# /******************************************************************************** +# * Copyright (c) 2025 Contributors to the Eclipse Foundation +# * +# * See the NOTICE file(s) distributed with this work for additional +# * information regarding copyright ownership. +# * +# * This program and the accompanying materials are made available under the +# * terms of the Apache License 2.0 which is available at +# * http://www.apache.org/licenses/LICENSE-2.0 +# * +# * SPDX-License-Identifier: Apache-2.0 +# ********************************************************************************/ +#!/usr/bin/env python3 +import grpc +import time +import os +import sys +import threading +import queue +import pytest + +from pytest_bdd import given, when, then,parsers + +from gen_proto.kuksa.val.v2 import val_pb2_grpc; +from gen_proto.kuksa.val.v2 import val_pb2; +from gen_proto.kuksa.val.v2 import types_pb2; + +class ProviderValV2: + def __init__(self,pytestconfig): + """ + Initializes the gRPC channel, stub, and internal data structures. + """ + self.pytestconfig=pytestconfig + #self.address = f"{host}:{port}" + #if secure: + # # Configure secure credentials as needed. + # credentials = grpc.ssl_channel_credentials() + # self.channel = grpc.secure_channel(self.address, credentials) + #else: + self.channel = grpc.insecure_channel(self.pytestconfig.getini('viss_grpc_base_url')) + self.stub = val_pb2_grpc.VALStub(self.channel) + + # For the bidirectional stream + self._request_queue = queue.Queue() + self._response_queue = queue.Queue() + self._shutdown = False + # Default frequency (iterations per second) for checking/sending messages. + self._frequency = 1000.0 + self._send_interval = 1.0 / self._frequency + + # Thread for handling the streaming call. + self._stream_thread = None + + # Map for storing metadata by int id. + self.metadata_map = {} + self.start_stream() + + def disconnect(self): + if self.channel: + self.channel.close() + time.sleep(2) + + def list_metadata(self, root_branch="Vehicle"): + request = val_pb2.ListMetadataRequest(root=root_branch) + try: + response = self.stub.ListMetadata(request) + for metadata in response.metadata: + # Assume metadata.id is an int field. + self.metadata_map[int(metadata.id)] = metadata + return self.metadata_map + except grpc.RpcError as e: + print(f"ListMetadata RPC failed: {e.code()} - {e.details()}") + return None + + def set_frequency(self, frequency): + """ + Adjusts the frequency (iterations per second) at which the stream loop + checks for new messages. + :param frequency: New frequency in Hz (iterations per second). + """ + if frequency <= 0: + raise ValueError("Frequency must be positive.") + self._frequency = frequency + self._send_interval = 1.0 / frequency + print(f"Frequency set to {frequency} Hz.") + + def _get_queued_message(self): + """ + Get queued message + """ + while not self._shutdown: + try: + # Wait for a message from the queue. + message = self._request_queue.get(timeout=self._send_interval) + yield message + except queue.Empty: + # No message available in the interval; continue waiting. + continue + + def _stream_loop(self): + """ + Runs the bidirectional streaming call in a loop. Sends messages from the queue + and processes responses. + """ + try: + stream = self.stub.OpenProviderStream(self._get_queued_message()) + for response in stream: + # Process responses from the server. + if response.HasField("provide_actuation_response"): + print("Received ProvideActuationResponse") + elif response.HasField("publish_values_response"): + print("Received PublishValuesResponse") + elif response.HasField("provide_signal_response"): + self._response_queue.put(response) + #print("Received ProvideSignalResponse {}", response) + elif response.HasField("batch_actuate_stream_request"): + self._response_queue.put(response) + elif response.HasField("update_filter_request"): + self._response_queue.put(response) + elif response.HasField("get_provider_value_request"): + self._response_queue.put(response) + else: + print("Received unknown response:", response) + # Check shutdown flag between responses. + if self._shutdown: + break + except grpc.RpcError as e: + print(f"OpenProviderStream RPC failed: {e.code()} - {e.details()}") + + def start_stream(self): + if self._stream_thread and self._stream_thread.is_alive(): + return + self._shutdown = False + self._stream_thread = threading.Thread(target=self._stream_loop, daemon=True) + self._stream_thread.start() + + def shutdown(self): + self._shutdown = True + if self._stream_thread: + self._stream_thread.join() + self.disconnect() + # --- Methods for sending various types of messages through the same stream --- + + def send_provide_actuation(self, actuator_identifiers): + req = val_pb2.ProvideActuationRequest( + actuator_identifiers=actuator_identifiers + ) + msg = val_pb2.OpenProviderStreamRequest(provide_actuation_request=req) + self._request_queue.put(msg) + + def send_publish_values(self, request_id, data_points): + req = val_pb2.PublishValuesRequest( + request_id=request_id, + data_points=data_points + ) + msg = val_pb2.OpenProviderStreamRequest(publish_values_request=req) + self._request_queue.put(msg) + + def send_provide_signal(self, signals_sample_intervals): + req = val_pb2.ProvideSignalRequest( + signals_sample_intervals=signals_sample_intervals + ) + msg = val_pb2.OpenProviderStreamRequest(provide_signal_request=req) + self._request_queue.put(msg) + + def send_provider_error_indication(self, provider_error): + req = val_pb2.ProviderErrorIndication( + provider_error=provider_error + ) + msg = val_pb2.OpenProviderStreamRequest(provider_error_indication=req) + self._request_queue.put(msg) + + # --- Methods for receiving various types of request through the same stream --- + + def received_provide_actuation_response(self, request): + response = self._response_queue.get() + if response.HasField("provide_actuation_response"): + return response + else: + None + + def received_publish_values_response(self, request): + response = self._response_queue.get() + if response.HasField("publish_values_response"): + return response + else: + None + + def received_provide_signal_response(self, request): + response = self._response_queue.get() + if response.HasField("provide_signal_response"): + return response + else: + None + + def received_provide_signal_response(self, request): + response = self._response_queue.get() + if response.HasField("provide_signal_response"): + return response + else: + None + + def received_batch_actuate_stream_request(self, request): + response = self._response_queue.get() + if response.HasField("batch_actuate_stream_request"): + return response + else: + None + + def received_update_filter_request(self, request): + response = self._response_queue.get() + if response.HasField("update_filter_request"): + return response + else: + None + + def received_get_provider_value_request(self, request): + response = self._response_queue.get() + if response.HasField("get_provider_value_request"): + return response + else: + None + +@pytest.fixture +def connected_provider(request,pytestconfig): + connected_provider = ProviderValV2(pytestconfig) + def cleanup(): + connected_provider.disconnect() + request.addfinalizer(cleanup) + return connected_provider + +@given("the Provider connected via gRPC") +def grpc_kuksa_provider_via_grpc(connected_provider): + metadata = connected_provider.list_metadata(root_branch="Vehicle.*") + #if metadata is not None: + # for key, meta in metadata.items(): + # print(f" ID: {key}, Metadata: {meta}") + +@when(parsers.parse('Provider claims the signal "{path}"')) +def claim_signal(connected_provider, path): + sample_intervals = {882: types_pb2.SampleInterval(interval_ms=10)} + connected_provider.send_provide_signal(signals_sample_intervals=sample_intervals) + +@when(parsers.parse('Provider disconnects')) +def disconnect(connected_provider): + connected_provider.disconnect() + +# Example usage: +def main(): + provider = ProviderValV2(host="localhost", port=55555, secure=False) + + # Retrieve and store metadata. + #metadata = client.list_metadata(root_branch="Vehicle") + #if metadata is not None: + #print("Stored metadata (by id):") + #for key, meta in metadata.items(): + #print(f" ID: {key}, Metadata: {meta}") + + # Step 2: Start the bidirectional stream + provider.start_stream() + + # Step 3: Set frequency to 2 Hz (2 messages per second) + provider.set_frequency(2.0) + + # Step 4: Send ProvideActuationRequest + #client.send_provide_actuation(actuator_identifiers=["example_actuator_1", "example_actuator_2"]) + + # Step 5: Send PublishValuesRequest + #datapoint = val_pb2.Datapoint() + #datapoint.value_str = "sensor_value" + #client.send_publish_values(request_id=1, data_points={1001: datapoint}) + + # Step 6: Send BatchActuateStreamResponse + #signal_id = val_pb2.SignalId(name="test_signal") + #error = val_pb2.Error(code=0, message="No error") + #client.send_batch_actuate_stream_response(signal_id=signal_id, error=error) + + # Step 7: Send ProvideSignalRequest + sample_intervals = {882: types_pb2.SampleInterval(interval_ms=10)} + provider.send_provide_signal(signals_sample_intervals=sample_intervals) + + # Step 8: Send UpdateFilterResponse + #filter_error = val_pb2.Error(code=1, message="Filter updated successfully") + #client.send_update_filter_response(request_id=2, filter_error=filter_error) + + # Step 9: Send GetProviderValueResponse + #response_value = val_pb2.GetValueResponse() + #response_value.value.value_str = "retrieved_value" + #client.send_get_provider_value_response(request_id=3, response=response_value) + + # Step 10: Send ProviderErrorIndication + #provider_error = val_pb2.ProviderError(code=2, message="Sample error occurred") + #client.send_provider_error_indication(provider_error=provider_error) + + # Step 11: Let the stream run for a while + time.sleep(10) + + # Step 12: Shutdown the stream + # client.shutdown() + +if __name__ == "__main__": + main() diff --git a/integration_test/viss/STEPS.md b/integration_test/viss/STEPS.md index 985695db..c7d21874 100644 --- a/integration_test/viss/STEPS.md +++ b/integration_test/viss/STEPS.md @@ -32,6 +32,10 @@ The client will connect to the VISS server using the WebSocket channel. The client will connect to the VISS server using the MQTT channel. +> Given the Provider connected via gRPC + +The provider kuksa val v2 will connect to the Databroker using gRPC. + > Given I have a subscription to "_VSS Path_" _VSS Path_: A VSS path, such as "Vehicle.Speed" @@ -98,6 +102,14 @@ Only receive updates to a VSS data point when the delta change of a value reache _datatable_: A table, rows are signals. First column is VSS Path, Second column is new data point value to set. +> When Provider claims the signal "_VSS Path_" + + Provider will provide the values for the claimed signal + +> When Provider disconnects + + Provider disconnects + ## Then > Then I should receive a valid response diff --git a/integration_test/viss/conftest.py b/integration_test/viss/conftest.py index c170949d..0783bd5b 100644 --- a/integration_test/viss/conftest.py +++ b/integration_test/viss/conftest.py @@ -37,5 +37,6 @@ def pytest_addoption(parser): parser.addini('viss_ws_base_url', 'URL to Databroker VISS WebSocket endpoint (ws://hostname:port)', type="string", default="ws://localhost:8090") parser.addini('viss_http_base_url', 'URL to Databroker VISS HTTP endpoint (http://hostname:port)', type="string", default="http://localhost:8090") parser.addini('viss_mqtt_base_url', 'URL to Databroker VISS MQTT endpoint (mqtt://hostname:port)', type="string", default="mqtt://localhost:1883") + parser.addini('viss_grpc_base_url', 'URL to Databroker VISS gRPC endpoint (http://hostname:port)', type="string", default="localhost:55555") parser.addini('viss_connect_timeout', 'Connect timeout for VISS clients (float in seconds)', type="string", default="1.0") parser.addini('viss_message_timeout', 'Connect timeout for VISS clients (float in seconds)', type="string", default="0.5") diff --git a/integration_test/viss/features/viss_v2_core_subscribe.feature b/integration_test/viss/features/viss_v2_core_subscribe.feature index b1197fa8..9af526ee 100644 --- a/integration_test/viss/features/viss_v2_core_subscribe.feature +++ b/integration_test/viss/features/viss_v2_core_subscribe.feature @@ -7,6 +7,17 @@ Feature: VISS v2 Compliance Testing - Core: Subscribe Background: Given the VISS server is running Given the VISS client is connected via WebSocket + Given the Provider connected via gRPC + + # 5.1.3 Subscribe request + # The VISS server must support errors when service not available + @MustHave + Scenario: Subscribe when service unavailable + When Provider claims the signal "Vehicle.Speed" + When I send a subscription request for "Vehicle.Speed" + Then I should receive a valid subscribe response + When Provider disconnects + Then I should receive a service unavailable subscribe error event # 5.1.3 Subscribe request # The VISS server must support subscriptions to receive data updates. diff --git a/integration_test/viss/pytest.ini b/integration_test/viss/pytest.ini index e36822c1..7333e954 100644 --- a/integration_test/viss/pytest.ini +++ b/integration_test/viss/pytest.ini @@ -17,5 +17,6 @@ bdd_features_base_dir = "features" viss_ws_base_url = ws://localhost:8090 viss_http_base_url = http://localhost:8090 viss_mqtt_base_url = mqtt://localhost:1883 +viss_grpc_base_url = localhost:55555 viss_connect_timeout = 1.0 viss_message_timeout = 0.1 diff --git a/integration_test/viss/tests/test_steps/viss_v2_steps.py b/integration_test/viss/tests/test_steps/viss_v2_steps.py index 3c78f90e..52167c2f 100644 --- a/integration_test/viss/tests/test_steps/viss_v2_steps.py +++ b/integration_test/viss/tests/test_steps/viss_v2_steps.py @@ -104,6 +104,10 @@ def find_messages(self, *, logger.debug(f"Found message: {response}") return response + def trigger_reading_websocket(self): + for _, client in self.clients.items(): + if client.is_connected(): + client.read_all_websocket_messages() @pytest.fixture def request_id(): @@ -439,7 +443,6 @@ def receive_ws_subscribe(connected_clients,request_id, subscription_id): def receive_ws_subscribe_error_event(connected_clients,request_id): envelope = connected_clients.find_message(request_id=request_id, action="subscribe") response = envelope["body"] - assert "requestId" in response assert "error" in response assert "ts" in response assert "subscriptionId" not in response @@ -454,6 +457,22 @@ def receive_ws_subscribe_error_event(connected_clients,request_id): # "reason": "unavailable_data", # "message": "The requested data was not found."} +@then("I should receive a service unavailable subscribe error event") +def receive_ws_subscribe_error_event(connected_clients,subscription_id): + # READ ERROR + connected_clients.trigger_reading_websocket() + envelope = connected_clients.find_message(subscription_id=subscription_id, request_id=None, action="subscription") + response = envelope["body"] + + assert "error" in response + assert "ts" in response + assert "subscriptionId" in response + + # Current implementation + assert response["error"] == {"number": 503, + "reason": "service_unavailable", + "message": "The server is temporarily unable to handle the request."} + @then("I should receive a set error event") def receive_ws_set_error_event(connected_clients,request_id): envelope = connected_clients.find_message(request_id=request_id, action="set") diff --git a/integration_test/viss/tests/test_steps/websockets_viss_client.py b/integration_test/viss/tests/test_steps/websockets_viss_client.py index 62e1218d..a716d80a 100644 --- a/integration_test/viss/tests/test_steps/websockets_viss_client.py +++ b/integration_test/viss/tests/test_steps/websockets_viss_client.py @@ -56,12 +56,12 @@ def send(self, request_id, message, authorization=None): } logger.debug(f"Storing sent envelope: {envelope}") self.sent_messages.insert(envelope) - self._read_all_websocket_messages() + self.read_all_websocket_messages() # TODO: message should be "consumed" by the test, e.g. only checked once and then removed from the list of # received messages. - def _read_all_websocket_messages(self): + def read_all_websocket_messages(self): self._ws.settimeout(float(self.pytestconfig.getini('viss_message_timeout'))) while True: try: @@ -77,6 +77,7 @@ def _read_all_websocket_messages(self): envelope['subscriptionId'] = message['subscriptionId'] if "action" in message: envelope['action'] = message['action'] + #print("RECEIVED ENVELOPE: ", envelope) self.received_messages.insert(envelope) logger.debug(f"Storing received envelope: {envelope}") except websocket.WebSocketTimeoutException: @@ -127,8 +128,11 @@ def find_messages(self, *, search_template &= condition # Combine with AND logger.debug(f"Query template: {search_template}") + + #print("QUERY: ", search_template) results = self.received_messages.search(search_template) logger.debug(f"Found messages: {results}") + #print("QUERY RESULT: ", search_template) return results def find_message(self, *, @@ -142,4 +146,5 @@ def find_message(self, *, authorization=authorization) result = max(results,key=lambda x: x["timestamp"], default=None) logger.debug(f"Found latest message: {result}") + #print("FINAL RESULT: ", result) return result diff --git a/integration_test/viss/tests/test_viss_v2.py b/integration_test/viss/tests/test_viss_v2.py index d889436c..f02128c2 100644 --- a/integration_test/viss/tests/test_viss_v2.py +++ b/integration_test/viss/tests/test_viss_v2.py @@ -14,6 +14,7 @@ from pytest_bdd import scenarios from test_steps.viss_v2_steps import * # Import step definitions +from provider import * # Import provider definition import os # Unset proxy settings, to make sure we're connecting to localhost diff --git a/proto/kuksa/val/v2/val.proto b/proto/kuksa/val/v2/val.proto index cf64db91..0960ca62 100644 --- a/proto/kuksa/val/v2/val.proto +++ b/proto/kuksa/val/v2/val.proto @@ -177,24 +177,30 @@ service VAL { // // Errors: // - Provider sends ProvideActuationRequest -> Databroker returns ProvideActuationResponse - // Returns (GRPC error code) and closes the stream call (strict case). - // NOT_FOUND if any of the signals are non-existant. - // PERMISSION_DENIED if access is denied for any of the signals. - // UNAUTHENTICATED if no credentials provided or credentials has expired - // ALREADY_EXISTS if a provider already claimed the ownership of an actuator + // - strict case + // Returns (GRPC error code) and closes the stream call + // NOT_FOUND if any of the signals are non-existant. + // PERMISSION_DENIED if access is denied for any of the signals. + // UNAUTHENTICATED if no credentials provided or credentials has expired + // ALREADY_EXISTS if a provider already claimed the ownership of an actuator // // - Provider sends PublishValuesRequest -> Databroker returns PublishValuesResponse upon error, and nothing upon success - // GRPC errors are returned as messages in the stream - // response with the signal id `map status = 2;` (permissive case) - // NOT_FOUND if a signal is non-existant. - // PERMISSION_DENIED - // - if access is denied for a signal. - // INVALID_ARGUMENT - // - if the data type used in the request does not match - // the data type of the addressed signal - // - if the published value is not accepted, - // e.g. if sending an unsupported enum value - // - if the published value is out of the min/max range specified + // - permissive case + // GRPC errors are returned as messages in the stream + // response with the signal id `map status = 2;` + // NOT_FOUND if a signal is non-existant. + // PERMISSION_DENIED + // - if access is denied for a signal. + // INVALID_ARGUMENT + // - if the data type used in the request does not match + // the data type of the addressed signal + // - if the published value is not accepted, + // e.g. if sending an unsupported enum value + // - if the published value is out of the min/max range specified + // - strict case + // Returns (GRPC error code) and closes the stream call. + // ALREADY_EXISTS if a provider already claimed the ownership of the signals + // ABORTED if provider has not claimed yet the signals // // - Databroker sends BatchActuateStreamRequest -> Provider shall return a BatchActuateStreamResponse, // for every signal requested to indicate if the request was accepted or not. @@ -202,11 +208,17 @@ service VAL { // as of today Databroker will not react on the received error message. // // - Provider sends ProvideSignalRequest -> Databroker returns ProvideSignalResponse - // Returns (GRPC error code) and closes the stream call (strict case). - // NOT_FOUND if any of the signals are non-existant. - // PERMISSION_DENIED if access is denied for any of the signals. - // UNAUTHENTICATED if no credentials provided or credentials has expired - // ALREADY_EXISTS if a provider already claimed the ownership of any signal. + // - strict case + // Returns (GRPC error code) and closes the stream call. + // NOT_FOUND if any of the signals are non-existant. + // PERMISSION_DENIED if access is denied for any of the signals. + // UNAUTHENTICATED if no credentials provided or credentials has expired + // ALREADY_EXISTS if a provider already claimed the ownership of any signal. + // + // - Provider sends ProviderErrorIndication + // - strict case + // Returns (GRPC error code) and closes the stream call. + // ABORTED if provider has not claimed yet the signals // rpc OpenProviderStream(stream OpenProviderStreamRequest) returns (stream OpenProviderStreamResponse); @@ -337,6 +349,7 @@ message UpdateFilterRequest { // Databroker sends filters to provider. // In case provider restarts, databroker will send local filters stored // to continue the provider sending same signals with same filter. + // Filter value could be None, meaning all subscriptions were removed for that SignalId map filters_update = 2; } @@ -352,13 +365,13 @@ message ProviderErrorIndication { } message GetProviderValueRequest { - uint32 request_id = 1; /// Unique request id for the stream that can be used to match the corresponding response. - GetValueRequest request = 2; + uint32 request_id = 1; /// Unique request id for the stream that can be used to match the corresponding response. + repeated int32 signal_ids = 2; } message GetProviderValueResponse { - uint32 request_id = 1; /// Unique request id for the stream that can be used to match the corresponding request. - GetValueResponse response = 2; + uint32 request_id = 1; /// Unique request id for the stream that can be used to match the corresponding request. + map entries = 2; } message OpenProviderStreamRequest { @@ -377,7 +390,7 @@ message OpenProviderStreamRequest { // GetValue response GetProviderValueResponse get_provider_value_response = 6; // Indication of error on provider side - ProviderErrorIndication provider_error_indication = 7; + ProviderErrorIndication provider_error_indication = 7; } }