You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
While this demo covers the use case of 1<sup>st</sup>-day churn, it is easy to replace the data, related features and training model and reuse the same workflow for different business cases.
203
+
204
+
These steps are covered by the following pipeline:
205
+
206
+
-**1. Data generator** — Generates events for the training and serving and Create an enrichment table (lookup values).
207
+
-**2. Event handler** - Receive data from the input. This is a common input stream for all the data. This way, one can easily replace the event source data (in this case we have a data generator) without affecting the rest of this flow. It also store all incoming data to parquet files.
208
+
-**3. Stream to features** - Enrich the stream using the enrichment table and Update aggregation features using the incoming event handler.
209
+
-**4. Optional model training steps -**
210
+
-**4.1 Get Data Snapshot** - Takes a snapshot of the feature table for training.
211
+
-**4.2 Describe the Dataset** - Runs common analysis on the datasets and produces plots suche as histogram, feature importance, corollation and more.
212
+
-**4.3 Training** - Runing training with multiple classification models.
213
+
-**4.4 Testing** - Testing the best performing model.
214
+
-**5. Serving** - Serve the model and process the data from the enriched stream and aggregation features.
215
+
-**6. Inference logger** - We use the same event handler function from above but only its capability to store incoming data to parquet files.
0 commit comments