|
16 | 16 | heading = c("Univariate Distribution Categorical"), |
17 | 17 | engine = c("r"), |
18 | 18 | exceptionHandlingFunction = c(as.character(substitute(genericPipelineException))), |
19 | | - isDataFunction = T, |
| 19 | + isDataFunction = TRUE, |
20 | 20 | firstArgClass = "", |
21 | 21 | stringsAsFactors = F) |
22 | 22 |
|
|
25 | 25 | # outAsIn = FALSE, |
26 | 26 | engine = "r", |
27 | 27 | exceptionHandlingFunction = c(as.character(substitute(genericPipelineException))), |
28 | | - isDataFunction = T, |
| 28 | + isDataFunction = TRUE, |
29 | 29 | firstArgClass = "") -> .batchPredefFunctions |
30 | 30 | .batchPredefFunctions %>>% dplyr::add_row(functionName = "multiVarOutlierPlot", |
31 | 31 | heading = "Multivariate Outlier", |
|
37 | 37 | heading = "Ignore Columns", |
38 | 38 | engine = "r", |
39 | 39 | exceptionHandlingFunction = c(as.character(substitute(genericPipelineException))), |
40 | | - isDataFunction = T, |
| 40 | + isDataFunction = TRUE, |
41 | 41 | firstArgClass = "") -> .batchPredefFunctions |
42 | 42 | .batchPredefFunctions %>>% dplyr::add_row(functionName = "getFeaturesForPyClassification", |
43 | 43 | heading = "", |
|
49 | 49 | heading = "", |
50 | 50 | engine = "r", |
51 | 51 | exceptionHandlingFunction = c(as.character(substitute(genericPipelineException))), |
52 | | - isDataFunction = T, |
| 52 | + isDataFunction = TRUE, |
53 | 53 | firstArgClass = "") -> .batchPredefFunctions |
54 | 54 |
|
55 | 55 | ################################################################################################## |
|
66 | 66 | heading = c("Cast Kafka stream to a string"), |
67 | 67 | engine = c("spark-structured-streaming"), |
68 | 68 | exceptionHandlingFunction = c(as.character(substitute(genericPipelineException))), |
69 | | - isDataFunction = T, |
| 69 | + isDataFunction = TRUE, |
70 | 70 | firstArgClass = "", |
71 | 71 | stringsAsFactors = F) |
72 | 72 |
|
73 | 73 | .streamingPredefFunctions %>>% dplyr::add_row(functionName = "convertKafkaValueFromJson", |
74 | 74 | heading = "Convert Kafka Value from JSON", |
75 | 75 | engine = c("spark-structured-streaming"), |
76 | 76 | exceptionHandlingFunction = c(as.character(substitute(genericPipelineException))), |
77 | | - isDataFunction = T, |
| 77 | + isDataFunction = TRUE, |
78 | 78 | firstArgClass = "" |
79 | 79 | ) -> .streamingPredefFunctions |
80 | 80 |
|
|
0 commit comments