You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+7-6Lines changed: 7 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -20,12 +20,13 @@ Spark-Redis also supports Spark Streaming (DStreams) and Structured Streaming.
20
20
The library has several branches, each corresponds to a different supported Spark version. For example, 'branch-2.3' works with any Spark 2.3.x version.
21
21
The master branch contains the recent development for the next release.
| model | defines the Redis model used to persist DataFrame, see [Persistence model](#persistence-model)|`enum [binary, hash]`|`hash`|
336
+
| filter.keys.by.type | make sure the underlying data structures match persistence model |`Boolean`|`false`|
337
+
| partitions.number | number of partitions (applies only when reading DataFrame) |`Int`|`3`|
338
338
| key.column | when writing - specifies unique column used as a Redis key, by default a key is auto-generated <br/> when reading - specifies column name to store hash key |`String`| - |
339
-
| ttl | data time to live in `seconds`. Data doesn't expire if `ttl` is less than `1`|`Int`|`0`|
340
-
| infer.schema | infer schema from random row, all columns will have `String` type |`Boolean`|`false`|
341
-
| max.pipeline.size | maximum number of commands per pipeline (used to batch commands) |`Int`| 100 |
342
-
| scan.count | count option of SCAN command (used to iterate over keys) |`Int`| 100 |
343
-
| iterator.grouping.size | the number of items to be grouped when iterating over underlying RDD partition |`Int`| 1000 |
344
-
| host | overrides `spark.redis.host` configured in SparkSession |`String`|`localhost`|
345
-
| port | overrides `spark.redis.port` configured in SparkSession |`Int`|`6379`|
| dbNum | overrides `spark.redis.db` configured in SparkSession |`Int`|`0`|
348
-
| timeout | overrides `spark.redis.timeout` configured in SparkSession |`Int`|`2000`|
339
+
| ttl | data time to live in `seconds`. Data doesn't expire if `ttl` is less than `1`|`Int`|`0`|
340
+
| infer.schema | infer schema from random row, all columns will have `String` type |`Boolean`|`false`|
341
+
| max.pipeline.size | maximum number of commands per pipeline (used to batch commands) |`Int`| 100 |
342
+
| scan.count | count option of SCAN command (used to iterate over keys) |`Int`| 100 |
343
+
| iterator.grouping.size | the number of items to be grouped when iterating over underlying RDD partition |`Int`| 1000 |
344
+
| host | overrides `spark.redis.host` configured in SparkSession (if set, any other connection setting from SparkSession is ignored ) |`String`|`localhost`|
345
+
| port | overrides `spark.redis.port` configured in SparkSession (if set, any other connection setting from SparkSession is ignored ) |`Int`|`6379`|
346
+
| user | overrides `spark.redis.user` configured in SparkSession (if set, any other connection setting from SparkSession is ignored ) |`String`| - |
347
+
| auth | overrides `spark.redis.auth` configured in SparkSession (if set, any other connection setting from SparkSession is ignored ) |`String`| - |
348
+
| dbNum | overrides `spark.redis.db` configured in SparkSession (if set, any other connection setting from SparkSession is ignored ) |`Int`|`0`|
349
+
| timeout | overrides `spark.redis.timeout` configured in SparkSession (if set, any other connection setting from SparkSession is ignored ) |`Int`|`2000`|
0 commit comments