Skip to content

Commit 0fc8206

Browse files
committed
Changing "MarkLogic Spark connector"
Depending on the context, changed it to either "MarkLogic connector for Apache Spark", "MarkLogic connector", or just "connector". The one exception is the config.yml file the docs, where "MarkLogic connector for Apache Spark" did not fit. So using "MarkLogic Apache Spark connector" there instead of just "MarkLogic Spark connector". 4b45ec
1 parent 63df947 commit 0fc8206

14 files changed

+26
-27
lines changed

CONTRIBUTING.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ have cloned this repository to your local workstation.
33

44
# Do this first!
55

6-
In order to develop and/or test the MarkLogic Spark connector, or to try out the PySpark instructions below, you first
6+
In order to develop and/or test the connector, or to try out the PySpark instructions below, you first
77
need to deploy the test application in this project to MarkLogic. You can do so either on your own installation of
88
MarkLogic, or you can use `docker-compose` to install a 3-node MarkLogic cluster with a load balancer in front of it.
99

@@ -75,7 +75,7 @@ You can then run the tests from within the Docker environment via the following
7575

7676
The documentation for this project
7777
[has instructions on using PySpark](https://marklogic.github.io/marklogic-spark-connector/getting-started-pyspark.html)
78-
with the MarkLogic Spark connector. The documentation instructs a user to obtain the connector from this repository's
78+
with the connector. The documentation instructs a user to obtain the connector from this repository's
7979
releases page. For development and testing, you will most likely want to build the connector yourself by running the
8080
following command from the root of this repository:
8181

README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
1-
# MarkLogic Spark Connector
1+
# MarkLogic Connector for Apache Spark
22

3-
The MarkLogic Spark connector is an [Apache Spark 3 connector](https://spark.apache.org/docs/latest/) that supports
4-
reading data from and writing data to MarkLogic.
3+
The MarkLogic connector for Apache Spark is an [Apache Spark 3 connector](https://spark.apache.org/docs/latest/) that
4+
supports reading data from and writing data to MarkLogic.
55

66
Please see [the User Guide](http://marklogic.github.io/marklogic-spark-connector) for more information.
77

docs/_config.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
title: MarkLogic Spark Connector
1+
title: MarkLogic Apache Spark Connector
22
remote_theme: just-the-docs/just-the-docs
33
plugins:
44
- jekyll-remote-theme

docs/configuration.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ title: Configuration Reference
44
nav_order: 5
55
---
66

7-
The MarkLogic Spark connector has 3 sets of configuration options - connection options, reading options, and writing
7+
The MarkLogic connector has 3 sets of configuration options - connection options, reading options, and writing
88
options. Each set of options is defined in a separate table below.
99

1010
## Connection options

docs/getting-started/getting-started.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ has_children: true
66
permalink: /docs/getting-started
77
---
88

9-
This guide provides instructions on using the MarkLogic Spark connector with multiple popular Spark environments.
9+
This guide provides instructions on using the MarkLogic connector with multiple popular Spark environments.
1010
Before trying the connector in any of these environments, please [follow the instructions in the Setup guide](setup.md)
1111
to obtain the connector and deploy an example application to MarkLogic.
1212

docs/getting-started/java.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,12 +5,12 @@ parent: Getting Started
55
nav_order: 4
66
---
77

8-
The MarkLogic Spark connector is published to [Maven Central](https://central.sonatype.com/namespace/com.marklogic) and
8+
The MarkLogic connector is published to [Maven Central](https://central.sonatype.com/namespace/com.marklogic) and
99
can thus be expressed as a regular dependency of a Java application that also depends on the Spark APIs.
1010

1111
As an example, please see the project configuration in the
1212
[java-dependency example project](https://github.com/marklogic/marklogic-spark-connector/blob/master/examples/java-dependency)
13-
for how to depend on the MarkLogic Spark connector as a library. The `org.example.App` class in the project demonstrates
13+
for how to depend on the MarkLogic connector as a library. The `org.example.App` class in the project demonstrates
1414
a very simple Spark Java program for accessing the data in the application deployed via the [Setup guide](setup.md).
1515

1616
Note - if you are using Java 11 or higher, you may run into a `NoClassDefFoundError` for a class in the `javax.xml.bind`

docs/getting-started/jupyter.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ nav_order: 3
66
---
77

88
[Project Jupyter](https://jupyter.org/) provides a set of tools for working with notebooks, code, and data. The
9-
MarkLogic Spark connector can be easily integrated into these tools to allow users to access and analyze data in
9+
MarkLogic connector can be easily integrated into these tools to allow users to access and analyze data in
1010
MarkLogic.
1111

1212
Before going further, be sure you've followed the instructions in the [setup guide](setup.md) for
@@ -28,7 +28,7 @@ corner of the Notebook interface and select "Python 3 (ipykernel)" to create a n
2828
## Using the connector
2929

3030
In the first cell in the notebook created above, enter the following to allow Jupyter Notebook to access the MarkLogic
31-
Spark connector and also to initialize Spark:
31+
connector and also to initialize Spark:
3232

3333
```
3434
import os

docs/getting-started/pyspark.md

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -6,8 +6,7 @@ nav_order: 2
66
---
77

88
[PySpark](https://spark.apache.org/docs/latest/api/python/index.html) is a Python API for Spark and an excellent choice
9-
for learning how to use Spark. This guide describes how to install PySpark and use it with the MarkLogic Spark
10-
connector.
9+
for learning how to use Spark. This guide describes how to install PySpark and use it with the MarkLogic connector.
1110

1211
Before going further, be sure you've followed the instructions in the [Getting Started](getting-started.md) guide for
1312
obtaining the connector and deploying an example application to MarkLogic.
@@ -34,15 +33,15 @@ Run PySpark from the directory that you downloaded the connector to per the [set
3433

3534
The `--jars` command line option is PySpark's method for utilizing Spark connectors. Each Spark environment should have
3635
a similar mechanism for including third party connectors; please see the documentation for your particular Spark
37-
environment. In the example above, the `--jars` option allows for the MarkLogic Spark connector to be used within
36+
environment. In the example above, the `--jars` option allows for the connector to be used within
3837
PySpark.
3938

4039
When PySpark starts, you should see information like this on how to configure logging:
4140

4241
Setting default log level to "WARN".
4342
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
4443

45-
Setting the default log level to `INFO` or `DEBUG` will show logging from the MarkLogic Spark connector. This will also
44+
Setting the default log level to `INFO` or `DEBUG` will show logging from the MarkLogic connector. This will also
4645
include potentially significant amounts of log messages from PySpark itself.
4746

4847
### Reading data with the connector

docs/getting-started/setup.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ environments, as those examples depend on an application being deployed to MarkL
1010

1111
## Obtaining the connector
1212

13-
The MarkLogic Spark connector can be downloaded from
13+
The MarkLogic connector can be downloaded from
1414
[this repository's Releases page](https://github.com/marklogic/marklogic-spark-connector/releases). Each Spark
1515
environment should have documentation on how to include third-party connectors; please consult your Spark
1616
environment's documentation on how to achieve this.

docs/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ title: Introduction
44
nav_order: 1
55
---
66

7-
The MarkLogic Spark connector is an [Apache Spark 3 connector](https://spark.apache.org/docs/latest/) that supports
7+
The MarkLogic connector for Apache Spark is an [Apache Spark 3 connector](https://spark.apache.org/docs/latest/) that supports
88
reading data from and writing data to MarkLogic. Within any Spark 3 environment, the connector enables users to easily
99
query for data in MarkLogic, manipulate it using widely-known Spark operations, and then write results back to
1010
MarkLogic or disseminate them to another system. Data can also be easily imported into MarkLogic by first reading it

docs/reading.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ title: Reading Data
44
nav_order: 3
55
---
66

7-
The MarkLogic Spark connector allows for data to be retrieved from MarkLogic as rows via an
7+
The MarkLogic connector allows for data to be retrieved from MarkLogic as rows via an
88
[Optic query](https://docs.marklogic.com/guide/app-dev/OpticAPI#id_46710). The
99
sections below provide more detail on configuring how data is retrieved and converted into a Spark DataFrame.
1010

@@ -150,7 +150,7 @@ repository.
150150

151151
The Spark connector framework supports pushing down multiple operations to the connector data source. This can
152152
often provide a significant performance boost by allowing the data source to perform the operation, which can result in
153-
both fewer rows returned to Spark and less work for Spark to perform. The MarkLogic Spark connector supports pushing
153+
both fewer rows returned to Spark and less work for Spark to perform. The MarkLogic connector supports pushing
154154
down the following operations to MarkLogic:
155155

156156
- `count`

docs/writing.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ title: Writing Data
44
nav_order: 4
55
---
66

7-
The MarkLogic Spark connector allows for writing rows in a Spark DataFrame to MarkLogic as documents.
7+
The MarkLogic connector allows for writing rows in a Spark DataFrame to MarkLogic as documents.
88
The sections below provide more detail about how this process works and how it can be controlled.
99

1010
## Basic write operation
@@ -104,7 +104,7 @@ temporal collection.
104104

105105
## Streaming support
106106

107-
The MarkLogic Spark connector supports
107+
The connector supports
108108
[streaming writes](https://spark.apache.org/docs/latest/structured-streaming-programming-guide.html) to MarkLogic.
109109
The connector configuration does not change; instead, different Spark APIs are used to read a stream of data and
110110
write that stream to MarkLogic.
@@ -134,7 +134,7 @@ spark.readStream \
134134
.processAllAvailable()
135135
```
136136

137-
The above example will stream the data in the `./data/csv-files/100-employees.csv` file through the MarkLogic Spark
137+
The above example will stream the data in the `./data/csv-files/100-employees.csv` file through the
138138
connector and into MarkLogic. This will result 100 new JSON documents in the `streaming-example` collection.
139139

140140
The ability to stream data into MarkLogic can make Spark an effective tool for obtaining data from a variety of data
@@ -158,7 +158,7 @@ assist with debugging the cause of the error.
158158

159159
## Tuning performance
160160

161-
The MarkLogic Spark connector uses MarkLogic's
161+
The connector uses MarkLogic's
162162
[Data Movement SDK](https://docs.marklogic.com/guide/java/data-movement) for writing documents to a database. The
163163
following options can be set to adjust how the connector performs when writing data:
164164

@@ -178,7 +178,7 @@ resource consumption and throughput from Spark to MarkLogic.
178178

179179
Spark supports
180180
[several save modes](https://spark.apache.org/docs/latest/sql-data-sources-load-save-functions.html#save-modes)
181-
when writing data. The MarkLogic Spark connector requires the `append` mode to be used. Because Spark defaults to
181+
when writing data. The MarkLogic connector requires the `append` mode to be used. Because Spark defaults to
182182
the `error` mode, you will need to set this to `append` each time you use the connector to write data.
183183

184184
`append` is the only supported mode due to MarkLogic not having the concept of a single "table" that a document

examples/getting-started/marklogic-spark-getting-started.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@
5151
}
5252
],
5353
"source": [
54-
"# Make the MarkLogic Spark connector available to the underlying PySpark application.\n",
54+
"# Make the MarkLogic connector available to the underlying PySpark application.\n",
5555
"import os\n",
5656
"os.environ['PYSPARK_SUBMIT_ARGS'] = '--jars \"marklogic-spark-connector-2.0.0.jar\" pyspark-shell'\n",
5757
"\n",

examples/java-dependency/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
This project is a simple example of creating a Spark application in Java that depends on the MarkLogic Spark
1+
This project is a simple example of creating a Spark application in Java that depends on the MarkLogic
22
connector as a normal dependency expressed through Gradle.
33

44
Please see the [Java setup guide](https://marklogic.github.io/marklogic-spark-connector/docs/java)

0 commit comments

Comments
 (0)