File tree Expand file tree Collapse file tree 4 files changed +21
-5
lines changed Expand file tree Collapse file tree 4 files changed +21
-5
lines changed Original file line number Diff line number Diff line change @@ -6,8 +6,8 @@ intersphinx = ["https://www.mongodb.com/docs/manual/objects.inv"]
6
6
toc_landing_pages = [" configuration" ]
7
7
8
8
[constants ]
9
- driver -short = " Spark Connector"
10
- driver -long = " MongoDB {+driver -short+}"
9
+ connector -short = " Spark Connector"
10
+ connector -long = " MongoDB {+connector -short+}"
11
11
current-version = " 10.2.0"
12
12
artifact-id-2-13 = " mongo-spark-connector_2.13"
13
13
artifact-id-2-12 = " mongo-spark-connector_2.12"
Original file line number Diff line number Diff line change @@ -133,7 +133,7 @@ You can configure the following properties to read from MongoDB:
133
133
Partitioner Configurations
134
134
~~~~~~~~~~~~~~~~~~~~~~~~~~
135
135
136
- Partitioners change the read behavior for batch reads with the {+driver -short+}.
136
+ Partitioners change the read behavior for batch reads with the {+connector -short+}.
137
137
They do not affect Structured Streaming because the data stream processing
138
138
engine produces a single stream with Structured Streaming.
139
139
@@ -330,9 +330,13 @@ Change Streams
330
330
- | Specifies whether to publish the changed document or the full
331
331
change stream document.
332
332
|
333
- | When set to ``true``, the connector filters out messages that
333
+ | When this setting is ``true``, the connector exhibits the following behavior:
334
+
335
+ - The connector filters out messages that
334
336
omit the ``fullDocument`` field and only publishes the value of the
335
337
field.
338
+ - If you don't specify a schema, the connector infers the schema
339
+ from the change stream document rather than from the underlying collection.
336
340
337
341
.. note::
338
342
Original file line number Diff line number Diff line change @@ -42,6 +42,18 @@ Overview
42
42
43
43
.. include:: /scala/filters.txt
44
44
45
+ .. important:: Inferring the Schema of a Change Stream
46
+
47
+ When the {+connector-short+} infers the schema of a data frame
48
+ read from a change stream, by default,
49
+ it will use the schema of the underlying collection rather than that
50
+ of the change stream. If you set the ``change.stream.publish.full.document.only``
51
+ option to ``true``, the connector uses the schema of the
52
+ change stream instead.
53
+
54
+ For more information on configuring a read operation, see the
55
+ :ref:`spark-change-stream-conf` section of the Read Configuration Options guide.
56
+
45
57
SQL Queries
46
58
-----------
47
59
Original file line number Diff line number Diff line change @@ -191,7 +191,7 @@ Configuring a Write Stream to MongoDB
191
191
192
192
Configuring a Read Stream from MongoDB
193
193
--------------------------------------
194
- When reading a stream from a MongoDB database, the {+driver -long+} supports both
194
+ When reading a stream from a MongoDB database, the {+connector -long+} supports both
195
195
*micro-batch processing* and
196
196
*continuous processing*. Micro-batch processing is the default processing engine, while
197
197
continuous processing is an experimental feature introduced in
You can’t perform that action at this time.
0 commit comments