A B C D E F G H I J L M N O P R S T U V W
All Classes All Packages
All Classes All Packages
All Classes All Packages
A
- abort(long, WriterCommitMessage[]) - Method in class com.mongodb.spark.sql.connector.write.MongoStreamingWrite
-
The write was aborted due to a failure.
- AGGREGATION_ALLOW_DISK_USE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
Allow disk use when running the aggregation.
- AGGREGATION_PIPELINE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
Provide a custom aggregation pipeline.
- AGGREGATION_PIPELINE_DEFAULT - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
- alterNamespace(String[], NamespaceChange...) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Altering databases is currently not supported, so
alterNamespace
will always throw an exception. - alterTable(Identifier, TableChange...) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Altering collections is currently not supported, so
alterTable
will always throw an exception. - Assertions - Class in com.mongodb.spark.sql.connector.assertions
-
Assertions to validate inputs
B
- BsonDocumentToRowConverter - Class in com.mongodb.spark.sql.connector.schema
-
The helper for conversion of BsonDocuments to GenericRowWithSchema instances.
- BsonDocumentToRowConverter() - Constructor for class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
-
Create a new instance with the default json writer settings
- BsonDocumentToRowConverter(StructType) - Constructor for class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
-
Create a new instance with the default json writer settings
- build() - Method in class com.mongodb.spark.sql.connector.read.MongoScanBuilder
- buildForBatch() - Method in class com.mongodb.spark.sql.connector.write.MongoWriteBuilder
-
Returns a
MongoBatchWrite
to write data to batch source. - buildForStreaming() - Method in class com.mongodb.spark.sql.connector.write.MongoWriteBuilder
-
Returns a
MongoStreamingWrite
to write data to streaming source.
C
- capabilities() - Method in class com.mongodb.spark.sql.connector.MongoTable
-
Returns the set of capabilities for this table.
- CLIENT_FACTORY_CONFIG - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The MongoClientFactory configuration key
- CLIENT_FACTORY_DEFAULT - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The default MongoClientFactory configuration value
- close() - Method in class com.mongodb.spark.sql.connector.read.MongoPartitionReader
-
Closes this stream and releases any system resources associated with it.
- close() - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
- COLLECTION_NAME_CONFIG - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The collection name config
- com.mongodb.spark.connector - package com.mongodb.spark.connector
- com.mongodb.spark.sql.connector - package com.mongodb.spark.sql.connector
- com.mongodb.spark.sql.connector.annotations - package com.mongodb.spark.sql.connector.annotations
- com.mongodb.spark.sql.connector.assertions - package com.mongodb.spark.sql.connector.assertions
- com.mongodb.spark.sql.connector.config - package com.mongodb.spark.sql.connector.config
- com.mongodb.spark.sql.connector.connection - package com.mongodb.spark.sql.connector.connection
- com.mongodb.spark.sql.connector.exceptions - package com.mongodb.spark.sql.connector.exceptions
- com.mongodb.spark.sql.connector.read - package com.mongodb.spark.sql.connector.read
-
This package contains the sql connector read support.
- com.mongodb.spark.sql.connector.read.partitioner - package com.mongodb.spark.sql.connector.read.partitioner
- com.mongodb.spark.sql.connector.schema - package com.mongodb.spark.sql.connector.schema
- com.mongodb.spark.sql.connector.write - package com.mongodb.spark.sql.connector.write
-
This package contains the sql connector write support.
- commit(long, WriterCommitMessage[]) - Method in class com.mongodb.spark.sql.connector.write.MongoStreamingWrite
-
Logs the that the write has been committed
- commit(Offset) - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
- ConfigException - Exception in com.mongodb.spark.sql.connector.exceptions
-
A class for all config exceptions
- ConfigException(String) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.ConfigException
-
Constructs a new instance.
- ConfigException(String, Object, String) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.ConfigException
-
Constructs a new instance.
- ConfigException(String, Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.ConfigException
-
Constructs a new instance.
- ConfigException(Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.ConfigException
-
Constructs a new instance.
- CONNECTION_STRING_CONFIG - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The connection string configuration key
- CONNECTION_STRING_DEFAULT - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The default connection string configuration value
- containsKey(String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Returns
true
if this map contains a mapping for the specified key. - create() - Method in class com.mongodb.spark.sql.connector.connection.DefaultMongoClientFactory
- create() - Method in interface com.mongodb.spark.sql.connector.connection.MongoClientFactory
- createConfig(Map<String, String>) - Static method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Create a Mongo Configuration that does not yet have a fixed use case
- createContinuousReaderFactory() - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
- createNamespace(String[], Map<String, String>) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Create a database.
- createPartitionBounds(BsonValue, BsonValue) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
-
Creates the upper and lower boundary query
- createPartitionPipeline(BsonDocument, List<BsonDocument>) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
-
Creates the aggregation pipeline for a partition.
- createReader(InputPartition) - Method in class com.mongodb.spark.sql.connector.read.MongoPartitionReaderFactory
-
Returns a row-based partition reader to read data from the given
MongoInputPartition
. - createReader(InputPartition) - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReaderFactory
- createReaderFactory() - Method in class com.mongodb.spark.sql.connector.read.MongoBatch
-
Returns a factory to create a
PartitionReader
for eachInputPartition
. - createStreamingWriterFactory(PhysicalWriteInfo) - Method in class com.mongodb.spark.sql.connector.write.MongoStreamingWrite
-
Creates the MongoDataWriterFactory instance will be serialized and sent to executors.
- createTable(Identifier, StructType, Transform[], Map<String, String>) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Create a collection.
D
- DATABASE_NAME_CONFIG - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The database name config
- DataException - Exception in com.mongodb.spark.sql.connector.exceptions
-
A class for all data exceptions.
- DataException(String) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.DataException
-
Constructs a new instance.
- DataException(String, Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.DataException
-
Constructs a new instance.
- DataException(Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.DataException
-
Constructs a new instance.
- DefaultMongoClientFactory - Class in com.mongodb.spark.sql.connector.connection
-
The default MongoClientFactory implementation.
- DefaultMongoClientFactory(MongoConfig) - Constructor for class com.mongodb.spark.sql.connector.connection.DefaultMongoClientFactory
-
Create a new instance of MongoClientFactory
- description() - Method in class com.mongodb.spark.sql.connector.read.MongoScan
-
A description string of this scan.
- deserializeOffset(String) - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
- dropNamespace(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Drop a database.
- dropTable(Identifier) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Drop a collection.
E
- ensureArgument(Supplier<Boolean>, Supplier<String>) - Static method in class com.mongodb.spark.sql.connector.assertions.Assertions
-
Ensures the validity of arguments
- ensureState(Supplier<Boolean>, Supplier<String>) - Static method in class com.mongodb.spark.sql.connector.assertions.Assertions
-
Ensures the validity of state
- equals(Object) - Method in class com.mongodb.spark.sql.connector.connection.DefaultMongoClientFactory
- equals(Object) - Method in class com.mongodb.spark.sql.connector.MongoTable
- equals(Object) - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
- equals(Object) - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenPartitionOffset
F
- fromRow(InternalRow) - Method in class com.mongodb.spark.sql.connector.schema.RowToBsonDocumentConverter
-
Converts a
Row
to aBsonDocument
- fromRow(Row) - Method in class com.mongodb.spark.sql.connector.schema.RowToBsonDocumentConverter
-
Converts a
Row
to aBsonDocument
G
- generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.PaginateBySizePartitioner
- generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.PaginateIntoPartitionsPartitioner
- generatePartitions(ReadConfig) - Method in interface com.mongodb.spark.sql.connector.read.partitioner.Partitioner
-
Generate the partitions for the collection based upon the read configuration
- generatePartitions(ReadConfig) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
- generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.SamplePartitioner
- generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.ShardedPartitioner
- generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.SinglePartitionPartitioner
- get() - Method in class com.mongodb.spark.sql.connector.read.MongoPartitionReader
-
Return the current record.
- get() - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
- get(String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Returns the value to which the specified key is mapped
- getAggregationAllowDiskUse() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- getAggregationPipeline() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- getBoolean(String, boolean) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Returns the boolean value to which the specified key is mapped, or
defaultValue
if there is no mapping for the key. - getCollectionName() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
- getConnectionString() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
- getDatabaseName() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
- getDouble(String, double) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Returns the double value to which the specified key is mapped, or
defaultValue
if there is no mapping for the key. - getIdFields() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
- getInferSchemaMapTypeMinimumKeySize() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- getInferSchemaSampleSize() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- getInt(String, int) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Returns the int value to which the specified key is mapped, or
defaultValue
if there is no mapping for the key. - getList(String, List<String>) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Returns a list of strings from a comma delimited string to which the specified key is mapped, or
defaultValue
if there is no mapping for the key. - getLong(String, long) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Returns the long value to which the specified key is mapped, or
defaultValue
if there is no mapping for the key. - getMaxBatchSize() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
- getMongoClient(MongoClientFactory) - Static method in class com.mongodb.spark.sql.connector.connection.LazyMongoClientCache
-
Returns a
MongoClient
from the cache. - getNamespace() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
- getOffset() - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
- getOperationType() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
- getOptions() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
- getOrDefault(String, String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Returns the value to which the specified key is mapped, or
defaultValue
if this config contains no mapping for the key. - getOriginals() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
- getPartitioner() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- getPartitionerOptions() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- getPartitionId() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
- getPipeline() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
- getPreferredLocations(ReadConfig) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
- getResumeToken() - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenOffset
- getResumeTokenPartitionOffset() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
- getSchema() - Method in class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
- getStreamFullDocument() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- getTable(StructType, Transform[], Map<String, String>) - Method in class com.mongodb.spark.sql.connector.MongoTableProvider
-
Return a
Table
instance with the specified table schema, partitioning and properties to do read/write. - getWriteConcern() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
H
- hashCode() - Method in class com.mongodb.spark.sql.connector.connection.DefaultMongoClientFactory
- hashCode() - Method in class com.mongodb.spark.sql.connector.MongoTable
- hashCode() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
- hashCode() - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenPartitionOffset
I
- ID_FIELD_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
-
A comma delimited field list used to identify a document
- INFER_SCHEMA_MAP_TYPE_ENABLED_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
Enable Map Types when inferring the schema.
- INFER_SCHEMA_MAP_TYPE_MINIMUM_KEY_SIZE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
The minimum size of a
StructType
before its inferred to aMapType
instead. - INFER_SCHEMA_SAMPLE_SIZE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
The size of the sample of documents from the collection to use when inferring the schema
- inferSchema(CaseInsensitiveStringMap) - Method in class com.mongodb.spark.sql.connector.MongoTableProvider
-
Infer the schema of the table identified by the given options.
- inferSchema(CaseInsensitiveStringMap) - Static method in class com.mongodb.spark.sql.connector.schema.InferSchema
-
Infer the schema for the collection
- InferSchema - Class in com.mongodb.spark.sql.connector.schema
-
A helper that determines the
StructType
for aBsonDocument
and finds the commonStructType
for a list of BsonDocuments. - inferSchemaMapType() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- initialize(String, CaseInsensitiveStringMap) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Initializes the MongoCatalog.
- initialOffset() - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
- INSERT - com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
-
Insert the whole document
- isOrdered() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
J
- json() - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenOffset
L
- LazyMongoClientCache - Class in com.mongodb.spark.sql.connector.connection
-
A lazily initialized
MongoClientCache
. - listNamespaces() - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
List namespaces (databases).
- listNamespaces(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
List namespaces (databases).
- listTables(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
List the collections in a namespace (database).
- loadNamespaceMetadata(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Currently only returns an empty map if the namespace (database) exists.
- loadTable(Identifier) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Load collection.
- LOGGER - Static variable in interface com.mongodb.spark.sql.connector.read.partitioner.Partitioner
M
- matchQuery(List<BsonDocument>) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
-
Returns the head
$match
aggregation stage or an empty document. - MAX_BATCH_SIZE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
-
The maximum batch size for the batch in the bulk operation.
- MAX_NUMBER_OF_PARTITIONS_CONFIG - Static variable in class com.mongodb.spark.sql.connector.read.partitioner.PaginateIntoPartitionsPartitioner
- mergeOffsets(PartitionOffset[]) - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
- MongoBatch - Class in com.mongodb.spark.sql.connector.read
-
MongoBatch defines how to read data from MongoDB.
- MongoBatch(StructType, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoBatch
-
Construct a new instance
- MongoCatalog - Class in com.mongodb.spark.sql.connector
-
Spark Catalog methods for working with namespaces (databases) and tables (collections).
- MongoCatalog() - Constructor for class com.mongodb.spark.sql.connector.MongoCatalog
- MongoClientFactory - Interface in com.mongodb.spark.sql.connector.connection
-
A factory for creating MongoClients
- MongoConfig - Interface in com.mongodb.spark.sql.connector.config
-
The MongoConfig interface.
- MongoContinuousStream - Class in com.mongodb.spark.sql.connector.read
-
MongoContinuousStream defines how to read a stream of data from MongoDB.
- MongoContinuousStream(StructType, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoContinuousStream
-
Construct a new instance
- MongoInputPartition - Class in com.mongodb.spark.sql.connector.read
-
The MongoInputPartition.
- MongoInputPartition(int, List<BsonDocument>) - Constructor for class com.mongodb.spark.sql.connector.read.MongoInputPartition
-
Construct a new instance
- MongoInputPartition(int, List<BsonDocument>, ResumeTokenPartitionOffset) - Constructor for class com.mongodb.spark.sql.connector.read.MongoInputPartition
-
Construct a new instance
- MongoInputPartition(int, List<BsonDocument>, List<String>) - Constructor for class com.mongodb.spark.sql.connector.read.MongoInputPartition
-
Construct a new instance
- MongoPartitionReader - Class in com.mongodb.spark.sql.connector.read
-
A partition reader returned by
MongoPartitionReaderFactory.createReader(org.apache.spark.sql.connector.read.InputPartition)
. - MongoPartitionReader(MongoInputPartition, BsonDocumentToRowConverter, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoPartitionReader
-
Construct a new instance
- MongoPartitionReaderFactory - Class in com.mongodb.spark.sql.connector.read
-
A factory used to create
MongoPartitionReader
instances. - MongoPartitionReaderFactory(BsonDocumentToRowConverter, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoPartitionReaderFactory
-
Construct a new instance
- MongoScan - Class in com.mongodb.spark.sql.connector.read
-
A logical representation of MongoDB data source scan.
- MongoScan(StructType, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoScan
-
Construct a new instance
- MongoScanBuilder - Class in com.mongodb.spark.sql.connector.read
-
A builder for a
MongoScan
. - MongoScanBuilder(StructType, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoScanBuilder
-
Construct a new instance
- MongoSparkException - Exception in com.mongodb.spark.sql.connector.exceptions
-
A base class for all mongo spark exceptions.
- MongoSparkException(String) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.MongoSparkException
-
Constructs a new instance.
- MongoSparkException(String, Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.MongoSparkException
-
Constructs a new instance.
- MongoSparkException(Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.MongoSparkException
-
Constructs a new instance.
- MongoStreamingWrite - Class in com.mongodb.spark.sql.connector.write
-
MongoStreamingWrite defines how to write the data to MongoDB when streaming data.
- MongoStreamPartitionReader - Class in com.mongodb.spark.sql.connector.read
-
A partition reader returned by
MongoStreamPartitionReaderFactory.createReader(org.apache.spark.sql.connector.read.InputPartition)
. - MongoStreamPartitionReader(MongoInputPartition, BsonDocumentToRowConverter, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
-
Construct a new instance
- MongoStreamPartitionReaderFactory - Class in com.mongodb.spark.sql.connector.read
-
A factory used to create
MongoStreamPartitionReader
instances. - MongoStreamPartitionReaderFactory(BsonDocumentToRowConverter, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReaderFactory
-
Construct a new instance for a continuous stream
- MongoTable - Class in com.mongodb.spark.sql.connector
-
Represents a MongoDB Collection.
- MongoTable(MongoConfig) - Constructor for class com.mongodb.spark.sql.connector.MongoTable
-
Construct a new instance
- MongoTable(StructType, MongoConfig) - Constructor for class com.mongodb.spark.sql.connector.MongoTable
-
Construct a new instance
- MongoTable(StructType, Transform[], MongoConfig) - Constructor for class com.mongodb.spark.sql.connector.MongoTable
-
Construct a new instance
- MongoTableProvider - Class in com.mongodb.spark.sql.connector
-
The MongoDB collection provider
- MongoTableProvider() - Constructor for class com.mongodb.spark.sql.connector.MongoTableProvider
-
Construct a new instance
- MongoWriteBuilder - Class in com.mongodb.spark.sql.connector.write
-
MongoWriteBuilder handles the creation of batch writer or streaming writers.
- MongoWriteBuilder(LogicalWriteInfo, WriteConfig) - Constructor for class com.mongodb.spark.sql.connector.write.MongoWriteBuilder
-
Construct a new instance
N
- name() - Method in class com.mongodb.spark.sql.connector.MongoCatalog
- name() - Method in class com.mongodb.spark.sql.connector.MongoTable
-
The name of this table.
- NAME - Static variable in class com.mongodb.spark.connector.Versions
- namespaceExists(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Test whether a namespace (database) exists.
- newScanBuilder(CaseInsensitiveStringMap) - Method in class com.mongodb.spark.sql.connector.MongoTable
-
Returns a
MongoScanBuilder
. - newWriteBuilder(LogicalWriteInfo) - Method in class com.mongodb.spark.sql.connector.MongoTable
-
Returns a
MongoWriteBuilder
. - next() - Method in class com.mongodb.spark.sql.connector.read.MongoPartitionReader
-
Proceed to next record, returns false if there is no more records.
- next() - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
- NotThreadSafe - Annotation Type in com.mongodb.spark.sql.connector.annotations
-
The class to which this annotation is applied is not thread-safe.
O
- OPERATION_TYPE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
-
The write operation type to perform
- ORDERED_BULK_OPERATION_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
-
Use ordered bulk operations
P
- PaginateBySizePartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
-
Paginate by size partitioner.
- PaginateBySizePartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.PaginateBySizePartitioner
-
Construct an instance
- PaginateIntoPartitionsPartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
-
Paginate into partitions partitioner.
- PaginateIntoPartitionsPartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.PaginateIntoPartitionsPartitioner
-
Construct an instance
- parse(String) - Static method in class com.mongodb.spark.sql.connector.read.ResumeTokenOffset
-
Create a new instance from a json representation of a resume token
- PARTITION_SIZE_MB_CONFIG - Static variable in class com.mongodb.spark.sql.connector.read.partitioner.PaginateBySizePartitioner
- PARTITION_SIZE_MB_CONFIG - Static variable in class com.mongodb.spark.sql.connector.read.partitioner.SamplePartitioner
- Partitioner - Interface in com.mongodb.spark.sql.connector.read.partitioner
-
The Partitioner provides the logic to partition a collection individual processable partitions.
- PARTITIONER_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
The partitioner full class name.
- PARTITIONER_DEFAULT - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
The default partitioner if none is set: "com.mongodb.spark.sql.connector.read.partitioner.SamplePartitioner"
- PARTITIONER_OPTIONS_PREFIX - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
The prefix for specific partitioner based configuration.
- PartitionerHelper - Class in com.mongodb.spark.sql.connector.read.partitioner
-
Partitioner helper class, contains various utility methods used by the partitioner instances.
- partitioning() - Method in class com.mongodb.spark.sql.connector.MongoTable
- planInputPartitions() - Method in class com.mongodb.spark.sql.connector.read.MongoBatch
-
Returns a list of partitions that split the collection into parts
- planInputPartitions(Offset) - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
- preferredLocations() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
-
The preferred locations for the read.
- PREFIX - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The prefix for all general Spark MongoDB configurations.
- properties() - Method in class com.mongodb.spark.sql.connector.MongoTable
- pruneColumns(StructType) - Method in class com.mongodb.spark.sql.connector.read.MongoScanBuilder
- pushedFilters() - Method in class com.mongodb.spark.sql.connector.read.MongoScanBuilder
- pushFilters(Filter[]) - Method in class com.mongodb.spark.sql.connector.read.MongoScanBuilder
-
Processes filters on the dataset.
R
- READ_PREFIX - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The prefix for specific output (read) based configurations.
- readConfig(Map<String, String>) - Static method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Create a Read Configuration
- ReadConfig - Class in com.mongodb.spark.sql.connector.config
-
The Read Configuration
- readSchema() - Method in class com.mongodb.spark.sql.connector.read.MongoScan
-
Returns the actual schema of this data source scan, which may be different from the physical schema of the underlying storage, as column pruning or other optimizations may happen.
- renameTable(Identifier, Identifier) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Renames a collection.
- REPLACE - com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
-
Replace the whole document or insert a new one if it doesn't exist
- ResumeTokenOffset - Class in com.mongodb.spark.sql.connector.read
-
An offset that contains a resume token from a change stream
- ResumeTokenOffset(BsonDocument) - Constructor for class com.mongodb.spark.sql.connector.read.ResumeTokenOffset
-
Construct a new instance
- ResumeTokenPartitionOffset - Class in com.mongodb.spark.sql.connector.read
-
A resume token partition offset
- ResumeTokenPartitionOffset(BsonDocument) - Constructor for class com.mongodb.spark.sql.connector.read.ResumeTokenPartitionOffset
-
Construct a new instance
- RowToBsonDocumentConverter - Class in com.mongodb.spark.sql.connector.schema
-
The helper for conversion of GenericRowWithSchema instances to BsonDocuments.
- RowToBsonDocumentConverter() - Constructor for class com.mongodb.spark.sql.connector.schema.RowToBsonDocumentConverter
-
Construct a new instance
- RowToBsonDocumentConverter(StructType) - Constructor for class com.mongodb.spark.sql.connector.schema.RowToBsonDocumentConverter
-
Construct a new instance
S
- SamplePartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
-
Sample Partitioner
- SamplePartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.SamplePartitioner
-
Construct an instance
- schema() - Method in class com.mongodb.spark.sql.connector.MongoTable
-
Returns the schema of this table.
- ShardedPartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
-
Sharded Partitioner
- ShardedPartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.ShardedPartitioner
-
Construct an instance
- shortName() - Method in class com.mongodb.spark.sql.connector.MongoTableProvider
- SINGLE_PARTITIONER - Static variable in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
- SinglePartitionPartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
-
Single Partition Partitioner
- SinglePartitionPartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.SinglePartitionPartitioner
-
Construct an instance
- stop() - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
- storageStats(ReadConfig) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
- STREAM_LOOKUP_FULL_DOCUMENT_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
Streaming full document configuration.
- STREAM_PUBLISH_FULL_DOCUMENT_ONLY_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
-
Publish Full Document only when streaming.
- streamPublishFullDocumentOnly() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- subConfiguration(String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Gets all configurations starting with a prefix.
- supportsExternalMetadata() - Method in class com.mongodb.spark.sql.connector.MongoTableProvider
-
Returns true if the source has the ability of accepting external table metadata when getting tables.
T
- tableExists(Identifier) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
-
Test whether a collection exists.
- ThreadSafe - Annotation Type in com.mongodb.spark.sql.connector.annotations
-
The class to which this annotation is applied is thread-safe.
- toBatch() - Method in class com.mongodb.spark.sql.connector.read.MongoScan
-
Returns the physical representation of this scan for batch query.
- toContinuousStream(String) - Method in class com.mongodb.spark.sql.connector.read.MongoScan
-
Returns the physical representation of this scan for streaming query with continuous mode.
- toInternalRow(BsonDocument) - Method in class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
-
Converts a
BsonDocument
into aInternalRow
. - toReadConfig() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
- toString() - Method in enum com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
- toString() - Method in class com.mongodb.spark.sql.connector.MongoTable
- toString() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
- toString() - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenPartitionOffset
- toString() - Method in class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
- toWriteConfig() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
- truncate() - Method in class com.mongodb.spark.sql.connector.write.MongoWriteBuilder
U
- UPDATE - com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
-
Update the document or insert a new one if it doesn't exist
V
- validateConfig(Supplier<T>, Supplier<String>) - Static method in class com.mongodb.spark.sql.connector.assertions.Assertions
-
Checks the validity of a value
- validateConfig(T, Predicate<T>, Supplier<String>) - Static method in class com.mongodb.spark.sql.connector.assertions.Assertions
-
Checks the validity of a value
- valueOf(String) - Static method in enum com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
-
Returns the enum constant of this type with the specified name.
- values() - Static method in enum com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
-
Returns an array containing the constants of this enum type, in the order they are declared.
- VERSION - Static variable in class com.mongodb.spark.connector.Versions
- Versions - Class in com.mongodb.spark.connector
W
- withOption(String, String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Return a
MongoConfig
instance with the extra options applied. - withOption(String, String) - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- withOption(String, String) - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
- withOptions(Map<String, String>) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Return a
MongoConfig
instance with the extra options applied. - withOptions(Map<String, String>) - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
- withOptions(Map<String, String>) - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
- WRITE_CONCERN_JOURNAL_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
-
The optional
WriteConcern
journal property. - WRITE_CONCERN_W_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
-
The optional
WriteConcern
w property. - WRITE_CONCERN_W_TIMEOUT_MS_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
-
The optional
WriteConcern
wTimeout property in milliseconds. - WRITE_PREFIX - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
The prefix for specific input (write) based configurations.
- writeConfig(Map<String, String>) - Static method in interface com.mongodb.spark.sql.connector.config.MongoConfig
-
Create a Write Configuration
- WriteConfig - Class in com.mongodb.spark.sql.connector.config
-
The Write Configuration
- WriteConfig.OperationType - Enum in com.mongodb.spark.sql.connector.config
-
The operation type for the write.
All Classes All Packages