A B C D E F G H I J L M N O P R S T U V W 
All Classes All Packages

A

abort(long, WriterCommitMessage[]) - Method in class com.mongodb.spark.sql.connector.write.MongoStreamingWrite
The write was aborted due to a failure.
AGGREGATION_ALLOW_DISK_USE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
Allow disk use when running the aggregation.
AGGREGATION_PIPELINE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
Provide a custom aggregation pipeline.
AGGREGATION_PIPELINE_DEFAULT - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
 
alterNamespace(String[], NamespaceChange...) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Altering databases is currently not supported, so alterNamespace will always throw an exception.
alterTable(Identifier, TableChange...) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Altering collections is currently not supported, so alterTable will always throw an exception.
Assertions - Class in com.mongodb.spark.sql.connector.assertions
Assertions to validate inputs

B

BsonDocumentToRowConverter - Class in com.mongodb.spark.sql.connector.schema
The helper for conversion of BsonDocuments to GenericRowWithSchema instances.
BsonDocumentToRowConverter() - Constructor for class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
Create a new instance with the default json writer settings
BsonDocumentToRowConverter(StructType) - Constructor for class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
Create a new instance with the default json writer settings
build() - Method in class com.mongodb.spark.sql.connector.read.MongoScanBuilder
 
buildForBatch() - Method in class com.mongodb.spark.sql.connector.write.MongoWriteBuilder
Returns a MongoBatchWrite to write data to batch source.
buildForStreaming() - Method in class com.mongodb.spark.sql.connector.write.MongoWriteBuilder
Returns a MongoStreamingWrite to write data to streaming source.

C

capabilities() - Method in class com.mongodb.spark.sql.connector.MongoTable
Returns the set of capabilities for this table.
CLIENT_FACTORY_CONFIG - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The MongoClientFactory configuration key
CLIENT_FACTORY_DEFAULT - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The default MongoClientFactory configuration value
close() - Method in class com.mongodb.spark.sql.connector.read.MongoPartitionReader
Closes this stream and releases any system resources associated with it.
close() - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
 
COLLECTION_NAME_CONFIG - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The collection name config
com.mongodb.spark.connector - package com.mongodb.spark.connector
 
com.mongodb.spark.sql.connector - package com.mongodb.spark.sql.connector
 
com.mongodb.spark.sql.connector.annotations - package com.mongodb.spark.sql.connector.annotations
 
com.mongodb.spark.sql.connector.assertions - package com.mongodb.spark.sql.connector.assertions
 
com.mongodb.spark.sql.connector.config - package com.mongodb.spark.sql.connector.config
 
com.mongodb.spark.sql.connector.connection - package com.mongodb.spark.sql.connector.connection
 
com.mongodb.spark.sql.connector.exceptions - package com.mongodb.spark.sql.connector.exceptions
 
com.mongodb.spark.sql.connector.read - package com.mongodb.spark.sql.connector.read
This package contains the sql connector read support.
com.mongodb.spark.sql.connector.read.partitioner - package com.mongodb.spark.sql.connector.read.partitioner
 
com.mongodb.spark.sql.connector.schema - package com.mongodb.spark.sql.connector.schema
 
com.mongodb.spark.sql.connector.write - package com.mongodb.spark.sql.connector.write
This package contains the sql connector write support.
commit(long, WriterCommitMessage[]) - Method in class com.mongodb.spark.sql.connector.write.MongoStreamingWrite
Logs the that the write has been committed
commit(Offset) - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
 
ConfigException - Exception in com.mongodb.spark.sql.connector.exceptions
A class for all config exceptions
ConfigException(String) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.ConfigException
Constructs a new instance.
ConfigException(String, Object, String) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.ConfigException
Constructs a new instance.
ConfigException(String, Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.ConfigException
Constructs a new instance.
ConfigException(Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.ConfigException
Constructs a new instance.
CONNECTION_STRING_CONFIG - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The connection string configuration key
CONNECTION_STRING_DEFAULT - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The default connection string configuration value
containsKey(String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Returns true if this map contains a mapping for the specified key.
create() - Method in class com.mongodb.spark.sql.connector.connection.DefaultMongoClientFactory
 
create() - Method in interface com.mongodb.spark.sql.connector.connection.MongoClientFactory
 
createConfig(Map<String, String>) - Static method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Create a Mongo Configuration that does not yet have a fixed use case
createContinuousReaderFactory() - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
 
createNamespace(String[], Map<String, String>) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Create a database.
createPartitionBounds(BsonValue, BsonValue) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
Creates the upper and lower boundary query
createPartitionPipeline(BsonDocument, List<BsonDocument>) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
Creates the aggregation pipeline for a partition.
createReader(InputPartition) - Method in class com.mongodb.spark.sql.connector.read.MongoPartitionReaderFactory
Returns a row-based partition reader to read data from the given MongoInputPartition.
createReader(InputPartition) - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReaderFactory
 
createReaderFactory() - Method in class com.mongodb.spark.sql.connector.read.MongoBatch
Returns a factory to create a PartitionReader for each InputPartition.
createStreamingWriterFactory(PhysicalWriteInfo) - Method in class com.mongodb.spark.sql.connector.write.MongoStreamingWrite
Creates the MongoDataWriterFactory instance will be serialized and sent to executors.
createTable(Identifier, StructType, Transform[], Map<String, String>) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Create a collection.

D

DATABASE_NAME_CONFIG - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The database name config
DataException - Exception in com.mongodb.spark.sql.connector.exceptions
A class for all data exceptions.
DataException(String) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.DataException
Constructs a new instance.
DataException(String, Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.DataException
Constructs a new instance.
DataException(Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.DataException
Constructs a new instance.
DefaultMongoClientFactory - Class in com.mongodb.spark.sql.connector.connection
The default MongoClientFactory implementation.
DefaultMongoClientFactory(MongoConfig) - Constructor for class com.mongodb.spark.sql.connector.connection.DefaultMongoClientFactory
Create a new instance of MongoClientFactory
description() - Method in class com.mongodb.spark.sql.connector.read.MongoScan
A description string of this scan.
deserializeOffset(String) - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
 
dropNamespace(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Drop a database.
dropTable(Identifier) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Drop a collection.

E

ensureArgument(Supplier<Boolean>, Supplier<String>) - Static method in class com.mongodb.spark.sql.connector.assertions.Assertions
Ensures the validity of arguments
ensureState(Supplier<Boolean>, Supplier<String>) - Static method in class com.mongodb.spark.sql.connector.assertions.Assertions
Ensures the validity of state
equals(Object) - Method in class com.mongodb.spark.sql.connector.connection.DefaultMongoClientFactory
 
equals(Object) - Method in class com.mongodb.spark.sql.connector.MongoTable
 
equals(Object) - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
 
equals(Object) - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenPartitionOffset
 

F

fromRow(InternalRow) - Method in class com.mongodb.spark.sql.connector.schema.RowToBsonDocumentConverter
Converts a Row to a BsonDocument
fromRow(Row) - Method in class com.mongodb.spark.sql.connector.schema.RowToBsonDocumentConverter
Converts a Row to a BsonDocument

G

generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.PaginateBySizePartitioner
 
generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.PaginateIntoPartitionsPartitioner
 
generatePartitions(ReadConfig) - Method in interface com.mongodb.spark.sql.connector.read.partitioner.Partitioner
Generate the partitions for the collection based upon the read configuration
generatePartitions(ReadConfig) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
 
generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.SamplePartitioner
 
generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.ShardedPartitioner
 
generatePartitions(ReadConfig) - Method in class com.mongodb.spark.sql.connector.read.partitioner.SinglePartitionPartitioner
 
get() - Method in class com.mongodb.spark.sql.connector.read.MongoPartitionReader
Return the current record.
get() - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
 
get(String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Returns the value to which the specified key is mapped
getAggregationAllowDiskUse() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
getAggregationPipeline() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
getBoolean(String, boolean) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Returns the boolean value to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getCollectionName() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
 
getConnectionString() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
 
getDatabaseName() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
 
getDouble(String, double) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Returns the double value to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getIdFields() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
 
getInferSchemaMapTypeMinimumKeySize() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
getInferSchemaSampleSize() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
getInt(String, int) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Returns the int value to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getList(String, List<String>) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Returns a list of strings from a comma delimited string to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getLong(String, long) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Returns the long value to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getMaxBatchSize() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
 
getMongoClient(MongoClientFactory) - Static method in class com.mongodb.spark.sql.connector.connection.LazyMongoClientCache
Returns a MongoClient from the cache.
getNamespace() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
 
getOffset() - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
 
getOperationType() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
 
getOptions() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
 
getOrDefault(String, String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Returns the value to which the specified key is mapped, or defaultValue if this config contains no mapping for the key.
getOriginals() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
 
getPartitioner() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
getPartitionerOptions() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
getPartitionId() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
 
getPipeline() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
 
getPreferredLocations(ReadConfig) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
 
getResumeToken() - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenOffset
 
getResumeTokenPartitionOffset() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
 
getSchema() - Method in class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
 
getStreamFullDocument() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
getTable(StructType, Transform[], Map<String, String>) - Method in class com.mongodb.spark.sql.connector.MongoTableProvider
Return a Table instance with the specified table schema, partitioning and properties to do read/write.
getWriteConcern() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
 

H

hashCode() - Method in class com.mongodb.spark.sql.connector.connection.DefaultMongoClientFactory
 
hashCode() - Method in class com.mongodb.spark.sql.connector.MongoTable
 
hashCode() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
 
hashCode() - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenPartitionOffset
 

I

ID_FIELD_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
A comma delimited field list used to identify a document
INFER_SCHEMA_MAP_TYPE_ENABLED_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
Enable Map Types when inferring the schema.
INFER_SCHEMA_MAP_TYPE_MINIMUM_KEY_SIZE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
The minimum size of a StructType before its inferred to a MapType instead.
INFER_SCHEMA_SAMPLE_SIZE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
The size of the sample of documents from the collection to use when inferring the schema
inferSchema(CaseInsensitiveStringMap) - Method in class com.mongodb.spark.sql.connector.MongoTableProvider
Infer the schema of the table identified by the given options.
inferSchema(CaseInsensitiveStringMap) - Static method in class com.mongodb.spark.sql.connector.schema.InferSchema
Infer the schema for the collection
InferSchema - Class in com.mongodb.spark.sql.connector.schema
A helper that determines the StructType for a BsonDocument and finds the common StructType for a list of BsonDocuments.
inferSchemaMapType() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
initialize(String, CaseInsensitiveStringMap) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Initializes the MongoCatalog.
initialOffset() - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
 
INSERT - com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
Insert the whole document
isOrdered() - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
 

J

json() - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenOffset
 

L

LazyMongoClientCache - Class in com.mongodb.spark.sql.connector.connection
A lazily initialized MongoClientCache.
listNamespaces() - Method in class com.mongodb.spark.sql.connector.MongoCatalog
List namespaces (databases).
listNamespaces(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
List namespaces (databases).
listTables(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
List the collections in a namespace (database).
loadNamespaceMetadata(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Currently only returns an empty map if the namespace (database) exists.
loadTable(Identifier) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Load collection.
LOGGER - Static variable in interface com.mongodb.spark.sql.connector.read.partitioner.Partitioner
 

M

matchQuery(List<BsonDocument>) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
Returns the head $match aggregation stage or an empty document.
MAX_BATCH_SIZE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
The maximum batch size for the batch in the bulk operation.
MAX_NUMBER_OF_PARTITIONS_CONFIG - Static variable in class com.mongodb.spark.sql.connector.read.partitioner.PaginateIntoPartitionsPartitioner
 
mergeOffsets(PartitionOffset[]) - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
 
MongoBatch - Class in com.mongodb.spark.sql.connector.read
MongoBatch defines how to read data from MongoDB.
MongoBatch(StructType, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoBatch
Construct a new instance
MongoCatalog - Class in com.mongodb.spark.sql.connector
Spark Catalog methods for working with namespaces (databases) and tables (collections).
MongoCatalog() - Constructor for class com.mongodb.spark.sql.connector.MongoCatalog
 
MongoClientFactory - Interface in com.mongodb.spark.sql.connector.connection
A factory for creating MongoClients
MongoConfig - Interface in com.mongodb.spark.sql.connector.config
The MongoConfig interface.
MongoContinuousStream - Class in com.mongodb.spark.sql.connector.read
MongoContinuousStream defines how to read a stream of data from MongoDB.
MongoContinuousStream(StructType, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoContinuousStream
Construct a new instance
MongoInputPartition - Class in com.mongodb.spark.sql.connector.read
The MongoInputPartition.
MongoInputPartition(int, List<BsonDocument>) - Constructor for class com.mongodb.spark.sql.connector.read.MongoInputPartition
Construct a new instance
MongoInputPartition(int, List<BsonDocument>, ResumeTokenPartitionOffset) - Constructor for class com.mongodb.spark.sql.connector.read.MongoInputPartition
Construct a new instance
MongoInputPartition(int, List<BsonDocument>, List<String>) - Constructor for class com.mongodb.spark.sql.connector.read.MongoInputPartition
Construct a new instance
MongoPartitionReader - Class in com.mongodb.spark.sql.connector.read
MongoPartitionReader(MongoInputPartition, BsonDocumentToRowConverter, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoPartitionReader
Construct a new instance
MongoPartitionReaderFactory - Class in com.mongodb.spark.sql.connector.read
A factory used to create MongoPartitionReader instances.
MongoPartitionReaderFactory(BsonDocumentToRowConverter, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoPartitionReaderFactory
Construct a new instance
MongoScan - Class in com.mongodb.spark.sql.connector.read
A logical representation of MongoDB data source scan.
MongoScan(StructType, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoScan
Construct a new instance
MongoScanBuilder - Class in com.mongodb.spark.sql.connector.read
A builder for a MongoScan.
MongoScanBuilder(StructType, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoScanBuilder
Construct a new instance
MongoSparkException - Exception in com.mongodb.spark.sql.connector.exceptions
A base class for all mongo spark exceptions.
MongoSparkException(String) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.MongoSparkException
Constructs a new instance.
MongoSparkException(String, Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.MongoSparkException
Constructs a new instance.
MongoSparkException(Throwable) - Constructor for exception com.mongodb.spark.sql.connector.exceptions.MongoSparkException
Constructs a new instance.
MongoStreamingWrite - Class in com.mongodb.spark.sql.connector.write
MongoStreamingWrite defines how to write the data to MongoDB when streaming data.
MongoStreamPartitionReader - Class in com.mongodb.spark.sql.connector.read
MongoStreamPartitionReader(MongoInputPartition, BsonDocumentToRowConverter, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
Construct a new instance
MongoStreamPartitionReaderFactory - Class in com.mongodb.spark.sql.connector.read
A factory used to create MongoStreamPartitionReader instances.
MongoStreamPartitionReaderFactory(BsonDocumentToRowConverter, ReadConfig) - Constructor for class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReaderFactory
Construct a new instance for a continuous stream
MongoTable - Class in com.mongodb.spark.sql.connector
Represents a MongoDB Collection.
MongoTable(MongoConfig) - Constructor for class com.mongodb.spark.sql.connector.MongoTable
Construct a new instance
MongoTable(StructType, MongoConfig) - Constructor for class com.mongodb.spark.sql.connector.MongoTable
Construct a new instance
MongoTable(StructType, Transform[], MongoConfig) - Constructor for class com.mongodb.spark.sql.connector.MongoTable
Construct a new instance
MongoTableProvider - Class in com.mongodb.spark.sql.connector
The MongoDB collection provider
MongoTableProvider() - Constructor for class com.mongodb.spark.sql.connector.MongoTableProvider
Construct a new instance
MongoWriteBuilder - Class in com.mongodb.spark.sql.connector.write
MongoWriteBuilder handles the creation of batch writer or streaming writers.
MongoWriteBuilder(LogicalWriteInfo, WriteConfig) - Constructor for class com.mongodb.spark.sql.connector.write.MongoWriteBuilder
Construct a new instance

N

name() - Method in class com.mongodb.spark.sql.connector.MongoCatalog
 
name() - Method in class com.mongodb.spark.sql.connector.MongoTable
The name of this table.
NAME - Static variable in class com.mongodb.spark.connector.Versions
 
namespaceExists(String[]) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Test whether a namespace (database) exists.
newScanBuilder(CaseInsensitiveStringMap) - Method in class com.mongodb.spark.sql.connector.MongoTable
Returns a MongoScanBuilder.
newWriteBuilder(LogicalWriteInfo) - Method in class com.mongodb.spark.sql.connector.MongoTable
next() - Method in class com.mongodb.spark.sql.connector.read.MongoPartitionReader
Proceed to next record, returns false if there is no more records.
next() - Method in class com.mongodb.spark.sql.connector.read.MongoStreamPartitionReader
 
NotThreadSafe - Annotation Type in com.mongodb.spark.sql.connector.annotations
The class to which this annotation is applied is not thread-safe.

O

OPERATION_TYPE_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
The write operation type to perform
ORDERED_BULK_OPERATION_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
Use ordered bulk operations

P

PaginateBySizePartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
Paginate by size partitioner.
PaginateBySizePartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.PaginateBySizePartitioner
Construct an instance
PaginateIntoPartitionsPartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
Paginate into partitions partitioner.
PaginateIntoPartitionsPartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.PaginateIntoPartitionsPartitioner
Construct an instance
parse(String) - Static method in class com.mongodb.spark.sql.connector.read.ResumeTokenOffset
Create a new instance from a json representation of a resume token
PARTITION_SIZE_MB_CONFIG - Static variable in class com.mongodb.spark.sql.connector.read.partitioner.PaginateBySizePartitioner
 
PARTITION_SIZE_MB_CONFIG - Static variable in class com.mongodb.spark.sql.connector.read.partitioner.SamplePartitioner
 
Partitioner - Interface in com.mongodb.spark.sql.connector.read.partitioner
The Partitioner provides the logic to partition a collection individual processable partitions.
PARTITIONER_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
The partitioner full class name.
PARTITIONER_DEFAULT - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
The default partitioner if none is set: "com.mongodb.spark.sql.connector.read.partitioner.SamplePartitioner"
PARTITIONER_OPTIONS_PREFIX - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
The prefix for specific partitioner based configuration.
PartitionerHelper - Class in com.mongodb.spark.sql.connector.read.partitioner
Partitioner helper class, contains various utility methods used by the partitioner instances.
partitioning() - Method in class com.mongodb.spark.sql.connector.MongoTable
 
planInputPartitions() - Method in class com.mongodb.spark.sql.connector.read.MongoBatch
Returns a list of partitions that split the collection into parts
planInputPartitions(Offset) - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
 
preferredLocations() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
The preferred locations for the read.
PREFIX - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The prefix for all general Spark MongoDB configurations.
properties() - Method in class com.mongodb.spark.sql.connector.MongoTable
 
pruneColumns(StructType) - Method in class com.mongodb.spark.sql.connector.read.MongoScanBuilder
 
pushedFilters() - Method in class com.mongodb.spark.sql.connector.read.MongoScanBuilder
 
pushFilters(Filter[]) - Method in class com.mongodb.spark.sql.connector.read.MongoScanBuilder
Processes filters on the dataset.

R

READ_PREFIX - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The prefix for specific output (read) based configurations.
readConfig(Map<String, String>) - Static method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Create a Read Configuration
ReadConfig - Class in com.mongodb.spark.sql.connector.config
The Read Configuration
readSchema() - Method in class com.mongodb.spark.sql.connector.read.MongoScan
Returns the actual schema of this data source scan, which may be different from the physical schema of the underlying storage, as column pruning or other optimizations may happen.
renameTable(Identifier, Identifier) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Renames a collection.
REPLACE - com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
Replace the whole document or insert a new one if it doesn't exist
ResumeTokenOffset - Class in com.mongodb.spark.sql.connector.read
An offset that contains a resume token from a change stream
ResumeTokenOffset(BsonDocument) - Constructor for class com.mongodb.spark.sql.connector.read.ResumeTokenOffset
Construct a new instance
ResumeTokenPartitionOffset - Class in com.mongodb.spark.sql.connector.read
A resume token partition offset
ResumeTokenPartitionOffset(BsonDocument) - Constructor for class com.mongodb.spark.sql.connector.read.ResumeTokenPartitionOffset
Construct a new instance
RowToBsonDocumentConverter - Class in com.mongodb.spark.sql.connector.schema
The helper for conversion of GenericRowWithSchema instances to BsonDocuments.
RowToBsonDocumentConverter() - Constructor for class com.mongodb.spark.sql.connector.schema.RowToBsonDocumentConverter
Construct a new instance
RowToBsonDocumentConverter(StructType) - Constructor for class com.mongodb.spark.sql.connector.schema.RowToBsonDocumentConverter
Construct a new instance

S

SamplePartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
Sample Partitioner
SamplePartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.SamplePartitioner
Construct an instance
schema() - Method in class com.mongodb.spark.sql.connector.MongoTable
Returns the schema of this table.
ShardedPartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
Sharded Partitioner
ShardedPartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.ShardedPartitioner
Construct an instance
shortName() - Method in class com.mongodb.spark.sql.connector.MongoTableProvider
 
SINGLE_PARTITIONER - Static variable in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
 
SinglePartitionPartitioner - Class in com.mongodb.spark.sql.connector.read.partitioner
Single Partition Partitioner
SinglePartitionPartitioner() - Constructor for class com.mongodb.spark.sql.connector.read.partitioner.SinglePartitionPartitioner
Construct an instance
stop() - Method in class com.mongodb.spark.sql.connector.read.MongoContinuousStream
 
storageStats(ReadConfig) - Static method in class com.mongodb.spark.sql.connector.read.partitioner.PartitionerHelper
 
STREAM_LOOKUP_FULL_DOCUMENT_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
Streaming full document configuration.
STREAM_PUBLISH_FULL_DOCUMENT_ONLY_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.ReadConfig
Publish Full Document only when streaming.
streamPublishFullDocumentOnly() - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
subConfiguration(String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Gets all configurations starting with a prefix.
supportsExternalMetadata() - Method in class com.mongodb.spark.sql.connector.MongoTableProvider
Returns true if the source has the ability of accepting external table metadata when getting tables.

T

tableExists(Identifier) - Method in class com.mongodb.spark.sql.connector.MongoCatalog
Test whether a collection exists.
ThreadSafe - Annotation Type in com.mongodb.spark.sql.connector.annotations
The class to which this annotation is applied is thread-safe.
toBatch() - Method in class com.mongodb.spark.sql.connector.read.MongoScan
Returns the physical representation of this scan for batch query.
toContinuousStream(String) - Method in class com.mongodb.spark.sql.connector.read.MongoScan
Returns the physical representation of this scan for streaming query with continuous mode.
toInternalRow(BsonDocument) - Method in class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
Converts a BsonDocument into a InternalRow.
toReadConfig() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
 
toString() - Method in enum com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
 
toString() - Method in class com.mongodb.spark.sql.connector.MongoTable
 
toString() - Method in class com.mongodb.spark.sql.connector.read.MongoInputPartition
 
toString() - Method in class com.mongodb.spark.sql.connector.read.ResumeTokenPartitionOffset
 
toString() - Method in class com.mongodb.spark.sql.connector.schema.BsonDocumentToRowConverter
 
toWriteConfig() - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
 
truncate() - Method in class com.mongodb.spark.sql.connector.write.MongoWriteBuilder
 

U

UPDATE - com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
Update the document or insert a new one if it doesn't exist

V

validateConfig(Supplier<T>, Supplier<String>) - Static method in class com.mongodb.spark.sql.connector.assertions.Assertions
Checks the validity of a value
validateConfig(T, Predicate<T>, Supplier<String>) - Static method in class com.mongodb.spark.sql.connector.assertions.Assertions
Checks the validity of a value
valueOf(String) - Static method in enum com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
Returns the enum constant of this type with the specified name.
values() - Static method in enum com.mongodb.spark.sql.connector.config.WriteConfig.OperationType
Returns an array containing the constants of this enum type, in the order they are declared.
VERSION - Static variable in class com.mongodb.spark.connector.Versions
 
Versions - Class in com.mongodb.spark.connector
 

W

withOption(String, String) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Return a MongoConfig instance with the extra options applied.
withOption(String, String) - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
withOption(String, String) - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
 
withOptions(Map<String, String>) - Method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Return a MongoConfig instance with the extra options applied.
withOptions(Map<String, String>) - Method in class com.mongodb.spark.sql.connector.config.ReadConfig
 
withOptions(Map<String, String>) - Method in class com.mongodb.spark.sql.connector.config.WriteConfig
 
WRITE_CONCERN_JOURNAL_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
The optional WriteConcern journal property.
WRITE_CONCERN_W_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
The optional WriteConcern w property.
WRITE_CONCERN_W_TIMEOUT_MS_CONFIG - Static variable in class com.mongodb.spark.sql.connector.config.WriteConfig
The optional WriteConcern wTimeout property in milliseconds.
WRITE_PREFIX - Static variable in interface com.mongodb.spark.sql.connector.config.MongoConfig
The prefix for specific input (write) based configurations.
writeConfig(Map<String, String>) - Static method in interface com.mongodb.spark.sql.connector.config.MongoConfig
Create a Write Configuration
WriteConfig - Class in com.mongodb.spark.sql.connector.config
The Write Configuration
WriteConfig.OperationType - Enum in com.mongodb.spark.sql.connector.config
The operation type for the write.
A B C D E F G H I J L M N O P R S T U V W 
All Classes All Packages