Index
All Classes and Interfaces|All Packages|Serialized Form
A
- add(RdfSource) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceCollection
- add(RdfSource) - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- aggregateUsingJavaCollector(JavaRDD<? extends T>, Collector<? super T, A, R>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOps
-
Aggregate a JavaRDD using a serializable Collector.
- allowOverwriteFiles - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- andThen(JavaPairRddFunction<K, V, KO, VO>) - Method in interface net.sansa_stack.spark.rdd.function.ToJavaPairRddFunction
- andThen(JavaPairRddFunction<KO, VO, KX, VX>) - Method in interface net.sansa_stack.spark.rdd.function.JavaPairRddFunction
- andThen(JavaRddFunction<O, X>) - Method in interface net.sansa_stack.spark.rdd.function.JavaRddFunction
- andThen(JavaRddFunction<O, X>) - Method in interface net.sansa_stack.spark.rdd.function.ToJavaRddFunction
- applyDefaults() - Method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- ARRAY - Enum constant in enum class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonSourceType
- asDatasets() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSource
-
A stream of datasets having one named graph each
- asDatasets() - Method in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- asDatasets() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- asDatasets() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- asJavaRdd(RDD<Tuple2<LongWritable, T>>) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaders
-
Tiny helper to get the desired JavaRDD
- asJavaRddFunction(RxFunction<I, O>) - Static method in class net.sansa_stack.spark.rdd.op.rx.JavaRddRxOps
-
Wrap a
RxFunction
as aJavaRddFunction
- asModels() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSource
- asModels() - Method in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- asModels() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- asModels() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- asQuads() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSource
- asQuads() - Method in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- asQuads() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- asQuads() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- asTriples() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSource
-
Return the backing loader; null if unknown
- asTriples() - Method in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- asTriples() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- asTriples() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- AsyncRdfParserHadoop - Class in net.sansa_stack.spark.io.rdf.loader
-
Async parsing RDF on a single node using hadoop
- AsyncRdfParserHadoop() - Constructor for class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop
- AsyncRdfParserHadoop.Builder<T> - Class in net.sansa_stack.spark.io.rdf.loader
B
- bindingMapper(Var) - Static method in class net.sansa_stack.spark.io.json.input.JsonDataSources
-
Convert a JavaRDD>JsonElement< into a JavaRDD>Binding< by means of converting JSON elements into Nodes (primitive JSON will become native RDF!) and adding them to bindings with the given outputVar.
- bindSecondArgument(A2, BiFunction<I, A2, O>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
-
Create a Function from a BiFunction by binding the second argument to a given value
- bindToExecCxt(ExecutionContext, BiFunction<I, ExecutionContext, O>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- buildBaseConfiguration(RDD<?>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddWriterUtils
- buildConfiguration(RDD<?>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter2
- Builder(Configuration, Path, InputFormat<?, T>, RefSupplier<ExecutorService>, StreamRDF, BiConsumer<T, StreamRDF>) - Constructor for class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
C
- clazz - Variable in class net.sansa_stack.spark.io.rdf.input.api.RddLoaderBase
- clearDataset() - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- clone() - Method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- close() - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- collectToTable() - Method in interface net.sansa_stack.query.spark.api.domain.JavaResultSetSpark
-
Load the whole result set into an in-memory Jena table
- ColumnNamingScheme - Interface in net.sansa_stack.spark.io.csv.input
- compileNodeTupleMapper(Query, Function<Template, Function<Binding, Stream<T>>>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
-
Compile a construct query into a function that can efficiently produce triples/quads from a given binding.
- compileTarqlMapper(List<SparqlStmt>, boolean) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- compileTarqlMapperGeneral(Collection<SparqlStmt>, boolean) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- componentCount - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- conf - Variable in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- conf - Variable in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- configuration - Variable in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- configure(Configuration) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter2
- configureFrom(RddRdfWriterSettings<?>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- configureFrom(RddRowSetWriterSettings<?>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterSettings
- configureFrom(RddWriterSettings<?>) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- configureHadoop(Configuration, String, UnivocityCsvwConf, List<String>, Function<String[][], Function<String[], Binding>>) - Static method in class net.sansa_stack.spark.io.csv.input.CsvDataSources
- consoleOutSupplier - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- convertToNode - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- convertToNode(JavaRDD<T>) - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfOps
- convertToNode(JavaRDD<T>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- convertToQuad - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- convertToQuad(JavaRDD<T>) - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfOps
- convertToQuad(JavaRDD<T>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- convertToTriple - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- convertToTriple(JavaRDD<T>) - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfOps
- convertToTriple(JavaRDD<T>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- copyFrom(RdfPostProcessingSettings) - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsMutable
- copyInto(RdfPostProcessingSettingsMutable) - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettings
- create() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- create() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterFactory
- create(int, SerializableBiConsumer<T, StreamRDF>, JavaRddFunction<T, Triple>, JavaRddFunction<T, Quad>, JavaRddFunction<T, Node>, SerializableFunction<? super T, Comparable<?>>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
-
Create method.
- create(Class<T>, Class<? extends FileInputFormat<LongWritable, T>>) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaders
-
Create a default RddRdfLoader
- create(String[]) - Method in interface net.sansa_stack.spark.io.csv.input.ColumnNamingScheme
- create(Configuration, SerializableSupplier<LinkSparqlUpdate>) - Static method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- create(Path, FileSystem, Lang) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFactory
- create(Path, FileSystem, Lang) - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactoryImpl
- create(InputFormat<?, X>, BiConsumer<X, StreamRDF>) - Static method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- createColumnHeadings(List<String>, String[], boolean) - Static method in interface net.sansa_stack.spark.io.csv.input.ColumnNamingScheme
- createDefault() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFormatRegistry
- createForDataset() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- createForDataset() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- createForDatasetGraph() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- createForDatasetGraph() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- createForGraph() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- createForGraph() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- createForModel() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- createForModel() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- createForQuad() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- createForQuad() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- createForTriple() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- createForTriple() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- createJavaRdd(SparkContext, String, Class<T>, Class<? extends FileInputFormat<LongWritable, T>>) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaders
- createRdd(JavaSparkContext, HadoopInputData<K, V, X>) - Static method in class net.sansa_stack.spark.io.rdf.input.api.InputFormatUtils
- createRdd(SparkContext, String, Class<T>, Class<? extends FileInputFormat<LongWritable, T>>) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaders
- createRddFromJson(JavaSparkContext, String, int, Var) - Static method in class net.sansa_stack.spark.io.json.input.JsonDataSources
- createRddOfBindings(JavaSparkContext, String, UnivocityCsvwConf) - Static method in class net.sansa_stack.spark.io.csv.input.CsvDataSources
- createRddOfBindings(JavaSparkContext, String, UnivocityCsvwConf, List<String>) - Static method in class net.sansa_stack.spark.io.csv.input.CsvDataSources
- createRddOfDatasetCore(SparkContext, String, Class<T>, Class<? extends FileInputFormat<LongWritable, T>>) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaders
- createStreamRDFFactory(RDFFormat, boolean, PrefixMapping) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
-
Create a function that can create a StreamRDF instance that is backed by the given OutputStream.
- CsvDataSources - Class in net.sansa_stack.spark.io.csv.input
- CsvDataSources() - Constructor for class net.sansa_stack.spark.io.csv.input.CsvDataSources
- CsvRowMapper - Interface in net.sansa_stack.spark.io.csv.input
- CsvRowMapperFactories - Class in net.sansa_stack.spark.io.csv.input
- CsvRowMapperFactories() - Constructor for class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
- CsvRowMapperFactory - Interface in net.sansa_stack.spark.io.csv.input
D
- datasets(RDD<DatasetOneNg>) - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfProcessor
- declaredPrefixes - Variable in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- deferOutputForUsedPrefixes - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
-
Only for console output: Instead of writing tuples out immediatly, collect up to this number of tuples in order to derive the used prefixes.
- delegate - Variable in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- delete() - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- delete(Node) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- deletePartitionFolderAfterMerge - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- detectedType - Variable in class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonProbeResult
- dispatcher - Variable in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- dispatcher - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
-
References the lambdas in RddRdfOpsImpl directly (saves one entry in the call stack per record)
- distinct - Variable in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- distinctPartitions - Variable in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
E
- enrichRddWithRowNum(JavaRDD<Binding>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- excel(int) - Static method in interface net.sansa_stack.spark.io.csv.input.ColumnNamingScheme
- exceptions - Variable in class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonProbeResult
- executorServiceRef - Variable in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
F
- fileInputFormatClass - Variable in class net.sansa_stack.spark.io.rdf.input.api.RddLoaderBase
- find(Lang, Class<T>) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RddRdfLoaderRegistry
-
Search the registry for a loader that can supply a resource of language 'lang' as records of type rdfType.
- find(Lang, Class<T>) - Method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderRegistryImpl
- flatMapToNamedModels(JavaRDD<? extends Dataset>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfDatasetsOps
- flatMapToQuads(JavaRDD<? extends Dataset>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfDatasetsOps
- flatMapToTriples(JavaRDD<? extends Dataset>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfDatasetsOps
-
Maps a dataset to triples - emits quads from named graphs as triples by dropping the named graph
- forDataset(JavaRDD<? extends DatasetOneNg>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- forDatasetGraph(JavaRDD<? extends DatasetGraphOneNg>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- forGraph(JavaRDD<? extends Graph>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- formatClass - Variable in class net.sansa_stack.spark.io.common.HadoopFormat
- forModel(JavaRDD<? extends Model>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- forQuad() - Static method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- forQuad(JavaRDD<? extends Quad>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- forRowSet(JavaResultSetSpark) - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterFactory
- forTriple() - Static method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- forTriple(JavaRDD<? extends Triple>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- from(SparkSession) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactoryImpl
-
Deprecated.
- fromRdd(JavaRDD<?>) - Static method in class net.sansa_stack.spark.util.JavaSparkContextUtils
G
- get() - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderRegistryImpl
- get() - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- get(String) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFactory
- get(String, FileSystem) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFactory
- get(String, Lang) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFactory
- get(Path, FileSystem, Lang) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFactory
- get(Path, Lang) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFactory
- get(Node) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- get(Lang) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFormatRegistry
- getComponentCount() - Method in interface net.sansa_stack.spark.io.rdf.input.api.NodeTupleSource
- getComponentCount() - Method in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- getComponentCount() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromResource
- getComponentCount() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- getComponentCount() - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfOps
-
Whether the operations provided by this class are based on quads
- getComponentCount() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- getConfiguration() - Method in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- getConsoleOutSupplier() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- getDataset() - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- getDefaultFileSystem() - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactoryImpl
- getDefaults() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFormatRegistry
-
The default registry for elephas output formats based on
Lang
. - getDelegate() - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- getDetectedType() - Method in class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonProbeResult
- getDistinct() - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettings
- getDistinct() - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- getDistinctPartitions() - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettings
- getDistinctPartitions() - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- getEffectiveRdd(RdfPostProcessingSettings) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
-
Create the effective RDD w.r.t.
- getExceptions() - Method in class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonProbeResult
- getFallbackOutputFormat() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- getFileInputFormatClass() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RddLoader
- getFileInputFormatClass() - Method in class net.sansa_stack.spark.io.rdf.input.api.RddLoaderBase
- getFormatClass() - Method in class net.sansa_stack.spark.io.common.HadoopFormat
- getGlobalPrefixMapping() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- getHadoopConfiguration() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- getInputFormat() - Method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- getInputFormatClass() - Method in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- getInstance() - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFormatRegistry
- getKeyClass() - Method in class net.sansa_stack.spark.io.common.HadoopFormat
- getKeyClass() - Method in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- getKeyFunction() - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfOps
- getKeyFunction() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- getLang() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromResource
-
Get the language of the source; may have been probed for by an RdfSourceFactory
- getLang() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- getMapper() - Method in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- getMembers() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceCollection
- getMembers() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- getOptimizePrefixes() - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettings
- getOptimizePrefixes() - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- getOutputFormat() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- getOutputLang() - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterSettings
- getPartitionFolder() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- getPartitionFolderFs() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- getPath() - Method in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- getPath() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- getPostProcessingSettings() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- getRdd() - Method in interface net.sansa_stack.query.spark.api.domain.JavaResultSetSpark
- getRdd() - Method in class net.sansa_stack.query.spark.api.domain.JavaResultSetSparkImpl
- getRdd() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- getResultVars() - Method in interface net.sansa_stack.query.spark.api.domain.JavaResultSetSpark
- getResultVars() - Method in class net.sansa_stack.query.spark.api.domain.JavaResultSetSparkImpl
- getRowSet() - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
- getSession(JavaSparkContext) - Static method in class net.sansa_stack.spark.util.JavaSparkContextUtils
- getSort() - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettings
- getSort() - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- getSortAscending() - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettings
- getSortAscending() - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- getSortPartitions() - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettings
- getSortPartitions() - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- getTargetFile() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- getTargetFileFs() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- getValueClass() - Method in class net.sansa_stack.spark.io.common.HadoopFormat
- getValueClass() - Method in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- getValueClass() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RddLoader
- getValueClass() - Method in class net.sansa_stack.spark.io.rdf.input.api.RddLoaderBase
- globalPrefixMapping - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- GRAPH - Enum constant in enum class net.sansa_stack.spark.io.rdf.input.api.RdfSourceRecordType
- groupBy(JavaRDD<Triple>, SerializableFunction<? super Triple, K>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
- groupByNamedGraph(JavaRDD<Quad>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfQuadsOps
-
group quads by graph IRI into a pairs (graphIri, Model)
- groupByObjectNodes(JavaRDD<Triple>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
- groupByObjects(JavaRDD<Triple>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
- groupByPredicateNodes(JavaRDD<Triple>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
-
This function mainly exists for completeness
- groupByPredicates(JavaRDD<Triple>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
-
This function mainly exists for completeness
- groupBySubjectNodes(JavaRDD<Triple>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
- groupBySubjects(JavaRDD<Triple>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
- groupKeysAndReduceValues(JavaPairRDD<K, V>, boolean, boolean, int, Function2<V, V, V>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOps
-
Convenience helper to group values by keys, optionally sort them and reduce the values.
- groupNamedGraphsByGraphIri(JavaRDD<? extends Dataset>, boolean, boolean, int) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfDatasetsOps
-
Group all graphs by their named graph IRIs.
- groupNamedModels(JavaPairRDD<K, Model>, boolean, boolean, int) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfNamedModelsOps
-
Group and/or sort named models by their graph iri
- groupTriplesIntoModels(JavaPairRDD<K, Triple>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
H
- hadoopConfiguration - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- hadoopConfiguration - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- HadoopFormat<T> - Class in net.sansa_stack.spark.io.common
-
This class bundles hadoop format information: keyClass, valueClass and formatClass.
- HadoopFormat(Class<?>, Class<?>, Class<? extends T>) - Constructor for class net.sansa_stack.spark.io.common.HadoopFormat
- HadoopInputData<K,
V, X> - Class in net.sansa_stack.spark.io.rdf.input.api -
A class to capture the arguments of
JavaSparkContext.newAPIHadoopFile(String, Class, Class, Class, Configuration)
. - HadoopInputData(String, Class<? extends InputFormat<K, V>>, Class<K>, Class<V>, Configuration, Function<JavaPairRDD<K, V>, X>) - Constructor for class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- HadoopInputFormat - Class in net.sansa_stack.spark.io.common
- HadoopInputFormat(Class<?>, Class<?>, Class<? extends InputFormat>) - Constructor for class net.sansa_stack.spark.io.common.HadoopInputFormat
- HadoopOutputFormat - Class in net.sansa_stack.spark.io.common
- HadoopOutputFormat(Class<?>, Class<?>, Class<? extends OutputFormat>) - Constructor for class net.sansa_stack.spark.io.common.HadoopOutputFormat
- headerToVars(String[][]) - Static method in class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
I
- identity() - Static method in interface net.sansa_stack.spark.rdd.function.JavaPairRddFunction
- identity() - Static method in interface net.sansa_stack.spark.rdd.function.JavaRddFunction
- inputFile - Variable in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- inputFormat - Variable in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- inputFormatClass - Variable in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- InputFormatUtils - Class in net.sansa_stack.spark.io.rdf.input.api
- InputFormatUtils() - Constructor for class net.sansa_stack.spark.io.rdf.input.api.InputFormatUtils
- isAllowOverwriteFiles() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- isClosed() - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- isConsoleOutput() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
-
If neither partition folder nor targe file is set the output goes to the console
- isDeletePartitionFolderAfterMerge() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- isEmpty() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceCollection
- isEmpty() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- isMapQuadsToTriplesForTripleLangs() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- isPartitionsAsIndependentFiles() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- isPartitionsAsIndependentFiles() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- isUseCoalesceOne() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- isValidLang(Lang) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterUtils
-
Check if the given lang has a registered hadoop writer
J
- JavaPairRddFunction<KI,
VI, KO, VO> - Interface in net.sansa_stack.spark.rdd.function -
Interface for building chains of transformations over JavaRDDs and JavaPairRDDs.
- JavaRddConsumer<T> - Interface in net.sansa_stack.spark.rdd.function
-
Consumer interface for JavaRDDs.
- JavaRddFunction<I,
O> - Interface in net.sansa_stack.spark.rdd.function -
Interface for building chains of transformations over JavaRDDs and JavaPairRDDs.
- JavaRddOfBindingsOps - Class in net.sansa_stack.spark.rdd.op.rdf
- JavaRddOfBindingsOps() - Constructor for class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- JavaRddOfDatasetsOps - Class in net.sansa_stack.spark.rdd.op.rdf
- JavaRddOfDatasetsOps() - Constructor for class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfDatasetsOps
- JavaRddOfNamedModelsOps - Class in net.sansa_stack.spark.rdd.op.rdf
-
Operations on the RDD[(String, Model)] type.
- JavaRddOfNamedModelsOps() - Constructor for class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfNamedModelsOps
- JavaRddOfQuadsOps - Class in net.sansa_stack.spark.rdd.op.rdf
- JavaRddOfQuadsOps() - Constructor for class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfQuadsOps
- JavaRddOfResourcesOps - Class in net.sansa_stack.spark.rdd.op.rdf
- JavaRddOfResourcesOps() - Constructor for class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfResourcesOps
- JavaRddOfTriplesOps - Class in net.sansa_stack.spark.rdd.op.rdf
- JavaRddOfTriplesOps() - Constructor for class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
- JavaRddOps - Class in net.sansa_stack.spark.rdd.op.rdf
- JavaRddOps() - Constructor for class net.sansa_stack.spark.rdd.op.rdf.JavaRddOps
- JavaRddRxOps - Class in net.sansa_stack.spark.rdd.op.rx
- JavaRddRxOps() - Constructor for class net.sansa_stack.spark.rdd.op.rx.JavaRddRxOps
- JavaResultSetSpark - Interface in net.sansa_stack.query.spark.api.domain
-
An interface to represent a SPARQL result set by bundling a (Java)RDD of Bindings together with a list of result variables.
- JavaResultSetSparkImpl - Class in net.sansa_stack.query.spark.api.domain
- JavaResultSetSparkImpl(List<Var>, JavaRDD<Binding>) - Constructor for class net.sansa_stack.query.spark.api.domain.JavaResultSetSparkImpl
- JavaSparkContextUtils - Class in net.sansa_stack.spark.util
- JavaSparkContextUtils() - Constructor for class net.sansa_stack.spark.util.JavaSparkContextUtils
- JenaKryoRegistrator - Class in net.sansa_stack.spark.io.rdf.kryo
- JenaKryoRegistrator() - Constructor for class net.sansa_stack.spark.io.rdf.kryo.JenaKryoRegistrator
- jsonArray(String, Configuration) - Static method in class net.sansa_stack.spark.io.json.input.JsonDataSources
- JsonDataSources - Class in net.sansa_stack.spark.io.json.input
- JsonDataSources() - Constructor for class net.sansa_stack.spark.io.json.input.JsonDataSources
- JsonDataSources.JsonProbeResult - Class in net.sansa_stack.spark.io.json.input
- JsonDataSources.JsonSourceType - Enum Class in net.sansa_stack.spark.io.json.input
- JsonProbeResult(JsonDataSources.JsonSourceType, Map<JsonDataSources.JsonSourceType, Throwable>) - Constructor for class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonProbeResult
- jsonSequence(String, Configuration) - Static method in class net.sansa_stack.spark.io.json.input.JsonDataSources
K
- keyClass - Variable in class net.sansa_stack.spark.io.common.HadoopFormat
- keyClass - Variable in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- keyFunction - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
L
- lang - Variable in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- LinkDatasetGraphSansa - Class in net.sansa_stack.spark.io.rdf.loader
-
A
LinkDatasetGraph
implementation that loads files via the sansa parser and sends the data (triples and quads) to a sink. - LinkDatasetGraphSansa(Configuration, SerializableSupplier<StreamRDF>, Transactional) - Constructor for class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- load(String) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- load(Graph) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- load(Node, String) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- load(Node, Graph) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- load(SparkContext, String) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RddLoader
- load(SparkContext, String) - Method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderImpl
- loadDataset(String) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- loadDataset(DatasetGraph) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- loadDefaults(RddRdfLoaderRegistry) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderRegistryImpl
M
- map(Function<? super X, Y>) - Method in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
-
Return a fresh
HadoopInputData
instance where "nextMapper" is applied to the result of the current mapper - mapIntoGraph(Node) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfQuadsOps
- mapIntoGraph(Node) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
- mapPartitions(JavaPairRDD<K, V>, RxFunction<Tuple2<K, V>, O>) - Static method in class net.sansa_stack.spark.rdd.op.rx.JavaRddRxOps
- mapPartitions(JavaPairRDD<K, V>, StreamFunction<Tuple2<K, V>, O>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOps
- mapPartitions(JavaRDD<I>, RxFunction<I, O>) - Static method in class net.sansa_stack.spark.rdd.op.rx.JavaRddRxOps
-
Map operation based on a flowable transformer
- mapPartitions(JavaRDD<I>, StreamFunction<I, O>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOps
-
Map operation based on a flowable transformer
- mapper - Variable in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- mapQuadsToTriplesForTripleLangs - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter2
- mapQuadsToTriplesForTripleLangs - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
-
Whether to convert quads to triples if a triple-based output format is requested
- mapToDataset(JavaRDD<Quad>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfQuadsOps
- mapToDatasets(JavaPairRDD<String, Model>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfNamedModelsOps
-
Map each (name, model) pair to a dataset with the same information
- mapToDatasets(JavaRDD<Resource>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfResourcesOps
-
Map every IRI resource to a dataset having a single named graph matching the IRI.
- mapToModel(JavaRDD<Triple>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
- mapToNamedModels(JavaRDD<Resource>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfResourcesOps
-
Map IRI resources to a named model.
- mapToResourceInDataset(JavaPairRDD<String, Model>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfNamedModelsOps
- mapToResources(JavaPairRDD<String, Model>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfNamedModelsOps
- mapToTriples(Node) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfQuadsOps
- mayProduceQuads(Collection<SparqlStmt>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- mayProduceQuads(SparqlStmt) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- members - Variable in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- mentionesRowNum(Collection<SparqlStmt>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- mentionesRowNum(SparqlStmt) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- merge(Collection<String[][]>) - Static method in interface net.sansa_stack.spark.io.csv.input.ColumnNamingScheme
-
Merges multiple namings such that always only the first one is retained.
- mergeFolder(Path, Path, String, Comparator<? super Path>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddWriterUtils
-
Merge all files in the given srcFolder into outFile (uses java nio abstraction which can be backed by hadoop paths)
- MIXED_QUAD - Static variable in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- MIXED_TRIPLE - Static variable in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- models(RDD<Model>) - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfProcessor
- mutate(Consumer<? super SELF>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
-
Pass this object to a consumer.
N
- NAMED_GRAPH - Enum constant in enum class net.sansa_stack.spark.io.rdf.input.api.RdfSourceRecordType
- net.sansa_stack.query.spark.api.domain - package net.sansa_stack.query.spark.api.domain
- net.sansa_stack.spark.io.common - package net.sansa_stack.spark.io.common
- net.sansa_stack.spark.io.csv.input - package net.sansa_stack.spark.io.csv.input
- net.sansa_stack.spark.io.json.input - package net.sansa_stack.spark.io.json.input
- net.sansa_stack.spark.io.rdf.input.api - package net.sansa_stack.spark.io.rdf.input.api
- net.sansa_stack.spark.io.rdf.input.impl - package net.sansa_stack.spark.io.rdf.input.impl
- net.sansa_stack.spark.io.rdf.kryo - package net.sansa_stack.spark.io.rdf.kryo
- net.sansa_stack.spark.io.rdf.loader - package net.sansa_stack.spark.io.rdf.loader
- net.sansa_stack.spark.io.rdf.output - package net.sansa_stack.spark.io.rdf.output
- net.sansa_stack.spark.rdd.function - package net.sansa_stack.spark.rdd.function
- net.sansa_stack.spark.rdd.op.rdf - package net.sansa_stack.spark.rdd.op.rdf
- net.sansa_stack.spark.rdd.op.rx - package net.sansa_stack.spark.rdd.op.rx
- net.sansa_stack.spark.util - package net.sansa_stack.spark.util
- newRdfSourceCollection() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSourceFactory
-
Return a collection to which RdfSources can be added from which a union rdd can be obtained
- newRdfSourceCollection() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactoryImpl
- NodeTupleSource - Interface in net.sansa_stack.spark.io.rdf.input.api
-
Source of fixed size tuples of RDF nodes
- number(int, int) - Static method in interface net.sansa_stack.spark.io.csv.input.ColumnNamingScheme
O
- of(Class<?>, Class<?>, Class<? extends InputFormat>) - Static method in class net.sansa_stack.spark.io.common.HadoopInputFormat
- of(Class<?>, Class<?>, Class<? extends OutputFormat>) - Static method in class net.sansa_stack.spark.io.common.HadoopOutputFormat
- of(SparkSession) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactories
- ofDatasets(JavaRDD<DatasetOneNg>) - Static method in class net.sansa_stack.spark.io.rdf.input.api.RdfSources
- ofModels(JavaRDD<Model>) - Static method in class net.sansa_stack.spark.io.rdf.input.api.RdfSources
- ofQuads(JavaRDD<Quad>) - Static method in class net.sansa_stack.spark.io.rdf.input.api.RdfSources
- ofTriples(JavaRDD<Triple>) - Static method in class net.sansa_stack.spark.io.rdf.input.api.RdfSources
- optimizePrefixes - Variable in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- outputFormat - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- outputLang - Variable in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterSettings
P
- parse(Path, RDFFormat, Configuration, StreamRDF) - Static method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop
- parseRaw(Path, Configuration, InputFormat<?, T>, ExecutorService, StreamRDF, BiConsumer<T, StreamRDF>) - Static method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop
-
The sink must be started beforehand!
- partitionFolder - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- partitionFolderFs - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- partitionMapperNQuads(Iterator<Quad>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- partitionMapperNTriples(Iterator<Triple>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- partitionMapperRDFStream(Function<OutputStream, StreamRDF>, BiConsumer<? super T, StreamRDF>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- partitionsAsIndependentFiles - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- path - Variable in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- path - Variable in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- peekDeclaredPrefixes() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfLikeSource
-
At present this creates a model holding an RDF sample based on a file's starting bytes.
- peekDeclaredPrefixes() - Method in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- peekDeclaredPrefixes() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- peekDeclaredPrefixes() - Method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- peekPrefixes(SparkContext, String) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RddRdfLoader
-
Peek prefixes w.r.t.
- peekPrefixes(SparkContext, String) - Method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderImpl
- postProcess(RddWriterSettings<?>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddWriterUtils
- postProcess(JavaRDD<Triple>, boolean, boolean, boolean, int) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
-
Sort quads by their string representation (relies on
NodeFmtLib.str(org.apache.jena.graph.Triple)
) - postProcess(JavaRDD<Quad>, boolean, boolean, boolean, int) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfQuadsOps
-
Post process RDF data - sort, distinct, repartition Sort quads by their string representation (relies on
NodeFmtLib.str(org.apache.jena.graph.Triple)
) - postProcessingSettings - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- prefixes - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter2
- prepare(RddWriterSettings<?>, Configuration) - Static method in class net.sansa_stack.spark.io.rdf.output.RddWriterUtils
- probeJsonFormat(Reader, Gson, int) - Static method in class net.sansa_stack.spark.io.json.input.JsonDataSources
-
Detect whether input is...
- probeJsonFormat(String, Configuration, int) - Static method in class net.sansa_stack.spark.io.json.input.JsonDataSources
- probeJsonInputFormat(String, Configuration, int) - Static method in class net.sansa_stack.spark.io.json.input.JsonDataSources
- probeLang(Path, FileSystem) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactoryImpl
- put(String) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- put(Graph) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- put(Node, String) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- put(Node, Graph) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- putDataset(String) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- putDataset(DatasetGraph) - Method in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
Q
- QUAD - Enum constant in enum class net.sansa_stack.spark.io.rdf.input.api.RdfSourceRecordType
- quadMapper(Collection<Query>, Supplier<ExecutionContext>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- quads(RDD<Quad>) - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfProcessor
R
- rdd - Variable in class net.sansa_stack.query.spark.api.domain.JavaResultSetSparkImpl
- rdd - Variable in class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- rdd - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- RddLoader<K,
T> - Interface in net.sansa_stack.spark.io.rdf.input.api - RddLoaderBase<K,
T> - Class in net.sansa_stack.spark.io.rdf.input.api - RddLoaderBase(Class<T>, Class<? extends FileInputFormat<K, T>>) - Constructor for class net.sansa_stack.spark.io.rdf.input.api.RddLoaderBase
- RddRdfLoader<T> - Interface in net.sansa_stack.spark.io.rdf.input.api
-
An RddRdfLoader provides rdf-related methods to operate on paths w.r.t.
- RddRdfLoaderImpl<T> - Class in net.sansa_stack.spark.io.rdf.input.impl
- RddRdfLoaderImpl(Class<T>, Class<? extends FileInputFormat<LongWritable, T>>) - Constructor for class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderImpl
- RddRdfLoaderRegistry - Interface in net.sansa_stack.spark.io.rdf.input.api
- RddRdfLoaderRegistryImpl - Class in net.sansa_stack.spark.io.rdf.input.impl
-
A registry for RddRdfLoaders that can supply input of a specific
Lang
to an RDD of a requested type (Triples, Quads, Datasets, ett). - RddRdfLoaderRegistryImpl() - Constructor for class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderRegistryImpl
- RddRdfLoaders - Class in net.sansa_stack.spark.io.rdf.input.impl
- RddRdfLoaders() - Constructor for class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaders
- RddRdfOps<T> - Interface in net.sansa_stack.spark.io.rdf.output
-
Interface that captures common operations on RDD<T@gt;
- RddRdfOpsImpl<T> - Class in net.sansa_stack.spark.io.rdf.output
- RddRdfOpsImpl(int, BiConsumer<T, StreamRDF>, JavaRddFunction<T, Triple>, JavaRddFunction<T, Quad>, JavaRddFunction<T, Node>, Function<? super T, Comparable<?>>) - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- RddRdfProcessor - Interface in net.sansa_stack.spark.io.rdf.output
- RddRdfWriter<T> - Class in net.sansa_stack.spark.io.rdf.output
-
Important: Instances of this class should only be created using
RddRdfWriterFactory
because the factory is RDD-independent and can validate settings at an early stage. - RddRdfWriter(RddRdfOpsImpl<T>) - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- RddRdfWriter2 - Class in net.sansa_stack.spark.io.rdf.output
-
Core class for configuration and execution of writing RDDs of RDF out using Hadaop.
- RddRdfWriter2(RDFFormat, boolean, PrefixMapping) - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRdfWriter2
- RddRdfWriterFactory - Class in net.sansa_stack.spark.io.rdf.output
-
A factory for
RddRdfWriter
instances which enables validation of settings at an early stage usingRddRdfWriterFactory.validate()
. - RddRdfWriterFactory() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- RddRdfWriterFormatRegistry - Class in net.sansa_stack.spark.io.rdf.output
-
Registry for mapping between jena's
RDFFormat
and hadoop'sOutputFormat
. - RddRdfWriterFormatRegistry() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFormatRegistry
- RddRdfWriterSettings<SELF extends RddRdfWriterSettings> - Class in net.sansa_stack.spark.io.rdf.output
- RddRdfWriterSettings() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- RddRowSetWriter - Class in net.sansa_stack.spark.io.rdf.output
- RddRowSetWriter() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
- RddRowSetWriterFactory - Class in net.sansa_stack.spark.io.rdf.output
- RddRowSetWriterFactory() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterFactory
- RddRowSetWriterSettings<SELF extends RddRowSetWriterSettings> - Class in net.sansa_stack.spark.io.rdf.output
- RddRowSetWriterSettings() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterSettings
- RddRowSetWriterUtils - Class in net.sansa_stack.spark.io.rdf.output
-
Static util methods to write
JavaResultSetSpark
instances out using hadoop. - RddRowSetWriterUtils() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterUtils
- RddWriterSettings<SELF extends RddWriterSettings> - Class in net.sansa_stack.spark.io.rdf.output
- RddWriterSettings() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- RddWriterUtils - Class in net.sansa_stack.spark.io.rdf.output
-
Utilities common to (but not limited to) Rdf and RowSet output
- RddWriterUtils() - Constructor for class net.sansa_stack.spark.io.rdf.output.RddWriterUtils
- rdfFormat - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter2
- RdfLikeSource - Interface in net.sansa_stack.spark.io.rdf.input.api
- RdfPostProcessingSettings - Interface in net.sansa_stack.spark.io.rdf.output
- RdfPostProcessingSettingsBase - Class in net.sansa_stack.spark.io.rdf.output
- RdfPostProcessingSettingsBase() - Constructor for class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- RdfPostProcessingSettingsBase(Boolean, Integer, Boolean, Boolean, Integer, Boolean) - Constructor for class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- RdfPostProcessingSettingsMutable - Interface in net.sansa_stack.spark.io.rdf.output
- RdfSource - Interface in net.sansa_stack.spark.io.rdf.input.api
-
An RdfSource is a NodeTupleSource with tuple size either 3 or 4.
- RdfSourceCollection - Interface in net.sansa_stack.spark.io.rdf.input.api
- RdfSourceCollectionImpl - Class in net.sansa_stack.spark.io.rdf.input.impl
- RdfSourceCollectionImpl(SparkSession) - Constructor for class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- RdfSourceCollectionImpl(SparkSession, Collection<RdfSource>) - Constructor for class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- RdfSourceFactories - Class in net.sansa_stack.spark.io.rdf.input.impl
- RdfSourceFactories() - Constructor for class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactories
- RdfSourceFactory - Interface in net.sansa_stack.spark.io.rdf.input.api
-
Turn source references into sources of RDF data.
- RdfSourceFactoryImpl - Class in net.sansa_stack.spark.io.rdf.input.impl
-
Implementation of a source factory based on spark/hadoop.
- RdfSourceFactoryImpl(SparkSession) - Constructor for class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactoryImpl
- RdfSourceFromRdd<T> - Class in net.sansa_stack.spark.io.rdf.input.api
- RdfSourceFromRdd(JavaRDD<T>, RddRdfOps<T>, PrefixMap) - Constructor for class net.sansa_stack.spark.io.rdf.input.api.RdfSourceFromRdd
- RdfSourceFromResource - Interface in net.sansa_stack.spark.io.rdf.input.api
-
Abstraction of a source of RDF.
- RdfSourceFromResourceImpl - Class in net.sansa_stack.spark.io.rdf.input.impl
- RdfSourceFromResourceImpl(SparkSession, Path, Lang) - Constructor for class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- RdfSourceRecordType - Enum Class in net.sansa_stack.spark.io.rdf.input.api
-
The record type of an rdf source
- RdfSources - Class in net.sansa_stack.spark.io.rdf.input.api
- RdfSources() - Constructor for class net.sansa_stack.spark.io.rdf.input.api.RdfSources
- register(Lang, Class<T>, RddRdfLoader<T>) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RddRdfLoaderRegistry
-
Registration where the (underlying hadoop inputformat-based) loader loads items of type X and the result type is also X
- register(Lang, Class<T>, RddRdfLoader<T>) - Method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderRegistryImpl
- register(Lang, HadoopOutputFormat) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFormatRegistry
- registerClasses(Kryo) - Method in class net.sansa_stack.spark.io.rdf.kryo.JenaKryoRegistrator
- registerMapped(Lang, Class<T>, RddRdfLoader<X>) - Method in interface net.sansa_stack.spark.io.rdf.input.api.RddRdfLoaderRegistry
-
Registration where the (underlying hadoop inputformat-based) loader loads items of type X (such as triples) but the result is mapped to type Y (such as quads)
- registerMapped(Lang, Class<T>, RddRdfLoader<X>) - Method in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderRegistryImpl
- registry - Variable in class net.sansa_stack.spark.io.rdf.input.impl.RddRdfLoaderRegistryImpl
- registry - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFormatRegistry
- requireLoader(Lang, Class<T>) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- requireValidLang(Lang) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterUtils
-
Raises an
IllegalArgumentException
if writing out using the given lang is unsupported. - resultVars - Variable in class net.sansa_stack.query.spark.api.domain.JavaResultSetSparkImpl
- row(String[]) - Static method in interface net.sansa_stack.spark.io.csv.input.ColumnNamingScheme
- rowMapperFactoryBinding(String[][]) - Static method in class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
- rowMapperFactoryBinding(Var[]) - Static method in class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
-
A RowMapperFactory that uses a predefined set of variables
- rowMapperFactoryJson(Binding, Var, SerializableBiFunction<String[][], String[], JsonObject>) - Static method in class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
-
Wrap a rowMapperFactoryJson such that it produces bindings
- ROWNUM - Static variable in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
-
The special ?ROWNUM variable supported by tarql
- rowSet - Variable in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
- rowToBinding(Var[][], String[]) - Static method in class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
-
Util method to create a binding from a list of variables and a list of strings.
- rowToBinding(Var[], String[]) - Static method in class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
- rowToJson(String[][], String[]) - Static method in class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
- run() - Method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- run() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- run() - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
- runActual() - Method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- runActual(RddWriterSettings<?>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- runActual(RddWriterSettings<?>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
- runOutputToConsole() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- runOutputToConsole() - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
- runSpark() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
-
Run the save action according to configuration
- runSpark() - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
- runUnchecked() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
-
Same as
RddRdfWriter.run()
but without the checked IOException - runUnchecked() - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
-
Same as
RddRowSetWriter.run()
but without the checked IOException
S
- safeDeletePartitionFolder(FileSystem, Path, Configuration) - Static method in class net.sansa_stack.spark.io.rdf.output.RddWriterUtils
-
This method first checks that all top-level files in the partition folder belong to hadoop.
- save(JavaPairRDD<?, ?>, HadoopOutputFormat, Path, Configuration) - Static method in class net.sansa_stack.spark.io.rdf.output.RddWriterUtils
- saveToFolder(JavaRDD<T>, String, RDFFormat, boolean, PrefixMapping, BiConsumer<T, StreamRDF>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
-
Deprecated.
- saveUsingElephas(JavaRDD<T>, Path, Lang, SerializableFunction<? super T, ?>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- self() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- self() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- self() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- sendRecordToStreamRDF - Variable in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- sendRecordToStreamRDF - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- sendRecordToStreamRDF(T, StreamRDF) - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfOps
- sendRecordToStreamRDF(T, StreamRDF) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfOpsImpl
- sendToStreamRDF(JavaRDD<T>, SerializableBiConsumer<T, StreamRDF>, SerializableSupplier<StreamRDF>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- SEQUENCE - Enum constant in enum class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonSourceType
- setAllowOverwriteFiles(boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setConf(Configuration) - Method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- setConsoleOutput() - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setConsoleOutSupplier(Supplier<OutputStream>) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setDeferOutputForUsedPrefixes(long) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- setDeletePartitionFolderAfterMerge(boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setDistinct(Boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- setDistinct(Boolean) - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsMutable
- setDistinctPartitions(Integer) - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- setDistinctPartitions(Integer) - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsMutable
- setGlobalPrefixMapping(Map<String, String>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- setGlobalPrefixMapping(PrefixMapping) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
-
Set a prefix mapping to be used "globally" across all partitions.
- setHadoopConfiguration(Configuration) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setInputFile(Path) - Method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- setMapQuadsToTriplesForTripleLangs(boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
-
Whether to convert quads to triples if a triple-based output format is requested Jena by default discards any quad outside of the default graph when writing to a triple format.
- setOptimizePrefixes(Boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- setOptimizePrefixes(Boolean) - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsMutable
- setOutputFormat(String) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
-
Raises an exception if the format is not found
- setOutputFormat(RDFFormat) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- setOutputLang(String) - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterSettings
- setOutputLang(Lang) - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterSettings
- setPartitionFolder(String) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setPartitionFolder(Path) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setPartitionFolderFs(FileSystem) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setPartitionsAsIndependentFiles(boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterSettings
- setPartitionsAsIndependentFiles(boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setPostProcessingSettings(RdfPostProcessingSettingsMutable) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setRdd(JavaRDD<? extends T>) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- setRowSet(JavaResultSetSpark) - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriter
- setSink(StreamRDF) - Method in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- setSort(Boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- setSort(Boolean) - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsMutable
- setSortAscending(Boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- setSortAscending(Boolean) - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsMutable
- setSortPartitions(Integer) - Method in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- setSortPartitions(Integer) - Method in interface net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsMutable
- setTargetFile(String) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setTargetFile(Path) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setTargetFileFs(FileSystem) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- setUseCoalesceOne(boolean) - Method in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- sink - Variable in class net.sansa_stack.spark.io.rdf.loader.AsyncRdfParserHadoop.Builder
- sinkFactory - Variable in class net.sansa_stack.spark.io.rdf.loader.LinkDatasetGraphSansa
- sort - Variable in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- sortAscending - Variable in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- sortPartitions - Variable in class net.sansa_stack.spark.io.rdf.output.RdfPostProcessingSettingsBase
- sparkContext - Variable in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- sparkSession - Variable in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- sparkSession - Variable in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFactoryImpl
- sparkSession - Variable in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceFromResourceImpl
- sparql(String[]) - Static method in interface net.sansa_stack.spark.io.csv.input.ColumnNamingScheme
-
Convert the column headings such that they are safe for use with sparql variables.
T
- targetFile - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- targetFileFs - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- tarqlDatasets(JavaRDD<Binding>, Collection<SparqlStmt>, boolean, SerializableFunction<DatasetGraph, Stream<T>>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
-
Turns each row into a dataset based on SPARQL update statements.
- tarqlDatasets(JavaRDD<Binding>, Query) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
-
Method for the typical case of mapping an RDD of bindings via a construct query to an RDD of datasets.
- tarqlOptimize(Op) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
-
Apply default optimizations for algebra expressions meant for tarql Combines EXTENDS
- tarqlQuads(JavaRDD<Binding>, Collection<SparqlStmt>, boolean, Supplier<ExecutionContext>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- tarqlQuads(JavaRDD<Binding>, Query, Supplier<ExecutionContext>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- tarqlTriples(JavaRDD<Binding>, Collection<SparqlStmt>, boolean, Supplier<ExecutionContext>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- templateMapperQuads(Template) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- templateMapperTriples(Template) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- toGraphName(Node) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfTriplesOps
-
Map a node losslessly to an IRI suitable for use as a graph name This is needed to e.g.
- ToJavaPairRddFunction<I,
K, V> - Interface in net.sansa_stack.spark.rdd.function -
Interface for building chains of transformations over JavaRDDs and JavaPairRDDs.
- ToJavaRddFunction<K,
V, O> - Interface in net.sansa_stack.spark.rdd.function -
Interface for building chains of transformations over JavaRDDs and JavaPairRDDs.
- toPairRdd(ToJavaPairRddFunction<O, K, V>) - Method in interface net.sansa_stack.spark.rdd.function.JavaRddFunction
- toPairRdd(ToJavaPairRddFunction<O, KX, VX>) - Method in interface net.sansa_stack.spark.rdd.function.ToJavaRddFunction
- toPairRdd(JavaRDD<T>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddWriterUtils
- toRdd(ToJavaRddFunction<K, V, O>) - Method in interface net.sansa_stack.spark.rdd.function.ToJavaPairRddFunction
- toRdd(ToJavaRddFunction<KO, VO, X>) - Method in interface net.sansa_stack.spark.rdd.function.JavaPairRddFunction
- toString() - Method in class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonProbeResult
- toString(PrefixMapping, RDFFormat) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
-
Convert a prefix mapping to a string
- transformHeader(String[][], Function<String, String>) - Static method in class net.sansa_stack.spark.io.csv.input.CsvRowMapperFactories
- TRIPLE - Enum constant in enum class net.sansa_stack.spark.io.rdf.input.api.RdfSourceRecordType
- tripleMapper(Collection<Query>, Supplier<ExecutionContext>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOfBindingsOps
- triples(RDD<Triple>) - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfProcessor
U
- union(SparkSession, Collection<I>, Function<I, RDD<T>>) - Static method in class net.sansa_stack.spark.io.rdf.input.impl.RdfSourceCollectionImpl
- unionIfNeeded(JavaSparkContext, Collection<JavaRDD<T>>) - Static method in class net.sansa_stack.spark.rdd.op.rdf.JavaRddOps
- UNKNOWN - Enum constant in enum class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonSourceType
- useCoalesceOne - Variable in class net.sansa_stack.spark.io.rdf.output.RddWriterSettings
- usesQuads() - Method in interface net.sansa_stack.spark.io.rdf.input.api.RdfSource
-
Whether this source is based on a quad model
- usesQuads() - Method in interface net.sansa_stack.spark.io.rdf.output.RddRdfOps
V
- validate() - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriterFactory
- validate() - Method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterFactory
- validate(RddRdfWriterSettings<?>) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter
- valueClass - Variable in class net.sansa_stack.spark.io.common.HadoopFormat
- valueClass - Variable in class net.sansa_stack.spark.io.rdf.input.api.HadoopInputData
- valueOf(String) - Static method in enum class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonSourceType
-
Returns the enum constant of this class with the specified name.
- valueOf(String) - Static method in enum class net.sansa_stack.spark.io.rdf.input.api.RdfSourceRecordType
-
Returns the enum constant of this class with the specified name.
- values() - Static method in enum class net.sansa_stack.spark.io.json.input.JsonDataSources.JsonSourceType
-
Returns an array containing the constants of this enum class, in the order they are declared.
- values() - Static method in enum class net.sansa_stack.spark.io.rdf.input.api.RdfSourceRecordType
-
Returns an array containing the constants of this enum class, in the order they are declared.
W
- wrapWithAnalyzer(HadoopInputData<?, ?, ?>) - Static method in class net.sansa_stack.spark.io.rdf.input.api.InputFormatUtils
-
Wrap an input format that is based on
RecordReaderGenericBase
with an analyzer that turns each split into parsing metadata rather than data. - write(JavaResultSetSpark, Path, Lang) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterUtils
- write(RDD<Binding>, Path, List<Var>, Lang) - Static method in class net.sansa_stack.spark.io.rdf.output.RddRowSetWriterUtils
- writeQuads(RDD<Quad>, Path) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter2
- writeTriples(RDD<Triple>, Path) - Method in class net.sansa_stack.spark.io.rdf.output.RddRdfWriter2
All Classes and Interfaces|All Packages|Serialized Form