A B C D E F G H I J L M N O P R S T W X
All Classes All Packages
All Classes All Packages
All Classes All Packages
A
- accumulate(A, T) - Method in interface net.sansa_stack.hadoop.core.Accumulating
- accumulate(DatasetOneNg, Quad) - Method in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset.AccumulatingDataset
- accumulatedValue(A) - Method in interface net.sansa_stack.hadoop.core.Accumulating
- accumulatedValue(DatasetOneNg) - Method in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset.AccumulatingDataset
- accumulating - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- Accumulating<T,G,A,U> - Interface in net.sansa_stack.hadoop.core
- AccumulatingDataset() - Constructor for class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset.AccumulatingDataset
- afterSeek() - Method in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- afterSeek() - Method in class net.sansa_stack.hadoop.util.SeekablePushbackInputStream
- aggregate(boolean, Flowable<U>, List<U>) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
-
Modify a flow to perform aggregation of items into records according to specification The complex part here is to correctly combine the two flows: - The first group of the splitAggregateFlow needs to be skipped as this in handled by the previous split's processor - If there are no further groups in splitFlow then no items are emitted at all (because all items belong to s previous split) - ONLY if the splitFlow owned at least one group: The first group in the tailFlow needs to be emitted
B
- BASE_IRI_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- baseIri - Variable in class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfBase
- baseIriKey - Variable in class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfBase
- buf - Variable in class net.sansa_stack.nio.util.ReadableByteChannelFromInputStream
- bytesRead - Variable in class net.sansa_stack.nio.util.ReadableByteChannelWithConditionalBound
C
- CELL_MAXLENGTH_DEFAULT_VALUE - Static variable in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- CELL_MAXLENGTH_DEFAULT_VALUE - Static variable in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- CELL_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
-
The maximum length of a CSV cell containing new lines
- CELL_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
-
The maximum length of a CSV cell containing new lines
- classify(Quad) - Method in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset.AccumulatingDataset
- classify(T) - Method in interface net.sansa_stack.hadoop.core.Accumulating
- close() - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- close() - Method in class net.sansa_stack.hadoop.util.InputStreamWithCloseLogging
- close() - Method in class net.sansa_stack.io.util.InputStreamWithCloseIgnore
- close() - Method in interface net.sansa_stack.nio.util.ReadableByteChannelDecorator
- close() - Method in class net.sansa_stack.nio.util.ReadableByteChannelFromInputStream
- close() - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
- codec - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- ConfigurationUtils - Class in net.sansa_stack.hadoop.util
- ConfigurationUtils() - Constructor for class net.sansa_stack.hadoop.util.ConfigurationUtils
- createAccumulator(G) - Method in interface net.sansa_stack.hadoop.core.Accumulating
- createAccumulator(Node) - Method in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset.AccumulatingDataset
- createFlow(Job, InputFormat<?, T>, InputSplit) - Static method in class net.sansa_stack.hadoop.util.FileSplitUtils
-
Create a flow of records for a given input split w.r.t.
- createRecordFlow() - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- createRecordFlow() - Method in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
-
Override createRecordFlow to skip the first record if the requested format demands so.
- createRecordFlow() - Method in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
-
Override createRecordFlow to skip the first record if the requested format demands so.
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.commons_csv.csv.FileInputFormatCsv
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.gson.json.FileInputFormatJsonArray
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.univocity.csv.csv.FileInputFormatCsv
- createRecordReaderActual(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- createRecordReaderActual(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.jena.trig.FileInputFormatRdfTrigDataset
- createRecordReaderActual(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.jena.trig.FileInputFormatRdfTrigQuad
- createRecordReaderActual(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.jena.turtle.FileInputFormatRdfTurtleTriple
- createStartOfCsvRecordPattern(long) - Static method in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
-
Create a regex for matching csv record starts.
- createStartOfCsvRecordPattern(long) - Static method in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
-
Create a regex for matching csv record starts.
- createTestParameters(Map<String, Range<Integer>>) - Static method in class net.sansa_stack.hadoop.util.FileSplitUtils
-
Util method typically for use with split-related unit tests
- creationStackTrace - Variable in class net.sansa_stack.hadoop.util.InputStreamWithCloseLogging
- CSV_FORMAT_RAW_KEY - Static variable in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
-
Key for the serialized bytes of a
CSVFormat
instance - CSV_FORMAT_RAW_KEY - Static variable in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
-
Key for the serialized bytes of a
CSVFormat
instance - CsvUtils - Class in net.sansa_stack.hadoop.format.commons_csv.csv
- CsvUtils - Class in net.sansa_stack.hadoop.format.univocity.csv.csv
- CsvUtils() - Constructor for class net.sansa_stack.hadoop.format.commons_csv.csv.CsvUtils
- CsvUtils() - Constructor for class net.sansa_stack.hadoop.format.univocity.csv.csv.CsvUtils
- currentKey - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- currentValue - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
D
- datasetFlow - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- decompressor - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- decoratee - Variable in class net.sansa_stack.nio.util.SeekableByteChannelDecoratorBase
- defaultSettings(boolean) - Static method in class net.sansa_stack.hadoop.format.univocity.csv.csv.CsvUtils
- DeferredSeekablePushbackInputStream - Class in net.sansa_stack.hadoop.util
-
A wrapper for hadoop input streams created from codecs in ReadMode.BY_BLOCK: Defers reading by one byte such that position changes are advertised on the byte BEFORE the block boundary rather than on the byte AFTER it.
- DeferredSeekablePushbackInputStream(InputStream) - Constructor for class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- DeferredSeekablePushbackInputStream(InputStream, Seekable) - Constructor for class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- delegate - Variable in class net.sansa_stack.nio.util.ReadableByteChannelDecoratorBase
- didHitSplitBound(Seekable, long) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- disableSkipHeaderRecord(CSVFormat) - Method in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- disableSkipHeaderRecord(CSVFormat) - Method in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
E
- effectiveCsvFormat - Variable in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- effectiveCsvFormat - Variable in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- effectiveInputStream(InputStream) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- effectiveInputStream(InputStream) - Method in class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
-
Always replace the first character (which is either a comma or open bracket) with an open bracket in order to mimick a JSON array start.
- effectiveInputStreamSupp(Seekable) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- EMPTY_BYTE_ARRAY - Static variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
F
- fallbackBuffer - Variable in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- FileInputFormatCsv - Class in net.sansa_stack.hadoop.format.commons_csv.csv
- FileInputFormatCsv - Class in net.sansa_stack.hadoop.format.univocity.csv.csv
- FileInputFormatCsv() - Constructor for class net.sansa_stack.hadoop.format.commons_csv.csv.FileInputFormatCsv
- FileInputFormatCsv() - Constructor for class net.sansa_stack.hadoop.format.univocity.csv.csv.FileInputFormatCsv
- FileInputFormatJsonArray - Class in net.sansa_stack.hadoop.format.gson.json
- FileInputFormatJsonArray() - Constructor for class net.sansa_stack.hadoop.format.gson.json.FileInputFormatJsonArray
- FileInputFormatRdfBase<T> - Class in net.sansa_stack.hadoop.format.jena.base
-
Base class for unit testing of reading an RDF file with an arbitrary number of splits.
- FileInputFormatRdfBase(Lang, String) - Constructor for class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- FileInputFormatRdfTrigDataset - Class in net.sansa_stack.hadoop.format.jena.trig
- FileInputFormatRdfTrigDataset() - Constructor for class net.sansa_stack.hadoop.format.jena.trig.FileInputFormatRdfTrigDataset
- FileInputFormatRdfTrigQuad - Class in net.sansa_stack.hadoop.format.jena.trig
- FileInputFormatRdfTrigQuad() - Constructor for class net.sansa_stack.hadoop.format.jena.trig.FileInputFormatRdfTrigQuad
- FileInputFormatRdfTurtleTriple - Class in net.sansa_stack.hadoop.format.jena.turtle
- FileInputFormatRdfTurtleTriple() - Constructor for class net.sansa_stack.hadoop.format.jena.turtle.FileInputFormatRdfTurtleTriple
- FileOutputFormatTrig2<TKey> - Class in net.sansa_stack.hadoop.output
-
Not yet used; the idea is to provide an improved version of elepha's TrigOutputFormat
- FileOutputFormatTrig2() - Constructor for class net.sansa_stack.hadoop.output.FileOutputFormatTrig2
- FileSplitUtils - Class in net.sansa_stack.hadoop.util
- FileSplitUtils() - Constructor for class net.sansa_stack.hadoop.util.FileSplitUtils
- fileSystem - Variable in class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
- findFirstPositionWithProbeSuccess(Seekable, Predicate<Long>, Matcher, boolean, Predicate<Seekable>) - Static method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
-
Uses the matcher to find candidate probing positions, and returns the first positoin where probing succeeds.
- findNextRecord(Pattern, Seekable, long, long, long, long, Predicate<Long>, Predicate<Seekable>) - Static method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
G
- getBytesRead() - Method in class net.sansa_stack.nio.util.ReadableByteChannelWithConditionalBound
- getCsvFormat(Configuration, CSVFormat) - Static method in class net.sansa_stack.hadoop.format.commons_csv.csv.FileInputFormatCsv
- getCurrentKey() - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- getCurrentValue() - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- getDecodedStreamFromSplit(FileSplit, Configuration) - Static method in class net.sansa_stack.hadoop.util.FileSplitUtils
-
Util method to open a decoded stream from a split.
- getDecoratee() - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
- getDecoratee() - Method in class net.sansa_stack.nio.util.SeekableByteChannelDecoratorBase
- getDelegate() - Method in interface net.sansa_stack.nio.util.ReadableByteChannelDecorator
- getDelegate() - Method in class net.sansa_stack.nio.util.ReadableByteChannelDecoratorBase
- getFileExtension() - Method in class net.sansa_stack.hadoop.output.FileOutputFormatTrig2
- getModel(Configuration) - Static method in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
-
Extract a Model from a hadoop conf using
FileInputFormatRdfBase.PREFIXES_KEY
- getModel(Configuration, String) - Static method in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
-
Extract a Model from a hadoop conf.
- getName() - Method in class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
- getPos() - Method in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- getPos() - Method in interface net.sansa_stack.hadoop.util.SeekableDecorator
- getPos(Seekable) - Static method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- getProgress() - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- getRecordWriter(StreamRDF, Writer, Configuration) - Method in class net.sansa_stack.hadoop.output.FileOutputFormatTrig2
- getSeekable() - Method in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- getSeekable() - Method in interface net.sansa_stack.hadoop.util.SeekableDecorator
- getSeekable() - Method in class net.sansa_stack.hadoop.util.SeekableInputStream
-
You should not change the position of the underlying seekable directly while this input stream is in use.
- getSeekable() - Method in class net.sansa_stack.hadoop.util.SeekablePushbackInputStream
- getSerializable(Configuration, String, T) - Static method in class net.sansa_stack.hadoop.util.ConfigurationUtils
-
Get a (non-null) string as a base64 url encoded serialized object
- getSplits(JobContext) - Method in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- getStream(Writer, Configuration) - Method in class net.sansa_stack.hadoop.output.FileOutputFormatTrig2
- gson - Variable in class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
H
- hashCode() - Method in class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
- headerBytesKey - Variable in class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfBase
I
- identity() - Static method in interface net.sansa_stack.hadoop.core.Accumulating
-
Identity accumulator - turns each item into a group that contains only the item and whose value is the item
- in - Variable in class net.sansa_stack.nio.util.ReadableByteChannelFromInputStream
- initialize(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
-
Read out config paramaters (prefixes, length thresholds, ...) and examine the codec in order to set an internal flag whether the stream will be encoded or not.
- initialize(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- initialize(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
- initialize(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfBase
- initialize(InputSplit, TaskAttemptContext) - Method in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- initRecordFlow() - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- InputStreamWithCloseIgnore - Class in net.sansa_stack.io.util
-
A wrapper whose close() method is a noop.
- InputStreamWithCloseIgnore(InputStream) - Constructor for class net.sansa_stack.io.util.InputStreamWithCloseIgnore
-
Constructs a new ProxyInputStream.
- InputStreamWithCloseLogging - Class in net.sansa_stack.hadoop.util
-
Util class to debug a stream already closed exception
- InputStreamWithCloseLogging(InputStream, BiConsumer<? super Throwable, ? super Throwable>) - Constructor for class net.sansa_stack.hadoop.util.InputStreamWithCloseLogging
- InputStreamWithZeroOffsetRead - Class in net.sansa_stack.io.util
-
Workaround for HADOOP-17453: read(bts, off, len) with off != 0 is broken in several version of BZip2Codec Invoking read with a non-zero offset creates a in intermediate buffer to which is read with a zero offset The content of the intermidate buffer is then copied to the requesting buffer bts at the appropriate offset.
- InputStreamWithZeroOffsetRead(InputStream) - Constructor for class net.sansa_stack.io.util.InputStreamWithZeroOffsetRead
-
Constructs a new ProxyInputStream.
- InterruptingSeekableByteChannel - Class in net.sansa_stack.nio.util
- InterruptingSeekableByteChannel(SeekableByteChannel, long) - Constructor for class net.sansa_stack.nio.util.InterruptingSeekableByteChannel
- interruptPos - Variable in class net.sansa_stack.nio.util.InterruptingSeekableByteChannel
- isEncoded - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- isFirstSplit - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- isInEofState - Variable in class net.sansa_stack.nio.util.ReadableByteChannelWithConditionalBound
- isOpen() - Method in class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- isOpen() - Method in interface net.sansa_stack.nio.util.ReadableByteChannelDecorator
- isOpen() - Method in class net.sansa_stack.nio.util.ReadableByteChannelFromInputStream
- isOpen() - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
- isSplitable(JobContext, Path) - Method in class net.sansa_stack.hadoop.format.commons_csv.csv.FileInputFormatCsv
- isSplitable(JobContext, Path) - Method in class net.sansa_stack.hadoop.format.gson.json.FileInputFormatJsonArray
- isSplitable(JobContext, Path) - Method in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- isSplitable(JobContext, Path) - Method in class net.sansa_stack.hadoop.format.univocity.csv.csv.FileInputFormatCsv
J
- jsonFwdPattern - Static variable in class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
L
- lang - Variable in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
-
Input language
- lang - Variable in class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfBase
- lines(Seekable) - Static method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- listFileSplits(Path, long, long) - Static method in class net.sansa_stack.hadoop.util.FileSplitUtils
- LocatorHdfs - Class in net.sansa_stack.hadoop.jena.locator
-
Support for resources using the "http:" and "https" schemes
- LocatorHdfs(FileSystem) - Constructor for class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
- LocatorHdfs(FileSystem, String[]) - Constructor for class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
- log() - Method in class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
- logClose(String) - Static method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- logUnexpectedClose(String) - Static method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
M
- maxRecordLength - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- maxRecordLengthKey - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- minRecordLength - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- minRecordLengthKey - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
N
- net.sansa_stack.hadoop.core - package net.sansa_stack.hadoop.core
- net.sansa_stack.hadoop.format.commons_csv.csv - package net.sansa_stack.hadoop.format.commons_csv.csv
- net.sansa_stack.hadoop.format.gson.json - package net.sansa_stack.hadoop.format.gson.json
- net.sansa_stack.hadoop.format.jena.base - package net.sansa_stack.hadoop.format.jena.base
- net.sansa_stack.hadoop.format.jena.trig - package net.sansa_stack.hadoop.format.jena.trig
- net.sansa_stack.hadoop.format.jena.turtle - package net.sansa_stack.hadoop.format.jena.turtle
- net.sansa_stack.hadoop.format.univocity.csv.csv - package net.sansa_stack.hadoop.format.univocity.csv.csv
- net.sansa_stack.hadoop.jena.locator - package net.sansa_stack.hadoop.jena.locator
- net.sansa_stack.hadoop.output - package net.sansa_stack.hadoop.output
- net.sansa_stack.hadoop.util - package net.sansa_stack.hadoop.util
- net.sansa_stack.io.util - package net.sansa_stack.io.util
- net.sansa_stack.nio.util - package net.sansa_stack.nio.util
- newCsvParser(Reader) - Method in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- newCsvParser(Reader) - Method in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- nextKeyValue() - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
O
- open - Variable in class net.sansa_stack.nio.util.ReadableByteChannelFromInputStream
P
- parse(Callable<InputStream>) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
-
Create a flowable from the input stream.
- parse(Callable<InputStream>) - Method in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- parse(Callable<InputStream>) - Method in class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
- parse(Callable<InputStream>) - Method in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset
- parse(Callable<InputStream>) - Method in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigQuad
- parse(Callable<InputStream>) - Method in class net.sansa_stack.hadoop.format.jena.turtle.RecordReaderRdfTurtleTriple
- parse(Callable<InputStream>) - Method in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- PARSED_PREFIXES_LENGTH_DEFAULT - Static variable in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- parseFromSeekable(Seekable) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- performOpen(String) - Method in class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
- position() - Method in class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- position() - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
- position(long) - Method in class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- position(long) - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
- postambleBytes - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- preambleBytes - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
-
Subclasses may initialize the pre/post-amble bytes in the
RecordReaderGenericBase.initialize(InputSplit, TaskAttemptContext)
method rather than the ctor! A (possibly empty) sequence of bytes to prepended to any stream passed to the parser. - PREFIXES_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- PREFIXES_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset
- PREFIXES_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigQuad
- PREFIXES_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.turtle.RecordReaderRdfTurtleTriple
- prefixesLengthMaxKey - Variable in class net.sansa_stack.hadoop.format.jena.base.FileInputFormatRdfBase
- prefixesMaxLengthKey - Variable in class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfBase
- prober(Seekable) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- probeRecordCount - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- probeRecordCountKey - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
R
- raisedThrowable - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- rawStream - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- read() - Method in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- read(byte[], int, int) - Method in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
-
This method essentially delays reads by one byte.
- read(byte[], int, int) - Method in class net.sansa_stack.io.util.InputStreamWithZeroOffsetRead
- read(ByteBuffer) - Method in class net.sansa_stack.nio.util.InterruptingSeekableByteChannel
- read(ByteBuffer) - Method in interface net.sansa_stack.nio.util.ReadableByteChannelDecorator
- read(ByteBuffer) - Method in class net.sansa_stack.nio.util.ReadableByteChannelFromInputStream
- read(ByteBuffer) - Method in class net.sansa_stack.nio.util.ReadableByteChannelWithConditionalBound
- read(ByteBuffer) - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
- ReadableByteChannelDecorator - Interface in net.sansa_stack.nio.util
- ReadableByteChannelDecoratorBase<T extends ReadableByteChannel> - Class in net.sansa_stack.nio.util
- ReadableByteChannelDecoratorBase(T) - Constructor for class net.sansa_stack.nio.util.ReadableByteChannelDecoratorBase
- ReadableByteChannelFromInputStream - Class in net.sansa_stack.nio.util
-
A replacement for Channels.newChannel with the following changes for interoperability with hadoop and our GenericRecordReader: - Does not close the underlying stream on interrupt - This implementation's read method just delegates to the input stream (without additional buffering / repeated reads).
- ReadableByteChannelFromInputStream(InputStream) - Constructor for class net.sansa_stack.nio.util.ReadableByteChannelFromInputStream
- ReadableByteChannelWithConditionalBound<T extends ReadableByteChannel> - Class in net.sansa_stack.nio.util
-
Readable byte channel wrapper that before every read checks for an end-of-file (eof) condition.
- ReadableByteChannelWithConditionalBound(T, Predicate<? super ReadableByteChannelWithConditionalBound<T>>) - Constructor for class net.sansa_stack.nio.util.ReadableByteChannelWithConditionalBound
- readCsvRecords(String, FileSystem, CsvParserSettings) - Static method in class net.sansa_stack.hadoop.format.univocity.csv.csv.CsvUtils
-
Create a flowable to a CSV file via hadoop.
- readCsvRecords(String, FileSystem, CSVFormat) - Static method in class net.sansa_stack.hadoop.format.commons_csv.csv.CsvUtils
-
Create a flowable to a CSV file via hadoop.
- readCsvRecords(Callable<? extends InputStream>, CsvParserSettings) - Static method in class net.sansa_stack.hadoop.format.univocity.csv.csv.CsvUtils
-
Create a flowable to a CSV file from a supplier of input streams
- readCsvRecords(Callable<? extends InputStream>, CSVFormat) - Static method in class net.sansa_stack.hadoop.format.commons_csv.csv.CsvUtils
-
Create a flowable to a CSV file from a supplier of input streams
- readCsvRecords(Path, FileSystem, CsvParserSettings) - Static method in class net.sansa_stack.hadoop.format.univocity.csv.csv.CsvUtils
-
Create a flowable to a CSV file via hadoop.
- readCsvRecords(Path, FileSystem, CSVFormat) - Static method in class net.sansa_stack.hadoop.format.commons_csv.csv.CsvUtils
-
Create a flowable to a CSV file via hadoop.
- readCsvRows(String, FileSystem, CsvParserSettings) - Static method in class net.sansa_stack.hadoop.format.univocity.csv.csv.CsvUtils
- readCsvRows(String, FileSystem, CSVFormat) - Static method in class net.sansa_stack.hadoop.format.commons_csv.csv.CsvUtils
- readCsvRows(Path, FileSystem, CsvParserSettings) - Static method in class net.sansa_stack.hadoop.format.univocity.csv.csv.CsvUtils
- readCsvRows(Path, FileSystem, CSVFormat) - Static method in class net.sansa_stack.hadoop.format.commons_csv.csv.CsvUtils
- readInternal(byte[], int, int) - Method in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
-
This method is assumed to be invoked with len >= 2
- RECORD_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- RECORD_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
- RECORD_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset
- RECORD_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigQuad
- RECORD_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.turtle.RecordReaderRdfTurtleTriple
- RECORD_MAXLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- RECORD_MINLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- RECORD_MINLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
- RECORD_MINLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset
- RECORD_MINLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigQuad
- RECORD_MINLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.turtle.RecordReaderRdfTurtleTriple
- RECORD_MINLENGTH_KEY - Static variable in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- RECORD_PROBECOUNT_KEY - Static variable in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- RECORD_PROBECOUNT_KEY - Static variable in class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
- RECORD_PROBECOUNT_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset
- RECORD_PROBECOUNT_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigQuad
- RECORD_PROBECOUNT_KEY - Static variable in class net.sansa_stack.hadoop.format.jena.turtle.RecordReaderRdfTurtleTriple
- RECORD_PROBECOUNT_KEY - Static variable in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- RecordReaderCsv - Class in net.sansa_stack.hadoop.format.commons_csv.csv
-
A generic parser implementation for CSV with the offset-seeking condition that CSV rows must all have the same length.
- RecordReaderCsv - Class in net.sansa_stack.hadoop.format.univocity.csv.csv
-
A generic parser implementation for CSV with the offset-seeking condition that CSV rows must all have the same length.
- RecordReaderCsv() - Constructor for class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- RecordReaderCsv() - Constructor for class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- RecordReaderCsv(String, String, String, Pattern) - Constructor for class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- RecordReaderCsv(String, String, String, Pattern) - Constructor for class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
- RecordReaderGenericBase<U,G,A,T> - Class in net.sansa_stack.hadoop.core
-
A generic record reader that uses a callback mechanism to detect a consecutive sequence of records that must start in the current split and which may extend over any number of successor splits.
- RecordReaderGenericBase(String, String, String, Pattern, Accumulating<U, G, A, T>) - Constructor for class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- RecordReaderGenericRdfAccumulatingBase<U,G,A,T> - Class in net.sansa_stack.hadoop.format.jena.base
- RecordReaderGenericRdfAccumulatingBase(String, String, String, String, Pattern, Lang, Accumulating<U, G, A, T>) - Constructor for class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfAccumulatingBase
- RecordReaderGenericRdfBase<U,G,A,T> - Class in net.sansa_stack.hadoop.format.jena.base
- RecordReaderGenericRdfBase(String, String, String, String, Pattern, Lang, Accumulating<U, G, A, T>) - Constructor for class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfBase
- RecordReaderGenericRdfNonAccumulatingBase<T> - Class in net.sansa_stack.hadoop.format.jena.base
- RecordReaderGenericRdfNonAccumulatingBase(String, String, String, String, Pattern, Lang) - Constructor for class net.sansa_stack.hadoop.format.jena.base.RecordReaderGenericRdfNonAccumulatingBase
- RecordReaderJsonArray - Class in net.sansa_stack.hadoop.format.gson.json
- RecordReaderJsonArray() - Constructor for class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
- RecordReaderJsonArray(Gson) - Constructor for class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
- RecordReaderJsonArray(String, String, String, Pattern, Gson) - Constructor for class net.sansa_stack.hadoop.format.gson.json.RecordReaderJsonArray
- RecordReaderRdfTrigDataset - Class in net.sansa_stack.hadoop.format.jena.trig
-
RecordReader for the Trig RDF format that groups consecutive quads having the same IRI for the graph component into Datasets.
- RecordReaderRdfTrigDataset() - Constructor for class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset
- RecordReaderRdfTrigDataset.AccumulatingDataset - Class in net.sansa_stack.hadoop.format.jena.trig
- RecordReaderRdfTrigQuad - Class in net.sansa_stack.hadoop.format.jena.trig
- RecordReaderRdfTrigQuad() - Constructor for class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigQuad
- RecordReaderRdfTurtleTriple - Class in net.sansa_stack.hadoop.format.jena.turtle
- RecordReaderRdfTurtleTriple() - Constructor for class net.sansa_stack.hadoop.format.jena.turtle.RecordReaderRdfTurtleTriple
- recordStartPattern - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
-
Regex pattern to search for candidate record starts used to avoid having to invoke the actual parser (which may start a new thread) on each single character
- requestedCsvFormat - Variable in class net.sansa_stack.hadoop.format.commons_csv.csv.RecordReaderCsv
- requestedCsvFormat - Variable in class net.sansa_stack.hadoop.format.univocity.csv.csv.RecordReaderCsv
S
- safeRead - Variable in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
-
Unsafe reads modify the byte after the reported number of read bytes in the read buffer.
- SCHEME_NAMES - Static variable in class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
- seek(long) - Method in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- seek(long) - Method in interface net.sansa_stack.hadoop.util.SeekableDecorator
- seek(long) - Method in class net.sansa_stack.hadoop.util.SeekablePushbackInputStream
- seekable - Variable in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- seekable - Variable in class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- seekable - Variable in class net.sansa_stack.hadoop.util.SeekableInputStream
- seekable - Variable in class net.sansa_stack.hadoop.util.SeekablePushbackInputStream
- SeekableByteChannelDecorator - Interface in net.sansa_stack.nio.util
- SeekableByteChannelDecoratorBase<T extends SeekableByteChannel> - Class in net.sansa_stack.nio.util
- SeekableByteChannelDecoratorBase(T) - Constructor for class net.sansa_stack.nio.util.SeekableByteChannelDecoratorBase
- SeekableByteChannelFromSeekableInputStream - Class in net.sansa_stack.hadoop.util
- SeekableByteChannelFromSeekableInputStream(InputStream) - Constructor for class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- SeekableByteChannelFromSeekableInputStream(InputStream, Seekable) - Constructor for class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- SeekableDecorator - Interface in net.sansa_stack.hadoop.util
- SeekableInputStream - Class in net.sansa_stack.hadoop.util
-
A basic wrapper that combines Hadoop's Seekable and InputStream into one class.
- SeekableInputStream(InputStream) - Constructor for class net.sansa_stack.hadoop.util.SeekableInputStream
- SeekableInputStream(InputStream, Seekable) - Constructor for class net.sansa_stack.hadoop.util.SeekableInputStream
-
Constructs a new ProxyInputStream.
- SeekablePushbackInputStream - Class in net.sansa_stack.hadoop.util
- SeekablePushbackInputStream(InputStream, int) - Constructor for class net.sansa_stack.hadoop.util.SeekablePushbackInputStream
- SeekablePushbackInputStream(InputStream, Seekable, int) - Constructor for class net.sansa_stack.hadoop.util.SeekablePushbackInputStream
- seekToNewSource(long) - Method in class net.sansa_stack.hadoop.util.DeferredSeekablePushbackInputStream
- seekToNewSource(long) - Method in interface net.sansa_stack.hadoop.util.SeekableDecorator
- seekToNewSource(long) - Method in class net.sansa_stack.hadoop.util.SeekablePushbackInputStream
- setCsvFormat(Configuration, CSVFormat) - Static method in class net.sansa_stack.hadoop.format.commons_csv.csv.FileInputFormatCsv
- setSerializable(Configuration, String, Serializable) - Static method in class net.sansa_stack.hadoop.util.ConfigurationUtils
-
Set a serializable object as a base64 url encoded string
- setStreamToInterval(long, long) - Method in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
-
Seek to a given offset and prepare to read up to the 'end' position (exclusive) For non-encoded streams this is just performs a seek on th stream and returns start/end unchanged.
- size() - Method in class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- size() - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
- split - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- splitEnd - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- splitLength - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- splitStart - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- stackTraceConsumer - Variable in class net.sansa_stack.hadoop.util.InputStreamWithCloseLogging
- stream - Variable in class net.sansa_stack.hadoop.core.RecordReaderGenericBase
- streamFileSplits(Path, long, long) - Static method in class net.sansa_stack.hadoop.util.FileSplitUtils
-
Utility method to create a specific number of splits for a file.
T
- testForEof - Variable in class net.sansa_stack.nio.util.ReadableByteChannelWithConditionalBound
- trigFwdPattern - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigDataset
- trigFwdPattern - Static variable in class net.sansa_stack.hadoop.format.jena.trig.RecordReaderRdfTrigQuad
- truncate(long) - Method in class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- truncate(long) - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
- turtleRecordStartPattern - Static variable in class net.sansa_stack.hadoop.format.jena.turtle.RecordReaderRdfTurtleTriple
-
Syntatic constructs in Turtle can start with: TODO Anything missing? base / @base prefix / @prefix @lt;foo;> - an IRI [ ] - a blank node foo: - a CURIE
W
- wrap(InputStream, Function<? super Throwable, String>, Consumer<? super String>) - Static method in class net.sansa_stack.hadoop.util.InputStreamWithCloseLogging
-
Convenience method for e.g.
- write(ByteBuffer) - Method in class net.sansa_stack.hadoop.util.SeekableByteChannelFromSeekableInputStream
- write(ByteBuffer) - Method in interface net.sansa_stack.nio.util.SeekableByteChannelDecorator
X
- xschemeNames - Variable in class net.sansa_stack.hadoop.jena.locator.LocatorHdfs
All Classes All Packages