Package | Description |
---|---|
org.apache.hadoop.hbase.backup | |
org.apache.hadoop.hbase.coprocessor |
Table of Contents
|
org.apache.hadoop.hbase.mob | |
org.apache.hadoop.hbase.mob.compactions | |
org.apache.hadoop.hbase.regionserver | |
org.apache.hadoop.hbase.regionserver.compactions |
Modifier and Type | Method and Description |
---|---|
static void |
HFileArchiver.archiveStoreFiles(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
HRegionInfo regionInfo,
org.apache.hadoop.fs.Path tableDir,
byte[] family,
Collection<StoreFile> compactedFiles)
Remove the store files, either by archiving them or outright deletion
|
Modifier and Type | Method and Description |
---|---|
default void |
RegionObserver.postCompact(ObserverContext<RegionCoprocessorEnvironment> c,
Store store,
StoreFile resultFile,
CompactionLifeCycleTracker tracker)
Called after compaction has completed and the new store file has been moved in to place.
|
default void |
RegionObserver.postFlush(ObserverContext<RegionCoprocessorEnvironment> c,
Store store,
StoreFile resultFile)
Called after a Store's memstore is flushed to disk.
|
Modifier and Type | Method and Description |
---|---|
default void |
RegionObserver.postCompactSelection(ObserverContext<RegionCoprocessorEnvironment> c,
Store store,
org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableList<StoreFile> selected,
CompactionLifeCycleTracker tracker)
Called after the
StoreFile s to compact have been selected from the available
candidates. |
default void |
RegionObserver.preCompactSelection(ObserverContext<RegionCoprocessorEnvironment> c,
Store store,
List<StoreFile> candidates,
CompactionLifeCycleTracker tracker)
Called prior to selecting the
StoreFiles to compact from the list of
available candidates. |
Modifier and Type | Method and Description |
---|---|
static void |
MobUtils.removeMobFiles(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
TableName tableName,
org.apache.hadoop.fs.Path tableDir,
byte[] family,
Collection<StoreFile> storeFiles)
Archives the mob files.
|
Constructor and Description |
---|
CachedMobFile(StoreFile sf) |
MobFile(StoreFile sf) |
Modifier and Type | Method and Description |
---|---|
List<StoreFile> |
PartitionedMobCompactionRequest.CompactionDelPartition.getStoreFiles() |
Modifier and Type | Method and Description |
---|---|
void |
PartitionedMobCompactionRequest.CompactionDelPartition.addStoreFile(StoreFile file) |
Modifier and Type | Class and Description |
---|---|
class |
HStoreFile
A Store data file.
|
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableCollection<StoreFile> |
StripeStoreFileManager.clearCompactedFiles() |
Collection<StoreFile> |
StoreFileManager.clearCompactedFiles()
Clears all the compacted files and returns them.
|
org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableCollection<StoreFile> |
StripeStoreFileManager.clearFiles() |
org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableCollection<StoreFile> |
StoreFileManager.clearFiles()
Clears all the files currently in use and returns them.
|
Collection<StoreFile> |
Store.close()
Close all the readers We don't need to worry about subsequent requests because the Region
holds a write lock that will prevent any more reads or writes.
|
org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableCollection<StoreFile> |
HStore.close() |
Map<byte[],List<StoreFile>> |
HRegion.close()
Close down this HRegion.
|
Map<byte[],List<StoreFile>> |
HRegion.close(boolean abort)
Close down this HRegion.
|
List<StoreFile> |
Store.compact(CompactionContext compaction,
ThroughputController throughputController)
Deprecated.
see compact(CompactionContext, ThroughputController, User)
|
List<StoreFile> |
HStore.compact(CompactionContext compaction,
ThroughputController throughputController)
Compact the StoreFiles.
|
List<StoreFile> |
Store.compact(CompactionContext compaction,
ThroughputController throughputController,
User user) |
List<StoreFile> |
HStore.compact(CompactionContext compaction,
ThroughputController throughputController,
User user) |
Iterator<StoreFile> |
StripeStoreFileManager.getCandidateFilesForRowKeyBefore(KeyValue targetKey)
See
StoreFileManager.getCandidateFilesForRowKeyBefore(KeyValue)
for details on this methods. |
Iterator<StoreFile> |
StoreFileManager.getCandidateFilesForRowKeyBefore(KeyValue targetKey)
Gets initial, full list of candidate store files to check for row-key-before.
|
Collection<StoreFile> |
StripeStoreFileManager.getCompactedfiles() |
Collection<StoreFile> |
StoreFileManager.getCompactedfiles()
List of compacted files inside this store that needs to be excluded in reads
because further new reads will be using only the newly created files out of compaction.
|
Collection<StoreFile> |
Store.getCompactedFiles() |
Collection<StoreFile> |
HStore.getCompactedFiles() |
Collection<StoreFile> |
StripeStoreFileManager.getFilesForScan(byte[] startRow,
boolean includeStartRow,
byte[] stopRow,
boolean includeStopRow) |
Collection<StoreFile> |
StoreFileManager.getFilesForScan(byte[] startRow,
boolean includeStartRow,
byte[] stopRow,
boolean includeStopRow)
Gets the store files to scan for a Scan or Get request.
|
List<StoreFile> |
StripeStoreFileManager.getLevel0Files() |
Comparator<StoreFile> |
StripeStoreFileManager.getStoreFileComparator() |
Comparator<StoreFile> |
StoreFileManager.getStoreFileComparator() |
Collection<StoreFile> |
StripeStoreFileManager.getStorefiles() |
Collection<StoreFile> |
StoreFileManager.getStorefiles()
Gets the snapshot of the store files currently in use.
|
Collection<StoreFile> |
Store.getStorefiles() |
Collection<StoreFile> |
HStore.getStorefiles() |
ArrayList<org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableList<StoreFile>> |
StripeStoreFileManager.getStripes() |
Collection<StoreFile> |
StripeStoreFileManager.getUnneededFiles(long maxTs,
List<StoreFile> filesCompacting) |
Collection<StoreFile> |
StoreFileManager.getUnneededFiles(long maxTs,
List<StoreFile> filesCompacting) |
Iterator<StoreFile> |
StripeStoreFileManager.updateCandidateFilesForRowKeyBefore(Iterator<StoreFile> candidateFiles,
KeyValue targetKey,
Cell candidate)
See
StoreFileManager.getCandidateFilesForRowKeyBefore(KeyValue) and
StoreFileManager.updateCandidateFilesForRowKeyBefore(Iterator, KeyValue, Cell)
for details on this methods. |
Iterator<StoreFile> |
StoreFileManager.updateCandidateFilesForRowKeyBefore(Iterator<StoreFile> candidateFiles,
KeyValue targetKey,
Cell candidate)
Updates the candidate list for finding row key before.
|
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.fs.Path |
HRegionFileSystem.mergeStoreFile(HRegionInfo mergedRegion,
String familyName,
StoreFile f,
org.apache.hadoop.fs.Path mergedDir)
Write out a merge reference under the given merges directory.
|
void |
RegionCoprocessorHost.postCompact(Store store,
StoreFile resultFile,
CompactionLifeCycleTracker tracker,
User user)
Called after the store compaction has completed.
|
void |
RegionCoprocessorHost.postFlush(Store store,
StoreFile storeFile)
Invoked after a memstore flush
|
org.apache.hadoop.fs.Path |
HRegionFileSystem.splitStoreFile(HRegionInfo hri,
String familyName,
StoreFile f,
byte[] splitRow,
boolean top,
RegionSplitPolicy splitPolicy)
Write out a split reference.
|
Modifier and Type | Method and Description |
---|---|
void |
StripeStoreFileManager.addCompactionResults(Collection<StoreFile> compactedFiles,
Collection<StoreFile> results) |
void |
StripeStoreFileManager.addCompactionResults(Collection<StoreFile> compactedFiles,
Collection<StoreFile> results) |
void |
StoreFileManager.addCompactionResults(Collection<StoreFile> compactedFiles,
Collection<StoreFile> results)
Adds only the new compaction results into the structure.
|
void |
StoreFileManager.addCompactionResults(Collection<StoreFile> compactedFiles,
Collection<StoreFile> results)
Adds only the new compaction results into the structure.
|
protected void |
HStore.completeCompaction(Collection<StoreFile> compactedFiles)
It works by processing a compaction that's been written to disk.
|
static OptionalInt |
StoreUtils.getDeterministicRandomSeed(Collection<StoreFile> files)
Creates a deterministic hash code for store file collection.
|
static long |
StoreUtils.getLowestTimestamp(Collection<StoreFile> candidates)
Gets lowest timestamp from candidate StoreFiles
|
static long |
StoreUtils.getMaxMemstoreTSInList(Collection<StoreFile> sfs)
Return the largest memstoreTS found across all storefiles in the given list.
|
static long |
StoreUtils.getMaxSequenceIdInList(Collection<StoreFile> sfs)
Return the highest sequence ID found across all storefiles in
the given list.
|
default List<KeyValueScanner> |
Store.getScanners(List<StoreFile> files,
boolean cacheBlocks,
boolean isGet,
boolean usePread,
boolean isCompaction,
ScanQueryMatcher matcher,
byte[] startRow,
byte[] stopRow,
long readPt,
boolean includeMemstoreScanner)
Create scanners on the given files and if needed on the memstore with no filtering based on TTL
(that happens further down the line).
|
List<KeyValueScanner> |
Store.getScanners(List<StoreFile> files,
boolean cacheBlocks,
boolean usePread,
boolean isCompaction,
ScanQueryMatcher matcher,
byte[] startRow,
boolean includeStartRow,
byte[] stopRow,
boolean includeStopRow,
long readPt,
boolean includeMemstoreScanner)
Create scanners on the given files and if needed on the memstore with no filtering based on TTL
(that happens further down the line).
|
List<KeyValueScanner> |
HStore.getScanners(List<StoreFile> files,
boolean cacheBlocks,
boolean usePread,
boolean isCompaction,
ScanQueryMatcher matcher,
byte[] startRow,
boolean includeStartRow,
byte[] stopRow,
boolean includeStopRow,
long readPt,
boolean includeMemstoreScanner) |
static List<StoreFileScanner> |
StoreFileScanner.getScannersForCompaction(Collection<StoreFile> files,
boolean canUseDropBehind,
long readPt)
Get scanners for compaction.
|
static List<StoreFileScanner> |
StoreFileScanner.getScannersForStoreFiles(Collection<StoreFile> files,
boolean cacheBlocks,
boolean usePread,
boolean isCompaction,
boolean useDropBehind,
long readPt)
Return an array of scanners corresponding to the given set of store files.
|
static List<StoreFileScanner> |
StoreFileScanner.getScannersForStoreFiles(Collection<StoreFile> files,
boolean cacheBlocks,
boolean usePread,
boolean isCompaction,
boolean canUseDrop,
ScanQueryMatcher matcher,
long readPt)
Return an array of scanners corresponding to the given set of store files, And set the
ScanQueryMatcher for each store file scanner for further optimization
|
static List<StoreFileScanner> |
StoreFileScanner.getScannersForStoreFiles(Collection<StoreFile> files,
boolean cacheBlocks,
boolean usePread,
long readPt)
Return an array of scanners corresponding to the given set of store files.
|
Collection<StoreFile> |
StripeStoreFileManager.getUnneededFiles(long maxTs,
List<StoreFile> filesCompacting) |
Collection<StoreFile> |
StoreFileManager.getUnneededFiles(long maxTs,
List<StoreFile> filesCompacting) |
static boolean |
StoreUtils.hasReferences(Collection<StoreFile> files)
Determines whether any files in the collection are references.
|
void |
StripeStoreFileManager.insertNewFiles(Collection<StoreFile> sfs) |
void |
StoreFileManager.insertNewFiles(Collection<StoreFile> sfs)
Adds new files, either for from MemStore flush or bulk insert, into the structure.
|
void |
StripeStoreFileManager.loadFiles(List<StoreFile> storeFiles) |
void |
StoreFileManager.loadFiles(List<StoreFile> storeFiles)
Loads the initial store files into empty StoreFileManager.
|
boolean |
StripeStoreEngine.needsCompaction(List<StoreFile> filesCompacting) |
abstract boolean |
StoreEngine.needsCompaction(List<StoreFile> filesCompacting) |
boolean |
DefaultStoreEngine.needsCompaction(List<StoreFile> filesCompacting) |
boolean |
DateTieredStoreEngine.needsCompaction(List<StoreFile> filesCompacting) |
void |
RegionCoprocessorHost.postCompactSelection(Store store,
org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableList<StoreFile> selected,
CompactionLifeCycleTracker tracker,
User user)
Called after the
StoreFile s to be compacted have been selected from the available
candidates. |
boolean |
RegionCoprocessorHost.preCompactSelection(Store store,
List<StoreFile> candidates,
CompactionLifeCycleTracker tracker,
User user)
Called prior to selecting the
StoreFile s for compaction from the list of currently
available candidates. |
void |
StripeStoreFileManager.removeCompactedFiles(Collection<StoreFile> compactedFiles) |
void |
StoreFileManager.removeCompactedFiles(Collection<StoreFile> compactedFiles)
Remove the compacted files
|
void |
HRegionFileSystem.removeStoreFiles(String familyName,
Collection<StoreFile> storeFiles)
Closes and archives the specified store files from the specified family.
|
Iterator<StoreFile> |
StripeStoreFileManager.updateCandidateFilesForRowKeyBefore(Iterator<StoreFile> candidateFiles,
KeyValue targetKey,
Cell candidate)
See
StoreFileManager.getCandidateFilesForRowKeyBefore(KeyValue) and
StoreFileManager.updateCandidateFilesForRowKeyBefore(Iterator, KeyValue, Cell)
for details on this methods. |
Iterator<StoreFile> |
StoreFileManager.updateCandidateFilesForRowKeyBefore(Iterator<StoreFile> candidateFiles,
KeyValue targetKey,
Cell candidate)
Updates the candidate list for finding row key before.
|
void |
StoreScanner.updateReaders(List<StoreFile> sfs,
List<KeyValueScanner> memStoreScanners) |
void |
ChangedReadersObserver.updateReaders(List<StoreFile> sfs,
List<KeyValueScanner> memStoreScanners)
Notify observers.
|
Modifier and Type | Method and Description |
---|---|
protected ArrayList<StoreFile> |
RatioBasedCompactionPolicy.applyCompactionPolicy(ArrayList<StoreFile> candidates,
boolean mayUseOffPeak,
boolean mayBeStuck)
-- Default minor compaction selection algorithm:
choose CompactSelection from candidates --
First exclude bulk-load files if indicated in configuration.
|
protected ArrayList<StoreFile> |
ExploringCompactionPolicy.applyCompactionPolicy(ArrayList<StoreFile> candidates,
boolean mayUseOffPeak,
boolean mightBeStuck) |
List<StoreFile> |
ExploringCompactionPolicy.applyCompactionPolicy(List<StoreFile> candidates,
boolean mightBeStuck,
boolean mayUseOffPeak,
int minFiles,
int maxFiles) |
protected ArrayList<StoreFile> |
SortedCompactionPolicy.checkMinFilesCriteria(ArrayList<StoreFile> candidates,
int minFiles) |
protected ArrayList<StoreFile> |
SortedCompactionPolicy.filterBulk(ArrayList<StoreFile> candidates) |
protected ArrayList<StoreFile> |
SortedCompactionPolicy.getCurrentEligibleFiles(ArrayList<StoreFile> candidateFiles,
List<StoreFile> filesCompacting) |
Collection<StoreFile> |
CompactionRequest.getFiles() |
List<StoreFile> |
StripeCompactionPolicy.StripeInformationProvider.getLevel0Files() |
Collection<StoreFile> |
StripeCompactionPolicy.StripeInformationProvider.getStorefiles() |
ArrayList<org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableList<StoreFile>> |
StripeCompactionPolicy.StripeInformationProvider.getStripes() |
abstract List<StoreFile> |
CompactionContext.preSelect(List<StoreFile> filesCompacting)
Called before coprocessor preCompactSelection and should filter the candidates
for coprocessor; i.e.
|
List<StoreFile> |
SortedCompactionPolicy.preSelectCompactionForCoprocessor(Collection<StoreFile> candidates,
List<StoreFile> filesCompacting) |
List<StoreFile> |
StripeCompactionPolicy.preSelectFilesForCoprocessor(StripeCompactionPolicy.StripeInformationProvider si,
List<StoreFile> filesCompacting) |
protected ArrayList<StoreFile> |
SortedCompactionPolicy.skipLargeFiles(ArrayList<StoreFile> candidates,
boolean mayUseOffpeak) |
Modifier and Type | Method and Description |
---|---|
protected ArrayList<StoreFile> |
RatioBasedCompactionPolicy.applyCompactionPolicy(ArrayList<StoreFile> candidates,
boolean mayUseOffPeak,
boolean mayBeStuck)
-- Default minor compaction selection algorithm:
choose CompactSelection from candidates --
First exclude bulk-load files if indicated in configuration.
|
protected ArrayList<StoreFile> |
ExploringCompactionPolicy.applyCompactionPolicy(ArrayList<StoreFile> candidates,
boolean mayUseOffPeak,
boolean mightBeStuck) |
List<StoreFile> |
ExploringCompactionPolicy.applyCompactionPolicy(List<StoreFile> candidates,
boolean mightBeStuck,
boolean mayUseOffPeak,
int minFiles,
int maxFiles) |
protected ArrayList<StoreFile> |
SortedCompactionPolicy.checkMinFilesCriteria(ArrayList<StoreFile> candidates,
int minFiles) |
List<org.apache.hadoop.fs.Path> |
DefaultCompactor.compactForTesting(Collection<StoreFile> filesToCompact,
boolean isMajor)
Compact a list of files for testing.
|
protected abstract CompactionRequest |
SortedCompactionPolicy.createCompactionRequest(ArrayList<StoreFile> candidateSelection,
boolean tryingMajor,
boolean mayUseOffPeak,
boolean mayBeStuck) |
protected CompactionRequest |
RatioBasedCompactionPolicy.createCompactionRequest(ArrayList<StoreFile> candidateSelection,
boolean tryingMajor,
boolean mayUseOffPeak,
boolean mayBeStuck) |
protected CompactionRequest |
DateTieredCompactionPolicy.createCompactionRequest(ArrayList<StoreFile> candidateSelection,
boolean tryingMajor,
boolean mayUseOffPeak,
boolean mayBeStuck) |
protected List<StoreFileScanner> |
Compactor.createFileScanners(Collection<StoreFile> filesToCompact,
long smallestReadPoint,
boolean useDropBehind)
Creates file scanners for compaction.
|
protected ArrayList<StoreFile> |
SortedCompactionPolicy.filterBulk(ArrayList<StoreFile> candidates) |
protected ArrayList<StoreFile> |
SortedCompactionPolicy.getCurrentEligibleFiles(ArrayList<StoreFile> candidateFiles,
List<StoreFile> filesCompacting) |
protected ArrayList<StoreFile> |
SortedCompactionPolicy.getCurrentEligibleFiles(ArrayList<StoreFile> candidateFiles,
List<StoreFile> filesCompacting) |
protected Compactor.FileDetails |
Compactor.getFileDetails(Collection<StoreFile> filesToCompact,
boolean allFiles)
Extracts some details about the files to compact that are commonly needed by compactors.
|
long |
SortedCompactionPolicy.getNextMajorCompactTime(Collection<StoreFile> filesToCompact) |
static long |
StripeCompactionPolicy.getTotalFileSize(Collection<StoreFile> candidates) |
abstract boolean |
SortedCompactionPolicy.needsCompaction(Collection<StoreFile> storeFiles,
List<StoreFile> filesCompacting) |
abstract boolean |
SortedCompactionPolicy.needsCompaction(Collection<StoreFile> storeFiles,
List<StoreFile> filesCompacting) |
boolean |
RatioBasedCompactionPolicy.needsCompaction(Collection<StoreFile> storeFiles,
List<StoreFile> filesCompacting)
A heuristic method to decide whether to schedule a compaction request
|
boolean |
RatioBasedCompactionPolicy.needsCompaction(Collection<StoreFile> storeFiles,
List<StoreFile> filesCompacting)
A heuristic method to decide whether to schedule a compaction request
|
boolean |
FIFOCompactionPolicy.needsCompaction(Collection<StoreFile> storeFiles,
List<StoreFile> filesCompacting) |
boolean |
FIFOCompactionPolicy.needsCompaction(Collection<StoreFile> storeFiles,
List<StoreFile> filesCompacting) |
boolean |
DateTieredCompactionPolicy.needsCompaction(Collection<StoreFile> storeFiles,
List<StoreFile> filesCompacting)
Heuristics for guessing whether we need minor compaction.
|
boolean |
DateTieredCompactionPolicy.needsCompaction(Collection<StoreFile> storeFiles,
List<StoreFile> filesCompacting)
Heuristics for guessing whether we need minor compaction.
|
boolean |
StripeCompactionPolicy.needsCompactions(StripeCompactionPolicy.StripeInformationProvider si,
List<StoreFile> filesCompacting) |
abstract List<StoreFile> |
CompactionContext.preSelect(List<StoreFile> filesCompacting)
Called before coprocessor preCompactSelection and should filter the candidates
for coprocessor; i.e.
|
List<StoreFile> |
SortedCompactionPolicy.preSelectCompactionForCoprocessor(Collection<StoreFile> candidates,
List<StoreFile> filesCompacting) |
List<StoreFile> |
SortedCompactionPolicy.preSelectCompactionForCoprocessor(Collection<StoreFile> candidates,
List<StoreFile> filesCompacting) |
List<StoreFile> |
StripeCompactionPolicy.preSelectFilesForCoprocessor(StripeCompactionPolicy.StripeInformationProvider si,
List<StoreFile> filesCompacting) |
protected void |
SortedCompactionPolicy.removeExcessFiles(ArrayList<StoreFile> candidates,
boolean isUserCompaction,
boolean isMajorCompaction) |
abstract boolean |
CompactionContext.select(List<StoreFile> filesCompacting,
boolean isUserCompaction,
boolean mayUseOffPeak,
boolean forceMajor)
Called to select files for compaction.
|
CompactionRequest |
SortedCompactionPolicy.selectCompaction(Collection<StoreFile> candidateFiles,
List<StoreFile> filesCompacting,
boolean isUserCompaction,
boolean mayUseOffPeak,
boolean forceMajor) |
CompactionRequest |
SortedCompactionPolicy.selectCompaction(Collection<StoreFile> candidateFiles,
List<StoreFile> filesCompacting,
boolean isUserCompaction,
boolean mayUseOffPeak,
boolean forceMajor) |
CompactionRequest |
FIFOCompactionPolicy.selectCompaction(Collection<StoreFile> candidateFiles,
List<StoreFile> filesCompacting,
boolean isUserCompaction,
boolean mayUseOffPeak,
boolean forceMajor) |
CompactionRequest |
FIFOCompactionPolicy.selectCompaction(Collection<StoreFile> candidateFiles,
List<StoreFile> filesCompacting,
boolean isUserCompaction,
boolean mayUseOffPeak,
boolean forceMajor) |
StripeCompactionPolicy.StripeCompactionRequest |
StripeCompactionPolicy.selectCompaction(StripeCompactionPolicy.StripeInformationProvider si,
List<StoreFile> filesCompacting,
boolean isOffpeak) |
CompactionRequest |
DateTieredCompactionPolicy.selectMajorCompaction(ArrayList<StoreFile> candidateSelection) |
CompactionRequest |
DateTieredCompactionPolicy.selectMinorCompaction(ArrayList<StoreFile> candidateSelection,
boolean mayUseOffPeak,
boolean mayBeStuck)
We receive store files sorted in ascending order by seqId then scan the list of files.
|
boolean |
StripeCompactionPolicy.shouldPerformMajorCompaction(Collection<StoreFile> filesToCompact) |
abstract boolean |
SortedCompactionPolicy.shouldPerformMajorCompaction(Collection<StoreFile> filesToCompact) |
boolean |
RatioBasedCompactionPolicy.shouldPerformMajorCompaction(Collection<StoreFile> filesToCompact) |
boolean |
FIFOCompactionPolicy.shouldPerformMajorCompaction(Collection<StoreFile> filesToCompact) |
boolean |
DateTieredCompactionPolicy.shouldPerformMajorCompaction(Collection<StoreFile> filesToCompact) |
abstract boolean |
CompactionPolicy.shouldPerformMajorCompaction(Collection<StoreFile> filesToCompact) |
protected ArrayList<StoreFile> |
SortedCompactionPolicy.skipLargeFiles(ArrayList<StoreFile> candidates,
boolean mayUseOffpeak) |
void |
CompactionRequest.updateFiles(Collection<StoreFile> files) |
Constructor and Description |
---|
CompactionRequest(Collection<StoreFile> files) |
DateTieredCompactionRequest(Collection<StoreFile> files,
List<Long> boundaryList) |
Copyright © 2007–2017 The Apache Software Foundation. All rights reserved.