| 程序包 | 说明 |
|---|---|
| org.apache.hadoop.hbase.mob | |
| org.apache.hadoop.hbase.regionserver | |
| org.apache.hadoop.hbase.regionserver.compactions | |
| org.apache.hadoop.hbase.regionserver.throttle |
| 限定符和类型 | 方法和说明 |
|---|---|
List<org.apache.hadoop.fs.Path> |
DefaultMobStoreCompactor.compact(CompactionRequestImpl request,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DefaultMobStoreFlusher.flushSnapshot(MemStoreSnapshot snapshot,
long cacheFlushId,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker)
Flushes the snapshot of the MemStore.
|
protected boolean |
DefaultMobStoreCompactor.performCompaction(Compactor.FileDetails fd,
InternalScanner scanner,
CellSink writer,
long smallestReadPoint,
boolean cleanSeqId,
ThroughputController throughputController,
boolean major,
int numofFilesToCompact)
Performs compaction on a column family with the mob flag enabled.
|
protected void |
DefaultMobStoreFlusher.performMobFlush(MemStoreSnapshot snapshot,
long cacheFlushId,
InternalScanner scanner,
StoreFileWriter writer,
MonitoredTask status,
ThroughputController throughputController)
Flushes the cells in the mob store.
|
| 限定符和类型 | 方法和说明 |
|---|---|
ThroughputController |
CompactSplit.getCompactionThroughputController() |
ThroughputController |
HRegionServer.getFlushThroughputController() |
ThroughputController |
RegionServerServices.getFlushThroughputController() |
| 限定符和类型 | 方法和说明 |
|---|---|
boolean |
HRegion.compact(CompactionContext compaction,
HStore store,
ThroughputController throughputController)
Called by compaction thread and after region is opened to compact the
HStores if necessary.
|
boolean |
HRegion.compact(CompactionContext compaction,
HStore store,
ThroughputController throughputController,
User user) |
List<HStoreFile> |
HStore.compact(CompactionContext compaction,
ThroughputController throughputController,
User user)
Compact the StoreFiles.
|
protected List<org.apache.hadoop.fs.Path> |
HStore.flushCache(long logCacheFlushId,
MemStoreSnapshot snapshot,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker)
Write out current snapshot.
|
List<org.apache.hadoop.fs.Path> |
StripeStoreFlusher.flushSnapshot(MemStoreSnapshot snapshot,
long cacheFlushSeqNum,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker) |
List<org.apache.hadoop.fs.Path> |
DefaultStoreFlusher.flushSnapshot(MemStoreSnapshot snapshot,
long cacheFlushId,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker) |
| 限定符和类型 | 方法和说明 |
|---|---|
protected List<org.apache.hadoop.fs.Path> |
Compactor.compact(CompactionRequestImpl request,
Compactor.InternalScannerFactory scannerFactory,
Compactor.CellSinkFactory<T> sinkFactory,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactor.compact(CompactionRequestImpl request,
int targetCount,
long targetSize,
byte[] left,
byte[] right,
byte[] majorRangeFromRow,
byte[] majorRangeToRow,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactor.compact(CompactionRequestImpl request,
List<byte[]> targetBoundaries,
byte[] majorRangeFromRow,
byte[] majorRangeToRow,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DateTieredCompactor.compact(CompactionRequestImpl request,
List<Long> lowerBoundaries,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DefaultCompactor.compact(CompactionRequestImpl request,
ThroughputController throughputController,
User user)
Do a minor/major compaction on an explicit set of storefiles from a Store.
|
abstract List<org.apache.hadoop.fs.Path> |
CompactionContext.compact(ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactionPolicy.StripeCompactionRequest.execute(StripeCompactor compactor,
ThroughputController throughputController) |
abstract List<org.apache.hadoop.fs.Path> |
StripeCompactionPolicy.StripeCompactionRequest.execute(StripeCompactor compactor,
ThroughputController throughputController,
User user)
Executes the request against compactor (essentially, just calls correct overload of
compact method), to simulate more dynamic dispatch.
|
protected boolean |
Compactor.performCompaction(Compactor.FileDetails fd,
InternalScanner scanner,
CellSink writer,
long smallestReadPoint,
boolean cleanSeqId,
ThroughputController throughputController,
boolean major,
int numofFilesToCompact)
Performs the compaction.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
NoLimitThroughputController |
class |
PressureAwareCompactionThroughputController
A throughput controller which uses the follow schema to limit throughput
If compaction pressure is greater than 1.0, no limitation.
|
class |
PressureAwareFlushThroughputController
A throughput controller which uses the follow schema to limit throughput
If flush pressure is greater than or equal to 1.0, no limitation.
|
class |
PressureAwareThroughputController |
| 限定符和类型 | 方法和说明 |
|---|---|
static ThroughputController |
FlushThroughputControllerFactory.create(RegionServerServices server,
org.apache.hadoop.conf.Configuration conf) |
static ThroughputController |
CompactionThroughputControllerFactory.create(RegionServerServices server,
org.apache.hadoop.conf.Configuration conf) |
| 限定符和类型 | 方法和说明 |
|---|---|
static Class<? extends ThroughputController> |
FlushThroughputControllerFactory.getThroughputControllerClass(org.apache.hadoop.conf.Configuration conf) |
static Class<? extends ThroughputController> |
CompactionThroughputControllerFactory.getThroughputControllerClass(org.apache.hadoop.conf.Configuration conf) |
Copyright © 2007–2019 The Apache Software Foundation. All rights reserved.