Uses of Class
org.apache.hadoop.hbase.client.HTable

Packages that use HTable
org.apache.hadoop.hbase.catalog   
org.apache.hadoop.hbase.client Provides HBase Client 
org.apache.hadoop.hbase.client.coprocessor Provides client classes for invoking Coprocessor RPC protocols 
org.apache.hadoop.hbase.mapred Provides HBase MapReduce Input/OutputFormats, a table indexing MapReduce job, and utility 
org.apache.hadoop.hbase.mapreduce Provides HBase MapReduce Input/OutputFormats, a table indexing MapReduce job, and utility 
org.apache.hadoop.hbase.thrift Provides an HBase Thrift service. 
org.apache.hadoop.hbase.util   
 

Uses of HTable in org.apache.hadoop.hbase.catalog
 

Methods in org.apache.hadoop.hbase.catalog with parameters of type HTable
static void MetaEditor.addRegionToMeta(HTable meta, HRegionInfo regionInfo)
          Adds a hbase:meta row for the specified new region to the given catalog table.
static void MetaEditor.addRegionToMeta(HTable meta, HRegionInfo regionInfo, HRegionInfo splitA, HRegionInfo splitB)
          Adds a (single) hbase:meta row for the specified new region and its daughters.
 

Uses of HTable in org.apache.hadoop.hbase.client
 

Methods in org.apache.hadoop.hbase.client with parameters of type HTable
static void HTableUtil.bucketRsBatch(HTable htable, List<Row> rows)
          Processes a List of Rows (Put, Delete) and writes them to an HTable instance in RegionServer buckets via the htable.batch method.
static void HTableUtil.bucketRsPut(HTable htable, List<Put> puts)
          Processes a List of Puts and writes them to an HTable instance in RegionServer buckets via the htable.put method.
 

Uses of HTable in org.apache.hadoop.hbase.client.coprocessor
 

Methods in org.apache.hadoop.hbase.client.coprocessor with parameters of type HTable
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message>
double
AggregationClient.avg(HTable table, ColumnInterpreter<R,S,P,Q,T> ci, Scan scan)
          This is the client side interface/handle for calling the average method for a given cf-cq combination.
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message>
R
AggregationClient.max(HTable table, ColumnInterpreter<R,S,P,Q,T> ci, Scan scan)
          It gives the maximum value of a column for a given column family for the given range.
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message>
R
AggregationClient.median(HTable table, ColumnInterpreter<R,S,P,Q,T> ci, Scan scan)
          This is the client side interface/handler for calling the median method for a given cf-cq combination.
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message>
R
AggregationClient.min(HTable table, ColumnInterpreter<R,S,P,Q,T> ci, Scan scan)
          It gives the minimum value of a column for a given column family for the given range.
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message>
long
AggregationClient.rowCount(HTable table, ColumnInterpreter<R,S,P,Q,T> ci, Scan scan)
          It gives the row count, by summing up the individual results obtained from regions.
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message>
double
AggregationClient.std(HTable table, ColumnInterpreter<R,S,P,Q,T> ci, Scan scan)
          This is the client side interface/handle for calling the std method for a given cf-cq combination.
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message>
S
AggregationClient.sum(HTable table, ColumnInterpreter<R,S,P,Q,T> ci, Scan scan)
          It sums up the value returned from various regions.
 

Constructors in org.apache.hadoop.hbase.client.coprocessor with parameters of type HTable
SecureBulkLoadClient(HTable table)
           
 

Uses of HTable in org.apache.hadoop.hbase.mapred
 

Methods in org.apache.hadoop.hbase.mapred that return HTable
protected  HTable TableInputFormatBase.getHTable()
          Deprecated. Allows subclasses to get the HTable.
 

Methods in org.apache.hadoop.hbase.mapred with parameters of type HTable
 void TableRecordReader.setHTable(HTable htable)
          Deprecated.  
protected  void TableInputFormatBase.setHTable(HTable table)
          Deprecated. Allows subclasses to set the HTable.
 void TableRecordReaderImpl.setHTable(HTable htable)
          Deprecated.  
 

Constructors in org.apache.hadoop.hbase.mapred with parameters of type HTable
TableOutputFormat.TableRecordWriter(HTable table)
          Instantiate a TableRecordWriter with the HBase HClient for writing.
 

Uses of HTable in org.apache.hadoop.hbase.mapreduce
 

Methods in org.apache.hadoop.hbase.mapreduce that return HTable
protected  HTable TableInputFormatBase.getHTable()
          Allows subclasses to get the HTable.
 

Methods in org.apache.hadoop.hbase.mapreduce with parameters of type HTable
protected  void LoadIncrementalHFiles.bulkLoadPhase(HTable table, HConnection conn, ExecutorService pool, Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue, com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups)
          This takes the LQI's grouped by likely regions and attempts to bulk load them.
static void HFileOutputFormat2.configureIncrementalLoad(org.apache.hadoop.mapreduce.Job job, HTable table)
          Configure a MapReduce Job to perform an incremental load into the given table.
static void HFileOutputFormat.configureIncrementalLoad(org.apache.hadoop.mapreduce.Job job, HTable table)
          Deprecated. Configure a MapReduce Job to perform an incremental load into the given table.
static void HFileOutputFormat2.configureIncrementalLoadMap(org.apache.hadoop.mapreduce.Job job, HTable table)
           
 void LoadIncrementalHFiles.doBulkLoad(org.apache.hadoop.fs.Path hfofDir, HTable table)
          Perform a bulk load of the given directory into the given pre-existing table.
protected  List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> LoadIncrementalHFiles.groupOrSplit(com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups, org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem item, HTable table, Pair<byte[][],byte[][]> startEndKeys)
          Attempt to assign the given load queue item into its target region group.
 void TableRecordReader.setHTable(HTable htable)
          Sets the HBase table.
protected  void TableInputFormatBase.setHTable(HTable table)
          Allows subclasses to set the HTable.
 void TableRecordReaderImpl.setHTable(HTable htable)
          Sets the HBase table.
protected  List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> LoadIncrementalHFiles.splitStoreFile(org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem item, HTable table, byte[] startKey, byte[] splitKey)
           
 

Constructors in org.apache.hadoop.hbase.mapreduce with parameters of type HTable
TableOutputFormat.TableRecordWriter(HTable table)
          Instantiate a TableRecordWriter with the HBase HClient for writing.
 

Uses of HTable in org.apache.hadoop.hbase.thrift
 

Methods in org.apache.hadoop.hbase.thrift that return HTable
 HTable ThriftServerRunner.HBaseHandler.getTable(byte[] tableName)
          Creates and returns an HTable instance from a given table name.
 HTable ThriftServerRunner.HBaseHandler.getTable(ByteBuffer tableName)
           
 

Uses of HTable in org.apache.hadoop.hbase.util
 

Constructors in org.apache.hadoop.hbase.util with parameters of type HTable
RegionSizeCalculator(HTable table)
          Computes size of each region for table and given column families.
 



Copyright © 2015 The Apache Software Foundation. All rights reserved.