Uses of Class
org.apache.hadoop.hbase.client.Scan

Packages that use Scan
org.apache.hadoop.hbase.client Provides HBase Client 
org.apache.hadoop.hbase.mapreduce Provides HBase MapReduce Input/OutputFormats, a table indexing MapReduce job, and utility 
org.apache.hadoop.hbase.rest.client   
 

Uses of Scan in org.apache.hadoop.hbase.client
 

Fields in org.apache.hadoop.hbase.client declared as Scan
protected  Scan ClientScanner.scan
           
 

Methods in org.apache.hadoop.hbase.client that return Scan
 Scan Scan.addColumn(byte[] family, byte[] qualifier)
          Get the column from the specified family with the specified qualifier.
 Scan Scan.addFamily(byte[] family)
          Get all columns from the specified family.
protected  Scan ClientScanner.getScan()
           
 Scan Scan.setFamilyMap(Map<byte[],NavigableSet<byte[]>> familyMap)
          Setting the familyMap
 Scan Scan.setFilter(Filter filter)
           
 Scan Scan.setMaxVersions()
          Get all available versions.
 Scan Scan.setMaxVersions(int maxVersions)
          Get up to the specified number of versions of each column.
 Scan Scan.setReversed(boolean reversed)
          Set whether this scan is a reversed one
 Scan Scan.setStartRow(byte[] startRow)
          Set the start row of the scan.
 Scan Scan.setStopRow(byte[] stopRow)
          Set the stop row.
 Scan Scan.setTimeRange(long minStamp, long maxStamp)
          Get versions of columns only within the specified timestamp range, [minStamp, maxStamp).
 Scan Scan.setTimeStamp(long timestamp)
          Get versions of columns with the specified timestamp.
 

Methods in org.apache.hadoop.hbase.client with parameters of type Scan
 ResultScanner HTable.getScanner(Scan scan)
          Returns a scanner on the current table as specified by the Scan object.
 ResultScanner HTableInterface.getScanner(Scan scan)
          Returns a scanner on the current table as specified by the Scan object.
 

Constructors in org.apache.hadoop.hbase.client with parameters of type Scan
ClientScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, byte[] tableName)
          Deprecated. Use ClientScanner.ClientScanner(Configuration, Scan, TableName)
ClientScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, byte[] tableName, HConnection connection)
          Deprecated. Use ClientScanner.ClientScanner(Configuration, Scan, TableName, HConnection)
ClientScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName)
          Deprecated. 
ClientScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName, HConnection connection)
          Create a new ClientScanner for the specified table Note that the passed Scan's start row maybe changed changed.
ClientScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName, HConnection connection, org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcFactory)
          Deprecated. Use ClientScanner.ClientScanner(Configuration, Scan, TableName, HConnection, RpcRetryingCallerFactory, RpcControllerFactory) instead
ClientScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName, HConnection connection, org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcFactory, org.apache.hadoop.hbase.ipc.RpcControllerFactory controllerFactory)
          Create a new ClientScanner for the specified table Note that the passed Scan's start row maybe changed changed.
ClientSmallReversedScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName, HConnection connection)
          Create a new ReversibleClientScanner for the specified table Note that the passed Scan's start row maybe changed.
ClientSmallScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName)
          Create a new ClientSmallScanner for the specified table.
ClientSmallScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName, HConnection connection)
          Create a new ClientSmallScanner for the specified table.
ClientSmallScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName, HConnection connection, org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcFactory, org.apache.hadoop.hbase.ipc.RpcControllerFactory controllerFactory)
          Create a new ShortClientScanner for the specified table Note that the passed Scan's start row maybe changed changed.
ReversedClientScanner(org.apache.hadoop.conf.Configuration conf, Scan scan, TableName tableName, HConnection connection)
          Create a new ReversibleClientScanner for the specified table Note that the passed Scan's start row maybe changed.
ReversedScannerCallable(HConnection connection, TableName tableName, Scan scan, org.apache.hadoop.hbase.client.metrics.ScanMetrics scanMetrics, byte[] locateStartRow)
          Deprecated. 
ReversedScannerCallable(HConnection connection, TableName tableName, Scan scan, org.apache.hadoop.hbase.client.metrics.ScanMetrics scanMetrics, byte[] locateStartRow, org.apache.hadoop.hbase.ipc.PayloadCarryingRpcController rpcFactory)
           
Scan(Scan scan)
          Creates a new instance of this class while copying all values.
TableSnapshotScanner(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path rootDir, org.apache.hadoop.fs.Path restoreDir, String snapshotName, Scan scan)
          Creates a TableSnapshotScanner.
TableSnapshotScanner(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path restoreDir, String snapshotName, Scan scan)
          Creates a TableSnapshotScanner.
 

Uses of Scan in org.apache.hadoop.hbase.mapreduce
 

Methods in org.apache.hadoop.hbase.mapreduce that return Scan
 Scan TableInputFormatBase.getScan()
          Gets the scan defining the actual details like columns etc.
 Scan TableSplit.getScan()
          Returns a Scan object from the stored string representation.
 

Methods in org.apache.hadoop.hbase.mapreduce that return types with arguments of type Scan
protected  List<Scan> MultiTableInputFormatBase.getScans()
          Allows subclasses to get the list of Scan objects.
 

Methods in org.apache.hadoop.hbase.mapreduce with parameters of type Scan
static void TableInputFormat.addColumns(Scan scan, byte[][] columns)
          Adds an array of columns specified using old format, family:qualifier.
static void IdentityTableMapper.initJob(String table, Scan scan, Class<? extends TableMapper> mapper, org.apache.hadoop.mapreduce.Job job)
          Use this before submitting a TableMap job.
static void GroupingTableMapper.initJob(String table, Scan scan, String groupColumns, Class<? extends TableMapper> mapper, org.apache.hadoop.mapreduce.Job job)
          Use this before submitting a TableMap job.
static void TableMapReduceUtil.initTableMapperJob(byte[] table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job)
          Use this before submitting a TableMap job.
static void TableMapReduceUtil.initTableMapperJob(byte[] table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars)
          Use this before submitting a TableMap job.
static void TableMapReduceUtil.initTableMapperJob(byte[] table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass)
          Use this before submitting a TableMap job.
static void TableMapReduceUtil.initTableMapperJob(String table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job)
          Use this before submitting a TableMap job.
static void TableMapReduceUtil.initTableMapperJob(String table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars)
          Use this before submitting a TableMap job.
static void TableMapReduceUtil.initTableMapperJob(String table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, boolean initCredentials, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass)
          Use this before submitting a TableMap job.
static void TableMapReduceUtil.initTableMapperJob(String table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass)
          Use this before submitting a TableMap job.
static void TableMapReduceUtil.initTableSnapshotMapperJob(String snapshotName, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir)
          Sets up the job for reading from a table snapshot.
 void TableRecordReader.setScan(Scan scan)
          Sets the scan defining the actual details like columns etc.
 void TableInputFormatBase.setScan(Scan scan)
          Sets the scan defining the actual details like columns etc.
 void TableRecordReaderImpl.setScan(Scan scan)
          Sets the scan defining the actual details like columns etc.
 

Method parameters in org.apache.hadoop.hbase.mapreduce with type arguments of type Scan
static void TableMapReduceUtil.initTableMapperJob(List<Scan> scans, Class<? extends TableMapper> mapper, Class<? extends org.apache.hadoop.io.WritableComparable> outputKeyClass, Class<? extends org.apache.hadoop.io.Writable> outputValueClass, org.apache.hadoop.mapreduce.Job job)
          Use this before submitting a Multi TableMap job.
static void TableMapReduceUtil.initTableMapperJob(List<Scan> scans, Class<? extends TableMapper> mapper, Class<? extends org.apache.hadoop.io.WritableComparable> outputKeyClass, Class<? extends org.apache.hadoop.io.Writable> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars)
          Use this before submitting a Multi TableMap job.
static void TableMapReduceUtil.initTableMapperJob(List<Scan> scans, Class<? extends TableMapper> mapper, Class<? extends org.apache.hadoop.io.WritableComparable> outputKeyClass, Class<? extends org.apache.hadoop.io.Writable> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, boolean initCredentials)
          Use this before submitting a Multi TableMap job.
protected  void MultiTableInputFormatBase.setScans(List<Scan> scans)
          Allows subclasses to set the list of Scan objects.
 

Constructors in org.apache.hadoop.hbase.mapreduce with parameters of type Scan
TableSplit(byte[] tableName, Scan scan, byte[] startRow, byte[] endRow, String location)
          Deprecated. Since 0.96.0; use TableSplit.TableSplit(TableName, byte[], byte[], String)
TableSplit(TableName tableName, Scan scan, byte[] startRow, byte[] endRow, String location)
          Creates a new instance while assigning all variables.
TableSplit(TableName tableName, Scan scan, byte[] startRow, byte[] endRow, String location, long length)
          Creates a new instance while assigning all variables.
 

Uses of Scan in org.apache.hadoop.hbase.rest.client
 

Methods in org.apache.hadoop.hbase.rest.client with parameters of type Scan
 ResultScanner RemoteHTable.getScanner(Scan scan)
           
 



Copyright © 2015 The Apache Software Foundation. All rights reserved.