|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use Scan | |
---|---|
org.apache.hadoop.hbase.client | Provides HBase Client |
org.apache.hadoop.hbase.mapred | Provides HBase MapReduce Input/OutputFormats, a table indexing MapReduce job, and utility |
org.apache.hadoop.hbase.mapreduce | Provides HBase MapReduce Input/OutputFormats, a table indexing MapReduce job, and utility |
org.apache.hadoop.hbase.rest.client |
Uses of Scan in org.apache.hadoop.hbase.client |
---|
Fields in org.apache.hadoop.hbase.client declared as Scan | |
---|---|
protected Scan |
ClientScanner.scan
|
Methods in org.apache.hadoop.hbase.client that return Scan | |
---|---|
Scan |
Scan.addColumn(byte[] family,
byte[] qualifier)
Get the column from the specified family with the specified qualifier. |
Scan |
Scan.addFamily(byte[] family)
Get all columns from the specified family. |
protected Scan |
ClientScanner.getScan()
|
Scan |
Scan.setFamilyMap(Map<byte[],NavigableSet<byte[]>> familyMap)
Setting the familyMap |
Scan |
Scan.setFilter(Filter filter)
|
Scan |
Scan.setMaxVersions()
Get all available versions. |
Scan |
Scan.setMaxVersions(int maxVersions)
Get up to the specified number of versions of each column. |
Scan |
Scan.setReversed(boolean reversed)
Set whether this scan is a reversed one |
Scan |
Scan.setStartRow(byte[] startRow)
Set the start row of the scan. |
Scan |
Scan.setStopRow(byte[] stopRow)
Set the stop row. |
Scan |
Scan.setTimeRange(long minStamp,
long maxStamp)
Get versions of columns only within the specified timestamp range, [minStamp, maxStamp). |
Scan |
Scan.setTimeStamp(long timestamp)
Get versions of columns with the specified timestamp. |
Methods in org.apache.hadoop.hbase.client with parameters of type Scan | |
---|---|
ResultScanner |
HTableInterface.getScanner(Scan scan)
Returns a scanner on the current table as specified by the Scan
object. |
ResultScanner |
HTable.getScanner(Scan scan)
Returns a scanner on the current table as specified by the Scan
object. |
Constructors in org.apache.hadoop.hbase.client with parameters of type Scan | |
---|---|
ClientScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
byte[] tableName)
Deprecated. Use ClientScanner.ClientScanner(Configuration, Scan, TableName) |
|
ClientScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
byte[] tableName,
HConnection connection)
Deprecated. Use ClientScanner.ClientScanner(Configuration, Scan, TableName, HConnection) |
|
ClientScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName)
Deprecated. |
|
ClientScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName,
HConnection connection)
Create a new ClientScanner for the specified table Note that the passed Scan 's start row maybe changed changed. |
|
ClientScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName,
HConnection connection,
org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcFactory)
Deprecated. Use ClientScanner.ClientScanner(Configuration, Scan, TableName, HConnection,
RpcRetryingCallerFactory, RpcControllerFactory)
instead |
|
ClientScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName,
HConnection connection,
org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcFactory,
org.apache.hadoop.hbase.ipc.RpcControllerFactory controllerFactory)
Create a new ClientScanner for the specified table Note that the passed Scan 's start
row maybe changed changed. |
|
ClientSmallReversedScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName,
HConnection connection)
Create a new ReversibleClientScanner for the specified table Note that the passed Scan 's start row maybe changed. |
|
ClientSmallScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName)
Create a new ClientSmallScanner for the specified table. |
|
ClientSmallScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName,
HConnection connection)
Create a new ClientSmallScanner for the specified table. |
|
ClientSmallScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName,
HConnection connection,
org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcFactory,
org.apache.hadoop.hbase.ipc.RpcControllerFactory controllerFactory)
Create a new ShortClientScanner for the specified table Note that the passed Scan 's start row maybe changed changed. |
|
ReversedClientScanner(org.apache.hadoop.conf.Configuration conf,
Scan scan,
TableName tableName,
HConnection connection)
Create a new ReversibleClientScanner for the specified table Note that the passed Scan 's start row maybe changed. |
|
ReversedScannerCallable(HConnection connection,
TableName tableName,
Scan scan,
org.apache.hadoop.hbase.client.metrics.ScanMetrics scanMetrics,
byte[] locateStartRow)
Deprecated. |
|
ReversedScannerCallable(HConnection connection,
TableName tableName,
Scan scan,
org.apache.hadoop.hbase.client.metrics.ScanMetrics scanMetrics,
byte[] locateStartRow,
org.apache.hadoop.hbase.ipc.PayloadCarryingRpcController rpcFactory)
|
|
Scan(Scan scan)
Creates a new instance of this class while copying all values. |
|
TableSnapshotScanner(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.Path rootDir,
org.apache.hadoop.fs.Path restoreDir,
String snapshotName,
Scan scan)
Creates a TableSnapshotScanner. |
|
TableSnapshotScanner(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.Path restoreDir,
String snapshotName,
Scan scan)
Creates a TableSnapshotScanner. |
Uses of Scan in org.apache.hadoop.hbase.mapred |
---|
Method parameters in org.apache.hadoop.hbase.mapred with type arguments of type Scan | |
---|---|
static void |
TableMapReduceUtil.initMultiTableSnapshotMapperJob(Map<String,Collection<Scan>> snapshotScans,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars,
org.apache.hadoop.fs.Path tmpRestoreDir)
Deprecated. Sets up the job for reading from one or more multiple table snapshots, with one or more scans per snapshot. |
static void |
MultiTableSnapshotInputFormat.setInput(org.apache.hadoop.conf.Configuration conf,
Map<String,Collection<Scan>> snapshotScans,
org.apache.hadoop.fs.Path restoreDir)
Configure conf to read from snapshotScans, with snapshots restored to a subdirectory of restoreDir. |
Uses of Scan in org.apache.hadoop.hbase.mapreduce |
---|
Methods in org.apache.hadoop.hbase.mapreduce that return Scan | |
---|---|
Scan |
TableSplit.getScan()
Returns a Scan object from the stored string representation. |
Scan |
TableInputFormatBase.getScan()
Gets the scan defining the actual details like columns etc. |
Methods in org.apache.hadoop.hbase.mapreduce that return types with arguments of type Scan | |
---|---|
protected List<Scan> |
MultiTableInputFormatBase.getScans()
Allows subclasses to get the list of Scan objects. |
Methods in org.apache.hadoop.hbase.mapreduce with parameters of type Scan | |
---|---|
static void |
TableInputFormat.addColumns(Scan scan,
byte[][] columns)
Adds an array of columns specified using old format, family:qualifier. |
static void |
IdentityTableMapper.initJob(String table,
Scan scan,
Class<? extends TableMapper> mapper,
org.apache.hadoop.mapreduce.Job job)
Use this before submitting a TableMap job. |
static void |
GroupingTableMapper.initJob(String table,
Scan scan,
String groupColumns,
Class<? extends TableMapper> mapper,
org.apache.hadoop.mapreduce.Job job)
Use this before submitting a TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(byte[] table,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job)
Use this before submitting a TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(byte[] table,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars)
Use this before submitting a TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(byte[] table,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars,
Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass)
Use this before submitting a TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(String table,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job)
Use this before submitting a TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(String table,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars)
Use this before submitting a TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(String table,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars,
boolean initCredentials,
Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass)
Use this before submitting a TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(String table,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars,
Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass)
Use this before submitting a TableMap job. |
static void |
TableMapReduceUtil.initTableSnapshotMapperJob(String snapshotName,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars,
org.apache.hadoop.fs.Path tmpRestoreDir)
Sets up the job for reading from a table snapshot. |
void |
TableRecordReaderImpl.setScan(Scan scan)
Sets the scan defining the actual details like columns etc. |
void |
TableInputFormatBase.setScan(Scan scan)
Sets the scan defining the actual details like columns etc. |
void |
TableRecordReader.setScan(Scan scan)
Sets the scan defining the actual details like columns etc. |
Method parameters in org.apache.hadoop.hbase.mapreduce with type arguments of type Scan | |
---|---|
static void |
TableMapReduceUtil.initMultiTableSnapshotMapperJob(Map<String,Collection<Scan>> snapshotScans,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars,
org.apache.hadoop.fs.Path tmpRestoreDir)
Sets up the job for reading from one or more table snapshots, with one or more scans per snapshot. |
static void |
TableMapReduceUtil.initTableMapperJob(List<Scan> scans,
Class<? extends TableMapper> mapper,
Class<? extends org.apache.hadoop.io.WritableComparable> outputKeyClass,
Class<? extends org.apache.hadoop.io.Writable> outputValueClass,
org.apache.hadoop.mapreduce.Job job)
Use this before submitting a Multi TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(List<Scan> scans,
Class<? extends TableMapper> mapper,
Class<? extends org.apache.hadoop.io.WritableComparable> outputKeyClass,
Class<? extends org.apache.hadoop.io.Writable> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars)
Use this before submitting a Multi TableMap job. |
static void |
TableMapReduceUtil.initTableMapperJob(List<Scan> scans,
Class<? extends TableMapper> mapper,
Class<? extends org.apache.hadoop.io.WritableComparable> outputKeyClass,
Class<? extends org.apache.hadoop.io.Writable> outputValueClass,
org.apache.hadoop.mapreduce.Job job,
boolean addDependencyJars,
boolean initCredentials)
Use this before submitting a Multi TableMap job. |
static void |
MultiTableSnapshotInputFormat.setInput(org.apache.hadoop.conf.Configuration configuration,
Map<String,Collection<Scan>> snapshotScans,
org.apache.hadoop.fs.Path tmpRestoreDir)
|
protected void |
MultiTableInputFormatBase.setScans(List<Scan> scans)
Allows subclasses to set the list of Scan objects. |
Constructors in org.apache.hadoop.hbase.mapreduce with parameters of type Scan | |
---|---|
TableSplit(byte[] tableName,
Scan scan,
byte[] startRow,
byte[] endRow,
String location)
Deprecated. Since 0.96.0; use TableSplit.TableSplit(TableName, byte[], byte[], String) |
|
TableSplit(TableName tableName,
Scan scan,
byte[] startRow,
byte[] endRow,
String location)
Creates a new instance while assigning all variables. |
|
TableSplit(TableName tableName,
Scan scan,
byte[] startRow,
byte[] endRow,
String location,
long length)
Creates a new instance while assigning all variables. |
Uses of Scan in org.apache.hadoop.hbase.rest.client |
---|
Methods in org.apache.hadoop.hbase.rest.client with parameters of type Scan | |
---|---|
ResultScanner |
RemoteHTable.getScanner(Scan scan)
|
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |