@Experimental(value=SOURCE_SINK) public class HBaseIO extends java.lang.Object
For more information, see the online documentation at HBase.
The HBase source returns a set of rows from a single table, returning a PCollection<Result>
.
To configure a HBase source, you must supply a table id and a Configuration
to
identify the HBase instance. By default, HBaseIO.Read
will read all rows in the table.
The row range to be read can optionally be restricted using with a Scan
object or using
the HBaseIO.Read.withKeyRange(org.apache.beam.sdk.io.range.ByteKeyRange)
, and a Filter
using HBaseIO.Read.withFilter(org.apache.hadoop.hbase.filter.Filter)
, for example:
// Scan the entire table.
p.apply("read",
HBaseIO.read()
.withConfiguration(configuration)
.withTableId("table"));
// Filter data using a HBaseIO Scan
Scan scan = ...
p.apply("read",
HBaseIO.read()
.withConfiguration(configuration)
.withTableId("table"))
.withScan(scan));
// Scan a prefix of the table.
ByteKeyRange keyRange = ...;
p.apply("read",
HBaseIO.read()
.withConfiguration(configuration)
.withTableId("table")
.withKeyRange(keyRange));
// Scan a subset of rows that match the specified row filter.
p.apply("filtered read",
HBaseIO.read()
.withConfiguration(configuration)
.withTableId("table")
.withFilter(filter));
readAll()
allows to execute multiple Scan
s to multiple Table
s.
These queries are encapsulated via an initial PCollection
of HBaseIO.Read
s and can be
used to create advanced compositional patterns like reading from a Source and then based on the
data create new HBase scans.
Note: HBaseIO.ReadAll
only works with runners that support
Splittable DoFn.
PCollection<Read> queries = ...;
queries.apply("readAll", HBaseIO.readAll().withConfiguration(configuration));
The HBase sink executes a set of row mutations on a single table. It takes as input a PCollection<Mutation>
, where each Mutation
represents an idempotent
transformation on a row.
To configure a HBase sink, you must supply a table id and a Configuration
to identify
the HBase instance, for example:
Configuration configuration = ...;
PCollection<Mutation> data = ...;
data.apply("write",
HBaseIO.write()
.withConfiguration(configuration)
.withTableId("table"));
The design of the API for HBaseIO is currently related to the BigtableIO one, it can evolve or be different in some aspects, but the idea is that users can easily migrate from one to the other .
Modifier and Type | Class and Description |
---|---|
static class |
HBaseIO.Read
A
PTransform that reads from HBase. |
static class |
HBaseIO.ReadAll
Implementation of
readAll() . |
static class |
HBaseIO.Write
A
PTransform that writes to HBase. |
Modifier and Type | Method and Description |
---|---|
static HBaseIO.Read |
read()
Creates an uninitialized
HBaseIO.Read . |
static HBaseIO.ReadAll |
readAll()
A
PTransform that works like read() , but executes read operations coming from a
PCollection of HBaseIO.Read . |
static HBaseIO.Write |
write()
Creates an uninitialized
HBaseIO.Write . |
public static HBaseIO.Read read()
HBaseIO.Read
. Before use, the Read
must be initialized
with a HBaseIO.Read.withConfiguration(Configuration)
that specifies the HBase instance,
and a tableId
that specifies which table to read. A Filter
may also optionally be specified using HBaseIO.Read.withFilter(org.apache.hadoop.hbase.filter.Filter)
.@Experimental(value=SPLITTABLE_DO_FN) public static HBaseIO.ReadAll readAll()
PTransform
that works like read()
, but executes read operations coming from a
PCollection
of HBaseIO.Read
.public static HBaseIO.Write write()
HBaseIO.Write
. Before use, the Write
must be
initialized with a HBaseIO.Write.withConfiguration(Configuration)
that specifies the
destination HBase instance, and a tableId
that specifies
which table to write.