public class CarbonInputFormat<T> extends org.apache.hadoop.mapreduce.lib.input.FileInputFormat<Void,T>
| Modifier and Type | Field and Description |
|---|---|
static String |
INPUT_SEGMENT_NUMBERS |
| Constructor and Description |
|---|
CarbonInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
protected long |
computeSplitSize(long blockSize,
long minSize,
long maxSize) |
org.apache.hadoop.mapreduce.RecordReader<Void,T> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit inputSplit,
org.apache.hadoop.mapreduce.TaskAttemptContext taskAttemptContext) |
protected int |
getBlockIndex(org.apache.hadoop.fs.BlockLocation[] blkLocations,
long offset) |
static CarbonTable |
getCarbonTable(org.apache.hadoop.conf.Configuration configuration) |
org.apache.hadoop.fs.PathFilter |
getDataFileFilter(org.apache.hadoop.mapreduce.JobContext job) |
FilterResolverIntf |
getResolvedFilter(org.apache.hadoop.conf.Configuration configuration,
Expression filterExpression)
Configurations FileInputFormat.INPUT_DIR, CarbonInputFormat.INPUT_SEGMENT_NUMBERS
are used to get table path to read.
|
long |
getRowCount(org.apache.hadoop.mapreduce.JobContext job)
get total number of rows.
|
List<org.apache.hadoop.mapreduce.InputSplit> |
getSplits(org.apache.hadoop.mapreduce.JobContext job)
Configurations FileInputFormat.INPUT_DIR
are used to get table path to read.
|
static CarbonTablePath |
getTablePath(org.apache.hadoop.conf.Configuration configuration) |
static CarbonTableIdentifier |
getTableToAccess(org.apache.hadoop.conf.Configuration configuration)
Get CarbonTableIdentifier from job configuration
|
protected boolean |
isSplitable(org.apache.hadoop.mapreduce.JobContext context,
org.apache.hadoop.fs.Path filename) |
protected List<org.apache.hadoop.fs.FileStatus> |
listStatus(org.apache.hadoop.mapreduce.JobContext job) |
static void |
setCarbonReadSupport(Class<? extends CarbonReadSupport> readSupportClass,
org.apache.hadoop.conf.Configuration configuration) |
static void |
setCarbonTable(org.apache.hadoop.conf.Configuration configuration,
CarbonTable carbonTable)
It is optional, if user does not set then it reads from store
|
static void |
setColumnProjection(CarbonProjection projection,
org.apache.hadoop.conf.Configuration configuration) |
static void |
setFilterPredicates(org.apache.hadoop.conf.Configuration configuration,
Expression filterExpression)
It sets unresolved filter expression.
|
static void |
setFilterPredicates(org.apache.hadoop.conf.Configuration configuration,
FilterResolverIntf filterExpression)
It sets the resolved filter expression
|
static void |
setTableToAccess(org.apache.hadoop.conf.Configuration configuration,
CarbonTableIdentifier tableIdentifier) |
addInputPath, addInputPathRecursively, addInputPaths, getFormatMinSplitSize, getInputDirRecursive, getInputPathFilter, getInputPaths, getMaxSplitSize, getMinSplitSize, makeSplit, setInputDirRecursive, setInputPathFilter, setInputPaths, setInputPaths, setMaxInputSplitSize, setMinInputSplitSizepublic static final String INPUT_SEGMENT_NUMBERS
public static void setTableToAccess(org.apache.hadoop.conf.Configuration configuration,
CarbonTableIdentifier tableIdentifier)
public static CarbonTableIdentifier getTableToAccess(org.apache.hadoop.conf.Configuration configuration)
public static void setCarbonTable(org.apache.hadoop.conf.Configuration configuration,
CarbonTable carbonTable)
throws IOException
configuration - carbonTable - IOExceptionpublic static CarbonTable getCarbonTable(org.apache.hadoop.conf.Configuration configuration) throws IOException
IOExceptionpublic static void setFilterPredicates(org.apache.hadoop.conf.Configuration configuration,
Expression filterExpression)
configuration - filterExpression - public static void setFilterPredicates(org.apache.hadoop.conf.Configuration configuration,
FilterResolverIntf filterExpression)
configuration - filterExpression - public static void setColumnProjection(CarbonProjection projection, org.apache.hadoop.conf.Configuration configuration)
public static void setCarbonReadSupport(Class<? extends CarbonReadSupport> readSupportClass, org.apache.hadoop.conf.Configuration configuration)
public static CarbonTablePath getTablePath(org.apache.hadoop.conf.Configuration configuration) throws IOException
IOExceptionpublic List<org.apache.hadoop.mapreduce.InputSplit> getSplits(org.apache.hadoop.mapreduce.JobContext job) throws IOException
getSplits in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat<Void,T>job - IOExceptionpublic long getRowCount(org.apache.hadoop.mapreduce.JobContext job)
throws IOException,
IndexBuilderException
IOExceptionIndexBuilderExceptionpublic FilterResolverIntf getResolvedFilter(org.apache.hadoop.conf.Configuration configuration, Expression filterExpression) throws IOException, IndexBuilderException, QueryExecutionException
IOExceptionIndexBuilderExceptionQueryExecutionExceptionpublic org.apache.hadoop.mapreduce.RecordReader<Void,T> createRecordReader(org.apache.hadoop.mapreduce.InputSplit inputSplit, org.apache.hadoop.mapreduce.TaskAttemptContext taskAttemptContext) throws IOException, InterruptedException
createRecordReader in class org.apache.hadoop.mapreduce.InputFormat<Void,T>IOExceptionInterruptedExceptionprotected long computeSplitSize(long blockSize,
long minSize,
long maxSize)
protected int getBlockIndex(org.apache.hadoop.fs.BlockLocation[] blkLocations,
long offset)
protected List<org.apache.hadoop.fs.FileStatus> listStatus(org.apache.hadoop.mapreduce.JobContext job) throws IOException
listStatus in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat<Void,T>IOExceptionprotected boolean isSplitable(org.apache.hadoop.mapreduce.JobContext context,
org.apache.hadoop.fs.Path filename)
public org.apache.hadoop.fs.PathFilter getDataFileFilter(org.apache.hadoop.mapreduce.JobContext job)
job - Copyright © 2016 The Apache Software Foundation. All rights reserved.