java.lang.Object
org.apache.hadoop.mapreduce.InputFormat<K,V>
org.apache.hadoop.mapreduce.lib.input.FileInputFormat<K,V>
org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<ShapeKey,PrimitiveShape>
org.apache.sedona.core.formatMapper.shapefileParser.shapes.ShapeInputFormat

public class ShapeInputFormat extends org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<ShapeKey,PrimitiveShape>
  • Nested Class Summary

    Nested classes/interfaces inherited from class org.apache.hadoop.mapreduce.lib.input.FileInputFormat

    org.apache.hadoop.mapreduce.lib.input.FileInputFormat.Counter
  • Field Summary

    Fields inherited from class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat

    SPLIT_MINSIZE_PERNODE, SPLIT_MINSIZE_PERRACK

    Fields inherited from class org.apache.hadoop.mapreduce.lib.input.FileInputFormat

    DEFAULT_LIST_STATUS_NUM_THREADS, INPUT_DIR, INPUT_DIR_NONRECURSIVE_IGNORE_SUBDIRS, INPUT_DIR_RECURSIVE, LIST_STATUS_NUM_THREADS, NUM_INPUT_FILES, PATHFILTER_CLASS, SPLIT_MAXSIZE, SPLIT_MINSIZE
  • Constructor Summary

    Constructors
    Constructor
    Description
     
  • Method Summary

    Modifier and Type
    Method
    Description
    org.apache.hadoop.mapreduce.RecordReader<ShapeKey,PrimitiveShape>
    createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context)
     
    List<org.apache.hadoop.mapreduce.InputSplit>
    getSplits(org.apache.hadoop.mapreduce.JobContext job)
     
    protected boolean
    isSplitable(org.apache.hadoop.mapreduce.JobContext context, org.apache.hadoop.fs.Path file)
    enforce isSplitable() to return false so that every getSplits() only return one InputSplit

    Methods inherited from class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat

    createPool, createPool, getFileBlockLocations, setMaxSplitSize, setMinSplitSizeNode, setMinSplitSizeRack

    Methods inherited from class org.apache.hadoop.mapreduce.lib.input.FileInputFormat

    addInputPath, addInputPathRecursively, addInputPaths, computeSplitSize, getBlockIndex, getFormatMinSplitSize, getInputDirRecursive, getInputPathFilter, getInputPaths, getMaxSplitSize, getMinSplitSize, listStatus, makeSplit, makeSplit, setInputDirRecursive, setInputPathFilter, setInputPaths, setInputPaths, setMaxInputSplitSize, setMinInputSplitSize, shrinkStatus

    Methods inherited from class java.lang.Object

    clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait
  • Constructor Details

    • ShapeInputFormat

      public ShapeInputFormat()
  • Method Details

    • createRecordReader

      public org.apache.hadoop.mapreduce.RecordReader<ShapeKey,PrimitiveShape> createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) throws IOException
      Specified by:
      createRecordReader in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<ShapeKey,PrimitiveShape>
      Throws:
      IOException
    • isSplitable

      protected boolean isSplitable(org.apache.hadoop.mapreduce.JobContext context, org.apache.hadoop.fs.Path file)
      enforce isSplitable() to return false so that every getSplits() only return one InputSplit
      Overrides:
      isSplitable in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<ShapeKey,PrimitiveShape>
      Parameters:
      context -
      file -
      Returns:
    • getSplits

      public List<org.apache.hadoop.mapreduce.InputSplit> getSplits(org.apache.hadoop.mapreduce.JobContext job) throws IOException
      Overrides:
      getSplits in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<ShapeKey,PrimitiveShape>
      Throws:
      IOException