| 程序包 | 说明 |
|---|---|
| org.apache.hadoop.conf |
Configuration of system parameters.
|
| org.apache.hadoop.contrib.index.example | |
| org.apache.hadoop.contrib.index.mapred | |
| org.apache.hadoop.contrib.utils.join | |
| org.apache.hadoop.examples |
Hadoop example code.
|
| org.apache.hadoop.fs |
An abstract file system API.
|
| org.apache.hadoop.fs.permission | |
| org.apache.hadoop.hdfs.protocol | |
| org.apache.hadoop.hdfs.security.token.block | |
| org.apache.hadoop.hdfs.server.common | |
| org.apache.hadoop.hdfs.server.namenode | |
| org.apache.hadoop.hdfs.server.protocol | |
| org.apache.hadoop.io |
Generic i/o code for use when reading and writing data to the network,
to databases, and to files.
|
| org.apache.hadoop.io.serializer |
This package provides a mechanism for using different serialization frameworks
in Hadoop.
|
| org.apache.hadoop.ipc |
Tools to help define network clients and servers.
|
| org.apache.hadoop.mapred |
A software framework for easily writing applications which process vast
amounts of data (multi-terabyte data-sets) parallelly on large clusters
(thousands of nodes) built of commodity hardware in a reliable, fault-tolerant
manner.
|
| org.apache.hadoop.mapred.join |
Given a set of sorted datasets keyed with the same class and yielding equal
partitions, it is possible to effect a join of those datasets prior to the map.
|
| org.apache.hadoop.mapred.lib |
Library of generally useful mappers, reducers, and partitioners.
|
| org.apache.hadoop.mapred.lib.aggregate |
Classes for performing various counting and aggregations.
|
| org.apache.hadoop.mapred.lib.db |
org.apache.hadoop.mapred.lib.db Package
This package contains a library to read records from a database as an
input to a mapreduce job, and write the output records to the database.
|
| org.apache.hadoop.mapreduce | |
| org.apache.hadoop.mapreduce.lib.db | |
| org.apache.hadoop.mapreduce.lib.output | |
| org.apache.hadoop.mapreduce.security.token | |
| org.apache.hadoop.mapreduce.security.token.delegation | |
| org.apache.hadoop.mapreduce.split | |
| org.apache.hadoop.record |
Hadoop record I/O contains classes and a record description language
translator for simplifying serialization and deserialization of records in a
language-neutral manner.
|
| org.apache.hadoop.record.meta | |
| org.apache.hadoop.security | |
| org.apache.hadoop.security.authorize | |
| org.apache.hadoop.security.token | |
| org.apache.hadoop.security.token.delegation | |
| org.apache.hadoop.streaming.io | |
| org.apache.hadoop.tools.rumen |
Rumen is a data extraction and analysis tool built for
Apache Hadoop.
|
| org.apache.hadoop.typedbytes |
Typed bytes are sequences of bytes in which the first byte is a type code.
|
| org.apache.hadoop.util |
Common utilities.
|
| org.apache.hadoop.util.bloom |
| 限定符和类型 | 类和说明 |
|---|---|
class |
Configuration
Provides access to configuration parameters.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
LineDocTextAndOp
This class represents an operation.
|
| 限定符和类型 | 接口和说明 |
|---|---|
interface |
ILocalAnalysis<K extends WritableComparable,V extends Writable>
Application specific local analysis.
|
class |
IndexUpdateMapper<K extends WritableComparable,V extends Writable>
This class applies local analysis on a key-value pair and then convert the
result docid-operation pair to a shard-and-intermediate form pair.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
DocumentAndOp
This class represents an indexing operation.
|
class |
DocumentID
The class represents a document id, which is of type text.
|
class |
IntermediateForm
An intermediate form for one or more parsed Lucene documents and/or
delete terms.
|
class |
Shard
This class represents the metadata of a shard.
|
| 限定符和类型 | 方法和说明 |
|---|---|
static Class<? extends Writable> |
IndexUpdateMapper.getMapOutputValueClass()
Get the map output value class.
|
static Class<? extends Writable> |
IndexUpdateReducer.getOutputValueClass()
Get the reduce output value class.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
TaggedMapOutput
This abstract class serves as the base class for the values that
flow from the mappers to the reducers in a data join job.
|
| 限定符和类型 | 方法和说明 |
|---|---|
abstract Writable |
TaggedMapOutput.getData() |
| 限定符和类型 | 类和说明 |
|---|---|
static class |
MultiFileWordCount.WordOffset
This record keeps <filename,offset> pairs.
|
static class |
SecondarySort.IntPair
Define a pair of integers that are writable.
|
static class |
SleepJob.EmptySplit |
| 限定符和类型 | 方法和说明 |
|---|---|
void |
PiEstimator.PiReducer.reduce(BooleanWritable isInside,
Iterator<LongWritable> values,
OutputCollector<WritableComparable<?>,Writable> output,
Reporter reporter)
Accumulate number of points inside/outside results from the mappers.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
BlockLocation |
class |
ContentSummary
Store the summary of a content (a directory or a file).
|
class |
FileChecksum
An abstract class representing file checksums for files.
|
class |
FileStatus
Interface that represents the client side information for a file.
|
class |
MD5MD5CRC32FileChecksum
MD5 of MD5 of CRC32.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
FsPermission
A class for file/directory permissions.
|
class |
PermissionStatus
Store permission related information.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
Block
A Block is a Hadoop FS primitive, identified by a
long.
|
class |
BlockLocalPathInfo
A block and the full path information to the block data file and
the metadata file stored on the local file system.
|
class |
DatanodeID
DatanodeID is composed of the data node
name (hostname:portNumber) and the data storage ID,
which it currently represents.
|
class |
DatanodeInfo
DatanodeInfo represents the status of a DataNode.
|
static class |
DataTransferProtocol.PipelineAck
reply
|
class |
DirectoryListing
This class defines a partial listing of a directory to support
iterative directory listing.
|
class |
HdfsFileStatus
Interface that represents the over the wire information for a file.
|
class |
LocatedBlock
A LocatedBlock is a pair of Block, DatanodeInfo[]
objects.
|
class |
LocatedBlocks
Collection of blocks with their locations and the file length.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
BlockKey
Key used for generating and verifying block tokens
|
class |
BlockTokenIdentifier |
class |
ExportedBlockKeys
Object for passing block keys
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
GenerationStamp
A GenerationStamp is a Hadoop FS primitive, identified by a long.
|
class |
UpgradeStatusReport
Base upgrade upgradeStatus class.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
CheckpointSignature
A unique signature intended to identify checkpoint transactions.
|
class |
DatanodeDescriptor
DatanodeDescriptor tracks stats on a given DataNode,
such as available storage capacity, last update time, etc.,
and maintains a set of blocks stored on the datanode.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
BalancerBandwidthCommand
Balancer bandwidth command instructs each datanode to change its value for
the max amount of network bandwidth it may use during the block balancing
operation.
|
class |
BlockCommand
A BlockCommand is an instruction to a datanode
regarding some blocks under its control.
|
class |
BlockMetaDataInfo
Meta data information for a block
|
class |
BlockRecoveryInfo |
class |
BlocksWithLocations
A class to implement an array of BlockLocations
It provide efficient customized serialization/deserialization methods
in stead of using the default array (de)serialization provided by RPC
|
static class |
BlocksWithLocations.BlockWithLocations
A class to keep track of a block and its locations
|
class |
DatanodeCommand |
class |
DatanodeRegistration
DatanodeRegistration class conatins all information the Namenode needs
to identify and verify a Datanode when it contacts the Namenode.
|
class |
KeyUpdateCommand |
class |
NamespaceInfo
NamespaceInfo is returned by the name-node in reply
to a data-node handshake.
|
class |
UpgradeCommand
This as a generic distributed upgrade command.
|
| 限定符和类型 | 接口和说明 |
|---|---|
interface |
WritableComparable<T>
A
Writable which is also Comparable. |
| 限定符和类型 | 类和说明 |
|---|---|
class |
AbstractMapWritable
Abstract base class for MapWritable and SortedMapWritable
Unlike org.apache.nutch.crawl.MapWritable, this class allows creation of
MapWritable<Writable, MapWritable> so the CLASS_TO_ID and ID_TO_CLASS
maps travel with the class instead of being static.
|
class |
ArrayWritable
A Writable for arrays containing instances of a class.
|
class |
BooleanWritable
A WritableComparable for booleans.
|
class |
BytesWritable
A byte sequence that is usable as a key or value.
|
class |
ByteWritable
A WritableComparable for a single byte.
|
class |
CompressedWritable
A base-class for Writables which store themselves compressed and lazily
inflate on field access.
|
class |
DoubleWritable
Writable for Double values.
|
class |
FloatWritable
A WritableComparable for floats.
|
class |
GenericWritable
A wrapper for Writable instances.
|
class |
IntWritable
A WritableComparable for ints.
|
class |
LongWritable
A WritableComparable for longs.
|
class |
MapWritable
A Writable Map.
|
class |
MD5Hash
A Writable for MD5 hash values.
|
class |
NullWritable
Singleton Writable with no data.
|
class |
ObjectWritable
A polymorphic Writable that writes an instance with it's class name.
|
static class |
SequenceFile.Metadata
The class encapsulating with the metadata of a file.
|
class |
SortedMapWritable
A Writable SortedMap.
|
class |
Text
This class stores text using standard UTF8 encoding.
|
class |
TwoDArrayWritable
A Writable for 2D arrays containing a matrix of instances of a class.
|
class |
UTF8
已过时。
replaced by Text
|
class |
VersionedWritable
A base class for Writables that provides version checking.
|
class |
VIntWritable
A WritableComparable for integer values stored in variable-length format.
|
class |
VLongWritable
A WritableComparable for longs in a variable-length format.
|
| 限定符和类型 | 方法和说明 |
|---|---|
static <T extends Writable> |
WritableUtils.clone(T orig,
Configuration conf)
Make a copy of a writable object using serialization to a buffer.
|
| 限定符和类型 | 方法和说明 |
|---|---|
Writable[][] |
TwoDArrayWritable.get() |
Writable |
GenericWritable.get()
Return the wrapped instance.
|
Writable[] |
ArrayWritable.get() |
Writable |
ArrayFile.Reader.get(long n,
Writable value)
Return the
nth value in the file. |
Writable |
SortedMapWritable.get(Object key) |
Writable |
MapWritable.get(Object key) |
Writable |
MapFile.Reader.get(WritableComparable key,
Writable val)
Return the value for the named key, or null if none exists.
|
Writable |
BloomMapFile.Reader.get(WritableComparable key,
Writable val)
Fast version of the
MapFile.Reader.get(WritableComparable, Writable) method. |
Writable |
WritableFactory.newInstance()
Return a new instance.
|
static Writable |
WritableFactories.newInstance(Class<? extends Writable> c)
Create a new instance of a class with a defined factory.
|
static Writable |
WritableFactories.newInstance(Class<? extends Writable> c,
Configuration conf)
Create a new instance of a class with a defined factory.
|
Writable |
ArrayFile.Reader.next(Writable value)
Read and return the next value in the file.
|
Writable |
SortedMapWritable.put(WritableComparable key,
Writable value) |
Writable |
MapWritable.put(Writable key,
Writable value) |
Writable |
SortedMapWritable.remove(Object key) |
Writable |
MapWritable.remove(Object key) |
| 限定符和类型 | 方法和说明 |
|---|---|
Set<Map.Entry<WritableComparable,Writable>> |
SortedMapWritable.entrySet() |
Set<Map.Entry<Writable,Writable>> |
MapWritable.entrySet() |
Set<Map.Entry<Writable,Writable>> |
MapWritable.entrySet() |
SortedMap<WritableComparable,Writable> |
SortedMapWritable.headMap(WritableComparable toKey) |
Set<Writable> |
MapWritable.keySet() |
SortedMap<WritableComparable,Writable> |
SortedMapWritable.subMap(WritableComparable fromKey,
WritableComparable toKey) |
SortedMap<WritableComparable,Writable> |
SortedMapWritable.tailMap(WritableComparable fromKey) |
Collection<Writable> |
SortedMapWritable.values() |
Collection<Writable> |
MapWritable.values() |
| 限定符和类型 | 方法和说明 |
|---|---|
void |
ArrayFile.Writer.append(Writable value)
Append a value to the file.
|
void |
MapFile.Writer.append(WritableComparable key,
Writable val)
Append a key/value pair to the map.
|
void |
BloomMapFile.Writer.append(WritableComparable key,
Writable val) |
void |
SequenceFile.Writer.append(Writable key,
Writable val)
Append a key/value pair.
|
static void |
WritableUtils.cloneInto(Writable dst,
Writable src)
已过时。
use ReflectionUtils.cloneInto instead.
|
protected void |
AbstractMapWritable.copy(Writable other)
Used by child copy constructors.
|
Writable |
ArrayFile.Reader.get(long n,
Writable value)
Return the
nth value in the file. |
Writable |
MapFile.Reader.get(WritableComparable key,
Writable val)
Return the value for the named key, or null if none exists.
|
Writable |
BloomMapFile.Reader.get(WritableComparable key,
Writable val)
Fast version of the
MapFile.Reader.get(WritableComparable, Writable) method. |
WritableComparable |
MapFile.Reader.getClosest(WritableComparable key,
Writable val)
Finds the record that is the closest match to the specified key.
|
WritableComparable |
MapFile.Reader.getClosest(WritableComparable key,
Writable val,
boolean before)
Finds the record that is the closest match to the specified key.
|
void |
SequenceFile.Reader.getCurrentValue(Writable val)
Get the 'value' corresponding to the last read 'key'.
|
boolean |
SequenceFile.Reader.next(Writable key)
Read the next key in the file into
key, skipping its
value. |
Writable |
ArrayFile.Reader.next(Writable value)
Read and return the next value in the file.
|
boolean |
MapFile.Reader.next(WritableComparable key,
Writable val)
Read the next key/value pair in the map into
key and
val. |
boolean |
SequenceFile.Reader.next(Writable key,
Writable val)
Read the next key/value pair in the file into
key and
val. |
Writable |
SortedMapWritable.put(WritableComparable key,
Writable value) |
Writable |
MapWritable.put(Writable key,
Writable value) |
void |
GenericWritable.set(Writable obj)
Set the instance that is wrapped.
|
void |
ArrayWritable.set(Writable[] values) |
void |
TwoDArrayWritable.set(Writable[][] values) |
static byte[] |
WritableUtils.toByteArray(Writable... writables)
Convert writables to a byte array
|
| 限定符和类型 | 方法和说明 |
|---|---|
static long |
MapFile.fix(FileSystem fs,
Path dir,
Class<? extends Writable> keyClass,
Class<? extends Writable> valueClass,
boolean dryrun,
Configuration conf)
This method attempts to fix a corrupt MapFile by re-creating its index.
|
static long |
MapFile.fix(FileSystem fs,
Path dir,
Class<? extends Writable> keyClass,
Class<? extends Writable> valueClass,
boolean dryrun,
Configuration conf)
This method attempts to fix a corrupt MapFile by re-creating its index.
|
static Writable |
WritableFactories.newInstance(Class<? extends Writable> c)
Create a new instance of a class with a defined factory.
|
static Writable |
WritableFactories.newInstance(Class<? extends Writable> c,
Configuration conf)
Create a new instance of a class with a defined factory.
|
void |
MapWritable.putAll(Map<? extends Writable,? extends Writable> t) |
void |
MapWritable.putAll(Map<? extends Writable,? extends Writable> t) |
void |
SortedMapWritable.putAll(Map<? extends WritableComparable,? extends Writable> t) |
| 构造器和说明 |
|---|
ArrayWritable(Class<? extends Writable> valueClass,
Writable[] values) |
TwoDArrayWritable(Class valueClass,
Writable[][] values) |
| 构造器和说明 |
|---|
ArrayFile.Writer(Configuration conf,
FileSystem fs,
String file,
Class<? extends Writable> valClass)
Create the named file for values of the named class.
|
ArrayFile.Writer(Configuration conf,
FileSystem fs,
String file,
Class<? extends Writable> valClass,
SequenceFile.CompressionType compress,
Progressable progress)
Create the named file for values of the named class.
|
ArrayWritable(Class<? extends Writable> valueClass) |
ArrayWritable(Class<? extends Writable> valueClass,
Writable[] values) |
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class<? extends Writable> valClass,
SequenceFile.CompressionType compress,
CompressionCodec codec,
Progressable progress) |
| 限定符和类型 | 方法和说明 |
|---|---|
Deserializer<Writable> |
WritableSerialization.getDeserializer(Class<Writable> c) |
Serializer<Writable> |
WritableSerialization.getSerializer(Class<Writable> c) |
| 限定符和类型 | 方法和说明 |
|---|---|
Deserializer<Writable> |
WritableSerialization.getDeserializer(Class<Writable> c) |
Serializer<Writable> |
WritableSerialization.getSerializer(Class<Writable> c) |
| 限定符和类型 | 方法和说明 |
|---|---|
abstract Writable |
Server.call(Class<?> protocol,
Writable param,
long receiveTime)
Called for each call.
|
Writable |
RPC.Server.call(Class<?> protocol,
Writable param,
long receivedTime) |
Writable[] |
Client.call(Writable[] params,
InetSocketAddress[] addresses)
|
Writable[] |
Client.call(Writable[] params,
InetSocketAddress[] addresses,
Class<?> protocol,
UserGroupInformation ticket)
|
Writable[] |
Client.call(Writable[] params,
InetSocketAddress[] addresses,
Class<?> protocol,
UserGroupInformation ticket,
Configuration conf)
Makes a set of calls in parallel.
|
Writable |
Client.call(Writable param,
org.apache.hadoop.ipc.Client.ConnectionId remoteId)
Make a call, passing
param, to the IPC server defined by
remoteId, returning the value. |
Writable |
Client.call(Writable param,
InetSocketAddress address)
已过时。
Use
Client.call(Writable, ConnectionId) instead |
Writable |
Client.call(Writable param,
InetSocketAddress addr,
Class<?> protocol,
UserGroupInformation ticket,
int rpcTimeout)
已过时。
Use
Client.call(Writable, ConnectionId) instead |
Writable |
Client.call(Writable param,
InetSocketAddress addr,
Class<?> protocol,
UserGroupInformation ticket,
int rpcTimeout,
Configuration conf)
Make a call, passing
param, to the IPC server running at
address which is servicing the protocol protocol,
with the ticket credentials, rpcTimeout as timeout
and conf as configuration for this connection, returning the
value. |
Writable |
Client.call(Writable param,
InetSocketAddress addr,
UserGroupInformation ticket)
已过时。
Use
Client.call(Writable, ConnectionId) instead |
Writable |
Server.call(Writable param,
long receiveTime)
已过时。
Use
Server.call(Class, Writable, long) instead |
| 限定符和类型 | 方法和说明 |
|---|---|
abstract Writable |
Server.call(Class<?> protocol,
Writable param,
long receiveTime)
Called for each call.
|
Writable |
RPC.Server.call(Class<?> protocol,
Writable param,
long receivedTime) |
Writable[] |
Client.call(Writable[] params,
InetSocketAddress[] addresses)
|
Writable[] |
Client.call(Writable[] params,
InetSocketAddress[] addresses,
Class<?> protocol,
UserGroupInformation ticket)
|
Writable[] |
Client.call(Writable[] params,
InetSocketAddress[] addresses,
Class<?> protocol,
UserGroupInformation ticket,
Configuration conf)
Makes a set of calls in parallel.
|
Writable |
Client.call(Writable param,
org.apache.hadoop.ipc.Client.ConnectionId remoteId)
Make a call, passing
param, to the IPC server defined by
remoteId, returning the value. |
Writable |
Client.call(Writable param,
InetSocketAddress address)
已过时。
Use
Client.call(Writable, ConnectionId) instead |
Writable |
Client.call(Writable param,
InetSocketAddress addr,
Class<?> protocol,
UserGroupInformation ticket,
int rpcTimeout)
已过时。
Use
Client.call(Writable, ConnectionId) instead |
Writable |
Client.call(Writable param,
InetSocketAddress addr,
Class<?> protocol,
UserGroupInformation ticket,
int rpcTimeout,
Configuration conf)
Make a call, passing
param, to the IPC server running at
address which is servicing the protocol protocol,
with the ticket credentials, rpcTimeout as timeout
and conf as configuration for this connection, returning the
value. |
Writable |
Client.call(Writable param,
InetSocketAddress addr,
UserGroupInformation ticket)
已过时。
Use
Client.call(Writable, ConnectionId) instead |
Writable |
Server.call(Writable param,
long receiveTime)
已过时。
Use
Server.call(Class, Writable, long) instead |
| 构造器和说明 |
|---|
Client(Class<? extends Writable> valueClass,
Configuration conf)
Construct an IPC client with the default SocketFactory
|
Client(Class<? extends Writable> valueClass,
Configuration conf,
SocketFactory factory)
Construct an IPC client whose values are of the given
Writable
class. |
Server(String bindAddress,
int port,
Class<? extends Writable> paramClass,
int handlerCount,
Configuration conf) |
Server(String bindAddress,
int port,
Class<? extends Writable> paramClass,
int handlerCount,
Configuration conf,
String serverName) |
Server(String bindAddress,
int port,
Class<? extends Writable> paramClass,
int handlerCount,
Configuration conf,
String serverName,
SecretManager<? extends TokenIdentifier> secretManager)
Constructs a server listening on the named port and address.
|
| 限定符和类型 | 接口和说明 |
|---|---|
interface |
InputSplit
InputSplit represents the data to be processed by an
individual Mapper. |
| 限定符和类型 | 类和说明 |
|---|---|
class |
ClusterStatus
Status information on the current state of the Map-Reduce cluster.
|
class |
Counters
A set of named counters.
|
static class |
Counters.Counter
A counter record, comprising its name and value.
|
class |
Counters.Group
Group of counters, comprising of counters from a particular
counter Enum class. |
class |
FileSplit
A section of an input file.
|
class |
JobConf
A map/reduce job configuration.
|
class |
JobProfile
A JobProfile is a MapReduce primitive.
|
class |
JobQueueInfo
Class that contains the information regarding the Job Queues which are
maintained by the Hadoop Map/Reduce framework.
|
class |
JobStatus
Describes the current status of a job.
|
class |
JvmTask |
class |
MapTaskCompletionEventsUpdate
A class that represents the communication between the tasktracker and child
tasks w.r.t the map task completion events.
|
class |
MultiFileSplit
已过时。
Use
CombineFileSplit instead |
class |
QueueAclsInfo
Class to encapsulate Queue ACLs for a particular
user.
|
class |
Task
Base class for tasks.
|
class |
TaskCompletionEvent
This is used to track task completion events on
job tracker.
|
class |
TaskReport
A report on the state of a task.
|
class |
TaskStatus
Describes the current status of a task.
|
class |
TaskTrackerStatus
A TaskTrackerStatus is a MapReduce primitive.
|
| 限定符和类型 | 方法和说明 |
|---|---|
static <K extends WritableComparable,V extends Writable> |
MapFileOutputFormat.getEntry(MapFile.Reader[] readers,
Partitioner<K,V> partitioner,
K key,
V value)
Get an entry from output generated by this class.
|
| 限定符和类型 | 方法和说明 |
|---|---|
static <K extends WritableComparable,V extends Writable> |
MapFileOutputFormat.getEntry(MapFile.Reader[] readers,
Partitioner<K,V> partitioner,
K key,
V value)
Get an entry from output generated by this class.
|
| 限定符和类型 | 方法和说明 |
|---|---|
RecordWriter<WritableComparable,Writable> |
MapFileOutputFormat.getRecordWriter(FileSystem ignored,
JobConf job,
String name,
Progressable progress) |
static Class<? extends Writable> |
SequenceFileAsBinaryOutputFormat.getSequenceFileOutputValueClass(JobConf conf)
Get the value class for the
SequenceFile |
| 限定符和类型 | 类和说明 |
|---|---|
class |
ArrayListBackedIterator<X extends Writable>
This class provides an implementation of ResetableIterator.
|
interface |
ComposableInputFormat<K extends WritableComparable,V extends Writable>
Refinement of InputFormat requiring implementors to provide
ComposableRecordReader instead of RecordReader.
|
interface |
ComposableRecordReader<K extends WritableComparable,V extends Writable>
Additional operations required of a RecordReader to participate in a join.
|
class |
CompositeRecordReader<K extends WritableComparable,V extends Writable,X extends Writable>
A RecordReader that can effect joins of RecordReaders sharing a common key
type and partitioning.
|
class |
CompositeRecordReader<K extends WritableComparable,V extends Writable,X extends Writable>
A RecordReader that can effect joins of RecordReaders sharing a common key
type and partitioning.
|
class |
MultiFilterRecordReader<K extends WritableComparable,V extends Writable>
Base class for Composite join returning values derived from multiple
sources, but generally not tuples.
|
class |
OverrideRecordReader<K extends WritableComparable,V extends Writable>
Prefer the "rightmost" data source for this key.
|
interface |
ResetableIterator<T extends Writable>
This defines an interface to a stateful Iterator that can replay elements
added to it directly.
|
static class |
ResetableIterator.EMPTY<U extends Writable> |
class |
StreamBackedIterator<X extends Writable>
This class provides an implementation of ResetableIterator.
|
class |
WrappedRecordReader<K extends WritableComparable,U extends Writable>
Proxy class for a RecordReader participating in the join framework.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
CompositeInputSplit
This InputSplit contains a set of child InputSplits.
|
class |
TupleWritable
Writable type storing multiple
Writables. |
| 限定符和类型 | 方法和说明 |
|---|---|
Writable |
TupleWritable.get(int i)
Get ith Writable from Tuple.
|
| 限定符和类型 | 方法和说明 |
|---|---|
Iterator<Writable> |
TupleWritable.iterator()
Return an iterator over the elements in this tuple.
|
| 构造器和说明 |
|---|
TupleWritable(Writable[] vals)
Initialize tuple with storage; unknown whether any of them contain
"written" values.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
CombineFileSplit
A sub-collection of input files.
|
| 限定符和类型 | 方法和说明 |
|---|---|
static Class<? extends Writable> |
MultipleOutputs.getNamedOutputValueClass(JobConf conf,
String namedOutput)
Returns the value class for a named output.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
ValueAggregatorCombiner<K1 extends WritableComparable,V1 extends Writable>
This class implements the generic combiner of Aggregate.
|
class |
ValueAggregatorJobBase<K1 extends WritableComparable,V1 extends Writable>
This abstract class implements some common functionalities of the
the generic mapper, reducer and combiner classes of Aggregate.
|
class |
ValueAggregatorMapper<K1 extends WritableComparable,V1 extends Writable>
This class implements the generic mapper of Aggregate.
|
class |
ValueAggregatorReducer<K1 extends WritableComparable,V1 extends Writable>
This class implements the generic reducer of Aggregate.
|
| 限定符和类型 | 类和说明 |
|---|---|
protected static class |
DBInputFormat.DBInputSplit
A InputSplit that spans a set of rows
|
static class |
DBInputFormat.NullDBWritable
A Class that does nothing, implementing DBWritable
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
ClusterMetrics
Status information on the current state of the Map-Reduce cluster.
|
class |
Counter
A named counter that tracks the progress of a map/reduce job.
|
class |
CounterGroup
A group of
Counters that logically belong together. |
class |
ID
A general identifier, which internally stores the id
as an integer.
|
class |
JobID
JobID represents the immutable and unique identifier for
the job.
|
class |
TaskAttemptID
TaskAttemptID represents the immutable and unique identifier for
a task attempt.
|
class |
TaskID
TaskID represents the immutable and unique identifier for
a Map or Reduce Task.
|
| 限定符和类型 | 类和说明 |
|---|---|
static class |
DataDrivenDBInputFormat.DataDrivenDBInputSplit
A InputSplit that spans a set of rows
|
| 限定符和类型 | 方法和说明 |
|---|---|
static Class<? extends Writable> |
SequenceFileAsBinaryOutputFormat.getSequenceFileOutputValueClass(JobContext job)
Get the value class for the
SequenceFile |
| 限定符和类型 | 类和说明 |
|---|---|
class |
JobTokenIdentifier
The token identifier for job token
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
DelegationTokenIdentifier
A delegation token identifier that is specific to MapReduce.
|
| 限定符和类型 | 类和说明 |
|---|---|
static class |
JobSplit.SplitMetaInfo
This represents the meta information about the task split.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
Record
Abstract class that is extended by generated classes.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
RecordTypeInfo
A record's Type Information object which can read/write itself.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
Credentials
A class that provides the facilities of reading and writing
secret keys and Tokens.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
AccessControlList
Class representing a configured access control list.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
Token<T extends TokenIdentifier>
The client-side form of the token.
|
class |
TokenIdentifier
An identifier that identifies a token, may contain public information
about a token, including its kind (or type).
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
AbstractDelegationTokenIdentifier |
class |
DelegationKey
Key used for generating and verifying delegation tokens
|
| 限定符和类型 | 方法和说明 |
|---|---|
void |
RawBytesInputWriter.writeKey(Writable key) |
void |
RawBytesInputWriter.writeValue(Writable value) |
| 限定符和类型 | 类和说明 |
|---|---|
class |
ResourceUsageMetrics
Captures the resource usage metrics.
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
TypedBytesWritable
Writable for typed bytes.
|
| 限定符和类型 | 方法和说明 |
|---|---|
Writable |
TypedBytesWritableInput.read() |
Writable |
TypedBytesWritableInput.readWritable() |
Writable |
TypedBytesWritableInput.readWritable(Writable writable) |
| 限定符和类型 | 方法和说明 |
|---|---|
Class<? extends Writable> |
TypedBytesWritableInput.readType() |
| 限定符和类型 | 方法和说明 |
|---|---|
Writable |
TypedBytesWritableInput.readWritable(Writable writable) |
void |
TypedBytesWritableOutput.write(Writable w) |
void |
TypedBytesWritableOutput.writeWritable(Writable w) |
| 限定符和类型 | 方法和说明 |
|---|---|
static void |
ReflectionUtils.cloneWritableInto(Writable dst,
Writable src)
已过时。
|
| 限定符和类型 | 类和说明 |
|---|---|
class |
BloomFilter
Implements a Bloom filter, as defined by Bloom in 1970.
|
class |
CountingBloomFilter
Implements a counting Bloom filter, as defined by Fan et al. in a ToN
2000 paper.
|
class |
DynamicBloomFilter
Implements a dynamic Bloom filter, as defined in the INFOCOM 2006 paper.
|
class |
Filter
Defines the general behavior of a filter.
|
class |
Key
The general behavior of a key that must be stored in a filter.
|
class |
RetouchedBloomFilter
Implements a retouched Bloom filter, as defined in the CoNEXT 2006 paper.
|
Copyright © 2009 The Apache Software Foundation