|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
AccessControlException
instead.RemoteException
.
AccessControlException
with the specified detail message.
allocate
allocate
Configuration.addDeprecation(String key, String newKey,
String customMessage)
instead
Configuration.addDeprecation(String key, String newKey)
instead
TimelinePutResponse.TimelinePutError
instance into the existing list
TimelinePutResponse.TimelinePutError
instances into the existing list
TimelineEvents.EventsOfOneEntity
instance into the existing list
TimelineEvents.EventsOfOneEntity
instances into the existing list
Service
,
add it to the list of services managed by this CompositeService
Path
to the list of inputs for the map-reduce job.
Path
with a custom InputFormat
to the list of
inputs for the map-reduce job.
Path
with a custom InputFormat
and
Mapper
to the list of inputs for the map-reduce job.
Path
to the list of inputs for the map-reduce job.
Path
with a custom InputFormat
to the list of
inputs for the map-reduce job.
Path
with a custom InputFormat
and
Mapper
to the list of inputs for the map-reduce job.
Mapper
class to the chain mapper.
Mapper
class to the chain reducer.
Service
to the list of services managed by this
CompositeService
ApplicationMaster
and the
ResourceManager
.
ApplicationMaster
to the
ResourceManager
to obtain resources in the cluster.ResourceManager
the
ApplicationMaster
during resource negotiation.AMRMClientAsync
handles communication with the ResourceManager
and provides asynchronous updates on events such as container allocations and
completions.ApplicationAttemptId
denotes the particular attempt
of an ApplicationMaster
for a given ApplicationId
.(GetApplicationAttemptReportRequest)
API when the Application Attempt doesn't exist in Application History ServerApplicationAttemptReport
is a report of an application attempt.URLClassLoader
for YARN application isolation.ResourceManager
to submit/abort jobs and to get information on applications, cluster metrics,
nodes, queues and ACLs.ApplicationHistoryServer
to
get the information of completed applications etc.ApplicationId
represents the globally unique
identifier for an application.ApplicationId
in ApplicationSubmissionContext
.ApplicationMaster
and the ResourceManager
.(GetApplicationReportRequest)
API
when the Application doesn't exist in RM and AHSApplicationReport
is a report of an application.ApplicationSubmissionContext
represents all of the
information needed by the ResourceManager
to launch
the ApplicationMaster
for an application.Event
s in a separate thread.FSDataInputStream
to Avro's SeekableInput interface.FSDataInputStream
and its length.
FileContext
and a Path
.
SecretManager
for AMs to extend and validate Client-RM tokens
issued to clients by the RM using the underlying master-key shared by RM to
the AMs on their launch.WritableComparable
types supporting ordering/permutation by a representative set of bytes.BinaryComparable
keys using a configurable part of
the bytes array returned by BinaryComparable.getBytes()
.BinaryComparable
keys using a configurable part of
the bytes array returned by BinaryComparable.getBytes()
.CompressorStream
which works
with 'block-based' based compression algorithms, as opposed to
'stream-based' compression algorithms.BlockCompressorStream
.
BlockCompressorStream
with given output-stream and
compressor.
DecompressorStream
which works
with 'block-based' based compression algorithms, as opposed to
'stream-based' compression algorithms.BlockDecompressorStream
.
BlockDecompressorStream
.
BlockLocation
that also adds VolumeId
volume
location information for each replica.MapFile
and provides very much the same
functionality.Token.cancel(org.apache.hadoop.conf.Configuration)
instead
Token.cancel(org.apache.hadoop.conf.Configuration)
instead
Token
.
ResourceManager
to cancel a
delegation token.ResourceManager
to a cancelDelegationToken
request.File.canExecute()
File.canRead()
File.canWrite()
position
.
IOException
or
null pointers.
OutputCommitter.commitJob(JobContext)
or
OutputCommitter.abortJob(JobContext, int)
instead.
OutputCommitter.commitJob(org.apache.hadoop.mapreduce.JobContext)
or OutputCommitter.abortJob(org.apache.hadoop.mapreduce.JobContext, org.apache.hadoop.mapreduce.JobStatus.State)
instead.
OutputCommitter.commitJob(JobContext)
and
OutputCommitter.abortJob(JobContext, JobStatus.State)
instead.
SecretManager
for AMs to validate Client-RM tokens issued to
clients by the RM using the underlying master-key shared by RM to the AMs on
their launch.JobClient
.
InputSplit
to future operations.
RecordWriter
to future operations.
Cluster
.
RecordWriter
to future operations.
AbstractService.stop()
IOException
IOException
.
MultiFilterRecordReader.emit(org.apache.hadoop.mapred.join.TupleWritable)
every Tuple from the
collector (the outer join of child RRs).
MultiFilterRecordReader.emit(org.apache.hadoop.mapreduce.lib.join.TupleWritable)
every Tuple from the
collector (the outer join of child RRs).
InputFormat
that returns CombineFileSplit
's
in InputFormat.getSplits(JobConf, int)
method.InputFormat
that returns CombineFileSplit
's in
InputFormat.getSplits(JobContext)
method.CombineFileSplit
.CombineFileSplit
.CombineFileInputFormat
-equivalent for
SequenceFileInputFormat
.CombineFileInputFormat
-equivalent for
SequenceFileInputFormat
.CombineFileInputFormat
-equivalent for
TextInputFormat
.CombineFileInputFormat
-equivalent for
TextInputFormat
.CompressionOutputStream
to compress data.Configuration
.JobConf
.
JobConf
.
Configuration
.Configuration
.
NetUtils.connect(java.net.Socket, java.net.SocketAddress, int)
if it times out while connecting to the remote host.Container
represents an allocated resource in the cluster.ContainerId
represents a globally unique identifier
for a Container
in the cluster.ContainerLaunchContext
represents all of the information
needed by the NodeManager
to launch a container.ApplicationMaster
and a
NodeManager
to start/stop containers and to get status
of running containers.(GetContainerReportRequest)
API when the container doesn't exist in AHSContainerReport
is a report of an container.Container
.ContainerStatus
represents the current status of a
Container
.RuntimeException
.
RuntimeException
.
Counter
s that logically belong together.Counters
holds per job/task counters, defined either by the
Map-Reduce framework or applications.Group
of counters, comprising of counters from a particular
counter Enum
class.FileContext.create(Path, EnumSet, Options.CreateOpts...)
except
that the Path f must be fully qualified and the permission is absolute
(i.e.
YarnClientApplication
for a new application,
which in turn contains the ApplicationSubmissionContext
and
GetNewApplicationResponse
objects.
Compressor
for use by this CompressionCodec
.
Compressor
for use by this CompressionCodec
.
Decompressor
for use by this CompressionCodec
.
Decompressor
for use by this CompressionCodec
.
DirectDecompressor
for use by this DirectDecompressionCodec
.
DirectDecompressor
for use by this DirectDecompressionCodec
.
FsPermission
object.
CompressionInputStream
that will read from the given
input stream and return a stream for uncompressed data.
CompressionInputStream
that will read from the given
InputStream
with the given Decompressor
, and return a
stream for uncompressed data.
CompressionInputStream
that will read from the given
input stream.
CompressionInputStream
that will read from the given
InputStream
with the given Decompressor
.
AbstractFileSystem.create(Path, EnumSet, Options.CreateOpts...)
except that the opts
have been declared explicitly.
IOException
.
CompressionOutputStream
that will write to the given
OutputStream
.
CompressionOutputStream
that will write to the given
OutputStream
with the given Compressor
.
CompressionOutputStream
that will write to the given
OutputStream
.
CompressionOutputStream
that will write to the given
OutputStream
with the given Compressor
.
CombineFileInputFormat.createPool(List)
.
CombineFileInputFormat.createPool(PathFilter...)
.
recordName
.
FileContext.createSymlink(Path, Path, boolean)
;
FileContext.createSymlink(Path, Path, boolean)
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
SequenceFile.createWriter(Configuration, Writer.Option...)
instead.
DBWritable
.CompressionInputStream
to compress data.Stringifier
interface which stringifies the objects using base64 encoding of the
serialized version of the objects.WritableComparable
implementation.
Record
implementation.
FileContext.delete(Path, boolean)
except that Path f must be for
this file system.
FileSystem.delete(Path, boolean)
instead.
Writer
The format of the output would be
{ "properties" : [ {key1,value1,key1.isFinal,key1.resource}, {key2,value2,
key2.isFinal,key2.resource}...
o
is a ByteWritable with the same value.
o
is a DoubleWritable with the same value.
o
is an EnumSetWritable with the same value,
or both are null.
o
is a FloatWritable with the same value.
o
is a IntWritable with the same value.
o
is a LongWritable with the same value.
o
is an MD5Hash whose digest contains the
same values.
o
is a ShortWritable with the same value.
o
is a Text with the same contents.
o
is a VIntWritable with the same value.
o
is a VLongWritable with the same value.
InputFormat
.InputFormat
s.OutputCommitter
that commits files specified
in job output directory i.e.OutputCommitter
that commits files specified
in job output directory i.e.OutputFormat
.OutputFormat
s that read from FileSystem
s.FilterFileSystem
contains
some other file system, which it uses as
its basic file system, possibly transforming
the data along the way or providing additional
functionality.Application
.what
in the backing
buffer, starting as position start
.
Counters.findCounter(String, String)
instead
ApplicationMaster
to notify the
ResourceManager
about its completion (success or failed).
ApplicationMaster
to
inform the ResourceManager
about its completion.ResourceManager
to a
ApplicationMaster
on it's completion.true
if the end of the decompressed
data output stream has been reached.
FileContext.fixRelativePart(org.apache.hadoop.fs.Path)
ResourceManager
to abort submitted application.
Counters.makeEscapedCompactString()
counter
representation into a counter object.
FSInputStream
in a DataInputStream
and buffers input through a BufferedInputStream
.OutputStream
in a DataOutputStream
.FsAction
.
FileSystem
.Throwable
into a Runtime Exception.FileSystem
backed by an FTP client provided by Apache Commons Net.FileSystem.delete(Path, boolean)
name
property, null
if
no such property exists.
name
.
BytesWritable.getBytes()
instead.
WritableComparable
implementation.
TimelineEvents.EventsOfOneEntity
instances
Cluster.getAllJobStatuses()
instead.
Container
by the
ResourceManager
.
Resource
of the container.
QueueInfo
) about all queues, recursively if there
is a hierarchy
NodeManager
.
ResourceManager
needs the
ApplicationMaster
to take some action then it will send an
AMCommand to the ApplicationMaster
.
ContainerId
of AMContainer for this attempt
ContainerLaunchContext
to describe the
Container
with which the ApplicationMaster
is
launched.
ApplicationACL
s for the application.
ApplicationACL
s for the application.
ApplicationAttemptId
of an application attempt.
ApplicationAttemptId
of an application attempt.
ApplicationAttemptId
of this attempt of the
application
ApplicationAttemptId
of the application to which the
Container
was assigned.
ApplicationReport
of an application.
ResourceManager
ApplicationHistoryServer
.
ApplicationAttemptReport
for the application attempt.
ResourceManager
to get an
ApplicationAttemptReport
for an application attempt.ResourceManager
to a client requesting
an application attempt report.ResourceManager
ApplicationHistoryServer
.
ResourceManager
.ResourceManager
to a client requesting
a list of ApplicationAttemptReport
for application attempts.ApplicationId
of an application
ApplicationId
of the application.
ApplicationId
allocated by the
ResourceManager
.
ApplicationId
of the application to be aborted.
ApplicationId
of the application to be moved.
ApplicationId
of the ApplicationAttempId
.
ApplicationId
of the application.
ApplicationId
of the submitted application.
ApplicationReport
for applications.
ResourceManager
.
ResourceManager
.
ApplicationReport
for the application.
ResourceManager
to
get an ApplicationReport
for an application.ResourceManager
to a client
requesting an application report.GetApplicationsRequest
in the cluster from the ResourceManager
.
ApplicationHistoryServer
.
ResourceManager
.ResourceManager
to a client
requesting an ApplicationReport
for applications.ApplicationSubmissionContext
for the application.
ResourceRequest
to update the
ResourceManager
about the application's resource requirements.
NodeId
where container is running.
attempt id
of the Application
.
SkipBadRecords.COUNTER_MAP_PROCESSED_RECORDS
is incremented
by MapRunner after invoking the map function.
SkipBadRecords.COUNTER_REDUCE_PROCESSED_GROUPS
is incremented
by framework after invoking the reduce function.
name
property as a boolean
.
Text.getLength()
is
valid.
Resource
on the node.
Resource
capability of the request.
QueueInfo
) about all the immediate children queues
of the given queue
name
property as a Class
.
name
property as a Class
implementing the interface specified by xface
.
Class
of the given object.
name
property
as an array of Class
.
ClassLoader
for this job.
ApplicationMaster
.
ResourceManager
.
YarnClusterMetrics
for the cluster.
ResourceManager
.ResourceManager
to a client
requesting cluster metrics.ResourceManager
.
ResourceManager
.ResourceManager
to a client
requesting a NodeReport
for all nodes.ResourceManager
which is
used to generate globally unique ApplicationId
.
WritableComparable
comparator for
grouping keys of inputs to the combiner.
RawComparator
comparator for
grouping keys of inputs to the combiner.
Compressor
for the given CompressionCodec
from the
pool or a new one.
Compressor
needed by this CompressionCodec
.
Compressor
needed by this CompressionCodec
.
name
.
Reader
attached to the configuration resource with the
given name
.
exit status
of the container.
ContainerId
of the Container.
ContainerId
of the container.
ContainerId
of the container.
ContainerId
s of containers for which to obtain
the ContainerStatus
.
ContainerId
s of the containers to be stopped.
ContainerLaunchContext
for the container to be started
by the NodeManager
.
ContainerReport
for all the containers of an
application attempt.
ResourceManager
ApplicationHistoryServer
.
ContainerReport
for the container.
ResourceManager
to get an
ContainerReport
for a container.ResourceManager
to a client requesting
a container report.ResourceManager
ApplciationHistoryServer
.
PreemptionContainer
specifying which containers
owned by the ApplicationMaster
that may be reclaimed by the
ResourceManager
.
PreemptionContainer
specifying containers owned by
the ApplicationMaster
that may be reclaimed by the
ResourceManager
.
ResourceManager
from previous application attempts.
ResourceManager
.ResourceManager
to a client requesting
a list of ContainerReport
for containers.ContainerState
of the container.
ApplicationMaster
to request for current
statuses of Container
s from the NodeManager
.
ContainerStatus
es of the requested containers.
ApplicationMaster
to the
NodeManager
to get ContainerStatus
of requested
containers.NodeManager
to the
ApplicationMaster
when asked to obtain the
ContainerStatus
of requested containers.ContainerToken
for the container.
ContentSummary
of a given Path
.
Counters.Group.findCounter(String)
instead
Counters.Counter
of the given group with the given name.
Counters.Counter
of the given group with the given name.
Counter
for the given counterName
.
Counter
for the given groupName
and
counterName
.
ApplicationAttemptId
of the current
attempt of the application
Decompressor
for the given CompressionCodec
from the
pool or a new one.
Decompressor
needed by this CompressionCodec
.
Decompressor
needed by this CompressionCodec
.
NodeManager
FileSystem.getDefaultBlockSize(Path)
instead
FileSystem.getDefaultReplication(Path)
instead
Credentials.getToken(org.apache.hadoop.io.Text)
instead, this method is included for compatibility against Hadoop-1
ResourceManager
.GetDelegationTokenRequest
request
from the client.name
property as a double
.
Runnable
that periodically empties the trash of all
users, intended to be run by the superuser.
Runnable
that periodically empties the trash of all
users, intended to be run by the superuser.
TimelinePutResponse.TimelinePutError
instances
Service.getFailureCause()
occurred.
FileContext.getFileBlockLocations(Path, long, long)
except that
Path f must be for this file system.
FileContext.getFileChecksum(Path)
except that Path f must be for
this file system.
FileContext.getFileLinkStatus(Path)
except that an UnresolvedLinkException may be thrown if a symlink is
encountered in the path leading up to the final path component.
FileContext.getFileLinkStatus(Path)
FileContext.getFileStatus(Path)
except that an UnresolvedLinkException may be thrown if a symlink is
encountered in the path.
ApplicationMaster
.
name
property as a float
.
FileContext.getFsStatus(Path)
except that Path f must be for this
file system.
FileContext.getFsStatus(Path)
.
FsAction
.
RawComparator
comparator for
grouping keys of inputs to the reduce.
ApplicationMaster
is
running.
ApplicationMaster
is running.
ApplicationMaster
is running.
ApplicationId
which is unique for all applications started by a particular instance
of the ResourceManager
.
ContainerId
.
ResourceManager
ContainerResourceIncreaseRequest
being sent by the
ApplicationMaster
InputFormat
implementation for the map-reduce job,
defaults to TextInputFormat
if not specified explicity.
InputFormat
class for the job.
Path
s for the map-reduce job.
Path
s for the map-reduce job.
InputSplit
object for a map.
Job
with no particular Cluster
.
Job
with no particular Cluster
and a
given Configuration
.
Job
with no particular Cluster
and a given jobName.
Job
with no particular Cluster
and given
Configuration
and JobStatus
.
Job.getInstance()
Job.getInstance(Configuration)
Job
with no particular Cluster
and given
Configuration
and JobStatus
.
name
property as a List
of objects implementing the interface specified by xface
.
name
property as an int
.
name
property as a set of comma-delimited
int
values.
RunningJob
object to track an ongoing job.
JobClient.getJob(JobID)
.
RunningJob.getID()
.
JobID
object that this task attempt belongs to
JobID
object that this tip belongs to
JobPriority
for this job.
JobStatus
, of the Job.
SequenceFileRecordReader.next(Object, Object)
..
KeyFieldBasedComparator
options
KeyFieldBasedComparator
options
KeyFieldBasedPartitioner
options
KeyFieldBasedPartitioner
options
Compressor
s for this
CompressionCodec
Decompressor
s for this
CompressionCodec
InputSplit
.
FileContext.getLinkTarget(Path)
JobContext.getCacheArchives()
.
JobContext.getCacheFiles()
.
LocalResource
required by the container.
name
property as a long
.
name
property as a long
or
human readable format.
WrappedMapper.Context
for custom implementations.
CompressionCodec
for compressing the map outputs.
Mapper
class for the job.
Mapper
class for the job.
MapRunnable
class for the job.
true
.
JobClient.getMapTaskReports(JobID)
ContainerRequest
s matching the given
parameters.
Resource
allocated by the
ResourceManager
in the cluster.
Resource
allocated by the
ResourceManager
in the cluster.
mapreduce.map.maxattempts
property.
mapred.map.max.attempts
property.
mapreduce.reduce.maxattempts
property.
mapred.reduce.max.attempts
property.
JobConf.getMemoryForMapTask()
and
JobConf.getMemoryForReduceTask()
Resource
.
ApplicationId
for
submitting new applications.
ApplicationId
for
submitting an application.ResourceManager
to the client for
a request to get a new ApplicationId
for submitting applications.AMRMClient
.
NMClient
.
NodeId
of the NodeManager
for which the NMToken
is used to authenticate.
NodeId
of the node.
NodeReport
for all nodes in the cluster.
NodeReport
) in the cluster.
NodeState
of the node.
NodeManager
s in the cluster.
FsAction
.
OutputCommitter
implementation for the map-reduce job,
defaults to FileOutputCommitter
if not specified explicitly.
OutputCommitter
for the task-attempt.
SequenceFile.CompressionType
for the output SequenceFile
.
SequenceFile.CompressionType
for the output SequenceFile
.
CompressionCodec
for compressing the job outputs.
CompressionCodec
for compressing the job outputs.
OutputFormat
implementation for the map-reduce job,
defaults to TextOutputFormat
if not specified explicity.
OutputFormat
class for the job.
RawComparator
comparator used to compare keys.
Path
to the output directory for the map-reduce job.
Path
to the output directory for the map-reduce job.
WritableComparable
comparator for
grouping keys of inputs to the reduce.
Object.hashCode()
to partition.
BinaryComparable.getBytes()
to partition.
Object.hashCode()
to partition.
Partitioner
used to partition Mapper
-outputs
to be sent to the Reducer
s.
Partitioner
class for the job.
TotalOrderPartitioner.getPartitionFile(Configuration)
instead
Path
for a file that is unique for
the task within the job output directory.
Path
for a file that is unique for
the task within the job output directory.
name
property as a Pattern
.
PATTERN
).
Priority
of the application.
Priority
at which the Container
was
allocated.
Priority
of the container.
Priority
of the request.
RecordReader
consumed i.e.
ResourceManager
.
QueueInfo
for the specified queue.
QueueInfo
) about a given queue.
ResourceManager
.ResourceManager
to a client
requesting information about queues in the system.QueueState
of the queue.
ResourceManager
.
ResourceManager
to
get queue acls for the current user.ResourceManager
to clients
seeking queue acls for the user.name
property, without doing
variable expansion.If the key is
deprecated, it returns the value of the first key which replaces
the deprecated key and is not null.
RecordReader
for the given InputSplit
.
RecordReader
for the given InputSplit
.
RecordWriter
for the given job.
RecordWriter
for the given job.
RecordWriter
for the given task.
RecordWriter
for the given task.
Reducer
class for the job.
Reducer
class for the job.
WrappedReducer.Context
for custom implementations.
true
.
JobClient.getReduceTaskReports(JobID)
ResourceRequest
.
ContainerId
of containers being
released by the ApplicationMaster
.
TaskType
Resource
.
URL
for the named resource.
ApplicationMaster
for this
application.
Resource
allocated to the container.
ResourceBlacklistRequest
being sent by the
ApplicationMaster
.
PreemptionContainer
s enumerated in PreemptionContract.getContainers()
should not be
evicted from the cluster.
AbstractDelegationTokenIdentifier
.
QueueInfo
) about top level queues.
ApplicationMaster
is responding.
ApplicationMaster
.
ApplicationMaster
.
ApplicationsRequestScope
of applications to be filtered.
SequenceFile
SequenceFile
SequenceFile
SequenceFile
FileSystem.getServerDefaults(Path)
instead
BytesWritable.getLength()
instead.
name
property as a
InetSocketAddress
.
name
property as a
InetSocketAddress
.
RawComparator
comparator used to compare keys.
true
.
FileInputFormat.listStatus(JobConf)
when
they're too big.
StartContainerRequest
to start containers.
ContainerState
of the container.
FileSystem.getAllStatistics()
instead
name
property as
a collection of String
s.
name
property as
an array of String
s.
name
property as
an array of String
s.
ContainerId
s of the containers that are
started successfully.
TaskCompletionEvent.getTaskAttemptId()
instead.
TaskID
object that this task attempt belongs to
TaskID.getTaskIDsPattern(String, Integer, TaskType,
Integer)
TaskType
corresponding to the character
Token
used for authenticating with NodeManager
ApplicationMaster
.
ApplicationMaster
.
name
property as a trimmed String
,
null
if no such property exists.
name
property as a trimmed String
,
defaultValue
if no such property exists.
name
property as
a collection of String
s, trimmed of the leading and trailing whitespace.
name
property as
an array of String
s, trimmed of the leading and trailing whitespace.
name
property as
an array of String
s, trimmed of the leading and trailing whitespace.
LocalResourceType
of the resource to be localized.
UMASK_LABEL
config param has umask value that is either symbolic
or octal.
NodeReport
s.
Resource
on the node.
Resource
.
QueueACL
for the given user.
QueueUserACLInfo
per queue for the user.
FsAction
.
SequenceFileRecordReader.next(Object, Object)
..
LocalResourceVisibility
of the resource to be
localized.
VolumeId
corresponding to the block's replicas.
Path
to the task's temporary output directory
for the map-reduce job
Path
to the task's temporary output directory
for the map-reduce job
YarnApplicationState
of the application.
YarnClusterMetrics
) about the cluster.
Groups
.HAServiceProtocol
RPC calls.Object.hashCode()
.Object.hashCode()
.VolumeId
.Enum
type, by the specified amount.
InputFormat
describes the input-specification for a
Map-Reduce job.InputFormat
describes the input-specification for a
Map-Reduce job.TotalOrderPartitioner
.InputSplit
represents the data to be processed by an
individual Mapper
.InputSplit
represents the data to be processed by an
individual Mapper
.Mapper
that swaps keys and values.Mapper
that swaps keys and values.key
is deprecated.
FileStatus.isFile()
,
FileStatus.isDirectory()
, and FileStatus.isSymlink()
instead.
DNSToSwitchMapping
instance being on a single
switch.
OutputCommitter.isRecoverySupported(JobContext)
instead.
OutputCommitter.isRecoverySupported(JobContext)
instead.
AbstractDNSToSwitchMapping.isMappingSingleSwitch(DNSToSwitchMapping)
CombineFileInputFormat.isSplitable(FileSystem, Path)
.
Iterator
to go through the list of String
key-value pairs in the configuration.
Serialization
for Java Serializable
classes.RawComparator
that uses a JavaSerialization
Deserializer
to deserialize objects that are then compared via
their Comparable
interfaces.JobClient
is the primary interface for the user-job to interact
with the cluster.JobConf
, and connect to the
default cluster
Configuration
,
and connect to the default cluster
KeyFieldBasedComparator
.KeyFieldBasedComparator
.InputFormat
for plain text files.InputFormat
for plain text files.ResourceManager
to abort a submitted application.ResourceManager
to the client aborting
a submitted application.RunningJob.killTask(TaskAttemptID, boolean)
File.list()
.
File.listFiles()
.
FileContext.listLocatedStatus(Path)
except that Path f
must be for this file system.
FileContext.Util.listStatus(Path)
except that Path f must be
for this file system.
f
is a file, this method will make a single call to S3.
FileContext.listStatus(Path)
except that Path f must be for this
file system.
Credentials.readTokenStorageFile(org.apache.hadoop.fs.Path, org.apache.hadoop.conf.Configuration)
instead,
this method is included for compatibility against Hadoop-1.
Credentials.readTokenStorageFile(org.apache.hadoop.fs.Path, org.apache.hadoop.conf.Configuration)
instead,
this method is included for compatibility against Hadoop-1.
LocalResource
represents a local resource required to
run a container.LocalResourceType
specifies the type
of a resource localized by the NodeManager
.LocalResourceVisibility
specifies the visibility
of a resource localized by the NodeManager
.Reducer
that sums long values.map(...)
methods of the Mappers in the chain.
Mapper
.OutputFormat
that writes MapFile
s.OutputFormat
that writes
MapFile
s.Level
for the map task.
Level
for the reduce task.
JobConf.MAPRED_MAP_TASK_ENV
or
JobConf.MAPRED_REDUCE_TASK_ENV
JobConf.MAPRED_MAP_TASK_JAVA_OPTS
or
JobConf.MAPRED_REDUCE_TASK_JAVA_OPTS
JobConf.MAPREDUCE_JOB_MAP_MEMORY_MB_PROPERTY
and
JobConf.MAPREDUCE_JOB_REDUCE_MEMORY_MB_PROPERTY
Mapper
and Reducer
implementations.Mapper
s.MapRunnable
implementation.MarkableIterator
is a wrapper iterator class that
implements the MarkableIteratorInterface
.MBeans.register(String, String, Object)
FileContext.mkdir(Path, FsPermission, boolean)
except that the Path
f must be fully qualified and the permission is absolute (i.e.
FileSystem.mkdirs(Path, FsPermission)
with default permission.
ResourceManager
to move a submitted application to a different queue.ResourceManager
to the client moving
a submitted application to a different queue.InputFormat
that returns MultiFileSplit
's
in MultiFileInputFormat.getSplits(JobConf, int)
method.InputFormat
and Mapper
for each pathInputFormat
and Mapper
for each pathIOException
into an IOException
OutputCollector
passed to
the map()
and reduce()
methods of the
Mapper
and Reducer
implementations.MutableQuantiles
for a metric that rolls itself
over on the specified time interval.
FileSystem
for reading and writing files stored on
Amazon S3.true
if a preset dictionary is needed for decompression.
true
if the input data buffer is empty and
Decompressor.setInput(byte[], int, int)
should be called to
provide more input.
ResourceManager
.
ResourceManager
.
ResourceManager
.
ResourceManager
.
ResourceManager
.
RegisterApplicationMasterRequest
.
WritableComparable
instance.
DBRecordReader.nextKeyValue()
NMClientAsync
handles communication with all the NodeManagers
and provides asynchronous updates on getting responses from them.NodeManager
ScriptBasedMapping.toString()
method if there is no string
"no script"
NodeId
is the unique identifier for a node.NodeReport
is a summary of runtime information of a
node in the cluster.Node
.FileContext.open(Path)
except that Path f must be for this
file system.
FileContext.open(Path, int)
except that Path f must be for this
file system.
FileSystem
that uses Amazon S3
as a backing store.FileSystem
for reading and writing files on
Amazon S3.JMXJsonServlet
class.<key, value>
pairs output by Mapper
s
and Reducer
s.OutputCommitter
describes the commit of task output for a
Map-Reduce job.OutputCommitter
describes the commit of task output for a
Map-Reduce job.OutputFormat
describes the output-specification for a
Map-Reduce job.OutputFormat
describes the output-specification for a
Map-Reduce job.OutputCommitter
that commits files specified
in job output directory i.e.OutputCommitter
implementing partial commit of task output, as during preemption.FileSystem
.ResourceManager
.ResourceManager
.PreemptionMessage
is part of the RM-AM protocol, and it is used by
the RM to specify resources that the RM wants to reclaim from this
ApplicationMaster
(AM).QueueACL
enumerates the various ACLs for queues.QueueUserACLInfo
provides information QueueACL
for
the given user.RawComparator
.Comparator
that operates directly on byte representations of
objects.FsPermission
from DataInput
.
EOFException
for getting logs of
all types for a single container.
in
.
in
.
ResultSet
.
in
.
in
.
CompressedWritable.readFields(DataInput)
.
FSDataInputStream.readFully(long, byte[], int, int)
.
Writable
, String
, primitive type, or an array of
the preceding.
Writable
, String
, primitive type, or an array of
the preceding.
Record
comparison implementation.
RecordReader
reads <key, value> pairs from an
InputSplit
.Mapper
.RecordWriter
writes the output <key, value> pairs
to an output file.RecordWriter
writes the output <key, value> pairs
to an output file.reduce(...)
method of the Reducer with the
map(...)
methods of the Mappers in the chain.
Reducer
.Mapper
that extracts text matching a regular expression.Mapper
that extracts text matching a regular expression.ApplicationMaster
to register with
the ResourceManager
.
ApplicationMaster
to
ResourceManager
on registration.ResourceManager
to a new
ApplicationMaster
on registration.FileContext.rename(Path, Path, Options.Rename...)
except that Path
f must be for this file system.
FileContext.rename(Path, Path, Options.Rename...)
except that Path
f must be for this file system and NO OVERWRITE is performed.
FileContext.rename(Path, Path, Options.Rename...)
except that Path
f must be for this file system.
Token.renew(org.apache.hadoop.conf.Configuration)
instead
Token.renew(org.apache.hadoop.conf.Configuration)
instead
Token
.
ResourceManager
.ResourceManager
.AbstractFileSystem.getLinkTarget(Path)
Resource
models a set of computer resources in the
cluster.ResourceBlacklistRequest
encapsulates the list of resource-names
which should be added or removed from the blacklist of resources
for the application.ResourceRequest
represents the request made by an
application to the ResourceManager
to obtain various
Container
allocations.Compressor
to the pool.
Decompressor
to the pool.
Reducer.run(org.apache.hadoop.mapreduce.Reducer.Context)
method to
control how the reduce task works.
Tool
by Tool.run(String[])
, after
parsing with the given generic arguments.
Tool
with its Configuration
.
RunningJob
is the user-interface to query for details on a
running Map-Reduce job.FileSystem
backed by
Amazon S3.S3FileSystem
.DNSToSwitchMapping
interface using a
script configured via the
CommonConfigurationKeysPublic.NET_TOPOLOGY_SCRIPT_FILE_NAME_KEY
option.SequenceFile
s are flat files consisting of binary key/value
pairs.OutputFormat
that writes keys, values to
SequenceFile
s in binary(raw) formatOutputFormat
that writes keys,
values to SequenceFile
s in binary(raw) formatInputFormat
for SequenceFile
s.InputFormat
for SequenceFile
s.OutputFormat
that writes SequenceFile
s.OutputFormat
that writes SequenceFile
s.RecordReader
for SequenceFile
s.RecordReader
for SequenceFile
s.Service.STATE.NOTINITED
state.
value
of the name
property.
value
of the name
property.
Container
by the
ResourceManager
.
NodeManager
.
ContainerLaunchContext
to describe the
Container
with which the ApplicationMaster
is
launched.
ApplicationACL
s for the application.
ApplicationACL
s for the application.
ApplicationAttemptId
of an application attempt
ApplicationAttemptId
of an application attempt
ApplicationReport
of an application.
ApplicationAttemptReport
for the application attempt.
ApplicationId
of an application
ApplicationId
of the application
ApplicationId
of the application to be moved.
ApplicationId
of the submitted application.
ApplicationSubmissionContext
for the application.
ResourceRequest
to update the
ResourceManager
about the application's resource requirements.
SkipBadRecords.COUNTER_MAP_PROCESSED_RECORDS
is incremented
by MapRunner after invoking the map function.
SkipBadRecords.COUNTER_REDUCE_PROCESSED_GROUPS
is incremented
by framework after invoking the reduce function.
name
property to a boolean
.
Resource
capability of the request
name
property to the name of a
theClass
implementing the given interface xface
.
RawComparator
comparator for
grouping keys in the input to the combiner.
Reducer.reduce(Object, Iterable,
org.apache.hadoop.mapreduce.Reducer.Context)
ContainerId
of the container
ContainerId
s of containers for which to obtain
the ContainerStatus
ContainerId
s of the containers to be stopped.
ContainerLaunchContext
for the container to be started
by the NodeManager
ContainerReport
for all the containers of an
application attempt.
ResourceManager
from previous application attempts.
ContainerStatus
es of the requested containers.
NodeManager
name
property to a double
.
name
property to the given type.
TimelinePutResponse.TimelinePutError
instances
TimelineEvents.EventsOfOneEntity
instances
File.setExecutable(boolean)
File#setExecutable does not work as expected on Windows.
ApplicationMaster
name
property to a float
.
Reducer.reduce(Object, Iterable,
org.apache.hadoop.mapreduce.Reducer.Context)
ApplicationMaster
is
running.
ResourceManager
ContainerResourceIncreaseRequest
to inform the
ResourceManager
about some container's resources need to be
increased
InputFormat
implementation for the map-reduce job.
InputFormat
for the job.
Path
s as the list of inputs
for the map-reduce job.
Path
s as the list of inputs
for the map-reduce job.
name
property to an int
.
JobPriority
for this job.
KeyFieldBasedComparator
options used to compare keys.
KeyFieldBasedComparator
options used to compare keys.
KeyFieldBasedPartitioner
options used for
Partitioner
KeyFieldBasedPartitioner
options used for
Partitioner
bytes[offset:]
in Python syntax.
LocalResource
required by the container.
name
property to a long
.
CompressionCodec
for the map outputs.
Mapper
class for the job.
Mapper
for the job.
MapRunnable
class for the job.
JobConf.setMemoryForMapTask(long mem)
and
Use JobConf.setMemoryForReduceTask(long mem)
NMTokenCache.getSingleton()
.
AMRMClient
.
NMClient
.
bytes[left:(right+1)]
in Python syntax.
OutputCommitter
implementation for the map-reduce job.
SequenceFile.CompressionType
for the output SequenceFile
.
SequenceFile.CompressionType
for the output SequenceFile
.
CompressionCodec
to be used to compress job outputs.
CompressionCodec
to be used to compress job outputs.
OutputFormat
implementation for the map-reduce job.
OutputFormat
for the job.
RawComparator
comparator used to compare keys.
Path
of the output directory for the map-reduce job.
Path
of the output directory for the map-reduce job.
RawComparator
comparator for
grouping keys in the input to the reduce.
FileContext.setOwner(Path, String, String)
except that Path f must
be for this file system.
Partitioner
class used to partition
Mapper
-outputs to be sent to the Reducer
s.
Partitioner
for the job.
TotalOrderPartitioner.setPartitionFile(Configuration, Path)
instead
Pattern
.
PATTERN
).
FileContext.setPermission(Path, FsPermission)
except that Path f
must be for this file system.
Priority
of the application.
Priority
of the request
File.setReadable(boolean)
File#setReadable does not work as expected on Windows.
Reducer
class to the chain job.
Reducer
class for the job.
Reducer
for the job.
ContainerId
of containers being
released by the ApplicationMaster
FileContext.setReplication(Path, short)
except that Path f must be
for this file system.
ApplicationMaster
for this
application.
ResourceBlacklistRequest
to inform the
ResourceManager
about the blacklist additions and removals
per the ApplicationMaster
.
bytes[:(offset+1)]
in Python syntax.
ApplicationMaster
is
responding.
ApplicationsRequestScope
of applications to filter.
SequenceFile
SequenceFile
SequenceFile
SequenceFile
name
property as
a host:port
.
Reducer
.
StartContainerRequest
to start containers.
name
property as
as comma delimited values.
TaskCompletionEvent.setTaskAttemptId(TaskAttemptID)
instead.
TaskCompletionEvent.setTaskAttemptId(TaskAttemptID)
instead.
name
to the given time duration.
FileContext.setTimes(Path, long, long)
except that Path f must be
for this file system.
Path
's last modified time only to the given
valid time.
ApplicationMaster
.
ApplicationMaster
while
it is running.
LocalResourceType
of the resource to be localized.
FileContext.setVerifyChecksum(boolean, Path)
except that Path f
must be for this file system.
LocalResourceVisibility
of the resource to be
localized.
Path
of the task's temporary output directory
for the map-reduce job.
File.setWritable(boolean)
File#setWritable does not work as expected on Windows.
AbstractCounters.countCounters()
instead
ApplicationMaster
to the
NodeManager
to start a container.ApplicationMaster
provides a list of
StartContainerRequest
s to a NodeManager
to
start Container
s allocated to it using this interface.
StartContainerRequest
sent by
the ApplicationMaster
to the NodeManager
to
start containers.NodeManager
to the
ApplicationMaster
when asked to start an allocated
container.fileName
attribute,
if specified.
ApplicationMaster
requests a NodeManager
to
stop a list of Container
s allocated to it using this
interface.
ApplicationMaster
to the
NodeManager
to stop containers.NodeManager
to the
ApplicationMaster
when asked to stop allocated
containers.ResourceManager.
YARN.
It is a blocking call - it
will not return ApplicationId
until the submitted application is
submitted successfully and accepted by the ResourceManager.
ResourceManager
.ResourceManager
to a client on
application submission.Submitter.runJob(JobConf)
AbstractFileSystem.supportsSymlinks()
Clock
that gives the current time from the system
clock in milliseconds.DNSToSwitchMapping
implementation that reads a 2 column text
file.TaskID
.
TaskAttemptID.TaskAttemptID(String, int, TaskType, int, int)
.
TaskID
.
TaskID.TaskID(String, int, TaskType, int)
TaskID.TaskID(org.apache.hadoop.mapreduce.JobID, TaskType,
int)
JobID
.
JobID
.
InputFormat
for plain text files.InputFormat
for plain text files.OutputFormat
that writes plain text files.OutputFormat
that writes plain text files.Token
is the security entity used by the framework
to verify authenticity of any resource.Mapper
that maps text values into Tool
s.Writable
s.Writable
s.name
property as a host:port
.
URL
represents a serializable URL
.S3FileSystem
.VersionedWritable.readFields(DataInput)
when the
version of an object being read does not match the current implementation
version as returned by VersionedWritable.getVersion()
.FileSystem.createFileSystem(URI, Configuration)
After this constructor is called initialize() is called.
Mapper
which wraps a given one to allow custom
WrappedMapper.Context
implementations.InputStream
.
Reducer
which wraps a given one to allow for custom
WrappedReducer.Context
implementations.DataInput
and DataOutput
.Writable
which is also Comparable
.WritableComparable
s.WritableComparable
implementation.
Serialization
for Writable
s that delegates to
Writable.write(java.io.DataOutput)
and
Writable.readFields(java.io.DataInput)
.out
.
PreparedStatement
.
out
.
CompressedWritable.write(DataOutput)
.
Writable
, String
, primitive type, or an array of
the preceding.
Writable
, String
, primitive type, or an array of
the preceding.
OutputStream
using UTF-8 encoding.
Writer
.
XAttr
is byte[], this class is to
covert byte[] to some kind of string representation or convert back.RMAppAttempt
.ApplicationMaster
.YarnClusterMetrics
represents cluster metrics.Thread.setDefaultUncaughtExceptionHandler(UncaughtExceptionHandler)
In the main entry point.
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |