This is an automated email from the ASF dual-hosted git repository. iwasakims pushed a commit to branch branch-2.10 in repository https://gitbox.apache.org/repos/asf/hadoop.git
The following commit(s) were added to refs/heads/branch-2.10 by this push: new 7b0dbcb Publishing the bits for release 2.10.1 (addendum) 7b0dbcb is described below commit 7b0dbcb65744bda1231457b568e6beccbbbb1cf3 Author: Masatake Iwasaki <iwasak...@apache.org> AuthorDate: Tue Sep 22 11:51:53 2020 +0900 Publishing the bits for release 2.10.1 (addendum) --- .../jdiff/Apache_Hadoop_HDFS_2.10.1.xml | 312 +++++++++++++++++++++ 1 file changed, 312 insertions(+) diff --git a/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.10.1.xml b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.10.1.xml new file mode 100644 index 0000000..daf0eba --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.10.1.xml @@ -0,0 +1,312 @@ +<?xml version="1.0" encoding="iso-8859-1" standalone="no"?> +<!-- Generated by the JDiff Javadoc doclet --> +<!-- (http://www.jdiff.org) --> +<!-- on Mon Sep 14 13:48:50 GMT 2020 --> + +<api + xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance' + xsi:noNamespaceSchemaLocation='api.xsd' + name="Apache Hadoop HDFS 2.10.1" + jdversion="1.0.9"> + +<!-- Command line arguments = -doclet org.apache.hadoop.classification.tools.IncludePublicAnnotationsJDiffDoclet -docletpath /build/source/hadoop-hdfs-project/hadoop-hdfs/target/hadoop-annotations.jar:/build/source/hadoop-hdfs-project/hadoop-hdfs/target/jdiff.jar -verbose -classpath /build/source/hadoop-hdfs-project/hadoop-hdfs/target/classes:/build/source/hadoop-common-project/hadoop-annotations/target/hadoop-annotations-2.10.1.jar:/usr/lib/jvm/java-7-openjdk-amd64/lib/tools.jar:/buil [...] +<package name="org.apache.hadoop.hdfs"> + <doc> + <![CDATA[<p>A distributed implementation of {@link +org.apache.hadoop.fs.FileSystem}. This is loosely modelled after +Google's <a href="http://research.google.com/archive/gfs.html">GFS</a>.</p> + +<p>The most important difference is that unlike GFS, Hadoop DFS files +have strictly one writer at any one time. Bytes are always appended +to the end of the writer's stream. There is no notion of "record appends" +or "mutations" that are then checked or reordered. Writers simply emit +a byte stream. That byte stream is guaranteed to be stored in the +order written.</p>]]> + </doc> +</package> +<package name="org.apache.hadoop.hdfs.net"> +</package> +<package name="org.apache.hadoop.hdfs.protocol"> +</package> +<package name="org.apache.hadoop.hdfs.protocol.datatransfer"> +</package> +<package name="org.apache.hadoop.hdfs.protocol.datatransfer.sasl"> +</package> +<package name="org.apache.hadoop.hdfs.protocolPB"> +</package> +<package name="org.apache.hadoop.hdfs.qjournal.client"> +</package> +<package name="org.apache.hadoop.hdfs.qjournal.protocol"> +</package> +<package name="org.apache.hadoop.hdfs.qjournal.protocolPB"> +</package> +<package name="org.apache.hadoop.hdfs.qjournal.server"> + <!-- start interface org.apache.hadoop.hdfs.qjournal.server.JournalNodeMXBean --> + <interface name="JournalNodeMXBean" abstract="true" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <method name="getJournalsStatus" return="java.lang.String" + abstract="false" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <doc> + <![CDATA[Get status information (e.g., whether formatted) of JournalNode's journals. + + @return A string presenting status for each journal]]> + </doc> + </method> + <doc> + <![CDATA[This is the JMX management interface for JournalNode information]]> + </doc> + </interface> + <!-- end interface org.apache.hadoop.hdfs.qjournal.server.JournalNodeMXBean --> +</package> +<package name="org.apache.hadoop.hdfs.security.token.block"> +</package> +<package name="org.apache.hadoop.hdfs.security.token.delegation"> +</package> +<package name="org.apache.hadoop.hdfs.server.balancer"> +</package> +<package name="org.apache.hadoop.hdfs.server.blockmanagement"> +</package> +<package name="org.apache.hadoop.hdfs.server.common"> +</package> +<package name="org.apache.hadoop.hdfs.server.datanode"> +</package> +<package name="org.apache.hadoop.hdfs.server.datanode.fsdataset"> +</package> +<package name="org.apache.hadoop.hdfs.server.datanode.fsdataset.impl"> +</package> +<package name="org.apache.hadoop.hdfs.server.datanode.metrics"> +</package> +<package name="org.apache.hadoop.hdfs.server.datanode.web"> +</package> +<package name="org.apache.hadoop.hdfs.server.datanode.web.webhdfs"> +</package> +<package name="org.apache.hadoop.hdfs.server.mover"> +</package> +<package name="org.apache.hadoop.hdfs.server.namenode"> + <!-- start interface org.apache.hadoop.hdfs.server.namenode.AuditLogger --> + <interface name="AuditLogger" abstract="true" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <method name="initialize" + abstract="false" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="conf" type="org.apache.hadoop.conf.Configuration"/> + <doc> + <![CDATA[Called during initialization of the logger. + + @param conf The configuration object.]]> + </doc> + </method> + <method name="logAuditEvent" + abstract="false" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="succeeded" type="boolean"/> + <param name="userName" type="java.lang.String"/> + <param name="addr" type="java.net.InetAddress"/> + <param name="cmd" type="java.lang.String"/> + <param name="src" type="java.lang.String"/> + <param name="dst" type="java.lang.String"/> + <param name="stat" type="org.apache.hadoop.fs.FileStatus"/> + <doc> + <![CDATA[Called to log an audit event. + <p> + This method must return as quickly as possible, since it's called + in a critical section of the NameNode's operation. + + @param succeeded Whether authorization succeeded. + @param userName Name of the user executing the request. + @param addr Remote address of the request. + @param cmd The requested command. + @param src Path of affected source file. + @param dst Path of affected destination file (if any). + @param stat File information for operations that change the file's + metadata (permissions, owner, times, etc).]]> + </doc> + </method> + <doc> + <![CDATA[Interface defining an audit logger.]]> + </doc> + </interface> + <!-- end interface org.apache.hadoop.hdfs.server.namenode.AuditLogger --> + <!-- start class org.apache.hadoop.hdfs.server.namenode.HdfsAuditLogger --> + <class name="HdfsAuditLogger" extends="java.lang.Object" + abstract="true" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <implements name="org.apache.hadoop.hdfs.server.namenode.AuditLogger"/> + <constructor name="HdfsAuditLogger" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + </constructor> + <method name="logAuditEvent" + abstract="false" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="succeeded" type="boolean"/> + <param name="userName" type="java.lang.String"/> + <param name="addr" type="java.net.InetAddress"/> + <param name="cmd" type="java.lang.String"/> + <param name="src" type="java.lang.String"/> + <param name="dst" type="java.lang.String"/> + <param name="status" type="org.apache.hadoop.fs.FileStatus"/> + </method> + <method name="logAuditEvent" + abstract="false" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="succeeded" type="boolean"/> + <param name="userName" type="java.lang.String"/> + <param name="addr" type="java.net.InetAddress"/> + <param name="cmd" type="java.lang.String"/> + <param name="src" type="java.lang.String"/> + <param name="dst" type="java.lang.String"/> + <param name="stat" type="org.apache.hadoop.fs.FileStatus"/> + <param name="callerContext" type="org.apache.hadoop.ipc.CallerContext"/> + <param name="ugi" type="org.apache.hadoop.security.UserGroupInformation"/> + <param name="dtSecretManager" type="org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenSecretManager"/> + <doc> + <![CDATA[Same as + {@link #logAuditEvent(boolean, String, InetAddress, String, String, String, + FileStatus)} with additional parameters related to logging delegation token + tracking IDs. + + @param succeeded Whether authorization succeeded. + @param userName Name of the user executing the request. + @param addr Remote address of the request. + @param cmd The requested command. + @param src Path of affected source file. + @param dst Path of affected destination file (if any). + @param stat File information for operations that change the file's metadata + (permissions, owner, times, etc). + @param callerContext Context information of the caller + @param ugi UserGroupInformation of the current user, or null if not logging + token tracking information + @param dtSecretManager The token secret manager, or null if not logging + token tracking information]]> + </doc> + </method> + <method name="logAuditEvent" + abstract="true" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="succeeded" type="boolean"/> + <param name="userName" type="java.lang.String"/> + <param name="addr" type="java.net.InetAddress"/> + <param name="cmd" type="java.lang.String"/> + <param name="src" type="java.lang.String"/> + <param name="dst" type="java.lang.String"/> + <param name="stat" type="org.apache.hadoop.fs.FileStatus"/> + <param name="ugi" type="org.apache.hadoop.security.UserGroupInformation"/> + <param name="dtSecretManager" type="org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenSecretManager"/> + <doc> + <![CDATA[Same as + {@link #logAuditEvent(boolean, String, InetAddress, String, String, + String, FileStatus, CallerContext, UserGroupInformation, + DelegationTokenSecretManager)} without {@link CallerContext} information.]]> + </doc> + </method> + <doc> + <![CDATA[Extension of {@link AuditLogger}.]]> + </doc> + </class> + <!-- end class org.apache.hadoop.hdfs.server.namenode.HdfsAuditLogger --> + <!-- start class org.apache.hadoop.hdfs.server.namenode.INodeAttributeProvider --> + <class name="INodeAttributeProvider" extends="java.lang.Object" + abstract="true" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <constructor name="INodeAttributeProvider" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + </constructor> + <method name="start" + abstract="true" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <doc> + <![CDATA[Initialize the provider. This method is called at NameNode startup + time.]]> + </doc> + </method> + <method name="stop" + abstract="true" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <doc> + <![CDATA[Shutdown the provider. This method is called at NameNode shutdown time.]]> + </doc> + </method> + <method name="getAttributes" return="org.apache.hadoop.hdfs.server.namenode.INodeAttributes" + abstract="false" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="fullPath" type="java.lang.String"/> + <param name="inode" type="org.apache.hadoop.hdfs.server.namenode.INodeAttributes"/> + </method> + <method name="getAttributes" return="org.apache.hadoop.hdfs.server.namenode.INodeAttributes" + abstract="true" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="pathElements" type="java.lang.String[]"/> + <param name="inode" type="org.apache.hadoop.hdfs.server.namenode.INodeAttributes"/> + </method> + <method name="getAttributes" return="org.apache.hadoop.hdfs.server.namenode.INodeAttributes" + abstract="false" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="components" type="byte[][]"/> + <param name="inode" type="org.apache.hadoop.hdfs.server.namenode.INodeAttributes"/> + </method> + <method name="getExternalAccessControlEnforcer" return="org.apache.hadoop.hdfs.server.namenode.INodeAttributeProvider.AccessControlEnforcer" + abstract="false" native="false" synchronized="false" + static="false" final="false" visibility="public" + deprecated="not deprecated"> + <param name="defaultEnforcer" type="org.apache.hadoop.hdfs.server.namenode.INodeAttributeProvider.AccessControlEnforcer"/> + <doc> + <![CDATA[Can be over-ridden by implementations to provide a custom Access Control + Enforcer that can provide an alternate implementation of the + default permission checking logic. + @param defaultEnforcer The Default AccessControlEnforcer + @return The AccessControlEnforcer to use]]> + </doc> + </method> + </class> + <!-- end class org.apache.hadoop.hdfs.server.namenode.INodeAttributeProvider --> +</package> +<package name="org.apache.hadoop.hdfs.server.namenode.ha"> +</package> +<package name="org.apache.hadoop.hdfs.server.namenode.metrics"> +</package> +<package name="org.apache.hadoop.hdfs.server.namenode.snapshot"> +</package> +<package name="org.apache.hadoop.hdfs.server.namenode.top"> +</package> +<package name="org.apache.hadoop.hdfs.server.namenode.top.metrics"> +</package> +<package name="org.apache.hadoop.hdfs.server.namenode.top.window"> +</package> +<package name="org.apache.hadoop.hdfs.server.namenode.web.resources"> +</package> +<package name="org.apache.hadoop.hdfs.server.protocol"> +</package> +<package name="org.apache.hadoop.hdfs.tools"> +</package> +<package name="org.apache.hadoop.hdfs.tools.offlineEditsViewer"> +</package> +<package name="org.apache.hadoop.hdfs.tools.offlineImageViewer"> +</package> +<package name="org.apache.hadoop.hdfs.tools.snapshot"> +</package> +<package name="org.apache.hadoop.hdfs.util"> +</package> +<package name="org.apache.hadoop.hdfs.web"> +</package> +<package name="org.apache.hadoop.hdfs.web.resources"> +</package> + +</api> --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org