Skip to content

Commit 44a5cba

Browse files
committed
HADOOP-19407: Update 3.4.1 docs landing page to highlight changes shipped in the release.
Closes #7338 Signed-off-by: Shilun Fan <[email protected]>
1 parent 1ba30d6 commit 44a5cba

File tree

1 file changed

+40
-122
lines changed

1 file changed

+40
-122
lines changed

hadoop-project/src/site/markdown/index.md.vm

+40-122
Original file line numberDiff line numberDiff line change
@@ -23,157 +23,75 @@ Overview of Changes
2323
Users are encouraged to read the full set of release notes.
2424
This page provides an overview of the major changes.
2525

26-
S3A: Upgrade AWS SDK to V2
26+
Bulk Delete API
2727
----------------------------------------
2828

29-
[HADOOP-18073](https://issues.apache.org/jira/browse/HADOOP-18073) S3A: Upgrade AWS SDK to V2
29+
[HADOOP-18679](https://issues.apache.org/jira/browse/HADOOP-18679) Bulk Delete API.
3030

31-
This release upgrade Hadoop's AWS connector S3A from AWS SDK for Java V1 to AWS SDK for Java V2.
32-
This is a significant change which offers a number of new features including the ability to work with Amazon S3 Express One Zone Storage - the new high performance, single AZ storage class.
31+
This release provides an API to perform bulk delete of files/objects
32+
in an object store or filesystem.
3333

34-
HDFS DataNode Split one FsDatasetImpl lock to volume grain locks
35-
----------------------------------------
36-
37-
[HDFS-15382](https://issues.apache.org/jira/browse/HDFS-15382) Split one FsDatasetImpl lock to volume grain locks.
38-
39-
Throughput is one of the core performance evaluation for DataNode instance.
40-
However, it does not reach the best performance especially for Federation deploy all the time although there are different improvement,
41-
because of the global coarse-grain lock.
42-
These series issues (include [HDFS-16534](https://issues.apache.org/jira/browse/HDFS-16534), [HDFS-16511](https://issues.apache.org/jira/browse/HDFS-16511), [HDFS-15382](https://issues.apache.org/jira/browse/HDFS-15382) and [HDFS-16429](https://issues.apache.org/jira/browse/HDFS-16429).)
43-
try to split the global coarse-grain lock to fine-grain lock which is double level lock for blockpool and volume,
44-
to improve the throughput and avoid lock impacts between blockpools and volumes.
45-
46-
YARN Federation improvements
47-
----------------------------------------
48-
49-
[YARN-5597](https://issues.apache.org/jira/browse/YARN-5597) YARN Federation improvements.
50-
51-
We have enhanced the YARN Federation functionality for improved usability. The enhanced features are as follows:
52-
1. YARN Router now boasts a full implementation of all interfaces including the ApplicationClientProtocol, ResourceManagerAdministrationProtocol, and RMWebServiceProtocol.
53-
2. YARN Router support for application cleanup and automatic offline mechanisms for subCluster.
54-
3. Code improvements were undertaken for the Router and AMRMProxy, along with enhancements to previously pending functionalities.
55-
4. Audit logs and Metrics for Router received upgrades.
56-
5. A boost in cluster security features was achieved, with the inclusion of Kerberos support.
57-
6. The page function of the router has been enhanced.
58-
7. A set of commands has been added to the Router side for operating on SubClusters and Policies.
59-
60-
YARN Capacity Scheduler improvements
61-
----------------------------------------
62-
63-
[YARN-10496](https://issues.apache.org/jira/browse/YARN-10496) Support Flexible Auto Queue Creation in Capacity Scheduler
64-
65-
Capacity Scheduler resource distribution mode was extended with a new allocation mode called weight mode.
66-
Defining queue capacities with weights allows the users to use the newly added flexible queue auto creation mode.
67-
Flexible mode now supports the dynamic creation of both **parent queues** and **leaf queues**, enabling the creation of
68-
complex queue hierarchies application submission time.
69-
70-
[YARN-10888](https://issues.apache.org/jira/browse/YARN-10888) New capacity modes for Capacity Scheduler
71-
72-
Capacity Scheduler's resource distribution was completely refactored to be more flexible and extensible. There is a new concept
73-
called Capacity Vectors, which allows the users to mix various resource types in the hierarchy, and also in a single queue. With
74-
this optionally enabled feature it is now possible to define different resources with different units, like memory with GBs, vcores with
75-
percentage values, and GPUs/FPGAs with weights, all in the same queue.
76-
77-
[YARN-10889](https://issues.apache.org/jira/browse/YARN-10889) Queue Creation in Capacity Scheduler - Various improvements
34+
New binary distribution
35+
-----------------------
7836

79-
In addition to the two new features above, there were a number of commits for improvements and bug fixes in Capacity Scheduler.
37+
[HADOOP-19083](https://issues.apache.org/jira/browse/HADOOP-19083) provide hadoop binary tarball without aws v2 sdk
8038

81-
HDFS RBF: Code Enhancements, New Features, and Bug Fixes
82-
----------------------------------------
83-
84-
The HDFS RBF functionality has undergone significant enhancements, encompassing over 200 commits for feature
85-
improvements, new functionalities, and bug fixes.
86-
Important features and improvements are as follows:
87-
88-
**Feature**
89-
90-
[HDFS-15294](https://issues.apache.org/jira/browse/HDFS-15294) HDFS Federation balance tool introduces one tool to balance data across different namespace.
39+
Hadoop has added a new variant of the binary distribution tarball, labeled with "lean" in the file
40+
name. This tarball excludes the full AWS SDK v2 bundle, resulting in approximately 50% reduction in
41+
file size.
9142

92-
[HDFS-13522](https://issues.apache.org/jira/browse/HDFS-13522), [HDFS-16767](https://issues.apache.org/jira/browse/HDFS-16767) Support observer node from Router-Based Federation.
43+
S3A improvements
44+
----------------
9345

9446
**Improvement**
9547

96-
[HADOOP-13144](https://issues.apache.org/jira/browse/HADOOP-13144), [HDFS-13274](https://issues.apache.org/jira/browse/HDFS-13274), [HDFS-15757](https://issues.apache.org/jira/browse/HDFS-15757)
97-
98-
These tickets have enhanced IPC throughput between Router and NameNode via multiple connections per user, and optimized connection management.
99-
100-
[HDFS-14090](https://issues.apache.org/jira/browse/HDFS-14090) RBF: Improved isolation for downstream name nodes. {Static}
101-
102-
Router supports assignment of the dedicated number of RPC handlers to achieve isolation for all downstream nameservices
103-
it is configured to proxy. Since large or busy clusters may have relatively higher RPC traffic to the namenode compared to other clusters namenodes,
104-
this feature if enabled allows admins to configure higher number of RPC handlers for busy clusters.
48+
[HADOOP-18886](https://issues.apache.org/jira/browse/HADOOP-18886) S3A: AWS SDK V2 Migration: stabilization and S3Express
10549

106-
[HDFS-17128](https://issues.apache.org/jira/browse/HDFS-17128) RBF: SQLDelegationTokenSecretManager should use version of tokens updated by other routers.
50+
This release completes stabilization efforts on the AWS SDK v2 migration and support of Amazon S3
51+
Express One Zone storage. S3 Select is no longer supported.
10752

108-
The SQLDelegationTokenSecretManager enhances performance by maintaining processed tokens in memory. However, there is
109-
a potential issue of router cache inconsistency due to token loading and renewal. This issue has been addressed by the
110-
resolution of HDFS-17128.
53+
[HADOOP-18993](https://issues.apache.org/jira/browse/HADOOP-18993) S3A: Add option fs.s3a.classloader.isolation (#6301)
11154

112-
[HDFS-17148](https://issues.apache.org/jira/browse/HDFS-17148) RBF: SQLDelegationTokenSecretManager must cleanup expired tokens in SQL.
55+
This introduces configuration property `fs.s3a.classloader.isolation`, which defaults to `true`.
56+
Set to `false` to disable S3A classloader isolation, which can be useful for installing custom
57+
credential providers in user-provided jars.
11358

114-
SQLDelegationTokenSecretManager, while fetching and temporarily storing tokens from SQL in a memory cache with a short TTL,
115-
faces an issue where expired tokens are not efficiently cleaned up, leading to a buildup of expired tokens in the SQL database.
116-
This issue has been addressed by the resolution of HDFS-17148.
59+
[HADOOP-19047](https://issues.apache.org/jira/browse/HADOOP-19047) Support InMemory Tracking Of S3A Magic Commits
11760

118-
**Others**
61+
The S3A magic committer now supports configuration property
62+
`fs.s3a.committer.magic.track.commits.in.memory.enabled`. Set this to `true` to track commits in
63+
memory instead of on the file system, which reduces the number of remote calls.
11964

120-
Other changes to HDFS RBF include WebUI, command line, and other improvements. Please refer to the release document.
65+
[HADOOP-19161](https://issues.apache.org/jira/browse/HADOOP-19161) S3A: option “fs.s3a.performance.flags” to take list of performance flags
12166

122-
HDFS EC: Code Enhancements and Bug Fixes
123-
----------------------------------------
124-
125-
HDFS EC has made code improvements and fixed some bugs.
67+
S3A now supports configuration property `fs.s3a.performance.flag` for controlling activation of
68+
multiple performance optimizations. Refer to the S3A performance documentation for details.
12669

127-
Important improvements and bugs are as follows:
70+
ABFS improvements
71+
-----------------
12872

12973
**Improvement**
13074

131-
[HDFS-16613](https://issues.apache.org/jira/browse/HDFS-16613) EC: Improve performance of decommissioning dn with many ec blocks.
75+
[HADOOP-18516](https://issues.apache.org/jira/browse/HADOOP-18516) [ABFS]: Support fixed SAS token config in addition to Custom SASTokenProvider Implementation
13276

133-
In a hdfs cluster with a lot of EC blocks, decommission a dn is very slow. The reason is unlike replication blocks can be replicated
134-
from any dn which has the same block replication, the ec block have to be replicated from the decommissioning dn.
135-
The configurations `dfs.namenode.replication.max-streams` and `dfs.namenode.replication.max-streams-hard-limit` will limit
136-
the replication speed, but increase these configurations will create risk to the whole cluster's network. So it should add a new
137-
configuration to limit the decommissioning dn, distinguished from the cluster wide max-streams limit.
77+
ABFS now supports authentication via a fixed Shared Access Signature token. Refer to ABFS
78+
documentation of configuration property `fs.azure.sas.fixed.token` for details.
13879

139-
[HDFS-16663](https://issues.apache.org/jira/browse/HDFS-16663) EC: Allow block reconstruction pending timeout refreshable to increase decommission performance.
80+
[HADOOP-19089](https://issues.apache.org/jira/browse/HADOOP-19089) [ABFS] Reverting Back Support of setXAttr() and getXAttr() on root path
14081

141-
In [HDFS-16613](https://issues.apache.org/jira/browse/HDFS-16613), increase the value of `dfs.namenode.replication.max-streams-hard-limit` would maximize the IO
142-
performance of the decommissioning DN, which has a lot of EC blocks. Besides this, we also need to decrease the value of
143-
`dfs.namenode.reconstruction.pending.timeout-sec`, default is 5 minutes, to shorten the interval time for checking
144-
pendingReconstructions. Or the decommissioning node would be idle to wait for copy tasks in most of this 5 minutes.
145-
In decommission progress, we may need to reconfigure these 2 parameters several times. In [HDFS-14560](https://issues.apache.org/jira/browse/HDFS-14560), the
146-
`dfs.namenode.replication.max-streams-hard-limit` can already be reconfigured dynamically without namenode restart. And
147-
the `dfs.namenode.reconstruction.pending.timeout-sec` parameter also need to be reconfigured dynamically.
148-
149-
**Bug**
82+
[HADOOP-18869](https://issues.apache.org/jira/browse/HADOOP-18869) previously implemented support for xattrs on the root path in the 3.4.0 release. Support for this has been removed in 3.4.1 to prevent the need for calling container APIs.
15083

151-
[HDFS-16456](https://issues.apache.org/jira/browse/HDFS-16456) EC: Decommission a rack with only on dn will fail when the rack number is equal with replication.
84+
[HADOOP-19178](https://issues.apache.org/jira/browse/HADOOP-19178) WASB Driver Deprecation and eventual removal
15285

153-
In below scenario, decommission will fail by `TOO_MANY_NODES_ON_RACK` reason:
154-
- Enable EC policy, such as RS-6-3-1024k.
155-
- The rack number in this cluster is equal with or less than the replication number(9)
156-
- A rack only has one DN, and decommission this DN.
157-
This issue has been addressed by the resolution of HDFS-16456.
86+
This release announces deprecation of the WASB file system in favor of ABFS. Refer to ABFS
87+
documentation for additional guidance.
15888

159-
[HDFS-17094](https://issues.apache.org/jira/browse/HDFS-17094) EC: Fix bug in block recovery when there are stale datanodes.
160-
During block recovery, the `RecoveryTaskStriped` in the datanode expects a one-to-one correspondence between
161-
`rBlock.getLocations()` and `rBlock.getBlockIndices()`. However, if there are stale locations during a NameNode heartbeat,
162-
this correspondence may be disrupted. Specifically, although there are no stale locations in `recoveryLocations`, the block indices
163-
array remains complete. This discrepancy causes `BlockRecoveryWorker.RecoveryTaskStriped#recover` to generate an incorrect
164-
internal block ID, leading to a failure in the recovery process as the corresponding datanode cannot locate the replica.
165-
This issue has been addressed by the resolution of HDFS-17094.
166-
167-
[HDFS-17284](https://issues.apache.org/jira/browse/HDFS-17284). EC: Fix int overflow in calculating numEcReplicatedTasks and numReplicationTasks during block recovery.
168-
Due to an integer overflow in the calculation of numReplicationTasks or numEcReplicatedTasks, the NameNode's configuration
169-
parameter `dfs.namenode.replication.max-streams-hard-limit` failed to take effect. This led to an excessive number of tasks
170-
being sent to the DataNodes, consequently occupying too much of their memory.
171-
172-
This issue has been addressed by the resolution of HDFS-17284.
89+
**Bug**
17390

174-
**Others**
91+
[HADOOP-18542](https://issues.apache.org/jira/browse/HADOOP-18542) Azure Token provider requires tenant and client IDs despite being optional
17592

176-
Other improvements and fixes for HDFS EC, Please refer to the release document.
93+
It is no longer necessary to specify a tenant and client ID in configuration for MSI authentication
94+
when running in an Azure instance.
17795

17896
Transitive CVE fixes
17997
--------------------

0 commit comments

Comments
 (0)