HADOOP-17760. Delete hadoop.ssl.enabled and dfs.https.enable from docs and core-default.xml (#3099)
Reviewed-by: Ayush Saxena <ayushsaxena@apache.org>
(cherry picked from commit 9e7c7ad129
)
This commit is contained in:
parent
3e03ecd680
commit
25138c98bf
@ -3077,14 +3077,6 @@
|
|||||||
</description>
|
</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
<property>
|
|
||||||
<name>hadoop.ssl.enabled</name>
|
|
||||||
<value>false</value>
|
|
||||||
<description>
|
|
||||||
Deprecated. Use dfs.http.policy and yarn.http.policy instead.
|
|
||||||
</description>
|
|
||||||
</property>
|
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>hadoop.ssl.enabled.protocols</name>
|
<name>hadoop.ssl.enabled.protocols</name>
|
||||||
<value>TLSv1.2</value>
|
<value>TLSv1.2</value>
|
||||||
|
@ -267,9 +267,8 @@ The following settings allow configuring SSL access to the NameNode web UI (opti
|
|||||||
|
|
||||||
| Parameter | Value | Notes |
|
| Parameter | Value | Notes |
|
||||||
|:-----------------------------|:------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
|:-----------------------------|:------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
||||||
| `dfs.http.policy` | `HTTP_ONLY` or `HTTPS_ONLY` or `HTTP_AND_HTTPS` | `HTTPS_ONLY` turns off http access. This option takes precedence over the deprecated configuration dfs.https.enable and hadoop.ssl.enabled. If using SASL to authenticate data transfer protocol instead of running DataNode as root and using privileged ports, then this property must be set to `HTTPS_ONLY` to guarantee authentication of HTTP servers. (See `dfs.data.transfer.protection`.) |
|
| `dfs.http.policy` | `HTTP_ONLY` or `HTTPS_ONLY` or `HTTP_AND_HTTPS` | `HTTPS_ONLY` turns off http access. If using SASL to authenticate data transfer protocol instead of running DataNode as root and using privileged ports, then this property must be set to `HTTPS_ONLY` to guarantee authentication of HTTP servers. (See `dfs.data.transfer.protection`.) |
|
||||||
| `dfs.namenode.https-address` | `0.0.0.0:9871` | This parameter is used in non-HA mode and without federation. See [HDFS High Availability](../hadoop-hdfs/HDFSHighAvailabilityWithNFS.html#Deployment) and [HDFS Federation](../hadoop-hdfs/Federation.html#Federation_Configuration) for details. |
|
| `dfs.namenode.https-address` | `0.0.0.0:9871` | This parameter is used in non-HA mode and without federation. See [HDFS High Availability](../hadoop-hdfs/HDFSHighAvailabilityWithNFS.html#Deployment) and [HDFS Federation](../hadoop-hdfs/Federation.html#Federation_Configuration) for details. |
|
||||||
| `dfs.https.enable` | `true` | This value is deprecated. `Use dfs.http.policy` |
|
|
||||||
|
|
||||||
### Secondary NameNode
|
### Secondary NameNode
|
||||||
|
|
||||||
|
@ -236,8 +236,6 @@ public void initializeMemberVariables() {
|
|||||||
// - org.apache.hadoop.net.NetUtils
|
// - org.apache.hadoop.net.NetUtils
|
||||||
xmlPropsToSkipCompare
|
xmlPropsToSkipCompare
|
||||||
.add("hadoop.rpc.socket.factory.class.ClientProtocol");
|
.add("hadoop.rpc.socket.factory.class.ClientProtocol");
|
||||||
// - Where is this used?
|
|
||||||
xmlPropsToSkipCompare.add("hadoop.ssl.enabled");
|
|
||||||
|
|
||||||
// Keys with no corresponding variable
|
// Keys with no corresponding variable
|
||||||
// - org.apache.hadoop.io.compress.bzip2.Bzip2Factory
|
// - org.apache.hadoop.io.compress.bzip2.Bzip2Factory
|
||||||
|
@ -4559,7 +4559,6 @@
|
|||||||
"hadoop.hdfs.configuration.version" : "1",
|
"hadoop.hdfs.configuration.version" : "1",
|
||||||
"dfs.datanode.balance.bandwidthPerSec" : "1048576",
|
"dfs.datanode.balance.bandwidthPerSec" : "1048576",
|
||||||
"mapreduce.reduce.shuffle.connect.timeout" : "180000",
|
"mapreduce.reduce.shuffle.connect.timeout" : "180000",
|
||||||
"hadoop.ssl.enabled" : "false",
|
|
||||||
"dfs.journalnode.rpc-address" : "0.0.0.0:8485",
|
"dfs.journalnode.rpc-address" : "0.0.0.0:8485",
|
||||||
"yarn.nodemanager.aux-services" : "mapreduce.shuffle",
|
"yarn.nodemanager.aux-services" : "mapreduce.shuffle",
|
||||||
"mapreduce.job.counters.max" : "120",
|
"mapreduce.job.counters.max" : "120",
|
||||||
@ -9626,7 +9625,6 @@
|
|||||||
"hadoop.hdfs.configuration.version" : "1",
|
"hadoop.hdfs.configuration.version" : "1",
|
||||||
"dfs.datanode.balance.bandwidthPerSec" : "1048576",
|
"dfs.datanode.balance.bandwidthPerSec" : "1048576",
|
||||||
"mapreduce.reduce.shuffle.connect.timeout" : "180000",
|
"mapreduce.reduce.shuffle.connect.timeout" : "180000",
|
||||||
"hadoop.ssl.enabled" : "false",
|
|
||||||
"dfs.journalnode.rpc-address" : "0.0.0.0:8485",
|
"dfs.journalnode.rpc-address" : "0.0.0.0:8485",
|
||||||
"yarn.nodemanager.aux-services" : "mapreduce.shuffle",
|
"yarn.nodemanager.aux-services" : "mapreduce.shuffle",
|
||||||
"mapreduce.job.counters.max" : "120",
|
"mapreduce.job.counters.max" : "120",
|
||||||
|
Loading…
Reference in New Issue
Block a user