HDFS-6780. Batch the encryption zones listing API. (wang)
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/fs-encryption@1615189 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
3d9e39e51f
commit
70c99278a9
@ -72,6 +72,8 @@ fs-encryption (Unreleased)
|
|||||||
|
|
||||||
HDFS-6692. Add more HDFS encryption tests. (wang)
|
HDFS-6692. Add more HDFS encryption tests. (wang)
|
||||||
|
|
||||||
|
HDFS-6780. Batch the encryption zones listing API. (wang)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
@ -150,6 +150,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.EncryptionZoneIterator;
|
||||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsBlocksMetadata;
|
import org.apache.hadoop.hdfs.protocol.HdfsBlocksMetadata;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||||
@ -2857,13 +2858,10 @@ public void createEncryptionZone(String src, String keyName)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public List<EncryptionZone> listEncryptionZones() throws IOException {
|
public RemoteIterator<EncryptionZone> listEncryptionZones()
|
||||||
|
throws IOException {
|
||||||
checkOpen();
|
checkOpen();
|
||||||
try {
|
return new EncryptionZoneIterator(namenode);
|
||||||
return namenode.listEncryptionZones();
|
|
||||||
} catch (RemoteException re) {
|
|
||||||
throw re.unwrapRemoteException(AccessControlException.class);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
public void setXAttr(String src, String name, byte[] value,
|
public void setXAttr(String src, String name, byte[] value,
|
||||||
|
@ -566,6 +566,8 @@ public class DFSConfigKeys extends CommonConfigurationKeys {
|
|||||||
public static final String DFS_TRUSTEDCHANNEL_RESOLVER_CLASS = "dfs.trustedchannel.resolver.class";
|
public static final String DFS_TRUSTEDCHANNEL_RESOLVER_CLASS = "dfs.trustedchannel.resolver.class";
|
||||||
public static final String DFS_DATA_TRANSFER_PROTECTION_KEY = "dfs.data.transfer.protection";
|
public static final String DFS_DATA_TRANSFER_PROTECTION_KEY = "dfs.data.transfer.protection";
|
||||||
public static final String DFS_DATA_TRANSFER_SASL_PROPS_RESOLVER_CLASS_KEY = "dfs.data.transfer.saslproperties.resolver.class";
|
public static final String DFS_DATA_TRANSFER_SASL_PROPS_RESOLVER_CLASS_KEY = "dfs.data.transfer.saslproperties.resolver.class";
|
||||||
|
public static final int DFS_NAMENODE_LIST_ENCRYPTION_ZONES_NUM_RESPONSES_DEFAULT = 100;
|
||||||
|
public static final String DFS_NAMENODE_LIST_ENCRYPTION_ZONES_NUM_RESPONSES = "dfs.namenode.list.encryption.zones.num.responses";
|
||||||
|
|
||||||
// Journal-node related configs. These are read on the JN side.
|
// Journal-node related configs. These are read on the JN side.
|
||||||
public static final String DFS_JOURNALNODE_EDITS_DIR_KEY = "dfs.journalnode.edits.dir";
|
public static final String DFS_JOURNALNODE_EDITS_DIR_KEY = "dfs.journalnode.edits.dir";
|
||||||
|
@ -1805,7 +1805,8 @@ public void createEncryptionZone(Path path, String keyName)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* HDFS only */
|
/* HDFS only */
|
||||||
public List<EncryptionZone> listEncryptionZones() throws IOException {
|
public RemoteIterator<EncryptionZone> listEncryptionZones()
|
||||||
|
throws IOException {
|
||||||
return dfs.listEncryptionZones();
|
return dfs.listEncryptionZones();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -248,16 +248,17 @@ public void createEncryptionZone(Path path, String keyName)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Return a list of all {@link EncryptionZone}s in the HDFS hierarchy which
|
* Returns a RemoteIterator which can be used to list the encryption zones
|
||||||
* are visible to the caller. If the caller is an HDFS superuser,
|
* in HDFS. For large numbers of encryption zones, the iterator will fetch
|
||||||
* then the key name of each encryption zone will also be provided.
|
* the list of zones in a number of small batches.
|
||||||
*
|
* <p/>
|
||||||
* @throws IOException if there was a general IO exception
|
* Since the list is fetched in batches, it does not represent a
|
||||||
*
|
* consistent snapshot of the entire list of encryption zones.
|
||||||
* @return List<EncryptionZone> the list of Encryption Zones that the caller has
|
* <p/>
|
||||||
* access to.
|
* This method can only be called by HDFS superusers.
|
||||||
*/
|
*/
|
||||||
public List<EncryptionZone> listEncryptionZones() throws IOException {
|
public RemoteIterator<EncryptionZone> listEncryptionZones()
|
||||||
|
throws IOException {
|
||||||
return dfs.listEncryptionZones();
|
return dfs.listEncryptionZones();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1275,16 +1275,15 @@ public void createEncryptionZone(String src, String keyName)
|
|||||||
throws IOException;
|
throws IOException;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Return a list of all {@EncryptionZone}s in the HDFS hierarchy which are
|
* Used to implement cursor-based batched listing of {@EncryptionZone}s.
|
||||||
* visible to the caller. If the caller is the HDFS admin, then the returned
|
*
|
||||||
* EncryptionZone instances will have the key id field filled in. If the
|
* @param prevId ID of the last item in the previous batch. If there is no
|
||||||
* caller is not the HDFS admin, then the EncryptionZone instances will only
|
* previous batch, a negative value can be used.
|
||||||
* have the path field filled in and only those zones that are visible to the
|
* @return Batch of encryption zones.
|
||||||
* user are returned.
|
|
||||||
*/
|
*/
|
||||||
@Idempotent
|
@Idempotent
|
||||||
public List<EncryptionZone> listEncryptionZones()
|
public BatchedEntries<EncryptionZoneWithId> listEncryptionZones(
|
||||||
throws IOException;
|
long prevId) throws IOException;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Set xattr of a file or directory.
|
* Set xattr of a file or directory.
|
||||||
|
@ -0,0 +1,51 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.hdfs.protocol;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.fs.RemoteIterator;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* EncryptionZoneIterator is a remote iterator that iterates over encryption
|
||||||
|
* zones. It supports retrying in case of namenode failover.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Evolving
|
||||||
|
public class EncryptionZoneIterator implements RemoteIterator<EncryptionZone> {
|
||||||
|
|
||||||
|
private final EncryptionZoneWithIdIterator iterator;
|
||||||
|
|
||||||
|
public EncryptionZoneIterator(ClientProtocol namenode) {
|
||||||
|
iterator = new EncryptionZoneWithIdIterator(namenode);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean hasNext() throws IOException {
|
||||||
|
return iterator.hasNext();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public EncryptionZone next() throws IOException {
|
||||||
|
EncryptionZoneWithId ezwi = iterator.next();
|
||||||
|
return ezwi.toEncryptionZone();
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,64 @@
|
|||||||
|
package org.apache.hadoop.hdfs.protocol;
|
||||||
|
|
||||||
|
import org.apache.commons.lang.builder.HashCodeBuilder;
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Internal class similar to an {@link EncryptionZone} which also holds a
|
||||||
|
* unique id. Used to implement batched listing of encryption zones.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
public class EncryptionZoneWithId extends EncryptionZone {
|
||||||
|
|
||||||
|
final long id;
|
||||||
|
|
||||||
|
public EncryptionZoneWithId(String path, String keyName, long id) {
|
||||||
|
super(path, keyName);
|
||||||
|
this.id = id;
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getId() {
|
||||||
|
return id;
|
||||||
|
}
|
||||||
|
|
||||||
|
EncryptionZone toEncryptionZone() {
|
||||||
|
return new EncryptionZone(getPath(), getKeyName());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int hashCode() {
|
||||||
|
return new HashCodeBuilder(17, 29)
|
||||||
|
.append(super.hashCode())
|
||||||
|
.append(id)
|
||||||
|
.toHashCode();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean equals(Object o) {
|
||||||
|
if (this == o) {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
if (o == null || getClass() != o.getClass()) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
if (!super.equals(o)) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
EncryptionZoneWithId that = (EncryptionZoneWithId) o;
|
||||||
|
|
||||||
|
if (id != that.id) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String toString() {
|
||||||
|
return "EncryptionZoneWithId [" +
|
||||||
|
"id=" + id +
|
||||||
|
", " + super.toString() +
|
||||||
|
']';
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,53 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.hdfs.protocol;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.fs.BatchedRemoteIterator;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Used on the client-side to iterate over the list of encryption zones
|
||||||
|
* stored on the namenode.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Evolving
|
||||||
|
public class EncryptionZoneWithIdIterator
|
||||||
|
extends BatchedRemoteIterator<Long, EncryptionZoneWithId> {
|
||||||
|
|
||||||
|
private final ClientProtocol namenode;
|
||||||
|
|
||||||
|
EncryptionZoneWithIdIterator(ClientProtocol namenode) {
|
||||||
|
super(Long.valueOf(0));
|
||||||
|
this.namenode = namenode;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public BatchedEntries<EncryptionZoneWithId> makeRequest(Long prevId)
|
||||||
|
throws IOException {
|
||||||
|
return namenode.listEncryptionZones(prevId);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Long elementToPrevKey(EncryptionZoneWithId entry) {
|
||||||
|
return entry.getId();
|
||||||
|
}
|
||||||
|
}
|
@ -32,6 +32,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
||||||
import org.apache.hadoop.hdfs.protocol.CorruptFileBlocks;
|
import org.apache.hadoop.hdfs.protocol.CorruptFileBlocks;
|
||||||
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.EncryptionZoneWithId;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsFileStatus;
|
import org.apache.hadoop.hdfs.protocol.HdfsFileStatus;
|
||||||
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
||||||
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
|
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
|
||||||
@ -1317,7 +1318,15 @@ public ListEncryptionZonesResponseProto listEncryptionZones(
|
|||||||
RpcController controller, ListEncryptionZonesRequestProto req)
|
RpcController controller, ListEncryptionZonesRequestProto req)
|
||||||
throws ServiceException {
|
throws ServiceException {
|
||||||
try {
|
try {
|
||||||
return PBHelper.convertListEZResponse(server.listEncryptionZones());
|
BatchedEntries<EncryptionZoneWithId> entries = server
|
||||||
|
.listEncryptionZones(req.getId());
|
||||||
|
ListEncryptionZonesResponseProto.Builder builder =
|
||||||
|
ListEncryptionZonesResponseProto.newBuilder();
|
||||||
|
builder.setHasMore(entries.hasMore());
|
||||||
|
for (int i=0; i<entries.size(); i++) {
|
||||||
|
builder.addZones(PBHelper.convert(entries.get(i)));
|
||||||
|
}
|
||||||
|
return builder.build();
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new ServiceException(e);
|
throw new ServiceException(e);
|
||||||
}
|
}
|
||||||
|
@ -24,6 +24,7 @@
|
|||||||
import java.util.EnumSet;
|
import java.util.EnumSet;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
|
import com.google.common.collect.Lists;
|
||||||
import org.apache.hadoop.classification.InterfaceAudience;
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.classification.InterfaceStability;
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.crypto.CipherSuite;
|
import org.apache.hadoop.crypto.CipherSuite;
|
||||||
@ -52,7 +53,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.DatanodeID;
|
import org.apache.hadoop.hdfs.protocol.DatanodeID;
|
||||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZoneWithId;
|
||||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants.DatanodeReportType;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants.DatanodeReportType;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants.RollingUpgradeAction;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants.RollingUpgradeAction;
|
||||||
@ -146,6 +147,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.SetTimesRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.SetTimesRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.UpdateBlockForPipelineRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.UpdateBlockForPipelineRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.UpdatePipelineRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.UpdatePipelineRequestProto;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos.CreateEncryptionZoneRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos.CreateEncryptionZoneRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos.ListEncryptionZonesRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos.ListEncryptionZonesRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto;
|
||||||
@ -174,6 +176,11 @@
|
|||||||
import com.google.protobuf.ByteString;
|
import com.google.protobuf.ByteString;
|
||||||
import com.google.protobuf.ServiceException;
|
import com.google.protobuf.ServiceException;
|
||||||
|
|
||||||
|
|
||||||
|
import static org.apache.hadoop.fs.BatchedRemoteIterator.BatchedListEntries;
|
||||||
|
import static org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos
|
||||||
|
.EncryptionZoneWithIdProto;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This class forwards NN's ClientProtocol calls as RPC calls to the NN server
|
* This class forwards NN's ClientProtocol calls as RPC calls to the NN server
|
||||||
* while translating from the parameter types used in ClientProtocol to the
|
* while translating from the parameter types used in ClientProtocol to the
|
||||||
@ -1317,11 +1324,22 @@ public void createEncryptionZone(String src, String keyName)
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public List<EncryptionZone> listEncryptionZones() throws IOException {
|
public BatchedEntries<EncryptionZoneWithId> listEncryptionZones(long id)
|
||||||
|
throws IOException {
|
||||||
final ListEncryptionZonesRequestProto req =
|
final ListEncryptionZonesRequestProto req =
|
||||||
ListEncryptionZonesRequestProto.newBuilder().build();
|
ListEncryptionZonesRequestProto.newBuilder()
|
||||||
|
.setId(id)
|
||||||
|
.build();
|
||||||
try {
|
try {
|
||||||
return PBHelper.convert(rpcProxy.listEncryptionZones(null, req));
|
EncryptionZonesProtos.ListEncryptionZonesResponseProto response =
|
||||||
|
rpcProxy.listEncryptionZones(null, req);
|
||||||
|
List<EncryptionZoneWithId> elements =
|
||||||
|
Lists.newArrayListWithCapacity(response.getZonesCount());
|
||||||
|
for (EncryptionZoneWithIdProto p : response.getZonesList()) {
|
||||||
|
elements.add(PBHelper.convert(p));
|
||||||
|
}
|
||||||
|
return new BatchedListEntries<EncryptionZoneWithId>(elements,
|
||||||
|
response.getHasMore());
|
||||||
} catch (ServiceException e) {
|
} catch (ServiceException e) {
|
||||||
throw ProtobufHelper.getRemoteException(e);
|
throw ProtobufHelper.getRemoteException(e);
|
||||||
}
|
}
|
||||||
|
@ -18,6 +18,8 @@
|
|||||||
package org.apache.hadoop.hdfs.protocolPB;
|
package org.apache.hadoop.hdfs.protocolPB;
|
||||||
|
|
||||||
import static com.google.common.base.Preconditions.checkNotNull;
|
import static com.google.common.base.Preconditions.checkNotNull;
|
||||||
|
import static org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos
|
||||||
|
.EncryptionZoneWithIdProto;
|
||||||
|
|
||||||
import java.io.EOFException;
|
import java.io.EOFException;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
@ -59,7 +61,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo.AdminStates;
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo.AdminStates;
|
||||||
import org.apache.hadoop.hdfs.protocol.DatanodeLocalInfo;
|
import org.apache.hadoop.hdfs.protocol.DatanodeLocalInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZoneWithId;
|
||||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||||
import org.apache.hadoop.fs.FileEncryptionInfo;
|
import org.apache.hadoop.fs.FileEncryptionInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.FsAclPermission;
|
import org.apache.hadoop.hdfs.protocol.FsAclPermission;
|
||||||
@ -111,8 +113,6 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.NNHAStatusHeartbeatProto;
|
import org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.NNHAStatusHeartbeatProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto;
|
import org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto;
|
import org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos.ListEncryptionZonesResponseProto;
|
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.EncryptionZonesProtos.EncryptionZoneProto;
|
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos;
|
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockKeyProto;
|
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockKeyProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto;
|
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto;
|
||||||
@ -2265,45 +2265,6 @@ public static List<XAttr> convertXAttrs(List<XAttrProto> xAttrSpec) {
|
|||||||
return xAttrs;
|
return xAttrs;
|
||||||
}
|
}
|
||||||
|
|
||||||
public static List<EncryptionZone> convert(ListEncryptionZonesResponseProto a) {
|
|
||||||
final List<EncryptionZoneProto> ezs = a.getPathsAndKeysList();
|
|
||||||
return convertEZ(ezs);
|
|
||||||
}
|
|
||||||
|
|
||||||
public static ListEncryptionZonesResponseProto convertListEZResponse(
|
|
||||||
List<EncryptionZone> ezs) {
|
|
||||||
final ListEncryptionZonesResponseProto.Builder builder =
|
|
||||||
ListEncryptionZonesResponseProto.newBuilder();
|
|
||||||
builder.addAllPathsAndKeys(convertEZProto(ezs));
|
|
||||||
return builder.build();
|
|
||||||
}
|
|
||||||
|
|
||||||
public static List<EncryptionZoneProto> convertEZProto(
|
|
||||||
List<EncryptionZone> ezs) {
|
|
||||||
final ArrayList<EncryptionZoneProto> ret =
|
|
||||||
Lists.newArrayListWithCapacity(ezs.size());
|
|
||||||
for (EncryptionZone a : ezs) {
|
|
||||||
final EncryptionZoneProto.Builder builder =
|
|
||||||
EncryptionZoneProto.newBuilder();
|
|
||||||
builder.setPath(a.getPath());
|
|
||||||
builder.setKeyName(a.getKeyName());
|
|
||||||
ret.add(builder.build());
|
|
||||||
}
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
public static List<EncryptionZone> convertEZ(
|
|
||||||
List<EncryptionZoneProto> ezs) {
|
|
||||||
final ArrayList<EncryptionZone> ret =
|
|
||||||
Lists.newArrayListWithCapacity(ezs.size());
|
|
||||||
for (EncryptionZoneProto a : ezs) {
|
|
||||||
final EncryptionZone ez =
|
|
||||||
new EncryptionZone(a.getPath(), a.getKeyName());
|
|
||||||
ret.add(ez);
|
|
||||||
}
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
public static List<XAttr> convert(GetXAttrsResponseProto a) {
|
public static List<XAttr> convert(GetXAttrsResponseProto a) {
|
||||||
List<XAttrProto> xAttrs = a.getXAttrsList();
|
List<XAttrProto> xAttrs = a.getXAttrsList();
|
||||||
return convertXAttrs(xAttrs);
|
return convertXAttrs(xAttrs);
|
||||||
@ -2334,6 +2295,18 @@ public static ListXAttrsResponseProto convertListXAttrsResponse(
|
|||||||
return builder.build();
|
return builder.build();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static EncryptionZoneWithIdProto convert(EncryptionZoneWithId zone) {
|
||||||
|
return EncryptionZoneWithIdProto.newBuilder()
|
||||||
|
.setId(zone.getId())
|
||||||
|
.setKeyName(zone.getKeyName())
|
||||||
|
.setPath(zone.getPath()).build();
|
||||||
|
}
|
||||||
|
|
||||||
|
public static EncryptionZoneWithId convert(EncryptionZoneWithIdProto proto) {
|
||||||
|
return new EncryptionZoneWithId(proto.getPath(), proto.getKeyName(),
|
||||||
|
proto.getId());
|
||||||
|
}
|
||||||
|
|
||||||
public static ShortCircuitShmSlotProto convert(SlotId slotId) {
|
public static ShortCircuitShmSlotProto convert(SlotId slotId) {
|
||||||
return ShortCircuitShmSlotProto.newBuilder().
|
return ShortCircuitShmSlotProto.newBuilder().
|
||||||
setShmId(convert(slotId.getShmId())).
|
setShmId(convert(slotId.getShmId())).
|
||||||
|
@ -2,22 +2,25 @@
|
|||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.EnumSet;
|
import java.util.EnumSet;
|
||||||
import java.util.HashMap;
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.NavigableMap;
|
||||||
|
import java.util.TreeMap;
|
||||||
|
|
||||||
import com.google.common.base.Preconditions;
|
import com.google.common.base.Preconditions;
|
||||||
import com.google.common.collect.Lists;
|
import com.google.common.collect.Lists;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.UnresolvedLinkException;
|
import org.apache.hadoop.fs.UnresolvedLinkException;
|
||||||
import org.apache.hadoop.fs.XAttr;
|
import org.apache.hadoop.fs.XAttr;
|
||||||
import org.apache.hadoop.fs.XAttrSetFlag;
|
import org.apache.hadoop.fs.XAttrSetFlag;
|
||||||
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
import org.apache.hadoop.hdfs.XAttrHelper;
|
import org.apache.hadoop.hdfs.XAttrHelper;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZoneWithId;
|
||||||
import org.apache.hadoop.hdfs.protocol.SnapshotAccessControlException;
|
import org.apache.hadoop.hdfs.protocol.SnapshotAccessControlException;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
|
||||||
|
import static org.apache.hadoop.fs.BatchedRemoteIterator.BatchedListEntries;
|
||||||
import static org.apache.hadoop.hdfs.server.common.HdfsServerConstants
|
import static org.apache.hadoop.hdfs.server.common.HdfsServerConstants
|
||||||
.CRYPTO_XATTR_ENCRYPTION_ZONE;
|
.CRYPTO_XATTR_ENCRYPTION_ZONE;
|
||||||
|
|
||||||
@ -57,17 +60,26 @@ long getINodeId() {
|
|||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private final Map<Long, EncryptionZoneInt> encryptionZones;
|
private final TreeMap<Long, EncryptionZoneInt> encryptionZones;
|
||||||
private final FSDirectory dir;
|
private final FSDirectory dir;
|
||||||
|
private final int maxListEncryptionZonesResponses;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Construct a new EncryptionZoneManager.
|
* Construct a new EncryptionZoneManager.
|
||||||
*
|
*
|
||||||
* @param dir Enclosing FSDirectory
|
* @param dir Enclosing FSDirectory
|
||||||
*/
|
*/
|
||||||
public EncryptionZoneManager(FSDirectory dir) {
|
public EncryptionZoneManager(FSDirectory dir, Configuration conf) {
|
||||||
this.dir = dir;
|
this.dir = dir;
|
||||||
encryptionZones = new HashMap<Long, EncryptionZoneInt>();
|
encryptionZones = new TreeMap<Long, EncryptionZoneInt>();
|
||||||
|
maxListEncryptionZonesResponses = conf.getInt(
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_LIST_ENCRYPTION_ZONES_NUM_RESPONSES,
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_LIST_ENCRYPTION_ZONES_NUM_RESPONSES_DEFAULT
|
||||||
|
);
|
||||||
|
Preconditions.checkArgument(maxListEncryptionZonesResponses >= 0,
|
||||||
|
DFSConfigKeys.DFS_NAMENODE_LIST_ENCRYPTION_ZONES_NUM_RESPONSES + " " +
|
||||||
|
"must be a positive integer."
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -236,17 +248,30 @@ XAttr createEncryptionZone(String src, String keyName)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Return the current list of encryption zones.
|
* Cursor-based listing of encryption zones.
|
||||||
* <p/>
|
* <p/>
|
||||||
* Called while holding the FSDirectory lock.
|
* Called while holding the FSDirectory lock.
|
||||||
*/
|
*/
|
||||||
List<EncryptionZone> listEncryptionZones() throws IOException {
|
BatchedListEntries<EncryptionZoneWithId> listEncryptionZones(long prevId)
|
||||||
|
throws IOException {
|
||||||
assert dir.hasReadLock();
|
assert dir.hasReadLock();
|
||||||
final List<EncryptionZone> ret =
|
NavigableMap<Long, EncryptionZoneInt> tailMap = encryptionZones.tailMap
|
||||||
Lists.newArrayListWithExpectedSize(encryptionZones.size());
|
(prevId, false);
|
||||||
for (EncryptionZoneInt ezi : encryptionZones.values()) {
|
final int numResponses = Math.min(maxListEncryptionZonesResponses,
|
||||||
ret.add(new EncryptionZone(getFullPathName(ezi), ezi.getKeyName()));
|
tailMap.size());
|
||||||
|
final List<EncryptionZoneWithId> zones =
|
||||||
|
Lists.newArrayListWithExpectedSize(numResponses);
|
||||||
|
|
||||||
|
int count = 0;
|
||||||
|
for (EncryptionZoneInt ezi : tailMap.values()) {
|
||||||
|
zones.add(new EncryptionZoneWithId(getFullPathName(ezi),
|
||||||
|
ezi.getKeyName(), ezi.getINodeId()));
|
||||||
|
count++;
|
||||||
|
if (count >= numResponses) {
|
||||||
|
break;
|
||||||
}
|
}
|
||||||
return ret;
|
}
|
||||||
|
final boolean hasMore = (numResponses < tailMap.size());
|
||||||
|
return new BatchedListEntries<EncryptionZoneWithId>(zones, hasMore);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -17,6 +17,7 @@
|
|||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hdfs.server.namenode;
|
package org.apache.hadoop.hdfs.server.namenode;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.fs.BatchedRemoteIterator.BatchedListEntries;
|
||||||
import static org.apache.hadoop.hdfs.server.common.HdfsServerConstants.CRYPTO_XATTR_ENCRYPTION_ZONE;
|
import static org.apache.hadoop.hdfs.server.common.HdfsServerConstants.CRYPTO_XATTR_ENCRYPTION_ZONE;
|
||||||
import static org.apache.hadoop.hdfs.server.common.HdfsServerConstants.CRYPTO_XATTR_FILE_ENCRYPTION_INFO;
|
import static org.apache.hadoop.hdfs.server.common.HdfsServerConstants.CRYPTO_XATTR_FILE_ENCRYPTION_INFO;
|
||||||
import static org.apache.hadoop.util.Time.now;
|
import static org.apache.hadoop.util.Time.now;
|
||||||
@ -58,7 +59,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.Block;
|
import org.apache.hadoop.hdfs.protocol.Block;
|
||||||
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
||||||
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZoneWithId;
|
||||||
import org.apache.hadoop.hdfs.protocol.FSLimitException.MaxDirectoryItemsExceededException;
|
import org.apache.hadoop.hdfs.protocol.FSLimitException.MaxDirectoryItemsExceededException;
|
||||||
import org.apache.hadoop.hdfs.protocol.FSLimitException.PathComponentTooLongException;
|
import org.apache.hadoop.hdfs.protocol.FSLimitException.PathComponentTooLongException;
|
||||||
import org.apache.hadoop.hdfs.protocol.FsAclPermission;
|
import org.apache.hadoop.hdfs.protocol.FsAclPermission;
|
||||||
@ -227,7 +228,7 @@ public int getWriteHoldCount() {
|
|||||||
nameCache = new NameCache<ByteArray>(threshold);
|
nameCache = new NameCache<ByteArray>(threshold);
|
||||||
namesystem = ns;
|
namesystem = ns;
|
||||||
|
|
||||||
ezManager = new EncryptionZoneManager(this);
|
ezManager = new EncryptionZoneManager(this, conf);
|
||||||
}
|
}
|
||||||
|
|
||||||
private FSNamesystem getFSNamesystem() {
|
private FSNamesystem getFSNamesystem() {
|
||||||
@ -2646,10 +2647,11 @@ XAttr createEncryptionZone(String src, String keyName)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
List<EncryptionZone> listEncryptionZones() throws IOException {
|
BatchedListEntries<EncryptionZoneWithId> listEncryptionZones(long prevId)
|
||||||
|
throws IOException {
|
||||||
readLock();
|
readLock();
|
||||||
try {
|
try {
|
||||||
return ezManager.listEncryptionZones();
|
return ezManager.listEncryptionZones(prevId);
|
||||||
} finally {
|
} finally {
|
||||||
readUnlock();
|
readUnlock();
|
||||||
}
|
}
|
||||||
|
@ -183,6 +183,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.EncryptionZoneWithId;
|
||||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants.DatanodeReportType;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants.DatanodeReportType;
|
||||||
@ -8559,7 +8560,8 @@ private void createEncryptionZoneInt(final String srcArg, String keyName,
|
|||||||
logAuditEvent(true, "createEncryptionZone", srcArg, null, resultingStat);
|
logAuditEvent(true, "createEncryptionZone", srcArg, null, resultingStat);
|
||||||
}
|
}
|
||||||
|
|
||||||
List<EncryptionZone> listEncryptionZones() throws IOException {
|
BatchedListEntries<EncryptionZoneWithId> listEncryptionZones(long prevId)
|
||||||
|
throws IOException {
|
||||||
boolean success = false;
|
boolean success = false;
|
||||||
checkSuperuserPrivilege();
|
checkSuperuserPrivilege();
|
||||||
checkOperation(OperationCategory.READ);
|
checkOperation(OperationCategory.READ);
|
||||||
@ -8567,7 +8569,8 @@ List<EncryptionZone> listEncryptionZones() throws IOException {
|
|||||||
try {
|
try {
|
||||||
checkSuperuserPrivilege();
|
checkSuperuserPrivilege();
|
||||||
checkOperation(OperationCategory.READ);
|
checkOperation(OperationCategory.READ);
|
||||||
final List<EncryptionZone> ret = dir.listEncryptionZones();
|
final BatchedListEntries<EncryptionZoneWithId> ret =
|
||||||
|
dir.listEncryptionZones(prevId);
|
||||||
success = true;
|
success = true;
|
||||||
return ret;
|
return ret;
|
||||||
} finally {
|
} finally {
|
||||||
|
@ -77,7 +77,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.DatanodeID;
|
import org.apache.hadoop.hdfs.protocol.DatanodeID;
|
||||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
import org.apache.hadoop.hdfs.protocol.DirectoryListing;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZoneWithId;
|
||||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||||
import org.apache.hadoop.hdfs.protocol.FSLimitException;
|
import org.apache.hadoop.hdfs.protocol.FSLimitException;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||||
@ -1432,8 +1432,9 @@ public void createEncryptionZone(String src, String keyName)
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public List<EncryptionZone> listEncryptionZones() throws IOException {
|
public BatchedEntries<EncryptionZoneWithId> listEncryptionZones(
|
||||||
return namesystem.listEncryptionZones();
|
long prevId) throws IOException {
|
||||||
|
return namesystem.listEncryptionZones(prevId);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -26,6 +26,7 @@
|
|||||||
import org.apache.hadoop.conf.Configured;
|
import org.apache.hadoop.conf.Configured;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.fs.RemoteIterator;
|
||||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
||||||
import org.apache.hadoop.tools.TableListing;
|
import org.apache.hadoop.tools.TableListing;
|
||||||
@ -201,8 +202,9 @@ public int run(Configuration conf, List<String> args) throws IOException {
|
|||||||
final TableListing listing = new TableListing.Builder()
|
final TableListing listing = new TableListing.Builder()
|
||||||
.addField("").addField("", true)
|
.addField("").addField("", true)
|
||||||
.wrapWidth(MAX_LINE_WIDTH).hideHeaders().build();
|
.wrapWidth(MAX_LINE_WIDTH).hideHeaders().build();
|
||||||
final List<EncryptionZone> ezs = dfs.listEncryptionZones();
|
final RemoteIterator<EncryptionZone> it = dfs.listEncryptionZones();
|
||||||
for (EncryptionZone ez : ezs) {
|
while (it.hasNext()) {
|
||||||
|
EncryptionZone ez = it.next();
|
||||||
listing.addRow(ez.getPath(), ez.getKeyName());
|
listing.addRow(ez.getPath(), ez.getKeyName());
|
||||||
}
|
}
|
||||||
System.out.println(listing.toString());
|
System.out.println(listing.toString());
|
||||||
|
@ -42,13 +42,16 @@ message CreateEncryptionZoneResponseProto {
|
|||||||
}
|
}
|
||||||
|
|
||||||
message ListEncryptionZonesRequestProto {
|
message ListEncryptionZonesRequestProto {
|
||||||
|
required int64 id = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
message EncryptionZoneProto {
|
message EncryptionZoneWithIdProto {
|
||||||
required string path = 1;
|
required string path = 1;
|
||||||
required string keyName = 2;
|
required string keyName = 2;
|
||||||
|
required int64 id = 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
message ListEncryptionZonesResponseProto {
|
message ListEncryptionZonesResponseProto {
|
||||||
repeated EncryptionZoneProto pathsAndKeys = 1;
|
repeated EncryptionZoneWithIdProto zones = 1;
|
||||||
|
required bool hasMore = 2;
|
||||||
}
|
}
|
||||||
|
@ -2052,4 +2052,13 @@
|
|||||||
</description>
|
</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>dfs.namenode.list.encryption.zones.num.responses</name>
|
||||||
|
<value>false</value>
|
||||||
|
<description>When listing encryption zones, the maximum number of zones
|
||||||
|
that will be returned in a batch. Fetching the list incrementally in
|
||||||
|
batches improves namenode performance.
|
||||||
|
</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
</configuration>
|
</configuration>
|
||||||
|
@ -44,6 +44,7 @@
|
|||||||
import org.apache.hadoop.fs.FileSystemTestHelper;
|
import org.apache.hadoop.fs.FileSystemTestHelper;
|
||||||
import org.apache.hadoop.fs.FileSystemTestWrapper;
|
import org.apache.hadoop.fs.FileSystemTestWrapper;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.fs.RemoteIterator;
|
||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
import org.apache.hadoop.fs.permission.FsPermission;
|
||||||
import org.apache.hadoop.hdfs.client.HdfsAdmin;
|
import org.apache.hadoop.hdfs.client.HdfsAdmin;
|
||||||
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
import org.apache.hadoop.hdfs.protocol.EncryptionZone;
|
||||||
@ -90,6 +91,9 @@ public void setup() throws Exception {
|
|||||||
conf.set(KeyProviderFactory.KEY_PROVIDER_PATH,
|
conf.set(KeyProviderFactory.KEY_PROVIDER_PATH,
|
||||||
JavaKeyStoreProvider.SCHEME_NAME + "://file" + testRootDir + "/test.jks"
|
JavaKeyStoreProvider.SCHEME_NAME + "://file" + testRootDir + "/test.jks"
|
||||||
);
|
);
|
||||||
|
// Lower the batch size for testing
|
||||||
|
conf.setInt(DFSConfigKeys.DFS_NAMENODE_LIST_ENCRYPTION_ZONES_NUM_RESPONSES,
|
||||||
|
2);
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1).build();
|
||||||
Logger.getLogger(EncryptionZoneManager.class).setLevel(Level.TRACE);
|
Logger.getLogger(EncryptionZoneManager.class).setLevel(Level.TRACE);
|
||||||
fs = cluster.getFileSystem();
|
fs = cluster.getFileSystem();
|
||||||
@ -114,9 +118,13 @@ public void teardown() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public void assertNumZones(final int numZones) throws IOException {
|
public void assertNumZones(final int numZones) throws IOException {
|
||||||
final List<EncryptionZone> zones = dfsAdmin.listEncryptionZones();
|
RemoteIterator<EncryptionZone> it = dfsAdmin.listEncryptionZones();
|
||||||
assertEquals("Unexpected number of encryption zones!", numZones,
|
int count = 0;
|
||||||
zones.size());
|
while (it.hasNext()) {
|
||||||
|
count++;
|
||||||
|
it.next();
|
||||||
|
}
|
||||||
|
assertEquals("Unexpected number of encryption zones!", numZones, count);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -126,9 +134,10 @@ public void assertNumZones(final int numZones) throws IOException {
|
|||||||
* @throws IOException if a matching zone could not be found
|
* @throws IOException if a matching zone could not be found
|
||||||
*/
|
*/
|
||||||
public void assertZonePresent(String keyName, String path) throws IOException {
|
public void assertZonePresent(String keyName, String path) throws IOException {
|
||||||
final List<EncryptionZone> zones = dfsAdmin.listEncryptionZones();
|
final RemoteIterator<EncryptionZone> it = dfsAdmin.listEncryptionZones();
|
||||||
boolean match = false;
|
boolean match = false;
|
||||||
for (EncryptionZone zone : zones) {
|
while (it.hasNext()) {
|
||||||
|
EncryptionZone zone = it.next();
|
||||||
boolean matchKey = (keyName == null);
|
boolean matchKey = (keyName == null);
|
||||||
boolean matchPath = (path == null);
|
boolean matchPath = (path == null);
|
||||||
if (keyName != null && zone.getKeyName().equals(keyName)) {
|
if (keyName != null && zone.getKeyName().equals(keyName)) {
|
||||||
@ -282,6 +291,16 @@ public Object run() throws Exception {
|
|||||||
dfsAdmin.createEncryptionZone(deepZone, TEST_KEY);
|
dfsAdmin.createEncryptionZone(deepZone, TEST_KEY);
|
||||||
assertNumZones(++numZones);
|
assertNumZones(++numZones);
|
||||||
assertZonePresent(null, deepZone.toString());
|
assertZonePresent(null, deepZone.toString());
|
||||||
|
|
||||||
|
// Create and list some zones to test batching of listEZ
|
||||||
|
for (int i=1; i<6; i++) {
|
||||||
|
final Path zonePath = new Path("/listZone" + i);
|
||||||
|
fsWrapper.mkdir(zonePath, FsPermission.getDirDefault(), false);
|
||||||
|
dfsAdmin.createEncryptionZone(zonePath, TEST_KEY);
|
||||||
|
numZones++;
|
||||||
|
assertNumZones(numZones);
|
||||||
|
assertZonePresent(null, zonePath.toString());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -369,9 +388,8 @@ public void testReadWrite() throws Exception {
|
|||||||
// Read them back in and compare byte-by-byte
|
// Read them back in and compare byte-by-byte
|
||||||
verifyFilesEqual(fs, baseFile, encFile1, len);
|
verifyFilesEqual(fs, baseFile, encFile1, len);
|
||||||
// Roll the key of the encryption zone
|
// Roll the key of the encryption zone
|
||||||
List<EncryptionZone> zones = dfsAdmin.listEncryptionZones();
|
assertNumZones(1);
|
||||||
assertEquals("Expected 1 EZ", 1, zones.size());
|
String keyName = dfsAdmin.listEncryptionZones().next().getKeyName();
|
||||||
String keyName = zones.get(0).getKeyName();
|
|
||||||
cluster.getNamesystem().getProvider().rollNewVersion(keyName);
|
cluster.getNamesystem().getProvider().rollNewVersion(keyName);
|
||||||
// Read them back in and compare byte-by-byte
|
// Read them back in and compare byte-by-byte
|
||||||
verifyFilesEqual(fs, baseFile, encFile1, len);
|
verifyFilesEqual(fs, baseFile, encFile1, len);
|
||||||
@ -457,14 +475,12 @@ public void testCipherSuiteNegotiation() throws Exception {
|
|||||||
|
|
||||||
@Test(timeout = 120000)
|
@Test(timeout = 120000)
|
||||||
public void testCreateEZWithNoProvider() throws Exception {
|
public void testCreateEZWithNoProvider() throws Exception {
|
||||||
|
// Unset the key provider and make sure EZ ops don't work
|
||||||
final Configuration clusterConf = cluster.getConfiguration(0);
|
final Configuration clusterConf = cluster.getConfiguration(0);
|
||||||
clusterConf.set(KeyProviderFactory.KEY_PROVIDER_PATH, "");
|
clusterConf.set(KeyProviderFactory.KEY_PROVIDER_PATH, "");
|
||||||
cluster.restartNameNode(true);
|
cluster.restartNameNode(true);
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
/* Test failure of create EZ on a directory that doesn't exist. */
|
|
||||||
final Path zone1 = new Path("/zone1");
|
final Path zone1 = new Path("/zone1");
|
||||||
/* Normal creation of an EZ */
|
|
||||||
fsWrapper.mkdir(zone1, FsPermission.getDirDefault(), true);
|
fsWrapper.mkdir(zone1, FsPermission.getDirDefault(), true);
|
||||||
try {
|
try {
|
||||||
dfsAdmin.createEncryptionZone(zone1, TEST_KEY);
|
dfsAdmin.createEncryptionZone(zone1, TEST_KEY);
|
||||||
@ -476,8 +492,7 @@ public void testCreateEZWithNoProvider() throws Exception {
|
|||||||
JavaKeyStoreProvider.SCHEME_NAME + "://file" + testRootDir + "/test.jks"
|
JavaKeyStoreProvider.SCHEME_NAME + "://file" + testRootDir + "/test.jks"
|
||||||
);
|
);
|
||||||
// Try listing EZs as well
|
// Try listing EZs as well
|
||||||
List<EncryptionZone> zones = dfsAdmin.listEncryptionZones();
|
assertNumZones(0);
|
||||||
assertEquals("Expected no zones", 0, zones.size());
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private class MyInjector extends EncryptionFaultInjector {
|
private class MyInjector extends EncryptionFaultInjector {
|
||||||
|
Loading…
Reference in New Issue
Block a user