Merge r1569890 through r1573119 from trunk.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-5535@1573120 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
commit
0a7db7f179
@ -311,9 +311,6 @@ Release 2.5.0 - UNRELEASED
|
|||||||
|
|
||||||
IMPROVEMENTS
|
IMPROVEMENTS
|
||||||
|
|
||||||
HADOOP-10278. Refactor to make CallQueue pluggable. (Chris Li via
|
|
||||||
Arpit Agarwal)
|
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
@ -349,6 +346,12 @@ Release 2.4.0 - UNRELEASED
|
|||||||
HADOOP-10374. InterfaceAudience annotations should have
|
HADOOP-10374. InterfaceAudience annotations should have
|
||||||
RetentionPolicy.RUNTIME (Enis Soztutar via Arpit Agarwal)
|
RetentionPolicy.RUNTIME (Enis Soztutar via Arpit Agarwal)
|
||||||
|
|
||||||
|
HADOOP-10278. Refactor to make CallQueue pluggable. (Chris Li via
|
||||||
|
Arpit Agarwal)
|
||||||
|
|
||||||
|
HADOOP-10285. Admin interface to swap callqueue at runtime. (Chris Li via
|
||||||
|
Arpit Agarwal)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
@ -301,6 +301,10 @@
|
|||||||
<!-- protobuf generated code -->
|
<!-- protobuf generated code -->
|
||||||
<Class name="~org\.apache\.hadoop\.ipc\.protobuf\.TestProtos.*"/>
|
<Class name="~org\.apache\.hadoop\.ipc\.protobuf\.TestProtos.*"/>
|
||||||
</Match>
|
</Match>
|
||||||
|
<Match>
|
||||||
|
<!-- protobuf generated code -->
|
||||||
|
<Class name="~org\.apache\.hadoop\.ipc\.proto\.RefreshCallQueueProtocolProtos.*"/>
|
||||||
|
</Match>
|
||||||
|
|
||||||
<!--
|
<!--
|
||||||
Manually checked, misses child thread manually syncing on parent's intrinsic lock.
|
Manually checked, misses child thread manually syncing on parent's intrinsic lock.
|
||||||
|
@ -337,6 +337,7 @@
|
|||||||
<include>GetUserMappingsProtocol.proto</include>
|
<include>GetUserMappingsProtocol.proto</include>
|
||||||
<include>RefreshAuthorizationPolicyProtocol.proto</include>
|
<include>RefreshAuthorizationPolicyProtocol.proto</include>
|
||||||
<include>RefreshUserMappingsProtocol.proto</include>
|
<include>RefreshUserMappingsProtocol.proto</include>
|
||||||
|
<include>RefreshCallQueueProtocol.proto</include>
|
||||||
</includes>
|
</includes>
|
||||||
</source>
|
</source>
|
||||||
<output>${project.build.directory}/generated-sources/java</output>
|
<output>${project.build.directory}/generated-sources/java</output>
|
||||||
|
@ -0,0 +1,48 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.ipc;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.fs.CommonConfigurationKeys;
|
||||||
|
import org.apache.hadoop.io.retry.Idempotent;
|
||||||
|
import org.apache.hadoop.security.KerberosInfo;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Protocol which is used to refresh the call queue in use currently.
|
||||||
|
*/
|
||||||
|
@KerberosInfo(
|
||||||
|
serverPrincipal=CommonConfigurationKeys.HADOOP_SECURITY_SERVICE_USER_NAME_KEY)
|
||||||
|
@InterfaceAudience.LimitedPrivate({"HDFS"})
|
||||||
|
@InterfaceStability.Evolving
|
||||||
|
public interface RefreshCallQueueProtocol {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Version 1: Initial version
|
||||||
|
*/
|
||||||
|
public static final long versionID = 1L;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Refresh the callqueue.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
@Idempotent
|
||||||
|
void refreshCallQueue() throws IOException;
|
||||||
|
}
|
@ -0,0 +1,74 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.ipc.protocolPB;
|
||||||
|
|
||||||
|
import java.io.Closeable;
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.ipc.ProtobufHelper;
|
||||||
|
import org.apache.hadoop.ipc.ProtocolMetaInterface;
|
||||||
|
import org.apache.hadoop.ipc.RPC;
|
||||||
|
import org.apache.hadoop.ipc.RpcClientUtil;
|
||||||
|
import org.apache.hadoop.ipc.RefreshCallQueueProtocol;
|
||||||
|
import org.apache.hadoop.ipc.proto.RefreshCallQueueProtocolProtos.RefreshCallQueueRequestProto;
|
||||||
|
import org.apache.hadoop.ipc.protocolPB.RefreshCallQueueProtocolPB;
|
||||||
|
|
||||||
|
import com.google.protobuf.RpcController;
|
||||||
|
import com.google.protobuf.ServiceException;
|
||||||
|
|
||||||
|
public class RefreshCallQueueProtocolClientSideTranslatorPB implements
|
||||||
|
ProtocolMetaInterface, RefreshCallQueueProtocol, Closeable {
|
||||||
|
|
||||||
|
/** RpcController is not used and hence is set to null */
|
||||||
|
private final static RpcController NULL_CONTROLLER = null;
|
||||||
|
private final RefreshCallQueueProtocolPB rpcProxy;
|
||||||
|
|
||||||
|
private final static RefreshCallQueueRequestProto
|
||||||
|
VOID_REFRESH_CALL_QUEUE_REQUEST =
|
||||||
|
RefreshCallQueueRequestProto.newBuilder().build();
|
||||||
|
|
||||||
|
public RefreshCallQueueProtocolClientSideTranslatorPB(
|
||||||
|
RefreshCallQueueProtocolPB rpcProxy) {
|
||||||
|
this.rpcProxy = rpcProxy;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void close() throws IOException {
|
||||||
|
RPC.stopProxy(rpcProxy);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void refreshCallQueue() throws IOException {
|
||||||
|
try {
|
||||||
|
rpcProxy.refreshCallQueue(NULL_CONTROLLER,
|
||||||
|
VOID_REFRESH_CALL_QUEUE_REQUEST);
|
||||||
|
} catch (ServiceException se) {
|
||||||
|
throw ProtobufHelper.getRemoteException(se);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean isMethodSupported(String methodName) throws IOException {
|
||||||
|
return RpcClientUtil.isMethodSupported(rpcProxy,
|
||||||
|
RefreshCallQueueProtocolPB.class,
|
||||||
|
RPC.RpcKind.RPC_PROTOCOL_BUFFER,
|
||||||
|
RPC.getProtocolVersion(RefreshCallQueueProtocolPB.class),
|
||||||
|
methodName);
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,37 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.ipc.protocolPB;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.fs.CommonConfigurationKeys;
|
||||||
|
import org.apache.hadoop.ipc.ProtocolInfo;
|
||||||
|
import org.apache.hadoop.security.KerberosInfo;
|
||||||
|
import org.apache.hadoop.ipc.proto.RefreshCallQueueProtocolProtos.RefreshCallQueueProtocolService;
|
||||||
|
|
||||||
|
@KerberosInfo(
|
||||||
|
serverPrincipal=CommonConfigurationKeys.HADOOP_SECURITY_SERVICE_USER_NAME_KEY)
|
||||||
|
@ProtocolInfo(
|
||||||
|
protocolName = "org.apache.hadoop.ipc.RefreshCallQueueProtocol",
|
||||||
|
protocolVersion = 1)
|
||||||
|
@InterfaceAudience.LimitedPrivate({"HDFS"})
|
||||||
|
@InterfaceStability.Evolving
|
||||||
|
public interface RefreshCallQueueProtocolPB extends
|
||||||
|
RefreshCallQueueProtocolService.BlockingInterface {
|
||||||
|
}
|
@ -0,0 +1,55 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.ipc.protocolPB;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.ipc.RefreshCallQueueProtocol;
|
||||||
|
import org.apache.hadoop.ipc.proto.RefreshCallQueueProtocolProtos.RefreshCallQueueRequestProto;
|
||||||
|
import org.apache.hadoop.ipc.proto.RefreshCallQueueProtocolProtos.RefreshCallQueueResponseProto;
|
||||||
|
|
||||||
|
import com.google.protobuf.RpcController;
|
||||||
|
import com.google.protobuf.ServiceException;
|
||||||
|
|
||||||
|
public class RefreshCallQueueProtocolServerSideTranslatorPB implements
|
||||||
|
RefreshCallQueueProtocolPB {
|
||||||
|
|
||||||
|
private final RefreshCallQueueProtocol impl;
|
||||||
|
|
||||||
|
private final static RefreshCallQueueResponseProto
|
||||||
|
VOID_REFRESH_CALL_QUEUE_RESPONSE = RefreshCallQueueResponseProto
|
||||||
|
.newBuilder().build();
|
||||||
|
|
||||||
|
public RefreshCallQueueProtocolServerSideTranslatorPB(
|
||||||
|
RefreshCallQueueProtocol impl) {
|
||||||
|
this.impl = impl;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public RefreshCallQueueResponseProto refreshCallQueue(
|
||||||
|
RpcController controller, RefreshCallQueueRequestProto request)
|
||||||
|
throws ServiceException {
|
||||||
|
try {
|
||||||
|
impl.refreshCallQueue();
|
||||||
|
} catch (IOException e) {
|
||||||
|
throw new ServiceException(e);
|
||||||
|
}
|
||||||
|
return VOID_REFRESH_CALL_QUEUE_RESPONSE;
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,52 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/**
|
||||||
|
* These .proto interfaces are private and stable.
|
||||||
|
* Please see http://wiki.apache.org/hadoop/Compatibility
|
||||||
|
* for what changes are allowed for a *stable* .proto interface.
|
||||||
|
*/
|
||||||
|
|
||||||
|
option java_package = "org.apache.hadoop.ipc.proto";
|
||||||
|
option java_outer_classname = "RefreshCallQueueProtocolProtos";
|
||||||
|
option java_generic_services = true;
|
||||||
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.common;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Refresh callqueue request.
|
||||||
|
*/
|
||||||
|
message RefreshCallQueueRequestProto {
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* void response.
|
||||||
|
*/
|
||||||
|
message RefreshCallQueueResponseProto {
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Protocol which is used to refresh the callqueue.
|
||||||
|
*/
|
||||||
|
service RefreshCallQueueProtocolService {
|
||||||
|
/**
|
||||||
|
* Refresh the callqueue.
|
||||||
|
*/
|
||||||
|
rpc refreshCallQueue(RefreshCallQueueRequestProto)
|
||||||
|
returns(RefreshCallQueueResponseProto);
|
||||||
|
}
|
@ -109,9 +109,6 @@ Trunk (Unreleased)
|
|||||||
HDFS-4052. BlockManager#invalidateWork should print log outside the lock.
|
HDFS-4052. BlockManager#invalidateWork should print log outside the lock.
|
||||||
(Jing Zhao via suresh)
|
(Jing Zhao via suresh)
|
||||||
|
|
||||||
HDFS-4200. Reduce the size of synchronized sections in PacketResponder.
|
|
||||||
(suresh)
|
|
||||||
|
|
||||||
HDFS-3358. Specify explicitly that the NN UI status total is talking
|
HDFS-3358. Specify explicitly that the NN UI status total is talking
|
||||||
of persistent objects on heap. (harsh)
|
of persistent objects on heap. (harsh)
|
||||||
|
|
||||||
@ -370,6 +367,9 @@ Release 2.4.0 - UNRELEASED
|
|||||||
HDFS-6030. Remove an unused constructor in INode.java. (yzhang via
|
HDFS-6030. Remove an unused constructor in INode.java. (yzhang via
|
||||||
cmccabe)
|
cmccabe)
|
||||||
|
|
||||||
|
HDFS-4200. Reduce the size of synchronized sections in PacketResponder.
|
||||||
|
(suresh)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
HDFS-5790. LeaseManager.findPath is very slow when many leases need recovery
|
HDFS-5790. LeaseManager.findPath is very slow when many leases need recovery
|
||||||
@ -505,6 +505,15 @@ Release 2.4.0 - UNRELEASED
|
|||||||
HDFS-5339. WebHDFS URI does not accept logical nameservices when security is
|
HDFS-5339. WebHDFS URI does not accept logical nameservices when security is
|
||||||
enabled. (Haohui Mai via jing9)
|
enabled. (Haohui Mai via jing9)
|
||||||
|
|
||||||
|
HDFS-6033. PBImageXmlWriter incorrectly handles processing cache
|
||||||
|
directives. (atm)
|
||||||
|
|
||||||
|
HDFS-5821. TestHDFSCLI fails for user names with the dash character.
|
||||||
|
(Gera Shegalov via Arpit Agarwal)
|
||||||
|
|
||||||
|
HDFS-5956. A file size is multiplied by the replication factor in 'hdfs oiv
|
||||||
|
-p FileDistribution' option. (Akira Ajisaka via wheat9)
|
||||||
|
|
||||||
BREAKDOWN OF HDFS-5698 SUBTASKS AND RELATED JIRAS
|
BREAKDOWN OF HDFS-5698 SUBTASKS AND RELATED JIRAS
|
||||||
|
|
||||||
HDFS-5717. Save FSImage header in protobuf. (Haohui Mai via jing9)
|
HDFS-5717. Save FSImage header in protobuf. (Haohui Mai via jing9)
|
||||||
|
@ -75,6 +75,9 @@
|
|||||||
import org.apache.hadoop.security.protocolPB.RefreshAuthorizationPolicyProtocolPB;
|
import org.apache.hadoop.security.protocolPB.RefreshAuthorizationPolicyProtocolPB;
|
||||||
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolClientSideTranslatorPB;
|
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolClientSideTranslatorPB;
|
||||||
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolPB;
|
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolPB;
|
||||||
|
import org.apache.hadoop.ipc.RefreshCallQueueProtocol;
|
||||||
|
import org.apache.hadoop.ipc.protocolPB.RefreshCallQueueProtocolPB;
|
||||||
|
import org.apache.hadoop.ipc.protocolPB.RefreshCallQueueProtocolClientSideTranslatorPB;
|
||||||
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
||||||
import org.apache.hadoop.tools.protocolPB.GetUserMappingsProtocolClientSideTranslatorPB;
|
import org.apache.hadoop.tools.protocolPB.GetUserMappingsProtocolClientSideTranslatorPB;
|
||||||
import org.apache.hadoop.tools.protocolPB.GetUserMappingsProtocolPB;
|
import org.apache.hadoop.tools.protocolPB.GetUserMappingsProtocolPB;
|
||||||
@ -252,13 +255,16 @@ public static <T> ProxyAndInfo<T> createNonHAProxy(
|
|||||||
} else if (xface == RefreshAuthorizationPolicyProtocol.class) {
|
} else if (xface == RefreshAuthorizationPolicyProtocol.class) {
|
||||||
proxy = (T) createNNProxyWithRefreshAuthorizationPolicyProtocol(nnAddr,
|
proxy = (T) createNNProxyWithRefreshAuthorizationPolicyProtocol(nnAddr,
|
||||||
conf, ugi);
|
conf, ugi);
|
||||||
|
} else if (xface == RefreshCallQueueProtocol.class) {
|
||||||
|
proxy = (T) createNNProxyWithRefreshCallQueueProtocol(nnAddr, conf, ugi);
|
||||||
} else {
|
} else {
|
||||||
String message = "Upsupported protocol found when creating the proxy " +
|
String message = "Unsupported protocol found when creating the proxy " +
|
||||||
"connection to NameNode: " +
|
"connection to NameNode: " +
|
||||||
((xface != null) ? xface.getClass().getName() : "null");
|
((xface != null) ? xface.getClass().getName() : "null");
|
||||||
LOG.error(message);
|
LOG.error(message);
|
||||||
throw new IllegalStateException(message);
|
throw new IllegalStateException(message);
|
||||||
}
|
}
|
||||||
|
|
||||||
return new ProxyAndInfo<T>(proxy, dtService);
|
return new ProxyAndInfo<T>(proxy, dtService);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -286,6 +292,14 @@ private static JournalProtocol createNNProxyWithJournalProtocol(
|
|||||||
return new RefreshUserMappingsProtocolClientSideTranslatorPB(proxy);
|
return new RefreshUserMappingsProtocolClientSideTranslatorPB(proxy);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private static RefreshCallQueueProtocol
|
||||||
|
createNNProxyWithRefreshCallQueueProtocol(InetSocketAddress address,
|
||||||
|
Configuration conf, UserGroupInformation ugi) throws IOException {
|
||||||
|
RefreshCallQueueProtocolPB proxy = (RefreshCallQueueProtocolPB)
|
||||||
|
createNameNodeProxy(address, conf, ugi, RefreshCallQueueProtocolPB.class, 0);
|
||||||
|
return new RefreshCallQueueProtocolClientSideTranslatorPB(proxy);
|
||||||
|
}
|
||||||
|
|
||||||
private static GetUserMappingsProtocol createNNProxyWithGetUserMappingsProtocol(
|
private static GetUserMappingsProtocol createNNProxyWithGetUserMappingsProtocol(
|
||||||
InetSocketAddress address, Configuration conf, UserGroupInformation ugi)
|
InetSocketAddress address, Configuration conf, UserGroupInformation ugi)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
|
@ -58,6 +58,7 @@
|
|||||||
import org.apache.hadoop.security.SecurityUtil;
|
import org.apache.hadoop.security.SecurityUtil;
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
import org.apache.hadoop.security.UserGroupInformation;
|
||||||
import org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol;
|
import org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol;
|
||||||
|
import org.apache.hadoop.ipc.RefreshCallQueueProtocol;
|
||||||
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
||||||
import org.apache.hadoop.util.ExitUtil.ExitException;
|
import org.apache.hadoop.util.ExitUtil.ExitException;
|
||||||
import org.apache.hadoop.util.GenericOptionsParser;
|
import org.apache.hadoop.util.GenericOptionsParser;
|
||||||
@ -230,6 +231,8 @@ public long getProtocolVersion(String protocol,
|
|||||||
return RefreshAuthorizationPolicyProtocol.versionID;
|
return RefreshAuthorizationPolicyProtocol.versionID;
|
||||||
} else if (protocol.equals(RefreshUserMappingsProtocol.class.getName())){
|
} else if (protocol.equals(RefreshUserMappingsProtocol.class.getName())){
|
||||||
return RefreshUserMappingsProtocol.versionID;
|
return RefreshUserMappingsProtocol.versionID;
|
||||||
|
} else if (protocol.equals(RefreshCallQueueProtocol.class.getName())) {
|
||||||
|
return RefreshCallQueueProtocol.versionID;
|
||||||
} else if (protocol.equals(GetUserMappingsProtocol.class.getName())){
|
} else if (protocol.equals(GetUserMappingsProtocol.class.getName())){
|
||||||
return GetUserMappingsProtocol.versionID;
|
return GetUserMappingsProtocol.versionID;
|
||||||
} else {
|
} else {
|
||||||
|
@ -141,6 +141,9 @@
|
|||||||
import org.apache.hadoop.security.protocolPB.RefreshAuthorizationPolicyProtocolServerSideTranslatorPB;
|
import org.apache.hadoop.security.protocolPB.RefreshAuthorizationPolicyProtocolServerSideTranslatorPB;
|
||||||
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolPB;
|
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolPB;
|
||||||
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolServerSideTranslatorPB;
|
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolServerSideTranslatorPB;
|
||||||
|
import org.apache.hadoop.ipc.protocolPB.RefreshCallQueueProtocolPB;
|
||||||
|
import org.apache.hadoop.ipc.protocolPB.RefreshCallQueueProtocolServerSideTranslatorPB;
|
||||||
|
import org.apache.hadoop.ipc.proto.RefreshCallQueueProtocolProtos.RefreshCallQueueProtocolService;
|
||||||
import org.apache.hadoop.security.token.SecretManager.InvalidToken;
|
import org.apache.hadoop.security.token.SecretManager.InvalidToken;
|
||||||
import org.apache.hadoop.security.token.Token;
|
import org.apache.hadoop.security.token.Token;
|
||||||
import org.apache.hadoop.tools.proto.GetUserMappingsProtocolProtos.GetUserMappingsProtocolService;
|
import org.apache.hadoop.tools.proto.GetUserMappingsProtocolProtos.GetUserMappingsProtocolService;
|
||||||
@ -218,6 +221,11 @@ public NameNodeRpcServer(Configuration conf, NameNode nn)
|
|||||||
BlockingService refreshUserMappingService = RefreshUserMappingsProtocolService
|
BlockingService refreshUserMappingService = RefreshUserMappingsProtocolService
|
||||||
.newReflectiveBlockingService(refreshUserMappingXlator);
|
.newReflectiveBlockingService(refreshUserMappingXlator);
|
||||||
|
|
||||||
|
RefreshCallQueueProtocolServerSideTranslatorPB refreshCallQueueXlator =
|
||||||
|
new RefreshCallQueueProtocolServerSideTranslatorPB(this);
|
||||||
|
BlockingService refreshCallQueueService = RefreshCallQueueProtocolService
|
||||||
|
.newReflectiveBlockingService(refreshCallQueueXlator);
|
||||||
|
|
||||||
GetUserMappingsProtocolServerSideTranslatorPB getUserMappingXlator =
|
GetUserMappingsProtocolServerSideTranslatorPB getUserMappingXlator =
|
||||||
new GetUserMappingsProtocolServerSideTranslatorPB(this);
|
new GetUserMappingsProtocolServerSideTranslatorPB(this);
|
||||||
BlockingService getUserMappingService = GetUserMappingsProtocolService
|
BlockingService getUserMappingService = GetUserMappingsProtocolService
|
||||||
@ -264,6 +272,9 @@ public NameNodeRpcServer(Configuration conf, NameNode nn)
|
|||||||
refreshAuthService, serviceRpcServer);
|
refreshAuthService, serviceRpcServer);
|
||||||
DFSUtil.addPBProtocol(conf, RefreshUserMappingsProtocolPB.class,
|
DFSUtil.addPBProtocol(conf, RefreshUserMappingsProtocolPB.class,
|
||||||
refreshUserMappingService, serviceRpcServer);
|
refreshUserMappingService, serviceRpcServer);
|
||||||
|
// We support Refreshing call queue here in case the client RPC queue is full
|
||||||
|
DFSUtil.addPBProtocol(conf, RefreshCallQueueProtocolPB.class,
|
||||||
|
refreshCallQueueService, serviceRpcServer);
|
||||||
DFSUtil.addPBProtocol(conf, GetUserMappingsProtocolPB.class,
|
DFSUtil.addPBProtocol(conf, GetUserMappingsProtocolPB.class,
|
||||||
getUserMappingService, serviceRpcServer);
|
getUserMappingService, serviceRpcServer);
|
||||||
|
|
||||||
@ -306,6 +317,8 @@ public NameNodeRpcServer(Configuration conf, NameNode nn)
|
|||||||
refreshAuthService, clientRpcServer);
|
refreshAuthService, clientRpcServer);
|
||||||
DFSUtil.addPBProtocol(conf, RefreshUserMappingsProtocolPB.class,
|
DFSUtil.addPBProtocol(conf, RefreshUserMappingsProtocolPB.class,
|
||||||
refreshUserMappingService, clientRpcServer);
|
refreshUserMappingService, clientRpcServer);
|
||||||
|
DFSUtil.addPBProtocol(conf, RefreshCallQueueProtocolPB.class,
|
||||||
|
refreshCallQueueService, clientRpcServer);
|
||||||
DFSUtil.addPBProtocol(conf, GetUserMappingsProtocolPB.class,
|
DFSUtil.addPBProtocol(conf, GetUserMappingsProtocolPB.class,
|
||||||
getUserMappingService, clientRpcServer);
|
getUserMappingService, clientRpcServer);
|
||||||
|
|
||||||
@ -1129,6 +1142,17 @@ public void refreshSuperUserGroupsConfiguration() {
|
|||||||
ProxyUsers.refreshSuperUserGroupsConfiguration();
|
ProxyUsers.refreshSuperUserGroupsConfiguration();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override // RefreshCallQueueProtocol
|
||||||
|
public void refreshCallQueue() {
|
||||||
|
LOG.info("Refreshing call queue.");
|
||||||
|
|
||||||
|
Configuration conf = new Configuration();
|
||||||
|
clientRpcServer.refreshCallQueue(conf);
|
||||||
|
if (this.serviceRpcServer != null) {
|
||||||
|
serviceRpcServer.refreshCallQueue(conf);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
@Override // GetUserMappingsProtocol
|
@Override // GetUserMappingsProtocol
|
||||||
public String[] getGroupsForUser(String user) throws IOException {
|
public String[] getGroupsForUser(String user) throws IOException {
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
|
@ -23,6 +23,7 @@
|
|||||||
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
||||||
import org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol;
|
import org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol;
|
||||||
import org.apache.hadoop.security.RefreshUserMappingsProtocol;
|
import org.apache.hadoop.security.RefreshUserMappingsProtocol;
|
||||||
|
import org.apache.hadoop.ipc.RefreshCallQueueProtocol;
|
||||||
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
||||||
|
|
||||||
/** The full set of RPC methods implemented by the Namenode. */
|
/** The full set of RPC methods implemented by the Namenode. */
|
||||||
@ -33,6 +34,7 @@ public interface NamenodeProtocols
|
|||||||
NamenodeProtocol,
|
NamenodeProtocol,
|
||||||
RefreshAuthorizationPolicyProtocol,
|
RefreshAuthorizationPolicyProtocol,
|
||||||
RefreshUserMappingsProtocol,
|
RefreshUserMappingsProtocol,
|
||||||
|
RefreshCallQueueProtocol,
|
||||||
GetUserMappingsProtocol,
|
GetUserMappingsProtocol,
|
||||||
HAServiceProtocol {
|
HAServiceProtocol {
|
||||||
}
|
}
|
||||||
|
@ -66,6 +66,7 @@
|
|||||||
import org.apache.hadoop.security.SecurityUtil;
|
import org.apache.hadoop.security.SecurityUtil;
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
import org.apache.hadoop.security.UserGroupInformation;
|
||||||
import org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol;
|
import org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol;
|
||||||
|
import org.apache.hadoop.ipc.RefreshCallQueueProtocol;
|
||||||
import org.apache.hadoop.util.StringUtils;
|
import org.apache.hadoop.util.StringUtils;
|
||||||
import org.apache.hadoop.util.ToolRunner;
|
import org.apache.hadoop.util.ToolRunner;
|
||||||
|
|
||||||
@ -652,6 +653,7 @@ private void printHelp(String cmd) {
|
|||||||
"\t[-refreshServiceAcl]\n" +
|
"\t[-refreshServiceAcl]\n" +
|
||||||
"\t[-refreshUserToGroupsMappings]\n" +
|
"\t[-refreshUserToGroupsMappings]\n" +
|
||||||
"\t[-refreshSuperUserGroupsConfiguration]\n" +
|
"\t[-refreshSuperUserGroupsConfiguration]\n" +
|
||||||
|
"\t[-refreshCallQueue]\n" +
|
||||||
"\t[-printTopology]\n" +
|
"\t[-printTopology]\n" +
|
||||||
"\t[-refreshNamenodes datanodehost:port]\n"+
|
"\t[-refreshNamenodes datanodehost:port]\n"+
|
||||||
"\t[-deleteBlockPool datanodehost:port blockpoolId [force]]\n"+
|
"\t[-deleteBlockPool datanodehost:port blockpoolId [force]]\n"+
|
||||||
@ -723,6 +725,8 @@ private void printHelp(String cmd) {
|
|||||||
String refreshSuperUserGroupsConfiguration =
|
String refreshSuperUserGroupsConfiguration =
|
||||||
"-refreshSuperUserGroupsConfiguration: Refresh superuser proxy groups mappings\n";
|
"-refreshSuperUserGroupsConfiguration: Refresh superuser proxy groups mappings\n";
|
||||||
|
|
||||||
|
String refreshCallQueue = "-refreshCallQueue: Reload the call queue from config\n";
|
||||||
|
|
||||||
String printTopology = "-printTopology: Print a tree of the racks and their\n" +
|
String printTopology = "-printTopology: Print a tree of the racks and their\n" +
|
||||||
"\t\tnodes as reported by the Namenode\n";
|
"\t\tnodes as reported by the Namenode\n";
|
||||||
|
|
||||||
@ -805,6 +809,8 @@ private void printHelp(String cmd) {
|
|||||||
System.out.println(refreshUserToGroupsMappings);
|
System.out.println(refreshUserToGroupsMappings);
|
||||||
} else if ("refreshSuperUserGroupsConfiguration".equals(cmd)) {
|
} else if ("refreshSuperUserGroupsConfiguration".equals(cmd)) {
|
||||||
System.out.println(refreshSuperUserGroupsConfiguration);
|
System.out.println(refreshSuperUserGroupsConfiguration);
|
||||||
|
} else if ("refreshCallQueue".equals(cmd)) {
|
||||||
|
System.out.println(refreshCallQueue);
|
||||||
} else if ("printTopology".equals(cmd)) {
|
} else if ("printTopology".equals(cmd)) {
|
||||||
System.out.println(printTopology);
|
System.out.println(printTopology);
|
||||||
} else if ("refreshNamenodes".equals(cmd)) {
|
} else if ("refreshNamenodes".equals(cmd)) {
|
||||||
@ -843,6 +849,7 @@ private void printHelp(String cmd) {
|
|||||||
System.out.println(refreshServiceAcl);
|
System.out.println(refreshServiceAcl);
|
||||||
System.out.println(refreshUserToGroupsMappings);
|
System.out.println(refreshUserToGroupsMappings);
|
||||||
System.out.println(refreshSuperUserGroupsConfiguration);
|
System.out.println(refreshSuperUserGroupsConfiguration);
|
||||||
|
System.out.println(refreshCallQueue);
|
||||||
System.out.println(printTopology);
|
System.out.println(printTopology);
|
||||||
System.out.println(refreshNamenodes);
|
System.out.println(refreshNamenodes);
|
||||||
System.out.println(deleteBlockPool);
|
System.out.println(deleteBlockPool);
|
||||||
@ -1034,6 +1041,27 @@ public int refreshSuperUserGroupsConfiguration() throws IOException {
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public int refreshCallQueue() throws IOException {
|
||||||
|
// Get the current configuration
|
||||||
|
Configuration conf = getConf();
|
||||||
|
|
||||||
|
// for security authorization
|
||||||
|
// server principal for this call
|
||||||
|
// should be NN's one.
|
||||||
|
conf.set(CommonConfigurationKeys.HADOOP_SECURITY_SERVICE_USER_NAME_KEY,
|
||||||
|
conf.get(DFSConfigKeys.DFS_NAMENODE_USER_NAME_KEY, ""));
|
||||||
|
|
||||||
|
// Create the client
|
||||||
|
RefreshCallQueueProtocol refreshProtocol =
|
||||||
|
NameNodeProxies.createProxy(conf, FileSystem.getDefaultUri(conf),
|
||||||
|
RefreshCallQueueProtocol.class).getProxy();
|
||||||
|
|
||||||
|
// Refresh the user-to-groups mappings
|
||||||
|
refreshProtocol.refreshCallQueue();
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Displays format of commands.
|
* Displays format of commands.
|
||||||
* @param cmd The command that is being executed.
|
* @param cmd The command that is being executed.
|
||||||
@ -1093,6 +1121,9 @@ private static void printUsage(String cmd) {
|
|||||||
} else if ("-refreshSuperUserGroupsConfiguration".equals(cmd)) {
|
} else if ("-refreshSuperUserGroupsConfiguration".equals(cmd)) {
|
||||||
System.err.println("Usage: java DFSAdmin"
|
System.err.println("Usage: java DFSAdmin"
|
||||||
+ " [-refreshSuperUserGroupsConfiguration]");
|
+ " [-refreshSuperUserGroupsConfiguration]");
|
||||||
|
} else if ("-refreshCallQueue".equals(cmd)) {
|
||||||
|
System.err.println("Usage: java DFSAdmin"
|
||||||
|
+ " [-refreshCallQueue]");
|
||||||
} else if ("-printTopology".equals(cmd)) {
|
} else if ("-printTopology".equals(cmd)) {
|
||||||
System.err.println("Usage: java DFSAdmin"
|
System.err.println("Usage: java DFSAdmin"
|
||||||
+ " [-printTopology]");
|
+ " [-printTopology]");
|
||||||
@ -1125,6 +1156,7 @@ private static void printUsage(String cmd) {
|
|||||||
System.err.println(" [-refreshServiceAcl]");
|
System.err.println(" [-refreshServiceAcl]");
|
||||||
System.err.println(" [-refreshUserToGroupsMappings]");
|
System.err.println(" [-refreshUserToGroupsMappings]");
|
||||||
System.err.println(" [-refreshSuperUserGroupsConfiguration]");
|
System.err.println(" [-refreshSuperUserGroupsConfiguration]");
|
||||||
|
System.err.println(" [-refreshCallQueue]");
|
||||||
System.err.println(" [-printTopology]");
|
System.err.println(" [-printTopology]");
|
||||||
System.err.println(" [-refreshNamenodes datanodehost:port]");
|
System.err.println(" [-refreshNamenodes datanodehost:port]");
|
||||||
System.err.println(" [-deleteBlockPool datanode-host:port blockpoolId [force]]");
|
System.err.println(" [-deleteBlockPool datanode-host:port blockpoolId [force]]");
|
||||||
@ -1315,6 +1347,8 @@ public int run(String[] argv) throws Exception {
|
|||||||
exitCode = refreshUserToGroupsMappings();
|
exitCode = refreshUserToGroupsMappings();
|
||||||
} else if ("-refreshSuperUserGroupsConfiguration".equals(cmd)) {
|
} else if ("-refreshSuperUserGroupsConfiguration".equals(cmd)) {
|
||||||
exitCode = refreshSuperUserGroupsConfiguration();
|
exitCode = refreshSuperUserGroupsConfiguration();
|
||||||
|
} else if ("-refreshCallQueue".equals(cmd)) {
|
||||||
|
exitCode = refreshCallQueue();
|
||||||
} else if ("-printTopology".equals(cmd)) {
|
} else if ("-printTopology".equals(cmd)) {
|
||||||
exitCode = printTopology();
|
exitCode = printTopology();
|
||||||
} else if ("-refreshNamenodes".equals(cmd)) {
|
} else if ("-refreshNamenodes".equals(cmd)) {
|
||||||
|
@ -123,10 +123,10 @@ private void run(InputStream in) throws IOException {
|
|||||||
totalBlocks += f.getBlocksCount();
|
totalBlocks += f.getBlocksCount();
|
||||||
long fileSize = 0;
|
long fileSize = 0;
|
||||||
for (BlockProto b : f.getBlocksList()) {
|
for (BlockProto b : f.getBlocksList()) {
|
||||||
fileSize += b.getNumBytes() * f.getReplication();
|
fileSize += b.getNumBytes();
|
||||||
}
|
}
|
||||||
maxFileSize = Math.max(fileSize, maxFileSize);
|
maxFileSize = Math.max(fileSize, maxFileSize);
|
||||||
totalSpace += fileSize;
|
totalSpace += fileSize * f.getReplication();
|
||||||
|
|
||||||
int bucket = fileSize > maxSize ? distribution.length - 1 : (int) Math
|
int bucket = fileSize > maxSize ? distribution.length - 1 : (int) Math
|
||||||
.ceil((double)fileSize / steps);
|
.ceil((double)fileSize / steps);
|
||||||
|
@ -157,7 +157,7 @@ private void dumpCacheManagerSection(InputStream is) throws IOException {
|
|||||||
.o("maxRelativeExpiry", p.getMaxRelativeExpiry());
|
.o("maxRelativeExpiry", p.getMaxRelativeExpiry());
|
||||||
out.print("</pool>\n");
|
out.print("</pool>\n");
|
||||||
}
|
}
|
||||||
for (int i = 0; i < s.getNumPools(); ++i) {
|
for (int i = 0; i < s.getNumDirectives(); ++i) {
|
||||||
CacheDirectiveInfoProto p = CacheDirectiveInfoProto
|
CacheDirectiveInfoProto p = CacheDirectiveInfoProto
|
||||||
.parseDelimitedFrom(is);
|
.parseDelimitedFrom(is);
|
||||||
out.print("<directive>");
|
out.print("<directive>");
|
||||||
|
@ -39,6 +39,8 @@
|
|||||||
import org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol;
|
import org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol;
|
||||||
import org.apache.hadoop.security.protocolPB.RefreshAuthorizationPolicyProtocolClientSideTranslatorPB;
|
import org.apache.hadoop.security.protocolPB.RefreshAuthorizationPolicyProtocolClientSideTranslatorPB;
|
||||||
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolClientSideTranslatorPB;
|
import org.apache.hadoop.security.protocolPB.RefreshUserMappingsProtocolClientSideTranslatorPB;
|
||||||
|
import org.apache.hadoop.ipc.protocolPB.RefreshCallQueueProtocolClientSideTranslatorPB;
|
||||||
|
import org.apache.hadoop.ipc.RefreshCallQueueProtocol;
|
||||||
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
import org.apache.hadoop.tools.GetUserMappingsProtocol;
|
||||||
import org.apache.hadoop.tools.protocolPB.GetUserMappingsProtocolClientSideTranslatorPB;
|
import org.apache.hadoop.tools.protocolPB.GetUserMappingsProtocolClientSideTranslatorPB;
|
||||||
import org.junit.AfterClass;
|
import org.junit.AfterClass;
|
||||||
@ -169,4 +171,15 @@ public void testRefreshUserMappingsProtocol() throws IOException {
|
|||||||
assertTrue(
|
assertTrue(
|
||||||
translator.isMethodSupported("refreshUserToGroupsMappings"));
|
translator.isMethodSupported("refreshUserToGroupsMappings"));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testRefreshCallQueueProtocol() throws IOException {
|
||||||
|
RefreshCallQueueProtocolClientSideTranslatorPB translator =
|
||||||
|
(RefreshCallQueueProtocolClientSideTranslatorPB)
|
||||||
|
NameNodeProxies.createNonHAProxy(conf, nnAddress,
|
||||||
|
RefreshCallQueueProtocol.class,
|
||||||
|
UserGroupInformation.getCurrentUser(), true).getProxy();
|
||||||
|
assertTrue(
|
||||||
|
translator.isMethodSupported("refreshCallQueue"));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -28,6 +28,8 @@
|
|||||||
import java.io.RandomAccessFile;
|
import java.io.RandomAccessFile;
|
||||||
import java.io.StringReader;
|
import java.io.StringReader;
|
||||||
import java.io.StringWriter;
|
import java.io.StringWriter;
|
||||||
|
import java.util.Collections;
|
||||||
|
import java.util.Comparator;
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
import java.util.regex.Matcher;
|
import java.util.regex.Matcher;
|
||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
@ -241,7 +243,7 @@ private void copyPartOfFile(File src, File dest) throws IOException {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testFileDistributionVisitor() throws IOException {
|
public void testFileDistributionCalculator() throws IOException {
|
||||||
StringWriter output = new StringWriter();
|
StringWriter output = new StringWriter();
|
||||||
PrintWriter o = new PrintWriter(output);
|
PrintWriter o = new PrintWriter(output);
|
||||||
new FileDistributionCalculator(new Configuration(), 0, 0, o)
|
new FileDistributionCalculator(new Configuration(), 0, 0, o)
|
||||||
@ -250,10 +252,29 @@ public void testFileDistributionVisitor() throws IOException {
|
|||||||
|
|
||||||
Pattern p = Pattern.compile("totalFiles = (\\d+)\n");
|
Pattern p = Pattern.compile("totalFiles = (\\d+)\n");
|
||||||
Matcher matcher = p.matcher(output.getBuffer());
|
Matcher matcher = p.matcher(output.getBuffer());
|
||||||
|
|
||||||
assertTrue(matcher.find() && matcher.groupCount() == 1);
|
assertTrue(matcher.find() && matcher.groupCount() == 1);
|
||||||
int totalFiles = Integer.parseInt(matcher.group(1));
|
int totalFiles = Integer.parseInt(matcher.group(1));
|
||||||
assertEquals(totalFiles, NUM_DIRS * FILES_PER_DIR);
|
assertEquals(NUM_DIRS * FILES_PER_DIR, totalFiles);
|
||||||
|
|
||||||
|
p = Pattern.compile("totalDirectories = (\\d+)\n");
|
||||||
|
matcher = p.matcher(output.getBuffer());
|
||||||
|
assertTrue(matcher.find() && matcher.groupCount() == 1);
|
||||||
|
int totalDirs = Integer.parseInt(matcher.group(1));
|
||||||
|
// totalDirs includes root directory
|
||||||
|
assertEquals(NUM_DIRS + 1, totalDirs);
|
||||||
|
|
||||||
|
FileStatus maxFile = Collections.max(writtenFiles.values(),
|
||||||
|
new Comparator<FileStatus>() {
|
||||||
|
@Override
|
||||||
|
public int compare(FileStatus first, FileStatus second) {
|
||||||
|
return first.getLen() < second.getLen() ? -1 :
|
||||||
|
((first.getLen() == second.getLen()) ? 0 : 1);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
p = Pattern.compile("maxFileSize = (\\d+)\n");
|
||||||
|
matcher = p.matcher(output.getBuffer());
|
||||||
|
assertTrue(matcher.find() && matcher.groupCount() == 1);
|
||||||
|
assertEquals(maxFile.getLen(), Long.parseLong(matcher.group(1)));
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
File diff suppressed because it is too large
Load Diff
Loading…
Reference in New Issue
Block a user