HDFS-12862. CacheDirective becomes invalid when NN restart or failover. Contributed by Wang XL.

This commit is contained in:
Wei-Chiu Chuang 2019-07-08 15:19:42 +08:00
parent ce91d35b29
commit ec851e4db2
2 changed files with 66 additions and 2 deletions

View File

@ -538,7 +538,7 @@ public static void writeCacheDirectiveInfo(DataOutputStream out,
writeString(directive.getPool(), out); writeString(directive.getPool(), out);
} }
if (directive.getExpiration() != null) { if (directive.getExpiration() != null) {
writeLong(directive.getExpiration().getMillis(), out); writeLong(directive.getExpiration().getAbsoluteMillis(), out);
} }
} }
@ -610,7 +610,7 @@ public static void writeCacheDirectiveInfo(ContentHandler contentHandler,
} }
if (directive.getExpiration() != null) { if (directive.getExpiration() != null) {
XMLUtils.addSaxString(contentHandler, "EXPIRATION", XMLUtils.addSaxString(contentHandler, "EXPIRATION",
"" + directive.getExpiration().getMillis()); "" + directive.getExpiration().getAbsoluteMillis());
} }
} }

View File

@ -43,6 +43,7 @@
import java.util.List; import java.util.List;
import org.apache.commons.lang3.time.DateUtils; import org.apache.commons.lang3.time.DateUtils;
import org.apache.hadoop.hdfs.server.namenode.ha.HATestUtil;
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
@ -60,9 +61,11 @@
import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.DFSTestUtil; import org.apache.hadoop.hdfs.DFSTestUtil;
import org.apache.hadoop.hdfs.DistributedFileSystem; import org.apache.hadoop.hdfs.DistributedFileSystem;
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
import org.apache.hadoop.hdfs.HdfsConfiguration; import org.apache.hadoop.hdfs.HdfsConfiguration;
import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.hdfs.protocol.CacheDirectiveEntry; import org.apache.hadoop.hdfs.protocol.CacheDirectiveEntry;
import org.apache.hadoop.hdfs.protocol.CacheDirective;
import org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo; import org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo;
import org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo.Expiration; import org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo.Expiration;
import org.apache.hadoop.hdfs.protocol.CacheDirectiveIterator; import org.apache.hadoop.hdfs.protocol.CacheDirectiveIterator;
@ -1605,4 +1608,65 @@ public void testAddingCacheDirectiveInfosWhenCachingIsDisabled()
waitForCachedBlocks(namenode, expected, 0, waitForCachedBlocks(namenode, expected, 0,
"testAddingCacheDirectiveInfosWhenCachingIsDisabled:2"); "testAddingCacheDirectiveInfosWhenCachingIsDisabled:2");
} }
@Test(timeout=120000)
public void testExpiryTimeConsistency() throws Exception {
conf.setInt(DFSConfigKeys.DFS_HA_LOGROLL_PERIOD_KEY, 1);
conf.setInt(DFSConfigKeys.DFS_HA_TAILEDITS_PERIOD_KEY, 1);
MiniDFSCluster dfsCluster =
new MiniDFSCluster.Builder(conf).numDataNodes(NUM_DATANODES)
.nnTopology(MiniDFSNNTopology.simpleHATopology())
.build();
dfsCluster.transitionToActive(0);
DistributedFileSystem fs = dfsCluster.getFileSystem(0);
final NameNode ann = dfsCluster.getNameNode(0);
final Path filename = new Path("/file");
final short replication = (short) 3;
DFSTestUtil.createFile(fs, filename, 1, replication, 0x0BAC);
fs.addCachePool(new CachePoolInfo("pool"));
long id = fs.addCacheDirective(
new CacheDirectiveInfo.Builder().setPool("pool").setPath(filename)
.setExpiration(CacheDirectiveInfo.Expiration.newRelative(86400000))
.setReplication(replication).build());
fs.modifyCacheDirective(new CacheDirectiveInfo.Builder()
.setId(id)
.setExpiration(CacheDirectiveInfo.Expiration.newRelative(172800000))
.build());
final NameNode sbn = dfsCluster.getNameNode(1);
final CacheManager annCachemanager = ann.getNamesystem().getCacheManager();
final CacheManager sbnCachemanager = sbn.getNamesystem().getCacheManager();
HATestUtil.waitForStandbyToCatchUp(ann, sbn);
GenericTestUtils.waitFor(() -> {
boolean isConsistence = false;
ann.getNamesystem().readLock();
try {
sbn.getNamesystem().readLock();
try {
Iterator<CacheDirective> annDirectivesIt = annCachemanager.
getCacheDirectives().iterator();
Iterator<CacheDirective> sbnDirectivesIt = sbnCachemanager.
getCacheDirectives().iterator();
if (annDirectivesIt.hasNext() && sbnDirectivesIt.hasNext()) {
CacheDirective annDirective = annDirectivesIt.next();
CacheDirective sbnDirective = sbnDirectivesIt.next();
if (annDirective.getExpiryTimeString().
equals(sbnDirective.getExpiryTimeString())) {
isConsistence = true;
}
}
} finally {
sbn.getNamesystem().readUnlock();
}
} finally {
ann.getNamesystem().readUnlock();
}
if (!isConsistence) {
LOG.info("testEexpiryTimeConsistency:"
+ "ANN CacheDirective Status is inconsistent with SBN");
}
return isConsistence;
}, 500, 120000);
}
} }