diff --git a/hadoop-tools/hadoop-azure/.gitignore b/hadoop-tools/hadoop-azure/.gitignore index 522210137e..a7d3296d0f 100644 --- a/hadoop-tools/hadoop-azure/.gitignore +++ b/hadoop-tools/hadoop-azure/.gitignore @@ -1,5 +1,6 @@ .checkstyle bin/ +src/test/resources/combinationConfigFiles src/test/resources/abfs-combination-test-configs.xml dev-support/testlogs src/test/resources/accountSettings/* diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AbfsConfiguration.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AbfsConfiguration.java index 0bcb97a849..e1bc8aa57b 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AbfsConfiguration.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AbfsConfiguration.java @@ -49,6 +49,7 @@ import org.apache.hadoop.fs.azurebfs.diagnostics.StringConfigurationBasicValidator; import org.apache.hadoop.fs.azurebfs.enums.Trilean; import org.apache.hadoop.fs.azurebfs.extensions.CustomTokenProviderAdaptee; +import org.apache.hadoop.fs.azurebfs.extensions.EncryptionContextProvider; import org.apache.hadoop.fs.azurebfs.extensions.SASTokenProvider; import org.apache.hadoop.fs.azurebfs.oauth2.AccessTokenProvider; import org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider; @@ -337,6 +338,10 @@ public class AbfsConfiguration{ FS_AZURE_ABFS_RENAME_RESILIENCE, DefaultValue = DEFAULT_ENABLE_ABFS_RENAME_RESILIENCE) private boolean renameResilience; + private String clientProvidedEncryptionKey; + + private String clientProvidedEncryptionKeySHA; + public AbfsConfiguration(final Configuration rawConfig, String accountName) throws IllegalAccessException, InvalidConfigurationValueException, IOException { this.rawConfig = ProviderUtils.excludeIncompatibleCredentialProviders( @@ -957,6 +962,32 @@ public SASTokenProvider getSASTokenProvider() throws AzureBlobFileSystemExceptio } } + public EncryptionContextProvider createEncryptionContextProvider() { + try { + String configKey = FS_AZURE_ENCRYPTION_CONTEXT_PROVIDER_TYPE; + if (get(configKey) == null) { + return null; + } + Class encryptionContextClass = + getAccountSpecificClass(configKey, null, + EncryptionContextProvider.class); + Preconditions.checkArgument(encryptionContextClass != null, String.format( + "The configuration value for %s is invalid, or config key is not account-specific", + configKey)); + + EncryptionContextProvider encryptionContextProvider = + ReflectionUtils.newInstance(encryptionContextClass, rawConfig); + Preconditions.checkArgument(encryptionContextProvider != null, + String.format("Failed to initialize %s", encryptionContextClass)); + + LOG.trace("{} init complete", encryptionContextClass.getName()); + return encryptionContextProvider; + } catch (Exception e) { + throw new IllegalArgumentException( + "Unable to load encryption context provider class: ", e); + } + } + public boolean isReadAheadEnabled() { return this.enabledReadAhead; } @@ -1068,9 +1099,22 @@ public boolean enableAbfsListIterator() { return this.enableAbfsListIterator; } - public String getClientProvidedEncryptionKey() { - String accSpecEncKey = accountConf(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY); - return rawConfig.get(accSpecEncKey, null); + public String getEncodedClientProvidedEncryptionKey() { + if (clientProvidedEncryptionKey == null) { + String accSpecEncKey = accountConf( + FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY); + clientProvidedEncryptionKey = rawConfig.get(accSpecEncKey, null); + } + return clientProvidedEncryptionKey; + } + + public String getEncodedClientProvidedEncryptionKeySHA() { + if (clientProvidedEncryptionKeySHA == null) { + String accSpecEncKey = accountConf( + FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY_SHA); + clientProvidedEncryptionKeySHA = rawConfig.get(accSpecEncKey, null); + } + return clientProvidedEncryptionKeySHA; } @VisibleForTesting diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AzureBlobFileSystemStore.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AzureBlobFileSystemStore.java index 4b356ceef0..2b7141e6cb 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AzureBlobFileSystemStore.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AzureBlobFileSystemStore.java @@ -55,7 +55,15 @@ import java.util.concurrent.TimeUnit; import org.apache.hadoop.classification.VisibleForTesting; +import org.apache.hadoop.fs.azurebfs.extensions.EncryptionContextProvider; +import org.apache.hadoop.fs.azurebfs.security.ContextProviderEncryptionAdapter; +import org.apache.hadoop.fs.azurebfs.security.ContextEncryptionAdapter; +import org.apache.hadoop.fs.azurebfs.security.NoContextEncryptionAdapter; +import org.apache.hadoop.fs.azurebfs.utils.EncryptionType; +import org.apache.hadoop.fs.azurebfs.utils.NamespaceUtil; import org.apache.hadoop.fs.impl.BackReference; +import org.apache.hadoop.fs.PathIOException; + import org.apache.hadoop.util.Preconditions; import org.apache.hadoop.thirdparty.com.google.common.base.Strings; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.Futures; @@ -149,6 +157,7 @@ import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.AZURE_ABFS_ENDPOINT; import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.FS_AZURE_BUFFERED_PREAD_DISABLE; import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.FS_AZURE_IDENTITY_TRANSFORM_CLASS; +import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_ENCRYPTION_CONTEXT; /** * Provides the bridging logic between Hadoop's abstract filesystem and Azure Storage. @@ -362,25 +371,7 @@ public boolean getIsNamespaceEnabled(TracingContext tracingContext) + " getAcl server call", e); } - LOG.debug("Get root ACL status"); - try (AbfsPerfInfo perfInfo = startTracking("getIsNamespaceEnabled", - "getAclStatus")) { - AbfsRestOperation op = client - .getAclStatus(AbfsHttpConstants.ROOT_PATH, tracingContext); - perfInfo.registerResult(op.getResult()); - isNamespaceEnabled = Trilean.getTrilean(true); - perfInfo.registerSuccess(true); - } catch (AbfsRestOperationException ex) { - // Get ACL status is a HEAD request, its response doesn't contain - // errorCode - // So can only rely on its status code to determine its account type. - if (HttpURLConnection.HTTP_BAD_REQUEST != ex.getStatusCode()) { - throw ex; - } - - isNamespaceEnabled = Trilean.getTrilean(false); - } - + isNamespaceEnabled = Trilean.getTrilean(NamespaceUtil.isNamespaceEnabled(client, tracingContext)); return isNamespaceEnabled.toBoolean(); } @@ -469,16 +460,22 @@ public void setFilesystemProperties( } public Hashtable getPathStatus(final Path path, - TracingContext tracingContext) throws AzureBlobFileSystemException { + TracingContext tracingContext) throws IOException { try (AbfsPerfInfo perfInfo = startTracking("getPathStatus", "getPathStatus")){ LOG.debug("getPathStatus for filesystem: {} path: {}", client.getFileSystem(), path); final Hashtable parsedXmsProperties; + final String relativePath = getRelativePath(path); + final ContextEncryptionAdapter contextEncryptionAdapter + = createEncryptionAdapterFromServerStoreContext(relativePath, + tracingContext); final AbfsRestOperation op = client - .getPathStatus(getRelativePath(path), true, tracingContext); + .getPathStatus(relativePath, true, tracingContext, + contextEncryptionAdapter); perfInfo.registerResult(op.getResult()); + contextEncryptionAdapter.destroy(); final String xMsProperties = op.getResult().getResponseHeader(HttpHeaderConfigurations.X_MS_PROPERTIES); @@ -490,9 +487,52 @@ public Hashtable getPathStatus(final Path path, } } + /** + * Creates an object of {@link ContextEncryptionAdapter} + * from a file path. It calls {@link org.apache.hadoop.fs.azurebfs.services.AbfsClient + * #getPathStatus(String, boolean, TracingContext, EncryptionAdapter)} method to get + * contextValue (x-ms-encryption-context) from the server. The contextValue is passed + * to the constructor of EncryptionAdapter to create the required object of + * EncryptionAdapter. + * @param path Path of the file for which the object of EncryptionAdapter is required. + * @return
    + *
  • + * {@link NoContextEncryptionAdapter}: if encryptionType is not of type + * {@link org.apache.hadoop.fs.azurebfs.utils.EncryptionType#ENCRYPTION_CONTEXT}. + *
  • + *
  • + * new object of {@link ContextProviderEncryptionAdapter} containing required encryptionKeys for the give file: + * if encryptionType is of type {@link org.apache.hadoop.fs.azurebfs.utils.EncryptionType#ENCRYPTION_CONTEXT}. + *
  • + *
+ */ + private ContextEncryptionAdapter createEncryptionAdapterFromServerStoreContext(final String path, + final TracingContext tracingContext) throws IOException { + if (client.getEncryptionType() != EncryptionType.ENCRYPTION_CONTEXT) { + return NoContextEncryptionAdapter.getInstance(); + } + final String responseHeaderEncryptionContext = client.getPathStatus(path, + false, tracingContext, null).getResult() + .getResponseHeader(X_MS_ENCRYPTION_CONTEXT); + if (responseHeaderEncryptionContext == null) { + throw new PathIOException(path, + "EncryptionContext not present in GetPathStatus response"); + } + byte[] encryptionContext = responseHeaderEncryptionContext.getBytes( + StandardCharsets.UTF_8); + + try { + return new ContextProviderEncryptionAdapter(client.getEncryptionContextProvider(), + new Path(path).toUri().getPath(), encryptionContext); + } catch (IOException e) { + LOG.debug("Could not initialize EncryptionAdapter"); + throw e; + } + } + public void setPathProperties(final Path path, final Hashtable properties, TracingContext tracingContext) - throws AzureBlobFileSystemException { + throws IOException { try (AbfsPerfInfo perfInfo = startTracking("setPathProperties", "setPathProperties")){ LOG.debug("setPathProperties for filesystem: {} path: {} with properties: {}", client.getFileSystem(), @@ -505,9 +545,14 @@ public void setPathProperties(final Path path, } catch (CharacterCodingException ex) { throw new InvalidAbfsRestOperationException(ex); } + final String relativePath = getRelativePath(path); + final ContextEncryptionAdapter contextEncryptionAdapter + = createEncryptionAdapterFromServerStoreContext(relativePath, + tracingContext); final AbfsRestOperation op = client .setPathProperties(getRelativePath(path), commaSeparatedProperties, - tracingContext); + tracingContext, contextEncryptionAdapter); + contextEncryptionAdapter.destroy(); perfInfo.registerResult(op.getResult()).registerSuccess(true); } } @@ -563,23 +608,30 @@ public OutputStream createFile(final Path path, triggerConditionalCreateOverwrite = true; } + final ContextEncryptionAdapter contextEncryptionAdapter; + if (client.getEncryptionType() == EncryptionType.ENCRYPTION_CONTEXT) { + contextEncryptionAdapter = new ContextProviderEncryptionAdapter( + client.getEncryptionContextProvider(), getRelativePath(path)); + } else { + contextEncryptionAdapter = NoContextEncryptionAdapter.getInstance(); + } AbfsRestOperation op; if (triggerConditionalCreateOverwrite) { op = conditionalCreateOverwriteFile(relativePath, statistics, - isNamespaceEnabled ? getOctalNotation(permission) : null, - isNamespaceEnabled ? getOctalNotation(umask) : null, + new Permissions(isNamespaceEnabled, permission, umask), isAppendBlob, + contextEncryptionAdapter, tracingContext ); } else { op = client.createPath(relativePath, true, overwrite, - isNamespaceEnabled ? getOctalNotation(permission) : null, - isNamespaceEnabled ? getOctalNotation(umask) : null, + new Permissions(isNamespaceEnabled, permission, umask), isAppendBlob, null, + contextEncryptionAdapter, tracingContext); } @@ -595,6 +647,7 @@ public OutputStream createFile(final Path path, statistics, relativePath, 0, + contextEncryptionAdapter, tracingContext)); } } @@ -604,32 +657,31 @@ public OutputStream createFile(final Path path, * only if there is match for eTag of existing file. * @param relativePath * @param statistics - * @param permission - * @param umask + * @param permissions contains permission and umask * @param isAppendBlob * @return * @throws AzureBlobFileSystemException */ private AbfsRestOperation conditionalCreateOverwriteFile(final String relativePath, final FileSystem.Statistics statistics, - final String permission, - final String umask, + final Permissions permissions, final boolean isAppendBlob, - TracingContext tracingContext) throws AzureBlobFileSystemException { + final ContextEncryptionAdapter contextEncryptionAdapter, + final TracingContext tracingContext) throws IOException { AbfsRestOperation op; try { // Trigger a create with overwrite=false first so that eTag fetch can be // avoided for cases when no pre-existing file is present (major portion // of create file traffic falls into the case of no pre-existing file). - op = client.createPath(relativePath, true, false, permission, umask, - isAppendBlob, null, tracingContext); + op = client.createPath(relativePath, true, false, permissions, + isAppendBlob, null, contextEncryptionAdapter, tracingContext); } catch (AbfsRestOperationException e) { if (e.getStatusCode() == HttpURLConnection.HTTP_CONFLICT) { // File pre-exists, fetch eTag try { - op = client.getPathStatus(relativePath, false, tracingContext); + op = client.getPathStatus(relativePath, false, tracingContext, null); } catch (AbfsRestOperationException ex) { if (ex.getStatusCode() == HttpURLConnection.HTTP_NOT_FOUND) { // Is a parallel access case, as file which was found to be @@ -647,8 +699,8 @@ private AbfsRestOperation conditionalCreateOverwriteFile(final String relativePa try { // overwrite only if eTag matches with the file properties fetched befpre - op = client.createPath(relativePath, true, true, permission, umask, - isAppendBlob, eTag, tracingContext); + op = client.createPath(relativePath, true, true, permissions, + isAppendBlob, eTag, contextEncryptionAdapter, tracingContext); } catch (AbfsRestOperationException ex) { if (ex.getStatusCode() == HttpURLConnection.HTTP_PRECON_FAILED) { // Is a parallel access case, as file with eTag was just queried @@ -691,6 +743,7 @@ private AbfsOutputStreamContext populateAbfsOutputStreamContext( FileSystem.Statistics statistics, String path, long position, + ContextEncryptionAdapter contextEncryptionAdapter, TracingContext tracingContext) { int bufferSize = abfsConfiguration.getWriteBufferSize(); if (isAppendBlob && bufferSize > FileSystemConfigurations.APPENDBLOB_MAX_WRITE_BUFFER_SIZE) { @@ -707,6 +760,7 @@ private AbfsOutputStreamContext populateAbfsOutputStreamContext( .withWriteMaxConcurrentRequestCount(abfsConfiguration.getWriteMaxConcurrentRequestCount()) .withMaxWriteRequestsToQueue(abfsConfiguration.getMaxWriteRequestsToQueue()) .withLease(lease) + .withEncryptionAdapter(contextEncryptionAdapter) .withBlockFactory(getBlockFactory()) .withBlockOutputActiveBlocks(blockOutputActiveBlocks) .withClient(client) @@ -722,7 +776,7 @@ private AbfsOutputStreamContext populateAbfsOutputStreamContext( public void createDirectory(final Path path, final FsPermission permission, final FsPermission umask, TracingContext tracingContext) - throws AzureBlobFileSystemException { + throws IOException { try (AbfsPerfInfo perfInfo = startTracking("createDirectory", "createPath")) { boolean isNamespaceEnabled = getIsNamespaceEnabled(tracingContext); LOG.debug("createDirectory filesystem: {} path: {} permission: {} umask: {} isNamespaceEnabled: {}", @@ -734,11 +788,10 @@ public void createDirectory(final Path path, final FsPermission permission, boolean overwrite = !isNamespaceEnabled || abfsConfiguration.isEnabledMkdirOverwrite(); + Permissions permissions = new Permissions(isNamespaceEnabled, + permission, umask); final AbfsRestOperation op = client.createPath(getRelativePath(path), - false, overwrite, - isNamespaceEnabled ? getOctalNotation(permission) : null, - isNamespaceEnabled ? getOctalNotation(umask) : null, false, null, - tracingContext); + false, overwrite, permissions, false, null, null, tracingContext); perfInfo.registerResult(op.getResult()).registerSuccess(true); } } @@ -764,7 +817,19 @@ public AbfsInputStream openFileForRead(Path path, String relativePath = getRelativePath(path); String resourceType, eTag; long contentLength; - if (fileStatus instanceof VersionedFileStatus) { + ContextEncryptionAdapter contextEncryptionAdapter = NoContextEncryptionAdapter.getInstance(); + /* + * GetPathStatus API has to be called in case of: + * 1. fileStatus is null or not an object of VersionedFileStatus: as eTag + * would not be there in the fileStatus object. + * 2. fileStatus is an object of VersionedFileStatus and the object doesn't + * have encryptionContext field when client's encryptionType is + * ENCRYPTION_CONTEXT. + */ + if ((fileStatus instanceof VersionedFileStatus) && ( + client.getEncryptionType() != EncryptionType.ENCRYPTION_CONTEXT + || ((VersionedFileStatus) fileStatus).getEncryptionContext() + != null)) { path = path.makeQualified(this.uri, path); Preconditions.checkArgument(fileStatus.getPath().equals(path), String.format( @@ -773,19 +838,37 @@ public AbfsInputStream openFileForRead(Path path, resourceType = fileStatus.isFile() ? FILE : DIRECTORY; contentLength = fileStatus.getLen(); eTag = ((VersionedFileStatus) fileStatus).getVersion(); - } else { - if (fileStatus != null) { - LOG.debug( - "Fallback to getPathStatus REST call as provided filestatus " - + "is not of type VersionedFileStatus"); + final String encryptionContext + = ((VersionedFileStatus) fileStatus).getEncryptionContext(); + if (client.getEncryptionType() == EncryptionType.ENCRYPTION_CONTEXT) { + contextEncryptionAdapter = new ContextProviderEncryptionAdapter( + client.getEncryptionContextProvider(), getRelativePath(path), + encryptionContext.getBytes(StandardCharsets.UTF_8)); } + } else { AbfsHttpOperation op = client.getPathStatus(relativePath, false, - tracingContext).getResult(); + tracingContext, null).getResult(); resourceType = op.getResponseHeader( HttpHeaderConfigurations.X_MS_RESOURCE_TYPE); contentLength = Long.parseLong( op.getResponseHeader(HttpHeaderConfigurations.CONTENT_LENGTH)); eTag = op.getResponseHeader(HttpHeaderConfigurations.ETAG); + /* + * For file created with ENCRYPTION_CONTEXT, client shall receive + * encryptionContext from header field: X_MS_ENCRYPTION_CONTEXT. + */ + if (client.getEncryptionType() == EncryptionType.ENCRYPTION_CONTEXT) { + final String fileEncryptionContext = op.getResponseHeader( + HttpHeaderConfigurations.X_MS_ENCRYPTION_CONTEXT); + if (fileEncryptionContext == null) { + LOG.debug("EncryptionContext missing in GetPathStatus response"); + throw new PathIOException(path.toString(), + "EncryptionContext not present in GetPathStatus response headers"); + } + contextEncryptionAdapter = new ContextProviderEncryptionAdapter( + client.getEncryptionContextProvider(), getRelativePath(path), + fileEncryptionContext.getBytes(StandardCharsets.UTF_8)); + } } if (parseIsDirectory(resourceType)) { @@ -801,13 +884,14 @@ public AbfsInputStream openFileForRead(Path path, // Add statistics for InputStream return new AbfsInputStream(client, statistics, relativePath, contentLength, populateAbfsInputStreamContext( - parameters.map(OpenFileParameters::getOptions)), + parameters.map(OpenFileParameters::getOptions), + contextEncryptionAdapter), eTag, tracingContext); } } private AbfsInputStreamContext populateAbfsInputStreamContext( - Optional options) { + Optional options, ContextEncryptionAdapter contextEncryptionAdapter) { boolean bufferedPreadDisabled = options .map(c -> c.getBoolean(FS_AZURE_BUFFERED_PREAD_DISABLE, false)) .orElse(false); @@ -824,6 +908,7 @@ private AbfsInputStreamContext populateAbfsInputStreamContext( abfsConfiguration.shouldReadBufferSizeAlways()) .withReadAheadBlockSize(abfsConfiguration.getReadAheadBlockSize()) .withBufferedPreadDisabled(bufferedPreadDisabled) + .withEncryptionAdapter(contextEncryptionAdapter) .withAbfsBackRef(fsBackRef) .build(); } @@ -840,7 +925,7 @@ public OutputStream openFileForWrite(final Path path, String relativePath = getRelativePath(path); final AbfsRestOperation op = client - .getPathStatus(relativePath, false, tracingContext); + .getPathStatus(relativePath, false, tracingContext, null); perfInfo.registerResult(op.getResult()); final String resourceType = op.getResult().getResponseHeader(HttpHeaderConfigurations.X_MS_RESOURCE_TYPE); @@ -864,6 +949,21 @@ public OutputStream openFileForWrite(final Path path, } AbfsLease lease = maybeCreateLease(relativePath, tracingContext); + final ContextEncryptionAdapter contextEncryptionAdapter; + if (client.getEncryptionType() == EncryptionType.ENCRYPTION_CONTEXT) { + final String encryptionContext = op.getResult() + .getResponseHeader( + HttpHeaderConfigurations.X_MS_ENCRYPTION_CONTEXT); + if (encryptionContext == null) { + throw new PathIOException(path.toString(), + "File doesn't have encryptionContext."); + } + contextEncryptionAdapter = new ContextProviderEncryptionAdapter( + client.getEncryptionContextProvider(), getRelativePath(path), + encryptionContext.getBytes(StandardCharsets.UTF_8)); + } else { + contextEncryptionAdapter = NoContextEncryptionAdapter.getInstance(); + } return new AbfsOutputStream( populateAbfsOutputStreamContext( @@ -873,6 +973,7 @@ public OutputStream openFileForWrite(final Path path, statistics, relativePath, offset, + contextEncryptionAdapter, tracingContext)); } } @@ -906,7 +1007,7 @@ public boolean rename(final Path source, final Path destination, final TracingContext tracingContext, final String sourceEtag) throws - AzureBlobFileSystemException { + IOException { final Instant startAggregate = abfsPerfTracker.getLatencyInstant(); long countAggregate = 0; boolean shouldContinue; @@ -1005,7 +1106,7 @@ public FileStatus getFileStatus(final Path path, } } else { perfInfo.registerCallee("getPathStatus"); - op = client.getPathStatus(getRelativePath(path), false, tracingContext); + op = client.getPathStatus(getRelativePath(path), false, tracingContext, null); } perfInfo.registerResult(op.getResult()); @@ -1015,6 +1116,7 @@ public FileStatus getFileStatus(final Path path, String eTag = extractEtagHeader(result); final String lastModified = result.getResponseHeader(HttpHeaderConfigurations.LAST_MODIFIED); final String permissions = result.getResponseHeader((HttpHeaderConfigurations.X_MS_PERMISSIONS)); + final String encryptionContext = op.getResult().getResponseHeader(X_MS_ENCRYPTION_CONTEXT); final boolean hasAcl = AbfsPermission.isExtendedAcl(permissions); final long contentLength; final boolean resourceIsDir; @@ -1051,7 +1153,8 @@ public FileStatus getFileStatus(final Path path, blockSize, DateTimeUtils.parseLastModifiedTime(lastModified), path, - eTag); + eTag, + encryptionContext); } } @@ -1129,6 +1232,7 @@ public String listStatus(final Path path, final String startFrom, for (ListResultEntrySchema entry : retrievedSchema.paths()) { final String owner = identityTransformer.transformIdentityForGetRequest(entry.owner(), true, userName); final String group = identityTransformer.transformIdentityForGetRequest(entry.group(), false, primaryUserGroup); + final String encryptionContext = entry.getXMsEncryptionContext(); final FsPermission fsPermission = entry.permissions() == null ? new AbfsPermission(FsAction.ALL, FsAction.ALL, FsAction.ALL) : AbfsPermission.valueOf(entry.permissions()); @@ -1157,7 +1261,8 @@ public String listStatus(final Path path, final String startFrom, blockSize, lastModifiedMillis, entryPath, - entry.eTag())); + entry.eTag(), + encryptionContext)); } perfInfo.registerSuccess(true); @@ -1627,16 +1732,38 @@ private void initializeClient(URI uri, String fileSystemName, abfsConfiguration.getRawConfiguration()); } + // Encryption setup + EncryptionContextProvider encryptionContextProvider = null; + if (isSecure) { + encryptionContextProvider = + abfsConfiguration.createEncryptionContextProvider(); + if (encryptionContextProvider != null) { + if (abfsConfiguration.getEncodedClientProvidedEncryptionKey() != null) { + throw new PathIOException(uri.getPath(), + "Both global key and encryption context are set, only one allowed"); + } + encryptionContextProvider.initialize( + abfsConfiguration.getRawConfiguration(), accountName, + fileSystemName); + } else if (abfsConfiguration.getEncodedClientProvidedEncryptionKey() != null) { + if (abfsConfiguration.getEncodedClientProvidedEncryptionKeySHA() == null) { + throw new PathIOException(uri.getPath(), + "Encoded SHA256 hash must be provided for global encryption"); + } + } + } + LOG.trace("Initializing AbfsClient for {}", baseUrl); if (tokenProvider != null) { this.client = new AbfsClient(baseUrl, creds, abfsConfiguration, - tokenProvider, + tokenProvider, encryptionContextProvider, populateAbfsClientContext()); } else { this.client = new AbfsClient(baseUrl, creds, abfsConfiguration, - sasTokenProvider, + sasTokenProvider, encryptionContextProvider, populateAbfsClientContext()); } + LOG.trace("AbfsClient init complete"); } @@ -1654,12 +1781,7 @@ private AbfsClientContext populateAbfsClientContext() { .build(); } - private String getOctalNotation(FsPermission fsPermission) { - Preconditions.checkNotNull(fsPermission, "fsPermission"); - return String.format(AbfsHttpConstants.PERMISSION_FORMAT, fsPermission.toOctal()); - } - - private String getRelativePath(final Path path) { + public String getRelativePath(final Path path) { Preconditions.checkNotNull(path, "path"); String relPath = path.toUri().getPath(); if (relPath.isEmpty()) { @@ -1682,7 +1804,14 @@ private boolean parseIsDirectory(final String resourceType) { && resourceType.equalsIgnoreCase(AbfsHttpConstants.DIRECTORY); } - private String convertXmsPropertiesToCommaSeparatedString(final Hashtable properties) throws + /** + * Convert properties stored in a Map into a comma separated string. For map + * , method would convert to: + * key1=value1,key2=value,...,keyN=valueN + * */ + @VisibleForTesting + String convertXmsPropertiesToCommaSeparatedString(final Map properties) throws CharacterCodingException { StringBuilder commaSeparatedProperties = new StringBuilder(); @@ -1780,7 +1909,7 @@ private AbfsPerfInfo startTracking(String callerName, String calleeName) { * in a LIST or HEAD request. * The etag is included in the java serialization. */ - private static final class VersionedFileStatus extends FileStatus + static final class VersionedFileStatus extends FileStatus implements EtagSource { /** @@ -1795,11 +1924,13 @@ private static final class VersionedFileStatus extends FileStatus */ private String version; + private String encryptionContext; + private VersionedFileStatus( final String owner, final String group, final FsPermission fsPermission, final boolean hasAcl, final long length, final boolean isdir, final int blockReplication, final long blocksize, final long modificationTime, final Path path, - String version) { + final String version, final String encryptionContext) { super(length, isdir, blockReplication, blocksize, modificationTime, 0, fsPermission, owner, @@ -1809,6 +1940,7 @@ private VersionedFileStatus( hasAcl, false, false); this.version = version; + this.encryptionContext = encryptionContext; } /** Compare if this object is equal to another object. @@ -1861,6 +1993,10 @@ public String getEtag() { return getVersion(); } + public String getEncryptionContext() { + return encryptionContext; + } + @Override public String toString() { final StringBuilder sb = new StringBuilder( @@ -1872,6 +2008,54 @@ public String toString() { } } + /** + * Permissions class contain provided permission and umask in octalNotation. + * If the object is created for namespace-disabled account, the permission and + * umask would be null. + * */ + public static final class Permissions { + private final String permission; + private final String umask; + + Permissions(boolean isNamespaceEnabled, FsPermission permission, + FsPermission umask) { + if (isNamespaceEnabled) { + this.permission = getOctalNotation(permission); + this.umask = getOctalNotation(umask); + } else { + this.permission = null; + this.umask = null; + } + } + + private String getOctalNotation(FsPermission fsPermission) { + Preconditions.checkNotNull(fsPermission, "fsPermission"); + return String.format(AbfsHttpConstants.PERMISSION_FORMAT, fsPermission.toOctal()); + } + + public Boolean hasPermission() { + return permission != null && !permission.isEmpty(); + } + + public Boolean hasUmask() { + return umask != null && !umask.isEmpty(); + } + + public String getPermission() { + return permission; + } + + public String getUmask() { + return umask; + } + + @Override + public String toString() { + return String.format("{\"permission\":%s, \"umask\":%s}", permission, + umask); + } + } + /** * A builder class for AzureBlobFileSystemStore. */ diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/AbfsHttpConstants.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/AbfsHttpConstants.java index 7e4ddfa675..91f6bddcc1 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/AbfsHttpConstants.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/AbfsHttpConstants.java @@ -119,6 +119,9 @@ public final class AbfsHttpConstants { public static final char CHAR_EQUALS = '='; public static final char CHAR_STAR = '*'; public static final char CHAR_PLUS = '+'; + public static final String DECEMBER_2019_API_VERSION = "2019-12-12"; + public static final String APRIL_2021_API_VERSION = "2021-04-10"; + /** * Value that differentiates categories of the http_status. *
diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/ConfigurationKeys.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/ConfigurationKeys.java
index 461e43a9f7..91f9eff532 100644
--- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/ConfigurationKeys.java
+++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/ConfigurationKeys.java
@@ -203,8 +203,14 @@ public final class ConfigurationKeys {
 
   /** Setting this true will make the driver use it's own RemoteIterator implementation */
   public static final String FS_AZURE_ENABLE_ABFS_LIST_ITERATOR = "fs.azure.enable.abfslistiterator";
-  /** Server side encryption key */
-  public static final String FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY = "fs.azure.client-provided-encryption-key";
+  /** Server side encryption key encoded in Base6format {@value}.*/
+  public static final String FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY =
+      "fs.azure.encryption.encoded.client-provided-key";
+  /** SHA256 hash of encryption key encoded in Base64format */
+  public static final String FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY_SHA =
+      "fs.azure.encryption.encoded.client-provided-key-sha";
+  /** Custom EncryptionContextProvider type */
+  public static final String FS_AZURE_ENCRYPTION_CONTEXT_PROVIDER_TYPE = "fs.azure.encryption.context.provider.type";
 
   /** End point of ABFS account: {@value}. */
   public static final String AZURE_ABFS_ENDPOINT = "fs.azure.abfs.endpoint";
diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/HttpHeaderConfigurations.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/HttpHeaderConfigurations.java
index b123e90170..c792e463c7 100644
--- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/HttpHeaderConfigurations.java
+++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/HttpHeaderConfigurations.java
@@ -65,6 +65,7 @@ public final class HttpHeaderConfigurations {
   public static final String X_MS_ENCRYPTION_ALGORITHM = "x-ms-encryption-algorithm";
   public static final String X_MS_REQUEST_SERVER_ENCRYPTED = "x-ms-request-server-encrypted";
   public static final String X_MS_SERVER_ENCRYPTED = "x-ms-server-encrypted";
+  public static final String X_MS_ENCRYPTION_CONTEXT = "x-ms-encryption-context";
   public static final String X_MS_LEASE_ACTION = "x-ms-lease-action";
   public static final String X_MS_LEASE_DURATION = "x-ms-lease-duration";
   public static final String X_MS_LEASE_ID = "x-ms-lease-id";
diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/contracts/services/ListResultEntrySchema.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/contracts/services/ListResultEntrySchema.java
index a9883dd2ce..77f52e4a2b 100644
--- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/contracts/services/ListResultEntrySchema.java
+++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/contracts/services/ListResultEntrySchema.java
@@ -77,6 +77,18 @@ public class ListResultEntrySchema {
   @JsonProperty(value = "permissions")
   private String permissions;
 
+  /**
+   *  The encryption context property
+   */
+  @JsonProperty(value = "EncryptionContext")
+  private String xMsEncryptionContext;
+
+  /**
+   * The customer-provided encryption-256 value
+   * */
+  @JsonProperty(value = "CustomerProvidedKeySha256")
+  private String customerProvidedKeySha256;
+
   /**
    * Get the name value.
    *
@@ -238,4 +250,19 @@ public ListResultEntrySchema withPermissions(final String permissions) {
     return this;
   }
 
-}
\ No newline at end of file
+  /**
+   * Get the x-ms-encryption-context value.
+   * @return the x-ms-encryption-context value.
+   * */
+  public String getXMsEncryptionContext() {
+    return xMsEncryptionContext;
+  }
+
+  /**
+   * Get the customer-provided sha-256 key
+   * @return the x-ms-encryption-key-sha256 value used by client.
+   * */
+  public String getCustomerProvidedKeySha256() {
+    return customerProvidedKeySha256;
+  }
+}
diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/extensions/EncryptionContextProvider.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/extensions/EncryptionContextProvider.java
new file mode 100644
index 0000000000..20825228ae
--- /dev/null
+++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/extensions/EncryptionContextProvider.java
@@ -0,0 +1,67 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.fs.azurebfs.extensions;
+
+import javax.security.auth.Destroyable;
+import java.io.IOException;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.azurebfs.security.ABFSKey;
+
+/**
+ * This interface has two roles:
+ *
    + *
  • + * To create new encryptionContext from a given path: To be used in case of + * create file as there is no encryptionContext in remote server to refer to + * for encryptionKey creation. + *
  • + *
  • To create encryptionKey using encryptionContext.
  • + *
+ */ +public interface EncryptionContextProvider extends Destroyable { + /** + * Initialize instance. + * + * @param configuration rawConfig instance + * @param accountName Account Name (with domain) + * @param fileSystem container name + * @throws IOException error in initialization + */ + void initialize(Configuration configuration, String accountName, String fileSystem) throws IOException; + + /** + * Fetch encryption context for a given path. + * + * @param path file path from filesystem root + * @return encryptionContext key + * @throws IOException error in fetching encryption context + */ + ABFSKey getEncryptionContext(String path) throws IOException; + + /** + * Fetch encryption key in-exchange for encryption context. + * + * @param path file path from filesystem root + * @param encryptionContext encryptionContext fetched from server + * @return Encryption key + * @throws IOException error in fetching encryption key + */ + ABFSKey getEncryptionKey(String path, ABFSKey encryptionContext) throws IOException; +} diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ABFSKey.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ABFSKey.java new file mode 100644 index 0000000000..92bb89a482 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ABFSKey.java @@ -0,0 +1,64 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + *

+ * http://www.apache.org/licenses/LICENSE-2.0 + *

+ * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.security; + +import javax.crypto.SecretKey; +import java.util.Arrays; + +/** + * Implementation of SecretKey that would be used by EncryptionAdapter object, + * implementations of encryptionContextProvider to maintain the byteArrays of + * encryptionContext and encryptionKey. + */ +public final class ABFSKey implements SecretKey { + private byte[] bytes; + + public ABFSKey(byte[] bytes) { + if (bytes != null) { + this.bytes = bytes.clone(); + } + } + + @Override + public String getAlgorithm() { + return null; + } + + @Override + public String getFormat() { + return null; + } + + /** + * This method to be called by implementations of EncryptionContextProvider interface. + * Method returns clone of the original bytes array to prevent findbugs flags. + */ + @Override + public byte[] getEncoded() { + if (bytes == null) { + return null; + } + return bytes.clone(); + } + + @Override + public void destroy() { + Arrays.fill(bytes, (byte) 0); + } +} diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ContextEncryptionAdapter.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ContextEncryptionAdapter.java new file mode 100644 index 0000000000..575da319cf --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ContextEncryptionAdapter.java @@ -0,0 +1,45 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.security; + +/** + * Provides APIs to get encryptionKey from encryptionContext for a given path. + */ +public abstract class ContextEncryptionAdapter { + + /** + * @return computed encryptionKey from server provided encryptionContext + */ + public abstract String getEncodedKey(); + + /** + * @return computed encryptionKeySHA from server provided encryptionContext + */ + public abstract String getEncodedKeySHA(); + + /** + * @return encryptionContext to be supplied in createPath API + */ + public abstract String getEncodedContext(); + + /** + * Destroys all the encapsulated fields which are used for creating keys. + */ + public abstract void destroy(); +} diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ContextProviderEncryptionAdapter.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ContextProviderEncryptionAdapter.java new file mode 100644 index 0000000000..9b5b39c2f1 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/ContextProviderEncryptionAdapter.java @@ -0,0 +1,121 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.security; + +import java.io.IOException; +import java.util.Arrays; +import java.util.Base64; +import java.util.Objects; + +import org.apache.hadoop.fs.azurebfs.extensions.EncryptionContextProvider; + +import static org.apache.hadoop.fs.azurebfs.security.EncodingHelper.getBase64EncodedString; + +/** + * Class manages the encryptionContext and encryptionKey that needs to be added + * to the headers to server request if Customer-Encryption-Context is enabled in + * the configuration. + *
+ * For fileCreation, the object helps in creating encryptionContext through the + * implementation of EncryptionContextProvider. + *
+ * For all operations, the object helps in converting encryptionContext to + * encryptionKey through the implementation of EncryptionContextProvider. + */ +public class ContextProviderEncryptionAdapter extends ContextEncryptionAdapter { + private final String path; + private final ABFSKey encryptionContext; + private ABFSKey encryptionKey; + private final EncryptionContextProvider provider; + + /** + * Following constructor called when the encryptionContext of file is known. + * The server shall send encryptionContext as a String, the constructor shall + * convert the string into a byte-array. The converted byte-array would be used + * by the implementation of EncryptionContextProvider to create byte-array of + * encryptionKey. + * @param provider developer's implementation of {@link EncryptionContextProvider} + * @param path Path for which encryptionContext and encryptionKeys to be stored + * in the object + * @param encryptionContext encryptionContext for the path stored in the backend + * @throws IOException throws back the exception it receives from the + * {@link ContextProviderEncryptionAdapter#computeKeys()} method call. + */ + public ContextProviderEncryptionAdapter(EncryptionContextProvider provider, String path, + byte[] encryptionContext) throws IOException { + this.provider = provider; + this.path = path; + Objects.requireNonNull(encryptionContext, + "Encryption context should not be null."); + this.encryptionContext = new ABFSKey(Base64.getDecoder().decode(encryptionContext)); + Arrays.fill(encryptionContext, (byte) 0); + computeKeys(); + } + + /** + * Following constructor called in case of createPath. Since, the path is not + * on the server, encryptionContext is not there for the path. Implementation + * of the EncryptionContextProvider would be used to create encryptionContext + * from the path. + * @param provider developer's implementation of {@link EncryptionContextProvider} + * @param path file path for which encryptionContext and encryptionKeys to be + * created and stored + * @throws IOException throws back the exception it receives from the method call + * to {@link EncryptionContextProvider} object. + */ + public ContextProviderEncryptionAdapter(EncryptionContextProvider provider, String path) + throws IOException { + this.provider = provider; + this.path = path; + encryptionContext = provider.getEncryptionContext(path); + Objects.requireNonNull(encryptionContext, + "Encryption context should not be null."); + computeKeys(); + } + + private void computeKeys() throws IOException { + encryptionKey = provider.getEncryptionKey(path, encryptionContext); + Objects.requireNonNull(encryptionKey, "Encryption key should not be null."); + } + + @Override + public String getEncodedKey() { + return getBase64EncodedString(encryptionKey.getEncoded()); + } + + @Override + public String getEncodedKeySHA() { + return getBase64EncodedString(EncodingHelper.getSHA256Hash(encryptionKey.getEncoded())); + } + + @Override + public String getEncodedContext() { + return getBase64EncodedString(encryptionContext.getEncoded()); + } + + @Override + public void destroy() { + if (encryptionContext != null) { + encryptionContext.destroy(); + } + if (encryptionKey != null) { + encryptionKey.destroy(); + } + } +} diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/EncodingHelper.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/EncodingHelper.java new file mode 100644 index 0000000000..0f9ecf0294 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/EncodingHelper.java @@ -0,0 +1,50 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.security; + +import java.security.MessageDigest; +import java.security.NoSuchAlgorithmException; +import java.util.Base64; + +/** + * Utility class for managing encryption of bytes or base64String conversion of + * bytes. + */ +public final class EncodingHelper { + + private EncodingHelper() { + + } + + public static byte[] getSHA256Hash(byte[] key) { + try { + final MessageDigest digester = MessageDigest.getInstance("SHA-256"); + return digester.digest(key); + } catch (NoSuchAlgorithmException noSuchAlgorithmException) { + /*This exception can be ignored. Reason being SHA-256 is a valid algorithm, + and it is constant for all method calls.*/ + throw new RuntimeException("SHA-256 algorithm not found in MessageDigest", + noSuchAlgorithmException); + } + } + + public static String getBase64EncodedString(byte[] bytes) { + return Base64.getEncoder().encodeToString(bytes); + } +} diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/NoContextEncryptionAdapter.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/NoContextEncryptionAdapter.java new file mode 100644 index 0000000000..d6638c99e9 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/NoContextEncryptionAdapter.java @@ -0,0 +1,52 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.security; + +public final class NoContextEncryptionAdapter extends ContextEncryptionAdapter { + + private NoContextEncryptionAdapter() { + + } + private static final NoContextEncryptionAdapter + INSTANCE = new NoContextEncryptionAdapter(); + + public static NoContextEncryptionAdapter getInstance() { + return INSTANCE; + } + + @Override + public String getEncodedKey() { + return null; + } + + @Override + public String getEncodedKeySHA() { + return null; + } + + @Override + public String getEncodedContext() { + return null; + } + + @Override + public void destroy() { + + } +} diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsClient.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsClient.java index 9c1f590da9..8eeb548f50 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsClient.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsClient.java @@ -25,11 +25,7 @@ import java.net.MalformedURLException; import java.net.URL; import java.net.URLEncoder; -import java.nio.charset.StandardCharsets; -import java.security.MessageDigest; -import java.security.NoSuchAlgorithmException; import java.util.ArrayList; -import java.util.Base64; import java.util.List; import java.util.Locale; import java.util.UUID; @@ -38,8 +34,13 @@ import java.util.concurrent.TimeUnit; import org.apache.hadoop.classification.VisibleForTesting; +import org.apache.hadoop.fs.azurebfs.utils.NamespaceUtil; import org.apache.hadoop.fs.store.LogExactlyOnce; -import org.apache.hadoop.util.Preconditions; +import org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.Permissions; +import org.apache.hadoop.fs.azurebfs.extensions.EncryptionContextProvider; +import org.apache.hadoop.fs.azurebfs.security.ContextEncryptionAdapter; +import org.apache.hadoop.fs.azurebfs.utils.EncryptionType; +import org.apache.hadoop.io.IOUtils; import org.apache.hadoop.thirdparty.com.google.common.base.Strings; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.FutureCallback; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.Futures; @@ -48,6 +49,7 @@ import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListeningScheduledExecutorService; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.MoreExecutors; import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; +import org.apache.hadoop.util.Preconditions; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -65,7 +67,6 @@ import org.apache.hadoop.fs.azurebfs.contracts.services.AppendRequestParameters; import org.apache.hadoop.fs.azurebfs.oauth2.AccessTokenProvider; import org.apache.hadoop.fs.azurebfs.utils.TracingContext; -import org.apache.hadoop.io.IOUtils; import org.apache.hadoop.security.ssl.DelegatingSSLSocketFactory; import org.apache.hadoop.util.concurrent.HadoopExecutors; @@ -91,23 +92,27 @@ public class AbfsClient implements Closeable { private final URL baseUrl; private final SharedKeyCredentials sharedKeyCredentials; - private final String xMsVersion = "2019-12-12"; + private String xMsVersion = DECEMBER_2019_API_VERSION; private final ExponentialRetryPolicy retryPolicy; private final String filesystem; private final AbfsConfiguration abfsConfiguration; private final String userAgent; private final AbfsPerfTracker abfsPerfTracker; - private final String clientProvidedEncryptionKey; - private final String clientProvidedEncryptionKeySHA; + private String clientProvidedEncryptionKey = null; + private String clientProvidedEncryptionKeySHA = null; private final String accountName; private final AuthType authType; private AccessTokenProvider tokenProvider; private SASTokenProvider sasTokenProvider; private final AbfsCounters abfsCounters; + private EncryptionContextProvider encryptionContextProvider = null; + private EncryptionType encryptionType = EncryptionType.NONE; private final AbfsThrottlingIntercept intercept; private final ListeningScheduledExecutorService executorService; + private Boolean isNamespaceEnabled; + private boolean renameResilience; @@ -116,10 +121,11 @@ public class AbfsClient implements Closeable { */ private static final LogExactlyOnce ABFS_METADATA_INCOMPLETE_RENAME_FAILURE = new LogExactlyOnce(LOG); - private AbfsClient(final URL baseUrl, final SharedKeyCredentials sharedKeyCredentials, - final AbfsConfiguration abfsConfiguration, - final AbfsClientContext abfsClientContext) - throws IOException { + private AbfsClient(final URL baseUrl, + final SharedKeyCredentials sharedKeyCredentials, + final AbfsConfiguration abfsConfiguration, + final EncryptionContextProvider encryptionContextProvider, + final AbfsClientContext abfsClientContext) throws IOException { this.baseUrl = baseUrl; this.sharedKeyCredentials = sharedKeyCredentials; String baseUrlString = baseUrl.toString(); @@ -131,15 +137,16 @@ private AbfsClient(final URL baseUrl, final SharedKeyCredentials sharedKeyCreden this.intercept = AbfsThrottlingInterceptFactory.getInstance(accountName, abfsConfiguration); this.renameResilience = abfsConfiguration.getRenameResilience(); - String encryptionKey = this.abfsConfiguration - .getClientProvidedEncryptionKey(); - if (encryptionKey != null) { - this.clientProvidedEncryptionKey = getBase64EncodedString(encryptionKey); - this.clientProvidedEncryptionKeySHA = getBase64EncodedString( - getSHA256Hash(encryptionKey)); - } else { - this.clientProvidedEncryptionKey = null; - this.clientProvidedEncryptionKeySHA = null; + if (encryptionContextProvider != null) { + this.encryptionContextProvider = encryptionContextProvider; + xMsVersion = APRIL_2021_API_VERSION; // will be default once server change deployed + encryptionType = EncryptionType.ENCRYPTION_CONTEXT; + } else if (abfsConfiguration.getEncodedClientProvidedEncryptionKey() != null) { + clientProvidedEncryptionKey = + abfsConfiguration.getEncodedClientProvidedEncryptionKey(); + this.clientProvidedEncryptionKeySHA = + abfsConfiguration.getEncodedClientProvidedEncryptionKeySHA(); + encryptionType = EncryptionType.GLOBAL_KEY; } String sslProviderName = null; @@ -170,42 +177,30 @@ private AbfsClient(final URL baseUrl, final SharedKeyCredentials sharedKeyCreden public AbfsClient(final URL baseUrl, final SharedKeyCredentials sharedKeyCredentials, final AbfsConfiguration abfsConfiguration, final AccessTokenProvider tokenProvider, + final EncryptionContextProvider encryptionContextProvider, final AbfsClientContext abfsClientContext) throws IOException { - this(baseUrl, sharedKeyCredentials, abfsConfiguration, abfsClientContext); + this(baseUrl, sharedKeyCredentials, abfsConfiguration, + encryptionContextProvider, abfsClientContext); this.tokenProvider = tokenProvider; } public AbfsClient(final URL baseUrl, final SharedKeyCredentials sharedKeyCredentials, final AbfsConfiguration abfsConfiguration, final SASTokenProvider sasTokenProvider, + final EncryptionContextProvider encryptionContextProvider, final AbfsClientContext abfsClientContext) throws IOException { - this(baseUrl, sharedKeyCredentials, abfsConfiguration, abfsClientContext); + this(baseUrl, sharedKeyCredentials, abfsConfiguration, + encryptionContextProvider, abfsClientContext); this.sasTokenProvider = sasTokenProvider; } - private byte[] getSHA256Hash(String key) throws IOException { - try { - final MessageDigest digester = MessageDigest.getInstance("SHA-256"); - return digester.digest(key.getBytes(StandardCharsets.UTF_8)); - } catch (NoSuchAlgorithmException e) { - throw new IOException(e); - } - } - - private String getBase64EncodedString(String key) { - return getBase64EncodedString(key.getBytes(StandardCharsets.UTF_8)); - } - - private String getBase64EncodedString(byte[] bytes) { - return Base64.getEncoder().encodeToString(bytes); - } - @Override public void close() throws IOException { if (tokenProvider instanceof Closeable) { - IOUtils.cleanupWithLogger(LOG, (Closeable) tokenProvider); + IOUtils.cleanupWithLogger(LOG, + (Closeable) tokenProvider); } HadoopExecutors.shutdown(executorService, LOG, 0, TimeUnit.SECONDS); } @@ -226,6 +221,14 @@ SharedKeyCredentials getSharedKeyCredentials() { return sharedKeyCredentials; } + public void setEncryptionType(EncryptionType encryptionType) { + this.encryptionType = encryptionType; + } + + public EncryptionType getEncryptionType() { + return encryptionType; + } + AbfsThrottlingIntercept getIntercept() { return intercept; } @@ -242,16 +245,56 @@ List createDefaultHeaders() { return requestHeaders; } - private void addCustomerProvidedKeyHeaders( - final List requestHeaders) { - if (clientProvidedEncryptionKey != null) { - requestHeaders.add( - new AbfsHttpHeader(X_MS_ENCRYPTION_KEY, clientProvidedEncryptionKey)); - requestHeaders.add(new AbfsHttpHeader(X_MS_ENCRYPTION_KEY_SHA256, - clientProvidedEncryptionKeySHA)); - requestHeaders.add(new AbfsHttpHeader(X_MS_ENCRYPTION_ALGORITHM, - SERVER_SIDE_ENCRYPTION_ALGORITHM)); + /** + * This method adds following headers: + *

    + *
  1. X_MS_ENCRYPTION_KEY
  2. + *
  3. X_MS_ENCRYPTION_KEY_SHA256
  4. + *
  5. X_MS_ENCRYPTION_ALGORITHM
  6. + *
+ * Above headers have to be added in following operations: + *
    + *
  1. createPath
  2. + *
  3. append
  4. + *
  5. flush
  6. + *
  7. setPathProperties
  8. + *
  9. getPathStatus for fs.setXAttr and fs.getXAttr
  10. + *
  11. read
  12. + *
+ */ + private void addEncryptionKeyRequestHeaders(String path, + List requestHeaders, boolean isCreateFileRequest, + ContextEncryptionAdapter contextEncryptionAdapter, TracingContext tracingContext) + throws AzureBlobFileSystemException { + if (!getIsNamespaceEnabled(tracingContext)) { + return; } + String encodedKey, encodedKeySHA256; + switch (encryptionType) { + case GLOBAL_KEY: + encodedKey = clientProvidedEncryptionKey; + encodedKeySHA256 = clientProvidedEncryptionKeySHA; + break; + + case ENCRYPTION_CONTEXT: + if (isCreateFileRequest) { + // get new context for create file request + requestHeaders.add(new AbfsHttpHeader(X_MS_ENCRYPTION_CONTEXT, + contextEncryptionAdapter.getEncodedContext())); + } + // else use cached encryption keys from input/output streams + encodedKey = contextEncryptionAdapter.getEncodedKey(); + encodedKeySHA256 = contextEncryptionAdapter.getEncodedKeySHA(); + break; + + default: return; // no client-provided encryption keys + } + + requestHeaders.add(new AbfsHttpHeader(X_MS_ENCRYPTION_KEY, encodedKey)); + requestHeaders.add( + new AbfsHttpHeader(X_MS_ENCRYPTION_KEY_SHA256, encodedKeySHA256)); + requestHeaders.add(new AbfsHttpHeader(X_MS_ENCRYPTION_ALGORITHM, + SERVER_SIDE_ENCRYPTION_ALGORITHM)); } AbfsUriQueryBuilder createDefaultUriQueryBuilder() { @@ -301,7 +344,7 @@ public AbfsRestOperation setFilesystemProperties(final String properties, public AbfsRestOperation listPath(final String relativePath, final boolean recursive, final int listMaxResults, final String continuation, TracingContext tracingContext) - throws AzureBlobFileSystemException { + throws IOException { final List requestHeaders = createDefaultHeaders(); final AbfsUriQueryBuilder abfsUriQueryBuilder = createDefaultUriQueryBuilder(); @@ -355,24 +398,60 @@ public AbfsRestOperation deleteFilesystem(TracingContext tracingContext) throws return op; } - public AbfsRestOperation createPath(final String path, final boolean isFile, final boolean overwrite, - final String permission, final String umask, - final boolean isAppendBlob, final String eTag, - TracingContext tracingContext) throws AzureBlobFileSystemException { + /** + * Method for calling createPath API to the backend. Method can be called from: + *
    + *
  1. create new file
  2. + *
  3. overwrite file
  4. + *
  5. create new directory
  6. + *
+ * + * @param path: path of the file / directory to be created / overwritten. + * @param isFile: defines if file or directory has to be created / overwritten. + * @param overwrite: defines if the file / directory to be overwritten. + * @param permissions: contains permission and umask + * @param isAppendBlob: defines if directory in the path is enabled for appendBlob + * @param eTag: required in case of overwrite of file / directory. Path would be + * overwritten only if the provided eTag is equal to the one present in backend for + * the path. + * @param contextEncryptionAdapter: object that contains the encryptionContext and + * encryptionKey created from the developer provided implementation of + * {@link org.apache.hadoop.fs.azurebfs.extensions.EncryptionContextProvider} + * @param tracingContext: Object of {@link org.apache.hadoop.fs.azurebfs.utils.TracingContext} + * correlating to the current fs.create() request. + * @return object of {@link AbfsRestOperation} which contain all the information + * about the communication with the server. The information is in + * {@link AbfsRestOperation#getResult()} + * @throws AzureBlobFileSystemException throws back the exception it receives from the + * {@link AbfsRestOperation#execute(TracingContext)} method call. + */ + public AbfsRestOperation createPath(final String path, + final boolean isFile, + final boolean overwrite, + final Permissions permissions, + final boolean isAppendBlob, + final String eTag, + final ContextEncryptionAdapter contextEncryptionAdapter, + final TracingContext tracingContext) + throws AzureBlobFileSystemException { final List requestHeaders = createDefaultHeaders(); if (isFile) { - addCustomerProvidedKeyHeaders(requestHeaders); + addEncryptionKeyRequestHeaders(path, requestHeaders, true, + contextEncryptionAdapter, tracingContext); } if (!overwrite) { requestHeaders.add(new AbfsHttpHeader(IF_NONE_MATCH, AbfsHttpConstants.STAR)); } - if (permission != null && !permission.isEmpty()) { - requestHeaders.add(new AbfsHttpHeader(HttpHeaderConfigurations.X_MS_PERMISSIONS, permission)); + if (permissions.hasPermission()) { + requestHeaders.add( + new AbfsHttpHeader(HttpHeaderConfigurations.X_MS_PERMISSIONS, + permissions.getPermission())); } - if (umask != null && !umask.isEmpty()) { - requestHeaders.add(new AbfsHttpHeader(HttpHeaderConfigurations.X_MS_UMASK, umask)); + if (permissions.hasUmask()) { + requestHeaders.add(new AbfsHttpHeader(HttpHeaderConfigurations.X_MS_UMASK, + permissions.getUmask())); } if (eTag != null && !eTag.isEmpty()) { @@ -491,7 +570,6 @@ public AbfsRestOperation breakLease(final String path, return op; } - /** * Rename a file or directory. * If a source etag is passed in, the operation will attempt to recover @@ -522,7 +600,7 @@ public AbfsClientRenameResult renamePath( String sourceEtag, boolean isMetadataIncompleteState, boolean isNamespaceEnabled) - throws AzureBlobFileSystemException { + throws IOException { final List requestHeaders = createDefaultHeaders(); final boolean hasEtag = !isEmpty(sourceEtag); @@ -534,7 +612,7 @@ public AbfsClientRenameResult renamePath( // fetch the source etag to be used later in recovery try { final AbfsRestOperation srcStatusOp = getPathStatus(source, - false, tracingContext); + false, tracingContext, null); if (srcStatusOp.hasResult()) { final AbfsHttpOperation result = srcStatusOp.getResult(); sourceEtag = extractEtagHeader(result); @@ -598,7 +676,8 @@ public AbfsClientRenameResult renamePath( // Doing a HEAD call resolves the incomplete metadata state and // then we can retry the rename operation. AbfsRestOperation sourceStatusOp = getPathStatus(source, false, - tracingContext); + tracingContext, null); + isMetadataIncompleteState = true; // Extract the sourceEtag, using the status Op, and set it // for future rename recovery. AbfsHttpOperation sourceStatusResult = sourceStatusOp.getResult(); @@ -688,7 +767,8 @@ public boolean renameIdempotencyCheckOp( LOG.info("rename {} to {} failed, checking etag of destination", source, destination); try { - final AbfsRestOperation destStatusOp = getPathStatus(destination, false, tracingContext); + final AbfsRestOperation destStatusOp = getPathStatus(destination, + false, tracingContext, null); final AbfsHttpOperation result = destStatusOp.getResult(); final boolean recovered = result.getStatusCode() == HttpURLConnection.HTTP_OK @@ -714,10 +794,12 @@ boolean isSourceDestEtagEqual(String sourceEtag, AbfsHttpOperation result) { } public AbfsRestOperation append(final String path, final byte[] buffer, - AppendRequestParameters reqParams, final String cachedSasToken, TracingContext tracingContext) + AppendRequestParameters reqParams, final String cachedSasToken, + ContextEncryptionAdapter contextEncryptionAdapter, TracingContext tracingContext) throws AzureBlobFileSystemException { final List requestHeaders = createDefaultHeaders(); - addCustomerProvidedKeyHeaders(requestHeaders); + addEncryptionKeyRequestHeaders(path, requestHeaders, false, + contextEncryptionAdapter, tracingContext); if (reqParams.isExpectHeaderEnabled()) { requestHeaders.add(new AbfsHttpHeader(EXPECT, HUNDRED_CONTINUE)); } @@ -782,7 +864,7 @@ public AbfsRestOperation append(final String path, final byte[] buffer, reqParams.setExpectHeaderEnabled(false); reqParams.setRetryDueToExpect(true); return this.append(path, buffer, reqParams, cachedSasToken, - tracingContext); + contextEncryptionAdapter, tracingContext); } // If we have no HTTP response, throw the original exception. if (!op.hasResult()) { @@ -825,10 +907,11 @@ private boolean checkUserError(int responseStatusCode) { // Hence, we pass/succeed the appendblob append call // in case we are doing a retry after checking the length of the file public boolean appendSuccessCheckOp(AbfsRestOperation op, final String path, - final long length, TracingContext tracingContext) throws AzureBlobFileSystemException { + final long length, TracingContext tracingContext) + throws AzureBlobFileSystemException { if ((op.isARetriedRequest()) && (op.getResult().getStatusCode() == HttpURLConnection.HTTP_BAD_REQUEST)) { - final AbfsRestOperation destStatusOp = getPathStatus(path, false, tracingContext); + final AbfsRestOperation destStatusOp = getPathStatus(path, false, tracingContext, null); if (destStatusOp.getResult().getStatusCode() == HttpURLConnection.HTTP_OK) { String fileLength = destStatusOp.getResult().getResponseHeader( HttpHeaderConfigurations.CONTENT_LENGTH); @@ -844,9 +927,11 @@ public boolean appendSuccessCheckOp(AbfsRestOperation op, final String path, public AbfsRestOperation flush(final String path, final long position, boolean retainUncommittedData, boolean isClose, final String cachedSasToken, final String leaseId, - TracingContext tracingContext) throws AzureBlobFileSystemException { + ContextEncryptionAdapter contextEncryptionAdapter, TracingContext tracingContext) + throws AzureBlobFileSystemException { final List requestHeaders = createDefaultHeaders(); - addCustomerProvidedKeyHeaders(requestHeaders); + addEncryptionKeyRequestHeaders(path, requestHeaders, false, + contextEncryptionAdapter, tracingContext); // JDK7 does not support PATCH, so to workaround the issue we will use // PUT and specify the real method in the X-Http-Method-Override header. requestHeaders.add(new AbfsHttpHeader(X_HTTP_METHOD_OVERRIDE, @@ -875,10 +960,11 @@ public AbfsRestOperation flush(final String path, final long position, } public AbfsRestOperation setPathProperties(final String path, final String properties, - TracingContext tracingContext) + final TracingContext tracingContext, final ContextEncryptionAdapter contextEncryptionAdapter) throws AzureBlobFileSystemException { final List requestHeaders = createDefaultHeaders(); - addCustomerProvidedKeyHeaders(requestHeaders); + addEncryptionKeyRequestHeaders(path, requestHeaders, false, + contextEncryptionAdapter, tracingContext); // JDK7 does not support PATCH, so to workaround the issue we will use // PUT and specify the real method in the X-Http-Method-Override header. requestHeaders.add(new AbfsHttpHeader(X_HTTP_METHOD_OVERRIDE, @@ -900,8 +986,10 @@ public AbfsRestOperation setPathProperties(final String path, final String prope return op; } - public AbfsRestOperation getPathStatus(final String path, final boolean includeProperties, - TracingContext tracingContext) throws AzureBlobFileSystemException { + public AbfsRestOperation getPathStatus(final String path, + final boolean includeProperties, final TracingContext tracingContext, + final ContextEncryptionAdapter contextEncryptionAdapter) + throws AzureBlobFileSystemException { final List requestHeaders = createDefaultHeaders(); final AbfsUriQueryBuilder abfsUriQueryBuilder = createDefaultUriQueryBuilder(); @@ -913,7 +1001,9 @@ public AbfsRestOperation getPathStatus(final String path, final boolean includeP abfsUriQueryBuilder.addQuery(HttpQueryParams.QUERY_PARAM_ACTION, AbfsHttpConstants.GET_STATUS); operation = SASTokenProvider.GET_STATUS_OPERATION; } else { - addCustomerProvidedKeyHeaders(requestHeaders); + addEncryptionKeyRequestHeaders(path, requestHeaders, false, + contextEncryptionAdapter, + tracingContext); } abfsUriQueryBuilder.addQuery(HttpQueryParams.QUERY_PARAM_UPN, String.valueOf(abfsConfiguration.isUpnUsed())); appendSASTokenToQuery(path, operation, abfsUriQueryBuilder); @@ -928,11 +1018,18 @@ public AbfsRestOperation getPathStatus(final String path, final boolean includeP return op; } - public AbfsRestOperation read(final String path, final long position, final byte[] buffer, final int bufferOffset, - final int bufferLength, final String eTag, String cachedSasToken, + public AbfsRestOperation read(final String path, + final long position, + final byte[] buffer, + final int bufferOffset, + final int bufferLength, + final String eTag, + String cachedSasToken, + ContextEncryptionAdapter contextEncryptionAdapter, TracingContext tracingContext) throws AzureBlobFileSystemException { final List requestHeaders = createDefaultHeaders(); - addCustomerProvidedKeyHeaders(requestHeaders); + addEncryptionKeyRequestHeaders(path, requestHeaders, false, + contextEncryptionAdapter, tracingContext); requestHeaders.add(new AbfsHttpHeader(RANGE, String.format("bytes=%d-%d", position, position + bufferLength - 1))); requestHeaders.add(new AbfsHttpHeader(IF_MATCH, eTag)); @@ -1295,10 +1392,23 @@ public synchronized String getAccessToken() throws IOException { } } + private synchronized Boolean getIsNamespaceEnabled(TracingContext tracingContext) + throws AzureBlobFileSystemException { + if (isNamespaceEnabled == null) { + setIsNamespaceEnabled(NamespaceUtil.isNamespaceEnabled(this, + tracingContext)); + } + return isNamespaceEnabled; + } + public AuthType getAuthType() { return authType; } + public EncryptionContextProvider getEncryptionContextProvider() { + return encryptionContextProvider; + } + @VisibleForTesting String initializeUserAgent(final AbfsConfiguration abfsConfiguration, final String sslProviderName) { @@ -1373,6 +1483,16 @@ public SASTokenProvider getSasTokenProvider() { return this.sasTokenProvider; } + @VisibleForTesting + void setEncryptionContextProvider(EncryptionContextProvider provider) { + encryptionContextProvider = provider; + } + + @VisibleForTesting + void setIsNamespaceEnabled(final Boolean isNamespaceEnabled) { + this.isNamespaceEnabled = isNamespaceEnabled; + } + /** * Getter for abfsCounters from AbfsClient. * @return AbfsCounters instance. @@ -1381,6 +1501,10 @@ protected AbfsCounters getAbfsCounters() { return abfsCounters; } + public String getxMsVersion() { + return xMsVersion; + } + /** * Getter for abfsConfiguration from AbfsClient. * @return AbfsConfiguration instance diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsInputStream.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsInputStream.java index 86442dac9a..00b48e25b5 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsInputStream.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsInputStream.java @@ -40,6 +40,7 @@ import org.apache.hadoop.fs.azurebfs.constants.FSOperationType; import org.apache.hadoop.fs.azurebfs.contracts.exceptions.AbfsRestOperationException; import org.apache.hadoop.fs.azurebfs.contracts.exceptions.AzureBlobFileSystemException; +import org.apache.hadoop.fs.azurebfs.security.ContextEncryptionAdapter; import org.apache.hadoop.fs.azurebfs.utils.CachedSASToken; import org.apache.hadoop.fs.azurebfs.utils.Listener; import org.apache.hadoop.fs.azurebfs.utils.TracingContext; @@ -99,6 +100,7 @@ public class AbfsInputStream extends FSInputStream implements CanUnbuffer, // of valid bytes in buffer) private boolean closed = false; private TracingContext tracingContext; + private final ContextEncryptionAdapter contextEncryptionAdapter; // Optimisations modify the pointer fields. // For better resilience the following fields are used to save the @@ -157,6 +159,7 @@ public AbfsInputStream( this.context = abfsInputStreamContext; readAheadBlockSize = abfsInputStreamContext.getReadAheadBlockSize(); this.fsBackRef = abfsInputStreamContext.getFsBackRef(); + contextEncryptionAdapter = abfsInputStreamContext.getEncryptionAdapter(); // Propagate the config values to ReadBufferManager so that the first instance // to initialize can set the readAheadBlockSize @@ -548,7 +551,8 @@ int readRemote(long position, byte[] b, int offset, int length, TracingContext t } LOG.trace("Trigger client.read for path={} position={} offset={} length={}", path, position, offset, length); op = client.read(path, position, b, offset, length, - tolerateOobAppends ? "*" : eTag, cachedSasToken.get(), tracingContext); + tolerateOobAppends ? "*" : eTag, cachedSasToken.get(), + contextEncryptionAdapter, tracingContext); cachedSasToken.update(op.getSasToken()); LOG.debug("issuing HTTP GET request params position = {} b.length = {} " + "offset = {} length = {}", position, b.length, offset, length); @@ -701,8 +705,11 @@ public boolean seekToNewSource(long l) throws IOException { public synchronized void close() throws IOException { LOG.debug("Closing {}", this); closed = true; - buffer = null; // de-reference the buffer so it can be GC'ed sooner ReadBufferManager.getBufferManager().purgeBuffersForStream(this); + buffer = null; // de-reference the buffer so it can be GC'ed sooner + if (contextEncryptionAdapter != null) { + contextEncryptionAdapter.destroy(); + } } /** diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsInputStreamContext.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsInputStreamContext.java index b78a899340..59352d174f 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsInputStreamContext.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsInputStreamContext.java @@ -24,6 +24,8 @@ import org.apache.hadoop.fs.impl.BackReference; import org.apache.hadoop.util.Preconditions; +import org.apache.hadoop.fs.azurebfs.security.ContextEncryptionAdapter; + /** * Class to hold extra input stream configs. */ @@ -56,6 +58,8 @@ public class AbfsInputStreamContext extends AbfsStreamContext { /** A BackReference to the FS instance that created this OutputStream. */ private BackReference fsBackRef; + private ContextEncryptionAdapter contextEncryptionAdapter = null; + public AbfsInputStreamContext(final long sasTokenRenewPeriodForStreamsInSeconds) { super(sasTokenRenewPeriodForStreamsInSeconds); } @@ -133,6 +137,12 @@ public AbfsInputStreamContext withAbfsBackRef( return this; } + public AbfsInputStreamContext withEncryptionAdapter( + ContextEncryptionAdapter contextEncryptionAdapter){ + this.contextEncryptionAdapter = contextEncryptionAdapter; + return this; + } + public AbfsInputStreamContext build() { if (readBufferSize > readAheadBlockSize) { LOG.debug( @@ -195,4 +205,8 @@ public boolean isBufferedPreadDisabled() { public BackReference getFsBackRef() { return fsBackRef; } + + public ContextEncryptionAdapter getEncryptionAdapter() { + return contextEncryptionAdapter; + } } diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsOutputStream.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsOutputStream.java index 89b0fe2040..5780e290a0 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsOutputStream.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsOutputStream.java @@ -26,6 +26,7 @@ import java.util.concurrent.Future; import java.util.UUID; +import org.apache.hadoop.fs.azurebfs.security.ContextEncryptionAdapter; import org.apache.hadoop.classification.VisibleForTesting; import org.apache.hadoop.fs.impl.BackReference; import org.apache.hadoop.util.Preconditions; @@ -95,6 +96,7 @@ public class AbfsOutputStream extends OutputStream implements Syncable, private final int maxRequestsThatCanBeQueued; private ConcurrentLinkedDeque writeOperations; + private final ContextEncryptionAdapter contextEncryptionAdapter; // SAS tokens can be re-used until they expire private CachedSASToken cachedSasToken; @@ -152,6 +154,7 @@ public AbfsOutputStream(AbfsOutputStreamContext abfsOutputStreamContext) this.writeOperations = new ConcurrentLinkedDeque<>(); this.outputStreamStatistics = abfsOutputStreamContext.getStreamStatistics(); this.fsBackRef = abfsOutputStreamContext.getFsBackRef(); + this.contextEncryptionAdapter = abfsOutputStreamContext.getEncryptionAdapter(); if (this.isAppendBlob) { this.maxConcurrentRequestCount = 1; @@ -335,9 +338,9 @@ private void uploadBlockAsync(DataBlocks.DataBlock blockToUpload, */ AppendRequestParameters reqParams = new AppendRequestParameters( offset, 0, bytesLength, mode, false, leaseId, isExpectHeaderEnabled); - AbfsRestOperation op = - client.append(path, blockUploadData.toByteArray(), reqParams, - cachedSasToken.get(), new TracingContext(tracingContext)); + AbfsRestOperation op = client.append(path, + blockUploadData.toByteArray(), reqParams, cachedSasToken.get(), + contextEncryptionAdapter, new TracingContext(tracingContext)); cachedSasToken.update(op.getSasToken()); perfInfo.registerResult(op.getResult()); perfInfo.registerSuccess(true); @@ -507,6 +510,9 @@ public synchronized void close() throws IOException { // See HADOOP-16785 throw wrapException(path, e.getMessage(), e); } finally { + if (contextEncryptionAdapter != null) { + contextEncryptionAdapter.destroy(); + } if (hasLease()) { lease.free(); lease = null; @@ -587,8 +593,9 @@ private void writeAppendBlobCurrentBufferToService() throws IOException { "writeCurrentBufferToService", "append")) { AppendRequestParameters reqParams = new AppendRequestParameters(offset, 0, bytesLength, APPEND_MODE, true, leaseId, isExpectHeaderEnabled); - AbfsRestOperation op = client.append(path, uploadData.toByteArray(), reqParams, - cachedSasToken.get(), new TracingContext(tracingContext)); + AbfsRestOperation op = client.append(path, uploadData.toByteArray(), + reqParams, cachedSasToken.get(), contextEncryptionAdapter, + new TracingContext(tracingContext)); cachedSasToken.update(op.getSasToken()); outputStreamStatistics.uploadSuccessful(bytesLength); @@ -648,8 +655,9 @@ private synchronized void flushWrittenBytesToServiceInternal(final long offset, AbfsPerfTracker tracker = client.getAbfsPerfTracker(); try (AbfsPerfInfo perfInfo = new AbfsPerfInfo(tracker, "flushWrittenBytesToServiceInternal", "flush")) { - AbfsRestOperation op = client.flush(path, offset, retainUncommitedData, isClose, - cachedSasToken.get(), leaseId, new TracingContext(tracingContext)); + AbfsRestOperation op = client.flush(path, offset, retainUncommitedData, + isClose, cachedSasToken.get(), leaseId, contextEncryptionAdapter, + new TracingContext(tracingContext)); cachedSasToken.update(op.getSasToken()); perfInfo.registerResult(op.getResult()).registerSuccess(true); } catch (AzureBlobFileSystemException ex) { diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsOutputStreamContext.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsOutputStreamContext.java index 1d1a99c7d9..4763c99c47 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsOutputStreamContext.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsOutputStreamContext.java @@ -21,6 +21,7 @@ import java.util.concurrent.ExecutorService; import org.apache.hadoop.fs.FileSystem; +import org.apache.hadoop.fs.azurebfs.security.ContextEncryptionAdapter; import org.apache.hadoop.fs.azurebfs.utils.TracingContext; import org.apache.hadoop.fs.impl.BackReference; import org.apache.hadoop.fs.store.DataBlocks; @@ -50,6 +51,8 @@ public class AbfsOutputStreamContext extends AbfsStreamContext { private AbfsLease lease; + private ContextEncryptionAdapter contextEncryptionAdapter; + private DataBlocks.BlockFactory blockFactory; private int blockOutputActiveBlocks; @@ -193,6 +196,12 @@ public AbfsOutputStreamContext withLease(final AbfsLease lease) { return this; } + public AbfsOutputStreamContext withEncryptionAdapter( + final ContextEncryptionAdapter contextEncryptionAdapter) { + this.contextEncryptionAdapter = contextEncryptionAdapter; + return this; + } + public int getWriteBufferSize() { return writeBufferSize; } @@ -240,6 +249,10 @@ public String getLeaseId() { return this.lease.getLeaseID(); } + public ContextEncryptionAdapter getEncryptionAdapter() { + return contextEncryptionAdapter; + } + public DataBlocks.BlockFactory getBlockFactory() { return blockFactory; } diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/EncryptionType.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/EncryptionType.java new file mode 100644 index 0000000000..3cd414a766 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/EncryptionType.java @@ -0,0 +1,33 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.utils; + +/** + * Enum EncryptionType to represent the level of encryption applied. + *
    + *
  1. GLOBAL_KEY: encrypt all files with the same client-provided key.
  2. + *
  3. ENCRYPTION_CONTEXT: uses client-provided implementation to generate keys.
  4. + *
  5. NONE: encryption handled entirely at server.
  6. + *
+ */ +public enum EncryptionType { + GLOBAL_KEY, + ENCRYPTION_CONTEXT, + NONE +} diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/NamespaceUtil.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/NamespaceUtil.java new file mode 100644 index 0000000000..67225efa14 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/NamespaceUtil.java @@ -0,0 +1,88 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + *

+ * http://www.apache.org/licenses/LICENSE-2.0 + *

+ * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.utils; + +import java.net.HttpURLConnection; + +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +import org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants; +import org.apache.hadoop.fs.azurebfs.contracts.exceptions.AbfsRestOperationException; +import org.apache.hadoop.fs.azurebfs.contracts.exceptions.AzureBlobFileSystemException; +import org.apache.hadoop.fs.azurebfs.services.AbfsClient; + +/** + * Utility class to provide method which can return if the account is namespace + * enabled or not. + */ +public final class NamespaceUtil { + + public static final Logger LOG = LoggerFactory.getLogger(NamespaceUtil.class); + + private NamespaceUtil() { + + } + + /** + * Return if the account used in the provided abfsClient object namespace enabled + * or not. + * It would call {@link org.apache.hadoop.fs.azurebfs.services.AbfsClient#getAclStatus(String, TracingContext)}. + *

    + *
  1. + * If the API call is successful, then the account is namespace enabled. + *
  2. + *
  3. + * If the server returns with {@link java.net.HttpURLConnection#HTTP_BAD_REQUEST}, the account is non-namespace enabled. + *
  4. + *
  5. + * If the server call gets some other exception, then the method would throw the exception. + *
  6. + *
+ * @param abfsClient client for which namespace-enabled to be checked. + * @param tracingContext object to correlate Store requests. + * @return if the account corresponding to the given client is namespace-enabled + * or not. + * @throws AzureBlobFileSystemException throws back the exception the method receives + * from the {@link AbfsClient#getAclStatus(String, TracingContext)}. In case it gets + * {@link AbfsRestOperationException}, it checks if the exception statusCode is + * BAD_REQUEST or not. If not, then it will pass the exception to the calling method. + */ + public static Boolean isNamespaceEnabled(final AbfsClient abfsClient, + final TracingContext tracingContext) + throws AzureBlobFileSystemException { + Boolean isNamespaceEnabled; + try { + LOG.debug("Get root ACL status"); + abfsClient.getAclStatus(AbfsHttpConstants.ROOT_PATH, tracingContext); + isNamespaceEnabled = true; + } catch (AbfsRestOperationException ex) { + // Get ACL status is a HEAD request, its response doesn't contain + // errorCode + // So can only rely on its status code to determine its account type. + if (HttpURLConnection.HTTP_BAD_REQUEST != ex.getStatusCode()) { + throw ex; + } + isNamespaceEnabled = false; + } catch (AzureBlobFileSystemException ex) { + throw ex; + } + return isNamespaceEnabled; + } +} diff --git a/hadoop-tools/hadoop-azure/src/site/markdown/abfs.md b/hadoop-tools/hadoop-azure/src/site/markdown/abfs.md index aff1e32b83..9021f3e3b1 100644 --- a/hadoop-tools/hadoop-azure/src/site/markdown/abfs.md +++ b/hadoop-tools/hadoop-azure/src/site/markdown/abfs.md @@ -908,6 +908,38 @@ specified SSL channel mode. Value should be of the enum DelegatingSSLSocketFactory.SSLChannelMode. The default value will be DelegatingSSLSocketFactory.SSLChannelMode.Default. +### Encryption Options +Only one of the following two options can be configured. If config values of +both types are set, ABFS driver will throw an exception. If using the global +key type, ensure both pre-computed values are provided. + +#### Customer-Provided Global Key +A global encryption key can be configured by providing the following +pre-computed values. The key will be applied to any new files created post +setting the configuration, and will be required in the requests to read ro +modify the contents of the files. + +`fs.azure.encryption.encoded.client-provided-key`: The Base64 encoded version +of the 256-bit encryption key. + +`fs.azure.encryption.encoded.client-provided-key-sha`: The Base64 encoded +version of the SHA256 has of the 256-bit encryption key. + +#### Encryption Context Provider + +ABFS driver supports an interface called `EncryptionContextProvider` that +can be used as a plugin for clients to provide custom implementations for +the encryption framework. This framework allows for an `encryptionContext` +and an `encryptionKey` to be generated by the EncryptionContextProvider for +a file to be created. The server keeps track of the encryptionContext for +each file. To perform subsequent operations such as read on the encrypted file, +ABFS driver will fetch the corresponding encryption key from the +EncryptionContextProvider implementation by providing the encryptionContext +string retrieved from a GetFileStatus request to the server. + +`fs.azure.encryption.context.provider.type`: The canonical name of the class +implementing EncryptionContextProvider. + ### Server Options When the config `fs.azure.io.read.tolerate.concurrent.append` is made true, the If-Match header sent to the server for read calls will be set as * otherwise the diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/AbstractAbfsIntegrationTest.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/AbstractAbfsIntegrationTest.java index 74655fd573..66a1b22da9 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/AbstractAbfsIntegrationTest.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/AbstractAbfsIntegrationTest.java @@ -40,6 +40,7 @@ import org.apache.hadoop.fs.azurebfs.oauth2.AccessTokenProvider; import org.apache.hadoop.fs.azurebfs.security.AbfsDelegationTokenManager; import org.apache.hadoop.fs.azurebfs.services.AbfsClient; +import org.apache.hadoop.fs.azurebfs.services.AbfsClientUtils; import org.apache.hadoop.fs.azurebfs.services.AbfsOutputStream; import org.apache.hadoop.fs.azurebfs.services.AuthType; import org.apache.hadoop.fs.azurebfs.services.ITestAbfsClient; @@ -211,6 +212,7 @@ public void setup() throws Exception { wasb = new NativeAzureFileSystem(azureNativeFileSystemStore); wasb.initialize(wasbUri, rawConfig); } + AbfsClientUtils.setIsNamespaceEnabled(abfs.getAbfsClient(), true); } @After diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAbfsCustomEncryption.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAbfsCustomEncryption.java new file mode 100644 index 0000000000..9bd023572c --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAbfsCustomEncryption.java @@ -0,0 +1,460 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs; + +import java.io.IOException; +import java.nio.charset.StandardCharsets; +import java.util.ArrayList; +import java.util.Arrays; +import java.util.Base64; +import java.util.HashSet; +import java.util.Hashtable; +import java.util.List; +import java.util.Random; + +import org.apache.hadoop.fs.FileStatus; +import org.apache.hadoop.fs.PathIOException; +import org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations; +import org.apache.hadoop.fs.azurebfs.security.EncodingHelper; +import org.apache.hadoop.fs.azurebfs.services.AbfsClientUtils; +import org.apache.hadoop.fs.azurebfs.utils.TracingContext; +import org.assertj.core.api.Assertions; +import org.junit.Assume; +import org.junit.Test; +import org.junit.runner.RunWith; +import org.junit.runners.Parameterized; + +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.FSDataInputStream; +import org.apache.hadoop.fs.FSDataOutputStream; +import org.apache.hadoop.fs.FileSystem; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.fs.azurebfs.constants.FSOperationType; +import org.apache.hadoop.fs.azurebfs.contracts.services.AppendRequestParameters; +import org.apache.hadoop.fs.azurebfs.extensions.EncryptionContextProvider; +import org.apache.hadoop.fs.azurebfs.extensions.MockEncryptionContextProvider; +import org.apache.hadoop.fs.azurebfs.security.ContextProviderEncryptionAdapter; +import org.apache.hadoop.fs.azurebfs.services.AbfsClient; +import org.apache.hadoop.fs.azurebfs.services.AbfsHttpOperation; +import org.apache.hadoop.fs.azurebfs.services.AbfsRestOperation; +import org.apache.hadoop.fs.azurebfs.utils.EncryptionType; +import org.apache.hadoop.fs.impl.OpenFileParameters; +import org.apache.hadoop.fs.permission.AclEntry; +import org.apache.hadoop.fs.permission.FsPermission; +import org.apache.hadoop.test.LambdaTestUtils; +import org.apache.hadoop.util.Lists; + +import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.FS_AZURE_ENCRYPTION_CONTEXT_PROVIDER_TYPE; +import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY; +import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY_SHA; +import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_ENCRYPTION_CONTEXT; +import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_ENCRYPTION_KEY_SHA256; +import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_REQUEST_SERVER_ENCRYPTED; +import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_SERVER_ENCRYPTED; +import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.ENCRYPTION_KEY_LEN; +import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.FS_AZURE_TEST_NAMESPACE_ENABLED_ACCOUNT; +import static org.apache.hadoop.fs.azurebfs.contracts.services.AppendRequestParameters.Mode.APPEND_MODE; +import static org.apache.hadoop.fs.azurebfs.utils.AclTestHelpers.aclEntry; +import static org.apache.hadoop.fs.azurebfs.utils.EncryptionType.ENCRYPTION_CONTEXT; +import static org.apache.hadoop.fs.azurebfs.utils.EncryptionType.GLOBAL_KEY; +import static org.apache.hadoop.fs.azurebfs.utils.EncryptionType.NONE; +import static org.apache.hadoop.fs.permission.AclEntryScope.ACCESS; +import static org.apache.hadoop.fs.permission.AclEntryType.USER; +import static org.apache.hadoop.fs.permission.FsAction.ALL; + +@RunWith(Parameterized.class) +public class ITestAbfsCustomEncryption extends AbstractAbfsIntegrationTest { + + public static final String SERVER_FILE_CONTENT = "123"; + + private final byte[] cpk = new byte[ENCRYPTION_KEY_LEN]; + private final String cpkSHAEncoded; + + private List fileSystemsOpenedInTest = new ArrayList<>(); + + // Encryption type used by filesystem while creating file + @Parameterized.Parameter + public EncryptionType fileEncryptionType; + + // Encryption type used by filesystem to call different operations + @Parameterized.Parameter(1) + public EncryptionType requestEncryptionType; + + @Parameterized.Parameter(2) + public FSOperationType operation; + + @Parameterized.Parameter(3) + public boolean responseHeaderServerEnc; + + @Parameterized.Parameter(4) + public boolean responseHeaderReqServerEnc; + + @Parameterized.Parameter(5) + public boolean isExceptionCase; + + /** + * Boolean value to indicate that the server response would have header related + * to CPK and the test would need to assert its value. + */ + @Parameterized.Parameter(6) + public boolean isCpkResponseHdrExpected; + + /** + * Boolean value to indicate that the server response would have fields related + * to CPK and the test would need to assert its value. + */ + @Parameterized.Parameter(7) + public Boolean isCpkResponseKeyExpected = false; + + @Parameterized.Parameter(8) + public Boolean fileSystemListStatusResultToBeUsedForOpeningFile = false; + + @Parameterized.Parameters(name = "{0} mode, {2}") + public static Iterable params() { + return Arrays.asList(new Object[][] { + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.READ, true, false, false, true, false, false}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.READ, true, false, false, true, false, true}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.WRITE, false, true, false, true, false, false}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.APPEND, false, true, false, true, false, false}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.SET_ACL, false, false, false, false, false, false}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.GET_ATTR, true, false, false, true, false, false}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.SET_ATTR, false, true, false, true, false, false}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.LISTSTATUS, false, false, false, false, true, false}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.RENAME, false, false, false, false, false, false}, + {ENCRYPTION_CONTEXT, ENCRYPTION_CONTEXT, FSOperationType.DELETE, false, false, false, false, false, false}, + + {ENCRYPTION_CONTEXT, NONE, FSOperationType.WRITE, false, false, true, false, false, false}, + {ENCRYPTION_CONTEXT, NONE, FSOperationType.GET_ATTR, true, false, true, false, false, false}, + {ENCRYPTION_CONTEXT, NONE, FSOperationType.READ, false, false, true, false, false, false}, + {ENCRYPTION_CONTEXT, NONE, FSOperationType.SET_ATTR, false, true, true, false, false, false}, + {ENCRYPTION_CONTEXT, NONE, FSOperationType.RENAME, false, false, false, false, false, false}, + {ENCRYPTION_CONTEXT, NONE, FSOperationType.LISTSTATUS, false, false, false, false, false, false}, + {ENCRYPTION_CONTEXT, NONE, FSOperationType.DELETE, false, false, false, false, false, false}, + {ENCRYPTION_CONTEXT, NONE, FSOperationType.SET_ACL, false, false, false, false, false, false}, + {ENCRYPTION_CONTEXT, NONE, FSOperationType.SET_PERMISSION, false, false, false, false, false, false}, + + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.READ, true, false, false, true, false, false}, + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.WRITE, false, true, false, true, false, false}, + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.APPEND, false, true, false, true, false, false}, + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.SET_ACL, false, false, false, false, false, false}, + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.LISTSTATUS, false, false, false, false, false, false}, + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.RENAME, false, false, false, false, false, false}, + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.DELETE, false, false, false, false, false, false}, + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.GET_ATTR, true, false, false, true, false, false}, + {GLOBAL_KEY, GLOBAL_KEY, FSOperationType.SET_ATTR, false, true, false, true, false, false}, + + {GLOBAL_KEY, NONE, FSOperationType.READ, true, false, true, true, false, false}, + {GLOBAL_KEY, NONE, FSOperationType.WRITE, false, true, true, true, false, false}, + {GLOBAL_KEY, NONE, FSOperationType.SET_ATTR, false, false, true, true, false, false}, + {GLOBAL_KEY, NONE, FSOperationType.SET_ACL, false, false, false, false, false, false}, + {GLOBAL_KEY, NONE, FSOperationType.RENAME, false, false, false, false, false, false}, + {GLOBAL_KEY, NONE, FSOperationType.LISTSTATUS, false, false, false, false, false, false}, + {GLOBAL_KEY, NONE, FSOperationType.DELETE, false, false, false, false, false, false}, + {GLOBAL_KEY, NONE, FSOperationType.SET_PERMISSION, false, false, false, false, false, false}, + }); + } + + public ITestAbfsCustomEncryption() throws Exception { + Assume.assumeTrue("Account should be HNS enabled for CPK", + getConfiguration().getBoolean(FS_AZURE_TEST_NAMESPACE_ENABLED_ACCOUNT, + false)); + new Random().nextBytes(cpk); + cpkSHAEncoded = EncodingHelper.getBase64EncodedString( + EncodingHelper.getSHA256Hash(cpk)); + } + + @Test + public void testCustomEncryptionCombinations() throws Exception { + AzureBlobFileSystem fs = getOrCreateFS(); + Path testPath = path("/testFile"); + String relativePath = fs.getAbfsStore().getRelativePath(testPath); + MockEncryptionContextProvider ecp = + (MockEncryptionContextProvider) createEncryptedFile(testPath); + AbfsRestOperation op = callOperation(fs, new Path(relativePath), ecp); + if (op == null) { + return; + } + AbfsHttpOperation httpOp = op.getResult(); + if (isCpkResponseHdrExpected) { + if (requestEncryptionType == ENCRYPTION_CONTEXT) { + String encryptionContext = ecp.getEncryptionContextForTest(relativePath); + String expectedKeySHA = EncodingHelper.getBase64EncodedString( + EncodingHelper.getSHA256Hash( + ecp.getEncryptionKeyForTest(encryptionContext))); + Assertions.assertThat(httpOp.getResponseHeader(X_MS_ENCRYPTION_KEY_SHA256)) + .isEqualTo(expectedKeySHA); + } else { // GLOBAL_KEY + Assertions.assertThat(httpOp.getResponseHeader(X_MS_ENCRYPTION_KEY_SHA256)) + .isEqualTo(cpkSHAEncoded); + } + } else { + if (isCpkResponseKeyExpected) { + if (requestEncryptionType == ENCRYPTION_CONTEXT) { + String encryptionContext = ecp.getEncryptionContextForTest(relativePath); + String expectedKeySHA = EncodingHelper.getBase64EncodedString( + EncodingHelper.getSHA256Hash( + ecp.getEncryptionKeyForTest(encryptionContext))); + Assertions.assertThat(httpOp.getListResultSchema().paths().get(0) + .getCustomerProvidedKeySha256()).isEqualTo(expectedKeySHA); + } + } else { + Assertions.assertThat( + httpOp.getResponseHeader(X_MS_ENCRYPTION_KEY_SHA256)) + .isEqualTo(null); + } + } + Assertions.assertThat(httpOp.getResponseHeader(X_MS_SERVER_ENCRYPTED)) + .isEqualTo(responseHeaderServerEnc? "true" : null); + Assertions.assertThat(httpOp.getResponseHeader(X_MS_REQUEST_SERVER_ENCRYPTED)) + .isEqualTo(responseHeaderReqServerEnc? "true" : null); + } + + /** + * Executes a given operation at the AbfsClient level and returns + * AbfsRestOperation instance to verify response headers. Asserts excetion + * for combinations that should not succeed. + * @param fs AzureBlobFileSystem instance + * @param testPath path of file + * @param ecp EncryptionContextProvider instance to support AbfsClient methods + * @return Rest op or null depending on whether the request is allowed + * @throws Exception error + */ + private AbfsRestOperation callOperation(AzureBlobFileSystem fs, + Path testPath, EncryptionContextProvider ecp) + throws Exception { + AbfsClient client = fs.getAbfsClient(); + AbfsClientUtils.setEncryptionContextProvider(client, ecp); + if (isExceptionCase) { + LambdaTestUtils.intercept(IOException.class, () -> { + switch (operation) { + case WRITE: try (FSDataOutputStream out = fs.append(testPath)) { + out.write("bytes".getBytes()); + } + break; + case READ: try (FSDataInputStream in = fs.open(testPath)) { + in.read(new byte[5]); + } + break; + case SET_ATTR: fs.setXAttr(testPath, "attribute", "value".getBytes()); + break; + case GET_ATTR: fs.getXAttr(testPath, "attribute"); + break; + default: throw new NoSuchFieldException(); + } + }); + return null; + } else { + ContextProviderEncryptionAdapter encryptionAdapter = null; + if (fileEncryptionType == ENCRYPTION_CONTEXT) { + encryptionAdapter = new ContextProviderEncryptionAdapter(ecp, + fs.getAbfsStore().getRelativePath(testPath), + Base64.getEncoder().encode( + ((MockEncryptionContextProvider) ecp).getEncryptionContextForTest(testPath.toString()) + .getBytes(StandardCharsets.UTF_8))); + } + String path = testPath.toString(); + switch (operation) { + case READ: + if (!fileSystemListStatusResultToBeUsedForOpeningFile + || fileEncryptionType != ENCRYPTION_CONTEXT) { + TracingContext tracingContext = getTestTracingContext(fs, true); + AbfsHttpOperation statusOp = client.getPathStatus(path, false, + tracingContext, null).getResult(); + return client.read(path, 0, new byte[5], 0, 5, + statusOp.getResponseHeader(HttpHeaderConfigurations.ETAG), + null, encryptionAdapter, tracingContext); + } else { + /* + * In this block, its tested scenario is: + * 1.Create a file. + * 2.Fetch List of VersionFileStatus objects by listStatus API of the AzureBlobFileSystem. + * 3.Use the context value in the VersionFileStatus object for making read API call to backend. + * 4.Assert for no exception and get response. + */ + FileStatus status = fs.listStatus(testPath)[0]; + Assertions.assertThat(status) + .isInstanceOf(AzureBlobFileSystemStore.VersionedFileStatus.class); + + Assertions.assertThat( + ((AzureBlobFileSystemStore.VersionedFileStatus) status).getEncryptionContext()) + .isNotNull(); + + try (FSDataInputStream in = fs.openFileWithOptions(testPath, + new OpenFileParameters().withMandatoryKeys(new HashSet<>()) + .withStatus(fs.listStatus(testPath)[0])).get()) { + byte[] readBuffer = new byte[3]; + Assertions.assertThat(in.read(readBuffer)).isGreaterThan(0); + Assertions.assertThat(readBuffer).isEqualTo(SERVER_FILE_CONTENT.getBytes()); + return null; + } + } + case WRITE: + return client.flush(path, 3, false, false, null, + null, encryptionAdapter, getTestTracingContext(fs, false)); + case APPEND: + return client.append(path, "val".getBytes(), + new AppendRequestParameters(3, 0, 3, APPEND_MODE, false, null, true), + null, encryptionAdapter, getTestTracingContext(fs, false)); + case SET_ACL: + return client.setAcl(path, AclEntry.aclSpecToString( + Lists.newArrayList(aclEntry(ACCESS, USER, ALL))), + getTestTracingContext(fs, false)); + case LISTSTATUS: + return client.listPath(path, false, 5, null, + getTestTracingContext(fs, true)); + case RENAME: + TracingContext tc = getTestTracingContext(fs, true); + return client.renamePath(path, new Path(path + "_2").toString(), + null, tc, null, false, fs.getIsNamespaceEnabled(tc)).getOp(); + case DELETE: + return client.deletePath(path, false, null, + getTestTracingContext(fs, false)); + case GET_ATTR: + return client.getPathStatus(path, true, + getTestTracingContext(fs, false), + createEncryptionAdapterFromServerStoreContext(path, + getTestTracingContext(fs, false), client)); + case SET_ATTR: + Hashtable properties = new Hashtable<>(); + properties.put("key", "{ value: valueTest }"); + return client.setPathProperties(path, fs.getAbfsStore() + .convertXmsPropertiesToCommaSeparatedString(properties), + getTestTracingContext(fs, false), + createEncryptionAdapterFromServerStoreContext(path, + getTestTracingContext(fs, false), client)); + case SET_PERMISSION: + return client.setPermission(path, FsPermission.getDefault().toString(), + getTestTracingContext(fs, false)); + default: throw new NoSuchFieldException(); + } + } + } + + private ContextProviderEncryptionAdapter createEncryptionAdapterFromServerStoreContext(final String path, + final TracingContext tracingContext, + final AbfsClient client) throws IOException { + if (client.getEncryptionType() != ENCRYPTION_CONTEXT) { + return null; + } + final String responseHeaderEncryptionContext = client.getPathStatus(path, + false, tracingContext, null).getResult() + .getResponseHeader(X_MS_ENCRYPTION_CONTEXT); + if (responseHeaderEncryptionContext == null) { + throw new PathIOException(path, + "EncryptionContext not present in GetPathStatus response"); + } + byte[] encryptionContext = responseHeaderEncryptionContext.getBytes( + StandardCharsets.UTF_8); + + return new ContextProviderEncryptionAdapter(client.getEncryptionContextProvider(), + new Path(path).toUri().getPath(), encryptionContext); + } + + private AzureBlobFileSystem getECProviderEnabledFS() throws Exception { + Configuration configuration = getRawConfiguration(); + configuration.set(FS_AZURE_ENCRYPTION_CONTEXT_PROVIDER_TYPE + "." + + getAccountName(), MockEncryptionContextProvider.class.getCanonicalName()); + configuration.unset(FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY + "." + + getAccountName()); + configuration.unset(FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY_SHA + "." + + getAccountName()); + AzureBlobFileSystem fs = (AzureBlobFileSystem) FileSystem.newInstance(configuration); + fileSystemsOpenedInTest.add(fs); + return fs; + } + + private AzureBlobFileSystem getCPKEnabledFS() throws IOException { + Configuration conf = getRawConfiguration(); + String cpkEncoded = EncodingHelper.getBase64EncodedString(cpk); + String cpkEncodedSHA = EncodingHelper.getBase64EncodedString( + EncodingHelper.getSHA256Hash(cpk)); + conf.set(FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY + "." + + getAccountName(), cpkEncoded); + conf.set(FS_AZURE_ENCRYPTION_ENCODED_CLIENT_PROVIDED_KEY_SHA + "." + + getAccountName(), cpkEncodedSHA); + conf.unset(FS_AZURE_ENCRYPTION_CONTEXT_PROVIDER_TYPE); + AzureBlobFileSystem fs = (AzureBlobFileSystem) FileSystem.newInstance(conf); + fileSystemsOpenedInTest.add(fs); + return fs; + } + + private AzureBlobFileSystem getOrCreateFS() throws Exception { + if (getFileSystem().getAbfsClient().getEncryptionType() == requestEncryptionType) { + return getFileSystem(); + } + if (requestEncryptionType == ENCRYPTION_CONTEXT) { + return getECProviderEnabledFS(); + } else if (requestEncryptionType == GLOBAL_KEY) { + return getCPKEnabledFS(); + } else { + Configuration conf = getRawConfiguration(); + conf.unset(FS_AZURE_ENCRYPTION_CONTEXT_PROVIDER_TYPE); + AzureBlobFileSystem fs = (AzureBlobFileSystem) FileSystem.newInstance(conf); + fileSystemsOpenedInTest.add(fs); + return fs; + } + } + + /** + * Creates a file in the server with values for the following keys: + *
    + *
  1. x-ms-encryption-key: for ENCRYPTION_CONTEXT, GLOBAL
  2. + *
  3. x-ms-encryption-key-sha256: for ENCRYPTION_CONTEXT, GLOBAL
  4. + *
  5. x-ms-encryption-context: for ENCRYPTION_CONTEXT
  6. + *
+ * Returns in case of ENCRYPTION_CONTEXT the encryptionProvider object which + * was used to create the x-ms-encryption-context value used for creating the file. + */ + private EncryptionContextProvider createEncryptedFile(Path testPath) throws Exception { + AzureBlobFileSystem fs; + if (getFileSystem().getAbfsClient().getEncryptionType() == fileEncryptionType) { + fs = getFileSystem(); + } else { + fs = fileEncryptionType == ENCRYPTION_CONTEXT + ? getECProviderEnabledFS() + : getCPKEnabledFS(); + } + String relativePath = fs.getAbfsStore().getRelativePath(testPath); + try (FSDataOutputStream out = fs.create(new Path(relativePath))) { + out.write(SERVER_FILE_CONTENT.getBytes()); + } + // verify file is encrypted by calling getPathStatus (with properties) + // without encryption headers in request + if (fileEncryptionType != EncryptionType.NONE) { + final AbfsClient abfsClient = fs.getAbfsClient(); + abfsClient.setEncryptionType(EncryptionType.NONE); + LambdaTestUtils.intercept(IOException.class, () -> + abfsClient.getPathStatus(relativePath, + true, + getTestTracingContext(fs, false), + createEncryptionAdapterFromServerStoreContext(relativePath, + getTestTracingContext(fs, false), abfsClient))); + fs.getAbfsClient().setEncryptionType(fileEncryptionType); + } + return fs.getAbfsClient().getEncryptionContextProvider(); + } + + @Override + public void teardown() throws Exception { + super.teardown(); + for (AzureBlobFileSystem azureBlobFileSystem : fileSystemsOpenedInTest) { + azureBlobFileSystem.close(); + } + } +} diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAbfsInputStreamStatistics.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAbfsInputStreamStatistics.java index d96f1a2836..e8cbeb1255 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAbfsInputStreamStatistics.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAbfsInputStreamStatistics.java @@ -280,7 +280,7 @@ public void testWithNullStreamStatistics() throws IOException { // AbfsRestOperation Instance required for eTag. AbfsRestOperation abfsRestOperation = fs.getAbfsClient() .getPathStatus(nullStatFilePath.toUri().getPath(), false, - getTestTracingContext(fs, false)); + getTestTracingContext(fs, false), null); // AbfsInputStream with no StreamStatistics. in = new AbfsInputStream(fs.getAbfsClient(), null, diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemCreate.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemCreate.java index d0b3ff2974..f972fb03b8 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemCreate.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemCreate.java @@ -35,6 +35,8 @@ import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants; +import org.apache.hadoop.fs.azurebfs.security.ContextEncryptionAdapter; +import org.apache.hadoop.fs.azurebfs.services.AbfsClientUtils; import org.apache.hadoop.fs.permission.FsAction; import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.test.GenericTestUtils; @@ -57,6 +59,7 @@ import static org.mockito.ArgumentMatchers.any; import static org.mockito.ArgumentMatchers.eq; +import static org.mockito.ArgumentMatchers.nullable; import static org.mockito.Mockito.doThrow; import static org.mockito.Mockito.mock; import static org.mockito.Mockito.when; @@ -278,6 +281,7 @@ public void testCreateFileOverwrite(boolean enableConditionalCreateOverwrite) final AzureBlobFileSystem fs = (AzureBlobFileSystem) FileSystem.newInstance(currentFs.getUri(), config); + AbfsClientUtils.setIsNamespaceEnabled(fs.getAbfsClient(), true); long totalConnectionMadeBeforeTest = fs.getInstrumentationMap() .get(CONNECTIONS_MADE.getStatName()); @@ -421,16 +425,16 @@ public void testNegativeScenariosForCreateOverwriteDisabled() serverErrorResponseEx) // Scn5: create overwrite=false fails with Http500 .when(mockClient) .createPath(any(String.class), eq(true), eq(false), - isNamespaceEnabled ? any(String.class) : eq(null), - isNamespaceEnabled ? any(String.class) : eq(null), - any(boolean.class), eq(null), any(TracingContext.class)); + any(AzureBlobFileSystemStore.Permissions.class), any(boolean.class), eq(null), any(), + any(TracingContext.class)); doThrow(fileNotFoundResponseEx) // Scn1: GFS fails with Http404 .doThrow(serverErrorResponseEx) // Scn2: GFS fails with Http500 .doReturn(successOp) // Scn3: create overwrite=true fails with Http412 .doReturn(successOp) // Scn4: create overwrite=true fails with Http500 .when(mockClient) - .getPathStatus(any(String.class), eq(false), any(TracingContext.class)); + .getPathStatus(any(String.class), eq(false), any(TracingContext.class), nullable( + ContextEncryptionAdapter.class)); // mock for overwrite=true doThrow( @@ -439,9 +443,8 @@ public void testNegativeScenariosForCreateOverwriteDisabled() serverErrorResponseEx) // Scn4: create overwrite=true fails with Http500 .when(mockClient) .createPath(any(String.class), eq(true), eq(true), - isNamespaceEnabled ? any(String.class) : eq(null), - isNamespaceEnabled ? any(String.class) : eq(null), - any(boolean.class), eq(null), any(TracingContext.class)); + any(AzureBlobFileSystemStore.Permissions.class), any(boolean.class), eq(null), any(), + any(TracingContext.class)); // Scn1: GFS fails with Http404 // Sequence of events expected: diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemRandomRead.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemRandomRead.java index c1f0e06439..940d56fecb 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemRandomRead.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemRandomRead.java @@ -508,7 +508,7 @@ public void testAlwaysReadBufferSizeConfig(boolean alwaysReadBufferSizeConfigVal 1 * MEGABYTE, config); String eTag = fs.getAbfsClient() .getPathStatus(testFile.toUri().getPath(), false, - getTestTracingContext(fs, false)) + getTestTracingContext(fs, false), null) .getResult() .getResponseHeader(ETAG); diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestCustomerProvidedKey.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestCustomerProvidedKey.java deleted file mode 100644 index 9ae3bf2ee2..0000000000 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestCustomerProvidedKey.java +++ /dev/null @@ -1,976 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 ("License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - *

- * http://www.apache.org/licenses/LICENSE-2.0 - *

- * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.fs.azurebfs; - -import java.io.FileNotFoundException; -import java.io.IOException; -import java.nio.CharBuffer; -import java.nio.charset.CharacterCodingException; -import java.nio.charset.Charset; -import java.nio.charset.CharsetEncoder; -import java.nio.charset.StandardCharsets; -import java.security.MessageDigest; -import java.security.NoSuchAlgorithmException; -import java.util.EnumSet; -import java.util.Hashtable; -import java.util.List; -import java.util.Map; -import java.util.Optional; -import java.util.Random; -import java.util.UUID; - -import org.apache.hadoop.fs.azurebfs.utils.TracingContext; -import org.apache.hadoop.fs.contract.ContractTestUtils; -import org.apache.hadoop.util.Preconditions; -import org.apache.hadoop.util.Lists; -import org.assertj.core.api.Assertions; -import org.junit.Assume; -import org.junit.Test; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; - -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FSDataInputStream; -import org.apache.hadoop.fs.FSDataOutputStream; -import org.apache.hadoop.fs.FileStatus; -import org.apache.hadoop.fs.FileSystem; -import org.apache.hadoop.fs.Path; -import org.apache.hadoop.fs.XAttrSetFlag; -import org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants; -import org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations; -import org.apache.hadoop.fs.azurebfs.contracts.services.AppendRequestParameters; -import org.apache.hadoop.fs.azurebfs.contracts.services.AppendRequestParameters.Mode; -import org.apache.hadoop.fs.azurebfs.services.AbfsAclHelper; -import org.apache.hadoop.fs.azurebfs.services.AbfsClient; -import org.apache.hadoop.fs.azurebfs.services.AbfsHttpHeader; -import org.apache.hadoop.fs.azurebfs.services.AbfsHttpOperation; -import org.apache.hadoop.fs.azurebfs.services.AbfsRestOperation; -import org.apache.hadoop.fs.azurebfs.services.AuthType; -import org.apache.hadoop.fs.azurebfs.utils.Base64; -import org.apache.hadoop.fs.permission.AclEntry; -import org.apache.hadoop.fs.permission.FsAction; -import org.apache.hadoop.fs.permission.FsPermission; -import org.apache.hadoop.test.LambdaTestUtils; - -import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.AZURE_CREATE_REMOTE_FILESYSTEM_DURING_INITIALIZATION; -import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY; -import static org.apache.hadoop.fs.azurebfs.constants.FileSystemConfigurations.ONE_MB; -import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_ENCRYPTION_ALGORITHM; -import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_ENCRYPTION_KEY; -import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_ENCRYPTION_KEY_SHA256; -import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_REQUEST_SERVER_ENCRYPTED; -import static org.apache.hadoop.fs.azurebfs.constants.HttpHeaderConfigurations.X_MS_SERVER_ENCRYPTED; -import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.FS_AZURE_ABFS_ACCOUNT_NAME; -import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.FS_AZURE_ACCOUNT_KEY; -import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.FS_AZURE_TEST_CPK_ENABLED; -import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.FS_AZURE_TEST_CPK_ENABLED_SECONDARY_ACCOUNT; -import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.FS_AZURE_TEST_CPK_ENABLED_SECONDARY_ACCOUNT_KEY; -import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.FS_AZURE_TEST_NAMESPACE_ENABLED_ACCOUNT; -import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.TEST_CONFIGURATION_FILE_NAME; -import static org.apache.hadoop.fs.azurebfs.utils.AclTestHelpers.aclEntry; -import static org.apache.hadoop.fs.permission.AclEntryScope.ACCESS; -import static org.apache.hadoop.fs.permission.AclEntryType.USER; -import static org.apache.hadoop.fs.permission.FsAction.ALL; - -public class ITestCustomerProvidedKey extends AbstractAbfsIntegrationTest { - private static final Logger LOG = LoggerFactory - .getLogger(ITestCustomerProvidedKey.class); - - private static final String XMS_PROPERTIES_ENCODING = "ISO-8859-1"; - private static final int INT_512 = 512; - private static final int INT_50 = 50; - private static final int ENCRYPTION_KEY_LEN = 32; - private static final int FILE_SIZE = 10 * ONE_MB; - private static final int FILE_SIZE_FOR_COPY_BETWEEN_ACCOUNTS = 24 * ONE_MB; - - private boolean isNamespaceEnabled; - - public ITestCustomerProvidedKey() throws Exception { - boolean isCPKTestsEnabled = getConfiguration() - .getBoolean(FS_AZURE_TEST_CPK_ENABLED, false); - Assume.assumeTrue(isCPKTestsEnabled); - isNamespaceEnabled = getConfiguration() - .getBoolean(FS_AZURE_TEST_NAMESPACE_ENABLED_ACCOUNT, false); - } - - private String getFileName() throws IOException { - return path("/" + methodName.getMethodName()).toUri().getPath(); - } - - @Test - public void testReadWithCPK() throws Exception { - final AzureBlobFileSystem fs = getAbfs(true); - String fileName = getFileName(); - createFileAndGetContent(fs, fileName, FILE_SIZE); - - AbfsClient abfsClient = fs.getAbfsClient(); - int length = FILE_SIZE; - byte[] buffer = new byte[length]; - TracingContext tracingContext = getTestTracingContext(fs, false); - final AbfsRestOperation op = abfsClient.getPathStatus(fileName, false, - tracingContext); - final String eTag = op.getResult() - .getResponseHeader(HttpHeaderConfigurations.ETAG); - AbfsRestOperation abfsRestOperation = abfsClient - .read(fileName, 0, buffer, 0, length, eTag, null, tracingContext); - assertCPKHeaders(abfsRestOperation, true); - assertResponseHeader(abfsRestOperation, true, X_MS_ENCRYPTION_KEY_SHA256, - getCPKSha(fs)); - assertResponseHeader(abfsRestOperation, true, X_MS_SERVER_ENCRYPTED, - "true"); - assertResponseHeader(abfsRestOperation, false, - X_MS_REQUEST_SERVER_ENCRYPTED, ""); - - // Trying to read with different CPK headers - Configuration conf = fs.getConf(); - String accountName = conf.get(FS_AZURE_ABFS_ACCOUNT_NAME); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - "different-1234567890123456789012"); - try (AzureBlobFileSystem fs2 = (AzureBlobFileSystem) FileSystem.newInstance(conf); - FSDataInputStream iStream = fs2.open(new Path(fileName))) { - int len = 8 * ONE_MB; - byte[] b = new byte[len]; - LambdaTestUtils.intercept(IOException.class, () -> { - iStream.read(b, 0, len); - }); - } - - // Trying to read with no CPK headers - conf.unset(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName); - try (AzureBlobFileSystem fs3 = (AzureBlobFileSystem) FileSystem - .get(conf); FSDataInputStream iStream = fs3.open(new Path(fileName))) { - int len = 8 * ONE_MB; - byte[] b = new byte[len]; - LambdaTestUtils.intercept(IOException.class, () -> { - iStream.read(b, 0, len); - }); - } - } - - @Test - public void testReadWithoutCPK() throws Exception { - final AzureBlobFileSystem fs = getAbfs(false); - String fileName = getFileName(); - createFileAndGetContent(fs, fileName, FILE_SIZE); - - AbfsClient abfsClient = fs.getAbfsClient(); - int length = INT_512; - byte[] buffer = new byte[length * 4]; - TracingContext tracingContext = getTestTracingContext(fs, false); - final AbfsRestOperation op = abfsClient - .getPathStatus(fileName, false, tracingContext); - final String eTag = op.getResult() - .getResponseHeader(HttpHeaderConfigurations.ETAG); - AbfsRestOperation abfsRestOperation = abfsClient - .read(fileName, 0, buffer, 0, length, eTag, null, tracingContext); - assertCPKHeaders(abfsRestOperation, false); - assertResponseHeader(abfsRestOperation, false, X_MS_ENCRYPTION_KEY_SHA256, - getCPKSha(fs)); - assertResponseHeader(abfsRestOperation, true, X_MS_SERVER_ENCRYPTED, - "true"); - assertResponseHeader(abfsRestOperation, false, - X_MS_REQUEST_SERVER_ENCRYPTED, ""); - - // Trying to read with CPK headers - Configuration conf = fs.getConf(); - String accountName = conf.get(FS_AZURE_ABFS_ACCOUNT_NAME); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - "12345678901234567890123456789012"); - - try (AzureBlobFileSystem fs2 = (AzureBlobFileSystem) FileSystem.newInstance(conf); - AbfsClient abfsClient2 = fs2.getAbfsClient()) { - LambdaTestUtils.intercept(IOException.class, () -> { - abfsClient2.read(fileName, 0, buffer, 0, length, eTag, null, - getTestTracingContext(fs, false)); - }); - } - } - - @Test - public void testAppendWithCPK() throws Exception { - final AzureBlobFileSystem fs = getAbfs(true); - final String fileName = getFileName(); - createFileAndGetContent(fs, fileName, FILE_SIZE); - - // Trying to append with correct CPK headers - AppendRequestParameters appendRequestParameters = - new AppendRequestParameters( - 0, 0, 5, Mode.APPEND_MODE, false, null, true); - byte[] buffer = getRandomBytesArray(5); - AbfsClient abfsClient = fs.getAbfsClient(); - AbfsRestOperation abfsRestOperation = abfsClient - .append(fileName, buffer, appendRequestParameters, null, getTestTracingContext(fs, false)); - assertCPKHeaders(abfsRestOperation, true); - assertResponseHeader(abfsRestOperation, true, X_MS_ENCRYPTION_KEY_SHA256, - getCPKSha(fs)); - assertResponseHeader(abfsRestOperation, false, X_MS_SERVER_ENCRYPTED, ""); - assertResponseHeader(abfsRestOperation, true, X_MS_REQUEST_SERVER_ENCRYPTED, - "true"); - - // Trying to append with different CPK headers - Configuration conf = fs.getConf(); - String accountName = conf.get(FS_AZURE_ABFS_ACCOUNT_NAME); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - "different-1234567890123456789012"); - try (AzureBlobFileSystem fs2 = (AzureBlobFileSystem) FileSystem.newInstance(conf); - AbfsClient abfsClient2 = fs2.getAbfsClient()) { - LambdaTestUtils.intercept(IOException.class, () -> { - abfsClient2.append(fileName, buffer, appendRequestParameters, null, - getTestTracingContext(fs, false)); - }); - } - - // Trying to append with no CPK headers - conf.unset(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName); - try (AzureBlobFileSystem fs3 = (AzureBlobFileSystem) FileSystem - .get(conf); AbfsClient abfsClient3 = fs3.getAbfsClient()) { - LambdaTestUtils.intercept(IOException.class, () -> { - abfsClient3.append(fileName, buffer, appendRequestParameters, null, - getTestTracingContext(fs, false)); - }); - } - } - - @Test - public void testAppendWithoutCPK() throws Exception { - final AzureBlobFileSystem fs = getAbfs(false); - final String fileName = getFileName(); - createFileAndGetContent(fs, fileName, FILE_SIZE); - - // Trying to append without CPK headers - AppendRequestParameters appendRequestParameters = - new AppendRequestParameters( - 0, 0, 5, Mode.APPEND_MODE, false, null, true); - byte[] buffer = getRandomBytesArray(5); - AbfsClient abfsClient = fs.getAbfsClient(); - AbfsRestOperation abfsRestOperation = abfsClient - .append(fileName, buffer, appendRequestParameters, null, - getTestTracingContext(fs, false)); - assertCPKHeaders(abfsRestOperation, false); - assertResponseHeader(abfsRestOperation, false, X_MS_ENCRYPTION_KEY_SHA256, - ""); - assertResponseHeader(abfsRestOperation, false, X_MS_SERVER_ENCRYPTED, ""); - assertResponseHeader(abfsRestOperation, true, X_MS_REQUEST_SERVER_ENCRYPTED, - "true"); - - // Trying to append with CPK headers - Configuration conf = fs.getConf(); - String accountName = conf.get(FS_AZURE_ABFS_ACCOUNT_NAME); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - "12345678901234567890123456789012"); - try (AzureBlobFileSystem fs2 = (AzureBlobFileSystem) FileSystem.newInstance(conf); - AbfsClient abfsClient2 = fs2.getAbfsClient()) { - LambdaTestUtils.intercept(IOException.class, () -> { - abfsClient2.append(fileName, buffer, appendRequestParameters, null, - getTestTracingContext(fs, false)); - }); - } - } - - @Test - public void testSetGetXAttr() throws Exception { - final AzureBlobFileSystem fs = getAbfs(true); - final String fileName = getFileName(); - createFileAndGetContent(fs, fileName, FILE_SIZE); - - String valSent = "testValue"; - String attrName = "testXAttr"; - - // set get and verify - fs.setXAttr(new Path(fileName), attrName, - valSent.getBytes(StandardCharsets.UTF_8), - EnumSet.of(XAttrSetFlag.CREATE)); - byte[] valBytes = fs.getXAttr(new Path(fileName), attrName); - String valRecieved = new String(valBytes); - assertEquals(valSent, valRecieved); - - // set new value get and verify - valSent = "new value"; - fs.setXAttr(new Path(fileName), attrName, - valSent.getBytes(StandardCharsets.UTF_8), - EnumSet.of(XAttrSetFlag.REPLACE)); - valBytes = fs.getXAttr(new Path(fileName), attrName); - valRecieved = new String(valBytes); - assertEquals(valSent, valRecieved); - - // Read without CPK header - LambdaTestUtils.intercept(IOException.class, () -> { - getAbfs(false).getXAttr(new Path(fileName), attrName); - }); - - // Wrong CPK - LambdaTestUtils.intercept(IOException.class, () -> { - getSameFSWithWrongCPK(fs).getXAttr(new Path(fileName), attrName); - }); - } - - @Test - public void testCopyBetweenAccounts() throws Exception { - String accountName = getRawConfiguration() - .get(FS_AZURE_TEST_CPK_ENABLED_SECONDARY_ACCOUNT); - String accountKey = getRawConfiguration() - .get(FS_AZURE_TEST_CPK_ENABLED_SECONDARY_ACCOUNT_KEY); - Assume.assumeTrue(accountName != null && !accountName.isEmpty()); - Assume.assumeTrue(accountKey != null && !accountKey.isEmpty()); - String fileSystemName = "cpkfs"; - - // Create fs1 and a file with CPK - AzureBlobFileSystem fs1 = getAbfs(true); - int fileSize = FILE_SIZE_FOR_COPY_BETWEEN_ACCOUNTS; - byte[] fileContent = getRandomBytesArray(fileSize); - Path testFilePath = createFileWithContent(fs1, - String.format("fs1-file%s.txt", UUID.randomUUID()), fileContent); - - // Create fs2 with different CPK - Configuration conf = new Configuration(); - conf.addResource(TEST_CONFIGURATION_FILE_NAME); - conf.setBoolean(AZURE_CREATE_REMOTE_FILESYSTEM_DURING_INITIALIZATION, true); - conf.unset(FS_AZURE_ABFS_ACCOUNT_NAME); - conf.set(FS_AZURE_ABFS_ACCOUNT_NAME, accountName); - conf.set(FS_AZURE_ACCOUNT_KEY + "." + accountName, accountKey); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - "123456789012345678901234567890ab"); - conf.set("fs.defaultFS", "abfs://" + fileSystemName + "@" + accountName); - AzureBlobFileSystem fs2 = (AzureBlobFileSystem) FileSystem.newInstance(conf); - - // Read from fs1 and write to fs2, fs1 and fs2 are having different CPK - Path fs2DestFilePath = new Path( - String.format("fs2-dest-file%s.txt", UUID.randomUUID())); - FSDataOutputStream ops = fs2.create(fs2DestFilePath); - try (FSDataInputStream iStream = fs1.open(testFilePath)) { - long totalBytesRead = 0; - do { - int length = 8 * ONE_MB; - byte[] buffer = new byte[length]; - int bytesRead = iStream.read(buffer, 0, length); - totalBytesRead += bytesRead; - ops.write(buffer); - } while (totalBytesRead < fileContent.length); - ops.close(); - } - - // Trying to read fs2DestFilePath with different CPK headers - conf.unset(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - "different-1234567890123456789012"); - try (AzureBlobFileSystem fs3 = (AzureBlobFileSystem) FileSystem - .get(conf); FSDataInputStream iStream = fs3.open(fs2DestFilePath)) { - int length = 8 * ONE_MB; - byte[] buffer = new byte[length]; - LambdaTestUtils.intercept(IOException.class, () -> { - iStream.read(buffer, 0, length); - }); - } - - // Trying to read fs2DestFilePath with no CPK headers - conf.unset(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName); - try (AzureBlobFileSystem fs4 = (AzureBlobFileSystem) FileSystem - .get(conf); FSDataInputStream iStream = fs4.open(fs2DestFilePath)) { - int length = 8 * ONE_MB; - byte[] buffer = new byte[length]; - LambdaTestUtils.intercept(IOException.class, () -> { - iStream.read(buffer, 0, length); - }); - } - - // Read fs2DestFilePath and verify the content with the initial random - // bytes created and wrote into the source file at fs1 - try (FSDataInputStream iStream = fs2.open(fs2DestFilePath)) { - long totalBytesRead = 0; - int pos = 0; - do { - int length = 8 * ONE_MB; - byte[] buffer = new byte[length]; - int bytesRead = iStream.read(buffer, 0, length); - totalBytesRead += bytesRead; - for (int i = 0; i < bytesRead; i++) { - assertEquals(fileContent[pos + i], buffer[i]); - } - pos = pos + bytesRead; - } while (totalBytesRead < fileContent.length); - } - } - - @Test - public void testListPathWithCPK() throws Exception { - testListPath(true); - } - - @Test - public void testListPathWithoutCPK() throws Exception { - testListPath(false); - } - - private void testListPath(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final Path testPath = path("/" + methodName.getMethodName()); - String testDirName = testPath.toUri().getPath(); - fs.mkdirs(testPath); - createFileAndGetContent(fs, testDirName + "/aaa", FILE_SIZE); - createFileAndGetContent(fs, testDirName + "/bbb", FILE_SIZE); - AbfsClient abfsClient = fs.getAbfsClient(); - AbfsRestOperation abfsRestOperation = abfsClient - .listPath(testDirName, false, INT_50, null, - getTestTracingContext(fs, false)); - assertListstatus(fs, abfsRestOperation, testPath); - - // Trying with different CPK headers - Configuration conf = fs.getConf(); - String accountName = conf.get(FS_AZURE_ABFS_ACCOUNT_NAME); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - "different-1234567890123456789012"); - AzureBlobFileSystem fs2 = (AzureBlobFileSystem) FileSystem.newInstance(conf); - AbfsClient abfsClient2 = fs2.getAbfsClient(); - TracingContext tracingContext = getTestTracingContext(fs, false); - abfsRestOperation = abfsClient2.listPath(testDirName, false, INT_50, - null, tracingContext); - assertListstatus(fs, abfsRestOperation, testPath); - - if (isWithCPK) { - // Trying with no CPK headers - conf.unset(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName); - AzureBlobFileSystem fs3 = (AzureBlobFileSystem) FileSystem.get(conf); - AbfsClient abfsClient3 = fs3.getAbfsClient(); - abfsRestOperation = abfsClient3 - .listPath(testDirName, false, INT_50, null, tracingContext); - assertListstatus(fs, abfsRestOperation, testPath); - } - } - - private void assertListstatus(AzureBlobFileSystem fs, - AbfsRestOperation abfsRestOperation, Path testPath) throws IOException { - assertCPKHeaders(abfsRestOperation, false); - assertNoCPKResponseHeadersPresent(abfsRestOperation); - - FileStatus[] listStatuses = fs.listStatus(testPath); - Assertions.assertThat(listStatuses.length) - .describedAs("listStatuses should have 2 entries").isEqualTo(2); - - listStatuses = getSameFSWithWrongCPK(fs).listStatus(testPath); - Assertions.assertThat(listStatuses.length) - .describedAs("listStatuses should have 2 entries").isEqualTo(2); - } - - @Test - public void testCreatePathWithCPK() throws Exception { - testCreatePath(true); - } - - @Test - public void testCreatePathWithoutCPK() throws Exception { - testCreatePath(false); - } - - private void testCreatePath(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - createFileAndGetContent(fs, testFileName, FILE_SIZE); - - AbfsClient abfsClient = fs.getAbfsClient(); - FsPermission permission = new FsPermission(FsAction.EXECUTE, - FsAction.EXECUTE, FsAction.EXECUTE); - FsPermission umask = new FsPermission(FsAction.NONE, FsAction.NONE, - FsAction.NONE); - TracingContext tracingContext = getTestTracingContext(fs, false); - boolean isNamespaceEnabled = fs.getIsNamespaceEnabled(tracingContext); - AbfsRestOperation abfsRestOperation = abfsClient - .createPath(testFileName, true, true, - isNamespaceEnabled ? getOctalNotation(permission) : null, - isNamespaceEnabled ? getOctalNotation(umask) : null, false, null, - tracingContext); - assertCPKHeaders(abfsRestOperation, isWithCPK); - assertResponseHeader(abfsRestOperation, isWithCPK, - X_MS_ENCRYPTION_KEY_SHA256, getCPKSha(fs)); - assertResponseHeader(abfsRestOperation, false, X_MS_SERVER_ENCRYPTED, ""); - assertResponseHeader(abfsRestOperation, true, X_MS_REQUEST_SERVER_ENCRYPTED, - "true"); - - FileStatus[] listStatuses = fs.listStatus(new Path(testFileName)); - Assertions.assertThat(listStatuses.length) - .describedAs("listStatuses should have 1 entry").isEqualTo(1); - - listStatuses = getSameFSWithWrongCPK(fs).listStatus(new Path(testFileName)); - Assertions.assertThat(listStatuses.length) - .describedAs("listStatuses should have 1 entry").isEqualTo(1); - } - - @Test - public void testRenamePathWithCPK() throws Exception { - testRenamePath(true); - } - - @Test - public void testRenamePathWithoutCPK() throws Exception { - testRenamePath(false); - } - - private void testRenamePath(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - createFileAndGetContent(fs, testFileName, FILE_SIZE); - - FileStatus fileStatusBeforeRename = fs - .getFileStatus(new Path(testFileName)); - - String newName = "/newName"; - AbfsClient abfsClient = fs.getAbfsClient(); - AbfsRestOperation abfsRestOperation = abfsClient - .renamePath(testFileName, newName, null, - getTestTracingContext(fs, false), null, false, isNamespaceEnabled) - .getOp(); - assertCPKHeaders(abfsRestOperation, false); - assertNoCPKResponseHeadersPresent(abfsRestOperation); - - LambdaTestUtils.intercept(FileNotFoundException.class, - (() -> fs.getFileStatus(new Path(testFileName)))); - - FileStatus fileStatusAfterRename = fs.getFileStatus(new Path(newName)); - Assertions.assertThat(fileStatusAfterRename.getLen()) - .describedAs("File size has to be same before and after rename") - .isEqualTo(fileStatusBeforeRename.getLen()); - } - - @Test - public void testFlushWithCPK() throws Exception { - testFlush(true); - } - - @Test - public void testFlushWithoutCPK() throws Exception { - testFlush(false); - } - - private void testFlush(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - fs.create(new Path(testFileName)).close(); - AbfsClient abfsClient = fs.getAbfsClient(); - String expectedCPKSha = getCPKSha(fs); - - byte[] fileContent = getRandomBytesArray(FILE_SIZE); - Path testFilePath = new Path(testFileName + "1"); - try (FSDataOutputStream oStream = fs.create(testFilePath)) { - oStream.write(fileContent); - } - - // Trying to read with different CPK headers - Configuration conf = fs.getConf(); - String accountName = conf.get(FS_AZURE_ABFS_ACCOUNT_NAME); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - "different-1234567890123456789012"); - try (AzureBlobFileSystem fs2 = (AzureBlobFileSystem) FileSystem.newInstance(conf); - AbfsClient abfsClient2 = fs2.getAbfsClient()) { - LambdaTestUtils.intercept(IOException.class, () -> { - abfsClient2.flush(testFileName, 0, false, false, null, null, - getTestTracingContext(fs, false)); - }); - } - - // Trying to read with no CPK headers - if (isWithCPK) { - conf.unset(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName); - try (AzureBlobFileSystem fs3 = (AzureBlobFileSystem) FileSystem - .get(conf); AbfsClient abfsClient3 = fs3.getAbfsClient()) { - LambdaTestUtils.intercept(IOException.class, () -> { - abfsClient3.flush(testFileName, 0, false, false, null, null, - getTestTracingContext(fs, false)); - }); - } - } - - // With correct CPK - AbfsRestOperation abfsRestOperation = abfsClient - .flush(testFileName, 0, false, false, null, null, - getTestTracingContext(fs, false)); - assertCPKHeaders(abfsRestOperation, isWithCPK); - assertResponseHeader(abfsRestOperation, isWithCPK, - X_MS_ENCRYPTION_KEY_SHA256, expectedCPKSha); - assertResponseHeader(abfsRestOperation, false, X_MS_SERVER_ENCRYPTED, ""); - assertResponseHeader(abfsRestOperation, true, X_MS_REQUEST_SERVER_ENCRYPTED, - isWithCPK + ""); - } - - @Test - public void testSetPathPropertiesWithCPK() throws Exception { - testSetPathProperties(true); - } - - @Test - public void testSetPathPropertiesWithoutCPK() throws Exception { - testSetPathProperties(false); - } - - private void testSetPathProperties(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - createFileAndGetContent(fs, testFileName, FILE_SIZE); - - AbfsClient abfsClient = fs.getAbfsClient(); - final Hashtable properties = new Hashtable<>(); - properties.put("key", "val"); - AbfsRestOperation abfsRestOperation = abfsClient - .setPathProperties(testFileName, - convertXmsPropertiesToCommaSeparatedString(properties), - getTestTracingContext(fs, false)); - assertCPKHeaders(abfsRestOperation, isWithCPK); - assertResponseHeader(abfsRestOperation, isWithCPK, - X_MS_ENCRYPTION_KEY_SHA256, getCPKSha(fs)); - assertResponseHeader(abfsRestOperation, false, X_MS_SERVER_ENCRYPTED, ""); - assertResponseHeader(abfsRestOperation, true, X_MS_REQUEST_SERVER_ENCRYPTED, - "true"); - } - - @Test - public void testGetPathStatusFileWithCPK() throws Exception { - testGetPathStatusFile(true); - } - - @Test - public void testGetPathStatusFileWithoutCPK() throws Exception { - testGetPathStatusFile(false); - } - - private void testGetPathStatusFile(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - createFileAndGetContent(fs, testFileName, FILE_SIZE); - - AbfsClient abfsClient = fs.getAbfsClient(); - TracingContext tracingContext = getTestTracingContext(fs, false); - AbfsRestOperation abfsRestOperation = abfsClient - .getPathStatus(testFileName, false, tracingContext); - assertCPKHeaders(abfsRestOperation, false); - assertResponseHeader(abfsRestOperation, isWithCPK, - X_MS_ENCRYPTION_KEY_SHA256, getCPKSha(fs)); - assertResponseHeader(abfsRestOperation, true, X_MS_SERVER_ENCRYPTED, - "true"); - assertResponseHeader(abfsRestOperation, false, - X_MS_REQUEST_SERVER_ENCRYPTED, ""); - - abfsRestOperation = abfsClient.getPathStatus(testFileName, true, tracingContext); - assertCPKHeaders(abfsRestOperation, isWithCPK); - assertResponseHeader(abfsRestOperation, isWithCPK, - X_MS_ENCRYPTION_KEY_SHA256, getCPKSha(fs)); - assertResponseHeader(abfsRestOperation, true, X_MS_SERVER_ENCRYPTED, - "true"); - assertResponseHeader(abfsRestOperation, false, - X_MS_REQUEST_SERVER_ENCRYPTED, ""); - } - - @Test - public void testDeletePathWithCPK() throws Exception { - testDeletePath(false); - } - - @Test - public void testDeletePathWithoutCPK() throws Exception { - testDeletePath(false); - } - - private void testDeletePath(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - createFileAndGetContent(fs, testFileName, FILE_SIZE); - - FileStatus[] listStatuses = fs.listStatus(new Path(testFileName)); - Assertions.assertThat(listStatuses.length) - .describedAs("listStatuses should have 1 entry").isEqualTo(1); - - AbfsClient abfsClient = fs.getAbfsClient(); - AbfsRestOperation abfsRestOperation = abfsClient - .deletePath(testFileName, false, null, - getTestTracingContext(fs, false)); - assertCPKHeaders(abfsRestOperation, false); - assertNoCPKResponseHeadersPresent(abfsRestOperation); - - Assertions.assertThatThrownBy(() -> fs.listStatus(new Path(testFileName))) - .isInstanceOf(FileNotFoundException.class); - } - - @Test - public void testSetPermissionWithCPK() throws Exception { - testSetPermission(true); - } - - @Test - public void testSetPermissionWithoutCPK() throws Exception { - testSetPermission(false); - } - - private void testSetPermission(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - Assume.assumeTrue(fs.getIsNamespaceEnabled(getTestTracingContext(fs, false))); - createFileAndGetContent(fs, testFileName, FILE_SIZE); - AbfsClient abfsClient = fs.getAbfsClient(); - FsPermission permission = new FsPermission(FsAction.EXECUTE, - FsAction.EXECUTE, FsAction.EXECUTE); - AbfsRestOperation abfsRestOperation = abfsClient - .setPermission(testFileName, permission.toString(), - getTestTracingContext(fs, false)); - assertCPKHeaders(abfsRestOperation, false); - assertNoCPKResponseHeadersPresent(abfsRestOperation); - } - - @Test - public void testSetAclWithCPK() throws Exception { - testSetAcl(true); - } - - @Test - public void testSetAclWithoutCPK() throws Exception { - testSetAcl(false); - } - - private void testSetAcl(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - TracingContext tracingContext = getTestTracingContext(fs, false); - Assume.assumeTrue(fs.getIsNamespaceEnabled(tracingContext)); - createFileAndGetContent(fs, testFileName, FILE_SIZE); - AbfsClient abfsClient = fs.getAbfsClient(); - - List aclSpec = Lists.newArrayList(aclEntry(ACCESS, USER, ALL)); - final Map aclEntries = AbfsAclHelper - .deserializeAclSpec(AclEntry.aclSpecToString(aclSpec)); - - AbfsRestOperation abfsRestOperation = abfsClient - .setAcl(testFileName, AbfsAclHelper.serializeAclSpec(aclEntries), - tracingContext); - assertCPKHeaders(abfsRestOperation, false); - assertNoCPKResponseHeadersPresent(abfsRestOperation); - } - - @Test - public void testGetAclWithCPK() throws Exception { - testGetAcl(true); - } - - @Test - public void testGetAclWithoutCPK() throws Exception { - testGetAcl(false); - } - - private void testGetAcl(final boolean isWithCPK) throws Exception { - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - TracingContext tracingContext = getTestTracingContext(fs, false); - Assume.assumeTrue(fs.getIsNamespaceEnabled(tracingContext)); - createFileAndGetContent(fs, testFileName, FILE_SIZE); - AbfsClient abfsClient = fs.getAbfsClient(); - AbfsRestOperation abfsRestOperation = - abfsClient.getAclStatus(testFileName, tracingContext); - assertCPKHeaders(abfsRestOperation, false); - assertNoCPKResponseHeadersPresent(abfsRestOperation); - } - - @Test - public void testCheckAccessWithCPK() throws Exception { - testCheckAccess(true); - } - - @Test - public void testCheckAccessWithoutCPK() throws Exception { - testCheckAccess(false); - } - - private void testCheckAccess(final boolean isWithCPK) throws Exception { - boolean isHNSEnabled = getConfiguration() - .getBoolean(FS_AZURE_TEST_NAMESPACE_ENABLED_ACCOUNT, false); - Assume.assumeTrue(FS_AZURE_TEST_NAMESPACE_ENABLED_ACCOUNT + " is false", - isHNSEnabled); - Assume.assumeTrue("AuthType has to be OAuth", - getAuthType() == AuthType.OAuth); - - final AzureBlobFileSystem fs = getAbfs(isWithCPK); - final String testFileName = getFileName(); - fs.create(new Path(testFileName)).close(); - AbfsClient abfsClient = fs.getAbfsClient(); - AbfsRestOperation abfsRestOperation = abfsClient - .checkAccess(testFileName, "rwx", getTestTracingContext(fs, false)); - assertCPKHeaders(abfsRestOperation, false); - assertNoCPKResponseHeadersPresent(abfsRestOperation); - } - - private byte[] createFileAndGetContent(AzureBlobFileSystem fs, - String fileName, int fileSize) throws IOException { - byte[] fileContent = getRandomBytesArray(fileSize); - Path testFilePath = createFileWithContent(fs, fileName, fileContent); - ContractTestUtils.verifyFileContents(fs, testFilePath, fileContent); - return fileContent; - } - - private void assertCPKHeaders(AbfsRestOperation abfsRestOperation, - boolean isCPKHeaderExpected) { - assertHeader(abfsRestOperation, X_MS_ENCRYPTION_KEY, isCPKHeaderExpected); - assertHeader(abfsRestOperation, X_MS_ENCRYPTION_KEY_SHA256, - isCPKHeaderExpected); - assertHeader(abfsRestOperation, X_MS_ENCRYPTION_ALGORITHM, - isCPKHeaderExpected); - } - - private void assertNoCPKResponseHeadersPresent( - AbfsRestOperation abfsRestOperation) { - assertResponseHeader(abfsRestOperation, false, X_MS_SERVER_ENCRYPTED, ""); - assertResponseHeader(abfsRestOperation, false, - X_MS_REQUEST_SERVER_ENCRYPTED, ""); - assertResponseHeader(abfsRestOperation, false, X_MS_ENCRYPTION_KEY_SHA256, - ""); - } - - private void assertResponseHeader(AbfsRestOperation abfsRestOperation, - boolean isHeaderExpected, String headerName, String expectedValue) { - final AbfsHttpOperation result = abfsRestOperation.getResult(); - final String value = result.getResponseHeader(headerName); - if (isHeaderExpected) { - Assertions.assertThat(value).isEqualTo(expectedValue); - } else { - Assertions.assertThat(value).isNull(); - } - } - - private void assertHeader(AbfsRestOperation abfsRestOperation, - String headerName, boolean isCPKHeaderExpected) { - assertTrue(abfsRestOperation != null); - Optional header = abfsRestOperation.getRequestHeaders() - .stream().filter(abfsHttpHeader -> abfsHttpHeader.getName() - .equalsIgnoreCase(headerName)).findFirst(); - String desc; - if (isCPKHeaderExpected) { - desc = - "CPK header " + headerName + " is expected, but the same is absent."; - } else { - desc = "CPK header " + headerName - + " is not expected, but the same is present."; - } - Assertions.assertThat(header.isPresent()).describedAs(desc) - .isEqualTo(isCPKHeaderExpected); - } - - private byte[] getSHA256Hash(String key) throws IOException { - try { - final MessageDigest digester = MessageDigest.getInstance("SHA-256"); - return digester.digest(key.getBytes(StandardCharsets.UTF_8)); - } catch (NoSuchAlgorithmException e) { - throw new IOException(e); - } - } - - private String getCPKSha(final AzureBlobFileSystem abfs) throws IOException { - Configuration conf = abfs.getConf(); - String accountName = conf.get(FS_AZURE_ABFS_ACCOUNT_NAME); - String encryptionKey = conf - .get(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName); - if (encryptionKey == null || encryptionKey.isEmpty()) { - return ""; - } - return getBase64EncodedString(getSHA256Hash(encryptionKey)); - } - - private String getBase64EncodedString(byte[] bytes) { - return java.util.Base64.getEncoder().encodeToString(bytes); - } - - private Path createFileWithContent(FileSystem fs, String fileName, - byte[] fileContent) throws IOException { - Path testFilePath = new Path(fileName); - try (FSDataOutputStream oStream = fs.create(testFilePath)) { - oStream.write(fileContent); - oStream.flush(); - } - return testFilePath; - } - - private String convertXmsPropertiesToCommaSeparatedString( - final Hashtable properties) - throws CharacterCodingException { - StringBuilder commaSeparatedProperties = new StringBuilder(); - final CharsetEncoder encoder = Charset.forName(XMS_PROPERTIES_ENCODING) - .newEncoder(); - for (Map.Entry propertyEntry : properties.entrySet()) { - String key = propertyEntry.getKey(); - String value = propertyEntry.getValue(); - Boolean canEncodeValue = encoder.canEncode(value); - if (!canEncodeValue) { - throw new CharacterCodingException(); - } - String encodedPropertyValue = Base64 - .encode(encoder.encode(CharBuffer.wrap(value)).array()); - commaSeparatedProperties.append(key).append(AbfsHttpConstants.EQUAL) - .append(encodedPropertyValue); - commaSeparatedProperties.append(AbfsHttpConstants.COMMA); - } - if (commaSeparatedProperties.length() != 0) { - commaSeparatedProperties - .deleteCharAt(commaSeparatedProperties.length() - 1); - } - return commaSeparatedProperties.toString(); - } - - private String getOctalNotation(FsPermission fsPermission) { - Preconditions.checkNotNull(fsPermission, "fsPermission"); - return String - .format(AbfsHttpConstants.PERMISSION_FORMAT, fsPermission.toOctal()); - } - - private byte[] getRandomBytesArray(int length) { - final byte[] b = new byte[length]; - new Random().nextBytes(b); - return b; - } - - private AzureBlobFileSystem getAbfs(boolean withCPK) throws IOException { - return getAbfs(withCPK, "12345678901234567890123456789012"); - } - - private AzureBlobFileSystem getAbfs(boolean withCPK, String cpk) - throws IOException { - Configuration conf = getRawConfiguration(); - if (withCPK) { - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + getAccountName(), - cpk); - } else { - conf.unset( - FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + getAccountName()); - } - return (AzureBlobFileSystem) FileSystem.newInstance(conf); - } - - private AzureBlobFileSystem getSameFSWithWrongCPK( - final AzureBlobFileSystem fs) throws IOException { - AbfsConfiguration abfsConf = fs.getAbfsStore().getAbfsConfiguration(); - Configuration conf = abfsConf.getRawConfiguration(); - String accountName = conf.get(FS_AZURE_ABFS_ACCOUNT_NAME); - String cpk = conf - .get(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName); - if (cpk == null || cpk.isEmpty()) { - cpk = "01234567890123456789012345678912"; - } - cpk = "different-" + cpk; - String differentCpk = cpk.substring(0, ENCRYPTION_KEY_LEN - 1); - conf.set(FS_AZURE_CLIENT_PROVIDED_ENCRYPTION_KEY + "." + accountName, - differentCpk); - conf.set("fs.defaultFS", - "abfs://" + getFileSystemName() + "@" + accountName); - AzureBlobFileSystem sameFSWithDifferentCPK = - (AzureBlobFileSystem) FileSystem.newInstance(conf); - return sameFSWithDifferentCPK; - } - -} diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/TestTracingContext.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/TestTracingContext.java index 2da530364c..0a19a24de3 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/TestTracingContext.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/TestTracingContext.java @@ -100,17 +100,14 @@ public void checkCorrelationConfigValidation(String clientCorrelationId, TracingHeaderFormat.ALL_ID_FORMAT, null); boolean isNamespaceEnabled = fs.getIsNamespaceEnabled(tracingContext); String path = getRelativePath(new Path("/testDir")); - String permission = isNamespaceEnabled - ? getOctalNotation(FsPermission.getDirDefault()) - : null; - String umask = isNamespaceEnabled - ? getOctalNotation(FsPermission.getUMask(fs.getConf())) - : null; + AzureBlobFileSystemStore.Permissions permissions + = new AzureBlobFileSystemStore.Permissions(isNamespaceEnabled, + FsPermission.getDefault(), FsPermission.getUMask(fs.getConf())); //request should not fail for invalid clientCorrelationID AbfsRestOperation op = fs.getAbfsClient() - .createPath(path, false, true, permission, umask, false, null, - tracingContext); + .createPath(path, false, true, permissions, false, null, null, + tracingContext); int statusCode = op.getResult().getStatusCode(); Assertions.assertThat(statusCode).describedAs("Request should not fail") diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/constants/TestConfigurationKeys.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/constants/TestConfigurationKeys.java index 9e40f22d23..c6a9b47d57 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/constants/TestConfigurationKeys.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/constants/TestConfigurationKeys.java @@ -64,6 +64,7 @@ public final class TestConfigurationKeys { public static final String TEST_CONFIGURATION_FILE_NAME = "azure-test.xml"; public static final String TEST_CONTAINER_PREFIX = "abfs-testcontainer-"; public static final int TEST_TIMEOUT = 15 * 60 * 1000; + public static final int ENCRYPTION_KEY_LEN = 32; private TestConfigurationKeys() {} } diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/extensions/MockEncryptionContextProvider.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/extensions/MockEncryptionContextProvider.java new file mode 100644 index 0000000000..2e94d2648e --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/extensions/MockEncryptionContextProvider.java @@ -0,0 +1,71 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.extensions; + +import java.io.IOException; +import java.nio.charset.StandardCharsets; +import java.util.HashMap; +import java.util.Random; +import java.util.UUID; + +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.azurebfs.security.ABFSKey; + +import static org.apache.hadoop.fs.azurebfs.constants.TestConfigurationKeys.ENCRYPTION_KEY_LEN; + +public class MockEncryptionContextProvider implements EncryptionContextProvider { + private HashMap pathToContextMap = new HashMap<>(); + private HashMap contextToKeyByteMap = new HashMap<>(); + @Override + public void initialize(Configuration configuration, String accountName, + String fileSystem) throws IOException { + } + + @Override + public ABFSKey getEncryptionContext(String path) + throws IOException { + String newContext = UUID.randomUUID().toString(); + pathToContextMap.put(path, newContext); + byte[] newKey = new byte[ENCRYPTION_KEY_LEN]; + new Random().nextBytes(newKey); + ABFSKey key = new ABFSKey(newKey); + contextToKeyByteMap.put(newContext, key.getEncoded()); + return new ABFSKey(newContext.getBytes(StandardCharsets.UTF_8)); + } + + @Override + public ABFSKey getEncryptionKey(String path, ABFSKey encryptionContext) throws IOException { + String encryptionContextString = + new String(encryptionContext.getEncoded(), StandardCharsets.UTF_8); + if (!encryptionContextString.equals(pathToContextMap.get(path))) { + throw new IOException("encryption context does not match path"); + } + return new ABFSKey(contextToKeyByteMap.get(encryptionContextString)); + } + + public byte[] getEncryptionKeyForTest(String encryptionContext) { + return contextToKeyByteMap.get(encryptionContext); + } + + public String getEncryptionContextForTest(String path) { + return pathToContextMap.get(path); + } +} + + diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/AbfsClientUtils.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/AbfsClientUtils.java new file mode 100644 index 0000000000..e7dbf208c9 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/AbfsClientUtils.java @@ -0,0 +1,34 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.services; + +import org.apache.hadoop.fs.azurebfs.extensions.EncryptionContextProvider; + +public final class AbfsClientUtils { + private AbfsClientUtils() { + + } + public static void setIsNamespaceEnabled(final AbfsClient abfsClient, final Boolean isNamespaceEnabled) { + abfsClient.setIsNamespaceEnabled(isNamespaceEnabled); + } + + public static void setEncryptionContextProvider(final AbfsClient abfsClient, final EncryptionContextProvider provider) { + abfsClient.setEncryptionContextProvider(provider); + } +} diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/ITestAbfsClient.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/ITestAbfsClient.java index 6707c593f5..4f87e02000 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/ITestAbfsClient.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/ITestAbfsClient.java @@ -135,7 +135,7 @@ private String getUserAgentString(AbfsConfiguration config, boolean includeSSLProvider) throws IOException { AbfsClientContext abfsClientContext = new AbfsClientContextBuilder().build(); AbfsClient client = new AbfsClient(new URL("https://azure.com"), null, - config, (AccessTokenProvider) null, abfsClientContext); + config, (AccessTokenProvider) null, null, abfsClientContext); String sslProviderName = null; if (includeSSLProvider) { sslProviderName = DelegatingSSLSocketFactory.getDefaultFactory() @@ -341,6 +341,7 @@ public static AbfsClient createTestClientFromCurrentContext( (currentAuthType == AuthType.OAuth ? abfsConfig.getTokenProvider() : null), + null, abfsClientContext); return testClient; @@ -384,6 +385,10 @@ public static AbfsClient getMockAbfsClient(AbfsClient baseAbfsClientInstance, client = ITestAbfsClient.setAbfsClientField(client, "baseUrl", baseAbfsClientInstance.getBaseUrl()); + // override xMsVersion + client = ITestAbfsClient.setAbfsClientField(client, "xMsVersion", + baseAbfsClientInstance.getxMsVersion()); + // override auth provider if (currentAuthType == AuthType.SharedKey) { client = ITestAbfsClient.setAbfsClientField(client, "sharedKeyCredentials", @@ -608,7 +613,7 @@ public void testExpectHundredContinue() throws Exception { .isTrue(); intercept(AzureBlobFileSystemException.class, - () -> testClient.append(finalTestPath, buffer, appendRequestParameters, null, tracingContext)); + () -> testClient.append(finalTestPath, buffer, appendRequestParameters, null, null, tracingContext)); // Verify that the request was not exponentially retried because of user error. Assertions.assertThat(tracingContext.getRetryCount()) diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/TestAbfsInputStream.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/TestAbfsInputStream.java index 0395c4183b..9027e56c9c 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/TestAbfsInputStream.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/TestAbfsInputStream.java @@ -39,9 +39,9 @@ import org.apache.hadoop.fs.azurebfs.AbstractAbfsIntegrationTest; import org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem; import org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore; -import org.apache.hadoop.fs.azurebfs.contracts.exceptions.AzureBlobFileSystemException; import org.apache.hadoop.fs.azurebfs.contracts.exceptions.TimeoutException; import org.apache.hadoop.fs.azurebfs.contracts.services.ReadBufferStatus; +import org.apache.hadoop.fs.azurebfs.security.ContextEncryptionAdapter; import org.apache.hadoop.fs.azurebfs.utils.TestCachedSASToken; import org.apache.hadoop.fs.azurebfs.utils.TracingContext; import org.apache.hadoop.fs.impl.OpenFileParameters; @@ -49,6 +49,7 @@ import static org.mockito.ArgumentMatchers.any; import static org.mockito.ArgumentMatchers.anyBoolean; import static org.mockito.ArgumentMatchers.anyString; +import static org.mockito.ArgumentMatchers.nullable; import static org.mockito.Mockito.doReturn; import static org.mockito.Mockito.doThrow; import static org.mockito.Mockito.mock; @@ -168,14 +169,14 @@ private void queueReadAheads(AbfsInputStream inputStream) { inputStream.getTracingContext()); } - private void verifyReadCallCount(AbfsClient client, int count) throws - AzureBlobFileSystemException, InterruptedException { + private void verifyReadCallCount(AbfsClient client, int count) + throws IOException, InterruptedException { // ReadAhead threads are triggered asynchronously. // Wait a second before verifying the number of total calls. Thread.sleep(1000); verify(client, times(count)).read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), - any(String.class), any(String.class), any(TracingContext.class)); + any(String.class), any(String.class), any(), any(TracingContext.class)); } private void checkEvictedStatus(AbfsInputStream inputStream, int position, boolean expectedToThrowException) @@ -241,7 +242,8 @@ private void checkGetPathStatusCalls(Path testFile, FileStatus fileStatus, .ofNullable(new OpenFileParameters().withStatus(fileStatus)), null, tracingContext); verify(mockClient, times(0).description((String.format( "FileStatus [from %s result] provided, GetFileStatus should not be invoked", - source)))).getPathStatus(anyString(), anyBoolean(), any(TracingContext.class)); + source)))).getPathStatus(anyString(), anyBoolean(), any(TracingContext.class), any( + ContextEncryptionAdapter.class)); // verify GetPathStatus invoked when FileStatus not provided abfsStore.openFileForRead(testFile, @@ -249,7 +251,8 @@ private void checkGetPathStatusCalls(Path testFile, FileStatus fileStatus, tracingContext); verify(mockClient, times(1).description( "GetPathStatus should be invoked when FileStatus not provided")) - .getPathStatus(anyString(), anyBoolean(), any(TracingContext.class)); + .getPathStatus(anyString(), anyBoolean(), any(TracingContext.class), nullable( + ContextEncryptionAdapter.class)); Mockito.reset(mockClient); //clears invocation count for next test case } @@ -330,7 +333,7 @@ public void testFailedReadAhead() throws Exception { .when(client) .read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), any(String.class), - any(String.class), any(TracingContext.class)); + any(String.class), any(), any(TracingContext.class)); AbfsInputStream inputStream = getAbfsInputStream(client, "testFailedReadAhead.txt"); @@ -364,7 +367,7 @@ public void testFailedReadAheadEviction() throws Exception { .when(client) .read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), any(String.class), - any(String.class), any(TracingContext.class)); + any(String.class), any(), any(TracingContext.class)); AbfsInputStream inputStream = getAbfsInputStream(client, "testFailedReadAheadEviction.txt"); @@ -409,7 +412,7 @@ public void testOlderReadAheadFailure() throws Exception { .when(client) .read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), any(String.class), - any(String.class), any(TracingContext.class)); + any(String.class), any(), any(TracingContext.class)); AbfsInputStream inputStream = getAbfsInputStream(client, "testOlderReadAheadFailure.txt"); @@ -463,7 +466,7 @@ public void testSuccessfulReadAhead() throws Exception { .when(client) .read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), any(String.class), - any(String.class), any(TracingContext.class)); + any(String.class), any(), any(TracingContext.class)); AbfsInputStream inputStream = getAbfsInputStream(client, "testSuccessfulReadAhead.txt"); int beforeReadCompletedListSize = ReadBufferManager.getBufferManager().getCompletedReadListSize(); @@ -518,7 +521,8 @@ public void testStreamPurgeDuringReadAheadCallExecuting() throws Exception { .when(client) .read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), any(String.class), - any(String.class), any(TracingContext.class)); + any(String.class), nullable(ContextEncryptionAdapter.class), + any(TracingContext.class)); final ReadBufferManager readBufferManager = ReadBufferManager.getBufferManager(); @@ -584,7 +588,7 @@ public void testReadAheadManagerForFailedReadAhead() throws Exception { .when(client) .read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), any(String.class), - any(String.class), any(TracingContext.class)); + any(String.class), any(), any(TracingContext.class)); AbfsInputStream inputStream = getAbfsInputStream(client, "testReadAheadManagerForFailedReadAhead.txt"); @@ -637,7 +641,7 @@ public void testReadAheadManagerForOlderReadAheadFailure() throws Exception { .when(client) .read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), any(String.class), - any(String.class), any(TracingContext.class)); + any(String.class), any(), any(TracingContext.class)); AbfsInputStream inputStream = getAbfsInputStream(client, "testReadAheadManagerForOlderReadAheadFailure.txt"); @@ -691,7 +695,7 @@ public void testReadAheadManagerForSuccessfulReadAhead() throws Exception { .when(client) .read(any(String.class), any(Long.class), any(byte[].class), any(Integer.class), any(Integer.class), any(String.class), - any(String.class), any(TracingContext.class)); + any(String.class), any(), any(TracingContext.class)); AbfsInputStream inputStream = getAbfsInputStream(client, "testSuccessfulReadAhead.txt"); diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/TestAbfsOutputStream.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/TestAbfsOutputStream.java index e26ba938cf..f0987b5fd7 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/TestAbfsOutputStream.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/services/TestAbfsOutputStream.java @@ -120,10 +120,11 @@ public void verifyShortWriteRequest() throws Exception { AbfsPerfTracker tracker = new AbfsPerfTracker("test", accountName1, abfsConf); when(client.getAbfsPerfTracker()).thenReturn(tracker); when(client.append(anyString(), any(byte[].class), - any(AppendRequestParameters.class), any(), any(TracingContext.class))) + any(AppendRequestParameters.class), any(), + any(), any(TracingContext.class))) .thenReturn(op); when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), any(), - isNull(), any(TracingContext.class))).thenReturn(op); + isNull(), any(), any(TracingContext.class))).thenReturn(op); AbfsOutputStream out = new AbfsOutputStream( populateAbfsOutputStreamContext( @@ -157,13 +158,13 @@ public void verifyShortWriteRequest() throws Exception { WRITE_SIZE, 0, 2 * WRITE_SIZE, APPEND_MODE, false, null, true); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), + eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(), any(TracingContext.class)); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(), any(TracingContext.class)); // confirm there were only 2 invocations in all verify(client, times(2)).append( - eq(PATH), any(byte[].class), any(), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), any(), any(), any(), any(TracingContext.class)); } /** @@ -184,8 +185,8 @@ public void verifyWriteRequest() throws Exception { TracingHeaderFormat.ALL_ID_FORMAT, null); when(client.getAbfsPerfTracker()).thenReturn(tracker); - when(client.append(anyString(), any(byte[].class), any(AppendRequestParameters.class), any(), any(TracingContext.class))).thenReturn(op); - when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), any(), isNull(), any(TracingContext.class))).thenReturn(op); + when(client.append(anyString(), any(byte[].class), any(AppendRequestParameters.class), any(), any(), any(TracingContext.class))).thenReturn(op); + when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), any(), isNull(), any(), any(TracingContext.class))).thenReturn(op); AbfsOutputStream out = new AbfsOutputStream( populateAbfsOutputStreamContext( @@ -211,16 +212,13 @@ public void verifyWriteRequest() throws Exception { AppendRequestParameters secondReqParameters = new AppendRequestParameters( BUFFER_SIZE, 0, 5*WRITE_SIZE-BUFFER_SIZE, APPEND_MODE, false, null, true); - verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), - any(TracingContext.class)); - verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), - any(TracingContext.class)); + verify(client, times(1)).append(eq(PATH), any(byte[].class), + refEq(firstReqParameters), any(), any(), any(TracingContext.class)); + verify(client, times(1)).append(eq(PATH), any(byte[].class), + refEq(secondReqParameters), any(), any(), any(TracingContext.class)); // confirm there were only 2 invocations in all - verify(client, times(2)).append( - eq(PATH), any(byte[].class), any(), any(), - any(TracingContext.class)); + verify(client, times(2)).append(eq(PATH), any(byte[].class), any(), any(), + any(), any(TracingContext.class)); ArgumentCaptor acFlushPath = ArgumentCaptor.forClass(String.class); ArgumentCaptor acFlushPosition = ArgumentCaptor.forClass(Long.class); @@ -231,7 +229,8 @@ public void verifyWriteRequest() throws Exception { ArgumentCaptor acFlushSASToken = ArgumentCaptor.forClass(String.class); verify(client, times(1)).flush(acFlushPath.capture(), acFlushPosition.capture(), acFlushRetainUnCommittedData.capture(), acFlushClose.capture(), - acFlushSASToken.capture(), isNull(), acTracingContext.capture()); + acFlushSASToken.capture(), isNull(), isNull(), + acTracingContext.capture()); assertThat(Arrays.asList(PATH)).describedAs("path").isEqualTo(acFlushPath.getAllValues()); assertThat(Arrays.asList(Long.valueOf(5*WRITE_SIZE))).describedAs("position").isEqualTo(acFlushPosition.getAllValues()); assertThat(Arrays.asList(false)).describedAs("RetainUnCommittedData flag").isEqualTo(acFlushRetainUnCommittedData.getAllValues()); @@ -257,8 +256,8 @@ public void verifyWriteRequestOfBufferSizeAndClose() throws Exception { FSOperationType.WRITE, abfsConf.getTracingHeaderFormat(), null); when(client.getAbfsPerfTracker()).thenReturn(tracker); - when(client.append(anyString(), any(byte[].class), any(AppendRequestParameters.class), any(), any(TracingContext.class))).thenReturn(op); - when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), any(), isNull(), any(TracingContext.class))).thenReturn(op); + when(client.append(anyString(), any(byte[].class), any(AppendRequestParameters.class), any(), any(), any(TracingContext.class))).thenReturn(op); + when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), any(), isNull(), any(), any(TracingContext.class))).thenReturn(op); when(op.getSasToken()).thenReturn("testToken"); when(op.getResult()).thenReturn(httpOp); @@ -287,12 +286,12 @@ public void verifyWriteRequestOfBufferSizeAndClose() throws Exception { BUFFER_SIZE, 0, BUFFER_SIZE, APPEND_MODE, false, null, true); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(), any(TracingContext.class)); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(), any(TracingContext.class)); // confirm there were only 2 invocations in all verify(client, times(2)).append( - eq(PATH), any(byte[].class), any(), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), any(), any(), any(), any(TracingContext.class)); ArgumentCaptor acFlushPath = ArgumentCaptor.forClass(String.class); ArgumentCaptor acFlushPosition = ArgumentCaptor.forClass(Long.class); @@ -303,7 +302,8 @@ public void verifyWriteRequestOfBufferSizeAndClose() throws Exception { ArgumentCaptor acFlushSASToken = ArgumentCaptor.forClass(String.class); verify(client, times(1)).flush(acFlushPath.capture(), acFlushPosition.capture(), acFlushRetainUnCommittedData.capture(), acFlushClose.capture(), - acFlushSASToken.capture(), isNull(), acTracingContext.capture()); + acFlushSASToken.capture(), isNull(), isNull(), + acTracingContext.capture()); assertThat(Arrays.asList(PATH)).describedAs("path").isEqualTo(acFlushPath.getAllValues()); assertThat(Arrays.asList(Long.valueOf(2*BUFFER_SIZE))).describedAs("position").isEqualTo(acFlushPosition.getAllValues()); assertThat(Arrays.asList(false)).describedAs("RetainUnCommittedData flag").isEqualTo(acFlushRetainUnCommittedData.getAllValues()); @@ -327,10 +327,10 @@ public void verifyWriteRequestOfBufferSize() throws Exception { when(client.getAbfsPerfTracker()).thenReturn(tracker); when(client.append(anyString(), any(byte[].class), - any(AppendRequestParameters.class), any(), any(TracingContext.class))) + any(AppendRequestParameters.class), any(), any(), any(TracingContext.class))) .thenReturn(op); when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), - any(), isNull(), any(TracingContext.class))).thenReturn(op); + any(), isNull(), any(), any(TracingContext.class))).thenReturn(op); when(op.getSasToken()).thenReturn("testToken"); when(op.getResult()).thenReturn(httpOp); @@ -361,12 +361,12 @@ public void verifyWriteRequestOfBufferSize() throws Exception { BUFFER_SIZE, 0, BUFFER_SIZE, APPEND_MODE, false, null, true); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(), any(TracingContext.class)); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(), any(TracingContext.class)); // confirm there were only 2 invocations in all verify(client, times(2)).append( - eq(PATH), any(byte[].class), any(), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), any(), any(), any(), any(TracingContext.class)); } /** @@ -385,10 +385,10 @@ public void verifyWriteRequestOfBufferSizeWithAppendBlob() throws Exception { when(client.getAbfsPerfTracker()).thenReturn(tracker); when(client.append(anyString(), any(byte[].class), - any(AppendRequestParameters.class), any(), any(TracingContext.class))) + any(AppendRequestParameters.class), any(), any(), any(TracingContext.class))) .thenReturn(op); when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), any(), - isNull(), any(TracingContext.class))).thenReturn(op); + isNull(), any(), any(TracingContext.class))).thenReturn(op); AbfsOutputStream out = new AbfsOutputStream( populateAbfsOutputStreamContext( @@ -417,12 +417,12 @@ public void verifyWriteRequestOfBufferSizeWithAppendBlob() throws Exception { BUFFER_SIZE, 0, BUFFER_SIZE, APPEND_MODE, true, null, true); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(), any(TracingContext.class)); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(), any(TracingContext.class)); // confirm there were only 2 invocations in all verify(client, times(2)).append( - eq(PATH), any(byte[].class), any(), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), any(), any(), any(), any(TracingContext.class)); } /** @@ -445,10 +445,10 @@ public void verifyWriteRequestOfBufferSizeAndHFlush() throws Exception { when(client.getAbfsPerfTracker()).thenReturn(tracker); when(client.append(anyString(), any(byte[].class), - any(AppendRequestParameters.class), any(), any(TracingContext.class))) + any(AppendRequestParameters.class), any(), any(), any(TracingContext.class))) .thenReturn(op); when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), any(), - isNull(), any(TracingContext.class))).thenReturn(op); + isNull(), any(), any(TracingContext.class))).thenReturn(op); AbfsOutputStream out = new AbfsOutputStream( populateAbfsOutputStreamContext( @@ -477,12 +477,12 @@ public void verifyWriteRequestOfBufferSizeAndHFlush() throws Exception { BUFFER_SIZE, 0, BUFFER_SIZE, APPEND_MODE, false, null, true); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(), any(TracingContext.class)); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(), any(TracingContext.class)); // confirm there were only 2 invocations in all verify(client, times(2)).append( - eq(PATH), any(byte[].class), any(), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), any(), any(), any(), any(TracingContext.class)); ArgumentCaptor acFlushPath = ArgumentCaptor.forClass(String.class); ArgumentCaptor acFlushPosition = ArgumentCaptor.forClass(Long.class); @@ -493,7 +493,7 @@ public void verifyWriteRequestOfBufferSizeAndHFlush() throws Exception { ArgumentCaptor acFlushSASToken = ArgumentCaptor.forClass(String.class); verify(client, times(1)).flush(acFlushPath.capture(), acFlushPosition.capture(), acFlushRetainUnCommittedData.capture(), acFlushClose.capture(), - acFlushSASToken.capture(), isNull(), acTracingContext.capture()); + acFlushSASToken.capture(), isNull(), isNull(), acTracingContext.capture()); assertThat(Arrays.asList(PATH)).describedAs("path").isEqualTo(acFlushPath.getAllValues()); assertThat(Arrays.asList(Long.valueOf(2*BUFFER_SIZE))).describedAs("position").isEqualTo(acFlushPosition.getAllValues()); assertThat(Arrays.asList(false)).describedAs("RetainUnCommittedData flag").isEqualTo(acFlushRetainUnCommittedData.getAllValues()); @@ -515,10 +515,10 @@ public void verifyWriteRequestOfBufferSizeAndFlush() throws Exception { AbfsPerfTracker tracker = new AbfsPerfTracker("test", accountName1, abfsConf); when(client.getAbfsPerfTracker()).thenReturn(tracker); when(client.append(anyString(), any(byte[].class), - any(AppendRequestParameters.class), any(), any(TracingContext.class))) + any(AppendRequestParameters.class), any(), any(), any(TracingContext.class))) .thenReturn(op); when(client.flush(anyString(), anyLong(), anyBoolean(), anyBoolean(), any(), - isNull(), any(TracingContext.class))).thenReturn(op); + isNull(), any(), any(TracingContext.class))).thenReturn(op); AbfsOutputStream out = new AbfsOutputStream( populateAbfsOutputStreamContext( @@ -549,12 +549,12 @@ public void verifyWriteRequestOfBufferSizeAndFlush() throws Exception { BUFFER_SIZE, 0, BUFFER_SIZE, APPEND_MODE, false, null, true); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(firstReqParameters), any(), any(), any(TracingContext.class)); verify(client, times(1)).append( - eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), refEq(secondReqParameters), any(), any(), any(TracingContext.class)); // confirm there were only 2 invocations in all verify(client, times(2)).append( - eq(PATH), any(byte[].class), any(), any(), any(TracingContext.class)); + eq(PATH), any(byte[].class), any(), any(), any(), any(TracingContext.class)); } /**