HADOOP-8686. Fix warnings in native code. Contributed by Colin Patrick McCabe
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1375301 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
d5f080825c
commit
357e990c40
@ -302,6 +302,8 @@ Branch-2 ( Unreleased changes )
|
||||
HADOOP-8700. Use enum to define the checksum constants in DataChecksum.
|
||||
(szetszwo)
|
||||
|
||||
HADOOP-8686. Fix warnings in native code. (Colin Patrick McCabe via eli)
|
||||
|
||||
BUG FIXES
|
||||
|
||||
HADOOP-8372. NetUtils.normalizeHostName() incorrectly handles hostname
|
||||
|
@ -24,7 +24,7 @@
|
||||
// Simple Functions
|
||||
//****************************
|
||||
|
||||
extern int LZ4_compress (char* source, char* dest, int isize);
|
||||
extern int LZ4_compress (const char* source, char* dest, int isize);
|
||||
|
||||
/*
|
||||
LZ4_compress() :
|
||||
|
@ -20,7 +20,7 @@
|
||||
#include "org_apache_hadoop.h"
|
||||
#include "org_apache_hadoop_io_compress_lz4_Lz4Decompressor.h"
|
||||
|
||||
int LZ4_uncompress_unknownOutputSize (char* source, char* dest, int isize, int maxOutputSize);
|
||||
int LZ4_uncompress_unknownOutputSize(const char* source, char* dest, int isize, int maxOutputSize);
|
||||
|
||||
/*
|
||||
LZ4_uncompress_unknownOutputSize() :
|
||||
|
@ -25,6 +25,8 @@
|
||||
#include "org_apache_hadoop_io_compress_snappy.h"
|
||||
#include "org_apache_hadoop_io_compress_snappy_SnappyCompressor.h"
|
||||
|
||||
#define JINT_MAX 0x7fffffff
|
||||
|
||||
static jfieldID SnappyCompressor_clazz;
|
||||
static jfieldID SnappyCompressor_uncompressedDirectBuf;
|
||||
static jfieldID SnappyCompressor_uncompressedDirectBufLen;
|
||||
@ -39,7 +41,7 @@ JNIEXPORT void JNICALL Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
|
||||
// Load libsnappy.so
|
||||
void *libsnappy = dlopen(HADOOP_SNAPPY_LIBRARY, RTLD_LAZY | RTLD_GLOBAL);
|
||||
if (!libsnappy) {
|
||||
char* msg = (char*)malloc(1000);
|
||||
char msg[1000];
|
||||
snprintf(msg, 1000, "%s (%s)!", "Cannot load " HADOOP_SNAPPY_LIBRARY, dlerror());
|
||||
THROW(env, "java/lang/UnsatisfiedLinkError", msg);
|
||||
return;
|
||||
@ -71,6 +73,7 @@ JNIEXPORT jint JNICALL Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
|
||||
jint uncompressed_direct_buf_len = (*env)->GetIntField(env, thisj, SnappyCompressor_uncompressedDirectBufLen);
|
||||
jobject compressed_direct_buf = (*env)->GetObjectField(env, thisj, SnappyCompressor_compressedDirectBuf);
|
||||
jint compressed_direct_buf_len = (*env)->GetIntField(env, thisj, SnappyCompressor_directBufferSize);
|
||||
size_t buf_len;
|
||||
|
||||
// Get the input direct buffer
|
||||
LOCK_CLASS(env, clazz, "SnappyCompressor");
|
||||
@ -78,7 +81,7 @@ JNIEXPORT jint JNICALL Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
|
||||
UNLOCK_CLASS(env, clazz, "SnappyCompressor");
|
||||
|
||||
if (uncompressed_bytes == 0) {
|
||||
return (jint)0;
|
||||
return 0;
|
||||
}
|
||||
|
||||
// Get the output direct buffer
|
||||
@ -87,15 +90,22 @@ JNIEXPORT jint JNICALL Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
|
||||
UNLOCK_CLASS(env, clazz, "SnappyCompressor");
|
||||
|
||||
if (compressed_bytes == 0) {
|
||||
return (jint)0;
|
||||
return 0;
|
||||
}
|
||||
|
||||
snappy_status ret = dlsym_snappy_compress(uncompressed_bytes, uncompressed_direct_buf_len, compressed_bytes, &compressed_direct_buf_len);
|
||||
/* size_t should always be 4 bytes or larger. */
|
||||
buf_len = (size_t)compressed_direct_buf_len;
|
||||
snappy_status ret = dlsym_snappy_compress(uncompressed_bytes,
|
||||
uncompressed_direct_buf_len, compressed_bytes, &buf_len);
|
||||
if (ret != SNAPPY_OK){
|
||||
THROW(env, "Ljava/lang/InternalError", "Could not compress data. Buffer length is too small.");
|
||||
return 0;
|
||||
}
|
||||
if (buf_len > JINT_MAX) {
|
||||
THROW(env, "Ljava/lang/InternalError", "Invalid return buffer length.");
|
||||
return 0;
|
||||
}
|
||||
|
||||
(*env)->SetIntField(env, thisj, SnappyCompressor_uncompressedDirectBufLen, 0);
|
||||
|
||||
return (jint)compressed_direct_buf_len;
|
||||
return (jint)buf_len;
|
||||
}
|
||||
|
@ -16,6 +16,8 @@
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
#define _GNU_SOURCE
|
||||
|
||||
#include <assert.h>
|
||||
#include <errno.h>
|
||||
#include <fcntl.h>
|
||||
@ -366,23 +368,15 @@ Java_org_apache_hadoop_io_nativeio_NativeIO_chmod(
|
||||
*/
|
||||
static void throw_ioe(JNIEnv* env, int errnum)
|
||||
{
|
||||
const char* message;
|
||||
char buffer[80];
|
||||
char message[80];
|
||||
jstring jstr_message;
|
||||
|
||||
buffer[0] = 0;
|
||||
#ifdef STRERROR_R_CHAR_P
|
||||
// GNU strerror_r
|
||||
message = strerror_r(errnum, buffer, sizeof(buffer));
|
||||
assert (message != NULL);
|
||||
#else
|
||||
int ret = strerror_r(errnum, buffer, sizeof(buffer));
|
||||
if (ret == 0) {
|
||||
message = buffer;
|
||||
if ((errnum >= 0) && (errnum < sys_nerr)) {
|
||||
snprintf(message, sizeof(message), "%s", sys_errlist[errnum]);
|
||||
} else {
|
||||
message = "Unknown error";
|
||||
snprintf(message, sizeof(message), "Unknown error %d", errnum);
|
||||
}
|
||||
#endif
|
||||
|
||||
jobject errno_obj = errno_to_enum(env, errnum);
|
||||
|
||||
if ((jstr_message = (*env)->NewStringUTF(env, message)) == NULL)
|
||||
|
@ -40,8 +40,8 @@ Java_org_apache_hadoop_security_JniBasedUnixGroupsMapping_getGroupForUser
|
||||
(JNIEnv *env, jobject jobj, jstring juser) {
|
||||
extern int getGroupIDList(const char *user, int *ngroups, gid_t **groups);
|
||||
extern int getGroupDetails(gid_t group, char **grpBuf);
|
||||
|
||||
jobjectArray jgroups;
|
||||
const char *cuser = NULL;
|
||||
jobjectArray jgroups = NULL;
|
||||
int error = -1;
|
||||
|
||||
if (emptyGroups == NULL) {
|
||||
@ -56,7 +56,7 @@ Java_org_apache_hadoop_security_JniBasedUnixGroupsMapping_getGroupForUser
|
||||
}
|
||||
}
|
||||
char *grpBuf = NULL;
|
||||
const char *cuser = (*env)->GetStringUTFChars(env, juser, NULL);
|
||||
cuser = (*env)->GetStringUTFChars(env, juser, NULL);
|
||||
if (cuser == NULL) {
|
||||
goto cleanup;
|
||||
}
|
||||
|
@ -45,6 +45,8 @@ typedef struct listElement UserList;
|
||||
JNIEXPORT jobjectArray JNICALL
|
||||
Java_org_apache_hadoop_security_JniBasedUnixGroupsNetgroupMapping_getUsersForNetgroupJNI
|
||||
(JNIEnv *env, jobject jobj, jstring jgroup) {
|
||||
UserList *userListHead = NULL;
|
||||
int userListSize = 0;
|
||||
|
||||
// pointers to free at the end
|
||||
const char *cgroup = NULL;
|
||||
@ -65,9 +67,6 @@ Java_org_apache_hadoop_security_JniBasedUnixGroupsNetgroupMapping_getUsersForNet
|
||||
// get users
|
||||
// see man pages for setnetgrent, getnetgrent and endnetgrent
|
||||
|
||||
UserList *userListHead = NULL;
|
||||
int userListSize = 0;
|
||||
|
||||
// set the name of the group for subsequent calls to getnetgrent
|
||||
// note that we want to end group lokup regardless whether setnetgrent
|
||||
// was successful or not (as long as it was called we need to call
|
||||
|
@ -18,6 +18,7 @@
|
||||
|
||||
#include <arpa/inet.h>
|
||||
#include <assert.h>
|
||||
#include <inttypes.h>
|
||||
#include <stdlib.h>
|
||||
#include <stdint.h>
|
||||
#include <string.h>
|
||||
@ -50,7 +51,7 @@ static void throw_checksum_exception(JNIEnv *env,
|
||||
|
||||
// Format error message
|
||||
snprintf(message, sizeof(message),
|
||||
"Checksum error: %s at %ld exp: %d got: %d",
|
||||
"Checksum error: %s at %"PRId64" exp: %"PRId32" got: %"PRId32,
|
||||
filename, pos, expected_crc, got_crc);
|
||||
if ((jstr_message = (*env)->NewStringUTF(env, message)) == NULL) {
|
||||
goto cleanup;
|
||||
|
@ -41,7 +41,7 @@ static uint32_t crc32c_sb8(uint32_t crc, const uint8_t *buf, size_t length);
|
||||
|
||||
#ifdef USE_PIPELINED
|
||||
static void pipelined_crc32c(uint32_t *crc1, uint32_t *crc2, uint32_t *crc3, const uint8_t *p_buf, size_t block_size, int num_blocks);
|
||||
#endif USE_PIPELINED
|
||||
#endif
|
||||
static int cached_cpu_supports_crc32; // initialized by constructor below
|
||||
static uint32_t crc32c_hardware(uint32_t crc, const uint8_t* data, size_t length);
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user