HADOOP-7379. Add the ability to serialize and deserialize protocol buffers in ObjectWritable. Contributed by Todd Lipcon.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1136222 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
de5bf9d5ca
commit
7728f03f08
@ -44,6 +44,9 @@ Trunk (unreleased changes)
|
|||||||
HADOOP-7144. Expose JMX metrics via JSON servlet. (Robert Joseph Evans via
|
HADOOP-7144. Expose JMX metrics via JSON servlet. (Robert Joseph Evans via
|
||||||
cdouglas)
|
cdouglas)
|
||||||
|
|
||||||
|
HADOOP-7379. Add the ability to serialize and deserialize protocol buffers
|
||||||
|
in ObjectWritable. (todd)
|
||||||
|
|
||||||
IMPROVEMENTS
|
IMPROVEMENTS
|
||||||
|
|
||||||
HADOOP-7042. Updates to test-patch.sh to include failed test names and
|
HADOOP-7042. Updates to test-patch.sh to include failed test names and
|
||||||
|
@ -0,0 +1,71 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.io;
|
||||||
|
|
||||||
|
import java.io.DataOutput;
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.io.OutputStream;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* OutputStream implementation that wraps a DataOutput.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Public
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public class DataOutputOutputStream extends OutputStream {
|
||||||
|
|
||||||
|
private final DataOutput out;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Construct an OutputStream from the given DataOutput. If 'out'
|
||||||
|
* is already an OutputStream, simply returns it. Otherwise, wraps
|
||||||
|
* it in an OutputStream.
|
||||||
|
* @param out the DataOutput to wrap
|
||||||
|
* @return an OutputStream instance that outputs to 'out'
|
||||||
|
*/
|
||||||
|
public static OutputStream constructOutputStream(DataOutput out) {
|
||||||
|
if (out instanceof OutputStream) {
|
||||||
|
return (OutputStream)out;
|
||||||
|
} else {
|
||||||
|
return new DataOutputOutputStream(out);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private DataOutputOutputStream(DataOutput out) {
|
||||||
|
this.out = out;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void write(int b) throws IOException {
|
||||||
|
out.writeByte(b);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void write(byte[] b, int off, int len) throws IOException {
|
||||||
|
out.write(b, off, len);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void write(byte[] b) throws IOException {
|
||||||
|
out.write(b);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
}
|
@ -19,6 +19,8 @@
|
|||||||
package org.apache.hadoop.io;
|
package org.apache.hadoop.io;
|
||||||
|
|
||||||
import java.lang.reflect.Array;
|
import java.lang.reflect.Array;
|
||||||
|
import java.lang.reflect.InvocationTargetException;
|
||||||
|
import java.lang.reflect.Method;
|
||||||
|
|
||||||
import java.io.*;
|
import java.io.*;
|
||||||
import java.util.*;
|
import java.util.*;
|
||||||
@ -26,6 +28,9 @@
|
|||||||
import org.apache.hadoop.classification.InterfaceAudience;
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.classification.InterfaceStability;
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.conf.*;
|
import org.apache.hadoop.conf.*;
|
||||||
|
import org.apache.hadoop.util.ProtoUtil;
|
||||||
|
|
||||||
|
import com.google.protobuf.Message;
|
||||||
|
|
||||||
/** A polymorphic Writable that writes an instance with it's class name.
|
/** A polymorphic Writable that writes an instance with it's class name.
|
||||||
* Handles arrays, strings and primitive types without a Writable wrapper.
|
* Handles arrays, strings and primitive types without a Writable wrapper.
|
||||||
@ -191,6 +196,9 @@ public static void writeObject(DataOutput out, Object instance,
|
|||||||
UTF8.writeString(out, instance.getClass().getName());
|
UTF8.writeString(out, instance.getClass().getName());
|
||||||
((Writable)instance).write(out);
|
((Writable)instance).write(out);
|
||||||
|
|
||||||
|
} else if (Message.class.isAssignableFrom(declaredClass)) {
|
||||||
|
((Message)instance).writeDelimitedTo(
|
||||||
|
DataOutputOutputStream.constructOutputStream(out));
|
||||||
} else {
|
} else {
|
||||||
throw new IOException("Can't write: "+instance+" as "+declaredClass);
|
throw new IOException("Can't write: "+instance+" as "+declaredClass);
|
||||||
}
|
}
|
||||||
@ -261,6 +269,8 @@ public static Object readObject(DataInput in, ObjectWritable objectWritable, Con
|
|||||||
instance = UTF8.readString(in);
|
instance = UTF8.readString(in);
|
||||||
} else if (declaredClass.isEnum()) { // enum
|
} else if (declaredClass.isEnum()) { // enum
|
||||||
instance = Enum.valueOf((Class<? extends Enum>) declaredClass, UTF8.readString(in));
|
instance = Enum.valueOf((Class<? extends Enum>) declaredClass, UTF8.readString(in));
|
||||||
|
} else if (Message.class.isAssignableFrom(declaredClass)) {
|
||||||
|
instance = tryInstantiateProtobuf(declaredClass, in);
|
||||||
} else { // Writable
|
} else { // Writable
|
||||||
Class instanceClass = null;
|
Class instanceClass = null;
|
||||||
String str = UTF8.readString(in);
|
String str = UTF8.readString(in);
|
||||||
@ -285,6 +295,67 @@ public static Object readObject(DataInput in, ObjectWritable objectWritable, Con
|
|||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Try to instantiate a protocol buffer of the given message class
|
||||||
|
* from the given input stream.
|
||||||
|
*
|
||||||
|
* @param protoClass the class of the generated protocol buffer
|
||||||
|
* @param dataIn the input stream to read from
|
||||||
|
* @return the instantiated Message instance
|
||||||
|
* @throws IOException if an IO problem occurs
|
||||||
|
*/
|
||||||
|
private static Message tryInstantiateProtobuf(
|
||||||
|
Class<?> protoClass,
|
||||||
|
DataInput dataIn) throws IOException {
|
||||||
|
|
||||||
|
try {
|
||||||
|
if (dataIn instanceof InputStream) {
|
||||||
|
// We can use the built-in parseDelimitedFrom and not have to re-copy
|
||||||
|
// the data
|
||||||
|
Method parseMethod = getStaticProtobufMethod(protoClass,
|
||||||
|
"parseDelimitedFrom", InputStream.class);
|
||||||
|
return (Message)parseMethod.invoke(null, (InputStream)dataIn);
|
||||||
|
} else {
|
||||||
|
// Have to read it into a buffer first, since protobuf doesn't deal
|
||||||
|
// with the DataInput interface directly.
|
||||||
|
|
||||||
|
// Read the size delimiter that writeDelimitedTo writes
|
||||||
|
int size = ProtoUtil.readRawVarint32(dataIn);
|
||||||
|
if (size < 0) {
|
||||||
|
throw new IOException("Invalid size: " + size);
|
||||||
|
}
|
||||||
|
|
||||||
|
byte[] data = new byte[size];
|
||||||
|
dataIn.readFully(data);
|
||||||
|
Method parseMethod = getStaticProtobufMethod(protoClass,
|
||||||
|
"parseFrom", byte[].class);
|
||||||
|
return (Message)parseMethod.invoke(null, data);
|
||||||
|
}
|
||||||
|
} catch (InvocationTargetException e) {
|
||||||
|
|
||||||
|
if (e.getCause() instanceof IOException) {
|
||||||
|
throw (IOException)e.getCause();
|
||||||
|
} else {
|
||||||
|
throw new IOException(e.getCause());
|
||||||
|
}
|
||||||
|
} catch (IllegalAccessException iae) {
|
||||||
|
throw new AssertionError("Could not access parse method in " +
|
||||||
|
protoClass);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static Method getStaticProtobufMethod(Class<?> declaredClass, String method,
|
||||||
|
Class<?> ... args) {
|
||||||
|
|
||||||
|
try {
|
||||||
|
return declaredClass.getMethod(method, args);
|
||||||
|
} catch (Exception e) {
|
||||||
|
// This is a bug in Hadoop - protobufs should all have this static method
|
||||||
|
throw new AssertionError("Protocol buffer class " + declaredClass +
|
||||||
|
" does not have an accessible parseFrom(InputStream) method!");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Find and load the class with given name <tt>className</tt> by first finding
|
* Find and load the class with given name <tt>className</tt> by first finding
|
||||||
* it in the specified <tt>conf</tt>. If the specified <tt>conf</tt> is null,
|
* it in the specified <tt>conf</tt>. If the specified <tt>conf</tt> is null,
|
||||||
|
66
common/src/java/org/apache/hadoop/util/ProtoUtil.java
Normal file
66
common/src/java/org/apache/hadoop/util/ProtoUtil.java
Normal file
@ -0,0 +1,66 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.util;
|
||||||
|
|
||||||
|
import java.io.DataInput;
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
public abstract class ProtoUtil {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Read a variable length integer in the same format that ProtoBufs encodes.
|
||||||
|
* @param in the input stream to read from
|
||||||
|
* @return the integer
|
||||||
|
* @throws IOException if it is malformed or EOF.
|
||||||
|
*/
|
||||||
|
public static int readRawVarint32(DataInput in) throws IOException {
|
||||||
|
byte tmp = in.readByte();
|
||||||
|
if (tmp >= 0) {
|
||||||
|
return tmp;
|
||||||
|
}
|
||||||
|
int result = tmp & 0x7f;
|
||||||
|
if ((tmp = in.readByte()) >= 0) {
|
||||||
|
result |= tmp << 7;
|
||||||
|
} else {
|
||||||
|
result |= (tmp & 0x7f) << 7;
|
||||||
|
if ((tmp = in.readByte()) >= 0) {
|
||||||
|
result |= tmp << 14;
|
||||||
|
} else {
|
||||||
|
result |= (tmp & 0x7f) << 14;
|
||||||
|
if ((tmp = in.readByte()) >= 0) {
|
||||||
|
result |= tmp << 21;
|
||||||
|
} else {
|
||||||
|
result |= (tmp & 0x7f) << 21;
|
||||||
|
result |= (tmp = in.readByte()) << 28;
|
||||||
|
if (tmp < 0) {
|
||||||
|
// Discard upper 32 bits.
|
||||||
|
for (int i = 0; i < 5; i++) {
|
||||||
|
if (in.readByte() >= 0) {
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
throw new IOException("Malformed varint");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
@ -0,0 +1,81 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.io;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import com.google.protobuf.DescriptorProtos;
|
||||||
|
import com.google.protobuf.Message;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test case for the use of Protocol Buffers within ObjectWritable.
|
||||||
|
*/
|
||||||
|
public class TestObjectWritableProtos {
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testProtoBufs() throws IOException {
|
||||||
|
doTest(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testProtoBufs2() throws IOException {
|
||||||
|
doTest(2);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testProtoBufs3() throws IOException {
|
||||||
|
doTest(3);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Write a protobuf to a buffer 'numProtos' times, and then
|
||||||
|
* read them back, making sure all data comes through correctly.
|
||||||
|
*/
|
||||||
|
private void doTest(int numProtos) throws IOException {
|
||||||
|
Configuration conf = new Configuration();
|
||||||
|
DataOutputBuffer out = new DataOutputBuffer();
|
||||||
|
|
||||||
|
// Write numProtos protobufs to the buffer
|
||||||
|
Message[] sent = new Message[numProtos];
|
||||||
|
for (int i = 0; i < numProtos; i++) {
|
||||||
|
// Construct a test protocol buffer using one of the
|
||||||
|
// protos that ships with the protobuf library
|
||||||
|
Message testProto = DescriptorProtos.EnumValueDescriptorProto.newBuilder()
|
||||||
|
.setName("test" + i).setNumber(i).build();
|
||||||
|
ObjectWritable.writeObject(out, testProto,
|
||||||
|
DescriptorProtos.EnumValueDescriptorProto.class, conf);
|
||||||
|
sent[i] = testProto;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read back the data
|
||||||
|
DataInputBuffer in = new DataInputBuffer();
|
||||||
|
in.reset(out.getData(), out.getLength());
|
||||||
|
|
||||||
|
for (int i = 0; i < numProtos; i++) {
|
||||||
|
Message received = (Message)ObjectWritable.readObject(in, conf);
|
||||||
|
|
||||||
|
assertEquals(sent[i], received);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
@ -40,6 +40,10 @@
|
|||||||
import org.apache.hadoop.security.authorize.PolicyProvider;
|
import org.apache.hadoop.security.authorize.PolicyProvider;
|
||||||
import org.apache.hadoop.security.authorize.Service;
|
import org.apache.hadoop.security.authorize.Service;
|
||||||
import org.apache.hadoop.security.AccessControlException;
|
import org.apache.hadoop.security.AccessControlException;
|
||||||
|
|
||||||
|
import com.google.protobuf.DescriptorProtos;
|
||||||
|
import com.google.protobuf.DescriptorProtos.EnumDescriptorProto;
|
||||||
|
|
||||||
import static org.apache.hadoop.test.MetricsAsserts.*;
|
import static org.apache.hadoop.test.MetricsAsserts.*;
|
||||||
|
|
||||||
import static org.mockito.Mockito.*;
|
import static org.mockito.Mockito.*;
|
||||||
@ -71,6 +75,9 @@ public interface TestProtocol extends VersionedProtocol {
|
|||||||
int error() throws IOException;
|
int error() throws IOException;
|
||||||
void testServerGet() throws IOException;
|
void testServerGet() throws IOException;
|
||||||
int[] exchange(int[] values) throws IOException;
|
int[] exchange(int[] values) throws IOException;
|
||||||
|
|
||||||
|
DescriptorProtos.EnumDescriptorProto exchangeProto(
|
||||||
|
DescriptorProtos.EnumDescriptorProto arg);
|
||||||
}
|
}
|
||||||
|
|
||||||
public static class TestImpl implements TestProtocol {
|
public static class TestImpl implements TestProtocol {
|
||||||
@ -136,6 +143,11 @@ public int[] exchange(int[] values) {
|
|||||||
}
|
}
|
||||||
return values;
|
return values;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public EnumDescriptorProto exchangeProto(EnumDescriptorProto arg) {
|
||||||
|
return arg;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
//
|
//
|
||||||
@ -315,6 +327,13 @@ public void testCalls(Configuration conf) throws Exception {
|
|||||||
intResult = proxy.add(new int[] {1, 2});
|
intResult = proxy.add(new int[] {1, 2});
|
||||||
assertEquals(intResult, 3);
|
assertEquals(intResult, 3);
|
||||||
|
|
||||||
|
// Test protobufs
|
||||||
|
EnumDescriptorProto sendProto =
|
||||||
|
EnumDescriptorProto.newBuilder().setName("test").build();
|
||||||
|
EnumDescriptorProto retProto = proxy.exchangeProto(sendProto);
|
||||||
|
assertEquals(sendProto, retProto);
|
||||||
|
assertNotSame(sendProto, retProto);
|
||||||
|
|
||||||
boolean caught = false;
|
boolean caught = false;
|
||||||
try {
|
try {
|
||||||
proxy.error();
|
proxy.error();
|
||||||
|
@ -0,0 +1,72 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.util;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
|
||||||
|
import java.io.ByteArrayInputStream;
|
||||||
|
import java.io.ByteArrayOutputStream;
|
||||||
|
import java.io.DataInputStream;
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import com.google.protobuf.CodedOutputStream;
|
||||||
|
|
||||||
|
public class TestProtoUtil {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Values to test encoding as variable length integers
|
||||||
|
*/
|
||||||
|
private static final int[] TEST_VINT_VALUES = new int[] {
|
||||||
|
0, 1, -1, 127, 128, 129, 255, 256, 257,
|
||||||
|
0x1234, -0x1234,
|
||||||
|
0x123456, -0x123456,
|
||||||
|
0x12345678, -0x12345678
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test that readRawVarint32 is compatible with the varints encoded
|
||||||
|
* by ProtoBuf's CodedOutputStream.
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testVarInt() throws IOException {
|
||||||
|
// Test a few manufactured values
|
||||||
|
for (int value : TEST_VINT_VALUES) {
|
||||||
|
doVarIntTest(value);
|
||||||
|
}
|
||||||
|
// Check 1-bits at every bit position
|
||||||
|
for (int i = 1; i != 0; i <<= 1) {
|
||||||
|
doVarIntTest(i);
|
||||||
|
doVarIntTest(-i);
|
||||||
|
doVarIntTest(i - 1);
|
||||||
|
doVarIntTest(~i);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void doVarIntTest(int value) throws IOException {
|
||||||
|
ByteArrayOutputStream baos = new ByteArrayOutputStream();
|
||||||
|
CodedOutputStream cout = CodedOutputStream.newInstance(baos);
|
||||||
|
cout.writeRawVarint32(value);
|
||||||
|
cout.flush();
|
||||||
|
|
||||||
|
DataInputStream dis = new DataInputStream(
|
||||||
|
new ByteArrayInputStream(baos.toByteArray()));
|
||||||
|
assertEquals(value, ProtoUtil.readRawVarint32(dis));
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue
Block a user