@java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto) obj; boolean result = true; result = result && (hasBlockPoolId() == other.hasBlockPoolId()); if (hasBlockPoolId()) { result = result && getBlockPoolId() .equals(other.getBlockPoolId()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; }
public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getBlockPoolIdBytes()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; }
finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); } else { finalizeCmdBuilder_.clear();
if (hasFinalizeCmd()) { result = result && getFinalizeCmd() .equals(other.getFinalizeCmd());
finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); } else { finalizeCmdBuilder_.clear();
if (hasFinalizeCmd()) { result = result && getFinalizeCmd() .equals(other.getFinalizeCmd());
hash = (53 * hash) + getFinalizeCmd().hashCode();
hash = (53 * hash) + getFinalizeCmd().hashCode();
finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); } else { finalizeCmdBuilder_.clear();
@java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto) obj; boolean result = true; result = result && (hasBlockPoolId() == other.hasBlockPoolId()); if (hasBlockPoolId()) { result = result && getBlockPoolId() .equals(other.getBlockPoolId()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; }
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder mergeFinalizeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto value) { if (finalizeCmdBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && finalizeCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.newBuilder(finalizeCmd_).mergeFrom(value).buildPartial(); } else { finalizeCmd_ = value; } onChanged(); } else { finalizeCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder mergeFinalizeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto value) { if (finalizeCmdBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && finalizeCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.newBuilder(finalizeCmd_).mergeFrom(value).buildPartial(); } else { finalizeCmd_ = value; } onChanged(); } else { finalizeCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
@java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto) obj; boolean result = true; result = result && (hasBlockPoolId() == other.hasBlockPoolId()); if (hasBlockPoolId()) { result = result && getBlockPoolId() .equals(other.getBlockPoolId()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; }
public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); }
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder clearFinalizeCmd() { if (finalizeCmdBuilder_ == null) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); onChanged(); } else { finalizeCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); return this; } /**
@java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasBlockPoolId()) { hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER; hash = (53 * hash) + getBlockPoolId().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; }
public static FinalizeCommand convert(FinalizeCommandProto finalizeCmd) { return new FinalizeCommand(finalizeCmd.getBlockPoolId()); }
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder clearFinalizeCmd() { if (finalizeCmdBuilder_ == null) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); onChanged(); } else { finalizeCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); return this; } /**
public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); }
@java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasBlockPoolId()) { hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER; hash = (53 * hash) + getBlockPoolId().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; }