if (subBuilder != null) { subBuilder.mergeFrom(cause_); cause_ = subBuilder.buildPartial();
if (subBuilder != null) { subBuilder.mergeFrom(exception_); exception_ = subBuilder.buildPartial();
if (subBuilder != null) { subBuilder.mergeFrom(exception_); exception_ = subBuilder.buildPartial();
if (subBuilder != null) { subBuilder.mergeFrom(exception_); exception_ = subBuilder.buildPartial();
if (subBuilder != null) { subBuilder.mergeFrom(exception_); exception_ = subBuilder.buildPartial();
/** * <code>optional .hadoop.yarn.SerializedExceptionProto exception = 5;</code> */ public Builder mergeException(org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto value) { if (exceptionBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && exception_ != org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.getDefaultInstance()) { exception_ = org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.newBuilder(exception_).mergeFrom(value).buildPartial(); } else { exception_ = value; } onChanged(); } else { exceptionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
/** * <code>optional .hadoop.yarn.SerializedExceptionProto cause = 4;</code> */ public Builder mergeCause(org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto value) { if (causeBuilder_ == null) { if (((bitField0_ & 0x00000008) == 0x00000008) && cause_ != org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.getDefaultInstance()) { cause_ = org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.newBuilder(cause_).mergeFrom(value).buildPartial(); } else { cause_ = value; } onChanged(); } else { causeBuilder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /**
/** * <code>optional .hadoop.yarn.SerializedExceptionProto exception = 5;</code> */ public Builder mergeException(org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto value) { if (exceptionBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && exception_ != org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.getDefaultInstance()) { exception_ = org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.newBuilder(exception_).mergeFrom(value).buildPartial(); } else { exception_ = value; } onChanged(); } else { exceptionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
/** * <code>optional .hadoop.yarn.SerializedExceptionProto exception = 5;</code> */ public Builder mergeException(org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto value) { if (exceptionBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && exception_ != org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.getDefaultInstance()) { exception_ = org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.newBuilder(exception_).mergeFrom(value).buildPartial(); } else { exception_ = value; } onChanged(); } else { exceptionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
/** * <code>optional .hadoop.yarn.SerializedExceptionProto exception = 5;</code> */ public Builder mergeException(org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto value) { if (exceptionBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && exception_ != org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.getDefaultInstance()) { exception_ = org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto.newBuilder(exception_).mergeFrom(value).buildPartial(); } else { exception_ = value; } onChanged(); } else { exceptionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
public Builder clone() { return create().mergeFrom(buildPartial()); }
public org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto build() { org.apache.hadoop.yarn.proto.YarnProtos.SerializedExceptionProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; }