.setSubmissionState(SubmissionStateProto.valueOf(submissionState.name())) .build(); return responseBuilder.setUniqueNodeId(daemonId.getUniqueNodeIdInCluster()) .setSubmissionState(SubmissionStateProto.valueOf(submissionState.name())) .build();
public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_;
.setSubmissionState(SubmissionStateProto.valueOf(submissionState.name())) .build(); return responseBuilder.setUniqueNodeId(daemonId.getUniqueNodeIdInCluster()) .setSubmissionState(SubmissionStateProto.valueOf(submissionState.name())) .build();
SubmitWorkResponseProto .newBuilder() .setSubmissionState(SubmissionStateProto.ACCEPTED) .build()); try { server.init(new Configuration());
private Builder() { maybeForceBuilderInitialization(); }
private static Builder create() { return new Builder(); }
public Builder clone() { return create().mergeFrom(buildPartial()); }
public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); }
public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance()) return this; if (other.hasSubmissionState()) { setSubmissionState(other.getSubmissionState()); } if (other.hasUniqueNodeId()) { bitField0_ |= 0x00000002; uniqueNodeId_ = other.uniqueNodeId_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; }
@java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /**
private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() {
public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto)other); } else { super.mergeFrom(other); return this; } }
public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; }