public org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto buildPartial() { org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto result = new org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0;
hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasApplicationId()) { hash = (37 * hash) + APPLICATION_ID_FIELD_NUMBER; hash = (53 * hash) + getApplicationId().hashCode(); if (hasApplicationName()) { hash = (37 * hash) + APPLICATION_NAME_FIELD_NUMBER; hash = (53 * hash) + getApplicationName().hashCode(); if (hasApplicationType()) { hash = (37 * hash) + APPLICATION_TYPE_FIELD_NUMBER; hash = (53 * hash) + getApplicationType().hashCode(); if (hasUser()) { hash = (37 * hash) + USER_FIELD_NUMBER; hash = (53 * hash) + getUser().hashCode(); if (hasQueue()) { hash = (37 * hash) + QUEUE_FIELD_NUMBER; hash = (53 * hash) + getQueue().hashCode(); if (hasSubmitTime()) { hash = (37 * hash) + SUBMIT_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getSubmitTime()); if (hasStartTime()) { hash = (37 * hash) + START_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getStartTime());
.computeBytesSize(2, getApplicationNameBytes()); .computeBytesSize(3, getApplicationTypeBytes()); .computeBytesSize(4, getUserBytes()); .computeBytesSize(5, getQueueBytes()); .computeInt64Size(7, startTime_); size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size;
result = result && (hasApplicationId() == other.hasApplicationId()); if (hasApplicationId()) { result = result && getApplicationId() .equals(other.getApplicationId()); result = result && (hasApplicationName() == other.hasApplicationName()); if (hasApplicationName()) { result = result && getApplicationName() .equals(other.getApplicationName()); result = result && (hasApplicationType() == other.hasApplicationType()); if (hasApplicationType()) { result = result && getApplicationType() .equals(other.getApplicationType()); result = result && (hasUser() == other.hasUser()); if (hasUser()) { result = result && getUser() .equals(other.getUser()); result = result && (hasQueue() == other.hasQueue()); if (hasQueue()) { result = result && getQueue() .equals(other.getQueue()); result = result && (hasSubmitTime() == other.hasSubmitTime()); if (hasSubmitTime()) { result = result && (getSubmitTime() == other.getSubmitTime());
result = result && (hasApplicationId() == other.hasApplicationId()); if (hasApplicationId()) { result = result && getApplicationId() .equals(other.getApplicationId()); result = result && (hasApplicationName() == other.hasApplicationName()); if (hasApplicationName()) { result = result && getApplicationName() .equals(other.getApplicationName()); result = result && (hasApplicationType() == other.hasApplicationType()); if (hasApplicationType()) { result = result && getApplicationType() .equals(other.getApplicationType()); result = result && (hasUser() == other.hasUser()); if (hasUser()) { result = result && getUser() .equals(other.getUser()); result = result && (hasQueue() == other.hasQueue()); if (hasQueue()) { result = result && getQueue() .equals(other.getQueue()); result = result && (hasSubmitTime() == other.hasSubmitTime()); if (hasSubmitTime()) { result = result && (getSubmitTime() == other.getSubmitTime());
public org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto buildPartial() { org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto result = new org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0;
public org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto buildPartial() { org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto result = new org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0;
result = result && (hasApplicationId() == other.hasApplicationId()); if (hasApplicationId()) { result = result && getApplicationId() .equals(other.getApplicationId()); result = result && (hasApplicationName() == other.hasApplicationName()); if (hasApplicationName()) { result = result && getApplicationName() .equals(other.getApplicationName()); result = result && (hasApplicationType() == other.hasApplicationType()); if (hasApplicationType()) { result = result && getApplicationType() .equals(other.getApplicationType()); result = result && (hasUser() == other.hasUser()); if (hasUser()) { result = result && getUser() .equals(other.getUser()); result = result && (hasQueue() == other.hasQueue()); if (hasQueue()) { result = result && getQueue() .equals(other.getQueue()); result = result && (hasSubmitTime() == other.hasSubmitTime()); if (hasSubmitTime()) { result = result && (getSubmitTime() == other.getSubmitTime());
result = result && (hasApplicationId() == other.hasApplicationId()); if (hasApplicationId()) { result = result && getApplicationId() .equals(other.getApplicationId()); result = result && (hasApplicationName() == other.hasApplicationName()); if (hasApplicationName()) { result = result && getApplicationName() .equals(other.getApplicationName()); result = result && (hasApplicationType() == other.hasApplicationType()); if (hasApplicationType()) { result = result && getApplicationType() .equals(other.getApplicationType()); result = result && (hasUser() == other.hasUser()); if (hasUser()) { result = result && getUser() .equals(other.getUser()); result = result && (hasQueue() == other.hasQueue()); if (hasQueue()) { result = result && getQueue() .equals(other.getQueue()); result = result && (hasSubmitTime() == other.hasSubmitTime()); if (hasSubmitTime()) { result = result && (getSubmitTime() == other.getSubmitTime());
public Builder mergeFrom(org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto other) { if (other == org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto.getDefaultInstance()) return this; if (other.hasApplicationId()) { mergeApplicationId(other.getApplicationId()); if (other.hasApplicationName()) { bitField0_ |= 0x00000002; applicationName_ = other.applicationName_; onChanged(); if (other.hasApplicationType()) { bitField0_ |= 0x00000004; applicationType_ = other.applicationType_; onChanged(); if (other.hasUser()) { bitField0_ |= 0x00000008; user_ = other.user_; onChanged(); if (other.hasQueue()) { bitField0_ |= 0x00000010; queue_ = other.queue_; onChanged(); if (other.hasSubmitTime()) { setSubmitTime(other.getSubmitTime()); if (other.hasStartTime()) { setStartTime(other.getStartTime());
public Builder mergeFrom(org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto other) { if (other == org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto.getDefaultInstance()) return this; if (other.hasApplicationId()) { mergeApplicationId(other.getApplicationId()); if (other.hasApplicationName()) { bitField0_ |= 0x00000002; applicationName_ = other.applicationName_; onChanged(); if (other.hasApplicationType()) { bitField0_ |= 0x00000004; applicationType_ = other.applicationType_; onChanged(); if (other.hasUser()) { bitField0_ |= 0x00000008; user_ = other.user_; onChanged(); if (other.hasQueue()) { bitField0_ |= 0x00000010; queue_ = other.queue_; onChanged(); if (other.hasSubmitTime()) { setSubmitTime(other.getSubmitTime()); if (other.hasStartTime()) { setStartTime(other.getStartTime());
.computeBytesSize(2, getApplicationNameBytes()); .computeBytesSize(3, getApplicationTypeBytes()); .computeBytesSize(4, getUserBytes()); .computeBytesSize(5, getQueueBytes()); .computeInt64Size(7, startTime_); size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size;
.computeBytesSize(2, getApplicationNameBytes()); .computeBytesSize(3, getApplicationTypeBytes()); .computeBytesSize(4, getUserBytes()); .computeBytesSize(5, getQueueBytes()); .computeInt64Size(7, startTime_); size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size;
public org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto buildPartial() { org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto result = new org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0;
public Builder mergeFrom(org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto other) { if (other == org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto.getDefaultInstance()) return this; if (other.hasApplicationId()) { mergeApplicationId(other.getApplicationId()); if (other.hasApplicationName()) { bitField0_ |= 0x00000002; applicationName_ = other.applicationName_; onChanged(); if (other.hasApplicationType()) { bitField0_ |= 0x00000004; applicationType_ = other.applicationType_; onChanged(); if (other.hasUser()) { bitField0_ |= 0x00000008; user_ = other.user_; onChanged(); if (other.hasQueue()) { bitField0_ |= 0x00000010; queue_ = other.queue_; onChanged(); if (other.hasSubmitTime()) { setSubmitTime(other.getSubmitTime()); if (other.hasStartTime()) { setStartTime(other.getStartTime());
.computeBytesSize(2, getApplicationNameBytes()); .computeBytesSize(3, getApplicationTypeBytes()); .computeBytesSize(4, getUserBytes()); .computeBytesSize(5, getQueueBytes()); .computeInt64Size(7, startTime_); size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size;
hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasApplicationId()) { hash = (37 * hash) + APPLICATION_ID_FIELD_NUMBER; hash = (53 * hash) + getApplicationId().hashCode(); if (hasApplicationName()) { hash = (37 * hash) + APPLICATION_NAME_FIELD_NUMBER; hash = (53 * hash) + getApplicationName().hashCode(); if (hasApplicationType()) { hash = (37 * hash) + APPLICATION_TYPE_FIELD_NUMBER; hash = (53 * hash) + getApplicationType().hashCode(); if (hasUser()) { hash = (37 * hash) + USER_FIELD_NUMBER; hash = (53 * hash) + getUser().hashCode(); if (hasQueue()) { hash = (37 * hash) + QUEUE_FIELD_NUMBER; hash = (53 * hash) + getQueue().hashCode(); if (hasSubmitTime()) { hash = (37 * hash) + SUBMIT_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getSubmitTime()); if (hasStartTime()) { hash = (37 * hash) + START_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getStartTime());
hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasApplicationId()) { hash = (37 * hash) + APPLICATION_ID_FIELD_NUMBER; hash = (53 * hash) + getApplicationId().hashCode(); if (hasApplicationName()) { hash = (37 * hash) + APPLICATION_NAME_FIELD_NUMBER; hash = (53 * hash) + getApplicationName().hashCode(); if (hasApplicationType()) { hash = (37 * hash) + APPLICATION_TYPE_FIELD_NUMBER; hash = (53 * hash) + getApplicationType().hashCode(); if (hasUser()) { hash = (37 * hash) + USER_FIELD_NUMBER; hash = (53 * hash) + getUser().hashCode(); if (hasQueue()) { hash = (37 * hash) + QUEUE_FIELD_NUMBER; hash = (53 * hash) + getQueue().hashCode(); if (hasSubmitTime()) { hash = (37 * hash) + SUBMIT_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getSubmitTime()); if (hasStartTime()) { hash = (37 * hash) + START_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getStartTime());
hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasApplicationId()) { hash = (37 * hash) + APPLICATION_ID_FIELD_NUMBER; hash = (53 * hash) + getApplicationId().hashCode(); if (hasApplicationName()) { hash = (37 * hash) + APPLICATION_NAME_FIELD_NUMBER; hash = (53 * hash) + getApplicationName().hashCode(); if (hasApplicationType()) { hash = (37 * hash) + APPLICATION_TYPE_FIELD_NUMBER; hash = (53 * hash) + getApplicationType().hashCode(); if (hasUser()) { hash = (37 * hash) + USER_FIELD_NUMBER; hash = (53 * hash) + getUser().hashCode(); if (hasQueue()) { hash = (37 * hash) + QUEUE_FIELD_NUMBER; hash = (53 * hash) + getQueue().hashCode(); if (hasSubmitTime()) { hash = (37 * hash) + SUBMIT_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getSubmitTime()); if (hasStartTime()) { hash = (37 * hash) + START_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getStartTime());
public Builder mergeFrom(org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto other) { if (other == org.apache.hadoop.yarn.proto.ApplicationHistoryServerProtos.ApplicationStartDataProto.getDefaultInstance()) return this; if (other.hasApplicationId()) { mergeApplicationId(other.getApplicationId()); if (other.hasApplicationName()) { bitField0_ |= 0x00000002; applicationName_ = other.applicationName_; onChanged(); if (other.hasApplicationType()) { bitField0_ |= 0x00000004; applicationType_ = other.applicationType_; onChanged(); if (other.hasUser()) { bitField0_ |= 0x00000008; user_ = other.user_; onChanged(); if (other.hasQueue()) { bitField0_ |= 0x00000010; queue_ = other.queue_; onChanged(); if (other.hasSubmitTime()) { setSubmitTime(other.getSubmitTime()); if (other.hasStartTime()) { setStartTime(other.getStartTime());