public final com.google.protobuf.Message getResponsePrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } }
@java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; }
public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); }
com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable();
public Builder toBuilder() { return newBuilder(this); }
public void refreshHFiles( com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesRequest request, com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse> done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance(), com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.class, org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance())); } }
public final com.google.protobuf.Message getResponsePrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } }
@Override public void refreshHFiles(RpcController controller, RefreshHFilesProtos.RefreshHFilesRequest request, RpcCallback<RefreshHFilesProtos.RefreshHFilesResponse> done) { try { for (Store store : env.getRegion().getStores()) { LOG.debug("Refreshing HFiles for region: " + store.getRegionInfo().getRegionNameAsString() + " and store: " + store.getColumnFamilyName() + "class:" + store.getClass()); store.refreshStoreFiles(); } } catch (IOException ioe) { LOG.error("Exception while trying to refresh store files: ", ioe); CoprocessorRpcUtils.setControllerException(controller, ioe); } done.run(RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance()); }
public final com.google.protobuf.Message getResponsePrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } }
public org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance(); }
@java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse other = (org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse) obj; boolean result = true; result = result && getUnknownFields().equals(other.getUnknownFields()); return result; }
public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse prototype) {
public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; }
public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; }
public org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse buildPartial() { org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse result = new org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse(this); onBuilt(); return result; }
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse other) { if (other == org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.getUnknownFields()); return this; }
public org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse refreshHFiles( com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesRequest request) throws com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse.getDefaultInstance()); }
public org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse build() { org.apache.hadoop.hbase.protobuf.generated.RefreshHFilesProtos.RefreshHFilesResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; }
public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); }