public Builder clear() { super.clear(); if (keyBuilder_ == null) { key_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); } else { keyBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; }
public Builder clear() { super.clear(); if (markerBuilder_ == null) { marker_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); } else { markerBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; }
public Builder clear() { super.clear(); if (blockBuilder_ == null) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); status_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus.RECEIVING; bitField0_ = (bitField0_ & ~0x00000002); deleteHint_ = ""; bitField0_ = (bitField0_ & ~0x00000004); return this; }
/** * <code>optional .hadoop.hdfs.BlockProto nextMarker = 2;</code> */ public Builder clearNextMarker() { if (nextMarkerBuilder_ == null) { nextMarker_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); onChanged(); } else { nextMarkerBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /**
/** * <code>optional .hadoop.hdfs.BlockProto marker = 1;</code> */ public Builder clearMarker() { if (markerBuilder_ == null) { marker_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); onChanged(); } else { markerBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /**
/** * <code>required .hadoop.hdfs.BlockProto block = 1;</code> */ public Builder clearBlock() { if (blockBuilder_ == null) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); onChanged(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /**
private void initFields() { key_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
private void initFields() { marker_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
private void initFields() { fileRegions_ = java.util.Collections.emptyList(); nextMarker_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
/** * <code>repeated .hadoop.hdfs.BlockProto blocks = 3;</code> */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder( int index) { return getBlocksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); } /**
/** * <code>repeated .hadoop.hdfs.BlockProto blocks = 3;</code> */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder() { return getBlocksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); } /**
/** * <code>repeated .hadoop.hdfs.BlockProto blocks = 5;</code> */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder() { return getBlocksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); } /**
private void initFields() { replicaFound_ = false; state_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ReplicaStateProto.FINALIZED; block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
private void initFields() { key_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); value_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ProvidedStorageLocationProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
/** * <code>repeated .hadoop.hdfs.BlockProto blocks = 5;</code> */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder( int index) { return getBlocksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); } /**
/** * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder( int index) { return getBlocksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); } /**
private void initFields() { newGenStamp_ = 0L; block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.getDefaultInstance(); truncateBlock_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); ecPolicy_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ErasureCodingPolicyProto.getDefaultInstance(); blockIndices_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); status_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus.RECEIVING; deleteHint_ = ""; } private byte memoizedIsInitialized = -1;
/** * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder() { return getBlocksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); } /**
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance(); datanodeUuids_ = com.google.protobuf.LazyStringArrayList.EMPTY; storageUuids_ = com.google.protobuf.LazyStringArrayList.EMPTY; storageTypes_ = java.util.Collections.emptyList(); indices_ = com.google.protobuf.ByteString.EMPTY; dataBlockNum_ = 0; cellSize_ = 0; } private byte memoizedIsInitialized = -1;