private static ExportProtos.ExportRequest getConfiguredRequest(Configuration conf, Path dir, final Scan scan, final Token<?> userToken) throws IOException { boolean compressed = conf.getBoolean(FileOutputFormat.COMPRESS, false); String compressionType = conf.get(FileOutputFormat.COMPRESS_TYPE, DEFAULT_TYPE.toString()); String compressionCodec = conf.get(FileOutputFormat.COMPRESS_CODEC, DEFAULT_CODEC.getName()); DelegationToken protoToken = null; if (userToken != null) { protoToken = DelegationToken.newBuilder() .setIdentifier(ByteStringer.wrap(userToken.getIdentifier())) .setPassword(ByteStringer.wrap(userToken.getPassword())) .setKind(userToken.getKind().toString()) .setService(userToken.getService().toString()).build(); } LOG.info("compressed=" + compressed + ", compression type=" + compressionType + ", compression codec=" + compressionCodec + ", userToken=" + userToken); ExportProtos.ExportRequest.Builder builder = ExportProtos.ExportRequest.newBuilder() .setScan(ProtobufUtil.toScan(scan)) .setOutputPath(dir.toString()) .setCompressed(compressed) .setCompressCodec(compressionCodec) .setCompressType(compressionType); if (protoToken != null) { builder.setFsToken(protoToken); } return builder.build(); }
private static ExportProtos.ExportRequest getConfiguredRequest(Configuration conf, Path dir, final Scan scan, final Token<?> userToken) throws IOException { boolean compressed = conf.getBoolean(FileOutputFormat.COMPRESS, false); String compressionType = conf.get(FileOutputFormat.COMPRESS_TYPE, DEFAULT_TYPE.toString()); String compressionCodec = conf.get(FileOutputFormat.COMPRESS_CODEC, DEFAULT_CODEC.getName()); DelegationToken protoToken = null; if (userToken != null) { protoToken = DelegationToken.newBuilder() .setIdentifier(ByteStringer.wrap(userToken.getIdentifier())) .setPassword(ByteStringer.wrap(userToken.getPassword())) .setKind(userToken.getKind().toString()) .setService(userToken.getService().toString()).build(); } LOG.info("compressed=" + compressed + ", compression type=" + compressionType + ", compression codec=" + compressionCodec + ", userToken=" + userToken); ExportProtos.ExportRequest.Builder builder = ExportProtos.ExportRequest.newBuilder() .setScan(ProtobufUtil.toScan(scan)) .setOutputPath(dir.toString()) .setCompressed(compressed) .setCompressCodec(compressionCodec) .setCompressType(compressionType); if (protoToken != null) { builder.setFsToken(protoToken); } return builder.build(); }
private static ExportProtos.ExportRequest getConfiguredRequest(Configuration conf, Path dir, final Scan scan, final Token<?> userToken) throws IOException { boolean compressed = conf.getBoolean(FileOutputFormat.COMPRESS, false); String compressionType = conf.get(FileOutputFormat.COMPRESS_TYPE, DEFAULT_TYPE.toString()); String compressionCodec = conf.get(FileOutputFormat.COMPRESS_CODEC, DEFAULT_CODEC.getName()); DelegationToken protoToken = null; if (userToken != null) { protoToken = DelegationToken.newBuilder() .setIdentifier(ByteStringer.wrap(userToken.getIdentifier())) .setPassword(ByteStringer.wrap(userToken.getPassword())) .setKind(userToken.getKind().toString()) .setService(userToken.getService().toString()).build(); } LOG.info("compressed=" + compressed + ", compression type=" + compressionType + ", compression codec=" + compressionCodec + ", userToken=" + userToken); ExportProtos.ExportRequest.Builder builder = ExportProtos.ExportRequest.newBuilder() .setScan(ProtobufUtil.toScan(scan)) .setOutputPath(dir.toString()) .setCompressed(compressed) .setCompressCodec(compressionCodec) .setCompressType(compressionType); if (protoToken != null) { builder.setFsToken(protoToken); } return builder.build(); }
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ExportProtos.ExportRequest other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ExportProtos.ExportRequest.getDefaultInstance()) return this; if (other.hasScan()) { mergeScan(other.getScan()); } if (other.hasOutputPath()) { bitField0_ |= 0x00000002; outputPath_ = other.outputPath_; onChanged(); } if (other.hasCompressed()) { setCompressed(other.getCompressed()); } if (other.hasCompressType()) { bitField0_ |= 0x00000008; compressType_ = other.compressType_; onChanged(); } if (other.hasCompressCodec()) { bitField0_ |= 0x00000010; compressCodec_ = other.compressCodec_; onChanged(); } if (other.hasFsToken()) { mergeFsToken(other.getFsToken()); } this.mergeUnknownFields(other.getUnknownFields()); return this; }
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ExportProtos.ExportRequest other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ExportProtos.ExportRequest.getDefaultInstance()) return this; if (other.hasScan()) { mergeScan(other.getScan()); } if (other.hasOutputPath()) { bitField0_ |= 0x00000002; outputPath_ = other.outputPath_; onChanged(); } if (other.hasCompressed()) { setCompressed(other.getCompressed()); } if (other.hasCompressType()) { bitField0_ |= 0x00000008; compressType_ = other.compressType_; onChanged(); } if (other.hasCompressCodec()) { bitField0_ |= 0x00000010; compressCodec_ = other.compressCodec_; onChanged(); } if (other.hasFsToken()) { mergeFsToken(other.getFsToken()); } this.mergeUnknownFields(other.getUnknownFields()); return this; }
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ExportProtos.ExportRequest other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ExportProtos.ExportRequest.getDefaultInstance()) return this; if (other.hasScan()) { mergeScan(other.getScan()); } if (other.hasOutputPath()) { bitField0_ |= 0x00000002; outputPath_ = other.outputPath_; onChanged(); } if (other.hasCompressed()) { setCompressed(other.getCompressed()); } if (other.hasCompressType()) { bitField0_ |= 0x00000008; compressType_ = other.compressType_; onChanged(); } if (other.hasCompressCodec()) { bitField0_ |= 0x00000010; compressCodec_ = other.compressCodec_; onChanged(); } if (other.hasFsToken()) { mergeFsToken(other.getFsToken()); } this.mergeUnknownFields(other.getUnknownFields()); return this; }