gpt4 book ai didi

org.apache.beam.sdk.io.WriteFilesResult.getPerDestinationOutputFilenames()方法的使用及代码示例

转载 作者:知者 更新时间:2024-03-27 07:35:05 28 4
gpt4 key购买 nike

本文整理了Java中org.apache.beam.sdk.io.WriteFilesResult.getPerDestinationOutputFilenames()方法的一些代码示例,展示了WriteFilesResult.getPerDestinationOutputFilenames()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。WriteFilesResult.getPerDestinationOutputFilenames()方法的具体详情如下:
包路径:org.apache.beam.sdk.io.WriteFilesResult
类名称:WriteFilesResult
方法名:getPerDestinationOutputFilenames

WriteFilesResult.getPerDestinationOutputFilenames介绍

[英]Returns a PCollection of all output filenames generated by this WriteFilesorganized by user destination type.
[中]返回此WriteFileOrganized by user destination type生成的所有输出文件名的PCollection。

代码示例

代码示例来源:origin: org.apache.beam/beam-runners-google-cloud-dataflow-java

private void testStreamingWriteOverride(PipelineOptions options, int expectedNumShards) {
 TestPipeline p = TestPipeline.fromOptions(options);
 StreamingShardedWriteFactory<Object, Void, Object> factory =
   new StreamingShardedWriteFactory<>(p.getOptions());
 WriteFiles<Object, Void, Object> original = WriteFiles.to(new TestSink(tmpFolder.toString()));
 PCollection<Object> objs = (PCollection) p.apply(Create.empty(VoidCoder.of()));
 AppliedPTransform<PCollection<Object>, WriteFilesResult<Void>, WriteFiles<Object, Void, Object>>
   originalApplication =
     AppliedPTransform.of("writefiles", objs.expand(), Collections.emptyMap(), original, p);
 WriteFiles<Object, Void, Object> replacement =
   (WriteFiles<Object, Void, Object>)
     factory.getReplacementTransform(originalApplication).getTransform();
 assertThat(replacement, not(equalTo((Object) original)));
 assertThat(replacement.getNumShardsProvider().get(), equalTo(expectedNumShards));
 WriteFilesResult<Void> originalResult = objs.apply(original);
 WriteFilesResult<Void> replacementResult = objs.apply(replacement);
 Map<PValue, ReplacementOutput> res =
   factory.mapOutputs(originalResult.expand(), replacementResult);
 assertEquals(1, res.size());
 assertEquals(
   originalResult.getPerDestinationOutputFilenames(),
   res.get(replacementResult.getPerDestinationOutputFilenames()).getOriginal().getValue());
}

代码示例来源:origin: org.apache.beam/beam-sdks-java-core

@Test
@Category(NeedsRunner.class)
public void testCustomShardedWrite() throws IOException {
 // Flag to validate that the pipeline options are passed to the Sink
 WriteOptions options = TestPipeline.testingPipelineOptions().as(WriteOptions.class);
 options.setTestFlag("test_value");
 Pipeline p = TestPipeline.create(options);
 List<String> inputs = new ArrayList<>();
 // Prepare timestamps for the elements.
 List<Long> timestamps = new ArrayList<>();
 for (long i = 0; i < 1000; i++) {
  inputs.add(Integer.toString(3));
  timestamps.add(i + 1);
 }
 SimpleSink<Void> sink = makeSimpleSink();
 WriteFiles<String, ?, String> write = WriteFiles.to(sink).withSharding(new LargestInt());
 p.apply(Create.timestamped(inputs, timestamps).withCoder(StringUtf8Coder.of()))
   .apply(IDENTITY_MAP)
   .apply(write)
   .getPerDestinationOutputFilenames()
   .apply(new VerifyFilesExist<>());
 p.run();
 checkFileContents(
   getBaseOutputFilename(), inputs, Optional.of(3), true /* expectRemovedTempDirectory */);
}

代码示例来源:origin: org.apache.beam/beam-sdks-java-core

.apply(transform)
  .apply(write)
  .getPerDestinationOutputFilenames()
  .apply(new VerifyFilesExist<>());
p.run();

代码示例来源:origin: org.apache.beam/beam-sdks-java-core

res = input.apply(writeFiles);
res.getPerDestinationOutputFilenames().apply(new VerifyFilesExist<>());
p.run();

代码示例来源:origin: org.apache.beam/beam-sdks-java-core

@Test
@Category(NeedsRunner.class)
public void testWriteViaSink() throws Exception {
 List<String> data = ImmutableList.of("a", "b", "c", "d", "e", "f");
 PAssert.that(
     p.apply(Create.of(data))
       .apply(
         FileIO.<String>write()
           .to(tempFolder.getRoot().toString())
           .withSuffix(".txt")
           .via(TextIO.sink())
           .withIgnoreWindowing())
       .getPerDestinationOutputFilenames()
       .apply(Values.create())
       .apply(TextIO.readAll()))
   .containsInAnyOrder(data);
 p.run();
}

代码示例来源:origin: org.apache.beam/beam-sdks-java-core

@Test
@Category(NeedsRunner.class)
public void testWindowedWritesWithOnceTrigger() throws Throwable {
 // Tests for https://issues.apache.org/jira/browse/BEAM-3169
 PCollection<String> data =
   p.apply(Create.of("0", "1", "2"))
     .apply(
       Window.<String>into(FixedWindows.of(Duration.standardSeconds(1)))
         // According to this trigger, all data should be written.
         // However, the continuation of this trigger is elementCountAtLeast(1),
         // so with a buggy implementation that used a GBK before renaming files,
         // only 1 file would be renamed.
         .triggering(AfterPane.elementCountAtLeast(3))
         .withAllowedLateness(Duration.standardMinutes(1))
         .discardingFiredPanes());
 PCollection<String> filenames =
   data.apply(
       TextIO.write()
         .to(new File(tempFolder.getRoot(), "windowed-writes").getAbsolutePath())
         .withNumShards(2)
         .withWindowedWrites()
         .<Void>withOutputFilenames())
     .getPerDestinationOutputFilenames()
     .apply(Values.create());
 PAssert.that(filenames.apply(TextIO.readAll())).containsInAnyOrder("0", "1", "2");
 p.run();
}

代码示例来源:origin: org.apache.beam/beam-sdks-java-io-xml

.withPrefix("birds")
    .withSuffix(".xml"))
.getPerDestinationOutputFilenames()
.apply(Values.create())
.apply(FileIO.matchAll())

代码示例来源:origin: GoogleCloudPlatform/DataflowTemplates

fileWriteResults.getPerDestinationOutputFilenames().apply(GroupByKey.create());

28 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com