Examples of ReduceOperator


Examples of org.apache.flink.api.java.record.operators.ReduceOperator

      .input(dataPoints)
      .name("Find Nearest Centers")
      .build();

    // create reducer recomputes the cluster centers as the  average of all associated data points
    ReduceOperator recomputeClusterCenter = ReduceOperator.builder(new RecomputeClusterCenter(), IntValue.class, 0)
      .input(findNearestClusterCenters)
      .name("Recompute Center Positions")
      .build();

    // create DataSinkContract for writing the new cluster positions
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

        JoinOperator.builder(JoinLiO.class, IntValue.class, 0, 0)
      .name("OrdersLineitemsJoin")
      .build();
      join.setDegreeOfParallelism(degreeOfParallelism);
   
    ReduceOperator aggregation =
        ReduceOperator.builder(CountAgg.class, StringValue.class, 0)
      .name("AggregateGroupBy")
      .build();
    aggregation.setDegreeOfParallelism(this.degreeOfParallelism);
   
    lineFilter.setInput(lineItems);
    ordersFilter.setInput(orders);
    join.setFirstInput(ordersFilter);
    join.setSecondInput(lineFilter);
    aggregation.setInput(join);
    result.setInput(aggregation);
   
     
    return new Plan(result, "TPC-H 4");
  }
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

    BulkIteration iteration = new BulkIteration("Loop");
    iteration.setInput(initialInput);
    iteration.setMaximumNumberOfIterations(2);

    ReduceOperator dummyReduce = ReduceOperator.builder(new DummyReducer(), IntValue.class, 0).input(iteration.getPartialSolution())
        .name("Reduce something").build();

    MapOperator dummyMap = MapOperator.builder(new IdentityMapper()).input(dummyReduce).build();
    iteration.setNextPartialSolution(dummyMap);
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

      .name("JoinCO")
      .build();
    joinCO.setDegreeOfParallelism(numSubtasks);

    // create ReduceOperator for aggregating the result
    ReduceOperator aggCO = ReduceOperator.builder(new AggCO(), StringValue.class, 1)
      .name("AggCo")
      .build();
    aggCO.setDegreeOfParallelism(numSubtasks);

    // create DataSinkContract for writing the result
    FileDataSink result = new FileDataSink(new CsvOutputFormat(), output, "Output");
    result.setDegreeOfParallelism(numSubtasks);
    CsvOutputFormat.configureRecordFormat(result)
      .recordDelimiter('\n')
      .fieldDelimiter('|')
      .field(IntValue.class, 0)
      .field(StringValue.class, 1);

    // assemble the plan
    result.setInput(aggCO);
    aggCO.setInput(joinCO);
    joinCO.setFirstInput(orders);
    joinCO.setSecondInput(customers);

    return new Plan(result, "TPCH Asterix");
  }
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

        .input2(edges)
        .name("Join Candidate Id With Neighbor")
        .build();

    // find for each neighbor the smallest of all candidates
    ReduceOperator minCandidateId = ReduceOperator.builder(new MinimumComponentIDReduce(), LongValue.class, 0)
        .input(joinWithNeighbors)
        .name("Find Minimum Candidate Id")
        .build();
   
    // join candidates with the solution set and update if the candidate component-id is smaller
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

        .input2(edges)
        .name("Join Candidate Id With Neighbor")
        .build();

    // find for each neighbor the smallest of all candidates
    ReduceOperator minCandidateId = ReduceOperator.builder(new MinimumComponentIDReduce(), LongValue.class, 0)
        .input(joinWithNeighbors)
        .name("Find Minimum Candidate Id")
        .build();
   
    // join candidates with the solution set and update if the candidate component-id is smaller
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

      MapOperator.builder(new LineItemFilter())
      .name("LineItem Filter")
      .build();
    lineItemFilter.setDegreeOfParallelism(this.degreeOfParallelism);
   
    ReduceOperator groupByReturnFlag =
      ReduceOperator.builder(new GroupByReturnFlag(), StringValue.class, 0)
      .name("groupyBy")
      .build();
   
    lineItemFilter.setInput(lineItems);
    groupByReturnFlag.setInput(lineItemFilter);
    result.setInput(groupByReturnFlag);
   
    return new Plan(result, "TPC-H 1");
  }
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

    BulkIteration iteration = new BulkIteration("Loop");
    iteration.setInput(initialInput);
    iteration.setMaximumNumberOfIterations(5);
    Assert.assertTrue(iteration.getMaximumNumberOfIterations() > 1);

    ReduceOperator sumReduce = ReduceOperator.builder(new SumReducer())
        .input(iteration.getPartialSolution())
        .name("Compute sum (Reduce)")
        .build();
   
    iteration.setNextPartialSolution(sumReduce);
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

    FileDataSource source = new FileDataSource(new TextInputFormat(), dataInput, "Input Lines");

    MapOperator mapper = MapOperator.builder(new TokenizeLine()).input(source).name("Tokenize Lines").build();
   
    ReduceOperator reducer = ReduceOperator.builder(CountWords.class, StringValue.class, 0).input(mapper)
        .name("Count Words").build();
   
    FileDataSink out = new FileDataSink(new CsvOutputFormat(), output, reducer, "Word Counts");
   
    CsvOutputFormat.configureRecordFormat(out).recordDelimiter('\n')
View Full Code Here

Examples of org.apache.flink.api.java.record.operators.ReduceOperator

   
    @SuppressWarnings("unchecked")
    CsvInputFormat format = new CsvInputFormat(',', IntValue.class, IntValue.class);
    FileDataSource source = new FileDataSource(format, this.textPath, "Source");
   
    ReduceOperator reducer = ReduceOperator.builder(CheckingReducer.class)
      .keyField(IntValue.class, 0)
      .input(source)
      .name("Ordered Reducer")
      .build();
    reducer.setGroupOrder(new Ordering(1, IntValue.class, Order.ASCENDING));
   
    FileDataSink sink = new FileDataSink(CsvOutputFormat.class, this.resultPath, reducer, "Sink");
    CsvOutputFormat.configureRecordFormat(sink)
      .recordDelimiter('\n')
      .fieldDelimiter(',')
View Full Code Here
TOP
Copyright © 2018 www.massapi.com. All rights reserved.
All source code are property of their respective owners. Java is a trademark of Sun Microsystems, Inc and owned by ORACLE Inc. Contact coftware#gmail.com.