Package org.apache.hadoop.mapreduce

Examples of org.apache.hadoop.mapreduce.Job


    }
    DefaultStringifier<Map<String,Integer>> mapStringifier = new DefaultStringifier<Map<String,Integer>>(
        conf, GenericsUtil.getClass(dictionary));
    conf.set("dictionary", mapStringifier.toString(dictionary));
   
    Job job = new Job(conf, "Generating dataset based from input"
                            + input);
    job.setJarByClass(VectorCreationJob.class);
   
    job.setOutputKeyClass(Text.class);
    job.setOutputValueClass(VectorWritable.class);
   
    FileInputFormat.addInputPath(job, input);
    FileOutputFormat.setOutputPath(job, output);
   
    HadoopUtil.delete(conf, output);
   
    job.setInputFormatClass(TextInputFormat.class);
    job.setMapperClass(VectorMapper.class);
    job.setCombinerClass(VectorReducer.class);
    job.setReducerClass(VectorReducer.class);
    job.setOutputFormatClass(SequenceFileOutputFormat.class);
   
    job.waitForCompletion(true);
  }
View Full Code Here


    conf.set("io.serializations",
      "org.apache.hadoop.io.serializer.JavaSerialization,"
          + "org.apache.hadoop.io.serializer.WritableSerialization");
   
    String input = params.get("input");
    Job job = new Job(conf, "Generating dataset based from input"
                            + input);
    job.setJarByClass(ByKeyGroupingJob.class);
   
    job.setOutputKeyClass(Text.class);
    job.setOutputValueClass(Text.class);
   
    FileInputFormat.addInputPath(job, new Path(input));
    Path outPath = new Path(params.get("output"));
    FileOutputFormat.setOutputPath(job, outPath);
   
    HadoopUtil.delete(conf, outPath);
   
    job.setInputFormatClass(TextInputFormat.class);
    job.setMapperClass(ByKeyMapper.class);
    job.setCombinerClass(ByKeyReducer.class);
    job.setReducerClass(ByKeyReducer.class);
    job.setOutputFormatClass(SequenceFileOutputFormat.class);
   
    job.waitForCompletion(true);
  }
View Full Code Here

  public static void generateDictionary(Path input, Path output) throws IOException,
                                                                InterruptedException,
                                                                ClassNotFoundException {
    Configuration conf = CreateNewConfiguration();
   
    Job job = new Job(conf, "Generating dataset based from input"
                            + input);
    job.setJarByClass(VectorCreationJob.class);
   
    job.setOutputKeyClass(Text.class);
    job.setOutputValueClass(IntWritable.class);
   
    FileInputFormat.addInputPath(job, input);
    FileOutputFormat.setOutputPath(job, output);
   
    HadoopUtil.delete(conf, output);
   
    job.setInputFormatClass(TextInputFormat.class);
    job.setMapperClass(DictionaryMapper.class);
    job.setCombinerClass(DictionaryReducer.class);
    job.setReducerClass(DictionaryReducer.class);
    job.setOutputFormatClass(SequenceFileOutputFormat.class);
   
    job.waitForCompletion(true);
  }
View Full Code Here

      UserGroupInformation ugi = UserGroupInformation.createProxyUser(user, UserGroupInformation.getLoginUser());
      ugi.doAs(new PrivilegedExceptionAction<Void>() {

        @Override
        public Void run() throws Exception {
          job = new Job(createStreamJob(cfg, argsList.toArray(new String[argsList.size()])));
          return null;
        }
      });
    }
    else {
      job = new Job(createStreamJob(cfg, argsList.toArray(new String[argsList.size()])));
    }

    job.setJobName(name);
  }
View Full Code Here

      UserGroupInformation ugi = UserGroupInformation.createProxyUser(user, UserGroupInformation.getLoginUser());
      ugi.doAs(new PrivilegedExceptionAction<Void>() {

        @Override
        public Void run() throws Exception {
          job = new Job(cfg);
          return null;
        }
      });
    }
    else {
      job = new Job(cfg);
    }

    ClassLoader loader = (beanClassLoader != null ? beanClassLoader : org.springframework.util.ClassUtils.getDefaultClassLoader());

    if (jar != null) {
View Full Code Here

       
        //InfiniDBInputFormat.setInput(jobconf, InfiniDoopRecord.class, "nation",
     //      null,  "n_nationkey", fields);
        InfiniDBOutputFormat.setOutput(jobconf, "t1", outFields);
        //InfiniDBInputFormat.setInput(jobconf, InfiniDBRecord.class, "select * from nation", "select count(*) from nation");
    Job job = new Job(jobconf);
    jobconf.set("mapred.input.dir", "output1");
    jobconf.set("mapred.output.dir", "output");
   
        jobconf.setMapperClass(InfiniDoopMapper.class);
        jobconf.setNumMapTasks(1);
View Full Code Here

       
        InfiniDBInputFormat.setInput(jobconf, InfiniDoopRecord.class, "nation",
           null,  "n_nationkey", fields);
       
        //InfiniDBInputFormat.setInput(jobconf, InfiniDBRecord.class, "select * from nation", "select count(*) from nation");
    Job job = new Job(jobconf);
    jobconf.set("mapred.output.dir", "output1");
        jobconf.setMapperClass(InfiniDoopInputMapper.class);
        jobconf.setNumMapTasks(4);
        jobconf.setNumReduceTasks(8);
        JobClient client = new JobClient();
View Full Code Here

       
        //InfiniDBInputFormat.setInput(jobconf, InfiniDoopRecord.class, "nation",
     //      null,  "n_nationkey", fields);
        InfiniDBOutputFormat.setOutput(jobconf, "zz1", outFields);
        //InfiniDBInputFormat.setInput(jobconf, InfiniDBRecord.class, "select * from nation", "select count(*) from nation");
    Job job = new Job(jobconf);
    jobconf.set("mapred.input.dir", "output1");
    jobconf.set("mapred.output.dir", "output");
   
        jobconf.setMapperClass(InfiniDoopMapper.class);
        jobconf.setNumMapTasks(1);
View Full Code Here

    } catch (HIHOException e1) {
      e1.printStackTrace();
      throw new IOException(e1);
    }

    Job job = new Job(conf);
    job.setJobName("OracleLoading");
    job.setMapperClass(OracleLoadMapper.class);
    job.setJarByClass(ExportToOracleDb.class);
    job.getConfiguration().setInt(MRJobConfig.NUM_MAPS,
        conf.getInt(HIHOConf.NUMBER_MAPPERS, 1));

    try {
      // we first create the external table definition
      String query = conf.get(HIHOConf.EXTERNAL_TABLE_DML);
      // create table if user has specified
      if (query != null) {
        this.runQuery(query, conf);
      }
    } catch (HIHOException e1) {

      e1.printStackTrace();
    }

    // verify required properties are loaded

    job.setNumReduceTasks(0);
    job.setInputFormatClass(FileStreamInputFormat.class);
    FileStreamInputFormat.addInputPath(job, new Path(inputPath));
    job.setMapOutputKeyClass(NullWritable.class);
    job.setMapOutputValueClass(NullWritable.class);
    // job.setJarByClass(com.mysql.jdbc.Driver.class);
    job.setOutputFormatClass(NullOutputFormat.class);

    int ret = 0;
    try {

      ret = job.waitForCompletion(true) ? 0 : 1;
    } catch (Exception e) {
      e.printStackTrace();
    }
    // run alter table query and add locations
    try {
View Full Code Here

  @Override
  public int run(String[] args) throws Exception {

    Configuration conf = getConf();
    Job job = new Job(conf);
    job.setJobName("UUID_Job");
    job.setMapperClass(UuidMapper.class);
    job.setJarByClass(UuidJob.class);
    for (Entry<String, String> entry : conf) {
      logger.debug("key, value " + entry.getKey() + "="
          + entry.getValue());
    }
    job.getConfiguration().setInt(HIHOConf.NUMBER_MAPPERS, 5);
    job.setNumReduceTasks(0);
    job.setInputFormatClass(DelimitedTextInputFormat.class);
    DelimitedTextInputFormat.addInputPath(job, new Path(args[0]));
    DelimitedTextInputFormat.setProperties(job, ",", 1);
    job.setMapOutputKeyClass(NullWritable.class);
    job.setMapOutputValueClass(Text.class);
    job.setOutputKeyClass(NullWritable.class);
    job.setOutputValueClass(Text.class);
    job.setOutputFormatClass(NoKeyOnlyValueOutputFormat.class);
    NoKeyOnlyValueOutputFormat.setOutputPath(job, new Path("output"));
    int ret = 0;
    try {
      ret = job.waitForCompletion(true) ? 0 : 1;
    } catch (Exception e) {
      e.printStackTrace();
    }
    return ret;
  }
View Full Code Here

TOP

Related Classes of org.apache.hadoop.mapreduce.Job

Copyright © 2018 www.massapicom. All rights reserved.
All source code are property of their respective owners. Java is a trademark of Sun Microsystems, Inc and owned by ORACLE Inc. Contact coftware#gmail.com.