Package org.apache.sqoop.job.etl

Examples of org.apache.sqoop.job.etl.Exporter


    request.setOutputFormatClass(SqoopNullOutputFormat.class);
    request.setOutputKeyClass(Data.class);
    request.setOutputValueClass(NullWritable.class);

    Exporter exporter = (Exporter)request.getConnectorCallbacks();

    // Set up framework context
    MutableMapContext context = request.getFrameworkContext();
    context.setString(JobConstants.JOB_ETL_PARTITIONER, HdfsExportPartitioner.class.getName());
    context.setString(JobConstants.JOB_ETL_LOADER, exporter.getLoader().getName());
    context.setString(JobConstants.JOB_ETL_DESTROYER, exporter.getDestroyer().getName());

    // Extractor that will be able to read all supported file types
    context.setString(JobConstants.JOB_ETL_EXTRACTOR, HdfsExportExtractor.class.getName());
    context.setString(JobConstants.HADOOP_INPUTDIR, jobConf.input.inputDirectory);
View Full Code Here


    request.setOutputFormatClass(SqoopNullOutputFormat.class);
    request.setOutputKeyClass(Data.class);
    request.setOutputValueClass(NullWritable.class);

    Exporter exporter = (Exporter)request.getConnectorCallbacks();

    // Set up framework context
    MutableMapContext context = request.getFrameworkContext();
    context.setString(JobConstants.JOB_ETL_PARTITIONER, HdfsExportPartitioner.class.getName());
    context.setString(JobConstants.JOB_ETL_LOADER, exporter.getLoader().getName());
    context.setString(JobConstants.JOB_ETL_DESTROYER, exporter.getDestroyer().getName());

    // We should make one extractor that will be able to read all supported file types
    context.setString(JobConstants.JOB_ETL_EXTRACTOR, HdfsTextExportExtractor.class.getName());
    context.setString(JobConstants.HADOOP_INPUTDIR, jobConf.input.inputDirectory);
View Full Code Here

    request.setOutputFormatClass(SqoopNullOutputFormat.class);
    request.setOutputKeyClass(Data.class);
    request.setOutputValueClass(NullWritable.class);

    Exporter exporter = (Exporter)request.getConnectorCallbacks();

    // Set up framework context
    MutableMapContext context = request.getFrameworkContext();
    context.setString(JobConstants.JOB_ETL_PARTITIONER, HdfsExportPartitioner.class.getName());
    context.setString(JobConstants.JOB_ETL_LOADER, exporter.getLoader().getName());
    context.setString(JobConstants.JOB_ETL_DESTROYER, exporter.getDestroyer().getName());

    // Extractor that will be able to read all supported file types
    context.setString(JobConstants.JOB_ETL_EXTRACTOR, HdfsExportExtractor.class.getName());
    context.setString(JobConstants.HADOOP_INPUTDIR, jobConf.input.inputDirectory);
View Full Code Here

TOP

Related Classes of org.apache.sqoop.job.etl.Exporter

Copyright © 2018 www.massapicom. All rights reserved.
All source code are property of their respective owners. Java is a trademark of Sun Microsystems, Inc and owned by ORACLE Inc. Contact coftware#gmail.com.