/* * Copyright 2013 Cloudera Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.kitesdk.cli.commands; import com.beust.jcommander.Parameter; import com.beust.jcommander.Parameters; import com.google.common.base.Preconditions; import com.google.common.collect.Lists; import java.io.IOException; import java.nio.charset.Charset; import java.util.Iterator; import java.util.List; import java.util.UUID; import org.apache.avro.Schema; import org.apache.crunch.DoFn; import org.apache.crunch.PipelineResult; import org.apache.crunch.Target; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.kitesdk.compat.DynConstructors; import org.kitesdk.data.DatasetDescriptor; import org.kitesdk.data.DatasetException; import org.kitesdk.data.View; import org.kitesdk.data.spi.ColumnMappingParser; import org.kitesdk.data.spi.PartitionStrategyParser; import org.kitesdk.data.spi.filesystem.CSVProperties; import org.kitesdk.data.spi.filesystem.FileSystemDataset; import org.kitesdk.data.spi.filesystem.TemporaryFileSystemDatasetRepository; import org.kitesdk.tools.CopyTask; import org.kitesdk.tools.TaskUtil; import org.kitesdk.tools.TransformTask; import org.slf4j.Logger; import static org.apache.avro.generic.GenericData.Record; @Parameters(commandDescription="Copy CSV records into a Dataset") public class CSVImportCommand extends BaseDatasetCommand { public CSVImportCommand(Logger console) { super(console); } @Parameter(description="<csv path> <dataset name>") List<String> targets; @Parameter(names="--delimiter", description="Delimiter character") String delimiter = ","; @Parameter(names="--escape", description="Escape character") String escape = "\\"; @Parameter(names="--quote", description="Quote character") String quote = "\""; @Parameter(names="--no-header", description="Don't use first line as CSV header") boolean noHeader = false; @Parameter(names="--skip-lines", description="Lines to skip before CSV start") int linesToSkip = 0; @Parameter(names="--charset", description="Character set name", hidden = true) String charsetName = Charset.defaultCharset().displayName(); @Parameter(names="--header", description="Line to use as a header. Must match the CSV settings.") String header; @Parameter(names={"--no-compaction"}, description="Copy to output directly, without compacting the data") boolean noCompaction = false; @Parameter(names={"--num-writers"}, description="The number of writer processes to use") int numWriters = -1; @Parameter(names={"--files-per-partition"}, description="The number of files per partition to create") int filesPerPartition = -1; @Parameter(names={"--transform"}, description="A transform DoFn class name") String transform = null; @Parameter(names="--jar", description="Add a jar to the runtime classpath") List<String> jars; @Parameter( names={"--overwrite"}, description="Remove any data already in the target view or dataset") boolean overwrite = false; @Override @SuppressWarnings("unchecked") public int run() throws IOException { Preconditions.checkArgument(targets != null && targets.size() == 2, "CSV path and target dataset name are required."); Path source = qualifiedPath(targets.get(0)); FileSystem sourceFS = source.getFileSystem(getConf()); Preconditions.checkArgument(sourceFS.exists(source), "CSV path does not exist: " + source); if (header != null) { // if a header is given on the command line, do assume one is in the file noHeader = true; } CSVProperties props = new CSVProperties.Builder() .delimiter(delimiter) .escape(escape) .quote(quote) .header(header) .hasHeader(!noHeader) .linesToSkip(linesToSkip) .charset(charsetName) .build(); String dataset = targets.get(1); View<Record> target = load(dataset, Record.class); Schema datasetSchema = target.getDataset().getDescriptor().getSchema(); // TODO: replace this with a temporary Dataset from a FS repo // TODO: CDK-92: always use GenericRecord? DatasetDescriptor csvDescriptor = new DatasetDescriptor.Builder() .location(source.toUri()) .schema(ColumnMappingParser.removeEmbeddedMapping( PartitionStrategyParser.removeEmbeddedStrategy(datasetSchema))) .format("csv") .build(); csvDescriptor = props.addToDescriptor(csvDescriptor); TemporaryFileSystemDatasetRepository repo = new TemporaryFileSystemDatasetRepository(getConf(), // ensure the same FS as the file source is used sourceFS.makeQualified(new Path("/tmp/" + UUID.randomUUID().toString())), target.getDataset().getNamespace(), UUID.randomUUID().toString()); try { FileSystemDataset<Record> csvDataset = (FileSystemDataset) repo.create("default", "csv", csvDescriptor); Iterator<Path> iter = csvDataset.pathIterator().iterator(); Preconditions.checkArgument(iter.hasNext(), "CSV path has no data files: " + source); TaskUtil.configure(getConf()).addJars(jars); TransformTask task; if (transform != null) { DoFn<Record, Record> transformFn; try { DynConstructors.Ctor<DoFn<Record, Record>> ctor = new DynConstructors.Builder(DoFn.class) .loader(loaderForJars(jars)) .impl(transform) .buildChecked(); transformFn = ctor.newInstance(); } catch (NoSuchMethodException e) { throw new DatasetException( "Cannot find no-arg constructor for class: " + transform, e); } task = new TransformTask<Record, Record>( csvDataset, target, transformFn); } else { task = new CopyTask<Record>(csvDataset, target); } task.setConf(getConf()); if (noCompaction) { task.noCompaction(); } if (numWriters >= 0) { task.setNumWriters(numWriters); } if (filesPerPartition > 0) { task.setFilesPerPartition(filesPerPartition); } if (overwrite) { task.setWriteMode(Target.WriteMode.OVERWRITE); } PipelineResult result = task.run(); if (result.succeeded()) { long count = task.getCount(); if (count > 0) { console.info("Added {} records to \"{}\"", count, dataset); } return 0; } else { return 1; } } finally { // clean up the temporary repository repo.delete(); } } @Override public List<String> getExamples() { return Lists.newArrayList( "# Copy the records from sample.csv to dataset \"sample\"", "csv-import path/to/sample.csv sample", "# Copy the records from sample.csv to a dataset URI", "csv-import path/to/sample.csv dataset:hdfs:/user/me/datasets/sample", "# Copy the records from an HDFS directory to \"sample\"", "csv-import hdfs:/data/path/samples/ sample" ); } }