/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ package org.apache.sysml.runtime.io; import java.io.BufferedWriter; import java.io.IOException; import java.io.InputStream; import java.io.OutputStream; import java.io.OutputStreamWriter; import java.util.ArrayList; import java.util.Collections; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.io.IOUtils; import org.apache.hadoop.mapred.JobConf; import org.apache.sysml.conf.ConfigurationManager; import org.apache.sysml.runtime.DMLRuntimeException; import org.apache.sysml.runtime.matrix.CSVReblockMR; import org.apache.sysml.runtime.matrix.data.CSVFileFormatProperties; import org.apache.sysml.runtime.matrix.data.MatrixBlock; import org.apache.sysml.runtime.matrix.data.SparseBlock; import org.apache.sysml.runtime.util.MapReduceTool; public class WriterTextCSV extends MatrixWriter { //blocksize for string concatenation in order to prevent write OOM //(can be set to very large value to disable blocking) public static final int BLOCKSIZE_J = 32; //32 cells (typically ~512B, should be less than write buffer of 1KB) protected CSVFileFormatProperties _props = null; public WriterTextCSV( CSVFileFormatProperties props ) { _props = props; } @Override public final void writeMatrixToHDFS(MatrixBlock src, String fname, long rlen, long clen, int brlen, int bclen, long nnz) throws IOException, DMLRuntimeException { //validity check matrix dimensions if( src.getNumRows() != rlen || src.getNumColumns() != clen ) { throw new IOException("Matrix dimensions mismatch with metadata: "+src.getNumRows()+"x"+src.getNumColumns()+" vs "+rlen+"x"+clen+"."); } //prepare file access JobConf job = new JobConf(ConfigurationManager.getCachedJobConf()); FileSystem fs = FileSystem.get(job); Path path = new Path( fname ); //if the file already exists on HDFS, remove it. MapReduceTool.deleteFileIfExistOnHDFS( fname ); //core write (sequential/parallel) writeCSVMatrixToHDFS(path, job, fs, src, _props); IOUtilFunctions.deleteCrcFilesFromLocalFileSystem(fs, path); } @Override public final void writeEmptyMatrixToHDFS(String fname, long rlen, long clen, int brlen, int bclen) throws IOException, DMLRuntimeException { JobConf job = new JobConf(ConfigurationManager.getCachedJobConf()); FileSystem fs = FileSystem.get(job); Path path = new Path( fname ); MatrixBlock src = new MatrixBlock((int)rlen, 1, true); writeCSVMatrixToHDFS(path, job, fs, src, _props); IOUtilFunctions.deleteCrcFilesFromLocalFileSystem(fs, path); } protected void writeCSVMatrixToHDFS(Path path, JobConf job, FileSystem fs, MatrixBlock src, CSVFileFormatProperties csvprops) throws IOException { //sequential write csv file writeCSVMatrixToFile(path, job, fs, src, 0, (int)src.getNumRows(), csvprops); } protected final void writeCSVMatrixToFile( Path path, JobConf job, FileSystem fs, MatrixBlock src, int rl, int ru, CSVFileFormatProperties props ) throws IOException { boolean sparse = src.isInSparseFormat(); int clen = src.getNumColumns(); //create buffered writer BufferedWriter br = new BufferedWriter(new OutputStreamWriter(fs.create(path,true))); try { //for obj reuse and preventing repeated buffer re-allocations StringBuilder sb = new StringBuilder(); props = (props==null)? new CSVFileFormatProperties() : props; String delim = props.getDelim(); boolean csvsparse = props.isSparse(); // Write header line, if needed if( props.hasHeader() && rl==0 ) { //write row chunk-wise to prevent OOM on large number of columns for( int bj=0; bj<clen; bj+=BLOCKSIZE_J ) { for( int j=bj; j < Math.min(clen,bj+BLOCKSIZE_J); j++) { sb.append("C"+ (j+1)); if ( j < clen-1 ) sb.append(delim); } br.write( sb.toString() ); sb.setLength(0); } sb.append('\n'); br.write( sb.toString() ); sb.setLength(0); } // Write data lines if( sparse ) //SPARSE { SparseBlock sblock = src.getSparseBlock(); for(int i=rl; i < ru; i++) { //write row chunk-wise to prevent OOM on large number of columns int prev_jix = -1; if( sblock!=null && i<sblock.numRows() && !sblock.isEmpty(i) ) { int pos = sblock.pos(i); int alen = sblock.size(i); int[] aix = sblock.indexes(i); double[] avals = sblock.values(i); for(int j=pos; j<pos+alen; j++) { int jix = aix[j]; // output empty fields, if needed for( int j2=prev_jix; j2<jix-1; j2++ ) { if( !csvsparse ) sb.append('0'); sb.append(delim); //flush buffered string if( j2%BLOCKSIZE_J==0 ){ br.write( sb.toString() ); sb.setLength(0); } } // output the value (non-zero) sb.append( avals[j] ); if( jix < clen-1) sb.append(delim); br.write( sb.toString() ); sb.setLength(0); //flush buffered string if( jix%BLOCKSIZE_J==0 ){ br.write( sb.toString() ); sb.setLength(0); } prev_jix = jix; } } // Output empty fields at the end of the row. // In case of an empty row, output (clen-1) empty fields for( int bj=prev_jix+1; bj<clen; bj+=BLOCKSIZE_J ) { for( int j = bj; j < Math.min(clen,bj+BLOCKSIZE_J); j++) { if( !csvsparse ) sb.append('0'); if( j < clen-1 ) sb.append(delim); } br.write( sb.toString() ); sb.setLength(0); } sb.append('\n'); br.write( sb.toString() ); sb.setLength(0); } } else //DENSE { for( int i=rl; i<ru; i++ ) { //write row chunk-wise to prevent OOM on large number of columns for( int bj=0; bj<clen; bj+=BLOCKSIZE_J ) { for( int j=bj; j<Math.min(clen,bj+BLOCKSIZE_J); j++ ) { double lvalue = src.getValueDenseUnsafe(i, j); if( lvalue != 0 ) //for nnz sb.append(lvalue); else if( !csvsparse ) sb.append('0'); if( j != clen-1 ) sb.append(delim); } br.write( sb.toString() ); sb.setLength(0); } sb.append('\n'); br.write( sb.toString() ); //same as append sb.setLength(0); } } } finally { IOUtilFunctions.closeSilently(br); } } @SuppressWarnings("unchecked") public final void addHeaderToCSV(String srcFileName, String destFileName, long rlen, long clen) throws IOException { Configuration conf = new Configuration(ConfigurationManager.getCachedJobConf()); Path srcFilePath = new Path(srcFileName); Path destFilePath = new Path(destFileName); FileSystem hdfs = FileSystem.get(conf); if ( !_props.hasHeader() ) { // simply move srcFile to destFile /* * TODO: Remove this roundabout way! * For example: destFilePath = /user/biadmin/csv/temp/out/file.csv * & the only path that exists already on HDFS is /user/biadmin/csv/. * In this case: the directory structure /user/biadmin/csv/temp/out must be created. * Simple hdfs.rename() does not seem to create this directory structure. */ // delete the destination file, if exists already //boolean ret1 = hdfs.delete(destFilePath, true); // Create /user/biadmin/csv/temp/out/file.csv so that ..../temp/out/ is created. //boolean ret2 = hdfs.createNewFile(destFilePath); // delete the file "file.csv" but preserve the directory structure /user/biadmin/csv/temp/out/ //boolean ret3 = hdfs.delete(destFilePath, true); // finally, move the data to destFilePath = /user/biadmin/csv/temp/out/file.csv //boolean ret4 = hdfs.rename(srcFilePath, destFilePath); //System.out.println("Return values = del:" + ret1 + ", createNew:" + ret2 + ", del:" + ret3 + ", rename:" + ret4); return; } // construct the header line StringBuilder sb = new StringBuilder(); for (int i = 0; i < clen; i++) { sb.append("C" + (i + 1)); if (i < clen - 1) sb.append(_props.getDelim()); } sb.append('\n'); if (hdfs.isDirectory(srcFilePath)) { // compute sorted order among part files ArrayList<Path> files=new ArrayList<Path>(); for(FileStatus stat: hdfs.listStatus(srcFilePath, CSVReblockMR.hiddenFileFilter)) files.add(stat.getPath()); Collections.sort(files); // first part file path Path firstpart = files.get(0); // create a temp file, and add header and contents of first part Path tmp = new Path(firstpart.toString() + ".tmp"); OutputStream out = hdfs.create(tmp, true); out.write(sb.toString().getBytes()); sb.setLength(0); // copy rest of the data from firstpart InputStream in = null; try { in = hdfs.open(firstpart); IOUtils.copyBytes(in, out, conf, true); } finally { IOUtilFunctions.closeSilently(in); IOUtilFunctions.closeSilently(out); } // rename tmp to firstpart hdfs.delete(firstpart, true); hdfs.rename(tmp, firstpart); // rename srcfile to destFile hdfs.delete(destFilePath, true); hdfs.createNewFile(destFilePath); // force the creation of directory structure hdfs.delete(destFilePath, true); // delete the file, but preserve the directory structure hdfs.rename(srcFilePath, destFilePath); // move the data } else if (hdfs.isFile(srcFilePath)) { // create destination file OutputStream out = hdfs.create(destFilePath, true); // write header out.write(sb.toString().getBytes()); sb.setLength(0); // copy the data from srcFile InputStream in = null; try { in = hdfs.open(srcFilePath); IOUtils.copyBytes(in, out, conf, true); } finally { IOUtilFunctions.closeSilently(in); IOUtilFunctions.closeSilently(out); } } else { throw new IOException(srcFilePath.toString() + ": No such file or directory"); } } }