/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.ambari.view.hive2.utils;
import org.apache.ambari.view.utils.hdfs.HdfsApi;
import org.apache.hadoop.fs.FSDataInputStream;
import java.io.IOException;
import java.nio.charset.Charset;
import java.util.Arrays;
import static java.lang.Math.ceil;
/**
* Pagination for HDFS file implementation
*/
public class FilePaginator {
public static int MB = 1024*1024;
public static int PAGE_SIZE = 1*MB;
private String filePath;
private HdfsApi hdfsApi;
/**
* Constructor
* @param filePath Path to file on HDFS
* @param hdfsApi hdfs api
*/
public FilePaginator(String filePath, HdfsApi hdfsApi) {
this.filePath = filePath;
this.hdfsApi = hdfsApi;
}
/**
* Set page size
* @param PAGE_SIZE size
*/
public static void setPageSize(int PAGE_SIZE) {
FilePaginator.PAGE_SIZE = PAGE_SIZE;
}
/**
* Get page count
* @return page count
* @throws java.io.IOException
* @throws InterruptedException
*/
public long pageCount() throws IOException, InterruptedException {
return (long)
ceil( hdfsApi.getFileStatus(filePath).getLen() / ((double)PAGE_SIZE) );
}
/**
* Read one page of size PAGE_SIZE
* @param page page index
* @return data in UTF-8
* @throws java.io.IOException
* @throws InterruptedException
*/
public String readPage(long page) throws IOException, InterruptedException {
FSDataInputStream stream = hdfsApi.open(filePath);
try {
stream.seek(page * PAGE_SIZE);
} catch (IOException e) {
throw new IllegalArgumentException("Page " + page + " does not exists");
}
byte[] buffer = new byte[PAGE_SIZE];
int readCount = 0;
int read = 0;
while(read < PAGE_SIZE) {
try {
readCount = stream.read(buffer, read, PAGE_SIZE-read);
} catch (IOException e) {
stream.close();
throw e;
}
if (readCount == -1)
break;
read += readCount;
}
if (read != 0) {
byte[] readData = Arrays.copyOfRange(buffer, 0, read);
return new String(readData, Charset.forName("UTF-8"));
} else {
if (page == 0) {
return "";
}
throw new IllegalArgumentException("Page " + page + " does not exists");
}
}
public String readFull(long sizeLimit) throws IOException, InterruptedException {
StringBuilder builder = new StringBuilder();
int i = 0;
while (true) {
try {
builder.append(readPage(i++));
} catch (IllegalArgumentException ex) {
break;
}
if (sizeLimit != -1 && (i+1)*PAGE_SIZE > sizeLimit)
break;
}
return builder.toString();
}
public String readFull() throws IOException, InterruptedException {
return readFull(-1);
}
}