/**
* Copyright 2011-2017 Asakusa Framework Team.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.asakusafw.dmdl.directio.hive.parquet;
import static org.hamcrest.Matchers.*;
import static org.junit.Assert.*;
import org.junit.Test;
import com.asakusafw.directio.hive.serde.DataModelMapping.ExceptionHandlingStrategy;
import com.asakusafw.directio.hive.serde.DataModelMapping.FieldMappingStrategy;
import com.asakusafw.dmdl.directio.hive.common.GeneratorTesterRoot;
import com.asakusafw.dmdl.directio.hive.common.HiveDataModelTrait;
import com.asakusafw.dmdl.semantics.ModelDeclaration;
import parquet.column.ParquetProperties.WriterVersion;
import parquet.hadoop.metadata.CompressionCodecName;
/**
* Test for {@link ParquetFileDriver}.
*/
public class ParquetFileDriverTest extends GeneratorTesterRoot {
/**
* simple case.
*/
@Test
public void simple() {
ModelDeclaration model = analyze(new String[] {
"@directio.hive.parquet",
"model = { simple : INT; };"
}).findModelDeclaration("model");
assertThat(model.getTrait(HiveDataModelTrait.class), is(notNullValue()));
ParquetFileTrait trait = model.getTrait(ParquetFileTrait.class);
assertThat(trait, is(notNullValue()));
assertThat(trait.getOriginalAst(), is(notNullValue()));
assertThat(trait.getTableName(), is(nullValue()));
assertThat(trait.configuration().getFieldMappingStrategy(), is(nullValue()));
assertThat(trait.configuration().getOnMissingSource(), is(nullValue()));
assertThat(trait.configuration().getOnMissingTarget(), is(nullValue()));
assertThat(trait.configuration().getOnIncompatibleType(), is(nullValue()));
assertThat(trait.configuration().getCompressionCodecName(), is(nullValue()));
assertThat(trait.configuration().getBlockSize(), is(nullValue()));
assertThat(trait.configuration().getDataPageSize(), is(nullValue()));
assertThat(trait.configuration().getDictionaryPageSize(), is(nullValue()));
assertThat(trait.configuration().getEnableDictionary(), is(nullValue()));
assertThat(trait.configuration().getEnableValidation(), is(nullValue()));
assertThat(trait.configuration().getWriterVersion(), is(nullValue()));
}
/**
* explicitly defines all attributes.
*/
@Test
public void attributes() {
ModelDeclaration model = analyze(new String[] {
"@directio.hive.parquet(",
" table_name = 'attributes_test',",
" field_mapping = 'name',",
" on_missing_source = 'ignore',",
" on_missing_target = 'fail',",
" on_incompatible_type = 'logging',",
" format_version = 'v2',",
" compression = 'uncompressed',",
" block_size = 100000000,",
" data_page_size = 1000001,",
" dictionary_page_size = 1000002,",
" enable_dictionary = FALSE,",
" enable_validation = TRUE,",
")",
"model = { simple : INT; };"
}).findModelDeclaration("model");
assertThat(model.getTrait(HiveDataModelTrait.class), is(notNullValue()));
ParquetFileTrait trait = model.getTrait(ParquetFileTrait.class);
assertThat(trait, is(notNullValue()));
assertThat(trait.getOriginalAst(), is(notNullValue()));
assertThat(trait.getTableName(), is("attributes_test"));
assertThat(trait.configuration().getFieldMappingStrategy(), is(FieldMappingStrategy.NAME));
assertThat(trait.configuration().getOnMissingSource(), is(ExceptionHandlingStrategy.IGNORE));
assertThat(trait.configuration().getOnMissingTarget(), is(ExceptionHandlingStrategy.FAIL));
assertThat(trait.configuration().getOnIncompatibleType(), is(ExceptionHandlingStrategy.LOGGING));
assertThat(trait.configuration().getCompressionCodecName(), is(CompressionCodecName.UNCOMPRESSED));
assertThat(trait.configuration().getBlockSize(), is(100000000));
assertThat(trait.configuration().getDataPageSize(), is(1000001));
assertThat(trait.configuration().getDictionaryPageSize(), is(1000002));
assertThat(trait.configuration().getEnableDictionary(), is(false));
assertThat(trait.configuration().getEnableValidation(), is(true));
assertThat(trait.configuration().getWriterVersion(), is(WriterVersion.PARQUET_2_0));
}
/**
* unsupported type but it is ignored.
*/
@Test
public void invalid_ignore() {
// ok.
analyze(new String[] {
"@directio.hive.parquet",
"model = {",
" value : INT;",
" @directio.hive.ignore",
" ignored : DATE;",
"};",
});
}
/**
* unknown format version.
*/
@Test
public void invalid_format_version() {
shouldSemanticError(new String[] {
"@directio.hive.parquet(",
" format_version = 'UNKNOWN',",
")",
"model = { simple : INT; };"
});
}
/**
* unknown compression type.
*/
@Test
public void invalid_compression() {
shouldSemanticError(new String[] {
"@directio.hive.parquet(",
" compression = 'UNKNOWN',",
")",
"model = { simple : INT; };"
});
}
}