/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package opennlp.tools.tokenize;
import org.junit.Assert;
import org.junit.Test;
/**
* Tests for the {@link WhitespaceTokenizer} class.
*/
public class WhitespaceTokenizerTest {
@Test
public void testOneToken() {
Assert.assertEquals("one", WhitespaceTokenizer.INSTANCE.tokenize("one")[0]);
Assert.assertEquals("one", WhitespaceTokenizer.INSTANCE.tokenize(" one")[0]);
Assert.assertEquals("one", WhitespaceTokenizer.INSTANCE.tokenize("one ")[0]);
}
/**
* Tests if it can tokenize whitespace separated tokens.
*/
@Test
public void testWhitespaceTokenization() {
String text = "a b c d e f ";
String[] tokenizedText = WhitespaceTokenizer.INSTANCE.tokenize(text);
Assert.assertTrue("a".equals(tokenizedText[0]));
Assert.assertTrue("b".equals(tokenizedText[1]));
Assert.assertTrue("c".equals(tokenizedText[2]));
Assert.assertTrue("d".equals(tokenizedText[3]));
Assert.assertTrue("e".equals(tokenizedText[4]));
Assert.assertTrue("f".equals(tokenizedText[5]));
Assert.assertTrue(tokenizedText.length == 6);
}
@Test
public void testTokenizationOfStringWithoutTokens() {
Assert.assertEquals(0, WhitespaceTokenizer.INSTANCE.tokenize("").length); // empty
Assert.assertEquals(0, WhitespaceTokenizer.INSTANCE.tokenize(" ").length); // space
Assert.assertEquals(0, WhitespaceTokenizer.INSTANCE.tokenize(" ").length); // tab
Assert.assertEquals(0, WhitespaceTokenizer.INSTANCE.tokenize(" ").length);
}
}