/*
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS HEADER.
*
* Copyright (c) 2014-2015 Oracle and/or its affiliates. All rights reserved.
*
* The contents of this file are subject to the terms of either the GNU
* General Public License Version 2 only ("GPL") or the Common Development
* and Distribution License("CDDL") (collectively, the "License"). You
* may not use this file except in compliance with the License. You can
* obtain a copy of the License at
* http://glassfish.java.net/public/CDDL+GPL_1_1.html
* or packager/legal/LICENSE.txt. See the License for the specific
* language governing permissions and limitations under the License.
*
* When distributing the software, include this License Header Notice in each
* file and include the License file at packager/legal/LICENSE.txt.
*
* GPL Classpath Exception:
* Oracle designates this particular file as subject to the "Classpath"
* exception as provided by Oracle in the GPL Version 2 section of the License
* file that accompanied this code.
*
* Modifications:
* If applicable, add the following below the License Header, with the fields
* enclosed by brackets [] replaced by your own identifying information:
* "Portions Copyright [year] [name of copyright owner]"
*
* Contributor(s):
* If you wish your version of this file to be governed by only the CDDL or
* only the GPL Version 2, indicate your decision by adding "[Contributor]
* elects to include this software in this distribution under the [CDDL or GPL
* Version 2] license." If you don't indicate a single choice of license, a
* recipient has the option to distribute your version of this file under
* either the CDDL, the GPL Version 2 or to extend the choice of license to
* its licensees as provided above. However, if you add GPL Version 2 code
* and therefore, elected the GPL Version 2 license, then the option applies
* only if the new code is made subject to such option by the copyright
* holder.
*/
package org.glassfish.jersey.tests.e2e.client.connector;
import java.io.IOException;
import java.io.InputStream;
import java.io.OutputStream;
import java.lang.annotation.Annotation;
import java.lang.reflect.Type;
import java.util.Arrays;
import java.util.List;
import java.util.concurrent.Future;
import javax.ws.rs.POST;
import javax.ws.rs.Path;
import javax.ws.rs.WebApplicationException;
import javax.ws.rs.client.Entity;
import javax.ws.rs.core.Application;
import javax.ws.rs.core.MediaType;
import javax.ws.rs.core.MultivaluedMap;
import javax.ws.rs.core.Response;
import javax.ws.rs.ext.MessageBodyWriter;
import org.glassfish.jersey.client.ClientConfig;
import org.glassfish.jersey.client.ClientProperties;
import org.glassfish.jersey.client.RequestEntityProcessing;
import org.glassfish.jersey.client.spi.ConnectorProvider;
import org.glassfish.jersey.grizzly.connector.GrizzlyConnectorProvider;
import org.glassfish.jersey.jdk.connector.JdkConnectorProvider;
import org.glassfish.jersey.server.ResourceConfig;
import org.glassfish.jersey.test.JerseyTest;
import org.junit.Test;
import org.junit.runner.RunWith;
import org.junit.runners.Parameterized;
import static org.hamcrest.CoreMatchers.equalTo;
import static org.hamcrest.MatcherAssert.assertThat;
/**
* Test to make sure huge entity gets chunk-encoded.
*
* @author Jakub Podlesak (jakub.podlesak at oracle.com)
*/
@RunWith(Parameterized.class)
public class HugeEntityTest extends JerseyTest {
private static final int BUFFER_LENGTH = 1024 * 1024; // 1M
private static final long HUGE_DATA_LENGTH = 20L * 1024L * 1024L * 1024L; // 20G seems sufficient
/**
* JERSEY-2337 reproducer. The resource is used to check the right amount of data
* is being received from the client and also gives us ability to check we receive
* correct data.
*/
@Path("/")
public static class ConsumerResource {
/**
* Return back the count of bytes received.
* This way, we should be able to consume a huge amount of data.
*/
@POST
@Path("size")
public String post(InputStream in) throws IOException {
long totalBytesRead = 0L;
byte[] buffer = new byte[BUFFER_LENGTH];
int read;
do {
read = in.read(buffer);
if (read > 0) {
totalBytesRead += read;
}
} while (read != -1);
return String.valueOf(totalBytesRead);
}
@POST
@Path("echo")
public String echo(String s) {
return s;
}
}
private final ConnectorProvider connectorProvider;
public HugeEntityTest(ConnectorProvider connectorProvider) {
this.connectorProvider = connectorProvider;
}
@Parameterized.Parameters
public static List<? extends ConnectorProvider> testData() {
return Arrays.asList(new GrizzlyConnectorProvider(), new JdkConnectorProvider());
}
@Override
protected Application configure() {
return new ResourceConfig(ConsumerResource.class);
}
@Override
protected void configureClient(ClientConfig config) {
config.register(TestEntityWriter.class);
config.property(ClientProperties.REQUEST_ENTITY_PROCESSING, RequestEntityProcessing.CHUNKED);
config.connectorProvider(connectorProvider);
}
public static class TestEntity {
final long size;
public TestEntity(long size) {
this.size = size;
}
}
/**
* Utility writer that generates that many zero bytes as given by the input entity size field.
*/
public static class TestEntityWriter implements MessageBodyWriter<TestEntity> {
@Override
public boolean isWriteable(Class<?> type, Type genericType, Annotation[] annotations, MediaType mediaType) {
return type == TestEntity.class;
}
@Override
public long getSize(TestEntity t, Class<?> type, Type genericType, Annotation[] annotations, MediaType mediaType) {
return -1; // no matter what we return here, the output will get chunk-encoded
}
@Override
public void writeTo(TestEntity t,
Class<?> type,
Type genericType,
Annotation[] annotations,
MediaType mediaType,
MultivaluedMap<String, Object> httpHeaders,
OutputStream entityStream) throws IOException, WebApplicationException {
final byte[] buffer = new byte[BUFFER_LENGTH];
final long bufferCount = t.size / BUFFER_LENGTH;
final int remainder = (int) (t.size % BUFFER_LENGTH);
for (long b = 0; b < bufferCount; b++) {
entityStream.write(buffer);
}
if (remainder > 0) {
entityStream.write(buffer, 0, remainder);
}
}
}
/**
* JERSEY-2337 reproducer. We are going to send huge amount of data over the wire.
* Should not the data have been chunk-encoded, we would easily run out of memory.
*
* @throws Exception in case of a test error.
*/
@Test
public void testPost() throws Exception {
Response response = target("/size").request()
.post(Entity.entity(new TestEntity(HUGE_DATA_LENGTH), MediaType.APPLICATION_OCTET_STREAM_TYPE));
String content = response.readEntity(String.class);
assertThat(Long.parseLong(content), equalTo(HUGE_DATA_LENGTH));
// just to check the right data have been transfered.
response = target("/echo").request().post(Entity.text("Hey Sync!"));
assertThat(response.readEntity(String.class), equalTo("Hey Sync!"));
}
/**
* JERSEY-2337 reproducer. We are going to send huge amount of data over the wire. This time in an async fashion.
* Should not the data have been chunk-encoded, we would easily run out of memory.
*
* @throws Exception in case of a test error.
*/
@Test
public void testAsyncPost() throws Exception {
Future<Response> response = target("/size").request().async()
.post(Entity.entity(new TestEntity(HUGE_DATA_LENGTH), MediaType.APPLICATION_OCTET_STREAM_TYPE));
final String content = response.get().readEntity(String.class);
assertThat(Long.parseLong(content), equalTo(HUGE_DATA_LENGTH));
// just to check the right data have been transfered.
response = target("/echo").request().async().post(Entity.text("Hey Async!"));
assertThat(response.get().readEntity(String.class), equalTo("Hey Async!"));
}
}