/*
* Copyright 2012-2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License"). You may not use this file except in compliance with
* the License. A copy of the License is located at
*
* http://aws.amazon.com/apache2.0
*
* or in the "license" file accompanying this file. This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
* CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions
* and limitations under the License.
*/
package com.amazonaws.services.rekognition.model;
import java.io.Serializable;
import javax.annotation.Generated;
@Generated("com.amazonaws:aws-java-sdk-code-generator")
public class DetectFacesResult extends com.amazonaws.AmazonWebServiceResult<com.amazonaws.ResponseMetadata> implements Serializable, Cloneable {
/**
* <p>
* Details of each face found in the image.
* </p>
*/
private java.util.List<FaceDetail> faceDetails;
/**
* <p>
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the degrees of
* rotation. If your application is displaying the image, you can use this value to adjust the orientation.
* </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects orientation,
* performs face detection, and then returns the faces. That is, the bounding box coordinates in the response are
* based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* </note>
*/
private String orientationCorrection;
/**
* <p>
* Details of each face found in the image.
* </p>
*
* @return Details of each face found in the image.
*/
public java.util.List<FaceDetail> getFaceDetails() {
return faceDetails;
}
/**
* <p>
* Details of each face found in the image.
* </p>
*
* @param faceDetails
* Details of each face found in the image.
*/
public void setFaceDetails(java.util.Collection<FaceDetail> faceDetails) {
if (faceDetails == null) {
this.faceDetails = null;
return;
}
this.faceDetails = new java.util.ArrayList<FaceDetail>(faceDetails);
}
/**
* <p>
* Details of each face found in the image.
* </p>
* <p>
* <b>NOTE:</b> This method appends the values to the existing list (if any). Use
* {@link #setFaceDetails(java.util.Collection)} or {@link #withFaceDetails(java.util.Collection)} if you want to
* override the existing values.
* </p>
*
* @param faceDetails
* Details of each face found in the image.
* @return Returns a reference to this object so that method calls can be chained together.
*/
public DetectFacesResult withFaceDetails(FaceDetail... faceDetails) {
if (this.faceDetails == null) {
setFaceDetails(new java.util.ArrayList<FaceDetail>(faceDetails.length));
}
for (FaceDetail ele : faceDetails) {
this.faceDetails.add(ele);
}
return this;
}
/**
* <p>
* Details of each face found in the image.
* </p>
*
* @param faceDetails
* Details of each face found in the image.
* @return Returns a reference to this object so that method calls can be chained together.
*/
public DetectFacesResult withFaceDetails(java.util.Collection<FaceDetail> faceDetails) {
setFaceDetails(faceDetails);
return this;
}
/**
* <p>
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the degrees of
* rotation. If your application is displaying the image, you can use this value to adjust the orientation.
* </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects orientation,
* performs face detection, and then returns the faces. That is, the bounding box coordinates in the response are
* based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* </note>
*
* @param orientationCorrection
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the
* degrees of rotation. If your application is displaying the image, you can use this value to adjust the
* orientation. </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects
* orientation, performs face detection, and then returns the faces. That is, the bounding box coordinates in
* the response are based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* @see OrientationCorrection
*/
public void setOrientationCorrection(String orientationCorrection) {
this.orientationCorrection = orientationCorrection;
}
/**
* <p>
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the degrees of
* rotation. If your application is displaying the image, you can use this value to adjust the orientation.
* </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects orientation,
* performs face detection, and then returns the faces. That is, the bounding box coordinates in the response are
* based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* </note>
*
* @return The algorithm detects the image orientation. If it detects that the image was rotated, it returns the
* degrees of rotation. If your application is displaying the image, you can use this value to adjust the
* orientation. </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects
* orientation, performs face detection, and then returns the faces. That is, the bounding box coordinates
* in the response are based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* @see OrientationCorrection
*/
public String getOrientationCorrection() {
return this.orientationCorrection;
}
/**
* <p>
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the degrees of
* rotation. If your application is displaying the image, you can use this value to adjust the orientation.
* </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects orientation,
* performs face detection, and then returns the faces. That is, the bounding box coordinates in the response are
* based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* </note>
*
* @param orientationCorrection
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the
* degrees of rotation. If your application is displaying the image, you can use this value to adjust the
* orientation. </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects
* orientation, performs face detection, and then returns the faces. That is, the bounding box coordinates in
* the response are based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* @return Returns a reference to this object so that method calls can be chained together.
* @see OrientationCorrection
*/
public DetectFacesResult withOrientationCorrection(String orientationCorrection) {
setOrientationCorrection(orientationCorrection);
return this;
}
/**
* <p>
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the degrees of
* rotation. If your application is displaying the image, you can use this value to adjust the orientation.
* </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects orientation,
* performs face detection, and then returns the faces. That is, the bounding box coordinates in the response are
* based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* </note>
*
* @param orientationCorrection
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the
* degrees of rotation. If your application is displaying the image, you can use this value to adjust the
* orientation. </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects
* orientation, performs face detection, and then returns the faces. That is, the bounding box coordinates in
* the response are based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* @see OrientationCorrection
*/
public void setOrientationCorrection(OrientationCorrection orientationCorrection) {
this.orientationCorrection = orientationCorrection.toString();
}
/**
* <p>
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the degrees of
* rotation. If your application is displaying the image, you can use this value to adjust the orientation.
* </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects orientation,
* performs face detection, and then returns the faces. That is, the bounding box coordinates in the response are
* based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* </note>
*
* @param orientationCorrection
* The algorithm detects the image orientation. If it detects that the image was rotated, it returns the
* degrees of rotation. If your application is displaying the image, you can use this value to adjust the
* orientation. </p>
* <p>
* For example, if the service detects that the input image was rotated by 90 degrees, it corrects
* orientation, performs face detection, and then returns the faces. That is, the bounding box coordinates in
* the response are based on the corrected orientation.
* </p>
* <note>
* <p>
* If the source image Exif metadata populates the orientation field, Amazon Rekognition does not perform
* orientation correction and the value of OrientationCorrection will be nil.
* </p>
* @return Returns a reference to this object so that method calls can be chained together.
* @see OrientationCorrection
*/
public DetectFacesResult withOrientationCorrection(OrientationCorrection orientationCorrection) {
setOrientationCorrection(orientationCorrection);
return this;
}
/**
* Returns a string representation of this object; useful for testing and debugging.
*
* @return A string representation of this object.
*
* @see java.lang.Object#toString()
*/
@Override
public String toString() {
StringBuilder sb = new StringBuilder();
sb.append("{");
if (getFaceDetails() != null)
sb.append("FaceDetails: ").append(getFaceDetails()).append(",");
if (getOrientationCorrection() != null)
sb.append("OrientationCorrection: ").append(getOrientationCorrection());
sb.append("}");
return sb.toString();
}
@Override
public boolean equals(Object obj) {
if (this == obj)
return true;
if (obj == null)
return false;
if (obj instanceof DetectFacesResult == false)
return false;
DetectFacesResult other = (DetectFacesResult) obj;
if (other.getFaceDetails() == null ^ this.getFaceDetails() == null)
return false;
if (other.getFaceDetails() != null && other.getFaceDetails().equals(this.getFaceDetails()) == false)
return false;
if (other.getOrientationCorrection() == null ^ this.getOrientationCorrection() == null)
return false;
if (other.getOrientationCorrection() != null && other.getOrientationCorrection().equals(this.getOrientationCorrection()) == false)
return false;
return true;
}
@Override
public int hashCode() {
final int prime = 31;
int hashCode = 1;
hashCode = prime * hashCode + ((getFaceDetails() == null) ? 0 : getFaceDetails().hashCode());
hashCode = prime * hashCode + ((getOrientationCorrection() == null) ? 0 : getOrientationCorrection().hashCode());
return hashCode;
}
@Override
public DetectFacesResult clone() {
try {
return (DetectFacesResult) super.clone();
} catch (CloneNotSupportedException e) {
throw new IllegalStateException("Got a CloneNotSupportedException from Object.clone() " + "even though we're Cloneable!", e);
}
}
}