blob: 3aa1c220ca338881f4b0739edc2386c91ec80170 [file] [log] [blame]
// Copyright 2019 Joe Drago. All rights reserved.
// SPDX-License-Identifier: BSD-2-Clause
#include "avif/avif.h"
#include "avifjpeg.h"
#include "avifpng.h"
#include "avifutil.h"
#include "y4m.h"
#include <assert.h>
#include <inttypes.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#if defined(_WIN32)
// for setmode()
#include <fcntl.h>
#include <io.h>
#endif
#define NEXTARG() \
if (((argIndex + 1) == argc) || (argv[argIndex + 1][0] == '-')) { \
fprintf(stderr, "%s requires an argument.", arg); \
goto cleanup; \
} \
arg = argv[++argIndex]
typedef struct avifInputFile
{
const char * filename;
uint64_t duration; // If 0, use the default duration
} avifInputFile;
static avifInputFile stdinFile;
typedef struct
{
int fileIndex;
avifImage * image;
uint32_t fileBitDepth;
avifBool fileIsRGB;
avifAppSourceTiming sourceTiming;
} avifInputCacheEntry;
typedef struct avifInput
{
avifInputFile * files;
int filesCount;
int fileIndex;
struct y4mFrameIterator * frameIter;
avifPixelFormat requestedFormat;
int requestedDepth;
avifBool useStdin;
avifBool cacheEnabled;
avifInputCacheEntry * cache;
int cacheCount;
} avifInput;
typedef struct
{
char ** keys;
char ** values;
int count;
} avifCodecSpecificOptions;
static void syntax(void)
{
printf("Syntax: avifenc [options] input.[jpg|jpeg|png|y4m] output.avif\n");
printf("Options:\n");
printf(" -h,--help : Show syntax help\n");
printf(" -V,--version : Show the version number\n");
printf(" -j,--jobs J : Number of jobs (worker threads, default: 1. Use \"all\" to use all available cores)\n");
printf(" -o,--output FILENAME : Instead of using the last filename given as output, use this filename\n");
printf(" -l,--lossless : Set all defaults to encode losslessly, and emit warnings when settings/input don't allow for it\n");
printf(" -d,--depth D : Output depth [8,10,12]. (JPEG/PNG only; For y4m or stdin, depth is retained)\n");
printf(" -y,--yuv FORMAT : Output format [default=auto, 444, 422, 420, 400]. Ignored for y4m or stdin (y4m format is retained)\n");
printf(" For JPEG, auto honors the JPEG's internal format, if possible. For all other cases, auto defaults to 444\n");
printf(" -p,--premultiply : Premultiply color by the alpha channel and signal this in the AVIF\n");
printf(" --sharpyuv : Use sharp RGB to YUV420 conversion (if supported). Ignored for y4m or if output is not 420.\n");
printf(" --stdin : Read y4m frames from stdin instead of files; no input filenames allowed, must set before offering output filename\n");
printf(" --cicp,--nclx P/T/M : Set CICP values (nclx colr box) (3 raw numbers, use -r to set range flag)\n");
printf(" P = color primaries\n");
printf(" T = transfer characteristics\n");
printf(" M = matrix coefficients\n");
printf(" (use 2 for any you wish to leave unspecified)\n");
printf(" -r,--range RANGE : YUV range [limited or l, full or f]. (JPEG/PNG only, default: full; For y4m or stdin, range is retained)\n");
printf(" -q,--qcolor Q : Set quality for color (%d-%d, where %d is lossless)\n",
AVIF_QUALITY_WORST,
AVIF_QUALITY_BEST,
AVIF_QUALITY_LOSSLESS);
printf(" --qalpha Q : Set quality for alpha (%d-%d, where %d is lossless)\n",
AVIF_QUALITY_WORST,
AVIF_QUALITY_BEST,
AVIF_QUALITY_LOSSLESS);
printf(" --tilerowslog2 R : Set log2 of number of tile rows (0-6, default: 0)\n");
printf(" --tilecolslog2 C : Set log2 of number of tile columns (0-6, default: 0)\n");
printf(" --autotiling : Set --tilerowslog2 and --tilecolslog2 automatically\n");
printf(" -g,--grid MxN : Encode a single-image grid AVIF with M cols & N rows. Either supply MxN identical W/H/D images, or a single\n");
printf(" image that can be evenly split into the MxN grid and follow AVIF grid image restrictions. The grid will adopt\n");
printf(" the color profile of the first image supplied.\n");
printf(" -s,--speed S : Encoder speed (%d-%d, slowest-fastest, 'default' or 'd' for codec internal defaults. default speed: 6)\n",
AVIF_SPEED_SLOWEST,
AVIF_SPEED_FASTEST);
printf(" -c,--codec C : AV1 codec to use (choose from versions list below)\n");
printf(" --exif FILENAME : Provide an Exif metadata payload to be associated with the primary item (implies --ignore-exif)\n");
printf(" --xmp FILENAME : Provide an XMP metadata payload to be associated with the primary item (implies --ignore-xmp)\n");
printf(" --icc FILENAME : Provide an ICC profile payload to be associated with the primary item (implies --ignore-icc)\n");
printf(" -a,--advanced KEY[=VALUE] : Pass an advanced, codec-specific key/value string pair directly to the codec. avifenc will warn on any not used by the codec.\n");
printf(" --duration D : Set all following frame durations (in timescales) to D; default 1. Can be set multiple times (before supplying each filename)\n");
printf(" --timescale,--fps V : Set the timescale to V. If all frames are 1 timescale in length, this is equivalent to frames per second (Default: 30)\n");
printf(" If neither duration nor timescale are set, avifenc will attempt to use the framerate stored in a y4m header, if present.\n");
printf(" -k,--keyframe INTERVAL : Set the forced keyframe interval (maximum frames between keyframes). Set to 0 to disable (default).\n");
printf(" --ignore-exif : If the input file contains embedded Exif metadata, ignore it (no-op if absent)\n");
printf(" --ignore-xmp : If the input file contains embedded XMP metadata, ignore it (no-op if absent)\n");
printf(" --ignore-icc : If the input file contains an embedded ICC profile, ignore it (no-op if absent)\n");
printf(" --pasp H,V : Add pasp property (aspect ratio). H=horizontal spacing, V=vertical spacing\n");
printf(" --crop CROPX,CROPY,CROPW,CROPH : Add clap property (clean aperture), but calculated from a crop rectangle\n");
printf(" --clap WN,WD,HN,HD,HON,HOD,VON,VOD: Add clap property (clean aperture). Width, Height, HOffset, VOffset (in num/denom pairs)\n");
printf(" --irot ANGLE : Add irot property (rotation). [0-3], makes (90 * ANGLE) degree rotation anti-clockwise\n");
printf(" --imir MODE : Add imir property (mirroring). 0=top-to-bottom, 1=left-to-right\n");
printf(" --repetition-count N or infinite : Number of times an animated image sequence will be repeated. Use 'infinite' for infinite repetitions (Default: infinite)\n");
printf(" --min QP : Set min quantizer for color (%d-%d, where %d is lossless)\n",
AVIF_QUANTIZER_BEST_QUALITY,
AVIF_QUANTIZER_WORST_QUALITY,
AVIF_QUANTIZER_LOSSLESS);
printf(" --max QP : Set max quantizer for color (%d-%d, where %d is lossless)\n",
AVIF_QUANTIZER_BEST_QUALITY,
AVIF_QUANTIZER_WORST_QUALITY,
AVIF_QUANTIZER_LOSSLESS);
printf(" --minalpha QP : Set min quantizer for alpha (%d-%d, where %d is lossless)\n",
AVIF_QUANTIZER_BEST_QUALITY,
AVIF_QUANTIZER_WORST_QUALITY,
AVIF_QUANTIZER_LOSSLESS);
printf(" --maxalpha QP : Set max quantizer for alpha (%d-%d, where %d is lossless)\n",
AVIF_QUANTIZER_BEST_QUALITY,
AVIF_QUANTIZER_WORST_QUALITY,
AVIF_QUANTIZER_LOSSLESS);
printf(" --target-size S : Set target file size in bytes (up to 7 times slower)\n");
printf(" --progressive : EXPERIMENTAL: Encode a progressive image\n");
printf(" -- : Signals the end of options. Everything after this is interpreted as file names.\n");
printf("\n");
if (avifCodecName(AVIF_CODEC_CHOICE_AOM, 0)) {
printf("aom-specific advanced options:\n");
printf(" 1. <key>=<value> applies to both the color (YUV) planes and the alpha plane (if present).\n");
printf(" 2. color:<key>=<value> or c:<key>=<value> applies only to the color (YUV) planes.\n");
printf(" 3. alpha:<key>=<value> or a:<key>=<value> applies only to the alpha plane (if present).\n");
printf(" Since the alpha plane is encoded as a monochrome image, the options that refer to the chroma planes,\n");
printf(" such as enable-chroma-deltaq=B, should not be used with the alpha plane. In addition, the film grain\n");
printf(" options are unlikely to make sense for the alpha plane.\n");
printf("\n");
printf(" When used with libaom 3.0.0 or later, any key-value pairs supported by the aom_codec_set_option() function\n");
printf(" can be used. When used with libaom 2.0.x or older, the following key-value pairs can be used:\n");
printf("\n");
printf(" aq-mode=M : Adaptive quantization mode (0: off (default), 1: variance, 2: complexity, 3: cyclic refresh)\n");
printf(" cq-level=Q : Constant/Constrained Quality level (0-63, end-usage must be set to cq or q)\n");
printf(" enable-chroma-deltaq=B : Enable delta quantization in chroma planes (0: disable (default), 1: enable)\n");
printf(" end-usage=MODE : Rate control mode (vbr, cbr, cq, or q)\n");
printf(" sharpness=S : Bias towards block sharpness in rate-distortion optimization of transform coefficients (0-7, default: 0)\n");
printf(" tune=METRIC : Tune the encoder for distortion metric (psnr or ssim, default: psnr)\n");
printf(" film-grain-test=TEST : Film grain test vectors (0: none (default), 1: test-1 2: test-2, ... 16: test-16)\n");
printf(" film-grain-table=FILENAME : Path to file containing film grain parameters\n");
printf("\n");
}
avifPrintVersions();
}
// This is *very* arbitrary, I just want to set people's expectations a bit
static const char * qualityString(int quality)
{
if (quality == AVIF_QUALITY_LOSSLESS) {
return "Lossless";
}
if (quality >= 80) {
return "High";
}
if (quality >= 50) {
return "Medium";
}
if (quality == AVIF_QUALITY_WORST) {
return "Worst";
}
return "Low";
}
static avifBool parseCICP(int cicp[3], const char * arg)
{
char buffer[128];
strncpy(buffer, arg, 127);
buffer[127] = 0;
int index = 0;
char * token = strtok(buffer, "/");
while (token != NULL) {
cicp[index] = atoi(token);
++index;
if (index >= 3) {
break;
}
token = strtok(NULL, "/");
}
if (index == 3) {
return AVIF_TRUE;
}
return AVIF_FALSE;
}
// Returns the count of uint32_t (up to 8)
static int parseU32List(uint32_t output[8], const char * arg)
{
char buffer[128];
strncpy(buffer, arg, 127);
buffer[127] = 0;
int index = 0;
char * token = strtok(buffer, ",x");
while (token != NULL) {
output[index] = (uint32_t)atoi(token);
++index;
if (index >= 8) {
break;
}
token = strtok(NULL, ",x");
}
return index;
}
static avifBool convertCropToClap(uint32_t srcW, uint32_t srcH, avifPixelFormat yuvFormat, uint32_t clapValues[8])
{
avifCleanApertureBox clap;
avifCropRect cropRect;
cropRect.x = clapValues[0];
cropRect.y = clapValues[1];
cropRect.width = clapValues[2];
cropRect.height = clapValues[3];
avifDiagnostics diag;
avifDiagnosticsClearError(&diag);
avifBool convertResult = avifCleanApertureBoxConvertCropRect(&clap, &cropRect, srcW, srcH, yuvFormat, &diag);
if (!convertResult) {
fprintf(stderr,
"ERROR: Impossible crop rect: imageSize:[%ux%u], pixelFormat:%s, cropRect:[%u,%u, %ux%u] - %s\n",
srcW,
srcH,
avifPixelFormatToString(yuvFormat),
cropRect.x,
cropRect.y,
cropRect.width,
cropRect.height,
diag.error);
return convertResult;
}
clapValues[0] = clap.widthN;
clapValues[1] = clap.widthD;
clapValues[2] = clap.heightN;
clapValues[3] = clap.heightD;
clapValues[4] = clap.horizOffN;
clapValues[5] = clap.horizOffD;
clapValues[6] = clap.vertOffN;
clapValues[7] = clap.vertOffD;
return AVIF_TRUE;
}
static avifBool avifInputAddCachedImage(avifInput * input)
{
avifImage * newImage = avifImageCreateEmpty();
if (!newImage) {
return AVIF_FALSE;
}
avifInputCacheEntry * newCachedImages = malloc((input->cacheCount + 1) * sizeof(*input->cache));
if (!newCachedImages) {
avifImageDestroy(newImage);
return AVIF_FALSE;
}
avifInputCacheEntry * oldCachedImages = input->cache;
input->cache = newCachedImages;
if (input->cacheCount) {
memcpy(input->cache, oldCachedImages, input->cacheCount * sizeof(*input->cache));
}
memset(&input->cache[input->cacheCount], 0, sizeof(input->cache[input->cacheCount]));
input->cache[input->cacheCount].fileIndex = input->fileIndex;
input->cache[input->cacheCount].image = newImage;
++input->cacheCount;
free(oldCachedImages);
return AVIF_TRUE;
}
static const avifInputFile * avifInputGetFile(const avifInput * input, int imageIndex)
{
if (imageIndex < input->cacheCount) {
return &input->files[input->cache[imageIndex].fileIndex];
}
if (input->useStdin) {
ungetc(fgetc(stdin), stdin); // Kick stdin to force EOF
if (feof(stdin)) {
return NULL;
}
return &stdinFile;
}
if (input->fileIndex >= input->filesCount) {
return NULL;
}
return &input->files[input->fileIndex];
}
static avifBool avifInputHasRemainingData(const avifInput * input, int imageIndex)
{
if (imageIndex < input->cacheCount) {
return AVIF_TRUE;
}
if (input->useStdin) {
return !feof(stdin);
}
return (input->fileIndex < input->filesCount);
}
static avifBool avifInputReadImage(avifInput * input,
int imageIndex,
avifBool ignoreICC,
avifBool ignoreExif,
avifBool ignoreXMP,
avifImage * image,
uint32_t * outDepth,
avifBool * sourceIsRGB,
avifAppSourceTiming * sourceTiming,
avifChromaDownsampling chromaDownsampling)
{
if (imageIndex < input->cacheCount) {
const avifInputCacheEntry * cached = &input->cache[imageIndex];
const avifCropRect rect = { 0, 0, cached->image->width, cached->image->height };
if (avifImageSetViewRect(image, cached->image, &rect) != AVIF_RESULT_OK) {
assert(AVIF_FALSE);
}
if (outDepth) {
*outDepth = cached->fileBitDepth;
}
if (sourceIsRGB) {
*sourceIsRGB = cached->fileIsRGB;
}
if (sourceTiming) {
*sourceTiming = cached->sourceTiming;
}
return AVIF_TRUE;
}
avifImage * dstImage = image;
uint32_t * dstDepth = outDepth;
avifBool * dstSourceIsRGB = sourceIsRGB;
avifAppSourceTiming * dstSourceTiming = sourceTiming;
if (input->cacheEnabled) {
if (!avifInputAddCachedImage(input)) {
fprintf(stderr, "ERROR: Out of memory");
return AVIF_FALSE;
}
assert(imageIndex + 1 == input->cacheCount);
dstImage = input->cache[imageIndex].image;
// Copy CICP, clap etc.
if (avifImageCopy(dstImage, image, /*planes=*/0) != AVIF_RESULT_OK) {
assert(AVIF_FALSE);
}
dstDepth = &input->cache[imageIndex].fileBitDepth;
dstSourceIsRGB = &input->cache[imageIndex].fileIsRGB;
dstSourceTiming = &input->cache[imageIndex].sourceTiming;
}
if (dstSourceTiming) {
// A source timing of all 0s is a sentinel value hinting that the value is unset / should be
// ignored. This is memset here as many of the paths in avifInputReadImage() do not set these
// values. See the declaration for avifAppSourceTiming for more information.
memset(dstSourceTiming, 0, sizeof(avifAppSourceTiming));
}
if (input->useStdin) {
if (feof(stdin)) {
return AVIF_FALSE;
}
if (!y4mRead(NULL, dstImage, dstSourceTiming, &input->frameIter)) {
fprintf(stderr, "ERROR: Cannot read y4m through standard input");
return AVIF_FALSE;
}
if (dstDepth) {
*dstDepth = dstImage->depth;
}
assert(dstImage->yuvFormat != AVIF_PIXEL_FORMAT_NONE);
if (dstSourceIsRGB) {
*dstSourceIsRGB = AVIF_FALSE;
}
} else {
if (input->fileIndex >= input->filesCount) {
return AVIF_FALSE;
}
const avifAppFileFormat inputFormat = avifReadImage(input->files[input->fileIndex].filename,
input->requestedFormat,
input->requestedDepth,
chromaDownsampling,
ignoreICC,
ignoreExif,
ignoreXMP,
dstImage,
dstDepth,
dstSourceTiming,
&input->frameIter);
if (inputFormat == AVIF_APP_FILE_FORMAT_UNKNOWN) {
fprintf(stderr, "Cannot determine input file format: %s\n", input->files[input->fileIndex].filename);
return AVIF_FALSE;
}
if (dstSourceIsRGB) {
*dstSourceIsRGB = (inputFormat != AVIF_APP_FILE_FORMAT_Y4M);
}
if (!input->frameIter) {
++input->fileIndex;
}
assert(dstImage->yuvFormat != AVIF_PIXEL_FORMAT_NONE);
}
if (input->cacheEnabled) {
// Reuse the just created cache entry.
assert(imageIndex < input->cacheCount);
return avifInputReadImage(input, imageIndex, ignoreICC, ignoreExif, ignoreXMP, image, outDepth, sourceIsRGB, sourceTiming, chromaDownsampling);
}
return AVIF_TRUE;
}
static avifBool readEntireFile(const char * filename, avifRWData * raw)
{
FILE * f = fopen(filename, "rb");
if (!f) {
return AVIF_FALSE;
}
fseek(f, 0, SEEK_END);
long pos = ftell(f);
if (pos <= 0) {
fclose(f);
return AVIF_FALSE;
}
size_t fileSize = (size_t)pos;
fseek(f, 0, SEEK_SET);
avifRWDataRealloc(raw, fileSize);
size_t bytesRead = fread(raw->data, 1, fileSize, f);
fclose(f);
if (bytesRead != fileSize) {
avifRWDataFree(raw);
return AVIF_FALSE;
}
return AVIF_TRUE;
}
// Returns NULL if a memory allocation failed.
static char * avifStrdup(const char * str)
{
size_t len = strlen(str);
char * dup = avifAlloc(len + 1);
if (!dup) {
return NULL;
}
memcpy(dup, str, len + 1);
return dup;
}
static avifBool avifCodecSpecificOptionsAdd(avifCodecSpecificOptions * options, const char * keyValue)
{
avifBool success = AVIF_FALSE;
char ** oldKeys = options->keys;
char ** oldValues = options->values;
options->keys = malloc((options->count + 1) * sizeof(*options->keys));
options->values = malloc((options->count + 1) * sizeof(*options->values));
if (!options->keys || !options->values) {
free(options->keys);
free(options->values);
options->keys = oldKeys;
options->values = oldValues;
return AVIF_FALSE;
}
if (options->count) {
memcpy(options->keys, oldKeys, options->count * sizeof(*options->keys));
memcpy(options->values, oldValues, options->count * sizeof(*options->values));
}
const char * value = strchr(keyValue, '=');
if (value) {
// Keep the parts on the left and on the right of the equal sign,
// but not the equal sign itself.
options->values[options->count] = avifStrdup(value + 1);
const size_t keyLength = strlen(keyValue) - strlen(value);
options->keys[options->count] = malloc(keyLength + 1);
if (!options->values[options->count] || !options->keys[options->count]) {
goto cleanup;
}
memcpy(options->keys[options->count], keyValue, keyLength);
options->keys[options->count][keyLength] = '\0';
} else {
// Pass in a non-NULL, empty string. Codecs can use the mere existence of a key as a boolean value.
options->values[options->count] = avifStrdup("");
options->keys[options->count] = avifStrdup(keyValue);
if (!options->values[options->count] || !options->keys[options->count]) {
goto cleanup;
}
}
success = AVIF_TRUE;
cleanup:
++options->count;
free(oldKeys);
free(oldValues);
return success;
}
// Returns the best cell size for a given horizontal or vertical dimension.
static avifBool avifGetBestCellSize(const char * dimensionStr, uint32_t numPixels, uint32_t numCells, avifBool isSubsampled, uint32_t * cellSize)
{
assert(numPixels);
assert(numCells);
// ISO/IEC 23008-12:2017, Section 6.6.2.3.1:
// The reconstructed image is formed by tiling the input images into a grid with a column width
// (potentially excluding the right-most column) equal to tile_width and a row height (potentially
// excluding the bottom-most row) equal to tile_height, without gap or overlap, and then
// trimming on the right and the bottom to the indicated output_width and output_height.
// The priority could be to use a cell size that is a multiple of 64, but there is not always a valid one,
// even though it is recommended by MIAF. Just use ceil(numPixels/numCells) for simplicity and to avoid
// as much padding in the right-most and bottom-most cells as possible.
// Use uint64_t computation to avoid a potential uint32_t overflow.
*cellSize = (uint32_t)(((uint64_t)numPixels + numCells - 1) / numCells);
// ISO/IEC 23000-22:2019, Section 7.3.11.4.2:
// - the tile_width shall be greater than or equal to 64, and should be a multiple of 64
// - the tile_height shall be greater than or equal to 64, and should be a multiple of 64
if (*cellSize < 64) {
*cellSize = 64;
if ((uint64_t)(numCells - 1) * *cellSize >= (uint64_t)numPixels) {
// Some cells would be entirely off-canvas.
fprintf(stderr, "ERROR: There are too many cells %s (%u) to have at least 64 pixels per cell.\n", dimensionStr, numCells);
return AVIF_FALSE;
}
}
// The maximum AV1 frame size is 65536 pixels inclusive.
if (*cellSize > 65536) {
fprintf(stderr, "ERROR: Cell size %u is bigger %s than the maximum AV1 frame size 65536.\n", *cellSize, dimensionStr);
return AVIF_FALSE;
}
// ISO/IEC 23000-22:2019, Section 7.3.11.4.2:
// - when the images are in the 4:2:2 chroma sampling format the horizontal tile offsets and widths,
// and the output width, shall be even numbers;
// - when the images are in the 4:2:0 chroma sampling format both the horizontal and vertical tile
// offsets and widths, and the output width and height, shall be even numbers.
if (isSubsampled && (*cellSize & 1)) {
++*cellSize;
if ((uint64_t)(numCells - 1) * *cellSize >= (uint64_t)numPixels) {
// Some cells would be entirely off-canvas.
fprintf(stderr, "ERROR: Odd cell size %u is forbidden on a %s subsampled image.\n", *cellSize - 1, dimensionStr);
return AVIF_FALSE;
}
}
// Each pixel is covered by exactly one cell, and each cell contains at least one pixel.
assert(((uint64_t)(numCells - 1) * *cellSize < (uint64_t)numPixels) && ((uint64_t)numCells * *cellSize >= (uint64_t)numPixels));
return AVIF_TRUE;
}
static avifBool avifImageSplitGrid(const avifImage * gridSplitImage, uint32_t gridCols, uint32_t gridRows, avifImage ** gridCells)
{
uint32_t cellWidth, cellHeight;
avifPixelFormatInfo formatInfo;
avifGetPixelFormatInfo(gridSplitImage->yuvFormat, &formatInfo);
const avifBool isSubsampledX = !formatInfo.monochrome && formatInfo.chromaShiftX;
const avifBool isSubsampledY = !formatInfo.monochrome && formatInfo.chromaShiftY;
if (!avifGetBestCellSize("horizontally", gridSplitImage->width, gridCols, isSubsampledX, &cellWidth) ||
!avifGetBestCellSize("vertically", gridSplitImage->height, gridRows, isSubsampledY, &cellHeight)) {
return AVIF_FALSE;
}
for (uint32_t gridY = 0; gridY < gridRows; ++gridY) {
for (uint32_t gridX = 0; gridX < gridCols; ++gridX) {
uint32_t gridIndex = gridX + (gridY * gridCols);
avifImage * cellImage = avifImageCreateEmpty();
if (!cellImage) {
fprintf(stderr, "ERROR: Cell creation failed: out of memory\n");
return AVIF_FALSE;
}
gridCells[gridIndex] = cellImage;
avifCropRect cellRect = { gridX * cellWidth, gridY * cellHeight, cellWidth, cellHeight };
if (cellRect.x + cellRect.width > gridSplitImage->width) {
cellRect.width = gridSplitImage->width - cellRect.x;
}
if (cellRect.y + cellRect.height > gridSplitImage->height) {
cellRect.height = gridSplitImage->height - cellRect.y;
}
const avifResult copyResult = avifImageSetViewRect(cellImage, gridSplitImage, &cellRect);
if (copyResult != AVIF_RESULT_OK) {
fprintf(stderr, "ERROR: Cell creation failed: %s\n", avifResultToString(copyResult));
return AVIF_FALSE;
}
}
}
return AVIF_TRUE;
}
typedef struct
{
avifCodecChoice codecChoice;
int jobs;
int quality;
avifBool qualityIsConstrained; // true if quality explicitly set by the user
int qualityAlpha;
avifBool qualityAlphaIsConstrained; // true if qualityAlpha explicitly set by the user
int minQuantizer;
int maxQuantizer;
int minQuantizerAlpha;
int maxQuantizerAlpha;
int targetSize;
int tileRowsLog2;
int tileColsLog2;
avifBool autoTiling;
avifBool progressive;
int speed;
int paspCount;
uint32_t paspValues[8]; // only the first two are used
int clapCount;
uint32_t clapValues[8];
int gridDimsCount;
uint32_t gridDims[8]; // only the first two are used
int repetitionCount;
int keyframeInterval;
avifBool ignoreExif;
avifBool ignoreXMP;
avifBool ignoreICC;
// This holds the output timing for image sequences. The timescale member in this struct will
// become the timescale set on avifEncoder, and the duration member will be the default duration
// for any frame that doesn't have a specific duration set on the commandline. See the
// declaration of avifAppSourceTiming for more documentation.
avifAppSourceTiming outputTiming;
avifColorPrimaries colorPrimaries;
avifTransferCharacteristics transferCharacteristics;
avifMatrixCoefficients matrixCoefficients;
avifChromaDownsampling chromaDownsampling;
avifCodecSpecificOptions codecSpecificOptions;
} avifSettings;
static avifBool avifEncodeRestOfImageSequence(avifEncoder * encoder,
const avifSettings * settings,
avifInput * input,
int imageIndex,
const avifImage * firstImage)
{
avifBool success = AVIF_FALSE;
avifImage * nextImage = NULL;
const avifInputFile * nextFile;
while ((nextFile = avifInputGetFile(input, imageIndex)) != NULL) {
uint64_t nextDurationInTimescales = nextFile->duration ? nextFile->duration : settings->outputTiming.duration;
printf(" * Encoding frame %d [%" PRIu64 "/%" PRIu64 " ts]: %s\n",
imageIndex,
nextDurationInTimescales,
settings->outputTiming.timescale,
nextFile->filename);
if (nextImage) {
avifImageDestroy(nextImage);
}
nextImage = avifImageCreateEmpty();
if (!nextImage) {
fprintf(stderr, "ERROR: Out of memory\n");
goto cleanup;
}
nextImage->colorPrimaries = firstImage->colorPrimaries;
nextImage->transferCharacteristics = firstImage->transferCharacteristics;
nextImage->matrixCoefficients = firstImage->matrixCoefficients;
nextImage->yuvRange = firstImage->yuvRange;
nextImage->alphaPremultiplied = firstImage->alphaPremultiplied;
// Ignore ICC, Exif and XMP because only the metadata of the first frame is taken into
// account by the libavif API.
if (!avifInputReadImage(input,
imageIndex,
/*ignoreICC=*/AVIF_TRUE,
/*ignoreExif=*/AVIF_TRUE,
/*ignoreXMP=*/AVIF_TRUE,
nextImage,
/*outDepth=*/NULL,
/*sourceIsRGB=*/NULL,
/*sourceTiming=*/NULL,
settings->chromaDownsampling)) {
goto cleanup;
}
// Verify that this frame's properties matches the first frame's properties
if ((firstImage->width != nextImage->width) || (firstImage->height != nextImage->height)) {
fprintf(stderr,
"ERROR: Image sequence dimensions mismatch, [%ux%u] vs [%ux%u]: %s\n",
firstImage->width,
firstImage->height,
nextImage->width,
nextImage->height,
nextFile->filename);
goto cleanup;
}
if (firstImage->depth != nextImage->depth) {
fprintf(stderr,
"ERROR: Image sequence depth mismatch, [%u] vs [%u]: %s\n",
firstImage->depth,
nextImage->depth,
nextFile->filename);
goto cleanup;
}
if ((firstImage->colorPrimaries != nextImage->colorPrimaries) ||
(firstImage->transferCharacteristics != nextImage->transferCharacteristics) ||
(firstImage->matrixCoefficients != nextImage->matrixCoefficients)) {
fprintf(stderr,
"ERROR: Image sequence CICP mismatch, [%u/%u/%u] vs [%u/%u/%u]: %s\n",
firstImage->colorPrimaries,
firstImage->matrixCoefficients,
firstImage->transferCharacteristics,
nextImage->colorPrimaries,
nextImage->transferCharacteristics,
nextImage->matrixCoefficients,
nextFile->filename);
goto cleanup;
}
if (firstImage->yuvRange != nextImage->yuvRange) {
fprintf(stderr,
"ERROR: Image sequence range mismatch, [%s] vs [%s]: %s\n",
(firstImage->yuvRange == AVIF_RANGE_FULL) ? "Full" : "Limited",
(nextImage->yuvRange == AVIF_RANGE_FULL) ? "Full" : "Limited",
nextFile->filename);
goto cleanup;
}
const avifResult nextImageResult = avifEncoderAddImage(encoder, nextImage, nextDurationInTimescales, AVIF_ADD_IMAGE_FLAG_NONE);
if (nextImageResult != AVIF_RESULT_OK) {
fprintf(stderr, "ERROR: Failed to encode image: %s\n", avifResultToString(nextImageResult));
goto cleanup;
}
++imageIndex;
}
success = AVIF_TRUE;
cleanup:
if (nextImage) {
avifImageDestroy(nextImage);
}
return success;
}
static avifBool avifEncodeRestOfLayeredImage(avifEncoder * encoder, const avifSettings * settings, int layerIndex, const avifImage * firstImage)
{
avifBool success = AVIF_FALSE;
int layers = encoder->extraLayerCount + 1;
int qualityIncrement = (settings->quality - encoder->quality) / encoder->extraLayerCount;
int qualityAlphaIncrement = (settings->qualityAlpha - encoder->qualityAlpha) / encoder->extraLayerCount;
while (layerIndex < layers) {
encoder->quality += qualityIncrement;
encoder->qualityAlpha += qualityAlphaIncrement;
if (layerIndex == layers - 1) {
encoder->quality = settings->quality;
encoder->qualityAlpha = settings->qualityAlpha;
}
printf(" * Encoding layer %d: color quality [%d (%s)], alpha quality [%d (%s)]\n",
layerIndex,
encoder->quality,
qualityString(encoder->quality),
encoder->qualityAlpha,
qualityString(encoder->qualityAlpha));
const avifResult result = avifEncoderAddImage(encoder, firstImage, settings->outputTiming.duration, AVIF_ADD_IMAGE_FLAG_NONE);
if (result != AVIF_RESULT_OK) {
fprintf(stderr, "ERROR: Failed to encode image: %s\n", avifResultToString(result));
goto cleanup;
}
++layerIndex;
}
success = AVIF_TRUE;
cleanup:
return success;
}
static avifBool avifEncodeImagesFixedQuality(const avifSettings * settings,
avifInput * input,
const avifInputFile * firstFile,
const avifImage * firstImage,
const avifImage * const * gridCells,
avifRWData * encoded,
avifIOStats * ioStats)
{
avifBool success = AVIF_FALSE;
avifRWDataFree(encoded);
avifEncoder * encoder = avifEncoderCreate();
if (!encoder) {
fprintf(stderr, "ERROR: Out of memory\n");
goto cleanup;
}
char manualTilingStr[128];
snprintf(manualTilingStr, sizeof(manualTilingStr), "tileRowsLog2 [%d], tileColsLog2 [%d]", settings->tileRowsLog2, settings->tileColsLog2);
printf("Encoding with AV1 codec '%s' speed [%d], color quality [%d (%s)], alpha quality [%d (%s)], %s, %d worker thread(s), please wait...\n",
avifCodecName(settings->codecChoice, AVIF_CODEC_FLAG_CAN_ENCODE),
settings->speed,
settings->quality,
qualityString(settings->quality),
settings->qualityAlpha,
qualityString(settings->qualityAlpha),
settings->autoTiling ? "automatic tiling" : manualTilingStr,
settings->jobs);
encoder->maxThreads = settings->jobs;
encoder->quality = settings->quality;
encoder->qualityAlpha = settings->qualityAlpha;
encoder->minQuantizer = settings->minQuantizer;
encoder->maxQuantizer = settings->maxQuantizer;
encoder->minQuantizerAlpha = settings->minQuantizerAlpha;
encoder->maxQuantizerAlpha = settings->maxQuantizerAlpha;
encoder->tileRowsLog2 = settings->tileRowsLog2;
encoder->tileColsLog2 = settings->tileColsLog2;
encoder->autoTiling = settings->autoTiling;
encoder->codecChoice = settings->codecChoice;
encoder->speed = settings->speed;
encoder->timescale = settings->outputTiming.timescale;
encoder->keyframeInterval = settings->keyframeInterval;
encoder->repetitionCount = settings->repetitionCount;
if (settings->progressive) {
// If the color quality or alpha quality is less than 10, the main()
// function overrides --progressive and sets settings->progressive to
// false.
assert((settings->quality >= 10) && (settings->qualityAlpha >= 10));
encoder->extraLayerCount = 1;
// Encode the base layer with a very low quality to ensure a small encoded size.
encoder->quality = 2;
if (firstImage->alphaPlane && firstImage->alphaRowBytes) {
encoder->qualityAlpha = 2;
}
printf(" * Encoding layer %d: color quality [%d (%s)], alpha quality [%d (%s)]\n",
0,
encoder->quality,
qualityString(encoder->quality),
encoder->qualityAlpha,
qualityString(encoder->qualityAlpha));
}
for (int i = 0; i < settings->codecSpecificOptions.count; ++i) {
if (avifEncoderSetCodecSpecificOption(encoder, settings->codecSpecificOptions.keys[i], settings->codecSpecificOptions.values[i]) !=
AVIF_RESULT_OK) {
fprintf(stderr,
"ERROR: Failed to set codec specific option: %s = %s\n",
settings->codecSpecificOptions.keys[i],
settings->codecSpecificOptions.values[i]);
goto cleanup;
}
}
if (settings->gridDimsCount > 0) {
const avifResult addImageResult =
avifEncoderAddImageGrid(encoder, settings->gridDims[0], settings->gridDims[1], gridCells, AVIF_ADD_IMAGE_FLAG_SINGLE);
if (addImageResult != AVIF_RESULT_OK) {
fprintf(stderr, "ERROR: Failed to encode image grid: %s\n", avifResultToString(addImageResult));
goto cleanup;
}
} else {
int imageIndex = 1; // firstImage with imageIndex 0 is already available.
avifAddImageFlags addImageFlags = AVIF_ADD_IMAGE_FLAG_NONE;
if (!avifInputHasRemainingData(input, imageIndex) && !settings->progressive) {
addImageFlags |= AVIF_ADD_IMAGE_FLAG_SINGLE;
}
uint64_t firstDurationInTimescales = firstFile->duration ? firstFile->duration : settings->outputTiming.duration;
if (input->useStdin || (input->filesCount > 1)) {
printf(" * Encoding frame %d [%" PRIu64 "/%" PRIu64 " ts]: %s\n",
0,
firstDurationInTimescales,
settings->outputTiming.timescale,
firstFile->filename);
}
const avifResult addImageResult = avifEncoderAddImage(encoder, firstImage, firstDurationInTimescales, addImageFlags);
if (addImageResult != AVIF_RESULT_OK) {
fprintf(stderr, "ERROR: Failed to encode image: %s\n", avifResultToString(addImageResult));
goto cleanup;
}
if (settings->progressive) {
if (!avifEncodeRestOfLayeredImage(encoder, settings, imageIndex, firstImage)) {
goto cleanup;
}
} else {
// Not generating a single-image grid: Use all remaining input files as subsequent
// frames.
if (!avifEncodeRestOfImageSequence(encoder, settings, input, imageIndex, firstImage)) {
goto cleanup;
}
}
}
const avifResult finishResult = avifEncoderFinish(encoder, encoded);
if (finishResult != AVIF_RESULT_OK) {
fprintf(stderr, "ERROR: Failed to finish encoding: %s\n", avifResultToString(finishResult));
goto cleanup;
}
success = AVIF_TRUE;
memcpy(ioStats, &encoder->ioStats, sizeof(*ioStats));
cleanup:
if (encoder) {
if (!success) {
avifDumpDiagnostics(&encoder->diag);
}
avifEncoderDestroy(encoder);
}
return success;
}
#define INVALID_QUALITY (-1)
#define DEFAULT_QUALITY 60 // Maps to a quantizer (QP) of 25.
#define DEFAULT_QUALITY_ALPHA AVIF_QUALITY_LOSSLESS
static avifBool avifEncodeImages(avifSettings * settings,
avifInput * input,
const avifInputFile * firstFile,
const avifImage * firstImage,
const avifImage * const * gridCells,
avifRWData * encoded,
avifIOStats * ioStats)
{
if (settings->targetSize == -1) {
return avifEncodeImagesFixedQuality(settings, input, firstFile, firstImage, gridCells, encoded, ioStats);
}
if (settings->qualityIsConstrained && settings->qualityAlphaIsConstrained) {
fprintf(stderr, "ERROR: --target_size is used with constrained --qcolor and --qalpha\n");
return AVIF_FALSE;
}
printf("Starting a binary search to find the %s generating the encoded image size closest to %d bytes, please wait...\n",
settings->qualityAlphaIsConstrained ? "color quality"
: (settings->qualityIsConstrained ? "alpha quality" : "color and alpha qualities"),
settings->targetSize);
const size_t targetSize = (size_t)settings->targetSize;
// TODO(yguyon): Use quantizer instead of quality because quantizer range is smaller (faster binary search).
int closestQuality = INVALID_QUALITY;
avifRWData closestEncoded = { NULL, 0 };
size_t closestSizeDiff = 0;
avifIOStats closestIoStats = { 0, 0 };
int minQuality = AVIF_QUALITY_WORST; // inclusive
int maxQuality = AVIF_QUALITY_BEST; // inclusive
while (minQuality <= maxQuality) {
const int quality = (minQuality + maxQuality) / 2;
if (!settings->qualityIsConstrained) {
settings->quality = quality;
}
if (!settings->qualityAlphaIsConstrained) {
settings->qualityAlpha = quality;
}
if (!avifEncodeImagesFixedQuality(settings, input, firstFile, firstImage, gridCells, encoded, ioStats)) {
avifRWDataFree(&closestEncoded);
return AVIF_FALSE;
}
printf("Encoded image of size %" AVIF_FMT_ZU " bytes.\n", encoded->size);
if (encoded->size == targetSize) {
return AVIF_TRUE;
}
size_t sizeDiff;
if (encoded->size > targetSize) {
sizeDiff = encoded->size - targetSize;
maxQuality = quality - 1;
} else {
sizeDiff = targetSize - encoded->size;
minQuality = quality + 1;
}
if ((closestQuality == INVALID_QUALITY) || (sizeDiff < closestSizeDiff)) {
closestQuality = quality;
avifRWDataFree(&closestEncoded);
closestEncoded = *encoded;
encoded->size = 0;
encoded->data = NULL;
closestSizeDiff = sizeDiff;
closestIoStats = *ioStats;
}
}
if (!settings->qualityIsConstrained) {
settings->quality = closestQuality;
}
if (!settings->qualityAlphaIsConstrained) {
settings->qualityAlpha = closestQuality;
}
avifRWDataFree(encoded);
*encoded = closestEncoded;
*ioStats = closestIoStats;
printf("Kept the encoded image of size %" AVIF_FMT_ZU " bytes generated with color quality %d and alpha quality %d.\n",
encoded->size,
settings->quality,
settings->qualityAlpha);
return AVIF_TRUE;
}
int main(int argc, char * argv[])
{
if (argc < 2) {
syntax();
return 1;
}
const char * outputFilename = NULL;
avifInput input;
memset(&input, 0, sizeof(input));
input.files = malloc(sizeof(avifInputFile) * argc);
input.requestedFormat = AVIF_PIXEL_FORMAT_NONE; // AVIF_PIXEL_FORMAT_NONE is used as a sentinel for "auto"
// See here for the discussion on the semi-arbitrary defaults for speed/min/max:
// https://github.com/AOMediaCodec/libavif/issues/440
int returnCode = 0;
avifSettings settings;
memset(&settings, 0, sizeof(settings));
settings.codecChoice = AVIF_CODEC_CHOICE_AUTO;
settings.jobs = 1;
settings.quality = INVALID_QUALITY;
settings.qualityAlpha = INVALID_QUALITY;
settings.minQuantizer = -1;
settings.maxQuantizer = -1;
settings.minQuantizerAlpha = -1;
settings.maxQuantizerAlpha = -1;
settings.targetSize = -1;
settings.tileRowsLog2 = -1;
settings.tileColsLog2 = -1;
settings.autoTiling = AVIF_FALSE;
settings.progressive = AVIF_FALSE;
settings.speed = 6;
settings.repetitionCount = AVIF_REPETITION_COUNT_INFINITE;
settings.keyframeInterval = 0;
settings.ignoreExif = AVIF_FALSE;
settings.ignoreXMP = AVIF_FALSE;
settings.ignoreICC = AVIF_FALSE;
avifBool cropConversionRequired = AVIF_FALSE;
uint8_t irotAngle = 0xff; // sentinel value indicating "unused"
uint8_t imirMode = 0xff; // sentinel value indicating "unused"
avifRange requestedRange = AVIF_RANGE_FULL;
avifBool lossless = AVIF_FALSE;
avifImage * image = NULL;
avifRWData raw = AVIF_DATA_EMPTY;
avifRWData exifOverride = AVIF_DATA_EMPTY;
avifRWData xmpOverride = AVIF_DATA_EMPTY;
avifRWData iccOverride = AVIF_DATA_EMPTY;
avifBool cicpExplicitlySet = AVIF_FALSE;
avifBool premultiplyAlpha = AVIF_FALSE;
uint32_t gridCellCount = 0;
avifImage ** gridCells = NULL;
avifImage * gridSplitImage = NULL; // used for cleanup tracking
// By default, the color profile itself is unspecified, so CP/TC are set (to 2) accordingly.
// However, if the end-user doesn't specify any CICP, we will convert to YUV using BT601
// coefficients anyway (as MC:2 falls back to MC:5/6), so we might as well signal it explicitly.
// See: ISO/IEC 23000-22:2019 Amendment 2, or the comment in avifCalcYUVCoefficients()
settings.colorPrimaries = AVIF_COLOR_PRIMARIES_UNSPECIFIED;
settings.transferCharacteristics = AVIF_TRANSFER_CHARACTERISTICS_UNSPECIFIED;
settings.matrixCoefficients = AVIF_MATRIX_COEFFICIENTS_BT601;
settings.chromaDownsampling = AVIF_CHROMA_DOWNSAMPLING_AUTOMATIC;
int argIndex = 1;
while (argIndex < argc) {
const char * arg = argv[argIndex];
if (!strcmp(arg, "--")) {
// Stop parsing flags, everything after this is positional arguments
++argIndex;
// Parse additional positional arguments if any
while (argIndex < argc) {
arg = argv[argIndex];
input.files[input.filesCount].filename = arg;
input.files[input.filesCount].duration = settings.outputTiming.duration;
++input.filesCount;
++argIndex;
}
break;
} else if (!strcmp(arg, "-h") || !strcmp(arg, "--help")) {
syntax();
goto cleanup;
} else if (!strcmp(arg, "-V") || !strcmp(arg, "--version")) {
avifPrintVersions();
goto cleanup;
} else if (!strcmp(arg, "-j") || !strcmp(arg, "--jobs")) {
NEXTARG();
if (!strcmp(arg, "all")) {
settings.jobs = avifQueryCPUCount();
} else {
settings.jobs = atoi(arg);
if (settings.jobs < 1) {
settings.jobs = 1;
}
}
} else if (!strcmp(arg, "--stdin")) {
input.useStdin = AVIF_TRUE;
} else if (!strcmp(arg, "-o") || !strcmp(arg, "--output")) {
NEXTARG();
outputFilename = arg;
} else if (!strcmp(arg, "-d") || !strcmp(arg, "--depth")) {
NEXTARG();
input.requestedDepth = atoi(arg);
if ((input.requestedDepth != 8) && (input.requestedDepth != 10) && (input.requestedDepth != 12)) {
fprintf(stderr, "ERROR: invalid depth: %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "-y") || !strcmp(arg, "--yuv")) {
NEXTARG();
if (!strcmp(arg, "444")) {
input.requestedFormat = AVIF_PIXEL_FORMAT_YUV444;
} else if (!strcmp(arg, "422")) {
input.requestedFormat = AVIF_PIXEL_FORMAT_YUV422;
} else if (!strcmp(arg, "420")) {
input.requestedFormat = AVIF_PIXEL_FORMAT_YUV420;
} else if (!strcmp(arg, "400")) {
input.requestedFormat = AVIF_PIXEL_FORMAT_YUV400;
} else {
fprintf(stderr, "ERROR: invalid format: %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "-k") || !strcmp(arg, "--keyframe")) {
NEXTARG();
settings.keyframeInterval = atoi(arg);
} else if (!strcmp(arg, "-q") || !strcmp(arg, "--qcolor")) {
NEXTARG();
settings.quality = atoi(arg);
if (settings.quality < AVIF_QUALITY_WORST) {
settings.quality = AVIF_QUALITY_WORST;
}
if (settings.quality > AVIF_QUALITY_BEST) {
settings.quality = AVIF_QUALITY_BEST;
}
settings.qualityIsConstrained = AVIF_TRUE;
} else if (!strcmp(arg, "--qalpha")) {
NEXTARG();
settings.qualityAlpha = atoi(arg);
if (settings.qualityAlpha < AVIF_QUALITY_WORST) {
settings.qualityAlpha = AVIF_QUALITY_WORST;
}
if (settings.qualityAlpha > AVIF_QUALITY_BEST) {
settings.qualityAlpha = AVIF_QUALITY_BEST;
}
settings.qualityAlphaIsConstrained = AVIF_TRUE;
} else if (!strcmp(arg, "--min")) {
NEXTARG();
settings.minQuantizer = atoi(arg);
if (settings.minQuantizer < AVIF_QUANTIZER_BEST_QUALITY) {
settings.minQuantizer = AVIF_QUANTIZER_BEST_QUALITY;
}
if (settings.minQuantizer > AVIF_QUANTIZER_WORST_QUALITY) {
settings.minQuantizer = AVIF_QUANTIZER_WORST_QUALITY;
}
} else if (!strcmp(arg, "--max")) {
NEXTARG();
settings.maxQuantizer = atoi(arg);
if (settings.maxQuantizer < AVIF_QUANTIZER_BEST_QUALITY) {
settings.maxQuantizer = AVIF_QUANTIZER_BEST_QUALITY;
}
if (settings.maxQuantizer > AVIF_QUANTIZER_WORST_QUALITY) {
settings.maxQuantizer = AVIF_QUANTIZER_WORST_QUALITY;
}
} else if (!strcmp(arg, "--minalpha")) {
NEXTARG();
settings.minQuantizerAlpha = atoi(arg);
if (settings.minQuantizerAlpha < AVIF_QUANTIZER_BEST_QUALITY) {
settings.minQuantizerAlpha = AVIF_QUANTIZER_BEST_QUALITY;
}
if (settings.minQuantizerAlpha > AVIF_QUANTIZER_WORST_QUALITY) {
settings.minQuantizerAlpha = AVIF_QUANTIZER_WORST_QUALITY;
}
} else if (!strcmp(arg, "--maxalpha")) {
NEXTARG();
settings.maxQuantizerAlpha = atoi(arg);
if (settings.maxQuantizerAlpha < AVIF_QUANTIZER_BEST_QUALITY) {
settings.maxQuantizerAlpha = AVIF_QUANTIZER_BEST_QUALITY;
}
if (settings.maxQuantizerAlpha > AVIF_QUANTIZER_WORST_QUALITY) {
settings.maxQuantizerAlpha = AVIF_QUANTIZER_WORST_QUALITY;
}
} else if (!strcmp(arg, "--target-size")) {
NEXTARG();
settings.targetSize = atoi(arg);
if (settings.targetSize < 0) {
settings.targetSize = -1;
}
} else if (!strcmp(arg, "--tilerowslog2")) {
NEXTARG();
settings.tileRowsLog2 = atoi(arg);
if (settings.tileRowsLog2 < 0) {
settings.tileRowsLog2 = 0;
}
if (settings.tileRowsLog2 > 6) {
settings.tileRowsLog2 = 6;
}
} else if (!strcmp(arg, "--tilecolslog2")) {
NEXTARG();
settings.tileColsLog2 = atoi(arg);
if (settings.tileColsLog2 < 0) {
settings.tileColsLog2 = 0;
}
if (settings.tileColsLog2 > 6) {
settings.tileColsLog2 = 6;
}
} else if (!strcmp(arg, "--autotiling")) {
settings.autoTiling = AVIF_TRUE;
} else if (!strcmp(arg, "--progressive")) {
settings.progressive = AVIF_TRUE;
} else if (!strcmp(arg, "-g") || !strcmp(arg, "--grid")) {
NEXTARG();
settings.gridDimsCount = parseU32List(settings.gridDims, arg);
if (settings.gridDimsCount != 2) {
fprintf(stderr, "ERROR: Invalid grid dims: %s\n", arg);
returnCode = 1;
goto cleanup;
}
if ((settings.gridDims[0] == 0) || (settings.gridDims[0] > 256) || (settings.gridDims[1] == 0) ||
(settings.gridDims[1] > 256)) {
fprintf(stderr, "ERROR: Invalid grid dims (valid dim range [1-256]): %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "--cicp") || !strcmp(arg, "--nclx")) {
NEXTARG();
int cicp[3];
if (!parseCICP(cicp, arg)) {
returnCode = 1;
goto cleanup;
}
settings.colorPrimaries = (avifColorPrimaries)cicp[0];
settings.transferCharacteristics = (avifTransferCharacteristics)cicp[1];
settings.matrixCoefficients = (avifMatrixCoefficients)cicp[2];
cicpExplicitlySet = AVIF_TRUE;
} else if (!strcmp(arg, "-r") || !strcmp(arg, "--range")) {
NEXTARG();
if (!strcmp(arg, "limited") || !strcmp(arg, "l")) {
requestedRange = AVIF_RANGE_LIMITED;
} else if (!strcmp(arg, "full") || !strcmp(arg, "f")) {
requestedRange = AVIF_RANGE_FULL;
} else {
fprintf(stderr, "ERROR: Unknown range: %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "-s") || !strcmp(arg, "--speed")) {
NEXTARG();
if (!strcmp(arg, "default") || !strcmp(arg, "d")) {
settings.speed = AVIF_SPEED_DEFAULT;
} else {
settings.speed = atoi(arg);
if (settings.speed > AVIF_SPEED_FASTEST) {
settings.speed = AVIF_SPEED_FASTEST;
}
if (settings.speed < AVIF_SPEED_SLOWEST) {
settings.speed = AVIF_SPEED_SLOWEST;
}
}
} else if (!strcmp(arg, "--exif")) {
NEXTARG();
if (!readEntireFile(arg, &exifOverride)) {
fprintf(stderr, "ERROR: Unable to read Exif metadata: %s\n", arg);
returnCode = 1;
goto cleanup;
}
settings.ignoreExif = AVIF_TRUE;
} else if (!strcmp(arg, "--xmp")) {
NEXTARG();
if (!readEntireFile(arg, &xmpOverride)) {
fprintf(stderr, "ERROR: Unable to read XMP metadata: %s\n", arg);
returnCode = 1;
goto cleanup;
}
settings.ignoreXMP = AVIF_TRUE;
} else if (!strcmp(arg, "--icc")) {
NEXTARG();
if (!readEntireFile(arg, &iccOverride)) {
fprintf(stderr, "ERROR: Unable to read ICC profile: %s\n", arg);
returnCode = 1;
goto cleanup;
}
settings.ignoreICC = AVIF_TRUE;
} else if (!strcmp(arg, "--duration")) {
NEXTARG();
int durationInt = atoi(arg);
if (durationInt < 1) {
fprintf(stderr, "ERROR: Invalid duration: %d\n", durationInt);
returnCode = 1;
goto cleanup;
}
settings.outputTiming.duration = (uint64_t)durationInt;
} else if (!strcmp(arg, "--timescale") || !strcmp(arg, "--fps")) {
NEXTARG();
int timescaleInt = atoi(arg);
if (timescaleInt < 1) {
fprintf(stderr, "ERROR: Invalid timescale: %d\n", timescaleInt);
returnCode = 1;
goto cleanup;
}
settings.outputTiming.timescale = (uint64_t)timescaleInt;
} else if (!strcmp(arg, "-c") || !strcmp(arg, "--codec")) {
NEXTARG();
settings.codecChoice = avifCodecChoiceFromName(arg);
if (settings.codecChoice == AVIF_CODEC_CHOICE_AUTO) {
fprintf(stderr, "ERROR: Unrecognized codec: %s\n", arg);
returnCode = 1;
goto cleanup;
} else {
const char * codecName = avifCodecName(settings.codecChoice, AVIF_CODEC_FLAG_CAN_ENCODE);
if (codecName == NULL) {
fprintf(stderr, "ERROR: AV1 Codec cannot encode: %s\n", arg);
returnCode = 1;
goto cleanup;
}
}
} else if (!strcmp(arg, "-a") || !strcmp(arg, "--advanced")) {
NEXTARG();
if (!avifCodecSpecificOptionsAdd(&settings.codecSpecificOptions, arg)) {
fprintf(stderr, "ERROR: Out of memory when setting codec specific option: %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "--ignore-exif")) {
settings.ignoreExif = AVIF_TRUE;
} else if (!strcmp(arg, "--ignore-xmp")) {
settings.ignoreXMP = AVIF_TRUE;
} else if (!strcmp(arg, "--ignore-icc")) {
settings.ignoreICC = AVIF_TRUE;
} else if (!strcmp(arg, "--pasp")) {
NEXTARG();
settings.paspCount = parseU32List(settings.paspValues, arg);
if (settings.paspCount != 2) {
fprintf(stderr, "ERROR: Invalid pasp values: %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "--crop")) {
NEXTARG();
settings.clapCount = parseU32List(settings.clapValues, arg);
if (settings.clapCount != 4) {
fprintf(stderr, "ERROR: Invalid crop values: %s\n", arg);
returnCode = 1;
goto cleanup;
}
cropConversionRequired = AVIF_TRUE;
} else if (!strcmp(arg, "--clap")) {
NEXTARG();
settings.clapCount = parseU32List(settings.clapValues, arg);
if (settings.clapCount != 8) {
fprintf(stderr, "ERROR: Invalid clap values: %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "--irot")) {
NEXTARG();
irotAngle = (uint8_t)atoi(arg);
if (irotAngle > 3) {
fprintf(stderr, "ERROR: Invalid irot angle: %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "--imir")) {
NEXTARG();
imirMode = (uint8_t)atoi(arg);
if (imirMode > 1) {
fprintf(stderr, "ERROR: Invalid imir mode: %s\n", arg);
returnCode = 1;
goto cleanup;
}
} else if (!strcmp(arg, "--repetition-count")) {
NEXTARG();
if (!strcmp(arg, "infinite")) {
settings.repetitionCount = AVIF_REPETITION_COUNT_INFINITE;
} else {
settings.repetitionCount = atoi(arg);
if (settings.repetitionCount < 0) {
fprintf(stderr, "ERROR: Invalid repetition count: %s\n", arg);
returnCode = 1;
goto cleanup;
}
}
} else if (!strcmp(arg, "-l") || !strcmp(arg, "--lossless")) {
lossless = AVIF_TRUE;
} else if (!strcmp(arg, "-p") || !strcmp(arg, "--premultiply")) {
premultiplyAlpha = AVIF_TRUE;
} else if (!strcmp(arg, "--sharpyuv")) {
settings.chromaDownsampling = AVIF_CHROMA_DOWNSAMPLING_SHARP_YUV;
} else if (arg[0] == '-') {
fprintf(stderr, "ERROR: unrecognized option %s\n\n", arg);
syntax();
returnCode = 1;
goto cleanup;
} else {
// Positional argument
input.files[input.filesCount].filename = arg;
input.files[input.filesCount].duration = settings.outputTiming.duration;
++input.filesCount;
}
++argIndex;
}
if ((settings.minQuantizer < 0) != (settings.maxQuantizer < 0)) {
fprintf(stderr, "--min and --max must be either both specified or both unspecified.\n");
returnCode = 1;
goto cleanup;
}
if ((settings.minQuantizerAlpha < 0) != (settings.maxQuantizerAlpha < 0)) {
fprintf(stderr, "--minalpha and --maxalpha must be either both specified or both unspecified.\n");
returnCode = 1;
goto cleanup;
}
// Check lossy/lossless parameters and set to default if needed.
if (lossless) {
// Pixel format.
if (input.requestedFormat != AVIF_PIXEL_FORMAT_NONE && input.requestedFormat != AVIF_PIXEL_FORMAT_YUV444) {
fprintf(stderr,
"When set, the pixel format can only be 444 in lossless "
"mode.\n");
returnCode = 1;
}
// Don't subsample when using AVIF_MATRIX_COEFFICIENTS_IDENTITY.
input.requestedFormat = AVIF_PIXEL_FORMAT_YUV444;
// Quality.
if ((settings.quality != INVALID_QUALITY && settings.quality != AVIF_QUALITY_LOSSLESS) ||
(settings.qualityAlpha != INVALID_QUALITY && settings.qualityAlpha != AVIF_QUALITY_LOSSLESS)) {
fprintf(stderr, "Quality cannot be set in lossless mode, except to %d.\n", AVIF_QUALITY_LOSSLESS);
returnCode = 1;
}
settings.quality = settings.qualityAlpha = AVIF_QUALITY_LOSSLESS;
// Quantizers.
if (settings.minQuantizer > 0 || settings.maxQuantizer > 0 || settings.minQuantizerAlpha > 0 || settings.maxQuantizerAlpha > 0) {
fprintf(stderr, "Quantizers cannot be set in lossless mode, except to 0.\n");
returnCode = 1;
}
settings.minQuantizer = settings.maxQuantizer = settings.minQuantizerAlpha = settings.maxQuantizerAlpha = AVIF_QUANTIZER_LOSSLESS;
// Codec.
if (settings.codecChoice != AVIF_CODEC_CHOICE_AUTO && settings.codecChoice != AVIF_CODEC_CHOICE_AOM) {
fprintf(stderr, "Codec can only be AOM in lossless mode.\n");
returnCode = 1;
}
// rav1e doesn't support lossless transform yet:
// https://github.com/xiph/rav1e/issues/151
// SVT-AV1 doesn't support lossless encoding yet:
// https://gitlab.com/AOMediaCodec/SVT-AV1/-/issues/1636
settings.codecChoice = AVIF_CODEC_CHOICE_AOM;
// Range.
if (requestedRange != AVIF_RANGE_FULL) {
fprintf(stderr, "Range has to be full in lossless mode.\n");
returnCode = 1;
}
// Matrix coefficients.
if (cicpExplicitlySet && settings.matrixCoefficients != AVIF_MATRIX_COEFFICIENTS_IDENTITY) {
fprintf(stderr, "Matrix coefficients have to be identity in lossless mode.\n");
returnCode = 1;
}
settings.matrixCoefficients = AVIF_MATRIX_COEFFICIENTS_IDENTITY;
if (returnCode == 1)
goto cleanup;
} else {
// Set lossy defaults.
if (settings.minQuantizer == -1) {
assert(settings.maxQuantizer == -1);
if (settings.quality == INVALID_QUALITY) {
settings.quality = DEFAULT_QUALITY;
}
settings.minQuantizer = AVIF_QUANTIZER_BEST_QUALITY;
settings.maxQuantizer = AVIF_QUANTIZER_WORST_QUALITY;
} else {
assert(settings.maxQuantizer != -1);
if (settings.quality == INVALID_QUALITY) {
const int quantizer = (settings.minQuantizer + settings.maxQuantizer) / 2;
settings.quality = ((63 - quantizer) * 100 + 31) / 63;
}
}
if (settings.minQuantizerAlpha == -1) {
assert(settings.maxQuantizerAlpha == -1);
if (settings.qualityAlpha == INVALID_QUALITY) {
settings.qualityAlpha = DEFAULT_QUALITY_ALPHA;
}
settings.minQuantizerAlpha = AVIF_QUANTIZER_BEST_QUALITY;
settings.maxQuantizerAlpha = AVIF_QUANTIZER_WORST_QUALITY;
} else {
assert(settings.maxQuantizerAlpha != -1);
if (settings.qualityAlpha == INVALID_QUALITY) {
const int quantizerAlpha = (settings.minQuantizerAlpha + settings.maxQuantizerAlpha) / 2;
settings.qualityAlpha = ((63 - quantizerAlpha) * 100 + 31) / 63;
}
}
}
assert(settings.quality != INVALID_QUALITY);
assert(settings.qualityAlpha != INVALID_QUALITY);
// In progressive encoding we use a very low quality (2) for the base layer to ensure a small
// encoded size. If the target quality is close to the quality of the base layer, don't bother
// with progressive encoding.
if (settings.progressive && ((settings.quality < 10) || (settings.qualityAlpha < 10))) {
settings.progressive = AVIF_FALSE;
printf("The --progressive option was ignored because the quality is below 10.\n");
}
stdinFile.filename = "(stdin)";
stdinFile.duration = settings.outputTiming.duration;
if (!outputFilename) {
if (((input.useStdin && (input.filesCount == 1)) || (!input.useStdin && (input.filesCount > 1)))) {
--input.filesCount;
outputFilename = input.files[input.filesCount].filename;
}
}
if (!outputFilename || (input.useStdin && (input.filesCount > 0)) || (!input.useStdin && (input.filesCount < 1))) {
syntax();
returnCode = 1;
goto cleanup;
}
#if defined(_WIN32)
if (input.useStdin) {
setmode(fileno(stdin), O_BINARY);
}
#endif
image = avifImageCreateEmpty();
if (!image) {
fprintf(stderr, "ERROR: Out of memory\n");
returnCode = 1;
goto cleanup;
}
// Set these in advance so any upcoming RGB -> YUV use the proper coefficients
image->colorPrimaries = settings.colorPrimaries;
image->transferCharacteristics = settings.transferCharacteristics;
image->matrixCoefficients = settings.matrixCoefficients;
image->yuvRange = requestedRange;
image->alphaPremultiplied = premultiplyAlpha;
if ((image->matrixCoefficients == AVIF_MATRIX_COEFFICIENTS_IDENTITY) && (input.requestedFormat != AVIF_PIXEL_FORMAT_NONE) &&
(input.requestedFormat != AVIF_PIXEL_FORMAT_YUV444)) {
// User explicitly asked for non YUV444 yuvFormat, while matrixCoefficients was likely
// set to AVIF_MATRIX_COEFFICIENTS_IDENTITY as a side effect of --lossless,
// and Identity is only valid with YUV444. Set matrixCoefficients back to the default.
image->matrixCoefficients = AVIF_MATRIX_COEFFICIENTS_BT601;
if (cicpExplicitlySet) {
// Only warn if someone explicitly asked for identity.
printf("WARNING: matrixCoefficients may not be set to identity (0) when subsampling. Resetting MC to defaults (%d).\n",
image->matrixCoefficients);
}
}
// --target-size requires multiple encodings of the same files. Cache the input images.
input.cacheEnabled = (settings.targetSize != -1);
const avifInputFile * firstFile = avifInputGetFile(&input, /*imageIndex=*/0);
uint32_t sourceDepth = 0;
avifBool sourceWasRGB = AVIF_FALSE;
avifAppSourceTiming firstSourceTiming;
if (!avifInputReadImage(&input,
/*imageIndex=*/0,
settings.ignoreICC,
settings.ignoreExif,
settings.ignoreXMP,
image,
&sourceDepth,
&sourceWasRGB,
&firstSourceTiming,
settings.chromaDownsampling)) {
returnCode = 1;
goto cleanup;
}
// Check again for y4m input (y4m input ignores input.requestedFormat and retains the format in file).
if ((image->matrixCoefficients == AVIF_MATRIX_COEFFICIENTS_IDENTITY) && (image->yuvFormat != AVIF_PIXEL_FORMAT_YUV444)) {
fprintf(stderr, "matrixCoefficients may not be set to identity (0) when subsampling.\n");
returnCode = 1;
goto cleanup;
}
printf("Successfully loaded: %s\n", firstFile->filename);
// Prepare image timings
if ((settings.outputTiming.duration == 0) && (settings.outputTiming.timescale == 0) && (firstSourceTiming.duration > 0) &&
(firstSourceTiming.timescale > 0)) {
// Set the default duration and timescale to the first image's timing.
settings.outputTiming = firstSourceTiming;
} else {
// Set output timing defaults to 30 fps
if (settings.outputTiming.duration == 0) {
settings.outputTiming.duration = 1;
}
if (settings.outputTiming.timescale == 0) {
settings.outputTiming.timescale = 30;
}
}
if (iccOverride.size) {
avifImageSetProfileICC(image, iccOverride.data, iccOverride.size);
}
if (exifOverride.size) {
avifImageSetMetadataExif(image, exifOverride.data, exifOverride.size);
}
if (xmpOverride.size) {
avifImageSetMetadataXMP(image, xmpOverride.data, xmpOverride.size);
}
if (!image->icc.size && !cicpExplicitlySet && (image->colorPrimaries == AVIF_COLOR_PRIMARIES_UNSPECIFIED) &&
(image->transferCharacteristics == AVIF_TRANSFER_CHARACTERISTICS_UNSPECIFIED)) {
// The final image has no ICC profile, the user didn't specify any CICP, and the source
// image didn't provide any CICP. Explicitly signal SRGB CP/TC here, as 2/2/x will be
// interpreted as SRGB anyway.
image->colorPrimaries = AVIF_COLOR_PRIMARIES_BT709;
image->transferCharacteristics = AVIF_TRANSFER_CHARACTERISTICS_SRGB;
}
if (settings.paspCount == 2) {
image->transformFlags |= AVIF_TRANSFORM_PASP;
image->pasp.hSpacing = settings.paspValues[0];
image->pasp.vSpacing = settings.paspValues[1];
}
if (cropConversionRequired) {
if (!convertCropToClap(image->width, image->height, image->yuvFormat, settings.clapValues)) {
returnCode = 1;
goto cleanup;
}
settings.clapCount = 8;
}
if (settings.clapCount == 8) {
image->transformFlags |= AVIF_TRANSFORM_CLAP;
image->clap.widthN = settings.clapValues[0];
image->clap.widthD = settings.clapValues[1];
image->clap.heightN = settings.clapValues[2];
image->clap.heightD = settings.clapValues[3];
image->clap.horizOffN = settings.clapValues[4];
image->clap.horizOffD = settings.clapValues[5];
image->clap.vertOffN = settings.clapValues[6];
image->clap.vertOffD = settings.clapValues[7];
// Validate clap
avifCropRect cropRect;
avifDiagnostics diag;
avifDiagnosticsClearError(&diag);
if (!avifCropRectConvertCleanApertureBox(&cropRect, &image->clap, image->width, image->height, image->yuvFormat, &diag)) {
fprintf(stderr,
"ERROR: Invalid clap: width:[%d / %d], height:[%d / %d], horizOff:[%d / %d], vertOff:[%d / %d] - %s\n",
(int32_t)image->clap.widthN,
(int32_t)image->clap.widthD,
(int32_t)image->clap.heightN,
(int32_t)image->clap.heightD,
(int32_t)image->clap.horizOffN,
(int32_t)image->clap.horizOffD,
(int32_t)image->clap.vertOffN,
(int32_t)image->clap.vertOffD,
diag.error);
returnCode = 1;
goto cleanup;
}
}
if (irotAngle != 0xff) {
image->transformFlags |= AVIF_TRANSFORM_IROT;
image->irot.angle = irotAngle;
}
if (imirMode != 0xff) {
image->transformFlags |= AVIF_TRANSFORM_IMIR;
image->imir.mode = imirMode;
}
avifBool usingAOM = AVIF_FALSE;
const char * codecName = avifCodecName(settings.codecChoice, AVIF_CODEC_FLAG_CAN_ENCODE);
if (codecName && !strcmp(codecName, "aom")) {
usingAOM = AVIF_TRUE;
}
avifBool hasAlpha = (image->alphaPlane && image->alphaRowBytes);
avifBool usingLosslessColor = (settings.quality == AVIF_QUALITY_LOSSLESS);
avifBool usingLosslessAlpha = (settings.qualityAlpha == AVIF_QUALITY_LOSSLESS);
avifBool depthMatches = (sourceDepth == image->depth);
avifBool using400 = (image->yuvFormat == AVIF_PIXEL_FORMAT_YUV400);
avifBool using444 = (image->yuvFormat == AVIF_PIXEL_FORMAT_YUV444);
avifBool usingFullRange = (image->yuvRange == AVIF_RANGE_FULL);
avifBool usingIdentityMatrix = (image->matrixCoefficients == AVIF_MATRIX_COEFFICIENTS_IDENTITY);
// Guess if the enduser is asking for lossless and enable it so that warnings can be emitted
if (!lossless && usingLosslessColor && (!hasAlpha || usingLosslessAlpha)) {
// The enduser is probably expecting lossless. Turn it on and emit warnings
printf("Quality set to %d, assuming --lossless to enable warnings on potential lossless issues.\n", AVIF_QUALITY_LOSSLESS);
lossless = AVIF_TRUE;
}
// Check for any reasons lossless will fail, and complain loudly
if (lossless) {
if (!usingAOM) {
fprintf(stderr, "WARNING: [--lossless] Only aom (-c) supports lossless transforms. Output might not be lossless.\n");
lossless = AVIF_FALSE;
}
if (!usingLosslessColor) {
fprintf(stderr,
"WARNING: [--lossless] Color quality (-q or --qcolor) not set to %d. Color output might not be lossless.\n",
AVIF_QUALITY_LOSSLESS);
lossless = AVIF_FALSE;
}
if (hasAlpha && !usingLosslessAlpha) {
fprintf(stderr,
"WARNING: [--lossless] Alpha present and alpha quality (--qalpha) not set to %d. Alpha output might not be lossless.\n",
AVIF_QUALITY_LOSSLESS);
lossless = AVIF_FALSE;
}
if (!depthMatches) {
fprintf(stderr,
"WARNING: [--lossless] Input depth (%d) does not match output depth (%d). Output might not be lossless.\n",
sourceDepth,
image->depth);
lossless = AVIF_FALSE;
}
if (sourceWasRGB) {
if (!using444 && !using400) {
fprintf(stderr, "WARNING: [--lossless] Input data was RGB and YUV subsampling (-y) isn't YUV444. Output might not be lossless.\n");
lossless = AVIF_FALSE;
}
if (!usingFullRange) {
fprintf(stderr, "WARNING: [--lossless] Input data was RGB and output range (-r) isn't full. Output might not be lossless.\n");
lossless = AVIF_FALSE;
}
if (!usingIdentityMatrix && !using400) {
fprintf(stderr, "WARNING: [--lossless] Input data was RGB and matrixCoefficients isn't set to identity (--cicp x/x/0); Output might not be lossless.\n");
lossless = AVIF_FALSE;
}
}
}
if (settings.gridDimsCount > 0) {
// Grid image!
gridCellCount = settings.gridDims[0] * settings.gridDims[1];
printf("Preparing to encode a %ux%u grid (%u cells)...\n", settings.gridDims[0], settings.gridDims[1], gridCellCount);
gridCells = calloc(gridCellCount, sizeof(avifImage *));
gridCells[0] = image; // take ownership of image
int imageIndex = 1; // The first grid cell was loaded into image (imageIndex 0).
const avifInputFile * nextFile;
while ((nextFile = avifInputGetFile(&input, imageIndex)) != NULL) {
if (imageIndex == 1) {
printf("Loading additional cells for grid image (%u cells)...\n", gridCellCount);
}
if (imageIndex >= (int)gridCellCount) {
// We have enough, warn and continue
fprintf(stderr,
"WARNING: [--grid] More than %u images were supplied for this %ux%u grid. The rest will be ignored.\n",
gridCellCount,
settings.gridDims[0],
settings.gridDims[1]);
break;
}
avifImage * cellImage = avifImageCreateEmpty();
if (!cellImage) {
fprintf(stderr, "ERROR: Out of memory\n");
returnCode = 1;
goto cleanup;
}
cellImage->colorPrimaries = image->colorPrimaries;
cellImage->transferCharacteristics = image->transferCharacteristics;
cellImage->matrixCoefficients = image->matrixCoefficients;
cellImage->yuvRange = image->yuvRange;
cellImage->alphaPremultiplied = image->alphaPremultiplied;
gridCells[imageIndex] = cellImage;
// Ignore ICC, Exif and XMP because only the metadata of the first frame is taken into
// account by the libavif API.
if (!avifInputReadImage(&input,
imageIndex,
/*ignoreICC=*/AVIF_TRUE,
/*ignoreExif=*/AVIF_TRUE,
/*ignoreXMP=*/AVIF_TRUE,
cellImage,
/*outDepth=*/NULL,
/*sourceIsRGB=*/NULL,
/*sourceTiming=*/NULL,
settings.chromaDownsampling)) {
returnCode = 1;
goto cleanup;
}
// Let avifEncoderAddImageGrid() verify the grid integrity (valid cell sizes, depths etc.).
++imageIndex;
}
if (imageIndex == 1) {
printf("Single image input for a grid image. Attempting to split into %u cells...\n", gridCellCount);
gridSplitImage = image;
gridCells[0] = NULL;
if (!avifImageSplitGrid(gridSplitImage, settings.gridDims[0], settings.gridDims[1], gridCells)) {
returnCode = 1;
goto cleanup;
}
} else if (imageIndex != (int)gridCellCount) {
fprintf(stderr, "ERROR: Not enough input files for grid image! (expecting %u, or a single image to be split)\n", gridCellCount);
returnCode = 1;
goto cleanup;
}
// TODO(yguyon): Check if it is possible to use frames from a single input file as grid cells. Maybe forbid it.
}
const char * lossyHint = " (Lossy)";
if (lossless) {
lossyHint = " (Lossless)";
}
printf("AVIF to be written:%s\n", lossyHint);
const avifImage * avif = gridCells ? gridCells[0] : image;
avifImageDump(avif,
settings.gridDims[0],
settings.gridDims[1],
settings.progressive ? AVIF_PROGRESSIVE_STATE_AVAILABLE : AVIF_PROGRESSIVE_STATE_UNAVAILABLE);
if (settings.autoTiling) {
if ((settings.tileRowsLog2 >= 0) || (settings.tileColsLog2 >= 0)) {
fprintf(stderr, "ERROR: --autotiling is specified but --tilerowslog2 or --tilecolslog2 is also specified\n");
returnCode = 1;
goto cleanup;
}
} else {
if (settings.tileRowsLog2 < 0) {
settings.tileRowsLog2 = 0;
}
if (settings.tileColsLog2 < 0) {
settings.tileColsLog2 = 0;
}
}
avifIOStats ioStats = { 0, 0 };
if (!avifEncodeImages(&settings, &input, firstFile, image, (const avifImage * const *)gridCells, &raw, &ioStats)) {
returnCode = 1;
goto cleanup;
}
printf("Encoded successfully.\n");
printf(" * Color AV1 total size: %" AVIF_FMT_ZU " bytes\n", ioStats.colorOBUSize);
printf(" * Alpha AV1 total size: %" AVIF_FMT_ZU " bytes\n", ioStats.alphaOBUSize);
const avifBool isImageSequence = (settings.gridDimsCount == 0) && (input.filesCount > 1);
if (isImageSequence) {
if (settings.repetitionCount == AVIF_REPETITION_COUNT_INFINITE) {
printf(" * Repetition Count: Infinite\n");
} else {
printf(" * Repetition Count: %d\n", settings.repetitionCount);
}
}
FILE * f = fopen(outputFilename, "wb");
if (!f) {
fprintf(stderr, "ERROR: Failed to open file for write: %s\n", outputFilename);
returnCode = 1;
goto cleanup;
}
if (fwrite(raw.data, 1, raw.size, f) != raw.size) {
fprintf(stderr, "Failed to write %" AVIF_FMT_ZU " bytes: %s\n", raw.size, outputFilename);
returnCode = 1;
} else {
printf("Wrote AVIF: %s\n", outputFilename);
}
fclose(f);
cleanup:
if (gridCells) {
for (uint32_t i = 0; i < gridCellCount; ++i) {
if (gridCells[i]) {
avifImageDestroy(gridCells[i]);
}
}
free(gridCells);
} else if (image) { // image is owned/cleaned up by gridCells if it exists
avifImageDestroy(image);
}
if (gridSplitImage) {
avifImageDestroy(gridSplitImage);
}
avifRWDataFree(&raw);
avifRWDataFree(&exifOverride);
avifRWDataFree(&xmpOverride);
avifRWDataFree(&iccOverride);
while (input.cacheCount) {
--input.cacheCount;
if (input.cache[input.cacheCount].image) {
avifImageDestroy(input.cache[input.cacheCount].image);
}
}
free(input.cache);
free(input.files);
while (settings.codecSpecificOptions.count) {
--settings.codecSpecificOptions.count;
free(settings.codecSpecificOptions.keys[settings.codecSpecificOptions.count]);
free(settings.codecSpecificOptions.values[settings.codecSpecificOptions.count]);
}
free(settings.codecSpecificOptions.keys);
free(settings.codecSpecificOptions.values);
return returnCode;
}