Files
libopenshot/src/Frame.cpp
Jonathan Thomas d70e80eac4 Harden playback/cache path for malformed media and concurrent timeline updates
- Invalidate timeline cache on ApplyJsonDiff() clip insert (remove affected frame range).
  - Add lock in Timeline::ClearAllCache() for safe concurrent access.
  - Make VideoCacheThread cross-thread state safe (atomics + seek-state mutex).
  - Lock CacheMemory::Contains() to avoid races.
  - Handle malformed audio streams in FFmpegReader by disabling invalid audio and continuing video-only.
  - Add FPS/timebase safety fallbacks in FFmpeg frame/PTS math.
  - Guard Frame::GetSamplesPerFrame() against invalid inputs.
  - Add/adjust regression tests for cache invalidation and invalid rate handling.
2026-02-11 20:11:47 -06:00

1004 lines
29 KiB
C++

/**
* @file
* @brief Source file for Frame class
* @author Jonathan Thomas <jonathan@openshot.org>
*
* @ref License
*/
// Copyright (c) 2008-2019 OpenShot Studios, LLC
//
// SPDX-License-Identifier: LGPL-3.0-or-later
#include <thread> // for std::this_thread::sleep_for
#include <chrono> // for std::chrono::milliseconds
#include <iomanip>
#include "Frame.h"
#include "AudioBufferSource.h"
#include "AudioResampler.h"
#include "QtUtilities.h"
#include <AppConfig.h>
#include <juce_audio_basics/juce_audio_basics.h>
#include <juce_audio_devices/juce_audio_devices.h>
#include <QApplication>
#include <QImage>
#include <QPixmap>
#include <QBitmap>
#include <QColor>
#include <QString>
#include <QVector>
#include <QPainter>
#include <QHBoxLayout>
#include <QWidget>
#include <QLabel>
#include <QPointF>
#include <QWidget>
using namespace std;
using namespace openshot;
// Constructor - image & audio
Frame::Frame(int64_t number, int width, int height, std::string color, int samples, int channels)
: audio(std::make_shared<juce::AudioBuffer<float>>(channels, samples)),
number(number), width(width), height(height),
pixel_ratio(1,1), color(color),
channels(channels), channel_layout(LAYOUT_STEREO),
sample_rate(44100),
has_audio_data(false), has_image_data(false),
max_audio_sample(0), audio_is_increasing(true)
{
// zero (fill with silence) the audio buffer
audio->clear();
}
// Delegating Constructor - blank frame
Frame::Frame() : Frame::Frame(1, 1, 1, "#000000", 0, 2) {}
// Delegating Constructor - image only
Frame::Frame(int64_t number, int width, int height, std::string color)
: Frame::Frame(number, width, height, color, 0, 2) {}
// Delegating Constructor - audio only
Frame::Frame(int64_t number, int samples, int channels)
: Frame::Frame(number, 1, 1, "#000000", samples, channels) {}
// Copy constructor
Frame::Frame ( const Frame &other )
{
// copy pointers and data
DeepCopy(other);
}
// Assignment operator
Frame& Frame::operator= (const Frame& other)
{
// copy pointers and data
DeepCopy(other);
return *this;
}
// Copy data and pointers from another Frame instance
void Frame::DeepCopy(const Frame& other)
{
number = other.number;
channels = other.channels;
width = other.width;
height = other.height;
channel_layout = other.channel_layout;
has_audio_data = other.has_audio_data;
has_image_data = other.has_image_data;
sample_rate = other.sample_rate;
pixel_ratio = Fraction(other.pixel_ratio.num, other.pixel_ratio.den);
color = other.color;
max_audio_sample = other.max_audio_sample;
audio_is_increasing = other.audio_is_increasing;
if (other.image)
image = std::make_shared<QImage>(*(other.image));
if (other.audio)
audio = std::make_shared<juce::AudioBuffer<float>>(*(other.audio));
if (other.wave_image)
wave_image = std::make_shared<QImage>(*(other.wave_image));
}
// Destructor
Frame::~Frame() {
// Clear all pointers
image.reset();
audio.reset();
#ifdef USE_OPENCV
imagecv.release();
#endif
}
// Display the frame image to the screen (primarily used for debugging reasons)
void Frame::Display()
{
if (!QApplication::instance()) {
// Only create the QApplication once
static int argc = 1;
static char* argv[1] = {NULL};
previewApp = std::make_shared<QApplication>(argc, argv);
}
// Get preview image
std::shared_ptr<QImage> previewImage = GetImage();
// Update the image to reflect the correct pixel aspect ration (i.e. to fix non-square pixels)
if (pixel_ratio.num != 1 || pixel_ratio.den != 1)
{
// Resize to fix DAR
previewImage = std::make_shared<QImage>(previewImage->scaled(
previewImage->size().width(), previewImage->size().height() * pixel_ratio.Reciprocal().ToDouble(),
Qt::IgnoreAspectRatio, Qt::SmoothTransformation));
}
// Create window
QWidget previewWindow;
previewWindow.setStyleSheet("background-color: #000000;");
QHBoxLayout layout;
// Create label with current frame's image
QLabel previewLabel;
previewLabel.setPixmap(QPixmap::fromImage(*previewImage));
previewLabel.setMask(QPixmap::fromImage(*previewImage).mask());
layout.addWidget(&previewLabel);
// Show the window
previewWindow.setLayout(&layout);
previewWindow.show();
previewApp->exec();
}
// Get an audio waveform image
std::shared_ptr<QImage> Frame::GetWaveform(int width, int height, int Red, int Green, int Blue, int Alpha)
{
// Clear any existing waveform image
ClearWaveform();
// Init a list of lines
QVector<QPointF> lines;
QVector<QPointF> labels;
// Calculate width of an image based on the # of samples
int total_samples = GetAudioSamplesCount();
if (total_samples > 0)
{
// If samples are present...
int new_height = 200 * audio->getNumChannels();
int height_padding = 20 * (audio->getNumChannels() - 1);
int total_height = new_height + height_padding;
int total_width = 0;
float zero_height = 1.0; // Used to clamp near-zero vales to this value to prevent gaps
// Loop through each audio channel
float Y = 100.0;
for (int channel = 0; channel < audio->getNumChannels(); channel++)
{
float X = 0.0;
// Get audio for this channel
const float *samples = audio->getReadPointer(channel);
for (int sample = 0; sample < GetAudioSamplesCount(); sample++, X++)
{
// Sample value (scaled to -100 to 100)
float value = samples[sample] * 100.0;
// Set threshold near zero (so we don't allow near-zero values)
// This prevents empty gaps from appearing in the waveform
if (value > -zero_height && value < 0.0) {
value = -zero_height;
} else if (value > 0.0 && value < zero_height) {
value = zero_height;
}
// Append a line segment for each sample
lines.push_back(QPointF(X, Y));
lines.push_back(QPointF(X, Y - value));
}
// Add Channel Label Coordinate
labels.push_back(QPointF(5.0, Y - 5.0));
// Increment Y
Y += (200 + height_padding);
total_width = X;
}
// Create blank image
wave_image = std::make_shared<QImage>(
total_width, total_height, QImage::Format_RGBA8888_Premultiplied);
wave_image->fill(QColor(0,0,0,0));
// Load QPainter with wave_image device
QPainter painter(wave_image.get());
// Set pen color
QPen pen;
pen.setColor(QColor(Red, Green, Blue, Alpha));
pen.setWidthF(1.0);
pen.setStyle(Qt::SolidLine);
painter.setPen(pen);
// Draw the waveform
painter.drawLines(lines);
painter.end();
}
else
{
// No audio samples present
wave_image = std::make_shared<QImage>(width, height, QImage::Format_RGBA8888_Premultiplied);
wave_image->fill(QColor(QString::fromStdString("#000000")));
}
// Resize Image (if needed)
if (wave_image->width() != width || wave_image->height() != height) {
QImage scaled_wave_image = wave_image->scaled(width, height, Qt::IgnoreAspectRatio, Qt::SmoothTransformation);
wave_image = std::make_shared<QImage>(scaled_wave_image);
}
// Return new image
return wave_image;
}
// Clear the waveform image (and deallocate its memory)
void Frame::ClearWaveform()
{
if (wave_image)
wave_image.reset();
}
// Get an audio waveform image pixels
const unsigned char* Frame::GetWaveformPixels(int width, int height, int Red, int Green, int Blue, int Alpha)
{
// Get audio wave form image
wave_image = GetWaveform(width, height, Red, Green, Blue, Alpha);
// Return array of pixel packets
return wave_image->constBits();
}
// Display the wave form
void Frame::DisplayWaveform()
{
// Get audio wave form image
GetWaveform(720, 480, 0, 123, 255, 255);
if (!QApplication::instance()) {
// Only create the QApplication once
static int argc = 1;
static char* argv[1] = {NULL};
previewApp = std::make_shared<QApplication>(argc, argv);
}
// Create window
QWidget previewWindow;
previewWindow.setStyleSheet("background-color: #000000;");
QHBoxLayout layout;
// Create label with current frame's waveform image
QLabel previewLabel;
previewLabel.setPixmap(QPixmap::fromImage(*wave_image));
previewLabel.setMask(QPixmap::fromImage(*wave_image).mask());
layout.addWidget(&previewLabel);
// Show the window
previewWindow.setLayout(&layout);
previewWindow.show();
previewApp->exec();
// Deallocate waveform image
ClearWaveform();
}
// Get magnitude of range of samples (if channel is -1, return average of all channels for that sample)
float Frame::GetAudioSample(int channel, int sample, int magnitude_range)
{
if (channel > 0) {
// return average magnitude for a specific channel/sample range
return audio->getMagnitude(channel, sample, magnitude_range);
} else {
// Return average magnitude for all channels
return audio->getMagnitude(sample, magnitude_range);
}
}
// Get an array of sample data (and optional reverse the sample values)
float* Frame::GetAudioSamples(int channel) {
// Copy audio data
juce::AudioBuffer<float> *buffer(audio.get());
// return JUCE audio data for this channel
return buffer->getWritePointer(channel);
}
// Get an array of sample data (all channels interleaved together), using any sample rate
float* Frame::GetInterleavedAudioSamples(int* sample_count)
{
// Copy audio data
juce::AudioBuffer<float> *buffer(audio.get());
float *output = NULL;
int num_of_channels = audio->getNumChannels();
int num_of_samples = GetAudioSamplesCount();
// INTERLEAVE all samples together (channel 1 + channel 2 + channel 1 + channel 2, etc...)
output = new float[num_of_channels * num_of_samples];
int position = 0;
// Loop through samples in each channel (combining them)
for (int sample = 0; sample < num_of_samples; sample++)
{
for (int channel = 0; channel < num_of_channels; channel++)
{
// Add sample to output array
output[position] = buffer->getReadPointer(channel)[sample];
// increment position
position++;
}
}
// Update sample count (since it might have changed due to resampling)
*sample_count = num_of_samples;
// return combined array
return output;
}
// Get number of audio channels
int Frame::GetAudioChannelsCount()
{
const std::lock_guard<std::recursive_mutex> lock(addingAudioMutex);
if (audio)
return audio->getNumChannels();
else
return 0;
}
// Get number of audio samples
int Frame::GetAudioSamplesCount()
{
const std::lock_guard<std::recursive_mutex> lock(addingAudioMutex);
return max_audio_sample;
}
juce::AudioBuffer<float> *Frame::GetAudioSampleBuffer()
{
return audio.get();
}
// Get the size in bytes of this frame (rough estimate)
int64_t Frame::GetBytes()
{
int64_t total_bytes = 0;
if (image) {
total_bytes += static_cast<int64_t>(
width * height * sizeof(char) * 4);
}
if (audio) {
// approximate audio size (sample rate / 24 fps)
total_bytes += (sample_rate / 24.0) * sizeof(float);
}
// return size of this frame
return total_bytes;
}
// Get pixel data (as packets)
const unsigned char* Frame::GetPixels()
{
// Check for blank image
if (!image)
// Fill with black
AddColor(width, height, color);
// Return array of pixel packets
return image->constBits();
}
// Get pixel data (for only a single scan-line)
const unsigned char* Frame::GetPixels(int row)
{
// Check for blank image
if (!image)
// Fill with black
AddColor(width, height, color);
// Return array of pixel packets
return image->constScanLine(row);
}
// Check a specific pixel color value (returns True/False)
bool Frame::CheckPixel(int row, int col, int red, int green, int blue, int alpha, int threshold) {
int col_pos = col * 4; // Find column array position
if (!image || row < 0 || row >= (height - 1) ||
col_pos < 0 || col_pos >= (width - 1) ) {
// invalid row / col
return false;
}
// Check pixel color
const unsigned char* pixels = GetPixels(row);
if (pixels[col_pos + 0] >= (red - threshold) && pixels[col_pos + 0] <= (red + threshold) &&
pixels[col_pos + 1] >= (green - threshold) && pixels[col_pos + 1] <= (green + threshold) &&
pixels[col_pos + 2] >= (blue - threshold) && pixels[col_pos + 2] <= (blue + threshold) &&
pixels[col_pos + 3] >= (alpha - threshold) && pixels[col_pos + 3] <= (alpha + threshold)) {
// Pixel color matches successfully
return true;
} else {
// Pixel color does not match
return false;
}
}
// Set Pixel Aspect Ratio
void Frame::SetPixelRatio(int num, int den)
{
pixel_ratio.num = num;
pixel_ratio.den = den;
}
// Set frame number
void Frame::SetFrameNumber(int64_t new_number)
{
number = new_number;
}
// Calculate the # of samples per video frame (for a specific frame number and frame rate)
int Frame::GetSamplesPerFrame(int64_t number, Fraction fps, int sample_rate, int channels)
{
// Directly return 0 for invalid audio/frame-rate parameters
// so that we do not need to deal with NaNs later
if (channels <= 0 || sample_rate <= 0 || fps.num <= 0 || fps.den <= 0) return 0;
// Get the total # of samples for the previous frame, and the current frame (rounded)
double fps_rate = fps.Reciprocal().ToDouble();
// Determine previous samples total, and make sure it's evenly divisible by the # of channels
double previous_samples = (sample_rate * fps_rate) * (number - 1);
double previous_samples_remainder = fmod(previous_samples, (double)channels); // subtract the remainder to the total (to make it evenly divisible)
previous_samples -= previous_samples_remainder;
// Determine the current samples total, and make sure it's evenly divisible by the # of channels
double total_samples = (sample_rate * fps_rate) * number;
double total_samples_remainder = fmod(total_samples, (double)channels); // subtract the remainder to the total (to make it evenly divisible)
total_samples -= total_samples_remainder;
// Subtract the previous frame's total samples with this frame's total samples. Not all sample rates can
// be evenly divided into frames, so each frame can have have different # of samples.
int samples_per_frame = round(total_samples - previous_samples);
if (samples_per_frame < 0)
samples_per_frame = 0;
return samples_per_frame;
}
// Calculate the # of samples per video frame (for the current frame number)
int Frame::GetSamplesPerFrame(Fraction fps, int sample_rate, int channels)
{
return GetSamplesPerFrame(number, fps, sample_rate, channels);
}
// Get height of image
int Frame::GetHeight()
{
return height;
}
// Get height of image
int Frame::GetWidth()
{
return width;
}
// Get the original sample rate of this frame's audio data
int Frame::SampleRate()
{
return sample_rate;
}
// Get the original sample rate of this frame's audio data
ChannelLayout Frame::ChannelsLayout()
{
return channel_layout;
}
// Save the frame image to the specified path. The image format is determined from the extension (i.e. image.PNG, image.JPEG)
void Frame::Save(std::string path, float scale, std::string format, int quality)
{
// Get preview image
std::shared_ptr<QImage> previewImage = GetImage();
// Update the image to reflect the correct pixel aspect ration (i.e. to fix non-square pixels)
if (pixel_ratio.num != 1 || pixel_ratio.den != 1)
{
// Resize to fix DAR
previewImage = std::make_shared<QImage>(previewImage->scaled(
previewImage->size().width(), previewImage->size().height() * pixel_ratio.Reciprocal().ToDouble(),
Qt::IgnoreAspectRatio, Qt::SmoothTransformation));
}
// scale image if needed
if (fabs(scale) > 1.001 || fabs(scale) < 0.999)
{
// Resize image
previewImage = std::make_shared<QImage>(previewImage->scaled(
previewImage->size().width() * scale, previewImage->size().height() * scale,
Qt::KeepAspectRatio, Qt::SmoothTransformation));
}
// Save image
previewImage->save(QString::fromStdString(path), format.c_str(), quality);
}
// Thumbnail the frame image to the specified path. The image format is determined from the extension (i.e. image.PNG, image.JPEG)
void Frame::Thumbnail(std::string path, int new_width, int new_height, std::string mask_path, std::string overlay_path,
std::string background_color, bool ignore_aspect, std::string format, int quality, float rotate) {
// Create blank thumbnail image & fill background color
auto thumbnail = std::make_shared<QImage>(
new_width, new_height, QImage::Format_RGBA8888_Premultiplied);
thumbnail->fill(QColor(QString::fromStdString(background_color)));
// Create painter
QPainter painter(thumbnail.get());
painter.setRenderHints(QPainter::Antialiasing | QPainter::SmoothPixmapTransform | QPainter::TextAntialiasing, true);
// Get preview image
std::shared_ptr<QImage> previewImage = GetImage();
// Update the image to reflect the correct pixel aspect ration (i.e. to fix non-squar pixels)
if (pixel_ratio.num != 1 || pixel_ratio.den != 1)
{
// Calculate correct DAR (display aspect ratio)
int aspect_width = previewImage->size().width();
int aspect_height = previewImage->size().height() * pixel_ratio.Reciprocal().ToDouble();
// Resize to fix DAR
previewImage = std::make_shared<QImage>(previewImage->scaled(
aspect_width, aspect_height,
Qt::IgnoreAspectRatio, Qt::SmoothTransformation));
}
// Resize frame image
if (ignore_aspect)
// Ignore aspect ratio
previewImage = std::make_shared<QImage>(previewImage->scaled(
new_width, new_height,
Qt::IgnoreAspectRatio, Qt::SmoothTransformation));
else
// Maintain aspect ratio
previewImage = std::make_shared<QImage>(previewImage->scaled(
new_width, new_height,
Qt::KeepAspectRatio, Qt::SmoothTransformation));
// Composite frame image onto background (centered)
int x = (new_width - previewImage->size().width()) / 2.0; // center
int y = (new_height - previewImage->size().height()) / 2.0; // center
painter.setCompositionMode(QPainter::CompositionMode_SourceOver);
// Create transform and rotate (if needed)
QTransform transform;
float origin_x = previewImage->width() / 2.0;
float origin_y = previewImage->height() / 2.0;
transform.translate(origin_x, origin_y);
transform.rotate(rotate);
transform.translate(-origin_x,-origin_y);
painter.setTransform(transform);
// Draw image onto QImage
painter.drawImage(x, y, *previewImage);
// Overlay Image (if any)
if (overlay_path != "") {
// Open overlay
auto overlay = std::make_shared<QImage>();
overlay->load(QString::fromStdString(overlay_path));
// Set pixel format
overlay = std::make_shared<QImage>(
overlay->convertToFormat(QImage::Format_RGBA8888_Premultiplied));
// Resize to fit
overlay = std::make_shared<QImage>(overlay->scaled(
new_width, new_height, Qt::IgnoreAspectRatio, Qt::SmoothTransformation));
// Composite onto thumbnail
painter.setCompositionMode(QPainter::CompositionMode_SourceOver);
painter.drawImage(0, 0, *overlay);
}
// Mask Image (if any)
if (mask_path != "") {
// Open mask
auto mask = std::make_shared<QImage>();
mask->load(QString::fromStdString(mask_path));
// Set pixel format
mask = std::make_shared<QImage>(
mask->convertToFormat(QImage::Format_RGBA8888_Premultiplied));
// Resize to fit
mask = std::make_shared<QImage>(mask->scaled(
new_width, new_height, Qt::IgnoreAspectRatio, Qt::SmoothTransformation));
// Negate mask
mask->invertPixels();
// Get pixels
unsigned char *pixels = static_cast<unsigned char *>(thumbnail->bits());
const unsigned char *mask_pixels = static_cast<const unsigned char *>(mask->constBits());
// Convert the mask image to grayscale
// Loop through pixels
for (int pixel = 0, byte_index=0; pixel < new_width * new_height; pixel++, byte_index+=4)
{
// Get the RGB values from the pixel
int gray_value = qGray(mask_pixels[byte_index], mask_pixels[byte_index] + 1, mask_pixels[byte_index] + 2);
int Frame_Alpha = pixels[byte_index + 3];
int Mask_Value = constrain(Frame_Alpha - gray_value);
// Set all alpha pixels to gray value
pixels[byte_index + 3] = Mask_Value;
}
}
// End painter
painter.end();
// Save image
thumbnail->save(QString::fromStdString(path), format.c_str(), quality);
}
// Constrain a color value from 0 to 255
int Frame::constrain(int color_value)
{
// Constrain new color from 0 to 255
if (color_value < 0)
color_value = 0;
else if (color_value > 255)
color_value = 255;
return color_value;
}
void Frame::AddColor(int new_width, int new_height, std::string new_color)
{
const std::lock_guard<std::recursive_mutex> lock(addingImageMutex);
// Update parameters
width = new_width;
height = new_height;
color = new_color;
AddColor(QColor(QString::fromStdString(new_color)));
}
// Add (or replace) pixel data to the frame (based on a solid color)
void Frame::AddColor(const QColor& new_color)
{
// Create new image object, and fill with pixel data
const std::lock_guard<std::recursive_mutex> lock(addingImageMutex);
image = std::make_shared<QImage>(width, height, QImage::Format_RGBA8888_Premultiplied);
// Fill with solid color
image->fill(new_color);
has_image_data = true;
}
// Add (or replace) pixel data to the frame
void Frame::AddImage(
int new_width, int new_height, int bytes_per_pixel,
QImage::Format type, const unsigned char *pixels_)
{
if (has_image_data) {
// Delete the previous QImage
image.reset();
}
// Create new image object from pixel data
auto new_image = std::make_shared<QImage>(
pixels_,
new_width, new_height,
new_width * bytes_per_pixel,
type,
(QImageCleanupFunction) &openshot::cleanUpBuffer,
(void*) pixels_
);
AddImage(new_image);
}
// Add (or replace) pixel data to the frame
void Frame::AddImage(std::shared_ptr<QImage> new_image)
{
// Ignore blank images
if (!new_image)
return;
// assign image data
const std::lock_guard<std::recursive_mutex> lock(addingImageMutex);
image = new_image;
// Always convert to Format_RGBA8888_Premultiplied (if different)
if (image->format() != QImage::Format_RGBA8888_Premultiplied)
*image = image->convertToFormat(QImage::Format_RGBA8888_Premultiplied);
// Update height and width
width = image->width();
height = image->height();
has_image_data = true;
}
// Add (or replace) pixel data to the frame (for only the odd or even lines)
void Frame::AddImage(std::shared_ptr<QImage> new_image, bool only_odd_lines)
{
// Ignore blank new_image
if (!new_image)
return;
// Check for blank source image
if (!image) {
// Replace the blank source image
AddImage(new_image);
} else {
// Ignore image of different sizes or formats
bool ret=false;
if (image == new_image || image->size() != new_image->size()) {
ret = true;
}
else if (new_image->format() != QImage::Format_RGBA8888_Premultiplied) {
new_image = std::make_shared<QImage>(
new_image->convertToFormat(QImage::Format_RGBA8888_Premultiplied));
}
if (ret) {
return;
}
// Get the frame's image
const std::lock_guard<std::recursive_mutex> lock(addingImageMutex);
unsigned char *pixels = image->bits();
const unsigned char *new_pixels = new_image->constBits();
// Loop through the scanlines of the image (even or odd)
int start = 0;
if (only_odd_lines)
start = 1;
for (int row = start; row < image->height(); row += 2) {
int offset = row * image->bytesPerLine();
memcpy(pixels + offset, new_pixels + offset, image->bytesPerLine());
}
// Update height and width
height = image->height();
width = image->width();
has_image_data = true;
}
}
// Resize audio container to hold more (or less) samples and channels
void Frame::ResizeAudio(int channels, int length, int rate, ChannelLayout layout)
{
const std::lock_guard<std::recursive_mutex> lock(addingAudioMutex);
// Resize JUCE audio buffer
audio->setSize(channels, length, true, true, false);
channel_layout = layout;
sample_rate = rate;
// Calculate max audio sample added
max_audio_sample = length;
}
/// Set the direction of the audio buffer of this frame
void Frame::SetAudioDirection(bool is_increasing) {
if (audio && !audio_is_increasing && is_increasing) {
// Forward audio buffer
audio->reverse(0, audio->getNumSamples());
} else if (audio && audio_is_increasing && !is_increasing) {
// Reverse audio buffer
audio->reverse(0, audio->getNumSamples());
}
audio_is_increasing = is_increasing;
}
// Add audio samples to a specific channel
void Frame::AddAudio(bool replaceSamples, int destChannel, int destStartSample, const float* source, int numSamples, float gainToApplyToSource = 1.0f) {
const std::lock_guard<std::recursive_mutex> lock(addingAudioMutex);
// Clamp starting sample to 0
int destStartSampleAdjusted = max(destStartSample, 0);
// Extend audio container to hold more (or less) samples and channels.. if needed
int new_length = destStartSampleAdjusted + numSamples;
int new_channel_length = audio->getNumChannels();
if (destChannel >= new_channel_length)
new_channel_length = destChannel + 1;
if (new_length > audio->getNumSamples() || new_channel_length > audio->getNumChannels())
audio->setSize(new_channel_length, new_length, true, true, false);
// Clear the range of samples first (if needed)
if (replaceSamples)
audio->clear(destChannel, destStartSampleAdjusted, numSamples);
// Add samples to frame's audio buffer
audio->addFrom(destChannel, destStartSampleAdjusted, source, numSamples, gainToApplyToSource);
has_audio_data = true;
// Calculate max audio sample added
if (new_length > max_audio_sample)
max_audio_sample = new_length;
// Reset audio direction
audio_is_increasing = true;
}
// Apply gain ramp (i.e. fading volume)
void Frame::ApplyGainRamp(int destChannel, int destStartSample, int numSamples, float initial_gain = 0.0f, float final_gain = 1.0f)
{
const std::lock_guard<std::recursive_mutex> lock(addingAudioMutex);
// Apply gain ramp
audio->applyGainRamp(destChannel, destStartSample, numSamples, initial_gain, final_gain);
}
// Get pointer to Magick++ image object
std::shared_ptr<QImage> Frame::GetImage()
{
// Check for blank image
if (!image)
// Fill with black
AddColor(width, height, color);
return image;
}
#ifdef USE_OPENCV
// Convert Qimage to Mat
cv::Mat Frame::Qimage2mat( std::shared_ptr<QImage>& qimage) {
cv::Mat mat = cv::Mat(qimage->height(), qimage->width(), CV_8UC4, (uchar*)qimage->constBits(), qimage->bytesPerLine()).clone();
cv::Mat mat2 = cv::Mat(mat.rows, mat.cols, CV_8UC3 );
int from_to[] = { 0,0, 1,1, 2,2 };
cv::mixChannels( &mat, 1, &mat2, 1, from_to, 3 );
cv::cvtColor(mat2, mat2, cv::COLOR_RGB2BGR);
return mat2;
}
// Get pointer to OpenCV image object
cv::Mat Frame::GetImageCV()
{
// Check for blank image
if (!image)
// Fill with black
AddColor(width, height, color);
// if (imagecv.empty())
// Convert Qimage to Mat
imagecv = Qimage2mat(image);
return imagecv;
}
std::shared_ptr<QImage> Frame::Mat2Qimage(cv::Mat img){
cv::cvtColor(img, img, cv::COLOR_BGR2RGB);
QImage qimg((uchar*) img.data, img.cols, img.rows, img.step, QImage::Format_RGB888);
std::shared_ptr<QImage> imgIn = std::make_shared<QImage>(qimg.copy());
// Always convert to RGBA8888 (if different)
if (imgIn->format() != QImage::Format_RGBA8888_Premultiplied)
*imgIn = imgIn->convertToFormat(QImage::Format_RGBA8888_Premultiplied);
return imgIn;
}
// Set pointer to OpenCV image object
void Frame::SetImageCV(cv::Mat _image)
{
imagecv = _image;
image = Mat2Qimage(_image);
}
#endif
// Play audio samples for this frame
void Frame::Play()
{
// Check if samples are present
if (!GetAudioSamplesCount())
return;
juce::AudioDeviceManager deviceManager;
juce::String error = deviceManager.initialise (
0, /* number of input channels */
2, /* number of output channels */
0, /* no XML settings.. */
true /* select default device on failure */);
// Output error (if any)
if (error.isNotEmpty()) {
cout << "Error on initialise(): " << error << endl;
}
juce::AudioSourcePlayer audioSourcePlayer;
deviceManager.addAudioCallback (&audioSourcePlayer);
std::unique_ptr<AudioBufferSource> my_source;
my_source.reset (new AudioBufferSource (audio.get()));
// Create TimeSliceThread for audio buffering
juce::TimeSliceThread my_thread("Audio buffer thread");
// Start thread
my_thread.startThread();
juce::AudioTransportSource transport1;
transport1.setSource (my_source.get(),
5000, // tells it to buffer this many samples ahead
&my_thread,
(double) sample_rate,
audio->getNumChannels()); // sample rate of source
transport1.setPosition (0);
transport1.setGain(1.0);
// Create MIXER
juce::MixerAudioSource mixer;
mixer.addInputSource(&transport1, false);
audioSourcePlayer.setSource (&mixer);
// Start transports
transport1.start();
while (transport1.isPlaying())
{
cout << "playing" << endl;
std::this_thread::sleep_for(std::chrono::seconds(1));
}
cout << "DONE!!!" << endl;
transport1.stop();
transport1.setSource (0);
audioSourcePlayer.setSource (0);
my_thread.stopThread(500);
deviceManager.removeAudioCallback (&audioSourcePlayer);
deviceManager.closeAudioDevice();
deviceManager.removeAllChangeListeners();
deviceManager.dispatchPendingMessages();
cout << "End of Play()" << endl;
}
// Add audio silence
void Frame::AddAudioSilence(int numSamples)
{
const std::lock_guard<std::recursive_mutex> lock(addingAudioMutex);
// Resize audio container
audio->setSize(channels, numSamples, false, true, false);
audio->clear();
has_audio_data = true;
// Calculate max audio sample added
max_audio_sample = numSamples;
// Reset audio direction
audio_is_increasing = true;
}