/** * @file * @brief Unit tests for openshot::Clip * @author Jonathan Thomas * * @ref License */ // Copyright (c) 2008-2019 OpenShot Studios, LLC // // SPDX-License-Identifier: LGPL-3.0-or-later #include #include #include "openshot_catch.h" #include #include #include #include "Clip.h" #include "DummyReader.h" #include "Enums.h" #include "Exceptions.h" #include "FFmpegReader.h" #include "Frame.h" #include "Fraction.h" #include "FrameMapper.h" #include "Timeline.h" #include "Json.h" #include "effects/Negate.h" using namespace openshot; TEST_CASE( "default constructor", "[libopenshot][clip]" ) { // Create a empty clip Clip c1; // Check basic settings CHECK(c1.anchor == ANCHOR_CANVAS); CHECK(c1.gravity == GRAVITY_CENTER); CHECK(c1.scale == SCALE_FIT); CHECK(c1.composite == COMPOSITE_SOURCE_OVER); CHECK(c1.Layer() == 0); CHECK(c1.Position() == Approx(0.0f).margin(0.00001)); CHECK(c1.Start() == Approx(0.0f).margin(0.00001)); CHECK(c1.End() == Approx(0.0f).margin(0.00001)); } TEST_CASE( "path string constructor", "[libopenshot][clip]" ) { // Create a empty clip std::stringstream path; path << TEST_MEDIA_PATH << "piano.wav"; Clip c1(path.str()); c1.Open(); // Check basic settings CHECK(c1.anchor == ANCHOR_CANVAS); CHECK(c1.gravity == GRAVITY_CENTER); CHECK(c1.scale == SCALE_FIT); CHECK(c1.composite == COMPOSITE_SOURCE_OVER); CHECK(c1.Layer() == 0); CHECK(c1.Position() == Approx(0.0f).margin(0.00001)); CHECK(c1.Start() == Approx(0.0f).margin(0.00001)); CHECK(c1.End() == Approx(4.39937f).margin(0.00001)); } TEST_CASE( "basic getters and setters", "[libopenshot][clip]" ) { // Create a empty clip Clip c1; // Check basic settings CHECK_THROWS_AS(c1.Open(), ReaderClosed); CHECK(c1.anchor == ANCHOR_CANVAS); CHECK(c1.gravity == GRAVITY_CENTER); CHECK(c1.scale == SCALE_FIT); CHECK(c1.composite == COMPOSITE_SOURCE_OVER); CHECK(c1.Layer() == 0); CHECK(c1.Position() == Approx(0.0f).margin(0.00001)); CHECK(c1.Start() == Approx(0.0f).margin(0.00001)); CHECK(c1.End() == Approx(0.0f).margin(0.00001)); // Change some properties c1.Layer(1); c1.Position(5.0); c1.Start(3.5); c1.End(10.5); CHECK(c1.Layer() == 1); CHECK(c1.Position() == Approx(5.0f).margin(0.00001)); CHECK(c1.Start() == Approx(3.5f).margin(0.00001)); CHECK(c1.End() == Approx(10.5f).margin(0.00001)); } TEST_CASE( "properties", "[libopenshot][clip]" ) { // Create a empty clip Clip c1; // Change some properties c1.Layer(1); c1.Position(5.0); c1.Start(3.5); c1.End(10.5); c1.alpha.AddPoint(1, 1.0); c1.alpha.AddPoint(500, 0.0); // Get properties JSON string at frame 1 std::string properties = c1.PropertiesJSON(1); // Parse JSON string into JSON objects Json::Value root; Json::CharReaderBuilder rbuilder; Json::CharReader* reader(rbuilder.newCharReader()); std::string errors; bool success = reader->parse( properties.c_str(), properties.c_str() + properties.size(), &root, &errors ); CHECK(success == true); // Check for specific things CHECK(root["alpha"]["value"].asDouble() == Approx(1.0f).margin(0.01)); CHECK(root["alpha"]["keyframe"].asBool() == true); // Get properties JSON string at frame 250 properties = c1.PropertiesJSON(250); // Parse JSON string into JSON objects root.clear(); success = reader->parse( properties.c_str(), properties.c_str() + properties.size(), &root, &errors ); CHECK(success == true); // Check for specific things CHECK(root["alpha"]["value"].asDouble() == Approx(0.5f).margin(0.01)); CHECK_FALSE(root["alpha"]["keyframe"].asBool()); // Get properties JSON string at frame 250 (again) properties = c1.PropertiesJSON(250); // Parse JSON string into JSON objects root.clear(); success = reader->parse( properties.c_str(), properties.c_str() + properties.size(), &root, &errors ); CHECK(success == true); // Check for specific things CHECK_FALSE(root["alpha"]["keyframe"].asBool()); // Get properties JSON string at frame 500 properties = c1.PropertiesJSON(500); // Parse JSON string into JSON objects root.clear(); success = reader->parse( properties.c_str(), properties.c_str() + properties.size(), &root, &errors ); CHECK(success == true); // Check for specific things CHECK(root["alpha"]["value"].asDouble() == Approx(0.0f).margin(0.00001)); CHECK(root["alpha"]["keyframe"].asBool() == true); // Free up the reader we allocated delete reader; } TEST_CASE( "effects", "[libopenshot][clip]" ) { // Load clip with video std::stringstream path; path << TEST_MEDIA_PATH << "sintel_trailer-720p.mp4"; Clip c10(path.str()); c10.Open(); Negate n; c10.AddEffect(&n); // Get frame 1 std::shared_ptr f = c10.GetFrame(500); // Get the image data const unsigned char* pixels = f->GetPixels(10); int pixel_index = 112 * 4; // pixel 112 (4 bytes per pixel) // Check image properties on scanline 10, pixel 112 CHECK((int)pixels[pixel_index] == 255); CHECK((int)pixels[pixel_index + 1] == 255); CHECK((int)pixels[pixel_index + 2] == 255); CHECK((int)pixels[pixel_index + 3] == 255); // Check the # of Effects CHECK((int)c10.Effects().size() == 1); // Add a 2nd negate effect Negate n1; c10.AddEffect(&n1); // Get frame 1 f = c10.GetFrame(500); // Get the image data pixels = f->GetPixels(10); pixel_index = 112 * 4; // pixel 112 (4 bytes per pixel) // Check image properties on scanline 10, pixel 112 CHECK((int)pixels[pixel_index] == 0); CHECK((int)pixels[pixel_index + 1] == 0); CHECK((int)pixels[pixel_index + 2] == 0); CHECK((int)pixels[pixel_index + 3] == 255); // Check the # of Effects CHECK((int)c10.Effects().size() == 2); } TEST_CASE( "verify parent Timeline", "[libopenshot][clip]" ) { Timeline t1(640, 480, Fraction(30,1), 44100, 2, LAYOUT_STEREO); // Load clip with video std::stringstream path; path << TEST_MEDIA_PATH << "sintel_trailer-720p.mp4"; Clip c1(path.str()); c1.Open(); // Check size of frame image CHECK(1280 == c1.GetFrame(1)->GetImage()->width()); CHECK(720 == c1.GetFrame(1)->GetImage()->height()); // Add clip to timeline t1.AddClip(&c1); // Check size of frame image (with an associated timeline) CHECK(640 == c1.GetFrame(1)->GetImage()->width()); CHECK(360 == c1.GetFrame(1)->GetImage()->height()); } TEST_CASE( "has_video", "[libopenshot][clip]" ) { std::stringstream path; path << TEST_MEDIA_PATH << "sintel_trailer-720p.mp4"; openshot::Clip c1(path.str()); c1.has_video.AddPoint(1.0, 0.0); c1.has_video.AddPoint(5.0, -1.0, openshot::CONSTANT); c1.has_video.AddPoint(10.0, 1.0, openshot::CONSTANT); c1.Open(); auto trans_color = QColor(Qt::transparent); auto f1 = c1.GetFrame(1); CHECK(f1->has_image_data); auto f2 = c1.GetFrame(5); CHECK(f2->has_image_data); auto f3 = c1.GetFrame(5); CHECK(f3->has_image_data); auto i1 = f1->GetImage(); QSize f1_size(f1->GetWidth(), f1->GetHeight()); CHECK(i1->size() == f1_size); CHECK(i1->pixelColor(20, 20) == trans_color); auto i2 = f2->GetImage(); QSize f2_size(f2->GetWidth(), f2->GetHeight()); CHECK(i2->size() == f2_size); CHECK(i2->pixelColor(20, 20) != trans_color); auto i3 = f3->GetImage(); QSize f3_size(f3->GetWidth(), f3->GetHeight()); CHECK(i3->size() == f3_size); CHECK(i3->pixelColor(20, 20) != trans_color); } TEST_CASE( "access frames past reader length", "[libopenshot][clip]" ) { // Create cache object to hold test frames openshot::CacheMemory cache; // Let's create some test frames for (int64_t frame_number = 1; frame_number <= 30; frame_number++) { // Create blank frame (with specific frame #, samples, and channels) // Sample count should be 44100 / 30 fps = 1470 samples per frame int sample_count = 1470; auto f = std::make_shared(frame_number, sample_count, 2); // Create test samples with incrementing value float *audio_buffer = new float[sample_count]; for (int64_t sample_number = 0; sample_number < sample_count; sample_number++) { // Generate an incrementing audio sample value (just as an example) audio_buffer[sample_number] = float(frame_number) + (float(sample_number) / float(sample_count)); } // Add custom audio samples to Frame (bool replaceSamples, int destChannel, int destStartSample, const float* source, f->AddAudio(true, 0, 0, audio_buffer, sample_count, 1.0); // add channel 1 f->AddAudio(true, 1, 0, audio_buffer, sample_count, 1.0); // add channel 2 // Add test frame to dummy reader cache.Add(f); delete[] audio_buffer; } // Create a dummy reader, with a pre-existing cache openshot::DummyReader r(openshot::Fraction(30, 1), 1920, 1080, 44100, 2, 1.0, &cache); r.Open(); // Open the reader openshot::Clip c1; c1.Reader(&r); c1.Open(); // Get the last valid frame # std::shared_ptr frame = c1.GetFrame(30); CHECK(frame->GetAudioSamples(0)[0] == Approx(30.0).margin(0.00001)); CHECK(frame->GetAudioSamples(0)[600] == Approx(30.4081631).margin(0.00001)); CHECK(frame->GetAudioSamples(0)[1200] == Approx(30.8163261).margin(0.00001)); // Get the +1 past the end of the reader (should be audio silence) frame = c1.GetFrame(31); CHECK(frame->GetAudioSamples(0)[0] == Approx(0.0).margin(0.00001)); CHECK(frame->GetAudioSamples(0)[600] == Approx(0.0).margin(0.00001)); CHECK(frame->GetAudioSamples(0)[1200] == Approx(0.0).margin(0.00001)); // Get the +2 past the end of the reader (should be audio silence) frame = c1.GetFrame(32); CHECK(frame->GetAudioSamples(0)[0] == Approx(0.0).margin(0.00001)); CHECK(frame->GetAudioSamples(0)[600] == Approx(0.0).margin(0.00001)); CHECK(frame->GetAudioSamples(0)[1200] == Approx(0.0).margin(0.00001)); } TEST_CASE( "setting and clobbering readers", "[libopenshot][clip]" ) { // Create a dummy reader #1, with a pre-existing cache openshot::DummyReader r1(openshot::Fraction(24, 1), 1920, 1080, 44100, 2, 1.0); r1.Open(); // Open the reader // Create a dummy reader #2, with a pre-existing cache openshot::DummyReader r2(openshot::Fraction(30, 1), 1920, 1080, 44100, 2, 1.0); r2.Open(); // Open the reader // Create a clip with constructor (and an allocated internal reader A) std::stringstream path; path << TEST_MEDIA_PATH << "piano.wav"; Clip c1(path.str()); c1.Open(); // Clobber allocated reader A with reader #1 c1.Reader(&r1); // Clobber reader #1 with reader #2 c1.Reader(&r2); // Clobber reader #2 with SetJson (allocated reader B) c1.SetJson("{\"reader\":{\"acodec\":\"raw\",\"audio_bit_rate\":0,\"audio_stream_index\":-1,\"audio_timebase\":{\"den\":1,\"num\":1},\"channel_layout\":4,\"channels\":2,\"display_ratio\":{\"den\":9,\"num\":16},\"duration\":1.0,\"file_size\":\"8294400\",\"fps\":{\"den\":1,\"num\":30},\"has_audio\":false,\"has_single_image\":false,\"has_video\":true,\"height\":1080,\"interlaced_frame\":false,\"metadata\":{},\"pixel_format\":-1,\"pixel_ratio\":{\"den\":1,\"num\":1},\"sample_rate\":44100,\"top_field_first\":true,\"type\":\"DummyReader\",\"vcodec\":\"raw\",\"video_bit_rate\":0,\"video_length\":\"30\",\"video_stream_index\":-1,\"video_timebase\":{\"den\":30,\"num\":1},\"width\":1920}}"); // Clobber allocated reader B with reader 2 c1.Reader(&r2); // Clobber reader 2 with reader 1 c1.Reader(&r1); } TEST_CASE( "time remapping", "[libopenshot][clip]" ) { Fraction fps(23,1); Timeline t1(640, 480, fps, 44100, 2, LAYOUT_STEREO); // Load clip with video std::stringstream path; path << TEST_MEDIA_PATH << "piano.wav"; Clip clip(path.str()); int original_video_length = clip.Reader()->info.video_length; clip.Position(0.0); clip.Start(0.0); // Set time keyframe (4X speed REVERSE) clip.time.AddPoint(1, original_video_length, openshot::LINEAR); clip.time.AddPoint(original_video_length, 1.0, openshot::LINEAR); // TODO: clip.Duration() != clip.Reader->info.duration // Set clip length based on time-values if (clip.time.GetLength() > 1) { clip.End(clip.time.GetLength() / fps.ToDouble()); } else { clip.End(clip.Reader()->info.duration); } // Add clip t1.AddClip(&clip); t1.Open(); // Get frame int64_t clip_start_frame = (clip.Position() * fps.ToDouble()) + 1; int64_t clip_end_frame = clip_start_frame + clip.time.GetLength(); if (clip.time.GetLength() == 1) { clip_end_frame = clip_start_frame + (clip.Duration() * fps.ToDouble()); } // Loop through frames for (int64_t frame = clip_start_frame; frame <= clip_end_frame; frame++) { int expected_sample_count = Frame::GetSamplesPerFrame(frame, t1.info.fps, t1.info.sample_rate, t1.info.channels); std::shared_ptr f = t1.GetFrame(frame); CHECK(expected_sample_count == f->GetAudioSamplesCount()); } // Clear cache t1.ClearAllCache(true); // Loop again through frames // Time-remapping should start over (detect a gap) for (int64_t frame = clip_start_frame; frame <= clip_end_frame; frame++) { int expected_sample_count = Frame::GetSamplesPerFrame(frame, t1.info.fps, t1.info.sample_rate, t1.info.channels); std::shared_ptr f = t1.GetFrame(frame); CHECK(expected_sample_count == f->GetAudioSamplesCount()); } t1.Close(); } TEST_CASE( "resample_audio_8000_to_48000_reverse", "[libopenshot][clip]" ) { // Create a reader std::stringstream path; path << TEST_MEDIA_PATH << "sine.wav"; openshot::FFmpegReader reader(path.str(), true); // Map to 24 fps, 2 channels stereo, 44100 sample rate FrameMapper map(&reader, Fraction(24,1), PULLDOWN_NONE, 48000, 2, LAYOUT_STEREO); map.Open(); Clip clip; clip.Reader(&map); clip.Open(); int original_video_length = clip.Reader()->info.video_length; clip.Position(0.0); clip.Start(0.0); // Set time keyframe (REVERSE direction using bezier curve) clip.time.AddPoint(1, original_video_length, openshot::LINEAR); clip.time.AddPoint(original_video_length, 1.0, openshot::BEZIER); // Loop again through frames // Time-remapping should start over (detect a gap) for (int64_t frame = 1; frame <= original_video_length; frame++) { int expected_sample_count = Frame::GetSamplesPerFrame(frame, map.info.fps, map.info.sample_rate, map.info.channels); std::shared_ptr f = clip.GetFrame(frame); CHECK(expected_sample_count == f->GetAudioSamplesCount()); } // Clear clip cache clip.GetCache()->Clear(); // Loop again through frames // Time-remapping should start over (detect a gap) for (int64_t frame = 1; frame < original_video_length; frame++) { int expected_sample_count = Frame::GetSamplesPerFrame(frame, map.info.fps, map.info.sample_rate, map.info.channels); std::shared_ptr f = clip.GetFrame(frame); CHECK(expected_sample_count == f->GetAudioSamplesCount()); } // Close mapper map.Close(); reader.Close(); clip.Close(); } // ----------------------------------------------------------------------------- // Additional tests validating PR changes: // - safe extension parsing (no dot in path) // - painter-based opacity behavior (no per-pixel mutation) // - transform/scaling path sanity (conditional render hint use) // ----------------------------------------------------------------------------- TEST_CASE( "safe_extension_parsing_no_dot", "[libopenshot][clip][pr]" ) { // Constructing a Clip with a path that has no dot used to risk UB in get_file_extension(); // This should now be safe and simply result in no reader being set. openshot::Clip c1("this_is_not_a_real_path_and_has_no_extension"); // Reader() should throw since no reader could be inferred. CHECK_THROWS_AS(c1.Reader(), openshot::ReaderClosed); // Opening also throws (consistent with other tests for unopened readers). CHECK_THROWS_AS(c1.Open(), openshot::ReaderClosed); } TEST_CASE( "painter_opacity_applied_no_per_pixel_mutation", "[libopenshot][clip][pr]" ) { // Build a red frame via DummyReader (no copies/assignments of DummyReader) openshot::CacheMemory cache; auto f = std::make_shared(1, 80, 60, "#000000", 0, 2); f->AddColor(QColor(Qt::red)); // opaque red cache.Add(f); openshot::DummyReader dummy(openshot::Fraction(30,1), 80, 60, 44100, 2, 1.0, &cache); dummy.Open(); // Clip that uses the dummy reader openshot::Clip clip; clip.Reader(&dummy); clip.Open(); // Alpha 0.5 at frame 1 (exercise painter.setOpacity path) clip.alpha.AddPoint(1, 0.5); clip.display = openshot::FRAME_DISPLAY_NONE; // avoid font/overlay variability // Render frame 1 (no timeline needed for this check) std::shared_ptr out_f = clip.GetFrame(1); auto img = out_f->GetImage(); REQUIRE(img); // must exist REQUIRE(img->format() == QImage::Format_RGBA8888_Premultiplied); // Pixel well inside the image should be "half-transparent red" over transparent bg. // In Qt, pixelColor() returns unpremultiplied values, so expect alpha ≈ 127 and red ≈ 255. QColor p = img->pixelColor(70, 50); CHECK(p.alpha() == Approx(127).margin(10)); CHECK(p.red() == Approx(255).margin(2)); CHECK(p.green() == Approx(0).margin(2)); CHECK(p.blue() == Approx(0).margin(2)); } TEST_CASE( "composite_over_opaque_background_blend", "[libopenshot][clip][pr]" ) { // Red source clip frame (fully opaque) openshot::CacheMemory cache; auto f = std::make_shared(1, 64, 64, "#000000", 0, 2); f->AddColor(QColor(Qt::red)); cache.Add(f); openshot::DummyReader dummy(openshot::Fraction(30,1), 64, 64, 44100, 2, 1.0, &cache); dummy.Open(); openshot::Clip clip; clip.Reader(&dummy); clip.Open(); // Make clip semi-transparent via alpha (0.5) clip.alpha.AddPoint(1, 0.5); clip.display = openshot::FRAME_DISPLAY_NONE; // no overlay here // Build a blue, fully-opaque background frame and composite into it auto bg = std::make_shared(1, 64, 64, "#000000", 0, 2); bg->AddColor(QColor(Qt::blue)); // blue background, opaque // Composite the clip onto bg std::shared_ptr out = clip.GetFrame(bg, /*clip_frame_number*/1); auto img = out->GetImage(); REQUIRE(img); // Center pixel should be purple-ish and fully opaque (red over blue @ 50% -> roughly (127,0,127), A=255) QColor center = img->pixelColor(32, 32); CHECK(center.alpha() == Approx(255).margin(0)); CHECK(center.red() == Approx(127).margin(12)); CHECK(center.green() == Approx(0).margin(6)); CHECK(center.blue() == Approx(127).margin(12)); } TEST_CASE( "transform_path_identity_vs_scaled", "[libopenshot][clip][pr]" ) { // Create a small checker-ish image to make scaling detectable const int W = 60, H = 40; QImage src(W, H, QImage::Format_RGBA8888_Premultiplied); src.fill(QColor(Qt::black)); { QPainter p(&src); p.setPen(QColor(Qt::white)); for (int x = 0; x < W; x += 4) p.drawLine(x, 0, x, H-1); for (int y = 0; y < H; y += 4) p.drawLine(0, y, W-1, y); } // Stuff the image into a Frame -> Cache -> DummyReader openshot::CacheMemory cache; auto f = std::make_shared(1, W, H, "#000000", 0, 2); f->AddImage(std::make_shared(src)); cache.Add(f); openshot::DummyReader dummy(openshot::Fraction(30,1), W, H, 44100, 2, 1.0, &cache); dummy.Open(); openshot::Clip clip; clip.Reader(&dummy); clip.Open(); // Helper lambda to count "near-white" pixels in a region (for debug/metrics) auto count_white = [](const QImage& im, int x0, int y0, int x1, int y1)->int { int cnt = 0; for (int y = y0; y <= y1; ++y) { for (int x = x0; x <= x1; ++x) { QColor c = im.pixelColor(x, y); if (c.red() > 240 && c.green() > 240 && c.blue() > 240) ++cnt; } } return cnt; }; // Helper lambda to compute per-pixel difference count between two images auto diff_count = [](const QImage& a, const QImage& b, int x0, int y0, int x1, int y1)->int { int cnt = 0; for (int y = y0; y <= y1; ++y) { for (int x = x0; x <= x1; ++x) { QColor ca = a.pixelColor(x, y); QColor cb = b.pixelColor(x, y); int dr = std::abs(ca.red() - cb.red()); int dg = std::abs(ca.green() - cb.green()); int db = std::abs(ca.blue() - cb.blue()); // treat any noticeable RGB change as a difference if ((dr + dg + db) > 24) ++cnt; } } return cnt; }; // Case A: Identity transform (no move/scale/rotate). Output should match source at a white grid point. std::shared_ptr out_identity; { clip.scale_x = openshot::Keyframe(1.0); clip.scale_y = openshot::Keyframe(1.0); clip.rotation = openshot::Keyframe(0.0); clip.location_x = openshot::Keyframe(0.0); clip.location_y = openshot::Keyframe(0.0); clip.display = openshot::FRAME_DISPLAY_NONE; out_identity = clip.GetFrame(1); auto img = out_identity->GetImage(); REQUIRE(img); // Pick a mid pixel that is white in the grid (multiple of 4) QColor c = img->pixelColor(20, 20); CHECK(c.red() >= 240); CHECK(c.green() >= 240); CHECK(c.blue() >= 240); } // Case B: Downscale (trigger transform path). Clear the clip cache so we don't // accidentally re-use the identity frame from final_cache. { clip.GetCache()->Clear(); // **critical fix** ensure recompute after keyframe changes // Force a downscale to half clip.scale_x = openshot::Keyframe(0.5); clip.scale_y = openshot::Keyframe(0.5); clip.rotation = openshot::Keyframe(0.0); clip.location_x = openshot::Keyframe(0.0); clip.location_y = openshot::Keyframe(0.0); clip.display = openshot::FRAME_DISPLAY_NONE; auto out_scaled = clip.GetFrame(1); auto img_scaled = out_scaled->GetImage(); REQUIRE(img_scaled); // Measure difference vs identity in a central region to avoid edges const int x0 = 8, y0 = 8, x1 = W - 9, y1 = H - 9; int changed = diff_count(*out_identity->GetImage(), *img_scaled, x0, y0, x1, y1); // After scaling, the image must not be identical to identity output. // Using a minimal check keeps this robust across Qt versions and platforms. CHECK(changed > 0); // Optional diagnostic: scaled typically yields <= number of pure whites vs identity. int white_id = count_white(*out_identity->GetImage(), x0, y0, x1, y1); int white_sc = count_white(*img_scaled, x0, y0, x1, y1); CHECK(white_sc <= white_id); } }