#include "../include/CVStabilization.h" CVStabilization::CVStabilization():smoothingWindow(30) {} CVStabilization::CVStabilization(int _smoothingWindow): smoothingWindow(_smoothingWindow){} // void CVStabilization::ProcessVideo(openshot::Clip &video){ // // Make sure Clip is opened // video.Open(); // // Get total number of frames // int videoLenght = video.Reader()->info.video_length; // // Get first Opencv image // std::shared_ptr f = video.GetFrame(0); // cv::Mat prev = f->GetImageCV(); // // OpticalFlow works with grayscale images // cv::cvtColor(prev, prev_grey, cv::COLOR_BGR2GRAY); // // Extract and track opticalflow features for each frame // for (long int frame_number = 1; frame_number <= videoLenght; frame_number++) // { // std::shared_ptr f = video.GetFrame(frame_number); // // Grab Mat image // cv::Mat cvimage = f->GetImageCV(); // cv::cvtColor(cvimage, cvimage, cv::COLOR_RGB2GRAY); // TrackFrameFeatures(cvimage, frame_number); // } // vector trajectory = ComputeFramesTrajectory(); // vector smoothed_trajectory = SmoothTrajectory(trajectory); // vector new_prev_to_cur_transform = GenNewCamPosition(smoothed_trajectory); // ApplyNewTrajectoryToClip(video, new_prev_to_cur_transform); // } // Track current frame features and find the relative transformation void CVStabilization::TrackFrameFeatures(cv::Mat frame, int frameNum){ if(prev_grey.empty()){ prev_grey = frame; return; } // OpticalFlow features vector vector prev_corner, cur_corner; vector prev_corner2, cur_corner2; vector status; vector err; // Extract new image teatures cv::goodFeaturesToTrack(prev_grey, prev_corner, 200, 0.01, 30); // Track features cv::calcOpticalFlowPyrLK(prev_grey, frame, prev_corner, cur_corner, status, err); // Remove untracked features for(size_t i=0; i < status.size(); i++) { if(status[i]) { prev_corner2.push_back(prev_corner[i]); cur_corner2.push_back(cur_corner[i]); } } // translation + rotation only cv::Mat T = estimateRigidTransform(prev_corner2, cur_corner2, false); // false = rigid transform, no scaling/shearing // If no transform is found. We'll just use the last known good transform. if(T.data == NULL) { last_T.copyTo(T); } T.copyTo(last_T); // decompose T double dx = T.at(0,2); double dy = T.at(1,2); double da = atan2(T.at(1,0), T.at(0,0)); prev_to_cur_transform.push_back(TransformParam(dx, dy, da)); // out_transform << frameNum << " " << dx << " " << dy << " " << da << endl; cur.copyTo(prev); frame.copyTo(prev_grey); cout << "Frame: " << frameNum << " - good optical flow: " << prev_corner2.size() << endl; } vector CVStabilization::ComputeFramesTrajectory(){ // Accumulated frame to frame transform double a = 0; double x = 0; double y = 0; vector trajectory; // trajectory at all frames // Compute global camera trajectory. First frame is the origin for(size_t i=0; i < prev_to_cur_transform.size(); i++) { x += prev_to_cur_transform[i].dx; y += prev_to_cur_transform[i].dy; a += prev_to_cur_transform[i].da; trajectory.push_back(CamTrajectory(x,y,a)); // out_trajectory << (i+1) << " " << x << " " << y << " " << a << endl; } return trajectory; } vector CVStabilization::SmoothTrajectory(vector &trajectory){ vector smoothed_trajectory; // trajectory at all frames for(size_t i=0; i < trajectory.size(); i++) { double sum_x = 0; double sum_y = 0; double sum_a = 0; int count = 0; for(int j=-smoothingWindow; j <= smoothingWindow; j++) { if(i+j >= 0 && i+j < trajectory.size()) { sum_x += trajectory[i+j].x; sum_y += trajectory[i+j].y; sum_a += trajectory[i+j].a; count++; } } double avg_a = sum_a / count; double avg_x = sum_x / count; double avg_y = sum_y / count; smoothed_trajectory.push_back(CamTrajectory(avg_x, avg_y, avg_a)); // out_smoothed_trajectory << (i+1) << " " << avg_x << " " << avg_y << " " << avg_a << endl; } return smoothed_trajectory; } // Generate new transformations parameters for each frame to follow the smoothed trajectory vector CVStabilization::GenNewCamPosition(vector & smoothed_trajectory){ vector new_prev_to_cur_transform; // Accumulated frame to frame transform double a = 0; double x = 0; double y = 0; for(size_t i=0; i < prev_to_cur_transform.size(); i++) { x += prev_to_cur_transform[i].dx; y += prev_to_cur_transform[i].dy; a += prev_to_cur_transform[i].da; // target - current double diff_x = smoothed_trajectory[i].x - x; double diff_y = smoothed_trajectory[i].y - y; double diff_a = smoothed_trajectory[i].a - a; double dx = prev_to_cur_transform[i].dx + diff_x; double dy = prev_to_cur_transform[i].dy + diff_y; double da = prev_to_cur_transform[i].da + diff_a; new_prev_to_cur_transform.push_back(TransformParam(dx, dy, da)); // out_new_transform << (i+1) << " " << dx << " " << dy << " " << da << endl; } return new_prev_to_cur_transform; } // // Send smoothed camera transformation to be applyed on clip // void CVStabilization::ApplyNewTrajectoryToClip(openshot::Clip &video, vector &new_prev_to_cur_transform){ // video.new_prev_to_cur_transform = new_prev_to_cur_transform; // video.hasStabilization = true; // }