Remote Editing Solutions

If you are an independent editor or the manager of a small to medium post facility, then you’ve likely wrestled with the WFH dilemma. Work-from-home, i.e. remote editing, has been on the minds of many. It’s been accelerated for sure by Covid-19, but that’s not the sole reason. There are numerous viable solutions and one size does not fit all. I take a closer look at various workflow options, along with a dive into the use of one popular and cost-effective solution – Jump Desktop. It’s all at Pro Video Coalition at the link below.

Real Remote Editing with Jump Desktop

©2021 Oliver Peters

Easy Resolve Grading with 6 Nodes

Spend any time watching Resolve tutorials and you’ll see many different ways in which colorists approach the creation of the same looks. Some create a look with just a few simple nodes. Others build a seemingly convoluted node tree designed to achieve the same goal. Neither approach is right or wrong.

Often what can all be done in a single node is spread across several in order to easily trace back through your steps when changes are needed. It also makes it easy to compare the impact of a correction by enabling and disabling a node. A series of nodes applied to a clip can be saved as a PowerGrade, which is a node preset. PowerGrades can be set up for a certain look or can be populated with blank (unaltered) nodes that are organized for how you like to work. Individual nodes can also be labeled, so that it’s easy to remember what operation you will do in each node.

The following is a simple PowerGrade (node sequence) that can be used as a starting point for most color grading work. It’s based on using log footage, but can also be modified for camera RAW or recordings in non-log color spaces, like Rec 709. These nodes are designed as a simple operational sequence to follow and each step can be used in a manner that works best with your footage. The sample ARRI clip was recorded with an ALEXA camera using the Log-C color profile.

Node 2 (LUT) – This is the starting point, because the first thing I want to do is apply the proper camera LUT to transform the image out of log. You could also do this with manual grading (no LUT). In that case the first three nodes would be rolled into one. Alternately you may use a Color Space Transform effect or even a Dehaze effect in some cases. But for the projects I grade, which largely use ARRI, Panasonic, Canon, and Sony cameras, adding the proper LUT seems to be the best starting point.

Node 1 (Contrast/Saturation) – With the LUT added to Node 2, I will go back to Node 1 to adjust contrast, pivot, and saturation. This changes the image going into the LUT and is a bit like adjusting the volume gain stage prior to applying an effect or filter when mixing sound. Since LUTs affect how color is treated, I will rarely adjust color balance or hue offsets (color wheels) in Node 1, as it may skew what the LUT is doing to the image in Node 2. The objective is to make subtle adjustments in Node 1 that improve the natural result coming out of Node 2.

Node 3 (Primary Correction) – This node is where you’ll want to correct color temperature/tint and use the color wheels, RGB curves,  and other controls to achieve a nice primary color correction. For example, you may need to shift color temperature warmer or cooler, lower black levels, apply a slight s-curve in the RGB curves, or adjust the overall level up or down.

Node 4 (Secondary Correction) – This node is for enhancement and the tools you’ll generally use are hue/sat curves. Let’s say you want to enhance skin tones, or the blue in the sky. Adjust the proper hue/sat curve in this node.

Node 5 (Windows) – You can add one or more “power windows” within the node (or use multiple nodes). Windows can be tracked to follow objects, but the main objective is a way to relight the scene. In most projects, I find that one window per shot is typically all I need, if any at all. Often this is to brighten up the lighting on the main talent in the shot. The use of windows is a way to direct the viewer’s attention. Often a simple soft-edged oval is all you’ll need to achieve a dramatic result.

Node 6 (Vignette) – The last node in this basic structure is to add a vignette, which I generally apply just to subtly darken the corners. This adds a bit of character to most shots. I’ll build the vignette manually with a circular window rather than apply a stock effect. The window is inverted so that the correction impacts the shot outside of the windowed area.

So there’s a simple node tree that works for many jobs. If you need to adjust parameters such as noise reduction, that’s best done in Node 1 or 2. Remember that Resolve grading works on two levels – clip and timeline. These are all clip-based nodes. If you want to apply a global effect, like adding film grain to the whole timeline, then you can change the grading mode from clip to timeline. In the timeline mode, any nodes you apply impact the whole timeline and are added on top of any clip-by-clip correction, so it works a bit like an adjustment layer.

©2021 Oliver Peters

Final Cut Pro at 10 and Other Musings

Recently Final Cut Pro (formerly Final Cut Pro X) hit its tenth anniversary.  Since I’ve been a bit quiet on this blog lately due to the workload, I thought it was a good time to reflect. I recently cut a set of involved commercials using FCP. While I’ve cut literally thousands of commercials in my career, my work in recent years tends to be corporate/branding/image content in the five to ten minute range. I work in a team and the tool of choice is Premiere Pro. It’s simply a better fit for us, since the bulk of staff and freelancers are very fluid in Adobe products and less so with Apple’s pro software. Sharing projects and elements also works better in the Adobe ecosystem.

Cutting the spots in Final Cut Pro

In the case of the four :60s, I originally budgeted about two days each, plus a few days for client revisions – eleven days in total. My objective was to complete the creative cut, but none of the finishing, since these spots involved extensive visual effects. I was covering for the client’s regular editor who had a scheduled vacation and would finish the project. The spots were shot with a Sony Venice, simultaneously recording 6K RAW and 4K XAVC (AVC-Intra) “proxy” files. The four spots totaled over 1200 clips with approximately an hour of footage per spot. My cutting options could be to work natively with the Sony RAW media in Premiere Pro or DaVinci Resolve, or to edit with the proxies in any NLE.

The Sony RAW files are large and don’t perform well playing from a shared storage system. I didn’t want to waste the time copying location drives to the NAS, partially for reasons of time. I also wanted to be able to access media to cut the spots whether at home or at the work facility. So I opted to use the proxies, which allowed me to cut the spots in FCP. Of course, if you think of proxies as low-res files, you’d be wrong. These Sony XAVC files are high-res, camera-original files on par with 4K ProRes HQ media. If it weren’t for VFX, these would actually be the high-quality source files used for the final edit.

I copied the proxy files to a 2TB Samsung T7 SSD portable drive. This gave me the freedom to edit wherever – either on my iMac at home or one of the iMac Pros at work. This is where Final Cut Pro comes in. When you wade through that much footage, it’s easy for an NLE to get bogged down by caching footage or for the editor to get lost in the volume of clips. Thanks to skimming and keyword collections, I was able to cut these spots far more quickly than using any of the other NLE options. I could go from copying proxy files to my first cut on a commercial within a single day. That’s half of the budgeted time.

The one wrinkle was that I had to turn over a Premiere Pro project linked to the RAW media files. There are various ways to do that, but automatic relinking is dicier with these RAW files, because each clip is within its own subfolder, similar to RED. This complicates Premiere’s ability to easily relink files. So rather than go through XtoCC, I opted to import the Sony RAW clips into Resolve, then import the FCPXML, which in turn automatically relinked to the RAW files in Resolve.

There are a few quirks in this method that you have to suss out, but once everything was correct in Resolve, I exported an XML for Premiere. In Premiere Pro, I imported that XML, made sure that Premiere linked to the RAW files, corrected any size and speed issues, removed any duplicate clips, and then the project was ready for turnover. While one could look at these steps and question the decision to not cut in Premiere in the first place, I can assure you that cutting with Final Cut was considerably faster and these roundtrip steps were minor.

Remote workflows

Over the past year, remote workflows and a general “work from home” movement has shifted how the industry moves forward. So much of what I do requires connection to shared storage, so totally working from home is impractical. These spots were the exception for me, but the client and director lived across the country. In years past, they used to fly in and work in supervised sessions with me. However, in more recent years, that work has been unattended using various review-and-approval solutions for client feedback and revisions. Lately that’s through Frame.io. In the case of these spots, my workflow wasn’t any different than it would have been two years ago.

On the other hand, since I have worked with these clients in supervised sessions, as well as remote projects, it’s easy to see what’s been lost in this shift. Remote workflows present two huge drawbacks. The first is turnaround time. It’s inherently an inefficient process. You’ll cut a new version, upload it for review, and then wait – often for hours or even the next day. Then make the tweaks, rinse, and repeat. This impacts not only the delivery schedule, but also your own ability to book sessions and determine fair billing.

Secondly, ideation takes a back seat. When a client is in the room, you can quickly go through options, show a rearranged cut, alternate takes, and so on. Final Cut’s audition function is great for this, but it’s a wasted feature in these modern workflows. During on-prem sessions, you could quickly show a client the options, evaluate, and move on. With remote workflows, that’s harder to show and is subject to the same latency of replying, so as a result, you have fewer options that can be properly vetted in the cut.

The elephant in the room is security. I know there are tons of solutions for “drilling” into your system from home that are supposed to be secure. In reality, the only true security is to have your system disconnected from the internet (but also not totally bulletproof). As Sony Pictures, QNAP owners, Colonial Pipeline, agencies of the US government, or multiple other corporations have found out, if a bad actor wants to get into your system, they can. No amount of encryption, firewalls, VPNs, multi-factor authentication, or anything else is going to be guaranteed to stop them. While remote access might have been a necessary evil due to COVID lockdowns, it’s not something that should be encouraged going forward.

However, I know that I’m swimming against the stream on this. Many editors/designers/colorists don’t seem to ever want to return to an office. This is at odds with surveys indicating the majority of producers and agencies are chomping to get back to working one-on-one. Real estate and commuting costs are factors that affect such decisions, so I suspect hybrids will evolve and the situation in the future may vary geographically.

Final Cut Pro’s future

I mention the WFH dilemma, because remote collaboration is one of the features that Apple has been encouraged to build into Final Cut Pro by some users. It’s clearly a direction Adobe has moved towards and where Avid already has a track record.

I’m not sure that’s in Apple’s best interest. For one thing, I don’t personally believe Apple does a good job of this. Access and synchronization performance of iCloud is terrible compared with Google’s solutions. Would a professional collaboration solution really be industry-leading and robust? I highly doubt it.

Naturally Apple wants to make money, but they are also interested in empowering the creative individual – be that a professional or an enthusiast. Define those terms in whatever way you like, but the emphasis is on the individual. That direction seems to be at odds with what “pro” users think should be the case for Apple ProApps software, based on their experiences in the late years of FCP 1-7/FCP Studio (pre-X).

I certainly have my own feature request list for Final Cut Pro, but ultimately the lack of these did not stop me from a rapid turnaround on the spots I just discussed. Nor on other projects when I turn to FCP as the tool of choice. I use all four major NLEs and probably never will settle on a single “best” NLE for all cases.

The term “YouTube content creator” or “influencer” is often used as a pejorative, but for many filmmakers and marketeers outlets like YouTube, facebook, and Instagram have become the new “broadcast.” I recently interviewed Alexander Fedorov for FCP.co. He’s a Russian photographer/filmmaker/vlogger who epitomizes the type of content creator for whom Apple is designing its professional products. I feel that Apple can indeed service multiple types of users, from the individual, self-taught filmmaker to the established broadcast pro. How Apple does that moving forward within a tool like Final Cut Pro is anyone’s guess. All I know is that using the measurements of what is and isn’t “pro” no longer works in so many different arenas.

©2021 Oliver Peters

Five Adobe Workflow Tips

Subscribers to Adobe Creative Cloud have a whole suite of creative tools at their fingertips. I believe most users often overlook some of the less promoted features. Here are five quick tips for your workflow. (Click on images to see an enlarged view.)

Camera Raw. Photographers know that the Adobe Camera Raw module is used to process camera raw images, such as .cr2 files. It’s a “develop” module that opens first when you import a camera raw file into Photoshop. It’s also used in Bridge and Lightroom. Many people use Photoshop for photo enhancement – working with the various filters and adjustment layer tools available. What may be overlooked is that you can use the Camera Raw Filter in Photoshop on any photo, even if the file is not raw, such as a JPEG or TIFF.

Select the layer containing the image and choose the Camera Raw Filter. This opens that image into this separate “develop” module. There you have all the photo and color enhancement tools in a single, comprehensive toolkit – the same as in Lightroom. Once you’re done and close the Camera Raw Filter, those adjustments are now “baked” into the image on that layer.

Remix. Audition is a powerful digital audio workstation application that many use in conjunction with Premiere Pro or separately for audio productions. One feature it has over Premiere Pro is the ability to use AI to automatically edit the length of music tracks. Let’s say you have a music track that’s 2:47 in length, but you want a :60 version to underscore a TV commercial. Yes, you could manually edit it, but Audition Remix turns this into an “automagic” task. This is especially useful for projects where you don’t need to have certain parts of the song time to specific visuals.

Open Audition, create a multitrack session, and place the music selection on any track in the timeline. Right-click the selection and enable Remix. Within the Remix dialogue box, set the target duration and parameters – for example, short versus long edits. Audition will calculate the number and location of edit points to seamlessly shorten the track to the approximate desired length.

Audition attempts to create edits at points that are musically logical. You won’t necessarily get an exact duration, since the value you entered is only a target. This is even more true with tracks that have a long musical fade-out. A little experimentation may be needed. For example, a target value of :59 will often yield significantly different results than a target of 1:02, thanks to the recalculation. Audition’s remix isn’t perfect, but will get you close enough that only minimal additional work is required. Once you are happy, bounce out the edited track for the shortened version to bring into Premiere Pro.

Photoshop Batch Processing. If you want to add interesting stylistic looks to a clip, then effects filters in Premiere Pro and/or After Effects usually fit the bill. Or you can go with expensive third party options like Continuum Complete or Sapphire from Boris FX. However, don’t forget Photoshop, which includes many stylized looks not offered in either of Adobe’s video applications, such as specific paint and brush filters. But, how do you apply those to a video clip?

The first step is to turn your clip into an image sequence using Adobe Media Encoder. Then open a representative frame in Photoshop to define the look. Create a Photoshop action using the filters and settings you desire. Save the action, but not the image. Then create a batch function to apply that stored action to the clean frames within the image sequence folder. The batch operation will automatically open each image, apply the effects, and save the stylized results to a new destination folder.

Open that new image sequence using any app that supports image sequences (including QuickTime) and save it as a ProRes (or other) movie file. Stylized effects, like oil paint, are applied to individual frames and will vary with the texture and lighting of each frame; therefore, the stitched movie will display an animated appearance to that effect.

After Effects for broadcast deliverables. After Effects is the proverbial Swiss Army knife for editors and designers. It’s my preferred conversion tool when I have 24p masters that need to be delivered as 60i broadcast files.

Import a 23.98 master and place it into a new composition. Scale, if needed (UHD to HD, for instance). Send to the Render Queue. Set the frame rate to 29.97, field render to Upper (for HD), and enable pulldown (any whole/split frame cadence is usually OK). Turn off Motion Blur and Frame Blending. Render for a proper interlaced broadcast deliverable file.

Photoshop motion graphics. One oft-ignored (or forgotten) feature of Photoshop is that you can do layer-based video animation and editing within. Essentially there’s a very rudimentary version of After Effects inside Photoshop. While you probably wouldn’t want to use it for video instead of using After Effects or Premiere Pro, Photoshop does have a value in creating animated lower thirds and other titles.

Photoshop provides much better text and graphic style options than Premiere Pro. The files are more lightweight than an After Effects comp on your Premiere timeline – or rendering animated ProRes 4444 movies. Since it’s still a Photoshop file (albeit a special version), the “edit in original” command opens the file in Photoshop for easy revisions. Let’s say you are working on a show that has 100 lower thirds that slide in and fade out. These can easily be prepped for the editor by the graphics department in Photoshop – no After Effects skills required.

Create a new file in Photoshop, turn on the timeline window, and add a new blank video layer. Add a still onto a layer for positioning reference, delete the video layer, and extend the layers and timeline to the desired length. Now build your text and graphic layers. Keyframe changes to opacity, position, and other settings for animation. Delete the reference image and save the file. This is now a keyable Photoshop file with embedded animation properties.

Import the Photoshop file into Premiere with Merged Layers. Add to your timeline. The style in Premiere should match the look created in Photoshop. It will animate based on the keyframe settings created in Photoshop.

©2021 Oliver Peters

Project organization

Leading into the new year, it’s time to take a fresh look at a perennial subject. Whether you work as a solo editor or part of a team, having a plan for organizing your projects – along with a workflow for moving media though your system – will lead to success in being able to find and restore material when needed at a future date. For a day-to-day workflow, I rely on five standard applications: Post Haste, Hedge, Better Rename, DiskCatalogMaker, and Kyno. I work on Macs, but there are Windows versions or alternatives for each.

Proper project organization. Regardless of your NLE, it’s a good idea to create a project “silo” for each job on your hard drive, RAID, or networked storage (NAS). That’s a main folder for the job, with subfolders for the edit project files, footage, audio, graphics, documents, exports, etc. I use Post Haste to create a new set of project folders for each new project.

Post Haste uses default or custom templates that can include Adobe project files. This provides a common starting point for each new project based on a template that I’ve created. Using this template, Post Haste generates a new project folder with common subfolders. A template Premiere Pro project file with my custom bin structure is contained within the Post Haste template. When each new set of folders is created, this Premiere file is also copied.

In order to track productions, each job is assigned a number, which becomes part of the name structure assigned within Post Haste. The same name is applied to the Premiere Pro project file. Typically, the master folder (and Premiere project) for a new job created through Post Haste will be labelled according to this schema: 9999_CLIENT_PROJECT_DATE.

Dealing with source footage, aka rushes or dailies. The first thing you have to deal with on a new project is the source media. Most of the location shoots for my projects come back to me with around 1TB of media for a day’s worth of filming. That’s often from two or three cameras, recorded in a variety of codecs at 4K/UHD resolution and 23.98fps. Someone on location (DIT, producer, DP, other) has copied the camera cards to working SSDs, which will be reused on later productions. Hedge is used to copy the cards, in order to provide checksum copy verification.

I receive those SSDs and not the camera cards. The first step is to copy that media “as is” into the source footage subfolder for that project on the editing RAID or NAS. Once my copy is complete, those same SSDs are separately copied “as is” via Hedge to one or more Western Digital or Seagate portable drives. Theoretically, this is for a deep archive, which hopefully will never be needed. Once we have at least two copies of the media, these working SSDs can be reformatted for the next production. The back-up drives should be stored in a safe location on-premises or better yet, offsite.

Since video cameras don’t use a standard folder structure on the cards, the next step is to reorganize the copied media in the footage folder according to date, camera, and roll. This means ripping media files out of their various camera subfolders. Within the footage folder, my subfolder hierarchy becomes shoot date (MMDDYY), then camera (A-CAM, B-CAM, etc), and then camera roll (A001, A002, etc). Media is located within the roll subfolder. Double-system audio recordings go into a SOUND folder for that date and follow this same hierarchy for sound rolls. When this reorganization is complete, I delete the leftover camera subfolders, such as Private, DCIM, etc.

It may be necessary to rename or append prefixes to file names in order to end up with completely unique file names within this project. That’s where Better Rename comes in. This is a Finder-level batch renaming tool. If a camera generates default names on a card, such as IMG_001, IMG_002 and so on, then renaming becomes essential. I try to preserve the original name in order to be able to trace the file back to back-up drives if I absolutely have to. Therefore, it’s best to append a prefix. I base this on project, date, camera, and roll. As an example, if IMG_001 was shot as part of the Bahamas project on December 20th, recorded by E-camera on roll seven, then the appended file would be named BAH1220E07_IMG_001.

Some camera codecs, like those used by drones and GoPros, are a beast for many NLEs to deal with. Proxy media is one way or you can transcode only the offending files. If you choose to transcode these files, then Compressor, Adobe Media Encoder, or Resolve are the best go-to applications. Transcode at the native file size and resolution into an optimized codec, like ProRes. Maintain log color spaces, because these optimized files become the new “camera” files in your edit. I will add separate folders for ORIG (camera original media) and PRORES (my transcoded, optimized files) within each camera roll folder. Only the ProRes media is to be imported into the NLE for editing.

Back-up! Do not proceed to GO! Now that you’ve spent all of this effort reorganizing, renaming, and transcoding media, you first want a back-up the files before starting to edit. I like to back up media to raw, removable, enterprise-grade HGST or Seagate hard drives. Over the years, I’ve accumulated a variety of drive sizes ranging from 2TB to now 8TB. Larger capacities are available, but 8TB is a cost-effective and manageable capacity. When placed into a Thunderbolt or USB drive dock, these function like any other local hard drive. 

When you’ve completed dealing with the media from the shoot, simply copy the whole job folder to a drive. You can store multiple projects on the same drive, depending on their capacity. This is an easy overnight process with most jobs, so it won’t impact your edit time. The point is to back up the newly organized version of your raw media. Once completed, you will have three copies of the source footage – the “as is” copy, the version on your RAID or NAS, and this back-up on the raw drive. After the project has been completed and delivered, load up the back-up drive and copy everything else from this job to that drive. This provides a “clone” of the complete job on both your RAID/NAS and the back-up drive.

In order to keep these back-up drives straight, you’ll need a catalog. At home, I’ve accumulated 12 drives thus far. At work we’ve accumulated over 200. I’ve found the easiest way to deal with this is an application called DiskCatalogMaker. It scans the drive and stores the file information in a catalog document. Each drive entry mimics what you see in the Finder, including folders, files, sizes, dates, and so on. The catalog document is searchable, which is why job numbers become important. It’s a good idea to periodically mount and spin up these drives to maintain reliability. Once a year is a minimum.

If you have sufficient capacity on your RAID or NAS, then you don’t want to immediately delete jobs and media when the work is done. In our case, once a job has been fully backed up, the job folder is moved into a BACKED UP folder on the NAS. This way we know when a job has been backed up, yet it is still easily retrieved should the client come back with revisions. Plus, you still have three total copies of the source media.

Other back-ups. I’ve talked a lot about backing up camera media, but what about other files? Generally files like graphics are supplied, so these are also backed up elsewhere. Plus they will get backed up on the raw drive when the job is done.

I also use Dropbox for interim back-ups of project files. Since a Premiere Pro project file is light and doesn’t carry media, it’s easy to back up in the cloud. At work, at the end of each day, each editor copies in-progress Premiere files to a company Dropbox folder. The idea is that in the event of some catastrophe, you could get your project back from Dropbox and then use the backed up camera drives to rebuild an edit. In addition, we also export and copy Resolve projects to Dropbox, as well as the DiskCatalogMaker catalog documents.

Whenever possible, audio stems and textless masters are exported for each completed job. These are stored with the final masters. Often it’s easier to make revisions using these elements, than to dive back into a complex job after it’s been deeply archived. Our NAS contains a separate top-level folder for all finished masters, in addition to the master subfolder within each project. When a production is done, the master file is copied into this other folder, resulting in two sets of the master files on the NAS. And by “master” I generally mean a final ProRes file along with a high-quality MP4 file. The MP4 is most often what the client will use as their “master,” since so much of our work these days is for the web. Therefore, both NAS locations hold a ProRes and an MP4. That’s in addition to the masters stored on the raw, back-up drive.

Final, Final revised, no really, this one is Final. Let’s address file naming conventions. Every editor knows the “danger” of calling something Final. Clients love to make changes until they no longer can. I work on projects that have running changes as adjustments are made for use in new presentations. Calling any of these “Final” never works. Broadcast commercials are usually assigned definitive ISCI codes, but that’s rarely the case with non-broadcast projects. The process that works for us is simply to use version numbers and dates. This makes sense and is what software developers use.

We use this convention: CLIENT_PROJECTNAME_VERSION_DATE_MODIFIER. As an example, if you are editing a McDonald’s Big Mac :60 commercial, then a final version might be labelled “MCD_Big Mac 60_v13_122620.” A slight change on that same day would become “MCD_Big Mac 60_v14_122620.” We use the “modifier” to designate variations from the norm. Our default master files are formatted as 1080p at 23.98 with stereo audio. So a variation exported as 4K/UHD or 720p or with a 5.1 surround mix would have the added suffix of “_4K” or “_720p” or “_51MIX.”

Some projects go through many updates and it’s often hard to know when a client (working remotely) considers a version truly done. They are supposed to tell you that, but they often just don’t. You sort of know, because the changes stop coming and a presentation deadline has been met. Whenever that happens, we export a ProRes master file plus high-quality MP4 files. The client may come back a week later with some revisions. Then, new ProRes and MP4 files are generated. Since version numbers are maintained, the ProRes master files will also have different version numbers and dates and, therefore, you can differentiate one from the other. Both variations may be valid and in use by the client.

Asset management. The last piece of software that comes in handy for us is Kyno. This is a lightweight asset management tool that we use to scan and find media on our NAS. Our method of organization makes it relatively easy to find things just by working in the Finder. However, if you are looking for that one piece of footage and need to be able to identify it visually, then that’s where Kyno is helpful. It’s like Adobe Bridge on steroids. One can organize and sort using the usual database tools, but it also has a very cool “drill down” feature. If you want to browse media within a folder without stepping through a series of subfolders, simply enable “drill down” and you can directly browse all media that’s contained therein. Kyno also features robust transcode and “send to” features designed with NLEs in mind. Prep media for an edit or create proxies? Simply use Kyno as an alternative to other options.

Hopefully this recap has provided some new workflow pointers for 2021. Good luck!

©2021 Oliver Peters