Skip to content

za-creature/eso

Repository files navigation

eso

A horizontally scalable high-throughput near-realtime multirate h264 & aac encoding engine built on top of ffmpeg and google cloud with a sprinkle of AWS. Produces adaptive bitrate MPEG-TS fragments suitable for use with a video player that supports HLS

How it works

The provided cloud function runs whenever a file is uploaded to an input bucket. On first invocation, it will generate a signed URL to the media file and pass that to a new instance of ffmpeg that splits the file into small consecutive video fragments into the local ramdisk, while continuously monitoring the stderr output for progress reports.

As soon as a fragment has been completed, it is uploaded to the same input bucket and deleted to free up memory. This process is lossless: video frame data is copied verbatim and audio is decoded to pcm (a frameless audio format is required in order to prevent audio / video desync) using ffmpeg's internal nut container for best compatibility.

Since a successfully uploaded fragment will re-trigger the cloud function, it is marked with custom metadata that informs the function to switch operation mode from input demuxing to segment transcoding. For efficiency, the transcoder operates in batches, producing multiple output bitrates from a single fragment decoding pass (constrained to a maximum total output bitrate per batch). If better latency is required, it is possible to adapt this step to perform a single transcode per invocation (for example by uploading multiple empty files just to spawn additional instances), though this is not currently implemented.

After all the generated fragments from the input file have been uploaded, a playlist is uploaded to the output bucket and, depending on the status of the asynchronous segment transcoding steps, may already be playable.

Installation

Prerequisites

  • a google cloud account with Functions enabled and two empty storage Buckets
  • an amazon web services account with an empty DynamoDB table (optional)
  • docker or a precompiled ffmpeg executable that is binary compatible with google's Node.js 10 cloud functions runtime
  • gcloud
  • make

Building

Edit .env.sample accordingly and then rename it to .env:

mv .env.sample .env

Copy your own ffmpeg binary to the bin/ directory or run (requires docker)

make

(Optional) configure your preferred multimedia settings in config.js

Deploying

Create a service key and save it to gcloud.json

Authenticate into google cloud SDK and set your default project ID:

gcloud auth activate-service-account --key-file=gcloud.json
gcloud config set project $(sed '/project_id/!d;s/.*"\(.*\)",/\1/' gcloud.json)

Run

make deploy

Usage

Once successfully deployed, eso will attempt to transcode all files uploaded to the input bucket into the configured output bucket (which is assumed to be publicly accessible for reading). Conversion status updates are sent to dynamodb if configured, with the following schema:

{
    key: String, // the key (filename) in the input bucket
    total_segments: Number, // total number of segments or -1 on fatal error
    ready_segments: Set, // a set of already transcoded segments
    thumb: Boolean // whether the video file has a thumbnail
}

A key is ready for on-demand seekable playback after ready_segments.size == total_segments, and a progress indicator for a sufficiently large file may be estimated linearly from the former ratio. All files transcoded from a key in the input bucket are stored in the output bucket using the input key as prefix. The exact mapping is:

(key) => `${key}/index.m3u8`  // master playlist (output video url)
(key) => `${key}/thumb.jpeg`  // thumbnail extracted from first video segment
(key, bandwidth) => `${key}/${bandwidth}k/index.m3u8` // profile playlist
(key, bandwidth, segment) => `${key}/${bandwidth}k/${segment}.ts` // ts segment

Where bandwidth is the estimated audio + video bitrate in kilobits per second, and segment is a monotonically increasing number between 0 and floor(stream_duration / segment_duration) (see config.js).

Non-seekable playback may be attempted as soon as the first few transcoded segments are uploaded, though continuous playback is not guaranteed in this scenario as it depends on many factors including player software, cloud function scheduling and input / output video complexity.

Limitations

  • can only split segments at an I-frame boundary; this means that some videos will scale (latency-wise) better than others and a DoS attack is possible in the current implementation
  • video output is limited by what x264 can support
  • audio output is limited to AAC HE, mono and stereo
  • since a cloud function can only run for at most 9 minutes, the practical limit for the size of an input file is around 30GiB (assumes 1Gbps network) and about 50 profiles, depending on complexity and segment length

Gotchas

  • for ease of deployment, eso makes liberal use of the word input as it also requires write access to the input bucket to cloud-fork itself by uploading intermediary video segments with custom metadata. These segments are deleted once successfully transcoded, however in order to deal with dangling references in case of failure, I strongly recommend configuring adequate lifecycle management policies on the input bucket
  • the profiles configured in config.js are treated as maximum values; depending on the quality of your input files, some profiles may be encoded with lower quality settings or omitted altogether
  • the meta entry is created before splitting the input file, and at that time it is only guaranteed to contain the thumb field; total_segments is created only after the file has been successfully split into segments and the ready_segments field is only created after the first segment has been successfully transcoded and uploaded to the output bucket
  • if you're redeploying after updating gcloud.json, all pending transcode operations will fail because the hashed private key is used to sign segment uploads; to avoid this, add SECRET=some_random_secret_string_here to .env and it will be used as a HMAC key instead, though if you're only seeing this after you've already went live, tough luck

License: MIT

This is a tech demo, and while functional, I wouldn't call it production ready. Reusing code from this repository is fine by me, but the generated binaries are not redistributable and some codecs may be protected by patents as well. Always double-check with your legal team before deploying multimedia transcoding technologies for commercial purposes!

About

Scalable h264 encoder for google cloud

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published