This library tries to expose the same API endpoints as that of processing-video library for PC. Any sketches using the processing-video library should work the same when using this library on Android.
Demo: https://www.youtube.com/watch?v=kjqcWGLH6Q8
A simple example :
import in.omerjerk.processing.video.android.*;
Capture cap;
void setup() {
size(720, 1280, P2D);
cap = new Capture(this, 720, 1280);
String[] list = Capture.list();
//Use this to print list of resolutions supported by the camera
Capture.printCompatibleResolutionsList(cap);
cap.setCamera(list[0]);
cap.start();
}
void draw() {
image(cap, 0, 0);
}
Behind the hood :
The idea is to render camera preview on a SurfaceTexture
and as soon as the new frame is received, copy the data from this SurfaceTexture
to a custom texture bound to target GL_TEXTURE_2D
(Note : We cannot directly render to a texture bound to GL_TEXTURE_2D
target, because for the preview to happen, the texture must be bound to GL_TEXTURE_EXTERNAL_OES
).
This custom texture is then rendered to a PGraphics object.
The backing texture of that PGraphics object is then used as the texture cache for our PImage file which stores the video.
gradle dist
will build the project, copy it to the user's Processing sketchbook folder, and prepare a .zip file ready for distribution.