Getting Started

To fetch the sources, build and install:

git clone https://git.libcamera.org/libcamera/libcamera.git
cd libcamera
meson setup build
ninja -C build install

Dependencies

The following Debian/Ubuntu packages are required for building libcamera. Other distributions may have differing package names:

A C++ toolchain: [required]
Either {g++, clang}
Meson Build system: [required]
meson (>= 0.60) ninja-build pkg-config
for the libcamera core: [required]
libyaml-dev python3-yaml python3-ply python3-jinja2
for IPA module signing: [recommended]

Either libgnutls28-dev or libssl-dev, openssl

Without IPA module signing, all IPA modules will be isolated in a separate process. This adds an unnecessary extra overhead at runtime.

for improved debugging: [optional]

libdw-dev libunwind-dev

libdw and libunwind provide backtraces to help debugging assertion failures. Their functions overlap, libdw provides the most detailed information, and libunwind is not needed if both libdw and the glibc backtrace() function are available.

for device hotplug enumeration: [optional]
libudev-dev
for documentation: [optional]
python3-sphinx doxygen graphviz texlive-latex-extra
for gstreamer: [optional]
libgstreamer1.0-dev libgstreamer-plugins-base1.0-dev
for Python bindings: [optional]
libpython3-dev pybind11-dev
for cam: [optional]

libevent-dev is required to support cam, however the following optional dependencies bring more functionality to the cam test tool:

  • libdrm-dev: Enables the KMS sink
  • libjpeg-dev: Enables MJPEG on the SDL sink
  • libsdl2-dev: Enables the SDL sink
for qcam: [optional]
libtiff-dev qt6-base-dev qt6-tools-dev-tools
for tracing with lttng: [optional]
liblttng-ust-dev python3-jinja2 lttng-tools
for android: [optional]
libexif-dev libjpeg-dev
for Python bindings: [optional]
pybind11-dev
for lc-compliance: [optional]
libevent-dev libgtest-dev
for abi-compat.sh: [optional]
abi-compliance-checker

Basic testing with cam utility

The cam utility can be used for basic testing. You can list the cameras detected on the system with cam -l, and capture ten frames from the first camera and save them to disk with cam -c 1 --capture=10 --file. See cam -h for more information about the cam tool.

In case of problems, a detailed debug log can be obtained from libcamera by setting the LIBCAMERA_LOG_LEVELS environment variable:

:~$ LIBCAMERA_LOG_LEVELS=*:DEBUG cam -l

Using GStreamer plugin

To use the GStreamer plugin from the source tree, use the meson devenv command. This will create a new shell instance with the GST_PLUGIN_PATH environment set accordingly.

meson devenv -C build

The debugging tool gst-launch-1.0 can be used to construct a pipeline and test it. The following pipeline will stream from the camera named “Camera 1” onto the OpenGL accelerated display element on your system.

gst-launch-1.0 libcamerasrc camera-name="Camera 1" ! queue ! glimagesink

To show the first camera found you can omit the camera-name property, or you can list the cameras and their capabilities using:

gst-device-monitor-1.0 Video

This will also show the supported stream sizes which can be manually selected if desired with a pipeline such as:

gst-launch-1.0 libcamerasrc ! 'video/x-raw,width=1280,height=720' ! \
     queue ! glimagesink

The libcamerasrc element has two log categories, named libcamera-provider (for the video device provider) and libcamerasrc (for the operation of the camera). All corresponding debug messages can be enabled by setting the GST_DEBUG environment variable to libcamera*:7.

Presently, to prevent element negotiation failures it is required to specify the colorimetry and framerate as part of your pipeline construction. For instance, to capture and encode as a JPEG stream and receive on another device the following example could be used as a starting point:

gst-launch-1.0 libcamerasrc ! \
     video/x-raw,colorimetry=bt709,format=NV12,width=1280,height=720,framerate=30/1 ! \
     queue ! jpegenc ! multipartmux ! \
     tcpserversink host=0.0.0.0 port=5000

Which can be received on another device over the network with:

gst-launch-1.0 tcpclientsrc host=$DEVICE_IP port=5000 ! \
     multipartdemux ! jpegdec ! autovideosink

The GStreamer element also supports multiple streams. This is achieved by requesting additional source pads. Downstream caps filters can be used to choose specific parameters like resolution and pixel format. The pad property stream-role can be used to select a role.

The following example displays a 640x480 view finder while streaming JPEG encoded 800x600 video. You can use the receiver pipeline above to view the remote stream from another device.

gst-launch-1.0 libcamerasrc name=cs src::stream-role=view-finder src_0::stream-role=video-recording \
    cs.src ! queue ! video/x-raw,width=640,height=480 ! videoconvert ! autovideosink \
    cs.src_0 ! queue ! video/x-raw,width=800,height=600 ! videoconvert ! \
    jpegenc ! multipartmux ! tcpserversink host=0.0.0.0 port=5000