Software - syncFLIR

A synchronized multiview video recording setup in C++ using FLIR cameras and Spinnaker SDK on Windows machines.

Welcome to syncFLIR

The project is hosted in GitLab.

What is this for?

Instead of using independent action cameras to record behavioral experiments and painfully trying to synchronize the different video tracks by e.g. hand-clapping cues or blinking LEDs, make your life easier using triggered cameras. Such an array of wired cameras is scalable without considerable extra effort, allowing for synchronized multi-view video recording and 3D triangulation. This project will help you solve half the problem, you only need to set it up.

SyncFLIR was tested with up to 8 BlackFlyS cameras from FLIR with wired GPIO connectors. One of the cameras is used as pacemaker to trigger all the rest. Alternatively, you can use an external TTL signal generator to trigger all cameras externally. NEW: The sourcecode is now compatible with Chameleon3 cameras, too!

How does it work?

Video recording is performed in separate parallel threads for each camera to reach higher framerate up to 170Hz, but the high data stream with increasing number of cameras may require special hardware specifications. For example, each camera should be connected to independent USB ports, requiring a USB 3.1 host controller card with as many independent channels. Additionally, SSD cards with high writing speeds of 800MB/s and capacity of 2TB will be necessary for longer recordings at higher framerate. If further upgrading is not an option, the different cameras can be distributed over separate machines while being still synchronized with all the others.

To further optimize processing speed, video frames are saved in a binary file as intermediate storage (RECord to BINary: RECtoBIN.cpp). The recording process is logged in a .csv file with exact time stamps and frame IDs. In a subsequent process, the binary files can be read out and converted to AVI video files with different video compressions (BINary to AVI: BINtoAVI.cpp). To keep track of recording consistancy (i.e., skipped frames, inter-frame-intervals, average fps, etc.) use the diagnostics tool provided ( to analyze the .csv recording log.


Install the Spinnaker SDK and test your camera connections with SpinView. Note: Some parts of the code seem to be compatible with Spinnaker library version only (see issue #3). The sourcecode is in C++, but if you don’t feel like compiling, download the executables provided RECtoBIN.exe and BINtoAVI.exe together with the corresponding myconfig.txt file. Configure your recording parameters such as camera serial number and framerate in the myconfig.txt file before running RECtoBIN.exe. To work on the sourcecode directly, replace the .cpp file in any existing /Spinnaker/src project and compile the file with the corresponding .h files.


  1. Set your parameters in the myconfig.txt file
  2. Record multiple synchronized videos using RECtoBIN.exe
  3. Update the generated metadata.txt file to configure video compression
  4. Convert the recorded binary files to raw or compressed video with BINtoAVI.exe
  5. Play your videos using the VideoPyToolbox
  6. Process your recording logfile running and look out for any skipped frames!


In the documentation you will find a list of the equipment needed to build your setup, instructions on how to wire everything up, as well as detailed installation and usage instructions to get yourself started. Brows the Wiki for more information and step by step guides. If there is something missing feel free to open a new issue.


SyncFLIR uses code examples and entire helper functions provided by FLIR Integrated Imaging Solutions, Inc. (FLIR) and the copyright for Spinnaker SDK applies, accordingly.


SyncFLIR is distributed under the MIT License. See LICENSE for more information.


The development of syncFLIR was carried out at the Institute of Cognitive Neuroscience (IKN), Department of Biopsychology at the Ruhr-University Bochum and supported by the German Research Foundation DFG in the context of funding the Research Training Group “Situated Cognition” (GRK 2185/1). Original: Gefördert durch die Deutsche Forschungsgemeinschaft (DFG) - Projektnummer GRK 2185/1.

Let me know on Twitter if you found this useful or would like to have a more detailed discussion on any of the methods used above.

Guillermo Hidalgo Gadea
Guillermo Hidalgo Gadea
PhD Student in Psychology

My research interests include embodied cognition, animal behavior and machine learning.