Timestamp Jitter Range
Hi, I came across this website: https://support.prophesee.ai/portal/en/kb/articles/evk-latency. It mentions jitter and latency here. The latency is clearly mentioned for IMX636; however, I didn't find any measurement about the jitter time. May I ask what
metavision_viewer unable to read camera_config.json correctly in KV260+imx636
this is print out and the camera_config file. metavision_viewer (4.6.2) is able to work well with the camera_config.json but when it got upgraded to 5.0.0, it become like this. Is there a change to the format of the camera_config.json file? if so, could
Building Active markers on kria ubuntu: Sophos hangs on the Kria embedded kit (imx636)
Hello, I am trying to get the event markers code to build & run on the kria ubuntu image, instead of the petalinux image. I am finding that building petalinux via the build tools is time consuming and requires a beefy machine, so I thought maybe switching
Help Getting Started: Streaming GenX320 Output with STM32F746 (No GUI Needed)
Hi everyone, I'm new to event-based vision and have just started working with a Prophesee GenX320 camera and an STM32F746 microcontroller. My initial goal is simple: Just open the camera and stream its output — no GUI, no visualization — just raw event
Stereo Calibration
Hello, I calibrated my two event cameras in a stereo setup to perform depth mapping but I am getting poor results and no where near the accuracy shown in the sample videos. I was wondering if you could provide the camera setup used for the courtyard stereo
Is there a Python version of the Simple Window using C++ Example?
I want to replicate this example using Python because I have very little experience working with C++. https://docs.prophesee.ai/stable/samples/modules/ui/simple_window.html Has someone done this before? Or is this something that Prophesee has provided
Camera not detected in Metavision Starter kit KV260 (IMX636)
Dear Prophesee Support Team, I am working with the Prophesee Metavision Starter Kit for AMD Kria KV260 and followed the instructions provided in the Quick Start Guide. After flashing the SD card and booting the board, I attempted to launch Metavision
UTC-Referenced Absolute Event Timing in Metavision Studio/GUI
For my application, I need to reference event times to UTC with microsecond precision. In Metavision Studio, the .raw recording names are only given to the nearest 1 second and the events are referenced to the camera connect time (not the recording start
EVT2.1 format valid bits meaning
Hello! According to the EVT 2.1 format description EVT 2.1 Format — Metavision SDK Docs 5.1.1 documentation , how should valid bits be understood correctly? 1. X and Y coordinates of the current valid event. 2. After receiving data from the image sensor,
Camera Time Sync
Hello Prophesee Team, I am using a C++ application to periodically call Metavision::Camera::get_last_timestamp() for the purpose of synchronizing and analyzing drift between system time and camera time. My expectation was that, by recording the value
The specific value of the refractory period
Hello, I would like to know what is the refractory period in microseconds for the IMX636 version of EVK3 when the bias parameter is set to the default value.
Upper bound of get_illumination hal api function
Hello! I was wondering what the upper bound is the get_illumination function in the HAL API. Experimentally, I've recorded 300 lux with the function, but I would be interested to know the upper bound on the sensing abilities.
About the temporal resolution of IMX636
I am currently using a high-speed projector with a minimum exposure time of 105 microseconds per frame.In my experiment, I project alternating all-white and all-black patterns continuously, with each frame having an exposure time of 105 µs. Under this
IMX636 internal clock PPM
Hello, I am trying to find information about expected internal clock accuracy as PPM compared to the "perfect" clock. online sources give average "Standard Crystal Oscillators" somewhere around 20 PPM, which means around 12 seconds drift per week. Does
Multiple DVS synchronization (more than 1 slave)
Hello everyone, Our team plans to synchronize more than one slave DVS (with one master and three slaves). We've followed the recommendations for multiple DVS synchronization, but during testing, we found that only one slave is precisely synced with the
How to set MV_FLAGS_EVT3_UNSAFE_DECODER in Python api
I'm currently utilizing the EventsIterator to read EVT3 files, but I've encountered some errors. I'd like to know how to set the MV_FLAGS_EVT3_UNSAFE_DECODER option in Python so that it can ignore these errors.
EVK4 Stereo Camera ROS Driver Installation
We would appreciate your support in helping us resolve the ROS driver compatibility issue for our EVK4 stereo camera setup. We are currently working on setting up an event-based stereo vision system using two EVK4 cameras (serial numbers: P50905 and P50898).
How do I know how the FrequencyMapAsyncAlgorithm works
Hello, I'm currently using the event camera to detect frequency information, but the FrequencyMapAsyncAlgorithm doesn't seem to be open to the public, how can I know how the algorithm implements frequency detection
Efficiently replaying sync data from sync camera
Dear community, Thank you for reading. I am currently working with two EVK4 camera synchronized with the sync module of the SDK. I recorded some synchronized raw and I wanted to play it back real-time, and still synchronized. I cannot use the sync module
Interfacing Prophesee USB EVK with RT Linux CRIO
Is it possible to use a USB EVK* with Linux based NI cRIO? Ideally I would like to get an event stream in my control hardware -a National Instruments cRIO9049. The cRIO uses Linux RT -not ubuntu- and I am not sure if there is a simple way of doing t
USB and MIPI camera different bias defaults and ranges
I am seeing different default bias values and allowable ranges between the MIPI and USB versions of the IMX636. The following figure are the bias values that are found in this article as well as through testing on the Metavision Studio application. For
Are Metavision SDK Cmake files missing KV260
Are the Metavision SDK CMake (.cmake) files missing from the KV260 Petalinux image? Same question for OpenCV. I am attempting to build some of the examples and CMake cannot find the MetavisionSDKConfig.cmake. I have attempted to locate any CMake files
Saving the results of the generic tracking sample to a numpy file
I'm attempting to adapt the generic tracking script to allow me to save the data in a numpy file, however, whenever I inspect the data, the timestamps when grouped by object_id are exactly the same for each object. That is, object 10 will always have
Slice Condition using Python
Hi, I'm trying to make use of the CameraStreamSlicer() function to change the duration of each slice. I've tried looking through all the documentation but I cant seem to figure out how to do this using the Python functions. Based on the information I
Vibration estimation seems to have a weird bottleneck
Hi, I've been using the vibration estimation code that is provided but im having some issues. The vibration estimation works fine even for detecting high frequencies however I want to run the detection it self at a higher frequency (so how often the algorithm
Issue of High latency and frame drop during event data capture for accumulation period less than 10 ms.
We are using - KV260 and imx636 mipi camera setup for our system under Kria Ubuntu, for one of the application with event camera. We are capturing data with high fps by setting the accumulation period of 1-10 msec using OpenEB Metavision SDK 4.6.0. For
Uncertainty in lux measurement from .get_illumination()?
Greetings, I was wondering if uncertainty in the lux measurement returned by `.get_illumination()` was known? Thank you.
Unable to locate the path to necessary DLLs for Metavision Python bindings
I am trying to follow the tutorial 'RAW File Loading using Python' (RAW File Loading using Python — Metavision SDK Docs 4.4.0 documentation (prophesee.ai)), but when I launch the program, it shows the following error: Unable to locate the path to necessary
No module named 'metavision_sdk_base_paths_internal'
Hi! When I tried to run the python tutorials, there is a problem: No module named 'metavision_sdk_base_paths_internal'. Please see the attached photo below: I found some related files but they are not named exactly the same with the module above: Could
IMX636 Bias Values for Automotive Driving Scences
Hello, I am planning to use the EVK4 with the IMX636 sensor in recording of an automotive dataset. Are there any existing experiences or recommendations regarding bias adjustment for driving scenes or outdoor environments? If not, what would be the best
EventsIterator can not work rightly
Dear Community, I am working on a project involving an event camera and have created a Python class to manage it. One of the class attributes, self.cam, is an initialized HAL device. The class includes a method, continuous_acquire_events, designed to
Recording Application for two EVK4's and an RGB Camera
Greetings! In order to do research on event based vision in the context auf autonomous driving I am currently working on developing an application that is used to record two EVK4's- as well as images of an industrial RGB camera. The application is written
About the IMX636 bias
I noticed in the documentation that for the IMX636 model, only bias_diff_on/off has an effect on its sensitivity. But in my tests, I found that bias_diff also had an impact. I want to know what the details should be.
The exact threshold value of IMX636
I am currently utilizing the imx636 sensor for my research.To proceed with my work, I need to determine the exact threshold value rather than the bias parameter. How could I get this information?
Event Camera for Optical Communication
Has anyone used an Event Camera for optical communication? I would like to use and LED to transmit data that can be received with an event camera. Some scenarios include: 1) stationary transmitter 2) moving transmitter 3) multiple transmitters
event camera noise
Hello, I have an EVK3-IMX636 camera, and my observation scenario involves detecting faint spatial targets in dark space. I would like to inquire about how to configure the camera so that it outputs as many events as possible generated by the sensor, without
Advice to combine spatter_tracking and vibration_estimation algorithms
Good afternoon, I wanted to ask how it is better to combine spatter_tracking and vibration_estimation algorithms in one code, so that when i run that code, both features appear at the same window, and work flowlessly parallel. I used this scripts provided
Spectral response curve for EVK4
Hello, I am trying to measure and verify the response probability S curve of the EVK4 camera. But I cannot find the quantum efficiency or detailed spectral response curve for the sensor in the preliminary datasheet. Only found a table with several wavelengths
How to play back data from raw files in real time?
When playing back data from file using OpenEB (5.0.0) using the code below (simplified), the data packets are delivered not in real-time (usually much faster). Is there anything that I'm doing wrong? ``` const auto cfg = Metavision::FileConfigHints().real_time_playback(true);
EVK4 Camera data transfer
The EVK4 camera with an USB3 cable is rated for hundreds of MB/s of data transfer; however, when running the EVK4 camera in a rapidly changing environment (where almost all pixels across the camera's resolution are changing) for around 10 seconds, an
Next Page