Refractory period/Pixel reset
Hello, I have a question regarding the TDRSTN pin of the EVK3 for GENX320. For caracterization, the datasheet indiates that (page 32) : Does this signal have the same effect as the bias_refr bias (https://docs.prophesee.ai/stable/hw/manuals/biases.html?highlight=refractory%20period)
Why do higher frequencies need brighter signals to be detected?
Hello dear Prophesee support team, In my experiments I realized in a constant brightness level higher frequencies signals (exp. 1KHz) cannot be detected as much as lower frequency signals (exp. 10~100Hz); meaning much less events are detected in higher
Pulse signal stopping problem during master-slave camera synchronization
Hello, I would like to ask you about the stop pulse signal: experiment procedure: 1. First synchronization: After compiling SDK: metavision_hal_sync, start the slave first and then the host. There is no problem with synchronization. Note: The USB data
Please ask metavision_sdk_get_started.cpp why the screen flips left and right after starting EVS to start collecting, thank you!
Please consult metavision_sdk_get_started.cpp. After starting EVS to start collecting, the screen flips left and right. For example: In the pictures and videos, the text on the side of the book is: QT5 development and examples. Comparing the actual object
Sensor Latency Gen 4.1
Hello I'm performing some experiments with PIV laser pulses and I see that even though here is specified that the max latency of the sensor is 200 us Temporal precision of Prophesee sensor's readout It seems that I'm having a higher latency. In the Image
Different definitions of contrast
Hello, I was wondering why contrast was defined differently between the IMX636 sensor and the GENX320 : For IMX636 : For GENX320 : If we want to plot the S-curve, what influences the choice between plotting pixel response probability in function of log-contrast
Quantifying Signal Strength
I'm trying to quantify the strength of a signal at different accumulation periods, from 100us up. Typically I would use SNR for this, but in the case of an event sensor, that metric is mostly irrelevant. Can someone recommend a strong candidate for quantifying
How is equivalent frame rate calculated?
Hello and good day, As I know frame rate is not a defined parameter in event-based cameras but is there a formula to estimate how fast of a change in the scene can be detected? In Prohphesee's website more than 10000 fps is mentioned as equivalent temporal
Flickering bars on event stream
I have a Prophesee EVK4 camera and I'm noticing a strange occurrence when attempting to record a stream with the camera face down. There is constant bar flickering across the stream. Most other orientations are completely fine; it only falls apart with
Request camera information
Hello, the picture is the information of my camera. I would like to know what the unit of the contrast threshold of this camera is? And what is the range of each parameter? Thank you.
RCT value of GENX320
Hello, I have a question regarding the RCT (Ramp Contrast Threshold) value of the GENX320 from the datasheet. I don't understand how RCT=50% if NCT=30% : Let's suppose that NCT=30%. Using the values of the experimental setup for the RCT measurement, if
Why does the sensor detect better changes in higher initial light level?
Hello, I have a question regarding CDP (Contrast Detection Probability) graphs and contrast sensitivity, Why does the sensor have a higher chance of detecting the changes (contrast) when the initial light level is brighter? What is the reasoning or mechanism
Delayed response of events in Low Light Conditions.
Hii, my experiment consist in a black rotating dot with a background light to generate contrast. When I keep all conditions the same only changing the light intensity of the background light I notice that there is like a time delay or a latency (I dont
Measuring Pixel array current GENX320
Hello, Is there a pin on the GENX320 EVK3 kit where it possible to measure the total current generated by the photodiodes of the pixel matrix ? I'm trying to measure dark current from the pixel's photodiode. For that, the closest i got to the value of
Help! how to deal with the problem?
OS: Windows 11 Professional (23H2) Error: Metavision studio internal error, code 3221225785 signal null
When the saturation takes place, do the remaining columns become more sensitive?
Hii, I have been doing an experiment to understand a bit more the behavior of the camera when the saturation takes place and my results show that the higher the saturation, the higher the average event rate per row. My experiment consists of several points
How to get a symmetry between bias_diff_off and bias_diff_on about IMX636
Hi, there is a note that it is possible to get a symmetry between "bias_diff_off" and "bias_diff_on" by setting those two bias parameters to the same value in the biases of metavision SDK docs. I want to know the same value is point to the current value
access to the introspect player
Hi there, I found out about the Introspect Player shown in the bias tunning video. https://support.prophesee.ai/portal/en/kb/articles/bias-tuning-region-of-interest-roi-31-8-2023 It would be very useful for our research if we could access it. I was wondering
Restarting the "Photodiode Voltage" for EVK4
I'm a student at Carnegie Mellon and we're working on a project that needs to measure the events when the illumination of a pixel increases from zero in a static scene (such as instantaneously taking the lens cap off). Currently, we have an EVK4 model.
Raw file to video
Hello, I noticed that there are some code samples available for converting RAW files to video in C++. However, I'm having difficulty finding an equivalent solution for Python. Could you please advise if there is any Python code available to do this ?
Accumulate events between triggers
Hi, Just wondering if it is possible to accumulate events between triggers to frames. For example, when an on trigger fires, then accumulate events until next trigger. Thanks!
Multi camera sync connection
I've connected two cameras by aligning the sync out and sync in as shown in the following photo. However, upon consulting the schematic for connecting the devices, I noticed that two sync ins are originating from a single sync out. In such a case, would
FPGA not properly configured
When trying to open my EVK1 with the Metavision Studio 3.1 i get the following error message in the console: FPGA not properly configrured. I have never changed any settings or send any FPGA commands. Does anybody have an idea what happened or how i can
Whether new events are generated during readout latency
Hello. I use the EVK4 event camera for some high activity cases, such as recording the rapid movement of many particles. In this case, due to the high number of particles, the timestamps of the generated events will have a maximum delay of 60 microseconds.
Recommendation for Frame based Camera for Integration with DVS
Hi, I plan to sync frame based camera with DVS, according to guideline the frame based camera should output sync out signal to DVS Trig in. my question is do you have any recommendation of frame based camera that has dedicated port to sync out signal?
get_illumination returning -1
I'm trying to compare the sensor lux readout to the events/sec at different bias settings, but the get_illumination readout from the I_monitoring interface only returns -1. Am I missing something about how to get this information? Thanks, Rich
EVK4 measurement error information
I want to track an object using a Kalman filter, and among the variables, there is an item called variance of the sensor's measurement error, which can be provided by the manufacturer, where can I request it? My model is evk4
Reset timestamps for multi-cam sync
Hello - I am utilizing 4 SilkyEvCam modules in a stereo setup. I have connected configured one of the cameras as the master and the other three as the slaves in software. I then connect the sync out of the master to the sync in of each slave. Distances
Active pixels behind the moving object - Pixel latency
Hi, I'm doing an experiment about a rotating black DOT in a white background with backlight to maximize contrast. In all the cases I have this ON events after the movement of the DOT which in this case is going clockwise, this actually looks like a tail
How to read camera data using fpga
We hope to use fpga to read the camera data instead of SDK. Is there any relevant information or documentation?
Data transmission & Processor
Hello, According to the data announced on your site, the equivalent temporal precision of the event-technology is over 10 000 fps. Is this value fixed? Or can it decrease again in the most unfavorable cases? For example, we have a VGA EVK1 with a resolution
No plugin found error on Jetson orin nano
Hi, I sussessfully compiled OPENEB-4.3.0 on my Jetson Orin Nano ( arm64 coretex, Ubuntu 20.4 ), and I lunched 'metavision_viewer' from the 'build' fowder. And I plugged EVK3 to my Jetson, but there comes the warning: 'no plugin found' and 'Metavision
Metavision interface for Mvtec Halcon
There is a problem to install the MVTec HALCON acquisition interface for prophesee camera. the last version of your SDK don't mutch with your dll file for the interface. These attachemements works only for 4.1 version and not for 4.3 is it possible to
Overheating Gen4.1 sensor
Hi, I've been doing experiments with the camera where the event rate is always around 100 Mev/s, with around 10 minutes of use, the camera starts to overheating and the Metavision Studio app stops suddenly. Can I know your opinion please about this? Since
Help understanding the difference in response between positive and negative events.
Hello everyone, I've been working with an EVK3 as part of my PHD thesis. In order to quantify the effect that the on and off biases have on the response of the camera, I have set up a simple experiment that uses the camera to measure responses to a pulsing
Bias Bandwidth
Hi, As far as I understand the Bias Bandwidth is a parameter that can be changed to avoid high or low lighting fluctuations, but I will also affects the bandwidth of the camera, which is restrictive to the maximum events per sec that can be processed.
What is the default bias value of IMX636? And how to calculate the event based the bias?
We read the default bias by the provided software and obtain the default pos_thres=102%,neg_thres=73%. It's very curious. If we set the intensity of two images as I1 and I2. We can get a positive event only if I2 - I1 > 102% * I1. It seems not very reasonable.
Band Pass filter bias units
I have an EVK3 and I'm trying to understand the units of the fall off and high pass filter settings. I'm hoping I can use this to target a specific frequency of motion if I know the expected frequency ahead of time. For instance, suppose I have a light
Relation of abolute bias values of Gen4.1 and IMX636
Hello, I read in the forum: "Setting the logging level to TRACE as described here : https://docs.prophesee.ai/3.0.1/faq.html#how-can-i-change-the-logging-level-at-runtime should help you to see absolute bias values on your Gen4.2 sensor (IMX636) and the
Where to get EVK4's Sensor Characterization?
Hi, Prophesee! Now I have an EVK4, but Sensor Characterization need to know, such as latency, shot noise rate, threshold and etc. In this link, https://support.prophesee.ai/portal/en/kb/prophesee-1/metavision-sensing/sensors, I found the documents of
Next Page