Not exactly. It is a two-dimensional pixel array like a traditional camera, but each pixel is independent of the others, and triggers an event whenever the pixel intensity increases or decreases by some threshold, say 5%. These events are triggered at a very fine time resolution, effectively on the order of 10,000 Hz. The events have no scalar value however, they only indicate a positive or negative change in intensity. So if the camera is stationary you’ll see events caused by moving objects, and if the camera is moving you’ll see events caused by moving objects and stationary edges.
Also, since each pixel is its own autonomous unit, these cameras exhibit dynamic ranges many orders of magnitude greater than traditional cameras. Intuitively it means if you pointed it at a bright light, only the pixels on the light would get saturated and the rest could function as normal, unlike a traditional camera where the global exposure control would make all of the other pixels very dark in response to the light.
These properties are similar to retinal sensing, in which all-or-nothing spike events are triggered in response to changes in illumination, and sensitivity is adapted near-independently across retinal “pixels”.
This is getting away from the original topic, but the intent was to indicate that input to the brain is often very sparse.