NOT READY to be read, I made an error to fix
example: ababab... is a also a no-gradient image
Lately I have been thinking about fundamental images.
The smallest image is a single sample, a point. An image cannot be smaller then a sample.
Such point can have a quantization precision (eg 2 bit, 8 bit, 32 bit float of luminance...)
Image are useless to us unless you can see them. As well at some point extra precision (more samples, more precision per sample) yields no more significance to us the humans with a particular perception spectrum. In 2009 the common man knows the difference between so-called standard and so-called high definition television, knows somewhat the difference between a camera that captures 1000 000 samples versus one that claims 20 000 000 ones...
So at some point the return on investment with extra precision for a particular image fades as you further increase the resolution.
Single Point Images
So a single point sample image (or a color solid) is the base of all imaging. Such can be arbitrarily resize from infinitely small to infinitely large without changing anything. So a first distinction between the field of all potential images is possibly the average color of all the samples in it (assuming we have to start a sampling grid where the distance between all samples is equal).
Already a first distinction has to be made here. Such is a called colorspace. In photometric modeling something called a white point would be the reflection of the sun on a white board. That value we call 1.0. From the standpoint of light and sensing media we can also say that 0.0 would be zero photon hitting the board during the capture interval. From there our language already easily semantically slip, as we are forced to distinguish between mathematical linearity (where 0.5 is 50% grey) and light power of 2 (also known as gamma 2) and our own perception of a greyscale gradient as linear. This we will discuss elsewhere, for our purpose here, defining fundamental images, we will stay in the mathematical world where 0.5 is the center between 0 and 1.0.
So, values under 0 and over 1, undershooting and overshooting, is quite common / possible in image processing. But let's ignore them as well for now. Our purpose here is to identify if there are basic images that tells us about real or natural images. So the first fundamental or moment of an image would be the average of all it's sample values.
Random Images
At the other extreme we could say that the set of all possible values of a sample times the resolution of the image creates a potential field of all possible images. For our purpose though what matters is a particular subset which fits some perceptual threshold where a minor tiny variations cannot produce a result we can identify as different. This essentially is what allow us to do analog to digital conversions that are useful. So it converts the problem from an infinite set to a very large pseudo-infinite one.
Since we started our infinity bounding with the idea of a black and white point, the base random images would be one where each pixel in it has the same chance of being anywhere between 0 and 1.0, thus would average to 0.5 as a solid average. That is the higher the number of samples you have in an image (resolution), the more the initial noise field will tend to average to 0.5
Constant Gradient Images
So if you take any image whose size is larger then one sample and average the sample now in 4 quadrants, you will have 4 values, a 2x2 image. There are various to calculate that, one would be nearest corner (are you in quadrant 1,2,3,4 of the image? and average the pixels from the same quadrant and write that as the pixel value in the 2x2 image. Another technique would be for each corner to assign 1.0 for the current corner (the one on the side of the 2x2 image pixel being computed) and say the corner of the source image has 1.0 of weight and the opposite diagonal corner has 0.0 of weight and as a function of distance this sample in the source image will have a weight in the overall computation of the destination pixel relative to it's distance to the matching corner. These two techniques will usually produce a slightly different answer. However they will produce 2x2 different pixels. In some case within the epsilon of our computation precision (within the smallest representable value) There exists a set of images that will return the same value for the 4 pixels and the same value as the average pixel. This already tells us something about the source image. Such 2x2 image we would say has zero gradient. As we increase the size of the image to 3x3, 4x4 pixels if the source image is not a constant solid color image we will get some pixels that are different. The 2x2 image should average to the same value that all the samples of the source image averages to.
Constant Gradient Images : Sample Size Grids
What is not a Constant Gradient Image?
Mapping the Potential Image Field to a 2x2 version of the image
about samples outside of the 2x2 image
Furthermore, this further clarify the noise potential field. Now not all images are possible if the result is a plane that passed through our 2x2 image. Only a slice through that statistical possibility. So because we started with a particular average color our noise image averaged a particular value - the noise field itself had a color which for now we assume is somewhere between 0.0 and 1.0. If we assume our samples are between 0.0 and 1.0 and say the average sample is 0.6 this is a bit like saying that half the samples would be between 0.6 and 1.0 and half between 0.0 and 0.6. That would be innacurate since if the average sample was then 0.8 on one side and 0.3 on the other side then the average sample would be 0.55 not 0.6. 0.6/0.55 is about 1.09 so it's actually 0.5 X 1.09 so 54.50% that would be equal or over 0.6. When we know 4 values instead of 1 then we can do something like calculate an interpolated value between to modulate our potential image based on the normalized distance (between 0 and 1) of that potential image. So on one side of the potential noise field the maybe it's 55% of being over 0.6 and on the opposite side maybe it's 53%...
Still a lot of way to go to arrive at a real image. As we explain the first volume in our semi-criticism of Wolframe model of universe, it would take a long time to tune in on a playboy centerfold even if we started with a 2x2 version of such image.
Sparse Higher Resolutions
Adaptive Sampling
It's now popular for ray-tracing rendering to perform what is known as adaptive sampling, that is where across a scene to render one varies the number of samples based on some critera which could be the curvature if you like. As we know that the edge of curved objects will need more samples to produce a properly anti-aliased result. Not to be confusing here but a common rendering way to describe things would be a number of samples are collected into a rendered pixel. Imagine a stochastic process that throws dots within a circle (each a ray tracer sample, a ray sent to get color back). If your first N samples come from different triangles or other decomposition then probably you need more samples (hence adaptive). By the same principle an image reconstruction filter could decide to work harder on certain portions of the image.
Point Spread Function
An imaging system produce a weighted sum over a 2D imaging area, the result pixel sample. Physics is such that photons don't interact. What is of interest to our discussion is in photographic term described as the circle of confusion. The nature of our optics is such that real lenses do not converge to a point but to a spot. The shape of such spot is the Point Spread function. While the Circle of confusion would be a particular number for a particular distance from the lens at which the element will be still perceived as a point. So is also called the largest blur circle.