Any analyst, at the beginning of his work, goes through the hated stage of determining the identification of distribution parameters. Then, with the accumulation of experience, for him the reconciliation of the residual scatter obtained means that a stage, in the analysis of Big Data, has been passed and you can move on. It is no longer necessary to check hundreds of models for consistency with different regression equations, to look for segments with transients, to compose models. To torment yourself with doubts: "Maybe there is some other model that is more suitable?"
I thought: “What if you go by the opposite. See what white noise can do. Can white noise create something that our attention compares with a significant object from our experience? ”
Fig.White noise (file taken from the network, size 448h235).
')
On this issue, argued as follows:
What is the probability that horizontal and vertical lines of noticeable length will appear?
If they can appear, what is the probability that they will coincide with their origin along one of the coordinates and form a rectangular figure?
Further in the text, I will explain how these tasks associated with the analysis of Big Data.
The practical consequence of the discovery of only one, such a consecutive contract, in the base under study, enables us to assume that all the data presented are homogeneous.
When I was working on this material, an observation was made about the following. Everything developed data analysis methods are made for technologies when, by small natural observations, it is necessary to determine the parameters of a much larger population, according to 100 observations, to determine the properties of the general population of 1 million or more. And for modern tasks, when it is necessary to decompose a huge database, the tools developed by statistics are very laborious.