Today, Many global companies show clearly their tendency to analyze the disposal data and exploit it in order to enlarge their money-making machines. Now, We attempt to control the Dark Data and turning it into a pre-processed Big Data with new strategies. We try to create the ability to process and classify the Dark Data inspired by Biological Phenomena...
A solution to controlling and dealing with the dark data inspired by biological phenomena
1. Dark Data
A solution to
controlling and
dealing with the
inspired by
Biological
Phenomen
2. Ali Alizade Haqiqi
Table of Contents
–– Abstract ………………………………..……………………………..…………………… 3
I. Introduction ……………………………..…………………………………………..… 4
II. Procedure for paper submission (Body) .……………………………………………… 4
II.A. Current achievements of mankind in the field of data storage ………… 4
II.B. Using micro data to develop an enormous business …………………… 4
II.C. Provides an inspirational model ……………………………………….. 5
II.D. Provides an adequate example to the purposed model …………………. 5
II.E. Generalize aforementioned example to the developed model ………..… 5
II.F. The structure of submitted request .…………………………………….. 6
II.G. The first function: Routing, The key to staying on path ………………. 6
II.H. The second function: Reproduction exponentially, A solution to overcome
the Big Data .……………………………..…………………………………… 6
II.I. How is routing done within the auxiliary requests? …………………… 7
II.J. The storage form of analyzed data ……………………………………. 7
II.K. The last word ………………………………………………………….. 7
III. Conclusion ……………………………..……………………………………………… 8
III. References ……………………………..………………………………………………. 9
3. Abstract ––
Today, Many global companies show clearly their tendency to analyze the disposal data and
exploit it in order to enlarge their money-making machines. Now, We attempt to control the Dark
Data and turning it into a pre-processed Big Data with new strategies. We try to create the ability
to process and classify the Dark Data inspired by Biological Phenomena. We send a request to
Data Center using molecular propagation methods; so that, as the molecular reproduction
performed the propagation operation and create requests as the same type. And what is the form
of sent or propagated request type? , We send an original request along with two auxiliary
requests to Data Center inspired by molecules constituting the air which have the ability to
perform two crucial functions. That two auxiliary requests can flow out the routing operation
using sending signals between each three requests until the original request was being inside the
Data Center and its data range and randomly nearing and sticking to one of the single data. But if
the original request comes out of the range for any reason, Including disorders in sent signals or
being empty path ahead of request from data, Or as soon as the original request found a data, The
connection between auxiliary requests will be closed and each of them turns into an original
request along with two auxiliary requests. And finally, classify the Data Centers data in shortest
time possible and processes and analyze them using the Data Centers own power and then
improves the finding, analyzing and classifying data operations using neural deep learning and
etc. We can even control the Dark Data effortlessly using the aforementioned method and
consider all of that raw data in the Big Data range.
4. I. INTRODUCTION
The position of the human being has been coming to a point that makes talking tougher on any
matter without having any abundant information and knowledge on the aforementioned field.
Now, When our folk talking afoul with evolution like this, We can express the problem about
statistical reports and recorded data within the corporate. The data existed as treasures in the
depth of the Data Centers and its real value will not be understandable until someone processes
and monitors them. What we're referring to is the same Big Data. Now, We had a larger range,
How? When we notice that the Big Data are the part of our ability to processes and heavy
analysis of a larger part called Dark Data, The story will be different. Imagine you’ve been a very
long road and a laser flashlight with long range is on your hand. How many meters can you light
up from the road? 100 meters? 200 meters? Or 1 kilometers? Whether the road is equal to the
same size? That's right, The rest of the road will be dark. The Dark Data are the same that road.
Now, Imagine that the half of the allowable lighting distance, Each 500 meters one person
stands with the same flashlight. And all over the road is the same. In this case, In addition to we
could lighten up all over the road using flashlights. We also removed A power of lightning that
was going out at the end of the path due to long range. The roads exist a lot on the companies
that lead to the convenience of users on the Internet. The corporate had the ability to lighten up
a part of aforementioned roads, But as the time goes forward and various startups and
companies are arising in different fields. More roads are building where are dark.
II. PROCEDURE FOR PAPER SUBMISSION ( BODY )
A. Current achievements of mankind in the field of data storage
In the current period, The human beings do things that they were very fictitious in the previous
periods, and even impossible. Up to 50 years ago, No one would ever think that the data
collection may be stored in one or more super computer, Now store in the memory that is smaller
than a knuckle. And this rapid progress is never stopped. Already in the world greatest
universities including MIT and etc, Some activities are performed in the field of storage on DNA
and similar structures. The Data Centers that the corporates are in their ownership, Each of them
has a data on their self that is larger than a galaxy which is attempting to use the data on their
own business effectively. You asked, “What is the relationship of processing aforementioned the
volume of data and converting to analyzed information between the businesses?”. When you
have a business that is strongly associated with members of a society directly or indirectly;
Know, If you don't start to store practical data associated with the business process, We should
tell you; Actually, You started a failed business.
B. Using micro data to develop an enormous business
For example, Consider an online retailer, If the retailer can store the smallest user behavior on
entering into website until they leave the website and somehow analyze and classify
aforementioned valuable and bulky data, they guaranteed their business easily 5 to 10 years. For
example, When a user stays longer on a page when logging into a certain category or the user
5. spends more time on a part of the page with the mouse or when the user hits the site at certain
times, We can easily analyze the user based on the basis of aforementioned rare behaviors. Now,
Imagine that all the services that a user using it physically or in the form of software can be able
to communicate with each other. Then we can communicate the data. Then we can send a much
smarter, more advanced and on time recommendations to the users. So the user provides the most
beneficial priority at the right time. If a business customers absorption flow has done in this way,
There's no doubt that it can reach its highest profitability.
C. Provides an inspirational model
Now when we talked more about the greater part of the data that is stored on a corporate Data
Centers, Let us explain more about our topics. When we talk about this massive volume of the
data; The variety, The volume and the velocity of the growth of data will going out of
expectations. We often faced with the problems of huge data storage and processing to control a
Data Center Big Data. Now if the factors of different aforementioned types such as the variety,
The volume and etc are going to many humongous several times, We will get involved with a
severe crisis. But let's not be so superficial. Many modern technologies that we use it as a daily
habit, Could be a crucial step in human beings progress inspired by natural and Biological
Phenomena. We can mention a very obvious example like wright brothers that we all know that
they made it inspired by what, The machine that all thought is a madness. They made it be an
integral part of the life of humanity. Although the invention is primarily developed by Mr. Da
Vinci’s. It might be a little better to explain our solution by talking indirectly about a Biological
Phenomenon.
D. Provides an adequate example to the purposed model
We all know that air is composed of two hydrogen atoms and one Oxygen atom. We know very
well that if we combine molecules of a material that are denser than other substance, A material
that is less dense will be defeated. In other words, If we take Carbon Dioxide into the air, The
Carbon Dioxide overcomes the air. It's true, Our inhales and exhales produce Carbon Dioxide. So
naturally, A room must be evacuated from the air. Imagine that you're wearing a space suit and
all its pores are closed. And no Oxygen bottle connects to you. As a result, You maybe breathe 7
to 8 minutes in that situation. Because returned Carbon Dioxide dominates the air of the clothes
and practically no air remains to breathe. This topic is true about a room that no air enters into it
from any vents, Whereas you're slowly converting the Oxygen to Carbon Dioxide. Now imagine
we will release a knuckle of Carbon Dioxide into the air in a city, What will happen? Nothing.
It's true that the air was polluted by only small quantities of Carbon Dioxide But it makes no
difference. And we won't feel it because the fresh air is much more than it. Now if the molecules
of Carbon Dioxide had the ability to reproduce and spawn, What would happen? In this case,
Even if a molecule of Carbon Dioxide absorbs into the air, We were losing all Oxygen of the city
by reproducing exponentially in the shortest time.
E. Generalize aforementioned example to the developed model
6. Well, Let's turn back to the main topic. If we will send a request as the molecules of Carbon
Dioxide which is connected to one or more Data Center or a data set that we can call it as Dark
Data, Just like that we released A bit of Carbon Dioxide into the air. So if our submitted request
does the act of reproduction and spawning as soon as possible that arrives at a data into the Data
Center randomly and then the request classifies, processes and analyzes the incoming data
through the processing power of the destination (The Data Center) real-time. So, After a short
time, They analyze the Data Centers data or Dark Data sources using Neural Deep Learning
approaches, machine learning, artificial intelligence and genetic algorithms that the system
basically developed according to them. And turn all the massive and dense data into pre-
processed Big Data. Dense clusters set of Big Data that processed and analyzed in place of
storage and it delivered the results to us in the fastest possible time.
F. The structure of submitted request
As we know, That two processes are very challenging in this method. The first is finding a
random data via the sent request and the second is the probability of not finding only one data
within a given volume of dense data. While the sent request got out of the scope and range of
finding random data within the Data Center. If we want to explain more the first, We come to the
conclusion that finding data might be Time-Consuming for the request in a Data Center
randomly. Therefore, The form of sent request and the requests that are created by reproduction
and spawning should be such a form that they can make the aforementioned operation much
easier. Also, We will get help from Biological Phenomena and we do the operation based on
molecules constituting the air. This means that any request that we send out is actually a set of
three requests, Including the original request and two auxiliary requests, That the two auxiliary
requests have two functions.
G. The first function: Routing, The key to staying on path
Their first function is routing until the original request is in the range of Data Center, And as long
as the original request would not arrive at a data, At first the two auxiliary requests find the
nearest data using the signal transmission between the requests around it selves. And then if they
found nothing in its range, They find the random data using nested signals transmission between
the other auxiliary requests in the range of Data Center. And then they help them to route and
achieve data by the original requests. Each auxiliary request has the ability to send and receive
routing request for the original request asynchronously. And the second function is quite related
to the second challenge that we mentioned in previous sections.
H. The second function: Reproduction exponentially, A solution to overcome the Big Data
Our second challenge or indeed the second function of auxiliary request that is the random data
start to classify and analyze according to the previous saying after routing for the original request
and arriving at a random data, And on the other hand the original request begin the reproduction
7. that creates a new request with the structure similar to what has been said, And the auxiliary
requests of the original request turn into a new original request and each of them create two
auxiliary requests for their selves. So actually we have the routine reproduction operation of each
request, And on the other hand, The auxiliary requests have a huge impact on the growth rate,
The speed of arriving at a data and the speed of their analysis.
I. How is routing done within the auxiliary requests?
The other item that needs to be explained is the routing function of auxiliary requests. The
routing means finding the address of data on the Data Center randomly. When the auxiliary
requests are routing, They access to the memory randomly and start the navigation inspired by
the shape of DNA. Therefore, We get help from the Biological Phenomena for the third time to
describe the shape of the auxiliary requests before they're turned into the original requests. We
consider the number of legs or indeed the Angstroms that are related to the DNA as the
sequences number of auxiliary requests. It should be noted that we're ignoring the DNA type
(Regardless of type A, B or Z) and Angstroms of DNA whether they're useless in the
implementation of auxiliary requests on the aforementioned topic completely. The number of
auxiliary requests sequences obtain by sending a signal between the auxiliary requests to getting
the number of original requests. Accordingly, It can be said that each request that creates, The
sequences of its auxiliary requests would be created based on the number of active original
requests in the range of Data Center. So it's natural that the initial request would be sent to the
Data Center with two single sequences of auxiliary requests. And the sequences form of
placement when they're routing is that each sequence placed in a single location of memory and
each sequence have the ability to send directions to the original request individually. So it's
obvious, if each of the single sequences that are placed in a single location of memory which is
navigating serial, faces a data; It would be sent the address of nearest data to the original request.
As mentioned above, The navigation form of sequences are serial, But the accessing form of
them to the memory or the range of Data Centers are random.
J. The storage form of analyzed data
After that the huge data sets have been processed and analyzed completely, The analysis and
classification results would be stored encrypted and shared on the Data Center and the shared
keys would be sent to the origin for accessing the results and also all of the learned data are
stored on the destination of initial request by default, When they're routing and analyzing. And
encrypted addressing would be used by each of original and auxiliary requests. So, The only
challenge that remains for us would be the location of data storage. However, If we used the
aforementioned approach in the field of Internet of Things, The challenge doesn't occur again.
Because when we are using it in this context, We can just amplify the power and the ability of
the aforementioned Artificial Intelligence after any new data that was created and proclaim the
analyzed data after the classification to the users immediately and no data is stored on memory.
K. The last word
8. Aside from the storage form and related issues, The aforementioned approach has the ability to
get the corporate vision with respect to its reserves data and related industries to an enormous
evolution.
III. CONCLUSION
As we explained about “How can we use the related data to a massive Big Data or on a larger
scale, A clusters of Dark Data?” in detailed. It can be said that the purposed solution can have a
major evolution of human life than the previous generations. The purposed solution formed,
Inspired by nature around us and its related Biological Phenomena and it would also overcome
the largest collections of data. So it's very clear that controlling a set of Big Data is an accessible
job to it. As noted above, We can create alternatives easily inspired by the surrounding
phenomenon with a closer look and deep insight that is related to the fields of activity like
services, products and even branches of science that the humanity never thought to it. Great
peoples came into this world and they were just a starter or a stater of a way in the history that
has been more matured by the other peoples in the decades or centuries later.
9. IV. REFERENCES
[1] Andrew Travers, Georgi Muskhelishvili, DNA Structure and Function, 2 June 2015
[2] R. Fausto, J. Laane, J. Lundell, S.A. Brandán, E. Aunan and M.Biczysko, Molecular
Structure, 7 April 2015
[3] G. Bertone, L. Amendola, S. Profumo, T. Tait and L. Verde, Physics of the Dark
Universe, 18 December 2016
[4] Hui-Huang Hsu, Chuan-Yu Chang and Ching-Hsien Hsu, Big Data Analytics for Sensor-
Network Collected Intelligence, 7 February 2017
[5] AssocProf. Dimitris Zissis, Prof. Dimitris Askounis, Dr. Luca Gazzantti, Prof. Minos
Garofalakis and Dr. Fabio Mazzarella, Special Issue on Spatiotemporal Big Data Challenges,
Approaches, and Solutions, 14 February 2016
[6] Jules Berman, Principle of Big Data, 30 May 2013