This paper discloses in synthesis a super-computation computer architecture (CA) model, presently a
provisional Patent Application at INPI (nº 116408). The outline is focused on a method to perform
computation at or near the speed of light, resorting to an inversion of the Princeton CA. It expands from
isomorphic binary/RGB (typical) digital “images”, in a network of (UTM)s over Turing-machines (M)s.
From the binary/RGB code, an arithmetic theory of (typical) digital images permits fully
synchronous/orthogonal calculus in parallelism, wherefrom an exponential surplus is achieved. One such
architecture depends on any “cell”-like exponential-prone basis such as the “pixel”, or rather the RGB
“octet-byte”, limited as it may be, once it is congruent with any wave-particle duality principle in
observable objects under the electromagnetic spectrum and reprogrammable designed. Well-ordered
instructions in binary/RGB modules are, further, programming composed to alter the structure of the
Internet, in virtual/virtuous eternal recursion/recurrence, under man-machine/machine-machine
communication ontology.
Word Recognition in Continuous Speech and Speaker Independent by Means of Rec...CSCJournals
Artificial neural networks have been applied successfully in many static systems but present some weaknesses if patterns involve a temporal component. Let’s note for example in speech recognition or contextual information, where different of the time interval, is crucial for comprehension. Speech, being a temporal form of sensory input, is a natural candidate for investigating temporal coding in neural networks. It is only through comprehension of the temporal relationship between different sounds which make up a spoken word or sentence that speech becomes intelligible. In fact we present in this paper presents three variants of self-organizing maps (SOM), the Leaky Integrators Neurons (LIN), the Spiking_SOM (SSOM) and the recurrent Spiking_SOM (RSSOM) models. The proposed variants is like the basic SOM, however it represents the characteristic to modify the learning function and the choice of the best matching unit (BMU). The case study of the proposed SOM variants is word recognition in continuous speech and speaker independent. The proposed SOM variants show good robustness and high word recognition rates.
An Efficient Block Matching Algorithm Using Logical ImageIJERA Editor
Motion estimation, which has been widely used in various image sequence coding schemes, plays a key role in the transmission and storage of video signals at reduced bit rates. There are two classes of motion estimation methods, Block matching algorithms (BMA) and Pel-recursive algorithms (PRA). Due to its implementation simplicity, block matching algorithms have been widely adopted by various video coding standards such as CCITT H.261, ITU-T H.263, and MPEG. In BMA, the current image frame is partitioned into fixed-size rectangular blocks. The motion vector for each block is estimated by finding the best matching block of pixels within the search window in the previous frame according to matching criteria. The goal of this work is to find a fast method for motion estimation and motion segmentation using proposed model. Recent day Communication between ends is facilitated by the development in the area of wired and wireless networks. And it is a challenge to transmit large data file over limited bandwidth channel. Block matching algorithms are very useful in achieving the efficient and acceptable compression. Block matching algorithm defines the total computation cost and effective bit budget. To efficiently obtain motion estimation different approaches can be followed but above constraints should be kept in mind. This paper presents a novel method using three step and diamond algorithms with modified search pattern based on logical image for the block based motion estimation. It has been found that, the improved PSNR value obtained from proposed algorithm shows a better computation time (faster) as compared to original Three step Search (3SS/TSS ) method .The experimental results based on the number of video sequences were presented to demonstrate the advantages of proposed motion estimation technique.
Design and implementation of video tracking system based on camera field of viewsipij
The basic idea of this paper is to design and implement of video tracking system based on Camera Field of
View (CFOV), Otsu’s method was used to detect targets such as vehicles and people. Whereas most
algorithms were spent a lot of time to execute the process, an algorithm was developed to achieve it in a
little time. The histogram projection was used in both directional to detect target from search region,
which is robust to various light conditions in Charge Couple Device (CCD) camera images and saves
computation time.
Our algorithm based on background subtraction, and normalize cross correlation operation from a series
of sequential sub images can estimate the motion vector. Camera field of view (CFOV) was determined and
calibrated to find the relation between real distance and image distance. The system was tested by
measuring the real position of object in the laboratory and compares it with the result of computed one. So
these results are promising to develop the system in future.
Word Recognition in Continuous Speech and Speaker Independent by Means of Rec...CSCJournals
Artificial neural networks have been applied successfully in many static systems but present some weaknesses if patterns involve a temporal component. Let’s note for example in speech recognition or contextual information, where different of the time interval, is crucial for comprehension. Speech, being a temporal form of sensory input, is a natural candidate for investigating temporal coding in neural networks. It is only through comprehension of the temporal relationship between different sounds which make up a spoken word or sentence that speech becomes intelligible. In fact we present in this paper presents three variants of self-organizing maps (SOM), the Leaky Integrators Neurons (LIN), the Spiking_SOM (SSOM) and the recurrent Spiking_SOM (RSSOM) models. The proposed variants is like the basic SOM, however it represents the characteristic to modify the learning function and the choice of the best matching unit (BMU). The case study of the proposed SOM variants is word recognition in continuous speech and speaker independent. The proposed SOM variants show good robustness and high word recognition rates.
An Efficient Block Matching Algorithm Using Logical ImageIJERA Editor
Motion estimation, which has been widely used in various image sequence coding schemes, plays a key role in the transmission and storage of video signals at reduced bit rates. There are two classes of motion estimation methods, Block matching algorithms (BMA) and Pel-recursive algorithms (PRA). Due to its implementation simplicity, block matching algorithms have been widely adopted by various video coding standards such as CCITT H.261, ITU-T H.263, and MPEG. In BMA, the current image frame is partitioned into fixed-size rectangular blocks. The motion vector for each block is estimated by finding the best matching block of pixels within the search window in the previous frame according to matching criteria. The goal of this work is to find a fast method for motion estimation and motion segmentation using proposed model. Recent day Communication between ends is facilitated by the development in the area of wired and wireless networks. And it is a challenge to transmit large data file over limited bandwidth channel. Block matching algorithms are very useful in achieving the efficient and acceptable compression. Block matching algorithm defines the total computation cost and effective bit budget. To efficiently obtain motion estimation different approaches can be followed but above constraints should be kept in mind. This paper presents a novel method using three step and diamond algorithms with modified search pattern based on logical image for the block based motion estimation. It has been found that, the improved PSNR value obtained from proposed algorithm shows a better computation time (faster) as compared to original Three step Search (3SS/TSS ) method .The experimental results based on the number of video sequences were presented to demonstrate the advantages of proposed motion estimation technique.
Design and implementation of video tracking system based on camera field of viewsipij
The basic idea of this paper is to design and implement of video tracking system based on Camera Field of
View (CFOV), Otsu’s method was used to detect targets such as vehicles and people. Whereas most
algorithms were spent a lot of time to execute the process, an algorithm was developed to achieve it in a
little time. The histogram projection was used in both directional to detect target from search region,
which is robust to various light conditions in Charge Couple Device (CCD) camera images and saves
computation time.
Our algorithm based on background subtraction, and normalize cross correlation operation from a series
of sequential sub images can estimate the motion vector. Camera field of view (CFOV) was determined and
calibrated to find the relation between real distance and image distance. The system was tested by
measuring the real position of object in the laboratory and compares it with the result of computed one. So
these results are promising to develop the system in future.
A Scheme for Joint Signal Reconstruction in Wireless Multimedia Sensor Networksijma
In context aware wireless multimedia sensor networks, scenarios are usually such that
signals of multiple distributed sensors contain a common sparse component and each individual
signal owns an innovation sparse component. So distributed compressive sensing based on joint
sparsity of a signal ensemble concept exploits both these intra- and inter- signal correlation structures
and compress signals to the extent possible. This paper proposes an optimized reconstruction
scheme based on joint sparsity model which is derived from the distributed compressive sensing. In
this regard, based on distributed compressive sensing, a joint reconstruction scheme is proposed to
compress and reconstruct ensemble of signals even in large scale data transmission. Furthermore,
simulation results show the effectiveness of the proposed method in diverse compression ratios and
processing times in comparison with the joint sparsity model and individual compressive sensing
reconstruction methods.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Constructing sierpinski gasket using gp us arraysDr Amira Bibo
A fractal is a mathematical set that typically displays self-similar
patterns, which means it is "the same from near as from far".
Fractals may be exactly the same at every scale, they may be
nearly the same at different scales. The concept of fractal extends
beyond trivial self-similarity and includes the idea of a detailed
pattern repeating itself. The algorithms to constructing different
fractal shapes in many cases typically involve large amounts of
floating point computation, to which modern GPUs are well
suited. In this paper we will construct Sierpinski Gasket using
GPUs arrays.
Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...Polytechnique Montreal
Cognitive radio allows unlicensed (cognitive) users to use licensed frequency bands by exploiting spectrum sensing techniques to detect whether or not the licensed (primary) users are present. In this paper, we present a compressed sensing applied to spectrum-occupancy detection in wide-band applications. The collected analog signals from each cognitive radio (CR) receiver at a fusion center are transformed to discrete-time signals by using analog-to-information converter (AIC) and then employed to calculate the autocorrelation. For signal reconstruction, we exploit a novel approach to solve the optimization problem consisting of minimizing both a quadratic (l2) error term and an l1-regularization term. In specific, we propose the Basic gradient projection (GP) and projected Barzilai-Borwein (PBB) algorithm to offer a better performance in terms of the mean squared error of the power spectrum density estimate and the detection probability of licensed signal occupancy.
A new block cipher for image encryption based on multi chaotic systemsTELKOMNIKA JOURNAL
In this paper, a new algorithm for image encryption is proposed based on three chaotic systems which are Chen system,logistic map and two-dimensional (2D) Arnold cat map. First, a permutation scheme is applied to the image, and then shuffled image is partitioned into blocks of pixels. For each block, Chen system is employed for confusion and then logistic map is employed for generating subsititution-box (S-box) to substitute image blocks. The S-box is dynamic, where it is shuffled for each image block using permutation operation. Then, 2D Arnold cat map is used for providing diffusion, after that XORing the result using Chen system to obtain the encrypted image.The high security of proposed algorithm is experimented using histograms, unified average changing intensity (UACI), number of pixels change rate (NPCR), entropy, correlation and keyspace analyses.
V. Caglioti, A. Giusti, A. Riva, M. Uberti: "Drawing Motion without Understanding It".
Proc. of International Symposium on Visual Computing (ISVC) 2009. Oral presentation (acceptance rate ~30%). Volt / Microsoft MSDN Best Paper Award.
Fixed-Point Code Synthesis for Neural Networksgerogepatton
Over the last few years, neural networks have started penetrating safety critical systems to take decisions in robots, rockets, autonomous driving car, etc. A problem is that these critical systems often have limited computing resources. Often, they use the fixed-point arithmetic for its many advantages (rapidity, compatibility with small memory devices.) In this article, a new technique is introduced to tune the formats (precision) of already trained neural networks using fixed-point arithmetic, which can be implemented using integer operations only. The new optimized neural network computes the output with fixed-point numbers without modifying the accuracy up to a threshold fixed by the user. A fixed-point code is synthesized for the new optimized neural network ensuring the respect of the threshold for any input vector belonging the range [xmin, xmax] determined during the analysis. From a technical point of view, we do a preliminary analysis of our floating neural network to determine the worst cases, then we generate a system of linear constraints among integer variables that we can solve by linear programming. The solution of this system is the new fixed-point format of each neuron. The experimental results obtained show the efficiency of our method which can ensure that the new fixed-point neural network has the same behavior as the initial floating-point neural network.
Improved Parallel Algorithm for Time Series Based Forecasting Using OTIS-MeshIDES Editor
Forecasting always plays an important role in
business, technology and many others and it helps
organizations to increase profits, reduce lost sales and more
efficient production planning. A parallel algorithm for
forecasting reported recently on OTIS-Mesh[9]. This parallel
algorithm requires 5( Root n– 1) electronic steps and 4 optical
steps. In this paper we present an improved parallel algorithm
for time series short term forecasting using OTIS-Mesh. This
parallel algorithm requires 5(-1) electronic steps and 1 optical
step using same number of I/O ports as considered in [9] and
shown to be an improvement over the parallel algorithm for
time series forecasting using OTIS-Mesh [9].
Vector algebra for Steep Slope Models analysisGeoCommunity
Natalia Kolecka: Vector algebra for Steep Slope Models analysis (poster), 9th International Symposium GIS Ostrava, VŠB – Technical Univerzity of Ostrava, from 23rd to 25th January 2012
Hardware progress has enabled solutions which were historically computationally intractable. This is particularly true in video analysis. This technological advance has opened a new frontier of problems. Within this expanse, we have chosen the classic problem of depth inference from images. Specifically, given a sequence of images captured over time, we output depth maps corresponding one-to-one with the input sequence. As a spatiotemporal problem, we were motivated to model it with convolutions (spatial) andLSTMs (temporal). These are used in a U-Net encoder-decoder architecture. The results indicate some potential in such an approach, the process by which we came to this conclusion is detailed below
Template matching is a basic method in image analysis to extract useful information from images. In this
paper, we suggest a new method for pattern matching. Our method transform the template image from two
dimensional image into one dimensional vector. Also all sub-windows (same size of template) in the
reference image will transform into one dimensional vectors. The three similarity measures SAD, SSD, and
Euclidean are used to compute the likeness between template and all sub-windows in the reference image
to find the best match. The experimental results show the superior performance of the proposed method
over the conventional methods on various template of different sizes.
To develop an Application to visualize the key board of computer with the concept of image processing. The virtual
keyboard should be accessible and functioning. The keyboard must give input to computer. With the help of camera, image
of keyboard will be fetched. The typing will be captured by camera, as we type on cardboard simply drawn on paper. Camera
will capture finger movement while typing. So basically this is giving the virtual keyboard.
As the technology advances, more and more systems are introduced which will look after the users comfort. Few
years before hard switches were used as keys. Traditional QWERTY keyboards are bulky and offer very little in terms of
enhancements. Now-a-days soft touch keypads are much popular in the market. These keypads give an elegant look and a
better feel. Currently keyboards are static and their interactivity and usability would increase if they were made dynamic and
adaptable. Various on-screen virtual keyboards are available but it is difficult to accommodate full sized keyboard on the
screen as it creates hindrance to see the documents being typed. Virtual Keyboard has no physical appearance. Although
other forms of Virtual Keyboards exist; they provide solutions using specialized devices such as 3D cameras. Due to this, a
practical implementation of such keyboards is not feasible. The Virtual Keyboard that we propose uses only a standard web
camera, with no additional hardware. Thus we see that the new technology always has more Benefits and is more userfriendly.
A Scheme for Joint Signal Reconstruction in Wireless Multimedia Sensor Networksijma
In context aware wireless multimedia sensor networks, scenarios are usually such that
signals of multiple distributed sensors contain a common sparse component and each individual
signal owns an innovation sparse component. So distributed compressive sensing based on joint
sparsity of a signal ensemble concept exploits both these intra- and inter- signal correlation structures
and compress signals to the extent possible. This paper proposes an optimized reconstruction
scheme based on joint sparsity model which is derived from the distributed compressive sensing. In
this regard, based on distributed compressive sensing, a joint reconstruction scheme is proposed to
compress and reconstruct ensemble of signals even in large scale data transmission. Furthermore,
simulation results show the effectiveness of the proposed method in diverse compression ratios and
processing times in comparison with the joint sparsity model and individual compressive sensing
reconstruction methods.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Constructing sierpinski gasket using gp us arraysDr Amira Bibo
A fractal is a mathematical set that typically displays self-similar
patterns, which means it is "the same from near as from far".
Fractals may be exactly the same at every scale, they may be
nearly the same at different scales. The concept of fractal extends
beyond trivial self-similarity and includes the idea of a detailed
pattern repeating itself. The algorithms to constructing different
fractal shapes in many cases typically involve large amounts of
floating point computation, to which modern GPUs are well
suited. In this paper we will construct Sierpinski Gasket using
GPUs arrays.
Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...Polytechnique Montreal
Cognitive radio allows unlicensed (cognitive) users to use licensed frequency bands by exploiting spectrum sensing techniques to detect whether or not the licensed (primary) users are present. In this paper, we present a compressed sensing applied to spectrum-occupancy detection in wide-band applications. The collected analog signals from each cognitive radio (CR) receiver at a fusion center are transformed to discrete-time signals by using analog-to-information converter (AIC) and then employed to calculate the autocorrelation. For signal reconstruction, we exploit a novel approach to solve the optimization problem consisting of minimizing both a quadratic (l2) error term and an l1-regularization term. In specific, we propose the Basic gradient projection (GP) and projected Barzilai-Borwein (PBB) algorithm to offer a better performance in terms of the mean squared error of the power spectrum density estimate and the detection probability of licensed signal occupancy.
A new block cipher for image encryption based on multi chaotic systemsTELKOMNIKA JOURNAL
In this paper, a new algorithm for image encryption is proposed based on three chaotic systems which are Chen system,logistic map and two-dimensional (2D) Arnold cat map. First, a permutation scheme is applied to the image, and then shuffled image is partitioned into blocks of pixels. For each block, Chen system is employed for confusion and then logistic map is employed for generating subsititution-box (S-box) to substitute image blocks. The S-box is dynamic, where it is shuffled for each image block using permutation operation. Then, 2D Arnold cat map is used for providing diffusion, after that XORing the result using Chen system to obtain the encrypted image.The high security of proposed algorithm is experimented using histograms, unified average changing intensity (UACI), number of pixels change rate (NPCR), entropy, correlation and keyspace analyses.
V. Caglioti, A. Giusti, A. Riva, M. Uberti: "Drawing Motion without Understanding It".
Proc. of International Symposium on Visual Computing (ISVC) 2009. Oral presentation (acceptance rate ~30%). Volt / Microsoft MSDN Best Paper Award.
Fixed-Point Code Synthesis for Neural Networksgerogepatton
Over the last few years, neural networks have started penetrating safety critical systems to take decisions in robots, rockets, autonomous driving car, etc. A problem is that these critical systems often have limited computing resources. Often, they use the fixed-point arithmetic for its many advantages (rapidity, compatibility with small memory devices.) In this article, a new technique is introduced to tune the formats (precision) of already trained neural networks using fixed-point arithmetic, which can be implemented using integer operations only. The new optimized neural network computes the output with fixed-point numbers without modifying the accuracy up to a threshold fixed by the user. A fixed-point code is synthesized for the new optimized neural network ensuring the respect of the threshold for any input vector belonging the range [xmin, xmax] determined during the analysis. From a technical point of view, we do a preliminary analysis of our floating neural network to determine the worst cases, then we generate a system of linear constraints among integer variables that we can solve by linear programming. The solution of this system is the new fixed-point format of each neuron. The experimental results obtained show the efficiency of our method which can ensure that the new fixed-point neural network has the same behavior as the initial floating-point neural network.
Improved Parallel Algorithm for Time Series Based Forecasting Using OTIS-MeshIDES Editor
Forecasting always plays an important role in
business, technology and many others and it helps
organizations to increase profits, reduce lost sales and more
efficient production planning. A parallel algorithm for
forecasting reported recently on OTIS-Mesh[9]. This parallel
algorithm requires 5( Root n– 1) electronic steps and 4 optical
steps. In this paper we present an improved parallel algorithm
for time series short term forecasting using OTIS-Mesh. This
parallel algorithm requires 5(-1) electronic steps and 1 optical
step using same number of I/O ports as considered in [9] and
shown to be an improvement over the parallel algorithm for
time series forecasting using OTIS-Mesh [9].
Vector algebra for Steep Slope Models analysisGeoCommunity
Natalia Kolecka: Vector algebra for Steep Slope Models analysis (poster), 9th International Symposium GIS Ostrava, VŠB – Technical Univerzity of Ostrava, from 23rd to 25th January 2012
Hardware progress has enabled solutions which were historically computationally intractable. This is particularly true in video analysis. This technological advance has opened a new frontier of problems. Within this expanse, we have chosen the classic problem of depth inference from images. Specifically, given a sequence of images captured over time, we output depth maps corresponding one-to-one with the input sequence. As a spatiotemporal problem, we were motivated to model it with convolutions (spatial) andLSTMs (temporal). These are used in a U-Net encoder-decoder architecture. The results indicate some potential in such an approach, the process by which we came to this conclusion is detailed below
Template matching is a basic method in image analysis to extract useful information from images. In this
paper, we suggest a new method for pattern matching. Our method transform the template image from two
dimensional image into one dimensional vector. Also all sub-windows (same size of template) in the
reference image will transform into one dimensional vectors. The three similarity measures SAD, SSD, and
Euclidean are used to compute the likeness between template and all sub-windows in the reference image
to find the best match. The experimental results show the superior performance of the proposed method
over the conventional methods on various template of different sizes.
To develop an Application to visualize the key board of computer with the concept of image processing. The virtual
keyboard should be accessible and functioning. The keyboard must give input to computer. With the help of camera, image
of keyboard will be fetched. The typing will be captured by camera, as we type on cardboard simply drawn on paper. Camera
will capture finger movement while typing. So basically this is giving the virtual keyboard.
As the technology advances, more and more systems are introduced which will look after the users comfort. Few
years before hard switches were used as keys. Traditional QWERTY keyboards are bulky and offer very little in terms of
enhancements. Now-a-days soft touch keypads are much popular in the market. These keypads give an elegant look and a
better feel. Currently keyboards are static and their interactivity and usability would increase if they were made dynamic and
adaptable. Various on-screen virtual keyboards are available but it is difficult to accommodate full sized keyboard on the
screen as it creates hindrance to see the documents being typed. Virtual Keyboard has no physical appearance. Although
other forms of Virtual Keyboards exist; they provide solutions using specialized devices such as 3D cameras. Due to this, a
practical implementation of such keyboards is not feasible. The Virtual Keyboard that we propose uses only a standard web
camera, with no additional hardware. Thus we see that the new technology always has more Benefits and is more userfriendly.
EFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGESijcnac
This project presents a new image compression technique for the coding of retinal and
fingerprint images. Retinal images are used to detect diseases like diabetes or
hypertension. Fingerprint images are used for the security purpose. In this work, the
contourlet transform of the retinal and fingerprint image is taken first. The coefficients of
the contourlet transform are quantized using adaptive multistage vector quantization
scheme. The number of code vectors in the adaptive vector quantization scheme depends
on the dynamic range of the input image.
Quality Measurements of Lossy Image Steganography Based on H-AMBTC Technique ...AM Publications,India
Steganography is a type of image information concealing technique which hides a secret message in a different media such as image, video and audio etc, called cover file. The main idea of steganography is to provide security to private or public data. In this paper we combined among Hadamard transformation and Absolute Moment Block Truncation Coding to make a new concept called (H-AMBTC), this concept used for compressing the cover file and conceal the secret data into the cover file. The H-AMBTC compression is not only image compression, but it is more than the AMBTC technique as only half of the number of pixels in the binary converted image are transmitted. In this paper, the comparison process of the H-AMBTC technique is done for 2x2, 4x4, 8x8 and 16x16 block sizes. H-AMBTC is a lossy technique as the cover image and the secret image can be recovered completely.
NUMBER PLATE IMAGE DETECTION FOR FAST MOTION VEHICLES USING BLUR KERNEL ESTIM...paperpublications3
Abstract: Many recent advancements have been introduced in both hardware and software technologies. Out of these technologies, we gain a lot of interest in Image Processing. As the eccentric identification of a vehicle, number plate is a key clue to discover theft and over-speed vehicles. The captured images from the camera are always in low resolution and suffer severe loss of edge information, which cast great challenge to existing blind deblurring methods. The blur kernel can be showed as linear uniform convolution and with angle and length estimation. In this paper, sparse representation is used to identify the blur kernel. Then, the length of the motion kernel has been estimated with Radon transform in Fourier domain. We evaluate our approach on real-world images and compare with several popular blind image deblurring algorithms. Based on the results obtained the supremacy of our proposed approach in terms of efficacy.
Keywords: Image Segmentation, Angle Estimation, Length Estimation, Deconvolution Algorithm, Artificial Neural Network.
Title: NUMBER PLATE IMAGE DETECTION FOR FAST MOTION VEHICLES USING BLUR KERNEL ESTIMATION AND ANN
Author: S.KALAIVANI, K.PRAVEENA, T.PREETHI, N.PUNITHA
ISSN 2350-1022
International Journal of Recent Research in Mathematics Computer Science and Information Technology
Paper Publications
A Real Time Image Processing Based Fire Safety Intensive Automatic Assistance...IJMTST Journal
Fire usually cause serious disasters. Thus, fire detection has been an important issue to protect human life
and property. In this project, I propose a fast and practical real-time image-based fire flame detection method
based on colour pair analysis and intensity level algorithm. Then, based on the above fire flame colour
features model, regions with fire-like colours are roughly separated from each frame of the test videos.
Besides segmenting fire flame regions, background objects with similar fire colours or caused by colour shift
resulted from the reflection of fire flames are also extracted from the image during the above colour
separation process. To remove these spurious fire-like regions, the image difference method and the invented
colour masking technique are applied. The device can detect fire by using Artificial Neural Network (ANN).
Finally device automatically control the fire safety assistance. This method was tested with Raspberry pi B+
Board interface with camera module.
A Smart Camera Processing Pipeline for Image Applications Utilizing Marching ...sipij
Image processing in machine vision is a challenging task because often real-time requirements have to be met in these systems. To accelerate the processing tasks in machine vision and to reduce data transfer latencies, new architectures for embedded systems in intelligent cameras are required. Furthermore, innovative processing approaches are necessary to realize these architectures efficiently. Marching Pixels are such a processing scheme, based on Organic Computing principles, and can be applied for example to determine object centroids in binary or gray-scale images. In this paper, we present a processing pipeline for smart camera systems utilizing such Marching Pixel algorithms. It consists of a buffering template for image pre-processing tasks in a FPGA to enhance captured images and an ASIC for the efficient realization of Marching Pixel approaches. The ASIC achieves a speedup of eight for the realization of Marching Pixel algorithms, compared to a common medium performance DSP platform.
6th International Conference on Machine Learning & Applications (CMLA 2024)ClaraZara1
6th International Conference on Machine Learning & Applications (CMLA 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of on Machine Learning & Applications.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
NUMERICAL SIMULATIONS OF HEAT AND MASS TRANSFER IN CONDENSING HEAT EXCHANGERS...ssuser7dcef0
Power plants release a large amount of water vapor into the
atmosphere through the stack. The flue gas can be a potential
source for obtaining much needed cooling water for a power
plant. If a power plant could recover and reuse a portion of this
moisture, it could reduce its total cooling water intake
requirement. One of the most practical way to recover water
from flue gas is to use a condensing heat exchanger. The power
plant could also recover latent heat due to condensation as well
as sensible heat due to lowering the flue gas exit temperature.
Additionally, harmful acids released from the stack can be
reduced in a condensing heat exchanger by acid condensation. reduced in a condensing heat exchanger by acid condensation.
Condensation of vapors in flue gas is a complicated
phenomenon since heat and mass transfer of water vapor and
various acids simultaneously occur in the presence of noncondensable
gases such as nitrogen and oxygen. Design of a
condenser depends on the knowledge and understanding of the
heat and mass transfer processes. A computer program for
numerical simulations of water (H2O) and sulfuric acid (H2SO4)
condensation in a flue gas condensing heat exchanger was
developed using MATLAB. Governing equations based on
mass and energy balances for the system were derived to
predict variables such as flue gas exit temperature, cooling
water outlet temperature, mole fraction and condensation rates
of water and sulfuric acid vapors. The equations were solved
using an iterative solution technique with calculations of heat
and mass transfer coefficients and physical properties.
Forklift Classes Overview by Intella PartsIntella Parts
Discover the different forklift classes and their specific applications. Learn how to choose the right forklift for your needs to ensure safety, efficiency, and compliance in your operations.
For more technical information, visit our website https://intellaparts.com
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Water billing management system project report.pdfKamal Acharya
Our project entitled “Water Billing Management System” aims is to generate Water bill with all the charges and penalty. Manual system that is employed is extremely laborious and quite inadequate. It only makes the process more difficult and hard.
The aim of our project is to develop a system that is meant to partially computerize the work performed in the Water Board like generating monthly Water bill, record of consuming unit of water, store record of the customer and previous unpaid record.
We used HTML/PHP as front end and MYSQL as back end for developing our project. HTML is primarily a visual design environment. We can create a android application by designing the form and that make up the user interface. Adding android application code to the form and the objects such as buttons and text boxes on them and adding any required support code in additional modular.
MySQL is free open source database that facilitates the effective management of the databases by connecting them to the software. It is a stable ,reliable and the powerful solution with the advanced features and advantages which are as follows: Data Security.MySQL is free open source database that facilitates the effective management of the databases by connecting them to the software.
Online aptitude test management system project report.pdfKamal Acharya
The purpose of on-line aptitude test system is to take online test in an efficient manner and no time wasting for checking the paper. The main objective of on-line aptitude test system is to efficiently evaluate the candidate thoroughly through a fully automated system that not only saves lot of time but also gives fast results. For students they give papers according to their convenience and time and there is no need of using extra thing like paper, pen etc. This can be used in educational institutions as well as in corporate world. Can be used anywhere any time as it is a web based application (user Location doesn’t matter). No restriction that examiner has to be present when the candidate takes the test.
Every time when lecturers/professors need to conduct examinations they have to sit down think about the questions and then create a whole new set of questions for each and every exam. In some cases the professor may want to give an open book online exam that is the student can take the exam any time anywhere, but the student might have to answer the questions in a limited time period. The professor may want to change the sequence of questions for every student. The problem that a student has is whenever a date for the exam is declared the student has to take it and there is no way he can take it at some other time. This project will create an interface for the examiner to create and store questions in a repository. It will also create an interface for the student to take examinations at his convenience and the questions and/or exams may be timed. Thereby creating an application which can be used by examiners and examinee’s simultaneously.
Examination System is very useful for Teachers/Professors. As in the teaching profession, you are responsible for writing question papers. In the conventional method, you write the question paper on paper, keep question papers separate from answers and all this information you have to keep in a locker to avoid unauthorized access. Using the Examination System you can create a question paper and everything will be written to a single exam file in encrypted format. You can set the General and Administrator password to avoid unauthorized access to your question paper. Every time you start the examination, the program shuffles all the questions and selects them randomly from the database, which reduces the chances of memorizing the questions.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesChristina Lin
Traditionally, dealing with real-time data pipelines has involved significant overhead, even for straightforward tasks like data transformation or masking. However, in this talk, we’ll venture into the dynamic realm of WebAssembly (WASM) and discover how it can revolutionize the creation of stateless streaming pipelines within a Kafka (Redpanda) broker. These pipelines are adept at managing low-latency, high-data-volume scenarios.
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
U-MENTALISM PATENT: THE BEGINNING OF CINEMATIC SUPERCOMPUTATION
1. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
DOI : 10.5121/ijwest.2021.12101 1
U-MENTALISM PATENT: THE BEGINNING OF
CINEMATIC SUPERCOMPUTATION
Luís Homem
Centro de Filosofia das Ciências da Universidade de Lisboa
ABSTRACT
This paper discloses in synthesis a super-computation computer architecture (CA) model, presently a
provisional Patent Application at INPI (nº 116408). The outline is focused on a method to perform
computation at or near the speed of light, resorting to an inversion of the Princeton CA. It expands from
isomorphic binary/RGB (typical) digital “images”, in a network of (UTM)s over Turing-machines (M)s.
From the binary/RGB code, an arithmetic theory of (typical) digital images permits fully
synchronous/orthogonal calculus in parallelism, wherefrom an exponential surplus is achieved. One such
architecture depends on any “cell”-like exponential-prone basis such as the “pixel”, or rather the RGB
“octet-byte”, limited as it may be, once it is congruent with any wave-particle duality principle in
observable objects under the electromagnetic spectrum and reprogrammable designed. Well-ordered
instructions in binary/RGB modules are, further, programming composed to alter the structure of the
Internet, in virtual/virtuous eternal recursion/recurrence, under man-machine/machine-machine
communication ontology.
KEYWORDS
U-Mentalism, Super-computation, Computer Architecture, Cybernetics, Programming Languages Design
1. INTRODUCTION
This document is intended to serve as white paper to describe in the most possible composed
details and in anticipation the technology of U-Mentalism. As referred beforehand “U-Mentalism
is a philosophical and programming idea that proposes a singular (one only and individual,
intensional) and universal (all and wholly comprehensive, extensional) programming language
which is, simultaneously, an inverted scheme of all the established computer architectures
(…)”[1], with this meaning a common ever-evolving Assembly Programming Language, giving
rise to a semantic explosion of programming languages, all throughout what can be described as
an inversion of the so called von Neumann or Princeton CA in network cybernetic fashion.
Protected as it may be by a provisional Patent Application at INPI (Portuguese Institute of
Industrial Property) (nº 116408), a fairly elaborated disclosure can be eloquent enough as to
describe its most basic settings. Although fundamentally expanded in technical computational
terms, it should always be attained that one such implementational, informatic and informational
method [U-Mentalism and the “C” approach in computation] is inextricable from a metaphysical
naturalistic method [U-Mentalism and the “O” approach in ontology]. In addition to this, it is also
to remember that the shades of relativistic and possibly technical contentious matter are all
related to the perdurable problem of the context and contingent matter of technology’s state-of-
the-art, yet never to the very core of the new utility general-purpose application or, better said,
the invention’s original idea. Lastly, it is worth mentioning that the technical drawing of the CA
herein disposed can also be found in the following divulging website: www.u-mentalism.com .
2. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
2
1.1. The Imagetic Frame of Reference of U-Mentalism in Relation to the
“O(Ntological)” and “C(Omputational)” Approaches
We shall begin by stating that the “C” method is to be performed by UTM(s) with controlled,
orthogonal and synchronous, camera-like digital images RGB sensing/processing and computing
binary isomorphic processors, thus with (typical) digital images “scanner” (impression) and
“printer” (emission) abilities. Universal Turing Machines are described in .6 “The universal
computing machine” of Alan Turing´s 1936 seminal paper On Computable Numbers, with an
Application to the Entscheidungsproblem [2]. In this paragraph the machine U is supplied with a
tape where is ahead written the S. D. (standard description) for any other machine M. In such
manner, U will compute exactly the same sequence as the machine M. The novelty herein in U-
Mentalism is that the U-machine or UTM(s) are camera image sensing (impression) and
processing (emission) RGB/binary isomorphic processing computers with likewise symbols and
m-configurations, and because UTM(s) are also M(s), every UTM is also able to compute any
other UTM(s)’ computable sequence, and in such a way that any such computable sequence in a
network of information, or the Internet, is equally likely to be computed. In UTM(s) the graphical
camera display is the forefront processor, and RGB, although synchronous to the binary code, is
the primeval symbolic feedback, in exact opposition to classical computation in M(s), wherein
control and communication is wholly set on the binary code.
Now, in U-Mentalism under the “O” approach, what is relevant is the constitution of the
physicalist most differentiated quanta of spacetime and observables, each of which to bear all
possible viewpoint “images” for every possible and most differentiated quanta of spacetime and
observables, wherein the latter observables are themselves included, as well as every all other,
thus conceivably measured, in every possible viewpoint “images” in spacetime, infinitely and
recursively. All in all, one such cogitation is produced by a pure imagetic frame of reference of
spacetime with bijective transformations of the state spacetime common to the different
observers, and wherein the proper metric/imagetic/recursive and observable state spacetime is the
observational reference frame of spacetime itself. All in all, in an analogy argument it goes as if
spacetime, not affected by the indeterminacy principle (Heisenberg) and in full entanglement,
could infinitely observe itself with the frame of reference being any constituted chosen metric
and noematic image. If, for example, we could, non-contradictorily, in inter-noumenal or inter-
monadic fashion, observe every viewpoint of every photon for the most differentiated quanta of
spacetime observables, recursively in a standard description (S.D.) frame of reference of imagetic
nature, herein presumed the pure imagetic viewpoint of photons, we are more inclined to
understand not only the observables incompleteness (roughly expanding Gödel´s theorems from
logic to “O”) and undefinability (roughly expanding Tarski´s theorem from logic to “O”), but
also, more conveniently, the consistent and effective passible intermediate states if the set of
boundaries or limits are much less forceful. One such case is, most definitely, the “(typical)
digital image”, a conveniently neutral physicalist viewpoint, further permitting the classes of
computable expressions and functions, such as the binary code at its core, to be mapped onto
images, as expressions of the RGB codomain. We choose the S.D. frame of reference to be,
according to the state-of-the-art technology, the 8K (≈8000 Pixels) 60 Frames per Second (FPS),
24 bits in depth images, choosing pixels-per-inch (PPI) as the standard resolution pattern, in
which case we are exhibiting a basic setting for U-Mentalism under the “C” approach. In this
wise, although far away from the philosophical crux of U-Mentalism “O” - “every possible image
in every possible spacetime composed in every possible mind and n-dimensionally by perceived
photons of light” [1] -we are resolutely bridging the chiasm by means of the presentation of a
simple object, i.e., an U.
3. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
3
1.2. U-Mentalism as a Method of (Typical) Digital Images Non-Standard Positional
Numeration base with Isomorphic Many-Bijective Modules Recursive Powers
Considering the general analogy between the “byte” and the “pixel”, with 28
(0-255) (256
tonal/chromatic values per each byte in the pixel with 3 pixels), it is known that a single RGB
pixel holds, according to the formula 2(8x3
), i.e., 16.777.216 tonal values or colors, which is
basically the same number of bytes overall combination. This value corresponds to 224
in prime
factorization, which is a measure very appurtenant due to the imperative of cryptography in the
system. The provisory value of the pixel (224
color/bytes combinations) is now our minimal
symbolic unit, in correlation with the standard binary code, also an (observable noematic)
wavelength impression. Now, the density of the pixel equation in agreement with the number of
total pixels provides the image resolution in PPI=
dp
di
[diagonal resolution in pixels=(dp);diagonal
size in inches = (di])]after the diagonal pixel resolution found through the use of the Pythagorean
theorem:
dp = √wp
2
+ hp
2
[diagonal resolution in pixels=(dp); width resolution in pixels =(wp);height resolution in
pixels=(hp)]. Needless to say, we are envisaging any possible variations of measures in the
overall structural and functional method.
Also relevant, both physically and symbolically, is the fact that the system in UTM(s) has
invariance by synchronicity in all positively-defined and non-accelerating frames of reference
(herein “Frames Per Second”= FPS) of the “(typical) digital image” in the system, and likewise
the speed of light in the vacuum is the invariant non plus ultra limit of the technology. Therefore,
under one such assertion, an 8K Ultra Full HD (7680*4320) has 33.177.600 pixels disposed (in a
16:9 ratio, i.e., 24
: 32
in prime factorization). In other terms, this means that each 8K RGB digital
color (FPS) image has 796.262.400 bits, or 99.532.800 bytes.
As we are referring to an RGB/binary synthesis within an isomorphic and bijective model, we are
most surely asserting a presumed less-to-the-furthest well-ordering recollection of typical digital
images, arriving either by general image sensing or general image processing to the UTM(s),
from a pool of very different kinds of typical digital images most generally found on the Internet:
photos, URL(s) and Web pages, all Turing-machine Frames (FPS) including e-mails and instant
messaging, kernel and system logs, OS environment FPS “films”, digital TV FPS “films”, and
every other sort of Turing-machines graphical interface FPS like outdoors and consoles, ATM(s)
and GPS(s), CCTV, camera drones, mobiles and tablets, ubiquitous computing things, etc.
Before anything else, the well-ordering of the RGB/binary graphical/digital coeval isomorphic
code should be preliminarily understood. Accordingly, below is shown a table of partial well-
ordering (16; 0-15) in a positional numeration base, with inherent many-bijective modules
recursive powers. The table is the correspondent to
256
16
bytes or colors, i.e.,
1
16
of the whole
symbolic power of one pixel only under the standard description of the 8K model which holds
33.177.600 pixels. It is to be noticed in the table below that the symbolic manipulation under this
pixelized part is, hence, only affecting the Blue Byte (in truth, rounded off to even numbers,
roughly only
7680∗3
16
i.e.,
1
1440
bytes parts of the total in one width or horizontal pixelized line
with 7680 pixels, in turn intersected with 4320 pixels in height or vertical lines, in a 16:9 ratio,
4. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
4
which sums up a total for each FPS, or “(typical) digital image”, of ≈33.000.000 pixels and,
therefore, of ≈99.000.000 bytes.
This is so due to the patterns of the RGB code industry convention, wherein WHITE is
255*65536+255*256+255 = #FFFFFF, and so, by order, RED is 255*65536+0*256+0 =
#FF0000, GREEN is 0*65536+255*256+0 = #00FF00, and BLUE is 0*65536+0*256+255 =
#0000FF. Coherently, we are generally pointing out to the IEC 61966-2-1:1999 sRGB (standard
Red Green Blue) color space associated (and all complementary norms, such as ITU-R BT-601
and ITU-R BT.709 for HDTV), which is, by itself, congruent and abridges other color models,
such as the inverse and subtractive CMY, and beyond the human-vision perceiving models
“Hexcone” HSV and “Double Hexone” HSL, or really any CIELAB color space, along with Y Cb
Cr and Y C’b C’r defined in terms of one luma component (Y′), and two chrominance components,
subsuming, in order, somehow overlapping color image pipeline U (blue projection) and V (red
projection), hence Y’UV. However, any other suitable well-ordering complies and falls as
predicted and logically accommodable in the technology.
Here we open a short parenthesis, for an assessment of future technologies. Under the exploratory
study of optical or photonic (projecting optical quantum systems) computing power, where U-
Mentalism technology and CA ought to be most perfectible, due to the inherent “cell”-like unit
packets of spectrum information, replicable of the RGB “cell”-like “pixel” in the state-of-the-art
technology, as long as it is made possible for pre-existing electro-optics devices processing and
interconnection, optical interconnection networks, and to a greater degree, U-Mentalism optical
computing system and architecture, to accurately and locally measure 256 different spectrum
wavelengths and frequency per each RGB color, thus measuring a total of 2563
= 16.777.216
possible different colors per “frame”, an unprecedent in depth new avenue for electro-optics or
all-photonic (projecting optical quantum) super computation, newly diverged from binary data, is
unlatched.
More precisely, under the appointed figures, it should be said that the U-Mentalism optical
computing system and architecture ought to be minimally implemented once the
camera/processing perceiving/observable UTM(s) vertex and focal-point (1-D) solution of
luminance - luminous flux per unit (solid, orthogonal and synchronous) angle, and unit source
area or luminous intensity per emitting area unit – thus in controlled, never-subjective but
photometric automatic brightness, and in articulation with the (2-D) vertices X, Y of
chromaticity, hence in full (3D) “film”, conceivably within a (3-D) X,Y,Z axis in the CIE (RGB)
color space, with perdurable constant and perceived/observable luminance in the UTM(s),
corresponds, in general terms, to the standard minimum requisite of realizable
spectra(illumination and imaging) various combinations, firstly considering electro-optics,
thereafter photonics (projecting optical quantum systems) or, at large, U-Mentalism optical
computing. In one such case, what arrives is the practical discarding of the binary code, i.e., the
conditions ought to be compulsory for the proper computer graphics palette (and gamut) with 256
shades for each RGB color to be, in obvious self-referential manner, a proper color graphics
symbolic index.
Likewise, human-perceiving color models, nevertheless conforming a mathematical model
wherein colors are represented as tuples of arithmetic representation, settled on neurophysiology
imperatives, the binary code too is, due to the very same neurophysiological causes, a human-
perceiving based mediation, moreover challenging enough of higher-level programming
languages symbolic design. Stressing the comparison, at a deep systems level, it could be said
that the trichromacy or three-different channels and cones of human vision, itself conditioned by
the stereoscopy two orbits or cavities binocular vision, is better representative of the symbolic
5. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
5
photo- and neuro-receptivity, in articulation with human-machine and machine-machine
communication, than it is the ON and OFF discrete states of the binary code all alone. But what is
more truthful is, instead, that one model only ought to be prevalent (from RGB/binary to single
RGB), even if both considered can, should, and ought to be concurrent until a full systems
symbolic design optimization is physically attainable.
However, a binary code table equivalent will always persist for the abstract drawing of semantic
equivalences in photonic computing, in both human-machine and machine-machine
communication, yet predicted to be, under one such scenario, relinquished, if not for anything
else, because of quantum computing emergence.
Anyhow, at this point we are on the fringe of both pure photonic computation and pure quantum
computation at symbolic systems level, insofar, if duly noted and enshrouding quantum
superposition, measurement, and entanglement, for instance, looking back to broad 8-bit and 16-
bit computer graphics ages, the minimal 3-bit RGB relative intensity palette, with 21x3
= 8 colors
in total, inherently disposes 4 color cube vertices only, which is, again in symbolic terms,
remindful of the two orthonormal basis states {|0⟩,|1⟩} of the quantum computational basis with
the physicalist entangled correspondents pairs into a cube “pixel”-like (in a Bell state) model.
Besides, the native determining triads |0⟩ = [
1
0
] and |1⟩ = [
0
1
] in Dirac´s notation for general
quantum states of two orthonormal basis states, considering the third-state of superposition which
diverges from binary logic, are emulated in the three channels Red, Green and Blue of the RGB
color model. Also relevant is, still in a pure imagetic provision, the idea that each vertex of the
cube herein under appraisal abstractly bounds a vertical and horizontal polarization under
measurement, herein represented by the edges of the cube, in a coherence state. This cogitation is
also closer to the consideration of the pure state as represented in a Bloch or Riemann sphere,
with mutually orthogonal linear algebra dependent state vectors. It could also be extended to
affirm, against this background, that each metric/imagetic/recursive/quantized “frame” would
have to be, under the U-Mentalism “O” approach, coincidentally and all together one three
pictures of quantum mechanics “frame” (meaning with this Schrödinger´s, Heisenberg´s, and
Dirac´s dynamic representations of quantum mechanics, depending on if the constant is,
respectively, the observable, both the ket state and density matrix, or none and all in interaction),
with great consequences in relation to time-dependency. This is not to be misunderstood in the
sense that quantum computation could ever, either trough a universal quantum simulator or a
quantum Turing-machine (QTM), or as it might be U-Mentalism QUTM(s), decide on the
validation of any arrow of time (from cosmological to perceptual) or any underlying physical
extra-theoretical principle, but instead and only that such M(s) would decide under multiple
equivalent and generally uncertainty quantum representational dynamic “pictures”, in one cell-
like “frame”. Maybe the only exception to this is the proper equivalent simulation of what can be
described as the collapse of time (beyond T-symmetry), and indeed of space time except its
representation, in the sense that this cogitation is equally equivalent, and thus all the well-ordered
arrows of time (from thermodynamic to quantum, from causal to cosmological) including any
UTM(s) or M(s) computable and recursive decidability, could all be perceptual. Yet, one such
perception could never be produced, as shown, by any sort of UTM(s) or M(s) simulation, neither
by any QTM(s), nor indeed by any assigned underlying physics determinacy principle, because,
at source, perceptual representation will always continue to be subject of indeterminacy, much
beyond model-theoretically time-or spacetime-dependency problems.
This is only to show that even under the exploratory study of optical or photonic (projecting
optical quantum systems) computing power, U-Mentalism technology “pixel”-like or, beyond,
“jot”- like, or even quantum “bits”-like model – “A bit is a bit. It is either 0 or it is 1, and that is
all there is to it.”[3] Chris Bernhardt rightly remarked – or, better said, quantized cell-like
6. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
6
“frame”, are definitely accommodable under the general U-Mentalism color, photon, or quanta
“cell”-like exponential-prone basis.
Now, refocusing to the table below, we shall remember that the 8K Ultra Full HD (7680 ∗ 4320)
with 33.177.600 pixels (disposed in a 16:9 ratio) with 99.532.800 bytes, the RGB convention
pattern well-orders RED, GREEN and BLUE accordingly, but in U-Mentalism what matters is
the binary module correspondent (imagetic and electro-optic operational) mirror. As we shall see,
even if U-Mentalism executes (typical) digital images in recollection, ought to fundamentally be
an emission RGB/binary technology, and not, though necessary, a binary/RGB mirror system.
For instance, all the different 256 BLUE combinations do not exactly all concur to all the
different final (perceptual) BLUE colours, but they unequivocally provide 256 (0-255) different
binary module combinations per channel, and RED, GREEN & BLUE in total provide the
different 256 × 256 × 256 = 16.777.216 possible and decidable effective modules combinations
per FPS.
The inference to the best explanation is simply that the balance and potential for escalation of the
binary system (0, 1) coupled with 796.262.400 bits, or 99.532.800 bytes, in the digital image, is
fairly inferior compared to the binary system (0, 1) coupled with the 16.777.216 RGB color/tonal
combination range in the pixel. The pixel is just the most accessible wavelength and color
isomorphic synthesis as of this stage of technological development.
Table 1. RGB 24 Bits Color Calculus, RGB Binary, Hexa, and Ordinal RGB modules
RGB 24 Bits Color
Calculus
RGB Binary Hexa Ordinal
module and
Decimal
(0*65536)+(0*256)+Blue (00000000,00000000,00000000) #000000 = 0 1st
= 0
(0*65536)+(0*256)+Blue (00000000,00000000,00000001) #000001 = 1 2nd,
=1
(0*65536)+(0*256)+Blue (00000000,00000000,00000010) #000002 = 2 3rd
= 2
(0*65536)+(0*256)+Blue (00000000,00000000,00000011) #000003 = 3 4th
=3
(0*65536)+(0*256)+Blue (00000000,00000000,00000100) #000004 = 4 5th
= 4
(0*65536)+(0*256)+Blue (00000000,00000000,00000101) #000005 = 5 6th
= 5
(0*65536)+(0*256)+Blue (00000000,00000000,00000110) #000006 = 6 7th
= 6
(0*65536)+(0*256)+Blue (00000000,00000000,00000111) #000007 = 7 8th
= 7
(0*65536)+(0*256)+Blue (00000000,00000000,00001000) #000008 = 8 9th
=8
(0*65536)+(0*256)+Blue (00000000,00000000,00001001) #000009 = 9 10th
= 9
(0*65536)+(0*256)+Blue (00000000,00000000,00001010) #00000a = 10 11th
= 10
(0*65536)+(0*256)+Blue (00000000,00000000,00001011) #00000b = 11 12th
= 11
(0*65536)+(0*256)+Blue (00000000,00000000,00001100) #00000c = 12 13th
= 12
(0*65536)+(0*256)+Blue (00000000,00000000,00001101) #00000d = 13 14th
= 13
(0*65536)+(0*256)+Blue (00000000,00000000,00001110) #00000e = 14 15th
= 14
(0*65536)+(0*256)+Blue (00000000,00000000,00001111) #00000f = 15 16th
= 15
(0*65536)+(0*256)+Blue (00000000,00000000,00010000) #000010 = 16 17th
= 16
The well-ordering herein disposed constitutes itself a major resemblance feature with the
fundamental theorem of arithmetic, but this time in graphic/digital format, and fundamentally
with inherent many-bijective recursive powers under super-computation, greatly emancipating
the products of the faculty of imagination, with much greater power of synthesis and scope.
Because it is not only analytical, but dialectic, or better said dynamic, it is not only arithmetic, but
essentially algorithmic.
7. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
7
Indeed, the unique-prime factorization theorem in arithmetic progression, as well as any chosen
unique or non-unique progression containing composites (not necessarily through 8 bits modules)
is spontaneously an algorithm of the system, moreover with canonical or non-canonical
operations and functions, either cognitive (man-machine noematic-representing) or practical
(man-machine evaluation-apt and choice-expressing), where from mathematical and
philosophical noemas and judgements are predicated in relation with UTM(s).
In what regards the positional system, enough is said if we declare that the binary radix of the
system, congruent with any other numeral system, works with the “octets-bytes”, or rather “3
octets – 3 bytes”, of the “pixel” itself as placed (RGB bytes and color) value notations, with
width and height, and ahead time-valued and inter placed-FPS combined index positions,
constituting any possible number or algorithm. Let us notice that by programming itself, non-
autonomously but spontaneously, a non-standard positional numeral system, in synthesis
(presumably also under the unity of apperception in man-machine communication) and
synchronicity with a place and time combined valued notations, is present, even if the system is a
standard positional numeral system.
At this point, with the value at hand of 99.532.800 bytes in an 8K Ultra HD digital image, this
value corresponds to the total isomorphic RGB digital image-to-binary code ready to be
processed by image sensors (impression), and also ahead processors (emission), being in this way
clear that the system has only to differentiate 28
per byte/per pixel, or 28*3
per pixel locally,
instead of the much harsher demand of the linear-dependent CPU, or non-linear CPU-dependent
GPGPU general amount of 99.532.800 bytes per FPS, in one “(typical) digital image”.
The next step is, thus, the assessment of the chosen value of 60 FPS (Frames per Second) in
conjunction with the present invention, a pretty conservative value to take into account, specially
if we consider that the INRS research team has, with the T-CUP, over passed the threshold of 10
trillion FPS, invading the femtosecond scale, i.e.,
1
1∗10−15 of one second (or quadrillionth of a
second). Now, 60 (22
*3*5 in prime factorization) FPS, as soon as it meets the second FPS or
frame, is defined as a “movement-image” or “film”, pointing out to a value of
60 (FPS) * 60 (´´) * 60(´) * 99.532.800 bytes
21.499.085.000.000 = 2.1499085e+13 bytes per hour in one film only in the technology herein
presented. The correspondent and conveniently converted value of bytes per hour in one film
only is, hence, 2.1499085e+25 Terabytes, or 21.499085*1015
Zettabyes.
It is pertinent to contend that this value is a dense discrete metric measure and, although it can be
put forward in synthesis in one film only, in programming algorithmic technical terms it might
eventually have been formed by the concurrence of many permuted and/or combined, rather than
composed, “films”, or as it might be FPS tunnels of “pixels” as wavelength
impressions/emissions all throughout every bit and at the full length of the movement-image per
hour in one film only under the technology.
Confronting anew with Alan Turing´s On Computable Numbers, with an Application to the
Entscheidungsproblem, definition – “The machine is supplied with a ‘tape’ (the analogue of
paper) running through it, and divided into sections (called ‘squares’) each capable of bearing a
‘symbol’”[2] – the shift to U-Mentalism in the “C” approach is easy to follow if we declare that
the “tape” is now film, “squares” are now (FPS) frames, and the “symbol” the graphical/digital
movement-image encompassing the necessary RGB/binary “r-th bearing of the symbol” in
network distributed in as many as possible partial computing UTM(s).
8. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
8
Attention should also be called to the fact that IDC and Seagate forecast that the global data
sphere, which was of 33 Zettabytes in 2018, will grow to 175 Zettabytes (175 ∗ 1021
bytes)by
the coming year of 2025 [4]. In other words, the figure found of 21.499085*1015
Zettabyes
for U-Mentalism one hour of one film only of processing power return from data is, on its own,
1.2285191e+14 times more than the expected global data for 2025. In point of fact, it would have
to elapse 120 years, with each year equally with 2025’s 175 Zettabytes + a 100% growth rate for
each year, so that the datasphere would approximate the return result of processing power of the
technology for one only film of one hour only thereof encapsulated. The growth rate of data is
hereof paramount, as considering if not, figures are that 17.500 years, each with equal 175
Zettabytes of data, would have to pass by to meet one hour only of U-Mentalism, with very
conservative parameters for the processing power. In view of this, and on the other way,
assuming the quadratic nature of quantum computing worst-case complexity in confront with
classical computation, if we envisage the system´s complexity complemented with quantum
computing complexity, and most specially, counting with the accelerator factor of U-Mentalism
on the production of data, the inception in years might be dramatically shorter. What is more,
supposing that the technology and CA would work on the full 175 Zettabytes of global data for
the year 2025, the super-computation involved would assume 48.611.111 Terabytes per second
(4.8611111 ∗ 1019
bytes per second), which sums up 265.39792
per second, already above the
capacity of a 64 bits architecture. In truth, in one such case, the technology is imminently
conjectured to improve above 1 Exbibyte (EiB) = 260
or 10246
.
Because the S.D. of the technology can be represented by a set forming the symmetric group of
the set, which is a bijection from S.D. to itself, and for which every placed octet-byte (binary)
element occurs exactly once as the corresponding placed (RGB) image value, S. D.nis the
symmetric group under permutation, a broad relation and it is also a function composition under
group theory. For the reason of implementation of less-to-the-furthest well-ordered recollection
(large numbers arithmetic) and further forward well-ordered collection (at large algorithmic) of
typical digital images, onward to be run by as many as possible UTM(s) in a network of UTM(s)
on the Internet, a calculus of permutations is needed and, complementarily, it is imperative to
calculate a fair assessment of U-Mentalism computational time complexity. With this procedure
we are already asseverating the inclusion of crucial demands of the system, such as the use of AI
& cryptography general image processing of typical digital images, apart from general image
sensing of typical digital images, as well as the run time complexity in relation with the data
memory involved.
1.3. U-Mentalism (Typical) Digital Images Permutations in Partial and Distributed
UTM(S) in a Network, on the Internet
Once the 8K resolution Ultra Full HD (typical) digital image (FPS) bears (7680 ∗ 4320) pixels,
which sums up 33.177.600 pixels (width * height), what follows is the application of the formula
of permutation having in mind the measurement in pixels
P(n, r)
n!
(n − r)!
Thus, n?
is, really, the number of colors per pixel, which is 2563
and the correspondent to the
combination range, which equals 16.777.216 color combinations per pixel, while r?
is, really, the
previous value of 8K Ultra Full HD (width * height), i.e., 33.177.600 pixels. It is very easy to
appreciate that the break-up of the two orders of factorials points out to unmanageable numbers,
directing both to countable infinite numbers, andO(n!)non-assessing time complexity, regardless
of the CA, recursion power or machine. One such calculus is, nevertheless, judiciously desirable
9. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
9
by cause of the intrinsically O(n!) factorial time complexity exposition of the system, dragging
O(2n) exponential time, and O(n^2) quadratic time hardness lines, as well as, in middle-way and
by order, O(n) linear and O(log n) logarithmic times- herein O(n!) grows faster as it abridges a
constant exponential base 2 -, but on the side of U-Mentalism overlapping solvability, not of
classical computation. Simply, the algorithms of U-Mentalism are yet unknown, and the infimum
complexity that solves a class of problems is of the same complexity as that of the problem. One
such assessment comes even beforehand newly fine-grained analysis, defining the possible class
of problems as the set of computational problems of related resource-based complexity, given
that time, processing, memory, and more so the relation between them is radically different in U-
Mentalism, however included in Turing-machines computability Application to the
Entscheidungsproblem [2].
On the edge, by nature of the intrinsic arithmetic system in U-Mentalism, we could even consider
each frame a large number image, reducing the composites factor in 8K of 33.177.600 pixels to
one FPS and, consequently, to each one (FPS) large number linear arithmetic progression image,
which and in turn transforms the calculus of the combination of colors per pixel per FPS, to one
onlylarge number arithmetic mirror per FPS. One such FPS large number image progression
would meet a factorial (FPS) table itself of O(n) linear time complexity, that is made exactly a
“C” imago mundi for the “movement-image” recursion in the system. It is true that the immediate
next polynomial running times (quadratic, cubic,nc
, etc) hold important classes of algorithms to
discern an unequivocal well-ordered (FPS) large numbers arithmetic progression mirrored (RGB)
images, in recollection (composition) and collection (permutations/combinations) of typical
digital images. However, U-Mentalism is not a system to solve one such linear progression FPS
problem, but instead to solve ahead any algorithm class problems newly defined by the system
itself, with inherent new space and time complexity powers, insofar one such linear FPS
progression is being expanded in the system.
It is important, under this context, to remember that decidability is based on the localist
decidability (even working with non-localist quantum computation by chance) of the pixel
isomorphic RGB image of every binary octet. Accordingly, the large number arithmetic (RGB)
image mirrors of the system have already exceedingly computing power, all throughout a system
where data and processing are positively “C” entangled: data capacity returns processing power,
and processing power returns data capacity.
All in all, and luckily, steadily paced performance up to constant verifiable factors is all that is
needed under the U-Mentalism system, in a deep and low-level performance requirement
RGB/binary enhancement only, wherein, unambiguously, arithmetic progression in FPS follows
locally each digit power of two in binary, once decidability in terms of a machine M or U is said
to be a decidable problem if there exists a corresponding M or U which halts on every input with
either 0 or 1, thus low-level feeding the FPS arithmetic progression of large number mirror RGB
images. This is, besides all, what makes it not constructive at all, and indeed counter-productive,
any glimpse whatsoever over a hypothetical solution based on (typical) digital image decreasing
measure overall pixel/bytes reconversion. An aforesaid presumptive choice of lesser resolution-
say, maintaining the 16:9 = (42
:32
) ratio, 500 ∗ 281.25 (width * height) in pixels -would naturally
decrease the computational power of the technology and, ergo, the overall scalability of the
technology in relation to the cybernetic network on the Internet, under which a minor convolution
of data and processing power altogether would impend on time complexity solvability.
What happens is exactly the contrary: the datasphere is too tiny when confronted with the power
of U-Mentalism, to the point where well-ordered recollection and collection are pivotal not only
to operability, but also to the progress of the technology.
10. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
10
More importantly, in the localist decidability of the pixel isomorphic RGB image of every binary
octet resides the fundamental criterion of difference and repetition at which underlies the XOR or
Exclusive OR argument at the root of progression of binary numbers or, indeed, mod 2 addition.
If noticed, the progression (00,01,10,11) corresponds to binary addition, after which completion
the next two bits on the left are triggered to shift by half-addition, the same is saying, the double
of the previous elements of the series of progression.
In other terms, it corresponds to a not equivalence NEQ difference and repetition operation in
binary/RGB isomorphic arithmetic progression, wherein the proper R(ed), G(reen) & B(lue) are
module operations. In this fashion, the whole (FPS) (typical) digital image becomes a truth-table,
for the reason that the lines and the columns (width * height) are themselves a sum operation,
needing only a carry-out color/bit to the left when the progression (00,01,10,11) ends.
This would equate having in the binary logic and image sensing (and processing) unit,
presumably, a two-color/bits XOR and AND adder. Naturally, because of the need of a full adder
circuit for the entire (FPS) (typical) digital image, the binary/RGB isomorphic nature of the
UTM(s) would rather, again presumably, be prone to use a two-bits/color XOR, AND and OR.
Inasmuch as, for instance, in propositional calculus, laying the foundational bedrock of logic
since Aristotle – considering the four different types of categorical propositions in the square of
opposition, withstanding the syllogism theory - there are three propositions for each place-valued
syllogism figure out of possible four. Thereupon, the possible total number of syllogism modes is
four times that number, i.e., 256 logically possible distinct types. Because 256 is the same
number of module 8 bits per RGB color in the (FPS) (typical) digital image, what this signifies is
that, for the sake of the argument and hypothetically, an (FPS) (typical) digital image is a
polysyllogism and a calculus ratiocinator M.
However, and fundamentally, in U-Mentalism the CA is inverse and, thereupon, it is not built on
binary/RGB, but instead RGB/binary. Without this judgement, it comes not to be transcendental.
It shall produce wavelength colors and forms synthesis, just like φύσις (nature).
Due to the localist nature of both the pixel and of the whole (FPS) (typical) digital image, the
RGB/binary isomorphic nature processed in the UTM(s) will be prone to use, not quite an
equivalent color summands adder, but instead an every n-ary color/bits RGB imagetic relations
instant mirror, filter and mixer, always remembering that in between different modules and (FPS)
films the exactly same holds true. What this means is that U-Mentalism is, at each UTM
processing, constant metric localist, either in a pixel, a module, or the entire FPS, with equal
time-dependent “film” computing power on the previous synchronous and orthogonal base for
each (typical) digital image, n-ary modules or bytes.
In truth, the RGB/binary relation in the CA is always affected by a special bottleneck related with
the communication with classic computation, as far as other much less grievous than the von
Neumann bottleneck, which basically corresponds to the arithmetic logic unit binary mirroring of
the RGB image codomain in case only of U-Mentalism scanner-to-printer or Eye-to-Brain, but
except for machine learning, not in the case of U-Mentalism printer-to-scanner or Brain-to-Eye.
If any bottleneck in the system exists that is worth mentioning it is, inevitably, what we choose to
designate the unfolding “O” and “C” philosophy of (time) history bottleneck. The reason behind
so is that computational means and resources up to this point of the “C” state-of-the-art do not
produce a reasonable amount of data as to test match the system, which is test halting (time)
history itself. Before having a chance of escalation from one hour of one only film in the
technology to a film of several hours, years, and even synthesis of the image metric distance, in
11. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
11
the “C” movement-image, in light-seconds, (time) history “O” has, simply, to elapse. In contrast,
as seen before, [“C”] computational time complexity in U-Mentalism has equally to elapse,
although much more tied with bondless mathematical and dynamical limits, precisely on the
grounds of the [“O”]constitutive transcendence on [“C”].
It is, ultimately, by virtue of this assertion that it is more appropriate to make mention of a
general U-Mentalism “C”-“O” bottleneck, which is, by and on itself, a rectification of the
classical von Neumann bottleneck. To alleviate any remaining doubts, it should be elucidated that
the system is an inversion of the von Neumann CA, not only because of the RGB graphical
primeval symbolic precedence, but also, amongst other aspects, of the inherent entanglement of
data and processing power, which drives high latency, on a U-Mentalism turn, from being orderly
unavoidable to well-orderly avoidable.
U-Mentalism, in this fashion, especially due to the diagonal method of computability beforehand,
and of well-ordering collection also on the basis of the diagonal method of cryptography and
permutation/combination, has a nature of transfinite (FPS) typical digital images, denumerable
“image-movement” sets, with inherent cardinals and ordinals equipollence. For the moment it
suffices to say that the nature of computable and definable numbers, as for the rest composite or
prime numbers in U-Mentalism as “3 octets – 3 bytes” modules, typical digital images, or “films”
is, intrinsically, a bijection of the well-ordered set of all finite ordinals in the system w0 to
cardinality א0
. Therefore, an algorithm for an well-ordered collection of typical digital images in
the system could easily resort to a typical diagonal on the binary basis and exceptionally
reductionist.
Inquisitively, one can picture also, in an Imitation Game [5] register, or in a Turing Test flair, a
different dialogical test. We shall call it for now both the U-Mentalist “O” inquiry and the “C”
test. On the grounds that any (human or machine) synthesis of electromagnetic wave-like
physical and symbolical impressions, susceptible of being, in turn, emitted in any body or
technology, are to be, in quantized electromagnetic wave-like impressions in continuous
spacetime, indiscernible in nature, the following questions arise:
Under U-Mentalism “O”, we inquire if it is possible to be an observer without photons or any
observable frame of reference in spacetime.
Under U-Mentalism “C” and likewise, we shall test if ever the prior impressed and, thus, emitted
UTM(s) wave-like synthesis or images, in spite of the foreign face-to-face relation within a body
of a presence in front of it, can be made discernible from the frame of reference of the observer.
12. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
12
Figure 1. U-Mentalism Computer Architecture Design
1.4. U-Mentalism UTM(S) Networkanalogy with Current Data Nº1 TOP500
Linpack
Super-computation is often measured or estimated according to the floating-point (FLOPS)
(additions and multiplications) computing power in the M, under the numeric and scientific
Fortran-derived linear equations in the LINPACK Benchmarks, taken as 64-bit floating-point
peak performance. Besides being quite a multi-dimensional problem to address, and essentially a
non-distributed supercomputer system ranking, it is the most reliable source for high-
performance computing so to have a basis of comparison with the U-Mentalism CA, either
distributed or non-distributed, in or out of a network, on or outside of the Internet. This is
important to refer, once the CA is developed to be on distributed massive parallel computation
with the most possible advanced cryptography methods on each UTM server/node (like open-
source blockchain), in a network, on the Internet. Any contrary application of the CA is
envisaged as jeopardizing and, indeed, potentially very harming if ill-fated. It shall, hence, be laid
open – patere (wherefrom the word patent derives) – for both public inspection and public
policies, laws and interests of governments and the people. Once these aspects bring in multi-
variables of which values is difficult to know, although being much easier to acknowledge
approximations, and most specially orientated-guise measures of central tendency, normal
distributions probability, and deterministic ranges, our method shall simply attain to the plausible
measure for the intensity of the required memory per unit of performance, along the standard of
FLOPS and bytes per FLOP (B/F).
Previously, though, in what has to do with the blockchain open-source technology, the
idealdesideratus of the technology is to have a (typical) digital modules, “frames” and “films”
13. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
13
(beyond hypertext) U-Mentalism “hypermedia” [6] world wide web service on the Internet itself,
wholly constituted as the proper inter-UTM(s) U-Mentalism nodes/servers (miners/theorem
solvers/record-keeping services) communication network of networks. In this fashion, U-
Mentalism “C” shall be developed to set off profoundly affecting social, cultural and economic
changes, such as social transparency and political trust, law and legal abiding contracts,
consensus breaking and activation, counter action to phenomena such as tax evasion and off-
shores, laundering money and corruption, defense of copyright claims, public registry, thereafter
with economic and financial crypto-currencies full securitization.
Blockchain, as follows, shall not to be implemented as an external feat technology, but instead its
decentralized, open and distributed ledger with recording of (typical) digital images modules,
“frames” and “films” hypermedia ought to account for the UTM(s) imagetic many-bijective
isomorphic powers of combinations/permutations in the system, on the Internet, in line with
privacy and data protection.
Comprehensively, cryptography in blockchain on the U-Mentalism system, herein understood
both as asymmetric-key algorithms and hash functions, and in perspective utter build out of
technical knowledge such as testing indistinguishability obfuscation i𝒪 on the U-Mentalism
system, are to beinferred as (typical) digital images RGB-binaryisomorphic modules, “frames”
and “films” plain-hypermedia, cipher-hypermedia, hypermedia authentication signatures, and,
consistently, imagetic and cinematic hypermedia public and private keys, inasmuch as
hypermedia sub-exponential modules in the case of indistinguishability obfuscationi𝒪, in the
pursuit of an universal black-box (machine-to-machine communication in the U-Mentalism
system).
However all the details and possible conjectures cannot be expanded in this paper, in what relates
to the coupling of the system bedrock of U-Mentalism with UTM(s) block(s) explorer services,
and all derived “C”-“O” exchanges in a new semantic web design, some philosophical and
technical appraisals are likely to be carried out. Along these lines, it is important to express that
the pairing of asymptotic 𝒪 in U-Mentalism on the Internet, meaning the order as the growth of
rate of a function, being commensurable with the basic well-ordering (arithmetic) recollection of
(typical) digital images non-standard base with isomorphic many-bijective modules recursive
powers, and also the well-ordered (algorithmic) collection of n-modules, -“frames” and –“films”,
besides the 3D cinematic diagonal method of composition/permutation on a network, all in all in
hypermedia design, offer, in general, distinctive cybernetic traits in what regards the relation of
cryptography with the system. Sharply, it is preferable to say that the aforementioned is true in
relation to encoding, hashing, cryptography and obfuscation, but we choose to address in succinct
way to cryptography, upon conclusion that, nonetheless possible different goals are sought after,
cryptography alone recapitulates the tripod podium of confidentiality, integrity, and authenticity.
Some relevant traits to be pointed out are the following, all of them radically different mainly due
to the excessive computing power and processing speed of the technology on one hand, and on
the other hand to the all-in-one nature of U-Mentalism on the Internet:
The compactness of the system disallows the typical range of attacks – from the byzantine
general problem to the 51 percent attack- and trivializes them, as there is not, in essence, any
other possible existing network capable of emulation, and subsequently there is not any rivaling
computational power. It is designed never to subside and always evolve insofar photonic and
quantum computation are scalable and expandable from the system, with QUTM(s) in U-
Mentalism (not binary, neither linear, but instead “frame”-imagetic and “film”-cinematic), on this
wise, being the ground zero for the uttermost radical new shift. Besides, the blockchain
technology shall also be made coetaneous with (both arithmetic and algorithmic) well-ordering in
14. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
14
the U-Mentalism system, in one such way that the proof-of-work and broadcasting of previous
hashes and nonces, along with the distributed timestamp on a peer-to-peer basis, are compelled to
be included in registrable and traceable meta-data on hypermedia basis. The compactness of the
system itself is better understood if it is declared that, because an UTM is also a Turing-Machine,
or as it might be a QUTM, for that reason UTM(s) on M(s), UTM(s) on UTM(s), or even perhaps
QUTM(s) on QUTM(s), are themselves permuted, meaning that computable diagonalization
(from computability proper to cryptography) is fully performed on massive parallel super-
computation hypermedia basis. In essence, because of entanglement of data and processing
power in the multi-dimensional U-Mentalism system, UTM(s) hypermedia are scalable to proper
hyper-physical means, as in the case of QUTM(s), but also DNA molecules in DNA computing,
or even Brain-Computer Interfaces. Chiefly, the multi-dimensional composition/permutation,
beyond 3D imaging of “films” in the network, allows cryptography methods also to grow
exponentially cryptoanalysis hypermedia hardness, on the grounds of multiplied number of
imaging and cinematic coordinates, such as under parameter-, configuration- or high-dimensional
computationally recognizable spaces.
By virtue of the great excessive computing power in the system, the exponential growth of
cryptography methods can be surveyed, and beyond, randomized “film”-modules exponentially
augmented, all the more on the basis of hardness decreasing techniques, such as, for example, the
use of optical computing aberration methods, or programming methods from computer vision and
multiple view geometry algorithms, now in hypermedia super-computation. As far as it is
correlated, LED-based advantages are herein indissociable, once energy and color efficiency, on-
off cycling and dimming, solid-components resistance and focus are contemplated, if no
prejudice comes from color rendition or area light source.
The appointed distinctive all-in-one trait of the U-Mentalism system in a network of networks, on
the Internet, is of paramount importance, beyond the orthodox bounds of cryptography and crypto
analysis understanding, the reason being that the excessive computing power of the technology
can possibly even impend a verging leverage and balance of the classical concepts. As an
exemplification, the case in point of obfuscation – “to compile programs into unintelligible ones
while preserving functionality“ [7] from well-founded assumptions – can, in an illustrative
scheme in U-Mentalism, be instead the shifting to deceivingly exponential manifold of
unintelligibly malfunctioning programs, as well as epistemic stances in a belief network. Some
other relevant ideas are, under the perspective of an entrusting and empowered philosophy of the
body, the association of asymmetric cryptography and DNA portrait private-key with cultural,
social and political value-exchanges. Again, the RGB/binary isomorphic encoding, matched with
the excessive computing power of the technology, opens the door to other conceivable
cryptography methods, on top of well-known methods, such as onion routing or secure multiparty
computation, most of them being in close relation with the multi-dimensional and cinematic
nature of “films”, or even cold secret cipher-hypermedia “frames” or “films”, but principally to
the all-in-one nature of the U-Mentalism technology in the network of networks, i.e., the Internet,
which is turned impossible to override.
Therefore, as much as we can be conclusive in relation to the proper informatic cybernetic
architecture of U-Mentalism in relation to cryptography and cryptographic techniques, ahead of
dealing directly with the chosen super-computation M(s) benchmark suite nº1 supercomputer, is
that the proper reach of hypermedia and cinematic RGB/binary encoding run by UTM(s) on M(s)
on the Internet, drastically alters the medium and range of classical limits, mostly because of the
substitution of proofs by targets, with short-lived approximative limits to such computational
processing power and system ipseity newly investigated bounds.
15. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
15
Currently, the nº1 position of the 55th
TOP 500 following the LINPACK benchmarks suite (June
2020) is the Fugaku petascale (1015
floating-point operations per second = 1 petaFLOPS, i.e., a
thousand million millions 64 bits operations per second) supercomputer. The Fugaku holds 415
petaFLOPS with a 158,976 (two types of) nodes Fujitsu TofuD, 6D mesh/torus Interconnect, in a
A64FX CPU (48+4 core) per node CA, with a second-generation High Bandwidth Memory
(HBM2) of 32 GiB/node.
Our next step is, thus, by multiplying Fugaku´s instance of cores * nodes (52*158,976
=8.266.752) find the equivalent processing power of the very same number of
UTM(s)cameras/computer processing nodes/servers and later, having in mind that the Internet
has around 50.000.000.000 nodes, well above the 1012
to 1024
FLOPS of all the existent
computers (2015), at the end reasonably cut the latter figure by reason of factors such as entrance
in the industry, price and energy, besides any hindering variables, thus obtaining a fair value for
the CA implementation in the network of networks, i.e., on the Internet.
First off, the value of 21.499085 ∗ 1015
Zettabytes of one only “film” of one hour in U-
Mentalism divided by the number of cores * nodes (52*158,976 = 8.266.752), equals
2.6006689e + 15 Zettabyes, which converted is 2.6006689e+15 Petabytes. Therefore, assuming
a 64 bits operation, we have roughly 325083615.064 PetaFLOPS per each server/node out of
8.266.752 in the technology, if U-Mentalism was to have the same number of nodes/servers on
the Internet, presumably settled for a start on an Internet with an even much bigger number of
nodes. And even if there would never be the same number of servers as nodes, in reality and at
present the Internet detains around 50.000.000.000 nodes.
In abstract, the value divided by the computer performance of the Fugaku supercomputer
indicates that one only hour of one only “film” in the technology would equate to
783334.012203 times the nº1 position of the TOP 500 Linpack benchmark as of October 2020.
In reverse manner, we could affirm that the 415 overall petaFLOPS of the Fugaku
supercomputer, compared with the 597196805556 Zettabytes/per second (or 597196.805556
Petabytes/per second) in U-Mentalism, stipulates that a presumed value in 64 bits
(74649.6006944) petaFLOPS/per second in U-Mentalism is, in itself, 179.87855589 times
more than the overall performance of the Sugaku supercomputer.
The extreme low latency of the invention is best shown if we divide the overall abstract
performance of U-Mentalism of petaFLOPS/per hour (2.6006689e+15) by the actual number of
Internet nodes (50.000.000.000). The result is 52013.378 petaFLOPS per each one out of
50.000.000.000 Internet nodes, when indeed, chances are not only that the divisor will be much
larger, but essentially that the number of typical digital images reaching the system will be
exponentially wider. Confronting with the Sugaku supercomputer, this would mean that at each
one of these 50.000.000.000 nodes, it would be instantiated 125.333440964 times more the
performance peak of the Fugaku supercomputer in PetaFLOPS.
Nevertheless, because the hindering variables are numerous and immense, even though we are
experimenting values with greatly sub-optimized inherent values (pixel resolution, FPS, Hz, and
subordinate processing-time of the technology to one-hour only), we shall now, thinking ahead
the barriers to the entrance in the industry in terms of price, energy, etc., cut the preliminary
values to around 20%. Thus, the result at hand is, under one such 20% cut under the very same
parameters, of 65016.7225 PetaFLOPS for each one out of 50.000.000.000 Internet nodes, which
equates to 100.266752771 times more the performance peak of the Fugaku supercomputer,
measured in PetaFLOPS for the very same value of each one out of the 50.000.000.000 Internet
nodes.
16. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
16
One such technology shall be exclusively scientifically-driven. In fact, in terms of the stored-
program concept, we can designate it, in indifferentiable manner, (φύσις) science, if granted that
the overall feedback and cybernetic loop mechanism, that we choose to call an “algorithmitron”,
i.e., an accelerating mechanism for all classes of algorithms, and thus a procedure on its own, is
itself well-ordered within the extensional and intensional self-image of man and the cosmos that
is (φύσις) nature. The extensional and intensional philosophical synthesis and programing
regulative idea shall be explored in recurrence and recursively.
Very concretely, the actual example provided by the current use of the nº1 TOP 500 Linpack
benchmark Fugaku Japanese supercomputer, which has presently been used for COVID-19
research, and the nº2 TOP 500 North-American supercomputer Summit, whose current work with
scientific impact is on various levels (deep learning for human systems biology, plasma fusion
simulation, combustion in turbulent environments, stellar astrophysics nuclear burning, cancer
treatment and surveillance planning, high-temperature superconductors) are nothing but just a
pale coup d’oeil of what can be, at a greater extent, achieved with the forthcoming fabrication. By
all means, U-Mentalism participative and all-engaging cultural-scientific accelerator, social and
technological, financial and political, inter-dependability and transparency, shall act as new
measures for the human. We have to remember, for that purpose, that a world with a 24 hours
“film” in the technology, however 1.2285191e + 14 times the value of 175 Zettabytes
(estimation of global data for 2025) with the necessary equivalent data input, thus dependent on
the U-Mentalism “C”-“O” bottleneck, would grant the CA with a processing power of
5.1597804e + 17 Zettabyes (21.499085 ∗ 24 ∗ 1015
). The same is valid for “films” of several
years, and even of the “image”-distance in light-seconds, most assuredly in prospective proper
physicalist-reductionist underpinnings, rather than merely technological.
Yet, more frequently than not in dialectic terms, time and history itself naturally supersede and
are transcendent in relation to any object (such as technology). In fact, the author suggests that
the contemporary crisis in philosophy of science (φύσις) points out to the extreme of that reality.
If any peak in civilization was possible to be found, it definitely was the period from the birth of
Classical Greece up to the end of the Hellenistic period, from classical Athens to the Hellenistic
Alexandria. It worked as a natural philosophy (φύσις) explosion or radiation, all throughout two
millennia with tergiversated and fragile-weaved, often sinuous paths on the edge of eradication,
as epiphenomenon’s echoes throughout the Roman and Byzantine, Islamic and Indian Empires,
before the turning from Medieval times to the European Renaissance, and thereupon Modernity
and Contemporaneity, that was afterwards, arguably, to see its dimmest resonate and last hour in
XXth
century Vienna. This serves to explain the non-mutual relation between progress and
technology. More so, sometimes the value of a technology is best evaluated if tested against the
worst demeanors and actions known to the history of civilizations. For example, in substitution of
a colossal computing and processing power, the technology of U-Mentalism could benefit more
from improvements in cryptography, or primarily human-agent decisions.
Coming to think of it, and bearing in mind imaging and sensing technology latent in U-
Mentalism - metal–oxide–semiconductor (MOS) based charged-coupled devices (CCD), and
active-pixel sensors (CMOS) in the present state-of-the-art, prior to the reductionist more general
“cell”-like synthesis–we can think of two unexpected, but conceivable and tenable breakthroughs
cognate with the technology that are worth to be referred.
One is, definitely, the use of biopolymers as paper for the use of electronic applications, namely
paper transistors recurring to metal oxide semiconductor (MOS), complementary (CMOS)
circuits, and eventually transparent conductive oxides, i.e., paper-based electronics or
papertronics [8,9]. Most importantly in the case of a simple and universal device architecture in
correlation with the novel U-Mentalism CA, it could literally be possible having always and ever
17. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
17
a paper copy of every book in the world in the same paper organic substrate, also electronic
component (dielectric), and charge storage media, an upturn revival of the inceptive idea of the
Great Library of Alexandria and Mouseion since Ptolemy Soter I, center of Hellenistic
civilization and epitome of Classical Greece, where the study of natural philosophy (φύσις) found
its ἀκμή (acme).
The other conceivable breakthrough is directly correlated with the possible use of transparent
oxide electronics as a backbone to U-Mentalism Assembly Language programming. Because in
U-Mentalism there will be the need to instruct in symbolic RGB/bytes machine code modules,
“frames” and “films” through instances of time in tunnels of “pixels” or, in fact, any other
instances of “cell”-like exponential-prone alike basis, having access to novel semi-conductor
amorphous oxides or applications with high transparency and electrical conductivity, can open
the gates to create an endless array of philosophical and programming short-cuts over typical
digital images. In reality, beyond Thin Film Transistors (“active matrix” TFT), Liquid Crystal
Display (“passive matrix” LCD), and Organic Light Emitting Diode (OLED), the transparent
semiconducting oxides (TSOs) and transparent conducting oxides (TCOs) can help the
technology to directly assemble the building block-structure-luminous response mechanism itself,
bridging optoelectronics with programming, and possibly breaching into Photo-Voltaic modules
(PV or solar panels) or electronics Organic Solar Cells (OSC), including polymer solar cells.
Lastly, the author would like to reiterate the U-Mentalism “O”-to-”C” cybernetic analogy with
photosynthesis, the very definition of “synthesis of light”, already expanded in a preliminary
paper, in all likelihood with improved understanding as of now close to the conclusion:
U-Mentalism is mainly intended to be a programming synthesis of light through typical (digital)
images, organized as symbolic-informational truth-equivalent programming language abstracts.
Photosynthesis puts together a synthesis of light, carbon dioxide and water into glucose at
reaction centre proteins with chlorophyll (digital images), wherein to the fore roots have absorbed
water (computability) from the soil, through the stem (programming language abstracts and
paradigms) and through the leaves (programming languages). This is why to the exact
chlorophyll complementary light (diagonalization) absorbance centre chloroplast organelle
(pixel) there is, at large, a leaf lamina (frame), as a surface area to capture the light, under light’s
every possible and each necessary time-image. There is, in the overall process of photosynthesis,
a light-dependent cycle and a light-independent cycle. In the light-dependent or light cycle
(scanner-kinescope), as an effect, short-term stores of energy are produced, enabling their transfer
to drive other reactions (computer vision & multiple-view geometry; U-Mentalism Recollection),
while in the light-independent cycle (printer-iconoscope; U-Mentalism Collection), the so called
Calvin cycle, the atmospheric carbon dioxide is incorporated into organic carbon compounds (U-
Mentalism Assembly Language Programming), and dependent on the previous light-dependent
reactions (semantic isomorphic correspondence), are then used to form further carbohydrates,
such as glucose, the most important source of energy metabolism in bioenergetics (cybernetics)
[1].
2. CONCLUSIONS
In the present study, following closely the theoretical and practical keystone of the provisional
Patent Application at INPI (Portuguese Institute of Industrial Property) (nº 116408), designated
as U-Mentalism, a method to perform computation at or near the speed of light, resorting to
“(typical) digital image” RGB-to-binary in singular or multiple nodes/servers in a network, on the
Internet, in its entirety a philosophically-meaningful new computer architecture, is displayed its
simplest baseline, adjustable for the research and industry communities. Foremost, the proper
discrepancy between the imagetic frame of reference of U-Mentalism in relation to the
18. International Journal of Web & Semantic Technology (IJWesT) Vol.12, No.1, January 2021
18
“O(ntological)” and “C(omputational)” approaches is elucidated. No substitute of the latter can
prepare ahead the in-depth comprehension of the intrinsic method of typical digital images
coincident with non-standard positional numeration base with isomorphic many-bijective
modules recursive powers. Ensuing, typical digital images permutations in partial and distributed
UTM(s) in a network, on the Internet, is shown to be the proper context for the technology to be
undertaken, which suits the passage to a vaguely prosaic, but matter-of-fact indisputable,
comparison of the fundamentally futuristic trait of the invention with the current data nº1
TOP500 Linpack supercomputer as of November 2020, the Fugaku supercomputer at the RIKEN
Center for Computational Science in Kobe, Japan.
ACKNOWLEDGEMENTS
This work is funded by national funds through FCT – Fundação para a Ciência e a Tecnologia,
I.P., in the frame of the project FCT/UIDB/00678/2020
REFERENCES
[1] Homem, Luís, (2019) “What is U-Mentalism?”, Journal of Advances in Computer Networks, Vol. 7,
No. 1, pp.18-24.
[2] Turing, Alan M., (1937) “On Computable Numbers, with an application to the
Entscheidungsproblem”, Proceedings of the London Mathematical Society, 2, 42 (1), pp. 230–65.
[3] Bernhardt, Chris, (2019) Quantum Computing for Everyone, The MIT Press.
[4] Reinsel, David & Gantz, John & Rydning, John, Data Age 2025,” The Digitalization of the World,
From Edge to Core”, An IDC White Paper - #US44413318, Sponsored by Seagate pp. 1-24.
[5] Turing, Alan M., (1950) “Computing Machinery and Intelligence”, Mind, LIX (236) pp. 433-460.
[6] Tim Berners-Lee (2000) Weaving the Web, The Original Design and Ultimate Destiny of the World
Wide Web, HarperCollins.
[7] Aayush Jain & Huija Lin & Amit Sahait (2020) “Indistinguishability Obfuscation from Well-
Founded Assumptions”, UCLA Center for Encrypted Functionalities, and NTT Research.
[8] Barquinha, Pedro & Martins, Rodrigo & Pereira, Luis & Fortunato, Elvira, (2012) “Transparent
Oxide Electronics, from Materials to Devices” Wiley, a John Wiley & Sons, Ltd., Publication
[9] Martins, Rodrigo & Gaspar, Diana & Mendes, Manuel J. & Pereira, Luis & Martins, Jorge &
Bahubalindruni, Pydi & Barquinha, Pedro & Fortunato, Elvira (2018), “Papertronics: Multigate paper
transístor for multifunction applications”, Applied Materials Today 12 (2018) pp.402-414
AUTHOR
Luís Homem (Lisbon, 21/12/78) has a degree in philosophy (2005) and a Ma degree in
natural and environmental philosophy (2008) at the University of Lisbon, having also
completed a Ma degree in logic and philosophy of science at the University of
Salamanca with the thesis “Topics in Programming Languages, a Philosophical
Analysis through the case of Prolog” (2012). He has also a PhD in logic and philosophy
of science at the University of Salamanca with the same title thesis “Topics in
Programming Languages, a Philosophical Analysis through the case of Prolog” (2018).
Being a doctorate integrated member of the Center for Philosophy of Sciences of the University of Lisbon
(CFCUL), the author has developed research mainly in philosophy of logic, science and language.