Neural Networks Pattern Classification for Certainty in Measurement of Position and Momentum with Heisenberg Uncertainty Principle
Examining the Role of Neural Networks in Resolving Heisenberg's Uncertainty Principle
by Mrs. Punam .*, Dr. Yashpal .,
- Published in Journal of Advances and Scholarly Researches in Allied Education, E-ISSN: 2230-7540
Volume 15, Issue No. 9, Oct 2018, Pages 136 - 141 (5)
Published by: Ignited Minds Journals
ABSTRACT
Heisenberg uncertainly principle can be discussed by using the pattern classification technique of Artificial Neutral Networks with Back propagation learning rue. In this process a suitable arrangement can be constructed that contains the control of two experiments, one of which is designed to measure the position and other one is designed to measure the momentum of a sub-atomic particle (the electron). The Control system will determine either the position of momentum of electron, which less uncertain for any wavelength of light at any instant and start measuring the less uncertain quantity with the corresponding experiment.
KEYWORD
Heisenberg uncertainty principle, pattern classification, neural networks, back propagation learning rule, position, momentum, measurement, sub-atomic particle, electron, control system
INTRODUCTION
Can a machine take some decision and set the environment or some experiment situation according to this decision? The answer of this question is in affirmative that a lump of silicon and metal can take the decision and behave according to this decision as we human can do. Human performs all decision processing task with biological neutral network structure (Muller and Reinhardt, 1991) We can simulate this structure in a serial fashion that provides the parallelism that exists in the brain, and we call this structure as artificial neutral network (ANN). ANN consists of numerous simple processing elements. (PE) (McCulloch and Pitts, 1943), which can be globally programmed for computation. We can train the ANN for some specific application and it also learns from examples. Once the network has been trained, it is capable of performing the task with the accuracy as the human can do (Simpson, 1990). The field of ANN came into prominence mainly because of our ability to deal with natural performance in terms of process and constitutes activation dynamics and of Neutral Network is governed by neuronal dynamics that constitutes activation dynamics and synaptic dynamics (Kosko, 1991). When we train the network then this training process encode the pattern synaptic weights (Hebb, 949). In the pattern recognition task (Kanal, 1987) we given an input to the network and we can analyze that this input will find a meaningful categorization or classification. Network can be trained with the appropriate classification corresponding to input-output pattern, with any learning law (Hassoun, 1995). These learning laws may be either unsupervised learning laws (Zurada, 1992) or supervised learning laws (Rumelhart, et. al., 1986) Various classification problems can be accomplished by the use of ANN which consists of layers of processing units, each one of which feeds the input to the next layer in a feedforward manner (Sontag, 1992) through set of connection weights. In the presence of pattern classification the connection weights are required to produce the appropriate classification for the various input patterns those are not be given in the training set. For such an updating of weights in super-history mode, it becomes necessary to know the desired output for each unit in the hidden and input layers so that the error the difference between the desired and actual outputs from each units may be used to guide the updation in weights. We may use the propagation learning rule (Werbos, 1974, Hush and Horne, 1993) in which differentiable non-linear output function for each unit in the hidden and output layers and propagated the error from output layer to the hidden layers for updating the weights. As we know that the Heisenberg uncertainty principle is the natural phenomenon that can arrive
For observing the particle, we must shine light of certain wavelength on it. If the wavelength of observing light is longer then there will be a great uncertainty in measuring of particle‘s position and if it shorter then there will be a great uncertainty in measuring the momentum of the particle i.e, the momentum and position of the particle cannot be simultaneously measured with certainty for any wavelength of observing light. It can also be shown that if any one of either position or momentum is likely to be certain with some measure then the other one become the uncertain with the same measure. In this present paper, we are using the pattern classification technique with artificial neutral networks to train a system, which is capable to identify for which wavelength of observing light the position or momentum will likely to be measured. The suitable system for this purpose can be constructed that contains the control of two experiments A and B, where experiment A measures the position and experiment B measures the momentum of a sub-atomic particle, The system will decide whether the position or momentum is less uncertain for a certain wavelength of observing light sat any instant and starts measuring the less uncertain quantity with the corresponding experiment. Hence to accomplish this task we can start with back propagation supervisory learning algorithm of artificial neutral networks for training the network for proper classification of measurement in position and measurement in momentum for any wavelength of observing light to decide the control mechanism of the experiments. The technique can also be used to determine the prediction for measuring the position and momentum of the particle for a different wavelength of the observing light for which the network has not been trained.
HEISENBERG‟S UNCERTAINTY PRINCIPLE
WITH PATTERN CLASSIFICATION:
Heisenberg proposed a very interesting that is direct consequence of dual nature of particle (Heisenberg, 1972). According to classical mechanics a moving particle has a definite momentum and occupies a definite position in space and it is possible to determine both position and momentum at any instant. In quantum mechanics a particle is described by a wave packet, which moves with group velocity. Quantum theory comes out into play when we associate the wavelength with the momentum of the corresponding particle. If the wave packet does not have a well defined wavelength, the particle does not have a well defined momenturn, This means that there is not only an uncertainty in its momentum, caused by the spread in spread wavelength. The two uncertainties are interrelated, because the spread in wavelength (i.e, the uncertainty position). If we want to localize the particle more precisely, that is, if we in the momentum of the particle. According to Bohr‘s probability interpretation (Bordm, et. al., 1926), the particle may be found any where within the wave packet. For a large wave packet with many crests the spreadness in velocity is very small so that the particle velocity can be accurately determined, but the position of position of particle become very uncertain. On the other hand for a small wave packet, the position of the particle can be more or less fixed but the velocity spreads of such a packet is very large so that it becomes uncertain. Hence it is impossible to measure simultaneously both the position and momentum of a particle with certainty at any instant. The precise mathematical form of this relation between the uncertainty in position and momentum of a particle is given by Heisenberg as: This equation indicates that the particle cannot possess a definite position as well as momentum at any instant and measurement process introduces the indeterminacy of the order ∆p ∆x. This indeterminacy is inherent in nature of a moving particle. Thus it can be understood by the Heisenberg uncertainty principle that in the sub atomic world both the momentum and position of a particle cannot be measure simultaneously with accuracy; but then will have to remain completely ignorant about the other one. Now in the process of measuring either the momentum or position, which is less uncertain for the observing light, we can consider a system, which consists of a light source with n adjustable slit apparatus as shown in Fig.1, in which the width of slit can be increased or decreased and the observing photon beam of a certain wavelength λ comes out from the slit that strikes the electron. We can observe the electron with the help of a microscope of high resolution. Each photon of light has the momentum
h
When these photons bounce off the electron, the original momentum will be changed. The exact amount of the change in momentum ∆p cannot be predicted but it will be of the same order as the photon momentum. i.e.
Because light has dual nature, wave nature as well s particle nature, we are not able to determine the exact location of electron in the substance regardless of high resolution instruments. An estimate of minimum uncertainty in measurement of position of electron may be of one photon wavelength. Hence the shorter the wavelength, the smaller the uncertainty in position. If we use light of short wavelength to increase the certainty in measurement of electron in the substance, there will be a corresponding decrease in the certainty in measurement of momentum of electron because the higher momentum of striking photon will perturb the motion of electron to a great extent. Light of longer wavelength will cause more certainty in measurement of electron with less certainty in measurement of position. The system also consists of two experiment A and B, where experiment A measures the position and experiment B measures the momentum of electron. The two experiments A ad B are controlled by the intelligent network in such a way that experiment A will work only when the position of electron is likely to be measured with less uncertainty than the momentum and experiment B will work ony when the experiment works, then other site idle, i.e, these two experiments cannot work simultaneously for a certain wavelength of striking photon. Thus for different wavelength of light photon, we can collect the input output patterns in which the different wavelength of observing light can be considered as the input patterns and corresponding the less uncertainty observation either in position or momentum can be considered as output pattern. Thus, the network can classify the input pattern the form of wavelength (λ1, λ2,…. λi…… λj….. λn) into two classes, Class A and Class B. If the network produces an output I for any input pattern for which the light has shorter wavelength (λ1, λ2,…. λi), then it belongs to class A and network will start experiment A. If the network produces an output 0 for any input pattern for which the slit has large width, say (λi,λj + 1…. Λn) then it belongs to class B and network will start experiment B. To accomplish the task of pattern classification, we consider a multiple layer feed forward Artificial neutral networks (Sontag, 1992) as in Fig.2 with the non-linear differentiable function in all processing units of output and all hidden layers. The number of processing units in the input layer corresponds to the dimensionality of the input pattern vectors, which are linear. The number of output processing units corresponds to the number of distinct classes in the pattern classification. The problem that we are considering, containing only the two pattern classes one for position and other for momentum, can be classified with linear separability using only one processing output unit. The number of hidden layers decide the shape of convex and in the problem that we are discussing here involves only the linear separable classes so there is no necessity of any hidden layer but for the
Now we can use the method of back propagation leaving rule (Werbos, 1974) so that a network can be trained to capture the mapping implicit in the set of input output pattern collected during an experiment and simultaneously expected to model the unknown system or function from which the prediction can be made for the new or untrained set of date (Bordm, et. al., 1926). Let us consider the input pattern (λj, λ2,…. λi) from the set of input patterns for which the experiment A should work. This pattern can be trained to the system by defining the error as squared difference between the desired output and actual output as: Here the target output should be I. i.e. where, EA = Error function for experiment A. QAj = Actual output for experiment A at node j Thus, we are here providing training in the network for learning the patter for experiment A, hence OAj should be near to 1 for providing the correct classification and it can be achieved by reducing the error (EA) with modification in weights of processing units of hidden and output layers as (Werbos, 1974). If the input pattern is (λj λj…. λn) then the experiment B should work. This pattern can also be trained by the system by defining the error for which the target output should be 0; where EB = Error function for experiment B QBj = Actual output for experiment B at node j. Thus, we are here providing training to the network for leaving the pattern experiment B, hence Onj should be near to O for producing the correct classification and it can be achieved by reducing the error (EB) with modification in weights of processing units of hidden and output layers as;
ALORITHM :
(1) Initialize the weights and threshold in the network for (w1, w2….. w1…. wj….. wn) and (1, 2….. i…. j….. n ) from the small random numbers. (2) Present the input pattern in the form of different wavelength of observing light to the network as; ( λ1, λ2….. λi…. λj….. λn ) Select the desire output for the observed input pattern as; For input pattern ( λ1, λ2….. λi ), the desired output is 1. Fr input pattern ( λj, λj + 1….. λn), the desired output is 0. (3) Do until the error is minimized up to a certain limit. Begin Compute the output function as : adjust the weight in output and hidden units as :
End (4) If the output is classified in class A then the system will start the experiment A otherwise experiment B. There is no situation that the two uncertainties are the same so there is no question of miss classification] From the above discussion it can be understood that the network can be trained to decide which experiment should operate at any instant. We can also use this arrangement for predicting the uncertainty in measurement of either position or momentum at any instant for any unknown input pattern in the form of light with different wavelengths, which has not been used during the training to the network. To meet this goal we can consider an unknown pattern λj + k and the network produced an interpolated version of the output class corresponding to the input learning pattern close to the given test input pattern. Thus, the network has been trained to capture the mapping implicit in the set of input output pattern and simultaneously expected to model the unknown function from which the prediction can be made about the new or untrained set of data. Hence the trained network will predict can be made about the new or untrained set of date. Hence the trained to be measured with certainty for the new unknown input pattern. Thus, if the network produces the interpolated version of output 1, then the position is less uncertain with respect to momentum. So the experiment A for measuring the position will conduct while setting other idle. And if the network produces the interpolated version of output O, then the position is less uncertain will respect to momentum. So the experiment B for measuring the momentum will conduct.
CONCLUSION:
Any sub – atomic particle such as an electron in a substance be observed without changing its momentum. To look at an electron, we must shine light of certain wavelength on it and the light photon strikes to the electron, cause the change in momentum of electron. Because of dual nature of particle there will be some indeterminacy in measurement of position and momentum of electron. To find out whether the position or light, we are using the neural network pattern classification technique with Back Propagation learning rule. In this process a suitable arrangement can be constructed that contains the control of two experiments, one of which whether position or momentum of electron is likely to be certain for any wavelength of striking light photon and starts measuring the less quantity with the corresponding experiment.
REFERENCES:
1. Muller, B. and I Reinhardt (1991). Neural Networks, An introduction, Physics of Neural Network, New York springer-Verlag. 2. McCulloch, W.S and W. Pitts (1943). Bull, Math, Bin, Phy, 5, 115-133. 3. Simpson, P.K. (1990). Artificial Neural Systems, Foundation, Paradigms, Application and Implemention Eimsford, N4: Pergamon Press. 4. Dorigo, M. Mariezzo, V. and A. Colomi (1996). IEEE Trans. Systems 26, p. 29. 5. Belew, R.K. (1989). Proc Third Int. Conf. UnG., George Manson Univ. 34 6. Adleman, I. (1994). Science 266, p. 1021. 7. Hunt, S.E. and D.E. Cooke (1996). J. Network Comp. Appl. 19, p. 189. 8. Kosko, B. (1991). Neural Network for Signal Processing PH, Englewood Chifts, N.J. (1991). 9. Hebb. D.O. (1949). The organization of Behavior, A Neuropsychological Theory, NY Wiley. 10. Kanal, I. (1987). IEEE Trans. Inform Theory, FF-20, 0697-722(1974); J Mantas Pattern Recognition 20(1) , pp. 1-6. 11. Hassoun, M.H. (1995). Fundamentals of Artificial Networks, Cambridge, MA, MIT Press. 12. Zurada, J.M. (1992). Introduction to Artificial Neural System, Singapore Info. Access and Distribution. 13. Rumelhart, D.E. Hinten, G.E. and R.J. Williams (1986). Parallel Distributed Processing 1, pp. 318-362. 14. Sontag, E.D. (1992). J. Computing Systems Science, 45, pp. 20-48 15. Werbos, P.J. (1974). Beyond Recognition, New Tools, For Predication and Analysis in
16. Hush, D.R and B.G Horne (1993). IEEE Signal Processing Magazine 10, pp. 8-39. 17. Heisenberg. W. (1972). Zeitscheift Fiir Physik, 43 172. 18. Bordm M., Heisenberg W. and P Jordan (1926). Zeitschrift Fiir Physik, 35, p. 557. 19. Haykin, S. (1994). Neural Network: A comprehensive Foundation, New York, Macmillan College Publish Company.
Corresponding Author Mrs. Punam*
Research Scholar, Department of Computer Science, OPJS University, Rajasthan