.

Saturday, March 2, 2019

Introduction to Computer Theory

CHAPTER 1 BACKGROUND The twentieth century has been filled with the most unconvincing shocks and surprises the theory of relativity, Communist revolutions, psychoanalysis, nuclear war, television, moon walks, ge light upic engineering, and so on. As astounding as any of these is the advent of the computing machine and its development from a mere calculating device into what seems like a thinking machine. The receive of the estimator was non wholly self-sufficing of the other events of this century.The history of the computer is a fascinating story however, it is non the issuing of this course. We atomic number 18 waste-to doe with with the supposition of computers, which means that we form several filch numeric casts that forget describe with varying degrees of accuracy move of computers and types of computers and similar machines. Our models will non be used to discuss the practical engineering details of the ironwargon of computers, exactly the to a greater e xtent abstract questions of the frontiers of capability of these mechanical devices.There are crumble courses that deal with circuits and switching theory (computer logic) and with instruction sets and register arrangements (computer ar-chitecture) and with data structures and algorithms and operational arrangements and compiler design and artificial intelligence and so forth. All of these courses expect a metaphysical comp superstarnt, just now they differ from our study in ii basic ways. First, they deal besides with computers that already exist our models, on 3 4 AUTOMATA THEORY the other hand, will encompass t turn out ensemble(prenominal) computers that do exist, will exist, and that keep ever be dreamed of.Second, they are kindle in how best to do things we sh altogether non be interested in optimality at all, but rather we shall be implicated with the question of possibility-what grass and what sewernot be done. We shall look at this from the perspective of what language structures the machines we describe fundament and cannot accept as remark, and what feasible meaning their output may have. This description of our intent is extremely familiar and perhaps a little misleading, but the mathematically precise definition of our study can be understood only by those who already know the opinions introduced in this course.This is often a characteristic of scholarshipafter years of study one can just begin to define the subject. We are now em runking on a typical example of such a journey. In our last chapter (Chapter 31) we shall eventually be able to define a computer. The history of computing device supposition is as well interesting. It was formed by fortunate coincidences, involving several seemingly orthogonal branches of intellectual endeavor. A small series of contemporaneous discoveries, by actually dissimilar battalion, separately motivated, flowed together to become our subject.Until we have established to a grea ter extent of a foundation, we can only describe in general damage the different schools of thought that have melded into this champaign. The most obvious component of Computer Theory is the theory of mathematical logic. As the twentieth century started, math was facing a dilemma. Georg Cantor (1845-1918) had recently invented the Theory of Sets (unions, inter branchs, inclusion, cardinality, etc. ). moreover at the same time he had discovered some genuinely uncomfortable paradoxes-he created things that looked like contradictions in what seemed to be harshly proven mathematical theorems.Some of his unusual bechanceings could be tolerated (such as that infinity comes in different sizes), but some could not (such as that some set is bigger than the world-wide set). This left a cloud over mathematics that needed to be re single-minded. David Hilbert (1862-1943) wanted all of mathematics put on the same die footing as Euclidean Geometry, which is characterized by precise defini tions, explicit axioms, and rigorous test copys. The format of a Euclidean proof is precisely specified. Every field is either an axiom, a previously proven theorem, or follows from the lines above it by one of a few innocent rules of inference.The mathematics that developed in the centuries since Euclid did not follow this standard of precision. Hilbert believed that if mathematics Xere put back on the Euclidean standard the Cantor paradoxes would go away. He was actually concerned with two ambitious projects low, to demonstrate that the new system was free of paradoxes second, to find methods that would pledge to enable humans to construct proofs of all the true reports in mathematics. Hilbert wanted something formulaic-a precise routine for producing results, like the directions in a cookbook.First escape all these lines, then write all these equations, then solve for all these take aims, and so on and so on and the proof is done-some approach that is certain(a) and sur e-fire without any reliance BACKGROUND 5 on maverick and undependable brilliant mathematical insight. We simply follow the rules and the answer moldiness come. This type of complete, guaranteed, easy-to-follow set of instructions is called an algorithm. He hoped that algorithms or procedures could be developed to solve whole classes of mathematical fusss.The collection of techniques called linear algebra provides just such an algorithm for solution all systems of linear equations. Hilbert wanted to develop algorithms for solving other mathematical problems, perhaps even an algorithm that could solve all mathematical problems of any kind in some finite human action of steps. Before starting to look for such an algorithm, an exact notion of what is and what is not a mathematical statement had to be developed. After that, there was the problem of defining exactly what can and what cannot be a step in an algorithm.The invents we have used procedure, formula, cookbook method, comple te instructions, are not part of mathematics and are no more meaningful than the word algorithm itself. Mathematical logicians, while trying to follow the suggestions of Hilbert and straighten out the predicament left by Cantor, found that they were able to prove mathematically that some of the desired algorithms cannot exist-not only at this time, but they can neer exist in the future, either. Their main I result was even more fantastic than that.Kurt Godel (1906-1978) not only showed that there was no algorithm that could guarantee to provide proofs for all the true statements in mathematics, but he turn up that not all the true statements even have a proof to be found. G6dels Incompleteness Theorem implies that in a specific mathematical system either there are some true statements without any possible proof or else there are some false statements that can be proven. This earth-shaking result made the mess in the philosophy of mathematics even worse, but very exciting.If not e very true statement has a proof, can we at least fulfill Hilberts weapons platform by finding a proof-generating algorithm to provide proofs whenever they do exist? Logicians began to crave the question Of what fundamental parts are all algorithms composed? The first general definition of an algorithm was proposed by Alonzo Church. Using his definition he and Stephen Cole Kleene and, independently, Emil Post were able to prove that there were problems that no algorithm could solve. While also solving this problem independently, Alan Mathison Turing (1912-1954) developed the concept of a theoretical universal-algorithm machine. Studying what was possible and what was not possible for such a machine to do, he discovered that some tasks that we might have expect this abstract omnipotent machine to be able to perform are impossible, even for it. Turings model for a universal-algorithm machine is directly connected to the maneuver of the computer. In fact, for completely different rea sons (wartime code-breaking) Turing himself had an important part in the social structure of the first computer, which he found on his work in abstract logic.On a wildly different front, two look intoers in neurophysiology, Warren 6 AUTOMATA THEORY Sturgis McCulloch and Walter Pitts (1923-1969), constructed a mathematical model for the way in which sensory receptor organs in animals behave. The model they constructed for a neural net was a theoretical machine of the same nature as the one Turing invented, but with certain limitations. Mathematical models of real and abstract machines took on more and more importance.Along with mathematical models for biological processes, models were introduced to study psychological, economic, and social stakes. Again, entirely independent of these considerations, the invention of the vacuum tube and the subsequent developments in electronics enabled engineers to build richly automatic electronic calculators. These developments fulfilled the age-old dream of Blaise Pascal (1623-1662), Gottfried Wilhelm von Leibniz (1646-1716), and Charles Babbage (1792-1871), all of whom built mechanical calculating devices as powerful as their several(prenominal) technologies would allow.In the 1940s, gifted engineers began building the first generation of computers the computer Colossus at Bletchley, England (Turings decoder), the ABC machine built by John Atanosoff in Iowa, the Harvard signal I built by Howard Aiken, and ENIAC built by John Presper Eckert, jr. and John William Mauchly (1907-1980) at the University of Pennsylvania. Shortly after the invention of the vacuum tube, the undreamt mathematician John von Neumann (1903-1957) developed the idea of a stored-program computer.The idea of storing the program inside the computer and allowing the computer to operate on (and modify) the program as well as the data was a tremendous advance. It may have been conceived decades earlier by Babbage and his co-worker Ada Augusta, Counte ss of Lovelace (1815-1853), but their technology was not adequate to explore this possibility. The ramifications of this idea, as pursued by von Neumann and Turing were quite profound. The early calculators could perform only one prede termined set of tasks at a time.To make changes in their procedures, the calculators had to be physically rebuilt either by rewiring, resetting, or reconnecting various parts. Von Neumann permanently wired certain trading operations into the machine and then designed a central control section that, after reading input data, could select which operation to perform based on a program or algorithm encoded in the input and stored in the computer along with the raw data to be processed. In this way, the inputs determined which operations were to be performed on themselves.Interestingly, current technology has progressed to the point where the ability to manufacture dedicated chips cheaply and easily has made the probability of rebuilding a computer for each program feasible again. However, by the last chapters of this book we will appreciate the significance of the difference in the midst of these two approaches. Von Neumanns goal was to convert the electronic calculator into a reallife model of one of the logicians ideal universal-algorithm machines, such as those Turing had described.Thus we have an unusual situation where the advanced theoretical work on the potential of the machine preceded the introduction that the machine could really exist. The people who first discussed BACKGROUND 7 these machines only dreamed they might ever be built. Many were very affect to find them actually working in their own lifetimes. Along with the concept of programming a computer came the question What is the best language in which to write programs?Many languages were invented, owing their distinction to the differences in the specific machines they were to be used on and to the differences in the types of problems for which they were desig ned. However, as more languages emerged, it became clear that they had some(prenominal) elements in common. They seemed to share the same possibilities and limitations. This observation was at first only intuitive, although Turing had already worked on much the same problem but from a different angle. At the time that a general theory of computer languages was being developed, another surprise occurred.Modem linguists, some influenced by the prevalent trends in mathematical logic and some by the emerging theories of developmental psychology, had been look into a very similar subject What is language in general? How could primitive humans have developed language? How do people empathize it? How do they learn it as children? What ideas can be expressed, and in what ways? How do people construct sentences from the ideas in their minds? Noam Chomsky created the subject of mathematical models for the description of languages to answer these questions.His theory grew to the point where it began to shed light on the study of computer languages. The languages humans invented to communicate with one another and the languages needed for humans to communicate with machines shared many basic properties. Although we do not know exactly how humans understand language, we do know how machines compiling what they are told. Thus, the formulations of mathematical logic became useful to linguistics, a previously nonmathematical subject. Metaphorically, we could hypothesise that the computer then took on linguistic abilities.It became a word processor, a translator, and an interpreter of simple grammar, as well as a compiler of computer languages. The software invented to interpret programming languages was applied to human languages as well. One point that will be made clear in our studies is why computer languages are easy for a computer to understand whereas human languages are very difficult. Because of the many influences on its development the subject of this book goe s by various names. It includes three major fundamental areas the Theory of Automata, the Theory of Formal Languages, and the Theory of Turing Machines.This book is divided into three parts corresponding to these topics. Our subject is sometimes called Computation Theory rather than Computer Theory, since the items that are central to it are the types of tasks (algorithms or programs) that can be performed, not the mechanical nature of the physical computer itself. However, the name computation is also misleading, since it popularly connotes arithmetical operations that are only a separate of what computers can do. The term computation is inaccurate when describing word AUTOMATA THEORY processing, form and searching and awkward in discussions of program verification. Just as the term Number Theory is not limited to a description of calligraphical displays of number systems but condensees on the question of which equations can be solved in integers, and the term Graph Theory does not include bar graphs, pie charts, and histograms, so too Computer Theory need not be limited to a description of physical machines but can focus on the question of which tasks are possible for which machines.We shall study different types of theoretical machines that are mathematical models for actual physical processes. By considering the possible inputs on which these machines can work, we can analyze their various strengths and weaknesses. We then arrive at what we may believe to be the most powerful machine possible. When we do, we shall be surprised to find tasks that even it cannot perform. This will be-our ultimate result, that no calculate what machine we build, there will always be questions that are simple to state that it cannot answer.Along the way, we shall begin to understand the concept of computability, which is the foundation of further research in this field. This is our goal. Computer Theory extends further to such topics as complexness and verification, but th ese are beyond our intended scope. Even for the topics we do cover-Automata, Languages, Turing Machines-much more is known than we present here. As intriguing and engaging as the field has proven so far, with any luck the most fascinating theorems are yet to be discovered.

No comments:

Post a Comment