Randomness is a concept with somewhat disparate meanings in several fields. It also has common meanings which may have loose connections with some of those more definite meanings. The Oxford English Dictionary defines "random" thus:
Having no definite aim or purpose; not sent or guided in a particular direction; made, done, occurring, etc., without method or conscious choice; haphazard.
Also, in statistics, as:
Governed by or involving equal chances for each of the actual or hypothetical members of a population; (also) produced or obtained by such a process, and therefore unpredictable in detail.
Closely connected, therefore, with the concepts of chance, probability, and information entropy, randomness implies a lack of predictability. More formally, in statistics, a random process is a repeating process whose outcomes follow no describable deterministic pattern, but follow a probability distribution, such that the relative probability of the occurrence of each outcome can be approximated or calculated. For example, the rolling of a fair sixsided die in neutral conditions may be said to produce random results, because one cannot compute, before a roll, what number will show up. However, the probability of rolling any one of the six rollable numbers can be calculated, assuming that each is equally likely.
The term is often used in statistics to signify welldefined statistical properties, such as a lack of bias or correlation. Monte Carlo Methods, which rely on random input, are important techniques in science, as, for instance, in computational science.^{[1]} Random selection is an official method to resolve tied elections in some jurisdictions^{[2]} and is even an ancient method of divination, as in tarot, the I Ching, and bibliomancy. Its use in politics is very old, as office holders in Ancient Athens were chosen by lot, there being no voting.
Contents 
In ancient history, the concepts of chance and randomness were intertwined with that of fate. Many ancient peoples threw dice to determine fate, and this later evolved into games of chance. Most ancient cultures used various methods of divination to attempt to circumvent randomness and fate.^{[3]}^{[4]}
The Chinese were perhaps the earliest people to formalize odds and chance 3,000 years ago. The Greek philosophers discussed randomness at length, but only in nonquantitative forms. It was only in the sixteenth century that Italian mathematicians began to formalize the odds associated with various games of chance. The invention of the calculus had a positive impact on the formal study of randomness. In the 1888 edition of his book The Logic of Chance John Venn wrote a chapter on "The conception of randomness" which included his view of the randomness of the digits of the number Pi by using them to construct a random walk in two dimensions.^{[5]}
The early part of the twentieth century saw a rapid growth in the formal analysis of randomness, as various approaches for a mathematical foundations of probability were introduced. In the mid to late twentieth century ideas of algorithmic information theory introduced new dimensions to the field via the concept of algorithmic randomness.
Although randomness had often been viewed as an obstacle and a nuisance for many centuries, in the twentieth century computer scientists computer scientist began to realize that the deliberate introduction of randomness into computations can be an effective tool for designing better algorithms. In some cases such randomized algorithms outperform the best deterministic methods.
Many scientific fields are concerned with randomness:
In the 19th century, scientists used the idea of random motions of molecules in the development of statistical mechanics in order to explain phenomena in thermodynamics and the properties of gases.
According to several standard interpretations of quantum mechanics, microscopic phenomena are objectively random^{[citation needed]}. That is, in an experiment where all causally relevant parameters are controlled, there will still be some aspects of the outcome which vary randomly. An example of such an experiment is placing a single unstable atom in a controlled environment; it cannot be predicted how long it will take for the atom to decay; only the probability of decay within a given time can be calculated.^{[6]} Thus, quantum mechanics does not specify the outcome of individual experiments but only the probabilities. Hidden variable theories are inconsistent with the view that nature contains irreducible randomness: such theories posit that in the processes that appear random, properties with a certain statistical distribution are somehow at work "behind the scenes" determining the outcome in each case.
The modern evolutionary synthesis ascribes the observed diversity of life to natural selection, in which some random genetic mutations are retained in the gene pool due to the nonrandom improved chance for survival and reproduction that those mutated genes confer on individuals who possess them.
The characteristics of an organism arise to some extent deterministically (e.g., under the influence of genes and the environment) and to some extent randomly. For example, the density of freckles that appear on a person's skin is controlled by genes and exposure to light; whereas the exact location of individual freckles seems to be random.^{[7]}
Randomness is important if an animal is to behave in a way that is unpredictable to others. For instance, insects in flight tend to move about with random changes in direction, making it difficult for pursuing predators to predict their trajectories.
The mathematical theory of probability arose from attempts to formulate mathematical descriptions of chance events, originally in the context of gambling, but later in connection with physics. Statistics is used to infer the underlying probability distribution of a collection of empirical observations. For the purposes of simulation, it is necessary to have a large supply of random numbers or means to generate them on demand.
Algorithmic information theory studies, among other topics, what constitutes a random sequence. The central idea is that a string of bits is random if and only if it is shorter than any computer program that can produce that string (Kolmogorov randomness)—this means that random strings are those that cannot be compressed. Pioneers of this field include Andrey Kolmogorov and his student Per MartinLöf, Ray Solomonoff, and Gregory Chaitin.
In mathematics, there must be an infinite expansion of information for randomness to exist. This can best be seen with an example. Given a random sequence of threebit numbers, each number can have only eight possible values:
000, 001, 010, 011, 100, 101, 110, 111
Therefore, as the random sequence progresses, it must recycle through the values it previously used. In order to increase the information space, another bit may be added to each possible number, giving 16 possible values from which to pick a random number. It could be said that the random fourbit number sequence is more random than the threebit one. This suggests that in order to have true randomness, there must be an infinite expansion of the information space.
Randomness is said to occur in numbers such as log (2) and Pi. The decimal digits of Pi constitute an infinite sequence and "never repeat in a cyclical fashion". Numbers like pi are also thought to be normal, which means that their digits are random in a certain statistical sense.
Pi certainly seems to behave this way. In the first six billion decimal places of pi, each of the digits from 0 through 9 shows up about six hundred million times. Yet such results, conceivably accidental, do not prove normality even in base 10, much less normality in other number bases.^{[8]}
In information science, irrelevant or meaningless data is considered to be noise. Noise consists of a large number of transient disturbances with a statistically randomized time distribution.
In communication theory, randomness in a signal is called "noise" and is opposed to that component of its variation that is causally attributable to the source, the signal.
The random walk hypothesis considers that asset prices in an organized market evolve at random.
Other socalled random factors intervene in trends and patterns to do with supplyanddemand distributions. As well as this, the random factor of the environment itself results in fluctuations in stock and broker markets.
Randomness, as opposed to unpredictability, is held to be an objective property  determinists believe it is an objective fact that randomness does not in fact exist. Also, what appears random to one observer may not appear random to another. Consider two observers of a sequence of bits, when only one of whom has the cryptographic key needed to turn the sequence of bits into a readable message. For that observer the message is not random, but it is unpredictable for the other.
One of the intriguing aspects of random processes is that it is hard to know whether a process is truly random. An observer may suspect that there is some "key" that unlocks the message. This is one of the foundations of superstition, and is also a motivation for discovery in science and mathematics.
Under the cosmological hypothesis of determinism, there is no randomness in the universe, only unpredictability, since there is only one possible outcome to all events in the universe. A follower of the narrow frequency interpretation of probability could assert that no event can be said to have probability, since there is only one universal outcome. On the other hand, under the rival Bayesian interpretation of probability there is no objection to the use of probabilities in order to represent a lack of complete knowledge of the outcomes.
Some mathematically defined sequences, such as the decimals of pi mentioned above, exhibit some of the same characteristics as random sequences, but because they are generated by a describable mechanism, they are called pseudorandom. To an observer who does not know the mechanism, a pseudorandom sequence is unpredictable.
Chaotic systems are unpredictable in practice due to their extreme sensitivity to initial conditions. Whether or not they are unpredictable in terms of computability theory is a subject of current research. At least in some disciplines of computability theory, the notion of randomness is identified with computational unpredictability.
Individual events that are random may still be precisely described en masse, usually in terms of probability or expected value. For instance, quantum mechanics allows a very precise calculation of the halflives of atoms even though the process of atomic decay is random. More simply, although a single toss of a fair coin cannot be predicted, its general behavior can be described by saying that if a large number of tosses are made, roughly half of them will show up heads. Ohm's law and the kinetic theory of gases are nonrandom macroscopic phenomena that are assumed to be random at the microscopic level.
Some theologians have attempted to resolve the apparent contradiction between an omniscient deity, or a first cause, and free will using randomness. Discordians have a strong belief in randomness and unpredictability. Buddhist philosophy states that any event is the result of previous events (karma), and as such, there is no such thing as a random event or a first event.
Martin Luther, the forefather of Protestantism, believed that there was nothing random based on his understanding of the Bible. As an outcome of his understanding of randomness, he strongly felt that free will was limited to lowlevel decision making by humans. Therefore, when someone sins against another, decision making is only limited to how one responds, preferably through forgiveness and loving actions. He believed, based on Biblical scripture, that humans cannot will themselves faith, salvation, sanctification, or other gifts from God. Additionally, the best people could do, according to his understanding, was not sin, but they fall short, and free will cannot achieve this objective. Thus, in his view, absolute free will and unbounded randomness are severely limited to the point that behaviors may even be patterned or ordered and not random. This is a point emphasized by the field of behavioral psychology.
These notions and more in Christianity often lend to a highly deterministic worldview and that the concept of random events is not possible. Especially, if purpose is part of this universe, then randomness, by definition, is not possible. This is also one of the rationales for religious opposition to evolution, where, according to theory, (nonrandom) selection is applied to the results of random genetic variation.
Donald Knuth, a Stanford computer scientist and Christian commentator, remarks that he finds pseudorandom numbers useful and applies them with purpose. He then extends this thought to God who may use randomness with purpose to allow free will to certain degrees. Knuth believes that God is interested in people's decisions and limited free will allows a certain degree of decision making. Knuth, based on his understanding of quantum computing and entanglement, comments that God exerts dynamic control over the world without violating any laws of physics, suggesting that what appears to be random to humans may not, in fact, be so random.^{[9]}
C. S. Lewis, a 20thcentury Christian philosopher, discussed free will at length. On the matter of human will, Lewis wrote: "God willed the free will of men and angels in spite of His knowledge that it could lead in some cases to sin and thence to suffering: i.e., He thought freedom worth creating even at that price." In his radio broadcast, Lewis indicated that God "gave [humans] free will. He gave them free will because a world of mere automata could never love..."
In some contexts, procedures that are commonly perceived as randomizers—drawing lots or the like —are used for divination, e.g., to reveal the will of the gods; see e.g. Cleromancy.
In most of its mathematical, political, social and religious use, randomness is used for its innate "fairness" and lack of bias.
Political: Greek Democracy was based on the concept of isonomia (equality of political rights) and used complex allotment machines to ensure that the positions on the ruling committees that ran Athens were fairly allocated. Allotment is now restricted to selecting jurors in AngloSaxon legal systems and in situations where "fairness" is approximated by randomization, such as selecting jurors and military draft lotteries.
Social: Random numbers were first investigated in the context of gambling, and many randomizing devices, such as dice, shuffling playing cards, and roulette wheels, were first developed for use in gambling. The ability to produce random numbers fairly is vital to electronic gambling, and, as such, the methods used to create them are usually regulated by government Gaming Control Boards. Random drawings are also used to determine lottery winners. Throughout history, randomness has been used for games of chance and to select out individuals for an unwanted task in a fair way (see drawing straws).
Sports: Some sports, including American Football, use coin tosses to randomly select starting conditions for games or seed tied teams for postseason play. The National Basketball Association uses a weighted lottery to order teams in its draft.
Mathematical: Random numbers are also used where their use is mathematically important, such as sampling for opinion polls and for statistical sampling in quality control systems. Computational solutions for some types of problems use random numbers extensively, such as in the Monte Carlo method and in genetic algorithms.
Medicine: Random allocation of a clinical intervention is used to reduce bias in controlled trials (e.g., randomized controlled trials).
Religious: Although not intended to be random, various forms of divination such as cleromancy see what appears to be a random event as a means for a divine being to communicate their will. (See also Free will and Determinism).
It is generally accepted that there exist three mechanisms responsible for (apparently) random behavior in systems:
The many applications of randomness have led to many different methods for generating random data. These methods may vary as to how unpredictable or statistically random they are, and how quickly they can generate random numbers.
Before the advent of computational random number generators, generating large amounts of sufficiently random numbers (important in statistics) required a lot of work. Results would sometimes be collected and distributed as random number tables.
There are many practical measures of randomness for a binary sequence. These include measures based on frequency, discrete transforms, and complexity, or a mixture of these. These include tests by Kak, Phillips, Yuen, Hopkins, Beth and Dai, Mund, and Marsaglia and Zaman.^{[10]}
Popular perceptions of randomness are frequently wrong, based on logical fallacies. The following is an attempt to identify the source of such fallacies and correct the logical errors.
This argument is that "in a random selection of numbers, since all numbers will eventually appear, those that have not come up yet are 'due', and thus more likely to come up soon." This logic is only correct if applied to a system where numbers that come up are removed from the system, such as when playing cards are drawn and not returned to the deck. In this case, once a jack is removed from the deck, the next draw is less likely to be a jack and more likely to be some other card. However, if the jack is returned to the deck, and the deck is thoroughly reshuffled, a jack is as likely to be drawn as any other card. The same applies in any other process where objects are selected independently, and none are removed after each event, such as the roll of a die, a coin toss, or most lottery number selection schemes. Truly random processes such as these do not have memory, making it impossible for past outcomes to affect future outcomes.
In a random sequence of numbers, a number may be said to be cursed because it has come up less often in the past, and so it is thought that it will occur less often in the future. A number may be assumed to be blessed because it has occurred more often than others in the past, and so it is thought to be likely to come up more often in the future. This logic is valid only if the randomisation is biased, for example with a loaded die. If the die is fair, then previous rolls give no indication of future events.
In nature, events rarely occur with perfectly equal frequency. So observing outcomes to determine which events are likely to have a higher probability, makes sense. It is fallacious to apply this logic to systems which are designed so that all outcomes are equally likely, such as shuffled cards, dice and roulette wheels.
Educational level: this is a nonformal education resource. 
Development status: this resource is experimental in nature. 
Learning and research project about randomness and wikis  feel free to join in.
A launching point for exploring Wikiversity  visit random links and see what you find....
RANDOM (older forms randon, randoun; from the French, cf. randir, to run quickly, impetuously; generally taken to be of Teutonic origin and connected with Ger. Rand, edge, brim, the idea being possibly of a brimming river), an adjective originally meaning impetuous, hasty, hence done without purpose or aim, haphazard. The term "random work" is used, in architecture, by the ragstone masons, for stones fitted together at random without any attempt at laying them in courses. "Random coursed work" is a like term applied to work coursed in horizontal beds, but the stones are of varying height, and fitted to one another (see Masonry).
Ranelagh >> 
Categories: RRAN
Random is a term used in mathematics, as well as in a more general use, for situations when there is no way to know the outcome (what will happen). Something that is chosen at random is not chosen for any special reason, but purely by chance. An example of a random act is a lottery.
A computer can make lists of random numbers. Humans are unable to do this, because the brain works in patterns. If someone is asked to keep saying "heads" or "tails" at random, a computer would soon be able to tell which one the person is likely to say next because the computer notices the patterns.
On a website such as English Wikipedia the user can click on "Random page" to get a random article. The chances of any one page appearing are exactly the same as for any other page.
Sometimes the word "random" is used more loosely. There are websites of random jokes, which just means: a variety of jokes about all sorts of things.
In recent years young people have started to use the word "random" even more loosely to describe anything which is rather strange or has no logic. Sentences such as "the mouldy cheese is escaping" or "I like pie and spam" might be described as "random", although this is not the proper dictionary or mathematical meaning of the word.
There are several ways in which a process or system can be seen as random:
The many applications of randomness have led to many different methods for generating random data. These methods may vary as to how unpredictable or statistically random they are, and how quickly they can generate random numbers.
Before there were computational random number generators, generating large amounts of sufficiently random numbers (important in statistics) required a lot of work. Results would sometimes be collected and distributed as random number tables.
Computers can make what appear to be random numbers. There are two ways to make them.
