26.01.2015 Views

ULTIMATE COMPUTING - Quantum Consciousness Studies

ULTIMATE COMPUTING - Quantum Consciousness Studies

ULTIMATE COMPUTING - Quantum Consciousness Studies

SHOW MORE
SHOW LESS
  • No tags were found...

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

From Brain to Cytoskeleton 69<br />

implied that at birth there existed a vast number of redundant and ineffective<br />

synaptic conditions which became “selected” during the individual’s lifetime of<br />

experience. An alternative view is that, at birth, excitations can pass between any<br />

two points of the CNS through a random network of connections. As maturation,<br />

experience, and learning occurred, synaptic activity gradually sculpted usable<br />

patterns by suppressing unwanted interconnections.<br />

Thus the connectionist brain/mind became viewed as one of two types of<br />

systems: a blank slate (“tabula rasa”) in which acquired learning and internal<br />

organization result from direct environmental imprinting, or a “selectionist”<br />

network chosen from a far vaster potential network. Selectionists believe that the<br />

brain/mind spontaneously generates variable patterns of connections during<br />

childhood periods of development referred to as “transient redundancy,” or from<br />

variable patterns of activity called prerepresentations in the adult. Environmental<br />

interactions merely select or selectively stabilize preexisting patterns of<br />

connections and/or neural firings which fit with the external input. Selectionists<br />

further believe that, as a correlate of learning, connections between neurons are<br />

eliminated (pruning) and/or the number of accessible firing patterns is reduced.<br />

Supporting a selectionist viewpoint is the observation that the number of neurons<br />

and apparent synapses decreases during certain important stages of development<br />

in children. However, this reduction could be masking an increase in complexity<br />

among dendritic arborizations, spines, synapses, and cytoskeleton. The<br />

selectionist view is also susceptible to the argument that new knowledge would<br />

appear difficult to incorporate.<br />

On the assumption that the basic mode of learning and consciousness within<br />

the brain is based on synaptic connections among neurons (connectionist view)<br />

several attempts to model learning at the level of large assemblies of<br />

interconnected neurons have been made. Hebb pioneered this field by proposing<br />

that learning occurred by strengthening of specific synaptic connections within a<br />

neuronal network. This led to a concept of functional groups of neurons<br />

connected by variable synapses. These functional groups as anatomical brain<br />

regions have been described by various authors as networks, assemblies, cartels,<br />

modules or crystals. These models are aided by the mathematics of statistical<br />

mechanics and have been rejuvenated due to the work of Hopfield (1982),<br />

Grossberg (1978), Kohonen (1984) and others who drew analogies between<br />

neural networks within the brain and properties of computers leading to<br />

applications for artificial intelligence. They emphasized that computational<br />

properties useful to biological organisms or to the construction of computers can<br />

emerge as collective properties of systems having a large number of simple<br />

equivalent components or neurons with a high degree of interconnection. Neural<br />

networks started as models of how the brain works and have now engendered<br />

chips and computers constructed with neural net connectionist architectures<br />

utilizing hundreds of computing units and linking them with many thousands of<br />

connections. Hopfield (1982) remarks that neural net chips can provide finely<br />

grained and massively parallel computing with:<br />

a brainlike tolerance for fuzzy facts and vague instructions. Some of<br />

the general properties you get in these systems are strikingly like ...<br />

properties we see in neurobiology ... . You don’t have to build them<br />

in; they’re just there ... .<br />

Neural networks had formally appeared in Rosenblatt’s (1962) “perceptron”<br />

model of the 1950’s and 1960’s. Perceptrons created enthusiasm, but failed to<br />

reach their potential due to limitations of the model and its mathematics. Al<br />

experts Marvin Minsky and Seymour Papert (1972) wrote a harshly critical

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!