Home of the original IBM PC emulator for browsers.
[PCjs Machine "ibm5170"]
Waiting for machine "ibm5170" to load....
╔════════════════════════════════════════════════════════╗
║ General Instructions: ║
║ While in any given chapter, ║
║ Screens may be changed in 2 ways: ║
║ >> "+","-",PgUp,PgDn = forward/backward 1 screen ║
║ >> F1 = enter screen number directly ║
║ F2 = chapter contents ║
║ F3 = this information ║
║ Alt/M = main menu ║
║ Alt/X = exit to DOS ║
║ any key to resume ║
╚════════════════════════════════════════════════════════╝
╔════════════════════════════════════════════════════════════════════╗
║ File Locations: (3.5",5.25") ║
║ ║
║ Main Menu --- 1,1 Scales ----------- 1,2 Meredith Monk - 2,3 ║
║ Preface ----- 1,1 Intervals -------- 2,3 Melody Writer - 2,4 ║
║ Schedule ---- 1,1 Musical Analysis - 2,3 Keybaord ------ 2,4 ║
║ Course Info -- 1,1 Harmony ---------- 2,3 Pitch Memory -- 2,4 ║
║ Acoustics ---- 1,1 Sound Polltion --- 2,3 SoundForms ---- 2,4 ║
║ Pitch Basics - 1,2 Harry Partch ----- 2,3 Editor -------- 2,4 ║
║ Rhythm ------- 1,2 Creativity ------- 2,3 ║
║ (any key to resume) ║
╚════════════════════════════════════════════════════════════════════╝
MUSICAL SPACE
That music is a temporal art is an easy notion to grasp.
It is, also, spatial, and in more than one way.
It is most clearly spatial in that sound has location in space.
Where a sound comes from can affect the listener. Thus the loca-
tion of a sound can be an integral aspect of the compositional
conceptioning...
This fact has often been ignored by composers,
but its use in antiphonal church music during
the Renaissance, and its "rediscovery" and de-
velopment by some composers (like Henry Brandt)
in the 20th century, illustrate its viability.
At this time, however, we are not going to consider this
kind of spatiality, but rather the metaphorical sense in
which the pitch of sounds is thought of as corresponding
to relative height, the "up and down" axis.
Pitch, as the frequency (rate of vibration) of a sound, has, of
course, no direct spatial higher or lower coordinate. Why we re-
present frequency in this way, and have throughout history, as
well as cross-culturally, is a matter of speculation.
Be that as it may, pitch is conceived of and represented
in spatial terms. And it is most destinctly a character-
istic of music, to be shaped and controlled. It is often
thought of as the chief incrediant of melody, although
the other parameters of sound also contribute their share.
Pitch is decidely the stuff of which intervals
and scales are made, and since these in turn
are primary organizing factors in most musics,
knowledge of how pitch is notated is of help in
understanding music.
Pitch, being a vibratory rate and perceived on an up versus down
axis, affects the human physiology in both clear and subtle ways.
Extremes of pitch, high or low, produce tensions, both muscular
and mental. Frequencies in the middle range of the audible
spectrum, expecially within the speech range, are more soothing.
These are generalities, of course, for other factors such as
loudness and timbre also affect the mind and body. Subtle
shadings of pitch, especially melodies employing microtones
(East Indian music, for instance, or the Blues), produce very
distinct reactions in us which can be felt bodily in changing
breath rate and heart contractions, among other things.
Every physical body resonates in response to sonic
vibrations, hence pitch has very direct effect on
material objects. The destructive uses to which
sound has been put, described in the chapter on
sound pollution, illustrate this. There is also,
in occult literature, speculation that particular
frequencies can be and have been used to produce
levitation, and are capable of healing properties.
However we approach the issue,
pitch is an essential ingredient
of music and knowledge of its
notation crucial for our better
comprehending how music works.
The Staff
Pitch is most commonly represented on some form of vertical axis.
Generalized pitch level could be notated around a single line
specified as a particular pitch. Or, as has occasionally been done
as a reminder of pitch relationships learned by rote, with up or
down gestures of the hand.
The extension of this idea into the use of more lines to stand for
other specific pitches, called a staff, is attributed to Guido
d'Arezzo (c. 990-1050 AD), who suggested the use of a 3- or 4-line
staff to denote specific pitches (the pitches we now call F2, A2,
and C3). The four-line staff is still employed for the notation
of Gregorian Chant, but as early as the 13th century, a 5-line
staff was used to notate polyphonic ("many-voiced") music, and it
is this staff which is in general use today.
The convention in referring to pitch placement on the modern 5-line
staff is to count from the bottom line or space upwards. Thus we say
"note such and such" is on the 3rd line up, or the 2nds space up, etc.
Pitch Names
In our investigation of acoustics, we discovered that each
doubling of frequency produces an octave. The term octave was
used because the most common number of pitches employed in
melodies in Western music within this doubling of frequencies has
been 7,the 8th pitch being twice the frequency above or below some
starting point--hence, octave.
PRELIMINARY EXERCISE
Hit "H" (to HEAR) to play "Doe A Deer" from The Sound of Music.
Try to determine how many distinctly different pitches there are
within the octave this tune encompasses. (You should be helped by
association with the "do-re-mi" format)! Proceed to next page to
see the tune written out. You will see notes on different lines
and spaces adding up to the number you heard from lowest to highest.
Careful observation of this tune should have lead you to hear and
see 8 different notes, from an extra line below the staff to the
third space up. These 8 notes, when arranged in sequential
ascending order, produce what is called a scale (from the Latin
scala, meaning "ladder"). The bottom note and the top note
constitute the frequency ratio of 2:1, the octave. Each of the
seven different pitches within the octave is designated by a
letter name, from A to G.
If a line or space on the staff is reserved for one of these
pitches, the location of the other pitches can easily be de-
termined by counting, in order, up or down (forward or back-
ward) the letter names of the pitches. For instance, on the
4-line staff of the Gregorian Chant tune you heard a while
ago, the stylized "C" at the top left of each staff desig-
nates that line as what we know via the piano as "middle"
C (261.6 hz). Counting, in order, down the lines and spaces,
one arrives at the pitch F as the first note of the piece.
The other notes can be determined accordingly.
The Clefs
Given a set of parallel horizontal lines on which to designate specific
pitches, all that is necessary is to indicate a starting point: one line
or space of the staff being a particular frequency.
Music notation, evolving at the same time as different styles and
genres of music, developed primarily to accomodate vocal music.
It was reasonable then that the pitches the staff
evolved to denote should essentially represent the
human voice range. We will accept this tradition,
even though instruments can sound frequencies well
outside the range of the voice.
The TREBLE CLEF was invented to represent the frequency range of female
voices. Originally called the "G" clef, it marks where the pitch "G"
within the SOPRANO and ALTO ranges is to be notated. This is on the
second line up, using the stylized "G" symbol.
This is, specifically, 392.2 hz.
The BASS CLEF was designed to accomodate the lower voice ranges (tenor and
bass). It originated as the "F" clef, which designates where F (174.6 Hz),
is written.
Originally, the Bass and Treble clefs were movable, so that their
placement on the staff indicated slightly higher or lower ranges.
Today, however, these clefs remain fixed. But there is a clef which is
movable. It originated to designate a range between treble and bass,
and is known as the "movable C' clef. It is used to locate "middle C."
LEDGER LINES
Notes outside the range of the five lines of the staff may be written
with the use of ledger lines. The same order of notes is continued
above or below the staff and the lines themselves are drawn with the
same distance between them as on the staff itself.
G A B C D E F G B C D E F G A B
D C B A G F E D F E D C B A G F
Notes several ledger lines above or below a staff exceed normal vocal
ranges, but are useful for instruments whose ranges are greater.
THE OTTAVA SIGN
Sometimes, to avoid excessive and confusing addition of ledger
lines, the ottava sign is employed, which shifts the entire
range of the notes under the bracketed sign by an octave. If
the sign is above the notes, the shift is obviously upward;
if below the notes, and often accompanied by the word basso,
the shift is down one octave. <RET>
Previous convention called for the abbreviation of ottava as 8va. Mod-
ern usage, however, accepts the numeral 8. In either case, the length
of time the octave displacement is to occur must be indicated by the
dotted lines.
Occasionally, some instruments (the piano, for one) play in extreme high
or low registers, which may call for so many ledger lines as not to be
accomodated by even one octave's shift. A double octave shift may then
be called for, the sign for which is the numeral 16. <RET>
Use of ledger lines, the ottava sign, and the double ottava sign,
is determined by common sense and custom. Flutists, for instance,
are trained to read notes which extend by five ledger lines above
the treble staff, while violinists, by training, are generally
more comfortable reading these same notes under an ottava sign.
In writing music for several instruments, use of these different
procedures is often a matter of what is appropriate for the
particular score format being used. If staves are too crowded
together, with parts extending above and below, the ottava sign
would be called for.
THE GRAND STAFF
Much music, including that for piano, is notated in a format known
as the grand staff, which conjoins the treble and bass staves
with middle C being the common link between them, one ledger line
below the treble staff, and on ledger line above the bass staff.
Specific Pitch Designation
Often, it is necessary to distinguish a specific pitch in a speci-
fic octave range. If you were asked, for instance, to notate "A"
on a grand staff, you would have to know which of the several oc-
tave displacements of "A" was meant. There are several systems for
designating specific pitches. With the advent of digital synthe-
sizers and especially the development of MIDI (Musical Instrument
Digital Interface), one method seems to be emerging as a standard.
It simply labels "Middle C" as C3 and every pitch within the
octave up to the next "C" with this number. Thus the octaves from
middle C to the highest note on the piano are labelled as follows:
Beyond Basic Notes
All of the pitches under discussion so far are known as "basic"
notes. In the chapter on acoustics, we discussed the fact that in
Western music the octave is normally divided into twelve equal
parts. Our pitch notation system enables us to denote 7 of these,
so far - the "basic" notes "A" to "G". Accidentals are used to
indicate the additional 5 pitches on the 5-line staff. Each of the
smallest increments between the twelve notes within an octave is
known as a half-step. A sharp sign (#) increases the frequency of
a basic note by a half-step:
A flat sign (b) decreases the frequency of a basic note by a half-
step:
Accidentals apply to an entire measure in which they are found.
The bar line "erases" the effect of an accidental unless the acci-
dentalized note is tied over the bar-line, or, for clarity's sake,
there has been an accidental just prior to the bar-line, whose
effect is to be cancelled out in the new measure. To cancel the
effect of a sharp or flat, a natural sign is used, which returns
the note to its "basic" frequency ( ).
Sometimes it is necessary to raise or lower a basic note by two
half-steps, in which case a double sharp or double flat is used:
PREFACE
Harry Partch, that quintessentially American individualist
composer, said that true education, as far as he was concerned, is
a matter chiefly compounded of investigation, investigation, and
investigation. That is the underlying premise of MUS 1/Music:
Technique and Imagination.
I assume that we are here to learn something that we didn't
previously know, and that we are here to question and challenge
anything and everything that we think we know. The etymological
roots of the word "education" mean "to draw out," or "to bring
out." The methods used in this course, and the approaches taken to
learning, are intended to be useful throughout one's life in the
continual process of bringing out of one's inner self insights
which range beyond the acquisition of facts. We use the study of
music as a means to acquire critical thinking skills, develop
imagination, and to become incurably curious, inveterate question-
ers, seekers after knowledge and wisdom.
A recurring theme in MUS 1 is the notion that creativity
can be and needs to be fostered and nurtured. Our goals
in this regard are to apply the insights provided for us
in this catalogue of the traits of creativity:
Creativity includes the ability to:
*WONDER, BE CURIOUS
*BE OPEN TO NEW EXPERIENCE, SEE THE FAMILIAR FROM AN UNFAMILIAR
POINT OF VIEW
*CONFRONT COMPLEXITY AND AMBIGUITY WITH INTEREST
*TAKE ADVANTAGE OF ACCIDENTAL EVENTS IN ORDER TO MAKE DESIRABLE
BUT UNSOUGHT DISCOVERIES (CALLED SERENDIPITY)
*MAKE ONE THING OUT OF ANOTHER BY SHIFTING ITS FUNCTIONS
*GENERALIZE IN ORDER TO SEE UNIVERSAL APPLICATIONS OF IDEAS
*SYNTHESIZE AND INTEGRATE, FIND ORDER IN DISORDER
*BE INTENSELY CONSIOUS YET IN TOUCH WITH UNCONSCIOUS SOURCES
*VISUALIZE OR IMAGINE NEW POSSIBILITIES
*BE ANALYTICAL AND CRITICAL
*KNOW ONESELF, AND HAVE THE COURAGE TO BE ONESELF IN THE FACE
OF OPPOSITION
*BE PERSISTENT, WORK HARD FOR LONG PERIODS, IN PURSUIT OF A
GOAL, WITHOUT GUARANTEED RESULTS
(from Art Creates Us Creates Art, Duane Preble,
Canfield Press, San Francisco, 1976, p.10)
This text itself is an attempt to apply some of these principles.
It is most certainly an experiment. While there are numerous
music theory programs on the market, and very useful ones,
especially for computers capable of multi-voice and multitimbral
effects, I know of no text like this one. Its incorporation of
audible examples, interactive exercises and experimental "labora-
tories" excites me, even given the definite limitations of present
sound and memory capabilities. At least we are able to hear some
things, which is more than the usual text provides. And in the
not distant future, with further advance of digital recording and
storage techniques, we will be able to have texts with aural
examples sounding like live orchestral concerts and providing full-
scale experimental and compositional control of sound.
Use this text as a tool. To learn about music. To learn about
learning itself. And take to heart this message, from the closing
lines of T.S. Eliot's Four Quartets:
We shall not cease from exploration
And the end of all our exploring
Will be to arrive where we started
And know the place for the first time.
Music Imagination & Technique by Norman Lowrey
Copyright (C), 1990. All rights reserved.
INTRODUCTION
MIT is a fully functional music instruction program that I developed for
use in an introductory music theory/composition course at Drew University.
Since every student at Drew acquires a computer as part of their matricula-
tion, I have been able to completely do away with a hard-copy text. All the
reading, exercises, and assignments are done on the computer and assignments
are transmitted over E-Mail, evaluated on screen, and returned via E-mail.
Since one of the intentions of creating this program was to have a 100%
software based text, it is essentially self-contained, needing very little
external documentation. It is menu driven, following mostly standard key-
stroke conventions. Its intended audience is high school age or above,
people interested in learning basic principles of music notation and
compositional processes and who have fairly good reading skills.
While the obvious advantages of using the computer for learning about
music (audibility, animation, interaction) make for a livelier approach
than the standard text, the presentation here is not intentionally "enter-
taining." Little has been done to hide the fact that acquiring knowledge
and applying it is rarely easy. This does not mean that you can't have fun
using this program! It's just to state that it is not intended as a computer
game.
=============================================================================
CONTENTS
The files included in the complete MIT program are divided into 4 cate-
gories, which appear as the main menu headings on bootup. These are:
Music Theory, Utilities, Discussion, and Help.
The topics included under Music Theory are:
1. Acoustics
2. Pitch Basics
3. Rhythm
4. Scales & Keys
5. Intervals
6. Analysis
7. Harmony
Each of these topics is a complete "chapter" devoted to the designated
subject, complete with table of contents, thorough discussion, notated
examples, animated illustrations, and drill/practice screens. The infor-
mation is presented beginning with the most basic material and leading to
increasingly more complex matters. Use of the materials does not have to
proceed in a sequential manner, however. Any screen is accessible through
directly entering a screen number while in any given chapter and return-
ing to the main menu for changing chapters.
The Utilities include:
1. Melody Writer
2. Keyboard
3. Pitch Memory
4. SoundForms
5. Editor
We use Melody Writer more than any other utility in the course for which
MIT was written. It is a simple music notation module for the writing of
single-line melodies that can be played back over the PC's internal
speaker or through an external MIDI synthesizer. This is like a word
processor for music, having a variety of editing features, and capable
of printing out fairly high quality music notation.
The Keyboard is accessible from both the main menu and from Melody
Writer. It can be used to turn the computer keyboard into a piano-
style instrument to play real-time tunes and also includes a feature
to accompany the tunes with rapidly arpeggiated chords. Unlike many
computer music "keyboards," I have written the program replacing the
standard key interrupt and thus tones can be sustained for as long as
a key is held down.
Pitch Memory is a little "game" in which you are presented with a random
pitch played by the computer and asked to try to match it as nearly as
possible by adjusting a pitch up and down on the computer.
SoundForms is like a musical Etch-A-Sketch. Up to 6 different screens
of pitch "shapes" can be drawn, with variable rates of playback. These
are treated like sound "objects" that are linkable on a "construction"
screen. I use this utility in the beginning weeks of my course for
students to create musical structures without their having to know
anything about traditional musical notation.
The Editor is also accessible from both the main menu and from Melody
Writer. It is a no-nonsense WordStar style text editor which contains
some extensions for writing Basic type music code to drive an external
4-voice synthesizer which plugs into the printer port. This unit is
still under development and will be available with a future upgrade of
MIT. I use the editor to do direct manipulation of the text files which
constitute the raw data generated by both Melody Writer and SoundForms.
The Discussion topics presently consist of:
1. Sound Pollution
2. The Music of Harry Partch
3. Creativity
4. The Music of Meredith Monk
I am leaving these texts in MIT for general distribution, not because
they directly pertain to the subject of music theory, but because they
are integral parts of the liberal arts orientation of my teaching method-
ology. Other topics may be added to the program as it evolves.
Under the menu heading of Help are:
1. Instructions
2. Print Texts
3. OS Shell
These items should be, I think, self-explanatory.
=============================================================================
INSTALLATION
MIT will run on an IBM or compatible computer with at least 256K of memory.
CGA card required.
On hard drive:
Make a subdirectory with whatever designation you wish (I use MIT).
Copy all files to this subdirectory.
Two floppies:
If you do not have an autoexec.bat file, make one which sets your
path=a:\;b:\. If you already have an autoexec.bat file, make sure
it includes this path statement.
Simply make backup copies of all MIT disks. The disk with the MIT.EXE
file needs to be in your 'A' drive and disks with filenames related to
the selected topic needs to be either in the 'A' drive after booting
into main menu or in the 'B' drive. You can run the program from one
drive, but a lot of disk swapping may be necessary.
=============================================================================
BOOTING
Move to whatever directory contains the MIT.EXE file and type MIT [Return].
Access to all of the other files of the program is gained through the main
menu that appears at this point.
If you want to print the graphics screens of any chapter or the
Melody Writer, you need to load the GRAPHICS.COM file (or equivalent)
that is part of your DOS utilities. If you want Melody Writer screens to
be printed down the length of an 8-1/2" x 11" paper, you will need some
other utility like HIGRAF-L.COM (available in DOS POWERTOOLS by Paul
Somerson, Bantom Books, 1988).
=============================================================================
REGISTRATION
Music Imagination & Technique has been developed over the past three years
and is, as you can tell by the number of disks required, an extensive
program. It is being offered as shareware with no limitations.
If you find the program useful, I encourage you to register your copy.
Your registation will entitle you to some minimal documentation,
including tutorials for the SoundForms and Melody Writer utilities, as
well as free phone support and low-cost upgrades. Your registration will
also help support continued development of this product. I am scheduled
for sabbatical leave from Drew University next fall during which time
I plan a major revision of MIT into a less text-based and more hyper-
text oriented program with development of an inexpensive external poly-
phonic synthesizer.
To register your copy of Music Imagintion & Technique, send a check or
money order for $50.00 to:
Norman Lowrey
37 Intervale Rd.
Boonton, NJ 07005
Phone (201) 316-8142
or:
Norman Lowrey, Chair
Music Department
Drew University
Madison, NJ 07940
Phone (201) 408-3421
=============================================================================
Thunder let loose upon the Void. The Voice of God. And trillions
of years later we are still rocking with the waves of that Big
Bang. For everything is still vibrating and all is vibration, and
in its broadest sense all vibration is a kind of "sound" whose
primacy has been recognized by virtually every culture, including
the Judeo-Christian heritage: "In the beginning was the Word, and
the Word was God..." The identification of the Word, which had to
be articulated through sound, with deity, the creative and
sustaining spirit of God... For other cultures, the primacy of
sound as a basis for existence has been even more emphatic.
Marius Schneider, in an article on "Primitive Music" in The New
Oxford History of Music, affirms that "sound represents the
original substance of the world" as far as the historian of
culture is concerned, and points out that the (East) Indian
tradition emphasizes the "luminous nature of sound" in the
similarity between svar (light) and svara (sound). [quoted in
McClain, Ernest G., The Myth of Invariance, p.7]
India's oldest sacred book, the Rg Veda, not only posits sound as
the "original substance of the world," but as discussed in a study
of the philosophical methodology of the Rg Veda by Antonio de
Nicolas [Four-Dimensional Man], the meanings of existence derived
by our sensorium, the whole sensory apparatus of the body, are
organized primarily on a model of sound. "Rgvedic man (sic) was
enveloped by sound, looked for centers of experience in the
experience of sound, found the model of complete, absolute
instantaneity and communication in sound." [McClain, ibid, p.2]
These same ideas, in different guises, can be found in the sacred
texts and mythologies of Babylon, Egypt, Greece and Palestine, in
the Egyptian Book of the Dead, the Bible, and Plato [ibid.,
p.xi], particularly as sound tuning systems and music are defined
by number, by mathematical relationships. And while, for most
people in the present day in Western culture, sound has lost this
primacy of meaning, there still occur discussions drawing us back
to this fundamental level, as in the scientist/inventor Itzhak
Bentov's book Stalking the Wild Pendulum: On the Mechanics of
Conciousness, in which it is speculated that the "orderly pattern
of atoms in matter" may be "the result of the interaction of some
kind of 'sound waves' in matter." [p.11] He further says that "we
could actually associate our whole reality with sound of one kind
or another because our reality is a vibratory reality, and there
is nothing static in it. Starting with the nucleus of an atom,
which vibrates at enormous rates, the electrons and the molecules
are all associated with characteristic vibratory rates. A most
important aspect of matter is vibratory energy."
And continuing:
"When we think, our brains produce rhythmic electric
currents. With their magnetic components, they
spread out into space at the velocity of light, as
do the electric waves or sounds produced by our
hearts. They all mingle to form enormous interference
patterns, spreading out and away from the planet.
"They are admittedly weak, but they are there. The
more finely our systems are tuned, the clearer a
signal we can pick out of the general noise and
jumble of 'sounds.'
"Our planet itself is producing shock waves in the
plasma that fills the solar system. These shock waves
interact with those caused by other planets and produce
resonances between the planets and the asteroids. In
short, our whole reality is based on one common factor,
and that is periodic change, or sound." [p.23]
It is in the context of the underlying significance of sound that
I wish for us to place our study of all aspects of sound's
emanations, from acoustics to the structure of a symphony or a pop
tune. In our most common understanding of sound as an auditory
phenomenon, sound is, as has been stated, a vibration. But no
simple vibration, this, for oscillographic analyses of a "large
number of musical, linguistic, and environmental sounds...reveal a
previously unrecognized sonic substructure of immense detail that
directly determines the nature of perceived sound." [Cogan, Robert
and Escot, Pozzi, Sonic Design, p.439]
COMPONENTS OF SOUND
In order for sound to exist, as we know it, at least three
mutually-supportive components are necessary. First, something
which vibrates, called the sound source. Second, a medium through
which the vibrations travel from the sound source to the third
element, a receptor capable of responding to the vibrations and
interpreting their significance.
SOURCE:
David Reck, in Music of the Whole Earth, says that "there are a
limited number of ways that sound can be produced, and it is
unmute testimony to the imagination of man that he has discovered
most of them and has used them with astounding variety." [p.61]
He lists the basic ways a sound source may be set into vibration
along with examples of musical instruments from various cultures
that use these methods: "A solid object may be hit (like a log
drum), scraped (like a comb), whirled through the air (like a bull-
roarer), shaken (like a rattle), plucked (like the metal prongs of
an mbira), or rubbed (like a glass harmonica...). A stretched skin
may be beaten, rubbed, or scraped (as in the drums of the world);
or stretched strings may be made to vibrate by plucking them (like
a guitar), by the friction of a bow or stick rubbed across them
(like a fiddle), or by striking them (like a hammer dulcimer).
Reeds set in an enclosed chamber with air forced across or through
them (by breath, bellows, or bag) will vibrate into sound (like a
harmonica, oboe, or bagpipe), as will breath (or air) split across
the edge of a hole and into an enclosed space (like blowing across
the top of a bottle or a flute). Air buzzed through the tightened
lips into a tube also causes sound vibrations (like trumpets,
horns, trombones). And finally, sound can be produced by
electronic means."
MEDIUM:
The medium through which vibrations are normally carried to our
ears is, of course, air. We know, however, of other materials
through which vibrations may be propagated: wood, metal,water....
any substance other than a vacuum. Water, in fact, is a far better
medium than air for transmitting vibrations. Whales can hear their
amazing "songs" over distances of hundreds of miles because of this.
[SONIC ACTIVITY #1: Tie the two ends of an arm's length of thread
to the bottom sides of a common wire coat hanger. Wrap the thread
around index fingers a couple times, leaving enough space between
hands to fit around head. Stick index fingers in ears, lean over
to allow hanger to be suspended, and swing hanger to strike
against any solid object, a desk, for instance. Listen carefully.
{A student in one of my classes once called this activity
"ridiculous." Aside from this failing, by thus labeling the
activity, to hear how incredible this sound actually is, he demon-
strated a deeper ignorance of the processes by which discoveries
are made. The mind at play, often even in seemingly "ridiculous"
or "silly" activities, is in, or borders on, that mode of
receptivity to experience in which previously unnoticed things may
be noticed, and more importantly, in which disparate "facts" may
be drawn together to arrive at unique and creative syntheses.}]
RECEPTOR:
A receptor is properly called a transducer, which is a device for
converting one form of energy to another. The chief receptor by
which we perceive sound is the ear. The ear converts acoustic
energy, the vibrations of air molecules, into electro-chemical
impulses which the brain in turn converts into sonic sense. A
microphone is another example of a sound receptor or transducer.
It functions similarly to the ear, only its electrical impulses
are directed to some other electronic processing device, like a
tape recorder or an amplifier.
Each one of these components, the sound source, the medium, and
the mechanism of the human ear, would provide fit study for a
complete book, but for the moment we want to take a closer look at
the sound source, the origin of the vibrations we eventually "hear."
VIBRATION
By vibration is meant some kind of back and forth motion (oscil-
lation). In order to vibrate, the sound source must be, as it is
known in the jargon of acoustics, an "elastic body." A good and
obvious example of an elastic body is a stretched rubber band.
[SONIC ACTIVITY #2: Try the index-finger-in-the ear trick with a
rubber band, for yet another amazing sound. (See p.8)]
To set the rubber band into vibration, it must be displaced from
its position of rest--usually by pulling it or plucking it. Being
elastic (and here don't confuse the word elastic only with
material like rubber--anything is "elastic" that returns to its
original state after being "disturbed" or displaced), when
released from displacement, its molecules seek to return to their
original place of "rest." But since a certain amount of energy
has been invested in its displacement, it is carried by the force
of momentum, a carrier of that initial energy investment, beyond
its original point of rest until the strictures of its molecular
structure balance the momentum and pull it back toward the
original position. The rubber band thus vibrates back and forth
until the forces of molecular cohesion and friction have complete-
ly absorbed the original energy, converting it into heat and sound.
When the rubber band is displaced and released, as it makes its first vigorous
snap back in the direction of rest, it pushes the air molecules surrounding it
in the direction of its movement, disturbing them, pushing them in fact, to-
gether, compacting them, making what is called a condensation. At the same
time, the air molecules behind the rubber band are dragged along with it, thus
spreading them out, causing a rarefaction. These disturbances in the air occur
with every back-and-forth motion of the rubber band, creating a series of con-
densations and rarefactions which are transmitted in all directions around
the rubber band as molecules of air knock together and pull apart, creating
waves of air. Thus a series of pressure waves is created in the air which
causes our eardrums to vibrate in response, and hence, sound. <RETURN>
FREQUENCY
We know that if we stretch the rubber band tauter, we'll hear a
different sound, which we describe as "higher," although this is
purely a metaphorical term. What we have just done is to make the
molecules of the rubber band seek their original point of rest
with greater intensity, causing the rubber band to vibrate faster.
Frequency is the technical term for rate of vibration. We
subjectively perceive the frequency as pitch, the "highness" or
"lowness" of a sound.
Frequency is measured by the number of times something vibrates
back and forth per some given time unit, which, when measuring
acoustic vibrations, is the second. The proper designation for
the frequency of a sound is cycles per second (abbrviated c.p.s)
or hz (pronounced Hertz, named for a 19th century physicist who
studied the nature of vibration). A cycle is one complete "trip"
of the vibration. This can be illustrated by observing the action
of a swinging pendulum.
The cyclic motion is usually measured in terms of movement from
the state of rest to maximum point of displacement, back through
the resting point to maximum point of displacement on the opposite
side, and back to the resting point. Actually, one may start from
any point and measure to the analogous point in the cycle, and
this is done in examining the phase of one vibration in relation
to another. <RETURN>
If a pen were to be attached to the bottom of the pendulum and a roll of
paper moved from right to left underneath the pen, a picture of the
pendulum's motion could be drawn. <RETURN>
If the distance between the starting line and ending line were to
represent one second, then the frequency here is 1 c.p.s., or 1 Hz. <RET>
If the distance between the starting line and ending line were to
represent one second, then the frequency here is 2 c.p.s., or 2 Hz.
The human ear responds to acoustic frequencies within the general range
of 20 Hz to 20,000 Hz (sometimes abbreviated 20K Hz). Some people can
hear somewhat beyond these ranges, and other animals are capable of
much wider response, which signifies that human perception of reality
is actually very limited.
You can now enter frequencies at the prompt to test your hearing, as
well as the frequency response of your computer. Hit 'E' to enter.
THE OCTAVE
[ <RETURN> to hear sound discussed below, any key to stop ]
The sound you are hearing as you read this is a sweep of frequencies from
20 Hz to 4K Hz and back down to the lowest frequency on a standard piano
(whose frequency you will be asked to calculate in a minute).
Then the frequency jumps by octaves up to 440 Hz, which, since the 1920's
has been used as the international tuning standard.
The distance from one frequency to another can be expressed
as a ratio. When we hear one frequency in relation to another,
their difference (or ratio) produces a perceived difference
in pitch, which is called a musical interval. The term octave
refers to what is perhaps the most important musical interval,
which is a ratio between two frequencies of 2:1.
The word octave is derived from the fact that most common tunes in Western
music employ 7 different pitches (labeled A,B,C,D,E,F,G), within the
frequency ratio of 2:1. The eighth pitch (thus octave) is the same
letter name (in this case "A") at a frequency ratio of 2:1. This ratio is
possibly the only universally common interval, forming the basis for the
world's diverse tuning systems. Since we'll be referring to the octave
again and again, a thorough understanding of this interval is important.
Here are a series of octaves, with their frequencies given. Note that
octaves have a similarity of sound. They are just "higher" or "lower".
They all form 2:1 ratios. <RETURN>
OCTAVE PRACTICE AND PROBLEMS 1 & 2
You may again enter frequencies as you did on page 16. This time focus
on listening to frequencies with rations of 2:1. ['E' TO ENTER,
'+' FOR NEXT PAGE, '-' FOR PRIOR. HIT 'R' TO GOTO TO PROBLEM #1].
ENTER FREQUENCY (no commas):
EQUAL TEMPERAMENT [ <RETURN> to hear sound mentioned below ]
As you read this sentence, the 7 basic pitches deployed within the
octave, which are common to Western music, are being sounded. You
have no doubt heard this scale before, and its particular pattern
will be described in greater detail later. An incredible variety
of music has been produced using only these pitches, but by adding
just a few more the possibilities for pitch combinations increase
geometrically. 5 additional pitches are common to Western music.
A tuning of pitches within an octave has been developed which is
called equal temperament. A temperament is another name for a way
of relating one pitch to another within an octave. Any equal
temperament is a tuning in which an octave is divided into equal
increments. The temperament of Western European music, which was
only codified within the past 300 years, divides the octave into
12 equal parts. To hear these 12 increments, hit 'H'.
What you have just heard is a pattern of intervals, each one of
which is called a half step. Whereas the frequency ratio of the
octave is 2:1, the ratio of a half step is much smaller. It can
be calculated as the 12th root of 2, which results in an
irrational number (1.0594531...). The starting frequency of the
series just sounded was 220 Hz. Each successive pitch was arrived
at by multiplying the previous pitch by this small amount.
If you consider that our hearing range is from 20 to 20K Hz, and
that this represents a little over 10 octaves, you can calculate
that 12-tone equal temperament makes available somewhat more than
120 distinctly different frequencies. It has been calculated that
"...the total of perceptible pitches...within our hearing
range...is about 1400..." [Cogan & Escott,p.442] We therefore are
capable of hearing many more pitches than we are used to hearing
in the music we are familiar with. Many cultures other than our
own employ much finer gradations of pitch difference. This is
true especially in the Orient, India, and Southeast Asia.
At least from the time of Pythagorus, scientists and musicians
have experimented with many tuning systems. We will hear later on
some of the music of Harry Partch, a composer in our own century
who used a tuning which employed up to 43 increments within an
octave. Just within the past year, the latest synthesizers
(including the "industry standard" Yamaha DX7II) have added the
feature of alternative tuning systems, so you will no doubt be
hearing more music containing smaller divisions of the octave.
You can now experiment with hearing varying divisions of the
octave. The standard way of dealing with small pitch differences
is to divide up the half step into 100 parts, each one of which is
called a cent. At the prompt, you can enter the number of cents,
from 1 to 100, you wish to hear as the smallest frequency change.
[ 'E' TO ENTER, <RETURN>. TO HEAR 1ST TIME, ALT/P TO HEAR AGAIN.
ENTER AS MANY TIMES AS YOU WANT WITH 'E'-<RETURN> SEQUENCE. ]
Frequency changes, along with varying lengths of tones, are the basis
for melody, which is the foundation of Western music. One final issue
regarding frequency before moving on:
A musical tone may be properly conceived of as vibration which is
periodic, that is, whose frequencies are consistent. If fluctuation
of frequency is very rapid, the vibration is aperiodic and results in
noise. Noise is technically defined as erratic, intermittent, or
statistically random oscillation. For an example, hit <RETURN>.
AMPLITUDE
The amount of energy invested into setting the vibrating source
into oscillation determines how loud a sound is. Loudness is how
we subjectively perceive amplitude. The farther something is
displaced from its resting state, the greater is the amplitude.
Whereas frequency is represented along the horizontal axis , as we
saw with the representation of a pendulum's swing, amplitude is
measured along the vertical axis. The farther out the swing, the
greater the displacement, thus the louder the sound.
Another term used to describe loudness is intensity. The common meas-
urement of intensity is the decibel. The decibel is a logarithmic num-
ber in which 0 decibels represents the threshold of hearing (for a
frequency of 1000 Hz). Every 3 decibels represents a doubling of
perceived sound intensity; every 10 decibels represents an increase of
pressure by a factor of 10. From 0 to around 40 db (as the decibel is
abbreviated), sounds are very quiet indeed. It is not until a sound
reaches this level that it said to cross the threshold of intelligibil-
ity. Sounds approaching and louder than 120 db achieve the distinction
of crossing the threshold of pain. The difference between 0 and 120
decibels is a one trillion (1,000,000,000,000) increase in intensity.
Since there is no programable control of loudness on this
computer, we cannot demonstrate changes in amplitude with good
aural examples. You may have noticed, however, that during the
frequency "sweep" in the unit on pitch, some frequencies seemed
louder than others. This is due to the whole computer vibrating
along with its tiny speaker, with some components (like side wall)
having particular frequencies more likely to vibrate than others.
This computer has its own resonance frequencies, and you'll have
the chance later on to try to find out what they are. For the
moment, we'll rely upon this phenomenon to illustrate differences
in loudness. By holding down the UP/DOWN arrows, you can sweep up
and down from approximately 50 to 4000 Hz. Listen for subtle varia-
tions in loudness. These are actually the result of the original
vibrations in the speaker being reinforced and not by any actual
increase of amplitude within the speaker itself. Nonetheless, you
should be able to hear what probably could be measured as a 3-4
decibel difference overall.
Sound loudness measurements are complex because the human ear
responds to different loudnesses at different frequency ranges and
with differing qualities of sound. Greater pressure is needed in
the extremities of the hearing range in order to produce a given
loudness, than in optimum response ranges of the ears. The
optimal response range is from 1000 to 4000 Hz. "Human hearing is
variable. It is affected, for example, by culture: some Africans
hear sounds that, to 20th-century urban North Americans, are
remarkably soft. In industrialized cultures, the young are able
to hear a wider range of frequencies than the old. We are still
ignorant of the exact roles that culture, habit, and environment
play in affecting human hearing equipment and ability." [Cogan &
Escott, p.442]
Sound Level Reference
0-40 db:.........barely perceptible sounds, not clearly
identified--distant wind sound, for instance
40-50 db:........whispering at 5-10 feet distance
rustling leaves, 10-20 feet away
cat purring, 5 feet distant
50-60 db:........normal conversation, 5-10 feet away
central air conditioning unit in big building
your own footsteps on concrete, hard-soled shoes
60-70 db:........orchestra, moderate passage, 30-40 feet
general din, dining hall
light traffic at 30 feet
70-80 db:........truck traffic, 30 feet
stereo, 3/4 gain, 10 feet
shouting
80-90 db:........lawnmower, 20 feet
loud orchestral passages, 20 feet
90-100 db:.......hammering nails, 5 feet
thunder, half mile to mile
100-110 db:......jackhammer, 10 feet
table saw, 5 feet
110-120 db:......Rock concert, 20-30 feet
airplane taking off, 200 feet
TIMBRE
The diagrams we've seen illustrating the swinging of a pendulum
are pictorial representations of wave forms. The pendulum swing
creates a simple curvilinear form because as it swings out to the
positive side of displacement it slows down until it reaches the
turn-around point, then regains speed as it passes the resting
point, slowing down on the opposite swing, and so on. This simple
motion is similar to the way a piano string or a violin string
vibrate over their overall length. The resultant wave form is a
sine wave. The tone such a vibration would produce is called a
sine tone.
No acoustic vibration is quite this simple, however. The vibration of a
piano string can illustrate this. When it is set in motion, it not only
oscillates back and forth over its whole length, but also vibrates in
parts, each one of which vibrates at its own rate. This occurs because
the molecules of the string are knocking against each other, waves of
kinetic energy are flowing along the string, being reflected back from
the stopped ends of the string, and reinforcing or cancelling each other.
You've seen this happen if you've ever thrown a pebble into a small pool
of water. A circle of waves radiates outward, reflects back off the sides
and crosses on-coming waves. Reinforcement and cancellation of the waves'
energies create an overall wave structure called an interference pattern.
The interference pattern in a vibrating string was discovered at
least as long ago as the 6th century BC, by Pythagorus. At
exactly whole-number (integer) divisions of the string, points of
no energy occur. These are called nodal points. The result is
that the string vibrates over its whole length at the same time as
it is vibrating over half its length, over a third of its length,
over a quarter of its length, and so on. Each of these subsidiary
vibrations occur at frequencies inversely proportionate to the
lengths and with varying amplitude relationships. The vibration
over the whole length is called the fundamental. It is the
frequency of this whole-length vibration which we have been
referring to in describing pitch. If this frequency is 100 Hz
then the half-length vibration of the string is 200 Hz, the third-
length vibration is 300 Hz, the quarter-length vibration is 400
Hz, and so on. ( <RETURN> for illustration ).
These subsidiary vibrations are variously described as overtones,
partials, or harmonics. They are audible, and every acoustic
vibration produces them. We generally do not perceive them
discretely, however, but rather coalesce them into a composite
sound whose quality varies from sound to sound, instrument to
instrument, depending upon their particular configurations. The
pattern of harmonics is different for virtually every sound, and
thus we can distinguish between one kind of sound and another.
This is what is called timbre (pronounced "tam-ber"), or "tone
color." Together with the manner in which a tone is initiated
(called onset), timbre allows us to distinguish whether we are
hearing a flute, oboe, human voice, and so on.
Depending upon overtone constituency, a sound may be described on
a continuum from relatively pure to relatively harsh. The fewer
the audible overtones, the more pure the sound, the greater the
number of partials the harsher the sound. The purest sound would
be a sine tone, and the closest instrumental sound to a sine tone
is a flute in its higher register. Electronic instruments can
produce sine tones within their circuitry such that they appear as
simple pendulum-swing curves on an oscilloscope, but when they are
made audible through a speaker, the membrane of the speaker itself
generates harmonics and thus a tone which is not entirely pure.
The timbre of the tone generated by this computer is not quite
pure and varies, like the amplitude, with different frequency
ranges. We again have no programable way of altering the timbre
here, but with some programming manipulation we can approximate a
variety of tone qualities. Listen first of all for the quality of
the unaltered tone and hear if you can perceive the slightly buzzy
quality indicative of the presence of overtones. Then imagine
that your ears are capable of focusing toward the bottom of the
frequency spectrum being sounded and, like a flashlight beam,
capable of being swept upward. Focus in on successively higher
frequency territories and try to isolate the pitches which cause
the buzzy sound. Do this with each example. With practice, you
should be able to hear discretely varying frequencies within the
framework of what, at first "glance", appears to be a single sound.
There are 7 examples. You may hear them as many times as you wish
by hitting 1-7.
Each of these sounds is distinctly different, and examples 5-7
sweep up and down through the harmonic series in the way you can
imagine doing when listening to a single sound.
Harmonics are so named because of the whole-number relationships
they bear to one another. Some sounds, like a cymbal crash,
generate partials whose relationships are inharmonic, non-whole
number ratios. When sounds produce harmonics (as opposed to in-
harmonics), their ratios are simply calculated. The fundamental
is harmonic number 1. The next highest is number 2, and so on. The
number in the series also indicates proportion. Number 2 is 2x1
(or 2:1). Number 3 is 3x1 (or 3:1). Ratios also may be determined
between any two harmonics. The ratio from harmonics 2 to 3 is 2:3,
and so on. Our pictorial representation of a vibrating string
only illustrated 4 divisions of the string, but actually the
divisions continue on to a theoretical infinity. We are capable
of hearing up to 16 harmonics, but as you can determine that the
proportions become smaller and smaller the higher one goes in the
series, hearing discrete upper partials becomes difficult. The
frequency sweeps in the preceding examples focused on the first 8
harmonics, which are easily perceived.
To facilitate calculation, all of the examples used a fundamental
of 100 Hz. Thus the second harmonic is 200 Hz, the 3rd is 300 Hz,
and to continue on: 4=400 Hz, 5=500 Hz, 6=600 Hz, 7=700 Hz, and
8=800 Hz. Listen now to these frequency relationships with each
pitch sounding for about one second. <RETURN>
When a single note on a piano is struck, these pitches are
component parts of what we normally designate as a single sound.
Some common sense analysis can lead to realization of why our ears
focus in on the fundamental as the indicator of frequency. First
of all, it is the loudest, having the greatest displacement. But
second, within the first 8 harmonics, it is replicated in octaves
3 more times. The second harmonic is an octave above the first
(2:1 ratio, remember!). The 4th harmonic is 2 octaves above the
fundamental, and the 8th harmonic is 3 octaves above the
fundamental. (There is yet one more octave relationship within
the first 8 harmonics--the relationship of the 3rd harmonic to the
6th forms a 2:1 ratio).
PROBLEM 3:
Given a fundamental of 55 Hz (3 octaves below tuning "A"), what
are the successive harmonics up to the 8th? <RETURN>
Wave Form
We have seen an image of a sine wave, with its simple curvilinear
form. If one started superimposing one sine wave upon another,
the resultant reinforcement and cancellation of oscillations would
produce other patterns. This is exactly what the 18th century
mathematician, Joseph Fourier, determined. He devised a theorem
which states that any wave can be written as a unique sum of sine
waves. This principle was employed in the early development of
electronic music, since all that was available to composers at
that time (the 1940's) were sine wave generators. Some
present-day synthesizers use this technique also, in a procedure
called additive synthesis. Many current synthesizers make
available other wave forms and present math theory contends that
"...almost any waveform family will work as the basic alphabet of
a wave language." [Quantum Reality, Nick Herbert, Anchor Books,
1987]
There are four commonly available wave forms. The sine, which
we've seen; a square wave (which is actually what this computer
generates); a sawtooth wave; and a triangle wave. <RETURN> to see
how these would appear as oscillographic images.
Sine wave additive synthesis is capable of producing virtually any
sound. With the added capability provided by other waveforms,
composers now have entire orchestras of sound contained in desktop
instruments no larger than the clavichords of the 17th century.
ENVELOPE
All sounds vary, to greater or lesser degree, with the passage of
time. The description of the overall changes in a sound over time
is called the sound's envelope. While timbre or frequency may
change over time and thus may be described in terms of timbral or
frequency envelope, the most usual application of the term is to
amplitude. Amplitude envelope is a description of the manner in
which a sound's loudness changes with the passage of time.
There are four standard components of amplitude envelope: Attack,
Decay, Sustain, and Release (often seen abbreviated as ADSR).
Attack is the amount of time it takes for a sound to reach its
maximum loudness level from the instant of its initiation. Decay
is the time it may take to level off to its Sustain level (if it
is cabable of sustaining). And release is the time it takes for
the sound to die out once "turned off." Some sounds bypass the
sustain segment, and the release cycle is equivalent to the decay.
A piano, for instance, whose envelope is called a "bell" shape
(since it has the same configuration as a struck bell), has a
nearly instantaneous attack cycle with gradual tapering off of the
sound (decay) until reaching 0 loudness. We cannot illustrate
amplitude envelope with this computer, but as various pictorial
representations appear, you will hear the envelope traced by the
analogous frequency changes. <RETURN>
Conclusion
This discussion has been somewhat technical. The greater our un-
standing of the nature of sound, the better we can hear. I can
think of no better way to enhance our hearing capabilities than to
develop the ability to hear harmonics. Used as an exercise,
focusing in on harmonics hones our hearing equipment to a fine
degree. Knowledge of other components of sound, like frequency,
amplitude, and envelope, also enhances our hearing acuity by
making it possible to focus on details of sound. The more detail
we can take in, the richer is our experience. In depth exploration
may also stimulate curiosity--the more we know, the more questions
we are able to ask. The more questions we ask, the more we come
to know. The more we know, the more questions....
There is, furthermore, a broader perspective here. For it is in
the rather mysterious property of vibrating things to vibrate in
complex ways, yet having simple harmonic proportions, that there
exists a model for the relationship of the heavenly bodies one to
another. This in fact led Pythagorus and others to devise
theories about the so-called Music or Harmony of the Spheres.
These theories manifest themselves in many ways, among the latest
being Superstring theory, in which reality is conceived to consist
of ten dimensions and "...the fundamental building blocks of
matter and energy aren't infinitesimal points but infinitesimal
strings." And "it's at this ultimate smallness that everything
exists as the dance of one-dimensional strings in a
ten-dimensional universe. A string vibrating and twitching in a
specific fashion might manifest itself in the real world as a
quark. Another string, shaking and rolling in a different
fashion, might appear as an electron, or a photon, or one of the
many other creatures of the subatomic bestiary. The strings are
the same; only the modes of vibration change." [Gary Taubes,
"Everything's Now Tied To Strings," Discover, November, 1986]
Also, in sound we have a representation of the idea of the One and
the Many. A single sound is all one sound, a whole, perceived as a
unified entity; but at the same time it is many-voiced, a
composite of an infinite number of vibrations, each unique and
discrete. Hence, when you listen to sound, listen carefully.
Listen not only to what is perceived at first as the fundamental
pitch, but listen also for the many voices, the overtones, which
coalesce into forming the overall sound. Listen with ears attuned
to the harmonic structure of the universe, and maybe you will come
to understand the importance of sound in the Rg Veda, sound as the
basis for all else, the primal vibration upon which the whole
phenomenological world floats, sound as Bentov suggested, giving
structure to the very atoms of which we are made.
Music Imagination & Technique by Norman Lowrey
Copyright (C) 1990. All rights reserved.
Detailed Program Description:
Music Imagination & Technique is a 100% software based course of instruc-
tion in the fundamentals of music theory. It is run from a main menu which
accesses four categories of "discourse:" Music Theory, Utilities, Discussion,
and Help.
The topics included under Music Theory are:
1. Acoustics
2. Pitch Basics
3. Rhythm
4. Scales & Keys
5. Intervals
6. Analysis
7. Harmony
Each of these topics is a complete "chapter" devoted to the designated
subject, complete with table of contents, thorough discussion, notated
examples, animated illustrations, and drill/practice screens. The infor-
mation is presented beginning with the most basic material and leading to
increasingly more complex matters. Use of the materials does not have to
proceed in a sequential manner, however. Any screen is accessible through
directly entering a screen number while in any given chapter and return-
ing to the main menu for changing chapters.
The Utilities include:
1. Melody Writer
2. Keyboard
3. Pitch Memory
4. SoundForms
5. Editor
We use Melody Writer more than any other utility in the course for which
MIT was written. It is a simple music notation module for the writing of
single-line melodies that can be played back over the PC's internal
speaker or through an external MIDI synthesizer. This is like a word
processor for music, having a variety of editing features, and capable
of printing out fairly high quality music notation.
The Keyboard is accessible from both the main menu and from Melody
Writer. It can be used to turn the computer keyboard into a piano-
style instrument to play real-time tunes and also includes a feature
to accompany the tunes with rapidly arpeggiated chords. Unlike many
computer music "keyboards," I have written the program replacing the
standard key interrupt and thus tones can be sustained for as long as
a key is held down.
Pitch Memory is a little "game" in which you are presented with a random
pitch played by the computer and asked to try to match it as nearly as
possible by adjusting a pitch up and down on the computer.
SoundForms is like a musical Etch-A-Sketch. Up to 6 different screens
of pitch "shapes" can be drawn, with variable rates of playback. These
are treated like sound "objects" that are linkable on a "construction"
screen. I use this utility in the beginning weeks of my course for
students to create musical structures without their having to know
anything about traditional musical notation.
The Editor is also accessible from both the main menu and from Melody
Writer. It is a no-nonsense WordStar style text editor which contains
some extensions for writing Basic type music code to drive an external
4-voice synthesizer which plugs into the printer port. This unit is
still under development and will be available with a future upgrade of
MIT. I use the editor to do direct manipulation of the text files which
constitute the raw data generated by both Melody Writer and SoundForms.
The Discussion topics presently consist of:
1. Sound Pollution
2. The Music of Harry Partch
3. Creativity
4. The Music of Meredith Monk
I am leaving these texts in MIT for general distribution, not because
they directly pertain to the subject of music theory, but because they
are integral parts of the liberal arts orientation of my teaching method-
ology. Other topics may be added to the program as it evolves.
Under the menu heading of Help are:
1. Instructions
2. Print Texts
3. OS Shell
These items should be, I think, self-explanatory.
=============================================================================
Unique features of the program:
MIT includes complete and thorough information pertaining to the funda-
mentals of music theory and is not just a "drill and practice" program.
Also, it contains several novel utilities and instructional tools to
make learning interesting and challenging.
=============================================================================
Program's capacity or limitations:
MIT is fully functional even without registration.
=============================================================================
Does your program require any special system requirements:
CGA card required.
=============================================================================
How to start the program:
Disk 1 in active drive: type MIT at prompt.
=============================================================================
What is the registration fee for your program:
$50.00
=============================================================================
Materials or services that come with registration:
Brief documentation, tutorials, free phone support, low-cost upgrades.
=============================================================================
List of program files and one-line description of each file:
Disk 1:
README.DOC - General description, installation instructions, booting.
MIT.EXE - Main menu system for all other files.
MIT.TXT - Text file for MIT.EXE
PREFACE.TBC - Prefatory discussion of Music Imagination & Technique;
PREFACE.TXT - Text file for PREFACE.TBC
SOUND.MNU - Table of contents for chapter on Acoustics.
SOUND.TBC - Overlay file for chapter on Acoustics.
SOUND.TXT - Text file for SOUND.TBC.
PRELUDE.EXE - Animated text introduction to chapter on Acoustics.
CGA.BGI - Graphics driver for PRELUDE.EXE.
TRIP.CHR - Font for PRELUDE.EXE
PITCH.MNU - Table of contents for chapter on Pitch Basics.
PITCH.TBC - Overlay file for chapter on Pitch Basics.
PITCH.TXT - Text file for PITCH.TBC.
SUNNY.SEQ - Example file for Melody Writer.
EXAMPLE.SNX - Example file for SoundForms.
Disk 2:
RHYTHM.MNU - Table of contents for chapter on Rhythm.
RHYTHM.TBC - Overlay file for chapter on Rhythm.
RHYTHM.TXT - Text file for RHYTHM.TBC.
SCALES.MNU - Table of contents for chapter on Scales.
SCALES.TBC - Overlay file for chapter on Scales.
SCALES.TXT - Text file for SCALES.TBC.
INTERVALS.MNU - Table of contents for chapter on Intervals.
INTERVALS.TBC - Overlay file for chapter on Intervals.
INTERVALS.TXT - Text file for INTERVALS.TBC
ANALYSIS.TBC - Overlay file for chapter on Analysis.
ANALYSIS.TXT - Text file for ANALYSIS.TBC.
Disk 3:
HARMONY.TBC - Overlay file for chapter on Harmony.
HARMONY.TXT - Text file for HARMONY.TBC.
POLLUTE.TBC - Overlay file for discussion of Sound Pollution.
POLLUTE.TXT - Text file for POLLUTE.TBC.
PARTCH.TBC - Overlay file for discussion of Harry Partch.
PARTCH.TXT - Text file for PARTCH.TBC.
CREATE.TBC - Overlay file for discussion of Creativity.
CREATE.TXT - Text file for CREATE.TBC.
MONK.TBC - Overlay file for discussion of Meredith Monk.
MONK.TXT - Text file for MONK.TBC.
Disk 4:
MW.TBC - Overlay file for Melody Writer.
KEYBOARD.EXE - Keyboard utility.
PITCHMEM.TBC - Overlay file for Pitch Memory utility.
SNDFORM.TBC - Overlay file for SoundForms utility.
SYNTHED.EXE - Text editor utility.
EDITERR.MSG - Error message file for SYNTHED.EXE
Volume in drive A has no label
Directory of A:\
MIT EXE 78504 12-07-90 5:53a
MIT TXT 1583 4-21-90 4:48p
README DOC 7981 12-07-90 5:45a
SUBMISSN DOC 7400 12-09-90 4:38p
PREFACE TBC 9538 8-17-89 2:33p
PREFACE TXT 3866 8-14-89 8:37a
SOUND MNU 1140 8-13-89 12:22p
SOUND TBC 51529 1-22-90 7:48p
SOUND TXT 39494 1-20-90 6:33p
PRELUDE EXE 19968 6-25-88 4:39p
PITCH MNU 1190 7-07-89 9:13a
PITCH TBC 60195 4-01-90 5:09p
PITCH TXT 12804 9-21-89 9:49a
CGA BGI 6029 12-16-87 4:00a
TRIP CHR 7213 12-16-87 4:00a
SUNNY SEQ 8136 2-17-90 3:28p
EXAMPLE SNX 4993 12-09-90 4:06p
GO BAT 38 4-13-93 9:00a
SHOW EXE 2040 9-12-88 10:48a
19 file(s) 323641 bytes
29696 bytes free