WO1997013240A1 - Audio synthesizer - Google Patents

Audio synthesizer Download PDF

Info

Publication number
WO1997013240A1
WO1997013240A1 PCT/GB1996/000331 GB9600331W WO9713240A1 WO 1997013240 A1 WO1997013240 A1 WO 1997013240A1 GB 9600331 W GB9600331 W GB 9600331W WO 9713240 A1 WO9713240 A1 WO 9713240A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
audio samples
synthesizer
control instructions
general purpose
Prior art date
Application number
PCT/GB1996/000331
Other languages
French (fr)
Inventor
Gal Ashour
Yoav Medan
Naftaly Sharir
Original Assignee
International Business Machines Corporation
Ibm United Kingdom Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=10781657&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=WO1997013240(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by International Business Machines Corporation, Ibm United Kingdom Limited filed Critical International Business Machines Corporation
Priority to EP96902371A priority Critical patent/EP0853802B1/en
Priority to DE69608292T priority patent/DE69608292T2/en
Priority to JP9514043A priority patent/JP2970689B2/en
Publication of WO1997013240A1 publication Critical patent/WO1997013240A1/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • G10H1/0041Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
    • G10H1/0058Transmission between separate instruments or between individual components of a musical system
    • G10H1/0066Transmission between separate instruments or between individual components of a musical system using a MIDI interface
    • G10H1/0075Transmission between separate instruments or between individual components of a musical system using a MIDI interface with translation or conversion means for unvailable commands, e.g. special tone colors
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H7/00Instruments in which the tones are synthesised from a data store, e.g. computer organs
    • G10H7/002Instruments in which the tones are synthesised from a data store, e.g. computer organs using a common processing for different operations or calculations, and a set of microinstructions (programme) to control the sequence thereof
    • G10H7/004Instruments in which the tones are synthesised from a data store, e.g. computer organs using a common processing for different operations or calculations, and a set of microinstructions (programme) to control the sequence thereof with one or more auxiliary processor in addition to the main processing unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2230/00General physical, ergonomic or hardware implementation of electrophonic musical tools or instruments, e.g. shape or architecture
    • G10H2230/025Computing or signal processing architecture features
    • G10H2230/041Processor load management, i.e. adaptation or optimization of computational load or data throughput in computationally intensive musical processes to avoid overload artifacts, e.g. by deliberately suppressing less audible or less relevant tones or decreasing their complexity

Definitions

  • the invention relates to audio synthesizers for generating digitally encoded audio samples in response to coded control instructions representing musical events, such as a MIDI data stream.
  • MIDI Musical Instrument Digital Interface
  • the MIDI data specifies performance information, as opposed to sound information. For example, which note or notes are being held down, if any additional pressure is applied to the note after being struck, when the key is released and any other adjustments made to the settings of the instrument.
  • MIDI data is communicated as a serial data stream organised into MIDI 'messages', which contain one MIDI command or event.
  • a MIDI synthesizer is controlled by a stream of MIDI messages.
  • the synthesizer receives and decodes the messages and operates accordingly. For example, a 'NOTE ON' event will cause the synthesizer to generate audio samples that correspond to a requested note and velocity that are supplied as parameters. Similarly, a 'NOTE OFF' event will cause the synthesizer to cease generating the audio samples.
  • Sof ware-only MIDI synthesizers have been proposed for use with general purpose computers. These can take advantage of all the resources of the computer, such as CPU power, memory, magnetic and CD-ROM storage, caching mechanisms and virtual memory and are easily customised, upgraded and maintained. However, they have the disadvantage that the sound quality can be degraded when the computing resources they require are not available due to other tasks the computer may be performing. Nevertheless, in recent years the power available from the processors used in personal computers has increased significantly and, in normal use, the load on the CPU can be quite low for much of the time.
  • This invention is directed to providing an audio synthesizer of the above defined type which combines the advantages of hardware and software implementations.
  • an audio synthesizer for generating an analogue or digital audio output m response to coded control instructions representing musical events, the synthesizer comprising: a general purpose computer portion having a CPU programmed to receive the control instructions and generate audio samples; a special purpose hardware portion for receiving the control instructions and generating the audio samples; a controller for directing the control instructions either so that the general purpose computer portion generates the audio samples or so that the hardware portion generates the audio samples; and means to combine the audio samples generated by the general purpose computer portion and the hardware portion to form an audio output which accords with the control instructions.
  • This provides a hybrid audio synthesizer which combines advantages from both software and hardware implementations and enables the spare computing power of the host CPU to be used to supplement that of the sound card.
  • the audio synthesizer comprises measurement means for repeatedly measuring the load on the CPU and the controller is arranged to direct the control instructions according to the measured load on the CPU.
  • the CPU usage can be measured, for example, by timing the synthesis internal loop, which is the computation intensive loop, although other methods for determining the CPU usage are possible.
  • the hybrid synthesizer can handle some of the MIDI events in software using the native CPU, while propagating some of the events to the hardware portion.
  • the decision as to how many simultaneous notes should be handled in software and how many in hardware is based on the current CPU usage.
  • the general purpose computer can be programmed to act as the controller and to either generate the audio samples itself or to transfer the instructions to the hardware portion.
  • the controller could be implemented as part of the special purpose hardware.
  • the controller is arranged to direct the intructions according to voice-type.
  • the audio synthesizer can be arranged so that a predefined set of voices are handled by the general purpose computer portion and thus the quality of the sound can be improved by directing intructions relating to particular instruments either to the software or the hardware synthesizers according to whether they are best suited to the particular synthesis method used.
  • the audio synthesizer can include delay means for delaying the direction of control instructions to the hardware portion, so that the audio samples generated by the hardware portion are synchronised with the audio samples generated by the general purpose computer.
  • the synthesis on the main CPU can be done using a wave-table synthesis, whilst using a relatively cheap sound card which only supports FM synthesis. In this case, both quality of the produced music and the overall performance is greatly improved.
  • the synthesizer includes an input for receiving audio samples and means for mixing audio samples received at said input with audio samples generated in response to the coded control instructions.
  • the audio synthesizer can take the form of a personal computer with an add-on sound card, although, of course, many other configurations are possible.
  • a method of operating an audio synthesizer to generate an analogue or digital audio output in response to coded control instructions representing musical events comprising: a general purpose computer portion having a CPU programmed to receive the control instructions and generate audio samples; and a special purpose hardware portion for receiving the control instructions and generating the audio samples, the method comprising directing the control instructions either so that the general purpose computer portion generates the audio samples or so that the hardware portion generates the audio samples so that the audio samples generated by the general purpose computer portion and the hardware portion can be combined to form an audio output which accords with the control instructions.
  • the audio samples can be combined to accord with the control instructions by delaying the direction of control instructions to the hardware portion, so that the audio samples generated by the hardware portion are synchronised with the audio samples generated by the general purpose computer.
  • the method comprises receiving a NOTE ON instruction; determining whether the NOTE ON instruction is for a predefined set of voices to be handled by the general purpose computer portion and, if not, directing the NOTE ON instruction so that the hardware portion generates corresponding audio samples; measuring the CPU load and if the CPU load exceeds a predefined threshold, directing the
  • the method further comprises: receiving a NOTE
  • the invention can be implemented in the form of an article of manufacture comprising a computer usable medium in which program code is embodied for causing a computer to perform the above described methods.
  • Fig 1 is a schematic diagram showing a personal computer
  • Fig 2 is a simplified schematic functional block diagram of a sound card
  • Fig 3 is a schematic diagram showing the software structure of an audio synthesizer
  • Fig 4 is a flow diagram showing the synthesis loop of a software synthesizer
  • Fig 5 is a flow diagram showing the process on receipt of a NOTE ON MIDI message
  • Fig 6 is a flow diagram showing the process on receipt of a NOTE OFF MIDI message.
  • Fig 1 is a schematic diagram showing a personal computer arranged to function as an audio synthesizer.
  • the computer comprises conventional components such as display device 100 and associated display adapter 110, and I/O interface 120 to which is attached a keyboard and a mouse.
  • the computer also comprises a CPU 130, RAM 140 and a magnetic storage device 150. These components are arranged to intercommunicate via a bus 160 in conventional manner.
  • the computer also comprises an audio adapter 160 which is capable of implementing a MIDI synthesizer by utilising a digital signal processor. Audio adapter 160 is shown connected one or more loudspeakers indicated at 170.
  • the system shown in Fig 1 may be implemented by using an IBM PS/2 computer available from IBM Corporation and a SoundBlaster 16 Value Edition card available from Creative Labs Ine (IBM and PS/2 are trademarks of IBM Corporation and SoundBlaster is a trade mark of Creative Technology Ine) .
  • IBM PS/2 available from IBM Corporation
  • SoundBlaster 16 Value Edition card available from Creative Labs Ine (IBM and PS/2 are trademarks of IBM Corporation and SoundBlaster is a trade mark of Creative Technology Ine) .
  • Fig 2 shows a simplified functional block diagram of the relevant parts of the adapter. It comprises bus interface logic 200, FM synthesizer 210, digital to analog converter 230 and audio amplifiers and mixers 240.
  • synthesizer 210 could be any other kind of synthesizer, eg a wavetable or waveguide synthesizer.
  • the card has an audio output indicated at 250, though the audio samples might equally be output in digital form for digital recording or processing via an external D/A converter. The structure and general operation of such a card will be well known to those skilled in the art and will not be discussed further herein.
  • the card is capable of simultaneously accepting and combining both MIDI events and wave samples sent from the computer.
  • MIDI data is communicated as a serial data stream organised into
  • MIDI 'messages' which contain one MIDI command or event.
  • MIDI commands are usually composed of one, two, or three bytes of data arranged and transmitted one after another.
  • the first byte sent in each command is called the 'status' byte and specifies an operation to be performed.
  • the next one or two bytes, if used, represent parameters of this command.
  • a NOTE ON command comprises three bytes, the first of which is the status byte.
  • This byte tells a synthesizer to play a note and specifies the channel number.
  • the channel number usually represents the type of sound to be played, ie which instrument of the synthesizer is to be used.
  • the second byte specifies the note to be played and the third byte specifies the velocity value for the note.
  • the bytes of the MIDI commands are specified in the MIDI standard.
  • a MIDI synthesizer can be controlled by a MIDI sequencer in the following way.
  • a standard MIDI file (SMF) contains a set of events, which are intended to be executed by a synthesizer at particular times. Generally, the events are not uniformly spaced in time.
  • a conventional MIDI sequencer parses the standard MIDI file, reads the present MIDI event and the time difference between it and the next event. The sequencer then sends the event m a MIDI message to a MIDI synthesizer at the time it is to be executed. The sequencer usually sets a timer and reads the next MIDI event after this time difference has elapsed.
  • a conventional MIDI synthesizer receives the MIDI message that the sequencer sends, decodes the message and operates accordingly. For example, a 'NOTE ON' event will cause the synthesizer to generate audio samples that correspond to a requested note and velocity that are supplied as parameters. Similarly, a 'NOTE OFF' event will cause the synthesizer to cease generating the audio samples.
  • DLL Dynamic Link Library
  • FIG. 3 A schematic diagram of the software components m the present implementation is shown in Fig 3.
  • the system comprises a MIDI sequencer application layer 300, an operating system layer 310, a DLL layer indicated at 320 and a sound card layer 330.
  • DLL layer 320 comprises 3 main blocks, controller 335, high level block 340 and low level block 350.
  • Sound card layer 330 comprises a sound card device driver 360 and the hardware resources of the sound card which are indicated at 370.
  • Also shown in Fig 3 is an output buffer 380.
  • the DLL layer 320 is triggered by MIDI messages which are sent by MIDI sequencer. These messages can be sent directly by application 300 or via the operating system services.
  • Controller 335 is responsible for the MIDI logic management. In other words, this block decides which notes are to be propagated to the sound card and which should be handled entirely by means of software synthesis. NOTE ON and NOTE OFF messages for notes to be handled on the sound card are sent, after a suitable delay to maintain synchronisation, directly to the sound card via line 345. Other messages are sent to both the sound card and the software synthesizer as will be described in more detail below.
  • the controller 335 is arranged to dynamically reduce the number of soft-voices, ie the number of notes that are currently synthesized by means of software, down to a limit of zero by directing any further NOTE ON messages to the sound card.
  • High level block 340 receives tne messages which are to be handled in software and updates the various internal tables which are used by the synthesizer. It handles conventional functions of a MIDI synthesizer such as managing voice allocation etc which are not directly relevant to the present invention and will be well understood by those skilled in the art.
  • Low level block 350 is triggered periodically and contains the synthesis engine which requires intensive computation. This block is responsible for the synthesis of notes that are not propagated to the sound card by controller 335.
  • the general operation of the low level block is shown in Fig 4.
  • the loop is generally triggered with a time period of between lmS and 5mS. Each time the loop is triggered it checks in step 400 for each voice which kind of note is to be produced and looks in the Wave Table database to find the appropriate waveform. The waveform for each active voice is then transformed if necessary to the correct pitch and the waveforms combined to generate output samples - step 410 - at a suitable sampling rate, for example 44100 samples per second, and places them in buffer 380 for retrieval by the device driver 360 in sound card layer 330.
  • a suitable sampling rate for example 44100 samples per second
  • the synthesis algorithm Since the synthesis algorithm is computationally very intensive, it can be used for monitoring the CPU usage. This measure of the CPU usage is available to controller 335 to enable it to decide whether notes should be propagated to the sound card.
  • the CPU usage can be monitored via the synthesis process. For example, a time measurement can be taken each time the low-level block 340 is triggered. If the difference between the current time and the time that was measured on the previous iteration is consistently larger than the requested period, then it can be assumed that the CPU is overloaded.
  • a time measurement can be taken at the beginning and it the end of the synthesis process m each iteration. The difference between these two measurements is approximated to be the CPU time which is used for the current number of soft voices. The CPU usage in percentage is then calculated as the ratio of the measured synthesis time and the time interval which is used as the triggering period for the low level block 350.
  • Controller 335 directs all MIDI messages, apart from NOTE ON and NOTE OFF messages to both the sound card and the h gh level block 340.
  • NOTE ON and NOTE OFF messages the processes shown in Figs 5 and 6 respectively are performed.
  • a determination of the CPU load is made in step 510 and if the CPU load is less than a fixed limit the NOTE ON message is directed to high level block 340 and the note is handled in software - step 520. If the CPU is greater than 50% loaded or the instrument to which the note relates can be adequately handled by the sound card, then the NOTE ON command is directed to the sound card - step 530.
  • step 600 in Fig 6 On receipt of a NOTE OFF command, it is determined, in step 600 in Fig 6, whether the voice concerned is presently being handled in software. This information is available in the tables maintained by high level block 340. If the note is being handled in software, then the NOTE OFF message is directed to high level block 340 - step 610, if not then the NOTE OFF message is directed to the sound card - step 620.
  • the present implementation has been written to work under the well known Windows operating system from Microsoft Corp (Windows is a trademark of Microsoft Corp.).
  • the general operation of the system is as follows. First the user invokes a Windows application, such as a sequencer, which requires a MIDI synthesizer.
  • the application opens the MIDI synthesizer DLL via the MIDI Mapper program, with the intervention of the MMSYSTEM driver.
  • the software MIDI Synthesizer DLL opens the sound card twice - once as a MIDI device to which MIDI messages will be propagated and once as a WAVE device to which the wave samples produced by the software MIDI synthesizer will be sent.
  • the software synthesizer Since the software synthesizer locks the WAVE input of the sound card, it is itself provided with a WAVE input to receive WAVE messages from other sources and handle the mixing of these with the wave data it generates.
  • MIDI messages for instruments for which the sound quality is acceptable using FM synthesis are propagated to the sound card and handled in hardware.
  • the software MIDI synthesizer handles the MIDI messages for instruments which require Wave table synthesis up to a fixed limit, such as 50%, of the available CPU power. In case of overload, the software synthesizer dynamically reduces the number of soft voices, ie the number of notes that are currently synthesized by means of software down to a limit of zero. As described above, this is achieved by directing any further NOTE ON messages to the sound card.
  • the present implementation takes the form of a computer program and can be distributed in the form of an article of manufacture comprising a computer usable medium in which suitable program code is embodied for causing a computer to perform the function of controller 335 described above.
  • the program may include the high level block 340 and 350 and low level blocks of the software synthesizer or could be implemented for use with a preexisting software synthesizer.
  • the controller might be implemented on the sound card itself, either in hardware or software, and arranged to receive MIDI messages from an external source, the sound card handling some notes itself and directing some notes over the PC bus to a software synthesizer.
  • the invention is applicable in the technical field of computers and digital audio systems.

Abstract

An audio synthesizer is disclosed for generating an analogue or digital audio output in response to coded control instructions representing musical events, such as a MIDI data stream. The synthesizer comprises: a general purpose computer portion (340, 350) having a CPU programmed to receive the control instructions and generate audio samples; a special purpose hardware portion (330) for receiving the control instructions and generating the audio samples; a controller (335) for directing the control instructions either so that the general purpose computer portion (340, 350) generates the audio samples or so that the hardware portion (330) generates the audio samples; and means to combine the audio samples generated by the general purpose computer portion (340, 350) and the hardware portion (330) to form an audio output which accords with the control instructions.

Description

AUDIO SYNTHESIZER
TECHNICAL FIELD
The invention relates to audio synthesizers for generating digitally encoded audio samples in response to coded control instructions representing musical events, such as a MIDI data stream.
BACKGROUND ART
Currently, most computers use sound cards or other dedicated hardware in order to produce music, for instance for games, composition and other multimedia applications. Typically, a dedicated sound card is plugged in the motherboard of the computer, although some recent personal computers have dedicated sound hardware located on the motherboard itself.
The main advantage of such hardware implementations is that even when the CPU of the computer is loaded to the extent of its power, the sound quality is not degraded.
However, an implementation entirely in hardware also has certain disadvantages. For example, the resources of the sound card are only used for the sound generation process and the machine cannot benefit from them in any other way. Software improvements on the computer, for example to the operating system, do not improve the sound quality or the performance since the sound card behaves as a semi independent entity.
MIDI (Musical Instrument Digital Interface) is an internationally recognised specification for data communication between digital electronic musical instruments and other devices, such as computers, lighting controllers, mixers or the like. The MIDI data specifies performance information, as opposed to sound information. For example, which note or notes are being held down, if any additional pressure is applied to the note after being struck, when the key is released and any other adjustments made to the settings of the instrument. MIDI data is communicated as a serial data stream organised into MIDI 'messages', which contain one MIDI command or event.
In a conventional MIDI playback system, a MIDI synthesizer is controlled by a stream of MIDI messages. The synthesizer receives and decodes the messages and operates accordingly. For example, a 'NOTE ON' event will cause the synthesizer to generate audio samples that correspond to a requested note and velocity that are supplied as parameters. Similarly, a 'NOTE OFF' event will cause the synthesizer to cease generating the audio samples.
Most commercially available sound cards have the capability of acting as MIDI synthesizers by receiving data from MIDI sources either though a MIDI port or via the PC bus.
Sof ware-only MIDI synthesizers have been proposed for use with general purpose computers. These can take advantage of all the resources of the computer, such as CPU power, memory, magnetic and CD-ROM storage, caching mechanisms and virtual memory and are easily customised, upgraded and maintained. However, they have the disadvantage that the sound quality can be degraded when the computing resources they require are not available due to other tasks the computer may be performing. Nevertheless, in recent years the power available from the processors used in personal computers has increased significantly and, in normal use, the load on the CPU can be quite low for much of the time.
DISCLOSURE OF THE INVENTION
This invention is directed to providing an audio synthesizer of the above defined type which combines the advantages of hardware and software implementations.
To achieve this aim, there is now provided an audio synthesizer for generating an analogue or digital audio output m response to coded control instructions representing musical events, the synthesizer comprising: a general purpose computer portion having a CPU programmed to receive the control instructions and generate audio samples; a special purpose hardware portion for receiving the control instructions and generating the audio samples; a controller for directing the control instructions either so that the general purpose computer portion generates the audio samples or so that the hardware portion generates the audio samples; and means to combine the audio samples generated by the general purpose computer portion and the hardware portion to form an audio output which accords with the control instructions.
This provides a hybrid audio synthesizer which combines advantages from both software and hardware implementations and enables the spare computing power of the host CPU to be used to supplement that of the sound card.
An embodiment of such a synthesizer has been developed for use with a MIDI data stream, however application of analogous techniques to other forms of control instructions is not excluded. In a preferred embodiment the audio synthesizer comprises measurement means for repeatedly measuring the load on the CPU and the controller is arranged to direct the control instructions according to the measured load on the CPU. The CPU usage can be measured, for example, by timing the synthesis internal loop, which is the computation intensive loop, although other methods for determining the CPU usage are possible.
In this case, when the CPU is not highly loaded, the hybrid synthesizer can handle some of the MIDI events in software using the native CPU, while propagating some of the events to the hardware portion. The decision as to how many simultaneous notes should be handled in software and how many in hardware is based on the current CPU usage.
The general purpose computer can be programmed to act as the controller and to either generate the audio samples itself or to transfer the instructions to the hardware portion. Alternatively, the controller could be implemented as part of the special purpose hardware.
Advantageously, the controller is arranged to direct the intructions according to voice-type. In this way, the audio synthesizer can be arranged so that a predefined set of voices are handled by the general purpose computer portion and thus the quality of the sound can be improved by directing intructions relating to particular instruments either to the software or the hardware synthesizers according to whether they are best suited to the particular synthesis method used.
The audio synthesizer can include delay means for delaying the direction of control instructions to the hardware portion, so that the audio samples generated by the hardware portion are synchronised with the audio samples generated by the general purpose computer.
In one particularly preferred embodiment, the synthesis on the main CPU can be done using a wave-table synthesis, whilst using a relatively cheap sound card which only supports FM synthesis. In this case, both quality of the produced music and the overall performance is greatly improved.
Preferably, the synthesizer includes an input for receiving audio samples and means for mixing audio samples received at said input with audio samples generated in response to the coded control instructions.
This allows wave messages to be received from other sources and mixed with the wave data generates by the synthesizer. In one implementation, the audio synthesizer can take the form of a personal computer with an add-on sound card, although, of course, many other configurations are possible.
Viewing the invention from a second aspect, there is provided a method of operating an audio synthesizer to generate an analogue or digital audio output in response to coded control instructions representing musical events, the synthesizer comprising: a general purpose computer portion having a CPU programmed to receive the control instructions and generate audio samples; and a special purpose hardware portion for receiving the control instructions and generating the audio samples, the method comprising directing the control instructions either so that the general purpose computer portion generates the audio samples or so that the hardware portion generates the audio samples so that the audio samples generated by the general purpose computer portion and the hardware portion can be combined to form an audio output which accords with the control instructions.
Suitably, it can be arranged that the audio samples can be combined to accord with the control instructions by delaying the direction of control instructions to the hardware portion, so that the audio samples generated by the hardware portion are synchronised with the audio samples generated by the general purpose computer.
In a preferred embodiment the method comprises receiving a NOTE ON instruction; determining whether the NOTE ON instruction is for a predefined set of voices to be handled by the general purpose computer portion and, if not, directing the NOTE ON instruction so that the hardware portion generates corresponding audio samples; measuring the CPU load and if the CPU load exceeds a predefined threshold, directing the
NOTE ON instruction so that the hardware portion generates corresponding audio samples; otherwise directing the NOTE ON instruction so that the general purpose computer portion generates the audio samples.
in this embodiment the method further comprises: receiving a NOTE
OFF instruction; determining whether the NOTE OFF instruction relates to a note being handled by the general purpose computer portion and, if not, directing the NOTE OFF instruction to the hardware portion; otherwise directing the NOTE OFF instruction to the general purpose computer portion.
The invention can be implemented in the form of an article of manufacture comprising a computer usable medium in which program code is embodied for causing a computer to perform the above described methods. BRIEF DESCRIPTION OF DRAWINGS
An embodiment of the invention will now be described by way of example only with reference to the accompanying drawings, wherem: Fig 1 is a schematic diagram showing a personal computer;
Fig 2 is a simplified schematic functional block diagram of a sound card;
Fig 3 is a schematic diagram showing the software structure of an audio synthesizer; Fig 4 is a flow diagram showing the synthesis loop of a software synthesizer;
Fig 5 is a flow diagram showing the process on receipt of a NOTE ON MIDI message;
Fig 6 is a flow diagram showing the process on receipt of a NOTE OFF MIDI message.
BEST MODE(S) FOR CARRYING OUT THE INVENTION
Fig 1 is a schematic diagram showing a personal computer arranged to function as an audio synthesizer. The computer comprises conventional components such as display device 100 and associated display adapter 110, and I/O interface 120 to which is attached a keyboard and a mouse. The computer also comprises a CPU 130, RAM 140 and a magnetic storage device 150. These components are arranged to intercommunicate via a bus 160 in conventional manner.
The computer also comprises an audio adapter 160 which is capable of implementing a MIDI synthesizer by utilising a digital signal processor. Audio adapter 160 is shown connected one or more loudspeakers indicated at 170.
The system shown in Fig 1 may be implemented by using an IBM PS/2 computer available from IBM Corporation and a SoundBlaster 16 Value Edition card available from Creative Labs Ine (IBM and PS/2 are trademarks of IBM Corporation and SoundBlaster is a trade mark of Creative Technology Ine) .
Fig 2 shows a simplified functional block diagram of the relevant parts of the adapter. It comprises bus interface logic 200, FM synthesizer 210, digital to analog converter 230 and audio amplifiers and mixers 240. Of course, in other embodiments synthesizer 210 could be any other kind of synthesizer, eg a wavetable or waveguide synthesizer. The card has an audio output indicated at 250, though the audio samples might equally be output in digital form for digital recording or processing via an external D/A converter. The structure and general operation of such a card will be well known to those skilled in the art and will not be discussed further herein. The card is capable of simultaneously accepting and combining both MIDI events and wave samples sent from the computer.
MIDI data is communicated as a serial data stream organised into
MIDI 'messages', which contain one MIDI command or event. MIDI commands are usually composed of one, two, or three bytes of data arranged and transmitted one after another. The first byte sent in each command is called the 'status' byte and specifies an operation to be performed. The next one or two bytes, if used, represent parameters of this command. For example, a NOTE ON command comprises three bytes, the first of which is the status byte. This byte tells a synthesizer to play a note and specifies the channel number. The channel number usually represents the type of sound to be played, ie which instrument of the synthesizer is to be used. The second byte specifies the note to be played and the third byte specifies the velocity value for the note. The bytes of the MIDI commands are specified in the MIDI standard.
In a conventional MIDI playback system a MIDI synthesizer can be controlled by a MIDI sequencer in the following way. A standard MIDI file (SMF) contains a set of events, which are intended to be executed by a synthesizer at particular times. Generally, the events are not uniformly spaced in time. A conventional MIDI sequencer parses the standard MIDI file, reads the present MIDI event and the time difference between it and the next event. The sequencer then sends the event m a MIDI message to a MIDI synthesizer at the time it is to be executed. The sequencer usually sets a timer and reads the next MIDI event after this time difference has elapsed.
A conventional MIDI synthesizer receives the MIDI message that the sequencer sends, decodes the message and operates accordingly. For example, a 'NOTE ON' event will cause the synthesizer to generate audio samples that correspond to a requested note and velocity that are supplied as parameters. Similarly, a 'NOTE OFF' event will cause the synthesizer to cease generating the audio samples.
In this system a software MIDI synthesizer is implemented as Dynamic Link Library (DLL) . This DLL handles some of the notes m software, while propagating the rest of the notes to the sound card according to performance and quality criteria.
A schematic diagram of the software components m the present implementation is shown in Fig 3. The system comprises a MIDI sequencer application layer 300, an operating system layer 310, a DLL layer indicated at 320 and a sound card layer 330. DLL layer 320 comprises 3 main blocks, controller 335, high level block 340 and low level block 350. Sound card layer 330 comprises a sound card device driver 360 and the hardware resources of the sound card which are indicated at 370. Also shown in Fig 3 is an output buffer 380.
The DLL layer 320 is triggered by MIDI messages which are sent by MIDI sequencer. These messages can be sent directly by application 300 or via the operating system services.
Controller 335 is responsible for the MIDI logic management. In other words, this block decides which notes are to be propagated to the sound card and which should be handled entirely by means of software synthesis. NOTE ON and NOTE OFF messages for notes to be handled on the sound card are sent, after a suitable delay to maintain synchronisation, directly to the sound card via line 345. Other messages are sent to both the sound card and the software synthesizer as will be described in more detail below.
The decision regarding which notes to handle by software, and which notes to propagate to the sound card takes the following considerations into account.
1. In the special case where the software DLL supports wave-table synthesis and the sound card supports only FM synthesis, there is an advantage in handling instruments which cannot be realistically synthesised by FM synthesis, such as acoustic instruments like violin, acoustic piano, clarinet etc, m software, whilst propagating to the sound card notes related to instruments that are not damaged by the FM synthesis, such as electric piano, electronic synthesizer effects, etc.
2. If there is insufficient CPU power to handle any further notes in software, then any further NOTE ON messages received should be propagated to the sound card. On the other hand if the CPU is relatively idle, it can participate in the synthesis process and thus improve the overall performance. In case of overload, the controller 335 is arranged to dynamically reduce the number of soft-voices, ie the number of notes that are currently synthesized by means of software, down to a limit of zero by directing any further NOTE ON messages to the sound card.
3. The number of empty slots on the sound card. Each empty slot is capable of playing one note. The number of slots supported by tne sound card represents the maximum nu oer of notes that the sound card can play simultaneously. High level block 340 receives tne messages which are to be handled in software and updates the various internal tables which are used by the synthesizer. It handles conventional functions of a MIDI synthesizer such as managing voice allocation etc which are not directly relevant to the present invention and will be well understood by those skilled in the art.
Low level block 350 is triggered periodically and contains the synthesis engine which requires intensive computation. This block is responsible for the synthesis of notes that are not propagated to the sound card by controller 335. The general operation of the low level block is shown in Fig 4. The loop is generally triggered with a time period of between lmS and 5mS. Each time the loop is triggered it checks in step 400 for each voice which kind of note is to be produced and looks in the Wave Table database to find the appropriate waveform. The waveform for each active voice is then transformed if necessary to the correct pitch and the waveforms combined to generate output samples - step 410 - at a suitable sampling rate, for example 44100 samples per second, and places them in buffer 380 for retrieval by the device driver 360 in sound card layer 330.
Since the synthesis algorithm is computationally very intensive, it can be used for monitoring the CPU usage. This measure of the CPU usage is available to controller 335 to enable it to decide whether notes should be propagated to the sound card.
There are a number of ways that the CPU usage can be monitored via the synthesis process. For example, a time measurement can be taken each time the low-level block 340 is triggered. If the difference between the current time and the time that was measured on the previous iteration is consistently larger than the requested period, then it can be assumed that the CPU is overloaded.
Alternatively, a time measurement can be taken at the beginning and it the end of the synthesis process m each iteration. The difference between these two measurements is approximated to be the CPU time which is used for the current number of soft voices. The CPU usage in percentage is then calculated as the ratio of the measured synthesis time and the time interval which is used as the triggering period for the low level block 350.
Controller 335 directs all MIDI messages, apart from NOTE ON and NOTE OFF messages to both the sound card and the h gh level block 340. In the case of NOTE ON and NOTE OFF messages, the processes shown in Figs 5 and 6 respectively are performed. in the case of a NOTE ON command it is determined in step 500 whether the instrument to which the note relates is one which a priori should be handled in software because of its nature. If so, then a determination of the CPU load is made in step 510 and if the CPU load is less than a fixed limit the NOTE ON message is directed to high level block 340 and the note is handled in software - step 520. If the CPU is greater than 50% loaded or the instrument to which the note relates can be adequately handled by the sound card, then the NOTE ON command is directed to the sound card - step 530.
On receipt of a NOTE OFF command, it is determined, in step 600 in Fig 6, whether the voice concerned is presently being handled in software. This information is available in the tables maintained by high level block 340. If the note is being handled in software, then the NOTE OFF message is directed to high level block 340 - step 610, if not then the NOTE OFF message is directed to the sound card - step 620.
it will be appreciated that there is a need to delay propagation of MIDI events to be handled in the sound card to maintain synchronisation between the samples produced in software and those produced in hardware. This will be a constant, but processor and/or sound card-dependant delay and m the present arrangement can be set manually by the user via a suitable menu presented to them. Determination of the correct delay may be easily established empirically by the user.
The present implementation has been written to work under the well known Windows operating system from Microsoft Corp (Windows is a trademark of Microsoft Corp.). The general operation of the system is as follows. First the user invokes a Windows application, such as a sequencer, which requires a MIDI synthesizer. The application opens the MIDI synthesizer DLL via the MIDI Mapper program, with the intervention of the MMSYSTEM driver.
The software MIDI Synthesizer DLL opens the sound card twice - once as a MIDI device to which MIDI messages will be propagated and once as a WAVE device to which the wave samples produced by the software MIDI synthesizer will be sent.
Since the software synthesizer locks the WAVE input of the sound card, it is itself provided with a WAVE input to receive WAVE messages from other sources and handle the mixing of these with the wave data it generates. MIDI messages for instruments for which the sound quality is acceptable using FM synthesis are propagated to the sound card and handled in hardware. The software MIDI synthesizer handles the MIDI messages for instruments which require Wave table synthesis up to a fixed limit, such as 50%, of the available CPU power. In case of overload, the software synthesizer dynamically reduces the number of soft voices, ie the number of notes that are currently synthesized by means of software down to a limit of zero. As described above, this is achieved by directing any further NOTE ON messages to the sound card.
As will be clear from the above description, the present implementation takes the form of a computer program and can be distributed in the form of an article of manufacture comprising a computer usable medium in which suitable program code is embodied for causing a computer to perform the function of controller 335 described above. The program may include the high level block 340 and 350 and low level blocks of the software synthesizer or could be implemented for use with a preexisting software synthesizer.
However, it will be appreciated that many variations are possible within the scope of the attached claims. For example, the controller might be implemented on the sound card itself, either in hardware or software, and arranged to receive MIDI messages from an external source, the sound card handling some notes itself and directing some notes over the PC bus to a software synthesizer.
INDUSTRIAL APPLICABILITY
The invention is applicable in the technical field of computers and digital audio systems.

Claims

1. An audio synthesizer for generating an analogue or digital audio output in response to coded control instructions representing musical events, the synthesizer comprising:
a general purpose computer portion having a CPU programmed to receive the control instructions and generate audio samples;
a special purpose hardware portion for receiving the control instructions and generating the audio samples;
a controller for directing the control instructions either so that the general purpose computer portion generates the audio samples or so that the hardware portion generates the audio samples; and
means to combine the audio samples generated by the general purpose computer portion and the hardware portion to form an audio output which accords with the control instructions.
2. An audio synthesizer as claimed m claim 1 comprising measurement means for repeatedly measuring the load on the CPU and wherein the controller is arranged to direct the control instructions according to the measured load on the CPU.
3. An audio synthesizer as claimed m claim 1 wherein the measurement means comprises timing means for timing a synthesis loop in the program for generating the audio samples.
4. An audio synthesizer as claimed in any claim 1 wherein the general purpose computer is programmed to act as the controller and to either generate the audio samples itself or to transfer the instructions to the hardware portion.
5. An audio synthesizer as claimed in claim 1 in which the controller is arranged to direct the instructions according to voice type.
6. An audio synthesizer as claimed in claim 1 in which the general purpose computer comprises a wave-table synthesizer for generating the audio samples and the hardware portion comprises an FM synthesizer for generating the audio samples.
7. An audio synthesizer as claimed m claim 1 including delay means for delaying the direction of control instructions to the hardware portion, so that the audio samples generated by the hardware portion are synchronised with the audio samples generated by the general purpose computer.
8. An audio synthesizer as claimed in claim 1 including an input for receiving audio samples and means for mixing audio samples received at said input with audio samples generated in response to the coded control instructions.
9. An audio synthesizer as claimed in claim 1 in the form of a personal computer with an add-on sound card.
10. A method of operating an audio synthesizer to generate an analogue or digital audio output in response to coded control instructions representing musical events, the synthesizer comprising: a general purpose computer portion having a CPU programmed to receive the control instructions and generate audio samples; and a special purpose hardware portion for receiving the control instructions and generating the audio samples, the method comprising directing the control instructions either so that the general purpose computer portion generates the audio samples or so that the hardware portion generates the audio samples so that the audio samples generated by the general purpose computer portion and the hardware portion can be combined to form an audio output which accords with the control instructions.
11. A method as claimed in claim 10 comprising delaying the direction of control instructions to the hardware portion, so that the audio samples generated by the hardware portion are synchronised with the audio samples generated by the general purpose computer.
12. A method as claimed in claim 10 comprising receiving a NOTE ON instruction; determining whether the NOTE ON instruction is for a predefined set of voices to be handled by the general purpose computer portion and, if not, directing the NOTE ON instruction so that the hardware portion generates corresponding audio samples; measuring the CPU load and if the CPU load exceeds a predefined threshold, directing the NOTE ON instruction so that the hardware portion generates corresponding audio samples; otherwise directing the NOTE ON instruction so that the general purpose computer portion generates the audio samples.
13. A method as claimed in claim 10 comprising receiving a NOTE OFF instruction; determining whether the NOTE OFF instruction relates to a note being handled by the general purpose computer portion and, if not, directing the NOTE OFF instruction to the hardware portion; otherwise directing the NOTE OFF instruction to the general purpose computer portion.
14. An article of manufacture comprising a computer usable medium in which program code is embodied for causing a computer to perform a method as claimed in claim 10.
PCT/GB1996/000331 1995-10-03 1996-02-15 Audio synthesizer WO1997013240A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP96902371A EP0853802B1 (en) 1995-10-03 1996-02-15 Audio synthesizer
DE69608292T DE69608292T2 (en) 1995-10-03 1996-02-15 AUDIO SYNTHESIZER
JP9514043A JP2970689B2 (en) 1995-10-03 1996-02-15 Audio synthesizer

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB9520124.0 1995-10-03
GB9520124A GB2306043A (en) 1995-10-03 1995-10-03 Audio synthesizer

Publications (1)

Publication Number Publication Date
WO1997013240A1 true WO1997013240A1 (en) 1997-04-10

Family

ID=10781657

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB1996/000331 WO1997013240A1 (en) 1995-10-03 1996-02-15 Audio synthesizer

Country Status (6)

Country Link
US (1) US5808221A (en)
EP (1) EP0853802B1 (en)
JP (1) JP2970689B2 (en)
DE (1) DE69608292T2 (en)
GB (1) GB2306043A (en)
WO (1) WO1997013240A1 (en)

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3152156B2 (en) * 1996-09-20 2001-04-03 ヤマハ株式会社 Music sound generation system, music sound generation apparatus and music sound generation method
US6051772A (en) * 1997-07-15 2000-04-18 Aureal Semiconductor, Inc. Method and apparatus for emulating a frequency modulation device
US6556560B1 (en) * 1997-12-04 2003-04-29 At&T Corp. Low-latency audio interface for packet telephony
US6357039B1 (en) * 1998-03-03 2002-03-12 Twelve Tone Systems, Inc Automatic code generation
FR2785711B1 (en) * 1998-11-06 2001-03-16 Jean Philippe Chevreau DEVICE THAT AUTOMATICALLY COMPOSES A MUSICAL PERFORMANCE BY INCORPORATING SOUND SAMPLES
US6647359B1 (en) * 1999-07-16 2003-11-11 Interval Research Corporation System and method for synthesizing music by scanning real or simulated vibrating object
US7078609B2 (en) 1999-10-19 2006-07-18 Medialab Solutions Llc Interactive digital music recorder and player
US7176372B2 (en) 1999-10-19 2007-02-13 Medialab Solutions Llc Interactive digital music recorder and player
US6392133B1 (en) 2000-10-17 2002-05-21 Dbtech Sarl Automatic soundtrack generator
US9818386B2 (en) 1999-10-19 2017-11-14 Medialab Solutions Corp. Interactive digital music recorder and player
JP3675362B2 (en) * 2000-08-18 2005-07-27 ヤマハ株式会社 Musical sound generating device and portable terminal device
US6671658B2 (en) * 2000-12-23 2003-12-30 Hewlett-Packard Development Company, L.P Method for service level estimation in an operating computer system
US7065380B2 (en) 2001-07-19 2006-06-20 Texas Instruments Incorporated Software partition of MIDI synthesizer for HOST/DSP (OMAP) architecture
EP1326228B1 (en) 2002-01-04 2016-03-23 MediaLab Solutions LLC Systems and methods for creating, modifying, interacting with and playing musical compositions
US7076035B2 (en) 2002-01-04 2006-07-11 Medialab Solutions Llc Methods for providing on-hold music using auto-composition
US9065931B2 (en) 2002-11-12 2015-06-23 Medialab Solutions Corp. Systems and methods for portable audio synthesis
US6977335B2 (en) 2002-11-12 2005-12-20 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US7928310B2 (en) 2002-11-12 2011-04-19 MediaLab Solutions Inc. Systems and methods for portable audio synthesis
US7169996B2 (en) 2002-11-12 2007-01-30 Medialab Solutions Llc Systems and methods for generating music using data/music data file transmitted/received via a network
US7038119B2 (en) * 2003-07-18 2006-05-02 Telefonaktiebolaget L M Ericsson (Publ) Dynamic control of processing load in a wavetable synthesizer
DE10339032A1 (en) * 2003-08-25 2005-05-25 Infineon Technologies Ag Software synthesizer e.g. for generating audio signal with variable amount of voices depending on operating status of processor, has data processing framework for processor unit and at time period, maximally generatable code is corrected
EP1571647A1 (en) * 2004-02-26 2005-09-07 Lg Electronics Inc. Apparatus and method for processing bell sound
US7285712B2 (en) * 2004-05-25 2007-10-23 Mediatek Incorporation Method of dynamically determining a maximum polyphony number according to operation mode and smoothly changing polyphony number when switching operation modes
US7807915B2 (en) 2007-03-22 2010-10-05 Qualcomm Incorporated Bandwidth control for retrieval of reference waveforms in an audio device
US8299347B2 (en) * 2010-05-21 2012-10-30 Gary Edward Johnson System and method for a simplified musical instrument
US9633642B2 (en) 2015-07-24 2017-04-25 Roland Corporation Electronic musical instrument system
US9536504B1 (en) 2015-11-30 2017-01-03 International Business Machines Corporation Automatic tuning floating bridge for electric stringed instruments

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0126975A2 (en) * 1983-05-21 1984-12-05 Reinhard Franz Electronic keyboard musical instrument
EP0377459A2 (en) * 1989-01-06 1990-07-11 Yamaha Corporation Electronic musical instrument having plural different tone generators
EP0484047A2 (en) * 1990-11-01 1992-05-06 International Business Machines Corporation Method and apparatus for simultaneous output of digital audio and midi synthesised music
US5354948A (en) * 1989-10-04 1994-10-11 Yamaha Corporation Tone signal generation device for generating complex tones by combining different tone sources

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5119710A (en) * 1986-03-09 1992-06-09 Nippon Gakki Seizo Kabushiki Kaisha Musical tone generator
JPH0631957B2 (en) * 1987-02-06 1994-04-27 ヤマハ株式会社 Electronic musical instrument
US5121667A (en) * 1989-11-06 1992-06-16 Emery Christopher L Electronic musical instrument with multiple voices responsive to mutually exclusive ram memory segments
JP2623878B2 (en) * 1989-12-21 1997-06-25 ヤマハ株式会社 Electronic musical instrument
JPH04128796A (en) * 1990-09-19 1992-04-30 Brother Ind Ltd Music reproduction device
US5119711A (en) * 1990-11-01 1992-06-09 International Business Machines Corporation Midi file translation
JP2626387B2 (en) * 1991-12-24 1997-07-02 ヤマハ株式会社 Electronic musical instrument
JP2850707B2 (en) * 1993-06-15 1999-01-27 ヤマハ株式会社 Music control device
US5541354A (en) * 1994-06-30 1996-07-30 International Business Machines Corporation Micromanipulation of waveforms in a sampling music synthesizer

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0126975A2 (en) * 1983-05-21 1984-12-05 Reinhard Franz Electronic keyboard musical instrument
EP0377459A2 (en) * 1989-01-06 1990-07-11 Yamaha Corporation Electronic musical instrument having plural different tone generators
US5354948A (en) * 1989-10-04 1994-10-11 Yamaha Corporation Tone signal generation device for generating complex tones by combining different tone sources
EP0484047A2 (en) * 1990-11-01 1992-05-06 International Business Machines Corporation Method and apparatus for simultaneous output of digital audio and midi synthesised music

Also Published As

Publication number Publication date
JP2970689B2 (en) 1999-11-02
EP0853802A1 (en) 1998-07-22
EP0853802B1 (en) 2000-05-10
GB2306043A (en) 1997-04-23
US5808221A (en) 1998-09-15
JPH10512376A (en) 1998-11-24
DE69608292T2 (en) 2001-01-04
DE69608292D1 (en) 2000-06-15
GB9520124D0 (en) 1995-12-06

Similar Documents

Publication Publication Date Title
EP0853802B1 (en) Audio synthesizer
JP3161561B2 (en) Multimedia system
US6646195B1 (en) Kernel-mode audio processing modules
US6100461A (en) Wavetable cache using simplified looping
USRE41757E1 (en) Sound source system based on computer software and method of generating acoustic waveform data
US7678986B2 (en) Musical instrument digital interface hardware instructions
JPH09179556A (en) Method and device for musical tone generation
US5847304A (en) PC audio system with frequency compensated wavetable data
US7718882B2 (en) Efficient identification of sets of audio parameters
US7663052B2 (en) Musical instrument digital interface hardware instruction set
JPH08160961A (en) Sound source device
JP3141789B2 (en) Sound source system using computer software
JP3269361B2 (en) Musical sound generating method and apparatus
JP3705203B2 (en) Music generation method
JP3735172B2 (en) Performance information processing method
Yim et al. A real-time MIDI music synthesis system using sinusoidal modeling on a TI TMS320C32 digital signal processor

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): JP US

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 1996902371

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1996902371

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 1996902371

Country of ref document: EP