US5150323A - Adaptive network for in-band signal separation - Google Patents

Adaptive network for in-band signal separation Download PDF

Info

Publication number
US5150323A
US5150323A US07/392,681 US39268189A US5150323A US 5150323 A US5150323 A US 5150323A US 39268189 A US39268189 A US 39268189A US 5150323 A US5150323 A US 5150323A
Authority
US
United States
Prior art keywords
neurons
signal
output
composite signal
composite
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US07/392,681
Inventor
Patrick F. Castelaz
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
DirecTV Group Inc
Raytheon Co
Original Assignee
Hughes Aircraft Co
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hughes Aircraft Co filed Critical Hughes Aircraft Co
Priority to US07/392,681 priority Critical patent/US5150323A/en
Assigned to HUGHES AIRCRAFT COMPANY, A CORP. OF DE reassignment HUGHES AIRCRAFT COMPANY, A CORP. OF DE ASSIGNMENT OF ASSIGNORS INTEREST. Assignors: CASTELAZ, PATRICK F.
Priority to JP2512242A priority patent/JPH04501330A/en
Priority to PCT/US1990/004488 priority patent/WO1991002324A1/en
Priority to EP90912714A priority patent/EP0439592A1/en
Application granted granted Critical
Publication of US5150323A publication Critical patent/US5150323A/en
Assigned to HUGHES ELECTRONICS CORPORATION reassignment HUGHES ELECTRONICS CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HE HOLDINGS INC., HUGHES ELECTRONICS, FORMERLY KNOWN AS HUGHES AIRCRAFT COMPANY
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0272Voice signal separating
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
    • G10L25/30Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks

Definitions

  • This invention relates to signal processing techniques and, more particularly, to a method and apparatus for separating in-band signals.
  • In-band signal separation of complex composite signals is an extremely difficult signal processing problem. In-band separation problems are encountered in situations where a single complex composite signal composed of individual constituent signals must be separated into their original components. In one example, known as the "cocktail party" problem, multiple speech signals may be received by a single sensor and it is desired to separate out the voices of individual speakers. Other examples include decomposition of complex composite radar or o sonar signals emitted from multiple sources and received at a single receiver. In addition, two-dimensional problems may present similar signal separation problems. These include object detection and identification of image data in which multiple overlapping (in-band) additive sources are present. In particular, with respect to image data, objects of interest may overlap within the intensity and frequency bands of the sensor, and it is desired to separate these overlapping images.
  • neural nets exhibit self-learning by changing their synaptic weights until the correct output is achieved in response to a particular input.
  • neural nets are capable of recognizing a target and producing a desired output even where the input is incomplete or hidden in background noise.
  • neural nets exhibit greater robustness, or fault tolerance, than Von Neumann sequential computers because there are many more processing nodes, each with primarily local connections. Damage to a few nodes or links need not impair overall performance significantly.
  • neural net models utilizing various topologies, neuron characteristics, and training or learning rules. Learning rules specify an internal set of weights and indicate how weights should be adapted during use, or training, to improve performance.
  • some of these neural net models include the Perceptron, described in U.S. Pat. No. 3,287,649 issued to F. Rosenblatt; the Hopfield Net, described in U.S. Pat. Nos. 4,660,166 and 4,719,591 issued to J. Hopfield; the Hamming Net and Kohohonen self-organizing maps, described in R.
  • in-band signal separation processor which can handle significant variations in the data and is also fault tolerant. It is further desirable to provide an in-band signal separation processor which can accept raw (e.g., time--amplitude) signal data with a minimum of preprocessing.
  • an adaptive network for in-band signal separation accepts as direct input, discrete portions of a composite signal.
  • the adaptive network is trained by presenting a training composite signal as input to the input neurons and by presenting a desired output to selected groups of its output neurons. This desired output consists of one or more of the constituent signals contained in the composite training input signal.
  • the training continues until the adaptive network produces the desired output in response to a known composite signal.
  • the adaptive network may then be used to separate constituent signals from an unknown composite signal, if one of the constituent signals has characteristics in common with the constituent signal used to train the network.
  • FIG. 1 is an overview of the in-band signal separation problem.
  • FIG. 2 (a-b) are graphs illustrating a conventional signal separation technique
  • FIG. 3 (a-b) are graphs of the conventional and neural network approaches to the in-band signal separation problem.
  • FIG. 4 is a graphical illustration of the adaptive network for in-band signal separation in accordance with the techniques of the present invention.
  • FIG. 5 is graphical illustration of the results of the adaptive network for in-band signal separation in accordance with the present invention after one training cycle
  • FIG. 6 is a graphical illustration of the results of the adaptive in-band network for a signal separation after ten training cycles
  • FIG. 7 is a graphical illustration of the results of the adaptive network for in-band signal separation after 100 training cycles
  • FIG. 8 is a graphical illustration of the adaptive network for in-band signal separation results after 300 training cycles
  • FIG. 9 is an illustration of the sampling technique in accordance with the preferred embodiment of the present invention.
  • FIG. 10 is an illustration of a multilayer perceptron in accordance with the prior art.
  • a method and apparatus for separating an in-band composite signal into its constituent signals.
  • FIG. 2A there is shown a composite signal 10 which includes two constituent signals 12 and 14.
  • the graph in FIG. 2A shows the signals 12 and 14 with the amplitude plotted as a function of frequency. It will be appreciated that the signals may also be represented in other ways, such as in the time, instead of the frequency domain.
  • composite signals can be separable, such as by different frequency bands, as shown in FIG. 2A, conventional filtering techniques can be used to separate the two signals.
  • FIG. 2B a filter 16 can be applied to the first and second signals 12 and 14 to accomplish signal separation.
  • the filter 16 has two discrete regions in which it permits signals to pass. In the first region, the first signal 12 is passed and all other signals having other frequencies are filtered out. In the second region, the second signal 14 is permitted to pass while other frequency regions are filtered out.
  • the composite signal 18 is comprised of constituent signals 20 and 22 which are overlapping (e.g., "in-band") in the frequency domain.
  • a conventional filter 24 will be unable to separate the signals 20 and 22.
  • More sophisticated filtering techniques would require an extensive analysis and design effort to develop feature extraction and filtering algorithms.
  • the execution of these algorithms would be slow even using state-of-the-art conventional signal processors.
  • conventional techniques often do not separate the two signals to a satisfactory degree, and the resulting composite signals are thus not true representations of the original constituents.
  • the present invention is based on the discovery that a neural network can adapt to the fine structure of a composite signal 18 to perform in-band signal separation when the neural network is presented and trained with the signals in accordance with certain teachings within the scope of the present invention.
  • Constituent signals 28 and 30 are combined into a composite signal 32 at the signal source.
  • these signals may comprise speech, radar, sonar, optical or other various signals.
  • the signals may be speech signals originating from two speakers and the composite signal is sensed by a microphone 34.
  • Microphone 34 responds to both the first signal 28 and the second signal 30 and generates a composite electrical signal 32.
  • This signal is sent to the adaptive network for in-band signal separation 26 which produces two outputs; the first output 36 is a faithful reproduction of the original first signal 28; and a second output signal 38, is a faithful reproduction of the original second input signal 30.
  • the neural network employed in the preferred embodiment utilizes a neural network known as a multilayer perceptron.
  • a multilayer perceptron includes a layer of input neurons 40, one or more layers of inner neurons 42, and a layer of output neurons 44.
  • each neuron in each layer is connected to each neuron in the adjacent layers by means of synaptic connections 43 as shown in FIG. 10.
  • the particular interconnection scheme and training algorithm employed may be according to a number of other neural network architectures including, but not limited to, the Boltzman machine, Counterprop, Hopfield net, Hamming net, etc. It is preferable that the neural network architecture and training algorithm employed belong to the class of supervised, as opposed to unsupervised nets.
  • the particular interconnection scheme and training algorithm employed with the multilayer perceptron and its associated learning algorithm, known as backward error propagation, are well known. Details of the multilayer perceptron are described in Rumelhart, Hinton, and Williams, "Learning Internal Representations by Error of Propagation", in D. E. Rumelhart and J. L. McClelland (Eds.), Parallel Distributed Processing; Explorations in the Microstructure of Cognition, Vol. 1 Foundations, M.I.I. Press (1986), which is incorporated herein by reference.
  • a low frequency composite signal 46 and a high frequency composite signal 48 are both transmitted to the input neurons 40 in the adaptive network 26.
  • the use of the low frequency 46 and high frequency 48 versions of the composite signal 32 permit a reduced number of input neurons 40 to be employed. A large number of inputs is generally considered to be necessary. This is because a high frequency representation of the signal is needed to get a faithful reproduction of the fine structure of the signal; and a broader or lower frequency representation is also needed to give the processor 26 information about more fundamental frequencies, (e.g., pitch) of the speaker. That is, the network should have available the high frequency structure which contains, for example, words and phonemes and the low frequency structure, which contains, for example, the pitch that is characteristic of a given speaker. Thus, one way to give the processor 26 this information is to employ a large number of input neurons 40. This may require, for example, two hundred or more input neurons 40 to give a broad enough sample of the speech data.
  • the necessary high and low frequency information can be given to the processor 26 by means of a filter circuit 50 shown in FIG. 9.
  • the filter circuit 50 accepts as input the composite speech signal 32 and generates a high frequency output 48 and a low frequency output 46.
  • the composite signal 32 is divided into 16 samples at a low frequency, for example, 640 Hertz(Hz).
  • a sample is taken every 25.6 milliseconds.
  • the high frequency samples are taken every 1.6 milliseconds, at a rate of 10 kilohertz. Sixteen of the high frequency samples, when combined, generate the high frequency input composite signal 48.
  • the low frequency composite signal 46 is fed along input line 52 to a series of sampling circuits 54 through an input buffer 55.
  • the low frequency composite signal 46 is fed through the input line 52, to a buffer circuit 55 and the sampling circuits 54 until each of the 16 samples reside in a single sampling circuit 54.
  • Each sampling circuit 54 is connected to an input neuron 40 in the input layer of the processor 26.
  • the high frequency composite signal 48 is transmitted through an input line 56 through input buffer 58 to a series of sampling circuits 60. It should be noted that while only eight sampling circuits 60 and 8 sample circuits 54 are shown in FIG. 4, there would actually be 16 of the high frequency sampling circuit 60 and 16 of low frequency sample circuits 54, each connected to an input neuron 40. It will be appreciated that depending on the specific application, more or less than 16 samples may be used. Also, the network could be configured to handle more than 2 speakers.
  • the composite signal 32 consists of a training input signal, which is composed of two known constituent signals such as the first and second constituent signal 28 and 30 shown in FIG. 1.
  • the processor 26 will produce an output at each of its output neurons 44.
  • the processor 26 is trained with a desired output consisting of high frequency representations of the two constituent signals 28 and 30.
  • the first 16 output neurons may be presented with the first constituent signal 28 and the next 16 output neurons may be presented with the second constituent signal 30 during training.
  • the actual output 62, 64 will approximate the desired output.
  • the first 16 output neurons 44 will approximate the first constituent signal 28 and the next 16 output neurons will approximate the second constituent signal 30.
  • the processor 26 could be trained with only a single training input such as constituent signal 28 for cases where only a single constituent signal is desired. However, where only a single signal is provided by the processor 26, it will be appreciated that this single signal may be separated from the composite and the remaining signal may yield a second constituent signal.
  • training may continue by repeating the training procedure for the next consecutive 16 high frequency samples, by shifting the high frequency signal over by an amount equal to the distance of 16 high frequency samples.
  • the processor 26 is again trained with this input, until the desired output is achieved to within a predetermined tolerance. This procedure can then be repeated a number of times which will depend on the complexity of the signal and the neural network architecture employed.
  • an unknown composite signal can be presented to the input neurons 40 in the same manner as the training composite signal. That is, a low frequency representation 46 and a high frequency representation 48 of the unknown composite signal is presented to 16 of the input neurons 40 respectively. If the unknown composite signal contains constituent signals 28 and 30, the output of the processor 26 will consist of the first constituent signal 28 from the first 16 output neurons 44 and the second constituent signal 30 from the next 16 output neurons 44. A slower but alternative technique which may be useful in certain applications would be to shift the data over by one high frequency sample at a time rather than 16 samples between training sessions.
  • FIGS. 5-8 illustrations of the output of the adaptive network 26 at various stages in training is shown.
  • the "composite input signal” is shown twice in the top row.
  • This composite signal is comprised of signals from two individual speakers, labelled “Speaker 1 Component” and “Speaker 2 Component", shown in the second row.
  • the composite signal may be signal from a microphone responding to two persons (speaker 1 and speaker 2) talking simultaneously.
  • the composite signal is fed to the adaptive network 26 and the network is trained with the known speaker 1 and speaker 2 examples, in accordance with the techniques described above.
  • the output of the adaptive network 26 appears as shown in the third row of FIG. 5. That is, the output neurons trained with the speaker 1 component produce output signals labelled "Speaker Network Output" and the output neurons trained with the speaker 2 component produce output signals labeled "Speaker 2 Network Output".
  • FIG. 6 shows the speaker 1 and speaker 2 network outputs after ten training cycles, as shown in FIG. 6, the speaker 1 and speaker 2 network outputs begin to show some significant distinguishing characteristics, particularly the speaker output.
  • FIG. 7 shows the outputs after 100 training cycles, and both the speaker 1 and 2 outputs begin to appear to resemble the original component signals. After 300 cycles, as shown in FIG. 8, the outputs become very good approximations of the original component signals.
  • the exact number of training cycles required will depend upon a number of factors, such as the complexity of the component signals and the desired fidelity of the outputs.
  • the trained adaptive network 26 can then be used to identify unknown composite signals to restore the original constituent signals if they are contained in the composite.
  • the speaker and speaker 2 components used for training will be from speech by the same person or persons whose speech is in the unknown composite. It is also possible, however, to use the adaptive processor 26 in accordance with the present invention, to separate speech from unknown speakers. That is, by training an adaptive network 26 of sufficient complexity, a sufficient number of times, it is possible for it to "learn" the general characteristics of human speech so as to separate two examples of such speech from a single composite signal. It will be appreciated that an unsupervised, as opposed to supervised neural net may be preferred for this kind of application.
  • the weight values developed through training could be transferred to the processor having its weights fixed to none values. In this way, mass production of processors 26 is possible without repeating the training procedure.
  • the composite signal 32 as shown in FIG. 1 consisted of the amplitudes of the raw signal in the time domain
  • the above techniques for the processor 26 could be employed in the frequency domain. That is, the input could be a frequency representation of the composite signal and the output also be some frequency representation. In this case, an inverse fourier transform could be used to restore the resultant signal.

Abstract

An Adaptive Network For In-Band Signal Separation (26) and method for providing in-band separation of a composite signal (32) into its constituent signals (28), (30). The input to the network (26) is a series of sampled portions of the composite signal (32). The network (26) is trained with at least one of said composite signals (28) (30) using a neural network training paradigm by presenting one or more of the constituent signals (28) (30) to said network (28). The network (26) may be used to separate multiple speech signals from a composite signal from a single sensor such as a microphone.

Description

BACKGROUND OF THE INVENTION
1. Technical Field
This invention relates to signal processing techniques and, more particularly, to a method and apparatus for separating in-band signals.
2. Discussion
In-band signal separation of complex composite signals is an extremely difficult signal processing problem. In-band separation problems are encountered in situations where a single complex composite signal composed of individual constituent signals must be separated into their original components. In one example, known as the "cocktail party" problem, multiple speech signals may be received by a single sensor and it is desired to separate out the voices of individual speakers. Other examples include decomposition of complex composite radar or o sonar signals emitted from multiple sources and received at a single receiver. In addition, two-dimensional problems may present similar signal separation problems. These include object detection and identification of image data in which multiple overlapping (in-band) additive sources are present. In particular, with respect to image data, objects of interest may overlap within the intensity and frequency bands of the sensor, and it is desired to separate these overlapping images.
Conventional approaches to in-band signal separation require extensive front end analysis and design in the development of feature extraction and filtering algorithms. Specifically, conventional techniques typically involve extensive preprocessing. Such preprocessing may require, for example, measuring pulse width, amplitude, rise and fall times, frequency, etc. Once these features are extracted, they can be matched with stored patterns for classification, identification and generation of the separated output signals. However, the software required to accomplish these steps is often complex and is time-consuming to develop. Moreover, conventional processors are often not capable of separating in-band signals satisfactorily. In addition, conventional digital signal processors are not able to tolerate certain variations in the input signal, such as changes in orientation of a visual pattern, or differences in speakers, in the case of speech recognition.
In recent years it has been realized that conventional Von Neumann computers, which operate serially, bear little resemblance to the parallel processing that takes place in biological systems such as the brain. It is not surprising, therefore, that conventional signal processing techniques should fail to adequately perform the tasks involved in human perception. Consequently, new methods based on neural models of the brain are being developed to perform perceptual tasks. These systems are known variously as neural networks, neuromorphic systems, learning machines, parallel distributed processors, self-organizing systems, or adaptive logic systems. Whatever the name, these models utilize numerous nonlinear computational elements operating in parallel and arranged in patterns reminiscent of biological neural networks. Each computational element or "neuron" is connected via weights or "synapses" that typically are adapted during training to improve performance. Thus, these systems exhibit self-learning by changing their synaptic weights until the correct output is achieved in response to a particular input. Once trained, neural nets are capable of recognizing a target and producing a desired output even where the input is incomplete or hidden in background noise. Also, neural nets exhibit greater robustness, or fault tolerance, than Von Neumann sequential computers because there are many more processing nodes, each with primarily local connections. Damage to a few nodes or links need not impair overall performance significantly.
There are a wide variety of neural net models utilizing various topologies, neuron characteristics, and training or learning rules. Learning rules specify an internal set of weights and indicate how weights should be adapted during use, or training, to improve performance. By way of illustration, some of these neural net models include the Perceptron, described in U.S. Pat. No. 3,287,649 issued to F. Rosenblatt; the Hopfield Net, described in U.S. Pat. Nos. 4,660,166 and 4,719,591 issued to J. Hopfield; the Hamming Net and Kohohonen self-organizing maps, described in R. Lippman, "An Introduction to Computing with Neural Nets", IEEE ASSP Magazine, April 1987, pages 4-22; and "The Generalized Delta Rule for Multilayered Perceptrons", described in Rumelhart, Hinton, and Williams, "Learning Internal Representations by Error Propagation", in D. E. Rumelhart and J. L. McClelland (Eds.), Parallel Distributed Processing; Explorations in the Microstructure of Cognition. Vol. 1: Foundation. MIT Press (1986).
While each of these models achieve varying degrees of success at the particular perceptual tasks to which it is best suited, the parallel inputs required by these systems are thought to necessitate special purpose preprocessors for real time hardware implementations. (See the above-mentioned article by R. Lippman.) For example, in Rosenblatt's Perceptron, (U.S. Pat. No. 3,287,649) each input receives a separate frequency band of an analog audio signal. Thus, while neural networks reduce the amount of algorithm development required to analyze a signal, the representation of the in-band signal separation problem to a neural network would still require extensive preprocessing to present the signal to the conventional neural network.
Thus, it would be desirable to provide a system for accomplishing in-band signal separation which does not require extensive algorithm and software development, but, which instead, can develop its own algorithm without requiring the algorithm to be explicitly defined in advance. It would also be desirable to provide an in-band signal separation processor which can handle significant variations in the data and is also fault tolerant. It is further desirable to provide an in-band signal separation processor which can accept raw (e.g., time--amplitude) signal data with a minimum of preprocessing.
SUMMARY OF THE INVENTION
In accordance with the teachings of the present invention, an adaptive network for in-band signal separation accepts as direct input, discrete portions of a composite signal. The adaptive network is trained by presenting a training composite signal as input to the input neurons and by presenting a desired output to selected groups of its output neurons. This desired output consists of one or more of the constituent signals contained in the composite training input signal. The training continues until the adaptive network produces the desired output in response to a known composite signal. The adaptive network may then be used to separate constituent signals from an unknown composite signal, if one of the constituent signals has characteristics in common with the constituent signal used to train the network.
BRIEF DESCRIPTION OF THE DRAWINGS
The various advantages of the present invention will become apparent to those skilled in the art after reading the following specifications and by reference to the drawings in which:
FIG. 1 is an overview of the in-band signal separation problem.
FIG. 2 (a-b) are graphs illustrating a conventional signal separation technique;
FIG. 3 (a-b) are graphs of the conventional and neural network approaches to the in-band signal separation problem;,
FIG. 4 is a graphical illustration of the adaptive network for in-band signal separation in accordance with the techniques of the present invention;
FIG. 5 is graphical illustration of the results of the adaptive network for in-band signal separation in accordance with the present invention after one training cycle;
FIG. 6 is a graphical illustration of the results of the adaptive in-band network for a signal separation after ten training cycles;
FIG. 7 is a graphical illustration of the results of the adaptive network for in-band signal separation after 100 training cycles;
FIG. 8 is a graphical illustration of the adaptive network for in-band signal separation results after 300 training cycles;
FIG. 9 is an illustration of the sampling technique in accordance with the preferred embodiment of the present invention; and
FIG. 10 is an illustration of a multilayer perceptron in accordance with the prior art.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
In accordance with the teachings of the present invention, a method and apparatus is provided for separating an in-band composite signal into its constituent signals. Referring now to FIG. 2A, there is shown a composite signal 10 which includes two constituent signals 12 and 14. The graph in FIG. 2A shows the signals 12 and 14 with the amplitude plotted as a function of frequency. It will be appreciated that the signals may also be represented in other ways, such as in the time, instead of the frequency domain. Where composite signals can be separable, such as by different frequency bands, as shown in FIG. 2A, conventional filtering techniques can be used to separate the two signals. As shown in FIG. 2B, a filter 16 can be applied to the first and second signals 12 and 14 to accomplish signal separation. In particular, the filter 16 has two discrete regions in which it permits signals to pass. In the first region, the first signal 12 is passed and all other signals having other frequencies are filtered out. In the second region, the second signal 14 is permitted to pass while other frequency regions are filtered out.
Referring now to FIG. 3A, the in-band signal separation problem is presented. In this example, the composite signal 18 is comprised of constituent signals 20 and 22 which are overlapping (e.g., "in-band") in the frequency domain. In this case a conventional filter 24 will be unable to separate the signals 20 and 22. More sophisticated filtering techniques would require an extensive analysis and design effort to develop feature extraction and filtering algorithms. Also, the execution of these algorithms would be slow even using state-of-the-art conventional signal processors. Moreover, conventional techniques often do not separate the two signals to a satisfactory degree, and the resulting composite signals are thus not true representations of the original constituents.
Referring now to FIG. 3B, the approach of present invention is illustrated. In particular, the present invention is based on the discovery that a neural network can adapt to the fine structure of a composite signal 18 to perform in-band signal separation when the neural network is presented and trained with the signals in accordance with certain teachings within the scope of the present invention.
Referring now to FIG. 1, the overall functions of an adaptive network for in-band signal separation 26 according to the present invention is shown. Constituent signals 28 and 30 are combined into a composite signal 32 at the signal source. For example, these signals may comprise speech, radar, sonar, optical or other various signals. In accordance with the preferred embodiment of the present invention, the signals may be speech signals originating from two speakers and the composite signal is sensed by a microphone 34. Microphone 34 responds to both the first signal 28 and the second signal 30 and generates a composite electrical signal 32. This signal is sent to the adaptive network for in-band signal separation 26 which produces two outputs; the first output 36 is a faithful reproduction of the original first signal 28; and a second output signal 38, is a faithful reproduction of the original second input signal 30.
Referring now to FIG. 4, the adaptive network for in-band signal separation 26 is shown in accordance with the preferred embodiment of the present invention. The neural network employed in the preferred embodiment utilizes a neural network known as a multilayer perceptron. As shown in FlG. 10, a multilayer perceptron includes a layer of input neurons 40, one or more layers of inner neurons 42, and a layer of output neurons 44. Ordinarily, in a multilayer perceptron each neuron in each layer is connected to each neuron in the adjacent layers by means of synaptic connections 43 as shown in FIG. 10. Alternatively, the particular interconnection scheme and training algorithm employed, may be according to a number of other neural network architectures including, but not limited to, the Boltzman machine, Counterprop, Hopfield net, Hamming net, etc. It is preferable that the neural network architecture and training algorithm employed belong to the class of supervised, as opposed to unsupervised nets. The particular interconnection scheme and training algorithm employed with the multilayer perceptron and its associated learning algorithm, known as backward error propagation, are well known. Details of the multilayer perceptron are described in Rumelhart, Hinton, and Williams, "Learning Internal Representations by Error of Propagation", in D. E. Rumelhart and J. L. McClelland (Eds.), Parallel Distributed Processing; Explorations in the Microstructure of Cognition, Vol. 1 Foundations, M.I.I. Press (1986), which is incorporated herein by reference.
In accordance with the preferred embodiment, a low frequency composite signal 46 and a high frequency composite signal 48 are both transmitted to the input neurons 40 in the adaptive network 26. The use of the low frequency 46 and high frequency 48 versions of the composite signal 32 permit a reduced number of input neurons 40 to be employed. A large number of inputs is generally considered to be necessary. This is because a high frequency representation of the signal is needed to get a faithful reproduction of the fine structure of the signal; and a broader or lower frequency representation is also needed to give the processor 26 information about more fundamental frequencies, (e.g., pitch) of the speaker. That is, the network should have available the high frequency structure which contains, for example, words and phonemes and the low frequency structure, which contains, for example, the pitch that is characteristic of a given speaker. Thus, one way to give the processor 26 this information is to employ a large number of input neurons 40. This may require, for example, two hundred or more input neurons 40 to give a broad enough sample of the speech data.
In accordance with the preferred embodiment of the present invention, the necessary high and low frequency information can be given to the processor 26 by means of a filter circuit 50 shown in FIG. 9. The filter circuit 50 accepts as input the composite speech signal 32 and generates a high frequency output 48 and a low frequency output 46. It will be appreciated by those skilled in the art that known filtering and sampling techniques may be employed to accomplish the functions of the filtering circuit 50. In accordance with the preferred embodiment, the composite signal 32 is divided into 16 samples at a low frequency, for example, 640 Hertz(Hz). Thus, a sample is taken every 25.6 milliseconds. When added together those samples generate the low frequency composite signal 46. The high frequency samples on the other hand are taken every 1.6 milliseconds, at a rate of 10 kilohertz. Sixteen of the high frequency samples, when combined, generate the high frequency input composite signal 48.
Referring again to FIG. 4, the low frequency composite signal 46 is fed along input line 52 to a series of sampling circuits 54 through an input buffer 55. The low frequency composite signal 46 is fed through the input line 52, to a buffer circuit 55 and the sampling circuits 54 until each of the 16 samples reside in a single sampling circuit 54. Each sampling circuit 54 is connected to an input neuron 40 in the input layer of the processor 26. In similar fashion, the high frequency composite signal 48 is transmitted through an input line 56 through input buffer 58 to a series of sampling circuits 60. It should be noted that while only eight sampling circuits 60 and 8 sample circuits 54 are shown in FIG. 4, there would actually be 16 of the high frequency sampling circuit 60 and 16 of low frequency sample circuits 54, each connected to an input neuron 40. It will be appreciated that depending on the specific application, more or less than 16 samples may be used. Also, the network could be configured to handle more than 2 speakers.
In order to train the processor 26 to perform in-band signal separation, the composite signal 32 consists of a training input signal, which is composed of two known constituent signals such as the first and second constituent signal 28 and 30 shown in FIG. 1. When 16 low frequency and 16 high frequency samples of the composite training signal 32 are fed from the sample circuits 54 and 60 to the input neurons 40, the processor 26 will produce an output at each of its output neurons 44. In accordance with the conventional back-prop training technique employed in the preferred embodiment, the processor 26 is trained with a desired output consisting of high frequency representations of the two constituent signals 28 and 30. In particular, the first 16 output neurons may be presented with the first constituent signal 28 and the next 16 output neurons may be presented with the second constituent signal 30 during training. After a sufficient number of training sessions, the actual output 62, 64 will approximate the desired output. In particular, the first 16 output neurons 44 will approximate the first constituent signal 28 and the next 16 output neurons will approximate the second constituent signal 30. Alternatively, the processor 26 could be trained with only a single training input such as constituent signal 28 for cases where only a single constituent signal is desired. However, where only a single signal is provided by the processor 26, it will be appreciated that this single signal may be separated from the composite and the remaining signal may yield a second constituent signal.
It should also be noted that once the processor 26 is trained for the first 16 high frequency samples and the first 16 low frequency samples, training may continue by repeating the training procedure for the next consecutive 16 high frequency samples, by shifting the high frequency signal over by an amount equal to the distance of 16 high frequency samples. The processor 26 is again trained with this input, until the desired output is achieved to within a predetermined tolerance. This procedure can then be repeated a number of times which will depend on the complexity of the signal and the neural network architecture employed.
Once the network is trained, an unknown composite signal can be presented to the input neurons 40 in the same manner as the training composite signal. That is, a low frequency representation 46 and a high frequency representation 48 of the unknown composite signal is presented to 16 of the input neurons 40 respectively. If the unknown composite signal contains constituent signals 28 and 30, the output of the processor 26 will consist of the first constituent signal 28 from the first 16 output neurons 44 and the second constituent signal 30 from the next 16 output neurons 44. A slower but alternative technique which may be useful in certain applications would be to shift the data over by one high frequency sample at a time rather than 16 samples between training sessions.
Referring now to FIGS. 5-8, illustrations of the output of the adaptive network 26 at various stages in training is shown. In FIG. 5 the "composite input signal" is shown twice in the top row. This composite signal is comprised of signals from two individual speakers, labelled "Speaker 1 Component" and "Speaker 2 Component", shown in the second row. For example, the composite signal may be signal from a microphone responding to two persons (speaker 1 and speaker 2) talking simultaneously. The composite signal is fed to the adaptive network 26 and the network is trained with the known speaker 1 and speaker 2 examples, in accordance with the techniques described above. After one training cycle, the output of the adaptive network 26 appears as shown in the third row of FIG. 5. That is, the output neurons trained with the speaker 1 component produce output signals labelled "Speaker Network Output" and the output neurons trained with the speaker 2 component produce output signals labeled "Speaker 2 Network Output".
After ten training cycles, as shown in FIG. 6, the speaker 1 and speaker 2 network outputs begin to show some significant distinguishing characteristics, particularly the speaker output. FIG. 7 shows the outputs after 100 training cycles, and both the speaker 1 and 2 outputs begin to appear to resemble the original component signals. After 300 cycles, as shown in FIG. 8, the outputs become very good approximations of the original component signals. The exact number of training cycles required will depend upon a number of factors, such as the complexity of the component signals and the desired fidelity of the outputs.
The trained adaptive network 26 can then be used to identify unknown composite signals to restore the original constituent signals if they are contained in the composite. In many cases the speaker and speaker 2 components used for training will be from speech by the same person or persons whose speech is in the unknown composite. It is also possible, however, to use the adaptive processor 26 in accordance with the present invention, to separate speech from unknown speakers. That is, by training an adaptive network 26 of sufficient complexity, a sufficient number of times, it is possible for it to "learn" the general characteristics of human speech so as to separate two examples of such speech from a single composite signal. It will be appreciated that an unsupervised, as opposed to supervised neural net may be preferred for this kind of application.
Once the processor 26 is trained, the weight values developed through training could be transferred to the processor having its weights fixed to none values. In this way, mass production of processors 26 is possible without repeating the training procedure.
It will be appreciated that while the composite signal 32 as shown in FIG. 1 consisted of the amplitudes of the raw signal in the time domain, the above techniques for the processor 26 could be employed in the frequency domain. That is, the input could be a frequency representation of the composite signal and the output also be some frequency representation. In this case, an inverse fourier transform could be used to restore the resultant signal.
In view of the foregoing, those skilled in the art should appreciate that the present invention provides an adaptive network for in-band signal separation 26 that can be used in a wide variety of applications. The various advantages should become apparent to those skilled in the art after having the benefit of studying specification, drawing and the following claims.

Claims (13)

What is claimed is:
1. A signal processor for separating a composite signal into at least one of its constituent signals, said processor comprising:
means for dividing said composite signal into discrete sampled portions;
a plurality of neurons capable of receiving signals and producing an output signal said neurons including input neurons adapted to receive said sampled portions of said composite signal;
a plurality of synaptic connection means providing a weighted interconnection between selected ones of said neurons;
means for training said processor to produce an output that approximates at least one of said constituent signals, said training means including:
(a) means for presenting a composite input training signal to selected ones of said neurons;
(b) means for presenting a desired output, consisting of at least one of said constituent signals, to selected ones of said neurons; and
(c) means for changing the strength of said synaptic connection means to cause said signal processor to produce said desired output in response to said training signal; and
filter means for generating both low frequency and high frequency representations of said composite signal to present to said input neurons, wherein said low frequency representation includes a larger portion of said composite signal than said high frequency representation; and
wherein said desired output consists of high frequency representation of at least one of said constituent signals, whereby said processor is capable of receiving both a high frequency and a low frequency portion of said composite signal during training and during processing.
2. The signal processor of claim 1 wherein said means for changing the strength of said synaptic connection means further comprises:
means for computing the difference between said desired output and the actual output of said neurons, and means for adjusting said synaptic connections so as to minimize the difference between said desired output and the actual output.
3. The signal processor of claim 2 wherein said neurons are arranged in a series of layers including input, inner and output neuron layers forming a multilayer perceptron neural network.
4. The signal processor of claim 3 wherein the total number of said input neurons is at least equal to the number of said low frequency plus high frequency samples, and wherein the total number of output neurons is at least equal to the number of said high frequency samples.
5. The signal processor of claim 3 further comprising means for advancing said input signal through said input neurons in a stepwise fashion so that the entire composite signal from beginning to end is directed to each of said input neurons and an output is produced during each step of said stepwise advancing.
6. The signal processor of claim 1 wherein said composite signal received by the processor is a signal from a sensor, wherein the amplitude of the signal varies over time.
7. The signal processor of claim 6 wherein said constituent signals comprise speech signals.
8. The signal processor of claim 1 wherein said composite signal received by the processor is a signal from a sensor which is represented in the frequency domain.
9. The signal processor of claim 1 wherein said composite signal has two constituent signals and said desired output consists of both of said two constituent signals.
10. The signal processor of claim 1 wherein said constituent signals overlap each other in the frequency spectrum.
11. A neural network for separating a composite signal into at least one of its constituent signals, said network comprising:
means for dividing said composite signal into discrete sampled portions;
a plurality of neurons capable of receiving signals and producing an output signal said neurons including input neurons adapted to receive said sampled portions of said composite signal;
a plurality of synaptic connection means providing a weighted interconnection between selected ones of said neurons, said weighted connections being fixed and derived from a separate processor said separate processor including;
means for dividing said composite signal into discrete sampled portion;
a plurality of neurons capable of receiving signals and producing an output signal, said neurons including input neurons adapted to receive said sampled portions of said composite signal;
a plurality of synaptic connection means providing a weighted interconnection between selected ones of said neurons;
means for training said processor to produce an output that approximates at least one of said constituent signals, said training means including:
(a) means for presenting a composite input training signal to said neurons;
(b) means for presenting a desired output, consisting of at least one of said constituent signals, to selected ones of said neurons; and
filter means for generating both low frequency and high frequency representations of said composite signal to present to said input neurons, wherein said low frequency representation includes a larger portion of said composite signal than said high frequency representation; and
wherein said desired output consists of high frequency representation of at least one of said constituent signals, whereby said network is capable of receiving both a high frequency and a low frequency portion of said composite signal during training and during processing.
12. A method for separating a composite signal into at least one of its constituent signals said method comprising:
dividing said composite signals into discrete sample portions;
receiving said sampled portions in a plurality of neurons, said neurons including input neurons adapted to receive said sampled portions of said composite signal; and
said neurons including output neurons adapted to produce output signals;
providing a weighted interconnection between selected ones of said neurons;
training said plurality of interconnected neurons to produce an output that approximates at least one of said constituent signals, said training including the steps of:
(a) presenting a composite input training signal to said neurons;
(b) presenting a desired output, consisting of at least one of said constituent signals, to selected ones of said neurons; and
(c) changing the strength of said synaptic connection means to cause said output neurons to produce said desired output in response to said training signal; and
generating both low frequency and high frequency representations of said composite signal to present to said input neurons, wherein said low frequency representation includes a larger portion of said composite signal than said high frequency representation; and
wherein said output neurons produce an output that consists of a high frequency representation of at least one of said constituent signals, whereby said input neurons are capable of receiving both a high frequency and a low frequency portion of said composite signal during training and during processing.
13. A neural network produced in accordance with the method of claim 12.
US07/392,681 1989-08-11 1989-08-11 Adaptive network for in-band signal separation Expired - Lifetime US5150323A (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US07/392,681 US5150323A (en) 1989-08-11 1989-08-11 Adaptive network for in-band signal separation
JP2512242A JPH04501330A (en) 1989-08-11 1990-08-09 Adaptive network for in-band signal separation
PCT/US1990/004488 WO1991002324A1 (en) 1989-08-11 1990-08-09 Adaptive network for in-band signal separation
EP90912714A EP0439592A1 (en) 1989-08-11 1990-08-09 Adaptive network for in-band signal separation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US07/392,681 US5150323A (en) 1989-08-11 1989-08-11 Adaptive network for in-band signal separation

Publications (1)

Publication Number Publication Date
US5150323A true US5150323A (en) 1992-09-22

Family

ID=23551587

Family Applications (1)

Application Number Title Priority Date Filing Date
US07/392,681 Expired - Lifetime US5150323A (en) 1989-08-11 1989-08-11 Adaptive network for in-band signal separation

Country Status (4)

Country Link
US (1) US5150323A (en)
EP (1) EP0439592A1 (en)
JP (1) JPH04501330A (en)
WO (1) WO1991002324A1 (en)

Cited By (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5265192A (en) * 1990-09-20 1993-11-23 Atlantic Richfield Company Method for the automated editing of seismic traces using an adaptive network
US5345539A (en) * 1990-11-02 1994-09-06 The Secretary Of State For Defence In Her Britannic Majesty's Government Of The United Kingdom Of Great Britain And Northern Ireland Radar apparatus using neural network for azimuth and elevation detection
US5408424A (en) * 1993-05-28 1995-04-18 Lo; James T. Optimal filtering by recurrent neural networks
US5448503A (en) * 1992-07-31 1995-09-05 The United States Of America As Represented By The Secretary Of The Navy Acoustic monitor
US5598508A (en) * 1991-10-18 1997-01-28 Goldman; Julian M. Real-time waveform analysis using artificial neural networks
US5630019A (en) * 1992-05-23 1997-05-13 Kabushiki Kaisha Topcon Waveform evaluating apparatus using neural network
US5638491A (en) * 1992-06-19 1997-06-10 United Parcel Service Of America, Inc. Method and apparatus for hierarchical input classification using a neural network
US5649065A (en) * 1993-05-28 1997-07-15 Maryland Technology Corporation Optimal filtering by neural networks with range extenders and/or reducers
US5737485A (en) * 1995-03-07 1998-04-07 Rutgers The State University Of New Jersey Method and apparatus including microphone arrays and neural networks for speech/speaker recognition systems
US5742740A (en) * 1990-09-20 1998-04-21 Atlantic Richfield Company Adaptive network for automated first break picking of seismic refraction events and method of operating the same
US5787393A (en) * 1992-03-30 1998-07-28 Seiko Epson Corporation Speech recognition apparatus using neural network, and learning method therefor
US5825671A (en) * 1994-03-16 1998-10-20 U.S. Philips Corporation Signal-source characterization system
US5963929A (en) * 1993-05-28 1999-10-05 Maryland Technology Corporation Recursive neural filters
US5987444A (en) * 1997-09-23 1999-11-16 Lo; James Ting-Ho Robust neutral systems
US6070139A (en) * 1995-08-21 2000-05-30 Seiko Epson Corporation Bifurcated speaker specific and non-speaker specific speech recognition method and apparatus
US6151592A (en) * 1995-06-07 2000-11-21 Seiko Epson Corporation Recognition apparatus using neural network, and learning method therefor
US6236862B1 (en) 1996-12-16 2001-05-22 Intersignal Llc Continuously adaptive dynamic signal separation and recovery system
US6304865B1 (en) 1998-10-27 2001-10-16 Dell U.S.A., L.P. Audio diagnostic system and method using frequency spectrum and neural network
EP1146200A1 (en) 2000-04-15 2001-10-17 Schlumberger Holdings Limited Drill bit design using neural networks
US6601051B1 (en) 1993-08-09 2003-07-29 Maryland Technology Corporation Neural systems with range reducers and/or extenders
US6654632B2 (en) 2000-07-06 2003-11-25 Algodyne, Ltd. System for processing a subject's electrical activity measurements
US6898582B2 (en) 1998-12-30 2005-05-24 Algodyne, Ltd. Method and apparatus for extracting low SNR transient signals from noise
US20070135726A1 (en) * 2005-12-08 2007-06-14 Shenzhen Mindray Bio-Medical Electronics Co., Ltd. Method for improving recognition rate of respiratory wave
US10249305B2 (en) 2016-05-19 2019-04-02 Microsoft Technology Licensing, Llc Permutation invariant training for talker-independent multi-talker speech separation
EP3469584A4 (en) * 2016-06-14 2020-01-08 The Trustees of Columbia University in the City of New York Neural decoding of attentional selection in multi-speaker environments
CN111984242A (en) * 2020-08-20 2020-11-24 中电科仪器仪表有限公司 Method and system for decomposing synthesized signal
US10957337B2 (en) 2018-04-11 2021-03-23 Microsoft Technology Licensing, Llc Multi-microphone speech separation
US11373672B2 (en) 2016-06-14 2022-06-28 The Trustees Of Columbia University In The City Of New York Systems and methods for speech separation and neural decoding of attentional selection in multi-speaker environments

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2679083A1 (en) * 1991-07-09 1993-01-15 Thomson Csf Automatic device and method for separating sources
FR2730881A1 (en) * 1995-02-22 1996-08-23 Philips Electronique Lab SYSTEM FOR ESTIMATING SIGNALS RECEIVED IN THE FORM OF MIXED SIGNALS

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3287649A (en) * 1963-09-09 1966-11-22 Research Corp Audio signal pattern perception device
US4660166A (en) * 1985-01-22 1987-04-21 Bell Telephone Laboratories, Incorporated Electronic network for collective decision based on large number of connections between signals
US4719591A (en) * 1985-11-07 1988-01-12 American Telephone And Telegraph Company, At&T Bell Labs. Optimization network for the decomposition of signals
US5003490A (en) * 1988-10-07 1991-03-26 Hughes Aircraft Company Neural network signal processor

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4885757A (en) * 1987-06-01 1989-12-05 Texas Instruments Incorporated Digital adaptive receiver employing maximum-likelihood sequence estimation with neural networks

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3287649A (en) * 1963-09-09 1966-11-22 Research Corp Audio signal pattern perception device
US4660166A (en) * 1985-01-22 1987-04-21 Bell Telephone Laboratories, Incorporated Electronic network for collective decision based on large number of connections between signals
US4719591A (en) * 1985-11-07 1988-01-12 American Telephone And Telegraph Company, At&T Bell Labs. Optimization network for the decomposition of signals
US5003490A (en) * 1988-10-07 1991-03-26 Hughes Aircraft Company Neural network signal processor

Non-Patent Citations (28)

* Cited by examiner, † Cited by third party
Title
"Brainmaker--Users Guide and Reference Manual", California Scientific Software, (1989), p. 303.
A. Bayle, "Learning in Neural Networks", PC AI, Nov./Dec. (1988), p. 40.
A. Bayle, Learning in Neural Networks , PC AI, Nov./Dec. (1988), p. 40. *
Brainmaker Users Guide and Reference Manual , California Scientific Software, (1989), p. 303. *
C. A. Will, "The DARPA Neural Network Study", Neural Network Review, vol. 2, No. 3, (1988), pp. 74-85.
C. A. Will, The DARPA Neural Network Study , Neural Network Review, vol. 2, No. 3, (1988), pp. 74 85. *
Chien, D., Featherston, M., Min, K., and Rogers, C., "Speech Separation with Neural Networks", 1988.
Chien, D., Featherston, M., Min, K., and Rogers, C., Speech Separation with Neural Networks , 1988. *
D. Schwartz, M. Jurik, "Neural Nets on a Personal Computer--PC Neural Network Products Come of Age", PC AI, Nov./Dec. (1988), p. 37.
D. Schwartz, M. Jurik, Neural Nets on a Personal Computer PC Neural Network Products Come of Age , PC AI, Nov./Dec. (1988), p. 37. *
J. Stanley, "Introduction to Neural Networks", pp. 119-133, 191-235 and 247, California Scientific Software, (1989).
J. Stanley, Introduction to Neural Networks , pp. 119 133, 191 235 and 247, California Scientific Software, (1989). *
Lippmann, Richard P., "An Introduction to Computing with Neural Nets", IEEE ASSP Magazine, Apr. 1987, pp. 4-22.
Lippmann, Richard P., An Introduction to Computing with Neural Nets , IEEE ASSP Magazine, Apr. 1987, pp. 4 22. *
M. Caudill, "Neural Networks Primer--Part II", AI Expert, p. 53, Jun. (1988).
M. Caudill, "Neural Networks Primer--Part III", AI Expert, p. 53, Jun. (1988).
M. Caudill, Neural Networks Primer Part II , AI Expert, p. 53, Jun. (1988). *
M. Caudill, Neural Networks Primer Part III , AI Expert, p. 53, Jun. (1988). *
NeuralWare Inc. Brochure, "NeuralWorks Explorer", (1988).
NeuralWare Inc. Brochure, NeuralWorks Explorer , (1988). *
P. D. Wasserman above, T. Schwartz, "Neural Networks, Part II: What are They and Why is Everybody so Interested in Them Now?", I.E.E.E. Expert, 3, (1988), No. 1, pp. 10-15.
P. D. Wasserman above, T. Schwartz, Neural Networks, Part II: What are They and Why is Everybody so Interested in Them Now , I.E.E.E. Expert, 3, (1988), No. 1, pp. 10 15. *
P. D. Wasserman, "Neural Computing Theory and Practice", Chapter 3, p. 43, (Van Nostrand Reinhold), Apr. (1989).
P. D. Wasserman, Neural Computing Theory and Practice , Chapter 3, p. 43, (Van Nostrand Reinhold), Apr. (1989). *
Rogers C., Chien, D., Featherston, M., and Min, K., "Neural Network Enhancement for a Two Speaker Separation System".
Rogers C., Chien, D., Featherston, M., and Min, K., Neural Network Enhancement for a Two Speaker Separation System . *
Rumelhardt, D. E., Hinton, G. E. and Williams, R. J., "Learning Internal Representations by Error Propagation", Parallel Distributed Processing Explorations in the Microstructure of Cognition, vol. 1, Chapter 8, pp. 318-362, date (unknown).
Rumelhardt, D. E., Hinton, G. E. and Williams, R. J., Learning Internal Representations by Error Propagation , Parallel Distributed Processing Explorations in the Microstructure of Cognition, vol. 1, Chapter 8, pp. 318 362, date (unknown). *

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5742740A (en) * 1990-09-20 1998-04-21 Atlantic Richfield Company Adaptive network for automated first break picking of seismic refraction events and method of operating the same
US5265192A (en) * 1990-09-20 1993-11-23 Atlantic Richfield Company Method for the automated editing of seismic traces using an adaptive network
US5345539A (en) * 1990-11-02 1994-09-06 The Secretary Of State For Defence In Her Britannic Majesty's Government Of The United Kingdom Of Great Britain And Northern Ireland Radar apparatus using neural network for azimuth and elevation detection
US5751911A (en) * 1991-10-18 1998-05-12 Goldman; Julian M. Real-time waveform analysis using artificial neural networks
US5598508A (en) * 1991-10-18 1997-01-28 Goldman; Julian M. Real-time waveform analysis using artificial neural networks
US5787393A (en) * 1992-03-30 1998-07-28 Seiko Epson Corporation Speech recognition apparatus using neural network, and learning method therefor
US5809461A (en) * 1992-03-30 1998-09-15 Seiko Epson Corporation Speech recognition apparatus using neural network and learning method therefor
US5630019A (en) * 1992-05-23 1997-05-13 Kabushiki Kaisha Topcon Waveform evaluating apparatus using neural network
US5664067A (en) * 1992-06-19 1997-09-02 United Parcel Service Of America, Inc. Method and apparatus for training a neural network
US5638491A (en) * 1992-06-19 1997-06-10 United Parcel Service Of America, Inc. Method and apparatus for hierarchical input classification using a neural network
US5974404A (en) * 1992-06-19 1999-10-26 United Parcel Service Of America, Inc. Method and apparatus for input classification using a neural network
US5448503A (en) * 1992-07-31 1995-09-05 The United States Of America As Represented By The Secretary Of The Navy Acoustic monitor
US5649065A (en) * 1993-05-28 1997-07-15 Maryland Technology Corporation Optimal filtering by neural networks with range extenders and/or reducers
US5408424A (en) * 1993-05-28 1995-04-18 Lo; James T. Optimal filtering by recurrent neural networks
US5963929A (en) * 1993-05-28 1999-10-05 Maryland Technology Corporation Recursive neural filters
US6601051B1 (en) 1993-08-09 2003-07-29 Maryland Technology Corporation Neural systems with range reducers and/or extenders
US5825671A (en) * 1994-03-16 1998-10-20 U.S. Philips Corporation Signal-source characterization system
US5737485A (en) * 1995-03-07 1998-04-07 Rutgers The State University Of New Jersey Method and apparatus including microphone arrays and neural networks for speech/speaker recognition systems
US6151592A (en) * 1995-06-07 2000-11-21 Seiko Epson Corporation Recognition apparatus using neural network, and learning method therefor
US6070139A (en) * 1995-08-21 2000-05-30 Seiko Epson Corporation Bifurcated speaker specific and non-speaker specific speech recognition method and apparatus
US6236862B1 (en) 1996-12-16 2001-05-22 Intersignal Llc Continuously adaptive dynamic signal separation and recovery system
US5987444A (en) * 1997-09-23 1999-11-16 Lo; James Ting-Ho Robust neutral systems
US6304865B1 (en) 1998-10-27 2001-10-16 Dell U.S.A., L.P. Audio diagnostic system and method using frequency spectrum and neural network
US6898582B2 (en) 1998-12-30 2005-05-24 Algodyne, Ltd. Method and apparatus for extracting low SNR transient signals from noise
EP1146200A1 (en) 2000-04-15 2001-10-17 Schlumberger Holdings Limited Drill bit design using neural networks
US6654632B2 (en) 2000-07-06 2003-11-25 Algodyne, Ltd. System for processing a subject's electrical activity measurements
US6768920B2 (en) 2000-07-06 2004-07-27 Algodyne, Ltd. System for delivering pain-reduction medication
US6826426B2 (en) 2000-07-06 2004-11-30 Algodyne, Ltd. Objective pain signal acquisition system and processed signal
US6751499B2 (en) 2000-07-06 2004-06-15 Algodyne, Ltd. Physiological monitor including an objective pain measurement
US20070135726A1 (en) * 2005-12-08 2007-06-14 Shenzhen Mindray Bio-Medical Electronics Co., Ltd. Method for improving recognition rate of respiratory wave
US8152733B2 (en) * 2005-12-08 2012-04-10 Shenzhen Mindray Bio-Medical Electronics Co., Ltd. Method for improving recognition rate of respiratory wave
US10249305B2 (en) 2016-05-19 2019-04-02 Microsoft Technology Licensing, Llc Permutation invariant training for talker-independent multi-talker speech separation
EP3469584A4 (en) * 2016-06-14 2020-01-08 The Trustees of Columbia University in the City of New York Neural decoding of attentional selection in multi-speaker environments
US11373672B2 (en) 2016-06-14 2022-06-28 The Trustees Of Columbia University In The City Of New York Systems and methods for speech separation and neural decoding of attentional selection in multi-speaker environments
US10957337B2 (en) 2018-04-11 2021-03-23 Microsoft Technology Licensing, Llc Multi-microphone speech separation
CN111984242A (en) * 2020-08-20 2020-11-24 中电科仪器仪表有限公司 Method and system for decomposing synthesized signal

Also Published As

Publication number Publication date
EP0439592A1 (en) 1991-08-07
WO1991002324A1 (en) 1991-02-21
JPH04501330A (en) 1992-03-05

Similar Documents

Publication Publication Date Title
US5150323A (en) Adaptive network for in-band signal separation
US5003490A (en) Neural network signal processor
US6038338A (en) Hybrid neural network for pattern recognition
US5095443A (en) Plural neural network system having a successive approximation learning method
US5285522A (en) Neural networks for acoustical pattern recognition
US3287649A (en) Audio signal pattern perception device
Waibel Modular construction of time-delay neural networks for speech recognition
CA2642041C (en) Spatio-temporal pattern recognition using a spiking neural network and processing thereof on a portable and/or distributed computer
Pandya et al. Pattern recognition with neural networks in C++
WO2006000103A1 (en) Spiking neural network and use thereof
GB2245401A (en) Neural network signal processor
RU2193797C2 (en) Content-addressable memory device (alternatives) and image identification method (alternatives)
WO1991002323A1 (en) Adaptive network for classifying time-varying data
KR100306848B1 (en) A selective attention method using neural networks
Polepalli et al. Digital neuromorphic design of a liquid state machine for real-time processing
Medhat et al. Automatic classification of music genre using masked conditional neural networks
Hu et al. An features extraction and recognition method for underwater acoustic target based on ATCNN
US6560582B1 (en) Dynamic memory processor
Chinmayi et al. Emotion Classification Using Deep Learning
JPH0581227A (en) Neuron system network signal processor and method of processing signal
EP0438573A1 (en) Pattern propagation neural network
Ghosh et al. Classification of spatiotemporal patterns with applications to recognition of sonar sequences
AU620959B2 (en) Neural network signal processor
Harpur et al. Experiments with simple Hebbian-based learning rules in pattern classification tasks
Vershkov et al. Optimization of computational complexity of an artificial neural network.

Legal Events

Date Code Title Description
AS Assignment

Owner name: HUGHES AIRCRAFT COMPANY, LOS ANGELES, CA A CORP. O

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST.;ASSIGNOR:CASTELAZ, PATRICK F.;REEL/FRAME:005112/0332

Effective date: 19890808

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

AS Assignment

Owner name: HUGHES ELECTRONICS CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HE HOLDINGS INC.;HUGHES ELECTRONICS, FORMERLY KNOWN AS HUGHES AIRCRAFT COMPANY;REEL/FRAME:009342/0796

Effective date: 19971217

FPAY Fee payment

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 12