• 沒有找到結果。

Chapter 7 Conclusions and Future researches

7.2. Future researches

To improve the current speech enhancement system, this dissertation proposes two

with the speech recognizer, and the second one is to combine multiple speakers’

location detection approach with the proposed SPFDBB or FDABB.

Currently, the proposed reference-signal-based frequency-domain beamformers are performed in two independent phases: speech purification and then recognition as shown in Figs. 1-5, 3-1 and 4-2. The proposed beamformers designed to reduce the speech distortion and suppress the noise effects assume that improving the quality of the speech waveform will result in better recognition performance and are independent of the recognition system. Although the proposed beamformers can conquer many practical issues, the beamformers still cannot compete with the microphone in a close distance in terms of the ASR rates. Generally, a speech recognizer is a statistical pattern classifier that operates on a sequence of features derived from the waveform. To increase the recognition accuracy in distant-talking environments, the architecture of connecting the proposed beamformers and the speech recognizer as shown in Fig. 7-1 is worth a further study in the further. This architecture enables the beamformer to use the data transmitted from the recognizer and ensures the beamformer enhances those signal components important for ASR. In other words, this architecture enables the designed filters not to undue emphasis on unimportant components.

For example, Seltzer et al. [124-125] proposed a likelihood-maximizing beamformer (LIMABEAM) that integrates the speech recognition system into the filter design process. They proved that incorporating the statistical models of the recognizer into the array processing stage can improve the ASR rates. The goal of the LIMABEAM is not to generate an enhanced output waveform but rather to generate a sequence of features which maximizes the likelihood of generating the correct hypothesis. In other words, the filter coefficient vectors are chosen to maximize the likelihood of the training signal as measured by the recognizer, rather than to improve its SNR or perceptual quality.

Nishiura et al. [128] also proposed a method which combines HMM model and microphone to enhance the speech recognition further.

M M M

Figure 7-1 Speech enhancement system with a combination of beamformer and recognizer

Furthermore, under the same architecture, it is highly possible to combine the proposed speaker’s location detection approach and SPFDBB or FDABB to significantly reduce the possibility of the wrong operation of SPFDBB or FDABB. The wrong operation means that the unmodeled or unexpected speech signal triggers the VAD to switch the beamformers to the speech stage, and thus obtain an error output.

According to the system architecture of frequency-domain beamformers in Chapters 3 and 4, this wrong operation is unavoidable due to the limitation of VAD. However, the proposed multiple speakers’ locations detection approach in Chapter 5, which can detect the unmodeled sound signals, is highly possible to avoid the wrong operation.

Therefore, the unmodeled speech signal could hardly cause the error output under the new system architecture in Fig. 7-2. Based on the system architecture, both the VAD result and speaker’s location detection system are combined to decide the stage of the SPFDBB or FDABB. It means that if the received sound signal is detected as containing speech signal and coming from one of the desired locations, then the system is switched to the speech stage. Consequently, it is also worthwhile to study how to construct the two components with a suitable integration procedure. Figure 7-3 shows

)

Figure 7-2 Overall system structure which integrates the speaker’s location detection approach and SPFDBB or FDABB

Start

Silent stage of the multiple speaker location detection system

Silent stage of SFPDBB or

FDABB Speech stage of the multiple

speaker location detection system

The speech signal comes from one of the

desired location?

Yes

Speech stage of SPFDBB or FDABB

Automatic speech recognizer Silent stage of

SPFDBB or FDABB

No

Figure 7-3 Flowchart of the architecture which integrates the speaker’s location detection approach and SPFDBB or FDABB

Reference

[1] S. F. Boll, “Suppression of acoustic noise in speech using spectral subtraction,”

IEEE Trans. on Acoust. Speech, Signal Processing, vol. ASSP-27, pp. 113-120, Apr. 1979.

[2] A. Kawamura, Y. Iiguni, and Y. Itoh, “A noise reduction method based on linear prediction with variable step-size,” IEICE Tran. Fundamentals, vol. E88-A, no. 4, pp. 855-861, April 2005.

[3] J. G. Ryan, R. A. Goubran, “Near-field beamforming for microphone arrays,”

IEEE Conference on Acoustics, Speech, and Signal Processing, vol. 1, pp. 21-24, Apr. 1997.

[4] P. L. Chu, “Superdirective microphone array for a set-top video conferencing system,” IEEE International Conference on Acoust., Speech, and Signal Processing, vol. I, pp. 235-238, Apr. 1999.

[5] M. Wax and T. Kailath, “Optimal localization of multiple sources by passive arrays,” IEEE Trans. on Acoust. Speech, Signal Processing, vol. ASSP-31, pp.

1210-1217, Oct. 1983.

[6] H. F. Silverman and S. E. Kirtman, “A two-stage algorithm for determining talker location from linear microphone-array data,” Computer, Speech, and Language, vol. 6, pp. 129-152, Apr. 1992.

[7] D. B. Ward, and R. C. Williamson, “Particle filter beamforming for acoustic source localization in a reverberant environment,” International Conference on Acoustics, Speech, and Signal Processing, vol. 2, pp. 1777-1780, May 2002.

[8] R. V. Balan and J. Rosca, “Apparatus and method for estimating the direction of Arrival of a source signal using a microphone array,” European Patent, No US2004013275, 2004.

[9] M. Wax, T. J. Shan, and T. Kailath. “ Spatio-temporal spectral analysis by eigenstructure methods,” IEEE Trans. on Acoust. Speech, Signal Processing, vol ASSP-32, pp. 817–827, Aug 1984.

[10] R. O. Schmidt, “Multiple emitter location and signal parameter estimation,”

IEEE Trans. on Antennas and Propagation, vol. 34, pp. 276-280. Mar 1986.

[11] H. Wang and M. Kaveh, “Coherent signal subspace processing for detection and estimation of angle of arrival of multiple wideband sources,” IEEE Trans. on Acoust. Speech, Signal Processing, vol ASSP-33, pp. 823–831, Aug. 1985.

[12] G. Bienvenu, “Eigensystem properties of the sampled space correlation matrix,”

IEEE International Conference on Acoust., Speech, and Signal Processing, pp.

332–335, 1983.

[13] K. M. Buckley and L. J. Griffiths, “Eigenstructure based broadband source location estimation,” IEEE International Conference on Acoust., Speech, and Signal Processing, pp. 1869–1872, 1986.

[14] M. A. Doron, A. J. Weiss, and H. Messer, “Maximum likelihood direction finding of wideband sources,” IEEE Trans. on Signal Processing, vol. 41, pp.

411–414, Jan 1993.

[15] M. Agarwal and S. Prasad, “DOA estimation of wideband sources using a harmonic source model and uniform linear array,” IEEE Trans. on Signal Processing, vol. 47, pp. 619-629, Mar. 1999.

[16] H. Messer, “The potential performance gain in using spectral information in passive detection/localization of wideband sources,” IEEE Trans. on Signal Processing, vol. 43, pp. 2964-2974, Dec. 1995.

[17] M. Agrawal and S. Prasad, “Broadband DOA estimation using spatial-only modeling of array data,” IEEE Trans. on Signal Processing, vol. 48, pp. 663-670, Mar. 2000.

[18] J.-H Lee, Y.-M Chen, and C.-C Yeh, “A covariance approximation method for near-field direction finding using a uniform linear array,” IEEE Trans. on Signal Processing, vol. 43, pp. 1293-1298, May 1995.

[19] K. Buckley and L. Griffiths, “Broad-band signal-subspace spatial-spectrum (BASS-ALE) estimation,” IEEE Trans. on Acoust. Speech, Signal Processing, vol.

36, pp. 953-964, July 1988.

[20] N. Strobel and R. Rabenstein, “Classification of time delay estimates for robust speaker localization,” International Conference on Acoustics, Speech, and Signal Processing, vol. 6, pp. 15-19, March 1999.

[21] G. C. Carter, A. H. Nuttall, and P. G. Cable, “The smoothed coherence transform(SCOT),” Naval Underwater Systems Center, New London Lab., New London, CT, Tech. Memo TC-159-72, Aug 8 1972.

[22] C. H. Knapp, and G. C. Carter, “The generalized correlation method for estimation of time delay,” IEEE Trans. on Acoust., Speech, Signal Processing, vol.

24, pp. 320-327, Aug 1976.

[23] G.. C. Carter, A. H. Nuttall, and P. G. Cable, “The smoothed coherence transform,” IEEE Signal Processing Letters, vol. 61, pp. 1497-1498, Oct 1973.

[24] J. Hu, T.M. Su, C.C. Cheng, W.H. Liu, and T.I. Wu, “A self-calibrated speaker tracking system using both audio and video data,” IEEE Conference on Control Applications, vol.2, pp. 731-735, Sept 2002.

[25] J. Hu, C.C. Cheng, W.H. Liu, and T.M. Su, “A speaker tracking system with distance estimation using microphone array,” IEEE/ASME International Conference on Advanced Manufacturing Technologies and Education, Aug 2002.

[26] S. Mavandadi, P. Aarabi, “Multichannel nonlinear phase analysis for time-frequency data fusion,” Proceedings of the SPIE, Architectures, Algorithms, and Applications VII (AeroSense 2003), vol. 5099, pp. 222-231, April 2003.

[27] P. Aarabi and S. Mavandadi, “Robust sound localization using conditional time–frequency histograms,” Information Fusion, vol. 4, pp. 111-122, June 2003.

[28] C. Winter, “Using continuous apertures discretely,” IEEE Trans. on Antennas and Propagation, vol. AP-25, pp. 695-700, Sep. 1977.

[29] G. W. Elko, T. C. Chou, R. J. Lustberg, M. M. Goodwin, “A constant-directivity beamforming microphone array,” J. Acoust. Soc. Amer., vol. 96, no. 5, pp.3244, Nov. 1994.

[30] J. H. Doles III and F. D. Benedict, “Broad-band array design using the asymptotic theory of unequally spaced arrays,” IEEE Trans. on on Antennas and Propagation, vol. 36, no. 1, pp. 27-33, Jan. 1988.

[31] A. Ishimaru and Y. S. Chen, “Thinning and broadbanding antenna arrays by unequal spacings,” IEEE Trans. on Antennas and Propagation, vol. AP-13, pp.

34-42, Jan. 1965.

[32] T. Chou, “Frequency-independent beamformer with low response error,” IEEE Trans. on Speech and Audio Processing, pp. 2995-2998, May 1995.

[33] B. D. Van Veen and K. M. Buckley, “Beamforming: a versatile approach to spatial filtering,” IEEE Acoustic, Speech, Signal Processing Magazine, pp. 4-24, Apr. 1988.

[34] M. Doerbecker, “Speech enhancement using small microphone array with optimized directivity,” International Workshop on Acoustic, Echo, and Noise Control, pp. 100-103, Sep. 1997.

[35] J. G. Ryan and R. A. Goubran, “Optimal near-field response for microphone arrays,” IEEE International Workshop on Application Signal Processing to Audio Acoustics, pp. 19-22, Oct. 1997.

[36] J. Bitzer, K. U. Simmer, and K. D. Kammeyer, “An alternative implementation of the superdierective beamformer,”in Proc. IEEE International Workshop on Application Signal Processing to Audio Acoustics, pp. 7-10, New Paltz, NY, USA, Oct. 1997.

[37] R. Zelinski, “Noise reduction based on microphone array with LMS adaptive post-filtering,” Electronic letter, vol. 26, pp. 2036-2037, Nov. 1990.

[38] M. Dorbecker and S. Ernst, “Combination of two-channel spectral subtraction and adaptive Wiener post-filtering for noise reduction and dereverberation,”

European Signal Processing Conference, pp. 995-998, Sep. 1996.

[39] C. Marro, Y. Mahieux, and K. U. simmer, “Analysis of noise reduction and dereverberation techniques based on microphone arrays with postfiltering,” IEEE Trans. on Acoustic, Speech, Signal Processing, vol. 6, pp. 240-259, May 1998.

[40] I. a. McCowan, C. Marro, and L. Mauuary, “Robust speech recognition using near-field superdirective beamforming with post-filtering,” IEEE International Conference on Acoust., Speech, and Signal Processing, pp. 1723-1726, Apr. 2000.

[41] S. Doclo and M. Moonen, “SVD-based optimal filtering with applications to noise reduction in speech signals,” IEEE International Workshop on Application Signal Processing to Audio Acoustics pp. 143-146, Oct. 1999.

[42] S. Doclo and M. Moonen, “Robustness of SVD-based optimal filtering noise reduction in multi-microphone speech signals,” IEEE International Workshop on Acoustic echo and noise control, pp. 80-83, Sept. 1999.

[43] F. Asano, S. Hayamizu, T. Yamada, and S. Nakamura, “Speech enhancement based on the subspace method,” IEEE Trans. on Acoust. Speech, Signal Processing, vol. 8, no. 5, pp. 497-507, Sept. 2000.

[44] S. Doclo, E. De Clippel, and M. Moonen, “Combined acoustic echo and noise reduction using GSVD-based optimal filtering,” IEEE International Conference on Acoust., Speech, and Signal Processing, vol. 2, pp. 1061-1064, June 2000.

[45] O.L. Frost, “An Algorithm for Linear Constrained Adaptive Array Processing,”

Proc. IEEE, vol.60, no.8, pp.926-935, Aug. 1972.

[46] L.J. Griffiths and C.W. Jim, “An alternative approach to linearly constrained adaptive beamforming,” IEEE Trans. on Antennas Propagation, vol. AP-30, pp.

27-34, Jan. 1982.

[47] N.K. Jablon, “Adaptive beamforming with the generalized sidelobe canceller in the presence of array imperfections,” IEEE Trans. on Antennas Propagation., vol.

AP-34, pp. 996–1012, Aug. 1986.

[48] I. Claessen, S. Nordholm, B. A. Begtsson, and P. Eriksson, “A multi-DSP implementation of a broad-band adaptive beamformer for use in a hands-free mobile radio telephone,” IEEE Trans. on Vehicular Tech., vol. 40, no. 1, pp.

194-202, Feb. 1991.

[49] Y. Grenier, “A microphone array for car environments,” Speech Communication, vol. 12, no. 1, pp. 25-39, Mar. 1993.

[50] M. Dahl, I. Claesson, and S. Nordebo, “Simultaneous echo cancellation and car noise suppression employing a microphone array,” IEEE International

Conference on Acoust., Speech, and Signal Processing, vol. I, pp. 239-242, Apr.

1997.

[51] P. M. Peterson, “Using linearly-constrained adaptive beamforming to reduce interference in hearing aids from competing talkers in reverberant rooms,” IEEE International Conference on Acoust., Speech, and Signal Processing, pp.

2364-2367, Apr. 1987.

[52] P. M. Zurek, J. E. Greenberg, and P. M. Peterson, “Sensitivity to design parameters in an adaptive-beamforming hearing aid,” IEEE International Conference on Acoust., Speech, and Signal Processing, A1.10, pp. 1129-1132, Apr. 1990.

[53] W. Soede, F. Bilson, and A. J. Berkhout, “Assignment of a directional microphone array for hearing-impaired listeners” J. Acoust. Soc. Amer., vol. 94, no. 2, pp. 799-808, Aug. 1993.

[54] J. M. Kates, “Superdirective arrays for hearing aids,” J. Acoust. Soc. Amer., vol.

94, no. 2, pp. 1930-1933, Oct. 1993.

[55] J. M. Kates, “A comparison of hearing-aid array-processing techniques,” J.

Acoust. Soc. Amer., vol. 99, no. 5, pp. 313803148, May 1996.

[56] A. Wang, K. Yao, R. E. Hudson, D. Korompis, S. F. Soli, and S. Gao, “A high performance microphone array system for hearing aid applications,” IEEE International Conference on Acoust., Speech, and Signal Processing, pp.

3197-3200, May 1996.

[57] K. L. Bell, Y. Ephraim, and H. L. Van Trees. “A Bayesian approach to A Bayesian approach to robust adaptive beamforming,” IEEE Trans. on Signal Processing, vol. 48, pp. 386–398, Feb. 2000.

[58] C. Henry, “Robust Adaptive Beamforming,” IEEE Trans. on Acoust. Speech, Signal Processing, vol. ASSP-35, pp. 1365-1376, Oct. 1987.

[59] O. Hoshuyama, A. Sugiyama, and A. Hirano, “A Robust Adaptive Beamformer for Microphone Arrays with Blocking Matrix Using Constrained Adaptive Filters,” IEEE Trans. on Signal Processing, vol. 47, no. 10, Oct 1999.

[60] S. A. Vorobyov, A. B. Gershman, Z. Q. Luo, “Robust adaptive beamforming using worst-case performance optimization: a solution to the signal mismatch problem,” IEEE Trans. on Signal Processing, vol. 51, pp. 313-324, Feb. 2003.

[61] M. Dahl, and I. Claesson, “Acoustic noise and echo canceling with microphone array”, IEEE Trans. on Vehicular Technology, vol. 48, pp.1518 -1526, Sept. 1999.

[62] Z. Yermeche, P. M. Garcia, N. Grbic, and I. Claesson, “A calibrated subband beamforming algorithm for speech enhancement,” Sensor Array and Multichannel Signal Processing Workshop Proceedings, pp. 4-6, Aug. 2002

[63] S.Y. Low, N. Grbic, and S. Nordholm, “Robust microphone array using subband adaptive beamformer and spectral subtraction,” IEEE International Conference on Communication Systems, vol. 2, pp. 25-28, Nov. 2002.

[64] J. S. Hu and Chieh-Cheng Cheng, “Frequency domain microphone array calibration and beamforming for automatic speech recognition,”, IEICE Trans.

Fundamentals, vol. E88-A, no. 9, pp. 2401-2411, Sep. 2005.

[65] L. J. Siegel and A. C. Bessey, “Voiced/Unvoiced/Mixed Excitation Classification of Speech,” IEEE Trans. on Acoustics, Speech, and Signal Processing, vol. ASSP-30, no. 3, June 1982.

[66] Y. Zhang and M. S. Scordilis, “Robust voiced/unvoiced/mixed/silence classifier with maximum a posteriori channel/background adaptation,” Proc. IEEE, pp.

229-232, 2005.

[67] L. F. Lamel, L. R. Rabiner, A. E. Rosenberg, and J. G.Wilson, “An improved endpoint detector for isolated word recognition” , IEEE Trans. on Acoust., Voice, Signal Processing, v29, pp. 777–785, Aug. 1981.

[68] Y. Ephraim and D. Malah, “Speech enhancement using a minimum mean-square error short-time spectral amplitude estimator,” IEEE Trans. on Acoust., speech, signal processing, vol. ASSP-32, pp. 1109-1121, Dec. 1984.

[69] Dong Enqing; Liu Guizhong; Zhou Yatong; Cai Yu, “activity detection based on short-time energy and noise spectrum adaptation,” IEEE International Conference on Signal processing, vol. 1, pp. 26-30, Aug. 2002.

[70] L. R. Rabiner and M. R. Sambur, “Voiced-unvoiced-silence detection using the Itakura LPC distance measure,” IEEE International Conference on Acoust., Speech, and Signal Processing, pp. 323–326, May 1977.

[71] Nemer, E., Goubran, R, and Mahmoud, S., “Robust voice activity detection using higher-order statistics in the LPC residual domain,” IEEE Trans. on Speech and audio processing, vol. 9, pp. 217-231, March 2001.

[72] J. Zhang, W. Ward, and B. Pellom, “Phone based voice activity detection using online Bayesian adaptation with conjugate normal distributions,” IEEE International Conference on Acoust., Speech, and Signal Processing, vol. 1, pp.

321-324, 2002.

[73] J. L. Shen, J. W. Hung, and L. S. Lee, “ Robust entropy based endpoint detection for voice recognition in noisy environments” , in Proc. ICSLP’96, 1996.

[74] Junqua, J.-C., Mak, B., and Reaves, B. A robust algorithm for word boundary detection in presence of noise. IEEE Trans. on Speech and Audio Processing, vol.

2(3):406–412, July 1994.

[75] R. Chengalvarayan, “Robust energy normalization using speech/nonspeech discriminator for German connected digit recognition,” in Proc. Eurospeech, pp.

61–64, Sep. 1999.

[76] P. Renevey and A. Drygajlo, “Entropy Based Voice Activity Detection in Very Noisy Conditions,” Proceedings of 7th European Conference on Speech Communication and Technology, EUROSPEECH'2001, pp. 1887-1890, Sep. 2001 [77] C. T. Lin, J. Y. Lin, and G.. D. Wu, “A robust word boundary detection

algorithm for variable noise-level environment in cars,” IEEE Trans. on Intelligent Transportation System, vol. 3, pp. 89-101, Mar. 2002.

[78] J. Sohn and W. Sung, “A voice activity detector employing soft decision based noise spectrum adaptation,” IEEE International Conference on Acoust., Speech, and Signal Processing, vol. 1, pp. 365-368, 1998.

[79] D. V. GOKHALE, Maximum Entropy Characterization of Some Distributions.

In Statistical Distributions in Scientific Work, Patil, Kotz and Ord. Eds., Boston, M.A. Reidel, Vol. 3, 299-304., 1975.

[80] R. Martin, “An efficient algorithm to estimate the instantaneous SNR of speech signals,” EUROSPEECH'1993, vol. 1, pp. 1093-1096, 1993.

[81] J. Ramírez, J.C. Segura, C. Benítez, d.l. Torre, Ángel, and R. Antonio, “Efficient voice activity detection algorithms using long-term speech information,” Speech Communication, vol. 42, pp. 271-287, April 2004.

[82] D. H. Johnson and D. E. Dudgeon, Array Signal Processing, Englewood Cliffs, NJ: Prentice-Hall, 1993.

[83] Y. J. Hong, C. C. Yeh, and D. R. Ucci, “The effect of finite-distance signal source on a far-field steering Applebaum array—Two dimensional array case,”

IEEE Trans. on Antennas Propagation, vol. 36, pp.468–475, Apr. 1988.

[84] J. Ringelstein, A. B. Gershman, and J. F. Böhme, “Direction finding in random inhomogeneous media in the presence of multiplicative noise,” IEEE Signal Processing Letter, vol. 7, pp. 269–272, Oct. 2000.

[85] D. Astely and B. Ottersten, “The effects of local scattering on direction of arrival estimation with MUSIC,” IEEE Trans. on Signal Processing, vol. 47, pp.

3220–3234, Dec. 1999.

[86] S. Gannot, D. Burshtein, and E. Weinstein, “Signal enhancement using beamforming and nonstationarity with applications to speech,” IEEE Trans. on Signal Processing, vol. 49, pp. 1614-1626, Aug. 2001.

[87] A. P. Oppenheim, R. W. Schafer, and J. R. Buck, Discrete-time signal processing, second edition, Prentice-Hall, 1999, chapter 9.

[88] S. S. Haykin, Adaptive Filter Theory. Prentice Hall, Englewood Cliffs, NJ, 3rd

[89] M. J. Grimble and A. E. SAYED, “Solution of the H∞ optimal linear filtering problem for discrete-time systems”, IEEE Trans. on Acoust., Speech, Signal Processing, vol. 38, no. 7, pp. 1092-1104, July 1990.

[90] K. M. Nagpal and P. P. Khargonekar, “Filtering and smoothing in an H∞ setting”, IEEE Trans. On Automatic control, vol. 37, no. 2, pp. 152-166, Feb. 1991.

[91] I. Yaesh and U. Shaked, “A transfer function approach to the problems of discrete-time systems H∞-optimal linear control and filtering”, IEEE Trans. On Automatic control, vol. 36, no. 11, pp. 1264-1271, Nov. 1991.

[92] U. Shaked and Y. Theodor, “H∞ optimal estimation a tutorial”, Proceeding of the 31st conference on decision and control, pp. 2278-2286, Dec. 1992.

[93] B. Hassibi, and T. Kailath, “H∞ adaptive filtering,” IEEE International Conference on Acoust., Speech, and Signal Processing, vol. 2, pp. 949-952, May 1995.

[94] B. Hassibi, A. H. Sayed and T. Kailath, “H-infinity optimality of the LMS algorithm”, IEEE Trans. on Signal Processing, vol. 44, pp. 267-280, Feb. 1996.

[95] W. Zhuang, “Adaptive H infinity channel equation for wireless personal communications,” IEEE Trans. on Vehicular Technology, vol. 48, no. 1, pp.

126-136, January 1999.

[96] X. Shen and L. Deng, “A dynamic system approach to speech enhancement using the H∞ filtering algorithm,” IEEE Trans. on Speech and audio processing, vol. 7, no. 4, July 1999.

[97] A.T. Erdogan, B. Hassibi and T. Kailath, “FIR H-infinity equalization of communication channels,” IEEE Trans. on Signal Processing, vol.81, no.5, pp.

907-17, May. 2001.

[98] R. L. Burden, J. D. Faires, A. C. Reynolds, Numerical analysis, PWS-Kent Pub.

Co., 1993.

[99] T. E. Shoup, Applied numerical methods for the microcomputer. Prentice-Hall, 1984, chapter 4.

[100] M. J. Maron and R. J. Lopez, Numerical analysis a practical approach, third edition. Wadsworth publishing, 1991, chapter 9.

[101] M. S. Brandstein and H. F. Silverman, “A robust method for speech signal time-delay estimation in reverberant rooms,” International Conference on Acoustics, Speech, and Signal Processing, vol. 1, pp. 375-378, April 1997.

[102] C. L. Nikas. and M. Shao, Signal Processing with Alpha-Stable Distributions and Applications. New York: Wiley, 1995.

[103] I. Potamitis, H. Chen, and G. Tremoulis., “Tracking of multiple moving speakers with multiple microphone arrays,” IEEE Trans. on Speech and Audio Processing, vol. 12, pp. 520-529, Sep. 2004.

[104] S. U. Pillai and B. H. Know, “Forward/Backward spatial smoothing techniques coherent signal identification,” IEEE Trans. on Acoust. Speech, Signal Processing, vol. 37, pp. 8-14, Jan. 1989.

[105] B. Mungamuru and P. Aarabi, “Enhanced sound localization,” IEEE Trans. on Systems, Man and Cybernetics, Part B, vol. 34, pp, 1526-1540, Jun. 2004.

[106] J. G. Ryan and R. A. Goubran, “Array optimization applied in the near field of a microphone array,” IEEE Trans. on Acoust. Speech, Signal Processing, vol. 8, pp.

173-176, Mar. 2000.

[107] Y. R. Zheng, R. A. Goubran, and M. K El-Tanany, “Robust near-field adaptive beamforming with distance discrimination”, IEEE Trans. on Acoust. Speech, Signal Processing, vol. 12, pp. 478-488, Sep. 2004.

[108] M. S. Brandstein, J. E. Adcock and H. F. Silverman, “A closed-form location estimator for use with room environment microphone arrays”, IEEE Trans. on Acoust. Speech, Signal Processing, vol. 5, pp. 45-50, Sep. 1997.

[109] M. Omologo and P. Svaizer, “Use of the crosspower-spectrum phase in acoustic event location,” IEEE Trans. on Acoust. Speech, Signal Processing, vol. 5, pp.

288-292, May 1997.

[110] N. B. Chong and C. M. S. See, “Sensor-array calibration using a maximum-likelihood approach,” IEEE Trans. on Antennas and Propagation, vol. 44, pp. 827-835, June, 1996.

[111] D. B. Ward, E. A. Lehmann, and R. C Williamson,, “Particle filtering algorithms for tracking an acoustic source in a reverberant environment,” IEEE Trans. on Speech and Audio Processing, vol. 11, pp. 826-836, Nov. 2003.

[112] H. Kuttruf, Room acoustics. London: Elsevier, 1991, chapter 3, pp. 56.

[113] D. D. Vries, E. M. Hulsebos, and J. Baan, “Spatial fluctuations in measures for

[113] D. D. Vries, E. M. Hulsebos, and J. Baan, “Spatial fluctuations in measures for