著者
Jianwu Dang Aijun Li Donna Erickson Atsuo Suemitsu Masato Akagi Kyoko Sakuraba Nobuaki Minematsu Keikichi Hirose
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.31, no.6, pp.394-402, 2010-11-01 (Released:2010-11-01)
参考文献数
16
被引用文献数
1 11

In this study, we conducted a comparative experiment on emotion perception among different cultures. Emotional components were perceived by subjects from Japan, the United States and China, all of whom had no experience living abroad. An emotional speech database without linguistic information was used in this study and evaluated using three- and/or six-emotional dimensions. Principal component analysis (PCA) indicates that the common factors could explain about 60% variance of the data among the three cultures by using a three-emotion description and about 50% variance between Japanese and Chinese cultures by using a six-emotion description. The effects of the emotion categories on perception results were investigated. The emotions of anger, joy and sadness (group 1) have consistent structures in PCA-based spaces when switching from three-emotion categories to six-emotion categories. Disgust, surprise, and fear (group 2) appeared as paired counterparts of anger, joy and sadness, respectively. When investigating the subspaces constructed by these two groups, the similarity between the two emotion groups was found to be fairly high in the two-dimensional space. The similarity becomes lower in 3- or higher dimensional spaces, but not significantly different. The results from this study suggest that a wide range of human emotions might fall into a small subspace of basic emotions.
著者
Hyun-yeol Chung Shozo Makino Ken'iti Kido
出版者
Acoustical Society of Japan
雑誌
Journal of the Acoustical Society of Japan (E) (ISSN:03882861)
巻号頁・発行日
vol.9, no.5, pp.225-232, 1988 (Released:2011-02-17)
参考文献数
11
被引用文献数
1 4

This paper describes analysis and recognition of eight Korean isolated vowels based on the distributions of formant frequencies. In the distributions of vowels on the F1-F2 plane, we found overlappings between the vowels /u/, /o/, /∂/ and /w/, /ε/ and /e/ overlapped heavily as if they are the same vowels. In the comparative investigation of the dialect of Seoul and Kyungsangdo, Kyungsangdo dialect showed a heavy overlapping between /∂/ and /w/. As a recognition experiment, the speaker-independent recognition with Bayes decision method was carried out for 1, 760 isolated vowels spoken by 20 speakers. A recognition rate of 81.2% correct was obtained with the formant frequencies F1 and F2while the rate with LPC cepstrum coefficients was only 76.4%. Normalization for eliminating inter-speaker differences showed good results for vowels uttered by speakers of the same dialect. The highest recognition rate of 99.8 % was obtained using a supervised learning method.
著者
Noriaki Katagiri Goh Kawai
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.34, no.2, pp.94-104, 2013-02-01 (Released:2013-03-01)
参考文献数
29
被引用文献数
1

Within the context of English language taught solely using English language at Japan's secondary schools, no research quantifies the differences between native instructors (first language English, may or may not speak Japanese) and non-native instructors (first language Japanese; second language English). We developed a video corpus of an English language classroom, and examined the speech of 3 native and 1 non-native instructors. The corpus contains 49 English lessons of 45 minutes each in a Japanese public high school with monolingual learners of English as a foreign language. The native and non-native instructors occasionally taught together. Almost all speech in the lessons was in English. We compared lexical tokens and types found in our transcriptions with a collection of typical classroom English dialogues, and a wordlist created from large bodies of written and spoken English. We obtained the distributions of words, and words preferred by either native or non-native instructors. Results suggest that (a) native and non-native instructors share a core vocabulary of classroom English, (b) native instructors teach vocabulary depth via open-ended conversations, (c) non-native instructors teach vocabulary breadth via textbook explanations, and (d) native and non-native instructors differ in teaching roles but not in language ability.
著者
Zwicker Eberhard Fastl Hugo Widnmann Ulrich Kurakata Kenji Kuwano Sonoko Namba Seiichiro
出版者
Acoustical Society of Japan
雑誌
Journal of the Acoustical Society of Japan (E) (ISSN:03882861)
巻号頁・発行日
vol.12, no.1, pp.39-42, 1991
被引用文献数
28 87

The method for calculating loudness level proposed by Zwicker is standardized in ISO 532B. This is a graphical procedure and it can be tedious to calculate loudness level by this procedure. Recently, DIN 45631 has been revised including a computer program for calculating loudness level in BASIC which runs on IBM-compatible PC's. Since the NEC PC-9801 series computers are popular in Japan, the program has been modified for the NEC PC-9801 series computers and is introduced in this paper.
著者
Sun Chao Yang Yi-xin
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical science and technology (ISSN:13463969)
巻号頁・発行日
vol.25, no.1, pp.2-8, 2004
参考文献数
9
被引用文献数
8

The problem of beamforming and related beamspace high resolution direction-of-arrival (DOA) estimation is studied in this paper. All beamspace processing methods are based on the beam outputs and the beampattern design plays an important role in providing high quality beam output data for further processing. Three typical situations which are frequently encountered in practical sonar system working environment and the most widely studied MUSIC algorithm are considered herein. First, when isotropic noise is the dominant noise at sensors, conventional beamforming techniques provide the optimum performance in the sense that DOA estimate is the ML estimate. Good DOA estimates are obtainable by applying MUSIC to the beam outputs directly. Then, uncorrelated interferes with much higher strength than the wanted signals are assumed to be present in the sidelobe region, and low sidelobe Dolph-Chebyshev and adaptive MVDR beampatterns are designed to guarantee the performance of MUSIC. And finally, the robustness of conventional techniques is combined with the adaptivity of MVDR beamforming to deal with the situation when the interfere in the sidelobe region is strongly correlated with one of the wanted sources. Performance in all three situations is studied with numerical examples.
著者
日比 正史
出版者
Acoustical Society of Japan
雑誌
Journal of the Acoustical Society of Japan (E) (ISSN:03882861)
巻号頁・発行日
vol.4, no.2, pp.83-95, 1983 (Released:2011-02-17)
参考文献数
12
被引用文献数
21 38

In perception of regular sequence of stimulus sounds that is neither too rapid nor too slow, we tend to perceive it as rhythmic. In the present study, we investigated the degree of temporal distortion intervened in the sequence which was needed in order for listeners to be able to report an irregularity, and we also investigated the timing patterns of the repetition of monosyllables as a response to the regular sound sequence. The degree of distortion which afforded 50% judgments of detection was found to be about 6% in the region of rates slower than 3 times per second, and about 7.6∼8.9% depending upon the type of distortion in the region of rates more rapid than 4 times per second. Another experiment showed that a negative correlation between the time intervals of neighboring vicinity which suggested an adjusting mechanism was found only in the region of rates slower than 3 times per second. From the results, the author tentatively concludes that the ongoing processing mechanism works in the region of rates slower than 3 times per second and the wholistic processing mechanism works in the region of rates more rapid than 3 times per second.
著者
Ryoichi Takashima Tetsuya Takiguchi Yasuo Ariki
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.34, no.3, pp.176-186, 2013-03-01 (Released:2013-05-01)
参考文献数
21

This paper presents a talker localization method using only a single microphone, where phoneme hidden Markov models (HMMs) of clean speech are introduced to estimate the acoustic transfer function from the user's position. In our previous work, we proposed a Gaussian mixture model (GMM) separation for estimation of the user's position, where the observed speech is separated into the acoustic transfer function and the clean speech GMM. In this paper, we propose an improved method using phoneme HMMs for separation of the acoustic transfer function. This method expresses the speech signal as a network of phoneme HMMs, while our previous method expresses it as a GMM without considering the temporal phonetic changes of the speech signal. The support vector machine (SVM) for classifying the user's position is trained using the separated frame sequences of the acoustic transfer function. Then, for each test data set, the acoustic transfer function is separated, and the position is estimated by discriminating the acoustic transfer function. The effectiveness of this method has been confirmed by talker localization experiments performed in a room environment.
著者
Ryota Nakasaki Hiroshi Hasegawa Masao Kasuga
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.34, no.3, pp.159-165, 2013-03-01 (Released:2013-05-01)
参考文献数
11
被引用文献数
1 5

In this paper, we investigated subjective impressions of air-conditioning sounds in a vehicle by using the psychoacoustic parameters of loudness and sharpness. First, we carried out a subjective evaluation using a rating scale method and investigated the relationships between the psychoacoustic parameters and nine evaluation words, quiet, refreshing, heavy, wide, muddy, violent, dry, warm, and cool, that represent impressions of air-conditioning sounds. As a result, we found that the impressions of ``violent'' and ``quiet'' strongly depended on the loudness, and the impressions of ``heavy,'' ``dry,'' ``warm,'' and ``cool'' strongly depended on the sharpness. Next, we performed a factor analysis. As a result, we found that the air-conditioning sounds can be explained by two factors, a volume factor and a thermal factor, which were strongly correlated with the loudness and sharpness, respectively. This result shows that the feelings of heating and cooling perceived from air-conditioning sounds are related to the sharpness. Therefore, a synergetic effect on the heating and cooling performance can be expected by improving the sharpness of air-conditioning sounds.
著者
Ryo Yoneda Masashi Yamada
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.34, no.3, pp.166-175, 2013-03-01 (Released:2013-05-01)
参考文献数
11
被引用文献数
1

Musical psychologists illustrated musical emotion with various numbers of dimensions ranging from two to eight. Most of them concentrated on classical music. Only a few researchers studied emotion in popular music, but the number of pieces they used was very small. In the present study, perceptual experiments were conducted using large sets of popular pieces. In Experiment 1, ten listeners rated musical emotion for 50 J-POP pieces using 17 SD scales. The results of factor analysis showed that the emotional space was constructed by three factors, ``evaluation,'' ``potency'' and ``activity.'' In Experiment 2, three musicians and eight non-musicians rated musical emotion for 169 popular pieces. The set of pieces included not only J-POP tunes but also Enka and western popular tunes. The listeners also rated the suitability for several listening situations. The results of factor analysis showed that the emotional space for the 169 pieces was spanned by the three factors, ``evaluation,'' ``potency'' and ``activity,'' again. The results of multiple-regression analyses suggested that the listeners like to listen to a ``beautiful'' tune with their lovers and a ``powerful'' and ``active'' tune in a situation where people were around them.
著者
Donna Erickson
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.34, no.3, pp.153-158, 2013-03-01 (Released:2013-05-01)
参考文献数
52
被引用文献数
2
著者
Hideo SUZUKI
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.21, no.3, pp.163-166, 2000 (Released:2001-01-31)
参考文献数
2
被引用文献数
2 4

A single-degree-of-freedom(SDOF) system with a mass(m), a spring(k), and a damping(c) is a basic mechanical system.It is well known that a complex mechanical system is represented by a combination of infinite number of SDOF systems.The modal analysis theory is based on this principle, but only one type among various types of SDOF systems is presently employed for a modeling of complex systems.When one needs to estimate m, k, and c of a SDOF system from the measurement of the resonance frequency and the loss factor, the relationships between them must be exactly known.Those relationships are known for commonly used SDOF systems, but those for rather unfamiliar types are not known.In this technical report, various types of SDOF systems and their equivalent electrical circuits are listed, and equations that relate the resonance frequencies to the undamped resonance(natural) frequencies and the loss factors are given.It is also shown that, for some type of SDOF systems, some cares must be taken how to interpret the loss factor.
著者
Hideki Banno Hiroaki Hata Masanori Morise Toru Takahashi Toshio Irino Hideki Kawahara
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.28, no.3, pp.140-146, 2007 (Released:2007-05-01)
参考文献数
19
被引用文献数
11 28

A very high quality speech analysis, modification and synthesis system—STRAIGHT—has now been implemented in C language and operated in realtime. This article first provides a brief summary of STRAIGHT components and then introduces the underlying principles that enabled realtime operation. In STRAIGHT, the built-in extended pitch synchronous analysis, which does not require analysis window alignment, plays an important role in realtime implementation. A detailed description of the processing steps, which are based on the so-called “just-in-time” architecture, is presented. Further, discussions on other issues related to realtime implementation and performance measures are also provided. The software will be available to researchers upon request.
著者
Ken’ichi Miyazaki
出版者
ACOUSTICAL SOCIETY OF JAPAN
雑誌
Acoustical Science and Technology (ISSN:13463969)
巻号頁・発行日
vol.25, no.6, pp.426-432, 2004 (Released:2004-11-01)
参考文献数
55
被引用文献数
14 19 25

Absolute pitch (AP) is the ability based on the fixed association between musical pitch and its verbal label. Experiments on AP identification demonstrated extreme accuracy of AP listeners in identifying pitch, influences of timbre and pitch range, and difference in accuracy between white-key notes and black-key notes. However, contrary to the common belief that AP is a component of musical ability, it was found that AP listeners have difficulty in perceiving pitch relations in different pitch contexts, and in recognizing transposed melodies, as compared to listeners having no AP. These results suggest that AP is irrelevant and even disadvantageous to music. Systematic music training in early childhood seems effective for acquiring AP. Possible genetic contributions to AP are undeniable, but evidence for them is inconclusive. There are several AP-like phenomena that do not reach consciousness: absolute tonality, long-term memory of pitch of repeatedly heard tunes, specific patterns of pitch comparison in the tritone paradox, and fixed pitch levels in speech. Contrary to true AP observed as a pitch naming ability, the implicit AP phenomena are widespread among general population.