Fig 3 - uploaded by Sayan Biswas
Content may be subject to copyright.
two-sided amplitude spectrum 

two-sided amplitude spectrum 

Source publication
Article
Full-text available
The term jitter and shimmer has long been used in the domain of speech and acoustic signal analysis as a parameter for speaker identification and other prosodic features. In this study, we look forward to use the same parameters in neural domain to identify and categorize emotional cues in different musical clips. For this, we chose two ragas of Hi...

Similar publications

Article
Full-text available
Any music origins in the society and develops with the changing realities of it. It accepts new and modified the existing in different periods of time. This process of acceptance and rejection makes any form of art exist for long. Similarly in various phases of transition Indian Classical music has embraced the elements which question its traits, e...
Preprint
Full-text available
Machine Learning models are capable of generating complex music across a range of genres from folk to classical music. However, current generative music AI models are typically difficult to understand and control in meaningful ways. Whilst research has started to explore how explainable AI (XAI) generative models might be created for music, no gene...
Article
Full-text available
Abstract for (18) “Equiheptatonic” Tuning in Thai Classical Music: Strict Propriety and Step Sizes. Available from: https://www.researchgate.net/publication/330429839_Equiheptatonic_Tuning_in_Thai_Classical_Music_Strict_Propriety_and_Step_Sizes: Tunings of Thai classical music have been a source of disagreement during the past century. Focusing on...
Article
Full-text available
In surveying the thirteen crisis-ridden years that Weimar democracy endured from its founding in 1919, perhaps none loom as large as the hyperinflation years spanning 1922–1923. According to many historians, the ‘Great Disorder’ not only destroyed the bonds between different social classes but also shattered Germans’ faith in and commitment to Weim...
Article
Full-text available
The purpose of this study was to propose an effective model for recognizing the detailed mood of classical music. First, in this study, the subject classical music was segmented via MFCC analysis by tone, which is one of the acoustic features. Short segments of 5 s or under, which are not easy to use in mood recognition or service, were merged with...

Citations

... Usually, MER becomes a matter of huge challenge especially because of its ambiguous emotional response. Studies involving non-linear techniques have been conducted in the recent past to understand this complex behavior of ICM and its manifestation in the human brain [38][39][40][41][42][43][44][45][46][47][48][49]. MER, particularly with ICM is still an uncharted territory for researchers especially because of the lack of structured datasets comprising of ICM clips. ...
Article
Music is often considered as the language of emotions. The way it stimulates the emotional appraisal across people from different communities, culture and demographics has long been known and hence categorizing on the basis of emotions is indeed an intriguing basic research area. Indian Classical Music (ICM) is famous for its ambiguous nature, i.e. its ability to evoke a number of mixed emotions through only a single musical narration, and hence classifying evoked emotions from ICM becomes a more challenging task. With the rapid advancements in the field of Deep Learning, this Music Emotion Recognition (MER) task is becoming more and more relevant and robust, hence can be applied to one of the most challenging test case i.e. classifying emotions elicited from ICM. In this paper we present a new dataset called JUMusEmoDB which presently has 1600 audio clips (approximately 30 s each) where 400 clips each correspond to happy, sad, calm and anxiety emotional scales. The initial annotations and emotional classification of the database was done based on an emotional rating test (5-point Likert scale) performed by 100 participants. The clips have been taken from different conventional ‘raga’ renditions played in two Indian stringed instruments – sitar and sarod by eminent maestros of ICM and digitized in 44.1 kHz frequency. The ragas, which are unique to ICM, are described as musical structures capable of inducing different moods or emotions. For supervised classification purposes, we have used Convolutional Neural Network (CNN) based architectures (resnet50, mobilenet v2.0, squeezenet v1.0 and a proposed ODE-Net) on corresponding music spectrograms of the 6400 sub-clips (where every clip was segmented into 4 sub-clips) which contain both time as well as frequency domain information. Along with emotion classification, instrument classification based response was also attempted on the same dataset using the CNN based architectures. In this context, a nonlinear technique, Multifractal Detrended Fluctuation Analysis (MFDFA) was also applied on the musical clips to classify them on the basis of complexity values extracted from the method. The initial classification accuracy obtained from the applied methods are quite inspiring and have been corroborated with ANOVA results to determine the statistical significance. This type of CNN based classification algorithm using a rich corpus of Indian Classical Music is unique even in the global perspective and can be replicated in other modalities of music also. The link to this newly developed dataset has been provided in the dataset description section of the paper. This dataset is still under development and we plan to include more data containing other emotional as well as instrumental entities into consideration.