Open access peer-reviewed chapter

Impulse Response Modeling of the Box Shaped Acoustic Guitar

Written By

Minakshi Pradeep Atre and Shaila Apte

Submitted: 23 July 2021 Reviewed: 01 August 2021 Published: 01 September 2021

DOI: 10.5772/intechopen.99757

From the Edited Volume

Music in Health and Diseases

Edited by Amit Agrawal, Roshan Sutar and Anvesh Jallapally

Chapter metrics overview

316 Chapter Downloads

View Full Metrics

Abstract

Music is the pulse of human lives and is an amazing tool to relieve and re-live. And when it comes to the signal processing, impulse is the pulse of the researchers. The work presented here is focused on impulse response modeling of noted produced by box shaped acoustic guitar. The impulse response is very fundamental behavior of any system. The music note is the convolution of the impulse response and the excitation signal of that guitar. The frequency of the generated music note follows the octave rule. The octave rule can be checked for impulse responses as well. If the excitation signal and impulse response are separated, then an impulse response of a single fret can be used to generate the impulse responses of other frets. Here the music notes are analyzed and synthesized on the basis of the plucking style and plucking expression of the guitar-player. If the impulse response of the musical instrument is known, the output music note can be synthesized in an unusual manner. Researchers have been able to estimate the impulse response by breaking the string of the guitar. Estimating the impulse response from the recorded music notes is possible using the methodology of cepstral domain window. By means of the Adaptive Cepstral Domain Window (ACDW) the author estimated the impulse response of guitar notes. The work has been further extended towards the classification of synthesized notes for plucking style and plucking expression using Neural Network and Machine Learning algorithms.

Keywords

  • impulse response
  • modeling
  • acoustic guitar
  • convolution
  • frets
  • octave rule
  • adaptive cepstral domain window

1. Introduction

The work starts with analysis of different ways of mathematical modeling of music instruments. The aim is to bridge the gap between the synthesized music note and the original note. Efforts have been taken to propose a model to preserve some of the oldest and ‘on the verge of obsolete’ music instruments. Past researchers developed various instruments’ models. One of the models included attack, decay, sustain and release (ADSR) parameter-based model. Figure 1 shows the ADSR graph. These parameters measure the time required for complete music note generation. The ADSR is also known as ‘timbre’ of the music note and is helpful for differentiating instrument families. Perception of the music note occurs with the timbre i.e. envelope and the fundamental frequency i.e. ‘pitch’ along with its harmonics. The fullness of the music notes is perceived by these harmonic frequencies. The Fast Fourier Transform (FFT) is the frequency domain representation of the any signal (here, the music notes). In the early days of modeling, the pitch and timbre set the foundation for the current development in the instrument modeling. ADSR synthesis has limitations of producing the required number of harmonics mathematically to reach that richness or fullness of the original music note.

Figure 1.

ADSR parameters. Source: https://www.a-mc.biz/.

The other method, digital waveguide (DWG), is also used for modeling the musical instruments. Figure 2 shows the schematic diagram of DWG technique. It is used to express the musical instruments in wave form guided between two fixed points. It puts forward the concept of music note as a wave traveling between two points. Bridge and Nut are the endpoints (rigid terminations) between which the music note wave is traveling. This DWG string modeling involves non-linear distortions and its post distortion gain needs to be adjusted to develop the model. DWG appears to be challenging due to more computational burden.

Figure 2.

DWG for acoustic guitar. Source: Ee.Columbia.Edu.

Then researchers also experimented with impulse response of the musical instrument. There has been competitive research to find the impulse response of musical instrument. Researchers have used hammer method for estimating the impulse response. The experimentation to find impulse response, also involved breaking the string of guitar as we discussed earlier. Well, keeping the limitations in mind, the research started with cepstral domain approach, well-known technique for speech processing.

Every instrument is unique and this uniqueness can be demonstrated with impulse response. It’s as unique as the fingerprint of a person. The experiments involved breaking the string of the guitar to record the impulse response. Certainly, this made the author to think further to improve this uniqueness of the instrument, i.e. impulse response using cepstral domain representation. Being a guitar lover, the author chose the box shaped acoustic guitar music notes for the research work.

A simple convolution operation is involved in the generation of the music note. The excitation signal, x(t) is convolved with the impulse response of the instrument to generate the music note, y(t). Figure 3 shows the block diagram for this signal processing operation. When the music note is recorded and x(t) is known, the impulse response can be estimated and recorded. This research demonstrated the extraction of the impulse response of the music note, based on adaptive cepstral domain window (ACDW) method. Figure 4 shows the outline of the cepstral domain window approach. The impulse response based synthesis is carried out and the listening tests are conducted on the guitar players to measure the mean opinion score (MoS) of synthesized notes. Further, the machine learning algorithms are used to classify the synthesized notes for playing expression.

Figure 3.

Block diagram for music note generation using convolution.

Figure 4.

Block schematic of cepstrum computation.

The outline of the chapter is as follows: the Section 2 will discuss the work by other researchers, Section 3 will discuss the methodology for the impulse response modeling while Section 4 will discuss the results in detail. The Section 5 will summarize the work and conclude the modeling work for acoustic guitar.

Advertisement

2. Literature review

This section reviews the modeling techniques of guitar as an instrument. The physical model of an acoustic guitar consists of three main parts: strings of guitar, the wooden sound box and the sound radiated by the soundboard. The review starts with the modeling techniques used for guitar strings, continues with the modeling techniques involving the sound box and finally the convolved signal i.e. radiated sound by the soundboard. It also covers the Neural Network (NN) and Deep Learning based classification techniques and verification of synthesized music instruments. The literature survey has been divided into: Physical or mathematical modeling and Impulse response methods.

The work by Gerald Schuller et al. in [1] has been used as reference for collection of acoustic guitar notes. They considered 5 plucking styles finger-style (FS), picked (PK), muted (MU), slap-thumb (ST), and slap-pluck (SP) and the 5 expression styles: normal (NO), vibrato (VI), bending (BE), harmonics (HA), and dead-note (DN) for feature extraction of plucking and expression styles of electric bass guitar. Anssi Klapuri et al. [2] have proposed a method for extracting the fingering configurations automatically from a recorded guitar performance. 330 different fingering configurations are considered, corresponding to different versions of the major, minor, major 7th, and minor 7th chords. Hidden Markov Model has been used.

Migneco et al. [3] proposed physical models for plucked string instruments that can produce high-quality tones using a computationally efficient implementation, but the estimation of model parameters through the analysis of audio remains challenging. Moreover, an accurate representation of the expressive aspects of a performance requires a separation of the performer’s articulation (source) from the instrument’s response (filter). This work explores a physically-inspired signal model for plucked guitar sounds. It facilitates the estimation of both string excitation and resonance parameters simultaneously. Julius O Smith [4] discussed the piano synthesis, focused on commuted synthesis. The instrument models can be treated as Linear Time Invariant systems and that’s why the commutation is possible. Commuted synthesis promotes implementation of enormous resonators inexpensively, three orders of magnitude less computation for other string instruments. The sound board and enclosure (i.e. guitar body) are commuted. It needs stored recording of their impulse responses. Otherwise it demands higher order digital filters.

Further, the work by Meng Koon Lee et al. in [5] talks about the physical modeling based on the interaction of the strings of the guitar with other parts of the guitar body. The researchers experimented with the sound generated by guitar with respect to soundboard and its relationship with the guitar body. The soundboard plays an increasingly important role compared to the sound hole, back plate, and the bridge at high frequencies. Design of bracings and their placements on the soundboard increase its structural stiffness as well as redistributing its deflection to non-braced regions and affecting its loudness as well as its response at low and high frequencies. The work is focused to increase the sound level with bracing designs and their placements. The analysis is being carried out for the archtop guitar.

The paper [6], written by Keith D Martin explains the classification technique based on physical properties. It is focused on the classification using pattern recognition. A statistical pattern-recognition technique is applied to instrument tones within a taxonomic hierarchy. The salient acoustic features related to physical properties of source excitation and resonance structure are measured from output of auditory model for 1,023 isolated tones over the full pitch ranges of 15 orchestral instruments. The data set included examples from the string (bowed and plucked), woodwind (single, double, and air reed), and brass families. Eric J. Henry et al. [7] proposed a model that can yield representations for the chords that require minimal prior knowledge to interpret. The model has been developed to address both challenges by modeling the physical constraints of a guitar to produce human-readable representations of music audio, i.e. guitar tablature via a deep convolutional network.

Jakob Abeßer et al. [8] worked on a feature-based approach for the classification of different playing techniques in bass guitar recordings. The applied audio features are chosen to capture typical instrument sounds induced by 10 different playing techniques. This work introduced a set of low-level features that allowed modeling the peculiarities of 10 different bass-related plucking and expression styles by capturing typical timbre related characteristics. The work further in [9] models the plectrum which is used for playing guitar notes. Here Francois Germain et al. proposed a model of the plectrum, a guitar pick, for use in physically inspired sound synthesis. The model is drawn from the mechanics of beams. The profile of the plectrum is computed in real time based on its interaction with the string, which depends on the movement impressed by the player and the equilibrium of dynamical forces. A condition for the release of the string is derived, which allows driving the digital waveguide simulating the string to the proper state at release time. The algorithm proposed by Henri Penttinen et al. [10] estimates the plucking point of guitar tones obtained with an under-saddle pickup. This problem is approached in the time domain by applying autocorrelation estimation. Onset detection has been improved in this proposed work. It enables a new way to control audio effect parameters in real time by simply changing the plucking point. The plucking position changes the timbre of the string’s tone, most notably the brightness. This effect is used as an expressive tool in music. By using the PPE (Plucking Point Estimation) algorithm to control an audio effect, change in the plucking position can affect the timbre even more dramatically than in the natural unprocessed case.

Gabriele Varieschi et al. [11] presented mathematical and physical models to be used in the analysis of the problem of intonation of musical instruments such as guitars, mandolins and similar instruments. The analysis is done by designing the fret’s placement on the fingerboard according to mathematical rules assuming an ideal string. The intonation of a string note gets affected when other string’s deformation and inharmonicity come into picture. To nullify the effects, the authors have designed some compensation procedures. V.E.Howle et al. [12] proposed a known tool of Eigenvalue to musical instruments. The work as its name “Eigenvalues and musical instruments” suggests is based on finding the eigenvalues of musical instruments. The instrument categories like strings, bars and drums fall under linear system’s class. It is focused on plotting the eigenvalues for different types of musical instrument giving pictorial view of change in the eigenvalues with change in different parameters like stiffness, friction or sound radiations.

Antoine Chaigne et al. [13] considered the end conditions of piano strings and proposed that it can be approximated by the input admittance at the bridge. A method of validation of admittance measurements on simple structures is proposed in this paper. High resolution signal analysis performed on string’s vibrations yields an estimate for the input admittance. This method is implemented on a simplified device composed of a piano string coupled to a thin steel beam.

A parametric modeling of string instruments is proposed by famous researchers, Matti Karjalainen et al. in [14]. Parametric modeling of musical instrument sounds again helps to re-synthesize the music sounds or morph them. This type of modeling can also be used to apply the parameters in physical and perceptual studies of acoustic instruments. It is typically based on pole-zero modeling technique applied to string instrument sounds. As proposed by Julius Smith the instruments are assumed to be linear time-invariant systems while using this parametric modeling. Our research work has been directed by the same principal as that of the work by Julius Smith.

The authors in work [15, 16] talk about sound separation. It is very important for developing the equalizers to balance the sounds of different musical instruments in music events. The method based on ‘anisotropic smoothness’ indicates that the harmonic instruments are smooth in the time domain whereas the percussive instruments are smooth in the frequency domain. The authors have worked on both the types of music notes. The spectrogram highlights this smoothness in time and the frequency domain and the method is implemented under some conditions. The work reduced the computational complexity for source separation as compared with the other methods as Monte-Carlo method and large-sized matrix multiplications. The results are discussed for the acoustic guitar and piano as the harmonic instruments and the drum as the percussive instruments. This paper again helped us to understand the spectrogram approach towards the source separation.

Further with reference to impulse response research, Nelson Lee et al. [17] proposed a method of decomposing a plucked string instrument into modular components. The model is based on parameter estimation of excitation signal, string vibration, body resonator and finally the radiated sound pressure. As the modeling progresses it becomes clear that for reaching close to body impulse responses, the order of the filter demands a hundred of poles and zeros. Inverse filtering is used to compensate for high orders of filters.

Friedrich Türckheim et al. [18] used the ‘Novel Approach of Impulse Response measurement’ as starting point for modeling approaches or to investigate the relationship between transfer functions and the instruments’ quality. This is done usually for the experimental determination of transfer function as the complete and reliable physical models are still to be developed. The impulse responses here have been compared with the commonly used impact hammer method. The work proposed in above two references have limitations of filter orders and determination of exact transfer functions. So the author thought of different approach for the calculation of impulse response of the guitar body.

Advertisement

3. Methodology

Methodology adopted by the author is different than other researchers. So this section is the summary of the research work carried out for impulse response modeling of acoustic guitar. The sections 3.1 and 3.2 discuss the structure of acoustic guitar body and the collection of music notes for two different plucking styles and plucking expressions. Section 3.3 describes how the Octave Rule is used for the frequency calculation and verification of all music notes along the fretboard. Lastly, the Section 3.4 gives the details of the impulse response modeling.

3.1 Structure of acoustic guitar

Let us understand the structure of the box shaped acoustic guitar. It has a resonance cavity, shaped like a butterfly. The wings are short at nut side and are bigger at the saddle side. As shown in Figure 5, there are six strings on the acoustic guitar. The strings are tied between the saddle and the nut of guitar on the fingerboard. The fingerboard is also known as fretboard.

Figure 5.

Structure of the box shaped acoustic guitar. Source: Gabriele Umberto Varieschi.

The fingerboard of guitar consists of 19 to 21 frets. The frets are the metal marks on the fretboard, arranged in logarithmic scale. They are shown with x1, x2, x3, …x19 here. The acoustic guitar model, FAW 802 is chosen for the research work. The music notes were recorded in an acoustic studio. The acoustic guitar chosen for this research work has twenty-one frets and six strings. Music notes on twenty frets are considered for the analysis and synthesis purpose.

3.2 Collection of guitar notes and database generation

The music note is recorded for each fret of each string (except the 21st fret). The guitar notes are collected based on: plucking style and the plucking expression. The plucking style indicates the object used for plucking the string of the acoustic Guitar. The plucking expression indicates whether the note is played by a naïve (beginner) person or the expert person. The two players, one naïve and the other, expert, are recorded for two plucking styles. The Figure 6 gives overview of the collection strategy of the Guitar notes.

Figure 6.

Collection of acoustic guitar notes.

When the string is plucked by finger, it will generate a music note and it is named as ‘plucked note’ here. Similarly, if the string is plucked by plectrum (or pick), the generated music note will be named as ‘picked note’. The total number of notes generated are calculated as: the number of frets multiplied by the number of strings, i.e. (20* 6 =) 120 notes. These all are called as “fretted notes”. If no fret is pressed, then the note played will be called as ‘open string’ note. There will be a set of 120 fretted notes with 6 open string notes for single player for each style. The notes played by same player are recorded for another plucking style i.e. picked.

The Figure 7 shows the frequency values for all strings and their frets. The frets are 1F to 20F. The first column gives the string numbers along with their names: string E, string B, string G, string D, string A, string E. The column, ‘OPEN’ gives the open string frequencies with 82 Hz as the lowest frequency value and 329 Hz as the highest open string frequency value. The frequency for fretted notes varies from 87 Hz for string 6 fret 1 to 1047 Hz for string 1 fret 20. This is the maximum frequency of the guitar note. All guitar notes are therefore recorded with 16 kHz sampling frequency in ‘.wav’ format. Software for sound analysis, named, ‘Audacity’ is used for noise removal of the guitar notes.

Figure 7.

Standard frequency for guitar notes: Guitar frets and their notes versus frequency.

Figure 8 shows the scatter plot of the frequencies of all frets of all strings. The numbers 1 to 21 on the x-axis represent the fret numbers of the strings. The y-axis depicts the frequencies of guitar notes. This scatter plot helps to understand the mathematical relationship as well as the minimum and the maximum frequency values for these notes. The maximum frequency for the notes is 2 kHz so the sampling frequency of 16 kHz is selected for recording of the music notes in acoustic studio.

Figure 8.

Frequency generation for music notes on all frets and strings.

The string E with 329 Hz frequency is string 1, string B with 247 Hz frequency is string 2, string G with 196 Hz frequency is string 3, string D with 147 Hz frequency is string 4, string A with 110 Hz frequency is string 5 and the string E with 82 Hz frequency is string 6. The strings are mentioned by the numbers (like string 1, string 2 …) in further discussion of research work. Total 504 sound notes are recorded including two plucking styles and plucking expressions. Another set of 504 notes is recorded for string modeling. The dataset generated has been published on Mendeley Repository, Elsevier.

3.3 Pythagoras fractions OR rule of 18 (octave rule)

Pythagoras fractions or Pythagorean tuning system is developed to study frequency ratios of all intervals which are based on the ratio of 3:2. It is the rule, given by Eq. (1), which indicates the mathematical relationship of all the music notes and is used for checking the frequency of generated notes. Here the ‘fn’ gives the frequency of the nth fret and ‘f0’ is the frequency of open string. The ratio 1/12 is called an ‘octave’. One can use this octave relationship for analysis and tuning of instruments. The frequency analysis of recorded notes is done on the basis of octave rule. The guitar used for this work is tuned each time before starting the recording of the notes. Once the instrument is tuned, the notes generated follow the octave pattern for frequency values. Here the open string frequencies of recorded notes are verified by autocorrelation formula and then verified by Pythagoras Fractions.

fn=f0/2n12E1

This will create a table for 20 frets of each string. Table 1 shows the sample calculations of 6 frets for all the six strings from the open string frequencies. In this table, after knowing the open string frequency, all the other frequencies are calculated by using the second column which gives the octave multiplier factor. First column gives the fret number, ‘n’. The values in third to eighth columns are calculated in the way explained below.

Table 1.

Frequency calculation for pick plucked guitar notes by octave method.

Consider the sample values in cells which are highlighted in orange. The calculation for the frequency for string 1 fret 1 is done as:

e.g. 329 Hz * 1.059463 = 349.199 Hz

Thus the octave rule is used to verify the music note’s frequency. The frequency graphs are plotted using FFT algorithms. After the verification of frequencies and observation of the spectrum of the music notes, it can be stated that picked music notes are closer (sharper) to ideal frequencies than the plucked music notes.

3.4 Impulse response modeling

The frequency analysis done in above sections helped to develop better understanding of the fullness of the music notes based on their number of harmonics. This is also helpful to get better understanding of the playing style and plucking expressions. The frequency analysis is now followed by the impulse response estimation. The next five subsections deal with the synthesis part of the research work and discusses the algorithmic approach towards the impulse response modeling.

3.4.1 Introduction

The cepstral domain approach is frequently used for speech signal processing but it is not so far used for music signal processing. This method is used for separation of vocal tract response and the excitation signal in speech signal processing. Based on the same principle, this modeling work is focused on separation of: 1) impulse response or body response from 2) excitation signal of the acoustic guitar notes.

The next section discusses the algorithm for Cepstral Domain Windowing (CDW) and its application for modeling of acoustic guitar notes using the same CDW method.

3.4.2 Cepstral domain window method

Let us focus on theoretical aspects of cepstral domain. Figure 9 shows the block schematic of the cepstral domain method used for speech analysis. The input to the system is speech signal. The speech signal consists of excitation signal convolved with the impulse response of the vocal tract. On similar principle, the music note is given as input the system. It gives the representative picture of impulse response and the excitation signal, characteristically separated. The excitation signal is periodic in nature and the impulse response is the slowly varying function. The signal is passed through a smooth window function, a Hamming window function and the spectrum is plotted by calculating the FFT of the block. When the logarithm of the magnitude of FFT output is calculated, the periodic excitation signal is clearly seen as rapidly varying function and the vocal tract response appears as the slowly varying function. By using the cepstral domain window, isolation of excitation signal from body response is possible. Thus cepstral domain method can be used for modeling of music note of guitar.

Figure 9.

Block schematic of the cepstral domain method.

3.4.3 Cepstral domain windowing method for acoustic guitar notes

The block schematic of cepstral domain windowing method for analysis of acoustic guitar music notes is given in Figure 10. The algorithm for CDW method discussed in Section 3.4.2 is used for calculation of body response or impulse response of the guitar box. The input to the system is the acoustic guitar note. The FFT block gives spectrum of the music signal and then the complex logarithm of magnitude of FFT output is taken. The periodic excitation is seen as a rapidly varying function and guitar body response appears as the envelope of the spectrum. The body response is a slowly varying function. After the IFFT of the signal is calculated it enters in the cepstral domain. The cepstral domain plot indicates a cluster near the origin that represents the body response and the periodic peaks after the cluster represent the periodic excitation generated due to plucking of the string by hand or by plectrum.

Figure 10.

Cepstral domain approach to synthesize the music note.

The string1 fret 2 note is taken as sample input to the different blocks in Figure 10. Figure 11 shows the time domain representation of this acoustic guitar note used as input to the system for isolation of body response and the excitation signal for string 1 fret 2 with finger plucking style. The sampling rate for the recorded note is 16 kHz.

Figure 11.

Time domain representation of input music note, string 1 fret 1 finger plucked note.

3.4.4 Synthesis of guitar note using isolated body response and the excitation signal

The body responses and the excitation signals are calculated for 252 guitar notes including the plucking style: finger and the plectrum plucked music notes. A note is then synthesized by convolving the estimated body response and the isolated excitation signal. The results are verified using the correlation coefficients and it is observed that the constant length window poses some limitations to give highly correlated synthesized guitar note. Figure 12 shows the plots for the original guitar note and the synthesized guitar note. But the results are not satisfactory because of the low correlation coefficient values. Table 2 presents the sample values of 6 frets of string 1 music notes.

Figure 12.

Synthesized guitar note for string 1 fret 1 using CDW method.

String 1Correlation coefficientMoS Score based on (0–1) scale
Fret 10.85640.85
Fret 20.82290.85
Fret 30.84860.85
Fret 40.68420.8
Fret 50.80610.8
Fret 60.83400.8

Table 2.

Correlation coefficient calculations for the string 1 with all frets with fixed sized window.

1Finger pluckOpen stringFret 1Fret 2Fret 3Fret 4Fret 5Fret 6
2Number of samples in the window70606050505050
3Correlation coefficient0.95120.95960.95170.94890.81960.88560.9008
4MOS score on the scale of(0–1)0.950.950.950.90.80.80.9
5Pick pluckOpen stringFret 1Fret 2Fret 3Fret 4Fret 5Fret 6
6Number of samples in the window501901001703006060
7Correlation coefficient0.94980.93940.94520.92880.92380.94420.9476
8MOS score on the scale of (0–1)0.90.90.90.90.90.90.9

Table 3.

Correlation coefficients for ACDW samples: Results for plucked & picked sound notes.

The synthesis results are improved by changing the length of the window. This ‘adaptiveness’ in the length of the window is named as the ‘Adaptive Cepstral Domain Window (ACDW)’. The ACDW method gives the best estimation of the impulse response and isolates the excitation signal from the impulse response. Once it’s isolated from the excitation signal, modeling of impulse response becomes the focus.

This section covers the discussion of the ACDW method along with the results of impulse response estimation and the synthesis of guitar notes. The length of the window in cepstral domain is changed in the range of 50 samples to 300 samples. The estimation of the best impulse response is done based on correlation coefficient. The correlation coefficient is the statistical parameter to indicate the degree of similarity. A lot of experimentation is done by varying the number of samples of the cepstral domain window. It is observed that the correlation coefficient drops when the number of samples in cepstral domain window are increased further. The range is finalized after studying the impulse response and the synthesis results.

Figure 13 plots the correlation coefficients versus the number of samples in the cepstral domain for the string 1 fret 1 finger plucked Guitar note. From the figure, it’s clear that when number of samples in the chosen length of cepstral window is 70, ACDW synthesis gives highest correlation coefficient. The graph shows the decaying nature of the impulse response for the selected guitar note. So the improvement in the synthesis is achieved with the help of ACDW method. Once this is achieved, the extracted impulse responses are analyzed further to observe their relationship. The isolated impulse responses for all the frets of a single string are plotted and it’s observed that these impulses are also following the important Octave relationship from one fret to the other.

Figure 13.

Comparison of correlation coefficients for different number of samples for string 1 fret 1 finger plucked guitar note.

This triggered the thought of using the impulse response of a single fret to generate the impulse responses for the other frets. The experimentation is carried out for string 1 with all its 20 frets for impulse generation. The generated impulses were convolved with the separated excitation signals to generate all the music notes along a single fret. This gives rise to generalized acoustic guitar model where a single impulse response can be stored and used to generate all other music notes of that guitar. Figure 14 shows the time domain graph of the impulse responses showcasing their octave relationship. It demonstrates the octave relationship followed by the impulse responses and a generalized model based on impulse response is developed.

Figure 14.

Impulse responses of finger plucked guitar notes.

In summary, Figure 15 shows the model for the ‘acoustic guitar notes synthesizer’. The string number and the fret number should be passed to the model and then it will isolate the impulse or the body response from the excitation signal. The estimated impulse response can be convolved with the excitation signal to synthesize the acoustic guitar note.

Figure 15.

Block schematic of the synthesized note using impulse response modeling.

Advertisement

4. Discussion of results

The synthesis results for the fixed sized windows is poor, given in Table 2, as verified with MoS (Mean Opinion Score), the subjective tests, taken from 10 guitar players. The MoS is carried out and also the correlation coefficient values are calculated. The results range from the value, 0.68 to 0.9 for the correlation coefficient. That’s the reason, the length of the cepstral domain window is kept variable to achieve best correlation coefficient value. This improved the results and helped to achieve the best possible impulse response which is further used to develop the generalized guitar model. Table 3 summarized the results of the adaptive cepstral domain window method. The column gives the plucking style and the parameters of the music note, such as, the number of samples in the window, correlation coefficients and MoS score on the scale of (0–1) for the synthesized music notes.

4.1 Synthesis results of the generated impulse response and excitation signal

The separated excitation signal and the estimated impulse response are convolved together to synthesize the acoustic Guitar note. The ACDW approach provides good scope for better synthesis. The synthesis has been carried out for the two playing styles, namely, finger plucked, and pick plucked notes. The highest correlation coefficient value is 0.98 for finger plucked guitar note. The MoS (Mean Opinion Score) is also best indication for the synthesized guitar note giving highest value as 0.95.

The NN and machine learning algorithms are used to classify the plucking style and plucking expression for the original recorded notes. Once the model is trained it is further deployed for cross-validation of the synthesized music notes based on the ACDW method. The results are discussed subsection 4.1.

The contribution of this impulse response modeling work is to isolate the body response and the excitation signal of acoustic guitar notes. An innovative synthesizer for acoustic guitar notes is implemented in this research work. The work has used innovative Adaptive Cepstral Domain Windowing (ACDW) method which is not used before for musical instrument modeling. The algorithm has used the cepstral domain approach. The other contribution of the present work is to develop the generalized model for Guitar Notes using impulse response of single fret. The reason why this is possible is ‘the octave relationship’ of frequency of all the frets along a string. The other name for the music is ‘harmony’ and the music notes generated by frets on strings follow this octave relationship to be in harmony. This research proved that ‘not only in frequency domain but also in time domain the music notes live in harmony’. This harmonic relationship in the time domain is used to generate the impulse response model using single fret. After plotting the graph of impulse response of all the frets for a string, fret 20 was chosen for modeling the impulse response for other frets.

To summarize the work, impulse response modeling is implemented with good accuracy. Further, the neural network (NN) is used for classification of naïve and expert player considering the expression in the note played. The classification is also done for plucking style i.e. finger plucking and plectrum plucking. The results of the two methods of synthesis i.e. ACDW and Generalized model are cross verified by NN model. The trained model of the classification is used for verification of the synthesized notes.

4.2 Discussion of the cross-validation results

The synthesis of the acoustic guitar notes is implemented using the ACDW method and a generalized model is developed using the body response of the single fret. The validation of the synthesized guitar notes is done by the subjective (listening) tests and the correlation coefficient values. NN model is used for the classification of guitar notes with respect to plucking style and plucking expression. Further this trained model is used for testing the synthesized guitar notes to identify plucking style and plucking expression. This is named as cross validation of the models.

Table 3 summarizes the validation of the synthesized guitar notes based on: 1) Mean opinion Score i.e. MOS as the subjective tests and 2) Correlation coefficients as the statistical parameter for finding the similarity between original and synthesized music notes using ACDW approach. The NN model is used for classification of the music notes based on: 1) plucking style and 2) plucking expressions.

Table 4 summarizes the cross-validation result for synthesized music notes, only of an expert player. The last two columns are highlighted to show the result of NN modeling to predict the class. The table values confirm the model validation as the classification results are greater than 80%. Only few sample results are shown in Table below.

Plucking StyleMusic Note(0–1)Scale for the synthesized noteCorrelation coefficientsExisting ClassPredicted Class
Plucked NoteString1fret190.80.8477ExpertExpert
String1fret180.80.8144ExpertExpert
String1fret130.70.79ExpertExpert
String1fret120.70.71ExpertBeginner
String1fret20.80.8909ExpertExpert
String1fret10.80.8120ExpertExpert
Picked NotesString1fret190.850.8784ExpertExpert
String1fret180.850.86ExpertExpert
String1fret130.750.81ExpertExpert
String1fret120.750.79ExpertExpert
String1fret20.750.78ExpertBeginner
String1fret10.80.71ExpertExpert

Table 4.

Cross-validation result for the expert’s synthesized music notes.

Similarly, a cross-validation has been carried out for the plucking style where the acoustic guitar notes played by the Expert Player have been passed to the trained model and the plucking style is predicted. The results from the Impulse Response modeling method are considered for the identification of the plucking style. Table 5 summarizes the results of the trained model for identification of plucking style of synthesized notes using NN classifiers. The cells highlighted with yellow indicate the wrong classification of the plucking style. The second column gives the names of the music notes while the 3rd and the 4th columns indicate the correlation coefficients for the subjective tests and the impulse response method.

Music Note(0–1) Scale for the synthesized noteCorrelation coefficients by Impulse Response ModelingOriginal StylePredicted Style
ExpertString1fret10.950.9283FingerFinger
String1fret20.950.9441FingerFinger
String1fret30.90.9324FingerFinger
String1fret40.80.7897FingerPick
String1fret50.850.8856FingerFinger
String1fret60.90.9008FingerFinger
ExpertString1fret10.950.9317PickPick
String1fret20.950.9452PickPick
String1fret30.90.9146PickFinger
String1fret40.90.9238PickPick
String1fret50.90.9127PickPick
String1fret60.950.9434PickPick

Table 5.

Cross-validation result for the plucking style of the expert player.

Advertisement

5. Conclusion

The limitation of the impulse response method using the hammer method and string-breaking method are overcome with the help of cepstral domain window method. The challenges of isolation of impulse response from the excitation signal are overcome using ACDW approach and a model is developed using the body features. The main contribution of the present research work is: 1) Physical model for Guitar as an instrument using Adaptive Cepstral Domain Window (ACDW) approach, 2) Generalized Model for Impulse Response of Acoustic Guitar for All Frets using a Response of Single Fret, and 3) Classification of guitar notes based on plucking style and plucking expression. The validation of the synthesized notes is done by using subjective listening tests i.e. Mean Opinion Score (MOS) and the correlation coefficients. The classification of plucking style and plucking expression is done using NN modeling techniques. The trained model is used for testing the plucking expression of the synthesized model. This model can be used to certify if the player is becoming an expert. If the score for expert identification is greater than 95% then player can be certified as expert.

References

  1. 1. Gerald Schuller, Jakob Abeßer, Christian Kehling. Parameter extraction for bass guitar sound models including playing styles”, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Year: 2015
  2. 2. Ana M. Barbancho; Anssi Klapuri; Lorenzo J. Tardon; Isabel Barbancho,“Automatic Transcription of Guitar Chords and Fingering From Audio”, IEEE Transactions on Audio, Speech, and Language Processing, Year: 2012
  3. 3. Raymond V. Migneco; Youngmoo E. Kim, “Modeling plucked guitar tones via joint source-filter estimation”, Digital Signal Processing Workshop and IEEE Signal Processing Education Workshop (DSP/SPE) 2011, 2011 IEEE.
  4. 4. Julius O Smith III, “Piano Synthesis”, Center for Computer Research in Music and Acoustics (CCRMA) Department of Music, Stanford University, Stanford, California 94305 USA, March 2014
  5. 5. Lee Meng Koon, Mohammad Hosseini Fouladi, and Satesh Narayana Namasivayam. “Mathematical modelling and acoustical analysis of classical guitars and their soundboards.” Advances in Acoustics and Vibration 2016
  6. 6. Martin, Keith D., and Youngmoo E. Kim. “Musical instrument identification: A pattern-recognition approach.” The Journal of the Acoustical Society of America 104.3 (1998): 1768-1768.
  7. 7. Eric J. Humphrey, Juan P. Bello, “From music audio to chord tablature: Teaching deep convolutional networks to play guitar”, 2014 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
  8. 8. Jacob Aberber, Hanna Lukashevich, Gerald Schuller, “Feature-Based Extraction of Plucking and Expression Styles of the Electric Bass Guitar”, Fraunhofer IDMT, Ilmenau, Germany, 2010.
  9. 9. Francois Germain, Gianpaolo Evangelista, “Synthesis of Guitar by Digital Waveguides: Modeling The Plectrum the Physical Interaction Of The Player With The Instrument”, October 2009
  10. 10. Henri Penttinen, Jaakko Siiskonen, and Vesa Valimaki, “Acoustic Guitar Plucking Point Estimation in Real Time”, Laboratory of Acoustics and Audio Signal Processing, Helsinki University of Technology P.O. Box 3000, FI-02015 HUT, Espoo, Finland (2005)
  11. 11. Gabriele U. Varieschi and Christina M. Gower, Intonation and Compensation of Fretted String Instruments, Department of Physics, Loyola Marymount University 5th September 2009
  12. 12. V.E. Howle , Lloyd N. Trefethen, “Eigenvalues and Musical Instruments”, Center for Applied Mathematics, 657 Frank H. T. Rhodes Hall, Cornell University, Ithaca, May 2000.
  13. 13. Kerem Ege, Antoine Caigne, “End conditions of Piano Strings”, Laboratory for Solid Mechanics, Ecole Polytechnique, UMR7649, 91128 Palaiseau Cedex, Unité de Mécanique, Ecole Nationale Supérieure de Techniques Avancées, Chemin de la Hunière.
  14. 14. Karjalainen, Matti, and Tuomas Paatero. “High-resolution parametric modeling of string instrument sounds” Signal Processing Conference, 2005 13th European. IEEE, 2005.
  15. 15. Tachibana, Hideyuki, et al. “Harmonic/percussive sound separation based on anisotropic smoothness of spectrograms.” IEEE/ACM Transactions on Audio, Speech, and Language Processing 22.12 (2014): 2059-2073.
  16. 16. Tachibana, Hideyuki, et al. “Comparative evaluations of various harmonic/percussive sound separation algorithms based on anisotropic continuity of spectrogram.” 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2012.
  17. 17. Lee, Nelson, Julius O. Smith, and Vesa Valimaki. “Analysis and synthesis of coupled vibrating strings using a hybrid modal-waveguide synthesis model.” IEEE transactions on audio, speech, and language processing 18.4 (2009): 833-842.
  18. 18. Friedrich Türckheim, Thorsten Smit, Carolin Hahne, and Robert Mores, “Novel Impulse Response Measurement Method for Stringed Instruments”, Proceedings of 20th International Congress on Acoustics, ICA 2010 23–27 August 2010, Sydney, Australia.

Written By

Minakshi Pradeep Atre and Shaila Apte

Submitted: 23 July 2021 Reviewed: 01 August 2021 Published: 01 September 2021