3pBA5 – Using Acoustic Levitation to Understand, Diagnose, and Treat Cancer and Other Diseases

Brian D. Patchett – brian.d.patchett@gmail.com
Natalie C. Sullivan – nhillsullivan@gmail.com
Timothy E. Doyle – Timothy.Doyle@uvu.edu

Department of Physics
Utah Valley University
800 West University Parkway, MS 179
Orem, Utah 84058

Popular version of paper 3pBA5, “Acoustic Levitation Device for Probing Biological Cells With High-Frequency Ultrasound”
Presented Wednesday afternoon, November 4, 2015
170th ASA Meeting, Jacksonville

Imagine a new medical advancement that would allow scientists to measure the physical characteristics of diseased cells involved in cancer, Alzheimer’s, and autoimmune diseases. Through the use of high-frequency ultrasonic waves, such an advancement will allow scientists to test the normal healthy range of virtually any cell type for density and stiffness, providing new capabilities for analyzing healthy cell development as well as insight into the changes that occur as diseases develop and the cells’ characteristics begin to change.

Prior research methods of probing cells with ultrasound have relied upon growing the cells on the bottom of a Petri dish, which distorts not only the cells’ shape and structure, butlso interfere with the ultrasonic signals. A new method was therefore needed to probe the cells without disturbing their natural form, and to “clean up” the signals received by the ultrasound device. Research presented at the 2015 ASA meeting in Jacksonville Florida will show that the use of acoustic levitation is effective in providing the ideal conditions for probing the cells.

Acoustic levitation is a phenomenon whereby pressure differences of stationary sound waves can be used to suspend small objects in gases or fluids such as air or water. We are currently exploring a new frontier in acoustic levitation of cellular structures in a fluid medium by perfecting a method by which we can manipulate the shape and frequency of sound waves inside of special containers. By manipulating these sound waves in just the right fashion it is possible to isolate a layer of cells in a fluid such as water, which can then be probed with an ultrasound device. The cells are then in a more natural form and environment, and the interference from the floor of the Petri dish is no longer a hindrance.

This method has proven effective in the laboratory with buoyancy neutral beads that are roughly the same size and shape as human blood cells, and a study is currently underway to test the effectiveness of this method with biological samples. If effective, this will give researchers new experimental methods by which to study cellular processes, thus leading to a better understanding of the development of certain diseases in the human body.

2pSCb11 – Effect of Menstrual Cycle Hormone Variations on Dichotic Listening Results

Richard Morris – Richard.morris@cci.fsu.edu
Alissa Smith

Florida State University
Tallahassee, Florida

Popular version of poster presentation 2pSCb11, “Effect of menstrual phase on dichotic listening”
Presented Tuesday afternoon, November 3, 2015, 3:30 PM, Grand Ballroom 8

How speech is processed by the brain has long been of interest to researchers and clinicians. One method to evaluate how the two sides of the brain work when hearing speech is called a dichotic listening task. In a dichotic listening task two words are presented simultaneously to a participant’s left and right ears via headphones. One word is presented to the left ear and a different one to the right ear. These words are spoken at the same pitch and loudness levels. The listener then indicates what word was heard. If the listener regularly reports hearing the words presented to one ear, then there is an ear advantage. Since most language processing occurs in the left hemisphere of the brain, most listeners attend more closely to the right ear. The regular selection of the word presented to the right ear is termed a right ear advantage (REA).

Previous researchers reported different responses from males and females to dichotic presentation of words. Those investigators found that males more consistently heard the word presented to the right ear and demonstrated a stronger REA. The female listeners in those studies exhibited more variability as to the ear of the word that was heard. Further research seemed to indicate that women exhibit different lateralization of speech processing at different phases of their menstrual cycle. In addition, data from recent studies indicate that the degree to which women can focus on the input to one ear or the other varies with their menstrual cycle.

However, the previous studies used a small number of participants. The purpose of the present study was to complete a dichotic listening study with a larger sample of female participants. In addition, the previous studies focused on women who did not take oral contraceptives as they were assumed to have smaller shifts in the lateralization of speech processing. Although this hypothesis is reasonable, it needs to be tested. For this study, it was hypothesized that the women would exhibit a greater REA during the days that they menstruate than during other days of their menstrual cycle. This hypothesis was based on the previous research reports. In addition, it was hypothesized that the women taking oral contraceptives will exhibit smaller fluctuations in the lateralization of their speech processing.

Participants in the study were 64 females, 19-25 years of age. Among the women 41 were taking oral contraceptives (OC) and 23 were not. The participants listened to the sound files during nine sessions that occurred once per week. All of the women were in good general health and had no speech, language, or hearing deficits.

The dichotic listening task was executed using the Alvin software package for speech perception research. The sound file consisted of consonant-vowel syllables comprised of the six plosive consonants /b/, /d/, /g/, /p/, /t/, and /k/ paired with the vowel “ah”. The listeners heard the syllables over stereo headphones. Each listener set the loudness of the syllables to a comfortable level.

At the beginning of the listening session, each participant wrote down the date of the initiation of her most recent menstrual period on a participant sheet identified by her participant number. Then, they heard the recorded syllables and indicated the consonant heard by striking that key on the computer keyboard. Each listening session consisted of three presentations of the syllables. There were different randomizations of the syllables for each presentation. In the first presentation, the stimuli will be presented in a non-forced condition. In this condition the listener indicted the plosive that she heard most clearly. After the first presentation, the experimental files were presented in a manner referred to as a forced left or right condition. In these two conditions the participant was directed to focus on the signal in the left or right ear. The sequence of focus on signal to the left ear or to the right ear was counterbalanced over the sessions.

The statistical analyses of the listeners’ responses revealed that no significant differences occurred between the women using oral contraceptives and those who did not. In addition, correlations between the day of the women’s menstrual cycle and their responses were consistently low. However, some patterns did emerge for the women’s responses across the experimental sessions as opposed to the days of their menstrual cycle. The participants in both groups exhibited a higher REA and lower percentage of errors for the final sessions in comparison to earlier sessions.

The results from the current subjects differ from those previously reported. Possibly the larger sample size of the current study, the additional month of data collection, or the data recording method affected the results. The larger sample size might have better represented how most women respond to dichotic listening tasks. The additional month of data collection may have allowed the women to learn how to respond to the task and then respond in a more consistent manner. The short data collection period may have confused the learning to respond to a novel task with a hormonally dependent response. Finally, previous studies had the experimenter record the subjects’ responses. That method of data recording may have added bias to the data collection. Further studies with large data sets and multiple months of data collection are needed to determine any sex and oral contraceptive use effects on REA.

2aAA9 – Quietly Staying Fit in the Multifamily Building

Paulette Nehemias Harvey – pendeavors@gmail.com
Kody Snow – ksnow@phoenixnv.com
Scott Harvey – sharvey@phoenixnv.com

Phoenix Noise & Vibration
5216 Chairmans Court, Suite 107
Frederick, Maryland 21703

Popular version of paper 2aAA9, “Challenges facing fitness center designers in multifamily buildings
Presented Tuesday morning, November 3, 2015, 11:00 AM, Grand Ballroom 3
Session 2aAA, Acoustics of Multifamily Dwellings
170th ASA Meeting, Jacksonville

Please keep in mind that the research described in this Lay Language Paper may not have yet been peer reviewed.

Harvey 1 Treadmill - fitness centerTransit centered living relies on amenities close to home; mixing multifamily residential units with professional, retail and commercial units on the same site. Use the nearby trains to get to work and out, but rely on the immediate neighborhood, even the lobby for errands and everyday needs. Transit centered living is appealing as it eliminates the need for sitting in traffic, seems good for the environment and adds a sense of security, aerobic health and time-saving convenience. Include an on-site fitness center and residents don’t even have to wear street clothes to get to their gym!

Developers know that a state-of-the-art fitness center is icing on their multifamily residence cake as far as attracting buyers. Gone is the little interior room with a couple treadmills and a stationary bike. Today’s designs include panoramic views, and enough ellipticals, free weights, weight & strength machines, and shower rooms to fill 2500-4000 square feet, not to mention the large classes offered with high energy music and an enthusiastic leader with a microphone. The increased focus on maintaining aerobic health, strength and mobility is fantastic, but the noise and vibration it generates? Not so great. Sometimes cooperative scheduling keeps the peace, but often residents will want to have access to their fitness center at all hours, so wise project leaders involve a noise control engineer early in the design process to develop a fitness center next to which everyone will want to live.

Remember the string and two empty cans? Stretch the string taut and conversations can travel the length of the string, but pinch the string and the system fails. As noise travels through all kinds of structures and through the air as well, it is the design goal of the noise and vibration experts to prevent that transmission. Airborne noise control can be effective using a combination of layered gypsum board, fiberglass batt insulation, concrete and resilient framing members that absorb the sound rather than transmit it through a wall or a floor/ceiling system. Controlling the structure borne noise and vibration can involve much thicker rubber mats, isolated concrete slabs and a design that incorporates the structural engineer’s input on stiffening the base building structure. And it’s not simply noise that the design is intended to restrict, it is silent, but annoying vibrations as well.

Harvey 2 Kettleball exercise - fitness center

Reducing the floor shaking impact of dropping barbells on the ground is the opposite of hearing a pin drop. Heavy, steel plates loaded on a barbell, lifted 6-8 feet off the ground and then dropped. Repeatedly. Nobody wants to live under that, so designers think location, location, location. But big windows are pointless in the basement, so something has to go under the fitness center. Garage space, storage units or mechanical rooms won’t mind the calisthenics above them. And sometimes the overall design of the building structure, whether it be high-rise with underground parking, Texas wrap building (u-shaped building with elevated parking garage on interior.), or a podium style building can offer an ideal location for this healthy necessity.

It’s not an acoustical trade secret that the best method of noise control is at the source so consider what makes the noise. Manufacturers have met the demand for replacing the old standard graduated barbell steel plates for free weight combinations with a rubber/urethane coated steel weight. These weights make much less noise when impacting each other, but are still capable of generating excessive structure-borne noise levels. This is a great example of controlling both air borne (plates clanking together) and structure borne (barbells impacting the floor) transmission paths. Speakers and sound systems and the wall/floor/ceiling systems can work together to offer clarity and quality to listeners and limitations for what the neighbors will hear, but it takes expertise and attention.

Disregarding the recommendations of noise and vibration professionals can result in an annoying, on-site gym that brings stressful tension and ongoing conflict, nothing that promotes healthy well-being.

Foresight in design and attention to acoustical specs on building materials, under the direction of a noise and vibration engineer, assures a fitness center that is a pleasant, effective space for fitness and social opportunities, an asset to the transit centered neighborhood. Do everyone a favor and pay attention to good design and product specification early on; that’s sound advice.

5aMU1 – The inner ear as a musical instrument

Brian Connolly – bconnolly1987@gmail.com
Music Department
Logic House
South Campus
Maynooth University
Co. Kildare
Ireland

Popular version of paper 5aMU1, “The inner ear as a musical instrument”
Presented Friday morning, November 6, 2015, 8:30 AM, Grand Ballroom 2
170th ASA meeting Jacksonville
See also: The inner ear as a musical instrument – POMA

(please use headphones for listening to all audio samples)

Did you know that your ears could sing? You may be surprised to hear that they, in fact, have the capacity to make particularly good performers and recent psychoacoustics research has revealed the true potential of the ears within musical creativity. ‘Psychoacoustics’ is loosely defined as the study of the perception of sound.

Figure 1: The Ear

inner ear

A good performer can carry out required tasks reliably and without errors. In many respects the very straight-forward nature of the ear’s responses to certain sounds results in the ear proving to be a very reliable performer as its behaviour can be predicted and so it is easily controlled. In the context of the listening system, the inner ear has the ability to behave as a highly effective instrument which can create its own sounds that many experimental musicians have been using to turn the listeners’ ears into participating performers in the realization of their music.

One of the most exciting avenues of musical creativity is the psychoacoustic phenomenon known as otoacoustic emissions. These are tones which are created within the inner ear when it is exposed to certain sounds. One such example of these emissions is ‘difference tones.’ When two clear frequencies enter the ear at, say 1,000Hz and 1,200Hz the listener will hear these two tones, as expected, but the inner ear will also create its own third frequency at 200Hz because this is the mathematical difference between the two original tones. The ear literally sends a 200Hz tone back out in reverse through the ear and this sound can be detected by an in-ear microphone, a process which doctors carrying out hearing tests on babies use as an integral part of their examinations. This means that composers can create certain tones within their work and predict that the listeners’ ears will also add their extra dimension to the music upon hearing it. Within certain loudness and frequency ranges, the listeners will also be able to feel their ears buzzing in response to these stimulus tones! This makes for a very exciting and new layer to contemporary music making and listening.

First listen to this tone. This is very close to the sound your ear will sing back during the second example.

Insert – 200.mp3

Here is the second sample containing just two tones at 1,000Hz and 1,200Hz. See if you can also hear the very low and buzzing difference tone which is not being sent into your ear, it is being created in your ear and sent back out towards your headphones!

Insert – 1000and1200.mp3

If you could hear the 200Hz difference tone in the previous example, have a listen to this much more complex demonstration which will make your ears sing a well known melody. It is important to try to not listen to the louder impulsive sounds and see if you can hear your ears humming along to perform the tune of Twinkle, Twinkle, Little Star at a much lower volume!

(NB: The difference tones will start after about 4 seconds of impulses)

Insert – Twinkle.mp3

Auditory beating is another phenomenon which has caught the interest of many contemporary composers. In the below example you will hear the following: 400Hz in your left ear and 405Hz in your right ear.

First play the below sample by placing the headphones into your ears just one at a time. Not together. You will hear two clear tones when you listen to them separately.

Insert – 400and405beating.mp3

Now try and see what happens when you place them into your ears simultaneously. You will be unable to hear these two tones together. Instead, you will hear a fused tone which beats five times per second. This is because each of your ears are sending electrical signals to the brain telling it what frequency it is responding to but these two frequencies are too close together and so a perceptual confusion occurs resulting in a combined frequency being perceived which beats at a rate which is the same as the mathematical difference between the two tones.

Auditory beating becomes particularly interesting in pieces of music written for surround sound environments when the proximity of the listener to the various speakers plays a key factor and so simply turning one’s head in these scenarios can often entirely change the colour of the sound as different layers of beating will alter the overall timbre of the sound.

So how can all of these be meaningful to composers and listeners alike? The examples shown here are intended to be basic and provide proofs of concept more so than anything else. In the much more complex world of music composition the scope for the employment of such material is seemingly endless. Considering the ear as a musical instrument gives the listener the opportunity to engage with sound and music in a more intimate way than ever before.

Brian Connolly’s compositions which explore such concepts in greater detail can be found at www.soundcloud.com/brianconnolly-1

4aEA10 – Preliminary evaluation of the sound absorption coefficient of a thin coconut coir fiber panel for automotive applications

Key F. Lima – keyflima@gmail.com
Pontifical Catholic University of Paraná
Curitiba, Paraná, Brazil

Popular version of paper 4aEA10, “Preliminary evaluation of the sound absorption coefficient of a thin coconut fiber panel for automotive applications”
Presented Thursday morning, November 5, 2015, 11:15 AM, Orlando Room
170th ASA Meeting, Jacksonville, Fl

Absorbents materials are fibrous or porous and must have the property of being good acoustic dissipaters. Sound propagation causes multiples reflections and friction of the air present in the absorbent medium converting sound energy to thermal energy. The acoustic surface treatment with absorbent material are widely used to reduce the reverberation in enclosed spaces or to increase the sound transmission loss of acoustics panels. In addition, these materials can be applied into acoustics filters with the purpose to increase their efficiencies. The sound absorption depends on the excitation frequency of the sound and it is more effective at high frequencies. Natural fibers such as coconut coir fiber have a great potential to be used like sound absorbent material. As natural fibers are agriculture waste, manufacturing this fiber is a natural product, therefore an economic and interesting option. This work compares the sound absorption coefficient between a thin coconut fiber panel and a composite panel made by fiberglass and expanded polyurethane foam, no-woven woven, and polyester woven, which are used in the automotive industry as a roof trim. The evaluation of sound absorption coefficient was carried out with the impedance tube technique.

In 1980, Chung and Blaser evaluated the normal incidence sound absorption coefficient through transfer function method.  The standard ASTM E1050-10 and ISO 10534-2 was based in Chung and Blaser’s method, Figure 1. In summary, this method uses an impedance tube with the sound source placed to one end and at another, the absorbent material backed in a rigid wall. The decomposition of the stationary sound wave pattern into forward and backward traveling components is achieved by measuring sound pressures. This evaluating is carried out simultaneously at two spaced locations in the tube’s sidewall where two microphones are located, Figure 1.

Impedance Tube Fig1

Figure 1. Impedance Tube

The wave decomposition allows to the determination of the complex reflection coefficient R(f) from which the complex acoustic impedance and the normal incidence sound absorption coefficient (a) of an absorbent material can be determined. Furthermore, the two coefficients R(f) and a are calculated by Transfer Function H12 between the two microphones through:fig1, where s is the distance between the microphones, x1 is the distance between the farthest microphone and the sample, i is the imaginary unity and k0 is the wave number of the air. If R(f) is known, the coefficient a is easily obtained by expression:fig2.

In this work, eight samples of coconut fiber and eight samples of composite panel made by fiberglass and expanded polyurethane foam, no-woven woven, and polyester woven used in the automotive industry, Figure 2 and 3. The material properties are shown in Table 1.

Sample Fig2 - coconut coir fiber

Figure 2. Samples

Composite Panel Fig3 - coconut coir fiber

Figure 3. Composite panel structure.

Table 1. Material Properties.
Coconut Fiber Composite Panel
Sample

diameter

[mm]

thickness

[mm]

mass

[g]

density

[kg/m3]

Sample

diameter

[mm]

thickness

[mm]

mass

[g]

density

[kg/m3]

1 28,25 5,17 0,67 649,5 1 28,05 5,78 0,41 360,6
2 28,20 5,04 0,62 618,8 2 28,08 5,66 0,42 376,6
3 28,20 4,93 0,60 612,6 3 28,15 5,59 0,42 379,6
4 28,35 5,09 0,69 674,7 4 28,23 5,54 0,44 398,8
5 100,43 4,98 8,89 708,0 5 99,55 5,86 5,40 371,9
6 100,43 4,84 9,73 797,7 6 99,55 6,20 5,54 360,9
7 100,73 5,34 9,64 712,1 7 99,68 6,06 5,57 370,4
8 100,45 4,79 9,13 755,2 8 99,55 5,99 5,62 378,9

The random noise signal with frequency band between 200 Hz and 5000 Hz was utilized to evaluate a.  The Figure 4 shows the mean normal incidence absorption coefficient obtained from the measurements.

Comparison absorption coeff Fig4

Figure 4. Comparison of normal absorption coefficient (a)

The results shows that the composite panel have a better sound absorption coefficient than coconut fiber panel. To improve the coconut fiber panel acoustical efficiency it is needed to add some filling material with the same effect of the polyurethane foam of the composite panel.

REFERENCES
Chung, J. Y. and Blaser D. A. (1980) “ Transfer function method of measuring  in-duct acoustic properties – I Theory,” J. Acoust. Soc. Am. 68, 907-913.

Chung, J. Y. and Blaser D. A. (1980) “ Transfer function method of measuring  in-duct acoustic properties – II Experiment,” J. Acoust. Soc. Am. 68, 913-921.

ASTM E1050:2012. “Standard test method for impedance and absorption of acoustical materials using a tube, two microphones and a digital frequency analysis system,” American Society for Testing and Materials, Philadelphia, PA, 2012.

ISO 10534-2:1998. “Determination of sound absorption coefficient and impedance in impedance tubes – Part 2: Transfer-function method”, International Organization for Standardization, Geneva, 1998.

4pEA4 – “See” subsurface soils using surface waves

Zhiqu Lu – zhiqulu@olemiss.edu
National Center for Physical Acoustics, The University of Mississippi,
1 Chucky Mullins,
University, MS, 38677

Lay language paper for 4pEA4
Presented Thursday afternoon, November 5, 2015
170th ASA Meeting, Jacksonville

Within a few meters beneath the earth surface, three distinctive soil layers are formed: a top dry and hard layer, a middle moist and soft region, and a deeper zone where the mechanical strength of the soil increases with depth.  The information of this subsurface soil is required for agricultural, environmental, civil engineering, and military applications. A seismic surface wave method has been recently developed to non-invasively obtain such information (Lu, 2014; Lu, 2015).  The method, known as the multichannel analysis of surface wave method (MASW) (Park, et al., 1999; Xia, et al., 1999), consists of three essential parts: surface wave generation and collection (Figure 1), spectrum analysis, and inversion process. The implement of the technique employs sophisticated sensor technology, wave propagation modeling, and inversion algorithm.

Lu1

Figure 1. The experimental setup for the MASW method

The technique makes use of the characteristic of one type of surface waves, the so-called Rayleigh waves that travel along the earth’s surface within a depth of one and a half wavelengths. Therefore the components of surface waves with short wavelength contain information of shallow soil, whereas the longer wavelength surface waves provide the properties of deep soil (Figure 2).

Lu2

Figure 2. Rayleigh wave propagation

The outcome of the MASW method is a soil vertical profile, i.e., the acoustic shear (S) wave velocity as a function of depth (Figure 3).

Lu3 - soil profile

Figure 3. A typical soil profile

By repeating the MASW measurements either spatially or temporarily, one can measure and “see” the spatial and temporal variations of the subsurface soils. Figure 4 shows a typical vertical cross-section image in which the intensity of the image represents the value of the shear wave velocity. From this image, three different layers mentioned above are identified.

Lu4 - soil vertical cross-section

Figure 4. A typical example of soil vertical cross-section image

Lu5

Figure 5. A vertical cross-section image showing the presence of a fragipan layer

Figure 5 displays another two-dimensional image in which a middle high velocity zone (red area) appears. This high velocity zone represents a geological anomaly, known as a fragipan, a naturally occurring dense and hard soil layer (Lu, et al., 2014). The detection of fragipan is important in agricultural land managements.

The MASW method can also be applied to monitor weather influence on soil properties (Lu 2014). Figure 6 shows the temporal variations of the underground soil.  This is a result of a long term survey conducted in 2012.  By drawing a vertical line and moving it from left side to right side, i.e., along the time index number axis, the evolution of the soil profile due to weather effects can be evaluated. In particular, the high velocity zones occurred in the summer of 2012, reflecting very dry soil conditions.

Lu6

Figure 6. The  temporal variations of soil profile due to weather effects

Lu,  Z., 2014.  Feasibility of using a seismic surface wave method to study seasonal and weather effects on shallow surface soils. Journal of Environmental & Engineering Geophysics, DOI: 10.2113/JEEG19.2.71, Vol.19, 71–85.

Lu, Z. 2015. Self-adaptive method for high frequency multi-channel analysis of surface wave method, Journal of Applied Geophysics, Vol. 121, 128-139. http://dx.doi.org/10.1016/j.jappgeo.2015.08.003

Lu, Z., Wilson, G.V., Hickey, C.J., 2014. Imaging a soil fragipan using a high-frequency MASW method. In Proceedings of the Symposium on the Application of Geophysics to Engineering and Environmental Problems (SAGEEP 2014), Boston, MA., Mar. 16-20.

Park, C.B., Miller, R.D., Xia, J., 1999. Multichannel analysis of surface waves. Geophysics, Vol. 64, 800-808.

Xia, J., Miller, R.D., Park, C.B., 1999. Estimation of near-surface shear-wave velocity by inversion of Rayleigh waves. Geophysics, Vol. 64, 691-700.