Sound Emitted by Plants
Sound Emitted by Plants
Sound Emitted by Plants
Correspondence
[email protected]
In brief
Plants emit species- and stress-specific
airborne sounds that can be detected in
acoustic chambers and greenhouses.
plant sounds
Highlights
d Plants emit ultrasonic airborne sounds when stressed
Article
Sounds emitted by plants under stress
are airborne and informative
Itzhak Khait,1,6,9 Ohad Lewin-Epstein,1,7,8,9 Raz Sharon,1,2 Kfir Saban,1 Revital Goldstein,1 Yehuda Anikster,1
Yarden Zeron,1 Chen Agassy,1 Shaked Nizan,1 Gayl Sharabi,1 Ran Perelman,1 Arjan Boonman,3 Nir Sade,1,4
Yossi Yovel,3,5,10 and Lilach Hadany1,5,10,11,*
1School of Plant Sciences and Food Security, The George S. Wise Faculty of Life Sciences, Tel-Aviv University, Tel-Aviv, Israel
2School of Mathematical Sciences, Tel-Aviv University, Tel-Aviv, Israel
3School of Zoology, The George S. Wise Faculty of Life Sciences, Tel-Aviv University, Tel-Aviv, Israel
4The Institute of Cereal Crop Improvement, The George S. Wise Faculty of Life Sciences, Tel Aviv University, Tel Aviv, Israel
5Sagol School of Neuroscience, Tel Aviv University, Tel-Aviv, Israel
6Present address: Centure Applications (Greeneye Technology), Tel Aviv, Israel
7Present address: Broad Institute of MIT and Harvard, Cambridge, MA, USA
8Present address: Center for Computational and Integrative Biology, Massachusetts General Hospital and Harvard Medical School, Boston,
MA, USA
9These authors contributed equally
10These authors contributed equally
11Lead contact
*Correspondence: [email protected]
https://doi.org/10.1016/j.cell.2023.03.009
SUMMARY
Stressed plants show altered phenotypes, including changes in color, smell, and shape. Yet, airborne sounds
emitted by stressed plants have not been investigated before. Here we show that stressed plants emit
airborne sounds that can be recorded from a distance and classified. We recorded ultrasonic sounds emitted
by tomato and tobacco plants inside an acoustic chamber, and in a greenhouse, while monitoring the plant’s
physiological parameters. We developed machine learning models that succeeded in identifying the condi-
tion of the plants, including dehydration level and injury, based solely on the emitted sounds. These informa-
tive sounds may also be detectable by other organisms. This work opens avenues for understanding plants
and their interactions with the environment and may have significant impact on agriculture.
1328 Cell 186, 1328–1336, March 30, 2023 ª 2023 The Authors. Published by Elsevier Inc.
This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
ll
Article OPEN ACCESS
Figure 1. Stressed plants emit remotely detectable ultrasounds that reveal plant condition and species
(A) Acoustic box setup. In each recording, three plants are placed inside a 50 3 100 3 150 cm3 acoustic box with two directional microphones oriented at each
plant. Using two microphones helps eliminating false detections resulting from electrical noise of the recording system and cross-plant interference.
(B) Mean number of sounds emitted during 1 h of recording by tomato and tobacco plants under two treatments: drought stress and cutting. Three control groups
were used: pot—pots with soil but without a plant; self-control—plant before treatment; and neighbor-control—untreated plants that shared the acoustic box with
treated plants. All treatment groups emitted significantly more sounds than all control groups (p < e6, Wilcoxon test with Bonferroni correction): less than 1
detection/hour for all plant-treatment combinations in the self-control and neighbor-control groups ; 19 % n % 51 plants for each group, and no sound detection in
the pot-control group (>500 recording hours). Error bars represent standard errors.
(C) Examples of time signals of sounds emitted by: a drought-stressed tomato, a drought-stressed tobacco, a cut tomato, and a cut tobacco, normalized. Peak
dBSPL values at 10 cms, relative to 20 mPa, are noted by the arrows (see STAR Methods).
the potential in studying plant bioacoustics, suggest that plant 10 cm, for tomato and tobacco, respectively, and the mean
acoustic emissions may play an important role in ecology and peak frequencies (frequency with maximal energy) of these
evolution, and may have direct implications for plant monitoring sounds was 49.6 ± 0.4 kHz and 54.8 ± 1.1 kHz, respectively.
in agriculture. The mean peak intensity of the sounds emitted by cut plants
was 65.6 ± 0.2 dBSPL and 63.3 ± 0.2 dBSPL at 10.0 cm, for
RESULTS tomato and tobacco, respectively, and the mean peak fre-
quency was 57.3 ± 0.7 kHz and 57.8 ± 0.7 kHz, respectively.
To investigate plants’ airborne sound emissions, we constructed Note that due to the directionality of the microphone, this is a
a reliable recording system, where each plant is recorded simul- lower bound of the intensity (see STAR Methods).
taneously by two microphones (see Figure 1A for illustration, and Spectrograms of raw recorded sounds from drought-stressed
STAR Methods for details). First, we recorded plants within an and cut plants are shown at Figure S1A, the mean spectra of the
acoustic box and developed machine learning algorithms to sounds are shown in Figure S1B, and an example of the
classify the recorded sounds. Then we tested the system in a electronic noise is in Figure S1C. An audio sample of the actual
greenhouse, while monitoring physiological parameters of the tomato recordings, after down sampling to the audible range
recorded plants. and condensation in time, is included in Audio S1.
We trained machine learning models to classify different plant
Stressed plants emit airborne sounds that reveal their conditions and species based on their sound emissions. We
condition: Acoustic box setup divided the sounds into four groups, corresponding to the four
We recorded tomato (Solanum lycopersicum) and tobacco combinations of two plant types (tomato or tobacco), and two
(Nicotiana tabacum) plants under different treatments—drought treatments (drought stress or cutting). A binary classifier was
stress, cutting (of the stem, before recording), and controls, trained to distinguish between two equal-size groups (‘‘pair’’) in
within an acoustically isolated box. We focused on the ultrasonic each comparison (Tomato-Dry vs. Tomato-Cut; Tobacco-Dry
sound range (20–150 kHz). vs. Tobacco-Cut; Tomato-Dry vs. Tobacco-Dry; Tomato-Cut
We found that plants emit sounds, and that stressed plants— vs. Tobacco-Cut). For cross validation, the model was tested
both drought-stressed (Dry) and cut plants (Cut; see STAR only on plants that were not a part of the training process (see
Methods)—emit significantly more sounds than plants of any of STAR Methods for more details). We used a support vector ma-
the control groups (p < e6, Wilcoxon test for each of the 12 chine (SVM) as the classifier with several methods of feature
comparisons with Holm-Bonferroni correction). Three controls extraction—basic,32,33 MFCC,34 and a scattering network.26
were used for each plant species and treatment: recording The SVM classifier with scattering network for feature extrac-
from the same plant before treatment (Self-control), recording tion achieved 70% accuracy for each of the four pairs (Figure 1F
from an untreated same-species neighbor plant (Neighbor-con- red line), significantly better than random (p < e12 for each pair,
trol), and recording a pot with soil but without a plant (Pot; see Wilcoxon rank-sum test with Holm-Bonferroni correction). The
STAR Methods). The mean number of sounds emitted by dry same classifier was trained to discriminate between the electri-
plants was 35.4 ± 6.1 and 11.0 ± 1.4 per hour for tomato and cal noise of the system (see STAR Methods) and the sounds
tobacco, respectively, and cut tomato and tobacco plants emitted by the plants (Tomato vs. Noise, Tobacco vs. Noise)
emitted 25.2 ± 3.2 and 15.2 ± 2.6 sounds per hour, respectively and achieved above 98% accuracy for both (Figure 1F). The re-
(Figure 1B). In contrast, the mean number of sounds emitted by sults were also robust to the dimension of the descriptors and
plants from all the control groups was lower than 1 per hour. Our the scattering network specific parameters (Figure S1D). When
system did not record any sound in the Pot control (Figure 1B) using the SVM classifier with MFCC as the input features, the re-
over >500 h of recordings. sults were still significantly better than random (black line,
What does a stressed plant sound like? Figures 1C and 1D p < e4 for each pair, corrected as above), and even when using
show examples of raw recorded time signals and their spectra only 4 ‘‘basic’’ features (see STAR Methods), the results were
as recorded from drought-stressed and cut plants, while the significantly better than random for 5 of the 6 pairs (p < e6 for
distributions of sound peak intensity and the maximum energy each of them, adjusted; Figure 1F gray line). However, scattering
frequency of drought-stressed and cut plants are shown at Fig- network performed better than either MFCC or Basic for all the
ure 1E. The mean peak sound intensity recorded from dry pairs (p < 0.05 and p < e6, respectively; Wilcoxon signed-
plants was 61.6 ± 0.1 dBSPL and 65.6 ± 0.4 dBSPL at rank test). Altogether, these results demonstrate that plant
the late morning-noon time. Nevertheless, we see that the daily through all the trachea in the cut stem. In accordance, sounds
transpiration rate decreases as dehydration continues, while were emitted by cut plants for a shorter period of time than by
the number of emitted sounds does not (Figure 4). dry plants. Of the sounds emitted by plants that were both cut
and dry, the majority were classified as ‘‘cut’’ in the first day,
DISCUSSION but the picture is reversed in the following days, with a majority
of sounds classified as ‘‘dry’’ (Figure S4C). (3) A 3D acoustic
Our results demonstrate that plants emit remotely detectable simulation shows that sounds emitted from the trachea would
and informative airborne sounds under stress (Figure 1). The radiate from the stem in all directions (Figure S4D). This is consis-
plant emissions that we report, in the ultrasonic range of 20– tent with the results of our two-microphone recording system,
100 kHz, could be detected from a distance of 3–5 m, by many which picked up sound on two sides of the stem they were
mammals and insects (given their hearing sensitivity, e.g., directed to (Figure 1A). (4) The frequency range of cavitation-
mice37 and moths27,38; see STAR Methods for details on the esti- related vibrations partially overlaps with the sounds we re-
mation of animal detection ranges). We succeeded in differenti- corded.17 When recorded with contact sensors, cavitation is
ating between sounds emitted under two different stress condi- usually also characterized by additional higher frequencies that
tions—dry and cut plants—with an accuracy of 70% using are beyond the sensitivity range of our microphones and that
supervised machine learning methods (Figure 1), and distin- would attenuate rapidly in air. Yet, only the vibrations that result
guishing between drought-stressed and control plants in a in airborne sounds (which we report here) are the ones that have
greenhouse, based only on the sounds they emit (accuracy of a potential of affecting other organisms and human-sensors that
84%, Figures 2B and 2D). We monitored the distribution of plant are not in direct contact with the plant. When considering groups
sounds with respect to days of dehydration, time of day, of plants, the advantage of tracking airborne sound may be even
soil moisture (Figure 3), and transpiration rate (Figure 4). Our re- greater, as a single artificial sensor for airborne sounds could
cordings revealed that the hourly pattern of sound emission detect sounds requiring multiple contact sensors.
correlates with the plant’s transpiration rate, while the daily num- A potential application of our results can be for monitoring
ber of sounds is a hump-shaped function of plant dehydration, plants in the field or greenhouse.40 Specifically, plant sound
increasing during the first days of dehydration, and declining emissions could offer a way for monitoring crops water and
as the plant dries up. The sounds emitted by the plants at high possibly disease states—questions of crucial importance in agri-
and low levels of dehydration are different, and we succeeded culture.41 More precise irrigation can save up to 50% of the water
classifying them with an accuracy of 81% (Figure 3). These find- expenditure and increase the yield, with dramatic economic im-
ings can alter the way we think about the plant kingdom, which plications.41,42 In times when more and more areas are exposed
has been considered to be almost silent until now.23 to drought due to climate change,43 efficient water use becomes
One potential mechanism that may be responsible for the even more critical, for both food security and ecology. Our re-
emission of at least part of the sounds we record is cavitation sults, demonstrating the ability to distinguish between drought-
in the stem.16 Several findings support this: (1) we found that stressed and control plants based on plant airborne sounds,
the frequencies of the sounds emitted by different plant species open an avenue of research in the field of precision agriculture.
correspond to their trachea diameter, with wider tracheas in the We have shown that plant sounds can be effectively classified
plants emitting lower sounds (Figures S4A and S4B), consistent by machine learning algorithms. We thus suggest that other or-
with the observed negative association between xylem diameter ganisms may have evolved to classify these sounds as well
and resonance frequency.39 (2) The different sounds emitted un- and respond to them. For instance, many moths—some of
der drying and cutting (Figure 1) are in accordance with the them using tomato and tobacco as hosts for their larvae44,45—
different gas dynamics of the plant in these two processes: while can hear and react to ultrasound in the frequencies and inten-
drying is gradual, with a low rate of air-seeding and reduced sities that we recorded.25–27 Nearby plants may also respond
pressure, cutting involves a rapid and significant air-seeding to the sounds emitted by plants. Plants were already shown to
9. Karban, R. (2008). Plant behaviour and communication. Ecol. Lett. 11, 32. Acevedo, M.A., Corrada-Bravo, C.J., Corrada-Bravo, H., Villanueva-Riv-
727–739. era, L.J., and Aide, T.M. (2009). Automated classification of bird and
10. Falik, O., Mordoch, Y., Quansah, L., Fait, A., and Novoplansky, A. (2011). amphibian calls using machine learning: A comparison of methods.
Rumor has it.: relay communication of stress cues in plants. PLoS One 6, Ecol. Inf. 4, 206–214.
e23625. https://doi.org/10.1371/journal.pone.0023625. 33. Giannakopoulos, T., and Pikrakis, A. (2014). Introduction to Audio Anal-
11. Chamovitz, D. (2012). What a Plant Knows: A Field Guide to the Senses of ysis: A MATLAB Approach (Academic Press).
Your Garden-And beyond (Scientific American/Farrar, Straus and Giroux). 34. Ellis, D.P. (2005). {PLP} and {RASTA}(and {MFCC}, and inversion) in
12. Lev-Yadun, S. (2016). Defensive (Anti-herbivory) Coloration in Land Plants {M} atlab.
(Springer). 35. Brodribb, T.J., and Holbrook, N.M. (2004). Diurnal depression of leaf hy-
13. Hassanien, R.H., HOU, T.-z., LI, Y.-f., and LI, B.-m. (2014). Advances in ef- draulic conductance in a tropical tree species. Plant Cell Environ. 27,
fects of sound waves on plants. J. Integr. Agric. 13, 335–348. 820–827.
14. Gagliano, M., Mancuso, S., and Robert, D. (2012). Towards understanding 36. Gosa, S.C., Koch, A., Shenhar, I., Hirschberg, J., Zamir, D., and Moshe-
plant bioacoustics. Trends Plant Sci. 17, 323–325. lion, M. (2022). The potential of dynamic physiological traits in young to-
mato plants to predict field-yield performance. Plant Sci. 315, 111122.
15. Tyree, M.T., and Sperry, J.S. (1989). Vulnerability of xylem to cavitation
and embolism. Annu. Rev. Plant Physiol. Plant Mol. Biol. 40, 19–36. 37. Heffner, H.E., and Heffner, R.S. (1985). Hearing in two cricetid rodents:
16. Cochard, H., Badel, E., Herbette, S., Delzon, S., Choat, B., and Jansen, S. Wood rat (Neotoma floridana) and grasshopper mouse (Onychomys leu-
(2013). Methods for measuring plant vulnerability to cavitation: a critical re- cogaster). J. Comp. Psychol. 99, 275–288.
view. J. Exp. Bot. 64, 4779–4791. 38. Moir, H.M., Jackson, J.C., and Windmill, J.F.C. (2013). Extremely high fre-
17. De Roo, L., Vergeynst, L., De Baerdemaeker, N., and Steppe, K. (2016). quency sensitivity in a ‘simple’ear. Biol. Lett. 9, 20130241.
Acoustic emissions to measure drought-induced cavitation in plants. 39. Dutta, S., Chen, Z., Kaiser, E., Matamoros, P.M., Steeneken, P.G., and
Appl. Sci. 6, 71. Verbiest, G.J. (2022). Ultrasound Pulse Emission Spectroscopy Method
18. Zweifel, R., and Zeugin, F. (2008). Ultrasonic acoustic emissions in to Characterize Xylem Conduits in Plant Stems. Research 2022.
drought-stressed trees – more than signals from cavitation? New Phytol. 40. Khait, I., Sharon, R., Yovel, Y., and Hadany, L. (2020). Plant-monitor.
179, 1070–1079. https://doi.org/10.1111/j.1469-8137.2008.02521.x. Patent Application Number WO2020039434A1. patent application
19. Ikeda, T., and Ohtsu, M. (1992). Detection of xylem cavitation in field-- WO2020039434A1.
grown pine trees using the acoustic emission technique. Ecol. Res. 7, 41. Playán, E., and Mateos, L. (2006). Modernization and optimization of irriga-
391–395. tion systems to increase water productivity. Agric. Water Manag. 80,
20. Jackson, G., and Grace, J. (1996). Field measurements of xylem cavita- 100–116.
tion: are acoustic emissions useful? J. Exp. Bot. 47, 1643–1650. 42. Sadler, E., Evans, R., Stone, K., and Camp, C. (2005). Opportunities for
21. Bailey, N.W., Fowler-Finn, K.D., Rebar, D., and Rodrı́guez, R.L. (2013). conservation with precision irrigation. J. Soil Water Conserv. 60, 371–378.
Green symphonies or wind in the willows? Testing acoustic communica- 43. Allen, C.D., and Breshears, D.D. (1998). Drought-induced shift of a forest–
tion in plants. Behav. Ecol. 24, 797–798. ars228. woodland ecotone: rapid landscape response to climate variation. Proc.
22. ten Cate, C. (2013). Acoustic communication in plants: do the woods really Natl. Acad. Sci. USA 95, 14839–14842.
sing? Behav. Ecol. 24, 799–800. 44. Specht, A., de Paula-Moraes, S.V., and Sosa-Gómez, D.R. (2015). Host
23. Gagliano, M. (2012). Green symphonies: a call for studies on acoustic plants of Chrysodeixis includens (Walker)(Lepidoptera, Noctuidae, Plusii-
communication in plants. Behav. Ecol. 24, 789–796. ars206. nae). Rev. Bras. Entomol. 59, 343–345.
24. Jung, J., Kim, S.-K., Kim, J.Y., Jeong, M.-J., and Ryu, C.-M. (2018). 45. Liu, Z., Li, D., Gong, P., and Wu, K. (2004). Life table studies of the cotton
Beyond chemical triggers: evidence for sound-evoked physiological reac- bollworm, Helicoverpa armigera (Hübner)(Lepidoptera: Noctuidae), on
tions in plants. Front. Plant Sci. 9, 25. different host plants. Environ. Entomol. 33, 1570–1576.
25. Miller, L.A., and Surlykke, A. (2001). How Some Insects Detect and Avoid Be- 46. Rodrigo-Moreno, A., Bazihizina, N., Azzarello, E., Masi, E., Tran, D., Bou-
ing Eaten by Bats: Tactics and Countertactics of Prey and Predator Evolu- teau, F., Baluska, F., and Mancuso, S. (2017). Root phonotropism: Early
tionarily speaking, insects have responded to selective pressure from bats signalling events following sound perception in Arabidopsis roots. Plant
with new evasive mechanisms, and these very responses in turn put pressure Sci. 264, 9–15.
on bats to ‘‘improve’’ their tactics. Bioscience 51, 570–581. 47. López-Ribera, I., and Vicient, C.M. (2017). Drought tolerance induced by
26. Spangler, H.G. (1988). Moth hearing, defense, and communication. Annu. sound in Arabidopsis plants. Plant Signal. Behav. 12, e1368938. https://
Rev. Entomol. 33, 59–81. doi.org/10.1080/15592324.2017.1368938.
27. Fullard, J.H., Dawson, J.W., and Jacobs, D.S. (2003). Auditory encoding 48. Jeong, M.-J., Cho, J.-I., Park, S.-H., Kim, K.-H., Lee, S.K., Kwon, T.-R.,
during the last moment of a moth’s life. J. Exp. Biol. 206, 281–294. Park, S.-C., and Siddiqui, Z.S. (2014). Sound frequencies induce drought
28. Mishra, R.C., Ghosh, R., and Bae, H. (2016). Plant acoustics: in the search tolerance in rice plant. Pak J Bot 46, 2015–2020.
of a sound mechanism for sound signaling in plants. J. Exp. Bot. 67, 49. Gibly, A., Bonshtien, A., Balaji, V., Debbie, P., Martin, G.B., and Sessa, G.
4483–4494. (2004). Identification and expression profiling of tomato genes differen-
29. Jeong, M.-J., Shim, C.-K., Lee, J.-O., Kwon, H.-B., Kim, Y.-H., Lee, S.-K., tially regulated during a resistance response to Xanthomonas campestris
Byun, M.-O., and Park, S.-C. (2008). Plant gene responses to frequency- pv. Mol. Plant Microbe Interact. 17, 1212–1222.
specific sound signals. Mol. Breeding 21, 217–226. 50. Gera, A., Loebenstein, G., and Shabtai, S. (1983). Enhanced tobacco
30. Ghosh, R., Mishra, R.C., Choi, B., Kwon, Y.S., Bae, D.W., Park, S.-C., mosaic virus production and suppressed synthesis of a virus inhibitor in
Jeong, M.-J., and Bae, H. (2016). Exposure to sound vibrations lead to protoplasts exposed to antibiotics. Virology 127, 475–478.
transcriptomic, proteomic and hormonal changes in Arabidopsis. Sci. 51. Sessa, G., Yang, X.-Q., Raz, V., Eyal, Y., and Fluhr, R. (1995). Dark induc-
Rep. 6, 33370–33417. tion and subcellular localization of the pathogenesis-related PRB-1b pro-
31. Veits, M., Khait, I., Obolski, U., Zinger, E., Boonman, A., Goldshtein, A., Sa- tein. Plant Mol. Biol. 28, 537–547.
ban, K., Seltzer, R., Ben-Dor, U., Estlein, P., et al. (2019). Flowers respond 52. Zinger, E., Gueijman, A., Obolski, U., Ram, Y., Ruby, E., Binder, M., Ye-
to pollinator sound within minutes by increasing nectar sugar concentra- chieli, N., Ohad, N., and Hadany, L. (2019). Less fit Lamium amplexicaule
tion. Ecol. Lett. 22, 1483–1492. plants produce more dispersible seeds. Sci. Rep. 9, 6299.
STAR+METHODS
RESOURCE AVAILABILITY
Lead contact
Further information and requests for resources should be directed to and will be fulfilled by the lead contact, Lilach Hadany (lilach.
[email protected]).
S. lycopersicum ‘Hawaii 7981’.49 The plants for the experiments presented in Figures 1, 2, and 3 were grown in a growth room at 25 C
and kept in long-day conditions (16 h day, 8 h night). The plants used in the transpiration experiment (Figure 4) were grown in a green-
house. The plants used in the infection experiment (Figure S3F) were grown in a greenhouse and infected with TMV.50 The plants were
tested in the experiments 5–8 weeks after germination.
N. tabacum ‘Samsun NN’.51 The plants were grown in a growth room at 25 C and kept in long-day conditions (16 h day, 8 h night).
T. aestivum (cv fielder), grown in a greenhouse.
METHOD DETAILS
Recording protocol
In the acoustic box
The recordings were performed in a 50 3 100 3 150 cm3 acoustically isolated custom-made 1.7cm thick wooden box tiled with 6cm
acoustic foam on all sides to minimize echoes. The walls, floor and ceiling of the acoustic box had a maximum Target Strength of 40
dB at 20–30 kHz, dropping off to 50 dB at 65–70 kHz, and the box was located in the quiet basement of the faculty of life sciences
TAU, behind thick walls. The box’s walls dampened sounds generate outside the box by at least 100 dB SPL in the relevant frequency
range (see below details regarding the estimation of the acoustic box acoustics). Two cable holes, 2 cm radius each, were located in
two corners of the box and covered with PVC and acoustic foam. Inside the acoustic box were only the recorded plants and 6 mi-
crophones connected to an UltraSoundGate 1216H A/D converter (Avisoft). The PC and all the electricity connections were in the
room outside the acoustic box. Two USB cables connected the PC to the 1216H device inside the box, through the holes. There
was no light inside the acoustic box, and the testing was performed in the dark.
The recordings were performed using a condenser CM16 ultrasound microphone (Avisoft), digitized using an UltraSoundGate
1216H A/D converter (Avisoft), and stored onto a PC. The sampling rate was 500 kHz per channel. We filtered the recordings using
20 kHz high-pass filter. A recording was saved only if triggered with a sound which exceeded 2% of the maximum dynamic range of
the microphone. In such cases, a recording of length 1.5 s was saved, starting half a second before the trigger and lasting 1 s after the
trigger. Each recording produced a 6-channel wav file (a channel for each microphone). Two microphones were directed at each
plant stem, keeping a clear line of sight with no dampening leaves between the stem and the microphones, from a distance of
10 cm. Only sounds that were recorded by both microphones directed at the same plant were considered as ‘‘plant sounds’’ in
the analysis afterward. The frequency responses of the microphones can be found in Avisoft website. No sounds were recorded
by both microphones from empty pots (with either dry or wet soil), over several weeks of recordings, demonstrating the effectiveness
of the box. No sounds were recorded from empty pots even when sharing a box with stressed plants, suggesting that cross talk
between microphones was not significant in our setting.
In the greenhouse
The recordings were performed in a greenhouse in Tel-Aviv University. During the greenhouse experiment, the only plants that were
present inside the greenhouse were the recorded plants. The recordings were performed using the same hardware and setting, as
mentioned in the acoustic box recording section, with three plants in each round and two microphones oriented at each plant, apart
from the acoustic box itself.
Data pre-processing
Data processing was performed offline using python 3.6 code we developed, with the following steps: 1. Identifying the microphone
that had recorded the highest intensity peak in each 1.5 s recording file. 2. Selecting only the sounds that were detected by two
microphones oriented at the same plant at the same time, and saving the one that triggered the recording for further analysis. 3.
Focusing on a short segment of 2ms around the peak. A processed recording includes a segment of 1ms before and 1ms after
the peak of a recorded sound that triggered the system to record. ‘‘Greenhouse noise’’ sounds were obtained when the greenhouse
included only acoustic equipment without plants or pots, by the two microphones (thus excluding ‘‘Electrical noise’’ registered on a
single microphone).
Microphone calibration
The Avisoft CM16 microphone’s sensitivity is 500 mV/Pa and is almost flat between 10 and 120 kHz (see http://www.avisoft.com/
ultrasound-microphones/cm16-cmpa/). The sensitivity of the microphones used in this experiment were calibrated using a calibrated
GRAS (40DP) microphone by emitting signals covering the relevant frequency range toward the two microphones when positioned at
the same distance relative to the speaker. For playback, an ultrasonic speaker (Vifa, connected to an Avisoft 116 D/A converted) was
used to play back sound sweeps in the relevant frequency range (1 ms tones at frequencies between 70 and 20kHz(. This procedure
is routinely used in the lab for various acoustic measurements (see e.g.,58). Our GRAS microphone is routinely calibrated using a Lar-
son Davis CAL150 calibrator with a 1kHz 94 dB SPL calibration tone. The CM16 microphone is much more directional than the GRAS
so its calibration is accurate only when the sound is arriving on-axis. CM16 directionality can be found here (http://www.avisoft.com/
ultrasound-microphones/cm16-cmpa/). Our sound intensity estimates are thus a lower bound on the actual intensity.
For statistical analysis of the number of sound emissions for the treatment and the control groups (Figure 1F) we used the Wilcoxon
rank-sum test. To compare our classifier to random result (Figure 2B), we used the binomial probability distribution function (PDF) and
calculate the probability to get the classifier accuracy or higher randomly for each group. To compare the results obtained when using
scattering network26 for feature extraction to the results obtained when using MFCC34 or basic feature29,30 extraction methods (Fig-
ure 1F), we used Wilcoxon sign rank test with Holm-Bonferroni correction. To test the success in distinguishing between drought-
stressed and control plants (Figure 2) we used Fisher’s exact test. To test the differences in the number of emitted sounds per
day (Figure 3A) we used Wilcoxon signed-ranks tests with Holm-Bonferroni adjustment for multiple comparisons.
Supplemental figures
A
D E
SVM-Scattering Network 1 Raw
Balanced Accuracy (%)
SVM-Scattering Network 2
SVM-Scattering Network 3 Normalized
SVM-Scattering Network 4
SVM-Scattering Network 5
SVM-Scattering Network 6
Figure S1. Plant sounds recorded in the acoustic box, related to Figure 1
(A) Plant sound spectrograms. Spectrograms of sounds emitted by stressed plants: a drought stressed tomato, a drought stressed tobacco, a cut tomato and a
cut tobacco. The corresponding time signals and spectra of these sounds are shown in Figures 1C and 1D.
(B) Mean spectra of the sounds recorded in the acoustic box. Average signal spectrum of five groups of normalized sounds: Emitted by dry tomatoes, cut to-
matoes, dry tobacco, cut tobacco, and electronic noise. The light blue areas around the line represent the standard error of the mean.
(C) Example of electronic noise. Top—time signal of the noise, normalized to the peak. Bottom—spectrum of the sound shown above. The electronic sound was
recorded in the acoustic box, but only by one microphone, and was thus not considered as ‘‘plant sound’’ for further analysis.
(D) Comparison of different scattering network configurations. The accuracy of sound classification with 6 different configurations for the scattering network,
using SVM as classifier. Each line represents a different feature set, all obtained by scattering network. The scattering networks had different time intervals,
different Q-factors, and a different number of filters, for exact values see Table S1.
(E) The accuracy of sound classification achieved by CNN models. The balanced accuracy of sound classification achieved by binary CNN models (see STAR
Methods), for training and testing on the raw sound-vectors (black) and for training and testing on sound-vectors that were normalized by dividing each value in a
vector by the vector’s maximal absolute value. Each balanced accuracy result is based on leave-one-plant-out cross validation (LOPOCV, see STAR Methods).
ll
Article OPEN ACCESS
Figure S2. Analysis of plant sounds recorded in the greenhouse, related to Figure 2
(A) Success of CNN models in distinguishing between tomato sounds and greenhouse (GH) noises. Confusion matrices showing the success of trained CNN
classifiers in distinguishing between tomato sounds and GH noises, with different corrections: raw sound-vectors (upper-left), sound vectors that were
normalized by dividing each value in a vector by the vector’s maximal absolute value (upper-right); raw sound vectors that were super-imposed with raw white
noise vectors (WN) from the opposite group (tomato sound + white noise from the GH, GH sound + white noise from the acoustic box, lower-left); normalized
sound vectors that were super-imposed with a normalized white noise vector from the opposite group (lower-right). These confusion matrices are the summation
of the confusion matrices that are produced in the LOPOCV process (see STAR Methods).
(B) Illustration of the construction of super-imposed sounds. First, libraries of white noises (WN) from the acoustic box and from the greenhouse are generated, by
taking ten 2ms segments from the time segment 200-400ms of each sound from the dry tomato library (acoustic box) and from the empty greenhouse noises.
These selected segments appear 100-300ms before the signal that triggered the recording (see STAR Methods in the main text), and they contain white noises
characteristic to the recording environment. We then generate super-imposed sounds, by summing (element-wise) the tomato sound vectors, each with a
randomly selected greenhouse white noise vector, and by summing the greenhouse sounds, each with a randomly selected acoustic box white noise vector. After
the summation we trim the vectors so that all values greater than 1 are set to 1, and all values smaller than 1 are set to 1.
(C) Tomato-classified sounds per hour. We recorded dry and irrigated tomato plants for 1 h in the green house. Then, using a trained Convolutional Neural
Network (CNN) we filtered the greenhouse noises and left only the tomato classified sounds. We here show the number of tomato-classified sounds, recorded
during 1 h in the greenhouse for dry and irrigated tomato plants. The y axis is truncated at 40 for better resolution; plant 22 emitted 95 sounds.
(D) Receiver operating characteristic (ROC) curve presenting the performance of the classification model used to distinguish between dry and irrigated tomato
plants. We plot the true positive rate as function of the false positive rate for all possible decision threshold, where a decision threshold is of the type: above a
certain number ðnc Þ of tomato-classified sounds per hour we classify the plant as dry, otherwise we classify the plant as irrigated.
ll
OPEN ACCESS Article
Figure S3. Acoustic manifestations of dehydration and other stress factors in multiple plant species, related to Figure 3
(A) The proportion of sounds per day along nine consecutive days of dehydration in tomato plants. The dots represent the average of 23 plants, while the error bars
represent the standard errors. We find significant differences in the proportion of emitted sounds between the first and second day, and between the second
and third day (p values are <0.01 and <0.001 respectively; p values where calculated using Wilcoxon signed-ranks tests, adjusted for 8 comparisons between
pairs of consecutive days using Holm-Bonferroni method). The soil volumetric water content (VWC) of the plants at the beginning of the recording was 0:21 ± 0:03
ðm3 =m3 ; mean ± std).
(B) The proportion of sounds per hour is plotted as a function of the time of day. Each dot represents the average proportion of sounds emitted in the relevant hour
over all 23 plants and nine days, while the error bars represent the standard errors. Dark gray areas represent the hours of complete darkness, light gray areas
show when the greenhouse lighting was on, and the white area represents the approximate hours of natural daylight.
(C) A bar plot showing the proportion of emitted sounds as a function of the plant VWC during sound emission. The bars present the averages of the 23 plants,
while the bars represent the standard errors.
(D) Confusion matrix showing the success of the trained CNN classifiers (see STAR Methods) in distinguishing between sounds emitted by a plant experiencing
VWC<0.05 and sounds emitted by a plant experiencing VWC>0.05, in a LOPOCV process (see STAR Methods). The balanced accuracy score is 72.5%.
(E) The success of CNN models in distinguishing between sounds emitted by tomato plants with low and higher VWC values without pre-classification. Confusion
matrices showing the success of CNN classifiers in distinguishing between sounds that are linked to a plant experiencing VWC<0.01, sounds that are linked to a
plant experiencing VWC>0.05 and sounds recorded in an empty greenhouse (GH sounds). The analysis in all four sub-panels is based on the recordings of the
multi-day experiment (Figure 3). The recorded sounds (either emitted by a plant, or greenhouse noises) were tagged according to the VWC of the plant at the time
of the sound emission. These sounds, along with the empty GH sounds library, were fed to 3-class CNN classifiers. Four pre-processing procedures were used,
and a confusion matrix was calculated for each procedure, while applying LOPOCV process (see STAR Methods): all sounds from the GH experiment were
included and the raw sound vectors were used (upper-left); all sounds from the GH experiment were included and the sound vectors were normalized (upper-
right); sounds from the GH plant-recordings were pre-classified by our CNN, and only sounds classified as tomato sounds were used, as raw sound vectors
(lower-left); sounds from the GH plants-recordings were pre-classified by our CNN, and only sounds classified as tomato sounds were used. All the sounds were
normalized (lower-right). (F) Examples of recorded sounds from different plants. Left - Examples of time signals of 6 plants (top to bottom): Triticum aestivum
(Wheat, dry), Zea mays (Corn, dry), Vitis vinifera (Grapevine, Cabernet Sauvignon variety, cut), Mammilaria spinosissima (Pincushion Cactus, cut), Lamium
amplexicaule (Henbit deadnettle, cut), and Solanum lycopersicum (Tomato, infected with TMV). Right - Spectra of the sounds on the left, normalized. Each of
these plants emitted a significant amount of sounds when stressed, suggesting that sound emission is common under stress among non-woody plants. We have
not succeeded recording from woody plants (almond tree, woody parts of grapevine) using our methodology.
ll
OPEN ACCESS Article
Figure S4. Additional investigation of the mechanism of plant sound emission, related to Figure 4
(A) Sound frequency and trachea cross-section width. Boxplots of the distribution of the plant’s max frequencies vs. its tracheas’ cross section width (estimated
as the average of the ellipse width and height, see examples below) for grapevine, tobacco, and tomato.
(B) Examples of three photos used for the estimation of trachea dimensions. Grapevine (Left) Tobacco (middle), and tomato (right). We performed horizontal cross
section of tomato, tobacco and grapevine stems to measure their trachea cross section area from pictures taken under the microscope, as shown, using
ImageJ.59
(C) Classification of sounds emitted by plants that experienced both dehydration and cutting. The sounds emitted by cut-and-dry plants in the acoustic box were
classified as either ‘‘cut’’ or ‘‘dry’’. We see that while the overall fraction of sounds classified as ‘‘dry’’ in that experiment is around 55%, it varies with time: right
after cutting, sounds classified as ‘‘cut’’ are more common, but from day 2 the majority of sounds are classified as ‘‘dry’’.
(D) Comsol simulations demonstrating sound propagating in all directions from within a stem-like structure. Presented is the 3D far-field sound field.
(E) Horizontal cross section through the sound field from (D) (different colors show different heights along the stem). The simulation shows that sound spreads
equally in all directions in the horizontal plain. The virtual model included two concentric silicon tubes (3.6cm diameter and 4 cm diameter) representing the xylem
vessels within the stem. Because the size was ca. 10 times larger than a real plant, we used a frequency ca. 10 times lower – 7500Hz, thus maintaining the ratio
similar to the real plant. The length of the stem model was 45cm to avoid any effects of length.