Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (13)

Search Parameters:
Keywords = data sonification

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
34 pages, 11454 KiB  
Article
Compassionate Care with Autonomous AI Humanoid Robots in Future Healthcare Delivery: A Multisensory Simulation of Next-Generation Models
by Joannes Paulus Tolentino Hernandez
Biomimetics 2024, 9(11), 687; https://doi.org/10.3390/biomimetics9110687 - 11 Nov 2024
Viewed by 846
Abstract
The integration of AI and robotics in healthcare raises concerns, and additional issues regarding autonomous systems are anticipated. Effective communication is crucial for robots to be seen as “caring”, necessitating advanced mechatronic design and natural language processing (NLP). This paper examines the potential [...] Read more.
The integration of AI and robotics in healthcare raises concerns, and additional issues regarding autonomous systems are anticipated. Effective communication is crucial for robots to be seen as “caring”, necessitating advanced mechatronic design and natural language processing (NLP). This paper examines the potential of humanoid robots to autonomously replicate compassionate care. The study employs computational simulations using mathematical and agent-based modeling to analyze human–robot interactions (HRIs) surpassing Tetsuya Tanioka’s TRETON. It incorporates stochastic elements (through neuromorphic computing) and quantum-inspired concepts (through the lens of Martha Rogers’ theory), running simulations over 100 iterations to analyze complex behaviors. Multisensory simulations (visual and audio) demonstrate the significance of “dynamic communication”, (relational) “entanglement”, and (healthcare system and robot’s function) “superpositioning” in HRIs. Quantum and neuromorphic computing may enable humanoid robots to empathetically respond to human emotions, based on Jean Watson’s ten caritas processes for creating transpersonal states. Autonomous AI humanoid robots will redefine the norms of “caring”. Establishing “pluralistic agreements” through open discussions among stakeholders worldwide is necessary to align innovations with the values of compassionate care within a “posthumanist” framework, where the compassionate care provided by Level 4 robots meets human expectations. Achieving compassionate care with autonomous AI humanoid robots involves translating nursing, communication, computer science, and engineering concepts into robotic care representations while considering ethical discourses through collaborative efforts. Nurses should lead the design and implementation of AI and robots guided by “technological knowing” in Rozzano Locsin’s TCCN theory. Full article
(This article belongs to the Special Issue Optimal Design Approaches of Bioinspired Robots)
Show Figures

Figure 1

Figure 1
<p>Interpretation of Tanioka’s [<a href="#B10-biomimetics-09-00687" class="html-bibr">10</a>] model according to cybernetic HRI communication [<a href="#B92-biomimetics-09-00687" class="html-bibr">92</a>].</p>
Full article ">Figure 2
<p>Communication in “Level 3” HRI [<a href="#B92-biomimetics-09-00687" class="html-bibr">92</a>].</p>
Full article ">Figure 3
<p>Model validation for “Level 3” HRI [<a href="#B92-biomimetics-09-00687" class="html-bibr">92</a>].</p>
Full article ">Figure 4
<p>The representation of dissonance with “Level 3” HRI [<a href="#B92-biomimetics-09-00687" class="html-bibr">92</a>]. (Download the file at <a href="https://github.com/jphernandezrn/Data-Sonification-Human-Robot-Interaction" target="_blank">https://github.com/jphernandezrn/Data-Sonification-Human-Robot-Interaction</a> (accessed on 25 August 2024).</p>
Full article ">Figure 5
<p>The representation of Level 4 HRI. (Note: The mathematics in quantum communication is referenced from Yuan and Cheng [<a href="#B94-biomimetics-09-00687" class="html-bibr">94</a>], when discussing fidelity).</p>
Full article ">Figure 6
<p>The communication, entanglement, and superpositioning of the three states.</p>
Full article ">Figure 7
<p>Model validation involving overlapping states.</p>
Full article ">Figure 8
<p>The sonification of frequencies between states exhibiting quantum relationships. (Download the file at <a href="https://github.com/jphernandezrn/Data-Sonification-Human-Robot-Interaction" target="_blank">https://github.com/jphernandezrn/Data-Sonification-Human-Robot-Interaction</a>).</p>
Full article ">Figure 9
<p>An intuitive, self-regulating, and agile robot system architecture through steps 1–9. Note: <sup>a</sup> Information processing must be dynamic, symbolically instantiated (unsupervised), and evolving (unbounded materially) through <sup>c</sup> “state transition” (the humanoid robot’s conditions based on actions or events). Unbounded transitions refer to a system’s capacity for an unlimited number of transitions between states, often occurring when the conditions for transitions are not strictly defined or when the system can respond to a wide variety of inputs. In the real world, second-order cybernetics [<a href="#B99-biomimetics-09-00687" class="html-bibr">99</a>] should allow the operation of artificial cognition that is fluid and capable of co-creating knowledge within the healthcare network. <sup>b</sup> Alternatively, it can involve the construction and decomposition of “information granules” (the chunks of information) [<a href="#B95-biomimetics-09-00687" class="html-bibr">95</a>], applicable to both algorithmic (deductive) and non-algorithmic (inductive and abductive) computing using quantum logic. This process evolves through machine learning with quantum logic.</p>
Full article ">Figure 10
<p>Care actions and intentionality construed from wave function collapse.</p>
Full article ">Figure 11
<p>Model validation using machine learning.</p>
Full article ">Figure 12
<p>The data sonification of simulated care actions. Download the file at <a href="https://github.com/jphernandezrn/Data-Sonification-Human-Robot-Interaction" target="_blank">https://github.com/jphernandezrn/Data-Sonification-Human-Robot-Interaction</a> (accessed on 25 August 2024).</p>
Full article ">Figure 13
<p>The spectrogram comparison of the three audio files.</p>
Full article ">Figure 14
<p>The mathematical model simulation of “stochasticity” and “intentionality” in the humanoid robot. Note: The blue line represents the relationship between “stochasticity” and “intentionality” in a neuromorphic circuit, as modeled by the equation <span class="html-italic">I</span> = 0.5278 + 0.0666<span class="html-italic">S</span> − 0.0565<span class="html-italic">S</span><sup>2</sup>.) The pattern exhibits three distinct phases: Initial Rise (0.0 to ~0.45); Peak Plateau (~0.45 to ~0.8); and Final Decline (~0.8 to 1.0).</p>
Full article ">Figure 15
<p>The mathematical model simulation of adaptive learning in the humanoid robot. Note: The blue line (“Initial”) shows the robot’s behavior before learning, characterized by jagged fluctuations due to varying levels of randomness (stochasticity). In contrast, the red line (“After Learning”) presents a smoother curve with less variability, indicating enhanced stability after learning. Both lines begin at around 0.5275 intentionality, peak at approximately 0.5475 at “medium stochasticity” (0.6), where there is a balanced mix of predictability and unpredictability, and then decline as stochasticity approaches 1.0. The main difference is that the red line represents a more optimized response, showing that adaptive learning has resulted in more controlled and predictable behavior while maintaining the relationship between “stochasticity” and “intentionality”.</p>
Full article ">Figure 16
<p>Neuromorphic circuit design.</p>
Full article ">Figure 17
<p>Quantum-neuromorphic circuit design.</p>
Full article ">Figure 18
<p>Quantum-neuromorphic circuit simulation.</p>
Full article ">Figure 19
<p>The data sonification of the quantum-neuromorphic circuit simulation. Note: The ‘x’ symbols in (<b>A</b>) mark the peak amplitudes of the quantum-neuromorphic circuit’s waveform, indicating moments of maximum oscillation in the system’s behavior. (Download the file at <a href="https://github.com/jphernandezrn/Data-Sonification-Human-Robot-Interaction" target="_blank">https://github.com/jphernandezrn/Data-Sonification-Human-Robot-Interaction</a>).</p>
Full article ">
28 pages, 4312 KiB  
Article
Application of Convolutional Neural Network for Decoding of 12-Lead Electrocardiogram from a Frequency-Modulated Audio Stream (Sonified ECG)
by Vessela Krasteva, Ivo Iliev and Serafim Tabakov
Sensors 2024, 24(6), 1883; https://doi.org/10.3390/s24061883 - 15 Mar 2024
Cited by 1 | Viewed by 1639
Abstract
Research of novel biosignal modalities with application to remote patient monitoring is a subject of state-of-the-art developments. This study is focused on sonified ECG modality, which can be transmitted as an acoustic wave and received by GSM (Global System for Mobile Communications) microphones. [...] Read more.
Research of novel biosignal modalities with application to remote patient monitoring is a subject of state-of-the-art developments. This study is focused on sonified ECG modality, which can be transmitted as an acoustic wave and received by GSM (Global System for Mobile Communications) microphones. Thus, the wireless connection between the patient module and the cloud server can be provided over an audio channel, such as a standard telephone call or audio message. Patients, especially the elderly or visually impaired, can benefit from ECG sonification because the wireless interface is readily available, facilitating the communication and transmission of secure ECG data from the patient monitoring device to the remote server. The aim of this study is to develop an AI-driven algorithm for 12-lead ECG sonification to support diagnostic reliability in the signal processing chain of the audio ECG stream. Our methods present the design of two algorithms: (1) a transformer (ECG-to-Audio) based on the frequency modulation (FM) of eight independent ECG leads in the very low frequency band (300–2700 Hz); and (2) a transformer (Audio-to-ECG) based on a four-layer 1D convolutional neural network (CNN) to decode the audio ECG stream (10 s @ 11 kHz) to the original eight-lead ECG (10 s @ 250 Hz). The CNN model is trained in unsupervised regression mode, searching for the minimum error between the transformed and original ECG signals. The results are reported using the PTB-XL 12-lead ECG database (21,837 recordings), split 50:50 for training and test. The quality of FM-modulated ECG audio is monitored by short-time Fourier transform, and examples are illustrated in this paper and supplementary audio files. The errors of the reconstructed ECG are estimated by a popular ECG diagnostic toolbox. They are substantially low in all ECG leads: amplitude error (quartile range RMSE = 3–7 μV, PRD = 2–5.2%), QRS detector (Se, PPV > 99.7%), P-QRS-T fiducial points’ time deviation (<2 ms). Low errors generalized across diverse patients and arrhythmias are a testament to the efficacy of the developments. They support 12-lead ECG sonification as a wireless interface to provide reliable data for diagnostic measurements by automated tools or medical experts. Full article
(This article belongs to the Special Issue Advances in ECG/EEG Monitoring)
Show Figures

Figure 1

Figure 1
<p>Generalized concept for using sonified ECG in remote patient monitoring.</p>
Full article ">Figure 2
<p>Design concept for training and testing the signal processing modules of the audio ECG stream, including Transformer (ECG-to-Audio) and Transformer (Audio-to-ECG) in a PC simulation platform. The magnifier indicates the observational points in the test process.</p>
Full article ">Figure 3
<p>Design diagram of two modules used as a frequency modulator (FM) Transformer (ECG-to-Audio) and a FM demodulator Transformer (Audio-to-ECG).</p>
Full article ">Figure 4
<p>Examples of original ECG recordings (leads I, II, V1–V6) (top), generated audio ECG signals (middle) and their short-time Fourier transform (STFT) spectra (bottom) for various arrhythmias found in the PTB-XL test dataset: (<b>a</b>) Recording id 10963 from a patient with normal sinus rhythm. The ECG audio file is recorded in <a href="#app1-sensors-24-01883" class="html-app">&lt;Supplement S1.wav&gt;</a>. (<b>b</b>) Recording id 10224 from a patient with atrial fibrillation and premature ventricular contractions. The ECG audio file is recorded in <a href="#app1-sensors-24-01883" class="html-app">&lt;Supplement S2.wav&gt;</a>.</p>
Full article ">Figure 5
<p>Examples of original ECG recordings (leads I, II, V1–V6) (top), generated audio ECG signals (middle) and their short-time Fourier transform (STFT) spectra (bottom) for various arrhythmias found in the PTB-XL test dataset: (<b>a</b>) Recording id 10967 from a patient with diagnostic labels for premature ventricular contraction(s), premature atrial contraction(s), sinus rhythm, left bundle branch block, and ischemic heart disease. The ECG audio file is recorded in <a href="#app1-sensors-24-01883" class="html-app">&lt;Supplement S3.wav&gt;</a>; (<b>b</b>) Recording id 10355 from a patient with diagnostic labels for premature ventricular contractions and bigeminy. The ECG audio file is recorded in <a href="#app1-sensors-24-01883" class="html-app">&lt;Supplement S4.wav&gt;</a>.</p>
Full article ">Figure 6
<p>Example of original ECG (<b>top</b>), transformed ECG (<b>middle</b>), and their absolute difference (<b>bottom</b>) used for computation of the amplitude error RMSE = 3.8 μV and PRD = 3.4%. The figure reproduces the test PTB-XL dataset recording (id 10224, lead V6) of a patient with atrial fibrillation and premature ventricular contractions. The audio ECG streams of the original and transformed ECG are shown in <a href="#sensors-24-01883-f004" class="html-fig">Figure 4</a>b.</p>
Full article ">Figure 7
<p>Statistical distributions of the amplitude errors measured for the transformed ECG vs. original ECG in separate leads (I, II, V1–V6) of the test set: (<b>a</b>) RMSE: root mean squared error; (<b>b</b>) PRD: percentage root-mean-square difference. The violin plot wrapping is proportional to the kernel density estimate of the underlying distribution. Median and quartile ranges are also denoted.</p>
Full article ">Figure 8
<p>Illustration of the QRS detector performance: example of transformed ECG signal (test PTB-XL dataset, recording id 10224, lead V3) and marked R-peak positions as detected from the reference and test measurements. TP: true positives; FN: false negatives.</p>
Full article ">Figure 9
<p>Illustration of the fiducial point measurement performance: example of transformed ECG signal (test PTB-XL dataset, recording id 10224, lead V3) and marked fiducial point positions as detected from the reference and test measurements. The absolute difference between the detection times of the reference vs. test measurements is shown next to each fiducial point.</p>
Full article ">Figure 10
<p>Normalized median PSD for lead II, considering original ECG, transformed ECG, and their error. On left: PSD estimation in the full frequency range 0–100 Hz. On right: zoomed zones of interest around PSD peak in the frequency range 0–10 Hz. The graphs represent the median PSD value for a specific frequency as a statistical estimate over all recordings in the test set.</p>
Full article ">Figure 11
<p>Normalized PSD error of the transformed ECG vs. original ECG, estimated in the frequency range 0–100 Hz for 8 ECG leads (I, II, V1–V6). PSD error trends are presented as median value and quartile range (Q25%, Q75%). The carrier frequency (F<sub>C</sub>) of each lead in the audio stream is additionally given in the legend to indicate that presented spectral errors correspond to the ECG after demodulation of specific audio frequency bands.</p>
Full article ">
9 pages, 277 KiB  
Opinion
Enhancing Perceptual—Motor Skills in Sports: The Role of Ecological Sounds
by Tiziano Agostini, Fabrizio Sors, Mauro Murgia and Alessandra Galmonte
J. Intell. 2024, 12(2), 15; https://doi.org/10.3390/jintelligence12020015 - 30 Jan 2024
Cited by 1 | Viewed by 2448
Abstract
Starting approximately from the beginning of the new millennium, a series of studies highlighted that auditory information deriving from biological motion can significantly influence the behavioral, cognitive and neurophysiological processes involved in the perception and execution of complex movements. In particular, it was [...] Read more.
Starting approximately from the beginning of the new millennium, a series of studies highlighted that auditory information deriving from biological motion can significantly influence the behavioral, cognitive and neurophysiological processes involved in the perception and execution of complex movements. In particular, it was observed that an appropriate use of sounds deriving from one’s own movement promotes improvements in the movement execution itself. Two main approaches can be used, namely the sonification one or the ecological sound one; the former is based on the conversion of physiological and/or physical movement data into sound, while the latter is based on the use of auditory recordings of movement sounds as models. In the present article, some of the main applications of both approaches—especially the latter—to the domains of sport and motor rehabilitation are reviewed, with the aim of addressing two questions: Is it possible to consider rhythm as a Gestalt of human movement? If so, is it possible to build up cognitive strategies to improve/standardize movement performance from this Gestalt? As with most topics in science, a definitive answer is not possible, yet the evidence leads us to lean toward a positive answer to both questions. Full article
(This article belongs to the Special Issue Grounding Cognition in Perceptual Experience)
16 pages, 5643 KiB  
Article
A Wearable Sonification System to Improve Movement Awareness: A Feasibility Study
by Frank Feltham, Thomas Connelly, Chi-Tsun Cheng and Toh Yen Pang
Appl. Sci. 2024, 14(2), 816; https://doi.org/10.3390/app14020816 - 18 Jan 2024
Cited by 1 | Viewed by 1691
Abstract
This paper presents the design, development, and feasibility testing of a wearable sonification system for real-time posture monitoring and feedback. The system utilizes inexpensive motion sensors integrated into a compact, wearable package to measure body movements and standing balance continuously. The sensor data [...] Read more.
This paper presents the design, development, and feasibility testing of a wearable sonification system for real-time posture monitoring and feedback. The system utilizes inexpensive motion sensors integrated into a compact, wearable package to measure body movements and standing balance continuously. The sensor data is processed through sonification algorithms to generate real-time auditory feedback cues indicating the user’s balance and posture. The system aims to improve movement awareness and physical conditioning, with potential applications in balance rehabilitation and physical therapy. Initial feasibility testing was conducted with a small group of healthy participants performing standing balance tasks with eyes open and closed. Results indicate that the real-time audio feedback improved participants’ ability to maintain balance, especially in the case of closed eyes. This preliminary study demonstrates the potential for wearable sonification systems to provide intuitive real-time feedback on posture and movement to improve motor skills and balance. Full article
(This article belongs to the Section Biomedical Engineering)
Show Figures

Figure 1

Figure 1
<p>An overview of the design and the signal flow diagram of the wearable auditory feedback system.</p>
Full article ">Figure 2
<p>A WT901BLECL sensor from WitMotion Shenzhen Co., Ltd.</p>
Full article ">Figure 3
<p>The CAF system takes sagittal rotation data to generate continuous and discrete tones.</p>
Full article ">Figure 4
<p>Diagram of the algorithm and auditory feedback system.</p>
Full article ">Figure 5
<p>Verification of the IMU sensor’s integration with MATLAB and Max for effective sonification and user auditory feedback.</p>
Full article ">Figure 6
<p>Time series data plot for Participant 3 across the four test conditions.</p>
Full article ">Figure 7
<p>Time series data plots and spectrogram for Participant 3 across test conditions with and without auditory feedback.</p>
Full article ">Figure 8
<p>Mean change in time series data plots for Participant 3 across four test conditions (+ significance difference <span class="html-italic">p</span> &lt; 0.05).</p>
Full article ">
12 pages, 1732 KiB  
Article
Parameter Mapping Sonification of Human Olfactory Thresholds
by Jean-Luc Boevé and Rudi Giot
Biology 2023, 12(5), 670; https://doi.org/10.3390/biology12050670 - 28 Apr 2023
Cited by 2 | Viewed by 1327
Abstract
An objective of chemical ecology is to understand the chemical diversity across and within species, as well as the bioactivity of chemical compounds. We previously studied defensive volatiles from phytophagous insects that were subjected to parameter mapping sonification. The created sounds contained information [...] Read more.
An objective of chemical ecology is to understand the chemical diversity across and within species, as well as the bioactivity of chemical compounds. We previously studied defensive volatiles from phytophagous insects that were subjected to parameter mapping sonification. The created sounds contained information about the repellent bioactivity of the volatiles, such as the repellence from the volatiles themselves when tested against live predators. Here, we applied a similar sonification process to data about human olfactory thresholds. Randomized mapping conditions were used and a peak sound pressure, Lpeak, was calculated from each audio file. The results indicate that Lpeak values were significantly correlated with the olfactory threshold values (e.g., rS = 0.72, t = 10.19, p < 0.001, Spearman rank-order correlation; standardized olfactory thresholds of 100 volatiles). Furthermore, multiple linear regressions used the olfactory threshold as a dependent variable. The regressions revealed that the molecular weight, the number of carbon and oxygen atoms, as well as the functional groups aldehyde, acid, and (remaining) double bond were significant determinants of the bioactivity, while the functional groups ester, ketone, and alcohol were not. We conclude that the presented sonification methodology that converts chemicals into sound data allows for the study of their bioactivities by integrating compound characteristics that are easily accessible. Full article
(This article belongs to the Section Bioinformatics)
Show Figures

Figure 1

Figure 1
<p>Schematic representations of the causal relationships and correlations between elements dealing with volatile compounds studied by sonification in Boevé and Giot [<a href="#B12-biology-12-00670" class="html-bibr">12</a>] (<b>A</b>) and in this study (<b>B</b>). In (<b>B</b>), data about bioassay on humans were gathered from the literature. Datasets are mentioned in a box, ‘objects’ in a circle. (<b>A</b>) causal relationship is shown by an arrow, and a statistical correlation is shown by a thick bar. Notice that in (<b>A</b>), the insect secretion was tested on ants indirectly by confronting them with live insects. Gaz chromatography-mass spectrometry (GC-MS). Molecular weight (MW). Standardized human olfactory threshold (SHOT). Peak sound pressure (Lpeak). For more information, see text.</p>
Full article ">Figure 2
<p>Screenshot of the layout interface used in an illustrative condition of parameter mapping by which chemical parameters are linked to sound parameters (green nodes). Letters and digits with a black background are added to the screenshot, as the mapping preset names refer to them in <a href="#biology-12-00670-t001" class="html-table">Table 1</a>. By clicking the “Record All” (below, right), the Processing script performs a batch export of one audio file per molecule listed in a CSV file. For more information, see text.</p>
Full article ">Figure 3
<p>Scatterplot and regression line of the SHOT d1 versus Lpeak values for 100 molecules. The molecules and their values are listed in <a href="#app1-biology-12-00670" class="html-app">Table S1</a>. The regression formula is Y = −34.474 + 1.9131 X.</p>
Full article ">Figure 4
<p>Peak sound pressure obtained by separately testing sound parameters. Five theoretical, small to large, molecules were created in a CSV file (named “setX”). They were used in parameter mapping conditions that included a reduced set of linking nodes. For more explanation, see text.</p>
Full article ">
21 pages, 26921 KiB  
Article
Medical Augmented Reality: Definition, Principle Components, Domain Modeling, and Design-Development-Validation Process
by Nassir Navab, Alejandro Martin-Gomez, Matthias Seibold, Michael Sommersperger, Tianyu Song, Alexander Winkler, Kevin Yu and Ulrich Eck
J. Imaging 2023, 9(1), 4; https://doi.org/10.3390/jimaging9010004 - 23 Dec 2022
Cited by 16 | Viewed by 5408
Abstract
Three decades after the first set of work on Medical Augmented Reality (MAR) was presented to the international community, and ten years after the deployment of the first MAR solutions into operating rooms, its exact definition, basic components, systematic design, and validation still [...] Read more.
Three decades after the first set of work on Medical Augmented Reality (MAR) was presented to the international community, and ten years after the deployment of the first MAR solutions into operating rooms, its exact definition, basic components, systematic design, and validation still lack a detailed discussion. This paper defines the basic components of any Augmented Reality (AR) solution and extends them to exemplary Medical Augmented Reality Systems (MARS). We use some of the original MARS applications developed at the Chair for Computer Aided Medical Procedures and deployed into medical schools for teaching anatomy and into operating rooms for telemedicine and surgical guidance throughout the last decades to identify the corresponding basic components. In this regard, the paper is not discussing all past or existing solutions but only aims at defining the principle components and discussing the particular domain modeling for MAR and its design-development-validation process, and providing exemplary cases through the past in-house developments of such solutions. Full article
Show Figures

Figure 1

Figure 1
<p><b>The apparatus of Brunelleschi</b> consists of (<b>A</b>) a mirror, and a painted or printed image. Both components inhibit a hole for the user to look through. (<b>B</b>) The apparatus creates a linear projection that shows the image inside the user’s view. (<b>C</b>) The illustrated third-person view visualizes the frustum that is covered by the projection. Many modern AR applications use the same mathematical basis for creating in situ visualizations.</p>
Full article ">Figure 2
<p><b>The Medical Augmented Reality Framework</b> consists of four primary components: Digital World, AR/VR Display, AR/VR User Interaction, and evaluation. An MARS perceives the Physical World with its sensors and processes in a medium that users may perceive and interact with through the AR/VR interfaces. Evaluation is integral to the system’s conception, development, and deployment.</p>
Full article ">Figure 3
<p><b>Augmented Reality Teleconsultation System for Medicine</b> (ArTekMed) combines point cloud reconstruction with Augmented and Virtual Reality. (<b>A</b>) Capturing the local site with the patient requires extrinsically calibrated RGB-D sensors from which the system computes a real-time point cloud reconstruction. (<b>B</b>) The local user interacts with the real world while perceiving additional virtual content delivered with AR. (<b>C</b>) The remote user dons a VR headset and controller for interacting with the acquired point cloud. (<b>D</b>) The reconstruction represents the digital world known to the computer and is displayed to the VR User. (<b>E</b>) AR annotations made by the VR user is shown in situ on the patient.</p>
Full article ">Figure 4
<p><b>Interaction Techniques Unique to ArTekMed</b>: (<b>A</b>) The Magnorama creates a dynamic 3D cutout from the real-time reconstruction and allows the user to interact and explore the space while intuitively creating annotations at the original region of interest within the duplicate. (<b>B</b>) The principle of Magnoramas translates well into AR. The resulting technique of Duplicated Reality allows co-located collaboration in tight spaces, even without a remote user. (<b>C</b>) For remote users to experience more details of the patient and their surroundings, ArTekMed deploys Projective Bisector Mirrors to bridge the gap between reality and reconstruction through the mirror metaphor.</p>
Full article ">Figure 5
<p>Typical setups in ophthalmic surgery consist of a complex operating area and multi-modal real-time imaging. Visual and auditive AR applications aim to improve perception and provide additional information while avoiding visual clutter and reducing the cognitive load of complex intraoperative data.</p>
Full article ">Figure 6
<p><b>CAMC</b> aims to reduce the need for ionizing radiations and to provide spatially aware, intuitive visualization of joint optical and fluoroscopic data. (<b>a</b>) Calibration of the C-arm with the patient and the technician and surgeon’s HMD enables efficient surgical procedures in a collaborative ecosystem. (<b>b</b>) Advanced AR interface aids in better planning trajectories on the X-ray acquisitions. (<b>c</b>) The adaptive UI and augmentations in intra-operative planning and execution support various image-guided procedures.</p>
Full article ">Figure 7
<p><b>The Magic Mirror</b> visualizes anatomical structures in situ on the <span class="html-italic">mirror</span> reflection of the user in front of the RGB-D camera. Additionally, our Magic Mirror system displays transverse slices of a CT volume on the right half of the monitor that matches the slice selected by the user with their right hand.</p>
Full article ">
20 pages, 2187 KiB  
Article
Participatory Design of Sonification Development for Learning about Molecular Structures in Virtual Reality
by Miguel Garcia-Ruiz, Pedro Cesar Santana-Mancilla, Laura Sanely Gaytan-Lugo and Adriana Iniguez-Carrillo
Multimodal Technol. Interact. 2022, 6(10), 89; https://doi.org/10.3390/mti6100089 - 12 Oct 2022
Cited by 2 | Viewed by 2438
Abstract
Background: Chemistry and biology students often have difficulty understanding molecular structures. Sonification (the rendition of data into non-speech sounds that convey information) can be used to support molecular understanding by complementing scientific visualization. A proper sonification design is important for its effective educational [...] Read more.
Background: Chemistry and biology students often have difficulty understanding molecular structures. Sonification (the rendition of data into non-speech sounds that convey information) can be used to support molecular understanding by complementing scientific visualization. A proper sonification design is important for its effective educational use. This paper describes a participatory design (PD) approach to designing and developing the sonification of a molecular structure model to be used in an educational setting. Methods: Biology, music, and computer science students and specialists designed a sonification of a model of an insulin molecule, following Spinuzzi’s PD methodology and involving evolutionary prototyping. The sonification was developed using open-source software tools used in digital music composition. Results and Conclusions: We tested our sonification played on a virtual reality headset with 15 computer science students. Questionnaire and observational results showed that multidisciplinary PD was useful and effective for developing an educational scientific sonification. PD allowed for speeding up and improving our sonification design and development. Making a usable (effective, efficient, and pleasant to use) sonification of molecular information requires the multidisciplinary participation of people with music, computer science, and molecular biology backgrounds. Full article
(This article belongs to the Special Issue Virtual Reality and Augmented Reality)
Show Figures

Figure 1

Figure 1
<p>Students discussing the molecular sonification design.</p>
Full article ">Figure 2
<p>Our PD methodology, based on Spinuzzi [<a href="#B51-mti-06-00089" class="html-bibr">51</a>].</p>
Full article ">Figure 3
<p>The sonification development process.</p>
Full article ">Figure 4
<p>A VR headset displaying a molecular model using Sketchfab.</p>
Full article ">Figure 5
<p>The molecular model of insulin shown on the Sketchfab website.</p>
Full article ">Figure 6
<p>Graph showing the Likert-scale results from <a href="#mti-06-00089-t002" class="html-table">Table 2</a>.</p>
Full article ">Figure A1
<p>Scatter plots of the five questions asked to participants, described in <a href="#mti-06-00089-t002" class="html-table">Table 2</a>.</p>
Full article ">
10 pages, 2264 KiB  
Article
Auditory Display of Fluorescence Image Data in an In Vivo Tumor Model
by Sheen-Woo Lee, Sang Hoon Lee, Zhen Cheng and Woon Seung Yeo
Diagnostics 2022, 12(7), 1728; https://doi.org/10.3390/diagnostics12071728 - 16 Jul 2022
Cited by 2 | Viewed by 1522
Abstract
Objectives: This research aims to apply an auditory display for tumor imaging using fluorescence data, discuss its feasibility for in vivo tumor evaluation, and check its potential for assisting enhanced cancer perception. Methods: Xenografted mice underwent fluorescence imaging after an injection [...] Read more.
Objectives: This research aims to apply an auditory display for tumor imaging using fluorescence data, discuss its feasibility for in vivo tumor evaluation, and check its potential for assisting enhanced cancer perception. Methods: Xenografted mice underwent fluorescence imaging after an injection of cy5.5-glucose. Spectral information from the raw data was parametrized to emphasize the near-infrared fluorescence information, and the resulting parameters were mapped to control a sound synthesis engine in order to provide the auditory display. Drag–click maneuvers using in-house data navigation software-generated sound from regions of interest (ROIs) in vivo. Results: Four different representations of the auditory display were acquired per ROI: (1) audio spectrum, (2) waveform, (3) numerical signal-to-noise ratio (SNR), and (4) sound itself. SNRs were compared for statistical analysis. Compared with the no-tumor area, the tumor area produced sounds with a heterogeneous spectrum and waveform, and featured a higher SNR as well (3.63 ± 8.41 vs. 0.42 ± 0.085, p < 0.05). Sound from the tumor was perceived by the naked ear as high-timbred and unpleasant. Conclusions: By accentuating the specific tumor spectrum, auditory display of fluorescence imaging data can generate sound which helps the listener to detect and discriminate small tumorous conditions in living animals. Despite some practical limitations, it can aid in the translation of fluorescent images by facilitating information transfer to the clinician in in vivo tumor imaging. Full article
(This article belongs to the Section Medical Imaging and Theranostics)
Show Figures

Figure 1

Figure 1
<p>Hyperspectral data, fluorescence only, and an augmented image from parameters for auditory display. Fluorescence images were acquired under an emission filter setting from 640 to 800 [nm] in 10 [nm] increments. Hyperspectral data were overlaid on grayscale photographs to identify the ROI (<b>a</b>). Markers placed on ROIs obtained the pattern of the hyperspectral data (<b>b</b>). Spectral unmixing and overlay of the unmixed data on the grayscale images showed the tumor area with the accentuated NIRF spectrum, confirming the tumor uptake of the NIRF probe (<b>c</b>). Tumor without visible NIRF uptake on fluorescence only (<b>d</b>) was shown to have signal aided by data for auditory display (<b>e</b>).</p>
Full article ">Figure 2
<p>Components of the preprocessor and the sound synthesizer built on Max: (<b>a</b>) shows the algorithm to produce a target value; (<b>b</b>) illustrates the normalization process to provide the parameter value for sound synthesis; and (<b>c</b>) depicts the FM sound synthesis patch used for this study as well as the spectrum and waveform of the result of auditory display.</p>
Full article ">Figure 3
<p>Samples of a session of auditory display. Shown in (<b>a</b>) is the navigation map (JPEG picture of a mouse during an image session). The tumor area is indicated by a single white curved line, the background skull by double white lines, and the non-tumor body by a dotted line. The rest of the figure depicts parameters for sonification as well as the spectrum/waveform of the resulting sounds from the tumor area (<b>b</b>,<b>c</b>), and those of the non-tumor area (<b>d</b>,<b>e</b>).</p>
Full article ">Figure 4
<p>Spectrum and waveforms of the sonified results from each mouse in the experiment. The left column shows the processed NIRF images from the fluorescence imaging system on grayscale photographs of each mouse (<b>a</b>–<b>f</b>), showing variable fluorescence signals from the tumor. The middle column shows the result of auditory display from the tumor area, with the spectrum (<b>above</b>) and waveform (<b>below</b>) from the tumor. The right column shows the spectrum and waveform of the sound from the background (non-tumor) area. Compared with the background, sounds from the tumor areas of all mice featured a heterogeneous spectrum and irregular waveform.</p>
Full article ">
20 pages, 5862 KiB  
Article
Haptic Feedback to Assist Blind People in Indoor Environment Using Vibration Patterns
by Shah Khusro, Babar Shah, Inayat Khan and Sumayya Rahman
Sensors 2022, 22(1), 361; https://doi.org/10.3390/s22010361 - 4 Jan 2022
Cited by 21 | Viewed by 4787
Abstract
Feedback is one of the significant factors for the mental mapping of an environment. It is the communication of spatial information to blind people to perceive the surroundings. The assistive smartphone technologies deliver feedback for different activities using several feedback mediums, including voice, [...] Read more.
Feedback is one of the significant factors for the mental mapping of an environment. It is the communication of spatial information to blind people to perceive the surroundings. The assistive smartphone technologies deliver feedback for different activities using several feedback mediums, including voice, sonification and vibration. Researchers 0have proposed various solutions for conveying feedback messages to blind people using these mediums. Voice and sonification feedback are effective solutions to convey information. However, these solutions are not applicable in a noisy environment and may occupy the most important auditory sense. The privacy of a blind user can also be compromised with speech feedback. The vibration feedback could effectively be used as an alternative approach to these mediums. This paper proposes a real-time feedback system specifically designed for blind people to convey information to them based on vibration patterns. The proposed solution has been evaluated through an empirical study by collecting data from 24 blind people through a mixed-mode survey using a questionnaire. Results show the average recognition accuracy for 10 different vibration patterns are 90%, 82%, 75%, 87%, 65%, and 70%. Full article
(This article belongs to the Special Issue Big Data Analytics in Internet of Things Environment)
Show Figures

Figure 1

Figure 1
<p>Overview of vibration patterns feedback for blind people.</p>
Full article ">Figure 2
<p>Classes of Taxonomy.</p>
Full article ">Figure 3
<p>Feedback sets after the first, second, and third iteration.</p>
Full article ">Figure 4
<p>Downstairs Patterns.</p>
Full article ">Figure 5
<p>Left Movement Patterns.</p>
Full article ">Figure 6
<p>Walking Patterns.</p>
Full article ">Figure 7
<p>Patterns bit combination using Morse code.</p>
Full article ">Figure 8
<p>Attitude towards usage of the proposed solution.</p>
Full article ">Figure 9
<p>Intention to Use.</p>
Full article ">Figure 10
<p>Perceived Usefulness after using the proposed vibration patterns.</p>
Full article ">Figure 11
<p>Understandability and Learnability.</p>
Full article ">Figure 12
<p>Ease of Use.</p>
Full article ">Figure 13
<p>Recognition Accuracy among the proposed patterns after a field test.</p>
Full article ">Figure 14
<p>Average Reaction time per action of the user.</p>
Full article ">Figure 15
<p>Errors occurring by participants while recognition.</p>
Full article ">
22 pages, 466 KiB  
Article
A Comparative Analysis of Modeling and Predicting Perceived and Induced Emotions in Sonification
by Faranak Abri, Luis Felipe Gutiérrez, Prerit Datta, David R. W. Sears, Akbar Siami Namin and Keith S. Jones
Electronics 2021, 10(20), 2519; https://doi.org/10.3390/electronics10202519 - 15 Oct 2021
Cited by 6 | Viewed by 2213
Abstract
Sonification is the utilization of sounds to convey information about data or events. There are two types of emotions associated with sounds: (1) “perceived” emotions, in which listeners recognize the emotions expressed by the sound, and (2) “induced” emotions, in which listeners feel [...] Read more.
Sonification is the utilization of sounds to convey information about data or events. There are two types of emotions associated with sounds: (1) “perceived” emotions, in which listeners recognize the emotions expressed by the sound, and (2) “induced” emotions, in which listeners feel emotions induced by the sound. Although listeners may widely agree on the perceived emotion for a given sound, they often do not agree about the induced emotion of a given sound, so it is difficult to model induced emotions. This paper describes the development of several machine and deep learning models that predict the perceived and induced emotions associated with certain sounds, and it analyzes and compares the accuracy of those predictions. The results revealed that models built for predicting perceived emotions are more accurate than ones built for predicting induced emotions. However, the gap in predictive power between such models can be narrowed substantially through the optimization of the machine and deep learning models. This research has several applications in automated configurations of hardware devices and their integration with software components in the context of the Internet of Things, for which security is of utmost importance. Full article
(This article belongs to the Special Issue Deep Learning for the Internet of Things)
Show Figures

Figure 1

Figure 1
<p>Visualization of EmoSoundscape and IADSE data sets. (<b>a</b>) Scatter plot of the normalized data points of EmoSoundscape in the AV space. (<b>b</b>) Scatter plot of normalized data points of IADSE in the AVD space.</p>
Full article ">
18 pages, 283 KiB  
Article
Stolen Voices Is a Slowly Unfolding Eavesdrop on the East Coast of the UK
by Rebecca Collins and Johanna Linsley
Arts 2019, 8(4), 140; https://doi.org/10.3390/arts8040140 - 23 Oct 2019
Cited by 4 | Viewed by 3635
Abstract
Stolen Voices is a research enquiry that uses listening as both methodology and material. Stolen Voices develops techniques for ‘listening in’ and eavesdropping to help articulate an epistemology of place through sonic frameworks. A core motivation for the listening is a semi-fictional story [...] Read more.
Stolen Voices is a research enquiry that uses listening as both methodology and material. Stolen Voices develops techniques for ‘listening in’ and eavesdropping to help articulate an epistemology of place through sonic frameworks. A core motivation for the listening is a semi-fictional story we tell ourselves (and anyone else who is listening): an ‘event’ has taken place along the East Coast of the United Kingdom (UK), and we have been tasked with figuring out what has happened. While the specifics of the event might be difficult to pin down, the urgency of the investigation is fuelled by concrete concerns found in the UK edgelands, at the border/margin of the country: the uncertain future of the UK’s relationship with Europe; the effects of climate change on coastal landscapes; the waning of industries like manufacturing and coal extraction; the oil industry in crisis; the rise of global shipping infrastructures. By using a semi-fictional framework, we move away from mapping techniques like data-sonification towards a methodology that embraces gaps and inventive excesses while insisting on the importance of making an account. Through listening, we foster attention to contingencies and indeterminacies and their relationships to prevailing structures and knowledge hierarchies. Stolen Voices asks: what is the relationship between a listener and what is heard? How can listening attune us to the complexities of contemporary political, economic, ecological and social processes? How did we get to where we are now, and how, through listening, can we seek out levers for change? What do the rhythms and atmospheres of specific geographic locations inform or reveal about history? Evolving over several years, in response to what we hear, the investigation necessarily proceeds slowly. In this article, we unfold our methodological processes for the detection of sound, voices, atmosphere and affect. We use creative-critical writing to evidence the construction of a research investigation focused on the act of listening as a spatial practice and necessarily collective endeavour. Full article
15 pages, 2734 KiB  
Article
Combining VR Visualization and Sonification for Immersive Exploration of Urban Noise Standards
by Markus Berger and Ralf Bill
Multimodal Technol. Interact. 2019, 3(2), 34; https://doi.org/10.3390/mti3020034 - 13 May 2019
Cited by 23 | Viewed by 4920
Abstract
Urban traffic noise situations are usually visualized as conventional 2D maps or 3D scenes. These representations are indispensable tools to inform decision makers and citizens about issues of health, safety, and quality of life but require expert knowledge in order to be properly [...] Read more.
Urban traffic noise situations are usually visualized as conventional 2D maps or 3D scenes. These representations are indispensable tools to inform decision makers and citizens about issues of health, safety, and quality of life but require expert knowledge in order to be properly understood and put into context. The subjectivity of how we perceive noise as well as the inaccuracies in common noise calculation standards are rarely represented. We present a virtual reality application that seeks to offer an audiovisual glimpse into the background workings of one of these standards, by employing a multisensory, immersive analytics approach that allows users to interactively explore and listen to an approximate rendering of the data in the same environment that the noise simulation occurs in. In order for this approach to be useful, it should manage complicated noise level calculations in a real time environment and run on commodity low-cost VR hardware. In a prototypical implementation, we utilized simple VR interactions common to current mobile VR headsets and combined them with techniques from data visualization and sonification to allow users to explore road traffic noise in an immersive real-time urban environment. The noise levels were calculated over CityGML LoD2 building geometries, in accordance with Common Noise Assessment Methods in Europe (CNOSSOS-EU) sound propagation methods. Full article
(This article belongs to the Special Issue Interactive 3D Cartography)
Show Figures

Figure 1

Figure 1
<p>Standard VR point and click teleportation in our city model.</p>
Full article ">Figure 2
<p>User perspective after scale change. Propagation lines are shown in the last real-scale position, so that propagation can be studied in a larger spatial context.</p>
Full article ">Figure 3
<p>Noise propagation in a street with one observer point and seven source points. Green lines are direct paths and path end segments. Blue represents paths that lead to diffraction around buildings, and yellow paths that cause reflections on vertical building surfaces. The white speaker symbol shows the image sources vertically projected to the ground. (Otherwise, some would end up above the camera.) The observer is represented by a camera, the source points are located at the beginning of the line segments.</p>
Full article ">Figure 4
<p>(<b>a</b>) View on the noise grid from the user’s perspective; (<b>b</b>) top–down view on the noise grid.</p>
Full article ">Figure 5
<p>The segment of CityGML data used for testing this prototype.</p>
Full article ">
15 pages, 1094 KiB  
Article
Dual Mode Gait Sonification for Rehabilitation After Unilateral Hip Arthroplasty
by Julia Reh, Tong-Hun Hwang, Gerd Schmitz and Alfred O. Effenberg
Brain Sci. 2019, 9(3), 66; https://doi.org/10.3390/brainsci9030066 - 19 Mar 2019
Cited by 16 | Viewed by 5618
Abstract
The pattern of gait after hip arthroplasty strongly affects regeneration and quality of life. Acoustic feedback could be a supportive method for patients to improve their walking ability and to regain a symmetric and steady gait. In this study, a new gait sonification [...] Read more.
The pattern of gait after hip arthroplasty strongly affects regeneration and quality of life. Acoustic feedback could be a supportive method for patients to improve their walking ability and to regain a symmetric and steady gait. In this study, a new gait sonification method with two different modes—real-time feedback (RTF) and instructive model sequences (IMS)—is presented. The impact of the method on gait symmetry and steadiness of 20 hip arthroplasty patients was investigated. Patients were either assigned to a sonification group (SG) (n = 10) or a control group (CG) (n = 10). All of them performed 10 gait training sessions (TS) lasting 20 min, in which kinematic data were measured using an inertial sensor system. Results demonstrate converging step lengths of the affected and unaffected leg over time in SG compared with a nearly parallel development of both legs in CG. Within the SG, a higher variability of stride length and stride time was found during the RTF training mode in comparison to the IMS mode. Therefore, the presented dual mode method provides the potential to support gait rehabilitation as well as home-based gait training of orthopedic patients with various restrictions. Full article
(This article belongs to the Section Neurotechnology and Neuroimaging)
Show Figures

Figure 1

Figure 1
<p>Process of intervention with ten training sessions (TS) on twelve days. The control group (CG) did not receive any acoustic feedback, while the sonification group (SG) received real-time feedback (RTF) alternating with instructive model sequences (IMS).</p>
Full article ">Figure 2
<p>Patient of the sonification group during gait training. The temporal course can be observed on the notebook screen.</p>
Full article ">Figure 3
<p>Step length of the affected and unaffected leg for SG (<span class="html-italic">n</span> = 10) (left) in week 1 and week 2, step length of the affected and unaffected leg for CG (<span class="html-italic">n</span> = 10) (right) in week 1 and week 2. Values are means ± standard deviation.</p>
Full article ">Figure 4
<p>Coefficient of variation (COV) of stride length for SG. Values are means ± standard error.</p>
Full article ">Figure 5
<p>Coefficient of variation (COV) of stride time for SG. Values are means ± standard error.</p>
Full article ">
Back to TopTop