Imperial College London

DrLorenzoPicinali

Faculty of EngineeringDyson School of Design Engineering

Reader in Audio Experience Design
 
 
 
//

Contact

 

l.picinali Website CV

 
 
//

Location

 

Level 1 staff officeDyson BuildingSouth Kensington Campus

//

Summary

 

Publications

Publication Type
Year
to

112 results found

Reyes-Lecuona A, Cuevas-Rodríguez M, González-Toledo D, Garre C, De-La-Rubia-Cuestas E, Molina-Tanco L, Rodríguez-Rivero Á, Picinali Let al., 2019, The 3D Tune-In Toolkit: A C++ library for binaural spatialisation, and hearing loss / hearing AIDS emulation, 23rd International Congress on Acoustics, Pages: 2217-2218, ISSN: 2226-7808

This contribution presents the 3D Tune-In (3DTI) Toolkit, an open source C++ library for binaural spatialisationwhich includes hearing loss and hearing aid emulators. Binaural spatialisation is performed through convolutionwith user-imported Head Related Transfer Functions (HRTFs) and Binaural Room Impulse Responses (BRIRs),including additional functionalities such as near- and far-field source simulation, customisation of InterauralTime Differences (ITDs), and Ambisonic-based binaural reverberation. Hearing loss is simulated through gammatone filters and multiband expanders/compressors, including advanced non-linear features such as frequencysmearing and temporal distortion. A generalised hearing aid simulator is also included, with functionalities suchas dynamic equalisation, calibration from user-inputted audiograms, and directional processing.

Conference paper

Scase M, Griffin E, Picinali L, 2019, Pre-Navigation via Interactive Audio Tactile Maps to Promote the Wellbeing of Visually Impaired People., Stud Health Technol Inform, Vol: 260, Pages: 170-177

BACKGROUND: Pre-navigational tools can assist visually impaired people when navigating unfamiliar environments. Assistive technology products (eg tactile maps or auditory simulations) can stimulate cognitive mapping processes to provide navigational assistance in these people. OBJECTIVES: We compared how well blind and visually impaired people could learn a map presented via a tablet computer auditory tactile map (ATM) in contrast to a conventional tactile map accompanied by a text description objectives. METHODS: Performance was assessed with a multiple choice test that quizzed participants on orientation and spatial awareness. Semi-structured interviews explored participant experiences and preferences. RESULTS: A statistically significant difference was found between the conditions with participants using the ATM performing much better than those who used a conventional tactile map and text description. Participants preferred the flexibility of learning of the ATM. CONCLUSION: This computer-based ATM provided an effective, easy to use and cost-effective way of enabling blind and partially sighted people learn a cognitive map and enhance their wellbeing.

Journal article

Comunita M, Picinali L, 2019, Estimating ear canal volume through electrical impedance measurements from in-ear headphones - initial results, AES International Conference on Headphone Technology, Publisher: AUDIO ENGINEERING SOC INC

Conference paper

Cuevas-Rodriguez M, Gonzalez-Toledo D, La Rubia-Cuestas ED, Garre C, Molina-Tanco L, Reyes-Lecuona A, Poirier-Quinot D, Picinali Let al., 2018, The 3D Tune-In Toolkit - 3D audio spatialiser, hearing loss and hearing aid simulations

The 3DTI Toolkit is a standard C++ library for audio spatialisation and simulation using loudspeakers or headphones developed within the 3D Tune-In (3DTI) project (http://www.3d-tune-in.eu), which aims at using 3D sound and simulating hearing loss and hearing aids within virtual environments and games. The Toolkit allows the design and rendering of highly realistic and immersive 3D audio, and the simulation of virtual hearing aid devices and of different typologies of hearing loss. The library includes a real-time 3D binaural audio renderer offering full 3D spatialization based on efficient Head Related Transfer Function (HRTF) convolution, including smooth interpolation among impulse responses, customization of listener head radius and specific simulation of far-distance and near-field effects. In addition, spatial reverberation is simulated in real time using a uniformly partitioned convolution with Binaural Room Impulse Responses (BRIRs) employing a virtual Ambisonic approach. The 3D Tune-In Toolkit includes also a loudspeaker-based spatialiser implemented using Ambisonic encoding/decoding. This poster presents a brief overview of the main features of the Toolkit, which is released open-source under GPL v3 license (the code is available in GitHub https://github.com/3DTune-In/3dti-AudioToolkit).

Conference paper

Sethi S, Ewers R, Jones N, Orme D, Picinali Let al., 2018, Robust, real-time and autonomous monitoring of ecosystems with an open, low-cost, networked device, Methods in Ecology and Evolution, Vol: 9, Pages: 2383-2387, ISSN: 2041-210X

1. Automated methods of monitoring ecosystems provide a cost-effective way to track changes in natural system's dynamics across temporal and spatial scales. However, methods of recording and storing data captured from the field still require significant manual effort. 2. Here we introduce an open source, inexpensive, fully autonomous ecosystem monitoring unit for capturing and remotely transmitting continuous data streams from field sites over long time-periods. We provide a modular software framework for deploying various sensors, together with implementations to demonstrate proof of concept for continuous audio monitoring and time-lapse photography. 3. We show how our system can outperform comparable technologies for fractions of the cost, provided a local mobile network link is available. The system is robust to unreliable network signals and has been shown to function in extreme environmental conditions, such as in the tropical rainforests of Sabah, Borneo. 4. We provide full details on how to assemble the hardware, and the open-source software. Paired with appropriate automated analysis techniques, this system could provide spatially dense, near real-time, continuous insights into ecosystem and biodiversity dynamics at a low cost.

Journal article

Vijayasingam A, Shah A, Simmonds NJ, Elston C, Frost E, Wilkins J, Picinali L, Premachandra P, Gillen L, Morris-Rosendahl D, Ni Met al., 2018, INTERIM RESULTS FROM A PROSPECTIVE STUDY OF TABLET AND WEB-BASED AUDIOMETRY TO DETECT OTOTOXICITY IN ADULTS WITH CYSTIC FIBROSIS, THORAX, Vol: 73, Pages: A87-A88, ISSN: 0040-6376

Journal article

Frangakis N, Lim V, Tanco LM, Smatana P, Hreno J, Picinali L, Simeone L, Amditis Aet al., 2018, PLUGGY: A pluggable social platform for cultural heritage awareness and participation, CEUR Workshop, Pages: 21-30, ISSN: 1613-0073

Copyright held by the author(s). One of the preconditions for genuine sustainability is a heritage that is present anywhere and anytime in everyday life. We present PLUGGY, a Pluggable Social Platform for Heritage Awareness and Participation. PLUGGY will address the need of society to be actively involved in cultural heritage activities, not only as an observer but also as a creator and a major influencing factor. With PLUGGY, we aim to bridge this gap by providing the tools needed to allow users to share their local knowledge and everyday experience with others, together with the contribution of cultural institutions. Users will be able to build extensive networks around a common area of interest, connecting the past, the present and the future. It will be powered by its users and puts people's values, aspirations and needs first. Users of PLUGGY will be the providers of information about cultural heritage in the everyday and ordinary, real life. Through its social platform and by using its innovative curation tools, designed to solely focus on a niche area in social media, citizens will be able to act as skilled storytellers by creating fascinating personalised stories and share them through social networking with friends, associates and professionals. In this paper, we describe a structured formative and summative evaluative approach of PLUGGY’s core concepts and the results will be used to inform and improve its design.

Conference paper

Kim C, Steadman M, Lestang JH, Goodman DFM, Picinali Let al., 2018, A VR-based mobile platform for training to non-individualized binaural 3D audio, 144th Audio Engineering Society Convention 2018

© 2018 Audio Engineering Society. All Rights Reserved. Delivery of immersive 3D audio with arbitrarily-positioned sound sources over headphones often requires processing of individual source signals through a set of Head-Related Transfer Functions (HRTFs). The individual morphological differences and the impracticality of HRTF measurement make it difficult to deliver completely individualized 3D audio, and instead lead to the use of previously-measured non-individual sets of HRTFs. In this study, a VR-based mobile sound localization training prototype system is introduced which uses HRTF sets for audio. It consists of a mobile phone as a head-mounted device, a hand-held Bluetooth controller, and a network-enabled laptop with a USB audio interface and a pair of headphones. The virtual environment was developed on the mobile phone such that the user can listen-to/navigate-in an acoustically neutral scene and locate invisible target sound sources presented at random directions using non-individualized HRTFs in repetitive sessions. Various training paradigms can be designed with this system, with performance-related feedback provided according to the user’s localization accuracy, including visual indication of the target location, and some aspects of a typical first-person shooting game, such as enemies, scoring, and level advancement. An experiment was conducted using this system, in which 11 subjects went through multiple training sessions, using non-individualized HRTF sets. The localization performance evaluations showed reduction of overall localization angle error over repeated training sessions, reflecting lower front-back confusion rates.

Conference paper

Lim V, Frangakis N, Molina Tanco L, Picinali Let al., 2018, PLUGGY: A Pluggable Social Platform for Cultural Heritage Awareness and Participation, International Workshop on Analysis in Digital Cultural Heritage 2017

Conference paper

D'Cruz M, Patel H, Hallewell M, Salanitri D, Velzen J, Picinali Let al., 2017, Novel 3D games for people with and without hearing loss, 9th International Conference on Virtual Worlds and Games for Serious Applications, VS-Games 2017, Pages: 175-176

© 2017 IEEE. Over 90 million people in Europe currently suffer with hearing loss and with an aging community this is expected to rise significantly. Digital Hearing AIDS (HAs) offer real opportunities to enhance hearing capability in different acoustic contexts, however understanding functionalities and calibration can seem overly complex. The 3D Tune-In project has developed a 3D toolkit including a sound spatialisation algorithm and hearing/hearing loss simulators as the basis of five novel digital games addressing challenges of hearing loss and hearing education for children and older adults. Early evaluations have demonstrated the opportunities for hearing impaired groups, as well as the digital games community.

Conference paper

Isaac Engel J, Picinali L, 2017, Long-term user adaptation to an audio augmented reality system, SOUND AND VIBRATION. INTERNATIONAL CONGRESS. 24TH 2017

Audio Augmented Reality (AAR) consists in extending a real auditory environment with virtual sound sources. This can be achieved using binaural earphones/microphones. The microphones, placed in the outer part of each earphone, record sounds from the user's environment, which are then mixed with virtual binaural audio, and the resulting signal is finally played back through the earphones. However, previous studies show that, with a system of this type, audio coming from the microphones (or hear-through audio) does not sound natural to the user. The goal of this study is to explore the capabilities of long-term user adaptation to an AAR system built with off-the-shelf components (a pair of binaural microphones/earphones and a smartphone), aiming at achieve perceived realism for the hear-through audio. To compensate the acoustical effects of ear canal occlusion, the recorded signal is equalised in the smartphone. In-out latency was minimised to avoid distortion caused by comb filtering effect. To evaluate the adaptation process of the users to the headset, two case studies were performed. The subjects wore an AAR headset for several days while performing daily tests to check the progress of the adaptation. Both quantitative and qualitative evaluations (i.e., localising real and virtual sound sources and analysing the perception of pre-recorded auditory scenes) were carried out, finding slight signs of adaptation, especially in the subjective tests. A demo will be available for the conference visitors, including also the integration of visual Augmented Reality functionalities.

Conference paper

Picinali L, Wallin A, Levtov Y, Poirier-Quinot Det al., 2017, Comparative perceptual evaluation between different methods for implementing Reverberation in a binaural context, AES 2017, Publisher: Audio Engineering Society

Reverberation has always been considered of primary importance in order to improve the realism, externalisation and immersiveness of binaurally spatialised sounds. Different techniques exist for implementing reverberation in a binaural context, each with a different level of computational complexity and spatial accuracy. A perceptual study has been performed in order to compare between the realism and localization accuracy achieved using 5 different binaural reverberation techniques. These included multichannel Ambisonic-based, stereo and mono reverberation methods. A custom web-based application has been developed implementing the testing procedures, and allowing participants to take the test remotely. Initial results with 54 participants show that no major difference in terms of perceived level of realism and spatialisation accuracy could be found between four of the five proposed reverberation methods, suggesting that a high level of complexity in the reverberation process does not always correspond to improved perceptual attributes.

Conference paper

Mascetti S, Gerino A, Bernareggi C, Picinali Let al., 2017, On the evaluation of novel sonification techniques for non visual shape exploration, ACM Transactions on Accessible Computing, Vol: 9, Pages: 13.1-13.28, ISSN: 1936-7228

There are several situations in which a person with visual impairment or blindness needs to extract infor-mation from an image. For example, graphical representations are often used in education, in particular inSTEM subjects. In this contribution we propose a set of 6 soni cation techniques to support individualswith visual impairment or blindness in recognizing shapes on touchscreen devices. These techniques arecompared among themselves and with 2 other soni cation techniques already proposed in the literature.UsingInvisible Puzzle, a mobile application which allows to conduct non-supervised evaluation sessions, weconducted tests with 49 subjects with visual impairment and blindness, and 178 sighted subjects. All sub-jects involved in the process successfully completed the evaluation session, showing high level of engagement,demonstrating therefore the e ectiveness of the evaluation procedure. Results give interesting insights intothe di erences among the soni cation techniques and, most importantly, show that after a short trainingsubjects are able to successfully identify several di erent shapes.

Journal article

Patel H, Cobb S, Hallewell M, D'Cruz M, Eastgate R, Picinali L, Tamascelli Set al., 2016, User involvement in design and application of virtual reality gamification to facilitate the use of hearing AIDS, Pages: 77-81

The 3D Tune-In project aims to create an innovative toolkit based on 3D sound, visuals and gamification techniques to facilitate different target audiences in understanding and using the varied settings of their hearing aid to attain optimum performance in different social contexts. In the early stages of project development, hearing aid (HA) users participated in activities to identify user requirements regarding the difficulties and issues they face in everyday situations due to their hearing loss. The findings from questionnaire and interview studies and identification of current personas and scenarios of use indicate that the project can clearly and distinctly s the requirements of people with hearing loss as well as improve the general public's understanding of hearing loss. Five Future Scenarios of use have been derived to describe how the technologies and games to be developed by the 3D Tune-In project will address these requirements.

Conference paper

Mascetti S, Rossetti C, Gerino A, Bernareggi C, Picinali L, Rizzi Aet al., 2016, Towards a Natural User Interface to Support People with Visual Impairments in Detecting Colors, 15th International Conference, ICCHP 2016, Publisher: Springer, Pages: 171-178, ISSN: 0302-9743

A mobile application that detects an item’s color is potentially very useful for visually impaired people. However, users could run into difficulties when centering the target item in the mobile device camera field of view. To address this problem, in this contribution we propose a mobile application that detects the color of the item pointed by the user with one finger. In its current version, the application requires the user to wear a marker on the finger used for pointing. A preliminary evaluation conducted with blind users confirms the usefulness of the application, and encourages further development.

Conference paper

Eastgate R, Picinali L, Patel H, D'Cruz Met al., 2016, 3D Games for Tuning and Learning about Hearing Aids, Hearing Journal, Vol: 69, Pages: 30-32, ISSN: 0745-7472

Journal article

Levtov Y, Picinali L, D'Cruz M, Simeone Let al., 2016, Audio Engineering Society Convention, 140th Audio Engineering Society Convention

Conference paper

Mascetti S, Picinali L, Gerino A, Ahmetovic D, Bernareggi Cet al., 2016, Sonification of guidance data during road crossing for people with visual impairments or blindness, International Journal of Human Computer Studies, Vol: 85, Pages: 16-26, ISSN: 1071-5819

In the last years several solutions were proposed to support people with visual impairments or blindness during road crossing. These solutions focus on computer vision techniques for recognizing pedestrian crosswalks and computing their relative position from the user. Instead, this contribution addresses a different problem; the design of an auditory interface that can effectively guide the user during road crossing. Two original auditory guiding modes based on data sonification are presented and compared with a guiding mode based on speech messages.Experimental evaluation shows that there is no guiding mode that is best suited for all test subjects. The average time to align and cross is not significantly different among the three guiding modes, and test subjects distribute their preferences for the best guiding mode almost uniformly among the three solutions. From the experiments it also emerges that higher effort is necessary for decoding the sonified instructions if compared to the speech instructions, and that test subjects require frequent 'hints' (in the form of speech messages). Despite this, more than 2/3 of test subjects prefer one of the two guiding modes based on sonification. There are two main reasons for this: firstly, with speech messages it is harder to hear the sound of the environment, and secondly sonified messages convey information about the "quantity" of the expected movement.

Journal article

Iliya S, Menzies D, Neri F, Cornelius P, Picinali Let al., 2015, Robust impaired speech segmentation using neural network mixture model, ISSPIT 2014, Pages: 000444-000449

This paper presents a signal processing technique for segmenting short speech utterances into unvoiced and voiced sections and identifying points where the spectrum becomes steady. The segmentation process is part of a system for deriving musculoskeletal articulation data from disordered utterances, in order to provide training feedback for people with speech articulation problem. The approach implement a novel and innovative segmentation scheme using artificial neural network mixture model (ANNMM) for identification and capturing of the various sections of the disordered (impaired) speech signals. This paper also identify some salient features that distinguish normal speech from impaired speech of the same utterances. This research aim at developing artificial speech therapist capable of providing reliable text and audiovisual feed back progress report to the patient.

Conference paper

O'Sullivan L, Picinali L, Gerino A, Cawthorne Det al., 2015, A Prototype Audio-Tactile Map System with an Advanced Auditory Display, INTERNATIONAL JOURNAL OF MOBILE HUMAN COMPUTER INTERACTION, Vol: 7, Pages: 53-75, ISSN: 1942-390X

Journal article

Picinali L, D'Cruz M, Simeone L, 2015, 3D-Tune-In: 3D sound, visuals and gamification to facilitate the use of hearing aids, EuroVR Conference 2015

Conference paper

Ulanicki B, Picinali L, Janus T, 2015, Measurements and analysis of cavitation in a pressure reducing valve during operation a case study, Computing and Control for the Water Industry (CCWI2015), Publisher: Elsevier, Pages: 270-279, ISSN: 1877-7058

This paper proposes a methodology and presents its practical application for evaluating whether a pressure reducing valve (PRV) is under cavitation during its operation in a water distribution system. The approach is based on collecting measurements over a 24-hour period such that high demand and low demand times are included. The collected measurements allow evaluation of four indicators related to cavitation, namely the hydraulic cavitation index, noise generated by the valve, acoustic cavitation index and the spectra of the noise. These four indicators provide sufficient information for diagnosis of cavitation with high certainty.

Conference paper

Picinali L, Gerino A, Bernareggi C, Alabastro N, Mascetti Set al., 2015, Towards Large Scale Evaluation of Novel Sonification Techniques for Non Visual Shape Exploration, ACM SIGACCESS Conference on Computers & Accessibility, Publisher: ACM, Pages: 13-21

Conference paper

Picinali L, Rona A, Cockrill M, Panday P, Tripathi R, Ayub Met al., 2015, Vibro-acoustic response of tympanic-membrane-like models, 22nd International Converence on Sound and Vibration

Conference paper

Iliya S, Neri F, Menzies D, Cornelius P, Picinali Let al., 2015, Differential evolution schemes for speech segmentation: A comparative study, 2014 IEEE Symposium on Differential Evolution (SDE), Publisher: IEEE, Pages: 1-8

This paper presents a signal processing technique for segmenting short speech utterances into unvoiced and voiced sections and identifying points where the spectrum becomes steady. The segmentation process is part of a system for deriving musculoskeletal articulation data from disordered utterances, in order to provide training feedback. The functioning of the signal processing technique has been optimized by selecting the parameters of the model. The optimization has been carried out by testing and comparing multiple Differential Evolution implementations, including a standard one, a memetic one, and a controlled randomized one. Numerical results have also been compared with a famous and efficient swarm intelligence algorithm. For the given problem, Differential Evolution schemes appear to display a very good performance as they can quickly reach a high quality solution. The binomial crossover appears, for the given problem, beneficial with respect to the exponential one. The controlled randomization appears to be the best choice in this case. The overall optimized system proved to segment well the speech utterances and efficiently detect its uninteresting parts

Conference paper

Battey B, Giannoukakis M, Picinali L, 2015, Haptic control of multistate generative music systems, Pages: 98-101

Force-feedback controllers have been considered as a solution to the lack of sonically coupled physical feedback in digital-music interfaces, with researchers focusing on instrument-like models of interaction. However, there has been little research applied to the use of force-feedback interfaces to the control of real-time generative-music systems. This paper proposes that haptic interfaces could enable performers to have a more fully embodied engagement with such systems, increasing expressive control and enabling new compositional and performance potentials. A proof-of-concept project is described, which entailed development of a core software toolkit and implementation of a series of test cases.

Conference paper

Gerino A, Picinali L, Bernareggi C, Mascetti Set al., 2015, Eyes-free Exploration of Shapes with Invisible Puzzle, 17th International ACM SIGACCESS Conference on Computers and Accessibility (ASSETS 2015), Publisher: ASSOC COMPUTING MACHINERY, Pages: 425-426

Conference paper

Picinali L, Afonso A, Denis M, Katz BFGet al., 2014, Corrigendum to “Exploration of architectural spaces by blind people using auditory virtual reality for the construction of spatial knowledge” [Int. J. Hum. Comput. Stud. 72 (4) (2014) 393–407], International Journal of Human-Computer Studies, Vol: 72, Pages: 875-875, ISSN: 1071-5819

The authors regret that the legend below the two diagrams in Fig. 7 (p. 405) contains an error. The blue data should correspond to the real navigation condition, and not to the virtual one. Vice versa, the red data should correspond to the virtual navigation condition, and not to the real one.The authors would like to apologise for any inconvenience caused.

Journal article

O'Sullivan L, Picinali L, Feakes C, Cawthorne Det al., 2014, Audio tactile maps (ATM) system for the exploration of digital heritage buildings by visually-impaired individuals - First prototype and preliminary evaluation, Forum Acusticum 2014, ISSN: 2221-3767

Navigation within historic spaces requires analysis of a variety of acoustic, proprioceptive and tactile cues; a task that is well-developed in many visually-impaired individuals but for which sighted individuals rely almost entirely on vision. For the visually-impaired, the creation of a cognitive map of a space can be a long process for which the individual may repeat various paths numerous times. While this action is typically performed by the individual on-site, it is of some interest to investigate to what degree this task can be performed off-site using a virtual simulator. We propose a tactile map navigation system with interactive auditory display. The system is based on a paper tactile map upon which the user's hands are tracked. Audio feedback provides; (i) information on user-selected map features, (ii) dynamic navigation information as the hand is moved, (iii) guidance on how to reach the location of one hand (arrival point) from the location of the other hand (departure point) and (iv) additional interactive 3D-audio cues useful for navigation. This paper presents an overview of the initial technical development stage, reporting observations from preliminary evaluations with a blind individual. The system will be beneficial to visuallyimpaired visitors to heritage sites; we describe one such site which is being used to further assess our prototype.

Conference paper

Menelas B-AJ, Picinali L, Bourdot P, Katz BFGet al., 2014, Non-visual identification, localization, and selection of entities of interest in a 3D environment, JOURNAL ON MULTIMODAL USER INTERFACES, Vol: 8, Pages: 243-256, ISSN: 1783-7677

Journal article

This data is extracted from the Web of Science and reproduced under a licence from Thomson Reuters. You may not copy or re-distribute this data in whole or in part without the written consent of the Science business of Thomson Reuters.

Request URL: http://wlsprd.imperial.ac.uk:80/respub/WEB-INF/jsp/search-html.jsp Request URI: /respub/WEB-INF/jsp/search-html.jsp Query String: id=00851530&limit=30&person=true&page=3&respub-action=search.html