DAFx-12 / Program
15th
International Conference |
|||||||||||
|
We are very happy to introduce our three keynote speakers for DAFx-12, all leaders in their respective fields:
Tuesday
evening's social event is a performance of Trevor's latest work, Encounters
in the Republic of Heaven. This is an 8-channel
surround-sound piece based on speaking voices recorded in the North
East of England where Trevor Wishart was Arts Council England
Composer-in-Residence from October 1st, 2006 to September 2009, based
at the University of Durham. The piece is constructed in 4 Acts of
approximately 20 minutes each, combining portraits of individual
speakers (accompanied by sounds and imaginary instruments derived from
the voices themselves) - speech that waltzes, speech that locks in
harmony, clouds of speech that circle the audience, culminating with
speech that transforms into song.
The piece was completed on January 1st 2011 and has been performed around the world, but this will be its first performance in York.
Echo
and reverb have been widely used as audio effects for decades. This
paper explores unusual places, both ancient and modern, with remarkable
natural sound effects including staircases that chirp, culverts that
whistle and the most reverberant place in the world. The physics of the
phenomena will be revealed, sometimes drawing on historical
explanations, at other times using predictions from modern numerical
models to visualise and better understand what is going on. The
perception of listeners will also be explored, including what these
spaces reveal about our hearing and our emotional response to
architectural spaces.
Trevor
Cox is Professor of Acoustic Engineering at the University of Salford,
a former Senior Media Fellow funded by the Engineering
and Physical Sciences Research Council, and immediate past
president of the Institute
of Acoustics (IOA). One major strand of his research is room
acoustics for intelligible speech and quality music production and
reproduction. Trevor's diffuser designs can be found in rooms around
the world. He has co-authored a research book entitled "acoustic
absorbers and diffusers". He was award the IOA's Tyndall Medal in 2004.
Trevor has a long track record of communicating acoustic engineering to
the public and has been involved in engagement projects worth over £1M.
He was given the IOA award for promoting acoustics to the public in
2009. He has developed and presented science shows to 15,000 pupils
including performing at the Royal Albert Hall, Purcell Rooms and the
Royal Institution.
Trevor has presented sixteen documentaries for BBC radio including: Life's soundtrack, Save our Sounds and Science vs the Strad. And is currently writing The Sound Book for Bodley Head.
The developments of broadcast
technology are very related to the field of audio effects. In the early
days a broadcast chain was more or less linear, only including
different types of amplifiers. From there it has developed to a signal
path with more and more signal processing. With the transformation into
the digital domain digital audio effects have become important for the
broadcaster on various levels.
Radio programmes like radio dramas or the BBC radiophonic workshop rely
on audio signal processing and have transformed the use of audio
effects into an art form. On the other side, processing driven by the
needs of FM transmission like multiband dynamic processing are
important for broadcasters to define the station sound and optimize
coverage.
With the transformation of broadcasting systems in the future it is
expected that also requirements and types of digital audio effects will
change and develop further.
This talk will present some highlights on the history of audio
processing in broadcasting and will try to highlight some potential
options and needs for future developments in digital audio effects.
Important aspects that will be discussed are the developments towards
spatial audio effects, perceptual audio effects and personalized audio
effects.
Frank Melchior is head of the
audio research team at BBC
R&D. He received a Dipl.-Ing. (equiv. to Master of
Science) in Media Technology from the Technical University Ilmenau and
a doctoral degree on the topic of spatial sound design and acoustics
from University of Technology, Delft, The Netherlands. Frank worked as
a researcher and project manager at the Fraunhofer
Institute for Digital Media Technology. He has spent his
professional career developing audio systems including wave field
synthesis based systems and leading R&D teams on various topics
including authoring, live sound applications, automotive audio and
motion picture sound. Before he joined the BBC he was the Chief
Technical Officer at IOSONO,
Erfurt, Germany, who develop spatial audio solutions. Frank's research
interests include spatial array signal processing, spatial audio
reproduction, user interface technology for audio systems and listening
experiments.
Dr Frank Melchior is very
passionate about finding the best way to bring the emotion and creative
ideas embedded in sound to the listener.
The main DAFx program is supported by invited workshop/tutorial sessions in key topics of interest to the DAFx audience on the Monday directly before, and Friday after the three days of oral presentations and poster sessions. These workshops are open to all DAFx12 delegates:
Current graphics processing units (GPUs) are massively parallel computation engines, and they have gained lots of popularity in high-performance computing in the recent years. This workshop aims to give a general understanding on the capabilities of GPUs, how the parallel computation is performed on a GPU, and what kind of problems are suitable for parallelization. Two commonly applied ways to utilize GPUs are introduced. While CUDA is the most typically applied programmer-level API, there is a good interface to use a GPU directly from Matlab, as well. The workshop will have several examples on utilization of GPUs on practical real-time audio signal processing tasks.
Lauri Savioja received the
degree of Doctor of Science in Technology, from the Helsinki University
of Technology (TKK), Espoo, Finland, in 1999. He majored in computer
science, and the topic of his doctoral thesis was room acoustic
modeling. He worked at the TKK Laboratory of Telecommunications
Software and Multimedia as a researcher, lecturer, and professor from
1995 until the formation of Aalto University
where he currently works as a professor in the Department of Media
Technology at the School of Science. The academic year 2009-2010, he
spent as a visiting researcher at NVIDIA
Research. His research interests include room
acoustics, virtual reality, and parallel computation.
Prof. Savioja is a member of the Association for Computing Machinery
(ACM), Audio Engineering Society (AES), and a life member of the
Acoustical Society of Finland. From 2010 he has been an Associate
Editor of the IEEE Transactions on Audio, Speech, and Language
Processing.
Sustainable and reusable software and data are becoming increasingly important in today's research environment. Methods for processing audio have become so complex they cannot fully be described in the printed paper. Researchers are becoming increasingly aware of the benefits of open access publication and reusable software and data in helping others to build on their work. However, many researchers find this difficult to achieve in practice. In this workshop we will explore ways to maximise the accessibility and impact of research through publication of research software, availability of research data, and open access paper publication. Topics will include: software version control; unit testing; writing maintainable code; code reviews; choice of language and platform; software licensing; data repositories and data management; levels of open access publication, negotiating publication agreements, and associating papers with software and data publications. The workshop will be of interest to researchers who wish to make immediate improvements to the effectiveness, sustainability and impact of their work, as well as to research group leaders with an interest in policy and management.
Mark Plumbley is Director of the
Centre
for Digital Music (C4DM) at Queen Mary University of London,
and leads the EPSRC-funded
soundsoftware.ac.uk
initiative. His work in audio signal analysis includes beat tracking,
music transcription, source separation and object coding, using
techniques such as neural networks, independent component analysis,
sparse representations and Bayesian modelling. Prof. Plumbley is a
member of the IEEE Machine Learning in Signal Processing Technical
Committee, and leads the UK Digital Music Research Network.
Mark will be leading this workshop accompanied by his colleagues:
This tutorial is intended as an introduction to time stepping methods for physical audio synthesis and effects processing, with some emphasis on how they fit into the larger picture of more standard audio signal processing/synthesis techniques. Various examples will be discussed and worked through in detail, spanning the range of acoustic musical instruments, electromechanical instruments and reverberation devices, and full 3D room acoustics simulation. Audio and video examples will be presented, along with Matlab code walk-throughs.
Stefan Bilbao is currently a
Senior Lecturer in Acoustics at the University of Edinburgh. He was
formerly a lecturer at the Sonic Arts Research Centre, at the Queen's
University Belfast, and a postdoctoral research associate at the
Stanford Space Telecommunications and Radioscience Laboratory. Prior to
that, he received the MSc and PhD degrees in Electrical Engineering at
Stanford University while working at the Centre for Computer Research
in Music and Acoustics, and a BA in Physics at Harvard University, and
in between spent two years in the real-time group at the Institut de Recherche et
Coordination Acoustique/Musique in Paris.
His main interests are in sound synthesis, numerical method design, and
parallel processing for audio on GPUs.
Singing is a way humans have expressed themselves from earliest times, and the range of sounds that they can achieve from what is basically a variable-pitch buzzer and a flexible short tube is immense. This workshop will explore the physiology of the human singing voice and relate its elements to the nature of the acoustic output and how it can be varied. Shape changes that are made to the flexible tube or 'vocal tract' will be explored based on magnetic resonance images (MRI) from which the airway itself can be reconstructed. This enables the shape changes to be investigated in terms of the physiology as well as the acoustics by means of digital waveguide synthesis. Reference will be made to choral singing; in particular tuning in unaccompanied or 'a cappella' singing.
David Howard is Head of the
Department of Electronics and Head of the Audio Laboratory at the
University of York. His research centres around the analysis and
synthesis of singing, speech and music and the use of real-time visual
displays in vocal pedagogy. David has been an Engineering
and Physical Sciences Research Council Senior Media Fellow
and he has been involved in a number of radio and TV programmes mainly
about the science of sound, particularly singing and music. In the
BBC4-TV programme 'Voice' he worked with Jeremy Hardy on his singing(!)
and Rory Bremner on his impression of Tony Blair. Recently he was
involved in Channel 4's 'Hidden Talent Show'.
David conducts the 'Vale of York Voice'; a local choir that sings
evensong in York Minster about once a month.
Ambisonics, with its channel based, speaker agnostic approach to surround sound encoding and decoding, has always been problematic to use in speaker layout obsessed Digital Audio Workstations. This session will explain what Ambisonics is and demonstrate how to setup and use the WigWare Ambisonic Plug-in Suite with the hierarchical routing structure presented by Reaper paying particular attention to the differences between first and higher order Ambisonics, and the lesser used near field compensation and distance filters. Example mixes created by 2nd year undergraduate students at the University of Derby will also be presented, along with newly developed teaching aids and visualisations that are soon to be released as open educational material for the dissemination of Ambisonics to an audience new to the subject.
Dr Bruce Wiggins graduated with
a 1st class honours in Music Technology and Audio System Design from
the University of Derby in 1999. His interest in audio signal
processing spurred him to continue at Derby completing his PhD entitled
"An Investigation into the Real-time Manipulation and Control of 3D
Sound Fields" in 2004 where he solved the problem of generating
Ambisonic decoders for irregular speaker arrays and also carried out
work on binaural/transaural reproduction systems.
He is now a lecturer in the Electronics and Sound subject group at the
University where he teaches electronics, audio programming and digital
signal processing which are all fed from his continuing research
interests in Ambisonic surround sound systems earning him excellence
awards for his promising research in 2005/6, his application of
technology in 2006/7 and excellence in learning, teaching and
assessment in 2007/8.
The SoundSoftware
and Software
Carpentry projects are organising a three-day bootcamp on
software development, during the week preceding the main DAFx-12
conference. The bootcamp will run from the 13th to 15th of September
2012 at the University of York.
The goal of the bootcamp is to help scientists and engineers become
more productive by teaching them basic software development skills like
program design, version control, testing, and task automation. In this
three-day boot camp, short tutorials will alternate with hands-on
practical exercises. Participants will be encouraged both to help one
another, and to apply what they have learned to their own research
problems during and between sessions. There will be online follow up
sessions for 6 to 8 weeks extending the material from the boot camp.
More information and separate registration for this event is available at the York Bootcamp Website. Note that this workshop is not included in the main DAFx-12 registration, but delegates may wish to take advantage of this event alongside their attendance at the DAFx conference.
The final program below is also available as a PDF here.
Monday 17th September
|
||
09:00 - 19:00 |
Registration |
[TFTV - Lobby Area] |
Tutorial 1: |
[TFTV - Digital Cinema] | |
13:00
- 14:30 |
GPUs for DAFx [Mon 17 - #1] Lauri Savioja |
|
Tutorial 2: |
[TFTV - Digital Cinema] | |
14:30
- 16:00 |
Pointers to More Effective
Software and Data in Audio Research [Mon 17 - #2] Mark Plumbley, Chris Cannam, Luis Figueira and Steve Welburn |
|
16:00
- 16:15 |
Coffee |
[TFTV - Lobby Area] |
Tutorial 3: |
[TFTV - Digital Cinema] | |
16:15
- 17:45 |
FDTD Methods for Physical
Modeling Sound Synthesis and Audio Effects Processing [Mon 17 - #3] Stefan Bilbao |
|
17:45
- 19:00 |
Tours/Demos |
[TFTV] |
19:00
- 21:00 |
Welcome Reception |
[Ron Cooke Hub] |
Tuesday 18th September |
||
08:30
- 18:00 |
Registration
|
[TFTV - Lobby Area] |
08:45
- 09:00 |
Welcome
to DAFx12 |
[TFTV - Digital Cinema] |
Keynote 1: |
[TFTV - Digital Cinema] | |
09:00
- 10:00 |
The Music of Speech Trevor Wishart |
|
Oral Session 1: |
Spectral Processing and Analysis |
[TFTV - Digital Cinema] |
10:00
- 10:20 |
Automatic Partial Extraction
From the Modal Distribution [#68] Thomas Lysaght, Joseph Timoney and Victor Lazzarini |
|
10:20
- 10:40 |
VST Plug-in Module Performing
Wavelet Transform in Real-time [#61] Pavel Rajmic, Zdenek Prusa and Robert Konczi |
|
10:40
- 11:00 |
Poster Presentations 1: |
|
Practical Empirical Mode
Decomposition For Audio Synthesis [#12] Niklas Klügel |
||
The Simplest Analysis Method
for Non-Stationary Sinusoidal Modeling [#35] Sylvain Marchand |
||
Metamorph: Real-Time
High-Level Sound Transformations Based on a Sinusoids Plus Noise Plus
Transients Model [#53] John Glover, Victor Lazzarini and Joseph Timoney |
||
Range-constrained Phase
Reconstruction for Recovering Time-domain Signal from Quantized
Amplitude & Phase Spectrogram [#66] Sho Sato, Nobutaka Ono, Yutaka Kamamoto and Shigeki Sagayama |
||
11:00
- 11:30 |
Coffee/Poster Session |
[TFTV - Lobby Area] |
Oral Session 2: |
Processing via Non-negative
Matrix Factorisation |
[TFTV - Digital Cinema] |
11:30
- 11:50 |
Effective Separation of
Low-Pitch Notes Using NMF Using Non-Power-of-2 Discrete Fourier
Transforms [#44] Ta-Chun Chen, Tien-Ming Wang, Ya-Han Kuo and Alvin Su |
|
11:50
- 12:10 |
Shifted NMF with Group
Sparsity for Clustering NMF Basis Functions [#38] Rajesh Jaiswal, Derry Fitzgerald, Eugene Coyle and Scott Rickard |
|
12:10
- 12:30 |
Sparse Decomposition,
Clustering and Noise for Fire Texture Sound Re-Synthesis [#65] Stefan Kersten and Hendrik Purwins |
|
12:30
- 12:45 |
Poster Presentations 2: |
|
A jump start for NMF with
N-FINDR and NNLS [#78] Joachim Ganseman and Paul Scheunders |
||
Drumkit Transcription via
Convolutive NMF [#39] Henry Lindsay-Smith, Skot McDonald and Mark Sandler |
||
12:45
- 14:00 |
Lunch/Poster Session |
[TFTV - Lobby Area] |
Oral Session 3: |
Physical Modelling |
[TFTV - Digital Cinema] |
14:00
- 14:20 |
Real-time Finite Difference
Physical Models of Musical Instruments on a Field Programmable Gate
Array (FPGA) [#47] Florian Pfeifle and Rolf Bader |
|
14:20
- 14:40 |
Soliton Sonification -
Experiments with the Kortweg-deVries Equation [#76] Rudolf Rabenstein |
|
14:40
- 15:00 |
A pickup model for the
Clavinet [#62] Luca Remaggi, Leonardo Gabrielli, Rafael Cauduro Dias de Paiva, Vesa Välimäki and Stefano Squartini |
|
15:00
- 15:20 |
A Physically-Constrained
Source Model for FDTD Acoustic Simulation [#48] Jonathan Sheaffer, Maarten Van Walstijn and Bruno Fazenda |
|
15:20
- 15:45 |
Poster Presentations 3: |
|
Analysis of Sound Field
Distribution for Room Acoustics: From the Point of View of Hardware
Implementation [#31] Yiyu Tan, Yasushi Inoguchi, Yukinori Sato, Makoto Otani, Yukio Iwaya, Hiroshi Matsuoka and Takao Tsuchiya |
||
Binaural simulations using
audio rate FDTD schemes and CUDA [#40] Craig Webb and Stefan Bilbao |
||
A Study on Dynamic Vocal
Tract Shaping for Dipthong Simulation Using a 2D Digital Wavegude Mesh
[#49] Anocha Rugchatjaroen and David Howard |
||
Achieving Convolution-based
Reverberation Through use of Geometric Acoustic Modelling Techniques
[#64] Stephen Oxnard and Damian Murphy |
||
15:45
- 16:15 |
Coffee/Poster Session |
[TFTV - Lobby Area] |
Oral Session 4: |
Time and Pitch Scaling |
[TFTV - Digital Cinema] |
16:15
- 16:35 |
Pitch Shifting of Audio
Signals Using the Constant-Q Transform [#81] Christian Schörkhuber, Anssi Klapuri and Alois Sontacchi |
|
16:35
- 16:55 |
User-Guided Variable-Rate
Time-Stretching Via Stiffness Control [#67] Nicholas J. Bryan, Jorge Herrera and Ge Wang |
|
16:55
- 17:15 |
Improved PVSOLA Time
Stretching and Pitch Shifting for Polyphonic Audio [#26] Sebastian Kraft, Martin Holters, Adrian Von Dem Knesebeck and Udo Zölzer |
|
17:15
- 17:35 |
On Stretching Gaussian Noises
with the Phase Vocoder [#41] Wei-Hsiang Liao, Axel Roebel and Alvin W.Y. Su |
|
18:30
- 19:30 |
Buffet | [TFTV - Lobby Area] |
19:30
- 22:00 |
Evening Concert |
[TFTV - Black Box Theatre] |
Encounters in the Republic of
Heaven: All the Colours of Speech Trevor Wishart |
||
Wednesday 19th September
|
||
08:30
- 17:00 |
Registration |
[TFTV
- Lobby Area] |
Keynote 2: |
[TFTV
- Digital Cinema] |
|
09:00
- 10:00 |
Back to Nature: Fantastical
Echoes and Reverberations Trevor Cox |
|
Oral Session 5: |
Auralisation |
[TFTV - Digital Cinema] |
10:00
- 10:20 |
Real-time Auralisation System
for Virtual Microphone Positioning [#75] Tom Barker, Guilherme Campos, Paulo Dias, José Vieira, Catarina Mendonca and Jorge Santos |
|
10:20
- 10:40 |
Spatial High Frequency
Extrapolation Method for Room Acoustic Auralization [#42] Alex Southern and Lauri Savioja |
|
10:40
- 11:00 |
Poster Presentations 4:
|
|
3D Binaural Audio Capture and
Reproduction Using A Miniature Microphone Array [#1] Shengkui Zhao, Ryan Rogowski, Reece Johnson and Douglas Jones |
||
Variable Source Radiation
Pattern Synthesis for use in Two-Dimensional Sound Reproduction [#16] Martin Morrell and Joshua Reiss |
||
Energy-based calibration of
virtual performance systems [#70] Iain Laird, Paul Chapman and Damian Murphy |
||
11:00
- 11:30 |
Coffee/Poster Session |
[TFTV - Lobby Area] |
Oral Session 6: |
Spatial Processing and
Presentation |
[TFTV
- Digital Cinema] |
11:30
- 11:50 |
Parametric Spatial Audio
Effects [#22] Archontis Politis, Tapani Pihlajamäki and Ville Pulkki |
|
11:50
- 12:10 |
Binaural In-Ear Monitoring of
Acoustic Instruments in Live Music Performance [#73] Elías Zea |
|
12:10
- 12:30 |
Poster Presentations 5: |
|
Acoustic Measurement Methods
for Outdoor Sites: A Comparative Study [#51] Mariana Lopez and Sandra Pauletto |
||
Simulating Microphone Bleed
and Tom-tom Resonance in Multisampled Drum Workstations [#5] Alice Clifford, Henry Lindsay Smith and Josh Reiss |
||
12:30
- 13:45 |
Lunch/Poster Session |
[TFTV
- Lobby Area] |
Oral Session 7: |
Virtual Analogue |
[TFTV - Digital Cinema] |
13:45
- 14:05 |
A Wave Digital Filter Model
of the Fairchild 670 Limiter [#9] Peter Raffensperger |
|
14:05
- 14:25 |
Harmonic Instability of
Digital Soft Clipping Algorithms [#45] Sean Enderby and Zlatko Baracskai |
|
14:25
- 14:45 |
Simulation of Fender Type
Guitar Preamp using Approximation and State-Space Model [#25] Jaromir Macak, Martin Holters and Jiri Schimmel |
|
14:45
- 15:05 |
Virtual Analog Oscillator
Hard Synchronisation: Fourier series and an efficient implementation
[#37] Joseph Timoney, Victor Lazzarini, Jari Kleimola, Jussi Pekonen and Vesa Valimaki |
|
15:05
- 15:30 |
Poster Presentations 6: |
|
Visualization of Signals and
Algorithms in Kronos [#11] Vesa Norilo |
||
Scattering Representation of
Modulated Sounds [#50] Joakim Andén and Stéphane Mallat |
||
Audio ADC dynamic range
matching by means of a DSP equalizer and dynamics processor combination
[#8] Alexander Potchinkov |
||
Uniform Noise Sequences for
Nonlinear System Identification [#79] Francois Germain, Jonathan Abel, Philippe Depalle and Marcelo Wanderley |
||
15:30
- 16:00 |
Coffee/Poster Session |
[TFTV - Lobby Area] |
Oral Session 8: |
Sound Synthesis |
[TFTV - Digital Cinema] |
16:00
- 16:20 |
Higher-Order Integrated
Wavetable Synthesis [#69] Andreas Franck and Vesa Välimäki |
|
16:20
- 16:40 |
Synthesis of Resonant Sounds
with a Heterodyne Model [#58] Victor Lazzarini and Joseph Timoney |
|
16:40
- 17:00 |
Nonlinear-Phase Basis
Functions in Quasi-Bandlimited Oscillator Algorithms [#15] Jussi Pekonen and Martin Holters |
|
17:00
- 17:20 |
Timpani Drum Synthesis in 3D
on GPGPUs [#36] Stefan Bilbao and Craig Webb |
|
17:30
- 18:30 |
DAFx
Scientific Committee Meeting |
[TFTV - Digital Cinema] |
19:30
- 20:00 |
Concert
at Guildhall and Mansion House - Arrive |
[York
Centre] |
20:00
- 22:00 |
Main
Programme Starts |
|
Thursday 20th September
|
||
08:30 - 17:00 | Registration |
[TFTV - Lobby Area] |
Keynote 3: | [TFTV - Digital Cinema] | |
09:00
- 10:00 |
Applications of Digital Audio
Effects in Broadcast: Yesterday, Today and Tomorrow Frank Melchior |
|
Oral
Session 9: |
Effects/Synthesis for Touchscreen
and Mobile Devices |
[TFTV - Digital Cinema] |
10:00 - 10:20 |
The Wablet: Scanned Synthesis
on a Multi-Touch Interface [#18] Robert Tubb, Anssi Klapuri and Simon Dixon |
|
10:20 - 10:40 |
Digital Audio Effects on Mobile Platforms [#2] Victor Lazzarini, Steven Yi and Joseph Timoney |
|
10:40
- 11:00 |
Poster
Presentations 7: |
|
Time-Domain Chroma Extraction [#24] Marco Fink, Martin Holters and Udo Zölzer |
||
Online Real-time Onset Detection with Recurrent
Neural Networks [#4] Sebastian Böck, Andreas Arzt, Florian Krebs and Markus Schedl |
||
Deploying Nonlinear Image Filters to Spectrogram
for Harmonic/Percussive Separation [#63] Aggelos Gkiokas, Vassilis Papavassiliou, Vassilis Katsouros and George Carayannis |
||
Multi-channel Audio Information Hiding [#54] Jonathan Blackledge, Ruairi Hickson and Abdul Al-Rawi |
||
11:00
- 11:30 |
Coffee/Poster
Session |
[TFTV - Lobby Area] |
Oral
Session 10: |
Computational
Auditory Scene Analysis and Processing |
[TFTV - Digital Cinema] |
11:30 - 11:50 |
Characterisation of Acoustic Scenes Using a
Temporally-constrained Shift-invariant Model [#30] Emmanouil Benetos, Mathieu Lagrange and Simon Dixon |
|
11:50 - 12:10 |
Audio-visual Multiple Active Speaker
Localization in Reverberant Environments [#29] Zhao Li, Thorsten Herfet, Martin Grochulla and Thorsten Thormählen |
|
12:10 - 12:30 |
An Autonomous Method for Multi-Track Dynamic
Range Compression [#6] Jacob Maddams, Saoirse Finn and Joshua Reiss |
|
12:30
- 12:45 |
Poster
Presentations 8: |
|
Voice Features For Control: A Vocalist Dependent
Method For Noise Measurement And Independent Signals Computation [#10] Stefano Fasciani |
||
Musical Aspects of Vowel Formants in the Extreme
Metal Voice [#55] Eric Smialek, Philippe Depalle and David Brackett |
||
12:45
- 14:00 |
Lunch/Poster
Session |
[TFTV - Lobby Area] |
Oral
Session 11: |
Feature
Extraction and Source Separation |
[TFTV - Digital Cinema] |
14:00 - 14:20 |
Phase-based informed source separation for
active listening of music [#57] Nicolas Sturmel, Laurent Daudet and Laurent Girin |
|
14:20 - 14:40 |
On the use of Masking Filters in Sound Source
Separation [#23] Derry Fitzgerald and Rajesh Jaiswal |
|
14:40 - 15:00 |
Unsupervised Feature Learning for Speech and
Music Detection in Radio Broadcasts [#32] Jan Schlüter and Reinhard Sonnleitner |
|
15:00 - 15:20 |
A Simple and Effective Spectral Feature for
Speech Detection in Mixed Audio Signals [#19] Reinhard Sonnleitner, Bernhard Niedermayer, Gerhard Widmer and Jan Schlüter |
|
15:20
- 15:45 |
Poster
Presentations 9: |
|
Towards Efficient Music Genre Classification
Using FastMap [#20] Franz de Leon and Kirk Martinez |
||
Music Emotion Classification: Dataset
Acquisition And Comparative Analysis [#77] Renato Panda and Rui Pedro Paiva |
||
15:45
- 16:15 |
Coffee/Poster
Session |
[TFTV - Lobby Area] |
Oral
Session 12: |
Resources
and Algorithms for Audio Signal Processing |
[TFTV - Digital Cinema] |
16:15 - 16:35 |
The development of an online course in DSP
eartraining [#13] Øyvind Brandtsegg, Sigurd Saue, Victor Lazzarini, John Pål Inderberg, Axel Tidemann, Håkon Kvidal, Jan Tro, Jøran Rudi and Notto J. W. Thelle |
|
16:35 - 16:55 |
Impact Of Personalized Equalization Curves On
Music Quality In Dichotic Listening [#3] Duo Zhang, Hongmei Xia, Tiffany Chua, Gerald Maguire, David Franklin, Daniel Huang, Hung Tran and Hongbin Chen |
|
16:55 - 17:15 |
The Helmholtz Resonator Tree [#33] Rafael Cauduro Dias de Paiva and Vesa Välimäki |
|
17:15
- 17:30 |
Handover to
DAFx13 |
|
18:00 |
Buses
to National Railway Museum |
[Outside
TFTV] |
18:30
- 22:00 |
Conference
Banquet |
[National
Railway Museum] |
Friday 21st
September |
||
Tutorial
4: |
[TFTV - Digital Cinema] | |
10:00 - 11:30 |
The Acoustics of the Singing Voice [Fri 21 - #1] David Howard |
|
Tutorial 5: |
[TFTV - Black Box Theatre] | |
11:30 - 13:00 |
Practical Ambisonics [Fri 21 - #2] Bruce Wiggins |
|
13:00 |
DAFx Closes -
See you in Maynooth in 2013! |