Hi Richard,

I’ve changed the title of this topic to something more relevant. 

I still prefer the term Delta Stereophony to describe this. It seems to date 
back to the mid 1980’s, and was described by Gerhard Steinke and Wolfgang 
Ahnert. They were working in East Germany behind the Iron Curtain, reputedly 
working with Sinclair ZX Spectrum computers and expensive AKG delay lines 
somehow imported from Austria.

It does make a great deal of sense. When digital delay lines became more 
generally available and affordable (1990’s ???) they were increasingly used in 
public address systems to improve coverage over a greater area, using speakers 
down the length of an auditorium to augment the usual left/right or LCR main 
frontal system. The feed to these was delayed by an amount that caused the time 
of arrival of sound from them to match that of the main frontal system . 
Sometimes the feed to a "front fill” system, arrayed along the front of the 
stage to increase clarity in the rows of seating near the stage, was also 
delayed to match the time of arrival of sound from its source. Amplitudes were 
usually adjusted by ear, as indeed were delay times after an initial 
calculation.

These systems were more “appropriately distributed mono” than spatial. It is 
impossible to get the delay/amplitude combination correct for every position in 
the space with a finite number of speakers and output channels, so compromises 
are inevitable. This became common practice, especially for large scale stadium 
events. Digital mixing desks now commonly incorporate delays on each output, 
making this simpler to implement.

Current products do not allow progress to true Delta Stereophony (DBADP), as 
the architecture does not provide delay as well as amplitude control on each 
matrix crosspoint, and the market doesn’t expect or demand it. DSP chips are 
now capable of providing it, as proved by TiMax, LISA, d&b’s Soundscape, Iosono 
, Astro, and Meyer’s relaunched system. The market is small, and the DSP boxes 
pricey. It becomes relatively more affordable for large multi-speaker systems 
with large budgets.

For the rest of us, it’s down to software. Ircam have a basic implementation of 
DBAP in Spat~ for Max/MSP (or you can roll your own), and adding the delay 
component is relatively simple. You can then scale the amplitude and delay 
separately for each source, as seems appropriate. Using delay alone is 
surprisingly effective. The variation of amplitude between widely spaced 
speakers can be excessive.

Of course you need a fast and powerful computer, and efficient programming to 
do this, but that is also true with any of the alternative algorithms 
(ambisonics, VBAP, DBAP, WFS etc.). None of these are perfect for every 
situation, and it is hard to envisage a combination of them that would work.

Ciao,

Dave Hunt


> On 14 Nov 2020, at 17:00, sursound-requ...@music.vt.edu wrote:
> 
> From: Richard Foss <rich...@immersivedsp.com>
> Subject: Re: [Sursound] Was: Recorder for ORTF-3D OUTDOOR SET
> Date: 14 November 2020 at 16:48:36 GMT
> To: sursound@music.vt.edu
> 
> 
> Dave, I have meant to follow up on your message  for some time, because your 
> ideas match what I am currently busy with - at last getting to it!
> 
> Our first immersive audio implementation uses networked PoE miniDSP speakers 
> which each incorporate a matrix mixer with volume and delay control at the 
> cross points. The delays were a later addition, and I certainly found that 
> the localization was enhanced by incorporating delays. We implemented DBAP 
> for the amplitude panning, but we have implemented and experimented with 
> VBAP. Given that our targeted applications will need irregular speaker 
> configurations, we have settled on DBAP for now.
> 
> We had an idea, similar to yours, to utilize the signal processing 
> capabilities of audio interfaces/mixers. Because we owned MOTU devices, we 
> tried this first on three of the MOTU devices, and have updated our ImmerGo 
> software to work with these interfaces. However, it was not possible to 
> implement a delay matrix on the MOTU devices, so they just have a DBAP 
> implementation, not DBADP (your innovative label:)).
> 
> I am now working on a mixing console implementation where I believe I can 
> have delay/EQ at matrix cross points for a few channels, where there is an 
> inverse relationship between number of speakers and number of channels with 
> delay/EQ, although all channels can have DBAP. One does need to have mix 
> buses to enable this, and also there often is the timing constraint, because 
> a lot of messages go to the mixer as the sound sources are spatialized. I 
> have found that the MOTU devices are very responsive in this regard.
> 
> Anyway, good to have a fellow DBADP enthusiast .
> 
> Regards,
> 
> Richard.
> 
> —
> Richard Foss (PhD)
> Software engineer/director
> 
> ImmersiveDSP
> 46 Silver Ranch Estate
> Keurbooms River Road
> Plettenberg Bay 6600
> South Africa

_______________________________________________
Sursound mailing list
Sursound@music.vt.edu
https://mail.music.vt.edu/mailman/listinfo/sursound - unsubscribe here, edit 
account or options, view archives and so on.

Reply via email to