On Sun, 24 Aug 2008, Eric Kidd wrote:
Greetings, Haskell folks!
I'm working on a toy program for drum synthesis. This has lead me to
explore the various sound-related libraries in Hackage. Along the way,
I've learned several things:
1. There's a lot of Haskell sound libraries, but no agreement on how
to represent buffers of audio data.
In my sound synthesis library
http://darcs.haskell.org/synthesizer/
I use mainly two representations:
1. Chunky sequences based on StorableVector. This is for efficient storage
and functions that need to process data in a non-causal way. They are not
good for fusion.
http://darcs.haskell.org/synthesizer/src/Synthesizer/Storable/Signal.hs
2. A Stream like sequences without wrappers. In fact, they are no data
storages but generator functions equipped with a seed. I use them for
causal signal processes. I think that most functions needed for drum
synthesis are causal. By inlining them, many subsequent causal signal
processes are fused to one.
http://darcs.haskell.org/synthesizer/src/Synthesizer/State/Signal.hs
2. Some of the most useful sound libraries aren't listed in Hackage's
"Sound" section, including HCodecs, SDL-mixer and hogg.
my library is also not at Hackage, it's too much experimental currently
3. The Haskell OpenAL bindings are great, but ALUT has been removed
from MacOS 10.5. Of course, a pure-Haskell version of ALUT would
be pretty easy to write, and it could be used as a fallback.
4. '[[a]]' is _not_ a good representation for multi-channel audio
data. Manipulating even 10 seconds of audio is noticeably slow.
Using the signal representations I mentioned above, I can synthesize some
songs in realtime. For stereo signals I use Signal (a,a). However, a
StereoFrame data type with Storable instance might be even better.
5. HCodecs is probably the best library for reading WAVE files. It
uses 'DiffUArray Int a' to represent audio data, and not '[[a]]'.
It also supports MIDI and SoundFont, which is quite handy.
6. YampaSynth looks really cool, and it might be possible to turn it
into a reusable MIDI synthesizer library for use with HCodecs.
This would be extremely useful: it would provide a basis for all
kinds of crazy MIDI-based programming in Haskell.
If you are after realtime synthesis you might also want to look into
SuperCollider. I posted some code to control SuperCollider via MIDI to
haskell-art. I can send it to you, if you like.
What would the ideal Haskell sound API look like? Personally, I would
love to see:
a. Something like HCodecs's Data.Audio as the basic sound format.
Data.Audio is an array-based API, and it supports converting
between a wide range of common sample formats. I don't know how
fast this is, but it already exists, and it's used by YampaSynth.
b. OpenAL for sound playback, with a portable version of ALUT. This
may require writing a pure-Haskell version of ALUT.
So far I used Sox' play command fed by a pipe with runInteractiveCommand.
Other nice-to-have features might include:
e. A standard MIDI format, based on either the HCodecs package or the
midi package. (HCodecs is used by YampaSynth, and the midi package
is used by alsa-midi.)
midi package is mainly used by Haskore
f. A modular version of YampaSynth which can convert MIDI data
structures into Data.Audio values.
It looks like Haskell could be a really sweet audio programming
environment with just a bit of integration work. What do folks think?
Are there other libraries I should look at more closely? Other features
that should be included in an ideal audio API?
What do you mean by 'API'? I think there should not be one library
handling all kind of music tasks. For interoperability common data
structures are nice. However, what signal representation concerns,
I think we need more than one representation, because different
applications have different needs.
Haskell audio libraries
I have advertised my libraries for years in Haskell-Wiki and HCAR, but it
seems that people do not look there. :-(
Reading and writing sound files:
HCodecs: (Audible a) => DiffUArray Int a
hsndfile: MArray with Float and Double
HSoundFile: [[Double]]
ALUT: OpenAL.AL.Buffer
WAVE: [[Int32]]
Playing sounds:
OpenAL: Ptr UInt8, Ptr Int16, 1-N channels
SDL-mixer: ForeignPtr ChunkStruct, ForeignPtr MusicStruct
Sox via pipe, using any data structure which can be written to a file
Sound processing libraries:
dsp: Array a, [a]
jack: Operates via mapping functions
mainMono :: (CFloat -> IO CFloat) -> IO ()
mainStereo :: ((CFloat, CFloat) -> IO (CFloat, CFloat)) -> IO ()
synthesizer: StorableVector, State, Generic (older: list, fusable list)
MIDI-based:
HCodecs: Reads and writes MIDI files
midi: Reads and writes MIDI files
alsa-midi: Uses midi library
YampaSynth: Stand-alone program
Haskore uses MIDI and other back-ends like Haskell Synthesizer, CSound,
SuperCollider
Special-purpose APIs only (FFIs, etc.):
hCsound: CsoundPtr
hsc3: UGen
HasSound
No public sound-buffer API:
hbeat: The relevant source files are missing!
hogg: Very low-level API for Ogg internals only
libmpd: No sound buffer API
sonic-visualizer: No sound buffer API
truelevel: Stand-alone program (uses WAVE)
wavconvert: Stand-alone program
_______________________________________________
Haskell-Cafe mailing list
Haskell-Cafe@haskell.org
http://www.haskell.org/mailman/listinfo/haskell-cafe