[Adding Doug in so he sees the feedback] Thanks Sepideh, I must admit I have not tried it myself but watch the demo in action with a blind user.
Steve Lee OpenDirective http://opendirective.com On 16 April 2015 at 15:38, Shahi, Sepideh <[email protected]> wrote: > Hi Steve, > This is a very interesting project. I was able to play with the screen > reader for the Pie Chart, however, the sonifier function did not work.It > creates a trend line but it does not play it... > > Just a few points to consider: > > - When it’s reading aloud the pie chart, it gives out too much > information at once including total, highest, lowest, average, median and > several actions that user can take, which may not be necessary for all > users. > - Focus is fixed around the chart and it does not move to different > sections as user tabs to different parts of the chart. This makes it > difficult to expect what is going to be read next. > - The trend line does not have any visual association with the pie > chart. I missed it the first time it was displayed and the next time I > assumed there was something wrong with my display until I realized that is > supposed be a trend line. > > > Thanks, > Sepideh > > > On Apr 10, 2015, at 9:35 AM, Steve Lee <[email protected]> wrote: > > FYI - Doug Schepers had a neat hallway demo at CSUN using pitch to > explore SVG charts > > > https://github.com/shepazu/describler > Steve Lee > OpenDirective http://opendirective.com > > > On 9 April 2015 at 20:48, [email protected] <[email protected]> > wrote: > > Hi everyone, > > As part of the Floe Project's efforts to create personalized, accessible > user interfaces that can be used across a variety of Open Educational > Resources, we've been working on a design framework and JavaScript toolkit > for authoring multimodal charts, graphs, and other data "visualizations." > One of the central goals of this effort is to make it easier for teachers, > students, and content authors to represent data in "layers" consisting of > different modalities--graphics, text, and audio. > > To start, we've been focusing a lot on sonification, the process of > representing data using sounds. We're in the midst of a very early > brainstorming, sketching and idea generation process. Our work is > documented > in the wiki: > > http://wiki.fluidproject.org/display/fluid/Floe+Sonification+Framework > http://wiki.fluidproject.org/display/fluid/Exploration+and+Early+Sketches > http://wiki.fluidproject.org/display/fluid/Sonification+Sketches > http://wiki.fluidproject.org/display/fluid/Use+Cases > > In order to start exploring the potential of data sonification in a way > that > allows us to experiment with different approaches and to iterate from > mockups to working implementations reasonably quickly, we've constrained > our > current design sketches to tools that will help authors produce multimodal > "pie charts." The goal of this tool is to enable authors to produce layered > representations of fairly simple data, and to give end-users the ability to > explore, remap, and share their own personalized sonifications and > visualizations. > > As part of this process, we've been exploring some new methods for how we > design sonifications strategies and evaluate their effectiveness. We've > started working with a small, informal group of people in a co-design > context, and will also be sharing our in-progress work here on the list. > What we've done so far is to prototype several different types of > sonifications using low-tech tools and then shared them with people using a > process of "progressive explanation." We start by having them listen to the > sonification with no additional cues or explanation, asking them to > describe > their impressions (including how they imagine the sounds map to some > underlying data set). From there, we progressively explain more about the > intentions behind sonification (such as describing the sound mapping using > an "audio legend"), and continue to gather impressions and ideas from our > listeners. We've found this to be a very helpful process for exploring how > much textual or explanatory supporting material to provide with a given > sonification approach. Sepideh has posted some great examples and > prototypes > in the wiki. > > Over the coming months, we'll expand this design effort to encompass more > complex data and to more interactive situations such as simulations, games, > and performances. > > We'll continue to share ideas, sketches, and works in progress here on the > mailing list, in the #fluid-design IRC channel, and in the wiki. > Constructive feedback and creative ideas are always appreciated during this > early stage in the process, as well as the understanding that we're still > experimenting and exploring the design space. Failures and half-baked ideas > are as useful at this stage in the design process as successes. > > Colin > > _______________________________________________________ > fluid-work mailing list - [email protected] > To unsubscribe, change settings or access archives, > see http://lists.idrc.ocad.ca/mailman/listinfo/fluid-work > > _______________________________________________________ > fluid-work mailing list - [email protected] > To unsubscribe, change settings or access archives, > see http://lists.idrc.ocad.ca/mailman/listinfo/fluid-work > > >
_______________________________________________________ fluid-work mailing list - [email protected] To unsubscribe, change settings or access archives, see http://lists.idrc.ocad.ca/mailman/listinfo/fluid-work
