On 10 Oct 2016, at 21:24, Julian Bradfield <jcb+unic...@inf.ed.ac.uk> wrote:
> 
>> We need reliable plain-text notation systems. Otherwise distinctions we wish 
>> to encode may be lost. 
> 
> We have no need to make such distinctions in "plain text”.

You mightn’t. 

> It's convenient to have major distinctions easily accessible without
> font hacking,

Yes, indeed. 

> but there's no need to have every notation one might dream up forcibly 
> incorporated into "plain text”.

Hyperbole.

> In particular, for super/subscripts, which is where we came in, even
> the benighted souls using Word still typically recognize and can use
> LaTeX notation.

I can’t use LaTeX notation. I don’t use that proprietary system. And don’t you 
dare tell me that I am benighted, or using Word. Neither applies.


On 10 Oct 2016, at 21:31, Julian Bradfield <jcb+unic...@inf.ed.ac.uk> wrote:

> On 2016-10-10, Hans Åberg <haber...@telia.com> wrote:
>> It is possible to write math just using ASCII and TeX, which was the 
>> original idea of TeX. Is that want you want for linguistics?
> 
> I don't see the need to do everything in plain text.

Of course not. You’re a programmer. 

(Mathematical typesetting is not my concern.)

> Because phonetics has a much small set of symbols, I do kwəɪt ləɪk
> biːɪŋ eɪbl tʊ duː ðɪs, and because they're also used in non-specialist
> writing, it's useful to have the symbols hacked into Unicode instead
> of hacked into specialist fonts.
> But subscripts? No need.

And yet we use such things. 

I have an edition of the Bible I’m setting. Big book. Verse numbers. I like 
these to be superscript so they’re unobtrusive. Damn right I use the 
superscript characters for these. I can process the text, export it for 
concordance processing, whatever, and those out-of-text notations DON’T get 
converted to regular digits, which I need.

Michael

Reply via email to