I don't know if this is normal since I've no experience with other masters. However I've seen an asimmetry in code for user and hard real time. 1) In user example // write application time to master clock_gettime(CLOCK_TO_USE, &time); ecrt_master_application_time(master, TIMESPEC2NS(time));
2) In rtai example. .... tv.tv_usec += 1000; if (tv.tv_usec >= 1000000) { tv.tv_usec -= 1000000; tv.tv_sec++; } ecrt_master_application_time(master, EC_TIMEVAL2NANO(tv));In the RTai example the clock is taken a first time before the main loop and after is calculated manually (incrementing 1000 mSec), in these way (it seems to me) the time I give to master is always with the (small enough) jitter of the RTai scheduler.
I'm wrong ? -- Dott. Gianluca Medini Eurosoft S.r.l. Viale Colombo 74 Fucecchio FI Qualora questo messaggio fosse da Voi ricevuto per errore, vogliate cortesemente darcene notizia a mezzo fax (oppure e-mail) e distruggere il messaggio stesso. Qualsiasi utilizzo e/o conservazione dei dati ricevuti per errore, costituisce violazione delle disposizioni del D.Lgs 196/03 sulla tutela dei Dati Personali.
<<attachment: gianluca.vcf>>
_______________________________________________ etherlab-users mailing list etherlab-users@etherlab.org http://lists.etherlab.org/mailman/listinfo/etherlab-users