For an important but infrequently used program based on FLTK, I have suddenly 
found that it is able to obtain only 5-bit color buffers (5+5+5 instead of 
8+8+8), causing my application to mis-perform.  (I read back the frame buffer, 
so the bitdepth is important to the proper function of the program.)

I don't believe I rebuilt the application recently, but I did upgrade my Macs 
from 10.6 (where I know the program worked) to 10.6.4 (in steps of course, but 
I didn't check the program at each step), and now it no longer works.  During 
this progression of upgrades, I did not re-compile the program.  If I now 
re-compile, the program still has the problem.  Any suggestions?

For a bit more information, I am currently using Mac 10.6.4, and fink's package 
for FLTK 1.1.7.  The program is compiled for 32-bit.  My graphics card is an 
NVIDIA 8800GT.

To try getting to the bottom of the problem, I tried invoking 
fl_window->mode(FL_RGB8), although I don't have to, and I check 
fl_window->can_do()), which returns true.  However, I then ask OpenGL for the 
depth of the returned window using glGetIntegerv(GL_RED_BITS,&redbits), it says 
5 bits.  It should say 8.

Remember that this code has been working for years, and I believe the bug 
cropped up without me recompiling the program.  Thus, I suspect it has arisen 
from one of the Mac upgrades, possibly in GLX, or X, or Quartz, or in the 
NVIDIA driver.  Any suggestions?

-Marc Levoy, Stanford University
_______________________________________________
fltk-opengl mailing list
fltk-opengl@easysw.com
http://lists.easysw.com/mailman/listinfo/fltk-opengl

Reply via email to