Hey, folks,

I'm reasonably new to Cocoa developing.  I'm playing around with a subclass
of NSView which generates some stuff based on the size of the view.  The
code basically looks like:

-----------------------------
@interface FooView : NSView {
    NSRect myBounds;
}
- (void)doSomeStuff;
@end

@implementation FooView
- (void)doSomeStuff
{
    // myBounds at this point is set to {0, 0, 0.0, 0.0}, even if this is
called after drawRect is fired

    // This will set myBounds to {0, 0, 0.0, 0.0}
    myBounds = [self bounds];
}

- (void)drawRect: (NSRect)aRect
{
    myBounds = [self bounds];
}
@end
-------------------------

The weird thing is, when drawRect gets called, [self bounds] returns a
reasonable looking bounding rectangle.  When doSomeStuff gets called, [self
bounds] returns {0, 0, 0.0, 0.0}.  The extra weird thing is, if doSomeStuff
is called *after* drawRect gets fired, myBounds is {0, 0, 0.0, 0.0},
*despite* it being set to something reasonable in drawRect.

Any ideas what's going on here?

Cheers,
-- 

Jim

Fear is the dark room where the Devil develops his negatives.

  - Gary Busey
_______________________________________________

Cocoa-dev mailing list (Cocoa-dev@lists.apple.com)

Please do not post admin requests or moderator comments to the list.
Contact the moderators at cocoa-dev-admins(at)lists.apple.com

Help/Unsubscribe/Update your Subscription:
http://lists.apple.com/mailman/options/cocoa-dev/archive%40mail-archive.com

This email sent to arch...@mail-archive.com

Reply via email to