[Public WebGL] gl.enable(gl.TEXTURE_2D)

Kenneth Russell [email protected]
Wed Jan 20 16:07:37 PST 2010


On Wed, Jan 20, 2010 at 3:48 PM, Kenneth Russell <[email protected]> wrote:
> On Tue, Jan 19, 2010 at 8:37 AM, Giles Thomas <[email protected]> wrote:
>> Hi all,
>>
>> A quick follow-up question:
>>
>> 2010/1/13 Kenneth Russell <[email protected]>
>>>
>>> TEXTURE_2D is not a valid enable bit in OpenGL ES 2.0 or, consequently,
>>> WebGL. The enum was in the EnableCap section only because it was there in
>>> the OpenGL ES 2.0 headers. I've moved its definition to the TextureTarget
>>> section.
>>
>> Have there been versions of Chrome (or perhaps WebKit?) where
>> gl.enable(gl.TEXTURE_2D) was erroneously required to enable textures?  A
>> reader of my blog reports that Chrome 4.0.295 seems to need it with a
>> specific OS/graphics card combination -- details here:
>> http://learningwebgl.com/blog/?p=684&cpage=1#comment-787
>
> If there is a sample which only works with that call in place, then it
> isn't Chrome or WebKit imposing that requirement but the OpenGL
> driver. It should not be necessary to enable the TEXTURE_2D bit on a
> particular texture unit in order to sample it in a shader. You should
> report this issue to the graphics card vendor.
>
> By the way, because the code is evolving quickly, at this point I only
> recommend using WebGL inside of the latest Chromium builds. See
> http://khronos.org/webgl/wiki/Getting_a_WebGL_Implementation for
> instructions on downloading and running them.

To clarify, I should have said, if you're trying to use WebGL inside
of Chrome, use the Chromium continuous builds rather than the Dev
Channel builds for the time being.

-Ken

-----------------------------------------------------------
You are currently subscribe to [email protected]
To unsubscribe, send an email to [email protected] with
the following command in the body of your email:




More information about the public_webgl mailing list