X-Git-Url: https://git.sesse.net/?p=movit;a=blobdiff_plain;f=init.h;h=9c371e0316b4b6987a4a2e70ab995127d399e9d2;hp=e0dbe63688fff77225e45e4a162b2ad7cf1003df;hb=c8c1721179a81ac5384fab773e59be544b57128b;hpb=34f5f331bd3b643949ccbb0c3488af2823634d59 diff --git a/init.h b/init.h index e0dbe63..9c371e0 100644 --- a/init.h +++ b/init.h @@ -2,8 +2,8 @@ #define _INIT_H // Initialize the library; in particular, will query the GPU for information -// that is needed by various components. (In time, for instance, we will query -// about extensions here.) +// that is needed by various components. For instance, it verifies that +// we have all the OpenGL extensions we need. void init_movit(); // GPU features. These are not intended for end-user use. @@ -11,11 +11,20 @@ void init_movit(); // Whether init_movit() has been called. extern bool movit_initialized; -// An estimate on the number of different levels the linear texture interpolation -// of the GPU can deliver. My Intel card seems to be limited to 2^6 levels here, -// while a modern nVidia card (GTX 550 Ti) seem to use 2^8. +// An estimate on the smallest values the linear texture interpolation +// of the GPU can distinguish between, i.e., for a GPU with N-bit +// texture subpixel precision, this value will be 2^-N. // -// We currently don't bother to test above 2^10. +// From reading the little specs that exist and through practical tests, +// the broad picture seems to be that Intel cards have 6-bit precision, +// nVidia cards have 8-bit, and Radeon cards have 6-bit before R6xx +// (at least when not using trilinear sampling), but can reach +// 8-bit precision on R6xx or newer in some (unspecified) cases. +// +// We currently don't bother to test for more than 1024 levels. extern float movit_texel_subpixel_precision; +// Whether the GPU in use supports GL_EXT_texture_sRGB. +extern bool movit_srgb_textures_supported; + #endif // !defined(_INIT_H)