What's wrong with writefln("%s", glGetString(GL_EXTENSIONS));

Chris Sauls ibisbasenji at gmail.com
Sat Mar 4 10:28:34 PST 2006


Cris wrote:
> it doesn't work that way: "src\engine\renderer.d(104): cannot implicitly 
> convert expression ((glGetString)(7939u)) of type ubyte* to char*"
> 

Given that its natural type is ubyte* you should be able to do this:

# private import std.string ;
#
# char[] toString (ubyte* foo) {
#   return std.string.toString(cast(char*) foo);
# }

I would think so, anyhow.  If it will not allow a direct cast to char*, you can always add 
a middle state cast to void* first.

-- Chris Nicholson-Sauls



More information about the Digitalmars-d-learn mailing list