1)How can i draw on screen, something which i created on pBuffer..ie an image is attached to a pBuffer…and if i want to render it …how can i do it.. i guess there are only 2 ways…using glReadPixels or using glCopyPixels.Am i right?
2) What does the api eglCreatePBufferFromClientBuffer() do?
Is the surface returned by it an offscreen buffer connected surface or framebuffer connected surface ? It seems it returns a surface which is connected to an offscreen buffer….
eglCreatePbufferFromClientBuffer allows you to take a buffer from an EGL client API and encapsulate it as a pbuffer surface. Currently the only supported buffer type is an OpenVG VGimage. This enables such use cases as rendering to a VGimage with OpenGL ES.
VGimage is an off-screen buffer, and so are pbuffers.
I’m not sure I understand your first question. Could you please rephrase it? Where does the image data come from and where do you want to put it?
Viewing 2 posts - 1 through 2 (of 2 total)
You must be logged in to reply to this topic.