static void virtio_gpu_gl_flushed(VirtIOGPUBase *b)
{
VirtIOGPU *g = VIRTIO_GPU(b);
- VirtIOGPUGL *gl = VIRTIO_GPU_GL(b);
- if (gl->renderer_reset) {
- gl->renderer_reset = false;
- virtio_gpu_virgl_reset(g);
- }
virtio_gpu_process_cmdq(g);
}
virtio_gpu_virgl_init(g);
gl->renderer_inited = true;
}
+ if (gl->renderer_reset) {
+ gl->renderer_reset = false;
+ virtio_gpu_virgl_reset(g);
+ }
cmd = virtqueue_pop(vq, sizeof(struct virtio_gpu_ctrl_command));
while (cmd) {
virtio_gpu_reset(vdev);
- if (gl->renderer_inited) {
- if (g->parent_obj.renderer_blocked) {
- gl->renderer_reset = true;
- } else {
- virtio_gpu_virgl_reset(g);
- }
+ /*
+ * GL functions must be called with the associated GL context in main
+ * thread, and when the renderer is unblocked.
+ */
+ if (gl->renderer_inited && !gl->renderer_reset) {
+ virtio_gpu_virgl_reset_scanout(g);
+ gl->renderer_reset = true;
}
}
virtio_gpu_fence_poll(g);
}
-void virtio_gpu_virgl_reset(VirtIOGPU *g)
+void virtio_gpu_virgl_reset_scanout(VirtIOGPU *g)
{
int i;
- virgl_renderer_reset();
for (i = 0; i < g->parent_obj.conf.max_outputs; i++) {
dpy_gfx_replace_surface(g->parent_obj.scanout[i].con, NULL);
dpy_gl_scanout_disable(g->parent_obj.scanout[i].con);
}
}
+void virtio_gpu_virgl_reset(VirtIOGPU *g)
+{
+ virgl_renderer_reset();
+}
+
int virtio_gpu_virgl_init(VirtIOGPU *g)
{
int ret;
void virtio_gpu_virgl_process_cmd(VirtIOGPU *g,
struct virtio_gpu_ctrl_command *cmd);
void virtio_gpu_virgl_fence_poll(VirtIOGPU *g);
+void virtio_gpu_virgl_reset_scanout(VirtIOGPU *g);
void virtio_gpu_virgl_reset(VirtIOGPU *g);
int virtio_gpu_virgl_init(VirtIOGPU *g);
int virtio_gpu_virgl_get_num_capsets(VirtIOGPU *g);