river/river/LockManager.zig
tiosgz 4726a6b0f1 Root: migrate {all,active}_outputs to wl.list
As discussed with ifreund on irc. This avoids extra allocation in case
of all_outputs and confusion in case of active_outputs (because with the
Output embedded in the Node, i thought its value was copied instead of
its pointer).
2023-08-13 11:10:46 +00:00

267 lines
8.8 KiB
Zig

// This file is part of river, a dynamic tiling wayland compositor.
//
// Copyright 2021 The River Developers
//
// This program is free software: you can redistribute it and/or modify
// it under the terms of the GNU General Public License as published by
// the Free Software Foundation, version 3.
//
// This program is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU General Public License for more details.
//
// You should have received a copy of the GNU General Public License
// along with this program. If not, see <https://www.gnu.org/licenses/>.
const LockManager = @This();
const std = @import("std");
const assert = std.debug.assert;
const build_options = @import("build_options");
const wlr = @import("wlroots");
const wl = @import("wayland").server.wl;
const server = &@import("main.zig").server;
const util = @import("util.zig");
const LockSurface = @import("LockSurface.zig");
const log = std.log.scoped(.session_lock);
state: enum {
/// No lock request has been made and the session is unlocked.
unlocked,
/// A lock request has been made and river is waiting for all outputs to have
/// rendered a lock surface before sending the locked event.
waiting_for_lock_surfaces,
/// A lock request has been made but waiting for a lock surface to be rendered
/// on all outputs timed out. Now river is waiting only for all outputs to at
/// least be blanked before sending the locked event.
waiting_for_blank,
/// All outputs are either blanked or have a lock surface rendered and the
/// locked event has been sent.
locked,
} = .unlocked,
lock: ?*wlr.SessionLockV1 = null,
/// Limit on how long the locked event will be delayed to wait for
/// lock surfaces to be created and rendered. If this times out, then
/// the locked event will be sent immediately after all outputs have
/// been blanked.
lock_surfaces_timer: *wl.EventSource,
new_lock: wl.Listener(*wlr.SessionLockV1) = wl.Listener(*wlr.SessionLockV1).init(handleLock),
unlock: wl.Listener(void) = wl.Listener(void).init(handleUnlock),
destroy: wl.Listener(void) = wl.Listener(void).init(handleDestroy),
new_surface: wl.Listener(*wlr.SessionLockSurfaceV1) =
wl.Listener(*wlr.SessionLockSurfaceV1).init(handleSurface),
pub fn init(manager: *LockManager) !void {
const event_loop = server.wl_server.getEventLoop();
const timer = try event_loop.addTimer(*LockManager, handleLockSurfacesTimeout, manager);
errdefer timer.remove();
manager.* = .{
.lock_surfaces_timer = timer,
};
const wlr_manager = try wlr.SessionLockManagerV1.create(server.wl_server);
wlr_manager.events.new_lock.add(&manager.new_lock);
}
pub fn deinit(manager: *LockManager) void {
// deinit() should only be called after wl.Server.destroyClients()
assert(manager.lock == null);
manager.lock_surfaces_timer.remove();
manager.new_lock.link.remove();
}
fn handleLock(listener: *wl.Listener(*wlr.SessionLockV1), lock: *wlr.SessionLockV1) void {
const manager = @fieldParentPtr(LockManager, "new_lock", listener);
if (manager.lock != null) {
log.info("denying new session lock client, an active one already exists", .{});
lock.destroy();
return;
}
manager.lock = lock;
if (manager.state == .unlocked) {
manager.state = .waiting_for_lock_surfaces;
if (build_options.xwayland) {
server.root.layers.xwayland_override_redirect.node.setEnabled(false);
}
manager.lock_surfaces_timer.timerUpdate(200) catch {
log.err("error setting lock surfaces timer, imperfect frames may be shown", .{});
manager.state = .waiting_for_blank;
// This call is necessary in the case that all outputs in the layout are disabled.
manager.maybeLock();
};
{
var it = server.input_manager.seats.first;
while (it) |node| : (it = node.next) {
const seat = &node.data;
seat.setFocusRaw(.none);
seat.cursor.updateState();
// Enter locked mode
seat.prev_mode_id = seat.mode_id;
seat.enterMode(1);
}
}
} else {
if (manager.state == .locked) {
lock.sendLocked();
}
log.info("new session lock client given control of already locked session", .{});
}
lock.events.new_surface.add(&manager.new_surface);
lock.events.unlock.add(&manager.unlock);
lock.events.destroy.add(&manager.destroy);
}
fn handleLockSurfacesTimeout(manager: *LockManager) c_int {
log.err("waiting for lock surfaces timed out, imperfect frames may be shown", .{});
assert(manager.state == .waiting_for_lock_surfaces);
manager.state = .waiting_for_blank;
{
var it = server.root.active_outputs.iterator(.forward);
while (it.next()) |output| {
output.normal_content.node.setEnabled(false);
output.locked_content.node.setEnabled(true);
}
}
// This call is necessary in the case that all outputs in the layout are disabled.
manager.maybeLock();
return 0;
}
pub fn maybeLock(manager: *LockManager) void {
var all_outputs_blanked = true;
var all_outputs_rendered_lock_surface = true;
{
var it = server.root.active_outputs.iterator(.forward);
while (it.next()) |output| {
switch (output.lock_render_state) {
.pending_unlock, .unlocked, .pending_blank, .pending_lock_surface => {
all_outputs_blanked = false;
all_outputs_rendered_lock_surface = false;
},
.blanked => {
all_outputs_rendered_lock_surface = false;
},
.lock_surface => {},
}
}
}
switch (manager.state) {
.waiting_for_lock_surfaces => if (all_outputs_rendered_lock_surface) {
log.info("session locked", .{});
// The lock client may have been destroyed, for example due to a protocol error.
if (manager.lock) |lock| lock.sendLocked();
manager.state = .locked;
manager.lock_surfaces_timer.timerUpdate(0) catch {};
},
.waiting_for_blank => if (all_outputs_blanked) {
log.info("session locked", .{});
// The lock client may have been destroyed, for example due to a protocol error.
if (manager.lock) |lock| lock.sendLocked();
manager.state = .locked;
},
.unlocked, .locked => unreachable,
}
}
fn handleUnlock(listener: *wl.Listener(void)) void {
const manager = @fieldParentPtr(LockManager, "unlock", listener);
// TODO(wlroots): this will soon be handled by the wlroots session lock implementation
if (manager.state != .locked) {
manager.lock.?.resource.postError(.invalid_unlock, "the locked event was never sent");
return;
}
manager.state = .unlocked;
log.info("session unlocked", .{});
{
var it = server.root.active_outputs.iterator(.forward);
while (it.next()) |output| {
assert(!output.normal_content.node.enabled);
output.normal_content.node.setEnabled(true);
assert(output.locked_content.node.enabled);
output.locked_content.node.setEnabled(false);
}
}
if (build_options.xwayland) {
server.root.layers.xwayland_override_redirect.node.setEnabled(true);
}
{
var it = server.input_manager.seats.first;
while (it) |node| : (it = node.next) {
const seat = &node.data;
seat.setFocusRaw(.none);
// Exit locked mode
seat.enterMode(seat.prev_mode_id);
}
}
handleDestroy(&manager.destroy);
server.root.applyPending();
}
fn handleDestroy(listener: *wl.Listener(void)) void {
const manager = @fieldParentPtr(LockManager, "destroy", listener);
log.debug("ext_session_lock_v1 destroyed", .{});
manager.new_surface.link.remove();
manager.unlock.link.remove();
manager.destroy.link.remove();
manager.lock = null;
if (manager.state == .waiting_for_lock_surfaces) {
manager.state = .waiting_for_blank;
manager.lock_surfaces_timer.timerUpdate(0) catch {};
}
}
fn handleSurface(
listener: *wl.Listener(*wlr.SessionLockSurfaceV1),
wlr_lock_surface: *wlr.SessionLockSurfaceV1,
) void {
const manager = @fieldParentPtr(LockManager, "new_surface", listener);
log.debug("new ext_session_lock_surface_v1 created", .{});
assert(manager.state != .unlocked);
assert(manager.lock != null);
LockSurface.create(wlr_lock_surface, manager.lock.?) catch {
log.err("out of memory", .{});
wlr_lock_surface.resource.postNoMemory();
};
}