4726a6b0f1
As discussed with ifreund on irc. This avoids extra allocation in case of all_outputs and confusion in case of active_outputs (because with the Output embedded in the Node, i thought its value was copied instead of its pointer).
267 lines
8.8 KiB
Zig
267 lines
8.8 KiB
Zig
// This file is part of river, a dynamic tiling wayland compositor.
|
|
//
|
|
// Copyright 2021 The River Developers
|
|
//
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, version 3.
|
|
//
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
|
|
const LockManager = @This();
|
|
|
|
const std = @import("std");
|
|
const assert = std.debug.assert;
|
|
|
|
const build_options = @import("build_options");
|
|
|
|
const wlr = @import("wlroots");
|
|
const wl = @import("wayland").server.wl;
|
|
|
|
const server = &@import("main.zig").server;
|
|
const util = @import("util.zig");
|
|
|
|
const LockSurface = @import("LockSurface.zig");
|
|
|
|
const log = std.log.scoped(.session_lock);
|
|
|
|
state: enum {
|
|
/// No lock request has been made and the session is unlocked.
|
|
unlocked,
|
|
/// A lock request has been made and river is waiting for all outputs to have
|
|
/// rendered a lock surface before sending the locked event.
|
|
waiting_for_lock_surfaces,
|
|
/// A lock request has been made but waiting for a lock surface to be rendered
|
|
/// on all outputs timed out. Now river is waiting only for all outputs to at
|
|
/// least be blanked before sending the locked event.
|
|
waiting_for_blank,
|
|
/// All outputs are either blanked or have a lock surface rendered and the
|
|
/// locked event has been sent.
|
|
locked,
|
|
} = .unlocked,
|
|
lock: ?*wlr.SessionLockV1 = null,
|
|
|
|
/// Limit on how long the locked event will be delayed to wait for
|
|
/// lock surfaces to be created and rendered. If this times out, then
|
|
/// the locked event will be sent immediately after all outputs have
|
|
/// been blanked.
|
|
lock_surfaces_timer: *wl.EventSource,
|
|
|
|
new_lock: wl.Listener(*wlr.SessionLockV1) = wl.Listener(*wlr.SessionLockV1).init(handleLock),
|
|
unlock: wl.Listener(void) = wl.Listener(void).init(handleUnlock),
|
|
destroy: wl.Listener(void) = wl.Listener(void).init(handleDestroy),
|
|
new_surface: wl.Listener(*wlr.SessionLockSurfaceV1) =
|
|
wl.Listener(*wlr.SessionLockSurfaceV1).init(handleSurface),
|
|
|
|
pub fn init(manager: *LockManager) !void {
|
|
const event_loop = server.wl_server.getEventLoop();
|
|
const timer = try event_loop.addTimer(*LockManager, handleLockSurfacesTimeout, manager);
|
|
errdefer timer.remove();
|
|
|
|
manager.* = .{
|
|
.lock_surfaces_timer = timer,
|
|
};
|
|
|
|
const wlr_manager = try wlr.SessionLockManagerV1.create(server.wl_server);
|
|
wlr_manager.events.new_lock.add(&manager.new_lock);
|
|
}
|
|
|
|
pub fn deinit(manager: *LockManager) void {
|
|
// deinit() should only be called after wl.Server.destroyClients()
|
|
assert(manager.lock == null);
|
|
|
|
manager.lock_surfaces_timer.remove();
|
|
|
|
manager.new_lock.link.remove();
|
|
}
|
|
|
|
fn handleLock(listener: *wl.Listener(*wlr.SessionLockV1), lock: *wlr.SessionLockV1) void {
|
|
const manager = @fieldParentPtr(LockManager, "new_lock", listener);
|
|
|
|
if (manager.lock != null) {
|
|
log.info("denying new session lock client, an active one already exists", .{});
|
|
lock.destroy();
|
|
return;
|
|
}
|
|
|
|
manager.lock = lock;
|
|
|
|
if (manager.state == .unlocked) {
|
|
manager.state = .waiting_for_lock_surfaces;
|
|
|
|
if (build_options.xwayland) {
|
|
server.root.layers.xwayland_override_redirect.node.setEnabled(false);
|
|
}
|
|
|
|
manager.lock_surfaces_timer.timerUpdate(200) catch {
|
|
log.err("error setting lock surfaces timer, imperfect frames may be shown", .{});
|
|
manager.state = .waiting_for_blank;
|
|
// This call is necessary in the case that all outputs in the layout are disabled.
|
|
manager.maybeLock();
|
|
};
|
|
|
|
{
|
|
var it = server.input_manager.seats.first;
|
|
while (it) |node| : (it = node.next) {
|
|
const seat = &node.data;
|
|
seat.setFocusRaw(.none);
|
|
seat.cursor.updateState();
|
|
|
|
// Enter locked mode
|
|
seat.prev_mode_id = seat.mode_id;
|
|
seat.enterMode(1);
|
|
}
|
|
}
|
|
} else {
|
|
if (manager.state == .locked) {
|
|
lock.sendLocked();
|
|
}
|
|
|
|
log.info("new session lock client given control of already locked session", .{});
|
|
}
|
|
|
|
lock.events.new_surface.add(&manager.new_surface);
|
|
lock.events.unlock.add(&manager.unlock);
|
|
lock.events.destroy.add(&manager.destroy);
|
|
}
|
|
|
|
fn handleLockSurfacesTimeout(manager: *LockManager) c_int {
|
|
log.err("waiting for lock surfaces timed out, imperfect frames may be shown", .{});
|
|
|
|
assert(manager.state == .waiting_for_lock_surfaces);
|
|
manager.state = .waiting_for_blank;
|
|
|
|
{
|
|
var it = server.root.active_outputs.iterator(.forward);
|
|
while (it.next()) |output| {
|
|
output.normal_content.node.setEnabled(false);
|
|
output.locked_content.node.setEnabled(true);
|
|
}
|
|
}
|
|
|
|
// This call is necessary in the case that all outputs in the layout are disabled.
|
|
manager.maybeLock();
|
|
|
|
return 0;
|
|
}
|
|
|
|
pub fn maybeLock(manager: *LockManager) void {
|
|
var all_outputs_blanked = true;
|
|
var all_outputs_rendered_lock_surface = true;
|
|
{
|
|
var it = server.root.active_outputs.iterator(.forward);
|
|
while (it.next()) |output| {
|
|
switch (output.lock_render_state) {
|
|
.pending_unlock, .unlocked, .pending_blank, .pending_lock_surface => {
|
|
all_outputs_blanked = false;
|
|
all_outputs_rendered_lock_surface = false;
|
|
},
|
|
.blanked => {
|
|
all_outputs_rendered_lock_surface = false;
|
|
},
|
|
.lock_surface => {},
|
|
}
|
|
}
|
|
}
|
|
|
|
switch (manager.state) {
|
|
.waiting_for_lock_surfaces => if (all_outputs_rendered_lock_surface) {
|
|
log.info("session locked", .{});
|
|
// The lock client may have been destroyed, for example due to a protocol error.
|
|
if (manager.lock) |lock| lock.sendLocked();
|
|
manager.state = .locked;
|
|
manager.lock_surfaces_timer.timerUpdate(0) catch {};
|
|
},
|
|
.waiting_for_blank => if (all_outputs_blanked) {
|
|
log.info("session locked", .{});
|
|
// The lock client may have been destroyed, for example due to a protocol error.
|
|
if (manager.lock) |lock| lock.sendLocked();
|
|
manager.state = .locked;
|
|
},
|
|
.unlocked, .locked => unreachable,
|
|
}
|
|
}
|
|
|
|
fn handleUnlock(listener: *wl.Listener(void)) void {
|
|
const manager = @fieldParentPtr(LockManager, "unlock", listener);
|
|
|
|
// TODO(wlroots): this will soon be handled by the wlroots session lock implementation
|
|
if (manager.state != .locked) {
|
|
manager.lock.?.resource.postError(.invalid_unlock, "the locked event was never sent");
|
|
return;
|
|
}
|
|
|
|
manager.state = .unlocked;
|
|
|
|
log.info("session unlocked", .{});
|
|
|
|
{
|
|
var it = server.root.active_outputs.iterator(.forward);
|
|
while (it.next()) |output| {
|
|
assert(!output.normal_content.node.enabled);
|
|
output.normal_content.node.setEnabled(true);
|
|
|
|
assert(output.locked_content.node.enabled);
|
|
output.locked_content.node.setEnabled(false);
|
|
}
|
|
}
|
|
|
|
if (build_options.xwayland) {
|
|
server.root.layers.xwayland_override_redirect.node.setEnabled(true);
|
|
}
|
|
|
|
{
|
|
var it = server.input_manager.seats.first;
|
|
while (it) |node| : (it = node.next) {
|
|
const seat = &node.data;
|
|
seat.setFocusRaw(.none);
|
|
|
|
// Exit locked mode
|
|
seat.enterMode(seat.prev_mode_id);
|
|
}
|
|
}
|
|
|
|
handleDestroy(&manager.destroy);
|
|
|
|
server.root.applyPending();
|
|
}
|
|
|
|
fn handleDestroy(listener: *wl.Listener(void)) void {
|
|
const manager = @fieldParentPtr(LockManager, "destroy", listener);
|
|
|
|
log.debug("ext_session_lock_v1 destroyed", .{});
|
|
|
|
manager.new_surface.link.remove();
|
|
manager.unlock.link.remove();
|
|
manager.destroy.link.remove();
|
|
|
|
manager.lock = null;
|
|
if (manager.state == .waiting_for_lock_surfaces) {
|
|
manager.state = .waiting_for_blank;
|
|
manager.lock_surfaces_timer.timerUpdate(0) catch {};
|
|
}
|
|
}
|
|
|
|
fn handleSurface(
|
|
listener: *wl.Listener(*wlr.SessionLockSurfaceV1),
|
|
wlr_lock_surface: *wlr.SessionLockSurfaceV1,
|
|
) void {
|
|
const manager = @fieldParentPtr(LockManager, "new_surface", listener);
|
|
|
|
log.debug("new ext_session_lock_surface_v1 created", .{});
|
|
|
|
assert(manager.state != .unlocked);
|
|
assert(manager.lock != null);
|
|
|
|
LockSurface.create(wlr_lock_surface, manager.lock.?) catch {
|
|
log.err("out of memory", .{});
|
|
wlr_lock_surface.resource.postNoMemory();
|
|
};
|
|
}
|