mirror of
https://git.robbyzambito.me/zits
synced 2026-02-04 03:34:48 +00:00
375 lines
12 KiB
Zig
375 lines
12 KiB
Zig
const std = @import("std");
|
|
const Allocator = std.mem.Allocator;
|
|
const ArrayList = std.ArrayList;
|
|
const AutoHashMapUnmanaged = std.AutoHashMapUnmanaged;
|
|
|
|
const Io = std.Io;
|
|
const Dir = Io.Dir;
|
|
const Group = Io.Group;
|
|
const IpAddress = std.Io.net.IpAddress;
|
|
const Mutex = Io.Mutex;
|
|
const Queue = Io.Queue;
|
|
const Stream = std.Io.net.Stream;
|
|
|
|
const message_parser = @import("./message_parser.zig");
|
|
pub const MessageType = message_parser.MessageType;
|
|
pub const Message = message_parser.Message;
|
|
const ServerInfo = Message.ServerInfo;
|
|
pub const Client = @import("./Client.zig");
|
|
const Msgs = Client.Msgs;
|
|
const Server = @This();
|
|
|
|
pub const Subscription = struct {
|
|
subject: []const u8,
|
|
client_id: usize,
|
|
sid: []const u8,
|
|
queue: *Queue(Msgs),
|
|
|
|
fn deinit(self: Subscription, alloc: Allocator) void {
|
|
alloc.free(self.subject);
|
|
alloc.free(self.sid);
|
|
}
|
|
};
|
|
|
|
const eql = std.mem.eql;
|
|
const log = std.log;
|
|
const panic = std.debug.panic;
|
|
|
|
info: ServerInfo,
|
|
clients: AutoHashMapUnmanaged(usize, *Client) = .empty,
|
|
|
|
subs_lock: Mutex = .init,
|
|
subscriptions: ArrayList(Subscription) = .empty,
|
|
|
|
pub fn deinit(server: *Server, io: Io, alloc: Allocator) void {
|
|
server.subs_lock.lockUncancelable(io);
|
|
defer server.subs_lock.unlock(io);
|
|
for (server.subscriptions.items) |sub| {
|
|
sub.deinit(alloc);
|
|
}
|
|
server.subscriptions.deinit(alloc);
|
|
server.clients.deinit(alloc);
|
|
}
|
|
|
|
pub fn start(server: *Server, io: Io, gpa: Allocator) !void {
|
|
var tcp_server = try IpAddress.listen(try IpAddress.parse(
|
|
server.info.host,
|
|
server.info.port,
|
|
), io, .{ .reuse_address = true });
|
|
defer tcp_server.deinit(io);
|
|
log.debug("Server headers: {s}", .{if (server.info.headers) "true" else "false"});
|
|
log.debug("Server max payload: {d}", .{server.info.max_payload});
|
|
log.info("Server ID: {s}", .{server.info.server_id});
|
|
log.info("Server name: {s}", .{server.info.server_name});
|
|
log.info("Server listening on {s}:{d}", .{ server.info.host, server.info.port });
|
|
|
|
var client_group: Group = .init;
|
|
defer client_group.cancel(io);
|
|
|
|
const read_buffer_size, const write_buffer_size = getBufferSizes(io);
|
|
log.debug("read buf: {d} write buf: {d}", .{ read_buffer_size, write_buffer_size });
|
|
|
|
var id: usize = 0;
|
|
while (true) : (id +%= 1) {
|
|
if (server.clients.contains(id)) continue;
|
|
log.debug("Accepting next client", .{});
|
|
const stream = try tcp_server.accept(io);
|
|
log.debug("Accepted connection {d}", .{id});
|
|
_ = client_group.concurrent(io, handleConnectionInfallible, .{
|
|
server,
|
|
gpa,
|
|
io,
|
|
id,
|
|
stream,
|
|
read_buffer_size,
|
|
write_buffer_size,
|
|
}) catch {
|
|
log.err("Could not start concurrent handler for {d}", .{id});
|
|
stream.close(io);
|
|
};
|
|
}
|
|
}
|
|
|
|
fn addClient(server: *Server, allocator: Allocator, id: usize, client: *Client) !void {
|
|
try server.clients.put(allocator, id, client);
|
|
}
|
|
|
|
fn removeClient(server: *Server, io: Io, allocator: Allocator, id: usize) void {
|
|
server.subs_lock.lockUncancelable(io);
|
|
defer server.subs_lock.unlock(io);
|
|
if (server.clients.remove(id)) {
|
|
const len = server.subscriptions.items.len;
|
|
for (0..len) |from_end| {
|
|
const i = len - from_end - 1;
|
|
const sub = server.subscriptions.items[i];
|
|
if (sub.client_id == id) {
|
|
sub.deinit(allocator);
|
|
_ = server.subscriptions.swapRemove(i);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
fn handleConnectionInfallible(
|
|
server: *Server,
|
|
server_allocator: Allocator,
|
|
io: Io,
|
|
id: usize,
|
|
stream: Stream,
|
|
r_buf_size: usize,
|
|
w_buf_size: usize,
|
|
) void {
|
|
handleConnection(server, server_allocator, io, id, stream, r_buf_size, w_buf_size) catch |err| {
|
|
log.err("Failed processing client {d}: {any}", .{ id, err });
|
|
};
|
|
}
|
|
|
|
fn handleConnection(
|
|
server: *Server,
|
|
server_allocator: Allocator,
|
|
io: Io,
|
|
id: usize,
|
|
stream: Stream,
|
|
r_buf_size: usize,
|
|
w_buf_size: usize,
|
|
) !void {
|
|
defer stream.close(io);
|
|
|
|
// TODO: use a client allocator for things that should only live for as long as the client?
|
|
// I had this before, but it seemed to have made lifetimes harder to track.
|
|
// Messages made sense to parse using a client allocator, but that makes it hard to free
|
|
// messages when done processing them (usually outside the client process, ie: publish).
|
|
|
|
// Set up client writer
|
|
const w_buffer: []u8 = try server_allocator.alloc(u8, w_buf_size);
|
|
defer server_allocator.free(w_buffer);
|
|
var writer = stream.writer(io, w_buffer);
|
|
const out = &writer.interface;
|
|
|
|
// Set up client reader
|
|
const r_buffer: []u8 = try server_allocator.alloc(u8, r_buf_size);
|
|
defer server_allocator.free(r_buffer);
|
|
var reader = stream.reader(io, r_buffer);
|
|
const in = &reader.interface;
|
|
|
|
// Set up buffer queue
|
|
const qbuf: []Message = try server_allocator.alloc(Message, 16);
|
|
defer server_allocator.free(qbuf);
|
|
var recv_queue: Queue(Message) = .init(qbuf);
|
|
defer recv_queue.close(io);
|
|
|
|
const mbuf: []Msgs = try server_allocator.alloc(Msgs, w_buf_size / (@sizeOf(Msgs) + 128));
|
|
defer server_allocator.free(mbuf);
|
|
var msgs_queue: Queue(Msgs) = .init(mbuf);
|
|
defer {
|
|
msgs_queue.close(io);
|
|
while (msgs_queue.getOne(io)) |msg| {
|
|
switch (msg) {
|
|
.MSG => |m| m.deinit(server_allocator),
|
|
.HMSG => |h| h.deinit(server_allocator),
|
|
}
|
|
} else |_| {}
|
|
}
|
|
|
|
// Create client
|
|
var client: Client = .init(null, &recv_queue, &msgs_queue, in, out);
|
|
defer client.deinit(server_allocator);
|
|
|
|
try server.addClient(server_allocator, id, &client);
|
|
defer server.removeClient(io, server_allocator, id);
|
|
|
|
// Do initial handshake with client
|
|
// try recv_queue.putOne(io, .PONG);
|
|
try recv_queue.putOne(io, .{ .INFO = server.info });
|
|
|
|
var client_task = try io.concurrent(Client.start, .{ &client, io, server_allocator });
|
|
defer client_task.cancel(io) catch {};
|
|
|
|
// Messages are owned by the server after they are received from the client
|
|
while (client.next(server_allocator)) |msg| {
|
|
switch (msg) {
|
|
.PING => {
|
|
// Respond to ping with pong.
|
|
try client.send(io, .PONG);
|
|
},
|
|
.PUB => |pb| {
|
|
@branchHint(.likely);
|
|
defer pb.deinit(server_allocator);
|
|
try server.publishMessage(io, server_allocator, &client, msg);
|
|
},
|
|
.HPUB => |hp| {
|
|
@branchHint(.likely);
|
|
defer hp.deinit(server_allocator);
|
|
try server.publishMessage(io, server_allocator, &client, msg);
|
|
},
|
|
.SUB => |sub| {
|
|
defer sub.deinit(server_allocator);
|
|
try server.subscribe(io, server_allocator, id, &msgs_queue, sub);
|
|
},
|
|
.UNSUB => |unsub| {
|
|
defer unsub.deinit(server_allocator);
|
|
try server.unsubscribe(io, server_allocator, id, unsub);
|
|
},
|
|
.CONNECT => |connect| {
|
|
if (client.connect) |*current| {
|
|
current.deinit(server_allocator);
|
|
}
|
|
client.connect = connect;
|
|
},
|
|
else => |e| {
|
|
panic("Unimplemented message: {any}\n", .{e});
|
|
},
|
|
}
|
|
} else |err| switch (err) {
|
|
error.EndOfStream => {
|
|
log.debug("Client {d} disconnected", .{id});
|
|
},
|
|
else => {
|
|
return err;
|
|
},
|
|
}
|
|
}
|
|
|
|
fn subjectMatches(sub_subject: []const u8, pub_subject: []const u8) bool {
|
|
// TODO: assert that sub_subject and pub_subject are valid.
|
|
var sub_iter = std.mem.splitScalar(u8, sub_subject, '.');
|
|
var pub_iter = std.mem.splitScalar(u8, pub_subject, '.');
|
|
|
|
while (sub_iter.next()) |st| {
|
|
const pt = pub_iter.next() orelse return false;
|
|
|
|
if (eql(u8, st, ">")) return true;
|
|
|
|
if (!eql(u8, st, "*") and !eql(u8, st, pt)) {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
return pub_iter.next() == null;
|
|
}
|
|
|
|
test subjectMatches {
|
|
const expect = std.testing.expect;
|
|
try expect(subjectMatches("foo", "foo"));
|
|
try expect(!subjectMatches("foo", "bar"));
|
|
|
|
try expect(subjectMatches("foo.*", "foo.bar"));
|
|
try expect(!subjectMatches("foo.*", "foo"));
|
|
try expect(!subjectMatches("foo.>", "foo"));
|
|
|
|
// the wildcard subscriptions foo.*.quux and foo.> both match foo.bar.quux, but only the latter matches foo.bar.baz.
|
|
try expect(subjectMatches("foo.*.quux", "foo.bar.quux"));
|
|
try expect(subjectMatches("foo.>", "foo.bar.quux"));
|
|
try expect(!subjectMatches("foo.*.quux", "foo.bar.baz"));
|
|
try expect(subjectMatches("foo.>", "foo.bar.baz"));
|
|
}
|
|
|
|
fn publishMessage(server: *Server, io: Io, alloc: Allocator, source_client: *Client, msg: Message) !void {
|
|
defer if (source_client.connect) |c| {
|
|
if (c.verbose) {
|
|
source_client.send(io, .@"+OK") catch {};
|
|
}
|
|
};
|
|
|
|
const subject = switch (msg) {
|
|
.PUB => |pb| pb.subject,
|
|
.HPUB => |hp| hp.@"pub".subject,
|
|
else => unreachable,
|
|
};
|
|
try server.subs_lock.lock(io);
|
|
defer server.subs_lock.unlock(io);
|
|
for (server.subscriptions.items) |subscription| {
|
|
if (subjectMatches(subscription.subject, subject)) {
|
|
switch (msg) {
|
|
.PUB => |pb| {
|
|
try subscription.queue.putOne(io, .{
|
|
.MSG = try pb.toMsg(alloc, subscription.sid),
|
|
});
|
|
},
|
|
.HPUB => |hp| {
|
|
try subscription.queue.putOne(io, .{
|
|
.HMSG = try hp.toHMsg(alloc, subscription.sid),
|
|
});
|
|
},
|
|
else => unreachable,
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
fn subscribe(
|
|
server: *Server,
|
|
io: Io,
|
|
gpa: Allocator,
|
|
id: usize,
|
|
queue: *Queue(Msgs),
|
|
msg: Message.Sub,
|
|
) !void {
|
|
try server.subs_lock.lock(io);
|
|
defer server.subs_lock.unlock(io);
|
|
const subject = try gpa.dupe(u8, msg.subject);
|
|
errdefer gpa.free(subject);
|
|
const sid = try gpa.dupe(u8, msg.sid);
|
|
errdefer gpa.free(sid);
|
|
try server.subscriptions.append(gpa, .{
|
|
.subject = subject,
|
|
.client_id = id,
|
|
.sid = sid,
|
|
.queue = queue,
|
|
});
|
|
}
|
|
|
|
fn unsubscribe(
|
|
server: *Server,
|
|
io: Io,
|
|
gpa: Allocator,
|
|
id: usize,
|
|
msg: Message.Unsub,
|
|
) !void {
|
|
try server.subs_lock.lock(io);
|
|
defer server.subs_lock.unlock(io);
|
|
const len = server.subscriptions.items.len;
|
|
for (0..len) |from_end| {
|
|
const i = len - from_end - 1;
|
|
const sub = server.subscriptions.items[i];
|
|
if (sub.client_id == id and eql(u8, sub.sid, msg.sid)) {
|
|
sub.deinit(gpa);
|
|
_ = server.subscriptions.swapRemove(i);
|
|
}
|
|
}
|
|
}
|
|
|
|
const parseUnsigned = std.fmt.parseUnsigned;
|
|
|
|
fn getBufferSizes(io: Io) struct { usize, usize } {
|
|
const default_size = 4 * 1024;
|
|
const default = .{ default_size, default_size };
|
|
|
|
const dir = Dir.openDirAbsolute(io, "/proc/sys/net/core", .{}) catch {
|
|
log.warn("couldn't open /proc/sys/net/core", .{});
|
|
return default;
|
|
};
|
|
|
|
var buf: [64]u8 = undefined;
|
|
|
|
const rmem_max = readBufferSize(io, dir, "rmem_max", &buf, default_size);
|
|
const wmem_max = readBufferSize(io, dir, "wmem_max", &buf, default_size);
|
|
|
|
return .{ rmem_max, wmem_max };
|
|
}
|
|
|
|
fn readBufferSize(io: Io, dir: anytype, filename: []const u8, buf: []u8, default: usize) usize {
|
|
const bytes = dir.readFile(io, filename, buf) catch |err| {
|
|
log.err("couldn't open {s}: {any}", .{ filename, err });
|
|
return default;
|
|
};
|
|
|
|
return parseUnsigned(usize, bytes[0 .. bytes.len - 1], 10) catch |err| {
|
|
log.err("couldn't parse {s}: {any}", .{ bytes[0 .. bytes.len - 1], err });
|
|
return default;
|
|
};
|
|
}
|
|
|
|
pub const default_id = "server-id-123";
|
|
pub const default_name = "Zits Server";
|