Compare commits

..

No commits in common. "64b13ad9f260e88b7f6bbf401a570fda37f98c56" and "501c3466bc0d44fa49c07284765463f28ff5b4f3" have entirely different histories.

30 changed files with 766 additions and 816 deletions

View file

@ -1,5 +0,0 @@
# Run this command to always ignore formatting commits in `git blame`
# git config blame.ignoreRevsFile .git-blame-ignore-revs
# Switch to nixpkgs-fmt formatting
fd138d45e6a2cad89fead6e9f246ba282070d6b7

View file

@ -1,9 +1,10 @@
{ config {
, pkgs config,
, lib pkgs,
, modulesPath lib,
, flake-inputs modulesPath,
, ... flake-inputs,
...
}: { }: {
imports = [ imports = [
flake-inputs.disko.nixosModules.disko flake-inputs.disko.nixosModules.disko
@ -46,15 +47,15 @@
''; '';
# Enable remote builds from tlater # Enable remote builds from tlater
settings.trusted-users = [ "@wheel" ]; settings.trusted-users = ["@wheel"];
}; };
nixpkgs.config.allowUnfreePredicate = pkg: nixpkgs.config.allowUnfreePredicate = pkg:
builtins.elem (lib.getName pkg) [ "steam-original" "steam-runtime" "steam-run" "steamcmd" ]; builtins.elem (lib.getName pkg) ["steam-original" "steam-runtime" "steam-run" "steamcmd"];
# Optimization for minecraft servers, see: # Optimization for minecraft servers, see:
# https://bugs.mojang.com/browse/MC-183518 # https://bugs.mojang.com/browse/MC-183518
boot.kernelParams = [ "highres=off" "nohz=off" ]; boot.kernelParams = ["highres=off" "nohz=off"];
networking = { networking = {
usePredictableInterfaceNames = false; usePredictableInterfaceNames = false;
@ -105,15 +106,15 @@
users.users.tlater = { users.users.tlater = {
isNormalUser = true; isNormalUser = true;
extraGroups = [ "wheel" ]; extraGroups = ["wheel"];
openssh.authorizedKeys.keyFiles = [ ../keys/tlater.pub ]; openssh.authorizedKeys.keyFiles = [../keys/tlater.pub];
}; };
services = { services = {
openssh = { openssh = {
enable = true; enable = true;
allowSFTP = false; allowSFTP = false;
ports = [ 2222 ]; ports = [2222];
startWhenNeeded = true; startWhenNeeded = true;
settings = { settings = {
@ -132,14 +133,14 @@
pam = { pam = {
sshAgentAuth = { sshAgentAuth = {
enable = true; enable = true;
authorizedKeysFiles = [ "/etc/ssh/authorized_keys.d/%u" ]; authorizedKeysFiles = ["/etc/ssh/authorized_keys.d/%u"];
}; };
services.sudo.sshAgentAuth = true; services.sudo.sshAgentAuth = true;
}; };
}; };
# Remove some unneeded packages # Remove some unneeded packages
environment.defaultPackages = [ ]; environment.defaultPackages = [];
system.stateVersion = "20.09"; system.stateVersion = "20.09";
} }

View file

@ -8,7 +8,7 @@
# disables it by default. # disables it by default.
# #
# TODO(tlater): See if would be useful for anything? # TODO(tlater): See if would be useful for anything?
boot.kernelParams = [ "nosgx" ]; boot.kernelParams = ["nosgx"];
networking.hostName = "hetzner-1"; networking.hostName = "hetzner-1";
services.nginx.domain = "tlater.net"; services.nginx.domain = "tlater.net";

View file

@ -1,84 +1,82 @@
{ {
disko.devices.disk = disko.devices.disk = let
let bootPartition = {
bootPartition = { size = "1M";
size = "1M"; type = "EF02";
type = "EF02"; };
swapPartition = {
# 8G is apparently recommended for this much RAM, but we set up
# 4G on both disks for mirroring purposes.
#
# That'll still be 8G during normal operation, and it's probably
# not too bad to have slightly less swap if a disk dies.
size = "4G";
content = {
type = "swap";
randomEncryption = true;
}; };
};
swapPartition = { mountOptions = ["compress=zstd" "noatime"];
# 8G is apparently recommended for this much RAM, but we set up in {
# 4G on both disks for mirroring purposes. sda = {
# type = "disk";
# That'll still be 8G during normal operation, and it's probably device = "/dev/sda";
# not too bad to have slightly less swap if a disk dies. content = {
size = "4G"; type = "gpt";
content = { partitions = {
type = "swap"; boot = bootPartition;
randomEncryption = true; swap = swapPartition;
};
};
mountOptions = [ "compress=zstd" "noatime" ]; disk1 = {
in size = "100%";
{ # Empty partition to combine in RAID0 with the other disk
sda = {
type = "disk";
device = "/dev/sda";
content = {
type = "gpt";
partitions = {
boot = bootPartition;
swap = swapPartition;
disk1 = {
size = "100%";
# Empty partition to combine in RAID0 with the other disk
};
}; };
}; };
}; };
};
sdb = { sdb = {
type = "disk"; type = "disk";
device = "/dev/sdb"; device = "/dev/sdb";
content = { content = {
type = "gpt"; type = "gpt";
partitions = { partitions = {
boot = bootPartition; boot = bootPartition;
swap = swapPartition; swap = swapPartition;
disk2 = { disk2 = {
size = "100%"; size = "100%";
content = { content = {
type = "btrfs"; type = "btrfs";
# Hack to get multi-device btrfs going # Hack to get multi-device btrfs going
# See https://github.com/nix-community/disko/issues/99 # See https://github.com/nix-community/disko/issues/99
extraArgs = [ "-d" "raid1" "-m" "raid1" "--runtime-features" "quota" "/dev/sda3" ]; extraArgs = ["-d" "raid1" "-m" "raid1" "--runtime-features" "quota" "/dev/sda3"];
subvolumes = { subvolumes = {
"/volume" = { }; "/volume" = {};
"/volume/root" = { "/volume/root" = {
inherit mountOptions; inherit mountOptions;
mountpoint = "/"; mountpoint = "/";
};
"/volume/home" = {
inherit mountOptions;
mountpoint = "/home";
};
"/volume/var" = {
inherit mountOptions;
mountpoint = "/var";
};
"/volume/nix-store" = {
inherit mountOptions;
mountpoint = "/nix";
};
"/snapshots" = { };
}; };
"/volume/home" = {
inherit mountOptions;
mountpoint = "/home";
};
"/volume/var" = {
inherit mountOptions;
mountpoint = "/var";
};
"/volume/nix-store" = {
inherit mountOptions;
mountpoint = "/nix";
};
"/snapshots" = {};
}; };
}; };
}; };
}; };
}; };
}; };
};
} }

View file

@ -1,8 +1,8 @@
{ lib, ... }: { {lib, ...}: {
users.users.tlater.password = "insecure"; users.users.tlater.password = "insecure";
# Disable graphical tty so -curses works # Disable graphical tty so -curses works
boot.kernelParams = [ "nomodeset" ]; boot.kernelParams = ["nomodeset"];
networking.hostName = "testvm"; networking.hostName = "testvm";
# Sets the base domain for nginx to a local domain so that we can # Sets the base domain for nginx to a local domain so that we can

View file

@ -1,6 +1,7 @@
{ config {
, lib config,
, ... lib,
...
}: { }: {
services.nginx = { services.nginx = {
enable = true; enable = true;
@ -26,33 +27,31 @@
# Override the default, just keep fewer logs # Override the default, just keep fewer logs
nginx.rotate = 6; nginx.rotate = 6;
} }
// lib.mapAttrs' // lib.mapAttrs' (virtualHost: _:
(virtualHost: _: lib.nameValuePair "/var/log/nginx/${virtualHost}/access.log" {
lib.nameValuePair "/var/log/nginx/${virtualHost}/access.log" { frequency = "daily";
frequency = "daily"; rotate = 2;
rotate = 2; compress = true;
compress = true; delaycompress = true;
delaycompress = true; su = "${config.services.nginx.user} ${config.services.nginx.group}";
su = "${config.services.nginx.user} ${config.services.nginx.group}"; postrotate = "[ ! -f /var/run/nginx/nginx.pid ] || kill -USR1 `cat /var/run/nginx/nginx.pid`";
postrotate = "[ ! -f /var/run/nginx/nginx.pid ] || kill -USR1 `cat /var/run/nginx/nginx.pid`"; })
}) config.services.nginx.virtualHosts;
config.services.nginx.virtualHosts;
systemd.tmpfiles.rules = systemd.tmpfiles.rules =
lib.mapAttrsToList lib.mapAttrsToList (
( virtualHost: _:
virtualHost: _: #
# "d /var/log/nginx/${virtualHost} 0750 ${config.services.nginx.user} ${config.services.nginx.group}"
"d /var/log/nginx/${virtualHost} 0750 ${config.services.nginx.user} ${config.services.nginx.group}" )
) config.services.nginx.virtualHosts;
config.services.nginx.virtualHosts;
security.acme = { security.acme = {
defaults.email = "tm@tlater.net"; defaults.email = "tm@tlater.net";
acceptTerms = true; acceptTerms = true;
certs."tlater.net" = { certs."tlater.net" = {
extraDomainNames = [ "*.tlater.net" ]; extraDomainNames = ["*.tlater.net"];
dnsProvider = "hetzner"; dnsProvider = "hetzner";
group = "nginx"; group = "nginx";
credentialFiles."HETZNER_API_KEY_FILE" = config.sops.secrets."hetzner-api".path; credentialFiles."HETZNER_API_KEY_FILE" = config.sops.secrets."hetzner-api".path;
@ -63,6 +62,6 @@
user = "acme"; user = "acme";
paths = paths =
lib.mapAttrsToList (virtualHost: _: "/var/lib/acme/${virtualHost}") lib.mapAttrsToList (virtualHost: _: "/var/lib/acme/${virtualHost}")
config.services.nginx.virtualHosts; config.services.nginx.virtualHosts;
}; };
} }

View file

@ -1,11 +1,12 @@
{ pkgs {
, config pkgs,
, ... config,
...
}: { }: {
systemd.services.afvalcalendar = { systemd.services.afvalcalendar = {
description = "Enschede afvalcalendar -> ical converter"; description = "Enschede afvalcalendar -> ical converter";
wantedBy = [ "multi-user.target" ]; wantedBy = ["multi-user.target"];
after = [ "network.target" ]; after = ["network.target"];
script = '' script = ''
${pkgs.local.afvalcalendar}/bin/afvalcalendar > /srv/afvalcalendar/afvalcalendar.ical ${pkgs.local.afvalcalendar}/bin/afvalcalendar > /srv/afvalcalendar/afvalcalendar.ical
@ -25,14 +26,14 @@
ProtectKernelModules = true; ProtectKernelModules = true;
ProtectKernelLogs = true; ProtectKernelLogs = true;
ProtectControlGroups = true; ProtectControlGroups = true;
RestrictAddressFamilies = [ "AF_UNIX" "AF_INET" "AF_INET6" ]; RestrictAddressFamilies = ["AF_UNIX" "AF_INET" "AF_INET6"];
RestrictNamespaces = true; RestrictNamespaces = true;
LockPersonality = true; LockPersonality = true;
MemoryDenyWriteExecute = true; MemoryDenyWriteExecute = true;
RestrictRealtime = true; RestrictRealtime = true;
RestrictSUIDSGID = true; RestrictSUIDSGID = true;
SystemCallArchitectures = "native"; SystemCallArchitectures = "native";
SystemCallFilter = [ "@system-service" "~@privileged @resources @setuid @keyring" ]; SystemCallFilter = ["@system-service" "~@privileged @resources @setuid @keyring"];
Umask = 0002; Umask = 0002;
SupplementaryGroups = "afvalcalendar-hosting"; SupplementaryGroups = "afvalcalendar-hosting";
@ -49,7 +50,7 @@
root = "/srv/afvalcalendar"; root = "/srv/afvalcalendar";
}; };
users.groups.afvalcalendar-hosting = { }; users.groups.afvalcalendar-hosting = {};
systemd.tmpfiles.settings."10-afvalcalendar" = { systemd.tmpfiles.settings."10-afvalcalendar" = {
"/srv/afvalcalendar".d = { "/srv/afvalcalendar".d = {
user = "nginx"; user = "nginx";

View file

@ -1,9 +1,9 @@
{ config {
, pkgs config,
, lib pkgs,
, ... lib,
}: ...
let }: let
inherit (lib) types optional singleton; inherit (lib) types optional singleton;
mkShutdownScript = service: mkShutdownScript = service:
pkgs.writeShellScript "backup-${service}-shutdown" '' pkgs.writeShellScript "backup-${service}-shutdown" ''
@ -42,17 +42,17 @@ let
RESTIC_REPOSITORY = "rclone:storagebox:backups"; RESTIC_REPOSITORY = "rclone:storagebox:backups";
RCLONE_CONFIG = rcloneConfig; RCLONE_CONFIG = rcloneConfig;
}; };
in in {
{
options = { options = {
services.backups = lib.mkOption { services.backups = lib.mkOption {
description = lib.mdDoc '' description = lib.mdDoc ''
Configure restic backups with a specific tag. Configure restic backups with a specific tag.
''; '';
type = types.attrsOf (types.submodule ({ config type = types.attrsOf (types.submodule ({
, name config,
, ... name,
}: { ...
}: {
options = { options = {
user = lib.mkOption { user = lib.mkOption {
type = types.str; type = types.str;
@ -76,7 +76,7 @@ in
preparation = { preparation = {
packages = lib.mkOption { packages = lib.mkOption {
type = types.listOf types.package; type = types.listOf types.package;
default = [ ]; default = [];
description = '' description = ''
The list of packages to make available in the The list of packages to make available in the
preparation script. preparation script.
@ -97,7 +97,7 @@ in
cleanup = { cleanup = {
packages = lib.mkOption { packages = lib.mkOption {
type = types.listOf types.package; type = types.listOf types.package;
default = [ ]; default = [];
description = '' description = ''
The list of packages to make available in the The list of packages to make available in the
cleanup script. cleanup script.
@ -116,7 +116,7 @@ in
}; };
pauseServices = lib.mkOption { pauseServices = lib.mkOption {
type = types.listOf types.str; type = types.listOf types.str;
default = [ ]; default = [];
description = '' description = ''
The systemd services that need to be shut down before The systemd services that need to be shut down before
the backup can run. Services will be restarted after the the backup can run. Services will be restarted after the
@ -131,7 +131,7 @@ in
}; };
}; };
config = lib.mkIf (config.services.backups != { }) { config = lib.mkIf (config.services.backups != {}) {
systemd.services = systemd.services =
{ {
restic-prune = { restic-prune = {
@ -164,81 +164,79 @@ in
}; };
}; };
} }
// lib.mapAttrs' // lib.mapAttrs' (name: backup:
(name: backup: lib.nameValuePair "backup-${name}" {
lib.nameValuePair "backup-${name}" { # Don't want to restart mid-backup
# Don't want to restart mid-backup restartIfChanged = false;
restartIfChanged = false;
environment = environment =
resticEnv resticEnv
// { // {
RESTIC_CACHE_DIR = "%C/backup-${name}"; RESTIC_CACHE_DIR = "%C/backup-${name}";
}; };
path = with pkgs; [ path = with pkgs; [
coreutils coreutils
openssh openssh
rclone rclone
restic restic
];
# TODO(tlater): If I ever add more than one repo, service
# shutdown/restarting will potentially break if multiple
# backups for the same service overlap. A more clever
# sentinel file with reference counts would probably solve
# this.
serviceConfig = {
User = backup.user;
Group = "backup";
RuntimeDirectory = "backup-${name}";
CacheDirectory = "backup-${name}";
CacheDirectoryMode = "0700";
PrivateTmp = true;
ExecStart = [
(lib.concatStringsSep " " (["${pkgs.restic}/bin/restic" "backup" "--tag" name] ++ backup.paths))
]; ];
# TODO(tlater): If I ever add more than one repo, service ExecStartPre =
# shutdown/restarting will potentially break if multiple map (service: "+${mkShutdownScript service}") backup.pauseServices
# backups for the same service overlap. A more clever ++ singleton (writeScript "backup-${name}-repo-init" [] ''
# sentinel file with reference counts would probably solve restic snapshots || restic init
# this. '')
serviceConfig = { ++ optional (backup.preparation.text != null)
User = backup.user; (writeScript "backup-${name}-prepare" backup.preparation.packages backup.preparation.text);
Group = "backup";
RuntimeDirectory = "backup-${name}";
CacheDirectory = "backup-${name}";
CacheDirectoryMode = "0700";
PrivateTmp = true;
ExecStart = [ # TODO(tlater): Add repo pruning/checking
(lib.concatStringsSep " " ([ "${pkgs.restic}/bin/restic" "backup" "--tag" name ] ++ backup.paths)) ExecStopPost =
]; map (service: "+${mkRestartScript service}") backup.pauseServices
++ optional (backup.cleanup.text != null)
ExecStartPre = (writeScript "backup-${name}-cleanup" backup.cleanup.packages backup.cleanup.text);
map (service: "+${mkShutdownScript service}") backup.pauseServices };
++ singleton (writeScript "backup-${name}-repo-init" [ ] '' })
restic snapshots || restic init config.services.backups;
'')
++ optional (backup.preparation.text != null)
(writeScript "backup-${name}-prepare" backup.preparation.packages backup.preparation.text);
# TODO(tlater): Add repo pruning/checking
ExecStopPost =
map (service: "+${mkRestartScript service}") backup.pauseServices
++ optional (backup.cleanup.text != null)
(writeScript "backup-${name}-cleanup" backup.cleanup.packages backup.cleanup.text);
};
})
config.services.backups;
systemd.timers = systemd.timers =
{ {
restic-prune = { restic-prune = {
wantedBy = [ "timers.target" ]; wantedBy = ["timers.target"];
timerConfig.OnCalendar = "Thursday 03:00:00 UTC"; timerConfig.OnCalendar = "Thursday 03:00:00 UTC";
# Don't make this persistent, in case the server was offline # Don't make this persistent, in case the server was offline
# for a while. This job cannot run at the same time as any # for a while. This job cannot run at the same time as any
# of the backup jobs. # of the backup jobs.
}; };
} }
// lib.mapAttrs' // lib.mapAttrs' (name: backup:
(name: backup: lib.nameValuePair "backup-${name}" {
lib.nameValuePair "backup-${name}" { wantedBy = ["timers.target"];
wantedBy = [ "timers.target" ]; timerConfig = {
timerConfig = { OnCalendar = "Wednesday 02:30:00 UTC";
OnCalendar = "Wednesday 02:30:00 UTC"; RandomizedDelaySec = "1h";
RandomizedDelaySec = "1h"; FixedRandomDelay = true;
FixedRandomDelay = true; Persistent = true;
Persistent = true; };
}; })
}) config.services.backups;
config.services.backups;
users = { users = {
# This user is only used to own the ssh key, because apparently # This user is only used to own the ssh key, because apparently
@ -247,7 +245,7 @@ in
group = "backup"; group = "backup";
isSystemUser = true; isSystemUser = true;
}; };
groups.backup = { }; groups.backup = {};
}; };
}; };
} }

View file

@ -1,6 +1,7 @@
{ config {
, flake-inputs config,
, ... flake-inputs,
...
}: { }: {
imports = [ imports = [
flake-inputs.sonnenshift.nixosModules.default flake-inputs.sonnenshift.nixosModules.default

View file

@ -1,16 +1,15 @@
{ pkgs {
, config pkgs,
, lib config,
, ... lib,
}: ...
let }: let
inherit (lib.strings) concatMapStringsSep; inherit (lib.strings) concatMapStringsSep;
cfg = config.services.matrix-conduit; cfg = config.services.matrix-conduit;
domain = "matrix.${config.services.nginx.domain}"; domain = "matrix.${config.services.nginx.domain}";
turn-realm = "turn.${config.services.nginx.domain}"; turn-realm = "turn.${config.services.nginx.domain}";
in in {
{
services.matrix-conduit = { services.matrix-conduit = {
enable = true; enable = true;
settings.global = { settings.global = {
@ -18,102 +17,98 @@ in
server_name = domain; server_name = domain;
database_backend = "rocksdb"; database_backend = "rocksdb";
turn_uris = turn_uris = let
let address = "${config.services.coturn.realm}:${toString config.services.coturn.listening-port}";
address = "${config.services.coturn.realm}:${toString config.services.coturn.listening-port}"; tls-address = "${config.services.coturn.realm}:${toString config.services.coturn.tls-listening-port}";
tls-address = "${config.services.coturn.realm}:${toString config.services.coturn.tls-listening-port}"; in [
in "turn:${address}?transport=udp"
[ "turn:${address}?transport=tcp"
"turn:${address}?transport=udp" "turns:${tls-address}?transport=udp"
"turn:${address}?transport=tcp" "turns:${tls-address}?transport=tcp"
"turns:${tls-address}?transport=udp" ];
"turns:${tls-address}?transport=tcp"
];
}; };
}; };
systemd.services.heisenbridge = systemd.services.heisenbridge = let
let replaceSecretBin = "${pkgs.replace-secret}/bin/replace-secret";
replaceSecretBin = "${pkgs.replace-secret}/bin/replace-secret"; registrationFile = builtins.toFile "heisenbridge-registration.yaml" (builtins.toJSON {
registrationFile = builtins.toFile "heisenbridge-registration.yaml" (builtins.toJSON { id = "heisenbridge";
id = "heisenbridge"; url = "http://127.0.0.1:9898";
url = "http://127.0.0.1:9898"; as_token = "@AS_TOKEN@";
as_token = "@AS_TOKEN@"; hs_token = "@HS_TOKEN@";
hs_token = "@HS_TOKEN@"; rate_limited = false;
rate_limited = false; sender_localpart = "heisenbridge";
sender_localpart = "heisenbridge"; namespaces = {
namespaces = { users = [
users = [ {
{ regex = "@irc_.*";
regex = "@irc_.*"; exclusive = true;
exclusive = true; }
} {
{ regex = "@heisenbridge:.*";
regex = "@heisenbridge:.*"; exclusive = true;
exclusive = true; }
} ];
]; aliases = [];
aliases = [ ]; rooms = [];
rooms = [ ];
};
});
# TODO(tlater): Starting with systemd 253 it will become possible
# to do the credential setup as part of ExecStartPre/preStart
# instead.
#
# This will also make it possible to actually set caps on the
# heisenbridge process using systemd, so that we can run the
# identd process.
execScript = pkgs.writeShellScript "heisenbridge" ''
cp ${registrationFile} "$RUNTIME_DIRECTORY/heisenbridge-registration.yaml"
chmod 600 $RUNTIME_DIRECTORY/heisenbridge-registration.yaml
${replaceSecretBin} '@AS_TOKEN@' "$CREDENTIALS_DIRECTORY/heisenbridge_as-token" "$RUNTIME_DIRECTORY/heisenbridge-registration.yaml"
${replaceSecretBin} '@HS_TOKEN@' "$CREDENTIALS_DIRECTORY/heisenbridge_hs-token" "$RUNTIME_DIRECTORY/heisenbridge-registration.yaml"
chmod 400 $RUNTIME_DIRECTORY/heisenbridge-registration.yaml
${pkgs.heisenbridge}/bin/heisenbridge \
--config $RUNTIME_DIRECTORY/heisenbridge-registration.yaml \
--owner @tlater:matrix.tlater.net \
'http://localhost:${toString cfg.settings.global.port}'
'';
in
{
description = "Matrix<->IRC bridge";
wantedBy = [ "multi-user.target" ];
after = [ "conduit.service" ];
serviceConfig = {
Type = "simple";
LoadCredential = "heisenbridge:/run/secrets/heisenbridge";
ExecStart = execScript;
DynamicUser = true;
RuntimeDirectory = "heisenbridge";
RuntimeDirectoryMode = "0700";
RestrictNamespaces = true;
PrivateUsers = true;
ProtectHostname = true;
ProtectClock = true;
ProtectKernelTunables = true;
ProtectKernelModules = true;
ProtectKernelLogs = true;
ProtectControlGroups = true;
RestrictAddressFamilies = [ "AF_INET AF_INET6" ];
LockPersonality = true;
RestrictRealtime = true;
ProtectProc = "invisible";
ProcSubset = "pid";
UMask = 0077;
# For the identd port
# CapabilityBoundingSet = ["CAP_NET_BIND_SERVICE"];
# AmbientCapabilities = ["CAP_NET_BIND_SERVICE"];
}; };
});
# TODO(tlater): Starting with systemd 253 it will become possible
# to do the credential setup as part of ExecStartPre/preStart
# instead.
#
# This will also make it possible to actually set caps on the
# heisenbridge process using systemd, so that we can run the
# identd process.
execScript = pkgs.writeShellScript "heisenbridge" ''
cp ${registrationFile} "$RUNTIME_DIRECTORY/heisenbridge-registration.yaml"
chmod 600 $RUNTIME_DIRECTORY/heisenbridge-registration.yaml
${replaceSecretBin} '@AS_TOKEN@' "$CREDENTIALS_DIRECTORY/heisenbridge_as-token" "$RUNTIME_DIRECTORY/heisenbridge-registration.yaml"
${replaceSecretBin} '@HS_TOKEN@' "$CREDENTIALS_DIRECTORY/heisenbridge_hs-token" "$RUNTIME_DIRECTORY/heisenbridge-registration.yaml"
chmod 400 $RUNTIME_DIRECTORY/heisenbridge-registration.yaml
${pkgs.heisenbridge}/bin/heisenbridge \
--config $RUNTIME_DIRECTORY/heisenbridge-registration.yaml \
--owner @tlater:matrix.tlater.net \
'http://localhost:${toString cfg.settings.global.port}'
'';
in {
description = "Matrix<->IRC bridge";
wantedBy = ["multi-user.target"];
after = ["conduit.service"];
serviceConfig = {
Type = "simple";
LoadCredential = "heisenbridge:/run/secrets/heisenbridge";
ExecStart = execScript;
DynamicUser = true;
RuntimeDirectory = "heisenbridge";
RuntimeDirectoryMode = "0700";
RestrictNamespaces = true;
PrivateUsers = true;
ProtectHostname = true;
ProtectClock = true;
ProtectKernelTunables = true;
ProtectKernelModules = true;
ProtectKernelLogs = true;
ProtectControlGroups = true;
RestrictAddressFamilies = ["AF_INET AF_INET6"];
LockPersonality = true;
RestrictRealtime = true;
ProtectProc = "invisible";
ProcSubset = "pid";
UMask = 0077;
# For the identd port
# CapabilityBoundingSet = ["CAP_NET_BIND_SERVICE"];
# AmbientCapabilities = ["CAP_NET_BIND_SERVICE"];
}; };
};
# Pass in the TURN secret via EnvironmentFile, not supported by # Pass in the TURN secret via EnvironmentFile, not supported by
# upstream module currently. # upstream module currently.
@ -254,6 +249,6 @@ in
]; ];
# Other services store their data in conduit, so no other services # Other services store their data in conduit, so no other services
# need to be shut down currently. # need to be shut down currently.
pauseServices = [ "conduit.service" ]; pauseServices = ["conduit.service"];
}; };
} }

View file

@ -1,7 +1,7 @@
{ pkgs, ... }: { {pkgs, ...}: {
services.fail2ban = { services.fail2ban = {
enable = true; enable = true;
extraPackages = [ pkgs.ipset ]; extraPackages = [pkgs.ipset];
banaction = "iptables-ipset-proto6-allports"; banaction = "iptables-ipset-proto6-allports";
bantime-increment.enable = true; bantime-increment.enable = true;
@ -21,7 +21,7 @@
}; };
# Allow metrics services to connect to the socket as well # Allow metrics services to connect to the socket as well
users.groups.fail2ban = { }; users.groups.fail2ban = {};
systemd.services.fail2ban.serviceConfig = { systemd.services.fail2ban.serviceConfig = {
ExecStartPost = ExecStartPost =
"+" "+"

View file

@ -1,13 +1,12 @@
{ lib
, config
, flake-inputs
, ...
}:
let
domain = "foundryvtt.${config.services.nginx.domain}";
in
{ {
imports = [ flake-inputs.foundryvtt.nixosModules.foundryvtt ]; lib,
config,
flake-inputs,
...
}: let
domain = "foundryvtt.${config.services.nginx.domain}";
in {
imports = [flake-inputs.foundryvtt.nixosModules.foundryvtt];
services.foundryvtt = { services.foundryvtt = {
enable = true; enable = true;
@ -19,28 +18,26 @@ in
# Want to start it manually when I need it, not have it constantly # Want to start it manually when I need it, not have it constantly
# running # running
systemd.services.foundryvtt.wantedBy = lib.mkForce [ ]; systemd.services.foundryvtt.wantedBy = lib.mkForce [];
services.nginx.virtualHosts."${domain}" = services.nginx.virtualHosts."${domain}" = let
let inherit (config.services.foundryvtt) port;
inherit (config.services.foundryvtt) port; in {
in forceSSL = true;
{ useACMEHost = "tlater.net";
forceSSL = true; enableHSTS = true;
useACMEHost = "tlater.net";
enableHSTS = true;
locations."/" = { locations."/" = {
proxyWebsockets = true; proxyWebsockets = true;
proxyPass = "http://localhost:${toString port}"; proxyPass = "http://localhost:${toString port}";
};
}; };
};
services.backups.foundryvtt = { services.backups.foundryvtt = {
user = "foundryvtt"; user = "foundryvtt";
paths = [ paths = [
config.services.foundryvtt.dataDir config.services.foundryvtt.dataDir
]; ];
pauseServices = [ "foundryvtt.service" ]; pauseServices = ["foundryvtt.service"];
}; };
} }

View file

@ -1,12 +1,11 @@
{ pkgs
, config
, lib
, ...
}:
let
domain = "gitea.${config.services.nginx.domain}";
in
{ {
pkgs,
config,
lib,
...
}: let
domain = "gitea.${config.services.nginx.domain}";
in {
services.forgejo = { services.forgejo = {
enable = true; enable = true;
database.type = "postgres"; database.type = "postgres";
@ -28,37 +27,33 @@ in
}; };
}; };
systemd.services.forgejo.serviceConfig.ExecStartPre = systemd.services.forgejo.serviceConfig.ExecStartPre = let
let replaceSecretBin = "${pkgs.replace-secret}/bin/replace-secret";
replaceSecretBin = "${pkgs.replace-secret}/bin/replace-secret"; secretPath = config.sops.secrets."forgejo/metrics-token".path;
secretPath = config.sops.secrets."forgejo/metrics-token".path; runConfig = "${config.services.forgejo.customDir}/conf/app.ini";
runConfig = "${config.services.forgejo.customDir}/conf/app.ini"; in [
in "+${replaceSecretBin} '#metricstoken#' '${secretPath}' '${runConfig}'"
[ ];
"+${replaceSecretBin} '#metricstoken#' '${secretPath}' '${runConfig}'"
];
# Set up SSL # Set up SSL
services.nginx.virtualHosts."${domain}" = services.nginx.virtualHosts."${domain}" = let
let httpAddress = config.services.forgejo.settings.server.HTTP_ADDR;
httpAddress = config.services.forgejo.settings.server.HTTP_ADDR; httpPort = config.services.forgejo.settings.server.HTTP_PORT;
httpPort = config.services.forgejo.settings.server.HTTP_PORT; in {
in forceSSL = true;
{ useACMEHost = "tlater.net";
forceSSL = true; enableHSTS = true;
useACMEHost = "tlater.net";
enableHSTS = true;
locations."/".proxyPass = "http://${httpAddress}:${toString httpPort}"; locations."/".proxyPass = "http://${httpAddress}:${toString httpPort}";
locations."/metrics" = { locations."/metrics" = {
extraConfig = '' extraConfig = ''
access_log off; access_log off;
allow 127.0.0.1; allow 127.0.0.1;
${lib.optionalString config.networking.enableIPv6 "allow ::1;"} ${lib.optionalString config.networking.enableIPv6 "allow ::1;"}
deny all; deny all;
''; '';
};
}; };
};
# Block repeated failed login attempts # Block repeated failed login attempts
# #
@ -88,13 +83,13 @@ in
# Conf is backed up via nix # Conf is backed up via nix
]; ];
preparation = { preparation = {
packages = [ config.services.postgresql.package ]; packages = [config.services.postgresql.package];
text = "pg_dump ${config.services.forgejo.database.name} --file=/var/lib/forgejo/forgejo-db.sql"; text = "pg_dump ${config.services.forgejo.database.name} --file=/var/lib/forgejo/forgejo-db.sql";
}; };
cleanup = { cleanup = {
packages = [ pkgs.coreutils ]; packages = [pkgs.coreutils];
text = "rm /var/lib/forgejo/forgejo-db.sql"; text = "rm /var/lib/forgejo/forgejo-db.sql";
}; };
pauseServices = [ "forgejo.service" ]; pauseServices = ["forgejo.service"];
}; };
} }

View file

@ -1,28 +1,25 @@
{ config
, pkgs
, lib
, ...
}:
let
yaml = pkgs.formats.yaml { };
in
{ {
config,
pkgs,
lib,
...
}: let
yaml = pkgs.formats.yaml {};
in {
services.prometheus = { services.prometheus = {
exporters = { exporters = {
# Periodically check domain registration status # Periodically check domain registration status
domain = { domain = {
enable = true; enable = true;
listenAddress = "127.0.0.1"; listenAddress = "127.0.0.1";
extraFlags = extraFlags = let
let conf.domains = [
conf.domains = [ "tlater.net"
"tlater.net" "tlater.com"
"tlater.com"
];
in
[
"--config=${yaml.generate "domains.yml" conf}"
]; ];
in [
"--config=${yaml.generate "domains.yml" conf}"
];
}; };
# System statistics # System statistics
@ -52,50 +49,47 @@ in
group = "nginx"; group = "nginx";
settings.namespaces = settings.namespaces =
lib.mapAttrsToList lib.mapAttrsToList (name: virtualHost: {
(name: virtualHost: { inherit name;
inherit name; metrics_override.prefix = "nginxlog";
metrics_override.prefix = "nginxlog"; namespace_label = "vhost";
namespace_label = "vhost";
format = lib.concatStringsSep " " [ format = lib.concatStringsSep " " [
"$remote_addr - $remote_user [$time_local]" "$remote_addr - $remote_user [$time_local]"
''"$request" $status $body_bytes_sent'' ''"$request" $status $body_bytes_sent''
''"$http_referer" "$http_user_agent"'' ''"$http_referer" "$http_user_agent"''
''rt=$request_time uct="$upstream_connect_time"'' ''rt=$request_time uct="$upstream_connect_time"''
''uht="$upstream_header_time" urt="$upstream_response_time"'' ''uht="$upstream_header_time" urt="$upstream_response_time"''
]; ];
source.files = [ source.files = [
"/var/log/nginx/${name}/access.log" "/var/log/nginx/${name}/access.log"
]; ];
}) })
config.services.nginx.virtualHosts; config.services.nginx.virtualHosts;
}; };
}; };
extraExporters = { extraExporters = {
fail2ban = fail2ban = let
let cfg = config.services.prometheus.extraExporters.fail2ban;
cfg = config.services.prometheus.extraExporters.fail2ban; in {
in port = 9191;
{ serviceOpts = {
port = 9191; after = ["fail2ban.service"];
serviceOpts = { requires = ["fail2ban.service"];
after = [ "fail2ban.service" ]; serviceConfig = {
requires = [ "fail2ban.service" ]; Group = "fail2ban";
serviceConfig = { RestrictAddressFamilies = ["AF_UNIX" "AF_INET" "AF_INET6"];
Group = "fail2ban"; ExecStart = lib.concatStringsSep " " [
RestrictAddressFamilies = [ "AF_UNIX" "AF_INET" "AF_INET6" ]; "${pkgs.local.prometheus-fail2ban-exporter}/bin/fail2ban-prometheus-exporter"
ExecStart = lib.concatStringsSep " " [ "--collector.f2b.socket=/var/run/fail2ban/fail2ban.sock"
"${pkgs.local.prometheus-fail2ban-exporter}/bin/fail2ban-prometheus-exporter" "--web.listen-address='${cfg.listenAddress}:${toString cfg.port}'"
"--collector.f2b.socket=/var/run/fail2ban/fail2ban.sock" "--collector.f2b.exit-on-socket-connection-error=true"
"--web.listen-address='${cfg.listenAddress}:${toString cfg.port}'" ];
"--collector.f2b.exit-on-socket-connection-error=true"
];
};
}; };
}; };
};
}; };
# TODO(tlater): # TODO(tlater):

View file

@ -1,8 +1,6 @@
{ config, ... }: {config, ...}: let
let
domain = "metrics.${config.services.nginx.domain}"; domain = "metrics.${config.services.nginx.domain}";
in in {
{
services.grafana = { services.grafana = {
enable = true; enable = true;
settings = { settings = {

View file

@ -1,13 +1,12 @@
{ pkgs
, config
, lib
, ...
}:
let
inherit (lib) types mkOption mkDefault;
yaml = pkgs.formats.yaml { };
in
{ {
pkgs,
config,
lib,
...
}: let
inherit (lib) types mkOption mkDefault;
yaml = pkgs.formats.yaml {};
in {
options = { options = {
services.prometheus = { services.prometheus = {
extraExporters = mkOption { extraExporters = mkOption {
@ -32,10 +31,11 @@ in
}; };
services.victoriametrics.scrapeConfigs = mkOption { services.victoriametrics.scrapeConfigs = mkOption {
type = types.attrsOf (types.submodule ({ name type = types.attrsOf (types.submodule ({
, self name,
, ... self,
}: { ...
}: {
options = { options = {
job_name = mkOption { job_name = mkOption {
type = types.str; type = types.str;
@ -47,7 +47,7 @@ in
description = '' description = ''
Other settings to set for this scrape config. Other settings to set for this scrape config.
''; '';
default = { }; default = {};
}; };
targets = mkOption { targets = mkOption {
@ -57,11 +57,11 @@ in
Shortcut for `static_configs = lib.singleton {targets = [<targets>];}` Shortcut for `static_configs = lib.singleton {targets = [<targets>];}`
''; '';
default = [ ]; default = [];
}; };
static_configs = mkOption { static_configs = mkOption {
default = [ ]; default = [];
type = types.listOf (types.submodule { type = types.listOf (types.submodule {
options = { options = {
targets = mkOption { targets = mkOption {
@ -77,7 +77,7 @@ in
description = lib.mdDoc '' description = lib.mdDoc ''
Labels to apply to all targets defined for this static config. Labels to apply to all targets defined for this static config.
''; '';
default = { }; default = {};
}; };
}; };
}); });
@ -89,125 +89,116 @@ in
config = { config = {
systemd.services = lib.mkMerge [ systemd.services = lib.mkMerge [
(lib.mapAttrs' (lib.mapAttrs' (name: exporter:
(name: exporter: lib.nameValuePair "prometheus-${name}-exporter" (lib.mkMerge [
lib.nameValuePair "prometheus-${name}-exporter" (lib.mkMerge [ {
{ # Shamelessly copied from upstream because the upstream
# Shamelessly copied from upstream because the upstream # module is an intractable mess
# module is an intractable mess wantedBy = ["multi-user.target"];
wantedBy = [ "multi-user.target" ]; after = ["network.target"];
after = [ "network.target" ]; serviceConfig.Restart = mkDefault "always";
serviceConfig.Restart = mkDefault "always"; serviceConfig.PrivateTmp = mkDefault true;
serviceConfig.PrivateTmp = mkDefault true; serviceConfig.WorkingDirectory = mkDefault /tmp;
serviceConfig.WorkingDirectory = mkDefault /tmp; serviceConfig.DynamicUser = mkDefault true;
serviceConfig.DynamicUser = mkDefault true; # Hardening
# Hardening serviceConfig.CapabilityBoundingSet = mkDefault [""];
serviceConfig.CapabilityBoundingSet = mkDefault [ "" ]; serviceConfig.DeviceAllow = [""];
serviceConfig.DeviceAllow = [ "" ]; serviceConfig.LockPersonality = true;
serviceConfig.LockPersonality = true; serviceConfig.MemoryDenyWriteExecute = true;
serviceConfig.MemoryDenyWriteExecute = true; serviceConfig.NoNewPrivileges = true;
serviceConfig.NoNewPrivileges = true; serviceConfig.PrivateDevices = mkDefault true;
serviceConfig.PrivateDevices = mkDefault true; serviceConfig.ProtectClock = mkDefault true;
serviceConfig.ProtectClock = mkDefault true; serviceConfig.ProtectControlGroups = true;
serviceConfig.ProtectControlGroups = true; serviceConfig.ProtectHome = true;
serviceConfig.ProtectHome = true; serviceConfig.ProtectHostname = true;
serviceConfig.ProtectHostname = true; serviceConfig.ProtectKernelLogs = true;
serviceConfig.ProtectKernelLogs = true; serviceConfig.ProtectKernelModules = true;
serviceConfig.ProtectKernelModules = true; serviceConfig.ProtectKernelTunables = true;
serviceConfig.ProtectKernelTunables = true; serviceConfig.ProtectSystem = mkDefault "strict";
serviceConfig.ProtectSystem = mkDefault "strict"; serviceConfig.RemoveIPC = true;
serviceConfig.RemoveIPC = true; serviceConfig.RestrictAddressFamilies = ["AF_INET" "AF_INET6"];
serviceConfig.RestrictAddressFamilies = [ "AF_INET" "AF_INET6" ]; serviceConfig.RestrictNamespaces = true;
serviceConfig.RestrictNamespaces = true; serviceConfig.RestrictRealtime = true;
serviceConfig.RestrictRealtime = true; serviceConfig.RestrictSUIDSGID = true;
serviceConfig.RestrictSUIDSGID = true; serviceConfig.SystemCallArchitectures = "native";
serviceConfig.SystemCallArchitectures = "native"; serviceConfig.UMask = "0077";
serviceConfig.UMask = "0077"; }
} exporter.serviceOpts
exporter.serviceOpts ]))
])) config.services.prometheus.extraExporters)
config.services.prometheus.extraExporters)
{ {
vmagent-scrape-exporters = vmagent-scrape-exporters = let
let listenAddress = config.services.victoriametrics.listenAddress;
listenAddress = config.services.victoriametrics.listenAddress; vmAddr = (lib.optionalString (lib.hasPrefix ":" listenAddress) "127.0.0.1") + listenAddress;
vmAddr = (lib.optionalString (lib.hasPrefix ":" listenAddress) "127.0.0.1") + listenAddress; promscrape = yaml.generate "prometheus.yml" {
promscrape = yaml.generate "prometheus.yml" { scrape_configs = lib.mapAttrsToList (_: scrape:
scrape_configs = lib.mapAttrsToList lib.recursiveUpdate {
(_: scrape: inherit (scrape) job_name;
lib.recursiveUpdate static_configs =
{ scrape.static_configs
inherit (scrape) job_name; ++ lib.optional (scrape.targets != []) {targets = scrape.targets;};
static_configs = }
scrape.static_configs scrape.extraSettings)
++ lib.optional (scrape.targets != [ ]) { targets = scrape.targets; }; config.services.victoriametrics.scrapeConfigs;
}
scrape.extraSettings)
config.services.victoriametrics.scrapeConfigs;
};
in
{
enable = true;
path = [ pkgs.victoriametrics ];
wantedBy = [ "multi-user.target" ];
after = [ "network.target" "victoriametrics.service" ];
serviceConfig = {
ExecStart = [
(lib.concatStringsSep " " [
"${pkgs.victoriametrics}/bin/vmagent"
"-promscrape.config=${promscrape}"
"-remoteWrite.url=http://${vmAddr}/api/v1/write"
"-remoteWrite.tmpDataPath=%t/vmagent"
])
];
SupplementaryGroups = "metrics";
DynamicUser = true;
RuntimeDirectory = "vmagent";
CapabilityBoundingSet = [ "" ];
DeviceAllow = [ "" ];
LockPersonality = true;
MemoryDenyWriteExecute = true;
NoNewPrivileges = true;
PrivateDevices = true;
ProtectClock = true;
ProtectControlGroups = true;
ProtectHome = true;
ProtectHostname = true;
ProtectKernelLogs = true;
ProtectKernelModules = true;
ProtectKernelTunables = true;
ProtectSystem = "strict";
RemoveIPC = true;
RestrictAddressFamilies = [ "AF_INET" "AF_INET6" ];
RestrictNamespaces = true;
RestrictRealtime = true;
RestrictSUIDSGID = true;
SystemCallArchitectures = "native";
UMask = "0077";
};
}; };
in {
enable = true;
path = [pkgs.victoriametrics];
wantedBy = ["multi-user.target"];
after = ["network.target" "victoriametrics.service"];
serviceConfig = {
ExecStart = [
(lib.concatStringsSep " " [
"${pkgs.victoriametrics}/bin/vmagent"
"-promscrape.config=${promscrape}"
"-remoteWrite.url=http://${vmAddr}/api/v1/write"
"-remoteWrite.tmpDataPath=%t/vmagent"
])
];
SupplementaryGroups = "metrics";
DynamicUser = true;
RuntimeDirectory = "vmagent";
CapabilityBoundingSet = [""];
DeviceAllow = [""];
LockPersonality = true;
MemoryDenyWriteExecute = true;
NoNewPrivileges = true;
PrivateDevices = true;
ProtectClock = true;
ProtectControlGroups = true;
ProtectHome = true;
ProtectHostname = true;
ProtectKernelLogs = true;
ProtectKernelModules = true;
ProtectKernelTunables = true;
ProtectSystem = "strict";
RemoveIPC = true;
RestrictAddressFamilies = ["AF_INET" "AF_INET6"];
RestrictNamespaces = true;
RestrictRealtime = true;
RestrictSUIDSGID = true;
SystemCallArchitectures = "native";
UMask = "0077";
};
};
} }
]; ];
users.groups.metrics = { }; users.groups.metrics = {};
services.victoriametrics.scrapeConfigs = services.victoriametrics.scrapeConfigs = let
let allExporters =
allExporters = lib.mapAttrs (name: exporter: {
lib.mapAttrs inherit (exporter) listenAddress port;
(name: exporter: { }) ((lib.filterAttrs (_: exporter: builtins.isAttrs exporter && exporter.enable)
inherit (exporter) listenAddress port; config.services.prometheus.exporters)
}) // config.services.prometheus.extraExporters);
((lib.filterAttrs (_: exporter: builtins.isAttrs exporter && exporter.enable) in
config.services.prometheus.exporters) lib.mapAttrs (_: exporter: {
// config.services.prometheus.extraExporters); targets = ["${exporter.listenAddress}:${toString exporter.port}"];
in })
lib.mapAttrs allExporters;
(_: exporter: {
targets = [ "${exporter.listenAddress}:${toString exporter.port}" ];
})
allExporters;
}; };
} }

View file

@ -1,4 +1,4 @@
{ config, ... }: { {config, ...}: {
config.services.victoriametrics = { config.services.victoriametrics = {
enable = true; enable = true;
extraOptions = [ extraOptions = [
@ -7,10 +7,10 @@
scrapeConfigs = { scrapeConfigs = {
forgejo = { forgejo = {
targets = [ "127.0.0.1:${toString config.services.forgejo.settings.server.HTTP_PORT}" ]; targets = ["127.0.0.1:${toString config.services.forgejo.settings.server.HTTP_PORT}"];
extraSettings.authorization.credentials_file = config.sops.secrets."forgejo/metrics-token".path; extraSettings.authorization.credentials_file = config.sops.secrets."forgejo/metrics-token".path;
}; };
coturn.targets = [ "127.0.0.1:9641" ]; coturn.targets = ["127.0.0.1:9641"];
}; };
}; };
} }

View file

@ -1,15 +1,14 @@
{ pkgs {
, config pkgs,
, ... config,
}: ...
let }: let
# Update pending on rewrite of nextcloud news, though there is an # Update pending on rewrite of nextcloud news, though there is an
# alpha to switch to if it becomes necessary: # alpha to switch to if it becomes necessary:
# https://github.com/nextcloud/news/issues/2610 # https://github.com/nextcloud/news/issues/2610
nextcloud = pkgs.nextcloud27; nextcloud = pkgs.nextcloud27;
hostName = "nextcloud.${config.services.nginx.domain}"; hostName = "nextcloud.${config.services.nginx.domain}";
in in {
{
services.nextcloud = { services.nextcloud = {
inherit hostName; inherit hostName;
@ -43,7 +42,7 @@ in
}; };
# Ensure that this service doesn't start before postgres is ready # Ensure that this service doesn't start before postgres is ready
systemd.services.nextcloud-setup.after = [ "postgresql.service" ]; systemd.services.nextcloud-setup.after = ["postgresql.service"];
# Set up SSL # Set up SSL
services.nginx.virtualHosts."${hostName}" = { services.nginx.virtualHosts."${hostName}" = {

View file

@ -1,4 +1,4 @@
{ pkgs, ... }: { {pkgs, ...}: {
services.postgresql = { services.postgresql = {
package = pkgs.postgresql_14; package = pkgs.postgresql_14;
enable = true; enable = true;

View file

@ -1,17 +1,16 @@
{ pkgs
, lib
, ...
}:
let
inherit (lib) concatStringsSep;
in
{ {
pkgs,
lib,
...
}: let
inherit (lib) concatStringsSep;
in {
# Sadly, steam-run requires some X libs # Sadly, steam-run requires some X libs
environment.noXlibs = false; environment.noXlibs = false;
systemd.services.starbound = { systemd.services.starbound = {
description = "Starbound"; description = "Starbound";
after = [ "network.target" ]; after = ["network.target"];
serviceConfig = { serviceConfig = {
ExecStart = "${pkgs.local.starbound}/bin/launch-starbound ${./configs/starbound.json}"; ExecStart = "${pkgs.local.starbound}/bin/launch-starbound ${./configs/starbound.json}";
@ -68,7 +67,7 @@ in
# Game servers shouldn't use cgroups themselves either # Game servers shouldn't use cgroups themselves either
ProtectControlGroups = true; ProtectControlGroups = true;
# Most game servers will never need other socket types # Most game servers will never need other socket types
RestrictAddressFamilies = [ "AF_UNIX AF_INET AF_INET6" ]; RestrictAddressFamilies = ["AF_UNIX AF_INET AF_INET6"];
# Also a no-brainer, no game server should ever need this # Also a no-brainer, no game server should ever need this
LockPersonality = true; LockPersonality = true;
# Some game servers will probably try to set this, but they # Some game servers will probably try to set this, but they
@ -117,6 +116,6 @@ in
paths = [ paths = [
"/var/lib/private/starbound/storage/universe/" "/var/lib/private/starbound/storage/universe/"
]; ];
pauseServices = [ "starbound.service" ]; pauseServices = ["starbound.service"];
}; };
} }

View file

@ -1,8 +1,6 @@
{ config, ... }: {config, ...}: let
let
domain = config.services.nginx.domain; domain = config.services.nginx.domain;
in in {
{
services.tlaternet-webserver = { services.tlaternet-webserver = {
enable = true; enable = true;
listen = { listen = {
@ -12,17 +10,15 @@ in
}; };
# Set up SSL # Set up SSL
services.nginx.virtualHosts."${domain}" = services.nginx.virtualHosts."${domain}" = let
let inherit (config.services.tlaternet-webserver.listen) addr port;
inherit (config.services.tlaternet-webserver.listen) addr port; in {
in serverAliases = ["www.${domain}"];
{
serverAliases = [ "www.${domain}" ];
forceSSL = true; forceSSL = true;
useACMEHost = "tlater.net"; useACMEHost = "tlater.net";
enableHSTS = true; enableHSTS = true;
locations."/".proxyPass = "http://${addr}:${toString port}"; locations."/".proxyPass = "http://${addr}:${toString port}";
}; };
} }

View file

@ -1,4 +1,4 @@
{ config, ... }: { {config, ...}: {
# iptables needs to permit forwarding from wg0 to wg0 # iptables needs to permit forwarding from wg0 to wg0
networking.firewall.extraCommands = '' networking.firewall.extraCommands = ''
iptables -A FORWARD -i wg0 -o wg0 -j ACCEPT iptables -A FORWARD -i wg0 -o wg0 -j ACCEPT
@ -26,7 +26,7 @@
{ {
# yui # yui
wireguardPeerConfig = { wireguardPeerConfig = {
AllowedIPs = [ "10.45.249.2/32" ]; AllowedIPs = ["10.45.249.2/32"];
PublicKey = "5mlnqEVJWks5OqgeFA2bLIrvST9TlCE81Btl+j4myz0="; PublicKey = "5mlnqEVJWks5OqgeFA2bLIrvST9TlCE81Btl+j4myz0=";
}; };
} }
@ -34,7 +34,7 @@
{ {
# yuanyuan # yuanyuan
wireguardPeerConfig = { wireguardPeerConfig = {
AllowedIPs = [ "10.45.249.10/32" ]; AllowedIPs = ["10.45.249.10/32"];
PublicKey = "0UsFE2atz/O5P3OKQ8UHyyyGQNJbp1MeIWUJLuoerwE="; PublicKey = "0UsFE2atz/O5P3OKQ8UHyyyGQNJbp1MeIWUJLuoerwE=";
}; };
} }

View file

@ -31,8 +31,8 @@
}; };
# Heisenbridge # Heisenbridge
"heisenbridge/as-token" = { }; "heisenbridge/as-token" = {};
"heisenbridge/hs-token" = { }; "heisenbridge/hs-token" = {};
"hetzner-api" = { "hetzner-api" = {
owner = "acme"; owner = "acme";
@ -62,10 +62,10 @@
}; };
# Steam # Steam
"steam/tlater" = { }; "steam/tlater" = {};
# Turn # Turn
"turn/env" = { }; "turn/env" = {};
"turn/secret" = { "turn/secret" = {
owner = "turnserver"; owner = "turnserver";
}; };

240
flake.nix
View file

@ -32,130 +32,126 @@
}; };
}; };
outputs = outputs = {
{ self self,
, nixpkgs nixpkgs,
, sops-nix sops-nix,
, nvfetcher nvfetcher,
, deploy-rs deploy-rs,
, ... ...
} @ inputs: } @ inputs: let
let system = "x86_64-linux";
system = "x86_64-linux"; pkgs = nixpkgs.legacyPackages.${system};
pkgs = nixpkgs.legacyPackages.${system}; in {
in ##################
{ # Configurations #
################## ##################
# Configurations # nixosConfigurations = {
################## # The actual system definition
nixosConfigurations = { hetzner-1 = nixpkgs.lib.nixosSystem {
# The actual system definition inherit system;
hetzner-1 = nixpkgs.lib.nixosSystem { specialArgs.flake-inputs = inputs;
inherit system;
specialArgs.flake-inputs = inputs;
modules = [ modules = [
./configuration ./configuration
./configuration/hardware-specific/hetzner ./configuration/hardware-specific/hetzner
];
};
};
############################
# Deployment configuration #
############################
deploy.nodes = {
hetzner-1 = {
hostname = "116.202.158.55";
profiles.system = {
user = "root";
path = deploy-rs.lib.${system}.activate.nixos self.nixosConfigurations.hetzner-1;
};
sshUser = "tlater";
sshOpts = [ "-p" "2222" "-o" "ForwardAgent=yes" ];
};
};
#########
# Tests #
#########
checks = builtins.mapAttrs (system: deployLib: deployLib.deployChecks self.deploy) deploy-rs.lib;
###################
# Utility scripts #
###################
apps.${system} = {
default = self.apps.${system}.run-vm;
run-vm = {
type = "app";
program =
let
vm = nixpkgs.lib.nixosSystem {
inherit system;
specialArgs.flake-inputs = inputs;
modules = [
./configuration
./configuration/hardware-specific/vm.nix
];
};
in
(pkgs.writeShellScript "" ''
${vm.config.system.build.vm.outPath}/bin/run-testvm-vm
'').outPath;
};
update-pkgs = {
type = "app";
program =
let
nvfetcher-bin = "${nvfetcher.packages.${system}.default}/bin/nvfetcher";
in
(pkgs.writeShellScript "update-pkgs" ''
cd "$(git rev-parse --show-toplevel)/pkgs"
${nvfetcher-bin} -o _sources_pkgs -c nvfetcher.toml
'').outPath;
};
update-nextcloud-apps = {
type = "app";
program =
let
nvfetcher-bin = "${nvfetcher.packages.${system}.default}/bin/nvfetcher";
in
(pkgs.writeShellScript "update-nextcloud-apps" ''
cd "$(git rev-parse --show-toplevel)/pkgs"
${nvfetcher-bin} -o _sources_nextcloud -c nextcloud-apps.toml
'').outPath;
};
};
###########################
# Development environment #
###########################
devShells.${system}.default = nixpkgs.legacyPackages.${system}.mkShell {
sopsPGPKeyDirs = [ "./keys/hosts/" "./keys/users/" ];
nativeBuildInputs = [
sops-nix.packages.${system}.sops-import-keys-hook
];
packages = with pkgs; [
sops-nix.packages.${system}.sops-init-gpg-key
deploy-rs.packages.${system}.default
nixpkgs-fmt
cargo
clippy
rustc
rustfmt
rust-analyzer
pkg-config
openssl
]; ];
}; };
}; };
############################
# Deployment configuration #
############################
deploy.nodes = {
hetzner-1 = {
hostname = "116.202.158.55";
profiles.system = {
user = "root";
path = deploy-rs.lib.${system}.activate.nixos self.nixosConfigurations.hetzner-1;
};
sshUser = "tlater";
sshOpts = ["-p" "2222" "-o" "ForwardAgent=yes"];
};
};
#########
# Tests #
#########
checks = builtins.mapAttrs (system: deployLib: deployLib.deployChecks self.deploy) deploy-rs.lib;
###################
# Utility scripts #
###################
apps.${system} = {
default = self.apps.${system}.run-vm;
run-vm = {
type = "app";
program = let
vm = nixpkgs.lib.nixosSystem {
inherit system;
specialArgs.flake-inputs = inputs;
modules = [
./configuration
./configuration/hardware-specific/vm.nix
];
};
in
(pkgs.writeShellScript "" ''
${vm.config.system.build.vm.outPath}/bin/run-testvm-vm
'')
.outPath;
};
update-pkgs = {
type = "app";
program = let
nvfetcher-bin = "${nvfetcher.packages.${system}.default}/bin/nvfetcher";
in
(pkgs.writeShellScript "update-pkgs" ''
cd "$(git rev-parse --show-toplevel)/pkgs"
${nvfetcher-bin} -o _sources_pkgs -c nvfetcher.toml
'')
.outPath;
};
update-nextcloud-apps = {
type = "app";
program = let
nvfetcher-bin = "${nvfetcher.packages.${system}.default}/bin/nvfetcher";
in
(pkgs.writeShellScript "update-nextcloud-apps" ''
cd "$(git rev-parse --show-toplevel)/pkgs"
${nvfetcher-bin} -o _sources_nextcloud -c nextcloud-apps.toml
'')
.outPath;
};
};
###########################
# Development environment #
###########################
devShells.${system}.default = nixpkgs.legacyPackages.${system}.mkShell {
sopsPGPKeyDirs = ["./keys/hosts/" "./keys/users/"];
nativeBuildInputs = [
sops-nix.packages.${system}.sops-import-keys-hook
];
packages = with pkgs; [
sops-nix.packages.${system}.sops-init-gpg-key
deploy-rs.packages.${system}.default
cargo
clippy
rustc
rustfmt
rust-analyzer
pkg-config
openssl
];
};
};
} }

View file

@ -1,7 +1,8 @@
{ config {
, pkgs config,
, lib pkgs,
, ... lib,
...
}: { }: {
options = { options = {
services.nginx.domain = lib.mkOption { services.nginx.domain = lib.mkOption {
@ -9,37 +10,36 @@
description = "The base domain name to append to virtual domain names"; description = "The base domain name to append to virtual domain names";
}; };
services.nginx.virtualHosts = services.nginx.virtualHosts = let
let extraVirtualHostOptions = {
extraVirtualHostOptions = name,
{ name config,
, config ...
, ... }: {
}: { options = {
options = { enableHSTS = lib.mkEnableOption "Enable HSTS";
enableHSTS = lib.mkEnableOption "Enable HSTS";
addAccessLog = lib.mkOption { addAccessLog = lib.mkOption {
type = lib.types.bool; type = lib.types.bool;
default = true; default = true;
description = '' description = ''
Add special logging to `/var/log/nginx/''${serverName}` Add special logging to `/var/log/nginx/''${serverName}`
''; '';
};
};
config = {
extraConfig = lib.concatStringsSep "\n" [
(lib.optionalString config.enableHSTS ''
add_header Strict-Transport-Security "max-age=15552000; includeSubDomains" always;
'')
(lib.optionalString config.addAccessLog ''
access_log /var/log/nginx/${name}/access.log upstream_time;
'')
];
};
}; };
in };
config = {
extraConfig = lib.concatStringsSep "\n" [
(lib.optionalString config.enableHSTS ''
add_header Strict-Transport-Security "max-age=15552000; includeSubDomains" always;
'')
(lib.optionalString config.addAccessLog ''
access_log /var/log/nginx/${name}/access.log upstream_time;
'')
];
};
};
in
lib.mkOption { lib.mkOption {
type = lib.types.attrsOf (lib.types.submodule extraVirtualHostOptions); type = lib.types.attrsOf (lib.types.submodule extraVirtualHostOptions);
}; };
@ -47,15 +47,13 @@
config = { config = {
# Don't attempt to run acme if the domain name is not tlater.net # Don't attempt to run acme if the domain name is not tlater.net
systemd.services = systemd.services = let
let confirm = ''[[ "tlater.net" = ${config.services.nginx.domain} ]]'';
confirm = ''[[ "tlater.net" = ${config.services.nginx.domain} ]]''; in
in lib.mapAttrs' (cert: _:
lib.mapAttrs' lib.nameValuePair "acme-${cert}" {
(cert: _: serviceConfig.ExecCondition = ''${pkgs.runtimeShell} -c '${confirm}' '';
lib.nameValuePair "acme-${cert}" { })
serviceConfig.ExecCondition = ''${pkgs.runtimeShell} -c '${confirm}' ''; config.security.acme.certs;
})
config.security.acme.certs;
}; };
} }

View file

@ -1,6 +1,7 @@
{ pkgs {
, rustPlatform pkgs,
, ... rustPlatform,
...
}: }:
rustPlatform.buildRustPackage { rustPlatform.buildRustPackage {
pname = "afvalcalendar"; pname = "afvalcalendar";

View file

@ -1,23 +1,22 @@
{ pkgs {
, lib pkgs,
, lib,
}: }: let
let
inherit (builtins) fromJSON mapAttrs readFile; inherit (builtins) fromJSON mapAttrs readFile;
inherit (pkgs) callPackage; inherit (pkgs) callPackage;
in in
{ {
starbound = callPackage ./starbound { }; starbound = callPackage ./starbound {};
prometheus-fail2ban-exporter = callPackage ./prometheus/fail2ban-exporter.nix { prometheus-fail2ban-exporter = callPackage ./prometheus/fail2ban-exporter.nix {
sources = pkgs.callPackage ./_sources_pkgs/generated.nix { }; sources = pkgs.callPackage ./_sources_pkgs/generated.nix {};
}; };
afvalcalendar = callPackage ./afvalcalendar { }; afvalcalendar = callPackage ./afvalcalendar {};
} }
// ( // (
# Add nextcloud apps # Add nextcloud apps
let let
mkNextcloudApp = pkgs.callPackage ./mkNextcloudApp.nix { }; mkNextcloudApp = pkgs.callPackage ./mkNextcloudApp.nix {};
sources = fromJSON (readFile ./_sources_nextcloud/generated.json); sources = fromJSON (readFile ./_sources_nextcloud/generated.json);
in in
mapAttrs (_: source: mkNextcloudApp source) sources mapAttrs (_: source: mkNextcloudApp source) sources
) )

View file

@ -1,6 +1,6 @@
{ fetchNextcloudApp {
, lib fetchNextcloudApp,
, lib,
}: source: }: source:
fetchNextcloudApp { fetchNextcloudApp {
url = source.src.url; url = source.src.url;

View file

@ -1,6 +1,6 @@
{ buildGoModule {
, sources buildGoModule,
, sources,
}: }:
buildGoModule { buildGoModule {
inherit (sources.prometheus-fail2ban-exporter) pname src version; inherit (sources.prometheus-fail2ban-exporter) pname src version;

View file

@ -1,35 +1,34 @@
{ stdenv {
, lib stdenv,
, makeWrapper lib,
, patchelf makeWrapper,
, steamPackages patchelf,
, replace-secret steamPackages,
, replace-secret,
}: }: let
let
# Use the directory in which starbound is installed so steamcmd # Use the directory in which starbound is installed so steamcmd
# doesn't have to be reinstalled constantly (we're using DynamicUser # doesn't have to be reinstalled constantly (we're using DynamicUser
# with StateDirectory to persist this). # with StateDirectory to persist this).
steamcmd = steamPackages.steamcmd.override { steamcmd = steamPackages.steamcmd.override {
steamRoot = "/var/lib/starbound/.steamcmd"; steamRoot = "/var/lib/starbound/.steamcmd";
}; };
wrapperPath = lib.makeBinPath [ patchelf steamcmd replace-secret ]; wrapperPath = lib.makeBinPath [patchelf steamcmd replace-secret];
in in
stdenv.mkDerivation { stdenv.mkDerivation {
name = "starbound-update-script"; name = "starbound-update-script";
nativeBuildInputs = [ makeWrapper ]; nativeBuildInputs = [makeWrapper];
dontUnpack = true; dontUnpack = true;
patchPhase = '' patchPhase = ''
interpreter="$(cat $NIX_CC/nix-support/dynamic-linker)" interpreter="$(cat $NIX_CC/nix-support/dynamic-linker)"
substitute ${./launch-starbound.sh} launch-starbound --subst-var interpreter substitute ${./launch-starbound.sh} launch-starbound --subst-var interpreter
''; '';
installPhase = '' installPhase = ''
mkdir -p $out/bin mkdir -p $out/bin
cp launch-starbound $out/bin/launch-starbound cp launch-starbound $out/bin/launch-starbound
chmod +x $out/bin/launch-starbound chmod +x $out/bin/launch-starbound
''; '';
postFixup = '' postFixup = ''
wrapProgram $out/bin/launch-starbound \ wrapProgram $out/bin/launch-starbound \
--prefix PATH : "${wrapperPath}" --prefix PATH : "${wrapperPath}"
''; '';
} }