650 lines
14 KiB
Go
650 lines
14 KiB
Go
package soju
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"log"
|
|
"mime"
|
|
"net"
|
|
"net/http"
|
|
"runtime/debug"
|
|
"sync"
|
|
"sync/atomic"
|
|
"time"
|
|
|
|
"github.com/SherClockHolmes/webpush-go"
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
"github.com/prometheus/client_golang/prometheus/promauto"
|
|
"gopkg.in/irc.v4"
|
|
"nhooyr.io/websocket"
|
|
|
|
"git.sr.ht/~emersion/soju/auth"
|
|
"git.sr.ht/~emersion/soju/config"
|
|
"git.sr.ht/~emersion/soju/database"
|
|
"git.sr.ht/~emersion/soju/identd"
|
|
)
|
|
|
|
// TODO: make configurable
|
|
var retryConnectMinDelay = time.Minute
|
|
var retryConnectMaxDelay = 10 * time.Minute
|
|
var retryConnectJitter = time.Minute
|
|
var connectTimeout = 15 * time.Second
|
|
var writeTimeout = 10 * time.Second
|
|
var upstreamMessageDelay = 2 * time.Second
|
|
var upstreamMessageBurst = 10
|
|
var backlogTimeout = 10 * time.Second
|
|
var handleDownstreamMessageTimeout = 10 * time.Second
|
|
var downstreamRegisterTimeout = 30 * time.Second
|
|
var chatHistoryLimit = 1000
|
|
var backlogLimit = 4000
|
|
|
|
var errWebPushSubscriptionExpired = fmt.Errorf("Web Push subscription expired")
|
|
|
|
type Logger interface {
|
|
Printf(format string, v ...interface{})
|
|
Debugf(format string, v ...interface{})
|
|
}
|
|
|
|
type logger struct {
|
|
*log.Logger
|
|
debug bool
|
|
}
|
|
|
|
func (l logger) Debugf(format string, v ...interface{}) {
|
|
if !l.debug {
|
|
return
|
|
}
|
|
l.Logger.Printf(format, v...)
|
|
}
|
|
|
|
func NewLogger(out io.Writer, debug bool) Logger {
|
|
return logger{
|
|
Logger: log.New(out, "", log.LstdFlags),
|
|
debug: debug,
|
|
}
|
|
}
|
|
|
|
type prefixLogger struct {
|
|
logger Logger
|
|
prefix string
|
|
}
|
|
|
|
var _ Logger = (*prefixLogger)(nil)
|
|
|
|
func (l *prefixLogger) Printf(format string, v ...interface{}) {
|
|
v = append([]interface{}{l.prefix}, v...)
|
|
l.logger.Printf("%v"+format, v...)
|
|
}
|
|
|
|
func (l *prefixLogger) Debugf(format string, v ...interface{}) {
|
|
v = append([]interface{}{l.prefix}, v...)
|
|
l.logger.Debugf("%v"+format, v...)
|
|
}
|
|
|
|
type int64Gauge struct {
|
|
v int64 // atomic
|
|
}
|
|
|
|
func (g *int64Gauge) Add(delta int64) {
|
|
atomic.AddInt64(&g.v, delta)
|
|
}
|
|
|
|
func (g *int64Gauge) Value() int64 {
|
|
return atomic.LoadInt64(&g.v)
|
|
}
|
|
|
|
func (g *int64Gauge) Float64() float64 {
|
|
return float64(g.Value())
|
|
}
|
|
|
|
type retryListener struct {
|
|
net.Listener
|
|
Logger Logger
|
|
|
|
delay time.Duration
|
|
}
|
|
|
|
func NewRetryListener(ln net.Listener) net.Listener {
|
|
return &retryListener{Listener: ln}
|
|
}
|
|
|
|
func (ln *retryListener) Accept() (net.Conn, error) {
|
|
for {
|
|
conn, err := ln.Listener.Accept()
|
|
if ne, ok := err.(net.Error); ok && ne.Temporary() {
|
|
if ln.delay == 0 {
|
|
ln.delay = 5 * time.Millisecond
|
|
} else {
|
|
ln.delay *= 2
|
|
}
|
|
if max := 1 * time.Second; ln.delay > max {
|
|
ln.delay = max
|
|
}
|
|
if ln.Logger != nil {
|
|
ln.Logger.Printf("accept error (retrying in %v): %v", ln.delay, err)
|
|
}
|
|
time.Sleep(ln.delay)
|
|
} else {
|
|
ln.delay = 0
|
|
return conn, err
|
|
}
|
|
}
|
|
}
|
|
|
|
type Config struct {
|
|
Hostname string
|
|
Title string
|
|
LogPath string
|
|
HTTPOrigins []string
|
|
AcceptProxyIPs config.IPSet
|
|
MaxUserNetworks int
|
|
MOTD string
|
|
UpstreamUserIPs []*net.IPNet
|
|
DisableInactiveUsersDelay time.Duration
|
|
EnableUsersOnAuth bool
|
|
Auth auth.Authenticator
|
|
}
|
|
|
|
type Server struct {
|
|
Logger Logger
|
|
Identd *identd.Identd // can be nil
|
|
MetricsRegistry prometheus.Registerer // can be nil
|
|
|
|
config atomic.Value // *Config
|
|
db database.Database
|
|
stopWG sync.WaitGroup
|
|
stopCh chan struct{}
|
|
|
|
lock sync.Mutex
|
|
listeners map[net.Listener]struct{}
|
|
users map[string]*user
|
|
shutdown bool
|
|
|
|
metrics struct {
|
|
downstreams int64Gauge
|
|
upstreams int64Gauge
|
|
|
|
upstreamOutMessagesTotal prometheus.Counter
|
|
upstreamInMessagesTotal prometheus.Counter
|
|
downstreamOutMessagesTotal prometheus.Counter
|
|
downstreamInMessagesTotal prometheus.Counter
|
|
|
|
upstreamConnectErrorsTotal prometheus.Counter
|
|
}
|
|
|
|
webPush *database.WebPushConfig
|
|
}
|
|
|
|
func NewServer(db database.Database) *Server {
|
|
srv := &Server{
|
|
Logger: NewLogger(log.Writer(), true),
|
|
db: db,
|
|
listeners: make(map[net.Listener]struct{}),
|
|
users: make(map[string]*user),
|
|
stopCh: make(chan struct{}),
|
|
}
|
|
srv.config.Store(&Config{
|
|
Hostname: "localhost",
|
|
MaxUserNetworks: -1,
|
|
Auth: auth.NewInternal(),
|
|
})
|
|
return srv
|
|
}
|
|
|
|
func (s *Server) prefix() *irc.Prefix {
|
|
return &irc.Prefix{Name: s.Config().Hostname}
|
|
}
|
|
|
|
func (s *Server) Config() *Config {
|
|
return s.config.Load().(*Config)
|
|
}
|
|
|
|
func (s *Server) SetConfig(cfg *Config) {
|
|
s.config.Store(cfg)
|
|
}
|
|
|
|
func (s *Server) Start() error {
|
|
s.registerMetrics()
|
|
|
|
if err := s.loadWebPushConfig(context.TODO()); err != nil {
|
|
return err
|
|
}
|
|
|
|
users, err := s.db.ListUsers(context.TODO())
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
s.lock.Lock()
|
|
for i := range users {
|
|
s.addUserLocked(&users[i])
|
|
}
|
|
s.lock.Unlock()
|
|
|
|
s.stopWG.Add(1)
|
|
go func() {
|
|
defer s.stopWG.Done()
|
|
s.disableInactiveUsersLoop()
|
|
}()
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *Server) registerMetrics() {
|
|
factory := promauto.With(s.MetricsRegistry)
|
|
|
|
factory.NewGaugeFunc(prometheus.GaugeOpts{
|
|
Name: "soju_users_active",
|
|
Help: "Current number of active users",
|
|
}, func() float64 {
|
|
s.lock.Lock()
|
|
n := len(s.users)
|
|
s.lock.Unlock()
|
|
return float64(n)
|
|
})
|
|
|
|
factory.NewGaugeFunc(prometheus.GaugeOpts{
|
|
Name: "soju_downstreams_active",
|
|
Help: "Current number of downstream connections",
|
|
}, s.metrics.downstreams.Float64)
|
|
|
|
factory.NewGaugeFunc(prometheus.GaugeOpts{
|
|
Name: "soju_upstreams_active",
|
|
Help: "Current number of upstream connections",
|
|
}, s.metrics.upstreams.Float64)
|
|
|
|
s.metrics.upstreamOutMessagesTotal = factory.NewCounter(prometheus.CounterOpts{
|
|
Name: "soju_upstream_out_messages_total",
|
|
Help: "Total number of outgoing messages sent to upstream servers",
|
|
})
|
|
|
|
s.metrics.upstreamInMessagesTotal = factory.NewCounter(prometheus.CounterOpts{
|
|
Name: "soju_upstream_in_messages_total",
|
|
Help: "Total number of incoming messages received from upstream servers",
|
|
})
|
|
|
|
s.metrics.downstreamOutMessagesTotal = factory.NewCounter(prometheus.CounterOpts{
|
|
Name: "soju_downstream_out_messages_total",
|
|
Help: "Total number of outgoing messages sent to downstream clients",
|
|
})
|
|
|
|
s.metrics.downstreamInMessagesTotal = factory.NewCounter(prometheus.CounterOpts{
|
|
Name: "soju_downstream_in_messages_total",
|
|
Help: "Total number of incoming messages received from downstream clients",
|
|
})
|
|
|
|
s.metrics.upstreamConnectErrorsTotal = factory.NewCounter(prometheus.CounterOpts{
|
|
Name: "soju_upstream_connect_errors_total",
|
|
Help: "Total number of upstream connection errors",
|
|
})
|
|
}
|
|
|
|
func (s *Server) loadWebPushConfig(ctx context.Context) error {
|
|
configs, err := s.db.ListWebPushConfigs(ctx)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to list Web push configs: %v", err)
|
|
}
|
|
|
|
if len(configs) > 1 {
|
|
return fmt.Errorf("expected zero or one Web push config, got %v", len(configs))
|
|
} else if len(configs) == 1 {
|
|
s.webPush = &configs[0]
|
|
return nil
|
|
}
|
|
|
|
s.Logger.Printf("generating Web push VAPID key pair")
|
|
priv, pub, err := webpush.GenerateVAPIDKeys()
|
|
if err != nil {
|
|
return fmt.Errorf("failed to generate Web push VAPID key pair: %v", err)
|
|
}
|
|
|
|
config := new(database.WebPushConfig)
|
|
config.VAPIDKeys.Public = pub
|
|
config.VAPIDKeys.Private = priv
|
|
if err := s.db.StoreWebPushConfig(ctx, config); err != nil {
|
|
return fmt.Errorf("failed to store Web push config: %v", err)
|
|
}
|
|
|
|
s.webPush = config
|
|
return nil
|
|
}
|
|
|
|
func (s *Server) sendWebPush(ctx context.Context, sub *webpush.Subscription, vapidPubKey string, msg *irc.Message) error {
|
|
ctx, cancel := context.WithTimeout(ctx, 15*time.Second)
|
|
defer cancel()
|
|
|
|
var urgency webpush.Urgency
|
|
switch msg.Command {
|
|
case "PRIVMSG", "NOTICE", "INVITE":
|
|
urgency = webpush.UrgencyHigh
|
|
default:
|
|
urgency = webpush.UrgencyNormal
|
|
}
|
|
|
|
options := webpush.Options{
|
|
VAPIDPublicKey: s.webPush.VAPIDKeys.Public,
|
|
VAPIDPrivateKey: s.webPush.VAPIDKeys.Private,
|
|
Subscriber: "https://soju.im",
|
|
TTL: 7 * 24 * 60 * 60, // seconds
|
|
Urgency: urgency,
|
|
RecordSize: 2048,
|
|
}
|
|
|
|
if vapidPubKey != options.VAPIDPublicKey {
|
|
return fmt.Errorf("unknown VAPID public key %q", vapidPubKey)
|
|
}
|
|
|
|
payload := []byte(msg.String())
|
|
resp, err := webpush.SendNotificationWithContext(ctx, payload, sub, &options)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
resp.Body.Close()
|
|
|
|
// 404 means the subscription has expired as per RFC 8030 section 7.3
|
|
if resp.StatusCode == http.StatusNotFound {
|
|
return errWebPushSubscriptionExpired
|
|
} else if resp.StatusCode/100 != 2 {
|
|
return fmt.Errorf("HTTP error: %v", resp.Status)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *Server) Shutdown() {
|
|
s.Logger.Printf("shutting down server")
|
|
|
|
close(s.stopCh)
|
|
|
|
s.lock.Lock()
|
|
s.shutdown = true
|
|
for ln := range s.listeners {
|
|
if err := ln.Close(); err != nil {
|
|
s.Logger.Printf("failed to stop listener: %v", err)
|
|
}
|
|
}
|
|
for _, u := range s.users {
|
|
u.events <- eventStop{}
|
|
}
|
|
s.lock.Unlock()
|
|
|
|
s.Logger.Printf("waiting for users to finish")
|
|
s.stopWG.Wait()
|
|
|
|
if err := s.db.Close(); err != nil {
|
|
s.Logger.Printf("failed to close DB: %v", err)
|
|
}
|
|
}
|
|
|
|
func (s *Server) createUser(ctx context.Context, user *database.User) (*user, error) {
|
|
s.lock.Lock()
|
|
defer s.lock.Unlock()
|
|
|
|
if _, ok := s.users[user.Username]; ok {
|
|
return nil, fmt.Errorf("user %q already exists", user.Username)
|
|
}
|
|
|
|
err := s.db.StoreUser(ctx, user)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("could not create user in db: %v", err)
|
|
}
|
|
|
|
return s.addUserLocked(user), nil
|
|
}
|
|
|
|
func (s *Server) forEachUser(f func(*user)) {
|
|
s.lock.Lock()
|
|
for _, u := range s.users {
|
|
f(u)
|
|
}
|
|
s.lock.Unlock()
|
|
}
|
|
|
|
func (s *Server) getUser(name string) *user {
|
|
s.lock.Lock()
|
|
u := s.users[name]
|
|
s.lock.Unlock()
|
|
return u
|
|
}
|
|
|
|
func (s *Server) addUserLocked(user *database.User) *user {
|
|
s.Logger.Printf("starting bouncer for user %q", user.Username)
|
|
u := newUser(s, user)
|
|
s.users[u.Username] = u
|
|
|
|
s.stopWG.Add(1)
|
|
|
|
go func() {
|
|
defer func() {
|
|
if err := recover(); err != nil {
|
|
s.Logger.Printf("panic serving user %q: %v\n%v", user.Username, err, string(debug.Stack()))
|
|
}
|
|
|
|
s.lock.Lock()
|
|
delete(s.users, u.Username)
|
|
s.lock.Unlock()
|
|
|
|
s.stopWG.Done()
|
|
}()
|
|
|
|
u.run()
|
|
}()
|
|
|
|
return u
|
|
}
|
|
|
|
var lastDownstreamID uint64
|
|
|
|
func (s *Server) handle(ic ircConn) {
|
|
defer func() {
|
|
if err := recover(); err != nil {
|
|
s.Logger.Printf("panic serving downstream %q: %v\n%v", ic.RemoteAddr(), err, string(debug.Stack()))
|
|
}
|
|
}()
|
|
|
|
s.lock.Lock()
|
|
shutdown := s.shutdown
|
|
s.lock.Unlock()
|
|
|
|
s.metrics.downstreams.Add(1)
|
|
id := atomic.AddUint64(&lastDownstreamID, 1)
|
|
dc := newDownstreamConn(s, ic, id)
|
|
if shutdown {
|
|
dc.SendMessage(&irc.Message{
|
|
Command: "ERROR",
|
|
Params: []string{"Server is shutting down"},
|
|
})
|
|
} else if err := dc.runUntilRegistered(); err != nil {
|
|
if !errors.Is(err, io.EOF) {
|
|
dc.logger.Printf("%v", err)
|
|
}
|
|
} else {
|
|
dc.user.events <- eventDownstreamConnected{dc}
|
|
if err := dc.readMessages(dc.user.events); err != nil {
|
|
dc.logger.Printf("%v", err)
|
|
}
|
|
dc.user.events <- eventDownstreamDisconnected{dc}
|
|
}
|
|
dc.Close()
|
|
s.metrics.downstreams.Add(-1)
|
|
}
|
|
|
|
func (s *Server) Serve(ln net.Listener) error {
|
|
ln = &retryListener{
|
|
Listener: ln,
|
|
Logger: &prefixLogger{logger: s.Logger, prefix: fmt.Sprintf("listener %v: ", ln.Addr())},
|
|
}
|
|
|
|
s.lock.Lock()
|
|
s.listeners[ln] = struct{}{}
|
|
s.lock.Unlock()
|
|
|
|
s.stopWG.Add(1)
|
|
|
|
defer func() {
|
|
s.lock.Lock()
|
|
delete(s.listeners, ln)
|
|
s.lock.Unlock()
|
|
|
|
s.stopWG.Done()
|
|
}()
|
|
|
|
for {
|
|
conn, err := ln.Accept()
|
|
if isErrClosed(err) {
|
|
return nil
|
|
} else if err != nil {
|
|
return fmt.Errorf("failed to accept connection: %v", err)
|
|
}
|
|
|
|
go s.handle(newNetIRCConn(conn))
|
|
}
|
|
}
|
|
|
|
func (s *Server) ServeHTTP(w http.ResponseWriter, req *http.Request) {
|
|
conn, err := websocket.Accept(w, req, &websocket.AcceptOptions{
|
|
Subprotocols: []string{"text.ircv3.net"}, // non-compliant, fight me
|
|
OriginPatterns: s.Config().HTTPOrigins,
|
|
})
|
|
if err != nil {
|
|
s.Logger.Printf("failed to serve HTTP connection: %v", err)
|
|
return
|
|
}
|
|
|
|
isProxy := false
|
|
if host, _, err := net.SplitHostPort(req.RemoteAddr); err == nil {
|
|
if ip := net.ParseIP(host); ip != nil {
|
|
isProxy = s.Config().AcceptProxyIPs.Contains(ip)
|
|
}
|
|
}
|
|
|
|
// Only trust the Forwarded header field if this is a trusted proxy IP
|
|
// to prevent users from spoofing the remote address
|
|
remoteAddr := req.RemoteAddr
|
|
if isProxy {
|
|
forwarded := parseForwarded(req.Header)
|
|
if forwarded["for"] != "" {
|
|
remoteAddr = forwarded["for"]
|
|
}
|
|
}
|
|
|
|
s.handle(newWebsocketIRCConn(conn, remoteAddr))
|
|
}
|
|
|
|
func parseForwarded(h http.Header) map[string]string {
|
|
forwarded := h.Get("Forwarded")
|
|
if forwarded == "" {
|
|
return map[string]string{
|
|
"for": h.Get("X-Forwarded-For"),
|
|
"proto": h.Get("X-Forwarded-Proto"),
|
|
"host": h.Get("X-Forwarded-Host"),
|
|
}
|
|
}
|
|
// Hack to easily parse header parameters
|
|
_, params, _ := mime.ParseMediaType("hack; " + forwarded)
|
|
return params
|
|
}
|
|
|
|
type ServerStats struct {
|
|
Users int
|
|
Downstreams int64
|
|
Upstreams int64
|
|
}
|
|
|
|
func (s *Server) Stats() *ServerStats {
|
|
var stats ServerStats
|
|
s.lock.Lock()
|
|
stats.Users = len(s.users)
|
|
s.lock.Unlock()
|
|
stats.Downstreams = s.metrics.downstreams.Value()
|
|
stats.Upstreams = s.metrics.upstreams.Value()
|
|
return &stats
|
|
}
|
|
|
|
func (s *Server) disableInactiveUsersLoop() {
|
|
ticker := time.NewTicker(4 * time.Hour)
|
|
defer ticker.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-s.stopCh:
|
|
return
|
|
case <-ticker.C:
|
|
}
|
|
|
|
if err := s.disableInactiveUsers(context.TODO()); err != nil {
|
|
s.Logger.Printf("failed to disable inactive users: %v", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
func (s *Server) disableInactiveUsers(ctx context.Context) error {
|
|
delay := s.Config().DisableInactiveUsersDelay
|
|
if delay == 0 {
|
|
return nil
|
|
}
|
|
|
|
ctx, cancel := context.WithTimeout(ctx, time.Minute)
|
|
defer cancel()
|
|
|
|
usernames, err := s.db.ListInactiveUsernames(ctx, time.Now().Add(-delay))
|
|
if err != nil {
|
|
return fmt.Errorf("failed to list inactive users: %v", err)
|
|
} else if len(usernames) == 0 {
|
|
return nil
|
|
}
|
|
|
|
// Filter out users with active downstream connections
|
|
var users []*user
|
|
s.lock.Lock()
|
|
for _, username := range usernames {
|
|
u := s.users[username]
|
|
if u == nil {
|
|
// TODO: disable the user in the DB
|
|
continue
|
|
}
|
|
|
|
if n := u.numDownstreamConns.Load(); n > 0 {
|
|
continue
|
|
}
|
|
|
|
users = append(users, u)
|
|
}
|
|
s.lock.Unlock()
|
|
|
|
if len(users) == 0 {
|
|
return nil
|
|
}
|
|
|
|
s.Logger.Printf("found %v inactive users", len(users))
|
|
for _, u := range users {
|
|
done := make(chan error, 1)
|
|
enabled := false
|
|
event := eventUserUpdate{
|
|
enabled: &enabled,
|
|
done: done,
|
|
}
|
|
select {
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
case u.events <- event:
|
|
// Event was sent, let's wait for the reply
|
|
}
|
|
|
|
select {
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
case err := <-done:
|
|
if err != nil {
|
|
return err
|
|
} else {
|
|
s.Logger.Printf("deleted inactive user %q", u.Username)
|
|
}
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|