mirror of
https://github.com/1f349/dendrite.git
synced 2024-12-23 08:44:11 +00:00
f8ae391a5b
* Add more logging for content fields * Fix fields
223 lines
6.6 KiB
Go
223 lines
6.6 KiB
Go
// Copyright 2017 Vector Creations Ltd
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
// Package input contains the code processes new room events
|
|
package input
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/Shopify/sarama"
|
|
"github.com/getsentry/sentry-go"
|
|
"github.com/matrix-org/dendrite/internal/hooks"
|
|
"github.com/matrix-org/dendrite/roomserver/acls"
|
|
"github.com/matrix-org/dendrite/roomserver/api"
|
|
"github.com/matrix-org/dendrite/roomserver/storage"
|
|
"github.com/matrix-org/gomatrixserverlib"
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
log "github.com/sirupsen/logrus"
|
|
"github.com/tidwall/gjson"
|
|
"go.uber.org/atomic"
|
|
)
|
|
|
|
var keyContentFields = map[string]string{
|
|
"m.room.join_rules": "join_rule",
|
|
"m.room.history_visibility": "history_visibility",
|
|
"m.room.member": "membership",
|
|
}
|
|
|
|
type Inputer struct {
|
|
DB storage.Database
|
|
Producer sarama.SyncProducer
|
|
ServerName gomatrixserverlib.ServerName
|
|
ACLs *acls.ServerACLs
|
|
OutputRoomEventTopic string
|
|
workers sync.Map // room ID -> *inputWorker
|
|
}
|
|
|
|
type inputTask struct {
|
|
ctx context.Context
|
|
event *api.InputRoomEvent
|
|
wg *sync.WaitGroup
|
|
err error // written back by worker, only safe to read when all tasks are done
|
|
}
|
|
|
|
type inputWorker struct {
|
|
r *Inputer
|
|
running atomic.Bool
|
|
input *fifoQueue
|
|
}
|
|
|
|
// Guarded by a CAS on w.running
|
|
func (w *inputWorker) start() {
|
|
defer w.running.Store(false)
|
|
for {
|
|
select {
|
|
case <-w.input.wait():
|
|
task, ok := w.input.pop()
|
|
if !ok {
|
|
continue
|
|
}
|
|
roomserverInputBackpressure.With(prometheus.Labels{
|
|
"room_id": task.event.Event.RoomID(),
|
|
}).Dec()
|
|
hooks.Run(hooks.KindNewEventReceived, task.event.Event)
|
|
_, task.err = w.r.processRoomEvent(task.ctx, task.event)
|
|
if task.err == nil {
|
|
hooks.Run(hooks.KindNewEventPersisted, task.event.Event)
|
|
} else {
|
|
sentry.CaptureException(task.err)
|
|
}
|
|
task.wg.Done()
|
|
case <-time.After(time.Second * 5):
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// WriteOutputEvents implements OutputRoomEventWriter
|
|
func (r *Inputer) WriteOutputEvents(roomID string, updates []api.OutputEvent) error {
|
|
messages := make([]*sarama.ProducerMessage, len(updates))
|
|
for i := range updates {
|
|
value, err := json.Marshal(updates[i])
|
|
if err != nil {
|
|
return err
|
|
}
|
|
logger := log.WithFields(log.Fields{
|
|
"room_id": roomID,
|
|
"type": updates[i].Type,
|
|
})
|
|
if updates[i].NewRoomEvent != nil {
|
|
eventType := updates[i].NewRoomEvent.Event.Type()
|
|
logger = logger.WithFields(log.Fields{
|
|
"event_type": eventType,
|
|
"event_id": updates[i].NewRoomEvent.Event.EventID(),
|
|
"adds_state": len(updates[i].NewRoomEvent.AddsStateEventIDs),
|
|
"removes_state": len(updates[i].NewRoomEvent.RemovesStateEventIDs),
|
|
"send_as_server": updates[i].NewRoomEvent.SendAsServer,
|
|
"sender": updates[i].NewRoomEvent.Event.Sender(),
|
|
})
|
|
if updates[i].NewRoomEvent.Event.StateKey() != nil {
|
|
logger = logger.WithField("state_key", *updates[i].NewRoomEvent.Event.StateKey())
|
|
}
|
|
contentKey := keyContentFields[eventType]
|
|
if contentKey != "" {
|
|
value := gjson.GetBytes(updates[i].NewRoomEvent.Event.Content(), contentKey)
|
|
if value.Exists() {
|
|
logger = logger.WithField("content_value", value.String())
|
|
}
|
|
}
|
|
|
|
if eventType == "m.room.server_acl" && updates[i].NewRoomEvent.Event.StateKeyEquals("") {
|
|
ev := updates[i].NewRoomEvent.Event.Unwrap()
|
|
defer r.ACLs.OnServerACLUpdate(ev)
|
|
}
|
|
}
|
|
logger.Infof("Producing to topic '%s'", r.OutputRoomEventTopic)
|
|
messages[i] = &sarama.ProducerMessage{
|
|
Topic: r.OutputRoomEventTopic,
|
|
Key: sarama.StringEncoder(roomID),
|
|
Value: sarama.ByteEncoder(value),
|
|
}
|
|
}
|
|
errs := r.Producer.SendMessages(messages)
|
|
if errs != nil {
|
|
for _, err := range errs.(sarama.ProducerErrors) {
|
|
log.WithError(err).WithField("message_bytes", err.Msg.Value.Length()).Error("Write to kafka failed")
|
|
}
|
|
}
|
|
return errs
|
|
}
|
|
|
|
func init() {
|
|
prometheus.MustRegister(roomserverInputBackpressure)
|
|
}
|
|
|
|
var roomserverInputBackpressure = prometheus.NewGaugeVec(
|
|
prometheus.GaugeOpts{
|
|
Namespace: "dendrite",
|
|
Subsystem: "roomserver",
|
|
Name: "input_backpressure",
|
|
Help: "How many events are queued for input for a given room",
|
|
},
|
|
[]string{"room_id"},
|
|
)
|
|
|
|
// InputRoomEvents implements api.RoomserverInternalAPI
|
|
func (r *Inputer) InputRoomEvents(
|
|
_ context.Context,
|
|
request *api.InputRoomEventsRequest,
|
|
response *api.InputRoomEventsResponse,
|
|
) {
|
|
// Create a wait group. Each task that we dispatch will call Done on
|
|
// this wait group so that we know when all of our events have been
|
|
// processed.
|
|
wg := &sync.WaitGroup{}
|
|
wg.Add(len(request.InputRoomEvents))
|
|
tasks := make([]*inputTask, len(request.InputRoomEvents))
|
|
|
|
for i, e := range request.InputRoomEvents {
|
|
// Work out if we are running per-room workers or if we're just doing
|
|
// it on a global basis (e.g. SQLite).
|
|
roomID := "global"
|
|
if r.DB.SupportsConcurrentRoomInputs() {
|
|
roomID = e.Event.RoomID()
|
|
}
|
|
|
|
// Look up the worker, or create it if it doesn't exist. This channel
|
|
// is buffered to reduce the chance that we'll be blocked by another
|
|
// room - the channel will be quite small as it's just pointer types.
|
|
w, _ := r.workers.LoadOrStore(roomID, &inputWorker{
|
|
r: r,
|
|
input: newFIFOQueue(),
|
|
})
|
|
worker := w.(*inputWorker)
|
|
|
|
// Create a task. This contains the input event and a reference to
|
|
// the wait group, so that the worker can notify us when this specific
|
|
// task has been finished.
|
|
tasks[i] = &inputTask{
|
|
ctx: context.Background(),
|
|
event: &request.InputRoomEvents[i],
|
|
wg: wg,
|
|
}
|
|
|
|
// Send the task to the worker.
|
|
if worker.running.CAS(false, true) {
|
|
go worker.start()
|
|
}
|
|
worker.input.push(tasks[i])
|
|
roomserverInputBackpressure.With(prometheus.Labels{
|
|
"room_id": roomID,
|
|
}).Inc()
|
|
}
|
|
|
|
// Wait for all of the workers to return results about our tasks.
|
|
wg.Wait()
|
|
|
|
// If any of the tasks returned an error, we should probably report
|
|
// that back to the caller.
|
|
for _, task := range tasks {
|
|
if task.err != nil {
|
|
response.ErrMsg = task.err.Error()
|
|
_, rejected := task.err.(*gomatrixserverlib.NotAllowed)
|
|
response.NotAllowed = rejected
|
|
return
|
|
}
|
|
}
|
|
}
|