mirror of
https://github.com/1f349/dendrite.git
synced 2024-11-25 21:21:35 +00:00
9f443317bc
* Initial graceful stop * Fix dendritejs * Use process context for outbound federation requests in destination queues * Reduce logging * Fix log level
105 lines
3.4 KiB
Go
105 lines
3.4 KiB
Go
// Copyright 2017 Vector Creations Ltd
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package consumers
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
|
|
"github.com/Shopify/sarama"
|
|
"github.com/matrix-org/dendrite/internal"
|
|
"github.com/matrix-org/dendrite/internal/eventutil"
|
|
"github.com/matrix-org/dendrite/setup/config"
|
|
"github.com/matrix-org/dendrite/setup/process"
|
|
"github.com/matrix-org/dendrite/syncapi/notifier"
|
|
"github.com/matrix-org/dendrite/syncapi/storage"
|
|
"github.com/matrix-org/dendrite/syncapi/types"
|
|
log "github.com/sirupsen/logrus"
|
|
)
|
|
|
|
// OutputClientDataConsumer consumes events that originated in the client API server.
|
|
type OutputClientDataConsumer struct {
|
|
clientAPIConsumer *internal.ContinualConsumer
|
|
db storage.Database
|
|
stream types.StreamProvider
|
|
notifier *notifier.Notifier
|
|
}
|
|
|
|
// NewOutputClientDataConsumer creates a new OutputClientData consumer. Call Start() to begin consuming from room servers.
|
|
func NewOutputClientDataConsumer(
|
|
process *process.ProcessContext,
|
|
cfg *config.SyncAPI,
|
|
kafkaConsumer sarama.Consumer,
|
|
store storage.Database,
|
|
notifier *notifier.Notifier,
|
|
stream types.StreamProvider,
|
|
) *OutputClientDataConsumer {
|
|
consumer := internal.ContinualConsumer{
|
|
Process: process,
|
|
ComponentName: "syncapi/clientapi",
|
|
Topic: string(cfg.Matrix.Kafka.TopicFor(config.TopicOutputClientData)),
|
|
Consumer: kafkaConsumer,
|
|
PartitionStore: store,
|
|
}
|
|
s := &OutputClientDataConsumer{
|
|
clientAPIConsumer: &consumer,
|
|
db: store,
|
|
notifier: notifier,
|
|
stream: stream,
|
|
}
|
|
consumer.ProcessMessage = s.onMessage
|
|
|
|
return s
|
|
}
|
|
|
|
// Start consuming from room servers
|
|
func (s *OutputClientDataConsumer) Start() error {
|
|
return s.clientAPIConsumer.Start()
|
|
}
|
|
|
|
// onMessage is called when the sync server receives a new event from the client API server output log.
|
|
// It is not safe for this function to be called from multiple goroutines, or else the
|
|
// sync stream position may race and be incorrectly calculated.
|
|
func (s *OutputClientDataConsumer) onMessage(msg *sarama.ConsumerMessage) error {
|
|
// Parse out the event JSON
|
|
var output eventutil.AccountData
|
|
if err := json.Unmarshal(msg.Value, &output); err != nil {
|
|
// If the message was invalid, log it and move on to the next message in the stream
|
|
log.WithError(err).Errorf("client API server output log: message parse failure")
|
|
return nil
|
|
}
|
|
|
|
log.WithFields(log.Fields{
|
|
"type": output.Type,
|
|
"room_id": output.RoomID,
|
|
}).Info("received data from client API server")
|
|
|
|
streamPos, err := s.db.UpsertAccountData(
|
|
context.TODO(), string(msg.Key), output.RoomID, output.Type,
|
|
)
|
|
if err != nil {
|
|
log.WithFields(log.Fields{
|
|
"type": output.Type,
|
|
"room_id": output.RoomID,
|
|
log.ErrorKey: err,
|
|
}).Panicf("could not save account data")
|
|
}
|
|
|
|
s.stream.Advance(streamPos)
|
|
s.notifier.OnNewAccountData(string(msg.Key), types.StreamingToken{AccountDataPosition: streamPos})
|
|
|
|
return nil
|
|
}
|