mirror of
https://github.com/1f349/dendrite.git
synced 2024-11-09 22:42:58 +00:00
a763cbb0e1
* Put federation client functions into their own file * Look for missing auth events in RS input * Remove retrieveMissingAuthEvents from federation API * Logging * Sorta transplanted the code over * Use event origin failing all else * Don't get stuck on mutexes: * Add verifier * Don't mark state events with zero snapshot NID as not existing * Check missing state if not an outlier before storing the event * Reject instead of soft-fail, don't copy roominfo so much * Use synchronous contexts, limit time to fetch missing events * Clean up some commented out bits * Simplify `/send` endpoint significantly * Submit async * Report errors on sending to RS input * Set max payload in NATS to 16MB * Tweak metrics * Add `workerForRoom` for tidiness * Try skipping unmarshalling errors for RespMissingEvents * Track missing prev events separately to avoid calculating state when not possible * Tweak logic around checking missing state * Care about state when checking missing prev events * Don't check missing state for create events * Try that again * Handle create events better * Send create room events as new * Use given event kind when sending auth/state events * Revert "Use given event kind when sending auth/state events" This reverts commit 089d64d271b5fca8c104e1554711187420dbebca. * Only search for missing prev events or state for new events * Tweaks * We only have missing prev if we don't supply state * Room version tweaks * Allow async inputs again * Apply backpressure to consumers/synchronous requests to hopefully stop things being overwhelmed * Set timeouts on roomserver input tasks (need to decide what timeout makes sense) * Use work queue policy, deliver all on restart * Reduce chance of duplicates being sent by NATS * Limit the number of servers we attempt to reduce backpressure * Some review comment fixes * Tidy up a couple things * Don't limit servers, randomise order using map * Some context refactoring * Update gmsl * Don't resend create events * Set stateIDs length correctly or else the roomserver thinks there are missing events when there aren't * Exclude our own servername * Try backing off servers * Make excluding self behaviour optional * Exclude self from g_m_e * Update sytest-whitelist * Update consumers for the roomserver output stream * Remember to send outliers for state returned from /gme * Make full HTTP tests less upsetti * Remove 'If a device list update goes missing, the server resyncs on the next one' from the sytest blacklist * Remove debugging test * Fix blacklist again, remove unnecessary duplicate context * Clearer contexts, don't use background in case there's something happening there * Don't queue up events more than once in memory * Correctly identify create events when checking for state * Fill in gaps again in /gme code * Remove `AuthEventIDs` from `InputRoomEvent` * Remove stray field Co-authored-by: Kegan Dougal <kegan@matrix.org>
98 lines
3.5 KiB
Go
98 lines
3.5 KiB
Go
package internal
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"time"
|
|
|
|
"github.com/matrix-org/dendrite/federationapi/api"
|
|
"github.com/matrix-org/gomatrixserverlib"
|
|
"github.com/matrix-org/util"
|
|
)
|
|
|
|
// QueryJoinedHostServerNamesInRoom implements api.FederationInternalAPI
|
|
func (f *FederationInternalAPI) QueryJoinedHostServerNamesInRoom(
|
|
ctx context.Context,
|
|
request *api.QueryJoinedHostServerNamesInRoomRequest,
|
|
response *api.QueryJoinedHostServerNamesInRoomResponse,
|
|
) (err error) {
|
|
joinedHosts, err := f.db.GetJoinedHostsForRooms(ctx, []string{request.RoomID}, request.ExcludeSelf)
|
|
if err != nil {
|
|
return
|
|
}
|
|
response.ServerNames = joinedHosts
|
|
|
|
return
|
|
}
|
|
|
|
func (a *FederationInternalAPI) fetchServerKeysDirectly(ctx context.Context, serverName gomatrixserverlib.ServerName) (*gomatrixserverlib.ServerKeys, error) {
|
|
ctx, cancel := context.WithTimeout(ctx, time.Second*30)
|
|
defer cancel()
|
|
ires, err := a.doRequestIfNotBackingOffOrBlacklisted(serverName, func() (interface{}, error) {
|
|
return a.federation.GetServerKeys(ctx, serverName)
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
sks := ires.(gomatrixserverlib.ServerKeys)
|
|
return &sks, nil
|
|
}
|
|
|
|
func (a *FederationInternalAPI) fetchServerKeysFromCache(
|
|
ctx context.Context, req *api.QueryServerKeysRequest,
|
|
) ([]gomatrixserverlib.ServerKeys, error) {
|
|
var results []gomatrixserverlib.ServerKeys
|
|
for keyID, criteria := range req.KeyIDToCriteria {
|
|
serverKeysResponses, _ := a.db.GetNotaryKeys(ctx, req.ServerName, []gomatrixserverlib.KeyID{keyID})
|
|
if len(serverKeysResponses) == 0 {
|
|
return nil, fmt.Errorf("failed to find server key response for key ID %s", keyID)
|
|
}
|
|
// we should only get 1 result as we only gave 1 key ID
|
|
sk := serverKeysResponses[0]
|
|
util.GetLogger(ctx).Infof("fetchServerKeysFromCache: minvalid:%v keys: %+v", criteria.MinimumValidUntilTS, sk)
|
|
if criteria.MinimumValidUntilTS != 0 {
|
|
// check if it's still valid. if they have the same value that's also valid
|
|
if sk.ValidUntilTS < criteria.MinimumValidUntilTS {
|
|
return nil, fmt.Errorf(
|
|
"found server response for key ID %s but it is no longer valid, min: %v valid_until: %v",
|
|
keyID, criteria.MinimumValidUntilTS, sk.ValidUntilTS,
|
|
)
|
|
}
|
|
}
|
|
results = append(results, sk)
|
|
}
|
|
return results, nil
|
|
}
|
|
|
|
func (a *FederationInternalAPI) QueryServerKeys(
|
|
ctx context.Context, req *api.QueryServerKeysRequest, res *api.QueryServerKeysResponse,
|
|
) error {
|
|
// attempt to satisfy the entire request from the cache first
|
|
results, err := a.fetchServerKeysFromCache(ctx, req)
|
|
if err == nil {
|
|
// satisfied entirely from cache, return it
|
|
res.ServerKeys = results
|
|
return nil
|
|
}
|
|
util.GetLogger(ctx).WithField("server", req.ServerName).WithError(err).Warn("notary: failed to satisfy keys request entirely from cache, hitting direct")
|
|
|
|
serverKeys, err := a.fetchServerKeysDirectly(ctx, req.ServerName)
|
|
if err != nil {
|
|
// try to load as much as we can from the cache in a best effort basis
|
|
util.GetLogger(ctx).WithField("server", req.ServerName).WithError(err).Warn("notary: failed to ask server for keys, returning best effort keys")
|
|
serverKeysResponses, dbErr := a.db.GetNotaryKeys(ctx, req.ServerName, req.KeyIDs())
|
|
if dbErr != nil {
|
|
return fmt.Errorf("notary: server returned %s, and db returned %s", err, dbErr)
|
|
}
|
|
res.ServerKeys = serverKeysResponses
|
|
return nil
|
|
}
|
|
// cache it!
|
|
if err = a.db.UpdateNotaryKeys(context.Background(), req.ServerName, *serverKeys); err != nil {
|
|
// non-fatal, still return the response
|
|
util.GetLogger(ctx).WithError(err).Warn("failed to UpdateNotaryKeys")
|
|
}
|
|
res.ServerKeys = []gomatrixserverlib.ServerKeys{*serverKeys}
|
|
return nil
|
|
}
|