mirror of
https://github.com/element-hq/synapse.git
synced 2024-12-27 09:58:41 +00:00
c5e89f5fae
Some checks are pending
Build docker images / build (push) Waiting to run
Deploy the documentation / Calculate variables for GitHub Pages deployment (push) Waiting to run
Deploy the documentation / GitHub Pages (push) Blocked by required conditions
Build release artifacts / Calculate list of debian distros (push) Waiting to run
Build release artifacts / Build .deb packages (push) Blocked by required conditions
Build release artifacts / Build wheels on ${{ matrix.os }} for ${{ matrix.arch }} (aarch64, ${{ startsWith(github.ref, 'refs/pull/') }}, ubuntu-22.04) (push) Waiting to run
Build release artifacts / Build wheels on ${{ matrix.os }} for ${{ matrix.arch }} (x86_64, ${{ startsWith(github.ref, 'refs/pull/') }}, macos-13) (push) Waiting to run
Build release artifacts / Build wheels on ${{ matrix.os }} for ${{ matrix.arch }} (x86_64, ${{ startsWith(github.ref, 'refs/pull/') }}, ubuntu-22.04) (push) Waiting to run
Build release artifacts / Build sdist (push) Waiting to run
Build release artifacts / Attach assets to release (push) Blocked by required conditions
Tests / changes (push) Waiting to run
Tests / lint-clippy-nightly (push) Blocked by required conditions
Tests / lint-rustfmt (push) Blocked by required conditions
Tests / lint-readme (push) Blocked by required conditions
Tests / check-sampleconfig (push) Blocked by required conditions
Tests / check-schema-delta (push) Blocked by required conditions
Tests / check-lockfile (push) Waiting to run
Tests / lint (push) Blocked by required conditions
Tests / Typechecking (push) Blocked by required conditions
Tests / lint-crlf (push) Waiting to run
Tests / lint-newsfile (push) Waiting to run
Tests / lint-pydantic (push) Blocked by required conditions
Tests / lint-clippy (push) Blocked by required conditions
Tests / linting-done (push) Blocked by required conditions
Tests / calculate-test-jobs (push) Blocked by required conditions
Tests / trial (push) Blocked by required conditions
Tests / trial-olddeps (push) Blocked by required conditions
Tests / trial-pypy (all, pypy-3.9) (push) Blocked by required conditions
Tests / sytest (push) Blocked by required conditions
Tests / export-data (push) Blocked by required conditions
Tests / portdb (11, 3.9) (push) Blocked by required conditions
Tests / portdb (17, 3.13) (push) Blocked by required conditions
Tests / complement (monolith, Postgres) (push) Blocked by required conditions
Tests / complement (monolith, SQLite) (push) Blocked by required conditions
Tests / complement (workers, Postgres) (push) Blocked by required conditions
Tests / cargo-test (push) Blocked by required conditions
Tests / cargo-bench (push) Blocked by required conditions
Tests / tests-done (push) Blocked by required conditions
To work around the fact that, pre-https://github.com/element-hq/synapse/pull/17903, our database may have old one-time-keys that the clients have long thrown away the private keys for, we want to delete OTKs that look like they came from libolm. To spread the load a bit, without holding up other background database updates, we use a scheduled task to do the work.
1899 lines
70 KiB
Python
1899 lines
70 KiB
Python
#
|
|
# This file is licensed under the Affero General Public License (AGPL) version 3.
|
|
#
|
|
# Copyright 2019 The Matrix.org Foundation C.I.C.
|
|
# Copyright 2016 OpenMarket Ltd
|
|
# Copyright (C) 2023 New Vector, Ltd
|
|
#
|
|
# This program is free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU Affero General Public License as
|
|
# published by the Free Software Foundation, either version 3 of the
|
|
# License, or (at your option) any later version.
|
|
#
|
|
# See the GNU Affero General Public License for more details:
|
|
# <https://www.gnu.org/licenses/agpl-3.0.html>.
|
|
#
|
|
# Originally licensed under the Apache License, Version 2.0:
|
|
# <http://www.apache.org/licenses/LICENSE-2.0>.
|
|
#
|
|
# [This file includes modifications made by New Vector Limited]
|
|
#
|
|
#
|
|
import time
|
|
from typing import Dict, Iterable
|
|
from unittest import mock
|
|
|
|
from parameterized import parameterized
|
|
from signedjson import key as key, sign as sign
|
|
|
|
from twisted.test.proto_helpers import MemoryReactor
|
|
|
|
from synapse.api.constants import RoomEncryptionAlgorithms
|
|
from synapse.api.errors import Codes, SynapseError
|
|
from synapse.appservice import ApplicationService
|
|
from synapse.handlers.device import DeviceHandler
|
|
from synapse.server import HomeServer
|
|
from synapse.storage.databases.main.appservice import _make_exclusive_regex
|
|
from synapse.types import JsonDict, UserID
|
|
from synapse.util import Clock
|
|
|
|
from tests import unittest
|
|
from tests.unittest import override_config
|
|
|
|
|
|
class E2eKeysHandlerTestCase(unittest.HomeserverTestCase):
|
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
|
self.appservice_api = mock.AsyncMock()
|
|
return self.setup_test_homeserver(application_service_api=self.appservice_api)
|
|
|
|
def prepare(self, reactor: MemoryReactor, clock: Clock, hs: HomeServer) -> None:
|
|
self.handler = hs.get_e2e_keys_handler()
|
|
self.store = self.hs.get_datastores().main
|
|
self.requester = UserID.from_string(f"@test_requester:{self.hs.hostname}")
|
|
|
|
def test_query_local_devices_no_devices(self) -> None:
|
|
"""If the user has no devices, we expect an empty list."""
|
|
local_user = "@boris:" + self.hs.hostname
|
|
res = self.get_success(self.handler.query_local_devices({local_user: None}))
|
|
self.assertDictEqual(res, {local_user: {}})
|
|
|
|
def test_reupload_one_time_keys(self) -> None:
|
|
"""we should be able to re-upload the same keys"""
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_id = "xyz"
|
|
keys: JsonDict = {
|
|
"alg1:k1": "key1",
|
|
"alg2:k2": {"key": "key2", "signatures": {"k1": "sig1"}},
|
|
"alg2:k3": {"key": "key3"},
|
|
}
|
|
|
|
# Note that "signed_curve25519" is always returned in key count responses. This is necessary until
|
|
# https://github.com/matrix-org/matrix-doc/issues/3298 is fixed.
|
|
res = self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"one_time_keys": keys}
|
|
)
|
|
)
|
|
self.assertDictEqual(
|
|
res, {"one_time_key_counts": {"alg1": 1, "alg2": 2, "signed_curve25519": 0}}
|
|
)
|
|
|
|
# we should be able to change the signature without a problem
|
|
keys["alg2:k2"]["signatures"]["k1"] = "sig2"
|
|
res = self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"one_time_keys": keys}
|
|
)
|
|
)
|
|
self.assertDictEqual(
|
|
res, {"one_time_key_counts": {"alg1": 1, "alg2": 2, "signed_curve25519": 0}}
|
|
)
|
|
|
|
def test_change_one_time_keys(self) -> None:
|
|
"""attempts to change one-time-keys should be rejected"""
|
|
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_id = "xyz"
|
|
keys = {
|
|
"alg1:k1": "key1",
|
|
"alg2:k2": {"key": "key2", "signatures": {"k1": "sig1"}},
|
|
"alg2:k3": {"key": "key3"},
|
|
}
|
|
|
|
res = self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"one_time_keys": keys}
|
|
)
|
|
)
|
|
self.assertDictEqual(
|
|
res, {"one_time_key_counts": {"alg1": 1, "alg2": 2, "signed_curve25519": 0}}
|
|
)
|
|
|
|
# Error when changing string key
|
|
self.get_failure(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"one_time_keys": {"alg1:k1": "key2"}}
|
|
),
|
|
SynapseError,
|
|
)
|
|
|
|
# Error when replacing dict key with string
|
|
self.get_failure(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"one_time_keys": {"alg2:k3": "key2"}}
|
|
),
|
|
SynapseError,
|
|
)
|
|
|
|
# Error when replacing string key with dict
|
|
self.get_failure(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id,
|
|
{"one_time_keys": {"alg1:k1": {"key": "key"}}},
|
|
),
|
|
SynapseError,
|
|
)
|
|
|
|
# Error when replacing dict key
|
|
self.get_failure(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id,
|
|
{
|
|
"one_time_keys": {
|
|
"alg2:k2": {"key": "key3", "signatures": {"k1": "sig1"}}
|
|
}
|
|
},
|
|
),
|
|
SynapseError,
|
|
)
|
|
|
|
def test_claim_one_time_key(self) -> None:
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_id = "xyz"
|
|
res = self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"one_time_keys": {"alg1:k1": "key1"}}
|
|
)
|
|
)
|
|
self.assertDictEqual(
|
|
res, {"one_time_key_counts": {"alg1": 1, "signed_curve25519": 0}}
|
|
)
|
|
|
|
# Keys should be returned in the order they were uploaded. To test, advance time
|
|
# a little, then upload a second key with an earlier key ID; it should get
|
|
# returned second.
|
|
self.reactor.advance(1)
|
|
res = self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"one_time_keys": {"alg1:k0": "key0"}}
|
|
)
|
|
)
|
|
self.assertDictEqual(
|
|
res, {"one_time_key_counts": {"alg1": 2, "signed_curve25519": 0}}
|
|
)
|
|
|
|
# now claim both keys back. They should be in the same order
|
|
res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
res,
|
|
{
|
|
"failures": {},
|
|
"one_time_keys": {local_user: {device_id: {"alg1:k1": "key1"}}},
|
|
},
|
|
)
|
|
res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
res,
|
|
{
|
|
"failures": {},
|
|
"one_time_keys": {local_user: {device_id: {"alg1:k0": "key0"}}},
|
|
},
|
|
)
|
|
|
|
def test_claim_one_time_key_bulk(self) -> None:
|
|
"""Like test_claim_one_time_key but claims multiple keys in one handler call."""
|
|
# Apologies to the reader. This test is a little too verbose. It is particularly
|
|
# tricky to make assertions neatly with all these nested dictionaries in play.
|
|
|
|
# Three users with two devices each. Each device uses two algorithms.
|
|
# Each algorithm is invoked with two keys.
|
|
alice = f"@alice:{self.hs.hostname}"
|
|
brian = f"@brian:{self.hs.hostname}"
|
|
chris = f"@chris:{self.hs.hostname}"
|
|
one_time_keys = {
|
|
alice: {
|
|
"alice_dev_1": {
|
|
"alg1:k1": {"dummy_id": 1},
|
|
"alg1:k2": {"dummy_id": 2},
|
|
"alg2:k3": {"dummy_id": 3},
|
|
"alg2:k4": {"dummy_id": 4},
|
|
},
|
|
"alice_dev_2": {
|
|
"alg1:k5": {"dummy_id": 5},
|
|
"alg1:k6": {"dummy_id": 6},
|
|
"alg2:k7": {"dummy_id": 7},
|
|
"alg2:k8": {"dummy_id": 8},
|
|
},
|
|
},
|
|
brian: {
|
|
"brian_dev_1": {
|
|
"alg1:k9": {"dummy_id": 9},
|
|
"alg1:k10": {"dummy_id": 10},
|
|
"alg2:k11": {"dummy_id": 11},
|
|
"alg2:k12": {"dummy_id": 12},
|
|
},
|
|
"brian_dev_2": {
|
|
"alg1:k13": {"dummy_id": 13},
|
|
"alg1:k14": {"dummy_id": 14},
|
|
"alg2:k15": {"dummy_id": 15},
|
|
"alg2:k16": {"dummy_id": 16},
|
|
},
|
|
},
|
|
chris: {
|
|
"chris_dev_1": {
|
|
"alg1:k17": {"dummy_id": 17},
|
|
"alg1:k18": {"dummy_id": 18},
|
|
"alg2:k19": {"dummy_id": 19},
|
|
"alg2:k20": {"dummy_id": 20},
|
|
},
|
|
"chris_dev_2": {
|
|
"alg1:k21": {"dummy_id": 21},
|
|
"alg1:k22": {"dummy_id": 22},
|
|
"alg2:k23": {"dummy_id": 23},
|
|
"alg2:k24": {"dummy_id": 24},
|
|
},
|
|
},
|
|
}
|
|
for user_id, devices in one_time_keys.items():
|
|
for device_id, keys_dict in devices.items():
|
|
counts = self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
user_id,
|
|
device_id,
|
|
{"one_time_keys": keys_dict},
|
|
)
|
|
)
|
|
# The upload should report 2 keys per algorithm.
|
|
expected_counts = {
|
|
"one_time_key_counts": {
|
|
# See count_e2e_one_time_keys for why this is hardcoded.
|
|
"signed_curve25519": 0,
|
|
"alg1": 2,
|
|
"alg2": 2,
|
|
},
|
|
}
|
|
self.assertEqual(counts, expected_counts)
|
|
|
|
# Claim a variety of keys.
|
|
# Raw format, easier to make test assertions about.
|
|
claims_to_make = {
|
|
(alice, "alice_dev_1", "alg1"): 1,
|
|
(alice, "alice_dev_1", "alg2"): 2,
|
|
(alice, "alice_dev_2", "alg2"): 1,
|
|
(brian, "brian_dev_1", "alg1"): 2,
|
|
(brian, "brian_dev_2", "alg2"): 9001,
|
|
(chris, "chris_dev_2", "alg2"): 1,
|
|
}
|
|
# Convert to the format the handler wants.
|
|
query: Dict[str, Dict[str, Dict[str, int]]] = {}
|
|
for (user_id, device_id, algorithm), count in claims_to_make.items():
|
|
query.setdefault(user_id, {}).setdefault(device_id, {})[algorithm] = count
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
query,
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
|
|
# No failures, please!
|
|
self.assertEqual(claim_res["failures"], {})
|
|
|
|
# Check that we get exactly the (user, device, algorithm)s we asked for.
|
|
got_otks = claim_res["one_time_keys"]
|
|
claimed_user_device_algorithms = {
|
|
(user_id, device_id, alg_key_id.split(":")[0])
|
|
for user_id, devices in got_otks.items()
|
|
for device_id, key_dict in devices.items()
|
|
for alg_key_id in key_dict
|
|
}
|
|
self.assertEqual(claimed_user_device_algorithms, set(claims_to_make))
|
|
|
|
# Now check the keys we got are what we expected.
|
|
def assertExactlyOneOtk(
|
|
user_id: str, device_id: str, *alg_key_pairs: str
|
|
) -> None:
|
|
key_dict = got_otks[user_id][device_id]
|
|
found = 0
|
|
for alg_key in alg_key_pairs:
|
|
if alg_key in key_dict:
|
|
expected_key_json = one_time_keys[user_id][device_id][alg_key]
|
|
self.assertEqual(key_dict[alg_key], expected_key_json)
|
|
found += 1
|
|
self.assertEqual(found, 1)
|
|
|
|
def assertAllOtks(user_id: str, device_id: str, *alg_key_pairs: str) -> None:
|
|
key_dict = got_otks[user_id][device_id]
|
|
for alg_key in alg_key_pairs:
|
|
expected_key_json = one_time_keys[user_id][device_id][alg_key]
|
|
self.assertEqual(key_dict[alg_key], expected_key_json)
|
|
|
|
# Expect a single arbitrary key to be returned.
|
|
assertExactlyOneOtk(alice, "alice_dev_1", "alg1:k1", "alg1:k2")
|
|
assertExactlyOneOtk(alice, "alice_dev_2", "alg2:k7", "alg2:k8")
|
|
assertExactlyOneOtk(chris, "chris_dev_2", "alg2:k23", "alg2:k24")
|
|
|
|
assertAllOtks(alice, "alice_dev_1", "alg2:k3", "alg2:k4")
|
|
assertAllOtks(brian, "brian_dev_1", "alg1:k9", "alg1:k10")
|
|
assertAllOtks(brian, "brian_dev_2", "alg2:k15", "alg2:k16")
|
|
|
|
# Now check the unused key counts.
|
|
for user_id, devices in one_time_keys.items():
|
|
for device_id in devices:
|
|
counts_by_alg = self.get_success(
|
|
self.store.count_e2e_one_time_keys(user_id, device_id)
|
|
)
|
|
# Somewhat fiddley to compute the expected count dict.
|
|
expected_counts_by_alg = {
|
|
"signed_curve25519": 0,
|
|
}
|
|
for alg in ["alg1", "alg2"]:
|
|
claim_count = claims_to_make.get((user_id, device_id, alg), 0)
|
|
remaining_count = max(0, 2 - claim_count)
|
|
if remaining_count > 0:
|
|
expected_counts_by_alg[alg] = remaining_count
|
|
|
|
self.assertEqual(
|
|
counts_by_alg, expected_counts_by_alg, f"{user_id}:{device_id}"
|
|
)
|
|
|
|
def test_claim_one_time_key_bulk_ordering(self) -> None:
|
|
"""Keys returned by the bulk claim call should be returned in the correct order"""
|
|
|
|
# Alice has lots of keys, uploaded in a specific order
|
|
alice = f"@alice:{self.hs.hostname}"
|
|
alice_dev = "alice_dev_1"
|
|
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
alice,
|
|
alice_dev,
|
|
{"one_time_keys": {"alg1:k20": 20, "alg1:k21": 21, "alg1:k22": 22}},
|
|
)
|
|
)
|
|
# Advance time by 1s, to ensure that there is a difference in upload time.
|
|
self.reactor.advance(1)
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
alice,
|
|
alice_dev,
|
|
{"one_time_keys": {"alg1:k10": 10, "alg1:k11": 11, "alg1:k12": 12}},
|
|
)
|
|
)
|
|
|
|
# Now claim some, and check we get the right ones.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{alice: {alice_dev: {"alg1": 2}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
# We should get the first-uploaded keys, even though they have later key ids.
|
|
# We should get a random set of two of k20, k21, k22.
|
|
self.assertEqual(claim_res["failures"], {})
|
|
claimed_keys = claim_res["one_time_keys"]["@alice:test"]["alice_dev_1"]
|
|
self.assertEqual(len(claimed_keys), 2)
|
|
for key_id in claimed_keys.keys():
|
|
self.assertIn(key_id, ["alg1:k20", "alg1:k21", "alg1:k22"])
|
|
|
|
def test_fallback_key(self) -> None:
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_id = "xyz"
|
|
fallback_key = {"alg1:k1": "fallback_key1"}
|
|
fallback_key2 = {"alg1:k2": "fallback_key2"}
|
|
fallback_key3 = {"alg1:k2": "fallback_key3"}
|
|
otk = {"alg1:k2": "key2"}
|
|
|
|
# we shouldn't have any unused fallback keys yet
|
|
res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(res, [])
|
|
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id,
|
|
{"fallback_keys": fallback_key},
|
|
)
|
|
)
|
|
|
|
# we should now have an unused alg1 key
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(fallback_res, ["alg1"])
|
|
|
|
# claiming an OTK when no OTKs are available should return the fallback
|
|
# key
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{"failures": {}, "one_time_keys": {local_user: {device_id: fallback_key}}},
|
|
)
|
|
|
|
# we shouldn't have any unused fallback keys again
|
|
unused_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(unused_res, [])
|
|
|
|
# claiming an OTK again should return the same fallback key
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{"failures": {}, "one_time_keys": {local_user: {device_id: fallback_key}}},
|
|
)
|
|
|
|
# re-uploading the same fallback key should still result in no unused fallback
|
|
# keys
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id,
|
|
{"fallback_keys": fallback_key},
|
|
)
|
|
)
|
|
|
|
unused_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(unused_res, [])
|
|
|
|
# uploading a new fallback key should result in an unused fallback key
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id,
|
|
{"fallback_keys": fallback_key2},
|
|
)
|
|
)
|
|
|
|
unused_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(unused_res, ["alg1"])
|
|
|
|
# if the user uploads a one-time key, the next claim should fetch the
|
|
# one-time key, and then go back to the fallback
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"one_time_keys": otk}
|
|
)
|
|
)
|
|
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{"failures": {}, "one_time_keys": {local_user: {device_id: otk}}},
|
|
)
|
|
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{"failures": {}, "one_time_keys": {local_user: {device_id: fallback_key2}}},
|
|
)
|
|
|
|
# using the unstable prefix should also set the fallback key
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id,
|
|
{"org.matrix.msc2732.fallback_keys": fallback_key3},
|
|
)
|
|
)
|
|
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{"failures": {}, "one_time_keys": {local_user: {device_id: fallback_key3}}},
|
|
)
|
|
|
|
def test_fallback_key_bulk(self) -> None:
|
|
"""Like test_fallback_key, but claims multiple keys in one handler call."""
|
|
alice = f"@alice:{self.hs.hostname}"
|
|
brian = f"@brian:{self.hs.hostname}"
|
|
chris = f"@chris:{self.hs.hostname}"
|
|
|
|
# Have three users upload fallback keys for two devices.
|
|
fallback_keys = {
|
|
alice: {
|
|
"alice_dev_1": {"alg1:k1": "fallback_key1"},
|
|
"alice_dev_2": {"alg2:k2": "fallback_key2"},
|
|
},
|
|
brian: {
|
|
"brian_dev_1": {"alg1:k3": "fallback_key3"},
|
|
"brian_dev_2": {"alg2:k4": "fallback_key4"},
|
|
},
|
|
chris: {
|
|
"chris_dev_1": {"alg1:k5": "fallback_key5"},
|
|
"chris_dev_2": {"alg2:k6": "fallback_key6"},
|
|
},
|
|
}
|
|
|
|
for user_id, devices in fallback_keys.items():
|
|
for device_id, key_dict in devices.items():
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
user_id,
|
|
device_id,
|
|
{"fallback_keys": key_dict},
|
|
)
|
|
)
|
|
|
|
# Each device should have an unused fallback key.
|
|
for user_id, devices in fallback_keys.items():
|
|
for device_id in devices:
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(user_id, device_id)
|
|
)
|
|
expected_algorithm_name = f"alg{device_id[-1]}"
|
|
self.assertEqual(fallback_res, [expected_algorithm_name])
|
|
|
|
# Claim the fallback key for one device per user.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{
|
|
alice: {"alice_dev_1": {"alg1": 1}},
|
|
brian: {"brian_dev_2": {"alg2": 1}},
|
|
chris: {"chris_dev_2": {"alg2": 1}},
|
|
},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
expected_claims = {
|
|
alice: {"alice_dev_1": {"alg1:k1": "fallback_key1"}},
|
|
brian: {"brian_dev_2": {"alg2:k4": "fallback_key4"}},
|
|
chris: {"chris_dev_2": {"alg2:k6": "fallback_key6"}},
|
|
}
|
|
self.assertEqual(
|
|
claim_res,
|
|
{"failures": {}, "one_time_keys": expected_claims},
|
|
)
|
|
|
|
for user_id, devices in fallback_keys.items():
|
|
for device_id in devices:
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(user_id, device_id)
|
|
)
|
|
# Claimed fallback keys should no longer show up as unused.
|
|
# Unclaimed fallback keys should still be unused.
|
|
if device_id in expected_claims[user_id]:
|
|
self.assertEqual(fallback_res, [])
|
|
else:
|
|
expected_algorithm_name = f"alg{device_id[-1]}"
|
|
self.assertEqual(fallback_res, [expected_algorithm_name])
|
|
|
|
def test_fallback_key_always_returned(self) -> None:
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_id = "xyz"
|
|
fallback_key = {"alg1:k1": "fallback_key1"}
|
|
otk = {"alg1:k2": "key2"}
|
|
|
|
# we shouldn't have any unused fallback keys yet
|
|
res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(res, [])
|
|
|
|
# Upload a OTK & fallback key.
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id,
|
|
{"one_time_keys": otk, "fallback_keys": fallback_key},
|
|
)
|
|
)
|
|
|
|
# we should now have an unused alg1 key
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(fallback_res, ["alg1"])
|
|
|
|
# Claiming an OTK and requesting to always return the fallback key should
|
|
# return both.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=True,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{
|
|
"failures": {},
|
|
"one_time_keys": {local_user: {device_id: {**fallback_key, **otk}}},
|
|
},
|
|
)
|
|
|
|
# This should not mark the key as used.
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(fallback_res, ["alg1"])
|
|
|
|
# Claiming an OTK again should return only the fallback key.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=True,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{"failures": {}, "one_time_keys": {local_user: {device_id: fallback_key}}},
|
|
)
|
|
|
|
# And mark it as used.
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id)
|
|
)
|
|
self.assertEqual(fallback_res, [])
|
|
|
|
def test_replace_master_key(self) -> None:
|
|
"""uploading a new signing key should make the old signing key unavailable"""
|
|
local_user = "@boris:" + self.hs.hostname
|
|
keys1 = {
|
|
"master_key": {
|
|
# private key: 2lonYOM6xYKdEsO+6KrC766xBcHnYnim1x/4LFGF8B0
|
|
"user_id": local_user,
|
|
"usage": ["master"],
|
|
"keys": {
|
|
"ed25519:nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk": "nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk"
|
|
},
|
|
}
|
|
}
|
|
self.get_success(self.handler.upload_signing_keys_for_user(local_user, keys1))
|
|
|
|
keys2 = {
|
|
"master_key": {
|
|
# private key: 4TL4AjRYwDVwD3pqQzcor+ez/euOB1/q78aTJ+czDNs
|
|
"user_id": local_user,
|
|
"usage": ["master"],
|
|
"keys": {
|
|
"ed25519:Hq6gL+utB4ET+UvD5ci0kgAwsX6qP/zvf8v6OInU5iw": "Hq6gL+utB4ET+UvD5ci0kgAwsX6qP/zvf8v6OInU5iw"
|
|
},
|
|
}
|
|
}
|
|
self.get_success(self.handler.upload_signing_keys_for_user(local_user, keys2))
|
|
|
|
devices = self.get_success(
|
|
self.handler.query_devices(
|
|
{"device_keys": {local_user: []}}, 0, local_user, "device123"
|
|
)
|
|
)
|
|
self.assertDictEqual(devices["master_keys"], {local_user: keys2["master_key"]})
|
|
|
|
def test_reupload_signatures(self) -> None:
|
|
"""re-uploading a signature should not fail"""
|
|
local_user = "@boris:" + self.hs.hostname
|
|
keys1 = {
|
|
"master_key": {
|
|
# private key: HvQBbU+hc2Zr+JP1sE0XwBe1pfZZEYtJNPJLZJtS+F8
|
|
"user_id": local_user,
|
|
"usage": ["master"],
|
|
"keys": {
|
|
"ed25519:EmkqvokUn8p+vQAGZitOk4PWjp7Ukp3txV2TbMPEiBQ": "EmkqvokUn8p+vQAGZitOk4PWjp7Ukp3txV2TbMPEiBQ"
|
|
},
|
|
},
|
|
"self_signing_key": {
|
|
# private key: 2lonYOM6xYKdEsO+6KrC766xBcHnYnim1x/4LFGF8B0
|
|
"user_id": local_user,
|
|
"usage": ["self_signing"],
|
|
"keys": {
|
|
"ed25519:nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk": "nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk"
|
|
},
|
|
},
|
|
}
|
|
master_signing_key = key.decode_signing_key_base64(
|
|
"ed25519",
|
|
"EmkqvokUn8p+vQAGZitOk4PWjp7Ukp3txV2TbMPEiBQ",
|
|
"HvQBbU+hc2Zr+JP1sE0XwBe1pfZZEYtJNPJLZJtS+F8",
|
|
)
|
|
sign.sign_json(keys1["self_signing_key"], local_user, master_signing_key)
|
|
signing_key = key.decode_signing_key_base64(
|
|
"ed25519",
|
|
"nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk",
|
|
"2lonYOM6xYKdEsO+6KrC766xBcHnYnim1x/4LFGF8B0",
|
|
)
|
|
self.get_success(self.handler.upload_signing_keys_for_user(local_user, keys1))
|
|
|
|
# upload two device keys, which will be signed later by the self-signing key
|
|
device_key_1: JsonDict = {
|
|
"user_id": local_user,
|
|
"device_id": "abc",
|
|
"algorithms": [
|
|
"m.olm.curve25519-aes-sha2",
|
|
RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2,
|
|
],
|
|
"keys": {
|
|
"ed25519:abc": "base64+ed25519+key",
|
|
"curve25519:abc": "base64+curve25519+key",
|
|
},
|
|
"signatures": {local_user: {"ed25519:abc": "base64+signature"}},
|
|
}
|
|
device_key_2: JsonDict = {
|
|
"user_id": local_user,
|
|
"device_id": "def",
|
|
"algorithms": [
|
|
"m.olm.curve25519-aes-sha2",
|
|
RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2,
|
|
],
|
|
"keys": {
|
|
"ed25519:def": "base64+ed25519+key",
|
|
"curve25519:def": "base64+curve25519+key",
|
|
},
|
|
"signatures": {local_user: {"ed25519:def": "base64+signature"}},
|
|
}
|
|
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, "abc", {"device_keys": device_key_1}
|
|
)
|
|
)
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, "def", {"device_keys": device_key_2}
|
|
)
|
|
)
|
|
|
|
# sign the first device key and upload it
|
|
del device_key_1["signatures"]
|
|
sign.sign_json(device_key_1, local_user, signing_key)
|
|
self.get_success(
|
|
self.handler.upload_signatures_for_device_keys(
|
|
local_user, {local_user: {"abc": device_key_1}}
|
|
)
|
|
)
|
|
|
|
# sign the second device key and upload both device keys. The server
|
|
# should ignore the first device key since it already has a valid
|
|
# signature for it
|
|
del device_key_2["signatures"]
|
|
sign.sign_json(device_key_2, local_user, signing_key)
|
|
self.get_success(
|
|
self.handler.upload_signatures_for_device_keys(
|
|
local_user, {local_user: {"abc": device_key_1, "def": device_key_2}}
|
|
)
|
|
)
|
|
|
|
device_key_1["signatures"][local_user]["ed25519:abc"] = "base64+signature"
|
|
device_key_2["signatures"][local_user]["ed25519:def"] = "base64+signature"
|
|
devices = self.get_success(
|
|
self.handler.query_devices(
|
|
{"device_keys": {local_user: []}}, 0, local_user, "device123"
|
|
)
|
|
)
|
|
del devices["device_keys"][local_user]["abc"]["unsigned"]
|
|
del devices["device_keys"][local_user]["def"]["unsigned"]
|
|
self.assertDictEqual(devices["device_keys"][local_user]["abc"], device_key_1)
|
|
self.assertDictEqual(devices["device_keys"][local_user]["def"], device_key_2)
|
|
|
|
def test_self_signing_key_doesnt_show_up_as_device(self) -> None:
|
|
"""signing keys should be hidden when fetching a user's devices"""
|
|
local_user = "@boris:" + self.hs.hostname
|
|
keys1 = {
|
|
"master_key": {
|
|
# private key: 2lonYOM6xYKdEsO+6KrC766xBcHnYnim1x/4LFGF8B0
|
|
"user_id": local_user,
|
|
"usage": ["master"],
|
|
"keys": {
|
|
"ed25519:nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk": "nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk"
|
|
},
|
|
}
|
|
}
|
|
self.get_success(self.handler.upload_signing_keys_for_user(local_user, keys1))
|
|
|
|
device_handler = self.hs.get_device_handler()
|
|
assert isinstance(device_handler, DeviceHandler)
|
|
e = self.get_failure(
|
|
device_handler.check_device_registered(
|
|
user_id=local_user,
|
|
device_id="nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk",
|
|
initial_device_display_name="new display name",
|
|
),
|
|
SynapseError,
|
|
)
|
|
res = e.value.code
|
|
self.assertEqual(res, 400)
|
|
|
|
query_res = self.get_success(
|
|
self.handler.query_local_devices({local_user: None})
|
|
)
|
|
self.assertDictEqual(query_res, {local_user: {}})
|
|
|
|
def test_upload_signatures(self) -> None:
|
|
"""should check signatures that are uploaded"""
|
|
# set up a user with cross-signing keys and a device. This user will
|
|
# try uploading signatures
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_id = "xyz"
|
|
# private key: OMkooTr76ega06xNvXIGPbgvvxAOzmQncN8VObS7aBA
|
|
device_pubkey = "NnHhnqiMFQkq969szYkooLaBAXW244ZOxgukCvm2ZeY"
|
|
device_key: JsonDict = {
|
|
"user_id": local_user,
|
|
"device_id": device_id,
|
|
"algorithms": [
|
|
"m.olm.curve25519-aes-sha2",
|
|
RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2,
|
|
],
|
|
"keys": {"curve25519:xyz": "curve25519+key", "ed25519:xyz": device_pubkey},
|
|
"signatures": {local_user: {"ed25519:xyz": "something"}},
|
|
}
|
|
device_signing_key = key.decode_signing_key_base64(
|
|
"ed25519", "xyz", "OMkooTr76ega06xNvXIGPbgvvxAOzmQncN8VObS7aBA"
|
|
)
|
|
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_id, {"device_keys": device_key}
|
|
)
|
|
)
|
|
|
|
# private key: 2lonYOM6xYKdEsO+6KrC766xBcHnYnim1x/4LFGF8B0
|
|
master_pubkey = "nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk"
|
|
master_key: JsonDict = {
|
|
"user_id": local_user,
|
|
"usage": ["master"],
|
|
"keys": {"ed25519:" + master_pubkey: master_pubkey},
|
|
}
|
|
master_signing_key = key.decode_signing_key_base64(
|
|
"ed25519", master_pubkey, "2lonYOM6xYKdEsO+6KrC766xBcHnYnim1x/4LFGF8B0"
|
|
)
|
|
usersigning_pubkey = "Hq6gL+utB4ET+UvD5ci0kgAwsX6qP/zvf8v6OInU5iw"
|
|
usersigning_key = {
|
|
# private key: 4TL4AjRYwDVwD3pqQzcor+ez/euOB1/q78aTJ+czDNs
|
|
"user_id": local_user,
|
|
"usage": ["user_signing"],
|
|
"keys": {"ed25519:" + usersigning_pubkey: usersigning_pubkey},
|
|
}
|
|
usersigning_signing_key = key.decode_signing_key_base64(
|
|
"ed25519", usersigning_pubkey, "4TL4AjRYwDVwD3pqQzcor+ez/euOB1/q78aTJ+czDNs"
|
|
)
|
|
sign.sign_json(usersigning_key, local_user, master_signing_key)
|
|
# private key: HvQBbU+hc2Zr+JP1sE0XwBe1pfZZEYtJNPJLZJtS+F8
|
|
selfsigning_pubkey = "EmkqvokUn8p+vQAGZitOk4PWjp7Ukp3txV2TbMPEiBQ"
|
|
selfsigning_key = {
|
|
"user_id": local_user,
|
|
"usage": ["self_signing"],
|
|
"keys": {"ed25519:" + selfsigning_pubkey: selfsigning_pubkey},
|
|
}
|
|
selfsigning_signing_key = key.decode_signing_key_base64(
|
|
"ed25519", selfsigning_pubkey, "HvQBbU+hc2Zr+JP1sE0XwBe1pfZZEYtJNPJLZJtS+F8"
|
|
)
|
|
sign.sign_json(selfsigning_key, local_user, master_signing_key)
|
|
cross_signing_keys = {
|
|
"master_key": master_key,
|
|
"user_signing_key": usersigning_key,
|
|
"self_signing_key": selfsigning_key,
|
|
}
|
|
self.get_success(
|
|
self.handler.upload_signing_keys_for_user(local_user, cross_signing_keys)
|
|
)
|
|
|
|
# set up another user with a master key. This user will be signed by
|
|
# the first user
|
|
other_user = "@otherboris:" + self.hs.hostname
|
|
other_master_pubkey = "fHZ3NPiKxoLQm5OoZbKa99SYxprOjNs4TwJUKP+twCM"
|
|
other_master_key: JsonDict = {
|
|
# private key: oyw2ZUx0O4GifbfFYM0nQvj9CL0b8B7cyN4FprtK8OI
|
|
"user_id": other_user,
|
|
"usage": ["master"],
|
|
"keys": {"ed25519:" + other_master_pubkey: other_master_pubkey},
|
|
}
|
|
self.get_success(
|
|
self.handler.upload_signing_keys_for_user(
|
|
other_user, {"master_key": other_master_key}
|
|
)
|
|
)
|
|
|
|
# test various signature failures (see below)
|
|
ret = self.get_success(
|
|
self.handler.upload_signatures_for_device_keys(
|
|
local_user,
|
|
{
|
|
local_user: {
|
|
# fails because the signature is invalid
|
|
# should fail with INVALID_SIGNATURE
|
|
device_id: {
|
|
"user_id": local_user,
|
|
"device_id": device_id,
|
|
"algorithms": [
|
|
"m.olm.curve25519-aes-sha2",
|
|
RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2,
|
|
],
|
|
"keys": {
|
|
"curve25519:xyz": "curve25519+key",
|
|
# private key: OMkooTr76ega06xNvXIGPbgvvxAOzmQncN8VObS7aBA
|
|
"ed25519:xyz": device_pubkey,
|
|
},
|
|
"signatures": {
|
|
local_user: {
|
|
"ed25519:" + selfsigning_pubkey: "something"
|
|
}
|
|
},
|
|
},
|
|
# fails because device is unknown
|
|
# should fail with NOT_FOUND
|
|
"unknown": {
|
|
"user_id": local_user,
|
|
"device_id": "unknown",
|
|
"signatures": {
|
|
local_user: {
|
|
"ed25519:" + selfsigning_pubkey: "something"
|
|
}
|
|
},
|
|
},
|
|
# fails because the signature is invalid
|
|
# should fail with INVALID_SIGNATURE
|
|
master_pubkey: {
|
|
"user_id": local_user,
|
|
"usage": ["master"],
|
|
"keys": {"ed25519:" + master_pubkey: master_pubkey},
|
|
"signatures": {
|
|
local_user: {"ed25519:" + device_pubkey: "something"}
|
|
},
|
|
},
|
|
},
|
|
other_user: {
|
|
# fails because the device is not the user's master-signing key
|
|
# should fail with NOT_FOUND
|
|
"unknown": {
|
|
"user_id": other_user,
|
|
"device_id": "unknown",
|
|
"signatures": {
|
|
local_user: {
|
|
"ed25519:" + usersigning_pubkey: "something"
|
|
}
|
|
},
|
|
},
|
|
other_master_pubkey: {
|
|
# fails because the key doesn't match what the server has
|
|
# should fail with UNKNOWN
|
|
"user_id": other_user,
|
|
"usage": ["master"],
|
|
"keys": {
|
|
"ed25519:" + other_master_pubkey: other_master_pubkey
|
|
},
|
|
"something": "random",
|
|
"signatures": {
|
|
local_user: {
|
|
"ed25519:" + usersigning_pubkey: "something"
|
|
}
|
|
},
|
|
},
|
|
},
|
|
},
|
|
)
|
|
)
|
|
|
|
user_failures = ret["failures"][local_user]
|
|
self.assertEqual(user_failures[device_id]["errcode"], Codes.INVALID_SIGNATURE)
|
|
self.assertEqual(
|
|
user_failures[master_pubkey]["errcode"], Codes.INVALID_SIGNATURE
|
|
)
|
|
self.assertEqual(user_failures["unknown"]["errcode"], Codes.NOT_FOUND)
|
|
|
|
other_user_failures = ret["failures"][other_user]
|
|
self.assertEqual(other_user_failures["unknown"]["errcode"], Codes.NOT_FOUND)
|
|
self.assertEqual(
|
|
other_user_failures[other_master_pubkey]["errcode"], Codes.UNKNOWN
|
|
)
|
|
|
|
# test successful signatures
|
|
del device_key["signatures"]
|
|
sign.sign_json(device_key, local_user, selfsigning_signing_key)
|
|
sign.sign_json(master_key, local_user, device_signing_key)
|
|
sign.sign_json(other_master_key, local_user, usersigning_signing_key)
|
|
ret = self.get_success(
|
|
self.handler.upload_signatures_for_device_keys(
|
|
local_user,
|
|
{
|
|
local_user: {device_id: device_key, master_pubkey: master_key},
|
|
other_user: {other_master_pubkey: other_master_key},
|
|
},
|
|
)
|
|
)
|
|
|
|
self.assertEqual(ret["failures"], {})
|
|
|
|
# fetch the signed keys/devices and make sure that the signatures are there
|
|
ret = self.get_success(
|
|
self.handler.query_devices(
|
|
{"device_keys": {local_user: [], other_user: []}},
|
|
0,
|
|
local_user,
|
|
"device123",
|
|
)
|
|
)
|
|
|
|
self.assertEqual(
|
|
ret["device_keys"][local_user]["xyz"]["signatures"][local_user][
|
|
"ed25519:" + selfsigning_pubkey
|
|
],
|
|
device_key["signatures"][local_user]["ed25519:" + selfsigning_pubkey],
|
|
)
|
|
self.assertEqual(
|
|
ret["master_keys"][local_user]["signatures"][local_user][
|
|
"ed25519:" + device_id
|
|
],
|
|
master_key["signatures"][local_user]["ed25519:" + device_id],
|
|
)
|
|
self.assertEqual(
|
|
ret["master_keys"][other_user]["signatures"][local_user][
|
|
"ed25519:" + usersigning_pubkey
|
|
],
|
|
other_master_key["signatures"][local_user]["ed25519:" + usersigning_pubkey],
|
|
)
|
|
|
|
def test_query_devices_remote_no_sync(self) -> None:
|
|
"""Tests that querying keys for a remote user that we don't share a room
|
|
with returns the cross signing keys correctly.
|
|
"""
|
|
|
|
remote_user_id = "@test:other"
|
|
local_user_id = "@test:test"
|
|
|
|
remote_master_key = "85T7JXPFBAySB/jwby4S3lBPTqY3+Zg53nYuGmu1ggY"
|
|
remote_self_signing_key = "QeIiFEjluPBtI7WQdG365QKZcFs9kqmHir6RBD0//nQ"
|
|
|
|
self.hs.get_federation_client().query_client_keys = mock.AsyncMock( # type: ignore[method-assign]
|
|
return_value={
|
|
"device_keys": {remote_user_id: {}},
|
|
"master_keys": {
|
|
remote_user_id: {
|
|
"user_id": remote_user_id,
|
|
"usage": ["master"],
|
|
"keys": {"ed25519:" + remote_master_key: remote_master_key},
|
|
},
|
|
},
|
|
"self_signing_keys": {
|
|
remote_user_id: {
|
|
"user_id": remote_user_id,
|
|
"usage": ["self_signing"],
|
|
"keys": {
|
|
"ed25519:"
|
|
+ remote_self_signing_key: remote_self_signing_key
|
|
},
|
|
}
|
|
},
|
|
}
|
|
)
|
|
|
|
e2e_handler = self.hs.get_e2e_keys_handler()
|
|
|
|
query_result = self.get_success(
|
|
e2e_handler.query_devices(
|
|
{
|
|
"device_keys": {remote_user_id: []},
|
|
},
|
|
timeout=10,
|
|
from_user_id=local_user_id,
|
|
from_device_id="some_device_id",
|
|
)
|
|
)
|
|
|
|
self.assertEqual(query_result["failures"], {})
|
|
self.assertEqual(
|
|
query_result["master_keys"],
|
|
{
|
|
remote_user_id: {
|
|
"user_id": remote_user_id,
|
|
"usage": ["master"],
|
|
"keys": {"ed25519:" + remote_master_key: remote_master_key},
|
|
},
|
|
},
|
|
)
|
|
self.assertEqual(
|
|
query_result["self_signing_keys"],
|
|
{
|
|
remote_user_id: {
|
|
"user_id": remote_user_id,
|
|
"usage": ["self_signing"],
|
|
"keys": {
|
|
"ed25519:" + remote_self_signing_key: remote_self_signing_key
|
|
},
|
|
}
|
|
},
|
|
)
|
|
|
|
def test_has_different_keys(self) -> None:
|
|
"""check that has_different_keys returns True when the keys provided are different to what
|
|
is in the database."""
|
|
local_user = "@boris:" + self.hs.hostname
|
|
keys1 = {
|
|
"master_key": {
|
|
# private key: 2lonYOM6xYKdEsO+6KrC766xBcHnYnim1x/4LFGF8B0
|
|
"user_id": local_user,
|
|
"usage": ["master"],
|
|
"keys": {
|
|
"ed25519:nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk": "nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unI9kDYcHwk"
|
|
},
|
|
}
|
|
}
|
|
self.get_success(self.handler.upload_signing_keys_for_user(local_user, keys1))
|
|
is_different = self.get_success(
|
|
self.handler.has_different_keys(
|
|
local_user,
|
|
{
|
|
"master_key": keys1["master_key"],
|
|
},
|
|
)
|
|
)
|
|
self.assertEqual(is_different, False)
|
|
# change the usage => different keys
|
|
keys1["master_key"]["usage"] = ["develop"]
|
|
is_different = self.get_success(
|
|
self.handler.has_different_keys(
|
|
local_user,
|
|
{
|
|
"master_key": keys1["master_key"],
|
|
},
|
|
)
|
|
)
|
|
self.assertEqual(is_different, True)
|
|
keys1["master_key"]["usage"] = ["master"] # reset
|
|
# change the key => different keys
|
|
keys1["master_key"]["keys"] = {
|
|
"ed25519:nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unIc0rncs": "nqOvzeuGWT/sRx3h7+MHoInYj3Uk2LD/unIc0rncs"
|
|
}
|
|
is_different = self.get_success(
|
|
self.handler.has_different_keys(
|
|
local_user,
|
|
{
|
|
"master_key": keys1["master_key"],
|
|
},
|
|
)
|
|
)
|
|
self.assertEqual(is_different, True)
|
|
|
|
def test_query_devices_remote_sync(self) -> None:
|
|
"""Tests that querying keys for a remote user that we share a room with,
|
|
but haven't yet fetched the keys for, returns the cross signing keys
|
|
correctly.
|
|
"""
|
|
|
|
remote_user_id = "@test:other"
|
|
local_user_id = "@test:test"
|
|
|
|
# Pretend we're sharing a room with the user we're querying. If not,
|
|
# `_query_devices_for_destination` will return early.
|
|
self.store.get_rooms_for_user = mock.AsyncMock(return_value={"some_room_id"})
|
|
|
|
remote_master_key = "85T7JXPFBAySB/jwby4S3lBPTqY3+Zg53nYuGmu1ggY"
|
|
remote_self_signing_key = "QeIiFEjluPBtI7WQdG365QKZcFs9kqmHir6RBD0//nQ"
|
|
|
|
self.hs.get_federation_client().query_user_devices = mock.AsyncMock( # type: ignore[method-assign]
|
|
return_value={
|
|
"user_id": remote_user_id,
|
|
"stream_id": 1,
|
|
"devices": [],
|
|
"master_key": {
|
|
"user_id": remote_user_id,
|
|
"usage": ["master"],
|
|
"keys": {"ed25519:" + remote_master_key: remote_master_key},
|
|
},
|
|
"self_signing_key": {
|
|
"user_id": remote_user_id,
|
|
"usage": ["self_signing"],
|
|
"keys": {
|
|
"ed25519:" + remote_self_signing_key: remote_self_signing_key
|
|
},
|
|
},
|
|
}
|
|
)
|
|
|
|
e2e_handler = self.hs.get_e2e_keys_handler()
|
|
|
|
query_result = self.get_success(
|
|
e2e_handler.query_devices(
|
|
{
|
|
"device_keys": {remote_user_id: []},
|
|
},
|
|
timeout=10,
|
|
from_user_id=local_user_id,
|
|
from_device_id="some_device_id",
|
|
)
|
|
)
|
|
|
|
self.assertEqual(query_result["failures"], {})
|
|
self.assertEqual(
|
|
query_result["master_keys"],
|
|
{
|
|
remote_user_id: {
|
|
"user_id": remote_user_id,
|
|
"usage": ["master"],
|
|
"keys": {"ed25519:" + remote_master_key: remote_master_key},
|
|
}
|
|
},
|
|
)
|
|
self.assertEqual(
|
|
query_result["self_signing_keys"],
|
|
{
|
|
remote_user_id: {
|
|
"user_id": remote_user_id,
|
|
"usage": ["self_signing"],
|
|
"keys": {
|
|
"ed25519:" + remote_self_signing_key: remote_self_signing_key
|
|
},
|
|
}
|
|
},
|
|
)
|
|
|
|
def test_query_devices_remote_down(self) -> None:
|
|
"""Tests that querying keys for a remote user on an unreachable server returns
|
|
results in the "failures" property
|
|
"""
|
|
|
|
remote_user_id = "@test:other"
|
|
local_user_id = "@test:test"
|
|
|
|
# The backoff code treats time zero as special
|
|
self.reactor.advance(5)
|
|
|
|
self.hs.get_federation_http_client().agent.request = mock.AsyncMock( # type: ignore[method-assign]
|
|
side_effect=Exception("boop")
|
|
)
|
|
|
|
e2e_handler = self.hs.get_e2e_keys_handler()
|
|
|
|
query_result = self.get_success(
|
|
e2e_handler.query_devices(
|
|
{
|
|
"device_keys": {remote_user_id: []},
|
|
},
|
|
timeout=10,
|
|
from_user_id=local_user_id,
|
|
from_device_id="some_device_id",
|
|
)
|
|
)
|
|
|
|
self.assertEqual(
|
|
query_result["failures"],
|
|
{
|
|
"other": {
|
|
"message": "Failed to send request: Exception: boop",
|
|
"status": 503,
|
|
}
|
|
},
|
|
)
|
|
|
|
# Do it again: we should hit the backoff
|
|
query_result = self.get_success(
|
|
e2e_handler.query_devices(
|
|
{
|
|
"device_keys": {remote_user_id: []},
|
|
},
|
|
timeout=10,
|
|
from_user_id=local_user_id,
|
|
from_device_id="some_device_id",
|
|
)
|
|
)
|
|
|
|
self.assertEqual(
|
|
query_result["failures"],
|
|
{"other": {"message": "Not ready for retry", "status": 503}},
|
|
)
|
|
|
|
@parameterized.expand(
|
|
[
|
|
# The remote homeserver's response indicates that this user has 0/1/2 devices.
|
|
([],),
|
|
(["device_1"],),
|
|
(["device_1", "device_2"],),
|
|
]
|
|
)
|
|
def test_query_all_devices_caches_result(self, device_ids: Iterable[str]) -> None:
|
|
"""Test that requests for all of a remote user's devices are cached.
|
|
|
|
We do this by asserting that only one call over federation was made, and that
|
|
the two queries to the local homeserver produce the same response.
|
|
"""
|
|
local_user_id = "@test:test"
|
|
remote_user_id = "@test:other"
|
|
request_body: JsonDict = {"device_keys": {remote_user_id: []}}
|
|
|
|
response_devices = [
|
|
{
|
|
"device_id": device_id,
|
|
"keys": {
|
|
"algorithms": ["dummy"],
|
|
"device_id": device_id,
|
|
"keys": {f"dummy:{device_id}": "dummy"},
|
|
"signatures": {device_id: {f"dummy:{device_id}": "dummy"}},
|
|
"unsigned": {},
|
|
"user_id": "@test:other",
|
|
},
|
|
}
|
|
for device_id in device_ids
|
|
]
|
|
|
|
response_body = {
|
|
"devices": response_devices,
|
|
"user_id": remote_user_id,
|
|
"stream_id": 12345, # an integer, according to the spec
|
|
}
|
|
|
|
e2e_handler = self.hs.get_e2e_keys_handler()
|
|
|
|
# Pretend we're sharing a room with the user we're querying. If not,
|
|
# `_query_devices_for_destination` will return early.
|
|
mock_get_rooms = mock.patch.object(
|
|
self.store,
|
|
"get_rooms_for_user",
|
|
new_callable=mock.AsyncMock,
|
|
return_value=["some_room_id"],
|
|
)
|
|
mock_get_users = mock.patch.object(
|
|
self.store,
|
|
"get_users_server_still_shares_room_with",
|
|
new_callable=mock.AsyncMock,
|
|
return_value={remote_user_id},
|
|
)
|
|
mock_request = mock.patch.object(
|
|
self.hs.get_federation_client(),
|
|
"query_user_devices",
|
|
new_callable=mock.AsyncMock,
|
|
return_value=response_body,
|
|
)
|
|
|
|
with mock_get_rooms, mock_get_users, mock_request as mocked_federation_request:
|
|
# Make the first query and sanity check it succeeds.
|
|
response_1 = self.get_success(
|
|
e2e_handler.query_devices(
|
|
request_body,
|
|
timeout=10,
|
|
from_user_id=local_user_id,
|
|
from_device_id="some_device_id",
|
|
)
|
|
)
|
|
self.assertEqual(response_1["failures"], {})
|
|
|
|
# We should have made a federation request to do so.
|
|
mocked_federation_request.assert_called_once()
|
|
|
|
# Reset the mock so we can prove we don't make a second federation request.
|
|
mocked_federation_request.reset_mock()
|
|
|
|
# Repeat the query.
|
|
response_2 = self.get_success(
|
|
e2e_handler.query_devices(
|
|
request_body,
|
|
timeout=10,
|
|
from_user_id=local_user_id,
|
|
from_device_id="some_device_id",
|
|
)
|
|
)
|
|
self.assertEqual(response_2["failures"], {})
|
|
|
|
# We should not have made a second federation request.
|
|
mocked_federation_request.assert_not_called()
|
|
|
|
# The two requests to the local homeserver should be identical.
|
|
self.assertEqual(response_1, response_2)
|
|
|
|
@override_config({"experimental_features": {"msc3983_appservice_otk_claims": True}})
|
|
def test_query_appservice(self) -> None:
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_id_1 = "xyz"
|
|
fallback_key = {"alg1:k1": "fallback_key1"}
|
|
device_id_2 = "abc"
|
|
otk = {"alg1:k2": "key2"}
|
|
|
|
# Inject an appservice interested in this user.
|
|
appservice = ApplicationService(
|
|
token="i_am_an_app_service",
|
|
id="1234",
|
|
namespaces={"users": [{"regex": r"@boris:.+", "exclusive": True}]},
|
|
# Note: this user does not have to match the regex above
|
|
sender="@as_main:test",
|
|
)
|
|
self.hs.get_datastores().main.services_cache = [appservice]
|
|
self.hs.get_datastores().main.exclusive_user_regex = _make_exclusive_regex(
|
|
[appservice]
|
|
)
|
|
|
|
# Setup a response, but only for device 2.
|
|
self.appservice_api.claim_client_keys.return_value = (
|
|
{local_user: {device_id_2: otk}},
|
|
[(local_user, device_id_1, "alg1", 1)],
|
|
)
|
|
|
|
# we shouldn't have any unused fallback keys yet
|
|
res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id_1)
|
|
)
|
|
self.assertEqual(res, [])
|
|
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id_1,
|
|
{"fallback_keys": fallback_key},
|
|
)
|
|
)
|
|
|
|
# we should now have an unused alg1 key
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id_1)
|
|
)
|
|
self.assertEqual(fallback_res, ["alg1"])
|
|
|
|
# claiming an OTK when no OTKs are available should ask the appservice, then
|
|
# query the fallback keys.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id_1: {"alg1": 1}, device_id_2: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=False,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{
|
|
"failures": {},
|
|
"one_time_keys": {
|
|
local_user: {device_id_1: fallback_key, device_id_2: otk}
|
|
},
|
|
},
|
|
)
|
|
|
|
@override_config({"experimental_features": {"msc3983_appservice_otk_claims": True}})
|
|
def test_query_appservice_with_fallback(self) -> None:
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_id_1 = "xyz"
|
|
fallback_key = {"alg1:k1": {"desc": "fallback_key1", "fallback": True}}
|
|
otk = {"alg1:k2": {"desc": "key2"}}
|
|
as_fallback_key = {"alg1:k3": {"desc": "fallback_key3", "fallback": True}}
|
|
as_otk = {"alg1:k4": {"desc": "key4"}}
|
|
|
|
# Inject an appservice interested in this user.
|
|
appservice = ApplicationService(
|
|
token="i_am_an_app_service",
|
|
id="1234",
|
|
namespaces={"users": [{"regex": r"@boris:.+", "exclusive": True}]},
|
|
# Note: this user does not have to match the regex above
|
|
sender="@as_main:test",
|
|
)
|
|
self.hs.get_datastores().main.services_cache = [appservice]
|
|
self.hs.get_datastores().main.exclusive_user_regex = _make_exclusive_regex(
|
|
[appservice]
|
|
)
|
|
|
|
# Setup a response.
|
|
response: Dict[str, Dict[str, Dict[str, JsonDict]]] = {
|
|
local_user: {device_id_1: {**as_otk, **as_fallback_key}}
|
|
}
|
|
self.appservice_api.claim_client_keys.return_value = (response, [])
|
|
|
|
# Claim OTKs, which will ask the appservice and do nothing else.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id_1: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=True,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{
|
|
"failures": {},
|
|
"one_time_keys": {
|
|
local_user: {device_id_1: {**as_otk, **as_fallback_key}}
|
|
},
|
|
},
|
|
)
|
|
|
|
# Now upload a fallback key.
|
|
res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id_1)
|
|
)
|
|
self.assertEqual(res, [])
|
|
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id_1,
|
|
{"fallback_keys": fallback_key},
|
|
)
|
|
)
|
|
|
|
# we should now have an unused alg1 key
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id_1)
|
|
)
|
|
self.assertEqual(fallback_res, ["alg1"])
|
|
|
|
# The appservice will return only the OTK.
|
|
self.appservice_api.claim_client_keys.return_value = (
|
|
{local_user: {device_id_1: as_otk}},
|
|
[],
|
|
)
|
|
|
|
# Claim OTKs, which should return the OTK from the appservice and the
|
|
# uploaded fallback key.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id_1: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=True,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{
|
|
"failures": {},
|
|
"one_time_keys": {
|
|
local_user: {device_id_1: {**as_otk, **fallback_key}}
|
|
},
|
|
},
|
|
)
|
|
|
|
# But the fallback key should not be marked as used.
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id_1)
|
|
)
|
|
self.assertEqual(fallback_res, ["alg1"])
|
|
|
|
# Now upload a OTK.
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user,
|
|
device_id_1,
|
|
{"one_time_keys": otk},
|
|
)
|
|
)
|
|
|
|
# Claim OTKs, which will return information only from the database.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id_1: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=True,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{
|
|
"failures": {},
|
|
"one_time_keys": {local_user: {device_id_1: {**otk, **fallback_key}}},
|
|
},
|
|
)
|
|
|
|
# But the fallback key should not be marked as used.
|
|
fallback_res = self.get_success(
|
|
self.store.get_e2e_unused_fallback_key_types(local_user, device_id_1)
|
|
)
|
|
self.assertEqual(fallback_res, ["alg1"])
|
|
|
|
# Finally, return only the fallback key from the appservice.
|
|
self.appservice_api.claim_client_keys.return_value = (
|
|
{local_user: {device_id_1: as_fallback_key}},
|
|
[],
|
|
)
|
|
|
|
# Claim OTKs, which will return only the fallback key from the database.
|
|
claim_res = self.get_success(
|
|
self.handler.claim_one_time_keys(
|
|
{local_user: {device_id_1: {"alg1": 1}}},
|
|
self.requester,
|
|
timeout=None,
|
|
always_include_fallback_keys=True,
|
|
)
|
|
)
|
|
self.assertEqual(
|
|
claim_res,
|
|
{
|
|
"failures": {},
|
|
"one_time_keys": {local_user: {device_id_1: as_fallback_key}},
|
|
},
|
|
)
|
|
|
|
@override_config({"experimental_features": {"msc3984_appservice_key_query": True}})
|
|
def test_query_local_devices_appservice(self) -> None:
|
|
"""Test that querying of appservices for keys overrides responses from the database."""
|
|
local_user = "@boris:" + self.hs.hostname
|
|
device_1 = "abc"
|
|
device_2 = "def"
|
|
device_3 = "ghi"
|
|
|
|
# There are 3 devices:
|
|
#
|
|
# 1. One which is uploaded to the homeserver.
|
|
# 2. One which is uploaded to the homeserver, but a newer copy is returned
|
|
# by the appservice.
|
|
# 3. One which is only returned by the appservice.
|
|
device_key_1: JsonDict = {
|
|
"user_id": local_user,
|
|
"device_id": device_1,
|
|
"algorithms": [
|
|
"m.olm.curve25519-aes-sha2",
|
|
RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2,
|
|
],
|
|
"keys": {
|
|
"ed25519:abc": "base64+ed25519+key",
|
|
"curve25519:abc": "base64+curve25519+key",
|
|
},
|
|
"signatures": {local_user: {"ed25519:abc": "base64+signature"}},
|
|
}
|
|
device_key_2a: JsonDict = {
|
|
"user_id": local_user,
|
|
"device_id": device_2,
|
|
"algorithms": [
|
|
"m.olm.curve25519-aes-sha2",
|
|
RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2,
|
|
],
|
|
"keys": {
|
|
"ed25519:def": "base64+ed25519+key",
|
|
"curve25519:def": "base64+curve25519+key",
|
|
},
|
|
"signatures": {local_user: {"ed25519:def": "base64+signature"}},
|
|
}
|
|
|
|
device_key_2b: JsonDict = {
|
|
"user_id": local_user,
|
|
"device_id": device_2,
|
|
"algorithms": [
|
|
"m.olm.curve25519-aes-sha2",
|
|
RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2,
|
|
],
|
|
# The device ID is the same (above), but the keys are different.
|
|
"keys": {
|
|
"ed25519:xyz": "base64+ed25519+key",
|
|
"curve25519:xyz": "base64+curve25519+key",
|
|
},
|
|
"signatures": {local_user: {"ed25519:xyz": "base64+signature"}},
|
|
}
|
|
device_key_3: JsonDict = {
|
|
"user_id": local_user,
|
|
"device_id": device_3,
|
|
"algorithms": [
|
|
"m.olm.curve25519-aes-sha2",
|
|
RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2,
|
|
],
|
|
"keys": {
|
|
"ed25519:jkl": "base64+ed25519+key",
|
|
"curve25519:jkl": "base64+curve25519+key",
|
|
},
|
|
"signatures": {local_user: {"ed25519:jkl": "base64+signature"}},
|
|
}
|
|
|
|
# Upload keys for devices 1 & 2a.
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_1, {"device_keys": device_key_1}
|
|
)
|
|
)
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
local_user, device_2, {"device_keys": device_key_2a}
|
|
)
|
|
)
|
|
|
|
# Inject an appservice interested in this user.
|
|
appservice = ApplicationService(
|
|
token="i_am_an_app_service",
|
|
id="1234",
|
|
namespaces={"users": [{"regex": r"@boris:.+", "exclusive": True}]},
|
|
# Note: this user does not have to match the regex above
|
|
sender="@as_main:test",
|
|
)
|
|
self.hs.get_datastores().main.services_cache = [appservice]
|
|
self.hs.get_datastores().main.exclusive_user_regex = _make_exclusive_regex(
|
|
[appservice]
|
|
)
|
|
|
|
# Setup a response.
|
|
self.appservice_api.query_keys.return_value = {
|
|
"device_keys": {
|
|
local_user: {device_2: device_key_2b, device_3: device_key_3}
|
|
}
|
|
}
|
|
|
|
# Request all devices.
|
|
res = self.get_success(self.handler.query_local_devices({local_user: None}))
|
|
self.assertIn(local_user, res)
|
|
for res_key in res[local_user].values():
|
|
res_key.pop("unsigned", None)
|
|
self.assertDictEqual(
|
|
res,
|
|
{
|
|
local_user: {
|
|
device_1: device_key_1,
|
|
device_2: device_key_2b,
|
|
device_3: device_key_3,
|
|
}
|
|
},
|
|
)
|
|
|
|
def test_check_cross_signing_setup(self) -> None:
|
|
# First check what happens with no master key.
|
|
alice = "@alice:test"
|
|
exists, replaceable_without_uia = self.get_success(
|
|
self.handler.check_cross_signing_setup(alice)
|
|
)
|
|
self.assertIs(exists, False)
|
|
self.assertIs(replaceable_without_uia, False)
|
|
|
|
# Upload a master key but don't specify a replacement timestamp.
|
|
dummy_key = {"keys": {"a": "b"}}
|
|
self.get_success(
|
|
self.store.set_e2e_cross_signing_key("@alice:test", "master", dummy_key)
|
|
)
|
|
|
|
# Should now find the key exists.
|
|
exists, replaceable_without_uia = self.get_success(
|
|
self.handler.check_cross_signing_setup(alice)
|
|
)
|
|
self.assertIs(exists, True)
|
|
self.assertIs(replaceable_without_uia, False)
|
|
|
|
# Set an expiry timestamp in the future.
|
|
self.get_success(
|
|
self.store.allow_master_cross_signing_key_replacement_without_uia(
|
|
alice,
|
|
1000,
|
|
)
|
|
)
|
|
|
|
# Should now be allowed to replace the key without UIA.
|
|
exists, replaceable_without_uia = self.get_success(
|
|
self.handler.check_cross_signing_setup(alice)
|
|
)
|
|
self.assertIs(exists, True)
|
|
self.assertIs(replaceable_without_uia, True)
|
|
|
|
# Wait 2 seconds, so that the timestamp is in the past.
|
|
self.reactor.advance(2.0)
|
|
|
|
# Should no longer be allowed to replace the key without UIA.
|
|
exists, replaceable_without_uia = self.get_success(
|
|
self.handler.check_cross_signing_setup(alice)
|
|
)
|
|
self.assertIs(exists, True)
|
|
self.assertIs(replaceable_without_uia, False)
|
|
|
|
def test_delete_old_one_time_keys(self) -> None:
|
|
"""Test the db migration that clears out old OTKs"""
|
|
|
|
# We upload two sets of keys, one just over a week ago, and one just less than
|
|
# a week ago. Each batch contains some keys that match the deletion pattern
|
|
# (key IDs of 6 chars), and some that do not.
|
|
#
|
|
# Finally, set the scheduled task going, and check what gets deleted.
|
|
|
|
user_id = "@user000:" + self.hs.hostname
|
|
device_id = "xyz"
|
|
|
|
# The scheduled task should be for "now" in real, wallclock time, so
|
|
# set the test reactor to just over a week ago.
|
|
self.reactor.advance(time.time() - 7.5 * 24 * 3600)
|
|
|
|
# Upload some keys
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
user_id,
|
|
device_id,
|
|
{
|
|
"one_time_keys": {
|
|
# some keys to delete
|
|
"alg1:AAAAAA": "key1",
|
|
"alg2:AAAAAB": {"key": "key2", "signatures": {"k1": "sig1"}},
|
|
# A key to *not* delete
|
|
"alg2:AAAAAAAAAA": {"key": "key3"},
|
|
}
|
|
},
|
|
)
|
|
)
|
|
|
|
# A day passes
|
|
self.reactor.advance(24 * 3600)
|
|
|
|
# Upload some more keys
|
|
self.get_success(
|
|
self.handler.upload_keys_for_user(
|
|
user_id,
|
|
device_id,
|
|
{
|
|
"one_time_keys": {
|
|
# some keys which match the pattern
|
|
"alg1:BAAAAA": "key1",
|
|
"alg2:BAAAAB": {"key": "key2", "signatures": {"k1": "sig1"}},
|
|
# A key to *not* delete
|
|
"alg2:BAAAAAAAAA": {"key": "key3"},
|
|
}
|
|
},
|
|
)
|
|
)
|
|
|
|
# The rest of the week passes, which should set the scheduled task going.
|
|
self.reactor.advance(6.5 * 24 * 3600)
|
|
|
|
# Check what we're left with in the database
|
|
remaining_key_ids = {
|
|
row[0]
|
|
for row in self.get_success(
|
|
self.handler.store.db_pool.simple_select_list(
|
|
"e2e_one_time_keys_json", None, ["key_id"]
|
|
)
|
|
)
|
|
}
|
|
self.assertEqual(
|
|
remaining_key_ids, {"AAAAAAAAAA", "BAAAAA", "BAAAAB", "BAAAAAAAAA"}
|
|
)
|