This repository has been archived by the owner on Apr 26, 2024. It is now read-only.
-
-
Notifications
You must be signed in to change notification settings - Fork 2.1k
Improve opentracing support for ResponseCache
#11607
Merged
Merged
Changes from 2 commits
Commits
Show all changes
9 commits
Select commit
Hold shift + click to select a range
1f6e744
Add a `ResponseCache.keys` method
richvdh e51c966
Wrap ResponseCache entries in an attrs class
richvdh 0d1dcb9
Wrap ResponseCache operations in opentracing spans
richvdh bee20eb
changelog
richvdh f71e964
fix docstring
richvdh c302d93
Fix behaviour when tracing is not enabled
richvdh 068ec3e
better typing
richvdh d4033b2
Update unit tests
richvdh 6a89434
Fixes to `AbstractObservableDeferred`
richvdh File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -12,15 +12,15 @@ | |
# See the License for the specific language governing permissions and | ||
# limitations under the License. | ||
import logging | ||
from typing import Any, Awaitable, Callable, Dict, Generic, Optional, TypeVar | ||
from typing import Any, Awaitable, Callable, Dict, Generic, Iterable, Optional, TypeVar | ||
|
||
import attr | ||
|
||
from twisted.internet import defer | ||
|
||
from synapse.logging.context import make_deferred_yieldable, run_in_background | ||
from synapse.util import Clock | ||
from synapse.util.async_helpers import ObservableDeferred | ||
from synapse.util.async_helpers import AbstractObservableDeferred, ObservableDeferred | ||
from synapse.util.caches import register_cache | ||
|
||
logger = logging.getLogger(__name__) | ||
|
@@ -54,6 +54,17 @@ class ResponseCacheContext(Generic[KV]): | |
""" | ||
|
||
|
||
@attr.s(auto_attribs=True) | ||
class ResponseCacheEntry: | ||
result: AbstractObservableDeferred | ||
"""The (possibly incomplete) result of the operation. | ||
|
||
Note that we continue to store an ObservableDeferred even after the operation | ||
completes (rather than switching to an immediate value), since that makes it | ||
easier to cache Failure results. | ||
""" | ||
|
||
|
||
class ResponseCache(Generic[KV]): | ||
""" | ||
This caches a deferred response. Until the deferred completes it will be | ||
|
@@ -63,10 +74,7 @@ class ResponseCache(Generic[KV]): | |
""" | ||
|
||
def __init__(self, clock: Clock, name: str, timeout_ms: float = 0): | ||
# This is poorly-named: it includes both complete and incomplete results. | ||
# We keep complete results rather than switching to absolute values because | ||
# that makes it easier to cache Failure results. | ||
self.pending_result_cache: Dict[KV, ObservableDeferred] = {} | ||
self._result_cache: Dict[KV, ResponseCacheEntry] = {} | ||
Comment on lines
-66
to
+98
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. since we're changing its type (so have to change all references), we might as well fix the name while we're here. |
||
|
||
self.clock = clock | ||
self.timeout_sec = timeout_ms / 1000.0 | ||
|
@@ -75,46 +83,48 @@ def __init__(self, clock: Clock, name: str, timeout_ms: float = 0): | |
self._metrics = register_cache("response_cache", name, self, resizable=False) | ||
|
||
def size(self) -> int: | ||
return len(self.pending_result_cache) | ||
return len(self._result_cache) | ||
|
||
def __len__(self) -> int: | ||
return self.size() | ||
|
||
def get(self, key: KV) -> Optional[defer.Deferred]: | ||
"""Look up the given key. | ||
def keys(self) -> Iterable[KV]: | ||
"""Get the keys currently in the result cache | ||
|
||
Returns a new Deferred (which also doesn't follow the synapse | ||
logcontext rules). You will probably want to make_deferred_yieldable the result. | ||
Returns both incomplete entries, and (if the timeout on this cache is non-zero), | ||
complete entries which are still in the cache. | ||
|
||
If there is no entry for the key, returns None. | ||
Note that the returned iterator is not safe in the face of concurrent execution: | ||
behaviour is undefined if `wrap` is called during iteration. | ||
""" | ||
return self._result_cache.keys() | ||
|
||
def _get(self, key: KV) -> Optional[ResponseCacheEntry]: | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
|
||
"""Look up the given key. | ||
|
||
Args: | ||
key: key to get/set in the cache | ||
key: key to get in the cache | ||
|
||
Returns: | ||
None if there is no entry for this key; otherwise a deferred which | ||
resolves to the result. | ||
The entry for this key, if any; else None. | ||
""" | ||
result = self.pending_result_cache.get(key) | ||
if result is not None: | ||
entry = self._result_cache.get(key) | ||
if entry is not None: | ||
self._metrics.inc_hits() | ||
return result.observe() | ||
return entry | ||
else: | ||
self._metrics.inc_misses() | ||
return None | ||
|
||
def _set( | ||
self, context: ResponseCacheContext[KV], deferred: "defer.Deferred[RV]" | ||
) -> "defer.Deferred[RV]": | ||
) -> ResponseCacheEntry: | ||
"""Set the entry for the given key to the given deferred. | ||
|
||
*deferred* should run its callbacks in the sentinel logcontext (ie, | ||
you should wrap normal synapse deferreds with | ||
synapse.logging.context.run_in_background). | ||
|
||
Returns a new Deferred (which also doesn't follow the synapse logcontext rules). | ||
You will probably want to make_deferred_yieldable the result. | ||
richvdh marked this conversation as resolved.
Show resolved
Hide resolved
|
||
|
||
Args: | ||
context: Information about the cache miss | ||
deferred: The deferred which resolves to the result. | ||
|
@@ -124,26 +134,27 @@ def _set( | |
""" | ||
result = ObservableDeferred(deferred, consumeErrors=True) | ||
key = context.cache_key | ||
self.pending_result_cache[key] = result | ||
entry = ResponseCacheEntry(result) | ||
self._result_cache[key] = entry | ||
|
||
def on_complete(r: RV) -> RV: | ||
# if this cache has a non-zero timeout, and the callback has not cleared | ||
# the should_cache bit, we leave it in the cache for now and schedule | ||
# its removal later. | ||
if self.timeout_sec and context.should_cache: | ||
self.clock.call_later( | ||
self.timeout_sec, self.pending_result_cache.pop, key, None | ||
self.timeout_sec, self._result_cache.pop, key, None | ||
) | ||
else: | ||
# otherwise, remove the result immediately. | ||
self.pending_result_cache.pop(key, None) | ||
self._result_cache.pop(key, None) | ||
return r | ||
|
||
# make sure we do this *after* adding the entry to pending_result_cache, | ||
# make sure we do this *after* adding the entry to result_cache, | ||
# in case the result is already complete (in which case flipping the order would | ||
# leave us with a stuck entry in the cache). | ||
result.addBoth(on_complete) | ||
return result.observe() | ||
return entry | ||
|
||
async def wrap( | ||
self, | ||
|
@@ -189,17 +200,20 @@ async def handle_request(request): | |
Returns: | ||
The result of the callback (from the cache, or otherwise) | ||
""" | ||
result = self.get(key) | ||
if not result: | ||
entry = self._get(key) | ||
if not entry: | ||
logger.debug( | ||
"[%s]: no cached result for [%s], calculating new one", self._name, key | ||
) | ||
context = ResponseCacheContext(cache_key=key) | ||
if cache_context: | ||
kwargs["cache_context"] = context | ||
d = run_in_background(callback, *args, **kwargs) | ||
result = self._set(context, d) | ||
elif not isinstance(result, defer.Deferred) or result.called: | ||
entry = self._set(context, d) | ||
return await make_deferred_yieldable(entry.result.observe()) | ||
|
||
result = entry.result.observe() | ||
if result.called: | ||
logger.info("[%s]: using completed cached result for [%s]", self._name, key) | ||
else: | ||
logger.info( | ||
|
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this is the only reference to
ResponseCache.pending_result_cache
outsideResponseCache
itself.