cache the encoded output instead
This commit is contained in:
parent
20a5aecfca
commit
60a59407d8
2 changed files with 21 additions and 22 deletions
|
@ -1,14 +1,12 @@
|
|||
import asyncio
|
||||
import json
|
||||
import struct
|
||||
import zlib
|
||||
from binascii import hexlify, unhexlify
|
||||
from decimal import Decimal
|
||||
from operator import itemgetter
|
||||
from typing import Optional, List, Iterable
|
||||
|
||||
from elasticsearch import AsyncElasticsearch, NotFoundError, ConnectionError
|
||||
from elasticsearch.exceptions import ConnectionTimeout
|
||||
from elasticsearch.helpers import async_streaming_bulk
|
||||
|
||||
from lbry.crypto.base58 import Base58
|
||||
|
@ -162,9 +160,14 @@ class SearchIndex:
|
|||
async def session_query(self, query_name, kwargs):
|
||||
offset, total = kwargs.get('offset', 0) if isinstance(kwargs, dict) else 0, 0
|
||||
total_referenced = []
|
||||
cache_item = None
|
||||
if query_name == 'resolve':
|
||||
total_referenced, response, censor = await self.resolve(*kwargs)
|
||||
else:
|
||||
cache_item = ResultCacheItem.from_cache(json.dumps(kwargs, sort_keys=True), self.search_cache)
|
||||
async with cache_item.lock:
|
||||
if cache_item.result:
|
||||
return cache_item.result
|
||||
censor = Censor(Censor.SEARCH)
|
||||
response, offset, total = await self.search(**kwargs)
|
||||
censor.apply(response)
|
||||
|
@ -172,7 +175,10 @@ class SearchIndex:
|
|||
if censor.censored:
|
||||
response, _, _ = await self.search(**kwargs, censor_type=0)
|
||||
total_referenced.extend(response)
|
||||
return Outputs.to_base64(response, await self._get_referenced_rows(total_referenced), offset, total, censor)
|
||||
result = Outputs.to_base64(response, await self._get_referenced_rows(total_referenced), offset, total, censor)
|
||||
if cache_item:
|
||||
cache_item.result = result
|
||||
return result
|
||||
|
||||
async def resolve(self, *urls):
|
||||
censor = Censor(Censor.RESOLVE)
|
||||
|
@ -219,18 +225,9 @@ class SearchIndex:
|
|||
return [], 0, 0
|
||||
kwargs['channel_id'] = result['claim_id']
|
||||
try:
|
||||
expanded = expand_query(**kwargs)
|
||||
cache_item = ResultCacheItem.from_cache(json.dumps(expanded, sort_keys=True), self.search_cache)
|
||||
async with cache_item.lock:
|
||||
if cache_item.result:
|
||||
result = json.loads(zlib.decompress(cache_item.result))
|
||||
else:
|
||||
result = await self.search_client.search(
|
||||
expand_query(**kwargs), index=self.index, track_total_hits=200
|
||||
)
|
||||
cache_item.result = zlib.compress(json.dumps(result).encode(), 1)
|
||||
except ConnectionTimeout:
|
||||
raise TimeoutError()
|
||||
except NotFoundError:
|
||||
# index has no docs, fixme: log something
|
||||
return [], 0, 0
|
||||
|
|
|
@ -17,6 +17,8 @@ from functools import partial
|
|||
|
||||
from binascii import hexlify
|
||||
from concurrent.futures import ProcessPoolExecutor, ThreadPoolExecutor
|
||||
|
||||
from elasticsearch import ConnectionTimeout
|
||||
from prometheus_client import Counter, Info, Histogram, Gauge
|
||||
|
||||
import lbry
|
||||
|
@ -1009,7 +1011,7 @@ class LBRYElectrumX(SessionBase):
|
|||
try:
|
||||
self.session_mgr.pending_query_metric.inc()
|
||||
return await self.db.search_index.session_query(query_name, kwargs)
|
||||
except (TimeoutError, asyncio.TimeoutError):
|
||||
except ConnectionTimeout:
|
||||
self.session_mgr.interrupt_count_metric.inc()
|
||||
raise RPCError(JSONRPC.QUERY_TIMEOUT, 'query timed out')
|
||||
finally:
|
||||
|
|
Loading…
Reference in a new issue