2018-06-14 06:53:38 +02:00
|
|
|
import struct
|
2020-01-03 04:18:49 +01:00
|
|
|
import logging
|
|
|
|
import typing
|
2018-08-04 18:10:41 +02:00
|
|
|
from binascii import hexlify, unhexlify
|
2020-01-03 04:18:49 +01:00
|
|
|
from typing import List, Iterable, Optional, Tuple
|
2018-06-14 06:53:38 +02:00
|
|
|
|
2021-12-19 22:07:01 +01:00
|
|
|
from coincurve import PublicKey as cPublicKey
|
|
|
|
from coincurve.ecdsa import deserialize_compact, cdata_to_der
|
2019-03-19 00:34:01 +01:00
|
|
|
from cryptography.hazmat.backends import default_backend
|
|
|
|
from cryptography.hazmat.primitives.serialization import load_der_public_key
|
|
|
|
from cryptography.hazmat.primitives import hashes
|
|
|
|
from cryptography.hazmat.primitives.asymmetric import ec
|
|
|
|
from cryptography.hazmat.primitives.asymmetric.utils import Prehashed
|
2019-05-05 04:15:25 +02:00
|
|
|
from cryptography.exceptions import InvalidSignature
|
2019-03-19 00:34:01 +01:00
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
from lbry.error import InsufficientFundsError
|
2019-12-31 20:52:57 +01:00
|
|
|
from lbry.crypto.hash import hash160, sha256
|
2020-01-03 04:18:49 +01:00
|
|
|
from lbry.crypto.base58 import Base58
|
|
|
|
from lbry.schema.url import normalize_name
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.schema.claim import Claim
|
2020-06-05 21:19:14 +02:00
|
|
|
from lbry.schema.base import Signable
|
2019-10-27 18:54:48 +01:00
|
|
|
from lbry.schema.purchase import Purchase
|
2020-06-05 21:19:14 +02:00
|
|
|
from lbry.schema.support import Support
|
2020-01-03 04:18:49 +01:00
|
|
|
|
|
|
|
from .script import InputScript, OutputScript
|
|
|
|
from .constants import COIN, NULL_HASH32
|
|
|
|
from .bcd_data_stream import BCDataStream
|
|
|
|
from .hash import TXRef, TXRefImmutable
|
|
|
|
from .util import ReadOnlyList
|
2021-12-19 22:07:01 +01:00
|
|
|
from .bip32 import PrivateKey, PublicKey
|
2020-01-03 04:18:49 +01:00
|
|
|
|
|
|
|
if typing.TYPE_CHECKING:
|
|
|
|
from lbry.wallet.account import Account
|
|
|
|
from lbry.wallet.ledger import Ledger
|
|
|
|
from lbry.wallet.wallet import Wallet
|
|
|
|
|
|
|
|
log = logging.getLogger()
|
|
|
|
|
|
|
|
|
|
|
|
class TXRefMutable(TXRef):
|
|
|
|
|
|
|
|
__slots__ = ('tx',)
|
|
|
|
|
|
|
|
def __init__(self, tx: 'Transaction') -> None:
|
|
|
|
super().__init__()
|
|
|
|
self.tx = tx
|
|
|
|
|
|
|
|
@property
|
|
|
|
def id(self):
|
|
|
|
if self._id is None:
|
|
|
|
self._id = hexlify(self.hash[::-1]).decode()
|
|
|
|
return self._id
|
|
|
|
|
|
|
|
@property
|
|
|
|
def hash(self):
|
|
|
|
if self._hash is None:
|
|
|
|
self._hash = sha256(sha256(self.tx.raw_sans_segwit))
|
|
|
|
return self._hash
|
|
|
|
|
|
|
|
@property
|
|
|
|
def height(self):
|
|
|
|
return self.tx.height
|
|
|
|
|
|
|
|
def reset(self):
|
|
|
|
self._id = None
|
|
|
|
self._hash = None
|
|
|
|
|
|
|
|
|
|
|
|
class TXORef:
|
|
|
|
|
|
|
|
__slots__ = 'tx_ref', 'position'
|
|
|
|
|
|
|
|
def __init__(self, tx_ref: TXRef, position: int) -> None:
|
|
|
|
self.tx_ref = tx_ref
|
|
|
|
self.position = position
|
|
|
|
|
|
|
|
@property
|
|
|
|
def id(self):
|
|
|
|
return f'{self.tx_ref.id}:{self.position}'
|
|
|
|
|
|
|
|
@property
|
|
|
|
def hash(self):
|
|
|
|
return self.tx_ref.hash + BCDataStream.uint32.pack(self.position)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def is_null(self):
|
|
|
|
return self.tx_ref.is_null
|
|
|
|
|
|
|
|
@property
|
|
|
|
def txo(self) -> Optional['Output']:
|
|
|
|
return None
|
|
|
|
|
|
|
|
|
|
|
|
class TXORefResolvable(TXORef):
|
|
|
|
|
|
|
|
__slots__ = ('_txo',)
|
|
|
|
|
|
|
|
def __init__(self, txo: 'Output') -> None:
|
|
|
|
assert txo.tx_ref is not None
|
|
|
|
assert txo.position is not None
|
|
|
|
super().__init__(txo.tx_ref, txo.position)
|
|
|
|
self._txo = txo
|
|
|
|
|
|
|
|
@property
|
|
|
|
def txo(self):
|
|
|
|
return self._txo
|
|
|
|
|
|
|
|
|
|
|
|
class InputOutput:
|
|
|
|
|
|
|
|
__slots__ = 'tx_ref', 'position'
|
|
|
|
|
|
|
|
def __init__(self, tx_ref: TXRef = None, position: int = None) -> None:
|
|
|
|
self.tx_ref = tx_ref
|
|
|
|
self.position = position
|
|
|
|
|
|
|
|
@property
|
|
|
|
def size(self) -> int:
|
|
|
|
""" Size of this input / output in bytes. """
|
|
|
|
stream = BCDataStream()
|
|
|
|
self.serialize_to(stream)
|
|
|
|
return len(stream.get_bytes())
|
|
|
|
|
|
|
|
def get_fee(self, ledger):
|
|
|
|
return self.size * ledger.fee_per_byte
|
|
|
|
|
|
|
|
def serialize_to(self, stream, alternate_script=None):
|
|
|
|
raise NotImplementedError
|
|
|
|
|
|
|
|
|
|
|
|
class Input(InputOutput):
|
|
|
|
|
|
|
|
NULL_SIGNATURE = b'\x00'*72
|
|
|
|
NULL_PUBLIC_KEY = b'\x00'*33
|
|
|
|
|
|
|
|
__slots__ = 'txo_ref', 'sequence', 'coinbase', 'script'
|
|
|
|
|
|
|
|
def __init__(self, txo_ref: TXORef, script: InputScript, sequence: int = 0xFFFFFFFF,
|
|
|
|
tx_ref: TXRef = None, position: int = None) -> None:
|
|
|
|
super().__init__(tx_ref, position)
|
|
|
|
self.txo_ref = txo_ref
|
|
|
|
self.sequence = sequence
|
|
|
|
self.coinbase = script if txo_ref.is_null else None
|
|
|
|
self.script = script if not txo_ref.is_null else None
|
|
|
|
|
|
|
|
@property
|
|
|
|
def is_coinbase(self):
|
|
|
|
return self.coinbase is not None
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def spend(cls, txo: 'Output') -> 'Input':
|
|
|
|
""" Create an input to spend the output."""
|
|
|
|
assert txo.script.is_pay_pubkey_hash, 'Attempting to spend unsupported output.'
|
|
|
|
script = InputScript.redeem_pubkey_hash(cls.NULL_SIGNATURE, cls.NULL_PUBLIC_KEY)
|
|
|
|
return cls(txo.ref, script)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def amount(self) -> int:
|
|
|
|
""" Amount this input adds to the transaction. """
|
|
|
|
if self.txo_ref.txo is None:
|
|
|
|
raise ValueError('Cannot resolve output to get amount.')
|
|
|
|
return self.txo_ref.txo.amount
|
|
|
|
|
|
|
|
@property
|
2020-03-20 06:11:05 +01:00
|
|
|
def is_my_input(self) -> Optional[bool]:
|
2020-01-03 04:18:49 +01:00
|
|
|
""" True if the output this input spends is yours. """
|
|
|
|
if self.txo_ref.txo is None:
|
|
|
|
return False
|
2020-03-20 06:11:05 +01:00
|
|
|
return self.txo_ref.txo.is_my_output
|
2020-01-03 04:18:49 +01:00
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def deserialize_from(cls, stream):
|
|
|
|
tx_ref = TXRefImmutable.from_hash(stream.read(32), -1)
|
|
|
|
position = stream.read_uint32()
|
|
|
|
script = stream.read_string()
|
|
|
|
sequence = stream.read_uint32()
|
|
|
|
return cls(
|
|
|
|
TXORef(tx_ref, position),
|
|
|
|
InputScript(script) if not tx_ref.is_null else script,
|
|
|
|
sequence
|
|
|
|
)
|
|
|
|
|
|
|
|
def serialize_to(self, stream, alternate_script=None):
|
|
|
|
stream.write(self.txo_ref.tx_ref.hash)
|
|
|
|
stream.write_uint32(self.txo_ref.position)
|
|
|
|
if alternate_script is not None:
|
|
|
|
stream.write_string(alternate_script)
|
|
|
|
else:
|
|
|
|
if self.is_coinbase:
|
|
|
|
stream.write_string(self.coinbase)
|
|
|
|
else:
|
|
|
|
stream.write_string(self.script.source)
|
|
|
|
stream.write_uint32(self.sequence)
|
2018-06-14 06:53:38 +02:00
|
|
|
|
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
class OutputEffectiveAmountEstimator:
|
2018-06-14 06:53:38 +02:00
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
__slots__ = 'txo', 'txi', 'fee', 'effective_amount'
|
2018-06-14 06:53:38 +02:00
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
def __init__(self, ledger: 'Ledger', txo: 'Output') -> None:
|
|
|
|
self.txo = txo
|
|
|
|
self.txi = Input.spend(txo)
|
|
|
|
self.fee: int = self.txi.get_fee(ledger)
|
|
|
|
self.effective_amount: int = txo.amount - self.fee
|
|
|
|
|
|
|
|
def __lt__(self, other):
|
|
|
|
return self.effective_amount < other.effective_amount
|
|
|
|
|
|
|
|
|
|
|
|
class Output(InputOutput):
|
2018-06-14 06:53:38 +02:00
|
|
|
|
2019-10-29 06:26:25 +01:00
|
|
|
__slots__ = (
|
2020-03-20 06:11:05 +01:00
|
|
|
'amount', 'script', 'is_internal_transfer', 'is_spent', 'is_my_output', 'is_my_input',
|
2020-03-22 06:13:26 +01:00
|
|
|
'channel', 'private_key', 'meta', 'sent_supports', 'sent_tips', 'received_tips',
|
2019-10-29 06:26:25 +01:00
|
|
|
'purchase', 'purchased_claim', 'purchase_receipt',
|
2020-06-05 21:19:14 +02:00
|
|
|
'reposted_claim', 'claims', '_signable'
|
2019-10-29 06:26:25 +01:00
|
|
|
)
|
2018-10-05 15:02:02 +02:00
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
def __init__(self, amount: int, script: OutputScript,
|
|
|
|
tx_ref: TXRef = None, position: int = None,
|
2020-03-20 06:11:05 +01:00
|
|
|
is_internal_transfer: Optional[bool] = None, is_spent: Optional[bool] = None,
|
|
|
|
is_my_output: Optional[bool] = None, is_my_input: Optional[bool] = None,
|
2020-03-22 06:13:26 +01:00
|
|
|
sent_supports: Optional[int] = None, sent_tips: Optional[int] = None,
|
|
|
|
received_tips: Optional[int] = None,
|
2021-12-19 22:07:01 +01:00
|
|
|
channel: Optional['Output'] = None,
|
|
|
|
private_key: Optional[PrivateKey] = None
|
2020-01-03 04:18:49 +01:00
|
|
|
) -> None:
|
|
|
|
super().__init__(tx_ref, position)
|
|
|
|
self.amount = amount
|
|
|
|
self.script = script
|
2020-03-20 06:11:05 +01:00
|
|
|
self.is_internal_transfer = is_internal_transfer
|
2020-03-07 02:12:38 +01:00
|
|
|
self.is_spent = is_spent
|
2020-03-20 06:11:05 +01:00
|
|
|
self.is_my_output = is_my_output
|
|
|
|
self.is_my_input = is_my_input
|
2020-03-22 06:13:26 +01:00
|
|
|
self.sent_supports = sent_supports
|
|
|
|
self.sent_tips = sent_tips
|
|
|
|
self.received_tips = received_tips
|
2018-10-05 15:02:02 +02:00
|
|
|
self.channel = channel
|
2021-12-19 22:07:01 +01:00
|
|
|
self.private_key: PrivateKey = private_key
|
2019-10-29 06:26:25 +01:00
|
|
|
self.purchase: 'Output' = None # txo containing purchase metadata
|
|
|
|
self.purchased_claim: 'Output' = None # resolved claim pointed to by purchase
|
|
|
|
self.purchase_receipt: 'Output' = None # txo representing purchase receipt for this claim
|
2019-11-18 21:48:52 +01:00
|
|
|
self.reposted_claim: 'Output' = None # txo representing claim being reposted
|
2019-11-13 23:50:35 +01:00
|
|
|
self.claims: List['Output'] = None # resolved claims for collection
|
2020-06-05 21:19:14 +02:00
|
|
|
self._signable: Optional[Signable] = None
|
2019-03-31 00:40:01 +01:00
|
|
|
self.meta = {}
|
2018-10-05 15:02:02 +02:00
|
|
|
|
2020-03-20 06:11:05 +01:00
|
|
|
def update_annotations(self, annotated: 'Output'):
|
2020-01-03 04:18:49 +01:00
|
|
|
if annotated is None:
|
2020-03-20 06:11:05 +01:00
|
|
|
self.is_internal_transfer = None
|
2020-03-07 02:12:38 +01:00
|
|
|
self.is_spent = None
|
2020-03-20 06:11:05 +01:00
|
|
|
self.is_my_output = None
|
|
|
|
self.is_my_input = None
|
2020-03-22 06:13:26 +01:00
|
|
|
self.sent_supports = None
|
|
|
|
self.sent_tips = None
|
|
|
|
self.received_tips = None
|
2020-01-03 04:18:49 +01:00
|
|
|
else:
|
2020-03-20 06:11:05 +01:00
|
|
|
self.is_internal_transfer = annotated.is_internal_transfer
|
2020-03-07 02:12:38 +01:00
|
|
|
self.is_spent = annotated.is_spent
|
2020-03-20 06:11:05 +01:00
|
|
|
self.is_my_output = annotated.is_my_output
|
|
|
|
self.is_my_input = annotated.is_my_input
|
2020-03-22 06:13:26 +01:00
|
|
|
self.sent_supports = annotated.sent_supports
|
|
|
|
self.sent_tips = annotated.sent_tips
|
|
|
|
self.received_tips = annotated.received_tips
|
2018-10-05 15:02:02 +02:00
|
|
|
self.channel = annotated.channel if annotated else None
|
2018-10-12 15:49:13 +02:00
|
|
|
self.private_key = annotated.private_key if annotated else None
|
2018-10-05 15:02:02 +02:00
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
@property
|
|
|
|
def ref(self):
|
|
|
|
return TXORefResolvable(self)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def id(self):
|
|
|
|
return self.ref.id
|
|
|
|
|
2021-06-02 16:14:17 +02:00
|
|
|
@property
|
|
|
|
def is_pubkey_hash(self):
|
|
|
|
return 'pubkey_hash' in self.script.values
|
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
@property
|
|
|
|
def pubkey_hash(self):
|
|
|
|
return self.script.values['pubkey_hash']
|
|
|
|
|
2021-06-02 16:14:17 +02:00
|
|
|
@property
|
|
|
|
def is_script_hash(self):
|
|
|
|
return 'script_hash' in self.script.values
|
|
|
|
|
2021-06-02 15:51:08 +02:00
|
|
|
@property
|
|
|
|
def script_hash(self):
|
|
|
|
return self.script.values['script_hash']
|
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
@property
|
|
|
|
def has_address(self):
|
2021-06-02 16:14:17 +02:00
|
|
|
return self.is_pubkey_hash or self.is_script_hash
|
2020-01-03 04:18:49 +01:00
|
|
|
|
|
|
|
def get_address(self, ledger):
|
2021-06-02 16:14:17 +02:00
|
|
|
if self.is_pubkey_hash:
|
2021-06-02 15:51:08 +02:00
|
|
|
return ledger.hash160_to_address(self.pubkey_hash)
|
2021-06-02 16:14:17 +02:00
|
|
|
elif self.is_script_hash:
|
2021-06-02 15:51:08 +02:00
|
|
|
return ledger.hash160_to_script_address(self.script_hash)
|
2020-01-03 04:18:49 +01:00
|
|
|
|
|
|
|
def get_estimator(self, ledger):
|
|
|
|
return OutputEffectiveAmountEstimator(ledger, self)
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def pay_pubkey_hash(cls, amount, pubkey_hash):
|
|
|
|
return cls(amount, OutputScript.pay_pubkey_hash(pubkey_hash))
|
|
|
|
|
2021-06-02 15:51:08 +02:00
|
|
|
@classmethod
|
|
|
|
def pay_script_hash(cls, amount, pubkey_hash):
|
|
|
|
return cls(amount, OutputScript.pay_script_hash(pubkey_hash))
|
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
@classmethod
|
|
|
|
def deserialize_from(cls, stream):
|
|
|
|
return cls(
|
|
|
|
amount=stream.read_uint64(),
|
|
|
|
script=OutputScript(stream.read_string())
|
|
|
|
)
|
|
|
|
|
|
|
|
def serialize_to(self, stream, alternate_script=None):
|
|
|
|
stream.write_uint64(self.amount)
|
|
|
|
stream.write_string(self.script.source)
|
|
|
|
|
2018-08-03 18:31:50 +02:00
|
|
|
def get_fee(self, ledger):
|
|
|
|
name_fee = 0
|
|
|
|
if self.script.is_claim_name:
|
|
|
|
name_fee = len(self.script.values['claim_name']) * ledger.fee_per_name_char
|
|
|
|
return max(name_fee, super().get_fee(ledger))
|
|
|
|
|
2018-11-28 22:28:30 +01:00
|
|
|
@property
|
|
|
|
def is_claim(self) -> bool:
|
|
|
|
return self.script.is_claim_name or self.script.is_update_claim
|
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
@property
|
|
|
|
def is_support(self) -> bool:
|
|
|
|
return self.script.is_support_claim
|
|
|
|
|
2020-06-05 21:19:14 +02:00
|
|
|
@property
|
|
|
|
def is_support_data(self) -> bool:
|
|
|
|
return self.script.is_support_claim_data
|
|
|
|
|
2018-08-04 18:10:41 +02:00
|
|
|
@property
|
2019-03-22 07:18:34 +01:00
|
|
|
def claim_hash(self) -> bytes:
|
2018-08-04 18:10:41 +02:00
|
|
|
if self.script.is_claim_name:
|
2019-03-22 07:18:34 +01:00
|
|
|
return hash160(self.tx_ref.hash + struct.pack('>I', self.position))
|
2018-08-04 18:10:41 +02:00
|
|
|
elif self.script.is_update_claim or self.script.is_support_claim:
|
2019-03-22 07:18:34 +01:00
|
|
|
return self.script.values['claim_id']
|
2018-08-04 18:10:41 +02:00
|
|
|
else:
|
|
|
|
raise ValueError('No claim_id associated.')
|
2019-03-22 07:18:34 +01:00
|
|
|
|
|
|
|
@property
|
|
|
|
def claim_id(self) -> str:
|
|
|
|
return hexlify(self.claim_hash[::-1]).decode()
|
2018-08-04 18:10:41 +02:00
|
|
|
|
|
|
|
@property
|
|
|
|
def claim_name(self) -> str:
|
|
|
|
if self.script.is_claim_involved:
|
|
|
|
return self.script.values['claim_name'].decode()
|
|
|
|
raise ValueError('No claim_name associated.')
|
|
|
|
|
2019-04-29 06:38:58 +02:00
|
|
|
@property
|
|
|
|
def normalized_name(self) -> str:
|
|
|
|
return normalize_name(self.claim_name)
|
|
|
|
|
2018-08-04 18:10:41 +02:00
|
|
|
@property
|
2019-03-18 05:59:13 +01:00
|
|
|
def claim(self) -> Claim:
|
2018-11-28 22:28:30 +01:00
|
|
|
if self.is_claim:
|
2019-03-20 06:46:23 +01:00
|
|
|
if not isinstance(self.script.values['claim'], Claim):
|
|
|
|
self.script.values['claim'] = Claim.from_bytes(self.script.values['claim'])
|
|
|
|
return self.script.values['claim']
|
2018-10-05 15:02:02 +02:00
|
|
|
raise ValueError('Only claim name and claim update have the claim payload.')
|
|
|
|
|
2019-06-07 05:42:31 +02:00
|
|
|
@property
|
|
|
|
def can_decode_claim(self):
|
|
|
|
try:
|
|
|
|
return self.claim
|
2020-06-05 21:19:14 +02:00
|
|
|
except Exception:
|
|
|
|
return False
|
|
|
|
|
|
|
|
@property
|
|
|
|
def support(self) -> Support:
|
|
|
|
if self.is_support_data:
|
|
|
|
if not isinstance(self.script.values['support'], Support):
|
|
|
|
self.script.values['support'] = Support.from_bytes(self.script.values['support'])
|
|
|
|
return self.script.values['support']
|
|
|
|
raise ValueError('Only supports with data can be represented as Supports.')
|
|
|
|
|
|
|
|
@property
|
|
|
|
def can_decode_support(self):
|
|
|
|
try:
|
|
|
|
return self.support
|
|
|
|
except Exception:
|
2019-06-07 05:42:31 +02:00
|
|
|
return False
|
|
|
|
|
2020-06-05 21:19:14 +02:00
|
|
|
@property
|
|
|
|
def signable(self) -> Signable:
|
|
|
|
if self._signable is None:
|
|
|
|
if self.is_claim:
|
|
|
|
self._signable = self.claim
|
|
|
|
elif self.is_support_data:
|
|
|
|
self._signable = self.support
|
|
|
|
return self._signable
|
|
|
|
|
2018-10-05 15:02:02 +02:00
|
|
|
@property
|
|
|
|
def permanent_url(self) -> str:
|
2018-08-04 18:10:41 +02:00
|
|
|
if self.script.is_claim_involved:
|
2021-05-07 17:31:28 +02:00
|
|
|
return f"lbry://{self.claim_name}#{self.claim_id}"
|
2018-08-04 18:10:41 +02:00
|
|
|
raise ValueError('No claim associated.')
|
|
|
|
|
2018-10-08 16:41:07 +02:00
|
|
|
@property
|
2018-10-12 15:49:13 +02:00
|
|
|
def has_private_key(self):
|
|
|
|
return self.private_key is not None
|
2018-10-08 16:41:07 +02:00
|
|
|
|
2019-05-26 05:06:22 +02:00
|
|
|
def get_signature_digest(self, ledger):
|
2020-06-05 21:19:14 +02:00
|
|
|
if self.signable.unsigned_payload:
|
2019-03-20 06:46:23 +01:00
|
|
|
pieces = [
|
2019-03-19 00:34:01 +01:00
|
|
|
Base58.decode(self.get_address(ledger)),
|
2020-06-05 21:19:14 +02:00
|
|
|
self.signable.unsigned_payload,
|
|
|
|
self.signable.signing_channel_hash[::-1]
|
2019-03-20 06:46:23 +01:00
|
|
|
]
|
2019-03-19 00:34:01 +01:00
|
|
|
else:
|
2019-03-20 06:46:23 +01:00
|
|
|
pieces = [
|
2019-04-19 14:49:50 +02:00
|
|
|
self.tx_ref.tx.inputs[0].txo_ref.hash,
|
2020-06-05 21:19:14 +02:00
|
|
|
self.signable.signing_channel_hash,
|
|
|
|
self.signable.to_message_bytes()
|
2019-03-20 06:46:23 +01:00
|
|
|
]
|
2019-05-26 05:06:22 +02:00
|
|
|
return sha256(b''.join(pieces))
|
|
|
|
|
|
|
|
@staticmethod
|
2021-12-19 22:07:01 +01:00
|
|
|
def is_signature_valid(signature, digest, public_key_bytes):
|
|
|
|
signature = cdata_to_der(deserialize_compact(signature))
|
|
|
|
public_key = cPublicKey(public_key_bytes)
|
|
|
|
is_valid = public_key.verify(signature, digest, None)
|
|
|
|
if not is_valid: # try old way
|
|
|
|
# ytsync signed claims don't seem to validate with coincurve
|
|
|
|
try:
|
|
|
|
pk = ec.EllipticCurvePublicKey.from_encoded_point(ec.SECP256K1(), public_key_bytes)
|
|
|
|
pk.verify(signature, digest, ec.ECDSA(Prehashed(hashes.SHA256())))
|
|
|
|
return True
|
|
|
|
except (ValueError, InvalidSignature):
|
|
|
|
pass
|
|
|
|
return is_valid
|
2019-03-20 06:46:23 +01:00
|
|
|
|
2019-05-26 05:06:22 +02:00
|
|
|
def is_signed_by(self, channel: 'Output', ledger=None):
|
|
|
|
return self.is_signature_valid(
|
2021-12-19 22:07:01 +01:00
|
|
|
self.signable.signature,
|
2019-05-26 05:06:22 +02:00
|
|
|
self.get_signature_digest(ledger),
|
|
|
|
channel.claim.channel.public_key_bytes
|
|
|
|
)
|
|
|
|
|
2019-03-20 06:46:23 +01:00
|
|
|
def sign(self, channel: 'Output', first_input_id=None):
|
2019-03-24 21:55:04 +01:00
|
|
|
self.channel = channel
|
2020-06-05 21:19:14 +02:00
|
|
|
self.signable.signing_channel_hash = channel.claim_hash
|
2019-03-19 00:34:01 +01:00
|
|
|
digest = sha256(b''.join([
|
2019-04-19 14:57:10 +02:00
|
|
|
first_input_id or self.tx_ref.tx.inputs[0].txo_ref.hash,
|
2020-06-05 21:19:14 +02:00
|
|
|
self.signable.signing_channel_hash,
|
|
|
|
self.signable.to_message_bytes()
|
2019-03-20 06:46:23 +01:00
|
|
|
]))
|
2021-12-19 22:07:01 +01:00
|
|
|
self.signable.signature = channel.private_key.sign_compact(digest)
|
2019-03-20 22:31:00 +01:00
|
|
|
self.script.generate()
|
2019-03-20 06:46:23 +01:00
|
|
|
|
2021-12-19 22:07:01 +01:00
|
|
|
def sign_data(self, data: bytes, timestamp: str) -> str:
|
2021-08-31 13:54:04 +02:00
|
|
|
pieces = [timestamp.encode(), self.claim_hash, data]
|
|
|
|
digest = sha256(b''.join(pieces))
|
2021-12-19 22:07:01 +01:00
|
|
|
signature = self.private_key.sign_compact(digest)
|
2021-08-31 13:54:04 +02:00
|
|
|
return hexlify(signature).decode()
|
|
|
|
|
2019-04-21 05:54:34 +02:00
|
|
|
def clear_signature(self):
|
|
|
|
self.channel = None
|
2020-06-05 21:19:14 +02:00
|
|
|
self.signable.clear_signature()
|
2019-04-21 05:54:34 +02:00
|
|
|
|
2021-12-19 22:07:01 +01:00
|
|
|
def set_channel_private_key(self, private_key: PrivateKey):
|
2021-10-13 16:56:10 +02:00
|
|
|
self.private_key = private_key
|
2021-12-19 22:07:01 +01:00
|
|
|
self.claim.channel.public_key_bytes = private_key.public_key.pubkey_bytes
|
2019-03-24 21:55:04 +01:00
|
|
|
self.script.generate()
|
2021-12-13 06:22:18 +01:00
|
|
|
return self.private_key
|
2019-03-20 06:46:23 +01:00
|
|
|
|
2021-12-19 22:07:01 +01:00
|
|
|
def is_channel_private_key(self, private_key: PrivateKey):
|
|
|
|
return self.claim.channel.public_key_bytes == private_key.public_key.pubkey_bytes
|
2019-03-19 00:34:01 +01:00
|
|
|
|
2018-06-14 06:53:38 +02:00
|
|
|
@classmethod
|
2018-08-04 18:10:41 +02:00
|
|
|
def pay_claim_name_pubkey_hash(
|
2019-03-19 00:34:01 +01:00
|
|
|
cls, amount: int, claim_name: str, claim: Claim, pubkey_hash: bytes) -> 'Output':
|
2020-01-03 04:18:49 +01:00
|
|
|
script = OutputScript.pay_claim_name_pubkey_hash(
|
2019-03-20 06:46:23 +01:00
|
|
|
claim_name.encode(), claim, pubkey_hash)
|
2020-01-03 04:18:49 +01:00
|
|
|
return cls(amount, script)
|
2018-08-14 22:16:29 +02:00
|
|
|
|
2018-08-04 18:10:41 +02:00
|
|
|
@classmethod
|
|
|
|
def pay_update_claim_pubkey_hash(
|
2019-03-19 00:34:01 +01:00
|
|
|
cls, amount: int, claim_name: str, claim_id: str, claim: Claim, pubkey_hash: bytes) -> 'Output':
|
2020-01-03 04:18:49 +01:00
|
|
|
script = OutputScript.pay_update_claim_pubkey_hash(
|
|
|
|
claim_name.encode(), unhexlify(claim_id)[::-1], claim, pubkey_hash
|
|
|
|
)
|
|
|
|
return cls(amount, script)
|
2018-06-14 06:53:38 +02:00
|
|
|
|
2018-08-29 21:43:05 +02:00
|
|
|
@classmethod
|
|
|
|
def pay_support_pubkey_hash(cls, amount: int, claim_name: str, claim_id: str, pubkey_hash: bytes) -> 'Output':
|
2020-01-03 04:18:49 +01:00
|
|
|
script = OutputScript.pay_support_pubkey_hash(
|
|
|
|
claim_name.encode(), unhexlify(claim_id)[::-1], pubkey_hash
|
|
|
|
)
|
2018-08-29 21:43:05 +02:00
|
|
|
return cls(amount, script)
|
|
|
|
|
2020-06-05 21:19:14 +02:00
|
|
|
@classmethod
|
|
|
|
def pay_support_data_pubkey_hash(
|
|
|
|
cls, amount: int, claim_name: str, claim_id: str, support: Support, pubkey_hash: bytes) -> 'Output':
|
|
|
|
script = OutputScript.pay_support_data_pubkey_hash(
|
|
|
|
claim_name.encode(), unhexlify(claim_id)[::-1], support, pubkey_hash
|
|
|
|
)
|
|
|
|
return cls(amount, script)
|
|
|
|
|
2019-03-19 00:34:01 +01:00
|
|
|
@classmethod
|
2019-10-27 18:54:48 +01:00
|
|
|
def add_purchase_data(cls, purchase: Purchase) -> 'Output':
|
2020-01-03 04:18:49 +01:00
|
|
|
script = OutputScript.return_data(purchase)
|
2019-10-27 18:54:48 +01:00
|
|
|
return cls(0, script)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def is_purchase_data(self) -> bool:
|
|
|
|
return self.script.is_return_data and (
|
2020-01-03 04:50:27 +01:00
|
|
|
isinstance(self.script.values['data'], Purchase) or
|
|
|
|
Purchase.has_start_byte(self.script.values['data'])
|
2019-10-27 18:54:48 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def purchase_data(self) -> Purchase:
|
|
|
|
if self.is_purchase_data:
|
|
|
|
if not isinstance(self.script.values['data'], Purchase):
|
|
|
|
self.script.values['data'] = Purchase.from_bytes(self.script.values['data'])
|
|
|
|
return self.script.values['data']
|
|
|
|
raise ValueError('Output does not have purchase data.')
|
|
|
|
|
|
|
|
@property
|
|
|
|
def can_decode_purchase_data(self):
|
|
|
|
try:
|
|
|
|
return self.purchase_data
|
2020-01-03 04:50:27 +01:00
|
|
|
except: # pylint: disable=bare-except
|
2019-10-27 18:54:48 +01:00
|
|
|
return False
|
2019-03-19 00:34:01 +01:00
|
|
|
|
2019-10-29 06:26:25 +01:00
|
|
|
@property
|
|
|
|
def purchased_claim_id(self):
|
|
|
|
if self.purchase is not None:
|
|
|
|
return self.purchase.purchase_data.claim_id
|
|
|
|
if self.purchased_claim is not None:
|
|
|
|
return self.purchased_claim.claim_id
|
|
|
|
|
|
|
|
@property
|
|
|
|
def has_price(self):
|
|
|
|
if self.can_decode_claim:
|
|
|
|
claim = self.claim
|
|
|
|
if claim.is_stream:
|
|
|
|
stream = claim.stream
|
|
|
|
return stream.has_fee and stream.fee.amount and stream.fee.amount > 0
|
|
|
|
return False
|
|
|
|
|
|
|
|
@property
|
|
|
|
def price(self):
|
|
|
|
return self.claim.stream.fee
|
|
|
|
|
2018-06-14 06:53:38 +02:00
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
class Transaction:
|
|
|
|
|
|
|
|
def __init__(self, raw=None, version: int = 1, locktime: int = 0, is_verified: bool = False,
|
2020-03-26 05:37:13 +01:00
|
|
|
height: int = -2, position: int = -1, julian_day: int = None) -> None:
|
2020-01-03 04:18:49 +01:00
|
|
|
self._raw = raw
|
|
|
|
self._raw_sans_segwit = None
|
2021-02-07 17:57:37 +01:00
|
|
|
self._raw_outputs = None
|
2020-01-03 04:18:49 +01:00
|
|
|
self.is_segwit_flag = 0
|
|
|
|
self.witnesses: List[bytes] = []
|
|
|
|
self.ref = TXRefMutable(self)
|
|
|
|
self.version = version
|
|
|
|
self.locktime = locktime
|
|
|
|
self._inputs: List[Input] = []
|
|
|
|
self._outputs: List[Output] = []
|
|
|
|
self.is_verified = is_verified
|
|
|
|
# Height Progression
|
|
|
|
# -2: not broadcast
|
|
|
|
# -1: in mempool but has unconfirmed inputs
|
|
|
|
# 0: in mempool and all inputs confirmed
|
|
|
|
# +num: confirmed in a specific block (height)
|
|
|
|
self.height = height
|
|
|
|
self.position = position
|
2020-03-26 05:37:13 +01:00
|
|
|
self._day = julian_day
|
2020-01-03 04:18:49 +01:00
|
|
|
if raw is not None:
|
|
|
|
self._deserialize()
|
|
|
|
|
|
|
|
@property
|
|
|
|
def is_broadcast(self):
|
|
|
|
return self.height > -2
|
|
|
|
|
|
|
|
@property
|
|
|
|
def is_mempool(self):
|
|
|
|
return self.height in (-1, 0)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def is_confirmed(self):
|
|
|
|
return self.height > 0
|
|
|
|
|
|
|
|
@property
|
|
|
|
def id(self):
|
|
|
|
return self.ref.id
|
|
|
|
|
|
|
|
@property
|
|
|
|
def hash(self):
|
|
|
|
return self.ref.hash
|
|
|
|
|
2020-03-26 05:37:13 +01:00
|
|
|
def get_julian_day(self, ledger):
|
|
|
|
if self._day is None and self.height > 0:
|
|
|
|
self._day = ledger.headers.estimated_julian_day(self.height)
|
|
|
|
return self._day
|
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
@property
|
|
|
|
def raw(self):
|
|
|
|
if self._raw is None:
|
|
|
|
self._raw = self._serialize()
|
|
|
|
return self._raw
|
|
|
|
|
|
|
|
@property
|
|
|
|
def raw_sans_segwit(self):
|
|
|
|
if self.is_segwit_flag:
|
|
|
|
if self._raw_sans_segwit is None:
|
|
|
|
self._raw_sans_segwit = self._serialize(sans_segwit=True)
|
|
|
|
return self._raw_sans_segwit
|
|
|
|
return self.raw
|
|
|
|
|
|
|
|
def _reset(self):
|
|
|
|
self._raw = None
|
|
|
|
self._raw_sans_segwit = None
|
2021-02-07 17:57:37 +01:00
|
|
|
self._raw_outputs = None
|
2020-01-03 04:18:49 +01:00
|
|
|
self.ref.reset()
|
|
|
|
|
|
|
|
@property
|
|
|
|
def inputs(self) -> ReadOnlyList[Input]:
|
|
|
|
return ReadOnlyList(self._inputs)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def outputs(self) -> ReadOnlyList[Output]:
|
|
|
|
return ReadOnlyList(self._outputs)
|
2018-06-14 06:53:38 +02:00
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
def _add(self, existing_ios: List, new_ios: Iterable[InputOutput], reset=False) -> 'Transaction':
|
|
|
|
for txio in new_ios:
|
|
|
|
txio.tx_ref = self.ref
|
|
|
|
txio.position = len(existing_ios)
|
|
|
|
existing_ios.append(txio)
|
|
|
|
if reset:
|
|
|
|
self._reset()
|
|
|
|
return self
|
2018-06-14 06:53:38 +02:00
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
def add_inputs(self, inputs: Iterable[Input]) -> 'Transaction':
|
|
|
|
return self._add(self._inputs, inputs, True)
|
|
|
|
|
|
|
|
def add_outputs(self, outputs: Iterable[Output]) -> 'Transaction':
|
|
|
|
return self._add(self._outputs, outputs, True)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def size(self) -> int:
|
|
|
|
""" Size in bytes of the entire transaction. """
|
|
|
|
return len(self.raw)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def base_size(self) -> int:
|
|
|
|
""" Size of transaction without inputs or outputs in bytes. """
|
|
|
|
return (
|
2020-01-03 04:50:27 +01:00
|
|
|
self.size
|
|
|
|
- sum(txi.size for txi in self._inputs)
|
|
|
|
- sum(txo.size for txo in self._outputs)
|
2020-01-03 04:18:49 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def input_sum(self):
|
|
|
|
return sum(i.amount for i in self.inputs if i.txo_ref.txo is not None)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def output_sum(self):
|
|
|
|
return sum(o.amount for o in self.outputs)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def net_account_balance(self) -> int:
|
|
|
|
balance = 0
|
|
|
|
for txi in self.inputs:
|
|
|
|
if txi.txo_ref.txo is None:
|
|
|
|
continue
|
2020-03-20 06:11:05 +01:00
|
|
|
if txi.is_my_input is True:
|
|
|
|
balance -= txi.amount
|
|
|
|
elif txi.is_my_input is None:
|
2020-01-03 04:18:49 +01:00
|
|
|
raise ValueError(
|
2020-03-20 06:11:05 +01:00
|
|
|
"Cannot access net_account_balance if inputs do not "
|
|
|
|
"have is_my_input set properly."
|
2020-01-03 04:18:49 +01:00
|
|
|
)
|
|
|
|
for txo in self.outputs:
|
2020-03-20 06:11:05 +01:00
|
|
|
if txo.is_my_output is True:
|
|
|
|
balance += txo.amount
|
|
|
|
elif txo.is_my_output is None:
|
2020-01-03 04:18:49 +01:00
|
|
|
raise ValueError(
|
2020-03-20 06:11:05 +01:00
|
|
|
"Cannot access net_account_balance if outputs do not "
|
|
|
|
"have is_my_output set properly."
|
2020-01-03 04:18:49 +01:00
|
|
|
)
|
|
|
|
return balance
|
|
|
|
|
|
|
|
@property
|
|
|
|
def fee(self) -> int:
|
|
|
|
return self.input_sum - self.output_sum
|
|
|
|
|
|
|
|
def get_base_fee(self, ledger) -> int:
|
|
|
|
""" Fee for base tx excluding inputs and outputs. """
|
|
|
|
return self.base_size * ledger.fee_per_byte
|
|
|
|
|
|
|
|
def get_effective_input_sum(self, ledger) -> int:
|
|
|
|
""" Sum of input values *minus* the cost involved to spend them. """
|
|
|
|
return sum(txi.amount - txi.get_fee(ledger) for txi in self._inputs)
|
|
|
|
|
|
|
|
def get_total_output_sum(self, ledger) -> int:
|
|
|
|
""" Sum of output values *plus* the cost involved to spend them. """
|
|
|
|
return sum(txo.amount + txo.get_fee(ledger) for txo in self._outputs)
|
|
|
|
|
|
|
|
def _serialize(self, with_inputs: bool = True, sans_segwit: bool = False) -> bytes:
|
|
|
|
stream = BCDataStream()
|
|
|
|
stream.write_uint32(self.version)
|
|
|
|
if with_inputs:
|
|
|
|
stream.write_compact_size(len(self._inputs))
|
|
|
|
for txin in self._inputs:
|
|
|
|
txin.serialize_to(stream)
|
2021-02-07 17:57:37 +01:00
|
|
|
self._serialize_outputs(stream)
|
2020-01-03 04:18:49 +01:00
|
|
|
stream.write_uint32(self.locktime)
|
|
|
|
return stream.get_bytes()
|
|
|
|
|
|
|
|
def _serialize_for_signature(self, signing_input: int) -> bytes:
|
|
|
|
stream = BCDataStream()
|
|
|
|
stream.write_uint32(self.version)
|
|
|
|
stream.write_compact_size(len(self._inputs))
|
|
|
|
for i, txin in enumerate(self._inputs):
|
|
|
|
if signing_input == i:
|
|
|
|
assert txin.txo_ref.txo is not None
|
|
|
|
txin.serialize_to(stream, txin.txo_ref.txo.script.source)
|
|
|
|
else:
|
|
|
|
txin.serialize_to(stream, b'')
|
2021-02-07 17:57:37 +01:00
|
|
|
self._serialize_outputs(stream)
|
2020-01-03 04:18:49 +01:00
|
|
|
stream.write_uint32(self.locktime)
|
|
|
|
stream.write_uint32(self.signature_hash_type(1)) # signature hash type: SIGHASH_ALL
|
|
|
|
return stream.get_bytes()
|
|
|
|
|
2021-02-07 17:57:37 +01:00
|
|
|
def _serialize_outputs(self, stream):
|
|
|
|
if self._raw_outputs is None:
|
|
|
|
self._raw_outputs = BCDataStream()
|
|
|
|
self._raw_outputs.write_compact_size(len(self._outputs))
|
|
|
|
for txout in self._outputs:
|
|
|
|
txout.serialize_to(self._raw_outputs)
|
|
|
|
stream.write(self._raw_outputs.get_bytes())
|
|
|
|
|
2020-01-03 04:18:49 +01:00
|
|
|
def _deserialize(self):
|
|
|
|
if self._raw is not None:
|
|
|
|
stream = BCDataStream(self._raw)
|
|
|
|
self.version = stream.read_uint32()
|
|
|
|
input_count = stream.read_compact_size()
|
|
|
|
if input_count == 0:
|
|
|
|
self.is_segwit_flag = stream.read_uint8()
|
|
|
|
input_count = stream.read_compact_size()
|
|
|
|
self._add(self._inputs, [
|
|
|
|
Input.deserialize_from(stream) for _ in range(input_count)
|
|
|
|
])
|
|
|
|
output_count = stream.read_compact_size()
|
|
|
|
self._add(self._outputs, [
|
|
|
|
Output.deserialize_from(stream) for _ in range(output_count)
|
|
|
|
])
|
|
|
|
if self.is_segwit_flag:
|
|
|
|
# drain witness portion of transaction
|
|
|
|
# too many witnesses for no crime
|
|
|
|
self.witnesses = []
|
|
|
|
for _ in range(input_count):
|
|
|
|
for _ in range(stream.read_compact_size()):
|
|
|
|
self.witnesses.append(stream.read(stream.read_compact_size()))
|
|
|
|
self.locktime = stream.read_uint32()
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def ensure_all_have_same_ledger_and_wallet(
|
|
|
|
cls, funding_accounts: Iterable['Account'],
|
|
|
|
change_account: 'Account' = None) -> Tuple['Ledger', 'Wallet']:
|
|
|
|
ledger = wallet = None
|
|
|
|
for account in funding_accounts:
|
|
|
|
if ledger is None:
|
|
|
|
ledger = account.ledger
|
|
|
|
wallet = account.wallet
|
|
|
|
if ledger != account.ledger:
|
|
|
|
raise ValueError(
|
|
|
|
'All funding accounts used to create a transaction must be on the same ledger.'
|
|
|
|
)
|
|
|
|
if wallet != account.wallet:
|
|
|
|
raise ValueError(
|
|
|
|
'All funding accounts used to create a transaction must be from the same wallet.'
|
|
|
|
)
|
|
|
|
if change_account is not None:
|
|
|
|
if change_account.ledger != ledger:
|
|
|
|
raise ValueError('Change account must use same ledger as funding accounts.')
|
|
|
|
if change_account.wallet != wallet:
|
|
|
|
raise ValueError('Change account must use same wallet as funding accounts.')
|
|
|
|
if ledger is None:
|
|
|
|
raise ValueError('No ledger found.')
|
|
|
|
if wallet is None:
|
|
|
|
raise ValueError('No wallet found.')
|
|
|
|
return ledger, wallet
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
async def create(cls, inputs: Iterable[Input], outputs: Iterable[Output],
|
|
|
|
funding_accounts: Iterable['Account'], change_account: 'Account',
|
|
|
|
sign: bool = True):
|
|
|
|
""" Find optimal set of inputs when only outputs are provided; add change
|
|
|
|
outputs if only inputs are provided or if inputs are greater than outputs. """
|
|
|
|
|
|
|
|
tx = cls() \
|
|
|
|
.add_inputs(inputs) \
|
|
|
|
.add_outputs(outputs)
|
|
|
|
|
|
|
|
ledger, _ = cls.ensure_all_have_same_ledger_and_wallet(funding_accounts, change_account)
|
|
|
|
|
|
|
|
# value of the outputs plus associated fees
|
|
|
|
cost = (
|
2020-01-03 04:50:27 +01:00
|
|
|
tx.get_base_fee(ledger) +
|
|
|
|
tx.get_total_output_sum(ledger)
|
2020-01-03 04:18:49 +01:00
|
|
|
)
|
|
|
|
# value of the inputs less the cost to spend those inputs
|
|
|
|
payment = tx.get_effective_input_sum(ledger)
|
|
|
|
|
|
|
|
try:
|
|
|
|
|
|
|
|
for _ in range(5):
|
|
|
|
|
|
|
|
if payment < cost:
|
|
|
|
deficit = cost - payment
|
|
|
|
spendables = await ledger.get_spendable_utxos(deficit, funding_accounts)
|
|
|
|
if not spendables:
|
|
|
|
raise InsufficientFundsError()
|
|
|
|
payment += sum(s.effective_amount for s in spendables)
|
|
|
|
tx.add_inputs(s.txi for s in spendables)
|
|
|
|
|
|
|
|
cost_of_change = (
|
2020-01-03 04:50:27 +01:00
|
|
|
tx.get_base_fee(ledger) +
|
|
|
|
Output.pay_pubkey_hash(COIN, NULL_HASH32).get_fee(ledger)
|
2020-01-03 04:18:49 +01:00
|
|
|
)
|
|
|
|
if payment > cost:
|
|
|
|
change = payment - cost
|
|
|
|
if change > cost_of_change:
|
|
|
|
change_address = await change_account.change.get_or_create_usable_address()
|
|
|
|
change_hash160 = change_account.ledger.address_to_hash160(change_address)
|
|
|
|
change_amount = change - cost_of_change
|
|
|
|
change_output = Output.pay_pubkey_hash(change_amount, change_hash160)
|
2020-03-20 06:11:05 +01:00
|
|
|
change_output.is_internal_transfer = True
|
2020-01-03 04:18:49 +01:00
|
|
|
tx.add_outputs([Output.pay_pubkey_hash(change_amount, change_hash160)])
|
|
|
|
|
|
|
|
if tx._outputs:
|
|
|
|
break
|
|
|
|
# this condition and the outer range(5) loop cover an edge case
|
|
|
|
# whereby a single input is just enough to cover the fee and
|
|
|
|
# has some change left over, but the change left over is less
|
|
|
|
# than the cost_of_change: thus the input is completely
|
|
|
|
# consumed and no output is added, which is an invalid tx.
|
|
|
|
# to be able to spend this input we must increase the cost
|
|
|
|
# of the TX and run through the balance algorithm a second time
|
|
|
|
# adding an extra input and change output, making tx valid.
|
|
|
|
# we do this 5 times in case the other UTXOs added are also
|
|
|
|
# less than the fee, after 5 attempts we give up and go home
|
|
|
|
cost += cost_of_change + 1
|
|
|
|
|
|
|
|
if sign:
|
|
|
|
await tx.sign(funding_accounts)
|
|
|
|
|
|
|
|
except Exception as e:
|
|
|
|
log.exception('Failed to create transaction:')
|
|
|
|
await ledger.release_tx(tx)
|
|
|
|
raise e
|
|
|
|
|
|
|
|
return tx
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def signature_hash_type(hash_type):
|
|
|
|
return hash_type
|
|
|
|
|
|
|
|
async def sign(self, funding_accounts: Iterable['Account']):
|
2021-03-01 16:18:47 +01:00
|
|
|
self._reset()
|
2020-01-03 04:18:49 +01:00
|
|
|
ledger, wallet = self.ensure_all_have_same_ledger_and_wallet(funding_accounts)
|
|
|
|
for i, txi in enumerate(self._inputs):
|
|
|
|
assert txi.script is not None
|
|
|
|
assert txi.txo_ref.txo is not None
|
|
|
|
txo_script = txi.txo_ref.txo.script
|
|
|
|
if txo_script.is_pay_pubkey_hash:
|
|
|
|
address = ledger.hash160_to_address(txo_script.values['pubkey_hash'])
|
|
|
|
private_key = await ledger.get_private_key_for_address(wallet, address)
|
|
|
|
assert private_key is not None, 'Cannot find private key for signing output.'
|
|
|
|
tx = self._serialize_for_signature(i)
|
|
|
|
txi.script.values['signature'] = \
|
|
|
|
private_key.sign(tx) + bytes((self.signature_hash_type(1),))
|
|
|
|
txi.script.values['pubkey'] = private_key.public_key.pubkey_bytes
|
|
|
|
txi.script.generate()
|
|
|
|
else:
|
|
|
|
raise NotImplementedError("Don't know how to spend this output.")
|
|
|
|
self._reset()
|
2019-03-24 21:55:04 +01:00
|
|
|
|
2018-08-23 05:19:04 +02:00
|
|
|
@classmethod
|
2020-01-03 04:18:49 +01:00
|
|
|
def pay(cls, amount: int, address: bytes, funding_accounts: List['Account'], change_account: 'Account'):
|
2020-01-03 04:50:27 +01:00
|
|
|
ledger, _ = cls.ensure_all_have_same_ledger_and_wallet(funding_accounts, change_account)
|
2018-08-23 05:19:04 +02:00
|
|
|
output = Output.pay_pubkey_hash(amount, ledger.address_to_hash160(address))
|
|
|
|
return cls.create([], [output], funding_accounts, change_account)
|
|
|
|
|
2018-06-14 21:18:36 +02:00
|
|
|
@classmethod
|
2019-03-24 21:55:04 +01:00
|
|
|
def claim_create(
|
|
|
|
cls, name: str, claim: Claim, amount: int, holding_address: str,
|
2020-01-03 04:18:49 +01:00
|
|
|
funding_accounts: List['Account'], change_account: 'Account', signing_channel: Output = None):
|
2020-01-03 04:50:27 +01:00
|
|
|
ledger, _ = cls.ensure_all_have_same_ledger_and_wallet(funding_accounts, change_account)
|
2018-06-14 21:18:36 +02:00
|
|
|
claim_output = Output.pay_claim_name_pubkey_hash(
|
2019-03-19 00:34:01 +01:00
|
|
|
amount, name, claim, ledger.address_to_hash160(holding_address)
|
2018-06-14 21:18:36 +02:00
|
|
|
)
|
2019-03-24 21:55:04 +01:00
|
|
|
if signing_channel is not None:
|
|
|
|
claim_output.sign(signing_channel, b'placeholder txid:nout')
|
|
|
|
return cls.create([], [claim_output], funding_accounts, change_account, sign=False)
|
2018-08-04 18:10:41 +02:00
|
|
|
|
2018-08-14 22:16:29 +02:00
|
|
|
@classmethod
|
2019-03-24 21:55:04 +01:00
|
|
|
def claim_update(
|
2019-04-21 05:54:34 +02:00
|
|
|
cls, previous_claim: Output, claim: Claim, amount: int, holding_address: str,
|
2020-01-03 04:18:49 +01:00
|
|
|
funding_accounts: List['Account'], change_account: 'Account', signing_channel: Output = None):
|
2020-01-03 04:50:27 +01:00
|
|
|
ledger, _ = cls.ensure_all_have_same_ledger_and_wallet(funding_accounts, change_account)
|
2019-03-24 21:55:04 +01:00
|
|
|
updated_claim = Output.pay_update_claim_pubkey_hash(
|
|
|
|
amount, previous_claim.claim_name, previous_claim.claim_id,
|
2019-04-21 05:54:34 +02:00
|
|
|
claim, ledger.address_to_hash160(holding_address)
|
2019-03-24 21:55:04 +01:00
|
|
|
)
|
|
|
|
if signing_channel is not None:
|
|
|
|
updated_claim.sign(signing_channel, b'placeholder txid:nout')
|
2019-04-21 05:54:34 +02:00
|
|
|
else:
|
|
|
|
updated_claim.clear_signature()
|
2019-03-24 21:55:04 +01:00
|
|
|
return cls.create(
|
|
|
|
[Input.spend(previous_claim)], [updated_claim], funding_accounts, change_account, sign=False
|
2018-08-14 22:16:29 +02:00
|
|
|
)
|
|
|
|
|
2018-08-04 18:10:41 +02:00
|
|
|
@classmethod
|
2019-03-24 21:55:04 +01:00
|
|
|
def support(cls, claim_name: str, claim_id: str, amount: int, holding_address: str,
|
2021-06-01 01:50:15 +02:00
|
|
|
funding_accounts: List['Account'], change_account: 'Account', signing_channel: Output = None,
|
|
|
|
comment: str = None):
|
2020-01-03 04:50:27 +01:00
|
|
|
ledger, _ = cls.ensure_all_have_same_ledger_and_wallet(funding_accounts, change_account)
|
2021-06-01 02:39:46 +02:00
|
|
|
if signing_channel is not None or comment is not None:
|
2020-06-05 21:19:14 +02:00
|
|
|
support = Support()
|
2021-06-01 02:39:46 +02:00
|
|
|
if comment is not None:
|
|
|
|
support.comment = comment
|
2020-06-05 21:19:14 +02:00
|
|
|
support_output = Output.pay_support_data_pubkey_hash(
|
|
|
|
amount, claim_name, claim_id, support, ledger.address_to_hash160(holding_address)
|
|
|
|
)
|
2021-06-01 02:39:46 +02:00
|
|
|
if signing_channel is not None:
|
|
|
|
support_output.sign(signing_channel, b'placeholder txid:nout')
|
2020-06-05 21:19:14 +02:00
|
|
|
else:
|
|
|
|
support_output = Output.pay_support_pubkey_hash(
|
|
|
|
amount, claim_name, claim_id, ledger.address_to_hash160(holding_address)
|
|
|
|
)
|
|
|
|
return cls.create([], [support_output], funding_accounts, change_account, sign=False)
|
2018-07-09 23:04:59 +02:00
|
|
|
|
2018-08-29 21:43:05 +02:00
|
|
|
@classmethod
|
2019-10-27 18:54:48 +01:00
|
|
|
def purchase(cls, claim_id: str, amount: int, merchant_address: bytes,
|
2020-01-03 04:18:49 +01:00
|
|
|
funding_accounts: List['Account'], change_account: 'Account'):
|
2020-01-03 04:50:27 +01:00
|
|
|
ledger, _ = cls.ensure_all_have_same_ledger_and_wallet(funding_accounts, change_account)
|
2019-10-27 18:54:48 +01:00
|
|
|
payment = Output.pay_pubkey_hash(amount, ledger.address_to_hash160(merchant_address))
|
|
|
|
data = Output.add_purchase_data(Purchase(claim_id))
|
|
|
|
return cls.create([], [payment, data], funding_accounts, change_account)
|
2018-08-29 21:43:05 +02:00
|
|
|
|
2018-11-05 06:09:30 +01:00
|
|
|
@property
|
|
|
|
def my_inputs(self):
|
|
|
|
for txi in self.inputs:
|
2020-03-20 06:11:05 +01:00
|
|
|
if txi.txo_ref.txo is not None and txi.txo_ref.txo.is_my_output:
|
2018-11-05 06:09:30 +01:00
|
|
|
yield txi
|
|
|
|
|
2018-09-26 04:40:52 +02:00
|
|
|
def _filter_my_outputs(self, f):
|
|
|
|
for txo in self.outputs:
|
2020-03-20 06:11:05 +01:00
|
|
|
if txo.is_my_output and f(txo.script):
|
2018-09-26 04:40:52 +02:00
|
|
|
yield txo
|
|
|
|
|
2018-11-05 06:09:30 +01:00
|
|
|
def _filter_other_outputs(self, f):
|
|
|
|
for txo in self.outputs:
|
2020-03-20 06:11:05 +01:00
|
|
|
if not txo.is_my_output and f(txo.script):
|
2018-11-05 06:09:30 +01:00
|
|
|
yield txo
|
|
|
|
|
2019-10-29 06:26:25 +01:00
|
|
|
def _filter_any_outputs(self, f):
|
|
|
|
for txo in self.outputs:
|
|
|
|
if f(txo):
|
|
|
|
yield txo
|
|
|
|
|
2018-09-26 04:40:52 +02:00
|
|
|
@property
|
|
|
|
def my_claim_outputs(self):
|
|
|
|
return self._filter_my_outputs(lambda s: s.is_claim_name)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def my_update_outputs(self):
|
|
|
|
return self._filter_my_outputs(lambda s: s.is_update_claim)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def my_support_outputs(self):
|
|
|
|
return self._filter_my_outputs(lambda s: s.is_support_claim)
|
|
|
|
|
2019-10-29 06:26:25 +01:00
|
|
|
@property
|
|
|
|
def any_purchase_outputs(self):
|
|
|
|
return self._filter_any_outputs(lambda o: o.purchase is not None)
|
|
|
|
|
2018-11-05 06:09:30 +01:00
|
|
|
@property
|
|
|
|
def other_support_outputs(self):
|
|
|
|
return self._filter_other_outputs(lambda s: s.is_support_claim)
|
|
|
|
|
2018-09-26 04:40:52 +02:00
|
|
|
@property
|
|
|
|
def my_abandon_outputs(self):
|
|
|
|
for txi in self.inputs:
|
|
|
|
abandon = txi.txo_ref.txo
|
2020-03-20 06:11:05 +01:00
|
|
|
if abandon is not None and abandon.is_my_output and abandon.script.is_claim_involved:
|
2018-09-26 04:40:52 +02:00
|
|
|
is_update = False
|
|
|
|
if abandon.script.is_claim_name or abandon.script.is_update_claim:
|
|
|
|
for update in self.my_update_outputs:
|
|
|
|
if abandon.claim_id == update.claim_id:
|
|
|
|
is_update = True
|
|
|
|
break
|
|
|
|
if not is_update:
|
|
|
|
yield abandon
|