mirror of
https://github.com/cypherstack/stack_wallet.git
synced 2024-12-26 05:19:22 +00:00
3216 lines
103 KiB
Dart
3216 lines
103 KiB
Dart
/*
|
|
* This file is part of Stack Wallet.
|
|
*
|
|
* Copyright (c) 2023 Cypher Stack
|
|
* All Rights Reserved.
|
|
* The code is distributed under GPLv3 license, see LICENSE file for details.
|
|
* Generated by Cypher Stack on 2023-05-26
|
|
*
|
|
*/
|
|
|
|
import 'dart:async';
|
|
import 'dart:io';
|
|
import 'dart:math';
|
|
|
|
import 'package:bech32/bech32.dart';
|
|
import 'package:bip32/bip32.dart' as bip32;
|
|
import 'package:bip39/bip39.dart' as bip39;
|
|
import 'package:bitbox/bitbox.dart' as bitbox;
|
|
import 'package:bitcoindart/bitcoindart.dart';
|
|
import 'package:bs58check/bs58check.dart' as bs58check;
|
|
import 'package:decimal/decimal.dart';
|
|
import 'package:flutter/foundation.dart';
|
|
import 'package:isar/isar.dart';
|
|
import 'package:stackwallet/db/isar/main_db.dart';
|
|
import 'package:stackwallet/electrumx_rpc/cached_electrumx.dart';
|
|
import 'package:stackwallet/electrumx_rpc/electrumx.dart';
|
|
import 'package:stackwallet/exceptions/electrumx/no_such_transaction.dart';
|
|
import 'package:stackwallet/models/balance.dart';
|
|
import 'package:stackwallet/models/isar/models/isar_models.dart' as isar_models;
|
|
import 'package:stackwallet/models/paymint/fee_object_model.dart';
|
|
import 'package:stackwallet/models/signing_data.dart';
|
|
import 'package:stackwallet/services/coins/coin_service.dart';
|
|
import 'package:stackwallet/services/event_bus/events/global/node_connection_status_changed_event.dart';
|
|
import 'package:stackwallet/services/event_bus/events/global/refresh_percent_changed_event.dart';
|
|
import 'package:stackwallet/services/event_bus/events/global/updated_in_background_event.dart';
|
|
import 'package:stackwallet/services/event_bus/events/global/wallet_sync_status_changed_event.dart';
|
|
import 'package:stackwallet/services/event_bus/global_event_bus.dart';
|
|
import 'package:stackwallet/services/mixins/coin_control_interface.dart';
|
|
import 'package:stackwallet/services/mixins/electrum_x_parsing.dart';
|
|
import 'package:stackwallet/services/mixins/wallet_cache.dart';
|
|
import 'package:stackwallet/services/mixins/wallet_db.dart';
|
|
import 'package:stackwallet/services/mixins/xpubable.dart';
|
|
import 'package:stackwallet/services/node_service.dart';
|
|
import 'package:stackwallet/services/transaction_notification_tracker.dart';
|
|
import 'package:stackwallet/utilities/address_utils.dart';
|
|
import 'package:stackwallet/utilities/amount/amount.dart';
|
|
import 'package:stackwallet/utilities/bip32_utils.dart';
|
|
import 'package:stackwallet/utilities/constants.dart';
|
|
import 'package:stackwallet/utilities/default_nodes.dart';
|
|
import 'package:stackwallet/utilities/enums/coin_enum.dart';
|
|
import 'package:stackwallet/utilities/enums/derive_path_type_enum.dart';
|
|
import 'package:stackwallet/utilities/enums/fee_rate_type_enum.dart';
|
|
import 'package:stackwallet/utilities/flutter_secure_storage_interface.dart';
|
|
import 'package:stackwallet/utilities/logger.dart';
|
|
import 'package:stackwallet/utilities/prefs.dart';
|
|
import 'package:stackwallet/widgets/crypto_notifications.dart';
|
|
import 'package:tuple/tuple.dart';
|
|
import 'package:uuid/uuid.dart';
|
|
|
|
const int MINIMUM_CONFIRMATIONS = 0;
|
|
|
|
const String GENESIS_HASH_MAINNET =
|
|
"000000000019d6689c085ae165831e934ff763ae46a2a6c172b3f1b60a8ce26f";
|
|
const String GENESIS_HASH_TESTNET =
|
|
"000000000933ea01ad0ee984209779baaec3ced90fa3f408719526f8d77f4943";
|
|
|
|
final Amount DUST_LIMIT = Amount(
|
|
rawValue: BigInt.from(546),
|
|
fractionDigits: Coin.eCash.decimals,
|
|
);
|
|
|
|
final eCashNetwork = NetworkType(
|
|
messagePrefix: '\x18Bitcoin Signed Message:\n',
|
|
bech32: 'bc',
|
|
bip32: Bip32Type(public: 0x0488b21e, private: 0x0488ade4),
|
|
pubKeyHash: 0x00,
|
|
scriptHash: 0x05,
|
|
wif: 0x80,
|
|
);
|
|
|
|
final eCashNetworkTestnet = NetworkType(
|
|
messagePrefix: '\x18Bitcoin Signed Message:\n',
|
|
bech32: 'tb',
|
|
bip32: Bip32Type(public: 0x043587cf, private: 0x04358394),
|
|
pubKeyHash: 0x6f,
|
|
scriptHash: 0xc4,
|
|
wif: 0xef,
|
|
);
|
|
|
|
String constructDerivePath({
|
|
required DerivePathType derivePathType,
|
|
required int networkWIF,
|
|
int account = 0,
|
|
required int chain,
|
|
required int index,
|
|
}) {
|
|
String coinType;
|
|
switch (networkWIF) {
|
|
case 0x80: // mainnet wif
|
|
switch (derivePathType) {
|
|
case DerivePathType.bip44:
|
|
coinType = "145";
|
|
break;
|
|
case DerivePathType.eCash44:
|
|
coinType = "899";
|
|
break;
|
|
default:
|
|
throw Exception(
|
|
"DerivePathType $derivePathType not supported for coinType");
|
|
}
|
|
break;
|
|
case 0xef: // testnet wif
|
|
throw Exception(
|
|
"DerivePathType $derivePathType not supported for coinType");
|
|
default:
|
|
throw Exception("Invalid ECash network wif used!");
|
|
}
|
|
|
|
int purpose;
|
|
switch (derivePathType) {
|
|
case DerivePathType.bip44:
|
|
case DerivePathType.eCash44:
|
|
purpose = 44;
|
|
break;
|
|
default:
|
|
throw Exception("DerivePathType $derivePathType not supported");
|
|
}
|
|
|
|
return "m/$purpose'/$coinType'/$account'/$chain/$index";
|
|
}
|
|
|
|
class ECashWallet extends CoinServiceAPI
|
|
with WalletCache, WalletDB, ElectrumXParsing, CoinControlInterface
|
|
implements XPubAble {
|
|
ECashWallet({
|
|
required String walletId,
|
|
required String walletName,
|
|
required Coin coin,
|
|
required ElectrumX client,
|
|
required CachedElectrumX cachedClient,
|
|
required TransactionNotificationTracker tracker,
|
|
required SecureStorageInterface secureStore,
|
|
MainDB? mockableOverride,
|
|
}) {
|
|
txTracker = tracker;
|
|
_walletId = walletId;
|
|
_walletName = walletName;
|
|
_coin = coin;
|
|
_electrumXClient = client;
|
|
_cachedElectrumXClient = cachedClient;
|
|
_secureStore = secureStore;
|
|
initCache(walletId, coin);
|
|
initWalletDB(mockableOverride: mockableOverride);
|
|
initCoinControlInterface(
|
|
walletId: walletId,
|
|
walletName: walletName,
|
|
coin: coin,
|
|
db: db,
|
|
getChainHeight: () => chainHeight,
|
|
refreshedBalanceCallback: (balance) async {
|
|
_balance = balance;
|
|
await updateCachedBalance(_balance!);
|
|
},
|
|
);
|
|
}
|
|
|
|
static const integrationTestFlag =
|
|
bool.fromEnvironment("IS_INTEGRATION_TEST");
|
|
|
|
final _prefs = Prefs.instance;
|
|
|
|
Timer? timer;
|
|
late final Coin _coin;
|
|
|
|
late final TransactionNotificationTracker txTracker;
|
|
|
|
NetworkType get _network {
|
|
switch (coin) {
|
|
case Coin.eCash:
|
|
return eCashNetwork;
|
|
// case Coin.bitcoinTestNet:
|
|
// return testnet;
|
|
default:
|
|
throw Exception("Invalid network type!");
|
|
}
|
|
}
|
|
|
|
@override
|
|
set isFavorite(bool markFavorite) {
|
|
_isFavorite = markFavorite;
|
|
updateCachedIsFavorite(markFavorite);
|
|
}
|
|
|
|
@override
|
|
bool get isFavorite => _isFavorite ??= getCachedIsFavorite();
|
|
|
|
bool? _isFavorite;
|
|
|
|
@override
|
|
Coin get coin => _coin;
|
|
|
|
@override
|
|
Future<List<isar_models.UTXO>> get utxos => db.getUTXOs(walletId).findAll();
|
|
|
|
@override
|
|
Future<List<isar_models.Transaction>> get transactions => db
|
|
.getTransactions(walletId)
|
|
.filter()
|
|
.not()
|
|
.group((q) => q
|
|
.subTypeEqualTo(isar_models.TransactionSubType.bip47Notification)
|
|
.and()
|
|
.typeEqualTo(isar_models.TransactionType.incoming))
|
|
.sortByTimestampDesc()
|
|
.findAll();
|
|
|
|
@override
|
|
Future<String> get currentReceivingAddress async =>
|
|
(await _currentReceivingAddress).value;
|
|
|
|
Future<isar_models.Address> get _currentReceivingAddress async =>
|
|
(await db
|
|
.getAddresses(walletId)
|
|
.filter()
|
|
.typeEqualTo(isar_models.AddressType.p2pkh)
|
|
.subTypeEqualTo(isar_models.AddressSubType.receiving)
|
|
.derivationPath((q) => q.valueStartsWith("m/44'/899"))
|
|
.sortByDerivationIndexDesc()
|
|
.findFirst()) ??
|
|
await _generateAddressForChain(0, 0, DerivePathTypeExt.primaryFor(coin));
|
|
|
|
Future<String> get currentChangeAddress async =>
|
|
(await _currentChangeAddress).value;
|
|
|
|
Future<isar_models.Address> get _currentChangeAddress async =>
|
|
(await db
|
|
.getAddresses(walletId)
|
|
.filter()
|
|
.typeEqualTo(isar_models.AddressType.p2pkh)
|
|
.subTypeEqualTo(isar_models.AddressSubType.change)
|
|
.derivationPath((q) => q.valueStartsWith("m/44'/899"))
|
|
.sortByDerivationIndexDesc()
|
|
.findFirst()) ??
|
|
await _generateAddressForChain(1, 0, DerivePathTypeExt.primaryFor(coin));
|
|
|
|
@override
|
|
Future<void> exit() async {
|
|
_hasCalledExit = true;
|
|
timer?.cancel();
|
|
timer = null;
|
|
stopNetworkAlivePinging();
|
|
}
|
|
|
|
bool _hasCalledExit = false;
|
|
|
|
@override
|
|
bool get hasCalledExit => _hasCalledExit;
|
|
|
|
@override
|
|
Future<FeeObject> get fees => _feeObject ??= _getFees();
|
|
Future<FeeObject>? _feeObject;
|
|
|
|
@override
|
|
Future<int> get maxFee async {
|
|
final fee = (await fees).fast as String;
|
|
final satsFee = Decimal.parse(fee) *
|
|
Decimal.fromInt(Constants.satsPerCoin(coin).toInt());
|
|
return satsFee.floor().toBigInt().toInt();
|
|
}
|
|
|
|
@override
|
|
Future<List<String>> get mnemonic => _getMnemonicList();
|
|
|
|
@override
|
|
Future<String?> get mnemonicString =>
|
|
_secureStore.read(key: '${_walletId}_mnemonic');
|
|
|
|
@override
|
|
Future<String?> get mnemonicPassphrase => _secureStore.read(
|
|
key: '${_walletId}_mnemonicPassphrase',
|
|
);
|
|
|
|
Future<int> get chainHeight async {
|
|
try {
|
|
final result = await _electrumXClient.getBlockHeadTip();
|
|
final height = result["height"] as int;
|
|
await updateCachedChainHeight(height);
|
|
if (height > storedChainHeight) {
|
|
GlobalEventBus.instance.fire(
|
|
UpdatedInBackgroundEvent(
|
|
"Updated current chain height in $walletId $walletName!",
|
|
walletId,
|
|
),
|
|
);
|
|
}
|
|
return height;
|
|
} catch (e, s) {
|
|
Logging.instance.log("Exception caught in chainHeight: $e\n$s",
|
|
level: LogLevel.Error);
|
|
return storedChainHeight;
|
|
}
|
|
}
|
|
|
|
@override
|
|
int get storedChainHeight => getCachedChainHeight();
|
|
|
|
DerivePathType addressType({required String address}) {
|
|
Uint8List? decodeBase58;
|
|
Segwit? decodeBech32;
|
|
try {
|
|
if (bitbox.Address.detectFormat(address) ==
|
|
bitbox.Address.formatCashAddr) {
|
|
if (validateCashAddr(address)) {
|
|
address = bitbox.Address.toLegacyAddress(address);
|
|
} else {
|
|
throw ArgumentError('$address is not currently supported');
|
|
}
|
|
}
|
|
} catch (_) {
|
|
// invalid cash addr format
|
|
}
|
|
try {
|
|
decodeBase58 = bs58check.decode(address);
|
|
} catch (err) {
|
|
// Base58check decode fail
|
|
}
|
|
if (decodeBase58 != null) {
|
|
if (decodeBase58[0] == _network.pubKeyHash) {
|
|
// P2PKH
|
|
return DerivePathType.bip44;
|
|
}
|
|
if (decodeBase58[0] == _network.scriptHash) {
|
|
// P2SH
|
|
return DerivePathType.bip49;
|
|
}
|
|
throw ArgumentError('Invalid version or Network mismatch');
|
|
} else {
|
|
try {
|
|
decodeBech32 = segwit.decode(address);
|
|
} catch (err) {
|
|
// Bech32 decode fail
|
|
}
|
|
if (_network.bech32 != decodeBech32!.hrp) {
|
|
throw ArgumentError('Invalid prefix or Network mismatch');
|
|
}
|
|
if (decodeBech32.version != 0) {
|
|
throw ArgumentError('Invalid address version');
|
|
}
|
|
// P2WPKH
|
|
return DerivePathType.bip84;
|
|
}
|
|
}
|
|
|
|
bool longMutex = false;
|
|
|
|
bool validateCashAddr(String cashAddr) {
|
|
String addr = cashAddr;
|
|
if (cashAddr.contains(":")) {
|
|
addr = cashAddr.split(":").last;
|
|
}
|
|
|
|
return addr.startsWith("q");
|
|
}
|
|
|
|
@override
|
|
bool validateAddress(String address) {
|
|
try {
|
|
// 0 for bitcoincash: address scheme, 1 for legacy address
|
|
final format = bitbox.Address.detectFormat(address);
|
|
if (kDebugMode) {
|
|
print("format $format");
|
|
}
|
|
|
|
// if (_coin == Coin.bitcoincashTestnet) {
|
|
// return true;
|
|
// }
|
|
|
|
if (format == bitbox.Address.formatCashAddr) {
|
|
return validateCashAddr(address);
|
|
} else {
|
|
return address.startsWith("1");
|
|
}
|
|
} catch (e) {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
@override
|
|
String get walletId => _walletId;
|
|
late final String _walletId;
|
|
|
|
@override
|
|
String get walletName => _walletName;
|
|
late String _walletName;
|
|
|
|
// setter for updating on rename
|
|
@override
|
|
set walletName(String newName) => _walletName = newName;
|
|
|
|
late ElectrumX _electrumXClient;
|
|
|
|
ElectrumX get electrumXClient => _electrumXClient;
|
|
|
|
late CachedElectrumX _cachedElectrumXClient;
|
|
|
|
CachedElectrumX get cachedElectrumXClient => _cachedElectrumXClient;
|
|
|
|
late SecureStorageInterface _secureStore;
|
|
|
|
@override
|
|
Future<void> updateNode(bool shouldRefresh) async {
|
|
final failovers = NodeService(secureStorageInterface: _secureStore)
|
|
.failoverNodesFor(coin: coin)
|
|
.map((e) => ElectrumXNode(
|
|
address: e.host,
|
|
port: e.port,
|
|
name: e.name,
|
|
id: e.id,
|
|
useSSL: e.useSSL,
|
|
))
|
|
.toList();
|
|
final newNode = await getCurrentNode();
|
|
_electrumXClient = ElectrumX.from(
|
|
node: newNode,
|
|
prefs: _prefs,
|
|
failovers: failovers,
|
|
);
|
|
_cachedElectrumXClient = CachedElectrumX.from(
|
|
electrumXClient: _electrumXClient,
|
|
);
|
|
|
|
if (shouldRefresh) {
|
|
unawaited(refresh());
|
|
}
|
|
}
|
|
|
|
Future<List<String>> _getMnemonicList() async {
|
|
final _mnemonicString = await mnemonicString;
|
|
if (_mnemonicString == null) {
|
|
return [];
|
|
}
|
|
final List<String> data = _mnemonicString.split(' ');
|
|
return data;
|
|
}
|
|
|
|
Future<ElectrumXNode> getCurrentNode() async {
|
|
final node = NodeService(secureStorageInterface: _secureStore)
|
|
.getPrimaryNodeFor(coin: coin) ??
|
|
DefaultNodes.getNodeFor(coin);
|
|
|
|
return ElectrumXNode(
|
|
address: node.host,
|
|
port: node.port,
|
|
name: node.name,
|
|
useSSL: node.useSSL,
|
|
id: node.id,
|
|
);
|
|
}
|
|
|
|
Future<List<isar_models.Address>> _fetchAllOwnAddresses() async {
|
|
final allAddresses = await db
|
|
.getAddresses(walletId)
|
|
.filter()
|
|
.not()
|
|
.typeEqualTo(isar_models.AddressType.nonWallet)
|
|
.and()
|
|
.not()
|
|
.subTypeEqualTo(isar_models.AddressSubType.nonWallet)
|
|
.findAll();
|
|
return allAddresses;
|
|
}
|
|
|
|
Future<FeeObject> _getFees() async {
|
|
try {
|
|
//TODO adjust numbers for different speeds?
|
|
const int f = 1, m = 5, s = 20;
|
|
|
|
final fast = await electrumXClient.estimateFee(blocks: f);
|
|
final medium = await electrumXClient.estimateFee(blocks: m);
|
|
final slow = await electrumXClient.estimateFee(blocks: s);
|
|
|
|
final feeObject = FeeObject(
|
|
numberOfBlocksFast: f,
|
|
numberOfBlocksAverage: m,
|
|
numberOfBlocksSlow: s,
|
|
fast: Amount.fromDecimal(
|
|
fast,
|
|
fractionDigits: coin.decimals,
|
|
).raw.toInt(),
|
|
medium: Amount.fromDecimal(
|
|
medium,
|
|
fractionDigits: coin.decimals,
|
|
).raw.toInt(),
|
|
slow: Amount.fromDecimal(
|
|
slow,
|
|
fractionDigits: coin.decimals,
|
|
).raw.toInt(),
|
|
);
|
|
|
|
Logging.instance.log("fetched fees: $feeObject", level: LogLevel.Info);
|
|
return feeObject;
|
|
} catch (e) {
|
|
Logging.instance
|
|
.log("Exception rethrown from _getFees(): $e", level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<void> _generateNewWallet() async {
|
|
Logging.instance
|
|
.log("IS_INTEGRATION_TEST: $integrationTestFlag", level: LogLevel.Info);
|
|
if (!integrationTestFlag) {
|
|
try {
|
|
final features = await electrumXClient
|
|
.getServerFeatures()
|
|
.timeout(const Duration(seconds: 3));
|
|
Logging.instance.log("features: $features", level: LogLevel.Info);
|
|
|
|
_serverVersion =
|
|
_parseServerVersion(features["server_version"] as String);
|
|
|
|
switch (coin) {
|
|
case Coin.eCash:
|
|
if (features['genesis_hash'] != GENESIS_HASH_MAINNET) {
|
|
throw Exception("genesis hash does not match main net!");
|
|
}
|
|
break;
|
|
// case Coin.e:
|
|
// if (features['genesis_hash'] != GENESIS_HASH_TESTNET) {
|
|
// throw Exception("genesis hash does not match test net!");
|
|
// }
|
|
// break;
|
|
default:
|
|
throw Exception(
|
|
"Attempted to generate a ECashWallet using a non eCash coin type: ${coin.name}");
|
|
}
|
|
} catch (e, s) {
|
|
Logging.instance.log("$e/n$s", level: LogLevel.Info);
|
|
}
|
|
}
|
|
|
|
// this should never fail
|
|
if ((await mnemonicString) != null || (await mnemonicPassphrase) != null) {
|
|
throw Exception(
|
|
"Attempted to overwrite mnemonic on generate new wallet!");
|
|
}
|
|
await _secureStore.write(
|
|
key: '${_walletId}_mnemonic',
|
|
value: bip39.generateMnemonic(strength: 128));
|
|
await _secureStore.write(key: '${_walletId}_mnemonicPassphrase', value: "");
|
|
|
|
const int startingIndex = 0;
|
|
const int receiveChain = 0;
|
|
const int changeChain = 1;
|
|
|
|
// Generate and add addresses to relevant arrays
|
|
final initialAddresses = await Future.wait([
|
|
// P2PKH
|
|
_generateAddressForChain(
|
|
receiveChain,
|
|
startingIndex,
|
|
DerivePathType.eCash44,
|
|
),
|
|
_generateAddressForChain(
|
|
changeChain,
|
|
startingIndex,
|
|
DerivePathType.eCash44,
|
|
),
|
|
// _generateAddressForChain(
|
|
// receiveChain,
|
|
// startingIndex,
|
|
// DerivePathType.bip44,
|
|
// ),
|
|
// _generateAddressForChain(
|
|
// changeChain,
|
|
// startingIndex,
|
|
// DerivePathType.bip44,
|
|
// ),
|
|
]);
|
|
|
|
await db.putAddresses(initialAddresses);
|
|
|
|
Logging.instance.log("_generateNewWalletFinished", level: LogLevel.Info);
|
|
}
|
|
|
|
/// Generates a new internal or external chain address for the wallet using a BIP84, BIP44, or BIP49 derivation path.
|
|
/// [chain] - Use 0 for receiving (external), 1 for change (internal). Should not be any other value!
|
|
/// [index] - This can be any integer >= 0
|
|
Future<isar_models.Address> _generateAddressForChain(
|
|
int chain,
|
|
int index,
|
|
DerivePathType derivePathType,
|
|
) async {
|
|
final _mnemonic = await mnemonicString;
|
|
final _mnemonicPassphrase = await mnemonicPassphrase;
|
|
if (_mnemonicPassphrase == null) {
|
|
Logging.instance.log(
|
|
"Exception in _generateAddressForChain: mnemonic passphrase null, possible migration issue; if using internal builds, delete wallet and restore from seed, if using a release build, please file bug report",
|
|
level: LogLevel.Error);
|
|
}
|
|
|
|
final derivePath = constructDerivePath(
|
|
derivePathType: derivePathType,
|
|
networkWIF: _network.wif,
|
|
chain: chain,
|
|
index: index,
|
|
);
|
|
final node = await Bip32Utils.getBip32Node(
|
|
_mnemonic!,
|
|
_mnemonicPassphrase!,
|
|
_network,
|
|
derivePath,
|
|
);
|
|
|
|
final data = PaymentData(pubkey: node.publicKey);
|
|
String address;
|
|
isar_models.AddressType addrType;
|
|
|
|
switch (derivePathType) {
|
|
case DerivePathType.bip44:
|
|
case DerivePathType.eCash44:
|
|
address = P2PKH(data: data, network: _network).data.address!;
|
|
addrType = isar_models.AddressType.p2pkh;
|
|
address = bitbox.Address.toECashAddress(address);
|
|
break;
|
|
default:
|
|
throw Exception("DerivePathType $derivePathType not supported");
|
|
}
|
|
|
|
// add generated address & info to derivations
|
|
// await addDerivation(
|
|
// chain: chain,
|
|
// address: address,
|
|
// pubKey: Format.uint8listToString(node.publicKey),
|
|
// wif: node.toWIF(),
|
|
// derivePathType: derivePathType,
|
|
// );
|
|
|
|
return isar_models.Address(
|
|
walletId: walletId,
|
|
value: address,
|
|
publicKey: node.publicKey,
|
|
type: addrType,
|
|
derivationIndex: index,
|
|
derivationPath: isar_models.DerivationPath()..value = derivePath,
|
|
subType: chain == 0
|
|
? isar_models.AddressSubType.receiving
|
|
: isar_models.AddressSubType.change,
|
|
);
|
|
}
|
|
|
|
/// Returns the latest receiving/change (external/internal) address for the wallet depending on [chain]
|
|
/// and
|
|
/// [chain] - Use 0 for receiving (external), 1 for change (internal). Should not be any other value!
|
|
Future<String> _getCurrentAddressForChain(
|
|
int chain,
|
|
DerivePathType derivePathType,
|
|
) async {
|
|
final subType = chain == 0 // Here, we assume that chain == 1 if it isn't 0
|
|
? isar_models.AddressSubType.receiving
|
|
: isar_models.AddressSubType.change;
|
|
|
|
isar_models.AddressType type;
|
|
String coinType;
|
|
String purpose;
|
|
switch (derivePathType) {
|
|
case DerivePathType.bip44:
|
|
type = isar_models.AddressType.p2pkh;
|
|
coinType = "0";
|
|
purpose = "44";
|
|
break;
|
|
case DerivePathType.eCash44:
|
|
type = isar_models.AddressType.p2pkh;
|
|
coinType = "899";
|
|
purpose = "44";
|
|
break;
|
|
default:
|
|
throw Exception("DerivePathType unsupported");
|
|
}
|
|
final address = await db
|
|
.getAddresses(walletId)
|
|
.filter()
|
|
.typeEqualTo(type)
|
|
.subTypeEqualTo(subType)
|
|
.derivationPath((q) => q.valueStartsWith("m/$purpose'/$coinType"))
|
|
.sortByDerivationIndexDesc()
|
|
.findFirst();
|
|
return address!.value;
|
|
}
|
|
|
|
Future<List<Map<String, dynamic>>> fastFetch(List<String> allTxHashes) async {
|
|
List<Map<String, dynamic>> allTransactions = [];
|
|
|
|
const futureLimit = 30;
|
|
List<Future<Map<String, dynamic>>> transactionFutures = [];
|
|
int currentFutureCount = 0;
|
|
for (final txHash in allTxHashes) {
|
|
Future<Map<String, dynamic>> transactionFuture =
|
|
cachedElectrumXClient.getTransaction(
|
|
txHash: txHash,
|
|
verbose: true,
|
|
coin: coin,
|
|
);
|
|
transactionFutures.add(transactionFuture);
|
|
currentFutureCount++;
|
|
if (currentFutureCount > futureLimit) {
|
|
currentFutureCount = 0;
|
|
await Future.wait(transactionFutures);
|
|
for (final fTx in transactionFutures) {
|
|
final tx = await fTx;
|
|
|
|
allTransactions.add(tx);
|
|
}
|
|
}
|
|
}
|
|
if (currentFutureCount != 0) {
|
|
currentFutureCount = 0;
|
|
await Future.wait(transactionFutures);
|
|
for (final fTx in transactionFutures) {
|
|
final tx = await fTx;
|
|
|
|
allTransactions.add(tx);
|
|
}
|
|
}
|
|
return allTransactions;
|
|
}
|
|
|
|
double? _serverVersion;
|
|
bool get serverCanBatch => _serverVersion != null && _serverVersion! >= 1.6;
|
|
|
|
// stupid + fragile
|
|
double? _parseServerVersion(String version) {
|
|
double? result;
|
|
try {
|
|
final list = version.split(" ");
|
|
if (list.isNotEmpty) {
|
|
final numberStrings = list.last.split(".");
|
|
final major = numberStrings.removeAt(0);
|
|
|
|
result = double.tryParse("$major.${numberStrings.join("")}");
|
|
}
|
|
} catch (_) {}
|
|
|
|
Logging.instance.log(
|
|
"$walletName _parseServerVersion($version) => $result",
|
|
level: LogLevel.Info,
|
|
);
|
|
return result;
|
|
}
|
|
|
|
/// attempts to convert a string to a valid scripthash
|
|
///
|
|
/// Returns the scripthash or throws an exception on invalid bch address
|
|
String _convertToScriptHash(String address, NetworkType network) {
|
|
try {
|
|
if (bitbox.Address.detectFormat(address) ==
|
|
bitbox.Address.formatCashAddr &&
|
|
validateCashAddr(address)) {
|
|
final addressLegacy = bitbox.Address.toLegacyAddress(address);
|
|
return AddressUtils.convertToScriptHash(addressLegacy, network);
|
|
}
|
|
return AddressUtils.convertToScriptHash(address, network);
|
|
} catch (e) {
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<void> _updateUTXOs() async {
|
|
final allAddresses = await _fetchAllOwnAddresses();
|
|
|
|
try {
|
|
final fetchedUtxoList = <List<Map<String, dynamic>>>[];
|
|
|
|
if (serverCanBatch) {
|
|
final Map<int, Map<String, List<dynamic>>> batches = {};
|
|
const batchSizeMax = 100;
|
|
int batchNumber = 0;
|
|
for (int i = 0; i < allAddresses.length; i++) {
|
|
if (batches[batchNumber] == null) {
|
|
batches[batchNumber] = {};
|
|
}
|
|
final scriptHash = _convertToScriptHash(
|
|
allAddresses[i].value,
|
|
_network,
|
|
);
|
|
batches[batchNumber]!.addAll({
|
|
scriptHash: [scriptHash]
|
|
});
|
|
if (i % batchSizeMax == batchSizeMax - 1) {
|
|
batchNumber++;
|
|
}
|
|
}
|
|
|
|
for (int i = 0; i < batches.length; i++) {
|
|
final response = await _electrumXClient.getBatchUTXOs(
|
|
args: batches[i]!,
|
|
);
|
|
for (final entry in response.entries) {
|
|
if (entry.value.isNotEmpty) {
|
|
fetchedUtxoList.add(entry.value);
|
|
}
|
|
}
|
|
}
|
|
} else {
|
|
for (int i = 0; i < allAddresses.length; i++) {
|
|
final scriptHash = _convertToScriptHash(
|
|
allAddresses[i].value,
|
|
_network,
|
|
);
|
|
|
|
final utxos = await electrumXClient.getUTXOs(scripthash: scriptHash);
|
|
if (utxos.isNotEmpty) {
|
|
fetchedUtxoList.add(utxos);
|
|
}
|
|
}
|
|
}
|
|
|
|
final List<isar_models.UTXO> outputArray = [];
|
|
|
|
for (int i = 0; i < fetchedUtxoList.length; i++) {
|
|
for (int j = 0; j < fetchedUtxoList[i].length; j++) {
|
|
final jsonUTXO = fetchedUtxoList[i][j];
|
|
|
|
final txn = await cachedElectrumXClient.getTransaction(
|
|
txHash: jsonUTXO["tx_hash"] as String,
|
|
verbose: true,
|
|
coin: coin,
|
|
);
|
|
|
|
bool shouldBlock = false;
|
|
String? blockReason;
|
|
String? label;
|
|
|
|
final vout = jsonUTXO["tx_pos"] as int;
|
|
|
|
final outputs = txn["vout"] as List;
|
|
|
|
String? utxoOwnerAddress;
|
|
// get UTXO owner address
|
|
for (final output in outputs) {
|
|
if (output["n"] == vout) {
|
|
utxoOwnerAddress =
|
|
output["scriptPubKey"]?["addresses"]?[0] as String? ??
|
|
output["scriptPubKey"]?["address"] as String?;
|
|
}
|
|
}
|
|
|
|
final utxo = isar_models.UTXO(
|
|
walletId: walletId,
|
|
txid: txn["txid"] as String,
|
|
vout: vout,
|
|
value: jsonUTXO["value"] as int,
|
|
name: label ?? "",
|
|
isBlocked: shouldBlock,
|
|
blockedReason: blockReason,
|
|
isCoinbase: txn["is_coinbase"] as bool? ?? false,
|
|
blockHash: txn["blockhash"] as String?,
|
|
blockHeight: jsonUTXO["height"] as int?,
|
|
blockTime: txn["blocktime"] as int?,
|
|
address: utxoOwnerAddress,
|
|
);
|
|
|
|
outputArray.add(utxo);
|
|
}
|
|
}
|
|
|
|
Logging.instance
|
|
.log('Outputs fetched: $outputArray', level: LogLevel.Info);
|
|
|
|
await db.updateUTXOs(walletId, outputArray);
|
|
|
|
// finally update balance
|
|
await _updateBalance();
|
|
} catch (e, s) {
|
|
Logging.instance
|
|
.log("Output fetch unsuccessful: $e\n$s", level: LogLevel.Error);
|
|
}
|
|
}
|
|
|
|
Future<void> _updateBalance() async {
|
|
await refreshBalance();
|
|
}
|
|
|
|
@override
|
|
Balance get balance => _balance ??= getCachedBalance();
|
|
Balance? _balance;
|
|
|
|
Future<int> getTxCount({required String address}) async {
|
|
String? scripthash;
|
|
try {
|
|
scripthash = _convertToScriptHash(address, _network);
|
|
final transactions =
|
|
await electrumXClient.getHistory(scripthash: scripthash);
|
|
return transactions.length;
|
|
} catch (e) {
|
|
Logging.instance.log(
|
|
"Exception rethrown in _getTxCount(address: $address, scripthash: $scripthash): $e",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<int> _getTxCount({required isar_models.Address address}) async {
|
|
try {
|
|
return await getTxCount(address: address.value);
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown in _getTxCount(address: $address: $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<Map<String, int>> _getBatchTxCount({
|
|
required Map<String, String> addresses,
|
|
}) async {
|
|
try {
|
|
final Map<String, List<dynamic>> args = {};
|
|
for (final entry in addresses.entries) {
|
|
args[entry.key] = [_convertToScriptHash(entry.value, _network)];
|
|
}
|
|
final response = await electrumXClient.getBatchHistory(args: args);
|
|
|
|
final Map<String, int> result = {};
|
|
for (final entry in response.entries) {
|
|
result[entry.key] = entry.value.length;
|
|
}
|
|
return result;
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown in _getBatchTxCount(address: $addresses: $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<void> _checkReceivingAddressForTransactions() async {
|
|
try {
|
|
final currentReceiving = await _currentReceivingAddress;
|
|
|
|
final int txCount = await getTxCount(address: currentReceiving.value);
|
|
Logging.instance.log(
|
|
'Number of txs for current receiving address $currentReceiving: $txCount',
|
|
level: LogLevel.Info);
|
|
|
|
if (txCount >= 1 || currentReceiving.derivationIndex < 0) {
|
|
// First increment the receiving index
|
|
final newReceivingIndex = currentReceiving.derivationIndex + 1;
|
|
|
|
// Use new index to derive a new receiving address
|
|
final newReceivingAddress = await _generateAddressForChain(
|
|
0, newReceivingIndex, DerivePathTypeExt.primaryFor(coin));
|
|
|
|
final existing = await db
|
|
.getAddresses(walletId)
|
|
.filter()
|
|
.valueEqualTo(newReceivingAddress.value)
|
|
.findFirst();
|
|
if (existing == null) {
|
|
// Add that new change address
|
|
await db.putAddress(newReceivingAddress);
|
|
} else {
|
|
// we need to update the address
|
|
await db.updateAddress(existing, newReceivingAddress);
|
|
}
|
|
// keep checking until address with no tx history is set as current
|
|
await _checkReceivingAddressForTransactions();
|
|
}
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown from _checkReceivingAddressForTransactions(${DerivePathTypeExt.primaryFor(coin)}): $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<void> _checkChangeAddressForTransactions() async {
|
|
try {
|
|
final currentChange = await _currentChangeAddress;
|
|
final int txCount = await getTxCount(address: currentChange.value);
|
|
Logging.instance.log(
|
|
'Number of txs for current change address $currentChange: $txCount',
|
|
level: LogLevel.Info);
|
|
|
|
if (txCount >= 1 || currentChange.derivationIndex < 0) {
|
|
// First increment the change index
|
|
final newChangeIndex = currentChange.derivationIndex + 1;
|
|
|
|
// Use new index to derive a new change address
|
|
final newChangeAddress = await _generateAddressForChain(
|
|
1, newChangeIndex, DerivePathTypeExt.primaryFor(coin));
|
|
|
|
final existing = await db
|
|
.getAddresses(walletId)
|
|
.filter()
|
|
.valueEqualTo(newChangeAddress.value)
|
|
.findFirst();
|
|
if (existing == null) {
|
|
// Add that new change address
|
|
await db.putAddress(newChangeAddress);
|
|
} else {
|
|
// we need to update the address
|
|
await db.updateAddress(existing, newChangeAddress);
|
|
}
|
|
// keep checking until address with no tx history is set as current
|
|
await _checkChangeAddressForTransactions();
|
|
}
|
|
} on SocketException catch (se, s) {
|
|
Logging.instance.log(
|
|
"SocketException caught in _checkReceivingAddressForTransactions(${DerivePathTypeExt.primaryFor(coin)}): $se\n$s",
|
|
level: LogLevel.Error);
|
|
return;
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown from _checkReceivingAddressForTransactions(${DerivePathTypeExt.primaryFor(coin)}): $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<void> _checkCurrentReceivingAddressesForTransactions() async {
|
|
try {
|
|
await _checkReceivingAddressForTransactions();
|
|
// await _checkReceivingAddressForTransactionsP2PKH();
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown from _checkCurrentReceivingAddressesForTransactions(): $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<void> _checkCurrentChangeAddressesForTransactions() async {
|
|
try {
|
|
await _checkChangeAddressForTransactions();
|
|
// await _checkP2PKHChangeAddressForTransactions();
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown from _checkCurrentChangeAddressesForTransactions(): $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<List<Map<String, dynamic>>> _fetchHistory(
|
|
List<String> allAddresses,
|
|
) async {
|
|
try {
|
|
List<Map<String, dynamic>> allTxHashes = [];
|
|
|
|
if (serverCanBatch) {
|
|
final Map<int, Map<String, List<dynamic>>> batches = {};
|
|
final Map<String, String> requestIdToAddressMap = {};
|
|
const batchSizeMax = 100;
|
|
int batchNumber = 0;
|
|
for (int i = 0; i < allAddresses.length; i++) {
|
|
if (batches[batchNumber] == null) {
|
|
batches[batchNumber] = {};
|
|
}
|
|
final scriptHash = _convertToScriptHash(
|
|
allAddresses[i],
|
|
_network,
|
|
);
|
|
final id = Logger.isTestEnv ? "$i" : const Uuid().v1();
|
|
requestIdToAddressMap[id] = allAddresses[i];
|
|
batches[batchNumber]!.addAll({
|
|
id: [scriptHash]
|
|
});
|
|
if (i % batchSizeMax == batchSizeMax - 1) {
|
|
batchNumber++;
|
|
}
|
|
}
|
|
|
|
for (int i = 0; i < batches.length; i++) {
|
|
final response =
|
|
await _electrumXClient.getBatchHistory(args: batches[i]!);
|
|
for (final entry in response.entries) {
|
|
for (int j = 0; j < entry.value.length; j++) {
|
|
entry.value[j]["address"] = requestIdToAddressMap[entry.key];
|
|
if (!allTxHashes.contains(entry.value[j])) {
|
|
allTxHashes.add(entry.value[j]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
} else {
|
|
for (int i = 0; i < allAddresses.length; i++) {
|
|
final scriptHash = _convertToScriptHash(
|
|
allAddresses[i],
|
|
_network,
|
|
);
|
|
|
|
final response = await electrumXClient.getHistory(
|
|
scripthash: scriptHash,
|
|
);
|
|
|
|
for (int j = 0; j < response.length; j++) {
|
|
response[j]["address"] = allAddresses[i];
|
|
if (!allTxHashes.contains(response[j])) {
|
|
allTxHashes.add(response[j]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return allTxHashes;
|
|
} catch (e, s) {
|
|
Logging.instance.log("_fetchHistory: $e\n$s", level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
bool _duplicateTxCheck(
|
|
List<Map<String, dynamic>> allTransactions, String txid) {
|
|
for (int i = 0; i < allTransactions.length; i++) {
|
|
if (allTransactions[i]["txid"] == txid) {
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
Future<void> _refreshTransactions() async {
|
|
List<isar_models.Address> allAddressesOld = await _fetchAllOwnAddresses();
|
|
|
|
Set<String> receivingAddresses = allAddressesOld
|
|
.where((e) => e.subType == isar_models.AddressSubType.receiving)
|
|
.map((e) {
|
|
if (bitbox.Address.detectFormat(e.value) == bitbox.Address.formatLegacy &&
|
|
(addressType(address: e.value) == DerivePathType.bip44 ||
|
|
addressType(address: e.value) == DerivePathType.eCash44)) {
|
|
return bitbox.Address.toECashAddress(e.value);
|
|
} else {
|
|
return e.value;
|
|
}
|
|
}).toSet();
|
|
|
|
Set<String> changeAddresses = allAddressesOld
|
|
.where((e) => e.subType == isar_models.AddressSubType.change)
|
|
.map((e) {
|
|
if (bitbox.Address.detectFormat(e.value) == bitbox.Address.formatLegacy &&
|
|
(addressType(address: e.value) == DerivePathType.bip44 ||
|
|
addressType(address: e.value) == DerivePathType.eCash44)) {
|
|
return bitbox.Address.toECashAddress(e.value);
|
|
} else {
|
|
return e.value;
|
|
}
|
|
}).toSet();
|
|
|
|
final List<Map<String, dynamic>> allTxHashes =
|
|
await _fetchHistory([...receivingAddresses, ...changeAddresses]);
|
|
|
|
List<Map<String, dynamic>> allTransactions = [];
|
|
|
|
for (final txHash in allTxHashes) {
|
|
final storedTx = await db
|
|
.getTransactions(walletId)
|
|
.filter()
|
|
.txidEqualTo(txHash["tx_hash"] as String)
|
|
.findFirst();
|
|
|
|
if (storedTx == null ||
|
|
storedTx.address.value == null ||
|
|
storedTx.height == null ||
|
|
(storedTx.height != null && storedTx.height! <= 0)) {
|
|
final tx = await cachedElectrumXClient.getTransaction(
|
|
txHash: txHash["tx_hash"] as String,
|
|
verbose: true,
|
|
coin: coin,
|
|
);
|
|
|
|
if (!_duplicateTxCheck(allTransactions, tx["txid"] as String)) {
|
|
tx["address"] = await db
|
|
.getAddresses(walletId)
|
|
.filter()
|
|
.valueEqualTo(txHash["address"] as String)
|
|
.findFirst();
|
|
tx["height"] = txHash["height"];
|
|
allTransactions.add(tx);
|
|
}
|
|
}
|
|
}
|
|
|
|
final List<Tuple2<isar_models.Transaction, isar_models.Address?>> txns = [];
|
|
|
|
for (final txData in allTransactions) {
|
|
Set<String> inputAddresses = {};
|
|
Set<String> outputAddresses = {};
|
|
|
|
Logging.instance.log(txData, level: LogLevel.Fatal);
|
|
|
|
Amount totalInputValue = Amount(
|
|
rawValue: BigInt.from(0),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
Amount totalOutputValue = Amount(
|
|
rawValue: BigInt.from(0),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
|
|
Amount amountSentFromWallet = Amount(
|
|
rawValue: BigInt.from(0),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
Amount amountReceivedInWallet = Amount(
|
|
rawValue: BigInt.from(0),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
Amount changeAmount = Amount(
|
|
rawValue: BigInt.from(0),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
|
|
// parse inputs
|
|
for (final input in txData["vin"] as List) {
|
|
final prevTxid = input["txid"] as String;
|
|
final prevOut = input["vout"] as int;
|
|
|
|
// fetch input tx to get address
|
|
final inputTx = await cachedElectrumXClient.getTransaction(
|
|
txHash: prevTxid,
|
|
coin: coin,
|
|
);
|
|
|
|
for (final output in inputTx["vout"] as List) {
|
|
// check matching output
|
|
if (prevOut == output["n"]) {
|
|
// get value
|
|
final value = Amount.fromDecimal(
|
|
Decimal.parse(output["value"].toString()),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
|
|
// add value to total
|
|
totalInputValue = totalInputValue + value;
|
|
|
|
// get input(prevOut) address
|
|
final address =
|
|
output["scriptPubKey"]?["addresses"]?[0] as String? ??
|
|
output["scriptPubKey"]?["address"] as String?;
|
|
|
|
if (address != null) {
|
|
inputAddresses.add(address);
|
|
|
|
// if input was from my wallet, add value to amount sent
|
|
if (receivingAddresses.contains(address) ||
|
|
changeAddresses.contains(address)) {
|
|
amountSentFromWallet = amountSentFromWallet + value;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// parse outputs
|
|
for (final output in txData["vout"] as List) {
|
|
// get value
|
|
final value = Amount.fromDecimal(
|
|
Decimal.parse(output["value"].toString()),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
|
|
// add value to total
|
|
totalOutputValue += value;
|
|
|
|
// get output address
|
|
final address = output["scriptPubKey"]?["addresses"]?[0] as String? ??
|
|
output["scriptPubKey"]?["address"] as String?;
|
|
if (address != null) {
|
|
outputAddresses.add(address);
|
|
|
|
// if output was to my wallet, add value to amount received
|
|
if (receivingAddresses.contains(address)) {
|
|
amountReceivedInWallet += value;
|
|
} else if (changeAddresses.contains(address)) {
|
|
changeAmount += value;
|
|
}
|
|
}
|
|
}
|
|
|
|
final mySentFromAddresses = [
|
|
...receivingAddresses.intersection(inputAddresses),
|
|
...changeAddresses.intersection(inputAddresses)
|
|
];
|
|
final myReceivedOnAddresses =
|
|
receivingAddresses.intersection(outputAddresses);
|
|
final myChangeReceivedOnAddresses =
|
|
changeAddresses.intersection(outputAddresses);
|
|
|
|
final fee = totalInputValue - totalOutputValue;
|
|
|
|
// this is the address initially used to fetch the txid
|
|
isar_models.Address transactionAddress =
|
|
txData["address"] as isar_models.Address;
|
|
|
|
isar_models.TransactionType type;
|
|
Amount amount;
|
|
if (mySentFromAddresses.isNotEmpty && myReceivedOnAddresses.isNotEmpty) {
|
|
// tx is sent to self
|
|
type = isar_models.TransactionType.sentToSelf;
|
|
amount =
|
|
amountSentFromWallet - amountReceivedInWallet - fee - changeAmount;
|
|
} else if (mySentFromAddresses.isNotEmpty) {
|
|
// outgoing tx
|
|
type = isar_models.TransactionType.outgoing;
|
|
amount = amountSentFromWallet - changeAmount - fee;
|
|
final possible =
|
|
outputAddresses.difference(myChangeReceivedOnAddresses).first;
|
|
|
|
if (transactionAddress.value != possible) {
|
|
transactionAddress = isar_models.Address(
|
|
walletId: walletId,
|
|
value: possible,
|
|
publicKey: [],
|
|
type: isar_models.AddressType.nonWallet,
|
|
derivationIndex: -1,
|
|
derivationPath: null,
|
|
subType: isar_models.AddressSubType.nonWallet,
|
|
);
|
|
}
|
|
} else {
|
|
// incoming tx
|
|
type = isar_models.TransactionType.incoming;
|
|
amount = amountReceivedInWallet;
|
|
}
|
|
|
|
List<isar_models.Input> inputs = [];
|
|
List<isar_models.Output> outputs = [];
|
|
|
|
for (final json in txData["vin"] as List) {
|
|
bool isCoinBase = json['coinbase'] != null;
|
|
final input = isar_models.Input(
|
|
txid: json['txid'] as String,
|
|
vout: json['vout'] as int? ?? -1,
|
|
scriptSig: json['scriptSig']?['hex'] as String?,
|
|
scriptSigAsm: json['scriptSig']?['asm'] as String?,
|
|
isCoinbase: isCoinBase ? isCoinBase : json['is_coinbase'] as bool?,
|
|
sequence: json['sequence'] as int?,
|
|
innerRedeemScriptAsm: json['innerRedeemscriptAsm'] as String?,
|
|
);
|
|
inputs.add(input);
|
|
}
|
|
|
|
for (final json in txData["vout"] as List) {
|
|
final output = isar_models.Output(
|
|
scriptPubKey: json['scriptPubKey']?['hex'] as String?,
|
|
scriptPubKeyAsm: json['scriptPubKey']?['asm'] as String?,
|
|
scriptPubKeyType: json['scriptPubKey']?['type'] as String?,
|
|
scriptPubKeyAddress:
|
|
json["scriptPubKey"]?["addresses"]?[0] as String? ??
|
|
json['scriptPubKey']['type'] as String,
|
|
value: Amount.fromDecimal(
|
|
Decimal.parse(json["value"].toString()),
|
|
fractionDigits: coin.decimals,
|
|
).raw.toInt(),
|
|
);
|
|
outputs.add(output);
|
|
}
|
|
|
|
final tx = isar_models.Transaction(
|
|
walletId: walletId,
|
|
txid: txData["txid"] as String,
|
|
timestamp: txData["blocktime"] as int? ??
|
|
(DateTime.now().millisecondsSinceEpoch ~/ 1000),
|
|
type: type,
|
|
subType: isar_models.TransactionSubType.none,
|
|
amount: amount.raw.toInt(),
|
|
amountString: amount.toJsonString(),
|
|
fee: fee.raw.toInt(),
|
|
height: txData["height"] as int?,
|
|
isCancelled: false,
|
|
isLelantus: false,
|
|
slateId: null,
|
|
otherData: null,
|
|
nonce: null,
|
|
inputs: inputs,
|
|
outputs: outputs,
|
|
numberOfMessages: null,
|
|
);
|
|
|
|
txns.add(Tuple2(tx, transactionAddress));
|
|
}
|
|
|
|
await db.addNewTransactionData(txns, walletId);
|
|
|
|
// quick hack to notify manager to call notifyListeners if
|
|
// transactions changed
|
|
if (txns.isNotEmpty) {
|
|
GlobalEventBus.instance.fire(
|
|
UpdatedInBackgroundEvent(
|
|
"Transactions updated/added for: $walletId $walletName ",
|
|
walletId,
|
|
),
|
|
);
|
|
}
|
|
}
|
|
|
|
int estimateTxFee({required int vSize, required int feeRatePerKB}) {
|
|
return vSize * (feeRatePerKB / 1000).ceil();
|
|
}
|
|
|
|
/// The coinselection algorithm decides whether or not the user is eligible to make the transaction
|
|
/// with [satoshiAmountToSend] and [selectedTxFeeRate]. If so, it will call buildTrasaction() and return
|
|
/// a map containing the tx hex along with other important information. If not, then it will return
|
|
/// an integer (1 or 2)
|
|
dynamic coinSelection({
|
|
required int satoshiAmountToSend,
|
|
required int selectedTxFeeRate,
|
|
required String recipientAddress,
|
|
required bool coinControl,
|
|
required bool isSendAll,
|
|
int? satsPerVByte,
|
|
int additionalOutputs = 0,
|
|
List<isar_models.UTXO>? utxos,
|
|
}) async {
|
|
Logging.instance
|
|
.log("Starting coinSelection ----------", level: LogLevel.Info);
|
|
final List<isar_models.UTXO> availableOutputs = utxos ?? await this.utxos;
|
|
final currentChainHeight = await chainHeight;
|
|
final List<isar_models.UTXO> spendableOutputs = [];
|
|
int spendableSatoshiValue = 0;
|
|
|
|
// Build list of spendable outputs and totaling their satoshi amount
|
|
for (final utxo in availableOutputs) {
|
|
if (utxo.isBlocked == false &&
|
|
utxo.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) &&
|
|
utxo.used != true) {
|
|
spendableOutputs.add(utxo);
|
|
spendableSatoshiValue += utxo.value;
|
|
}
|
|
}
|
|
|
|
if (coinControl) {
|
|
if (spendableOutputs.length < availableOutputs.length) {
|
|
throw ArgumentError("Attempted to use an unavailable utxo");
|
|
}
|
|
}
|
|
|
|
// don't care about sorting if using all utxos
|
|
if (!coinControl) {
|
|
// sort spendable by age (oldest first)
|
|
spendableOutputs.sort((a, b) => b.blockTime!.compareTo(a.blockTime!));
|
|
}
|
|
|
|
Logging.instance.log("spendableOutputs.length: ${spendableOutputs.length}",
|
|
level: LogLevel.Info);
|
|
Logging.instance.log("availableOutputs.length: ${availableOutputs.length}",
|
|
level: LogLevel.Info);
|
|
Logging.instance
|
|
.log("spendableOutputs: $spendableOutputs", level: LogLevel.Info);
|
|
Logging.instance.log("spendableSatoshiValue: $spendableSatoshiValue",
|
|
level: LogLevel.Info);
|
|
Logging.instance
|
|
.log("satoshiAmountToSend: $satoshiAmountToSend", level: LogLevel.Info);
|
|
// If the amount the user is trying to send is smaller than the amount that they have spendable,
|
|
// then return 1, which indicates that they have an insufficient balance.
|
|
if (spendableSatoshiValue < satoshiAmountToSend) {
|
|
return 1;
|
|
// If the amount the user wants to send is exactly equal to the amount they can spend, then return
|
|
// 2, which indicates that they are not leaving enough over to pay the transaction fee
|
|
} else if (spendableSatoshiValue == satoshiAmountToSend && !isSendAll) {
|
|
return 2;
|
|
}
|
|
// If neither of these statements pass, we assume that the user has a spendable balance greater
|
|
// than the amount they're attempting to send. Note that this value still does not account for
|
|
// the added transaction fee, which may require an extra input and will need to be checked for
|
|
// later on.
|
|
|
|
// Possible situation right here
|
|
int satoshisBeingUsed = 0;
|
|
int inputsBeingConsumed = 0;
|
|
List<isar_models.UTXO> utxoObjectsToUse = [];
|
|
|
|
if (!coinControl) {
|
|
for (var i = 0;
|
|
satoshisBeingUsed < satoshiAmountToSend &&
|
|
i < spendableOutputs.length;
|
|
i++) {
|
|
utxoObjectsToUse.add(spendableOutputs[i]);
|
|
satoshisBeingUsed += spendableOutputs[i].value;
|
|
inputsBeingConsumed += 1;
|
|
}
|
|
for (int i = 0;
|
|
i < additionalOutputs &&
|
|
inputsBeingConsumed < spendableOutputs.length;
|
|
i++) {
|
|
utxoObjectsToUse.add(spendableOutputs[inputsBeingConsumed]);
|
|
satoshisBeingUsed += spendableOutputs[inputsBeingConsumed].value;
|
|
inputsBeingConsumed += 1;
|
|
}
|
|
} else {
|
|
satoshisBeingUsed = spendableSatoshiValue;
|
|
utxoObjectsToUse = spendableOutputs;
|
|
inputsBeingConsumed = spendableOutputs.length;
|
|
}
|
|
|
|
Logging.instance
|
|
.log("satoshisBeingUsed: $satoshisBeingUsed", level: LogLevel.Info);
|
|
Logging.instance
|
|
.log("inputsBeingConsumed: $inputsBeingConsumed", level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('utxoObjectsToUse: $utxoObjectsToUse', level: LogLevel.Info);
|
|
|
|
// numberOfOutputs' length must always be equal to that of recipientsArray and recipientsAmtArray
|
|
List<String> recipientsArray = [recipientAddress];
|
|
List<int> recipientsAmtArray = [satoshiAmountToSend];
|
|
|
|
// gather required signing data
|
|
final utxoSigningData = await fetchBuildTxData(utxoObjectsToUse);
|
|
|
|
if (isSendAll) {
|
|
Logging.instance
|
|
.log("Attempting to send all $coin", level: LogLevel.Info);
|
|
|
|
final int vSizeForOneOutput = (await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: [recipientAddress],
|
|
satoshiAmounts: [satoshisBeingUsed - 1],
|
|
))["vSize"] as int;
|
|
int feeForOneOutput = satsPerVByte != null
|
|
? (satsPerVByte * vSizeForOneOutput)
|
|
: estimateTxFee(
|
|
vSize: vSizeForOneOutput,
|
|
feeRatePerKB: selectedTxFeeRate,
|
|
);
|
|
|
|
if (satsPerVByte == null) {
|
|
final int roughEstimate = roughFeeEstimate(
|
|
spendableOutputs.length,
|
|
1,
|
|
selectedTxFeeRate,
|
|
).raw.toInt();
|
|
if (feeForOneOutput < roughEstimate) {
|
|
feeForOneOutput = roughEstimate;
|
|
}
|
|
}
|
|
|
|
final int amount = satoshiAmountToSend - feeForOneOutput;
|
|
dynamic txn = await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: recipientsArray,
|
|
satoshiAmounts: [amount],
|
|
);
|
|
Map<String, dynamic> transactionObject = {
|
|
"hex": txn["hex"],
|
|
"recipient": recipientsArray[0],
|
|
"recipientAmt": Amount(
|
|
rawValue: BigInt.from(amount),
|
|
fractionDigits: coin.decimals,
|
|
),
|
|
"fee": feeForOneOutput,
|
|
"vSize": txn["vSize"],
|
|
"usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
};
|
|
return transactionObject;
|
|
}
|
|
|
|
final int vSizeForOneOutput;
|
|
try {
|
|
vSizeForOneOutput = (await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: [recipientAddress],
|
|
satoshiAmounts: [satoshisBeingUsed - 1],
|
|
))["vSize"] as int;
|
|
} catch (e) {
|
|
Logging.instance.log("vSizeForOneOutput: $e", level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
|
|
final int vSizeForTwoOutPuts;
|
|
try {
|
|
vSizeForTwoOutPuts = (await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: [
|
|
recipientAddress,
|
|
await _getCurrentAddressForChain(
|
|
1, DerivePathTypeExt.primaryFor(coin)),
|
|
],
|
|
satoshiAmounts: [
|
|
satoshiAmountToSend,
|
|
max(0, satoshisBeingUsed - satoshiAmountToSend - 1),
|
|
],
|
|
))["vSize"] as int;
|
|
} catch (e) {
|
|
Logging.instance.log("vSizeForTwoOutPuts: $e", level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
|
|
// Assume 1 output, only for recipient and no change
|
|
final feeForOneOutput = satsPerVByte != null
|
|
? (satsPerVByte * vSizeForOneOutput)
|
|
: estimateTxFee(
|
|
vSize: vSizeForOneOutput,
|
|
feeRatePerKB: selectedTxFeeRate,
|
|
);
|
|
// Assume 2 outputs, one for recipient and one for change
|
|
final feeForTwoOutputs = satsPerVByte != null
|
|
? (satsPerVByte * vSizeForTwoOutPuts)
|
|
: estimateTxFee(
|
|
vSize: vSizeForTwoOutPuts,
|
|
feeRatePerKB: selectedTxFeeRate,
|
|
);
|
|
|
|
Logging.instance
|
|
.log("feeForTwoOutputs: $feeForTwoOutputs", level: LogLevel.Info);
|
|
Logging.instance
|
|
.log("feeForOneOutput: $feeForOneOutput", level: LogLevel.Info);
|
|
|
|
if (satoshisBeingUsed - satoshiAmountToSend > feeForOneOutput) {
|
|
if (satoshisBeingUsed - satoshiAmountToSend >
|
|
feeForOneOutput + DUST_LIMIT.raw.toInt()) {
|
|
// Here, we know that theoretically, we may be able to include another output(change) but we first need to
|
|
// factor in the value of this output in satoshis.
|
|
int changeOutputSize =
|
|
satoshisBeingUsed - satoshiAmountToSend - feeForTwoOutputs;
|
|
// We check to see if the user can pay for the new transaction with 2 outputs instead of one. If they can and
|
|
// the second output's size > DUST_LIMIT satoshis, we perform the mechanics required to properly generate and use a new
|
|
// change address.
|
|
if (changeOutputSize > DUST_LIMIT.raw.toInt() &&
|
|
satoshisBeingUsed - satoshiAmountToSend - changeOutputSize ==
|
|
feeForTwoOutputs) {
|
|
// generate new change address if current change address has been used
|
|
await _checkChangeAddressForTransactions();
|
|
final String newChangeAddress = await _getCurrentAddressForChain(
|
|
1, DerivePathTypeExt.primaryFor(coin));
|
|
|
|
int feeBeingPaid =
|
|
satoshisBeingUsed - satoshiAmountToSend - changeOutputSize;
|
|
|
|
recipientsArray.add(newChangeAddress);
|
|
recipientsAmtArray.add(changeOutputSize);
|
|
// At this point, we have the outputs we're going to use, the amounts to send along with which addresses
|
|
// we intend to send these amounts to. We have enough to send instructions to build the transaction.
|
|
Logging.instance.log('2 outputs in tx', level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log('Change Output Size: $changeOutputSize',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log(
|
|
'Difference (fee being paid): $feeBeingPaid sats',
|
|
level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('Estimated fee: $feeForTwoOutputs', level: LogLevel.Info);
|
|
dynamic txn = await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: recipientsArray,
|
|
satoshiAmounts: recipientsAmtArray,
|
|
);
|
|
|
|
// make sure minimum fee is accurate if that is being used
|
|
if (txn["vSize"] - feeBeingPaid == 1) {
|
|
int changeOutputSize =
|
|
satoshisBeingUsed - satoshiAmountToSend - (txn["vSize"] as int);
|
|
feeBeingPaid =
|
|
satoshisBeingUsed - satoshiAmountToSend - changeOutputSize;
|
|
recipientsAmtArray.removeLast();
|
|
recipientsAmtArray.add(changeOutputSize);
|
|
Logging.instance.log('Adjusted Input size: $satoshisBeingUsed',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log(
|
|
'Adjusted Recipient output size: $satoshiAmountToSend',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log(
|
|
'Adjusted Change Output Size: $changeOutputSize',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log(
|
|
'Adjusted Difference (fee being paid): $feeBeingPaid sats',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log('Adjusted Estimated fee: $feeForTwoOutputs',
|
|
level: LogLevel.Info);
|
|
txn = await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: recipientsArray,
|
|
satoshiAmounts: recipientsAmtArray,
|
|
);
|
|
}
|
|
|
|
Map<String, dynamic> transactionObject = {
|
|
"hex": txn["hex"],
|
|
"recipient": recipientsArray[0],
|
|
"recipientAmt": Amount(
|
|
rawValue: BigInt.from(recipientsAmtArray[0]),
|
|
fractionDigits: coin.decimals,
|
|
),
|
|
"fee": feeBeingPaid,
|
|
"vSize": txn["vSize"],
|
|
"usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
};
|
|
return transactionObject;
|
|
} else {
|
|
// Something went wrong here. It either overshot or undershot the estimated fee amount or the changeOutputSize
|
|
// is smaller than or equal to DUST_LIMIT. Revert to single output transaction.
|
|
Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log(
|
|
'Difference (fee being paid): ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
dynamic txn = await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: recipientsArray,
|
|
satoshiAmounts: recipientsAmtArray,
|
|
);
|
|
Map<String, dynamic> transactionObject = {
|
|
"hex": txn["hex"],
|
|
"recipient": recipientsArray[0],
|
|
"recipientAmt": Amount(
|
|
rawValue: BigInt.from(recipientsAmtArray[0]),
|
|
fractionDigits: coin.decimals,
|
|
),
|
|
"fee": satoshisBeingUsed - satoshiAmountToSend,
|
|
"vSize": txn["vSize"],
|
|
"usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
};
|
|
return transactionObject;
|
|
}
|
|
} else {
|
|
// No additional outputs needed since adding one would mean that it'd be smaller than DUST_LIMIT sats
|
|
// which makes it uneconomical to add to the transaction. Here, we pass data directly to instruct
|
|
// the wallet to begin crafting the transaction that the user requested.
|
|
Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log(
|
|
'Difference (fee being paid): ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
dynamic txn = await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: recipientsArray,
|
|
satoshiAmounts: recipientsAmtArray,
|
|
);
|
|
Map<String, dynamic> transactionObject = {
|
|
"hex": txn["hex"],
|
|
"recipient": recipientsArray[0],
|
|
"recipientAmt": Amount(
|
|
rawValue: BigInt.from(recipientsAmtArray[0]),
|
|
fractionDigits: coin.decimals,
|
|
),
|
|
"fee": satoshisBeingUsed - satoshiAmountToSend,
|
|
"vSize": txn["vSize"],
|
|
"usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
};
|
|
return transactionObject;
|
|
}
|
|
} else if (satoshisBeingUsed - satoshiAmountToSend == feeForOneOutput) {
|
|
// In this scenario, no additional change output is needed since inputs - outputs equal exactly
|
|
// what we need to pay for fees. Here, we pass data directly to instruct the wallet to begin
|
|
// crafting the transaction that the user requested.
|
|
Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
level: LogLevel.Info);
|
|
Logging.instance.log(
|
|
'Fee being paid: ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
level: LogLevel.Info);
|
|
Logging.instance
|
|
.log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
dynamic txn = await buildTransaction(
|
|
utxosToUse: utxoObjectsToUse,
|
|
utxoSigningData: utxoSigningData,
|
|
recipients: recipientsArray,
|
|
satoshiAmounts: recipientsAmtArray,
|
|
);
|
|
Map<String, dynamic> transactionObject = {
|
|
"hex": txn["hex"],
|
|
"recipient": recipientsArray[0],
|
|
"recipientAmt": Amount(
|
|
rawValue: BigInt.from(recipientsAmtArray[0]),
|
|
fractionDigits: coin.decimals,
|
|
),
|
|
"fee": feeForOneOutput,
|
|
"vSize": txn["vSize"],
|
|
"usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
};
|
|
return transactionObject;
|
|
} else {
|
|
// Remember that returning 2 indicates that the user does not have a sufficient balance to
|
|
// pay for the transaction fee. Ideally, at this stage, we should check if the user has any
|
|
// additional outputs they're able to spend and then recalculate fees.
|
|
Logging.instance.log(
|
|
'Cannot pay tx fee - checking for more outputs and trying again',
|
|
level: LogLevel.Warning);
|
|
// try adding more outputs
|
|
if (spendableOutputs.length > inputsBeingConsumed) {
|
|
return coinSelection(
|
|
satoshiAmountToSend: satoshiAmountToSend,
|
|
selectedTxFeeRate: selectedTxFeeRate,
|
|
satsPerVByte: satsPerVByte,
|
|
recipientAddress: recipientAddress,
|
|
isSendAll: isSendAll,
|
|
additionalOutputs: additionalOutputs + 1,
|
|
utxos: utxos,
|
|
coinControl: coinControl,
|
|
);
|
|
}
|
|
return 2;
|
|
}
|
|
}
|
|
|
|
Future<List<SigningData>> fetchBuildTxData(
|
|
List<isar_models.UTXO> utxosToUse,
|
|
) async {
|
|
// return data
|
|
List<SigningData> signingData = [];
|
|
|
|
try {
|
|
// Populating the addresses to check
|
|
for (var i = 0; i < utxosToUse.length; i++) {
|
|
final txid = utxosToUse[i].txid;
|
|
final tx = await _cachedElectrumXClient.getTransaction(
|
|
txHash: txid,
|
|
coin: coin,
|
|
);
|
|
|
|
for (final output in tx["vout"] as List) {
|
|
final n = output["n"];
|
|
if (n != null && n == utxosToUse[i].vout) {
|
|
String address =
|
|
output["scriptPubKey"]?["addresses"]?[0] as String? ??
|
|
output["scriptPubKey"]["address"] as String;
|
|
if (bitbox.Address.detectFormat(address) !=
|
|
bitbox.Address.formatCashAddr) {
|
|
try {
|
|
address = bitbox.Address.toECashAddress(address);
|
|
} catch (_) {
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
utxosToUse[i] = utxosToUse[i].copyWith(address: address);
|
|
}
|
|
}
|
|
|
|
final derivePathType = addressType(address: utxosToUse[i].address!);
|
|
|
|
signingData.add(
|
|
SigningData(
|
|
derivePathType: derivePathType,
|
|
utxo: utxosToUse[i],
|
|
),
|
|
);
|
|
}
|
|
|
|
final root = await Bip32Utils.getBip32Root(
|
|
(await mnemonicString)!,
|
|
(await mnemonicPassphrase)!,
|
|
_network,
|
|
);
|
|
|
|
for (final sd in signingData) {
|
|
final address = await db.getAddress(walletId, sd.utxo.address!);
|
|
final node = await Bip32Utils.getBip32NodeFromRoot(
|
|
root,
|
|
address!.derivationPath!.value,
|
|
);
|
|
|
|
final paymentData = PaymentData(pubkey: node.publicKey);
|
|
|
|
final PaymentData data;
|
|
final Uint8List? redeemScript;
|
|
|
|
switch (sd.derivePathType) {
|
|
case DerivePathType.bip44:
|
|
case DerivePathType.eCash44:
|
|
data = P2PKH(
|
|
data: paymentData,
|
|
network: _network,
|
|
).data;
|
|
redeemScript = null;
|
|
break;
|
|
|
|
default:
|
|
throw Exception("DerivePathType unsupported");
|
|
}
|
|
|
|
final keyPair = ECPair.fromWIF(
|
|
node.toWIF(),
|
|
network: _network,
|
|
);
|
|
|
|
sd.redeemScript = redeemScript;
|
|
sd.output = data.output;
|
|
sd.keyPair = keyPair;
|
|
}
|
|
|
|
return signingData;
|
|
} catch (e, s) {
|
|
Logging.instance
|
|
.log("fetchBuildTxData() threw: $e,\n$s", level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
/// Builds and signs a transaction
|
|
Future<Map<String, dynamic>> buildTransaction({
|
|
required List<isar_models.UTXO> utxosToUse,
|
|
required List<SigningData> utxoSigningData,
|
|
required List<String> recipients,
|
|
required List<int> satoshiAmounts,
|
|
}) async {
|
|
final builder = bitbox.Bitbox.transactionBuilder(
|
|
testnet: false, //coin == Coin.bitcoincashTestnet,
|
|
);
|
|
|
|
// retrieve address' utxos from the rest api
|
|
List<bitbox.Utxo> _utxos =
|
|
[]; // await Bitbox.Address.utxo(address) as List<Bitbox.Utxo>;
|
|
for (var element in utxosToUse) {
|
|
_utxos.add(bitbox.Utxo(
|
|
element.txid,
|
|
element.vout,
|
|
bitbox.BitcoinCash.fromSatoshi(element.value),
|
|
element.value,
|
|
0,
|
|
MINIMUM_CONFIRMATIONS + 1));
|
|
}
|
|
Logger.print("bch utxos: $_utxos");
|
|
|
|
// placeholder for input signatures
|
|
final List<Map<dynamic, dynamic>> signatures = [];
|
|
|
|
// placeholder for total input balance
|
|
// int totalBalance = 0;
|
|
|
|
// iterate through the list of address _utxos and use them as inputs for the
|
|
// withdrawal transaction
|
|
for (var utxo in _utxos) {
|
|
// add the utxo as an input for the transaction
|
|
builder.addInput(utxo.txid, utxo.vout);
|
|
final ec =
|
|
utxoSigningData.firstWhere((e) => e.utxo.txid == utxo.txid).keyPair!;
|
|
|
|
final bitboxEC = bitbox.ECPair.fromWIF(ec.toWIF());
|
|
|
|
// add a signature to the list to be used later
|
|
signatures.add({
|
|
"vin": signatures.length,
|
|
"key_pair": bitboxEC,
|
|
"original_amount": utxo.satoshis
|
|
});
|
|
|
|
// totalBalance += utxo.satoshis;
|
|
}
|
|
|
|
// calculate the fee based on number of inputs and one expected output
|
|
// final fee =
|
|
// bitbox.BitcoinCash.getByteCount(signatures.length, recipients.length);
|
|
|
|
// calculate how much balance will be left over to spend after the fee
|
|
// final sendAmount = totalBalance - fee;
|
|
|
|
// add the output based on the address provided in the testing data
|
|
for (int i = 0; i < recipients.length; i++) {
|
|
String recipient = recipients[i];
|
|
int satoshiAmount = satoshiAmounts[i];
|
|
builder.addOutput(recipient, satoshiAmount);
|
|
}
|
|
|
|
// sign all inputs
|
|
for (var signature in signatures) {
|
|
builder.sign(
|
|
signature["vin"] as int,
|
|
signature["key_pair"] as bitbox.ECPair,
|
|
signature["original_amount"] as int);
|
|
}
|
|
|
|
// build the transaction
|
|
final tx = builder.build();
|
|
final txHex = tx.toHex();
|
|
final vSize = tx.virtualSize();
|
|
|
|
return {"hex": txHex, "vSize": vSize};
|
|
}
|
|
|
|
@override
|
|
Future<void> fullRescan(
|
|
int maxUnusedAddressGap,
|
|
int maxNumberOfIndexesToCheck,
|
|
) async {
|
|
Logging.instance.log("Starting full rescan!", level: LogLevel.Info);
|
|
longMutex = true;
|
|
GlobalEventBus.instance.fire(
|
|
WalletSyncStatusChangedEvent(
|
|
WalletSyncStatus.syncing,
|
|
walletId,
|
|
coin,
|
|
),
|
|
);
|
|
|
|
// clear cache
|
|
await _cachedElectrumXClient.clearSharedTransactionCache(coin: coin);
|
|
|
|
// back up data
|
|
// await _rescanBackup();
|
|
|
|
await db.deleteWalletBlockchainData(walletId);
|
|
|
|
try {
|
|
final _mnemonic = await mnemonicString;
|
|
final _mnemonicPassphrase = await mnemonicPassphrase;
|
|
if (_mnemonicPassphrase == null) {
|
|
Logging.instance.log(
|
|
"Exception in fullRescan: mnemonic passphrase null, possible migration issue; if using internal builds, delete wallet and restore from seed, if using a release build, please file bug report",
|
|
level: LogLevel.Error);
|
|
}
|
|
|
|
await _recoverWalletFromBIP32SeedPhrase(
|
|
mnemonic: _mnemonic!,
|
|
mnemonicPassphrase: _mnemonicPassphrase!,
|
|
maxUnusedAddressGap: maxUnusedAddressGap,
|
|
maxNumberOfIndexesToCheck: maxNumberOfIndexesToCheck,
|
|
isRescan: true,
|
|
);
|
|
|
|
longMutex = false;
|
|
await refresh();
|
|
Logging.instance.log("Full rescan complete!", level: LogLevel.Info);
|
|
GlobalEventBus.instance.fire(
|
|
WalletSyncStatusChangedEvent(
|
|
WalletSyncStatus.synced,
|
|
walletId,
|
|
coin,
|
|
),
|
|
);
|
|
} catch (e, s) {
|
|
GlobalEventBus.instance.fire(
|
|
WalletSyncStatusChangedEvent(
|
|
WalletSyncStatus.unableToSync,
|
|
walletId,
|
|
coin,
|
|
),
|
|
);
|
|
|
|
// restore from backup
|
|
// await _rescanRestore();
|
|
|
|
longMutex = false;
|
|
Logging.instance.log("Exception rethrown from fullRescan(): $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<void> _recoverWalletFromBIP32SeedPhrase({
|
|
required String mnemonic,
|
|
required String mnemonicPassphrase,
|
|
int maxUnusedAddressGap = 20,
|
|
int maxNumberOfIndexesToCheck = 1000,
|
|
bool isRescan = false,
|
|
}) async {
|
|
longMutex = true;
|
|
|
|
final root = await Bip32Utils.getBip32Root(
|
|
mnemonic,
|
|
mnemonicPassphrase,
|
|
_network,
|
|
);
|
|
|
|
final deriveTypes = [
|
|
DerivePathType.eCash44,
|
|
DerivePathType.bip44,
|
|
];
|
|
|
|
final List<Future<Tuple2<List<isar_models.Address>, DerivePathType>>>
|
|
receiveFutures = [];
|
|
final List<Future<Tuple2<List<isar_models.Address>, DerivePathType>>>
|
|
changeFutures = [];
|
|
|
|
const txCountBatchSize = 12;
|
|
const receiveChain = 0;
|
|
const changeChain = 1;
|
|
const indexZero = 0;
|
|
|
|
try {
|
|
// receiving addresses
|
|
Logging.instance.log(
|
|
"checking receiving addresses...",
|
|
level: LogLevel.Info,
|
|
);
|
|
|
|
if (serverCanBatch) {
|
|
for (final type in deriveTypes) {
|
|
receiveFutures.add(
|
|
_checkGapsBatched(
|
|
maxNumberOfIndexesToCheck,
|
|
maxUnusedAddressGap,
|
|
txCountBatchSize,
|
|
root,
|
|
type,
|
|
receiveChain,
|
|
),
|
|
);
|
|
}
|
|
} else {
|
|
for (final type in deriveTypes) {
|
|
receiveFutures.add(
|
|
_checkGaps(
|
|
maxNumberOfIndexesToCheck,
|
|
maxUnusedAddressGap,
|
|
root,
|
|
type,
|
|
receiveChain,
|
|
),
|
|
);
|
|
}
|
|
}
|
|
|
|
Logging.instance.log(
|
|
"checking change addresses...",
|
|
level: LogLevel.Info,
|
|
);
|
|
// change addresses
|
|
|
|
if (serverCanBatch) {
|
|
for (final type in deriveTypes) {
|
|
changeFutures.add(
|
|
_checkGapsBatched(
|
|
maxNumberOfIndexesToCheck,
|
|
maxUnusedAddressGap,
|
|
txCountBatchSize,
|
|
root,
|
|
type,
|
|
changeChain,
|
|
),
|
|
);
|
|
}
|
|
} else {
|
|
for (final type in deriveTypes) {
|
|
changeFutures.add(
|
|
_checkGaps(
|
|
maxNumberOfIndexesToCheck,
|
|
maxUnusedAddressGap,
|
|
root,
|
|
type,
|
|
changeChain,
|
|
),
|
|
);
|
|
}
|
|
}
|
|
|
|
// io limitations may require running these linearly instead
|
|
final futuresResult = await Future.wait([
|
|
Future.wait(receiveFutures),
|
|
Future.wait(changeFutures),
|
|
]);
|
|
|
|
final receiveResults = futuresResult[0];
|
|
final changeResults = futuresResult[1];
|
|
|
|
final List<isar_models.Address> addressesToStore = [];
|
|
|
|
// If restoring a wallet that never received any funds, then set receivingArray manually
|
|
// If we didn't do this, it'd store an empty array
|
|
for (final tuple in receiveResults) {
|
|
if (tuple.item1.isEmpty) {
|
|
final address = await _generateAddressForChain(
|
|
receiveChain,
|
|
indexZero,
|
|
tuple.item2,
|
|
);
|
|
addressesToStore.add(address);
|
|
} else {
|
|
addressesToStore.addAll(tuple.item1);
|
|
}
|
|
}
|
|
|
|
// If restoring a wallet that never sent any funds with change, then set changeArray
|
|
// manually. If we didn't do this, it'd store an empty array.
|
|
for (final tuple in changeResults) {
|
|
if (tuple.item1.isEmpty) {
|
|
final address = await _generateAddressForChain(
|
|
changeChain,
|
|
indexZero,
|
|
tuple.item2,
|
|
);
|
|
addressesToStore.add(address);
|
|
} else {
|
|
addressesToStore.addAll(tuple.item1);
|
|
}
|
|
}
|
|
|
|
if (isRescan) {
|
|
await db.updateOrPutAddresses(addressesToStore);
|
|
} else {
|
|
await db.putAddresses(addressesToStore);
|
|
}
|
|
|
|
await _updateUTXOs();
|
|
|
|
await Future.wait([
|
|
updateCachedId(walletId),
|
|
updateCachedIsFavorite(false),
|
|
]);
|
|
|
|
longMutex = false;
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown from _recoverWalletFromBIP32SeedPhrase(): $e\n$s",
|
|
level: LogLevel.Info);
|
|
|
|
longMutex = false;
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<Tuple2<List<isar_models.Address>, DerivePathType>> _checkGaps(
|
|
int maxNumberOfIndexesToCheck,
|
|
int maxUnusedAddressGap,
|
|
bip32.BIP32 root,
|
|
DerivePathType type,
|
|
int chain,
|
|
) async {
|
|
List<isar_models.Address> addressArray = [];
|
|
int gapCounter = 0;
|
|
for (int index = 0;
|
|
index < maxNumberOfIndexesToCheck && gapCounter < maxUnusedAddressGap;
|
|
index++) {
|
|
Logging.instance.log(
|
|
"index: $index, \t GapCounter chain=$chain ${type.name}: $gapCounter",
|
|
level: LogLevel.Info);
|
|
|
|
final derivePath = constructDerivePath(
|
|
derivePathType: type,
|
|
networkWIF: root.network.wif,
|
|
chain: chain,
|
|
index: index,
|
|
);
|
|
final node = await Bip32Utils.getBip32NodeFromRoot(root, derivePath);
|
|
|
|
String addressString;
|
|
final data = PaymentData(pubkey: node.publicKey);
|
|
isar_models.AddressType addressType;
|
|
switch (type) {
|
|
case DerivePathType.bip44:
|
|
case DerivePathType.eCash44:
|
|
addressString = P2PKH(data: data, network: _network).data.address!;
|
|
addressType = isar_models.AddressType.p2pkh;
|
|
addressString = bitbox.Address.toECashAddress(addressString);
|
|
break;
|
|
default:
|
|
throw Exception("DerivePathType $type not supported");
|
|
}
|
|
|
|
final address = isar_models.Address(
|
|
walletId: walletId,
|
|
value: addressString,
|
|
publicKey: node.publicKey,
|
|
type: addressType,
|
|
derivationIndex: index,
|
|
derivationPath: isar_models.DerivationPath()..value = derivePath,
|
|
subType: chain == 0
|
|
? isar_models.AddressSubType.receiving
|
|
: isar_models.AddressSubType.change,
|
|
);
|
|
|
|
// get address tx count
|
|
final count = await _getTxCount(address: address);
|
|
|
|
// check and add appropriate addresses
|
|
if (count > 0) {
|
|
// add address to array
|
|
addressArray.add(address);
|
|
// reset counter
|
|
gapCounter = 0;
|
|
// add info to derivations
|
|
} else {
|
|
// increase counter when no tx history found
|
|
gapCounter++;
|
|
}
|
|
}
|
|
|
|
return Tuple2(addressArray, type);
|
|
}
|
|
|
|
Future<Tuple2<List<isar_models.Address>, DerivePathType>> _checkGapsBatched(
|
|
int maxNumberOfIndexesToCheck,
|
|
int maxUnusedAddressGap,
|
|
int txCountBatchSize,
|
|
bip32.BIP32 root,
|
|
DerivePathType type,
|
|
int chain,
|
|
) async {
|
|
List<isar_models.Address> addressArray = [];
|
|
int gapCounter = 0;
|
|
for (int index = 0;
|
|
index < maxNumberOfIndexesToCheck && gapCounter < maxUnusedAddressGap;
|
|
index += txCountBatchSize) {
|
|
List<String> iterationsAddressArray = [];
|
|
Logging.instance.log(
|
|
"index: $index, \t GapCounter $chain ${type.name}: $gapCounter",
|
|
level: LogLevel.Info);
|
|
|
|
final _id = "k_$index";
|
|
Map<String, String> txCountCallArgs = {};
|
|
final Map<String, dynamic> receivingNodes = {};
|
|
|
|
for (int j = 0; j < txCountBatchSize; j++) {
|
|
final derivePath = constructDerivePath(
|
|
derivePathType: type,
|
|
networkWIF: root.network.wif,
|
|
chain: chain,
|
|
index: index + j,
|
|
);
|
|
final node = await Bip32Utils.getBip32NodeFromRoot(root, derivePath);
|
|
|
|
String addressString;
|
|
final data = PaymentData(pubkey: node.publicKey);
|
|
isar_models.AddressType addrType;
|
|
switch (type) {
|
|
case DerivePathType.bip44:
|
|
case DerivePathType.eCash44:
|
|
addressString = P2PKH(data: data, network: _network).data.address!;
|
|
addrType = isar_models.AddressType.p2pkh;
|
|
addressString = bitbox.Address.toECashAddress(addressString);
|
|
break;
|
|
default:
|
|
throw Exception("DerivePathType $type not supported");
|
|
}
|
|
|
|
final address = isar_models.Address(
|
|
walletId: walletId,
|
|
value: addressString,
|
|
publicKey: node.publicKey,
|
|
type: addrType,
|
|
derivationIndex: index + j,
|
|
derivationPath: isar_models.DerivationPath()..value = derivePath,
|
|
subType: chain == 0
|
|
? isar_models.AddressSubType.receiving
|
|
: isar_models.AddressSubType.change,
|
|
);
|
|
|
|
receivingNodes.addAll({
|
|
"${_id}_$j": {
|
|
"node": node,
|
|
"address": address,
|
|
}
|
|
});
|
|
txCountCallArgs.addAll({
|
|
"${_id}_$j": addressString,
|
|
});
|
|
}
|
|
|
|
// get address tx counts
|
|
final counts = await _getBatchTxCount(addresses: txCountCallArgs);
|
|
|
|
// check and add appropriate addresses
|
|
for (int k = 0; k < txCountBatchSize; k++) {
|
|
int count = counts["${_id}_$k"]!;
|
|
if (count > 0) {
|
|
final node = receivingNodes["${_id}_$k"];
|
|
final address = node["address"] as isar_models.Address;
|
|
// add address to array
|
|
addressArray.add(address);
|
|
iterationsAddressArray.add(address.value);
|
|
// set current index
|
|
// reset counter
|
|
gapCounter = 0;
|
|
// add info to derivations
|
|
}
|
|
|
|
// increase counter when no tx history found
|
|
if (count == 0) {
|
|
gapCounter++;
|
|
}
|
|
}
|
|
// cache all the transactions while waiting for the current function to finish.
|
|
unawaited(getTransactionCacheEarly(iterationsAddressArray));
|
|
}
|
|
|
|
return Tuple2(addressArray, type);
|
|
}
|
|
|
|
Future<void> getTransactionCacheEarly(List<String> allAddresses) async {
|
|
try {
|
|
final List<Map<String, dynamic>> allTxHashes =
|
|
await _fetchHistory(allAddresses);
|
|
for (final txHash in allTxHashes) {
|
|
try {
|
|
unawaited(cachedElectrumXClient.getTransaction(
|
|
txHash: txHash["tx_hash"] as String,
|
|
verbose: true,
|
|
coin: coin,
|
|
));
|
|
} catch (e) {
|
|
continue;
|
|
}
|
|
}
|
|
} catch (e) {
|
|
//
|
|
}
|
|
}
|
|
|
|
bool _shouldAutoSync = false;
|
|
|
|
@override
|
|
bool get shouldAutoSync => _shouldAutoSync;
|
|
|
|
@override
|
|
set shouldAutoSync(bool shouldAutoSync) {
|
|
if (_shouldAutoSync != shouldAutoSync) {
|
|
_shouldAutoSync = shouldAutoSync;
|
|
if (!shouldAutoSync) {
|
|
timer?.cancel();
|
|
timer = null;
|
|
stopNetworkAlivePinging();
|
|
} else {
|
|
startNetworkAlivePinging();
|
|
refresh();
|
|
}
|
|
}
|
|
}
|
|
|
|
bool isActive = false;
|
|
|
|
@override
|
|
void Function(bool)? get onIsActiveWalletChanged =>
|
|
(isActive) => this.isActive = isActive;
|
|
|
|
@override
|
|
Future<Amount> estimateFeeFor(Amount amount, int feeRate) async {
|
|
final available = balance.spendable;
|
|
|
|
if (available == amount) {
|
|
return amount - (await sweepAllEstimate(feeRate));
|
|
} else if (amount <= Amount.zero || amount > available) {
|
|
return roughFeeEstimate(1, 2, feeRate);
|
|
}
|
|
|
|
Amount runningBalance = Amount(
|
|
rawValue: BigInt.zero,
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
int inputCount = 0;
|
|
for (final output in (await utxos)) {
|
|
if (!output.isBlocked) {
|
|
runningBalance += Amount(
|
|
rawValue: BigInt.from(output.value),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
inputCount++;
|
|
if (runningBalance > amount) {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
final oneOutPutFee = roughFeeEstimate(inputCount, 1, feeRate);
|
|
final twoOutPutFee = roughFeeEstimate(inputCount, 2, feeRate);
|
|
|
|
if (runningBalance - amount > oneOutPutFee) {
|
|
if (runningBalance - amount > oneOutPutFee + DUST_LIMIT) {
|
|
final change = runningBalance - amount - twoOutPutFee;
|
|
if (change > DUST_LIMIT &&
|
|
runningBalance - amount - change == twoOutPutFee) {
|
|
return runningBalance - amount - change;
|
|
} else {
|
|
return runningBalance - amount;
|
|
}
|
|
} else {
|
|
return runningBalance - amount;
|
|
}
|
|
} else if (runningBalance - amount == oneOutPutFee) {
|
|
return oneOutPutFee;
|
|
} else {
|
|
return twoOutPutFee;
|
|
}
|
|
}
|
|
|
|
// TODO: correct formula for ecash?
|
|
Amount roughFeeEstimate(int inputCount, int outputCount, int feeRatePerKB) {
|
|
return Amount(
|
|
rawValue: BigInt.from(((181 * inputCount) + (34 * outputCount) + 10) *
|
|
(feeRatePerKB / 1000).ceil()),
|
|
fractionDigits: coin.decimals,
|
|
);
|
|
}
|
|
|
|
Future<Amount> sweepAllEstimate(int feeRate) async {
|
|
int available = 0;
|
|
int inputCount = 0;
|
|
for (final output in (await utxos)) {
|
|
if (!output.isBlocked &&
|
|
output.isConfirmed(storedChainHeight, MINIMUM_CONFIRMATIONS)) {
|
|
available += output.value;
|
|
inputCount++;
|
|
}
|
|
}
|
|
|
|
// transaction will only have 1 output minus the fee
|
|
final estimatedFee = roughFeeEstimate(inputCount, 1, feeRate);
|
|
|
|
return Amount(
|
|
rawValue: BigInt.from(available),
|
|
fractionDigits: coin.decimals,
|
|
) -
|
|
estimatedFee;
|
|
}
|
|
|
|
@override
|
|
Future<bool> generateNewAddress() async {
|
|
try {
|
|
final currentReceiving = await _currentReceivingAddress;
|
|
|
|
final newReceivingIndex = currentReceiving.derivationIndex + 1;
|
|
|
|
// Use new index to derive a new receiving address
|
|
final newReceivingAddress = await _generateAddressForChain(
|
|
0, newReceivingIndex, DerivePathTypeExt.primaryFor(coin));
|
|
|
|
// Add that new receiving address
|
|
await db.putAddress(newReceivingAddress);
|
|
|
|
return true;
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown from generateNewAddress(): $e\n$s",
|
|
level: LogLevel.Error);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
@override
|
|
Future<String> get xpub async {
|
|
final node = await Bip32Utils.getBip32Root(
|
|
(await mnemonic).join(" "),
|
|
await mnemonicPassphrase ?? "",
|
|
_network,
|
|
);
|
|
|
|
return node.neutered().toBase58();
|
|
}
|
|
|
|
@override
|
|
Future<Map<String, dynamic>> prepareSend({
|
|
required String address,
|
|
required Amount amount,
|
|
Map<String, dynamic>? args,
|
|
}) async {
|
|
try {
|
|
final feeRateType = args?["feeRate"];
|
|
final customSatsPerVByte = args?["satsPerVByte"] as int?;
|
|
final feeRateAmount = args?["feeRateAmount"];
|
|
final utxos = args?["UTXOs"] as Set<isar_models.UTXO>?;
|
|
|
|
if (customSatsPerVByte != null) {
|
|
// check for send all
|
|
bool isSendAll = false;
|
|
if (amount == balance.spendable) {
|
|
isSendAll = true;
|
|
}
|
|
|
|
final bool coinControl = utxos != null;
|
|
|
|
final result = await coinSelection(
|
|
satoshiAmountToSend: amount.raw.toInt(),
|
|
selectedTxFeeRate: -1,
|
|
satsPerVByte: customSatsPerVByte,
|
|
recipientAddress: address,
|
|
isSendAll: isSendAll,
|
|
utxos: utxos?.toList(),
|
|
coinControl: coinControl,
|
|
);
|
|
|
|
Logging.instance
|
|
.log("PREPARE SEND RESULT: $result", level: LogLevel.Info);
|
|
if (result is int) {
|
|
switch (result) {
|
|
case 1:
|
|
throw Exception("Insufficient balance!");
|
|
case 2:
|
|
throw Exception("Insufficient funds to pay for transaction fee!");
|
|
default:
|
|
throw Exception("Transaction failed with error code $result");
|
|
}
|
|
} else {
|
|
final hex = result["hex"];
|
|
if (hex is String) {
|
|
final fee = result["fee"] as int;
|
|
final vSize = result["vSize"] as int;
|
|
|
|
Logging.instance.log("txHex: $hex", level: LogLevel.Info);
|
|
Logging.instance.log("fee: $fee", level: LogLevel.Info);
|
|
Logging.instance.log("vsize: $vSize", level: LogLevel.Info);
|
|
// fee should never be less than vSize sanity check
|
|
if (fee < vSize) {
|
|
throw Exception(
|
|
"Error in fee calculation: Transaction fee cannot be less than vSize");
|
|
}
|
|
return result as Map<String, dynamic>;
|
|
} else {
|
|
throw Exception("sent hex is not a String!!!");
|
|
}
|
|
}
|
|
} else if (feeRateType is FeeRateType || feeRateAmount is int) {
|
|
late final int rate;
|
|
if (feeRateType is FeeRateType) {
|
|
int fee = 0;
|
|
final feeObject = await fees;
|
|
switch (feeRateType) {
|
|
case FeeRateType.fast:
|
|
fee = feeObject.fast;
|
|
break;
|
|
case FeeRateType.average:
|
|
fee = feeObject.medium;
|
|
break;
|
|
case FeeRateType.slow:
|
|
fee = feeObject.slow;
|
|
break;
|
|
|
|
default:
|
|
throw ArgumentError("Invalid use of custom fee");
|
|
}
|
|
rate = fee;
|
|
} else {
|
|
rate = feeRateAmount as int;
|
|
}
|
|
|
|
// check for send all
|
|
bool isSendAll = false;
|
|
if (amount == balance.spendable) {
|
|
isSendAll = true;
|
|
}
|
|
|
|
final bool coinControl = utxos != null;
|
|
|
|
final txData = await coinSelection(
|
|
satoshiAmountToSend: amount.raw.toInt(),
|
|
selectedTxFeeRate: rate,
|
|
recipientAddress: address,
|
|
isSendAll: isSendAll,
|
|
utxos: utxos?.toList(),
|
|
coinControl: coinControl,
|
|
);
|
|
|
|
Logging.instance.log("prepare send: $txData", level: LogLevel.Info);
|
|
try {
|
|
if (txData is int) {
|
|
switch (txData) {
|
|
case 1:
|
|
throw Exception("Insufficient balance!");
|
|
case 2:
|
|
throw Exception(
|
|
"Insufficient funds to pay for transaction fee!");
|
|
default:
|
|
throw Exception("Transaction failed with error code $txData");
|
|
}
|
|
} else {
|
|
final hex = txData["hex"];
|
|
|
|
if (hex is String) {
|
|
final fee = txData["fee"] as int;
|
|
final vSize = txData["vSize"] as int;
|
|
|
|
Logging.instance
|
|
.log("prepared txHex: $hex", level: LogLevel.Info);
|
|
Logging.instance.log("prepared fee: $fee", level: LogLevel.Info);
|
|
Logging.instance
|
|
.log("prepared vSize: $vSize", level: LogLevel.Info);
|
|
|
|
// fee should never be less than vSize sanity check
|
|
if (fee < vSize) {
|
|
throw Exception(
|
|
"Error in fee calculation: Transaction fee cannot be less than vSize");
|
|
}
|
|
|
|
return txData as Map<String, dynamic>;
|
|
} else {
|
|
throw Exception("prepared hex is not a String!!!");
|
|
}
|
|
}
|
|
} catch (e, s) {
|
|
Logging.instance.log("Exception rethrown from prepareSend(): $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
} else {
|
|
throw ArgumentError("Invalid fee rate argument provided!");
|
|
}
|
|
} catch (e, s) {
|
|
Logging.instance.log("Exception rethrown from prepareSend(): $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
@override
|
|
Future<String> confirmSend({required Map<String, dynamic> txData}) async {
|
|
try {
|
|
Logging.instance.log("confirmSend txData: $txData", level: LogLevel.Info);
|
|
|
|
final hex = txData["hex"] as String;
|
|
|
|
final txHash = await _electrumXClient.broadcastTransaction(rawTx: hex);
|
|
Logging.instance.log("Sent txHash: $txHash", level: LogLevel.Info);
|
|
|
|
final utxos = txData["usedUTXOs"] as List<isar_models.UTXO>;
|
|
|
|
// mark utxos as used
|
|
await db.putUTXOs(utxos.map((e) => e.copyWith(used: true)).toList());
|
|
|
|
return txHash;
|
|
} catch (e, s) {
|
|
Logging.instance.log("Exception rethrown from confirmSend(): $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
@override
|
|
Future<bool> testNetworkConnection() async {
|
|
try {
|
|
final result = await _electrumXClient.ping();
|
|
return result;
|
|
} catch (_) {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
Timer? _networkAliveTimer;
|
|
|
|
void startNetworkAlivePinging() {
|
|
// call once on start right away
|
|
_periodicPingCheck();
|
|
|
|
// then periodically check
|
|
_networkAliveTimer = Timer.periodic(
|
|
Constants.networkAliveTimerDuration,
|
|
(_) async {
|
|
_periodicPingCheck();
|
|
},
|
|
);
|
|
}
|
|
|
|
void _periodicPingCheck() async {
|
|
bool hasNetwork = await testNetworkConnection();
|
|
|
|
if (_isConnected != hasNetwork) {
|
|
NodeConnectionStatus status = hasNetwork
|
|
? NodeConnectionStatus.connected
|
|
: NodeConnectionStatus.disconnected;
|
|
GlobalEventBus.instance
|
|
.fire(NodeConnectionStatusChangedEvent(status, walletId, coin));
|
|
|
|
_isConnected = hasNetwork;
|
|
if (hasNetwork) {
|
|
unawaited(refresh());
|
|
}
|
|
}
|
|
}
|
|
|
|
void stopNetworkAlivePinging() {
|
|
_networkAliveTimer?.cancel();
|
|
_networkAliveTimer = null;
|
|
}
|
|
|
|
bool _isConnected = false;
|
|
|
|
@override
|
|
bool get isConnected => _isConnected;
|
|
|
|
@override
|
|
Future<void> initializeNew() async {
|
|
Logging.instance
|
|
.log("Generating new ${coin.prettyName} wallet.", level: LogLevel.Info);
|
|
|
|
if (getCachedId() != null) {
|
|
throw Exception(
|
|
"Attempted to initialize a new wallet using an existing wallet ID!");
|
|
}
|
|
|
|
await _prefs.init();
|
|
try {
|
|
await _generateNewWallet();
|
|
} catch (e, s) {
|
|
Logging.instance.log("Exception rethrown from initializeNew(): $e\n$s",
|
|
level: LogLevel.Fatal);
|
|
rethrow;
|
|
}
|
|
await Future.wait([
|
|
updateCachedId(walletId),
|
|
updateCachedIsFavorite(false),
|
|
]);
|
|
}
|
|
|
|
@override
|
|
Future<void> initializeExisting() async {
|
|
Logging.instance.log("initializeExisting() ${coin.prettyName} wallet.",
|
|
level: LogLevel.Info);
|
|
|
|
try {
|
|
final features = await electrumXClient.getServerFeatures();
|
|
_serverVersion =
|
|
_parseServerVersion(features["server_version"] as String);
|
|
} catch (_) {
|
|
// catch nothing as failure here means we just do not batch certain rpc
|
|
// calls
|
|
}
|
|
|
|
if (getCachedId() == null) {
|
|
throw Exception(
|
|
"Attempted to initialize an existing wallet using an unknown wallet ID!");
|
|
}
|
|
|
|
await _prefs.init();
|
|
// await _checkCurrentChangeAddressesForTransactions();
|
|
// await _checkCurrentReceivingAddressesForTransactions();
|
|
}
|
|
|
|
// hack to add tx to txData before refresh completes
|
|
// required based on current app architecture where we don't properly store
|
|
// transactions locally in a good way
|
|
@override
|
|
Future<void> updateSentCachedTxData(Map<String, dynamic> txData) async {
|
|
final transaction = isar_models.Transaction(
|
|
walletId: walletId,
|
|
txid: txData["txid"] as String,
|
|
timestamp: DateTime.now().millisecondsSinceEpoch ~/ 1000,
|
|
type: isar_models.TransactionType.outgoing,
|
|
subType: isar_models.TransactionSubType.none,
|
|
// precision may be lost here hence the following amountString
|
|
amount: (txData["recipientAmt"] as Amount).raw.toInt(),
|
|
amountString: (txData["recipientAmt"] as Amount).toJsonString(),
|
|
fee: txData["fee"] as int,
|
|
height: null,
|
|
isCancelled: false,
|
|
isLelantus: false,
|
|
otherData: null,
|
|
slateId: null,
|
|
nonce: null,
|
|
inputs: [],
|
|
outputs: [],
|
|
numberOfMessages: null,
|
|
);
|
|
|
|
final address = txData["address"] is String
|
|
? await db.getAddress(walletId, txData["address"] as String)
|
|
: null;
|
|
|
|
await db.addNewTransactionData(
|
|
[
|
|
Tuple2(transaction, address),
|
|
],
|
|
walletId,
|
|
);
|
|
}
|
|
|
|
@override
|
|
bool get isRefreshing => refreshMutex;
|
|
|
|
bool refreshMutex = false;
|
|
|
|
//TODO Show percentages properly/more consistently
|
|
/// Refreshes display data for the wallet
|
|
@override
|
|
Future<void> refresh() async {
|
|
if (refreshMutex) {
|
|
Logging.instance.log("$walletId $walletName refreshMutex denied",
|
|
level: LogLevel.Info);
|
|
return;
|
|
} else {
|
|
refreshMutex = true;
|
|
}
|
|
|
|
try {
|
|
GlobalEventBus.instance.fire(
|
|
WalletSyncStatusChangedEvent(
|
|
WalletSyncStatus.syncing,
|
|
walletId,
|
|
coin,
|
|
),
|
|
);
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.0, walletId));
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.1, walletId));
|
|
|
|
final currentHeight = await chainHeight;
|
|
const storedHeight = 1; //await storedChainHeight;
|
|
|
|
Logging.instance
|
|
.log("chain height: $currentHeight", level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log("cached height: $storedHeight", level: LogLevel.Info);
|
|
|
|
if (currentHeight != storedHeight) {
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.2, walletId));
|
|
await _checkCurrentChangeAddressesForTransactions();
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.3, walletId));
|
|
await _checkCurrentReceivingAddressesForTransactions();
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.4, walletId));
|
|
|
|
final fetchFuture = _refreshTransactions();
|
|
GlobalEventBus.instance
|
|
.fire(RefreshPercentChangedEvent(0.50, walletId));
|
|
|
|
final feeObj = _getFees();
|
|
GlobalEventBus.instance
|
|
.fire(RefreshPercentChangedEvent(0.60, walletId));
|
|
|
|
GlobalEventBus.instance
|
|
.fire(RefreshPercentChangedEvent(0.70, walletId));
|
|
_feeObject = Future(() => feeObj);
|
|
|
|
await fetchFuture;
|
|
GlobalEventBus.instance
|
|
.fire(RefreshPercentChangedEvent(0.80, walletId));
|
|
|
|
await _updateUTXOs();
|
|
await getAllTxsToWatch();
|
|
GlobalEventBus.instance
|
|
.fire(RefreshPercentChangedEvent(0.90, walletId));
|
|
}
|
|
|
|
refreshMutex = false;
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(1.0, walletId));
|
|
GlobalEventBus.instance.fire(
|
|
WalletSyncStatusChangedEvent(
|
|
WalletSyncStatus.synced,
|
|
walletId,
|
|
coin,
|
|
),
|
|
);
|
|
|
|
if (shouldAutoSync) {
|
|
timer ??= Timer.periodic(const Duration(seconds: 30), (timer) async {
|
|
Logging.instance.log(
|
|
"Periodic refresh check for $walletId $walletName in object instance: $hashCode",
|
|
level: LogLevel.Info);
|
|
// chain height check currently broken
|
|
// if ((await chainHeight) != (await storedChainHeight)) {
|
|
if (await refreshIfThereIsNewData()) {
|
|
await refresh();
|
|
GlobalEventBus.instance.fire(UpdatedInBackgroundEvent(
|
|
"New data found in $walletId $walletName in background!",
|
|
walletId));
|
|
}
|
|
// }
|
|
});
|
|
}
|
|
} catch (error, strace) {
|
|
refreshMutex = false;
|
|
GlobalEventBus.instance.fire(
|
|
NodeConnectionStatusChangedEvent(
|
|
NodeConnectionStatus.disconnected,
|
|
walletId,
|
|
coin,
|
|
),
|
|
);
|
|
GlobalEventBus.instance.fire(
|
|
WalletSyncStatusChangedEvent(
|
|
WalletSyncStatus.unableToSync,
|
|
walletId,
|
|
coin,
|
|
),
|
|
);
|
|
Logging.instance.log(
|
|
"Caught exception in refreshWalletData(): $error\n$strace",
|
|
level: LogLevel.Error);
|
|
}
|
|
}
|
|
|
|
Future<bool> refreshIfThereIsNewData() async {
|
|
if (longMutex) return false;
|
|
if (_hasCalledExit) return false;
|
|
Logging.instance.log("refreshIfThereIsNewData", level: LogLevel.Info);
|
|
|
|
try {
|
|
bool needsRefresh = false;
|
|
Set<String> txnsToCheck = {};
|
|
|
|
for (final String txid in txTracker.pendings) {
|
|
if (!txTracker.wasNotifiedConfirmed(txid)) {
|
|
txnsToCheck.add(txid);
|
|
}
|
|
}
|
|
|
|
for (String txid in txnsToCheck) {
|
|
final txn = await electrumXClient.getTransaction(txHash: txid);
|
|
int confirmations = txn["confirmations"] as int? ?? 0;
|
|
bool isUnconfirmed = confirmations < MINIMUM_CONFIRMATIONS;
|
|
if (!isUnconfirmed) {
|
|
// unconfirmedTxs = {};
|
|
needsRefresh = true;
|
|
break;
|
|
}
|
|
}
|
|
if (!needsRefresh) {
|
|
final allOwnAddresses = await _fetchAllOwnAddresses();
|
|
List<Map<String, dynamic>> allTxs = await _fetchHistory(
|
|
allOwnAddresses.map((e) => e.value).toList(growable: false));
|
|
for (Map<String, dynamic> transaction in allTxs) {
|
|
final txid = transaction['tx_hash'] as String;
|
|
if ((await db
|
|
.getTransactions(walletId)
|
|
.filter()
|
|
.txidMatches(txid)
|
|
.findFirst()) ==
|
|
null) {
|
|
Logging.instance.log(
|
|
" txid not found in address history already ${transaction['tx_hash']}",
|
|
level: LogLevel.Info);
|
|
needsRefresh = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
return needsRefresh;
|
|
} on NoSuchTransactionException catch (e) {
|
|
// TODO: move direct transactions elsewhere
|
|
await db.isar.writeTxn(() async {
|
|
await db.isar.transactions.deleteByTxidWalletId(e.txid, walletId);
|
|
});
|
|
await txTracker.deleteTransaction(e.txid);
|
|
return true;
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception caught in refreshIfThereIsNewData: $e\n$s",
|
|
level: LogLevel.Error);
|
|
rethrow;
|
|
}
|
|
}
|
|
|
|
Future<void> getAllTxsToWatch() async {
|
|
if (_hasCalledExit) return;
|
|
List<isar_models.Transaction> unconfirmedTxnsToNotifyPending = [];
|
|
List<isar_models.Transaction> unconfirmedTxnsToNotifyConfirmed = [];
|
|
|
|
final currentChainHeight = await chainHeight;
|
|
|
|
final txCount = await db.getTransactions(walletId).count();
|
|
|
|
const paginateLimit = 50;
|
|
|
|
for (int i = 0; i < txCount; i += paginateLimit) {
|
|
final transactions = await db
|
|
.getTransactions(walletId)
|
|
.offset(i)
|
|
.limit(paginateLimit)
|
|
.findAll();
|
|
for (final tx in transactions) {
|
|
if (tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) {
|
|
// get all transactions that were notified as pending but not as confirmed
|
|
if (txTracker.wasNotifiedPending(tx.txid) &&
|
|
!txTracker.wasNotifiedConfirmed(tx.txid)) {
|
|
unconfirmedTxnsToNotifyConfirmed.add(tx);
|
|
}
|
|
} else {
|
|
// get all transactions that were not notified as pending yet
|
|
if (!txTracker.wasNotifiedPending(tx.txid)) {
|
|
unconfirmedTxnsToNotifyPending.add(tx);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// notify on unconfirmed transactions
|
|
for (final tx in unconfirmedTxnsToNotifyPending) {
|
|
final confirmations = tx.getConfirmations(currentChainHeight);
|
|
|
|
if (tx.type == isar_models.TransactionType.incoming) {
|
|
CryptoNotificationsEventBus.instance.fire(
|
|
CryptoNotificationEvent(
|
|
title: "Incoming transaction",
|
|
walletId: walletId,
|
|
walletName: walletName,
|
|
date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
|
shouldWatchForUpdates: confirmations < MINIMUM_CONFIRMATIONS,
|
|
coin: coin,
|
|
txid: tx.txid,
|
|
confirmations: confirmations,
|
|
requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
),
|
|
);
|
|
await txTracker.addNotifiedPending(tx.txid);
|
|
} else if (tx.type == isar_models.TransactionType.outgoing) {
|
|
CryptoNotificationsEventBus.instance.fire(
|
|
CryptoNotificationEvent(
|
|
title: "Sending transaction",
|
|
walletId: walletId,
|
|
date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
|
shouldWatchForUpdates: confirmations < MINIMUM_CONFIRMATIONS,
|
|
txid: tx.txid,
|
|
confirmations: confirmations,
|
|
requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
walletName: walletName,
|
|
coin: coin,
|
|
),
|
|
);
|
|
await txTracker.addNotifiedPending(tx.txid);
|
|
}
|
|
}
|
|
|
|
// notify on confirmed
|
|
for (final tx in unconfirmedTxnsToNotifyConfirmed) {
|
|
if (tx.type == isar_models.TransactionType.incoming) {
|
|
CryptoNotificationsEventBus.instance.fire(
|
|
CryptoNotificationEvent(
|
|
title: "Incoming transaction confirmed",
|
|
walletId: walletId,
|
|
date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
|
shouldWatchForUpdates: false,
|
|
txid: tx.txid,
|
|
requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
walletName: walletName,
|
|
coin: coin,
|
|
),
|
|
);
|
|
await txTracker.addNotifiedConfirmed(tx.txid);
|
|
} else if (tx.type == isar_models.TransactionType.outgoing) {
|
|
CryptoNotificationsEventBus.instance.fire(
|
|
CryptoNotificationEvent(
|
|
title: "Outgoing transaction confirmed",
|
|
walletId: walletId,
|
|
date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
|
shouldWatchForUpdates: false,
|
|
txid: tx.txid,
|
|
requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
walletName: walletName,
|
|
coin: coin,
|
|
),
|
|
);
|
|
await txTracker.addNotifiedConfirmed(tx.txid);
|
|
}
|
|
}
|
|
}
|
|
|
|
@override
|
|
Future<void> recoverFromMnemonic({
|
|
required String mnemonic,
|
|
String? mnemonicPassphrase,
|
|
required int maxUnusedAddressGap,
|
|
required int maxNumberOfIndexesToCheck,
|
|
required int height,
|
|
}) async {
|
|
longMutex = true;
|
|
final start = DateTime.now();
|
|
try {
|
|
Logging.instance.log("IS_INTEGRATION_TEST: $integrationTestFlag",
|
|
level: LogLevel.Info);
|
|
if (!integrationTestFlag) {
|
|
final features = await electrumXClient.getServerFeatures();
|
|
Logging.instance.log("features: $features", level: LogLevel.Info);
|
|
_serverVersion =
|
|
_parseServerVersion(features["server_version"] as String);
|
|
switch (coin) {
|
|
case Coin.eCash:
|
|
if (features['genesis_hash'] != GENESIS_HASH_MAINNET) {
|
|
throw Exception("genesis hash does not match main net!");
|
|
}
|
|
break;
|
|
// case Coin.bitcoinTestNet:
|
|
// if (features['genesis_hash'] != GENESIS_HASH_TESTNET) {
|
|
// throw Exception("genesis hash does not match test net!");
|
|
// }
|
|
// break;
|
|
default:
|
|
throw Exception(
|
|
"Attempted to generate a ECashWallet using a non eCash coin type: ${coin.name}");
|
|
}
|
|
}
|
|
// check to make sure we aren't overwriting a mnemonic
|
|
// this should never fail
|
|
if ((await mnemonicString) != null ||
|
|
(await this.mnemonicPassphrase) != null) {
|
|
longMutex = false;
|
|
throw Exception("Attempted to overwrite mnemonic on restore!");
|
|
}
|
|
await _secureStore.write(
|
|
key: '${_walletId}_mnemonic', value: mnemonic.trim());
|
|
await _secureStore.write(
|
|
key: '${_walletId}_mnemonicPassphrase',
|
|
value: mnemonicPassphrase ?? "",
|
|
);
|
|
|
|
await _recoverWalletFromBIP32SeedPhrase(
|
|
mnemonic: mnemonic.trim(),
|
|
mnemonicPassphrase: mnemonicPassphrase ?? "",
|
|
maxUnusedAddressGap: maxUnusedAddressGap,
|
|
maxNumberOfIndexesToCheck: maxNumberOfIndexesToCheck,
|
|
);
|
|
} catch (e, s) {
|
|
Logging.instance.log(
|
|
"Exception rethrown from recoverFromMnemonic(): $e\n$s",
|
|
level: LogLevel.Error);
|
|
longMutex = false;
|
|
rethrow;
|
|
}
|
|
longMutex = false;
|
|
|
|
final end = DateTime.now();
|
|
Logging.instance.log(
|
|
"$walletName recovery time: ${end.difference(start).inMilliseconds} millis",
|
|
level: LogLevel.Info);
|
|
}
|
|
}
|