import 'dart:async'; import 'dart:convert'; import 'dart:io'; import 'dart:isolate'; import 'dart:math'; import 'package:bip32/bip32.dart' as bip32; import 'package:bip39/bip39.dart' as bip39; import 'package:bitcoindart/bitcoindart.dart'; import 'package:decimal/decimal.dart'; import 'package:flutter/foundation.dart'; import 'package:isar/isar.dart'; import 'package:lelantus/lelantus.dart'; import 'package:stackwallet/electrumx_rpc/cached_electrumx.dart'; import 'package:stackwallet/electrumx_rpc/electrumx.dart'; import 'package:stackwallet/models/balance.dart'; import 'package:stackwallet/models/isar/models/isar_models.dart' as isar_models; import 'package:stackwallet/models/lelantus_coin.dart'; import 'package:stackwallet/models/lelantus_fee_data.dart'; import 'package:stackwallet/models/paymint/fee_object_model.dart'; import 'package:stackwallet/services/coins/coin_service.dart'; import 'package:stackwallet/services/event_bus/events/global/node_connection_status_changed_event.dart'; import 'package:stackwallet/services/event_bus/events/global/refresh_percent_changed_event.dart'; import 'package:stackwallet/services/event_bus/events/global/updated_in_background_event.dart'; import 'package:stackwallet/services/event_bus/events/global/wallet_sync_status_changed_event.dart'; import 'package:stackwallet/services/event_bus/global_event_bus.dart'; import 'package:stackwallet/services/mixins/wallet_cache.dart'; import 'package:stackwallet/services/mixins/wallet_db.dart'; import 'package:stackwallet/services/node_service.dart'; import 'package:stackwallet/services/notifications_api.dart'; import 'package:stackwallet/services/transaction_notification_tracker.dart'; import 'package:stackwallet/utilities/address_utils.dart'; import 'package:stackwallet/utilities/assets.dart'; import 'package:stackwallet/utilities/constants.dart'; import 'package:stackwallet/utilities/default_nodes.dart'; import 'package:stackwallet/utilities/enums/coin_enum.dart'; import 'package:stackwallet/utilities/enums/fee_rate_type_enum.dart'; import 'package:stackwallet/utilities/flutter_secure_storage_interface.dart'; import 'package:stackwallet/utilities/format.dart'; import 'package:stackwallet/utilities/logger.dart'; import 'package:stackwallet/utilities/prefs.dart'; import 'package:tuple/tuple.dart'; import 'package:uuid/uuid.dart'; import '../../mixins/firo_hive.dart'; const DUST_LIMIT = 1000; const MINIMUM_CONFIRMATIONS = 1; const MINT_LIMIT = 100100000000; const int LELANTUS_VALUE_SPEND_LIMIT_PER_TRANSACTION = 5001 * 100000000; const JMINT_INDEX = 5; const MINT_INDEX = 2; const TRANSACTION_LELANTUS = 8; const ANONYMITY_SET_EMPTY_ID = 0; const String GENESIS_HASH_MAINNET = "4381deb85b1b2c9843c222944b616d997516dcbd6a964e1eaf0def0830695233"; const String GENESIS_HASH_TESTNET = "aa22adcc12becaf436027ffe62a8fb21b234c58c23865291e5dc52cf53f64fca"; final firoNetwork = NetworkType( messagePrefix: '\x18Zcoin Signed Message:\n', bech32: 'bc', bip32: Bip32Type(public: 0x0488b21e, private: 0x0488ade4), pubKeyHash: 0x52, scriptHash: 0x07, wif: 0xd2); final firoTestNetwork = NetworkType( messagePrefix: '\x18Zcoin Signed Message:\n', bech32: 'bc', bip32: Bip32Type(public: 0x043587cf, private: 0x04358394), pubKeyHash: 0x41, scriptHash: 0xb2, wif: 0xb9); // isolate Map isolates = {}; Future getIsolate(Map arguments) async { ReceivePort receivePort = ReceivePort(); //port for isolate to receive messages. arguments['sendPort'] = receivePort.sendPort; Logging.instance .log("starting isolate ${arguments['function']}", level: LogLevel.Info); Isolate isolate = await Isolate.spawn(executeNative, arguments); Logging.instance.log("isolate spawned!", level: LogLevel.Info); isolates[receivePort] = isolate; return receivePort; } Future executeNative(Map arguments) async { await Logging.instance.initInIsolate(); final sendPort = arguments['sendPort'] as SendPort; final function = arguments['function'] as String; try { if (function == "createJoinSplit") { final spendAmount = arguments['spendAmount'] as int; final address = arguments['address'] as String; final subtractFeeFromAmount = arguments['subtractFeeFromAmount'] as bool; final mnemonic = arguments['mnemonic'] as String; final index = arguments['index'] as int; final lelantusEntries = arguments['lelantusEntries'] as List; final coin = arguments['coin'] as Coin; final network = arguments['network'] as NetworkType?; final locktime = arguments['locktime'] as int; final anonymitySets = arguments['_anonymity_sets'] as List?; if (!(network == null || anonymitySets == null)) { var joinSplit = await isolateCreateJoinSplitTransaction( spendAmount, address, subtractFeeFromAmount, mnemonic, index, lelantusEntries, locktime, coin, network, anonymitySets, ); sendPort.send(joinSplit); return; } } else if (function == "estimateJoinSplit") { final spendAmount = arguments['spendAmount'] as int; final subtractFeeFromAmount = arguments['subtractFeeFromAmount'] as bool?; final lelantusEntries = arguments['lelantusEntries'] as List; final coin = arguments['coin'] as Coin; if (!(subtractFeeFromAmount == null)) { var feeData = await isolateEstimateJoinSplitFee( spendAmount, subtractFeeFromAmount, lelantusEntries, coin); sendPort.send(feeData); return; } } else if (function == "restore") { final latestSetId = arguments['latestSetId'] as int; final setDataMap = arguments['setDataMap'] as Map; final usedSerialNumbers = arguments['usedSerialNumbers'] as List?; final mnemonic = arguments['mnemonic'] as String; final coin = arguments['coin'] as Coin; final network = arguments['network'] as NetworkType?; if (!(usedSerialNumbers == null || network == null)) { var restoreData = await isolateRestore( mnemonic, coin, latestSetId, setDataMap, usedSerialNumbers, network, ); sendPort.send(restoreData); return; } } else if (function == "isolateDerive") { final mnemonic = arguments['mnemonic'] as String; final from = arguments['from'] as int; final to = arguments['to'] as int; final network = arguments['network'] as NetworkType?; if (!(network == null)) { var derived = await isolateDerive(mnemonic, from, to, network); sendPort.send(derived); return; } } Logging.instance.log( "Error Arguments for $function not formatted correctly", level: LogLevel.Fatal); sendPort.send("Error"); } catch (e, s) { Logging.instance.log( "An error was thrown in this isolate $function: $e\n$s", level: LogLevel.Error); sendPort.send("Error"); } finally { await Logging.instance.isar?.close(); } } void stop(ReceivePort port) { Isolate? isolate = isolates.remove(port); if (isolate != null) { Logging.instance.log('Stopping Isolate...', level: LogLevel.Info); isolate.kill(priority: Isolate.immediate); isolate = null; } } Future> isolateDerive( String mnemonic, int from, int to, NetworkType _network) async { Map result = {}; Map allReceive = {}; Map allChange = {}; final root = getBip32Root(mnemonic, _network); for (int i = from; i < to; i++) { var currentNode = getBip32NodeFromRoot(0, i, root); var address = P2PKH( network: _network, data: PaymentData(pubkey: currentNode.publicKey)) .data .address!; allReceive["$i"] = { "publicKey": Format.uint8listToString(currentNode.publicKey), "wif": currentNode.toWIF(), "address": address, }; currentNode = getBip32NodeFromRoot(1, i, root); address = P2PKH( network: _network, data: PaymentData(pubkey: currentNode.publicKey)) .data .address!; allChange["$i"] = { "publicKey": Format.uint8listToString(currentNode.publicKey), "wif": currentNode.toWIF(), "address": address, }; if (i % 50 == 0) { Logging.instance.log("thread at $i", level: LogLevel.Info); } } result['receive'] = allReceive; result['change'] = allChange; return result; } Future> isolateRestore( String mnemonic, Coin coin, int _latestSetId, Map _setDataMap, List _usedSerialNumbers, NetworkType network, ) async { List jindexes = []; List> lelantusCoins = []; final List spendTxIds = []; var lastFoundIndex = 0; var currentIndex = 0; try { final usedSerialNumbers = _usedSerialNumbers; Set usedSerialNumbersSet = {}; for (int ind = 0; ind < usedSerialNumbers.length; ind++) { usedSerialNumbersSet.add(usedSerialNumbers[ind]); } final root = getBip32Root(mnemonic, network); while (currentIndex < lastFoundIndex + 50) { final mintKeyPair = getBip32NodeFromRoot(MINT_INDEX, currentIndex, root); final mintTag = CreateTag( Format.uint8listToString(mintKeyPair.privateKey!), currentIndex, Format.uint8listToString(mintKeyPair.identifier), isTestnet: coin == Coin.firoTestNet); for (var setId = 1; setId <= _latestSetId; setId++) { final setData = _setDataMap[setId]; final foundCoin = setData["coins"].firstWhere( (dynamic e) => e[1] == mintTag, orElse: () => []); if (foundCoin.length == 4) { lastFoundIndex = currentIndex; if (foundCoin[2] is int) { final amount = foundCoin[2] as int; final serialNumber = GetSerialNumber(amount, Format.uint8listToString(mintKeyPair.privateKey!), currentIndex, isTestnet: coin == Coin.firoTestNet); String publicCoin = foundCoin[0] as String; String txId = foundCoin[3] as String; bool isUsed = usedSerialNumbersSet.contains(serialNumber); final duplicateCoin = lelantusCoins.firstWhere((element) { final coin = element.values.first; return coin.txId == txId && coin.index == currentIndex && coin.anonymitySetId != setId; }, orElse: () => {}); if (duplicateCoin.isNotEmpty) { //todo: check if print needed // debugPrint("removing duplicate: $duplicateCoin"); lelantusCoins.remove(duplicateCoin); } lelantusCoins.add({ publicCoin: LelantusCoin( currentIndex, amount, publicCoin, txId, setId, isUsed, ) }); Logging.instance .log("amount $amount used $isUsed", level: LogLevel.Info); } else { final keyPath = GetAesKeyPath(foundCoin[0] as String); final aesKeyPair = getBip32NodeFromRoot(JMINT_INDEX, keyPath, root); if (aesKeyPair.privateKey != null) { final aesPrivateKey = Format.uint8listToString(aesKeyPair.privateKey!); final amount = decryptMintAmount( aesPrivateKey, foundCoin[2] as String, ); final serialNumber = GetSerialNumber( amount, Format.uint8listToString(mintKeyPair.privateKey!), currentIndex, isTestnet: coin == Coin.firoTestNet); String publicCoin = foundCoin[0] as String; String txId = foundCoin[3] as String; bool isUsed = usedSerialNumbersSet.contains(serialNumber); final duplicateCoin = lelantusCoins.firstWhere((element) { final coin = element.values.first; return coin.txId == txId && coin.index == currentIndex && coin.anonymitySetId != setId; }, orElse: () => {}); if (duplicateCoin.isNotEmpty) { //todo: check if print needed // debugPrint("removing duplicate: $duplicateCoin"); lelantusCoins.remove(duplicateCoin); } lelantusCoins.add({ '${foundCoin[3]!}': LelantusCoin( currentIndex, amount, publicCoin, txId, setId, isUsed, ) }); jindexes.add(currentIndex); spendTxIds.add(foundCoin[3] as String); } } } } currentIndex++; } } catch (e, s) { Logging.instance.log("Exception rethrown from isolateRestore(): $e\n$s", level: LogLevel.Info); rethrow; } Map result = {}; // Logging.instance.log("mints $lelantusCoins", addToDebugMessagesDB: false); // Logging.instance.log("jmints $spendTxIds", addToDebugMessagesDB: false); result['_lelantus_coins'] = lelantusCoins; result['mintIndex'] = lastFoundIndex + 1; result['jindex'] = jindexes; result['spendTxIds'] = spendTxIds; return result; } Future> staticProcessRestore( List txns, Map result, int currentHeight, ) async { List? _l = result['_lelantus_coins'] as List?; final List> lelantusCoins = []; for (var el in _l ?? []) { lelantusCoins.add({el.keys.first: el.values.first as LelantusCoin}); } // Edit the receive transactions with the mint fees. Map editedTransactions = {}; for (var item in lelantusCoins) { item.forEach((key, value) { String txid = value.txId; isar_models.Transaction? tx; try { tx = txns.firstWhere((e) => e.txid == txid); } catch (_) { tx = null; } if (tx == null) { // This is a jmint. return; } List inputs = []; for (var element in tx.inputs) { isar_models.Transaction? input; try { input = txns.firstWhere((e) => e.txid == element.txid); } catch (_) { input = null; } if (input != null) { inputs.add(input); } } if (inputs.isEmpty) { //some error. return; } int mintFee = tx.fee; int sharedFee = mintFee ~/ inputs.length; for (var element in inputs) { editedTransactions[element.txid] = isar_models.Transaction( walletId: element.walletId, txid: element.txid, timestamp: element.timestamp, type: element.type, subType: isar_models.TransactionSubType.mint, amount: element.amount, fee: sharedFee, height: element.height, isCancelled: false, isLelantus: true, slateId: null, otherData: txid, ) ..inputs.addAll(element.inputs) ..outputs.addAll(element.outputs) ..address.value = element.address.value; } }); } // Logging.instance.log(editedTransactions, addToDebugMessagesDB: false); Map transactionMap = {}; for (final e in txns) { transactionMap[e.txid] = e; } // Logging.instance.log(transactionMap, addToDebugMessagesDB: false); editedTransactions.forEach((key, value) { transactionMap.update(key, (_value) => value); }); transactionMap.removeWhere((key, value) => lelantusCoins.any((element) => element.containsKey(key)) || ((value.height == -1 || value.height == null) && !value.isConfirmed(currentHeight, MINIMUM_CONFIRMATIONS))); result['newTxMap'] = transactionMap; return result; } Future isolateEstimateJoinSplitFee( int spendAmount, bool subtractFeeFromAmount, List lelantusEntries, Coin coin) async { Logging.instance.log("estimateJoinsplit fee", level: LogLevel.Info); // for (int i = 0; i < lelantusEntries.length; i++) { // Logging.instance.log(lelantusEntries[i], addToDebugMessagesDB: false); // } Logging.instance .log("$spendAmount $subtractFeeFromAmount", level: LogLevel.Info); List changeToMint = List.empty(growable: true); List spendCoinIndexes = List.empty(growable: true); // Logging.instance.log(lelantusEntries, addToDebugMessagesDB: false); final fee = estimateFee( spendAmount, subtractFeeFromAmount, lelantusEntries, changeToMint, spendCoinIndexes, isTestnet: coin == Coin.firoTestNet, ); final estimateFeeData = LelantusFeeData(changeToMint[0], fee, spendCoinIndexes); Logging.instance.log( "estimateFeeData ${estimateFeeData.changeToMint} ${estimateFeeData.fee} ${estimateFeeData.spendCoinIndexes}", level: LogLevel.Info); return estimateFeeData; } Future isolateCreateJoinSplitTransaction( int spendAmount, String address, bool subtractFeeFromAmount, String mnemonic, int index, List lelantusEntries, int locktime, Coin coin, NetworkType _network, List> anonymitySetsArg, ) async { final estimateJoinSplitFee = await isolateEstimateJoinSplitFee( spendAmount, subtractFeeFromAmount, lelantusEntries, coin); var changeToMint = estimateJoinSplitFee.changeToMint; var fee = estimateJoinSplitFee.fee; var spendCoinIndexes = estimateJoinSplitFee.spendCoinIndexes; Logging.instance .log("$changeToMint $fee $spendCoinIndexes", level: LogLevel.Info); if (spendCoinIndexes.isEmpty) { Logging.instance.log("Error, Not enough funds.", level: LogLevel.Error); return 1; } final tx = TransactionBuilder(network: _network); tx.setLockTime(locktime); tx.setVersion(3 | (TRANSACTION_LELANTUS << 16)); tx.addInput( '0000000000000000000000000000000000000000000000000000000000000000', 4294967295, 4294967295, Uint8List(0), ); final jmintKeyPair = getBip32Node(MINT_INDEX, index, mnemonic, _network); final String jmintprivatekey = Format.uint8listToString(jmintKeyPair.privateKey!); final keyPath = getMintKeyPath(changeToMint, jmintprivatekey, index, isTestnet: coin == Coin.firoTestNet); final aesKeyPair = getBip32Node(JMINT_INDEX, keyPath, mnemonic, _network); final aesPrivateKey = Format.uint8listToString(aesKeyPair.privateKey!); final jmintData = createJMintScript( changeToMint, Format.uint8listToString(jmintKeyPair.privateKey!), index, Format.uint8listToString(jmintKeyPair.identifier), aesPrivateKey, isTestnet: coin == Coin.firoTestNet, ); tx.addOutput( Format.stringToUint8List(jmintData), 0, ); int amount = spendAmount; if (subtractFeeFromAmount) { amount -= fee; } tx.addOutput( address, amount, ); final extractedTx = tx.buildIncomplete(); extractedTx.setPayload(Uint8List(0)); final txHash = extractedTx.getId(); final List setIds = []; final List> anonymitySets = []; final List anonymitySetHashes = []; final List groupBlockHashes = []; for (var i = 0; i < lelantusEntries.length; i++) { final anonymitySetId = lelantusEntries[i].anonymitySetId; if (!setIds.contains(anonymitySetId)) { setIds.add(anonymitySetId); final anonymitySet = anonymitySetsArg.firstWhere( (element) => element["setId"] == anonymitySetId, orElse: () => {}); if (anonymitySet.isNotEmpty) { anonymitySetHashes.add(anonymitySet['setHash'] as String); groupBlockHashes.add(anonymitySet['blockHash'] as String); List list = []; for (int i = 0; i < (anonymitySet['coins'] as List).length; i++) { list.add(anonymitySet['coins'][i][0] as String); } anonymitySets.add(list); } } } final String spendScript = createJoinSplitScript( txHash, spendAmount, subtractFeeFromAmount, Format.uint8listToString(jmintKeyPair.privateKey!), index, lelantusEntries, setIds, anonymitySets, anonymitySetHashes, groupBlockHashes, isTestnet: coin == Coin.firoTestNet); final finalTx = TransactionBuilder(network: _network); finalTx.setLockTime(locktime); finalTx.setVersion(3 | (TRANSACTION_LELANTUS << 16)); finalTx.addOutput( Format.stringToUint8List(jmintData), 0, ); finalTx.addOutput( address, amount, ); final extTx = finalTx.buildIncomplete(); extTx.addInput( Format.stringToUint8List( '0000000000000000000000000000000000000000000000000000000000000000'), 4294967295, 4294967295, Format.stringToUint8List("c9"), ); debugPrint("spendscript: $spendScript"); extTx.setPayload(Format.stringToUint8List(spendScript)); final txHex = extTx.toHex(); final txId = extTx.getId(); Logging.instance.log("txid $txId", level: LogLevel.Info); Logging.instance.log("txHex: $txHex", level: LogLevel.Info); return { "txid": txId, "txHex": txHex, "value": amount, "fees": Format.satoshisToAmount(fee, coin: coin).toDouble(), "fee": fee, "vSize": extTx.virtualSize(), "jmintValue": changeToMint, "publicCoin": "jmintData.publicCoin", "spendCoinIndexes": spendCoinIndexes, "height": locktime, "txType": "Sent", "confirmed_status": false, "amount": Format.satoshisToAmount(amount, coin: coin).toDouble(), "recipientAmt": amount, "address": address, "timestamp": DateTime.now().millisecondsSinceEpoch ~/ 1000, "subType": "join", }; } Future getBlockHead(ElectrumX client) async { try { final tip = await client.getBlockHeadTip(); return tip["height"] as int; } catch (e) { Logging.instance .log("Exception rethrown in getBlockHead(): $e", level: LogLevel.Error); rethrow; } } // end of isolates bip32.BIP32 getBip32Node( int chain, int index, String mnemonic, NetworkType network) { final root = getBip32Root(mnemonic, network); final node = getBip32NodeFromRoot(chain, index, root); return node; } /// wrapper for compute() bip32.BIP32 getBip32NodeWrapper( Tuple4 args, ) { return getBip32Node( args.item1, args.item2, args.item3, args.item4, ); } bip32.BIP32 getBip32NodeFromRoot(int chain, int index, bip32.BIP32 root) { String coinType; switch (root.network.wif) { case 0xd2: // firo mainnet wif coinType = "136"; // firo mainnet break; case 0xb9: // firo testnet wif coinType = "1"; // firo testnet break; default: throw Exception("Invalid Bitcoin network type used!"); } final node = root.derivePath("m/44'/$coinType'/0'/$chain/$index"); return node; } /// wrapper for compute() bip32.BIP32 getBip32NodeFromRootWrapper( Tuple3 args, ) { return getBip32NodeFromRoot( args.item1, args.item2, args.item3, ); } bip32.BIP32 getBip32Root(String mnemonic, NetworkType network) { final seed = bip39.mnemonicToSeed(mnemonic); final firoNetworkType = bip32.NetworkType( wif: network.wif, bip32: bip32.Bip32Type( public: network.bip32.public, private: network.bip32.private, ), ); final root = bip32.BIP32.fromSeed(seed, firoNetworkType); return root; } /// wrapper for compute() bip32.BIP32 getBip32RootWrapper(Tuple2 args) { return getBip32Root(args.item1, args.item2); } Future _getMintScriptWrapper( Tuple5 data) async { String mintHex = getMintScript(data.item1, data.item2, data.item3, data.item4, isTestnet: data.item5); return mintHex; } Future _setTestnetWrapper(bool isTestnet) async { // setTestnet(isTestnet); } /// Handles a single instance of a firo wallet class FiroWallet extends CoinServiceAPI with WalletCache, WalletDB, FiroHive { static const integrationTestFlag = bool.fromEnvironment("IS_INTEGRATION_TEST"); final _prefs = Prefs.instance; Timer? timer; late final Coin _coin; bool _shouldAutoSync = false; @override bool get shouldAutoSync => _shouldAutoSync; @override set shouldAutoSync(bool shouldAutoSync) { if (_shouldAutoSync != shouldAutoSync) { _shouldAutoSync = shouldAutoSync; if (!shouldAutoSync) { timer?.cancel(); timer = null; stopNetworkAlivePinging(); } else { startNetworkAlivePinging(); refresh(); } } } NetworkType get _network { switch (coin) { case Coin.firo: return firoNetwork; case Coin.firoTestNet: return firoTestNetwork; default: throw Exception("Invalid network type!"); } } @override set isFavorite(bool markFavorite) { _isFavorite = markFavorite; updateCachedIsFavorite(markFavorite); } @override bool get isFavorite => _isFavorite ??= getCachedIsFavorite(); bool? _isFavorite; @override Coin get coin => _coin; @override Future> get mnemonic => _getMnemonicList(); @override bool validateAddress(String address) { return Address.validateAddress(address, _network); } /// Holds wallet transaction data Future> get _txnData => db .getTransactions(walletId) .filter() .isLelantusIsNull() .or() .isLelantusEqualTo(false) .findAll(); // _transactionData ??= _refreshTransactions(); // models.TransactionData? cachedTxData; // hack to add tx to txData before refresh completes // required based on current app architecture where we don't properly store // transactions locally in a good way @override Future updateSentCachedTxData(Map txData) async { // final currentPrice = await firoPrice; // final locale = // Platform.isWindows ? "en_US" : await Devicelocale.currentLocale; // final String worthNow = Format.localizedStringAsFixed( // value: // ((currentPrice * Decimal.fromInt(txData["recipientAmt"] as int)) / // Decimal.fromInt(Constants.satsPerCoin(coin))) // .toDecimal(scaleOnInfinitePrecision: 2), // decimalPlaces: 2, // locale: locale!); // // final tx = models.Transaction( // txid: txData["txid"] as String, // confirmedStatus: false, // timestamp: DateTime.now().millisecondsSinceEpoch ~/ 1000, // txType: "Sent", // amount: txData["recipientAmt"] as int, // worthNow: worthNow, // worthAtBlockTimestamp: worthNow, // fees: txData["fee"] as int, // inputSize: 0, // outputSize: 0, // inputs: [], // outputs: [], // address: txData["address"] as String, // height: -1, // confirmations: 0, // ); // // if (cachedTxData == null) { // final data = await _fetchTransactionData(); // _transactionData = Future(() => data); // } // // final transactions = cachedTxData!.getAllTransactions(); // transactions[tx.txid] = tx; // cachedTxData = models.TransactionData.fromMap(transactions); // _transactionData = Future(() => cachedTxData!); } /// Holds the max fee that can be sent Future? _maxFee; @override Future get maxFee => _maxFee ??= _fetchMaxFee(); Future? _feeObject; @override Future get fees => _feeObject ??= _getFees(); @override Future get currentReceivingAddress async => (await _currentReceivingAddress).value; Future get _currentReceivingAddress async => (await db .getAddresses(walletId) .filter() .typeEqualTo(isar_models.AddressType.p2pkh) .subTypeEqualTo(isar_models.AddressSubType.receiving) .sortByDerivationIndexDesc() .findFirst())!; Future get currentChangeAddress async => (await _currentChangeAddress).value; Future get _currentChangeAddress async => (await db .getAddresses(walletId) .filter() .typeEqualTo(isar_models.AddressType.p2pkh) .subTypeEqualTo(isar_models.AddressSubType.change) .sortByDerivationIndexDesc() .findFirst())!; late String _walletName; @override String get walletName => _walletName; // setter for updating on rename @override set walletName(String newName) => _walletName = newName; /// unique wallet id late final String _walletId; @override String get walletId => _walletId; @override Future testNetworkConnection() async { try { final result = await _electrumXClient.ping(); return result; } catch (_) { return false; } } Timer? _networkAliveTimer; void startNetworkAlivePinging() { // call once on start right away _periodicPingCheck(); // then periodically check _networkAliveTimer = Timer.periodic( Constants.networkAliveTimerDuration, (_) async { _periodicPingCheck(); }, ); } void _periodicPingCheck() async { bool hasNetwork = await testNetworkConnection(); _isConnected = hasNetwork; if (_isConnected != hasNetwork) { NodeConnectionStatus status = hasNetwork ? NodeConnectionStatus.connected : NodeConnectionStatus.disconnected; GlobalEventBus.instance .fire(NodeConnectionStatusChangedEvent(status, walletId, coin)); } } void stopNetworkAlivePinging() { _networkAliveTimer?.cancel(); _networkAliveTimer = null; } bool _isConnected = false; @override bool get isConnected => _isConnected; Future> prepareSendPublic({ required String address, required int satoshiAmount, Map? args, }) async { try { final feeRateType = args?["feeRate"]; final feeRateAmount = args?["feeRateAmount"]; if (feeRateType is FeeRateType || feeRateAmount is int) { late final int rate; if (feeRateType is FeeRateType) { int fee = 0; final feeObject = await fees; switch (feeRateType) { case FeeRateType.fast: fee = feeObject.fast; break; case FeeRateType.average: fee = feeObject.medium; break; case FeeRateType.slow: fee = feeObject.slow; break; } rate = fee; } else { rate = feeRateAmount as int; } // check for send all bool isSendAll = false; final balance = Format.decimalAmountToSatoshis(availablePublicBalance(), coin); if (satoshiAmount == balance) { isSendAll = true; } final txData = await coinSelection(satoshiAmount, rate, address, isSendAll); Logging.instance.log("prepare send: $txData", level: LogLevel.Info); try { if (txData is int) { switch (txData) { case 1: throw Exception("Insufficient balance!"); case 2: throw Exception( "Insufficient funds to pay for transaction fee!"); default: throw Exception("Transaction failed with error code $txData"); } } else { final hex = txData["hex"]; if (hex is String) { final fee = txData["fee"] as int; final vSize = txData["vSize"] as int; Logging.instance .log("prepared txHex: $hex", level: LogLevel.Info); Logging.instance.log("prepared fee: $fee", level: LogLevel.Info); Logging.instance .log("prepared vSize: $vSize", level: LogLevel.Info); // fee should never be less than vSize sanity check if (fee < vSize) { throw Exception( "Error in fee calculation: Transaction fee cannot be less than vSize"); } return txData as Map; } else { throw Exception("prepared hex is not a String!!!"); } } } catch (e, s) { Logging.instance.log("Exception rethrown from prepareSend(): $e\n$s", level: LogLevel.Error); rethrow; } } else { throw ArgumentError("Invalid fee rate argument provided!"); } } catch (e, s) { Logging.instance.log("Exception rethrown from prepareSend(): $e\n$s", level: LogLevel.Error); rethrow; } } Future confirmSendPublic({dynamic txData}) async { try { Logging.instance.log("confirmSend txData: $txData", level: LogLevel.Info); final txHash = await _electrumXClient.broadcastTransaction( rawTx: txData["hex"] as String); Logging.instance.log("Sent txHash: $txHash", level: LogLevel.Info); txData["txid"] = txHash; // dirty ui update hack await updateSentCachedTxData(txData as Map); return txHash; } catch (e, s) { Logging.instance.log("Exception rethrown from confirmSend(): $e\n$s", level: LogLevel.Error); rethrow; } } @override Future> prepareSend({ required String address, required int satoshiAmount, Map? args, }) async { try { // check for send all bool isSendAll = false; final balance = Format.decimalAmountToSatoshis(availablePrivateBalance(), coin); if (satoshiAmount == balance) { // print("is send all"); isSendAll = true; } dynamic txHexOrError = await _createJoinSplitTransaction(satoshiAmount, address, isSendAll); Logging.instance.log("txHexOrError $txHexOrError", level: LogLevel.Error); if (txHexOrError is int) { // Here, we assume that transaction crafting returned an error switch (txHexOrError) { case 1: throw Exception("Insufficient balance!"); default: throw Exception("Error Creating Transaction!"); } } else { final fee = txHexOrError["fee"] as int; final vSize = txHexOrError["vSize"] as int; Logging.instance.log("prepared fee: $fee", level: LogLevel.Info); Logging.instance.log("prepared vSize: $vSize", level: LogLevel.Info); // fee should never be less than vSize sanity check if (fee < vSize) { throw Exception( "Error in fee calculation: Transaction fee cannot be less than vSize"); } return txHexOrError as Map; } } catch (e, s) { Logging.instance.log("Exception rethrown in firo prepareSend(): $e\n$s", level: LogLevel.Error); rethrow; } } @override Future confirmSend({required Map txData}) async { if (await _submitLelantusToNetwork(txData)) { try { final txid = txData["txid"] as String; // temporarily update apdate available balance until a full refresh is done // TODO: something here causes an exception to be thrown giving user false info that the tx failed // Decimal sendTotal = // Format.satoshisToAmount(txData["value"] as int, coin: coin); // sendTotal += Decimal.parse(txData["fees"].toString()); // TODO: is this needed? // final bals = await balances; // bals[0] -= sendTotal; // _balances = Future(() => bals); return txid; } catch (e, s) { //todo: come back to this debugPrint("$e $s"); return txData["txid"] as String; // don't throw anything here or it will tell the user that th tx // failed even though it was successfully broadcast to network // throw Exception("Transaction failed."); } } else { //TODO provide more info throw Exception("Transaction failed."); } } // /// returns txid on successful send // /// // /// can throw // @override // Future send({ // required String toAddress, // required int amount, // Map args = const {}, // }) async { // try { // dynamic txHexOrError = // await _createJoinSplitTransaction(amount, toAddress, false); // Logging.instance.log("txHexOrError $txHexOrError", level: LogLevel.Error); // if (txHexOrError is int) { // // Here, we assume that transaction crafting returned an error // switch (txHexOrError) { // case 1: // throw Exception("Insufficient balance!"); // default: // throw Exception("Error Creating Transaction!"); // } // } else { // if (await _submitLelantusToNetwork( // txHexOrError as Map)) { // final txid = txHexOrError["txid"] as String; // // // temporarily update apdate available balance until a full refresh is done // Decimal sendTotal = // Format.satoshisToAmount(txHexOrError["value"] as int, coin: coin); // sendTotal += Decimal.parse(txHexOrError["fees"].toString()); // final bals = await balances; // bals[0] -= sendTotal; // _balances = Future(() => bals); // // return txid; // } else { // //TODO provide more info // throw Exception("Transaction failed."); // } // } // } catch (e, s) { // Logging.instance.log("Exception rethrown in firo send(): $e\n$s", // level: LogLevel.Error); // rethrow; // } // } Future> _getMnemonicList() async { final mnemonicString = await _secureStore.read(key: '${_walletId}_mnemonic'); if (mnemonicString == null) { return []; } final List data = mnemonicString.split(' '); return data; } late ElectrumX _electrumXClient; ElectrumX get electrumXClient => _electrumXClient; late CachedElectrumX _cachedElectrumXClient; CachedElectrumX get cachedElectrumXClient => _cachedElectrumXClient; late SecureStorageInterface _secureStore; late TransactionNotificationTracker txTracker; // Constructor FiroWallet({ required String walletId, required String walletName, required Coin coin, required ElectrumX client, required CachedElectrumX cachedClient, required TransactionNotificationTracker tracker, required SecureStorageInterface secureStore, }) { txTracker = tracker; _walletId = walletId; _walletName = walletName; _coin = coin; _electrumXClient = client; _cachedElectrumXClient = cachedClient; _secureStore = secureStore; initCache(walletId, coin); initFiroHive(walletId); Logging.instance.log("$walletName isolates length: ${isolates.length}", level: LogLevel.Info); // investigate possible issues killing shared isolates between multiple firo instances for (final isolate in isolates.values) { isolate.kill(priority: Isolate.immediate); } isolates.clear(); } int estimateTxFee({required int vSize, required int feeRatePerKB}) { return vSize * (feeRatePerKB / 1000).ceil(); } /// The coinselection algorithm decides whether or not the user is eligible to make the transaction /// with [satoshiAmountToSend] and [selectedTxFeeRate]. If so, it will call buildTrasaction() and return /// a map containing the tx hex along with other important information. If not, then it will return /// an integer (1 or 2) dynamic coinSelection( int satoshiAmountToSend, int selectedTxFeeRate, String _recipientAddress, bool isSendAll, { int additionalOutputs = 0, List? utxos, }) async { Logging.instance .log("Starting coinSelection ----------", level: LogLevel.Info); final List availableOutputs = utxos ?? await this.utxos; final currentChainHeight = await chainHeight; final List spendableOutputs = []; int spendableSatoshiValue = 0; // Build list of spendable outputs and totaling their satoshi amount for (var i = 0; i < availableOutputs.length; i++) { if (availableOutputs[i].isBlocked == false && availableOutputs[i] .isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) == true) { spendableOutputs.add(availableOutputs[i]); spendableSatoshiValue += availableOutputs[i].value; } } // sort spendable by age (oldest first) spendableOutputs.sort((a, b) => b.blockTime!.compareTo(a.blockTime!)); Logging.instance.log("spendableOutputs.length: ${spendableOutputs.length}", level: LogLevel.Info); Logging.instance .log("spendableOutputs: $spendableOutputs", level: LogLevel.Info); Logging.instance.log("spendableSatoshiValue: $spendableSatoshiValue", level: LogLevel.Info); Logging.instance .log("satoshiAmountToSend: $satoshiAmountToSend", level: LogLevel.Info); // If the amount the user is trying to send is smaller than the amount that they have spendable, // then return 1, which indicates that they have an insufficient balance. if (spendableSatoshiValue < satoshiAmountToSend) { return 1; // If the amount the user wants to send is exactly equal to the amount they can spend, then return // 2, which indicates that they are not leaving enough over to pay the transaction fee } else if (spendableSatoshiValue == satoshiAmountToSend && !isSendAll) { return 2; } // If neither of these statements pass, we assume that the user has a spendable balance greater // than the amount they're attempting to send. Note that this value still does not account for // the added transaction fee, which may require an extra input and will need to be checked for // later on. // Possible situation right here int satoshisBeingUsed = 0; int inputsBeingConsumed = 0; List utxoObjectsToUse = []; for (var i = 0; satoshisBeingUsed <= satoshiAmountToSend && i < spendableOutputs.length; i++) { utxoObjectsToUse.add(spendableOutputs[i]); satoshisBeingUsed += spendableOutputs[i].value; inputsBeingConsumed += 1; } for (int i = 0; i < additionalOutputs && inputsBeingConsumed < spendableOutputs.length; i++) { utxoObjectsToUse.add(spendableOutputs[inputsBeingConsumed]); satoshisBeingUsed += spendableOutputs[inputsBeingConsumed].value; inputsBeingConsumed += 1; } Logging.instance .log("satoshisBeingUsed: $satoshisBeingUsed", level: LogLevel.Info); Logging.instance .log("inputsBeingConsumed: $inputsBeingConsumed", level: LogLevel.Info); Logging.instance .log('utxoObjectsToUse: $utxoObjectsToUse', level: LogLevel.Info); // numberOfOutputs' length must always be equal to that of recipientsArray and recipientsAmtArray List recipientsArray = [_recipientAddress]; List recipientsAmtArray = [satoshiAmountToSend]; // gather required signing data final utxoSigningData = await fetchBuildTxData(utxoObjectsToUse); if (isSendAll) { Logging.instance .log("Attempting to send all $coin", level: LogLevel.Info); final int vSizeForOneOutput = (await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: [_recipientAddress], satoshiAmounts: [satoshisBeingUsed - 1], ))["vSize"] as int; int feeForOneOutput = estimateTxFee( vSize: vSizeForOneOutput, feeRatePerKB: selectedTxFeeRate, ); if (feeForOneOutput < vSizeForOneOutput + 1) { feeForOneOutput = vSizeForOneOutput + 1; } final int amount = satoshiAmountToSend - feeForOneOutput; dynamic txn = await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: recipientsArray, satoshiAmounts: [amount], ); Map transactionObject = { "hex": txn["hex"], "recipient": recipientsArray[0], "recipientAmt": amount, "fee": feeForOneOutput, "vSize": txn["vSize"], }; return transactionObject; } final int vSizeForOneOutput = (await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: [_recipientAddress], satoshiAmounts: [satoshisBeingUsed - 1], ))["vSize"] as int; final int vSizeForTwoOutPuts = (await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: [ _recipientAddress, await _getCurrentAddressForChain(1), ], satoshiAmounts: [ satoshiAmountToSend, satoshisBeingUsed - satoshiAmountToSend - 1, ], // dust limit is the minimum amount a change output should be ))["vSize"] as int; //todo: check if print needed debugPrint("vSizeForOneOutput $vSizeForOneOutput"); debugPrint("vSizeForTwoOutPuts $vSizeForTwoOutPuts"); // Assume 1 output, only for recipient and no change var feeForOneOutput = estimateTxFee( vSize: vSizeForOneOutput, feeRatePerKB: selectedTxFeeRate, ); // Assume 2 outputs, one for recipient and one for change var feeForTwoOutputs = estimateTxFee( vSize: vSizeForTwoOutPuts, feeRatePerKB: selectedTxFeeRate, ); Logging.instance .log("feeForTwoOutputs: $feeForTwoOutputs", level: LogLevel.Info); Logging.instance .log("feeForOneOutput: $feeForOneOutput", level: LogLevel.Info); if (feeForOneOutput < (vSizeForOneOutput + 1)) { feeForOneOutput = (vSizeForOneOutput + 1); } if (feeForTwoOutputs < ((vSizeForTwoOutPuts + 1))) { feeForTwoOutputs = ((vSizeForTwoOutPuts + 1)); } Logging.instance .log("feeForTwoOutputs: $feeForTwoOutputs", level: LogLevel.Info); Logging.instance .log("feeForOneOutput: $feeForOneOutput", level: LogLevel.Info); if (satoshisBeingUsed - satoshiAmountToSend > feeForOneOutput) { if (satoshisBeingUsed - satoshiAmountToSend > feeForOneOutput + DUST_LIMIT) { // Here, we know that theoretically, we may be able to include another output(change) but we first need to // factor in the value of this output in satoshis. int changeOutputSize = satoshisBeingUsed - satoshiAmountToSend - feeForTwoOutputs; // We check to see if the user can pay for the new transaction with 2 outputs instead of one. If they can and // the second output's size > DUST_LIMIT satoshis, we perform the mechanics required to properly generate and use a new // change address. if (changeOutputSize > DUST_LIMIT && satoshisBeingUsed - satoshiAmountToSend - changeOutputSize == feeForTwoOutputs) { // generate new change address if current change address has been used await checkChangeAddressForTransactions(); final String newChangeAddress = await _getCurrentAddressForChain(1); int feeBeingPaid = satoshisBeingUsed - satoshiAmountToSend - changeOutputSize; recipientsArray.add(newChangeAddress); recipientsAmtArray.add(changeOutputSize); // At this point, we have the outputs we're going to use, the amounts to send along with which addresses // we intend to send these amounts to. We have enough to send instructions to build the transaction. Logging.instance.log('2 outputs in tx', level: LogLevel.Info); Logging.instance .log('Input size: $satoshisBeingUsed', level: LogLevel.Info); Logging.instance.log('Recipient output size: $satoshiAmountToSend', level: LogLevel.Info); Logging.instance.log('Change Output Size: $changeOutputSize', level: LogLevel.Info); Logging.instance.log( 'Difference (fee being paid): $feeBeingPaid sats', level: LogLevel.Info); Logging.instance .log('Estimated fee: $feeForTwoOutputs', level: LogLevel.Info); dynamic txn = await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: recipientsArray, satoshiAmounts: recipientsAmtArray, ); // make sure minimum fee is accurate if that is being used if (txn["vSize"] - feeBeingPaid == 1) { int changeOutputSize = satoshisBeingUsed - satoshiAmountToSend - (txn["vSize"] as int); feeBeingPaid = satoshisBeingUsed - satoshiAmountToSend - changeOutputSize; recipientsAmtArray.removeLast(); recipientsAmtArray.add(changeOutputSize); Logging.instance.log('Adjusted Input size: $satoshisBeingUsed', level: LogLevel.Info); Logging.instance.log( 'Adjusted Recipient output size: $satoshiAmountToSend', level: LogLevel.Info); Logging.instance.log( 'Adjusted Change Output Size: $changeOutputSize', level: LogLevel.Info); Logging.instance.log( 'Adjusted Difference (fee being paid): $feeBeingPaid sats', level: LogLevel.Info); Logging.instance.log('Adjusted Estimated fee: $feeForTwoOutputs', level: LogLevel.Info); txn = await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: recipientsArray, satoshiAmounts: recipientsAmtArray, ); } Map transactionObject = { "hex": txn["hex"], "recipient": recipientsArray[0], "recipientAmt": recipientsAmtArray[0], "fee": feeBeingPaid, "vSize": txn["vSize"], }; return transactionObject; } else { // Something went wrong here. It either overshot or undershot the estimated fee amount or the changeOutputSize // is smaller than or equal to [DUST_LIMIT]. Revert to single output transaction. Logging.instance.log('1 output in tx', level: LogLevel.Info); Logging.instance .log('Input size: $satoshisBeingUsed', level: LogLevel.Info); Logging.instance.log('Recipient output size: $satoshiAmountToSend', level: LogLevel.Info); Logging.instance.log( 'Difference (fee being paid): ${satoshisBeingUsed - satoshiAmountToSend} sats', level: LogLevel.Info); Logging.instance .log('Estimated fee: $feeForOneOutput', level: LogLevel.Info); dynamic txn = await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: recipientsArray, satoshiAmounts: recipientsAmtArray, ); Map transactionObject = { "hex": txn["hex"], "recipient": recipientsArray[0], "recipientAmt": recipientsAmtArray[0], "fee": satoshisBeingUsed - satoshiAmountToSend, "vSize": txn["vSize"], }; return transactionObject; } } else { // No additional outputs needed since adding one would mean that it'd be smaller than 546 sats // which makes it uneconomical to add to the transaction. Here, we pass data directly to instruct // the wallet to begin crafting the transaction that the user requested. Logging.instance.log('1 output in tx', level: LogLevel.Info); Logging.instance .log('Input size: $satoshisBeingUsed', level: LogLevel.Info); Logging.instance.log('Recipient output size: $satoshiAmountToSend', level: LogLevel.Info); Logging.instance.log( 'Difference (fee being paid): ${satoshisBeingUsed - satoshiAmountToSend} sats', level: LogLevel.Info); Logging.instance .log('Estimated fee: $feeForOneOutput', level: LogLevel.Info); dynamic txn = await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: recipientsArray, satoshiAmounts: recipientsAmtArray, ); Map transactionObject = { "hex": txn["hex"], "recipient": recipientsArray[0], "recipientAmt": recipientsAmtArray[0], "fee": satoshisBeingUsed - satoshiAmountToSend, "vSize": txn["vSize"], }; return transactionObject; } } else if (satoshisBeingUsed - satoshiAmountToSend == feeForOneOutput) { // In this scenario, no additional change output is needed since inputs - outputs equal exactly // what we need to pay for fees. Here, we pass data directly to instruct the wallet to begin // crafting the transaction that the user requested. Logging.instance.log('1 output in tx', level: LogLevel.Info); Logging.instance .log('Input size: $satoshisBeingUsed', level: LogLevel.Info); Logging.instance.log('Recipient output size: $satoshiAmountToSend', level: LogLevel.Info); Logging.instance.log( 'Fee being paid: ${satoshisBeingUsed - satoshiAmountToSend} sats', level: LogLevel.Info); Logging.instance .log('Estimated fee: $feeForOneOutput', level: LogLevel.Info); dynamic txn = await buildTransaction( utxosToUse: utxoObjectsToUse, utxoSigningData: utxoSigningData, recipients: recipientsArray, satoshiAmounts: recipientsAmtArray, ); Map transactionObject = { "hex": txn["hex"], "recipient": recipientsArray[0], "recipientAmt": recipientsAmtArray[0], "fee": feeForOneOutput, "vSize": txn["vSize"], }; return transactionObject; } else { // Remember that returning 2 indicates that the user does not have a sufficient balance to // pay for the transaction fee. Ideally, at this stage, we should check if the user has any // additional outputs they're able to spend and then recalculate fees. Logging.instance.log( 'Cannot pay tx fee - checking for more outputs and trying again', level: LogLevel.Warning); // try adding more outputs if (spendableOutputs.length > inputsBeingConsumed) { return coinSelection(satoshiAmountToSend, selectedTxFeeRate, _recipientAddress, isSendAll, additionalOutputs: additionalOutputs + 1, utxos: utxos); } return 2; } } Future> fetchBuildTxData( List utxosToUse, ) async { // return data Map results = {}; Map> addressTxid = {}; // addresses to check List addresses = []; try { // Populating the addresses to check for (var i = 0; i < utxosToUse.length; i++) { final txid = utxosToUse[i].txid; final tx = await _cachedElectrumXClient.getTransaction( txHash: txid, coin: coin, ); for (final output in tx["vout"] as List) { final n = output["n"]; if (n != null && n == utxosToUse[i].vout) { final address = output["scriptPubKey"]["addresses"][0] as String; if (!addressTxid.containsKey(address)) { addressTxid[address] = []; } (addressTxid[address] as List).add(txid); addresses.add(address); } } } // p2pkh / bip44 final addressesLength = addresses.length; if (addressesLength > 0) { final receiveDerivationsString = await _secureStore.read(key: "${walletId}_receiveDerivations"); final receiveDerivations = Map.from( jsonDecode(receiveDerivationsString ?? "{}") as Map); final changeDerivationsString = await _secureStore.read(key: "${walletId}_changeDerivations"); final changeDerivations = Map.from( jsonDecode(changeDerivationsString ?? "{}") as Map); for (int i = 0; i < addressesLength; i++) { // receives dynamic receiveDerivation; for (int j = 0; j < receiveDerivations.length; j++) { if (receiveDerivations["$j"]["address"] == addresses[i]) { receiveDerivation = receiveDerivations["$j"]; } } // receiveDerivation = receiveDerivations[addresses[i]]; // if a match exists it will not be null if (receiveDerivation != null) { final data = P2PKH( data: PaymentData( pubkey: Format.stringToUint8List( receiveDerivation["publicKey"] as String)), network: _network, ).data; for (String tx in addressTxid[addresses[i]]!) { results[tx] = { "output": data.output, "keyPair": ECPair.fromWIF( receiveDerivation["wif"] as String, network: _network, ), }; } } else { // if its not a receive, check change dynamic changeDerivation; for (int j = 0; j < changeDerivations.length; j++) { if (changeDerivations["$j"]["address"] == addresses[i]) { changeDerivation = changeDerivations["$j"]; } } // final changeDerivation = changeDerivations[addresses[i]]; // if a match exists it will not be null if (changeDerivation != null) { final data = P2PKH( data: PaymentData( pubkey: Format.stringToUint8List( changeDerivation["publicKey"] as String)), network: _network, ).data; for (String tx in addressTxid[addresses[i]]!) { results[tx] = { "output": data.output, "keyPair": ECPair.fromWIF( changeDerivation["wif"] as String, network: _network, ), }; } } } } } return results; } catch (e, s) { Logging.instance .log("fetchBuildTxData() threw: $e,\n$s", level: LogLevel.Error); rethrow; } } /// Builds and signs a transaction Future> buildTransaction({ required List utxosToUse, required Map utxoSigningData, required List recipients, required List satoshiAmounts, }) async { Logging.instance .log("Starting buildTransaction ----------", level: LogLevel.Info); final txb = TransactionBuilder(network: _network); txb.setVersion(1); // Add transaction inputs for (var i = 0; i < utxosToUse.length; i++) { final txid = utxosToUse[i].txid; txb.addInput(txid, utxosToUse[i].vout, null, utxoSigningData[txid]["output"] as Uint8List); } // Add transaction output for (var i = 0; i < recipients.length; i++) { txb.addOutput(recipients[i], satoshiAmounts[i]); } try { // Sign the transaction accordingly for (var i = 0; i < utxosToUse.length; i++) { final txid = utxosToUse[i].txid; txb.sign( vin: i, keyPair: utxoSigningData[txid]["keyPair"] as ECPair, witnessValue: utxosToUse[i].value, redeemScript: utxoSigningData[txid]["redeemScript"] as Uint8List?, ); } } catch (e, s) { Logging.instance.log("Caught exception while signing transaction: $e\n$s", level: LogLevel.Error); rethrow; } final builtTx = txb.build(); final vSize = builtTx.virtualSize(); return {"hex": builtTx.toHex(), "vSize": vSize}; } @override Future updateNode(bool shouldRefresh) async { final failovers = NodeService(secureStorageInterface: _secureStore) .failoverNodesFor(coin: coin) .map( (e) => ElectrumXNode( address: e.host, port: e.port, name: e.name, id: e.id, useSSL: e.useSSL, ), ) .toList(); final newNode = await _getCurrentNode(); _cachedElectrumXClient = CachedElectrumX.from( node: newNode, prefs: _prefs, failovers: failovers, ); _electrumXClient = ElectrumX.from( node: newNode, prefs: _prefs, failovers: failovers, ); if (shouldRefresh) { unawaited(refresh()); } } @override Future initializeNew() async { Logging.instance .log("Generating new ${coin.prettyName} wallet.", level: LogLevel.Info); if (getCachedId() != null) { throw Exception( "Attempted to initialize a new wallet using an existing wallet ID!"); } await _prefs.init(); try { await _generateNewWallet(); } catch (e, s) { Logging.instance.log("Exception rethrown from initializeNew(): $e\n$s", level: LogLevel.Fatal); rethrow; } await Future.wait([ updateCachedId(walletId), updateCachedIsFavorite(false), ]); } @override Future initializeExisting() async { Logging.instance.log( "Opening existing $_walletId ${coin.prettyName} wallet.", level: LogLevel.Info); if (getCachedId() == null) { throw Exception( "Attempted to initialize an existing wallet using an unknown wallet ID!"); } await _prefs.init(); await isarInit(walletId); } Future refreshIfThereIsNewData() async { if (longMutex) return false; if (_hasCalledExit) return false; Logging.instance .log("$walletName refreshIfThereIsNewData", level: LogLevel.Info); try { bool needsRefresh = false; Set txnsToCheck = {}; for (final String txid in txTracker.pendings) { if (!txTracker.wasNotifiedConfirmed(txid)) { txnsToCheck.add(txid); } } for (String txid in txnsToCheck) { final txn = await electrumXClient.getTransaction(txHash: txid); int confirmations = txn["confirmations"] as int? ?? 0; bool isUnconfirmed = confirmations < MINIMUM_CONFIRMATIONS; if (!isUnconfirmed) { needsRefresh = true; break; } } if (!needsRefresh) { final allOwnAddresses = await _fetchAllOwnAddresses(); List> allTxs = await _fetchHistory( allOwnAddresses.map((e) => e.value).toList(growable: false)); for (Map transaction in allTxs) { final txid = transaction['tx_hash'] as String; if ((await db .getTransactions(walletId) .filter() .txidMatches(txid) .findFirst()) == null) { Logging.instance.log( " txid not found in address history already ${transaction['tx_hash']}", level: LogLevel.Info); needsRefresh = true; break; } } } return needsRefresh; } catch (e, s) { Logging.instance.log( "Exception caught in refreshIfThereIsNewData: $e\n$s", level: LogLevel.Error); rethrow; } } Future getAllTxsToWatch() async { if (_hasCalledExit) return; Logging.instance.log("$walletName periodic", level: LogLevel.Info); List unconfirmedTxnsToNotifyPending = []; List unconfirmedTxnsToNotifyConfirmed = []; final currentChainHeight = await chainHeight; final txTxns = await db .getTransactions(walletId) .filter() .isLelantusIsNull() .or() .isLelantusEqualTo(false) .findAll(); final ltxTxns = await db .getTransactions(walletId) .filter() .isLelantusEqualTo(true) .findAll(); for (isar_models.Transaction tx in txTxns) { isar_models.Transaction? lTx; try { lTx = ltxTxns.firstWhere((e) => e.txid == tx.txid); } catch (_) { lTx = null; } if (tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) { if (txTracker.wasNotifiedPending(tx.txid) && !txTracker.wasNotifiedConfirmed(tx.txid)) { // get all transactions that were notified as pending but not as confirmed unconfirmedTxnsToNotifyConfirmed.add(tx); } if (lTx != null && (lTx.inputs.isEmpty || lTx.inputs.first.txid.isEmpty) && lTx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) == false && tx.type == isar_models.TransactionType.incoming) { // If this is a received that is past 1 or more confirmations and has not been minted, if (!txTracker.wasNotifiedPending(tx.txid)) { unconfirmedTxnsToNotifyPending.add(tx); } } } else { if (!txTracker.wasNotifiedPending(tx.txid)) { // get all transactions that were not notified as pending yet unconfirmedTxnsToNotifyPending.add(tx); } } } for (isar_models.Transaction tx in txTxns) { if (!tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) && tx.inputs.first.txid.isNotEmpty) { // Get all normal txs that are at 0 confirmations unconfirmedTxnsToNotifyPending .removeWhere((e) => e.txid == tx.inputs.first.txid); Logging.instance.log("removed tx: ${tx.txid}", level: LogLevel.Info); } } for (isar_models.Transaction lTX in ltxTxns) { isar_models.Transaction? tx; try { tx = ltxTxns.firstWhere((e) => e.txid == lTX.txid); } catch (_) { tx = null; } if (tx == null) { // if this is a ltx transaction that is unconfirmed and not represented in the normal transaction set. if (!lTX.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) { if (!txTracker.wasNotifiedPending(lTX.txid)) { unconfirmedTxnsToNotifyPending.add(lTX); } } else { if (txTracker.wasNotifiedPending(lTX.txid) && !txTracker.wasNotifiedConfirmed(lTX.txid)) { unconfirmedTxnsToNotifyConfirmed.add(lTX); } } } } Logging.instance.log( "unconfirmedTxnsToNotifyPending $unconfirmedTxnsToNotifyPending", level: LogLevel.Info); Logging.instance.log( "unconfirmedTxnsToNotifyConfirmed $unconfirmedTxnsToNotifyConfirmed", level: LogLevel.Info); for (final tx in unconfirmedTxnsToNotifyPending) { final confirmations = tx.getConfirmations(currentChainHeight); switch (tx.type) { case isar_models.TransactionType.incoming: unawaited( NotificationApi.showNotification( title: "Incoming transaction", body: walletName, walletId: walletId, iconAssetName: Assets.svg.iconFor(coin: coin), date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000), shouldWatchForUpdates: confirmations < MINIMUM_CONFIRMATIONS, coinName: coin.name, txid: tx.txid, confirmations: confirmations, requiredConfirmations: MINIMUM_CONFIRMATIONS, ), ); await txTracker.addNotifiedPending(tx.txid); break; case isar_models.TransactionType.outgoing: unawaited( NotificationApi.showNotification( title: tx.subType == isar_models.TransactionSubType.mint ? "Anonymizing" : "Outgoing transaction", body: walletName, walletId: walletId, iconAssetName: Assets.svg.iconFor(coin: coin), date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000), shouldWatchForUpdates: confirmations < MINIMUM_CONFIRMATIONS, coinName: coin.name, txid: tx.txid, confirmations: confirmations, requiredConfirmations: MINIMUM_CONFIRMATIONS, ), ); await txTracker.addNotifiedPending(tx.txid); break; default: break; } } for (final tx in unconfirmedTxnsToNotifyConfirmed) { if (tx.type == isar_models.TransactionType.incoming) { unawaited( NotificationApi.showNotification( title: "Incoming transaction confirmed", body: walletName, walletId: walletId, iconAssetName: Assets.svg.iconFor(coin: coin), date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000), shouldWatchForUpdates: false, coinName: coin.name, ), ); await txTracker.addNotifiedConfirmed(tx.txid); } else if (tx.type == isar_models.TransactionType.outgoing && tx.subType == isar_models.TransactionSubType.join) { unawaited( NotificationApi.showNotification( title: tx.subType == isar_models.TransactionSubType.mint // redundant check? ? "Anonymized" : "Outgoing transaction confirmed", body: walletName, walletId: walletId, iconAssetName: Assets.svg.iconFor(coin: coin), date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000), shouldWatchForUpdates: false, coinName: coin.name, ), ); await txTracker.addNotifiedConfirmed(tx.txid); } } } /// Generates initial wallet values such as mnemonic, chain (receive/change) arrays and indexes. Future _generateNewWallet() async { Logging.instance .log("IS_INTEGRATION_TEST: $integrationTestFlag", level: LogLevel.Info); if (!integrationTestFlag) { try { final features = await electrumXClient .getServerFeatures() .timeout(const Duration(seconds: 3)); Logging.instance.log("features: $features", level: LogLevel.Info); switch (coin) { case Coin.firo: if (features['genesis_hash'] != GENESIS_HASH_MAINNET) { throw Exception("genesis hash does not match main net!"); } break; case Coin.firoTestNet: if (features['genesis_hash'] != GENESIS_HASH_TESTNET) { throw Exception("genesis hash does not match test net!"); } break; default: throw Exception( "Attempted to generate a FiroWallet using a non firo coin type: ${coin.name}"); } } catch (e, s) { Logging.instance.log("$e/n$s", level: LogLevel.Info); } } // this should never fail as overwriting a mnemonic is big bad assert((await _secureStore.read(key: '${_walletId}_mnemonic')) == null); await _secureStore.write( key: '${_walletId}_mnemonic', value: bip39.generateMnemonic(strength: 256)); await firoUpdateJIndex([]); // Generate and add addresses to relevant arrays final initialReceivingAddress = await _generateAddressForChain(0, 0); final initialChangeAddress = await _generateAddressForChain(1, 0); await isarInit(walletId); await db.putAddresses([ initialReceivingAddress, initialChangeAddress, ]); } bool refreshMutex = false; @override bool get isRefreshing => refreshMutex; /// Refreshes display data for the wallet @override Future refresh() async { if (refreshMutex) { Logging.instance.log("$walletId $walletName refreshMutex denied", level: LogLevel.Info); return; } else { refreshMutex = true; } Logging.instance .log("PROCESSORS ${Platform.numberOfProcessors}", level: LogLevel.Info); try { GlobalEventBus.instance.fire( WalletSyncStatusChangedEvent( WalletSyncStatus.syncing, walletId, coin, ), ); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.0, walletId)); final receiveDerivationsString = await _secureStore.read(key: "${walletId}_receiveDerivations"); if (receiveDerivationsString == null || receiveDerivationsString == "{}") { GlobalEventBus.instance .fire(RefreshPercentChangedEvent(0.05, walletId)); final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic'); await fillAddresses(mnemonic!, numberOfThreads: Platform.numberOfProcessors - isolates.length - 1); } await checkReceivingAddressForTransactions(); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.1, walletId)); await _refreshUTXOs(); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.2, walletId)); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.25, walletId)); await _refreshTransactions(); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.35, walletId)); final feeObj = _getFees(); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.50, walletId)); _feeObject = Future(() => feeObj); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.60, walletId)); final lelantusCoins = getLelantusCoinMap(); Logging.instance.log("_lelantus_coins at refresh: $lelantusCoins", level: LogLevel.Warning, printFullLength: true); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.70, walletId)); await _refreshLelantusData(); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.80, walletId)); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.90, walletId)); await _refreshBalance(); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.95, walletId)); await getAllTxsToWatch(); GlobalEventBus.instance.fire(RefreshPercentChangedEvent(1.0, walletId)); GlobalEventBus.instance.fire( WalletSyncStatusChangedEvent( WalletSyncStatus.synced, walletId, coin, ), ); refreshMutex = false; if (isActive || shouldAutoSync) { timer ??= Timer.periodic(const Duration(seconds: 30), (timer) async { bool shouldNotify = await refreshIfThereIsNewData(); if (shouldNotify) { await refresh(); GlobalEventBus.instance.fire(UpdatedInBackgroundEvent( "New data found in $walletId $walletName in background!", walletId)); } }); } } catch (error, strace) { refreshMutex = false; GlobalEventBus.instance.fire( NodeConnectionStatusChangedEvent( NodeConnectionStatus.disconnected, walletId, coin, ), ); GlobalEventBus.instance.fire( WalletSyncStatusChangedEvent( WalletSyncStatus.unableToSync, walletId, coin, ), ); Logging.instance.log( "Caught exception in refreshWalletData(): $error\n$strace", level: LogLevel.Warning); } } Future _fetchMaxFee() async { final balance = availablePrivateBalance(); int spendAmount = (balance * Decimal.fromInt(Constants.satsPerCoin(coin))) .toBigInt() .toInt(); int fee = await estimateJoinSplitFee(spendAmount); return fee; } Future> _getLelantusEntry() async { final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic'); final List lelantusCoins = await _getUnspentCoins(); final root = await compute( getBip32RootWrapper, Tuple2( mnemonic!, _network, ), ); final waitLelantusEntries = lelantusCoins.map((coin) async { final keyPair = await compute( getBip32NodeFromRootWrapper, Tuple3( MINT_INDEX, coin.index, root, ), ); if (keyPair.privateKey == null) { Logging.instance.log("error bad key", level: LogLevel.Error); return DartLelantusEntry(1, 0, 0, 0, 0, ''); } final String privateKey = Format.uint8listToString(keyPair.privateKey!); return DartLelantusEntry(coin.isUsed ? 1 : 0, 0, coin.anonymitySetId, coin.value, coin.index, privateKey); }).toList(); final lelantusEntries = await Future.wait(waitLelantusEntries); if (lelantusEntries.isNotEmpty) { lelantusEntries.removeWhere((element) => element.amount == 0); } return lelantusEntries; } List> getLelantusCoinMap() { final _l = firoGetLelantusCoins(); final List> lelantusCoins = []; for (var el in _l ?? []) { lelantusCoins.add({el.keys.first: el.values.first as LelantusCoin}); } return lelantusCoins; } Future> _getUnspentCoins() async { final List> lelantusCoins = getLelantusCoinMap(); if (lelantusCoins.isNotEmpty) { lelantusCoins.removeWhere((element) => element.values.any((elementCoin) => elementCoin.value == 0)); } final jindexes = firoGetJIndex(); final transactions = await _txnData; final lelantusTransactionsd = await db .getTransactions(walletId) .filter() .isLelantusEqualTo(true) .findAll(); List coins = []; List lelantusCoinsList = lelantusCoins.fold([], (previousValue, element) { previousValue.add(element.values.first); return previousValue; }); final currentChainHeight = await chainHeight; for (int i = 0; i < lelantusCoinsList.length; i++) { // Logging.instance.log("lelantusCoinsList[$i]: ${lelantusCoinsList[i]}"); final txn = await cachedElectrumXClient.getTransaction( txHash: lelantusCoinsList[i].txId, verbose: true, coin: coin, ); final confirmations = txn["confirmations"]; bool isUnconfirmed = confirmations is int && confirmations < 1; if (!jindexes!.contains(lelantusCoinsList[i].index) && transactions .where((e) => e.txid == lelantusCoinsList[i].txId) .isEmpty) { isUnconfirmed = true; } // TODO: optimize the following if ((transactions .where((e) => e.txid == lelantusCoinsList[i].txId) .isNotEmpty && !transactions .where((e) => e.txid == lelantusCoinsList[i].txId) .first .isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) || (lelantusTransactionsd .where((e) => e.txid == lelantusCoinsList[i].txId) .isNotEmpty && !lelantusTransactionsd .where((e) => e.txid == lelantusCoinsList[i].txId) .first .isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS))) { continue; } if (!lelantusCoinsList[i].isUsed && lelantusCoinsList[i].anonymitySetId != ANONYMITY_SET_EMPTY_ID && !isUnconfirmed) { coins.add(lelantusCoinsList[i]); } } return coins; } // index 0 and 1 for the funds available to spend. // index 2 and 3 for all the funds in the wallet (including the undependable ones) // Future> _refreshBalance() async { Future _refreshBalance() async { try { final utxosUpdateFuture = _refreshUTXOs(); final List> lelantusCoins = getLelantusCoinMap(); if (lelantusCoins.isNotEmpty) { lelantusCoins.removeWhere((element) => element.values.any((elementCoin) => elementCoin.value == 0)); } final data = await _txnData; final lData = await db .getTransactions(walletId) .filter() .isLelantusEqualTo(true) .findAll(); final currentChainHeight = await chainHeight; final jindexes = firoGetJIndex(); int intLelantusBalance = 0; int unconfirmedLelantusBalance = 0; for (var element in lelantusCoins) { element.forEach((key, value) { isar_models.Transaction? tx; try { tx == data.firstWhere((e) => e.txid == value.txId); } catch (_) { tx = null; } isar_models.Transaction? ltx; try { ltx = lData.firstWhere((e) => e.txid == value.txId); } catch (_) { ltx = null; } // Logging.instance.log("$value $tx $ltx"); if (!jindexes!.contains(value.index) && tx == null) { // This coin is not confirmed and may be replaced } else if (jindexes.contains(value.index) && tx == null && !value.isUsed && ltx != null && !ltx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) { unconfirmedLelantusBalance += value.value; } else if (jindexes.contains(value.index) && !value.isUsed) { intLelantusBalance += value.value; } else if (!value.isUsed && (tx == null ? true : tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) != false)) { intLelantusBalance += value.value; } else if (tx != null && tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) == false) { unconfirmedLelantusBalance += value.value; } }); } // final int utxosIntValue = utxos.satoshiBalance; // final Decimal utxosValue = // Format.satoshisToAmount(utxosIntValue, coin: coin); // List balances = List.empty(growable: true); // // Decimal lelantusBalance = // Format.satoshisToAmount(intLelantusBalance, coin: coin); // balances.add(lelantusBalance); 0 // // balances.add(lelantusBalance * price); 1 // Decimal _unconfirmedLelantusBalance = // Format.satoshisToAmount(unconfirmedLelantusBalance, coin: coin); // balances.add(lelantusBalance + utxosValue + _unconfirmedLelantusBalance); 2 // // balances.add( // (lelantusBalance + utxosValue + _unconfirmedLelantusBalance) * price); 3 // int availableSats = // utxos.satoshiBalance - utxos.satoshiBalanceUnconfirmed; // if (availableSats < 0) { // availableSats = 0; // } // balances.add(Format.satoshisToAmount(availableSats, coin: coin)); 4 // wait for updated uxtos to get updated public balance await utxosUpdateFuture; // todo: shared total between private and public balances? _balancePrivate = Balance( coin: coin, total: intLelantusBalance + unconfirmedLelantusBalance + balance.total, spendable: intLelantusBalance, blockedTotal: 0, pendingSpendable: unconfirmedLelantusBalance + balance.total, ); await updateCachedBalanceSecondary(_balancePrivate!); // _balance = Balance( // coin: coin, // total: utxos.satoshiBalance, // spendable: availableSats, // blockedTotal: 0, // pendingSpendable: utxos.satoshiBalanceUnconfirmed, // ); // Logging.instance.log("balances $balances", level: LogLevel.Info); // await DB.instance.put( // boxName: walletId, // key: 'totalBalance', // value: balances[2].toString()); // return balances; } catch (e, s) { Logging.instance.log("Exception rethrown in getFullBalance(): $e\n$s", level: LogLevel.Error); rethrow; } } Future anonymizeAllPublicFunds() async { try { var mintResult = await _mintSelection(); if (mintResult.isEmpty) { Logging.instance.log("nothing to mint", level: LogLevel.Info); return; } await _submitLelantusToNetwork(mintResult); unawaited(refresh()); } catch (e, s) { Logging.instance.log( "Exception caught in anonymizeAllPublicFunds(): $e\n$s", level: LogLevel.Warning); rethrow; } } /// Returns the mint transaction hex to mint all of the available funds. Future> _mintSelection() async { final currentChainHeight = await chainHeight; final List availableOutputs = await utxos; final List spendableOutputs = []; // Build list of spendable outputs and totaling their satoshi amount for (var i = 0; i < availableOutputs.length; i++) { if (availableOutputs[i].isBlocked == false && availableOutputs[i] .isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) == true && !(availableOutputs[i].isCoinbase && availableOutputs[i].getConfirmations(currentChainHeight) <= 101)) { spendableOutputs.add(availableOutputs[i]); } } final List> lelantusCoins = getLelantusCoinMap(); if (lelantusCoins.isNotEmpty) { lelantusCoins.removeWhere((element) => element.values.any((elementCoin) => elementCoin.value == 0)); } final data = await _txnData; for (final value in data) { if (value.inputs.isNotEmpty) { for (var element in value.inputs) { if (lelantusCoins .any((element) => element.keys.contains(value.txid)) && spendableOutputs.firstWhere( (output) => output?.txid == element.txid, orElse: () => null) != null) { spendableOutputs .removeWhere((output) => output!.txid == element.txid); } } } } // If there is no Utxos to mint then stop the function. if (spendableOutputs.isEmpty) { Logging.instance.log("_mintSelection(): No spendable outputs found", level: LogLevel.Info); return {}; } int satoshisBeingUsed = 0; List utxoObjectsToUse = []; for (var i = 0; i < spendableOutputs.length; i++) { final spendable = spendableOutputs[i]; if (spendable != null) { utxoObjectsToUse.add(spendable); satoshisBeingUsed += spendable.value; } } var mintsWithoutFee = await createMintsFromAmount(satoshisBeingUsed); var tmpTx = await buildMintTransaction( utxoObjectsToUse, satoshisBeingUsed, mintsWithoutFee); int vSize = (tmpTx['transaction'] as Transaction).virtualSize(); final Decimal dvSize = Decimal.fromInt(vSize); final feesObject = await fees; final Decimal fastFee = Format.satoshisToAmount(feesObject.fast, coin: coin); int firoFee = (dvSize * fastFee * Decimal.fromInt(100000)).toDouble().ceil(); // int firoFee = (vSize * feesObject.fast * (1 / 1000.0) * 100000000).ceil(); if (firoFee < vSize) { firoFee = vSize + 1; } firoFee = firoFee + 10; int satoshiAmountToSend = satoshisBeingUsed - firoFee; var mintsWithFee = await createMintsFromAmount(satoshiAmountToSend); Map transaction = await buildMintTransaction( utxoObjectsToUse, satoshiAmountToSend, mintsWithFee); transaction['transaction'] = ""; Logging.instance.log(transaction.toString(), level: LogLevel.Info); Logging.instance.log(transaction['txHex'], level: LogLevel.Info); return transaction; } Future>> createMintsFromAmount(int total) async { var tmpTotal = total; var index = 0; var mints = >[]; final nextFreeMintIndex = firoGetMintIndex()!; while (tmpTotal > 0) { final mintValue = min(tmpTotal, MINT_LIMIT); final mint = await _getMintHex( mintValue, nextFreeMintIndex + index, ); mints.add({ "value": mintValue, "script": mint, "index": nextFreeMintIndex + index, "publicCoin": "", }); tmpTotal = tmpTotal - MINT_LIMIT; index++; } return mints; } /// returns a valid txid if successful Future submitHexToNetwork(String hex) async { try { final txid = await electrumXClient.broadcastTransaction(rawTx: hex); return txid; } catch (e, s) { Logging.instance.log( "Caught exception in submitHexToNetwork(\"$hex\"): $e $s", printFullLength: true, level: LogLevel.Info); // return an invalid tx return "transaction submission failed"; } } /// Builds and signs a transaction Future> buildMintTransaction( List utxosToUse, int satoshisPerRecipient, List> mintsMap) async { //todo: check if print needed // debugPrint(utxosToUse.toString()); List addressesToDerive = []; // Populating the addresses to derive for (var i = 0; i < utxosToUse.length; i++) { final txid = utxosToUse[i].txid; final outputIndex = utxosToUse[i].vout; // txid may not work for this as txid may not always be the same as tx_hash? final tx = await cachedElectrumXClient.getTransaction( txHash: txid, verbose: true, coin: coin, ); final vouts = tx["vout"] as List?; if (vouts != null && outputIndex < vouts.length) { final address = vouts[outputIndex]["scriptPubKey"]["addresses"][0] as String?; if (address != null) { addressesToDerive.add(address); } } } List elipticCurvePairArray = []; List outputDataArray = []; final receiveDerivationsString = await _secureStore.read(key: "${walletId}_receiveDerivations"); final changeDerivationsString = await _secureStore.read(key: "${walletId}_changeDerivations"); final receiveDerivations = Map.from( jsonDecode(receiveDerivationsString ?? "{}") as Map); final changeDerivations = Map.from( jsonDecode(changeDerivationsString ?? "{}") as Map); for (var i = 0; i < addressesToDerive.length; i++) { final addressToCheckFor = addressesToDerive[i]; for (var i = 0; i < receiveDerivations.length; i++) { final receive = receiveDerivations["$i"]; final change = changeDerivations["$i"]; if (receive['address'] == addressToCheckFor) { Logging.instance .log('Receiving found on loop $i', level: LogLevel.Info); // Logging.instance.log( // 'decoded receive[\'wif\'] version: ${wif.decode(receive['wif'] as String)}, _network: $_network'); elipticCurvePairArray .add(ECPair.fromWIF(receive['wif'] as String, network: _network)); outputDataArray.add(P2PKH( network: _network, data: PaymentData( pubkey: Format.stringToUint8List( receive['publicKey'] as String))) .data .output!); break; } if (change['address'] == addressToCheckFor) { Logging.instance.log('Change found on loop $i', level: LogLevel.Info); // Logging.instance.log( // 'decoded change[\'wif\'] version: ${wif.decode(change['wif'] as String)}, _network: $_network'); elipticCurvePairArray .add(ECPair.fromWIF(change['wif'] as String, network: _network)); outputDataArray.add(P2PKH( network: _network, data: PaymentData( pubkey: Format.stringToUint8List( change['publicKey'] as String))) .data .output!); break; } } } final txb = TransactionBuilder(network: _network); txb.setVersion(2); int height = await getBlockHead(electrumXClient); txb.setLockTime(height); int amount = 0; // Add transaction inputs for (var i = 0; i < utxosToUse.length; i++) { txb.addInput( utxosToUse[i].txid, utxosToUse[i].vout, null, outputDataArray[i]); amount += utxosToUse[i].value; } final index = firoGetMintIndex()!; Logging.instance.log("index of mint $index", level: LogLevel.Info); for (var mintsElement in mintsMap) { Logging.instance.log("using $mintsElement", level: LogLevel.Info); Uint8List mintu8 = Format.stringToUint8List(mintsElement['script'] as String); txb.addOutput(mintu8, mintsElement['value'] as int); } for (var i = 0; i < utxosToUse.length; i++) { txb.sign( vin: i, keyPair: elipticCurvePairArray[i], witnessValue: utxosToUse[i].value, ); } var incomplete = txb.buildIncomplete(); var txId = incomplete.getId(); var txHex = incomplete.toHex(); int fee = amount - incomplete.outs[0].value!; var builtHex = txb.build(); // return builtHex; // final locale = // Platform.isWindows ? "en_US" : await Devicelocale.currentLocale; return { "transaction": builtHex, "txid": txId, "txHex": txHex, "value": amount - fee, "fees": Format.satoshisToAmount(fee, coin: coin).toDouble(), "publicCoin": "", "height": height, "txType": "Sent", "confirmed_status": false, "amount": Format.satoshisToAmount(amount, coin: coin).toDouble(), "timestamp": DateTime.now().millisecondsSinceEpoch ~/ 1000, "subType": "mint", "mintsMap": mintsMap, }; } Future _refreshLelantusData() async { final List> lelantusCoins = getLelantusCoinMap(); final jindexes = firoGetJIndex(); // Get all joinsplit transaction ids final listLelantusTxData = await db .getTransactions(walletId) .filter() .isLelantusEqualTo(true) .findAll(); List joinsplits = []; for (final tx in listLelantusTxData) { if (tx.subType == isar_models.TransactionSubType.join) { joinsplits.add(tx.txid); } } for (final coin in lelantusCoins.fold([], (previousValue, element) { (previousValue as List).add(element.values.first); return previousValue; })) { if (jindexes != null) { if (jindexes.contains(coin.index) && !joinsplits.contains(coin.txId)) { joinsplits.add(coin.txId); } } } Map> data = {}; for (final entry in listLelantusTxData) { data[entry.txid] = Tuple2(entry.address.value, entry); } // Grab the most recent information on all the joinsplits final updatedJSplit = await getJMintTransactions( cachedElectrumXClient, joinsplits, coin, ); final currentChainHeight = await chainHeight; // update all of joinsplits that are now confirmed. for (final tx in updatedJSplit.entries) { isar_models.Transaction? currentTx; try { currentTx = listLelantusTxData.firstWhere((e) => e.txid == tx.value.txid); } catch (_) { currentTx = null; } if (currentTx == null) { // this send was accidentally not included in the list tx.value.isLelantus = true; data[tx.value.txid] = Tuple2(tx.value.address.value ?? tx.key, tx.value); continue; } if (currentTx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) != tx.value.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) { tx.value.isLelantus = true; data[tx.value.txid] = Tuple2(tx.value.address.value ?? tx.key, tx.value); } } // Logging.instance.log(txData.txChunks); final listTxData = await _txnData; for (final value in listTxData) { // ignore change addresses // bool hasAtLeastOneReceive = false; // int howManyReceiveInputs = 0; // for (var element in value.inputs) { // if (listLelantusTxData.containsKey(element.txid) && // listLelantusTxData[element.txid]!.txType == "Received" // // && // // listLelantusTxData[element.txid].subType != "mint" // ) { // // hasAtLeastOneReceive = true; // // howManyReceiveInputs++; // } // } if (value.type == isar_models.TransactionType.incoming && value.subType != isar_models.TransactionSubType.mint) { // Every receive other than a mint should be shown. Mints will be collected and shown from the send side value.isLelantus = true; data[value.txid] = Tuple2(value.address.value, value); } else if (value.type == isar_models.TransactionType.outgoing) { // all sends should be shown, mints will be displayed correctly in the ui value.isLelantus = true; data[value.txid] = Tuple2(value.address.value, value); } } // TODO: optimize this whole lelantus process final List< Tuple4, List, isar_models.Address?>> txnsData = []; for (final value in data.values) { // allow possible null address on mints as we don't display address // this should normally never be null anyways but old (dbVersion up to 4) // migrated transactions may not have had an address (full rescan should // fix this) final transactionAddress = value.item2.subType == isar_models.TransactionSubType.mint ? value.item1 : value.item1!; final outs = value.item2.outputs.where((_) => true).toList(growable: false); final ins = value.item2.inputs.where((_) => true).toList(growable: false); txnsData.add(Tuple4(value.item2, outs, ins, transactionAddress)); } await addNewTransactionData(txnsData, walletId); // // update the _lelantusTransactionData // final models.TransactionData newTxData = // models.TransactionData.fromMap(listLelantusTxData); // // Logging.instance.log(newTxData.txChunks); // _lelantusTransactionData = Future(() => newTxData); // await DB.instance.put( // boxName: walletId, key: 'latest_lelantus_tx_model', value: newTxData); // return newTxData; } Future _getMintHex(int amount, int index) async { final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic'); final mintKeyPair = await compute( getBip32NodeWrapper, Tuple4( MINT_INDEX, index, mnemonic!, _network, ), ); String keydata = Format.uint8listToString(mintKeyPair.privateKey!); String seedID = Format.uint8listToString(mintKeyPair.identifier); String mintHex = await compute( _getMintScriptWrapper, Tuple5( amount, keydata, index, seedID, coin == Coin.firoTestNet, ), ); return mintHex; } Future _submitLelantusToNetwork( Map transactionInfo) async { final latestSetId = await getLatestSetId(); final txid = await submitHexToNetwork(transactionInfo['txHex'] as String); // success if txid matches the generated txid Logging.instance.log( "_submitLelantusToNetwork txid: ${transactionInfo['txid']}", level: LogLevel.Info); if (txid == transactionInfo['txid']) { final index = firoGetMintIndex(); final List> lelantusCoins = getLelantusCoinMap(); List> coins; if (lelantusCoins.isEmpty) { coins = []; } else { coins = [...lelantusCoins]; } if (transactionInfo['spendCoinIndexes'] != null) { // This is a joinsplit // Update all of the coins that have been spent. for (final lCoinMap in coins) { final lCoin = lCoinMap.values.first; if ((transactionInfo['spendCoinIndexes'] as List) .contains(lCoin.index)) { lCoinMap[lCoinMap.keys.first] = LelantusCoin( lCoin.index, lCoin.value, lCoin.publicCoin, lCoin.txId, lCoin.anonymitySetId, true); } } // if a jmint was made add it to the unspent coin index LelantusCoin jmint = LelantusCoin( index!, transactionInfo['jmintValue'] as int? ?? 0, transactionInfo['publicCoin'] as String, transactionInfo['txid'] as String, latestSetId, false); if (jmint.value > 0) { coins.add({jmint.txId: jmint}); final jindexes = firoGetJIndex()!; jindexes.add(index); await firoUpdateJIndex(jindexes); await firoUpdateMintIndex(index + 1); } await firoUpdateLelantusCoins(coins); // add the send transaction final transaction = isar_models.Transaction( walletId: walletId, txid: transactionInfo['txid'] as String, timestamp: transactionInfo['timestamp'] as int? ?? (DateTime.now().millisecondsSinceEpoch ~/ 1000), type: transactionInfo['txType'] == "Received" ? isar_models.TransactionType.incoming : isar_models.TransactionType.outgoing, subType: transactionInfo["subType"] == "mint" ? isar_models.TransactionSubType.mint : transactionInfo["subType"] == "join" ? isar_models.TransactionSubType.join : isar_models.TransactionSubType.none, amount: Format.decimalAmountToSatoshis( Decimal.parse(transactionInfo["amount"].toString()), coin, ), fee: Format.decimalAmountToSatoshis( Decimal.parse(transactionInfo["fees"].toString()), coin, ), height: transactionInfo["height"] as int?, isCancelled: false, isLelantus: true, slateId: null, otherData: transactionInfo["otherData"] as String?, ); final transactionAddress = await db .getAddresses(walletId) .filter() .valueEqualTo(transactionInfo["address"] as String) .findFirst() ?? isar_models.Address( walletId: walletId, value: transactionInfo["address"] as String, derivationIndex: -1, type: isar_models.AddressType.nonWallet, subType: isar_models.AddressSubType.nonWallet, publicKey: [], ); final List< Tuple4, List, isar_models.Address?>> txnsData = []; txnsData.add(Tuple4(transaction, [], [], transactionAddress)); await addNewTransactionData(txnsData, walletId); // final models.TransactionData newTxData = // models.TransactionData.fromMap(transactions); // await DB.instance.put( // boxName: walletId, // key: 'latest_lelantus_tx_model', // value: newTxData); // final ldata = DB.instance.get( // boxName: walletId, // key: 'latest_lelantus_tx_model') as models.TransactionData; // _lelantusTransactionData = Future(() => ldata); } else { // This is a mint Logging.instance.log("this is a mint", level: LogLevel.Info); // TODO: transactionInfo['mintsMap'] for (final mintMap in transactionInfo['mintsMap'] as List>) { final index = mintMap['index'] as int?; LelantusCoin mint = LelantusCoin( index!, mintMap['value'] as int, mintMap['publicCoin'] as String, transactionInfo['txid'] as String, latestSetId, false, ); if (mint.value > 0) { coins.add({mint.txId: mint}); await firoUpdateMintIndex(index + 1); } } // Logging.instance.log(coins); await firoUpdateLelantusCoins(coins); } return true; } else { // Failed to send to network return false; } } Future _getFees() async { try { //TODO adjust numbers for different speeds? const int f = 1, m = 5, s = 20; final fast = await electrumXClient.estimateFee(blocks: f); final medium = await electrumXClient.estimateFee(blocks: m); final slow = await electrumXClient.estimateFee(blocks: s); final feeObject = FeeObject( numberOfBlocksFast: f, numberOfBlocksAverage: m, numberOfBlocksSlow: s, fast: Format.decimalAmountToSatoshis(fast, coin), medium: Format.decimalAmountToSatoshis(medium, coin), slow: Format.decimalAmountToSatoshis(slow, coin), ); Logging.instance.log("fetched fees: $feeObject", level: LogLevel.Info); return feeObject; // final result = await electrumXClient.getFeeRate(); // // final locale = await Devicelocale.currentLocale; // final String fee = // Format.satoshiAmountToPrettyString(result["rate"] as int, locale!); // // final fees = { // "fast": fee, // "average": fee, // "slow": fee, // }; // final FeeObject feeObject = FeeObject.fromJson(fees); // return feeObject; } catch (e) { Logging.instance .log("Exception rethrown from _getFees(): $e", level: LogLevel.Error); rethrow; } } Future _getCurrentNode() async { final node = NodeService(secureStorageInterface: _secureStore) .getPrimaryNodeFor(coin: coin) ?? DefaultNodes.getNodeFor(coin); return ElectrumXNode( address: node.host, port: node.port, name: node.name, useSSL: node.useSSL, id: node.id, ); } //TODO call get transaction and check each tx to see if it is a "received" tx? Future _getReceivedTxCount({required String address}) async { try { final scripthash = AddressUtils.convertToScriptHash(address, _network); final transactions = await electrumXClient.getHistory(scripthash: scripthash); return transactions.length; } catch (e) { Logging.instance.log( "Exception rethrown in _getReceivedTxCount(address: $address): $e", level: LogLevel.Error); rethrow; } } Future checkReceivingAddressForTransactions() async { try { final currentReceiving = await _currentReceivingAddress; final int txCount = await _getReceivedTxCount(address: currentReceiving.value); Logging.instance.log( 'Number of txs for current receiving address $currentReceiving: $txCount', level: LogLevel.Info); if (txCount >= 1) { // First increment the receiving index final newReceivingIndex = currentReceiving.derivationIndex + 1; // Use new index to derive a new receiving address final newReceivingAddress = await _generateAddressForChain( 0, newReceivingIndex, ); final existing = await db .getAddresses(walletId) .filter() .valueEqualTo(newReceivingAddress.value) .findFirst(); if (existing == null) { // Add that new change address await db.putAddress(newReceivingAddress); } else { // we need to update the address await db.updateAddress(existing, newReceivingAddress); // since we updated an existing address there is a chance it has // some tx history. To prevent address reuse we will call check again // recursively await checkReceivingAddressForTransactions(); } } } on SocketException catch (se, s) { Logging.instance.log( "SocketException caught in checkReceivingAddressForTransactions(): $se\n$s", level: LogLevel.Error); return; } catch (e, s) { Logging.instance.log( "Exception rethrown from checkReceivingAddressForTransactions(): $e\n$s", level: LogLevel.Error); rethrow; } } Future checkChangeAddressForTransactions() async { try { final currentChange = await _currentChangeAddress; final int txCount = await _getReceivedTxCount(address: currentChange.value); Logging.instance.log( 'Number of txs for current change address: $currentChange: $txCount', level: LogLevel.Info); if (txCount >= 1) { // First increment the change index final newChangeIndex = currentChange.derivationIndex + 1; // Use new index to derive a new change address final newChangeAddress = await _generateAddressForChain( 1, newChangeIndex, ); final existing = await db .getAddresses(walletId) .filter() .valueEqualTo(newChangeAddress.value) .findFirst(); if (existing == null) { // Add that new change address await db.putAddress(newChangeAddress); } else { // we need to update the address await db.updateAddress(existing, newChangeAddress); // since we updated an existing address there is a chance it has // some tx history. To prevent address reuse we will call check again // recursively await checkChangeAddressForTransactions(); } } } on SocketException catch (se, s) { Logging.instance.log( "SocketException caught in checkChangeAddressForTransactions(): $se\n$s", level: LogLevel.Error); return; } catch (e, s) { Logging.instance.log( "Exception rethrown from checkChangeAddressForTransactions(): $e\n$s", level: LogLevel.Error); rethrow; } } Future> _fetchAllOwnAddresses() async { final allAddresses = await db .getAddresses(walletId) .filter() .not() .typeEqualTo(isar_models.AddressType.nonWallet) .and() .group((q) => q .subTypeEqualTo(isar_models.AddressSubType.receiving) .or() .subTypeEqualTo(isar_models.AddressSubType.change)) .findAll(); // final List allAddresses = []; // final receivingAddresses = // DB.instance.get(boxName: walletId, key: 'receivingAddresses') // as List; // final changeAddresses = // DB.instance.get(boxName: walletId, key: 'changeAddresses') // as List; // // for (var i = 0; i < receivingAddresses.length; i++) { // allAddresses.add(receivingAddresses[i] as String); // } // for (var i = 0; i < changeAddresses.length; i++) { // allAddresses.add(changeAddresses[i] as String); // } return allAddresses; } Future>> _fetchHistory( List allAddresses) async { try { List> allTxHashes = []; final Map>> batches = {}; final Map requestIdToAddressMap = {}; const batchSizeMax = 100; int batchNumber = 0; for (int i = 0; i < allAddresses.length; i++) { if (batches[batchNumber] == null) { batches[batchNumber] = {}; } final scripthash = AddressUtils.convertToScriptHash(allAddresses[i], _network); final id = Logger.isTestEnv ? "$i" : const Uuid().v1(); requestIdToAddressMap[id] = allAddresses[i]; batches[batchNumber]!.addAll({ id: [scripthash] }); if (i % batchSizeMax == batchSizeMax - 1) { batchNumber++; } } for (int i = 0; i < batches.length; i++) { final response = await _electrumXClient.getBatchHistory(args: batches[i]!); for (final entry in response.entries) { for (int j = 0; j < entry.value.length; j++) { entry.value[j]["address"] = requestIdToAddressMap[entry.key]; if (!allTxHashes.contains(entry.value[j])) { allTxHashes.add(entry.value[j]); } } } } return allTxHashes; } catch (e, s) { Logging.instance.log("_fetchHistory: $e\n$s", level: LogLevel.Error); rethrow; } } bool _duplicateTxCheck( List> allTransactions, String txid) { for (int i = 0; i < allTransactions.length; i++) { if (allTransactions[i]["txid"] == txid) { return true; } } return false; } Future _refreshTransactions() async { final List allAddresses = await _fetchAllOwnAddresses(); final List> allTxHashes = await _fetchHistory(allAddresses.map((e) => e.value).toList()); List> allTransactions = []; final currentHeight = await chainHeight; for (final txHash in allTxHashes) { final storedTx = await db .getTransactions(walletId) .filter() .txidEqualTo(txHash["tx_hash"] as String) .findFirst(); if (storedTx == null || !storedTx.isConfirmed(currentHeight, MINIMUM_CONFIRMATIONS)) { final tx = await cachedElectrumXClient.getTransaction( txHash: txHash["tx_hash"] as String, verbose: true, coin: coin, ); if (!_duplicateTxCheck(allTransactions, tx["txid"] as String)) { tx["address"] = await db .getAddresses(walletId) .filter() .valueEqualTo(txHash["address"] as String) .findFirst(); tx["height"] = txHash["height"]; allTransactions.add(tx); } } } final List< Tuple4, List, isar_models.Address?>> txnsData = []; Set changeAddresses = allAddresses .where((e) => e.subType == isar_models.AddressSubType.change) .map((e) => e.value) .toSet(); for (final txObject in allTransactions) { // Logging.instance.log(txObject); List sendersArray = []; List recipientsArray = []; // Usually only has value when txType = 'Send' int inputAmtSentFromWallet = 0; // Usually has value regardless of txType due to change addresses int outputAmtAddressedToWallet = 0; for (final input in txObject["vin"] as List) { final address = input["address"] as String?; if (address != null) { sendersArray.add(address); } } // Logging.instance.log("sendersArray: $sendersArray"); for (final output in txObject["vout"] as List) { final address = output["scriptPubKey"]?["addresses"]?[0] as String? ?? output["scriptPubKey"]?["address"] as String?; if (address != null) { recipientsArray.add(address); } } // Logging.instance.log("recipientsArray: $recipientsArray"); final foundInSenders = allAddresses.any((element) => sendersArray.contains(element.value)); // Logging.instance.log("foundInSenders: $foundInSenders"); String outAddress = ""; int fees = 0; // If txType = Sent, then calculate inputAmtSentFromWallet, calculate who received how much in aliens array (check outputs) if (foundInSenders) { int outAmount = 0; int inAmount = 0; bool nFeesUsed = false; for (final input in txObject["vin"] as List) { final nFees = input["nFees"]; if (nFees != null) { nFeesUsed = true; fees = (Decimal.parse(nFees.toString()) * Decimal.fromInt(Constants.satsPerCoin(coin))) .toBigInt() .toInt(); } final address = input["address"] as String?; final value = input["valueSat"] as int?; if (address != null && value != null) { if (allAddresses.where((e) => e.value == address).isNotEmpty) { inputAmtSentFromWallet += value; } } if (value != null) { inAmount += value; } } for (final output in txObject["vout"] as List) { final address = output["scriptPubKey"]?["addresses"]?[0] as String? ?? output["scriptPubKey"]?["address"] as String?; final value = output["value"]; if (value != null) { outAmount += (Decimal.parse(value.toString()) * Decimal.fromInt(Constants.satsPerCoin(coin))) .toBigInt() .toInt(); if (address != null) { if (changeAddresses.contains(address)) { inputAmtSentFromWallet -= (Decimal.parse(value.toString()) * Decimal.fromInt(Constants.satsPerCoin(coin))) .toBigInt() .toInt(); } else { outAddress = address; } } } } fees = nFeesUsed ? fees : inAmount - outAmount; inputAmtSentFromWallet -= inAmount - outAmount; } else { for (final input in txObject["vin"] as List) { final nFees = input["nFees"]; if (nFees != null) { fees += (Decimal.parse(nFees.toString()) * Decimal.fromInt(Constants.satsPerCoin(coin))) .toBigInt() .toInt(); } } for (final output in txObject["vout"] as List) { final addresses = output["scriptPubKey"]["addresses"] as List?; if (addresses != null && addresses.isNotEmpty) { final address = addresses[0] as String; final value = output["value"] ?? 0; // Logging.instance.log(address + value.toString()); if (allAddresses.where((e) => e.value == address).isNotEmpty) { outputAmtAddressedToWallet += (Decimal.parse(value.toString()) * Decimal.fromInt(Constants.satsPerCoin(coin))) .toBigInt() .toInt(); outAddress = address; } } } } isar_models.TransactionType type; isar_models.TransactionSubType subType = isar_models.TransactionSubType.none; int amount; if (foundInSenders) { type = isar_models.TransactionType.outgoing; amount = inputAmtSentFromWallet; if (txObject["vout"][0]["scriptPubKey"]["type"] == "lelantusmint") { subType = isar_models.TransactionSubType.mint; } } else { type = isar_models.TransactionType.incoming; amount = outputAmtAddressedToWallet; } final transactionAddress = allAddresses.firstWhere((e) => e.value == outAddress, orElse: () => isar_models.Address( walletId: walletId, value: outAddress, derivationIndex: -1, type: isar_models.AddressType.nonWallet, subType: isar_models.AddressSubType.nonWallet, publicKey: [], )); final tx = isar_models.Transaction( walletId: walletId, txid: txObject["txid"] as String, timestamp: txObject["blocktime"] as int? ?? (DateTime.now().millisecondsSinceEpoch ~/ 1000), type: type, subType: subType, amount: amount, fee: fees, height: txObject["height"] as int? ?? 0, isCancelled: false, isLelantus: false, slateId: null, otherData: null, ); List outs = []; List ins = []; for (final json in txObject["vin"] as List) { bool isCoinBase = json['coinbase'] != null; final input = isar_models.Input( walletId: walletId, txid: json['txid'] as String? ?? "", vout: json['vout'] as int? ?? -1, scriptSig: json['scriptSig']?['hex'] as String?, scriptSigAsm: json['scriptSig']?['asm'] as String?, isCoinbase: isCoinBase ? isCoinBase : json['is_coinbase'] as bool?, sequence: json['sequence'] as int?, innerRedeemScriptAsm: json['innerRedeemscriptAsm'] as String?, ); ins.add(input); } for (final json in txObject["vout"] as List) { final output = isar_models.Output( walletId: walletId, scriptPubKey: json['scriptPubKey']?['hex'] as String?, scriptPubKeyAsm: json['scriptPubKey']?['asm'] as String?, scriptPubKeyType: json['scriptPubKey']?['type'] as String?, scriptPubKeyAddress: json["scriptPubKey"]?["addresses"]?[0] as String? ?? json['scriptPubKey']['type'] as String, value: Format.decimalAmountToSatoshis( Decimal.parse(json["value"].toString()), coin, ), ); outs.add(output); } txnsData.add(Tuple4(tx, outs, ins, transactionAddress)); } await addNewTransactionData(txnsData, walletId); // quick hack to notify manager to call notifyListeners if // transactions changed if (txnsData.isNotEmpty) { GlobalEventBus.instance.fire( UpdatedInBackgroundEvent( "Transactions updated/added for: $walletId $walletName ", walletId, ), ); } } Future _refreshUTXOs() async { final allAddresses = await _fetchAllOwnAddresses(); try { final fetchedUtxoList = >>[]; final Map>> batches = {}; const batchSizeMax = 100; int batchNumber = 0; for (int i = 0; i < allAddresses.length; i++) { if (batches[batchNumber] == null) { batches[batchNumber] = {}; } final scripthash = AddressUtils.convertToScriptHash(allAddresses[i].value, _network); batches[batchNumber]!.addAll({ scripthash: [scripthash] }); if (i % batchSizeMax == batchSizeMax - 1) { batchNumber++; } } for (int i = 0; i < batches.length; i++) { final response = await _electrumXClient.getBatchUTXOs(args: batches[i]!); for (final entry in response.entries) { if (entry.value.isNotEmpty) { fetchedUtxoList.add(entry.value); } } } final currentChainHeight = await chainHeight; final List outputArray = []; int satoshiBalanceTotal = 0; int satoshiBalancePending = 0; int satoshiBalanceSpendable = 0; int satoshiBalanceBlocked = 0; for (int i = 0; i < fetchedUtxoList.length; i++) { for (int j = 0; j < fetchedUtxoList[i].length; j++) { final txn = await cachedElectrumXClient.getTransaction( txHash: fetchedUtxoList[i][j]["tx_hash"] as String, verbose: true, coin: coin, ); // todo check here if we should mark as blocked final utxo = isar_models.UTXO( walletId: walletId, txid: txn["txid"] as String, vout: fetchedUtxoList[i][j]["tx_pos"] as int, value: fetchedUtxoList[i][j]["value"] as int, name: "", isBlocked: false, blockedReason: null, isCoinbase: txn["is_coinbase"] as bool? ?? false, blockHash: txn["blockhash"] as String?, blockHeight: fetchedUtxoList[i][j]["height"] as int?, blockTime: txn["blocktime"] as int?, ); satoshiBalanceTotal += utxo.value; if (utxo.isBlocked) { satoshiBalanceBlocked += utxo.value; } else { if (utxo.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) { satoshiBalanceSpendable += utxo.value; } else { satoshiBalancePending += utxo.value; } } outputArray.add(utxo); } } Logging.instance .log('Outputs fetched: $outputArray', level: LogLevel.Info); // TODO move this out of here and into IDB await db.isar.writeTxn(() async { await db.isar.utxos.clear(); await db.isar.utxos.putAll(outputArray); }); // finally update public balance _balance = Balance( coin: coin, total: satoshiBalanceTotal, spendable: satoshiBalanceSpendable, blockedTotal: satoshiBalanceBlocked, pendingSpendable: satoshiBalancePending, ); await updateCachedBalance(_balance!); } catch (e, s) { Logging.instance .log("Output fetch unsuccessful: $e\n$s", level: LogLevel.Error); } } /// Returns the latest receiving/change (external/internal) address for the wallet depending on [chain] /// [chain] - Use 0 for receiving (external), 1 for change (internal). Should not be any other value! Future _getCurrentAddressForChain(int chain) async { final subType = chain == 0 // Here, we assume that chain == 1 if it isn't 0 ? isar_models.AddressSubType.receiving : isar_models.AddressSubType.change; isar_models.Address? address = await db .getAddresses(walletId) .filter() .typeEqualTo(isar_models.AddressType.p2pkh) .subTypeEqualTo(subType) .sortByDerivationIndexDesc() .findFirst(); return address!.value; } Future fillAddresses(String suppliedMnemonic, {int perBatch = 50, int numberOfThreads = 4}) async { if (numberOfThreads <= 0) { numberOfThreads = 1; } if (Platform.environment["FLUTTER_TEST"] == "true" || integrationTestFlag) { perBatch = 10; numberOfThreads = 4; } final receiveDerivationsString = await _secureStore.read(key: "${walletId}_receiveDerivations"); final changeDerivationsString = await _secureStore.read(key: "${walletId}_changeDerivations"); var receiveDerivations = Map.from( jsonDecode(receiveDerivationsString ?? "{}") as Map); var changeDerivations = Map.from( jsonDecode(changeDerivationsString ?? "{}") as Map); final int start = receiveDerivations.length; List ports = List.empty(growable: true); for (int i = 0; i < numberOfThreads; i++) { ReceivePort receivePort = await getIsolate({ "function": "isolateDerive", "mnemonic": suppliedMnemonic, "from": start + i * perBatch, "to": start + (i + 1) * perBatch, "network": _network, }); ports.add(receivePort); } for (int i = 0; i < numberOfThreads; i++) { ReceivePort receivePort = ports.elementAt(i); var message = await receivePort.first; if (message is String) { Logging.instance.log("this is a string", level: LogLevel.Error); stop(receivePort); throw Exception("isolateDerive isolate failed"); } stop(receivePort); Logging.instance.log('Closing isolateDerive!', level: LogLevel.Info); receiveDerivations.addAll(message['receive'] as Map); changeDerivations.addAll(message['change'] as Map); } Logging.instance.log("isolate derives", level: LogLevel.Info); // Logging.instance.log(receiveDerivations); // Logging.instance.log(changeDerivations); final newReceiveDerivationsString = jsonEncode(receiveDerivations); final newChangeDerivationsString = jsonEncode(changeDerivations); await _secureStore.write( key: "${walletId}_receiveDerivations", value: newReceiveDerivationsString); await _secureStore.write( key: "${walletId}_changeDerivations", value: newChangeDerivationsString); } /// Generates a new internal or external chain address for the wallet using a BIP84 derivation path. /// [chain] - Use 0 for receiving (external), 1 for change (internal). Should not be any other value! /// [index] - This can be any integer >= 0 Future _generateAddressForChain( int chain, int index) async { // final wallet = await Hive.openBox(this._walletId); final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic'); Map? derivations; if (chain == 0) { final receiveDerivationsString = await _secureStore.read(key: "${walletId}_receiveDerivations"); derivations = Map.from( jsonDecode(receiveDerivationsString ?? "{}") as Map); } else if (chain == 1) { final changeDerivationsString = await _secureStore.read(key: "${walletId}_changeDerivations"); derivations = Map.from( jsonDecode(changeDerivationsString ?? "{}") as Map); } if (derivations!.isNotEmpty) { if (derivations["$index"] == null) { await fillAddresses(mnemonic!, numberOfThreads: Platform.numberOfProcessors - isolates.length - 1); Logging.instance.log("calling _generateAddressForChain recursively", level: LogLevel.Info); return _generateAddressForChain(chain, index); } return isar_models.Address( walletId: walletId, value: derivations["$index"]['address'] as String, publicKey: Format.stringToUint8List( derivations["$index"]['publicKey'] as String), type: isar_models.AddressType.p2pkh, derivationIndex: index, subType: chain == 0 ? isar_models.AddressSubType.receiving : isar_models.AddressSubType.change, ); } else { final node = await compute( getBip32NodeWrapper, Tuple4(chain, index, mnemonic!, _network)); final address = P2PKH(network: _network, data: PaymentData(pubkey: node.publicKey)) .data .address!; return isar_models.Address( walletId: walletId, value: address, publicKey: node.publicKey, type: isar_models.AddressType.p2pkh, derivationIndex: index, subType: chain == 0 ? isar_models.AddressSubType.receiving : isar_models.AddressSubType.change, ); } } // /// Takes in a list of isar_models.UTXOs and adds a name (dependent on object index within list) // /// and checks for the txid associated with the utxo being blocked and marks it accordingly. // /// Now also checks for output labeling. // Future _sortOutputs(List utxos) async { // final blockedHashArray = // DB.instance.get(boxName: walletId, key: 'blocked_tx_hashes') // as List?; // final List lst = []; // if (blockedHashArray != null) { // for (var hash in blockedHashArray) { // lst.add(hash as String); // } // } // final labels = // DB.instance.get(boxName: walletId, key: 'labels') as Map? ?? // {}; // // _outputsList = []; // // for (var i = 0; i < utxos.length; i++) { // if (labels[utxos[i].txid] != null) { // utxos[i].txName = labels[utxos[i].txid] as String? ?? ""; // } else { // utxos[i].txName = 'Output #$i'; // } // // if (utxos[i].status.confirmed == false) { // _outputsList.add(utxos[i]); // } else { // if (lst.contains(utxos[i].txid)) { // utxos[i].blocked = true; // _outputsList.add(utxos[i]); // } else if (!lst.contains(utxos[i].txid)) { // _outputsList.add(utxos[i]); // } // } // } // } @override Future fullRescan( int maxUnusedAddressGap, int maxNumberOfIndexesToCheck, ) async { Logging.instance.log("Starting full rescan!", level: LogLevel.Info); // timer?.cancel(); // for (final isolate in isolates.values) { // isolate.kill(priority: Isolate.immediate); // } // isolates.clear(); longMutex = true; GlobalEventBus.instance.fire( WalletSyncStatusChangedEvent( WalletSyncStatus.syncing, walletId, coin, ), ); // clear cache await _cachedElectrumXClient.clearSharedTransactionCache(coin: coin); // back up data // await _rescanBackup(); // clear blockchain info await db.deleteWalletBlockchainData(walletId); await _deleteDerivations(); try { final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic'); await _recoverWalletFromBIP32SeedPhrase(mnemonic!, maxUnusedAddressGap); longMutex = false; await refresh(); Logging.instance.log("Full rescan complete!", level: LogLevel.Info); GlobalEventBus.instance.fire( WalletSyncStatusChangedEvent( WalletSyncStatus.synced, walletId, coin, ), ); } catch (e, s) { GlobalEventBus.instance.fire( WalletSyncStatusChangedEvent( WalletSyncStatus.unableToSync, walletId, coin, ), ); // restore from backup // await _rescanRestore(); longMutex = false; Logging.instance.log("Exception rethrown from fullRescan(): $e\n$s", level: LogLevel.Error); rethrow; } } Future _deleteDerivations() async { // P2PKH derivations await _secureStore.delete(key: "${walletId}_receiveDerivations"); await _secureStore.delete(key: "${walletId}_changeDerivations"); } // Future _rescanBackup() async { // Logging.instance.log("starting rescan backup", level: LogLevel.Info); // // // backup current and clear data // final tempReceivingAddresses = // DB.instance.get(boxName: walletId, key: 'receivingAddresses'); // await DB.instance.delete( // key: 'receivingAddresses', // boxName: walletId, // ); // await DB.instance.put( // boxName: walletId, // key: 'receivingAddresses_BACKUP', // value: tempReceivingAddresses); // // final tempChangeAddresses = // DB.instance.get(boxName: walletId, key: 'changeAddresses'); // await DB.instance.delete( // key: 'changeAddresses', // boxName: walletId, // ); // await DB.instance.put( // boxName: walletId, // key: 'changeAddresses_BACKUP', // value: tempChangeAddresses); // // final tempReceivingIndex = // DB.instance.get(boxName: walletId, key: 'receivingIndex'); // await DB.instance.delete( // key: 'receivingIndex', // boxName: walletId, // ); // await DB.instance.put( // boxName: walletId, // key: 'receivingIndex_BACKUP', // value: tempReceivingIndex); // // final tempChangeIndex = // DB.instance.get(boxName: walletId, key: 'changeIndex'); // await DB.instance.delete( // key: 'changeIndex', // boxName: walletId, // ); // await DB.instance.put( // boxName: walletId, key: 'changeIndex_BACKUP', value: tempChangeIndex); // // final receiveDerivationsString = // await _secureStore.read(key: "${walletId}_receiveDerivations"); // final changeDerivationsString = // await _secureStore.read(key: "${walletId}_changeDerivations"); // // await _secureStore.write( // key: "${walletId}_receiveDerivations_BACKUP", // value: receiveDerivationsString); // await _secureStore.write( // key: "${walletId}_changeDerivations_BACKUP", // value: changeDerivationsString); // // await _secureStore.write( // key: "${walletId}_receiveDerivations", value: null); // await _secureStore.write(key: "${walletId}_changeDerivations", value: null); // // // back up but no need to delete // final tempMintIndex = // DB.instance.get(boxName: walletId, key: 'mintIndex'); // await DB.instance.put( // boxName: walletId, key: 'mintIndex_BACKUP', value: tempMintIndex); // // final tempLelantusCoins = // DB.instance.get(boxName: walletId, key: '_lelantus_coins'); // await DB.instance.put( // boxName: walletId, // key: '_lelantus_coins_BACKUP', // value: tempLelantusCoins); // // final tempJIndex = // DB.instance.get(boxName: walletId, key: 'jindex'); // await DB.instance.put( // boxName: walletId, key: 'jindex_BACKUP', value: tempJIndex); // // final tempLelantusTxModel = DB.instance // .get(boxName: walletId, key: 'latest_lelantus_tx_model'); // await DB.instance.put( // boxName: walletId, // key: 'latest_lelantus_tx_model_BACKUP', // value: tempLelantusTxModel); // // Logging.instance.log("rescan backup complete", level: LogLevel.Info); // } // // Future _rescanRestore() async { // Logging.instance.log("starting rescan restore", level: LogLevel.Info); // // // restore from backup // final tempReceivingAddresses = DB.instance // .get(boxName: walletId, key: 'receivingAddresses_BACKUP'); // final tempChangeAddresses = DB.instance // .get(boxName: walletId, key: 'changeAddresses_BACKUP'); // final tempReceivingIndex = DB.instance // .get(boxName: walletId, key: 'receivingIndex_BACKUP'); // final tempChangeIndex = // DB.instance.get(boxName: walletId, key: 'changeIndex_BACKUP'); // final tempMintIndex = // DB.instance.get(boxName: walletId, key: 'mintIndex_BACKUP'); // final tempLelantusCoins = DB.instance // .get(boxName: walletId, key: '_lelantus_coins_BACKUP'); // final tempJIndex = // DB.instance.get(boxName: walletId, key: 'jindex_BACKUP'); // final tempLelantusTxModel = DB.instance.get( // boxName: walletId, key: 'latest_lelantus_tx_model_BACKUP'); // // final receiveDerivationsString = // await _secureStore.read(key: "${walletId}_receiveDerivations_BACKUP"); // final changeDerivationsString = // await _secureStore.read(key: "${walletId}_changeDerivations_BACKUP"); // // await _secureStore.write( // key: "${walletId}_receiveDerivations", value: receiveDerivationsString); // await _secureStore.write( // key: "${walletId}_changeDerivations", value: changeDerivationsString); // // await DB.instance.put( // boxName: walletId, // key: 'receivingAddresses', // value: tempReceivingAddresses); // await DB.instance.put( // boxName: walletId, key: 'changeAddresses', value: tempChangeAddresses); // await DB.instance.put( // boxName: walletId, key: 'receivingIndex', value: tempReceivingIndex); // await DB.instance.put( // boxName: walletId, key: 'changeIndex', value: tempChangeIndex); // await DB.instance.put( // boxName: walletId, key: 'mintIndex', value: tempMintIndex); // await DB.instance.put( // boxName: walletId, key: '_lelantus_coins', value: tempLelantusCoins); // await DB.instance // .put(boxName: walletId, key: 'jindex', value: tempJIndex); // await DB.instance.put( // boxName: walletId, // key: 'latest_lelantus_tx_model', // value: tempLelantusTxModel); // // Logging.instance.log("rescan restore complete", level: LogLevel.Info); // } /// wrapper for _recoverWalletFromBIP32SeedPhrase() @override Future recoverFromMnemonic({ required String mnemonic, required int maxUnusedAddressGap, required int maxNumberOfIndexesToCheck, required int height, }) async { try { await compute( _setTestnetWrapper, coin == Coin.firoTestNet, ); Logging.instance.log("IS_INTEGRATION_TEST: $integrationTestFlag", level: LogLevel.Info); if (!integrationTestFlag) { final features = await electrumXClient.getServerFeatures(); Logging.instance.log("features: $features", level: LogLevel.Info); switch (coin) { case Coin.firo: if (features['genesis_hash'] != GENESIS_HASH_MAINNET) { throw Exception("genesis hash does not match main net!"); } break; case Coin.firoTestNet: if (features['genesis_hash'] != GENESIS_HASH_TESTNET) { throw Exception("genesis hash does not match test net!"); } break; default: throw Exception( "Attempted to generate a FiroWallet using a non firo coin type: ${coin.name}"); } // if (_networkType == BasicNetworkType.main) { // if (features['genesis_hash'] != GENESIS_HASH_MAINNET) { // throw Exception("genesis hash does not match main net!"); // } // } else if (_networkType == BasicNetworkType.test) { // if (features['genesis_hash'] != GENESIS_HASH_TESTNET) { // throw Exception("genesis hash does not match test net!"); // } // } } // this should never fail if ((await _secureStore.read(key: '${_walletId}_mnemonic')) != null) { throw Exception("Attempted to overwrite mnemonic on restore!"); } await _secureStore.write( key: '${_walletId}_mnemonic', value: mnemonic.trim()); await _recoverWalletFromBIP32SeedPhrase( mnemonic.trim(), maxUnusedAddressGap); await compute( _setTestnetWrapper, false, ); } catch (e, s) { await compute( _setTestnetWrapper, false, ); Logging.instance.log( "Exception rethrown from recoverFromMnemonic(): $e\n$s", level: LogLevel.Error); rethrow; } } bool longMutex = false; Future> getSetDataMap(int latestSetId) async { final Map setDataMap = {}; final anonymitySets = await fetchAnonymitySets(); for (int setId = 1; setId <= latestSetId; setId++) { final setData = anonymitySets .firstWhere((element) => element["setId"] == setId, orElse: () => {}); if (setData.isNotEmpty) { setDataMap[setId] = setData; } } return setDataMap; } Future _makeDerivations( String suppliedMnemonic, int maxUnusedAddressGap, ) async { List receivingAddressArray = []; List changeAddressArray = []; int receivingIndex = -1; int changeIndex = -1; // The gap limit will be capped at 20 int receivingGapCounter = 0; int changeGapCounter = 0; await fillAddresses(suppliedMnemonic, numberOfThreads: Platform.numberOfProcessors - isolates.length - 1); final receiveDerivationsString = await _secureStore.read(key: "${walletId}_receiveDerivations"); final changeDerivationsString = await _secureStore.read(key: "${walletId}_changeDerivations"); final receiveDerivations = Map.from( jsonDecode(receiveDerivationsString ?? "{}") as Map); final changeDerivations = Map.from( jsonDecode(changeDerivationsString ?? "{}") as Map); // log("rcv: $receiveDerivations"); // log("chg: $changeDerivations"); // Deriving and checking for receiving addresses for (var i = 0; i < receiveDerivations.length; i++) { // Break out of loop when receivingGapCounter hits maxUnusedAddressGap // Same gap limit for change as for receiving, breaks when it hits maxUnusedAddressGap if (receivingGapCounter >= maxUnusedAddressGap && changeGapCounter >= maxUnusedAddressGap) { break; } final receiveDerivation = receiveDerivations["$i"]; final address = receiveDerivation['address'] as String; final changeDerivation = changeDerivations["$i"]; final _address = changeDerivation['address'] as String; Future? futureNumTxs; Future? _futureNumTxs; if (receivingGapCounter < maxUnusedAddressGap) { futureNumTxs = _getReceivedTxCount(address: address); } if (changeGapCounter < maxUnusedAddressGap) { _futureNumTxs = _getReceivedTxCount(address: _address); } try { if (futureNumTxs != null) { int numTxs = await futureNumTxs; if (numTxs >= 1) { receivingIndex = i; final addr = isar_models.Address( walletId: walletId, value: address, publicKey: Format.stringToUint8List( receiveDerivation['publicKey'] as String), type: isar_models.AddressType.p2pkh, derivationIndex: i, subType: isar_models.AddressSubType.receiving, ); receivingAddressArray.add(addr); } else if (numTxs == 0) { receivingGapCounter += 1; } } } catch (e, s) { Logging.instance.log( "Exception rethrown from recoverWalletFromBIP32SeedPhrase(): $e\n$s", level: LogLevel.Error); rethrow; } try { if (_futureNumTxs != null) { int numTxs = await _futureNumTxs; if (numTxs >= 1) { changeIndex = i; final addr = isar_models.Address( walletId: walletId, value: _address, publicKey: Format.stringToUint8List( changeDerivation['publicKey'] as String), type: isar_models.AddressType.p2pkh, derivationIndex: i, subType: isar_models.AddressSubType.change, ); changeAddressArray.add(addr); } else if (numTxs == 0) { changeGapCounter += 1; } } } catch (e, s) { Logging.instance.log( "Exception rethrown from recoverWalletFromBIP32SeedPhrase(): $e\n$s", level: LogLevel.Error); rethrow; } } // If restoring a wallet that never received any funds, then set receivingArray manually // If we didn't do this, it'd store an empty array if (receivingIndex == -1) { final receivingAddress = await _generateAddressForChain(0, 0); receivingAddressArray.add(receivingAddress); } // If restoring a wallet that never sent any funds with change, then set changeArray // manually. If we didn't do this, it'd store an empty array. if (changeIndex == -1) { final changeAddress = await _generateAddressForChain(1, 0); changeAddressArray.add(changeAddress); } await db.putAddresses([ ...receivingAddressArray, ...changeAddressArray, ]); } /// Recovers wallet from [suppliedMnemonic]. Expects a valid mnemonic. Future _recoverWalletFromBIP32SeedPhrase( String suppliedMnemonic, int maxUnusedAddressGap) async { longMutex = true; Logging.instance .log("PROCESSORS ${Platform.numberOfProcessors}", level: LogLevel.Info); try { await isarInit(walletId); final latestSetId = await getLatestSetId(); final setDataMap = getSetDataMap(latestSetId); final usedSerialNumbers = getUsedCoinSerials(); final makeDerivations = _makeDerivations(suppliedMnemonic, maxUnusedAddressGap); await Future.wait([ updateCachedId(walletId), updateCachedIsFavorite(false), ]); await Future.wait([usedSerialNumbers, setDataMap, makeDerivations]); await _restore(latestSetId, await setDataMap, await usedSerialNumbers); longMutex = false; } catch (e, s) { longMutex = false; Logging.instance.log( "Exception rethrown from recoverWalletFromBIP32SeedPhrase(): $e\n$s", level: LogLevel.Error); rethrow; } } Future _restore(int latestSetId, Map setDataMap, dynamic usedSerialNumbers) async { final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic'); final dataFuture = _refreshTransactions(); ReceivePort receivePort = await getIsolate({ "function": "restore", "mnemonic": mnemonic, "coin": coin, "latestSetId": latestSetId, "setDataMap": setDataMap, "usedSerialNumbers": usedSerialNumbers, "network": _network, }); await Future.wait([dataFuture]); var result = await receivePort.first; if (result is String) { Logging.instance .log("restore() ->> this is a string", level: LogLevel.Error); stop(receivePort); throw Exception("isolate restore failed."); } stop(receivePort); final message = await staticProcessRestore( (await _txnData), result as Map, await chainHeight, ); await Future.wait([ firoUpdateMintIndex(message['mintIndex'] as int), firoUpdateLelantusCoins(message['_lelantus_coins'] as List), firoUpdateJIndex(message['jindex'] as List), ]); final transactionMap = message["newTxMap"] as Map; Map> data = {}; for (final entry in transactionMap.entries) { data[entry.key] = Tuple2(entry.value.address.value, entry.value); } // Create the joinsplit transactions. final spendTxs = await getJMintTransactions( _cachedElectrumXClient, message["spendTxIds"] as List, coin, ); Logging.instance.log(spendTxs, level: LogLevel.Info); for (var element in spendTxs.entries) { final address = element.value.address.value ?? data[element.value.txid]?.item1 ?? element.key; // isar_models.Address( // walletId: walletId, // value: transactionInfo["address"] as String, // derivationIndex: -1, // type: isar_models.AddressType.nonWallet, // subType: isar_models.AddressSubType.nonWallet, // publicKey: [], // ); data[element.value.txid] = Tuple2(address, element.value); } final List< Tuple4, List, isar_models.Address?>> txnsData = []; for (final value in data.values) { final transactionAddress = value.item1!; final outs = value.item2.outputs.where((_) => true).toList(growable: false); final ins = value.item2.inputs.where((_) => true).toList(growable: false); txnsData.add(Tuple4(value.item2, outs, ins, transactionAddress)); } await addNewTransactionData(txnsData, walletId); } Future>> fetchAnonymitySets() async { try { final latestSetId = await getLatestSetId(); final List> sets = []; List>> anonFutures = []; for (int i = 1; i <= latestSetId; i++) { final set = cachedElectrumXClient.getAnonymitySet( groupId: "$i", coin: coin, ); anonFutures.add(set); } await Future.wait(anonFutures); for (int i = 1; i <= latestSetId; i++) { Map set = (await anonFutures[i - 1]); set["setId"] = i; sets.add(set); } return sets; } catch (e, s) { Logging.instance.log( "Exception rethrown from refreshAnonymitySets: $e\n$s", level: LogLevel.Error); rethrow; } } Future _createJoinSplitTransaction( int spendAmount, String address, bool subtractFeeFromAmount) async { // final price = await firoPrice; final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic'); final index = firoGetMintIndex(); final lelantusEntry = await _getLelantusEntry(); final anonymitySets = await fetchAnonymitySets(); final locktime = await getBlockHead(electrumXClient); // final locale = // Platform.isWindows ? "en_US" : await Devicelocale.currentLocale; ReceivePort receivePort = await getIsolate({ "function": "createJoinSplit", "spendAmount": spendAmount, "address": address, "subtractFeeFromAmount": subtractFeeFromAmount, "mnemonic": mnemonic, "index": index, // "price": price, "lelantusEntries": lelantusEntry, "locktime": locktime, "coin": coin, "network": _network, "_anonymity_sets": anonymitySets, // "locale": locale, }); var message = await receivePort.first; if (message is String) { Logging.instance .log("Error in CreateJoinSplit: $message", level: LogLevel.Error); stop(receivePort); return 3; } if (message is int) { stop(receivePort); return message; } stop(receivePort); Logging.instance.log('Closing createJoinSplit!', level: LogLevel.Info); return message; } Future getLatestSetId() async { try { final id = await electrumXClient.getLatestCoinId(); return id; } catch (e, s) { Logging.instance.log("Exception rethrown in firo_wallet.dart: $e\n$s", level: LogLevel.Error); rethrow; } } Future> getUsedCoinSerials() async { try { final response = await cachedElectrumXClient.getUsedCoinSerials( coin: coin, ); return response; } catch (e, s) { Logging.instance.log("Exception rethrown in firo_wallet.dart: $e\n$s", level: LogLevel.Error); rethrow; } } @override Future exit() async { _hasCalledExit = true; timer?.cancel(); timer = null; stopNetworkAlivePinging(); for (final isolate in isolates.values) { isolate.kill(priority: Isolate.immediate); } isolates.clear(); Logging.instance .log("$walletName firo_wallet exit finished", level: LogLevel.Info); } bool _hasCalledExit = false; @override bool get hasCalledExit => _hasCalledExit; bool isActive = false; @override void Function(bool)? get onIsActiveWalletChanged => (isActive) async { timer?.cancel(); timer = null; if (isActive) { await compute( _setTestnetWrapper, coin == Coin.firoTestNet, ); } else { await compute( _setTestnetWrapper, false, ); } this.isActive = isActive; }; Future getCoinsToJoinSplit( int required, ) async { List coins = await _getLelantusEntry(); if (required > LELANTUS_VALUE_SPEND_LIMIT_PER_TRANSACTION) { return false; } int availableBalance = coins.fold( 0, (previousValue, element) => previousValue + element.amount); if (required > availableBalance) { return false; } // sort by biggest amount. if it is same amount we will prefer the older block coins.sort((a, b) => (a.amount != b.amount ? a.amount > b.amount : a.height < b.height) ? -1 : 1); int spendVal = 0; List coinsToSpend = []; while (spendVal < required) { if (coins.isEmpty) { break; } DartLelantusEntry? chosen; int need = required - spendVal; var itr = coins.first; if (need >= itr.amount) { chosen = itr; coins.remove(itr); } else { for (int index = coins.length - 1; index != 0; index--) { var coinIt = coins[index]; var nextItr = coins[index - 1]; if (coinIt.amount >= need && (index - 1 == 0 || nextItr.amount != coinIt.amount)) { chosen = coinIt; coins.remove(chosen); break; } } } // TODO: investigate the bug here where chosen is null, conditions, given one mint spendVal += chosen!.amount; coinsToSpend.insert(coinsToSpend.length, chosen); } // sort by group id ay ascending order. it is mandatory for creating proper joinsplit coinsToSpend.sort((a, b) => a.anonymitySetId < b.anonymitySetId ? 1 : -1); int changeToMint = spendVal - required; List indices = []; for (var l in coinsToSpend) { indices.add(l.index); } List coinsToBeSpentOut = []; coinsToBeSpentOut.addAll(coinsToSpend); return {"changeToMint": changeToMint, "coinsToSpend": coinsToBeSpentOut}; } Future estimateJoinSplitFee( int spendAmount, ) async { var lelantusEntry = await _getLelantusEntry(); final balance = availablePrivateBalance(); int spendAmount = (balance * Decimal.fromInt(Constants.satsPerCoin(coin))) .toBigInt() .toInt(); if (spendAmount == 0 || lelantusEntry.isEmpty) { return LelantusFeeData(0, 0, []).fee; } ReceivePort receivePort = await getIsolate({ "function": "estimateJoinSplit", "spendAmount": spendAmount, "subtractFeeFromAmount": true, "lelantusEntries": lelantusEntry, "coin": coin, }); final message = await receivePort.first; if (message is String) { Logging.instance.log("this is a string", level: LogLevel.Error); stop(receivePort); throw Exception("_fetchMaxFee isolate failed"); } stop(receivePort); Logging.instance.log('Closing estimateJoinSplit!', level: LogLevel.Info); return (message as LelantusFeeData).fee; } // int fee; // int size; // // for (fee = 0;;) { // int currentRequired = spendAmount; // // TODO: investigate the bug here // var map = await getCoinsToJoinSplit(currentRequired); // if (map is bool && !map) { // return 0; // } // // List coinsToBeSpent = // map['coinsToSpend'] as List; // // // 1054 is constant part, mainly Schnorr and Range proofs, 2560 is for each sigma/aux data // // 179 other parts of tx, assuming 1 utxo and 1 jmint // size = 1054 + 2560 * coinsToBeSpent.length + 180; // // uint64_t feeNeeded = GetMinimumFee(size, DEFAULT_TX_CONFIRM_TARGET); // int feeNeeded = // size; //TODO(Levon) temporary, use real estimation methods here // // if (fee >= feeNeeded) { // break; // } // // fee = feeNeeded; // } // // return fee; @override Future estimateFeeFor(int satoshiAmount, int feeRate) async { int fee = await estimateJoinSplitFee(satoshiAmount); return fee; } Future estimateFeeForPublic(int satoshiAmount, int feeRate) async { final available = balance.spendable; if (available == satoshiAmount) { return satoshiAmount - (await sweepAllEstimate(feeRate)); } else if (satoshiAmount <= 0 || satoshiAmount > available) { return roughFeeEstimate(1, 2, feeRate); } int runningBalance = 0; int inputCount = 0; for (final output in (await utxos)) { if (!output.isBlocked) { runningBalance += output.value; inputCount++; if (runningBalance > satoshiAmount) { break; } } } final oneOutPutFee = roughFeeEstimate(inputCount, 1, feeRate); final twoOutPutFee = roughFeeEstimate(inputCount, 2, feeRate); if (runningBalance - satoshiAmount > oneOutPutFee) { if (runningBalance - satoshiAmount > oneOutPutFee + DUST_LIMIT) { final change = runningBalance - satoshiAmount - twoOutPutFee; if (change > DUST_LIMIT && runningBalance - satoshiAmount - change == twoOutPutFee) { return runningBalance - satoshiAmount - change; } else { return runningBalance - satoshiAmount; } } else { return runningBalance - satoshiAmount; } } else if (runningBalance - satoshiAmount == oneOutPutFee) { return oneOutPutFee; } else { return twoOutPutFee; } } // TODO: correct formula for firo? int roughFeeEstimate(int inputCount, int outputCount, int feeRatePerKB) { return ((181 * inputCount) + (34 * outputCount) + 10) * (feeRatePerKB / 1000).ceil(); } Future sweepAllEstimate(int feeRate) async { int available = 0; int inputCount = 0; for (final output in (await utxos)) { if (!output.isBlocked && output.isConfirmed(storedChainHeight, MINIMUM_CONFIRMATIONS)) { available += output.value; inputCount++; } } // transaction will only have 1 output minus the fee final estimatedFee = roughFeeEstimate(inputCount, 1, feeRate); return available - estimatedFee; } Future>> fastFetch(List allTxHashes) async { List> allTransactions = []; const futureLimit = 30; List>> transactionFutures = []; int currentFutureCount = 0; for (final txHash in allTxHashes) { Future> transactionFuture = cachedElectrumXClient.getTransaction( txHash: txHash, verbose: true, coin: coin, ); transactionFutures.add(transactionFuture); currentFutureCount++; if (currentFutureCount > futureLimit) { currentFutureCount = 0; await Future.wait(transactionFutures); for (final fTx in transactionFutures) { final tx = await fTx; // delete unused large parts tx.remove("hex"); tx.remove("lelantusData"); allTransactions.add(tx); } } } if (currentFutureCount != 0) { currentFutureCount = 0; await Future.wait(transactionFutures); for (final fTx in transactionFutures) { final tx = await fTx; // delete unused large parts tx.remove("hex"); tx.remove("lelantusData"); allTransactions.add(tx); } } return allTransactions; } Future> getJMintTransactions( CachedElectrumX cachedClient, List transactions, // String currency, Coin coin, // Decimal currentPrice, // String locale, ) async { try { Map txs = {}; List> allTransactions = await fastFetch(transactions); for (int i = 0; i < allTransactions.length; i++) { try { final tx = allTransactions[i]; var sendIndex = 1; if (tx["vout"][0]["value"] != null && Decimal.parse(tx["vout"][0]["value"].toString()) > Decimal.zero) { sendIndex = 0; } tx["amount"] = tx["vout"][sendIndex]["value"]; tx["address"] = tx["vout"][sendIndex]["scriptPubKey"]["addresses"][0]; tx["fees"] = tx["vin"][0]["nFees"]; final txn = isar_models.Transaction( walletId: walletId, txid: tx["txid"] as String, timestamp: tx["time"] as int? ?? (DateTime.now().millisecondsSinceEpoch ~/ 1000), type: isar_models.TransactionType.outgoing, subType: isar_models.TransactionSubType.join, amount: Format.decimalAmountToSatoshis( Decimal.parse(tx["amount"].toString()), coin, ), fee: Format.decimalAmountToSatoshis( Decimal.parse(tx["fees"].toString()), coin, ), height: tx["height"] as int?, isCancelled: false, isLelantus: true, slateId: null, otherData: null, ); final address = await db .getAddresses(walletId) .filter() .valueEqualTo(tx["address"] as String) .findFirst() ?? isar_models.Address( walletId: walletId, value: tx["address"] as String, derivationIndex: -2, type: isar_models.AddressType.nonWallet, subType: isar_models.AddressSubType.unknown, publicKey: [], ); txs[address] = txn; } catch (e, s) { Logging.instance.log( "Exception caught in getJMintTransactions(): $e\n$s", level: LogLevel.Info); rethrow; } } return txs; } catch (e, s) { Logging.instance.log( "Exception rethrown in getJMintTransactions(): $e\n$s", level: LogLevel.Info); rethrow; } } @override Future generateNewAddress() async { try { final currentReceiving = await _currentReceivingAddress; final newReceivingIndex = currentReceiving.derivationIndex + 1; // Use new index to derive a new receiving address final newReceivingAddress = await _generateAddressForChain( 0, newReceivingIndex, ); // Add that new receiving address await db.putAddress(newReceivingAddress); return true; } catch (e, s) { Logging.instance.log( "Exception rethrown from generateNewAddress(): $e\n$s", level: LogLevel.Error); return false; } } Decimal availablePrivateBalance() { return balancePrivate.getSpendable(); } Decimal availablePublicBalance() { return balance.getSpendable(); } Future get chainHeight async { try { final result = await _electrumXClient.getBlockHeadTip(); final height = result["height"] as int; await updateCachedChainHeight(height); return height; } catch (e, s) { Logging.instance.log("Exception caught in chainHeight: $e\n$s", level: LogLevel.Error); return storedChainHeight; } } @override int get storedChainHeight => getCachedChainHeight(); @override Balance get balance => _balance ??= getCachedBalance(); Balance? _balance; Balance get balancePrivate => _balancePrivate ??= getCachedBalanceSecondary(); Balance? _balancePrivate; @override Future> get utxos => db.getUTXOs(walletId).findAll(); @override Future> get transactions => db.getTransactions(walletId).findAll(); }