mirror of
https://github.com/cypherstack/stack_wallet.git
synced 2024-12-25 12:59:24 +00:00
3027 lines
110 KiB
Dart
3027 lines
110 KiB
Dart
// /*
|
|
// * This file is part of Stack Wallet.
|
|
// *
|
|
// * Copyright (c) 2023 Cypher Stack
|
|
// * All Rights Reserved.
|
|
// * The code is distributed under GPLv3 license, see LICENSE file for details.
|
|
// * Generated by Cypher Stack on 2023-05-26
|
|
// *
|
|
// */
|
|
//
|
|
// import 'dart:async';
|
|
// import 'dart:convert';
|
|
// import 'dart:io';
|
|
//
|
|
// import 'package:bech32/bech32.dart';
|
|
// import 'package:bip32/bip32.dart' as bip32;
|
|
// import 'package:bip39/bip39.dart' as bip39;
|
|
// import 'package:bitcoindart/bitcoindart.dart';
|
|
// import 'package:bitcoindart/bitcoindart.dart' as btc_dart;
|
|
// import 'package:bs58check/bs58check.dart' as bs58check;
|
|
// import 'package:flutter/foundation.dart';
|
|
// import 'package:isar/isar.dart';
|
|
// import 'package:stackwallet/db/isar/main_db.dart';
|
|
// import 'package:stackwallet/electrumx_rpc/cached_electrumx.dart';
|
|
// import 'package:stackwallet/electrumx_rpc/electrumx.dart';
|
|
// import 'package:stackwallet/exceptions/electrumx/no_such_transaction.dart';
|
|
// import 'package:stackwallet/models/balance.dart';
|
|
// import 'package:stackwallet/models/isar/models/isar_models.dart' as isar_models;
|
|
// import 'package:stackwallet/models/paymint/fee_object_model.dart';
|
|
// import 'package:stackwallet/models/signing_data.dart';
|
|
// import 'package:stackwallet/services/coins/coin_service.dart';
|
|
// import 'package:stackwallet/services/event_bus/events/global/node_connection_status_changed_event.dart';
|
|
// import 'package:stackwallet/services/event_bus/events/global/refresh_percent_changed_event.dart';
|
|
// import 'package:stackwallet/services/event_bus/events/global/updated_in_background_event.dart';
|
|
// import 'package:stackwallet/services/event_bus/events/global/wallet_sync_status_changed_event.dart';
|
|
// import 'package:stackwallet/services/event_bus/global_event_bus.dart';
|
|
// import 'package:stackwallet/services/mixins/coin_control_interface.dart';
|
|
// import 'package:stackwallet/services/mixins/electrum_x_parsing.dart';
|
|
// import 'package:stackwallet/services/mixins/wallet_cache.dart';
|
|
// import 'package:stackwallet/services/mixins/wallet_db.dart';
|
|
// import 'package:stackwallet/services/mixins/xpubable.dart';
|
|
// import 'package:stackwallet/services/node_service.dart';
|
|
// import 'package:stackwallet/services/transaction_notification_tracker.dart';
|
|
// import 'package:stackwallet/utilities/address_utils.dart';
|
|
// import 'package:stackwallet/utilities/amount/amount.dart';
|
|
// import 'package:stackwallet/utilities/bip32_utils.dart';
|
|
// import 'package:stackwallet/utilities/constants.dart';
|
|
// import 'package:stackwallet/utilities/default_nodes.dart';
|
|
// import 'package:stackwallet/utilities/enums/coin_enum.dart';
|
|
// import 'package:stackwallet/utilities/enums/derive_path_type_enum.dart';
|
|
// import 'package:stackwallet/utilities/enums/fee_rate_type_enum.dart';
|
|
// import 'package:stackwallet/utilities/flutter_secure_storage_interface.dart';
|
|
// import 'package:stackwallet/utilities/format.dart';
|
|
// import 'package:stackwallet/utilities/logger.dart';
|
|
// import 'package:stackwallet/utilities/prefs.dart';
|
|
// import 'package:stackwallet/widgets/crypto_notifications.dart';
|
|
// import 'package:tuple/tuple.dart';
|
|
// import 'package:uuid/uuid.dart';
|
|
//
|
|
// class DogecoinWallet extends CoinServiceAPI
|
|
// with WalletCache, WalletDB, ElectrumXParsing, CoinControlInterface
|
|
// implements XPubAble {
|
|
// DogecoinWallet({
|
|
// required String walletId,
|
|
// required String walletName,
|
|
// required Coin coin,
|
|
// required ElectrumX client,
|
|
// required CachedElectrumX cachedClient,
|
|
// required TransactionNotificationTracker tracker,
|
|
// required SecureStorageInterface secureStore,
|
|
// MainDB? mockableOverride,
|
|
// }) {
|
|
// txTracker = tracker;
|
|
// _walletId = walletId;
|
|
// _walletName = walletName;
|
|
// _coin = coin;
|
|
// _electrumXClient = client;
|
|
// _cachedElectrumXClient = cachedClient;
|
|
// _secureStore = secureStore;
|
|
// initCache(walletId, coin);
|
|
// initWalletDB(mockableOverride: mockableOverride);
|
|
// initCoinControlInterface(
|
|
// walletId: walletId,
|
|
// walletName: walletName,
|
|
// coin: coin,
|
|
// db: db,
|
|
// getChainHeight: () => chainHeight,
|
|
// refreshedBalanceCallback: (balance) async {
|
|
// _balance = balance;
|
|
// await updateCachedBalance(_balance!);
|
|
// },
|
|
// );
|
|
//
|
|
// // paynym stuff
|
|
// // initPaynymWalletInterface(
|
|
// // walletId: walletId,
|
|
// // walletName: walletName,
|
|
// // network: network,
|
|
// // coin: coin,
|
|
// // db: db,
|
|
// // electrumXClient: electrumXClient,
|
|
// // getMnemonic: () => mnemonic,
|
|
// // getChainHeight: () => chainHeight,
|
|
// // getCurrentChangeAddress: () => currentChangeAddress,
|
|
// // estimateTxFee: estimateTxFee,
|
|
// // prepareSend: prepareSend,
|
|
// // getTxCount: getTxCount,
|
|
// // fetchBuildTxData: fetchBuildTxData,
|
|
// // refresh: refresh,
|
|
// // checkChangeAddressForTransactions: checkChangeAddressForTransactions,
|
|
// // addDerivation: addDerivation,
|
|
// // addDerivations: addDerivations,
|
|
// // );
|
|
// }
|
|
//
|
|
// static const integrationTestFlag =
|
|
// bool.fromEnvironment("IS_INTEGRATION_TEST");
|
|
// final _prefs = Prefs.instance;
|
|
//
|
|
// Timer? timer;
|
|
// late final Coin _coin;
|
|
//
|
|
// late final TransactionNotificationTracker txTracker;
|
|
//
|
|
// NetworkType get network {
|
|
// switch (coin) {
|
|
// case Coin.dogecoin:
|
|
// return dogecoin;
|
|
// case Coin.dogecoinTestNet:
|
|
// return dogecointestnet;
|
|
// default:
|
|
// throw Exception("Dogecoin network type not set!");
|
|
// }
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<List<isar_models.UTXO>> get utxos => db.getUTXOs(walletId).findAll();
|
|
//
|
|
// @override
|
|
// Future<List<isar_models.Transaction>> get transactions => db
|
|
// .getTransactions(walletId)
|
|
// .filter()
|
|
// .not()
|
|
// .group((q) => q
|
|
// .subTypeEqualTo(isar_models.TransactionSubType.bip47Notification)
|
|
// .and()
|
|
// .typeEqualTo(isar_models.TransactionType.incoming))
|
|
// .sortByTimestampDesc()
|
|
// .findAll();
|
|
//
|
|
// @override
|
|
// Coin get coin => _coin;
|
|
//
|
|
// @override
|
|
// Future<String> get currentReceivingAddress async =>
|
|
// (await _currentReceivingAddress).value;
|
|
//
|
|
// Future<isar_models.Address> get _currentReceivingAddress async =>
|
|
// (await db
|
|
// .getAddresses(walletId)
|
|
// .filter()
|
|
// .typeEqualTo(isar_models.AddressType.p2pkh)
|
|
// .subTypeEqualTo(isar_models.AddressSubType.receiving)
|
|
// .sortByDerivationIndexDesc()
|
|
// .findFirst()) ??
|
|
// await _generateAddressForChain(0, 0, DerivePathTypeExt.primaryFor(coin));
|
|
//
|
|
// // @override
|
|
// Future<String> get currentChangeAddress async =>
|
|
// (await _currentChangeAddress).value;
|
|
//
|
|
// Future<isar_models.Address> get _currentChangeAddress async =>
|
|
// (await db
|
|
// .getAddresses(walletId)
|
|
// .filter()
|
|
// .typeEqualTo(isar_models.AddressType.p2pkh)
|
|
// .subTypeEqualTo(isar_models.AddressSubType.change)
|
|
// .sortByDerivationIndexDesc()
|
|
// .findFirst()) ??
|
|
// await _generateAddressForChain(1, 0, DerivePathTypeExt.primaryFor(coin));
|
|
//
|
|
// @override
|
|
// Future<void> exit() async {
|
|
// _hasCalledExit = true;
|
|
// timer?.cancel();
|
|
// timer = null;
|
|
// stopNetworkAlivePinging();
|
|
// }
|
|
//
|
|
// bool _hasCalledExit = false;
|
|
//
|
|
// @override
|
|
// bool get hasCalledExit => _hasCalledExit;
|
|
//
|
|
// @override
|
|
// Future<FeeObject> get fees => _feeObject ??= _getFees();
|
|
// Future<FeeObject>? _feeObject;
|
|
//
|
|
// @override
|
|
// Future<int> get maxFee async {
|
|
// throw UnimplementedError("Not used in dogecoin");
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<List<String>> get mnemonic => _getMnemonicList();
|
|
//
|
|
// @override
|
|
// Future<String?> get mnemonicString =>
|
|
// _secureStore.read(key: '${_walletId}_mnemonic');
|
|
//
|
|
// @override
|
|
// Future<String?> get mnemonicPassphrase => _secureStore.read(
|
|
// key: '${_walletId}_mnemonicPassphrase',
|
|
// );
|
|
//
|
|
// Future<int> get chainHeight async {
|
|
// try {
|
|
// final result = await _electrumXClient.getBlockHeadTip();
|
|
// final height = result["height"] as int;
|
|
// await updateCachedChainHeight(height);
|
|
// if (height > storedChainHeight) {
|
|
// GlobalEventBus.instance.fire(
|
|
// UpdatedInBackgroundEvent(
|
|
// "Updated current chain height in $walletId $walletName!",
|
|
// walletId,
|
|
// ),
|
|
// );
|
|
// }
|
|
// return height;
|
|
// } catch (e, s) {
|
|
// Logging.instance.log("Exception caught in chainHeight: $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// return storedChainHeight;
|
|
// }
|
|
// }
|
|
//
|
|
// @override
|
|
// int get storedChainHeight => getCachedChainHeight();
|
|
//
|
|
// DerivePathType addressType({required String address}) {
|
|
// Uint8List? decodeBase58;
|
|
// Segwit? decodeBech32;
|
|
// try {
|
|
// decodeBase58 = bs58check.decode(address);
|
|
// } catch (err) {
|
|
// // Base58check decode fail
|
|
// }
|
|
// if (decodeBase58 != null) {
|
|
// if (decodeBase58[0] == network.pubKeyHash) {
|
|
// // P2PKH
|
|
// return DerivePathType.bip44;
|
|
// }
|
|
// throw ArgumentError('Invalid version or Network mismatch');
|
|
// } else {
|
|
// try {
|
|
// decodeBech32 = segwit.decode(address);
|
|
// } catch (err) {
|
|
// // Bech32 decode fail
|
|
// }
|
|
// if (network.bech32 != decodeBech32!.hrp) {
|
|
// throw ArgumentError('Invalid prefix or Network mismatch');
|
|
// }
|
|
// if (decodeBech32.version != 0) {
|
|
// throw ArgumentError('Invalid address version');
|
|
// }
|
|
// }
|
|
// throw ArgumentError('$address has no matching Script');
|
|
// }
|
|
//
|
|
// bool longMutex = false;
|
|
//
|
|
// @override
|
|
// Future<void> recoverFromMnemonic({
|
|
// required String mnemonic,
|
|
// String? mnemonicPassphrase,
|
|
// required int maxUnusedAddressGap,
|
|
// required int maxNumberOfIndexesToCheck,
|
|
// required int height,
|
|
// }) async {
|
|
// longMutex = true;
|
|
// final start = DateTime.now();
|
|
// try {
|
|
// Logging.instance.log("IS_INTEGRATION_TEST: $integrationTestFlag",
|
|
// level: LogLevel.Info);
|
|
// if (!integrationTestFlag) {
|
|
// final features = await electrumXClient.getServerFeatures();
|
|
// Logging.instance.log("features: $features", level: LogLevel.Info);
|
|
// switch (coin) {
|
|
// case Coin.dogecoin:
|
|
// if (features['genesis_hash'] != GENESIS_HASH_MAINNET) {
|
|
// throw Exception("genesis hash does not match main net!");
|
|
// }
|
|
// break;
|
|
// case Coin.dogecoinTestNet:
|
|
// if (features['genesis_hash'] != GENESIS_HASH_TESTNET) {
|
|
// throw Exception("genesis hash does not match test net!");
|
|
// }
|
|
// break;
|
|
// default:
|
|
// throw Exception(
|
|
// "Attempted to generate a DogecoinWallet using a non dogecoin coin type: ${coin.name}");
|
|
// }
|
|
// }
|
|
// // check to make sure we aren't overwriting a mnemonic
|
|
// // this should never fail
|
|
// if ((await mnemonicString) != null ||
|
|
// (await this.mnemonicPassphrase) != null) {
|
|
// longMutex = false;
|
|
// throw Exception("Attempted to overwrite mnemonic on restore!");
|
|
// }
|
|
// await _secureStore.write(
|
|
// key: '${_walletId}_mnemonic', value: mnemonic.trim());
|
|
// await _secureStore.write(
|
|
// key: '${_walletId}_mnemonicPassphrase',
|
|
// value: mnemonicPassphrase ?? "",
|
|
// );
|
|
// await _recoverWalletFromBIP32SeedPhrase(
|
|
// mnemonic: mnemonic.trim(),
|
|
// mnemonicPassphrase: mnemonicPassphrase ?? "",
|
|
// maxUnusedAddressGap: maxUnusedAddressGap,
|
|
// maxNumberOfIndexesToCheck: maxNumberOfIndexesToCheck,
|
|
// );
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown from recoverFromMnemonic(): $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// longMutex = false;
|
|
// rethrow;
|
|
// }
|
|
// longMutex = false;
|
|
//
|
|
// final end = DateTime.now();
|
|
// Logging.instance.log(
|
|
// "$walletName recovery time: ${end.difference(start).inMilliseconds} millis",
|
|
// level: LogLevel.Info);
|
|
// }
|
|
//
|
|
// Future<Map<String, dynamic>> _checkGaps(
|
|
// int maxNumberOfIndexesToCheck,
|
|
// int maxUnusedAddressGap,
|
|
// int txCountBatchSize,
|
|
// bip32.BIP32 root,
|
|
// DerivePathType type,
|
|
// int chain) async {
|
|
// List<isar_models.Address> addressArray = [];
|
|
// int returningIndex = -1;
|
|
// Map<String, Map<String, String>> derivations = {};
|
|
// int gapCounter = 0;
|
|
// for (int index = 0;
|
|
// index < maxNumberOfIndexesToCheck && gapCounter < maxUnusedAddressGap;
|
|
// index += txCountBatchSize) {
|
|
// List<String> iterationsAddressArray = [];
|
|
// Logging.instance.log(
|
|
// "index: $index, \t GapCounter $chain ${type.name}: $gapCounter",
|
|
// level: LogLevel.Info);
|
|
//
|
|
// final _id = "k_$index";
|
|
// Map<String, String> txCountCallArgs = {};
|
|
// final Map<String, dynamic> receivingNodes = {};
|
|
//
|
|
// for (int j = 0; j < txCountBatchSize; j++) {
|
|
// final derivePath = constructDerivePath(
|
|
// derivePathType: type,
|
|
// networkWIF: root.network.wif,
|
|
// chain: chain,
|
|
// index: index + j,
|
|
// );
|
|
// final node = await Bip32Utils.getBip32NodeFromRoot(root, derivePath);
|
|
//
|
|
// isar_models.Address address;
|
|
// switch (type) {
|
|
// case DerivePathType.bip44:
|
|
// final addressString = P2PKH(
|
|
// data: PaymentData(pubkey: node.publicKey), network: network)
|
|
// .data
|
|
// .address!;
|
|
// address = isar_models.Address(
|
|
// walletId: walletId,
|
|
// value: addressString,
|
|
// publicKey: node.publicKey,
|
|
// type: isar_models.AddressType.p2pkh,
|
|
// derivationIndex: index + j,
|
|
// derivationPath: isar_models.DerivationPath()..value = derivePath,
|
|
// subType: chain == 0
|
|
// ? isar_models.AddressSubType.receiving
|
|
// : isar_models.AddressSubType.change,
|
|
// );
|
|
// break;
|
|
// default:
|
|
// throw Exception("DerivePathType $type not supported");
|
|
// }
|
|
// receivingNodes.addAll({
|
|
// "${_id}_$j": {
|
|
// "node": node,
|
|
// "address": address,
|
|
// }
|
|
// });
|
|
// txCountCallArgs.addAll({
|
|
// "${_id}_$j": address.value,
|
|
// });
|
|
// }
|
|
//
|
|
// // get address tx counts
|
|
// final counts = await _getBatchTxCount(addresses: txCountCallArgs);
|
|
//
|
|
// // check and add appropriate addresses
|
|
// for (int k = 0; k < txCountBatchSize; k++) {
|
|
// int count = counts["${_id}_$k"]!;
|
|
// if (count > 0) {
|
|
// final node = receivingNodes["${_id}_$k"];
|
|
// final address = node["address"] as isar_models.Address;
|
|
// // add address to array
|
|
// addressArray.add(address);
|
|
// iterationsAddressArray.add(address.value);
|
|
// // set current index
|
|
// returningIndex = index + k;
|
|
// // reset counter
|
|
// gapCounter = 0;
|
|
// // add info to derivations
|
|
// derivations[address.value] = {
|
|
// "pubKey": Format.uint8listToString(
|
|
// (node["node"] as bip32.BIP32).publicKey),
|
|
// "wif": (node["node"] as bip32.BIP32).toWIF(),
|
|
// };
|
|
// }
|
|
//
|
|
// // increase counter when no tx history found
|
|
// if (count == 0) {
|
|
// gapCounter++;
|
|
// }
|
|
// }
|
|
// // cache all the transactions while waiting for the current function to finish.
|
|
// unawaited(getTransactionCacheEarly(iterationsAddressArray));
|
|
// }
|
|
// return {
|
|
// "addressArray": addressArray,
|
|
// "index": returningIndex,
|
|
// "derivations": derivations
|
|
// };
|
|
// }
|
|
//
|
|
// Future<void> getTransactionCacheEarly(List<String> allAddresses) async {
|
|
// try {
|
|
// final List<Map<String, dynamic>> allTxHashes =
|
|
// await _fetchHistory(allAddresses);
|
|
// for (final txHash in allTxHashes) {
|
|
// try {
|
|
// unawaited(cachedElectrumXClient.getTransaction(
|
|
// txHash: txHash["tx_hash"] as String,
|
|
// verbose: true,
|
|
// coin: coin,
|
|
// ));
|
|
// } catch (e) {
|
|
// continue;
|
|
// }
|
|
// }
|
|
// } catch (e) {
|
|
// //
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> _recoverWalletFromBIP32SeedPhrase({
|
|
// required String mnemonic,
|
|
// required String mnemonicPassphrase,
|
|
// int maxUnusedAddressGap = 20,
|
|
// int maxNumberOfIndexesToCheck = 1000,
|
|
// bool isRescan = false,
|
|
// }) async {
|
|
// longMutex = true;
|
|
//
|
|
// Map<String, Map<String, String>> p2pkhReceiveDerivations = {};
|
|
// Map<String, Map<String, String>> p2pkhChangeDerivations = {};
|
|
//
|
|
// final root = await Bip32Utils.getBip32Root(
|
|
// mnemonic,
|
|
// mnemonicPassphrase,
|
|
// network,
|
|
// );
|
|
//
|
|
// List<isar_models.Address> p2pkhReceiveAddressArray = [];
|
|
// int p2pkhReceiveIndex = -1;
|
|
//
|
|
// List<isar_models.Address> p2pkhChangeAddressArray = [];
|
|
// int p2pkhChangeIndex = -1;
|
|
//
|
|
// // actual size is 12 due to p2pkh so 12x1
|
|
// const txCountBatchSize = 12;
|
|
//
|
|
// try {
|
|
// // receiving addresses
|
|
// Logging.instance
|
|
// .log("checking receiving addresses...", level: LogLevel.Info);
|
|
// final resultReceive44 = _checkGaps(maxNumberOfIndexesToCheck,
|
|
// maxUnusedAddressGap, txCountBatchSize, root, DerivePathType.bip44, 0);
|
|
//
|
|
// Logging.instance
|
|
// .log("checking change addresses...", level: LogLevel.Info);
|
|
// // change addresses
|
|
// final resultChange44 = _checkGaps(maxNumberOfIndexesToCheck,
|
|
// maxUnusedAddressGap, txCountBatchSize, root, DerivePathType.bip44, 1);
|
|
//
|
|
// await Future.wait([
|
|
// resultReceive44,
|
|
// resultChange44,
|
|
// ]);
|
|
//
|
|
// p2pkhReceiveAddressArray =
|
|
// (await resultReceive44)['addressArray'] as List<isar_models.Address>;
|
|
// p2pkhReceiveIndex = (await resultReceive44)['index'] as int;
|
|
// p2pkhReceiveDerivations = (await resultReceive44)['derivations']
|
|
// as Map<String, Map<String, String>>;
|
|
//
|
|
// p2pkhChangeAddressArray =
|
|
// (await resultChange44)['addressArray'] as List<isar_models.Address>;
|
|
// p2pkhChangeIndex = (await resultChange44)['index'] as int;
|
|
// p2pkhChangeDerivations = (await resultChange44)['derivations']
|
|
// as Map<String, Map<String, String>>;
|
|
//
|
|
// // save the derivations (if any)
|
|
// if (p2pkhReceiveDerivations.isNotEmpty) {
|
|
// await addDerivations(
|
|
// chain: 0,
|
|
// derivePathType: DerivePathType.bip44,
|
|
// derivationsToAdd: p2pkhReceiveDerivations);
|
|
// }
|
|
// if (p2pkhChangeDerivations.isNotEmpty) {
|
|
// await addDerivations(
|
|
// chain: 1,
|
|
// derivePathType: DerivePathType.bip44,
|
|
// derivationsToAdd: p2pkhChangeDerivations);
|
|
// }
|
|
//
|
|
// // If restoring a wallet that never received any funds, then set receivingArray manually
|
|
// // If we didn't do this, it'd store an empty array
|
|
// if (p2pkhReceiveIndex == -1) {
|
|
// final address =
|
|
// await _generateAddressForChain(0, 0, DerivePathType.bip44);
|
|
// p2pkhReceiveAddressArray.add(address);
|
|
// }
|
|
//
|
|
// // If restoring a wallet that never sent any funds with change, then set changeArray
|
|
// // manually. If we didn't do this, it'd store an empty array.
|
|
// if (p2pkhChangeIndex == -1) {
|
|
// final address =
|
|
// await _generateAddressForChain(1, 0, DerivePathType.bip44);
|
|
// p2pkhChangeAddressArray.add(address);
|
|
// }
|
|
// if (isRescan) {
|
|
// await db.updateOrPutAddresses([
|
|
// ...p2pkhReceiveAddressArray,
|
|
// ...p2pkhChangeAddressArray,
|
|
// ]);
|
|
// } else {
|
|
// await db.putAddresses([
|
|
// ...p2pkhReceiveAddressArray,
|
|
// ...p2pkhChangeAddressArray,
|
|
// ]);
|
|
// }
|
|
//
|
|
// // paynym stuff
|
|
// // // generate to ensure notification address is in db before refreshing transactions
|
|
// // await getMyNotificationAddress(DerivePathType.bip44);
|
|
// //
|
|
// // // refresh transactions to pick up any received notification transactions
|
|
// // await _refreshTransactions();
|
|
// //
|
|
// // // restore paynym transactions
|
|
// // await restoreAllHistory(
|
|
// // maxUnusedAddressGap: maxUnusedAddressGap,
|
|
// // maxNumberOfIndexesToCheck: maxNumberOfIndexesToCheck,
|
|
// // );
|
|
//
|
|
// await _updateUTXOs();
|
|
//
|
|
// await Future.wait([
|
|
// updateCachedId(walletId),
|
|
// updateCachedIsFavorite(false),
|
|
// ]);
|
|
//
|
|
// longMutex = false;
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown from _recoverWalletFromBIP32SeedPhrase(): $e\n$s",
|
|
// level: LogLevel.Info);
|
|
//
|
|
// longMutex = false;
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<bool> refreshIfThereIsNewData() async {
|
|
// if (longMutex) return false;
|
|
// if (_hasCalledExit) return false;
|
|
// Logging.instance.log("refreshIfThereIsNewData", level: LogLevel.Info);
|
|
//
|
|
// try {
|
|
// bool needsRefresh = false;
|
|
// Logging.instance.log(
|
|
// "notified unconfirmed transactions: ${txTracker.pendings}",
|
|
// level: LogLevel.Info);
|
|
// Set<String> txnsToCheck = {};
|
|
//
|
|
// for (final String txid in txTracker.pendings) {
|
|
// if (!txTracker.wasNotifiedConfirmed(txid)) {
|
|
// txnsToCheck.add(txid);
|
|
// }
|
|
// }
|
|
//
|
|
// for (String txid in txnsToCheck) {
|
|
// final txn = await electrumXClient.getTransaction(txHash: txid);
|
|
// int confirmations = txn["confirmations"] as int? ?? 0;
|
|
// bool isUnconfirmed = confirmations < MINIMUM_CONFIRMATIONS;
|
|
// if (!isUnconfirmed) {
|
|
// // unconfirmedTxs = {};
|
|
// needsRefresh = true;
|
|
// break;
|
|
// }
|
|
// }
|
|
// if (!needsRefresh) {
|
|
// final allOwnAddresses = await _fetchAllOwnAddresses();
|
|
// List<Map<String, dynamic>> allTxs = await _fetchHistory(
|
|
// allOwnAddresses.map((e) => e.value).toList(growable: false));
|
|
// for (Map<String, dynamic> transaction in allTxs) {
|
|
// final txid = transaction['tx_hash'] as String;
|
|
// if ((await db
|
|
// .getTransactions(walletId)
|
|
// .filter()
|
|
// .txidMatches(txid)
|
|
// .findFirst()) ==
|
|
// null) {
|
|
// Logging.instance.log(
|
|
// " txid not found in address history already ${transaction['tx_hash']}",
|
|
// level: LogLevel.Info);
|
|
// needsRefresh = true;
|
|
// break;
|
|
// }
|
|
// }
|
|
// }
|
|
// return needsRefresh;
|
|
// } on NoSuchTransactionException catch (e) {
|
|
// // TODO: move direct transactions elsewhere
|
|
// await db.isar.writeTxn(() async {
|
|
// await db.isar.transactions.deleteByTxidWalletId(e.txid, walletId);
|
|
// });
|
|
// await txTracker.deleteTransaction(e.txid);
|
|
// return true;
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception caught in refreshIfThereIsNewData: $e\n$s",
|
|
// level: LogLevel.Info);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> getAllTxsToWatch() async {
|
|
// if (_hasCalledExit) return;
|
|
// List<isar_models.Transaction> unconfirmedTxnsToNotifyPending = [];
|
|
// List<isar_models.Transaction> unconfirmedTxnsToNotifyConfirmed = [];
|
|
//
|
|
// final currentChainHeight = await chainHeight;
|
|
//
|
|
// final txCount = await db.getTransactions(walletId).count();
|
|
//
|
|
// const paginateLimit = 50;
|
|
//
|
|
// for (int i = 0; i < txCount; i += paginateLimit) {
|
|
// final transactions = await db
|
|
// .getTransactions(walletId)
|
|
// .offset(i)
|
|
// .limit(paginateLimit)
|
|
// .findAll();
|
|
// for (final tx in transactions) {
|
|
// if (tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) {
|
|
// if (txTracker.wasNotifiedPending(tx.txid) &&
|
|
// !txTracker.wasNotifiedConfirmed(tx.txid)) {
|
|
// unconfirmedTxnsToNotifyConfirmed.add(tx);
|
|
// }
|
|
// } else {
|
|
// if (!txTracker.wasNotifiedPending(tx.txid)) {
|
|
// unconfirmedTxnsToNotifyPending.add(tx);
|
|
// }
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// // notify on new incoming transaction
|
|
// for (final tx in unconfirmedTxnsToNotifyPending) {
|
|
// final confirmations = tx.getConfirmations(currentChainHeight);
|
|
//
|
|
// if (tx.type == isar_models.TransactionType.incoming) {
|
|
// CryptoNotificationsEventBus.instance.fire(
|
|
// CryptoNotificationEvent(
|
|
// title: "Incoming transaction",
|
|
// walletId: walletId,
|
|
// date: DateTime.now(),
|
|
// shouldWatchForUpdates: confirmations < MINIMUM_CONFIRMATIONS,
|
|
// txid: tx.txid,
|
|
// confirmations: confirmations,
|
|
// requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
// walletName: walletName,
|
|
// coin: coin,
|
|
// ),
|
|
// );
|
|
//
|
|
// await txTracker.addNotifiedPending(tx.txid);
|
|
// } else if (tx.type == isar_models.TransactionType.outgoing) {
|
|
// CryptoNotificationsEventBus.instance.fire(
|
|
// CryptoNotificationEvent(
|
|
// title: "Sending transaction",
|
|
// walletId: walletId,
|
|
// date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
|
// shouldWatchForUpdates: confirmations < MINIMUM_CONFIRMATIONS,
|
|
// txid: tx.txid,
|
|
// confirmations: confirmations,
|
|
// requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
// walletName: walletName,
|
|
// coin: coin,
|
|
// ),
|
|
// );
|
|
//
|
|
// await txTracker.addNotifiedPending(tx.txid);
|
|
// }
|
|
// }
|
|
//
|
|
// // notify on confirmed
|
|
// for (final tx in unconfirmedTxnsToNotifyConfirmed) {
|
|
// if (tx.type == isar_models.TransactionType.incoming) {
|
|
// CryptoNotificationsEventBus.instance.fire(
|
|
// CryptoNotificationEvent(
|
|
// title: "Incoming transaction confirmed",
|
|
// walletId: walletId,
|
|
// date: DateTime.now(),
|
|
// shouldWatchForUpdates: false,
|
|
// txid: tx.txid,
|
|
// requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
// walletName: walletName,
|
|
// coin: coin,
|
|
// ),
|
|
// );
|
|
//
|
|
// await txTracker.addNotifiedConfirmed(tx.txid);
|
|
// } else if (tx.type == isar_models.TransactionType.outgoing) {
|
|
// CryptoNotificationsEventBus.instance.fire(
|
|
// CryptoNotificationEvent(
|
|
// title: "Outgoing transaction confirmed",
|
|
// walletId: walletId,
|
|
// date: DateTime.now(),
|
|
// shouldWatchForUpdates: false,
|
|
// txid: tx.txid,
|
|
// requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
// walletName: walletName,
|
|
// coin: coin,
|
|
// ),
|
|
// );
|
|
//
|
|
// await txTracker.addNotifiedConfirmed(tx.txid);
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// bool refreshMutex = false;
|
|
//
|
|
// bool _shouldAutoSync = false;
|
|
//
|
|
// @override
|
|
// bool get shouldAutoSync => _shouldAutoSync;
|
|
//
|
|
// @override
|
|
// set shouldAutoSync(bool shouldAutoSync) {
|
|
// if (_shouldAutoSync != shouldAutoSync) {
|
|
// _shouldAutoSync = shouldAutoSync;
|
|
// if (!shouldAutoSync) {
|
|
// timer?.cancel();
|
|
// timer = null;
|
|
// stopNetworkAlivePinging();
|
|
// } else {
|
|
// startNetworkAlivePinging();
|
|
// refresh();
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// //TODO Show percentages properly/more consistently
|
|
// /// Refreshes display data for the wallet
|
|
// @override
|
|
// Future<void> refresh() async {
|
|
// if (refreshMutex) {
|
|
// Logging.instance.log("$walletId $walletName refreshMutex denied",
|
|
// level: LogLevel.Info);
|
|
// return;
|
|
// } else {
|
|
// refreshMutex = true;
|
|
// }
|
|
//
|
|
// try {
|
|
// GlobalEventBus.instance.fire(
|
|
// WalletSyncStatusChangedEvent(
|
|
// WalletSyncStatus.syncing,
|
|
// walletId,
|
|
// coin,
|
|
// ),
|
|
// );
|
|
//
|
|
// GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.0, walletId));
|
|
//
|
|
// GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.1, walletId));
|
|
//
|
|
// final currentHeight = await chainHeight;
|
|
// const storedHeight = 1; //await storedChainHeight;
|
|
//
|
|
// Logging.instance
|
|
// .log("chain height: $currentHeight", level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log("cached height: $storedHeight", level: LogLevel.Info);
|
|
//
|
|
// if (currentHeight != storedHeight) {
|
|
// GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.2, walletId));
|
|
// await checkChangeAddressForTransactions();
|
|
//
|
|
// GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.3, walletId));
|
|
// await _checkCurrentReceivingAddressesForTransactions();
|
|
//
|
|
// // paynym stuff
|
|
// // await checkAllCurrentReceivingPaynymAddressesForTransactions();
|
|
//
|
|
// final fetchFuture = _refreshTransactions();
|
|
// final utxosRefreshFuture = _updateUTXOs();
|
|
// GlobalEventBus.instance
|
|
// .fire(RefreshPercentChangedEvent(0.50, walletId));
|
|
//
|
|
// final feeObj = _getFees();
|
|
// GlobalEventBus.instance
|
|
// .fire(RefreshPercentChangedEvent(0.60, walletId));
|
|
//
|
|
// GlobalEventBus.instance
|
|
// .fire(RefreshPercentChangedEvent(0.70, walletId));
|
|
// _feeObject = Future(() => feeObj);
|
|
//
|
|
// await utxosRefreshFuture;
|
|
// GlobalEventBus.instance
|
|
// .fire(RefreshPercentChangedEvent(0.80, walletId));
|
|
//
|
|
// await fetchFuture;
|
|
// await getAllTxsToWatch();
|
|
// GlobalEventBus.instance
|
|
// .fire(RefreshPercentChangedEvent(0.90, walletId));
|
|
// }
|
|
//
|
|
// GlobalEventBus.instance.fire(RefreshPercentChangedEvent(1.0, walletId));
|
|
// GlobalEventBus.instance.fire(
|
|
// WalletSyncStatusChangedEvent(
|
|
// WalletSyncStatus.synced,
|
|
// walletId,
|
|
// coin,
|
|
// ),
|
|
// );
|
|
// refreshMutex = false;
|
|
//
|
|
// if (shouldAutoSync) {
|
|
// timer ??= Timer.periodic(const Duration(seconds: 30), (timer) async {
|
|
// // chain height check currently broken
|
|
// // if ((await chainHeight) != (await storedChainHeight)) {
|
|
// if (await refreshIfThereIsNewData()) {
|
|
// await refresh();
|
|
// GlobalEventBus.instance.fire(UpdatedInBackgroundEvent(
|
|
// "New data found in $walletId $walletName in background!",
|
|
// walletId));
|
|
// }
|
|
// // }
|
|
// });
|
|
// }
|
|
// } catch (error, strace) {
|
|
// refreshMutex = false;
|
|
// GlobalEventBus.instance.fire(
|
|
// NodeConnectionStatusChangedEvent(
|
|
// NodeConnectionStatus.disconnected,
|
|
// walletId,
|
|
// coin,
|
|
// ),
|
|
// );
|
|
// GlobalEventBus.instance.fire(
|
|
// WalletSyncStatusChangedEvent(
|
|
// WalletSyncStatus.unableToSync,
|
|
// walletId,
|
|
// coin,
|
|
// ),
|
|
// );
|
|
// Logging.instance.log(
|
|
// "Caught exception in refreshWalletData(): $error\n$strace",
|
|
// level: LogLevel.Error);
|
|
// }
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<Map<String, dynamic>> prepareSend({
|
|
// required String address,
|
|
// required Amount amount,
|
|
// Map<String, dynamic>? args,
|
|
// }) async {
|
|
// try {
|
|
// final feeRateType = args?["feeRate"];
|
|
// final customSatsPerVByte = args?["satsPerVByte"] as int?;
|
|
// final feeRateAmount = args?["feeRateAmount"];
|
|
// final utxos = args?["UTXOs"] as Set<isar_models.UTXO>?;
|
|
//
|
|
// if (customSatsPerVByte != null) {
|
|
// // check for send all
|
|
// bool isSendAll = false;
|
|
// if (amount == balance.spendable) {
|
|
// isSendAll = true;
|
|
// }
|
|
//
|
|
// final bool coinControl = utxos != null;
|
|
//
|
|
// final result = await coinSelection(
|
|
// satoshiAmountToSend: amount.raw.toInt(),
|
|
// selectedTxFeeRate: -1,
|
|
// satsPerVByte: customSatsPerVByte,
|
|
// recipientAddress: address,
|
|
// isSendAll: isSendAll,
|
|
// utxos: utxos?.toList(),
|
|
// coinControl: coinControl,
|
|
// );
|
|
//
|
|
// Logging.instance
|
|
// .log("PREPARE SEND RESULT: $result", level: LogLevel.Info);
|
|
// if (result is int) {
|
|
// switch (result) {
|
|
// case 1:
|
|
// throw Exception("Insufficient balance!");
|
|
// case 2:
|
|
// throw Exception("Insufficient funds to pay for transaction fee!");
|
|
// default:
|
|
// throw Exception("Transaction failed with error code $result");
|
|
// }
|
|
// } else {
|
|
// final hex = result["hex"];
|
|
// if (hex is String) {
|
|
// final fee = result["fee"] as int;
|
|
// final vSize = result["vSize"] as int;
|
|
//
|
|
// Logging.instance.log("txHex: $hex", level: LogLevel.Info);
|
|
// Logging.instance.log("fee: $fee", level: LogLevel.Info);
|
|
// Logging.instance.log("vsize: $vSize", level: LogLevel.Info);
|
|
// // fee should never be less than vSize sanity check
|
|
// if (fee < vSize) {
|
|
// throw Exception(
|
|
// "Error in fee calculation: Transaction fee cannot be less than vSize");
|
|
// }
|
|
// return result as Map<String, dynamic>;
|
|
// } else {
|
|
// throw Exception("sent hex is not a String!!!");
|
|
// }
|
|
// }
|
|
// } else if (feeRateType is FeeRateType || feeRateAmount is int) {
|
|
// late final int rate;
|
|
// if (feeRateType is FeeRateType) {
|
|
// int fee = 0;
|
|
// final feeObject = await fees;
|
|
// switch (feeRateType) {
|
|
// case FeeRateType.fast:
|
|
// fee = feeObject.fast;
|
|
// break;
|
|
// case FeeRateType.average:
|
|
// fee = feeObject.medium;
|
|
// break;
|
|
// case FeeRateType.slow:
|
|
// fee = feeObject.slow;
|
|
// break;
|
|
// default:
|
|
// throw ArgumentError("Invalid use of custom fee");
|
|
// }
|
|
// rate = fee;
|
|
// } else {
|
|
// rate = feeRateAmount as int;
|
|
// }
|
|
// // check for send all
|
|
// bool isSendAll = false;
|
|
// if (amount == balance.spendable) {
|
|
// isSendAll = true;
|
|
// }
|
|
//
|
|
// final bool coinControl = utxos != null;
|
|
//
|
|
// final result = await coinSelection(
|
|
// satoshiAmountToSend: amount.raw.toInt(),
|
|
// selectedTxFeeRate: rate,
|
|
// recipientAddress: address,
|
|
// isSendAll: isSendAll,
|
|
// utxos: utxos?.toList(),
|
|
// coinControl: coinControl,
|
|
// );
|
|
//
|
|
// Logging.instance
|
|
// .log("PREPARE SEND RESULT: $result", level: LogLevel.Info);
|
|
// if (result is int) {
|
|
// switch (result) {
|
|
// case 1:
|
|
// throw Exception("Insufficient balance!");
|
|
// case 2:
|
|
// throw Exception("Insufficient funds to pay for transaction fee!");
|
|
// default:
|
|
// throw Exception("Transaction failed with error code $result");
|
|
// }
|
|
// } else {
|
|
// final hex = result["hex"];
|
|
// if (hex is String) {
|
|
// final fee = result["fee"] as int;
|
|
// final vSize = result["vSize"] as int;
|
|
//
|
|
// Logging.instance.log("txHex: $hex", level: LogLevel.Info);
|
|
// Logging.instance.log("fee: $fee", level: LogLevel.Info);
|
|
// Logging.instance.log("vsize: $vSize", level: LogLevel.Info);
|
|
// // fee should never be less than vSize sanity check
|
|
// if (fee < vSize) {
|
|
// throw Exception(
|
|
// "Error in fee calculation: Transaction fee cannot be less than vSize");
|
|
// }
|
|
// return result as Map<String, dynamic>;
|
|
// } else {
|
|
// throw Exception("sent hex is not a String!!!");
|
|
// }
|
|
// }
|
|
// } else {
|
|
// throw ArgumentError("Invalid fee rate argument provided!");
|
|
// }
|
|
// } catch (e, s) {
|
|
// Logging.instance.log("Exception rethrown from prepareSend(): $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<String> confirmSend({dynamic txData}) async {
|
|
// try {
|
|
// Logging.instance.log("confirmSend txData: $txData", level: LogLevel.Info);
|
|
// final txHash = await _electrumXClient.broadcastTransaction(
|
|
// rawTx: txData["hex"] as String);
|
|
// Logging.instance.log("Sent txHash: $txHash", level: LogLevel.Info);
|
|
//
|
|
// final utxos = txData["usedUTXOs"] as List<isar_models.UTXO>;
|
|
//
|
|
// // mark utxos as used
|
|
// await db.putUTXOs(utxos.map((e) => e.copyWith(used: true)).toList());
|
|
//
|
|
// return txHash;
|
|
// } catch (e, s) {
|
|
// Logging.instance.log("Exception rethrown from confirmSend(): $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<bool> testNetworkConnection() async {
|
|
// try {
|
|
// final result = await _electrumXClient.ping();
|
|
// return result;
|
|
// } catch (_) {
|
|
// return false;
|
|
// }
|
|
// }
|
|
//
|
|
// Timer? _networkAliveTimer;
|
|
//
|
|
// void startNetworkAlivePinging() {
|
|
// // call once on start right away
|
|
// _periodicPingCheck();
|
|
//
|
|
// // then periodically check
|
|
// _networkAliveTimer = Timer.periodic(
|
|
// Constants.networkAliveTimerDuration,
|
|
// (_) async {
|
|
// _periodicPingCheck();
|
|
// },
|
|
// );
|
|
// }
|
|
//
|
|
// void _periodicPingCheck() async {
|
|
// bool hasNetwork = await testNetworkConnection();
|
|
//
|
|
// if (_isConnected != hasNetwork) {
|
|
// NodeConnectionStatus status = hasNetwork
|
|
// ? NodeConnectionStatus.connected
|
|
// : NodeConnectionStatus.disconnected;
|
|
// GlobalEventBus.instance
|
|
// .fire(NodeConnectionStatusChangedEvent(status, walletId, coin));
|
|
//
|
|
// _isConnected = hasNetwork;
|
|
// if (hasNetwork) {
|
|
// unawaited(refresh());
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// void stopNetworkAlivePinging() {
|
|
// _networkAliveTimer?.cancel();
|
|
// _networkAliveTimer = null;
|
|
// }
|
|
//
|
|
// bool _isConnected = false;
|
|
//
|
|
// @override
|
|
// bool get isConnected => _isConnected;
|
|
//
|
|
// @override
|
|
// Future<void> initializeNew(
|
|
// ({String mnemonicPassphrase, int wordCount})? data,
|
|
// ) async {
|
|
// Logging.instance
|
|
// .log("Generating new ${coin.prettyName} wallet.", level: LogLevel.Info);
|
|
//
|
|
// if (getCachedId() != null) {
|
|
// throw Exception(
|
|
// "Attempted to initialize a new wallet using an existing wallet ID!");
|
|
// }
|
|
// await _prefs.init();
|
|
// try {
|
|
// await _generateNewWallet(data);
|
|
// } catch (e, s) {
|
|
// Logging.instance.log("Exception rethrown from initializeNew(): $e\n$s",
|
|
// level: LogLevel.Fatal);
|
|
// rethrow;
|
|
// }
|
|
//
|
|
// await Future.wait([
|
|
// updateCachedId(walletId),
|
|
// updateCachedIsFavorite(false),
|
|
// ]);
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<void> initializeExisting() async {
|
|
// Logging.instance.log("initializeExisting() ${coin.prettyName} wallet.",
|
|
// level: LogLevel.Info);
|
|
//
|
|
// if (getCachedId() == null) {
|
|
// throw Exception(
|
|
// "Attempted to initialize an existing wallet using an unknown wallet ID!");
|
|
// }
|
|
//
|
|
// await _prefs.init();
|
|
// // await _checkCurrentChangeAddressesForTransactions();
|
|
// // await _checkCurrentReceivingAddressesForTransactions();
|
|
// }
|
|
//
|
|
// // hack to add tx to txData before refresh completes
|
|
// // required based on current app architecture where we don't properly store
|
|
// // transactions locally in a good way
|
|
// @override
|
|
// Future<void> updateSentCachedTxData(Map<String, dynamic> txData) async {
|
|
// final transaction = isar_models.Transaction(
|
|
// walletId: walletId,
|
|
// txid: txData["txid"] as String,
|
|
// timestamp: DateTime.now().millisecondsSinceEpoch ~/ 1000,
|
|
// type: isar_models.TransactionType.outgoing,
|
|
// subType: isar_models.TransactionSubType.none,
|
|
// // precision may be lost here hence the following amountString
|
|
// amount: (txData["recipientAmt"] as Amount).raw.toInt(),
|
|
// amountString: (txData["recipientAmt"] as Amount).toJsonString(),
|
|
// fee: txData["fee"] as int,
|
|
// height: null,
|
|
// isCancelled: false,
|
|
// isLelantus: false,
|
|
// otherData: null,
|
|
// slateId: null,
|
|
// nonce: null,
|
|
// inputs: [],
|
|
// outputs: [],
|
|
// numberOfMessages: null,
|
|
// );
|
|
//
|
|
// final address = txData["address"] is String
|
|
// ? await db.getAddress(walletId, txData["address"] as String)
|
|
// : null;
|
|
//
|
|
// await db.addNewTransactionData(
|
|
// [
|
|
// Tuple2(transaction, address),
|
|
// ],
|
|
// walletId,
|
|
// );
|
|
// }
|
|
//
|
|
// @override
|
|
// bool validateAddress(String address) {
|
|
// return btc_dart.Address.validateAddress(address, network);
|
|
// }
|
|
//
|
|
// @override
|
|
// String get walletId => _walletId;
|
|
// late final String _walletId;
|
|
//
|
|
// @override
|
|
// String get walletName => _walletName;
|
|
// late String _walletName;
|
|
//
|
|
// // setter for updating on rename
|
|
// @override
|
|
// set walletName(String newName) => _walletName = newName;
|
|
//
|
|
// late ElectrumX _electrumXClient;
|
|
//
|
|
// ElectrumX get electrumXClient => _electrumXClient;
|
|
//
|
|
// late CachedElectrumX _cachedElectrumXClient;
|
|
//
|
|
// CachedElectrumX get cachedElectrumXClient => _cachedElectrumXClient;
|
|
//
|
|
// late SecureStorageInterface _secureStore;
|
|
//
|
|
// @override
|
|
// Future<void> updateNode(bool shouldRefresh) async {
|
|
// final failovers = NodeService(secureStorageInterface: _secureStore)
|
|
// .failoverNodesFor(coin: coin)
|
|
// .map((e) => ElectrumXNode(
|
|
// address: e.host,
|
|
// port: e.port,
|
|
// name: e.name,
|
|
// id: e.id,
|
|
// useSSL: e.useSSL,
|
|
// ))
|
|
// .toList();
|
|
// final newNode = await getCurrentNode();
|
|
// _electrumXClient = ElectrumX.from(
|
|
// node: newNode,
|
|
// prefs: _prefs,
|
|
// failovers: failovers,
|
|
// );
|
|
// _cachedElectrumXClient = CachedElectrumX.from(
|
|
// electrumXClient: _electrumXClient,
|
|
// );
|
|
//
|
|
// if (shouldRefresh) {
|
|
// unawaited(refresh());
|
|
// }
|
|
// }
|
|
//
|
|
// Future<List<String>> _getMnemonicList() async {
|
|
// final _mnemonicString = await mnemonicString;
|
|
// if (_mnemonicString == null) {
|
|
// return [];
|
|
// }
|
|
// final List<String> data = _mnemonicString.split(' ');
|
|
// return data;
|
|
// }
|
|
//
|
|
// Future<ElectrumXNode> getCurrentNode() async {
|
|
// final node = NodeService(secureStorageInterface: _secureStore)
|
|
// .getPrimaryNodeFor(coin: coin) ??
|
|
// DefaultNodes.getNodeFor(coin);
|
|
//
|
|
// return ElectrumXNode(
|
|
// address: node.host,
|
|
// port: node.port,
|
|
// name: node.name,
|
|
// useSSL: node.useSSL,
|
|
// id: node.id,
|
|
// );
|
|
// }
|
|
//
|
|
// Future<List<isar_models.Address>> _fetchAllOwnAddresses() async {
|
|
// final allAddresses = await db
|
|
// .getAddresses(walletId)
|
|
// .filter()
|
|
// .not()
|
|
// .typeEqualTo(isar_models.AddressType.nonWallet)
|
|
// .and()
|
|
// .not()
|
|
// .subTypeEqualTo(isar_models.AddressSubType.nonWallet)
|
|
// .findAll();
|
|
// return allAddresses;
|
|
// }
|
|
//
|
|
// Future<FeeObject> _getFees() async {
|
|
// try {
|
|
// //TODO adjust numbers for different speeds?
|
|
// const int f = 1, m = 5, s = 20;
|
|
//
|
|
// final fast = await electrumXClient.estimateFee(blocks: f);
|
|
// final medium = await electrumXClient.estimateFee(blocks: m);
|
|
// final slow = await electrumXClient.estimateFee(blocks: s);
|
|
//
|
|
// final feeObject = FeeObject(
|
|
// numberOfBlocksFast: f,
|
|
// numberOfBlocksAverage: m,
|
|
// numberOfBlocksSlow: s,
|
|
// fast: Amount.fromDecimal(
|
|
// fast,
|
|
// fractionDigits: coin.decimals,
|
|
// ).raw.toInt(),
|
|
// medium: Amount.fromDecimal(
|
|
// medium,
|
|
// fractionDigits: coin.decimals,
|
|
// ).raw.toInt(),
|
|
// slow: Amount.fromDecimal(
|
|
// slow,
|
|
// fractionDigits: coin.decimals,
|
|
// ).raw.toInt(),
|
|
// );
|
|
//
|
|
// Logging.instance.log("fetched fees: $feeObject", level: LogLevel.Info);
|
|
// return feeObject;
|
|
// } catch (e) {
|
|
// Logging.instance
|
|
// .log("Exception rethrown from _getFees(): $e", level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> _generateNewWallet(
|
|
// ({String mnemonicPassphrase, int wordCount})? data,
|
|
// ) async {
|
|
// Logging.instance
|
|
// .log("IS_INTEGRATION_TEST: $integrationTestFlag", level: LogLevel.Info);
|
|
// if (!integrationTestFlag) {
|
|
// try {
|
|
// final features = await electrumXClient
|
|
// .getServerFeatures()
|
|
// .timeout(const Duration(seconds: 3));
|
|
// Logging.instance.log("features: $features", level: LogLevel.Info);
|
|
// switch (coin) {
|
|
// case Coin.dogecoin:
|
|
// if (features['genesis_hash'] != GENESIS_HASH_MAINNET) {
|
|
// throw Exception("genesis hash does not match main net!");
|
|
// }
|
|
// break;
|
|
// case Coin.dogecoinTestNet:
|
|
// if (features['genesis_hash'] != GENESIS_HASH_TESTNET) {
|
|
// throw Exception("genesis hash does not match test net!");
|
|
// }
|
|
// break;
|
|
// default:
|
|
// throw Exception(
|
|
// "Attempted to generate a BitcoinWallet using a non bitcoin coin type: ${coin.name}");
|
|
// }
|
|
// } catch (e, s) {
|
|
// Logging.instance.log("$e/n$s", level: LogLevel.Info);
|
|
// }
|
|
// }
|
|
//
|
|
// // this should never fail
|
|
// if ((await mnemonicString) != null || (await mnemonicPassphrase) != null) {
|
|
// throw Exception(
|
|
// "Attempted to overwrite mnemonic on generate new wallet!");
|
|
// }
|
|
// final int strength;
|
|
// if (data == null || data.wordCount == 12) {
|
|
// strength = 128;
|
|
// } else if (data.wordCount == 24) {
|
|
// strength = 256;
|
|
// } else {
|
|
// throw Exception("Invalid word count");
|
|
// }
|
|
// await _secureStore.write(
|
|
// key: '${_walletId}_mnemonic',
|
|
// value: bip39.generateMnemonic(strength: strength));
|
|
// await _secureStore.write(
|
|
// key: '${_walletId}_mnemonicPassphrase',
|
|
// value: data?.mnemonicPassphrase ?? "",
|
|
// );
|
|
//
|
|
// // Generate and add addresses
|
|
// final initialReceivingAddressP2PKH =
|
|
// await _generateAddressForChain(0, 0, DerivePathType.bip44);
|
|
// final initialChangeAddressP2PKH =
|
|
// await _generateAddressForChain(1, 0, DerivePathType.bip44);
|
|
//
|
|
// await db.putAddresses([
|
|
// initialReceivingAddressP2PKH,
|
|
// initialChangeAddressP2PKH,
|
|
// ]);
|
|
//
|
|
// Logging.instance.log("_generateNewWalletFinished", level: LogLevel.Info);
|
|
// }
|
|
//
|
|
// /// Generates a new internal or external chain address for the wallet using a BIP44 derivation path.
|
|
// /// [chain] - Use 0 for receiving (external), 1 for change (internal). Should not be any other value!
|
|
// /// [index] - This can be any integer >= 0
|
|
// Future<isar_models.Address> _generateAddressForChain(
|
|
// int chain,
|
|
// int index,
|
|
// DerivePathType derivePathType,
|
|
// ) async {
|
|
// final _mnemonic = await mnemonicString;
|
|
// final _mnemonicPassphrase = await mnemonicPassphrase;
|
|
// if (_mnemonicPassphrase == null) {
|
|
// Logging.instance.log(
|
|
// "Exception in _generateAddressForChain: mnemonic passphrase null, possible migration issue; if using internal builds, delete wallet and restore from seed, if using a release build, please file bug report",
|
|
// level: LogLevel.Error);
|
|
// }
|
|
//
|
|
// final derivePath = constructDerivePath(
|
|
// derivePathType: derivePathType,
|
|
// networkWIF: network.wif,
|
|
// chain: chain,
|
|
// index: index,
|
|
// );
|
|
// final node = await Bip32Utils.getBip32Node(
|
|
// _mnemonic!,
|
|
// _mnemonicPassphrase!,
|
|
// network,
|
|
// derivePath,
|
|
// );
|
|
//
|
|
// final data = PaymentData(pubkey: node.publicKey);
|
|
// String address;
|
|
//
|
|
// switch (derivePathType) {
|
|
// case DerivePathType.bip44:
|
|
// address = P2PKH(data: data, network: network).data.address!;
|
|
// break;
|
|
// default:
|
|
// throw Exception("Unsupported DerivePathType");
|
|
// }
|
|
//
|
|
// // add generated address & info to derivations
|
|
// await addDerivation(
|
|
// chain: chain,
|
|
// address: address,
|
|
// pubKey: Format.uint8listToString(node.publicKey),
|
|
// wif: node.toWIF(),
|
|
// derivePathType: derivePathType,
|
|
// );
|
|
// return isar_models.Address(
|
|
// walletId: walletId,
|
|
// value: address,
|
|
// publicKey: node.publicKey,
|
|
// type: isar_models.AddressType.p2pkh,
|
|
// derivationIndex: index,
|
|
// derivationPath: isar_models.DerivationPath()..value = derivePath,
|
|
// subType: chain == 0
|
|
// ? isar_models.AddressSubType.receiving
|
|
// : isar_models.AddressSubType.change,
|
|
// );
|
|
// }
|
|
//
|
|
// /// Returns the latest receiving/change (external/internal) address for the wallet depending on [chain]
|
|
// /// and
|
|
// /// [chain] - Use 0 for receiving (external), 1 for change (internal). Should not be any other value!
|
|
// Future<String> _getCurrentAddressForChain(
|
|
// int chain,
|
|
// DerivePathType derivePathType,
|
|
// ) async {
|
|
// final subType = chain == 0 // Here, we assume that chain == 1 if it isn't 0
|
|
// ? isar_models.AddressSubType.receiving
|
|
// : isar_models.AddressSubType.change;
|
|
//
|
|
// isar_models.Address? address;
|
|
// switch (derivePathType) {
|
|
// case DerivePathType.bip44:
|
|
// address = await db
|
|
// .getAddresses(walletId)
|
|
// .filter()
|
|
// .typeEqualTo(isar_models.AddressType.p2pkh)
|
|
// .subTypeEqualTo(subType)
|
|
// .sortByDerivationIndexDesc()
|
|
// .findFirst();
|
|
// break;
|
|
// default:
|
|
// throw Exception("Unsupported DerivePathType");
|
|
// }
|
|
// return address!.value;
|
|
// }
|
|
//
|
|
// String _buildDerivationStorageKey(
|
|
// {required int chain, required DerivePathType derivePathType}) {
|
|
// String key;
|
|
// String chainId = chain == 0 ? "receive" : "change";
|
|
// switch (derivePathType) {
|
|
// case DerivePathType.bip44:
|
|
// key = "${walletId}_${chainId}DerivationsP2PKH";
|
|
// break;
|
|
// default:
|
|
// throw Exception("Unsupported DerivePathType");
|
|
// }
|
|
// return key;
|
|
// }
|
|
//
|
|
// Future<Map<String, dynamic>> _fetchDerivations(
|
|
// {required int chain, required DerivePathType derivePathType}) async {
|
|
// // build lookup key
|
|
// final key = _buildDerivationStorageKey(
|
|
// chain: chain, derivePathType: derivePathType);
|
|
//
|
|
// // fetch current derivations
|
|
// final derivationsString = await _secureStore.read(key: key);
|
|
// return Map<String, dynamic>.from(
|
|
// jsonDecode(derivationsString ?? "{}") as Map);
|
|
// }
|
|
//
|
|
// /// Add a single derivation to the local secure storage for [chain] and
|
|
// /// [derivePathType] where [chain] must either be 1 for change or 0 for receive.
|
|
// /// This will overwrite a previous entry where the address of the new derivation
|
|
// /// matches a derivation currently stored.
|
|
// Future<void> addDerivation({
|
|
// required int chain,
|
|
// required String address,
|
|
// required String pubKey,
|
|
// required String wif,
|
|
// required DerivePathType derivePathType,
|
|
// }) async {
|
|
// // build lookup key
|
|
// final key = _buildDerivationStorageKey(
|
|
// chain: chain, derivePathType: derivePathType);
|
|
//
|
|
// // fetch current derivations
|
|
// final derivationsString = await _secureStore.read(key: key);
|
|
// final derivations =
|
|
// Map<String, dynamic>.from(jsonDecode(derivationsString ?? "{}") as Map);
|
|
//
|
|
// // add derivation
|
|
// derivations[address] = {
|
|
// "pubKey": pubKey,
|
|
// "wif": wif,
|
|
// };
|
|
//
|
|
// // save derivations
|
|
// final newReceiveDerivationsString = jsonEncode(derivations);
|
|
// await _secureStore.write(key: key, value: newReceiveDerivationsString);
|
|
// }
|
|
//
|
|
// /// Add multiple derivations to the local secure storage for [chain] and
|
|
// /// [derivePathType] where [chain] must either be 1 for change or 0 for receive.
|
|
// /// This will overwrite any previous entries where the address of the new derivation
|
|
// /// matches a derivation currently stored.
|
|
// /// The [derivationsToAdd] must be in the format of:
|
|
// /// {
|
|
// /// addressA : {
|
|
// /// "pubKey": <the pubKey string>,
|
|
// /// "wif": <the wif string>,
|
|
// /// },
|
|
// /// addressB : {
|
|
// /// "pubKey": <the pubKey string>,
|
|
// /// "wif": <the wif string>,
|
|
// /// },
|
|
// /// }
|
|
// Future<void> addDerivations({
|
|
// required int chain,
|
|
// required DerivePathType derivePathType,
|
|
// required Map<String, dynamic> derivationsToAdd,
|
|
// }) async {
|
|
// // build lookup key
|
|
// final key = _buildDerivationStorageKey(
|
|
// chain: chain, derivePathType: derivePathType);
|
|
//
|
|
// // fetch current derivations
|
|
// final derivationsString = await _secureStore.read(key: key);
|
|
// final derivations =
|
|
// Map<String, dynamic>.from(jsonDecode(derivationsString ?? "{}") as Map);
|
|
//
|
|
// // add derivation
|
|
// derivations.addAll(derivationsToAdd);
|
|
//
|
|
// // save derivations
|
|
// final newReceiveDerivationsString = jsonEncode(derivations);
|
|
// await _secureStore.write(key: key, value: newReceiveDerivationsString);
|
|
// }
|
|
//
|
|
// Future<List<Map<String, dynamic>>> fastFetch(List<String> allTxHashes) async {
|
|
// List<Map<String, dynamic>> allTransactions = [];
|
|
//
|
|
// const futureLimit = 30;
|
|
// List<Future<Map<String, dynamic>>> transactionFutures = [];
|
|
// int currentFutureCount = 0;
|
|
// for (final txHash in allTxHashes) {
|
|
// Future<Map<String, dynamic>> transactionFuture =
|
|
// cachedElectrumXClient.getTransaction(
|
|
// txHash: txHash,
|
|
// verbose: true,
|
|
// coin: coin,
|
|
// );
|
|
// transactionFutures.add(transactionFuture);
|
|
// currentFutureCount++;
|
|
// if (currentFutureCount > futureLimit) {
|
|
// currentFutureCount = 0;
|
|
// await Future.wait(transactionFutures);
|
|
// for (final fTx in transactionFutures) {
|
|
// final tx = await fTx;
|
|
//
|
|
// allTransactions.add(tx);
|
|
// }
|
|
// }
|
|
// }
|
|
// if (currentFutureCount != 0) {
|
|
// currentFutureCount = 0;
|
|
// await Future.wait(transactionFutures);
|
|
// for (final fTx in transactionFutures) {
|
|
// final tx = await fTx;
|
|
//
|
|
// allTransactions.add(tx);
|
|
// }
|
|
// }
|
|
// return allTransactions;
|
|
// }
|
|
//
|
|
// Future<void> _updateUTXOs() async {
|
|
// final allAddresses = await _fetchAllOwnAddresses();
|
|
//
|
|
// try {
|
|
// final fetchedUtxoList = <List<Map<String, dynamic>>>[];
|
|
//
|
|
// final Map<int, Map<String, List<dynamic>>> batches = {};
|
|
// const batchSizeMax = 100;
|
|
// int batchNumber = 0;
|
|
// for (int i = 0; i < allAddresses.length; i++) {
|
|
// if (batches[batchNumber] == null) {
|
|
// batches[batchNumber] = {};
|
|
// }
|
|
// final scripthash =
|
|
// AddressUtils.convertToScriptHash(allAddresses[i].value, network);
|
|
// batches[batchNumber]!.addAll({
|
|
// scripthash: [scripthash]
|
|
// });
|
|
// if (i % batchSizeMax == batchSizeMax - 1) {
|
|
// batchNumber++;
|
|
// }
|
|
// }
|
|
//
|
|
// for (int i = 0; i < batches.length; i++) {
|
|
// final response =
|
|
// await _electrumXClient.getBatchUTXOs(args: batches[i]!);
|
|
// for (final entry in response.entries) {
|
|
// if (entry.value.isNotEmpty) {
|
|
// fetchedUtxoList.add(entry.value);
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// final List<isar_models.UTXO> outputArray = [];
|
|
//
|
|
// for (int i = 0; i < fetchedUtxoList.length; i++) {
|
|
// for (int j = 0; j < fetchedUtxoList[i].length; j++) {
|
|
// final jsonUTXO = fetchedUtxoList[i][j];
|
|
//
|
|
// final txn = await cachedElectrumXClient.getTransaction(
|
|
// txHash: jsonUTXO["tx_hash"] as String,
|
|
// verbose: true,
|
|
// coin: coin,
|
|
// );
|
|
//
|
|
// // fetch stored tx to see if paynym notification tx and block utxo
|
|
// final storedTx = await db.getTransaction(
|
|
// walletId,
|
|
// jsonUTXO["tx_hash"] as String,
|
|
// );
|
|
//
|
|
// bool shouldBlock = false;
|
|
// String? blockReason;
|
|
// String? label;
|
|
//
|
|
// if (storedTx?.subType ==
|
|
// isar_models.TransactionSubType.bip47Notification) {
|
|
// if (storedTx?.type == isar_models.TransactionType.incoming) {
|
|
// shouldBlock = true;
|
|
// blockReason = "Incoming paynym notification transaction.";
|
|
// } else if (storedTx?.type == isar_models.TransactionType.outgoing) {
|
|
// shouldBlock = false;
|
|
// blockReason = "Paynym notification change output. Incautious "
|
|
// "handling of change outputs from notification transactions "
|
|
// "may cause unintended loss of privacy.";
|
|
// label = blockReason;
|
|
// }
|
|
// }
|
|
//
|
|
// final vout = jsonUTXO["tx_pos"] as int;
|
|
//
|
|
// final outputs = txn["vout"] as List;
|
|
//
|
|
// String? utxoOwnerAddress;
|
|
// // get UTXO owner address
|
|
// for (final output in outputs) {
|
|
// if (output["n"] == vout) {
|
|
// utxoOwnerAddress =
|
|
// output["scriptPubKey"]?["addresses"]?[0] as String? ??
|
|
// output["scriptPubKey"]?["address"] as String?;
|
|
// }
|
|
// }
|
|
//
|
|
// final utxo = isar_models.UTXO(
|
|
// walletId: walletId,
|
|
// txid: txn["txid"] as String,
|
|
// vout: vout,
|
|
// value: jsonUTXO["value"] as int,
|
|
// name: label ?? "",
|
|
// isBlocked: shouldBlock,
|
|
// blockedReason: blockReason,
|
|
// isCoinbase: txn["is_coinbase"] as bool? ?? false,
|
|
// blockHash: txn["blockhash"] as String?,
|
|
// blockHeight: jsonUTXO["height"] as int?,
|
|
// blockTime: txn["blocktime"] as int?,
|
|
// address: utxoOwnerAddress,
|
|
// );
|
|
//
|
|
// outputArray.add(utxo);
|
|
// }
|
|
// }
|
|
//
|
|
// Logging.instance
|
|
// .log('Outputs fetched: $outputArray', level: LogLevel.Info);
|
|
//
|
|
// await db.updateUTXOs(walletId, outputArray);
|
|
//
|
|
// // finally update balance
|
|
// await _updateBalance();
|
|
// } catch (e, s) {
|
|
// Logging.instance
|
|
// .log("Output fetch unsuccessful: $e\n$s", level: LogLevel.Error);
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> _updateBalance() async {
|
|
// await refreshBalance();
|
|
// }
|
|
//
|
|
// @override
|
|
// Balance get balance => _balance ??= getCachedBalance();
|
|
// Balance? _balance;
|
|
//
|
|
// // /// Takes in a list of UtxoObjects and adds a name (dependent on object index within list)
|
|
// // /// and checks for the txid associated with the utxo being blocked and marks it accordingly.
|
|
// // /// Now also checks for output labeling.
|
|
// // Future<void> _sortOutputs(List<UtxoObject> utxos) async {
|
|
// // final blockedHashArray =
|
|
// // DB.instance.get<dynamic>(boxName: walletId, key: 'blocked_tx_hashes')
|
|
// // as List<dynamic>?;
|
|
// // final List<String> lst = [];
|
|
// // if (blockedHashArray != null) {
|
|
// // for (var hash in blockedHashArray) {
|
|
// // lst.add(hash as String);
|
|
// // }
|
|
// // }
|
|
// // final labels =
|
|
// // DB.instance.get<dynamic>(boxName: walletId, key: 'labels') as Map? ??
|
|
// // {};
|
|
// //
|
|
// // outputsList = [];
|
|
// //
|
|
// // for (var i = 0; i < utxos.length; i++) {
|
|
// // if (labels[utxos[i].txid] != null) {
|
|
// // utxos[i].txName = labels[utxos[i].txid] as String? ?? "";
|
|
// // } else {
|
|
// // utxos[i].txName = 'Output #$i';
|
|
// // }
|
|
// //
|
|
// // if (utxos[i].status.confirmed == false) {
|
|
// // outputsList.add(utxos[i]);
|
|
// // } else {
|
|
// // if (lst.contains(utxos[i].txid)) {
|
|
// // utxos[i].blocked = true;
|
|
// // outputsList.add(utxos[i]);
|
|
// // } else if (!lst.contains(utxos[i].txid)) {
|
|
// // outputsList.add(utxos[i]);
|
|
// // }
|
|
// // }
|
|
// // }
|
|
// // }
|
|
//
|
|
// Future<int> getTxCount({required String address}) async {
|
|
// String? scripthash;
|
|
// try {
|
|
// scripthash = AddressUtils.convertToScriptHash(address, network);
|
|
// final transactions =
|
|
// await electrumXClient.getHistory(scripthash: scripthash);
|
|
// return transactions.length;
|
|
// } catch (e) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown in _getTxCount(address: $address, scripthash: $scripthash): $e",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<Map<String, int>> _getBatchTxCount({
|
|
// required Map<String, String> addresses,
|
|
// }) async {
|
|
// try {
|
|
// final Map<String, List<dynamic>> args = {};
|
|
// for (final entry in addresses.entries) {
|
|
// args[entry.key] = [
|
|
// AddressUtils.convertToScriptHash(entry.value, network)
|
|
// ];
|
|
// }
|
|
// final response = await electrumXClient.getBatchHistory(args: args);
|
|
//
|
|
// final Map<String, int> result = {};
|
|
// for (final entry in response.entries) {
|
|
// result[entry.key] = entry.value.length;
|
|
// }
|
|
// return result;
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown in _getBatchTxCount(address: $addresses: $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> _checkReceivingAddressForTransactions() async {
|
|
// try {
|
|
// final currentReceiving = await _currentReceivingAddress;
|
|
//
|
|
// final int txCount = await getTxCount(address: currentReceiving.value);
|
|
// Logging.instance.log(
|
|
// 'Number of txs for current receiving address $currentReceiving: $txCount',
|
|
// level: LogLevel.Info);
|
|
//
|
|
// if (txCount >= 1 || currentReceiving.derivationIndex < 0) {
|
|
// // First increment the receiving index
|
|
// final newReceivingIndex = currentReceiving.derivationIndex + 1;
|
|
//
|
|
// // Use new index to derive a new receiving address
|
|
// final newReceivingAddress = await _generateAddressForChain(
|
|
// 0, newReceivingIndex, DerivePathTypeExt.primaryFor(coin));
|
|
//
|
|
// final existing = await db
|
|
// .getAddresses(walletId)
|
|
// .filter()
|
|
// .valueEqualTo(newReceivingAddress.value)
|
|
// .findFirst();
|
|
// if (existing == null) {
|
|
// // Add that new change address
|
|
// await db.putAddress(newReceivingAddress);
|
|
// } else {
|
|
// // we need to update the address
|
|
// await db.updateAddress(existing, newReceivingAddress);
|
|
// }
|
|
// // keep checking until address with no tx history is set as current
|
|
// await _checkReceivingAddressForTransactions();
|
|
// }
|
|
// } on SocketException catch (se, s) {
|
|
// Logging.instance.log(
|
|
// "SocketException caught in _checkReceivingAddressForTransactions(${DerivePathTypeExt.primaryFor(coin)}): $se\n$s",
|
|
// level: LogLevel.Error);
|
|
// return;
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown from _checkReceivingAddressForTransactions(${DerivePathTypeExt.primaryFor(coin)}): $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> checkChangeAddressForTransactions() async {
|
|
// try {
|
|
// final currentChange = await _currentChangeAddress;
|
|
// final int txCount = await getTxCount(address: currentChange.value);
|
|
// Logging.instance.log(
|
|
// 'Number of txs for current change address $currentChange: $txCount',
|
|
// level: LogLevel.Info);
|
|
//
|
|
// if (txCount >= 1 || currentChange.derivationIndex < 0) {
|
|
// // First increment the change index
|
|
// final newChangeIndex = currentChange.derivationIndex + 1;
|
|
//
|
|
// // Use new index to derive a new change address
|
|
// final newChangeAddress = await _generateAddressForChain(
|
|
// 1, newChangeIndex, DerivePathTypeExt.primaryFor(coin));
|
|
//
|
|
// final existing = await db
|
|
// .getAddresses(walletId)
|
|
// .filter()
|
|
// .valueEqualTo(newChangeAddress.value)
|
|
// .findFirst();
|
|
// if (existing == null) {
|
|
// // Add that new change address
|
|
// await db.putAddress(newChangeAddress);
|
|
// } else {
|
|
// // we need to update the address
|
|
// await db.updateAddress(existing, newChangeAddress);
|
|
// }
|
|
// // keep checking until address with no tx history is set as current
|
|
// await checkChangeAddressForTransactions();
|
|
// }
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown from _checkChangeAddressForTransactions(${DerivePathTypeExt.primaryFor(coin)}): $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> _checkCurrentReceivingAddressesForTransactions() async {
|
|
// try {
|
|
// // for (final type in DerivePathType.values) {
|
|
// await _checkReceivingAddressForTransactions();
|
|
// // }
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown from _checkCurrentReceivingAddressesForTransactions(): $e\n$s",
|
|
// level: LogLevel.Info);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// /// public wrapper because dart can't test private...
|
|
// Future<void> checkCurrentReceivingAddressesForTransactions() async {
|
|
// if (Platform.environment["FLUTTER_TEST"] == "true") {
|
|
// try {
|
|
// return _checkCurrentReceivingAddressesForTransactions();
|
|
// } catch (_) {
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> _checkCurrentChangeAddressesForTransactions() async {
|
|
// try {
|
|
// // for (final type in DerivePathType.values) {
|
|
// await checkChangeAddressForTransactions();
|
|
// // }
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown from _checkCurrentChangeAddressesForTransactions(): $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// /// public wrapper because dart can't test private...
|
|
// Future<void> checkCurrentChangeAddressesForTransactions() async {
|
|
// if (Platform.environment["FLUTTER_TEST"] == "true") {
|
|
// try {
|
|
// return _checkCurrentChangeAddressesForTransactions();
|
|
// } catch (_) {
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// Future<List<Map<String, dynamic>>> _fetchHistory(
|
|
// List<String> allAddresses) async {
|
|
// try {
|
|
// List<Map<String, dynamic>> allTxHashes = [];
|
|
//
|
|
// final Map<int, Map<String, List<dynamic>>> batches = {};
|
|
// final Map<String, String> requestIdToAddressMap = {};
|
|
// const batchSizeMax = 100;
|
|
// int batchNumber = 0;
|
|
// for (int i = 0; i < allAddresses.length; i++) {
|
|
// if (batches[batchNumber] == null) {
|
|
// batches[batchNumber] = {};
|
|
// }
|
|
// final scripthash =
|
|
// AddressUtils.convertToScriptHash(allAddresses[i], network);
|
|
// final id = Logger.isTestEnv ? "$i" : const Uuid().v1();
|
|
// requestIdToAddressMap[id] = allAddresses[i];
|
|
// batches[batchNumber]!.addAll({
|
|
// id: [scripthash]
|
|
// });
|
|
// if (i % batchSizeMax == batchSizeMax - 1) {
|
|
// batchNumber++;
|
|
// }
|
|
// }
|
|
//
|
|
// for (int i = 0; i < batches.length; i++) {
|
|
// final response =
|
|
// await _electrumXClient.getBatchHistory(args: batches[i]!);
|
|
// for (final entry in response.entries) {
|
|
// for (int j = 0; j < entry.value.length; j++) {
|
|
// entry.value[j]["address"] = requestIdToAddressMap[entry.key];
|
|
// if (!allTxHashes.contains(entry.value[j])) {
|
|
// allTxHashes.add(entry.value[j]);
|
|
// }
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// return allTxHashes;
|
|
// } catch (e, s) {
|
|
// Logging.instance.log("_fetchHistory: $e\n$s", level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// bool _duplicateTxCheck(
|
|
// List<Map<String, dynamic>> allTransactions, String txid) {
|
|
// for (int i = 0; i < allTransactions.length; i++) {
|
|
// if (allTransactions[i]["txid"] == txid) {
|
|
// return true;
|
|
// }
|
|
// }
|
|
// return false;
|
|
// }
|
|
//
|
|
// // Future<void> migrate() async {
|
|
// // final receivingAddressesP2PKH = DB.instance.get<dynamic>(
|
|
// // boxName: walletId, key: 'receivingAddressesP2PKH') as List<dynamic>;
|
|
// //
|
|
// // final changeAddressesP2PKH =
|
|
// // DB.instance.get<dynamic>(boxName: walletId, key: 'changeAddressesP2PKH')
|
|
// // as List<dynamic>;
|
|
// //
|
|
// // await isar.writeTxn(() async {
|
|
// // for (var i = 0; i < receivingAddressesP2PKH.length; i++) {
|
|
// // await isar.address.put(
|
|
// // isar_models.Address()
|
|
// // ..type = isar_models.AddressType.p2pkh
|
|
// // ..subType = isar_models.AddressSubType.receiving
|
|
// // ..publicKey = []
|
|
// // ..derivationIndex = i
|
|
// // ..value = receivingAddressesP2PKH[i] as String,
|
|
// // );
|
|
// // }
|
|
// // for (var i = 0; i < changeAddressesP2PKH.length; i++) {
|
|
// // await isar.address.put(
|
|
// // isar_models.Address()
|
|
// // ..type = isar_models.AddressType.p2pkh
|
|
// // ..subType = isar_models.AddressSubType.change
|
|
// // ..publicKey = []
|
|
// // ..derivationIndex = i
|
|
// // ..value = changeAddressesP2PKH[i] as String,
|
|
// // );
|
|
// // }
|
|
// // });
|
|
// //
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId, key: "receivingAddressesP2PKH", value: []);
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId, key: "changeAddressesP2PKH", value: []);
|
|
// // }
|
|
//
|
|
// Future<void> _refreshTransactions() async {
|
|
// final List<isar_models.Address> allAddresses =
|
|
// await _fetchAllOwnAddresses();
|
|
//
|
|
// final List<Map<String, dynamic>> allTxHashes =
|
|
// await _fetchHistory(allAddresses.map((e) => e.value).toList());
|
|
//
|
|
// List<String> hashes =
|
|
// allTxHashes.map((e) => e['tx_hash'] as String).toList(growable: false);
|
|
//
|
|
// await fastFetch(hashes);
|
|
// List<Map<String, dynamic>> allTransactions = [];
|
|
// final currentHeight = await chainHeight;
|
|
//
|
|
// for (final txHash in allTxHashes) {
|
|
// final storedTx = await db
|
|
// .getTransactions(walletId)
|
|
// .filter()
|
|
// .txidEqualTo(txHash["tx_hash"] as String)
|
|
// .findFirst();
|
|
//
|
|
// if (storedTx == null ||
|
|
// !storedTx.isConfirmed(currentHeight, MINIMUM_CONFIRMATIONS)) {
|
|
// final tx = await cachedElectrumXClient.getTransaction(
|
|
// txHash: txHash["tx_hash"] as String,
|
|
// verbose: true,
|
|
// coin: coin,
|
|
// );
|
|
//
|
|
// if (!_duplicateTxCheck(allTransactions, tx["txid"] as String)) {
|
|
// tx["address"] = await db
|
|
// .getAddresses(walletId)
|
|
// .filter()
|
|
// .valueEqualTo(txHash["address"] as String)
|
|
// .findFirst();
|
|
// tx["height"] = txHash["height"];
|
|
// allTransactions.add(tx);
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// Set<String> vHashes = {};
|
|
// for (final txObject in allTransactions) {
|
|
// for (int i = 0; i < (txObject["vin"] as List).length; i++) {
|
|
// final input = txObject["vin"]![i] as Map;
|
|
// final prevTxid = input["txid"] as String;
|
|
// vHashes.add(prevTxid);
|
|
// }
|
|
// }
|
|
// await fastFetch(vHashes.toList());
|
|
//
|
|
// final List<Tuple2<isar_models.Transaction, isar_models.Address?>> txns = [];
|
|
//
|
|
// for (final txObject in allTransactions) {
|
|
// final txn = await parseTransaction(
|
|
// txObject,
|
|
// cachedElectrumXClient,
|
|
// allAddresses,
|
|
// coin,
|
|
// MINIMUM_CONFIRMATIONS,
|
|
// walletId,
|
|
// );
|
|
//
|
|
// txns.add(txn);
|
|
// }
|
|
//
|
|
// await db.addNewTransactionData(txns, walletId);
|
|
//
|
|
// // quick hack to notify manager to call notifyListeners if
|
|
// // transactions changed
|
|
// if (txns.isNotEmpty) {
|
|
// GlobalEventBus.instance.fire(
|
|
// UpdatedInBackgroundEvent(
|
|
// "Transactions updated/added for: $walletId $walletName ",
|
|
// walletId,
|
|
// ),
|
|
// );
|
|
// }
|
|
// }
|
|
//
|
|
// int estimateTxFee({required int vSize, required int feeRatePerKB}) {
|
|
// return vSize * (feeRatePerKB / 1000).ceil();
|
|
// }
|
|
//
|
|
// /// The coinselection algorithm decides whether or not the user is eligible to make the transaction
|
|
// /// with [satoshiAmountToSend] and [selectedTxFeeRate]. If so, it will call buildTrasaction() and return
|
|
// /// a map containing the tx hex along with other important information. If not, then it will return
|
|
// /// an integer (1 or 2)
|
|
// dynamic coinSelection({
|
|
// required int satoshiAmountToSend,
|
|
// required int selectedTxFeeRate,
|
|
// required String recipientAddress,
|
|
// required bool coinControl,
|
|
// required bool isSendAll,
|
|
// int? satsPerVByte,
|
|
// int additionalOutputs = 0,
|
|
// List<isar_models.UTXO>? utxos,
|
|
// }) async {
|
|
// Logging.instance
|
|
// .log("Starting coinSelection ----------", level: LogLevel.Info);
|
|
// final List<isar_models.UTXO> availableOutputs = utxos ?? await this.utxos;
|
|
// final currentChainHeight = await chainHeight;
|
|
// final List<isar_models.UTXO> spendableOutputs = [];
|
|
// int spendableSatoshiValue = 0;
|
|
//
|
|
// // Build list of spendable outputs and totaling their satoshi amount
|
|
// for (final utxo in availableOutputs) {
|
|
// if (utxo.isBlocked == false &&
|
|
// utxo.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) &&
|
|
// utxo.used != true) {
|
|
// spendableOutputs.add(utxo);
|
|
// spendableSatoshiValue += utxo.value;
|
|
// }
|
|
// }
|
|
//
|
|
// if (coinControl) {
|
|
// if (spendableOutputs.length < availableOutputs.length) {
|
|
// throw ArgumentError("Attempted to use an unavailable utxo");
|
|
// }
|
|
// }
|
|
//
|
|
// // don't care about sorting if using all utxos
|
|
// if (!coinControl) {
|
|
// // sort spendable by age (oldest first)
|
|
// spendableOutputs.sort((a, b) => b.blockTime!.compareTo(a.blockTime!));
|
|
// }
|
|
//
|
|
// Logging.instance.log("spendableOutputs.length: ${spendableOutputs.length}",
|
|
// level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log("spendableOutputs: $spendableOutputs", level: LogLevel.Info);
|
|
// Logging.instance.log("spendableSatoshiValue: $spendableSatoshiValue",
|
|
// level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log("satoshiAmountToSend: $satoshiAmountToSend", level: LogLevel.Info);
|
|
// // If the amount the user is trying to send is smaller than the amount that they have spendable,
|
|
// // then return 1, which indicates that they have an insufficient balance.
|
|
// if (spendableSatoshiValue < satoshiAmountToSend) {
|
|
// return 1;
|
|
// // If the amount the user wants to send is exactly equal to the amount they can spend, then return
|
|
// // 2, which indicates that they are not leaving enough over to pay the transaction fee
|
|
// } else if (spendableSatoshiValue == satoshiAmountToSend && !isSendAll) {
|
|
// return 2;
|
|
// }
|
|
// // If neither of these statements pass, we assume that the user has a spendable balance greater
|
|
// // than the amount they're attempting to send. Note that this value still does not account for
|
|
// // the added transaction fee, which may require an extra input and will need to be checked for
|
|
// // later on.
|
|
//
|
|
// // Possible situation right here
|
|
// int satoshisBeingUsed = 0;
|
|
// int inputsBeingConsumed = 0;
|
|
// List<isar_models.UTXO> utxoObjectsToUse = [];
|
|
//
|
|
// if (!coinControl) {
|
|
// for (var i = 0;
|
|
// satoshisBeingUsed < satoshiAmountToSend &&
|
|
// i < spendableOutputs.length;
|
|
// i++) {
|
|
// utxoObjectsToUse.add(spendableOutputs[i]);
|
|
// satoshisBeingUsed += spendableOutputs[i].value;
|
|
// inputsBeingConsumed += 1;
|
|
// }
|
|
// for (int i = 0;
|
|
// i < additionalOutputs &&
|
|
// inputsBeingConsumed < spendableOutputs.length;
|
|
// i++) {
|
|
// utxoObjectsToUse.add(spendableOutputs[inputsBeingConsumed]);
|
|
// satoshisBeingUsed += spendableOutputs[inputsBeingConsumed].value;
|
|
// inputsBeingConsumed += 1;
|
|
// }
|
|
// } else {
|
|
// satoshisBeingUsed = spendableSatoshiValue;
|
|
// utxoObjectsToUse = spendableOutputs;
|
|
// inputsBeingConsumed = spendableOutputs.length;
|
|
// }
|
|
//
|
|
// Logging.instance
|
|
// .log("satoshisBeingUsed: $satoshisBeingUsed", level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log("inputsBeingConsumed: $inputsBeingConsumed", level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('utxoObjectsToUse: $utxoObjectsToUse', level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('satoshiAmountToSend $satoshiAmountToSend', level: LogLevel.Info);
|
|
//
|
|
// // numberOfOutputs' length must always be equal to that of recipientsArray and recipientsAmtArray
|
|
// List<String> recipientsArray = [recipientAddress];
|
|
// List<int> recipientsAmtArray = [satoshiAmountToSend];
|
|
//
|
|
// // gather required signing data
|
|
// final utxoSigningData = await fetchBuildTxData(utxoObjectsToUse);
|
|
//
|
|
// if (isSendAll) {
|
|
// Logging.instance
|
|
// .log("Attempting to send all $coin", level: LogLevel.Info);
|
|
//
|
|
// final int vSizeForOneOutput = (await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: [recipientAddress],
|
|
// satoshiAmounts: [satoshisBeingUsed - 1],
|
|
// ))["vSize"] as int;
|
|
// int feeForOneOutput = satsPerVByte != null
|
|
// ? (satsPerVByte * vSizeForOneOutput)
|
|
// : estimateTxFee(
|
|
// vSize: vSizeForOneOutput,
|
|
// feeRatePerKB: selectedTxFeeRate,
|
|
// );
|
|
// if (feeForOneOutput < (vSizeForOneOutput + 1) * 1000) {
|
|
// feeForOneOutput = (vSizeForOneOutput + 1) * 1000;
|
|
// }
|
|
//
|
|
// final int amount = satoshiAmountToSend - feeForOneOutput;
|
|
// dynamic txn = await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: recipientsArray,
|
|
// satoshiAmounts: [amount],
|
|
// );
|
|
// Map<String, dynamic> transactionObject = {
|
|
// "hex": txn["hex"],
|
|
// "recipient": recipientsArray[0],
|
|
// "recipientAmt": Amount(
|
|
// rawValue: BigInt.from(amount),
|
|
// fractionDigits: coin.decimals,
|
|
// ),
|
|
// "fee": feeForOneOutput,
|
|
// "vSize": txn["vSize"],
|
|
// "usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
// };
|
|
// return transactionObject;
|
|
// }
|
|
//
|
|
// final int vSizeForOneOutput = (await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: [recipientAddress],
|
|
// satoshiAmounts: [satoshisBeingUsed - 1],
|
|
// ))["vSize"] as int;
|
|
// final int vSizeForTwoOutPuts = (await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: [
|
|
// recipientAddress,
|
|
// await _getCurrentAddressForChain(1, DerivePathTypeExt.primaryFor(coin)),
|
|
// ],
|
|
// satoshiAmounts: [
|
|
// satoshiAmountToSend,
|
|
// satoshisBeingUsed - satoshiAmountToSend - 1,
|
|
// ], // dust limit is the minimum amount a change output should be
|
|
// ))["vSize"] as int;
|
|
// //todo: check if print needed
|
|
// debugPrint("vSizeForOneOutput $vSizeForOneOutput");
|
|
// debugPrint("vSizeForTwoOutPuts $vSizeForTwoOutPuts");
|
|
//
|
|
// // Assume 1 output, only for recipient and no change
|
|
// var feeForOneOutput = satsPerVByte != null
|
|
// ? (satsPerVByte * vSizeForOneOutput)
|
|
// : estimateTxFee(
|
|
// vSize: vSizeForOneOutput,
|
|
// feeRatePerKB: selectedTxFeeRate,
|
|
// );
|
|
// // Assume 2 outputs, one for recipient and one for change
|
|
// var feeForTwoOutputs = satsPerVByte != null
|
|
// ? (satsPerVByte * vSizeForTwoOutPuts)
|
|
// : estimateTxFee(
|
|
// vSize: vSizeForTwoOutPuts,
|
|
// feeRatePerKB: selectedTxFeeRate,
|
|
// );
|
|
//
|
|
// Logging.instance
|
|
// .log("feeForTwoOutputs: $feeForTwoOutputs", level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log("feeForOneOutput: $feeForOneOutput", level: LogLevel.Info);
|
|
// if (feeForOneOutput < (vSizeForOneOutput + 1) * 1000) {
|
|
// feeForOneOutput = (vSizeForOneOutput + 1) * 1000;
|
|
// }
|
|
// if (feeForTwoOutputs < ((vSizeForTwoOutPuts + 1) * 1000)) {
|
|
// feeForTwoOutputs = ((vSizeForTwoOutPuts + 1) * 1000);
|
|
// }
|
|
//
|
|
// Logging.instance
|
|
// .log("feeForTwoOutputs: $feeForTwoOutputs", level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log("feeForOneOutput: $feeForOneOutput", level: LogLevel.Info);
|
|
//
|
|
// if (satoshisBeingUsed - satoshiAmountToSend > feeForOneOutput) {
|
|
// if (satoshisBeingUsed - satoshiAmountToSend >
|
|
// feeForOneOutput + DUST_LIMIT.raw.toInt()) {
|
|
// // Here, we know that theoretically, we may be able to include another output(change) but we first need to
|
|
// // factor in the value of this output in satoshis.
|
|
// int changeOutputSize =
|
|
// satoshisBeingUsed - satoshiAmountToSend - feeForTwoOutputs;
|
|
// // We check to see if the user can pay for the new transaction with 2 outputs instead of one. If they can and
|
|
// // the second output's size > 546 satoshis, we perform the mechanics required to properly generate and use a new
|
|
// // change address.
|
|
// if (changeOutputSize > DUST_LIMIT.raw.toInt() &&
|
|
// satoshisBeingUsed - satoshiAmountToSend - changeOutputSize ==
|
|
// feeForTwoOutputs) {
|
|
// // generate new change address if current change address has been used
|
|
// await checkChangeAddressForTransactions();
|
|
// final String newChangeAddress = await _getCurrentAddressForChain(
|
|
// 1, DerivePathTypeExt.primaryFor(coin));
|
|
//
|
|
// int feeBeingPaid =
|
|
// satoshisBeingUsed - satoshiAmountToSend - changeOutputSize;
|
|
//
|
|
// recipientsArray.add(newChangeAddress);
|
|
// recipientsAmtArray.add(changeOutputSize);
|
|
// // At this point, we have the outputs we're going to use, the amounts to send along with which addresses
|
|
// // we intend to send these amounts to. We have enough to send instructions to build the transaction.
|
|
// Logging.instance.log('2 outputs in tx', level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
// Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log('Change Output Size: $changeOutputSize',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log(
|
|
// 'Difference (fee being paid): $feeBeingPaid sats',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('Estimated fee: $feeForTwoOutputs', level: LogLevel.Info);
|
|
// dynamic txn = await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: recipientsArray,
|
|
// satoshiAmounts: recipientsAmtArray,
|
|
// );
|
|
//
|
|
// // make sure minimum fee is accurate if that is being used
|
|
// if (txn["vSize"] - feeBeingPaid == 1) {
|
|
// int changeOutputSize =
|
|
// satoshisBeingUsed - satoshiAmountToSend - (txn["vSize"] as int);
|
|
// feeBeingPaid =
|
|
// satoshisBeingUsed - satoshiAmountToSend - changeOutputSize;
|
|
// recipientsAmtArray.removeLast();
|
|
// recipientsAmtArray.add(changeOutputSize);
|
|
// Logging.instance.log('Adjusted Input size: $satoshisBeingUsed',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log(
|
|
// 'Adjusted Recipient output size: $satoshiAmountToSend',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log(
|
|
// 'Adjusted Change Output Size: $changeOutputSize',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log(
|
|
// 'Adjusted Difference (fee being paid): $feeBeingPaid sats',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log('Adjusted Estimated fee: $feeForTwoOutputs',
|
|
// level: LogLevel.Info);
|
|
// txn = await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: recipientsArray,
|
|
// satoshiAmounts: recipientsAmtArray,
|
|
// );
|
|
// }
|
|
//
|
|
// Map<String, dynamic> transactionObject = {
|
|
// "hex": txn["hex"],
|
|
// "recipient": recipientsArray[0],
|
|
// "recipientAmt": Amount(
|
|
// rawValue: BigInt.from(recipientsAmtArray[0]),
|
|
// fractionDigits: coin.decimals,
|
|
// ),
|
|
// "fee": feeBeingPaid,
|
|
// "vSize": txn["vSize"],
|
|
// "usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
// };
|
|
// return transactionObject;
|
|
// } else {
|
|
// // Something went wrong here. It either overshot or undershot the estimated fee amount or the changeOutputSize
|
|
// // is smaller than or equal to [DUST_LIMIT]. Revert to single output transaction.
|
|
// Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
// Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log(
|
|
// 'Difference (fee being paid): ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
// dynamic txn = await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: recipientsArray,
|
|
// satoshiAmounts: recipientsAmtArray,
|
|
// );
|
|
// Map<String, dynamic> transactionObject = {
|
|
// "hex": txn["hex"],
|
|
// "recipient": recipientsArray[0],
|
|
// "recipientAmt": Amount(
|
|
// rawValue: BigInt.from(recipientsAmtArray[0]),
|
|
// fractionDigits: coin.decimals,
|
|
// ),
|
|
// "fee": satoshisBeingUsed - satoshiAmountToSend,
|
|
// "vSize": txn["vSize"],
|
|
// "usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
// };
|
|
// return transactionObject;
|
|
// }
|
|
// } else {
|
|
// // No additional outputs needed since adding one would mean that it'd be smaller than 546 sats
|
|
// // which makes it uneconomical to add to the transaction. Here, we pass data directly to instruct
|
|
// // the wallet to begin crafting the transaction that the user requested.
|
|
// Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
// Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log(
|
|
// 'Difference (fee being paid): ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
// dynamic txn = await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: recipientsArray,
|
|
// satoshiAmounts: recipientsAmtArray,
|
|
// );
|
|
// Map<String, dynamic> transactionObject = {
|
|
// "hex": txn["hex"],
|
|
// "recipient": recipientsArray[0],
|
|
// "recipientAmt": Amount(
|
|
// rawValue: BigInt.from(recipientsAmtArray[0]),
|
|
// fractionDigits: coin.decimals,
|
|
// ),
|
|
// "fee": satoshisBeingUsed - satoshiAmountToSend,
|
|
// "vSize": txn["vSize"],
|
|
// "usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
// };
|
|
// return transactionObject;
|
|
// }
|
|
// } else if (satoshisBeingUsed - satoshiAmountToSend == feeForOneOutput) {
|
|
// // In this scenario, no additional change output is needed since inputs - outputs equal exactly
|
|
// // what we need to pay for fees. Here, we pass data directly to instruct the wallet to begin
|
|
// // crafting the transaction that the user requested.
|
|
// Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
// Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance.log(
|
|
// 'Fee being paid: ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
// level: LogLevel.Info);
|
|
// Logging.instance
|
|
// .log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
// dynamic txn = await buildTransaction(
|
|
// utxoSigningData: utxoSigningData,
|
|
// recipients: recipientsArray,
|
|
// satoshiAmounts: recipientsAmtArray,
|
|
// );
|
|
// Map<String, dynamic> transactionObject = {
|
|
// "hex": txn["hex"],
|
|
// "recipient": recipientsArray[0],
|
|
// "recipientAmt": Amount(
|
|
// rawValue: BigInt.from(recipientsAmtArray[0]),
|
|
// fractionDigits: coin.decimals,
|
|
// ),
|
|
// "fee": feeForOneOutput,
|
|
// "vSize": txn["vSize"],
|
|
// "usedUTXOs": utxoSigningData.map((e) => e.utxo).toList(),
|
|
// };
|
|
// return transactionObject;
|
|
// } else {
|
|
// // Remember that returning 2 iTndicates that the user does not have a sufficient balance to
|
|
// // pay for the transaction fee. Ideally, at this stage, we should check if the user has any
|
|
// // additional outputs they're able to spend and then recalculate fees.
|
|
// Logging.instance.log(
|
|
// 'Cannot pay tx fee - checking for more outputs and trying again',
|
|
// level: LogLevel.Warning);
|
|
// // try adding more outputs
|
|
// if (spendableOutputs.length > inputsBeingConsumed) {
|
|
// return coinSelection(
|
|
// satoshiAmountToSend: satoshiAmountToSend,
|
|
// selectedTxFeeRate: selectedTxFeeRate,
|
|
// satsPerVByte: satsPerVByte,
|
|
// recipientAddress: recipientAddress,
|
|
// isSendAll: isSendAll,
|
|
// additionalOutputs: additionalOutputs + 1,
|
|
// utxos: utxos,
|
|
// coinControl: coinControl,
|
|
// );
|
|
// }
|
|
// return 2;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<List<SigningData>> fetchBuildTxData(
|
|
// List<isar_models.UTXO> utxosToUse,
|
|
// ) async {
|
|
// // return data
|
|
// List<SigningData> signingData = [];
|
|
//
|
|
// try {
|
|
// // Populating the addresses to check
|
|
// for (var i = 0; i < utxosToUse.length; i++) {
|
|
// if (utxosToUse[i].address == null) {
|
|
// final txid = utxosToUse[i].txid;
|
|
// final tx = await _cachedElectrumXClient.getTransaction(
|
|
// txHash: txid,
|
|
// coin: coin,
|
|
// );
|
|
// for (final output in tx["vout"] as List) {
|
|
// final n = output["n"];
|
|
// if (n != null && n == utxosToUse[i].vout) {
|
|
// utxosToUse[i] = utxosToUse[i].copyWith(
|
|
// address: output["scriptPubKey"]?["addresses"]?[0] as String? ??
|
|
// output["scriptPubKey"]["address"] as String,
|
|
// );
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// final derivePathType = addressType(address: utxosToUse[i].address!);
|
|
//
|
|
// signingData.add(
|
|
// SigningData(
|
|
// derivePathType: derivePathType,
|
|
// utxo: utxosToUse[i],
|
|
// ),
|
|
// );
|
|
// }
|
|
//
|
|
// Map<DerivePathType, Map<String, dynamic>> receiveDerivations = {};
|
|
// Map<DerivePathType, Map<String, dynamic>> changeDerivations = {};
|
|
//
|
|
// for (final sd in signingData) {
|
|
// String? pubKey;
|
|
// String? wif;
|
|
//
|
|
// // fetch receiving derivations if null
|
|
// receiveDerivations[sd.derivePathType] ??= await _fetchDerivations(
|
|
// chain: 0,
|
|
// derivePathType: sd.derivePathType,
|
|
// );
|
|
// final receiveDerivation =
|
|
// receiveDerivations[sd.derivePathType]![sd.utxo.address!];
|
|
//
|
|
// if (receiveDerivation != null) {
|
|
// pubKey = receiveDerivation["pubKey"] as String;
|
|
// wif = receiveDerivation["wif"] as String;
|
|
// } else {
|
|
// // fetch change derivations if null
|
|
// changeDerivations[sd.derivePathType] ??= await _fetchDerivations(
|
|
// chain: 1,
|
|
// derivePathType: sd.derivePathType,
|
|
// );
|
|
// final changeDerivation =
|
|
// changeDerivations[sd.derivePathType]![sd.utxo.address!];
|
|
// if (changeDerivation != null) {
|
|
// pubKey = changeDerivation["pubKey"] as String;
|
|
// wif = changeDerivation["wif"] as String;
|
|
// }
|
|
// }
|
|
//
|
|
// if (wif == null || pubKey == null) {
|
|
// final address = await db.getAddress(walletId, sd.utxo.address!);
|
|
// if (address?.derivationPath != null) {
|
|
// final node = await Bip32Utils.getBip32Node(
|
|
// (await mnemonicString)!,
|
|
// (await mnemonicPassphrase)!,
|
|
// network,
|
|
// address!.derivationPath!.value,
|
|
// );
|
|
//
|
|
// wif = node.toWIF();
|
|
// pubKey = Format.uint8listToString(node.publicKey);
|
|
// }
|
|
// }
|
|
//
|
|
// if (wif != null && pubKey != null) {
|
|
// final PaymentData data;
|
|
// final Uint8List? redeemScript;
|
|
//
|
|
// switch (sd.derivePathType) {
|
|
// case DerivePathType.bip44:
|
|
// data = P2PKH(
|
|
// data: PaymentData(
|
|
// pubkey: Format.stringToUint8List(pubKey),
|
|
// ),
|
|
// network: network,
|
|
// ).data;
|
|
// redeemScript = null;
|
|
// break;
|
|
//
|
|
// default:
|
|
// throw Exception("DerivePathType unsupported");
|
|
// }
|
|
//
|
|
// final keyPair = ECPair.fromWIF(
|
|
// wif,
|
|
// network: network,
|
|
// );
|
|
//
|
|
// sd.redeemScript = redeemScript;
|
|
// sd.output = data.output;
|
|
// sd.keyPair = keyPair;
|
|
// }
|
|
// }
|
|
//
|
|
// return signingData;
|
|
// } catch (e, s) {
|
|
// Logging.instance
|
|
// .log("fetchBuildTxData() threw: $e,\n$s", level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// /// Builds and signs a transaction
|
|
// Future<Map<String, dynamic>> buildTransaction({
|
|
// required List<SigningData> utxoSigningData,
|
|
// required List<String> recipients,
|
|
// required List<int> satoshiAmounts,
|
|
// }) async {
|
|
// Logging.instance
|
|
// .log("Starting buildTransaction ----------", level: LogLevel.Info);
|
|
//
|
|
// final txb = TransactionBuilder(
|
|
// network: network,
|
|
// maximumFeeRate: 2500000, // 1000x default value in bitcoindart lib
|
|
// );
|
|
// txb.setVersion(1);
|
|
//
|
|
// // Add transaction inputs
|
|
// for (var i = 0; i < utxoSigningData.length; i++) {
|
|
// final txid = utxoSigningData[i].utxo.txid;
|
|
// txb.addInput(
|
|
// txid,
|
|
// utxoSigningData[i].utxo.vout,
|
|
// null,
|
|
// utxoSigningData[i].output!,
|
|
// );
|
|
// }
|
|
//
|
|
// // Add transaction output
|
|
// for (var i = 0; i < recipients.length; i++) {
|
|
// txb.addOutput(recipients[i], satoshiAmounts[i]);
|
|
// }
|
|
//
|
|
// try {
|
|
// // Sign the transaction accordingly
|
|
// for (var i = 0; i < utxoSigningData.length; i++) {
|
|
// txb.sign(
|
|
// vin: i,
|
|
// keyPair: utxoSigningData[i].keyPair!,
|
|
// witnessValue: utxoSigningData[i].utxo.value,
|
|
// redeemScript: utxoSigningData[i].redeemScript,
|
|
// );
|
|
// }
|
|
// } catch (e, s) {
|
|
// Logging.instance.log("Caught exception while signing transaction: $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
//
|
|
// final builtTx = txb.build();
|
|
// final vSize = builtTx.virtualSize();
|
|
//
|
|
// return {"hex": builtTx.toHex(), "vSize": vSize};
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<void> fullRescan(
|
|
// int maxUnusedAddressGap,
|
|
// int maxNumberOfIndexesToCheck,
|
|
// ) async {
|
|
// Logging.instance.log("Starting full rescan!", level: LogLevel.Info);
|
|
// longMutex = true;
|
|
// GlobalEventBus.instance.fire(
|
|
// WalletSyncStatusChangedEvent(
|
|
// WalletSyncStatus.syncing,
|
|
// walletId,
|
|
// coin,
|
|
// ),
|
|
// );
|
|
//
|
|
// // clear cache
|
|
// await _cachedElectrumXClient.clearSharedTransactionCache(coin: coin);
|
|
//
|
|
// // back up data
|
|
// // await _rescanBackup();
|
|
//
|
|
// // clear blockchain info
|
|
// await db.deleteWalletBlockchainData(walletId);
|
|
// await _deleteDerivations();
|
|
//
|
|
// try {
|
|
// final _mnemonic = await mnemonicString;
|
|
// final _mnemonicPassphrase = await mnemonicPassphrase;
|
|
// if (_mnemonicPassphrase == null) {
|
|
// Logging.instance.log(
|
|
// "Exception in fullRescan: mnemonic passphrase null, possible migration issue; if using internal builds, delete wallet and restore from seed, if using a release build, please file bug report",
|
|
// level: LogLevel.Error);
|
|
// }
|
|
//
|
|
// await _recoverWalletFromBIP32SeedPhrase(
|
|
// mnemonic: _mnemonic!,
|
|
// mnemonicPassphrase: _mnemonicPassphrase!,
|
|
// maxUnusedAddressGap: maxUnusedAddressGap,
|
|
// maxNumberOfIndexesToCheck: maxNumberOfIndexesToCheck,
|
|
// isRescan: true,
|
|
// );
|
|
//
|
|
// longMutex = false;
|
|
// await refresh();
|
|
// Logging.instance.log("Full rescan complete!", level: LogLevel.Info);
|
|
// GlobalEventBus.instance.fire(
|
|
// WalletSyncStatusChangedEvent(
|
|
// WalletSyncStatus.synced,
|
|
// walletId,
|
|
// coin,
|
|
// ),
|
|
// );
|
|
// } catch (e, s) {
|
|
// GlobalEventBus.instance.fire(
|
|
// WalletSyncStatusChangedEvent(
|
|
// WalletSyncStatus.unableToSync,
|
|
// walletId,
|
|
// coin,
|
|
// ),
|
|
// );
|
|
//
|
|
// // restore from backup
|
|
// // await _rescanRestore();
|
|
//
|
|
// longMutex = false;
|
|
// Logging.instance.log("Exception rethrown from fullRescan(): $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// rethrow;
|
|
// }
|
|
// }
|
|
//
|
|
// Future<void> _deleteDerivations() async {
|
|
// // P2PKH derivations
|
|
// await _secureStore.delete(key: "${walletId}_receiveDerivationsP2PKH");
|
|
// await _secureStore.delete(key: "${walletId}_changeDerivationsP2PKH");
|
|
// }
|
|
//
|
|
// // Future<void> _rescanRestore() async {
|
|
// // Logging.instance.log("starting rescan restore", level: LogLevel.Info);
|
|
// //
|
|
// // // restore from backup
|
|
// // // p2pkh
|
|
// // final tempReceivingAddressesP2PKH = DB.instance
|
|
// // .get<dynamic>(boxName: walletId, key: 'receivingAddressesP2PKH_BACKUP');
|
|
// // final tempChangeAddressesP2PKH = DB.instance
|
|
// // .get<dynamic>(boxName: walletId, key: 'changeAddressesP2PKH_BACKUP');
|
|
// // final tempReceivingIndexP2PKH = DB.instance
|
|
// // .get<dynamic>(boxName: walletId, key: 'receivingIndexP2PKH_BACKUP');
|
|
// // final tempChangeIndexP2PKH = DB.instance
|
|
// // .get<dynamic>(boxName: walletId, key: 'changeIndexP2PKH_BACKUP');
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId,
|
|
// // key: 'receivingAddressesP2PKH',
|
|
// // value: tempReceivingAddressesP2PKH);
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId,
|
|
// // key: 'changeAddressesP2PKH',
|
|
// // value: tempChangeAddressesP2PKH);
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId,
|
|
// // key: 'receivingIndexP2PKH',
|
|
// // value: tempReceivingIndexP2PKH);
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId,
|
|
// // key: 'changeIndexP2PKH',
|
|
// // value: tempChangeIndexP2PKH);
|
|
// // await DB.instance.delete<dynamic>(
|
|
// // key: 'receivingAddressesP2PKH_BACKUP', boxName: walletId);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'changeAddressesP2PKH_BACKUP', boxName: walletId);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'receivingIndexP2PKH_BACKUP', boxName: walletId);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'changeIndexP2PKH_BACKUP', boxName: walletId);
|
|
// //
|
|
// // // P2PKH derivations
|
|
// // final p2pkhReceiveDerivationsString = await _secureStore.read(
|
|
// // key: "${walletId}_receiveDerivationsP2PKH_BACKUP");
|
|
// // final p2pkhChangeDerivationsString = await _secureStore.read(
|
|
// // key: "${walletId}_changeDerivationsP2PKH_BACKUP");
|
|
// //
|
|
// // await _secureStore.write(
|
|
// // key: "${walletId}_receiveDerivationsP2PKH",
|
|
// // value: p2pkhReceiveDerivationsString);
|
|
// // await _secureStore.write(
|
|
// // key: "${walletId}_changeDerivationsP2PKH",
|
|
// // value: p2pkhChangeDerivationsString);
|
|
// //
|
|
// // await _secureStore.delete(
|
|
// // key: "${walletId}_receiveDerivationsP2PKH_BACKUP");
|
|
// // await _secureStore.delete(key: "${walletId}_changeDerivationsP2PKH_BACKUP");
|
|
// //
|
|
// // // UTXOs
|
|
// // final utxoData = DB.instance
|
|
// // .get<dynamic>(boxName: walletId, key: 'latest_utxo_model_BACKUP');
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId, key: 'latest_utxo_model', value: utxoData);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'latest_utxo_model_BACKUP', boxName: walletId);
|
|
// //
|
|
// // Logging.instance.log("rescan restore complete", level: LogLevel.Info);
|
|
// // }
|
|
// //
|
|
// // Future<void> _rescanBackup() async {
|
|
// // Logging.instance.log("starting rescan backup", level: LogLevel.Info);
|
|
// //
|
|
// // // backup current and clear data
|
|
// // // p2pkh
|
|
// // final tempReceivingAddressesP2PKH = DB.instance
|
|
// // .get<dynamic>(boxName: walletId, key: 'receivingAddressesP2PKH');
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId,
|
|
// // key: 'receivingAddressesP2PKH_BACKUP',
|
|
// // value: tempReceivingAddressesP2PKH);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'receivingAddressesP2PKH', boxName: walletId);
|
|
// //
|
|
// // final tempChangeAddressesP2PKH = DB.instance
|
|
// // .get<dynamic>(boxName: walletId, key: 'changeAddressesP2PKH');
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId,
|
|
// // key: 'changeAddressesP2PKH_BACKUP',
|
|
// // value: tempChangeAddressesP2PKH);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'changeAddressesP2PKH', boxName: walletId);
|
|
// //
|
|
// // final tempReceivingIndexP2PKH =
|
|
// // DB.instance.get<dynamic>(boxName: walletId, key: 'receivingIndexP2PKH');
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId,
|
|
// // key: 'receivingIndexP2PKH_BACKUP',
|
|
// // value: tempReceivingIndexP2PKH);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'receivingIndexP2PKH', boxName: walletId);
|
|
// //
|
|
// // final tempChangeIndexP2PKH =
|
|
// // DB.instance.get<dynamic>(boxName: walletId, key: 'changeIndexP2PKH');
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId,
|
|
// // key: 'changeIndexP2PKH_BACKUP',
|
|
// // value: tempChangeIndexP2PKH);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'changeIndexP2PKH', boxName: walletId);
|
|
// //
|
|
// // // P2PKH derivations
|
|
// // final p2pkhReceiveDerivationsString =
|
|
// // await _secureStore.read(key: "${walletId}_receiveDerivationsP2PKH");
|
|
// // final p2pkhChangeDerivationsString =
|
|
// // await _secureStore.read(key: "${walletId}_changeDerivationsP2PKH");
|
|
// //
|
|
// // await _secureStore.write(
|
|
// // key: "${walletId}_receiveDerivationsP2PKH_BACKUP",
|
|
// // value: p2pkhReceiveDerivationsString);
|
|
// // await _secureStore.write(
|
|
// // key: "${walletId}_changeDerivationsP2PKH_BACKUP",
|
|
// // value: p2pkhChangeDerivationsString);
|
|
// //
|
|
// // await _secureStore.delete(key: "${walletId}_receiveDerivationsP2PKH");
|
|
// // await _secureStore.delete(key: "${walletId}_changeDerivationsP2PKH");
|
|
// //
|
|
// // // UTXOs
|
|
// // final utxoData =
|
|
// // DB.instance.get<dynamic>(boxName: walletId, key: 'latest_utxo_model');
|
|
// // await DB.instance.put<dynamic>(
|
|
// // boxName: walletId, key: 'latest_utxo_model_BACKUP', value: utxoData);
|
|
// // await DB.instance
|
|
// // .delete<dynamic>(key: 'latest_utxo_model', boxName: walletId);
|
|
// //
|
|
// // Logging.instance.log("rescan backup complete", level: LogLevel.Info);
|
|
// // }
|
|
//
|
|
// @override
|
|
// set isFavorite(bool markFavorite) {
|
|
// _isFavorite = markFavorite;
|
|
// updateCachedIsFavorite(markFavorite);
|
|
// }
|
|
//
|
|
// @override
|
|
// bool get isFavorite => _isFavorite ??= getCachedIsFavorite();
|
|
//
|
|
// bool? _isFavorite;
|
|
//
|
|
// @override
|
|
// bool get isRefreshing => refreshMutex;
|
|
//
|
|
// bool isActive = false;
|
|
//
|
|
// @override
|
|
// void Function(bool)? get onIsActiveWalletChanged =>
|
|
// (isActive) => this.isActive = isActive;
|
|
//
|
|
// @override
|
|
// Future<Amount> estimateFeeFor(Amount amount, int feeRate) async {
|
|
// final available = balance.spendable;
|
|
//
|
|
// if (available == amount) {
|
|
// return amount - (await sweepAllEstimate(feeRate));
|
|
// } else if (amount <= Amount.zero || amount > available) {
|
|
// return roughFeeEstimate(1, 2, feeRate);
|
|
// }
|
|
//
|
|
// Amount runningBalance = Amount(
|
|
// rawValue: BigInt.zero,
|
|
// fractionDigits: coin.decimals,
|
|
// );
|
|
// int inputCount = 0;
|
|
// for (final output in (await utxos)) {
|
|
// if (!output.isBlocked) {
|
|
// runningBalance = runningBalance +
|
|
// Amount(
|
|
// rawValue: BigInt.from(output.value),
|
|
// fractionDigits: coin.decimals,
|
|
// );
|
|
// inputCount++;
|
|
// if (runningBalance > amount) {
|
|
// break;
|
|
// }
|
|
// }
|
|
// }
|
|
//
|
|
// final oneOutPutFee = roughFeeEstimate(inputCount, 1, feeRate);
|
|
// final twoOutPutFee = roughFeeEstimate(inputCount, 2, feeRate);
|
|
//
|
|
// if (runningBalance - amount > oneOutPutFee) {
|
|
// if (runningBalance - amount > oneOutPutFee + DUST_LIMIT) {
|
|
// final change = runningBalance - amount - twoOutPutFee;
|
|
// if (change > DUST_LIMIT &&
|
|
// runningBalance - amount - change == twoOutPutFee) {
|
|
// return runningBalance - amount - change;
|
|
// } else {
|
|
// return runningBalance - amount;
|
|
// }
|
|
// } else {
|
|
// return runningBalance - amount;
|
|
// }
|
|
// } else if (runningBalance - amount == oneOutPutFee) {
|
|
// return oneOutPutFee;
|
|
// } else {
|
|
// return twoOutPutFee;
|
|
// }
|
|
// }
|
|
//
|
|
// // TODO: correct formula for doge?
|
|
// Amount roughFeeEstimate(int inputCount, int outputCount, int feeRatePerKB) {
|
|
// return Amount(
|
|
// rawValue: BigInt.from(((181 * inputCount) + (34 * outputCount) + 10) *
|
|
// (feeRatePerKB / 1000).ceil()),
|
|
// fractionDigits: coin.decimals,
|
|
// );
|
|
// }
|
|
//
|
|
// Future<Amount> sweepAllEstimate(int feeRate) async {
|
|
// int available = 0;
|
|
// int inputCount = 0;
|
|
// for (final output in (await utxos)) {
|
|
// if (!output.isBlocked &&
|
|
// output.isConfirmed(storedChainHeight, MINIMUM_CONFIRMATIONS)) {
|
|
// available += output.value;
|
|
// inputCount++;
|
|
// }
|
|
// }
|
|
//
|
|
// // transaction will only have 1 output minus the fee
|
|
// final estimatedFee = roughFeeEstimate(inputCount, 1, feeRate);
|
|
//
|
|
// return Amount(
|
|
// rawValue: BigInt.from(available),
|
|
// fractionDigits: coin.decimals,
|
|
// ) -
|
|
// estimatedFee;
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<bool> generateNewAddress() async {
|
|
// try {
|
|
// final currentReceiving = await _currentReceivingAddress;
|
|
//
|
|
// final newReceivingIndex = currentReceiving.derivationIndex + 1;
|
|
//
|
|
// // Use new index to derive a new receiving address
|
|
// final newReceivingAddress = await _generateAddressForChain(
|
|
// 0, newReceivingIndex, DerivePathTypeExt.primaryFor(coin));
|
|
//
|
|
// // Add that new receiving address
|
|
// await db.putAddress(newReceivingAddress);
|
|
//
|
|
// return true;
|
|
// } catch (e, s) {
|
|
// Logging.instance.log(
|
|
// "Exception rethrown from generateNewAddress(): $e\n$s",
|
|
// level: LogLevel.Error);
|
|
// return false;
|
|
// }
|
|
// }
|
|
//
|
|
// @override
|
|
// Future<String> get xpub async {
|
|
// final node = await Bip32Utils.getBip32Root(
|
|
// (await mnemonic).join(" "),
|
|
// await mnemonicPassphrase ?? "",
|
|
// network,
|
|
// );
|
|
//
|
|
// return node.neutered().toBase58();
|
|
// }
|
|
// }
|