2022-08-26 08:11:35 +00:00
|
|
|
import 'dart:async';
|
|
|
|
import 'dart:convert';
|
|
|
|
import 'dart:io';
|
|
|
|
import 'dart:isolate';
|
|
|
|
import 'dart:math';
|
|
|
|
|
|
|
|
import 'package:bip32/bip32.dart' as bip32;
|
|
|
|
import 'package:bip39/bip39.dart' as bip39;
|
|
|
|
import 'package:bitcoindart/bitcoindart.dart';
|
|
|
|
import 'package:decimal/decimal.dart';
|
|
|
|
import 'package:flutter/foundation.dart';
|
2023-01-12 00:16:17 +00:00
|
|
|
import 'package:isar/isar.dart';
|
2022-08-26 08:11:35 +00:00
|
|
|
import 'package:lelantus/lelantus.dart';
|
|
|
|
import 'package:stackwallet/electrumx_rpc/cached_electrumx.dart';
|
|
|
|
import 'package:stackwallet/electrumx_rpc/electrumx.dart';
|
2023-01-12 00:16:17 +00:00
|
|
|
import 'package:stackwallet/models/balance.dart';
|
|
|
|
import 'package:stackwallet/models/isar/models/isar_models.dart' as isar_models;
|
2022-08-26 08:11:35 +00:00
|
|
|
import 'package:stackwallet/models/lelantus_coin.dart';
|
|
|
|
import 'package:stackwallet/models/lelantus_fee_data.dart';
|
|
|
|
import 'package:stackwallet/models/paymint/fee_object_model.dart';
|
2023-01-12 00:16:17 +00:00
|
|
|
import 'package:stackwallet/services/coins/coin_paynym_extension.dart';
|
2022-08-26 08:11:35 +00:00
|
|
|
import 'package:stackwallet/services/coins/coin_service.dart';
|
|
|
|
import 'package:stackwallet/services/event_bus/events/global/node_connection_status_changed_event.dart';
|
|
|
|
import 'package:stackwallet/services/event_bus/events/global/refresh_percent_changed_event.dart';
|
|
|
|
import 'package:stackwallet/services/event_bus/events/global/updated_in_background_event.dart';
|
|
|
|
import 'package:stackwallet/services/event_bus/events/global/wallet_sync_status_changed_event.dart';
|
|
|
|
import 'package:stackwallet/services/event_bus/global_event_bus.dart';
|
2023-01-12 18:46:01 +00:00
|
|
|
import 'package:stackwallet/services/mixins/wallet_cache.dart';
|
|
|
|
import 'package:stackwallet/services/mixins/wallet_db.dart';
|
2022-08-26 08:11:35 +00:00
|
|
|
import 'package:stackwallet/services/node_service.dart';
|
|
|
|
import 'package:stackwallet/services/notifications_api.dart';
|
|
|
|
import 'package:stackwallet/services/transaction_notification_tracker.dart';
|
|
|
|
import 'package:stackwallet/utilities/address_utils.dart';
|
|
|
|
import 'package:stackwallet/utilities/assets.dart';
|
|
|
|
import 'package:stackwallet/utilities/constants.dart';
|
|
|
|
import 'package:stackwallet/utilities/default_nodes.dart';
|
|
|
|
import 'package:stackwallet/utilities/enums/coin_enum.dart';
|
2022-09-07 15:56:10 +00:00
|
|
|
import 'package:stackwallet/utilities/enums/fee_rate_type_enum.dart';
|
2022-08-26 08:11:35 +00:00
|
|
|
import 'package:stackwallet/utilities/flutter_secure_storage_interface.dart';
|
|
|
|
import 'package:stackwallet/utilities/format.dart';
|
|
|
|
import 'package:stackwallet/utilities/logger.dart';
|
|
|
|
import 'package:stackwallet/utilities/prefs.dart';
|
|
|
|
import 'package:tuple/tuple.dart';
|
2022-09-09 15:07:27 +00:00
|
|
|
import 'package:uuid/uuid.dart';
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
import '../../mixins/firo_hive.dart';
|
|
|
|
|
2022-09-07 15:56:10 +00:00
|
|
|
const DUST_LIMIT = 1000;
|
2022-08-26 08:11:35 +00:00
|
|
|
const MINIMUM_CONFIRMATIONS = 1;
|
|
|
|
const MINT_LIMIT = 100100000000;
|
|
|
|
const int LELANTUS_VALUE_SPEND_LIMIT_PER_TRANSACTION = 5001 * 100000000;
|
|
|
|
|
|
|
|
const JMINT_INDEX = 5;
|
|
|
|
const MINT_INDEX = 2;
|
|
|
|
const TRANSACTION_LELANTUS = 8;
|
|
|
|
const ANONYMITY_SET_EMPTY_ID = 0;
|
|
|
|
|
|
|
|
const String GENESIS_HASH_MAINNET =
|
|
|
|
"4381deb85b1b2c9843c222944b616d997516dcbd6a964e1eaf0def0830695233";
|
|
|
|
const String GENESIS_HASH_TESTNET =
|
|
|
|
"aa22adcc12becaf436027ffe62a8fb21b234c58c23865291e5dc52cf53f64fca";
|
|
|
|
|
|
|
|
final firoNetwork = NetworkType(
|
|
|
|
messagePrefix: '\x18Zcoin Signed Message:\n',
|
|
|
|
bech32: 'bc',
|
|
|
|
bip32: Bip32Type(public: 0x0488b21e, private: 0x0488ade4),
|
|
|
|
pubKeyHash: 0x52,
|
|
|
|
scriptHash: 0x07,
|
|
|
|
wif: 0xd2);
|
|
|
|
|
|
|
|
final firoTestNetwork = NetworkType(
|
|
|
|
messagePrefix: '\x18Zcoin Signed Message:\n',
|
|
|
|
bech32: 'bc',
|
|
|
|
bip32: Bip32Type(public: 0x043587cf, private: 0x04358394),
|
|
|
|
pubKeyHash: 0x41,
|
|
|
|
scriptHash: 0xb2,
|
|
|
|
wif: 0xb9);
|
|
|
|
|
|
|
|
// isolate
|
|
|
|
|
|
|
|
Map<ReceivePort, Isolate> isolates = {};
|
|
|
|
|
|
|
|
Future<ReceivePort> getIsolate(Map<String, dynamic> arguments) async {
|
|
|
|
ReceivePort receivePort =
|
|
|
|
ReceivePort(); //port for isolate to receive messages.
|
|
|
|
arguments['sendPort'] = receivePort.sendPort;
|
|
|
|
Logging.instance
|
|
|
|
.log("starting isolate ${arguments['function']}", level: LogLevel.Info);
|
|
|
|
Isolate isolate = await Isolate.spawn(executeNative, arguments);
|
|
|
|
Logging.instance.log("isolate spawned!", level: LogLevel.Info);
|
|
|
|
isolates[receivePort] = isolate;
|
|
|
|
return receivePort;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<void> executeNative(Map<String, dynamic> arguments) async {
|
|
|
|
await Logging.instance.initInIsolate();
|
|
|
|
final sendPort = arguments['sendPort'] as SendPort;
|
|
|
|
final function = arguments['function'] as String;
|
|
|
|
try {
|
|
|
|
if (function == "createJoinSplit") {
|
|
|
|
final spendAmount = arguments['spendAmount'] as int;
|
|
|
|
final address = arguments['address'] as String;
|
|
|
|
final subtractFeeFromAmount = arguments['subtractFeeFromAmount'] as bool;
|
|
|
|
final mnemonic = arguments['mnemonic'] as String;
|
|
|
|
final index = arguments['index'] as int;
|
|
|
|
final price = arguments['price'] as Decimal;
|
|
|
|
final lelantusEntries =
|
|
|
|
arguments['lelantusEntries'] as List<DartLelantusEntry>;
|
|
|
|
final coin = arguments['coin'] as Coin;
|
|
|
|
final network = arguments['network'] as NetworkType?;
|
|
|
|
final locktime = arguments['locktime'] as int;
|
|
|
|
final anonymitySets = arguments['_anonymity_sets'] as List<Map>?;
|
|
|
|
final locale = arguments["locale"] as String;
|
|
|
|
if (!(network == null || anonymitySets == null)) {
|
|
|
|
var joinSplit = await isolateCreateJoinSplitTransaction(
|
|
|
|
spendAmount,
|
|
|
|
address,
|
|
|
|
subtractFeeFromAmount,
|
|
|
|
mnemonic,
|
|
|
|
index,
|
|
|
|
price,
|
|
|
|
lelantusEntries,
|
|
|
|
locktime,
|
|
|
|
coin,
|
|
|
|
network,
|
|
|
|
anonymitySets,
|
|
|
|
locale);
|
|
|
|
sendPort.send(joinSplit);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
} else if (function == "estimateJoinSplit") {
|
|
|
|
final spendAmount = arguments['spendAmount'] as int;
|
|
|
|
final subtractFeeFromAmount = arguments['subtractFeeFromAmount'] as bool?;
|
|
|
|
final lelantusEntries =
|
|
|
|
arguments['lelantusEntries'] as List<DartLelantusEntry>;
|
|
|
|
final coin = arguments['coin'] as Coin;
|
|
|
|
|
|
|
|
if (!(subtractFeeFromAmount == null)) {
|
|
|
|
var feeData = await isolateEstimateJoinSplitFee(
|
|
|
|
spendAmount, subtractFeeFromAmount, lelantusEntries, coin);
|
|
|
|
sendPort.send(feeData);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
} else if (function == "restore") {
|
|
|
|
final latestSetId = arguments['latestSetId'] as int;
|
|
|
|
final setDataMap = arguments['setDataMap'] as Map;
|
|
|
|
final usedSerialNumbers = arguments['usedSerialNumbers'] as List?;
|
|
|
|
final mnemonic = arguments['mnemonic'] as String;
|
|
|
|
final coin = arguments['coin'] as Coin;
|
|
|
|
final network = arguments['network'] as NetworkType?;
|
|
|
|
if (!(usedSerialNumbers == null || network == null)) {
|
|
|
|
var restoreData = await isolateRestore(
|
2022-09-09 15:07:27 +00:00
|
|
|
mnemonic,
|
|
|
|
coin,
|
|
|
|
latestSetId,
|
|
|
|
setDataMap,
|
|
|
|
usedSerialNumbers,
|
|
|
|
network,
|
|
|
|
);
|
2022-08-26 08:11:35 +00:00
|
|
|
sendPort.send(restoreData);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
} else if (function == "isolateDerive") {
|
|
|
|
final mnemonic = arguments['mnemonic'] as String;
|
|
|
|
final from = arguments['from'] as int;
|
|
|
|
final to = arguments['to'] as int;
|
|
|
|
final network = arguments['network'] as NetworkType?;
|
|
|
|
if (!(network == null)) {
|
|
|
|
var derived = await isolateDerive(mnemonic, from, to, network);
|
|
|
|
sendPort.send(derived);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Logging.instance.log(
|
|
|
|
"Error Arguments for $function not formatted correctly",
|
|
|
|
level: LogLevel.Fatal);
|
|
|
|
sendPort.send("Error");
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"An error was thrown in this isolate $function: $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
sendPort.send("Error");
|
|
|
|
} finally {
|
2022-09-06 15:34:39 +00:00
|
|
|
await Logging.instance.isar?.close();
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void stop(ReceivePort port) {
|
|
|
|
Isolate? isolate = isolates.remove(port);
|
|
|
|
if (isolate != null) {
|
|
|
|
Logging.instance.log('Stopping Isolate...', level: LogLevel.Info);
|
|
|
|
isolate.kill(priority: Isolate.immediate);
|
|
|
|
isolate = null;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<Map<String, dynamic>> isolateDerive(
|
|
|
|
String mnemonic, int from, int to, NetworkType _network) async {
|
|
|
|
Map<String, dynamic> result = {};
|
|
|
|
Map<String, dynamic> allReceive = {};
|
|
|
|
Map<String, dynamic> allChange = {};
|
|
|
|
final root = getBip32Root(mnemonic, _network);
|
|
|
|
for (int i = from; i < to; i++) {
|
|
|
|
var currentNode = getBip32NodeFromRoot(0, i, root);
|
|
|
|
var address = P2PKH(
|
|
|
|
network: _network, data: PaymentData(pubkey: currentNode.publicKey))
|
|
|
|
.data
|
|
|
|
.address!;
|
|
|
|
allReceive["$i"] = {
|
|
|
|
"publicKey": Format.uint8listToString(currentNode.publicKey),
|
|
|
|
"wif": currentNode.toWIF(),
|
|
|
|
"address": address,
|
|
|
|
};
|
|
|
|
|
|
|
|
currentNode = getBip32NodeFromRoot(1, i, root);
|
|
|
|
address = P2PKH(
|
|
|
|
network: _network, data: PaymentData(pubkey: currentNode.publicKey))
|
|
|
|
.data
|
|
|
|
.address!;
|
|
|
|
allChange["$i"] = {
|
|
|
|
"publicKey": Format.uint8listToString(currentNode.publicKey),
|
|
|
|
"wif": currentNode.toWIF(),
|
|
|
|
"address": address,
|
|
|
|
};
|
|
|
|
if (i % 50 == 0) {
|
|
|
|
Logging.instance.log("thread at $i", level: LogLevel.Info);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
result['receive'] = allReceive;
|
|
|
|
result['change'] = allChange;
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<Map<String, dynamic>> isolateRestore(
|
|
|
|
String mnemonic,
|
|
|
|
Coin coin,
|
|
|
|
int _latestSetId,
|
|
|
|
Map<dynamic, dynamic> _setDataMap,
|
|
|
|
List<dynamic> _usedSerialNumbers,
|
|
|
|
NetworkType network,
|
|
|
|
) async {
|
|
|
|
List<int> jindexes = [];
|
|
|
|
List<Map<dynamic, LelantusCoin>> lelantusCoins = [];
|
|
|
|
|
|
|
|
final List<String> spendTxIds = [];
|
|
|
|
var lastFoundIndex = 0;
|
|
|
|
var currentIndex = 0;
|
|
|
|
|
|
|
|
try {
|
|
|
|
final usedSerialNumbers = _usedSerialNumbers;
|
|
|
|
Set<dynamic> usedSerialNumbersSet = {};
|
|
|
|
for (int ind = 0; ind < usedSerialNumbers.length; ind++) {
|
|
|
|
usedSerialNumbersSet.add(usedSerialNumbers[ind]);
|
|
|
|
}
|
|
|
|
|
|
|
|
final root = getBip32Root(mnemonic, network);
|
2022-10-18 17:28:14 +00:00
|
|
|
while (currentIndex < lastFoundIndex + 50) {
|
2022-08-26 08:11:35 +00:00
|
|
|
final mintKeyPair = getBip32NodeFromRoot(MINT_INDEX, currentIndex, root);
|
|
|
|
final mintTag = CreateTag(
|
|
|
|
Format.uint8listToString(mintKeyPair.privateKey!),
|
|
|
|
currentIndex,
|
|
|
|
Format.uint8listToString(mintKeyPair.identifier),
|
|
|
|
isTestnet: coin == Coin.firoTestNet);
|
|
|
|
|
|
|
|
for (var setId = 1; setId <= _latestSetId; setId++) {
|
|
|
|
final setData = _setDataMap[setId];
|
|
|
|
final foundCoin = setData["coins"].firstWhere(
|
|
|
|
(dynamic e) => e[1] == mintTag,
|
|
|
|
orElse: () => <Object>[]);
|
|
|
|
|
|
|
|
if (foundCoin.length == 4) {
|
|
|
|
lastFoundIndex = currentIndex;
|
|
|
|
if (foundCoin[2] is int) {
|
|
|
|
final amount = foundCoin[2] as int;
|
|
|
|
final serialNumber = GetSerialNumber(amount,
|
|
|
|
Format.uint8listToString(mintKeyPair.privateKey!), currentIndex,
|
|
|
|
isTestnet: coin == Coin.firoTestNet);
|
|
|
|
String publicCoin = foundCoin[0] as String;
|
|
|
|
String txId = foundCoin[3] as String;
|
|
|
|
bool isUsed = usedSerialNumbersSet.contains(serialNumber);
|
|
|
|
final duplicateCoin = lelantusCoins.firstWhere((element) {
|
|
|
|
final coin = element.values.first;
|
|
|
|
return coin.txId == txId &&
|
|
|
|
coin.index == currentIndex &&
|
|
|
|
coin.anonymitySetId != setId;
|
|
|
|
}, orElse: () => {});
|
|
|
|
if (duplicateCoin.isNotEmpty) {
|
2022-12-13 00:17:02 +00:00
|
|
|
//todo: check if print needed
|
|
|
|
// debugPrint("removing duplicate: $duplicateCoin");
|
2022-08-26 08:11:35 +00:00
|
|
|
lelantusCoins.remove(duplicateCoin);
|
|
|
|
}
|
|
|
|
lelantusCoins.add({
|
|
|
|
publicCoin: LelantusCoin(
|
|
|
|
currentIndex,
|
|
|
|
amount,
|
|
|
|
publicCoin,
|
|
|
|
txId,
|
|
|
|
setId,
|
|
|
|
isUsed,
|
|
|
|
)
|
|
|
|
});
|
|
|
|
Logging.instance
|
|
|
|
.log("amount $amount used $isUsed", level: LogLevel.Info);
|
|
|
|
} else {
|
|
|
|
final keyPath = GetAesKeyPath(foundCoin[0] as String);
|
|
|
|
final aesKeyPair = getBip32NodeFromRoot(JMINT_INDEX, keyPath, root);
|
|
|
|
if (aesKeyPair.privateKey != null) {
|
|
|
|
final aesPrivateKey =
|
|
|
|
Format.uint8listToString(aesKeyPair.privateKey!);
|
|
|
|
final amount = decryptMintAmount(
|
|
|
|
aesPrivateKey,
|
|
|
|
foundCoin[2] as String,
|
|
|
|
);
|
|
|
|
|
|
|
|
final serialNumber = GetSerialNumber(
|
|
|
|
amount,
|
|
|
|
Format.uint8listToString(mintKeyPair.privateKey!),
|
|
|
|
currentIndex,
|
|
|
|
isTestnet: coin == Coin.firoTestNet);
|
|
|
|
String publicCoin = foundCoin[0] as String;
|
|
|
|
String txId = foundCoin[3] as String;
|
|
|
|
bool isUsed = usedSerialNumbersSet.contains(serialNumber);
|
|
|
|
final duplicateCoin = lelantusCoins.firstWhere((element) {
|
|
|
|
final coin = element.values.first;
|
|
|
|
return coin.txId == txId &&
|
|
|
|
coin.index == currentIndex &&
|
|
|
|
coin.anonymitySetId != setId;
|
|
|
|
}, orElse: () => {});
|
|
|
|
if (duplicateCoin.isNotEmpty) {
|
2022-12-13 00:17:02 +00:00
|
|
|
//todo: check if print needed
|
|
|
|
// debugPrint("removing duplicate: $duplicateCoin");
|
2022-08-26 08:11:35 +00:00
|
|
|
lelantusCoins.remove(duplicateCoin);
|
|
|
|
}
|
|
|
|
lelantusCoins.add({
|
|
|
|
'${foundCoin[3]!}': LelantusCoin(
|
|
|
|
currentIndex,
|
|
|
|
amount,
|
|
|
|
publicCoin,
|
|
|
|
txId,
|
|
|
|
setId,
|
|
|
|
isUsed,
|
|
|
|
)
|
|
|
|
});
|
|
|
|
jindexes.add(currentIndex);
|
|
|
|
|
|
|
|
spendTxIds.add(foundCoin[3] as String);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
currentIndex++;
|
|
|
|
}
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown from isolateRestore(): $e\n$s",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
|
|
|
|
Map<String, dynamic> result = {};
|
|
|
|
// Logging.instance.log("mints $lelantusCoins", addToDebugMessagesDB: false);
|
|
|
|
// Logging.instance.log("jmints $spendTxIds", addToDebugMessagesDB: false);
|
|
|
|
|
|
|
|
result['_lelantus_coins'] = lelantusCoins;
|
|
|
|
result['mintIndex'] = lastFoundIndex + 1;
|
|
|
|
result['jindex'] = jindexes;
|
2022-09-09 15:07:27 +00:00
|
|
|
result['spendTxIds'] = spendTxIds;
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<Map<dynamic, dynamic>> staticProcessRestore(
|
2023-01-12 00:16:17 +00:00
|
|
|
List<isar_models.Transaction> txns,
|
2022-09-09 15:07:27 +00:00
|
|
|
Map<dynamic, dynamic> result,
|
2023-01-12 00:16:17 +00:00
|
|
|
int currentHeight,
|
2022-09-09 15:07:27 +00:00
|
|
|
) async {
|
|
|
|
List<dynamic>? _l = result['_lelantus_coins'] as List?;
|
|
|
|
final List<Map<dynamic, LelantusCoin>> lelantusCoins = [];
|
|
|
|
for (var el in _l ?? []) {
|
|
|
|
lelantusCoins.add({el.keys.first: el.values.first as LelantusCoin});
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
// Edit the receive transactions with the mint fees.
|
2023-01-12 00:16:17 +00:00
|
|
|
Map<String, isar_models.Transaction> editedTransactions =
|
|
|
|
<String, isar_models.Transaction>{};
|
2022-09-06 15:34:39 +00:00
|
|
|
for (var item in lelantusCoins) {
|
2022-08-26 08:11:35 +00:00
|
|
|
item.forEach((key, value) {
|
|
|
|
String txid = value.txId;
|
2023-01-12 00:16:17 +00:00
|
|
|
isar_models.Transaction? tx;
|
|
|
|
try {
|
|
|
|
tx = txns.firstWhere((e) => e.txid == txid);
|
|
|
|
} catch (_) {
|
|
|
|
tx = null;
|
|
|
|
}
|
|
|
|
|
2022-08-26 08:11:35 +00:00
|
|
|
if (tx == null) {
|
|
|
|
// This is a jmint.
|
|
|
|
return;
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
List<isar_models.Transaction> inputs = [];
|
2022-08-26 08:11:35 +00:00
|
|
|
for (var element in tx.inputs) {
|
2023-01-12 00:16:17 +00:00
|
|
|
isar_models.Transaction? input;
|
|
|
|
try {
|
|
|
|
input = txns.firstWhere((e) => e.txid == element.txid);
|
|
|
|
} catch (_) {
|
|
|
|
input = null;
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
if (input != null) {
|
|
|
|
inputs.add(input);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (inputs.isEmpty) {
|
|
|
|
//some error.
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
int mintFee = tx.fee;
|
|
|
|
int sharedFee = mintFee ~/ inputs.length;
|
2022-08-26 08:11:35 +00:00
|
|
|
for (var element in inputs) {
|
2023-01-12 00:16:17 +00:00
|
|
|
editedTransactions[element.txid] = isar_models.Transaction()
|
|
|
|
..txid = element.txid
|
|
|
|
..timestamp = element.timestamp
|
|
|
|
..type = element.type
|
|
|
|
..amount = element.amount
|
|
|
|
..fee = sharedFee
|
|
|
|
..inputs.addAll(element.inputs)
|
|
|
|
..outputs.addAll(element.outputs)
|
2023-01-13 17:11:41 +00:00
|
|
|
..address.value = element.address.value
|
2023-01-12 00:16:17 +00:00
|
|
|
..height = element.height
|
|
|
|
..subType = isar_models.TransactionSubType.mint
|
|
|
|
..otherData = txid
|
|
|
|
..isLelantus = true
|
|
|
|
..isCancelled = false;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
});
|
2022-09-06 15:34:39 +00:00
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
// Logging.instance.log(editedTransactions, addToDebugMessagesDB: false);
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Map<String, isar_models.Transaction> transactionMap = {};
|
|
|
|
for (final e in txns) {
|
|
|
|
transactionMap[e.txid] = e;
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
// Logging.instance.log(transactionMap, addToDebugMessagesDB: false);
|
|
|
|
|
|
|
|
editedTransactions.forEach((key, value) {
|
|
|
|
transactionMap.update(key, (_value) => value);
|
|
|
|
});
|
|
|
|
|
|
|
|
transactionMap.removeWhere((key, value) =>
|
|
|
|
lelantusCoins.any((element) => element.containsKey(key)) ||
|
2023-01-12 00:16:17 +00:00
|
|
|
((value.height == -1 || value.height == null) &&
|
|
|
|
!value.isConfirmed(currentHeight, MINIMUM_CONFIRMATIONS)));
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
result['newTxMap'] = transactionMap;
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<LelantusFeeData> isolateEstimateJoinSplitFee(
|
|
|
|
int spendAmount,
|
|
|
|
bool subtractFeeFromAmount,
|
|
|
|
List<DartLelantusEntry> lelantusEntries,
|
|
|
|
Coin coin) async {
|
|
|
|
Logging.instance.log("estimateJoinsplit fee", level: LogLevel.Info);
|
|
|
|
// for (int i = 0; i < lelantusEntries.length; i++) {
|
|
|
|
// Logging.instance.log(lelantusEntries[i], addToDebugMessagesDB: false);
|
|
|
|
// }
|
|
|
|
Logging.instance
|
|
|
|
.log("$spendAmount $subtractFeeFromAmount", level: LogLevel.Info);
|
|
|
|
|
|
|
|
List<int> changeToMint = List.empty(growable: true);
|
|
|
|
List<int> spendCoinIndexes = List.empty(growable: true);
|
|
|
|
// Logging.instance.log(lelantusEntries, addToDebugMessagesDB: false);
|
|
|
|
final fee = estimateFee(
|
|
|
|
spendAmount,
|
|
|
|
subtractFeeFromAmount,
|
|
|
|
lelantusEntries,
|
|
|
|
changeToMint,
|
|
|
|
spendCoinIndexes,
|
|
|
|
isTestnet: coin == Coin.firoTestNet,
|
|
|
|
);
|
|
|
|
|
|
|
|
final estimateFeeData =
|
|
|
|
LelantusFeeData(changeToMint[0], fee, spendCoinIndexes);
|
|
|
|
Logging.instance.log(
|
|
|
|
"estimateFeeData ${estimateFeeData.changeToMint} ${estimateFeeData.fee} ${estimateFeeData.spendCoinIndexes}",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
return estimateFeeData;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<dynamic> isolateCreateJoinSplitTransaction(
|
|
|
|
int spendAmount,
|
|
|
|
String address,
|
|
|
|
bool subtractFeeFromAmount,
|
|
|
|
String mnemonic,
|
|
|
|
int index,
|
|
|
|
Decimal price,
|
|
|
|
List<DartLelantusEntry> lelantusEntries,
|
|
|
|
int locktime,
|
|
|
|
Coin coin,
|
|
|
|
NetworkType _network,
|
|
|
|
List<Map<dynamic, dynamic>> anonymitySetsArg,
|
|
|
|
String locale,
|
|
|
|
) async {
|
|
|
|
final estimateJoinSplitFee = await isolateEstimateJoinSplitFee(
|
|
|
|
spendAmount, subtractFeeFromAmount, lelantusEntries, coin);
|
|
|
|
var changeToMint = estimateJoinSplitFee.changeToMint;
|
|
|
|
var fee = estimateJoinSplitFee.fee;
|
|
|
|
var spendCoinIndexes = estimateJoinSplitFee.spendCoinIndexes;
|
|
|
|
Logging.instance
|
|
|
|
.log("$changeToMint $fee $spendCoinIndexes", level: LogLevel.Info);
|
|
|
|
if (spendCoinIndexes.isEmpty) {
|
|
|
|
Logging.instance.log("Error, Not enough funds.", level: LogLevel.Error);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
final tx = TransactionBuilder(network: _network);
|
|
|
|
tx.setLockTime(locktime);
|
|
|
|
|
|
|
|
tx.setVersion(3 | (TRANSACTION_LELANTUS << 16));
|
|
|
|
|
|
|
|
tx.addInput(
|
|
|
|
'0000000000000000000000000000000000000000000000000000000000000000',
|
|
|
|
4294967295,
|
|
|
|
4294967295,
|
|
|
|
Uint8List(0),
|
|
|
|
);
|
|
|
|
|
|
|
|
final jmintKeyPair = getBip32Node(MINT_INDEX, index, mnemonic, _network);
|
|
|
|
|
|
|
|
final String jmintprivatekey =
|
|
|
|
Format.uint8listToString(jmintKeyPair.privateKey!);
|
|
|
|
|
|
|
|
final keyPath = getMintKeyPath(changeToMint, jmintprivatekey, index,
|
|
|
|
isTestnet: coin == Coin.firoTestNet);
|
|
|
|
|
|
|
|
final aesKeyPair = getBip32Node(JMINT_INDEX, keyPath, mnemonic, _network);
|
|
|
|
final aesPrivateKey = Format.uint8listToString(aesKeyPair.privateKey!);
|
|
|
|
|
|
|
|
final jmintData = createJMintScript(
|
|
|
|
changeToMint,
|
|
|
|
Format.uint8listToString(jmintKeyPair.privateKey!),
|
|
|
|
index,
|
|
|
|
Format.uint8listToString(jmintKeyPair.identifier),
|
|
|
|
aesPrivateKey,
|
|
|
|
isTestnet: coin == Coin.firoTestNet,
|
|
|
|
);
|
|
|
|
|
|
|
|
tx.addOutput(
|
|
|
|
Format.stringToUint8List(jmintData),
|
|
|
|
0,
|
|
|
|
);
|
|
|
|
|
|
|
|
int amount = spendAmount;
|
|
|
|
if (subtractFeeFromAmount) {
|
|
|
|
amount -= fee;
|
|
|
|
}
|
|
|
|
tx.addOutput(
|
|
|
|
address,
|
|
|
|
amount,
|
|
|
|
);
|
|
|
|
|
|
|
|
final extractedTx = tx.buildIncomplete();
|
|
|
|
extractedTx.setPayload(Uint8List(0));
|
|
|
|
final txHash = extractedTx.getId();
|
|
|
|
|
|
|
|
final List<int> setIds = [];
|
|
|
|
final List<List<String>> anonymitySets = [];
|
|
|
|
final List<String> anonymitySetHashes = [];
|
|
|
|
final List<String> groupBlockHashes = [];
|
|
|
|
for (var i = 0; i < lelantusEntries.length; i++) {
|
|
|
|
final anonymitySetId = lelantusEntries[i].anonymitySetId;
|
|
|
|
if (!setIds.contains(anonymitySetId)) {
|
|
|
|
setIds.add(anonymitySetId);
|
|
|
|
final anonymitySet = anonymitySetsArg.firstWhere(
|
|
|
|
(element) => element["setId"] == anonymitySetId,
|
|
|
|
orElse: () => <String, dynamic>{});
|
|
|
|
if (anonymitySet.isNotEmpty) {
|
|
|
|
anonymitySetHashes.add(anonymitySet['setHash'] as String);
|
|
|
|
groupBlockHashes.add(anonymitySet['blockHash'] as String);
|
|
|
|
List<String> list = [];
|
|
|
|
for (int i = 0; i < (anonymitySet['coins'] as List).length; i++) {
|
|
|
|
list.add(anonymitySet['coins'][i][0] as String);
|
|
|
|
}
|
|
|
|
anonymitySets.add(list);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
final String spendScript = createJoinSplitScript(
|
|
|
|
txHash,
|
|
|
|
spendAmount,
|
|
|
|
subtractFeeFromAmount,
|
|
|
|
Format.uint8listToString(jmintKeyPair.privateKey!),
|
|
|
|
index,
|
|
|
|
lelantusEntries,
|
|
|
|
setIds,
|
|
|
|
anonymitySets,
|
|
|
|
anonymitySetHashes,
|
|
|
|
groupBlockHashes,
|
|
|
|
isTestnet: coin == Coin.firoTestNet);
|
|
|
|
|
|
|
|
final finalTx = TransactionBuilder(network: _network);
|
|
|
|
finalTx.setLockTime(locktime);
|
|
|
|
|
|
|
|
finalTx.setVersion(3 | (TRANSACTION_LELANTUS << 16));
|
|
|
|
|
|
|
|
finalTx.addOutput(
|
|
|
|
Format.stringToUint8List(jmintData),
|
|
|
|
0,
|
|
|
|
);
|
|
|
|
|
|
|
|
finalTx.addOutput(
|
|
|
|
address,
|
|
|
|
amount,
|
|
|
|
);
|
|
|
|
|
|
|
|
final extTx = finalTx.buildIncomplete();
|
|
|
|
extTx.addInput(
|
|
|
|
Format.stringToUint8List(
|
|
|
|
'0000000000000000000000000000000000000000000000000000000000000000'),
|
|
|
|
4294967295,
|
|
|
|
4294967295,
|
|
|
|
Format.stringToUint8List("c9"),
|
|
|
|
);
|
|
|
|
debugPrint("spendscript: $spendScript");
|
|
|
|
extTx.setPayload(Format.stringToUint8List(spendScript));
|
|
|
|
|
|
|
|
final txHex = extTx.toHex();
|
|
|
|
final txId = extTx.getId();
|
|
|
|
Logging.instance.log("txid $txId", level: LogLevel.Info);
|
|
|
|
Logging.instance.log("txHex: $txHex", level: LogLevel.Info);
|
|
|
|
return {
|
|
|
|
"txid": txId,
|
|
|
|
"txHex": txHex,
|
|
|
|
"value": amount,
|
2022-11-23 18:31:31 +00:00
|
|
|
"fees": Format.satoshisToAmount(fee, coin: coin).toDouble(),
|
2022-08-26 08:11:35 +00:00
|
|
|
"fee": fee,
|
2022-09-24 16:16:56 +00:00
|
|
|
"vSize": extTx.virtualSize(),
|
2022-08-26 08:11:35 +00:00
|
|
|
"jmintValue": changeToMint,
|
|
|
|
"publicCoin": "jmintData.publicCoin",
|
|
|
|
"spendCoinIndexes": spendCoinIndexes,
|
|
|
|
"height": locktime,
|
|
|
|
"txType": "Sent",
|
|
|
|
"confirmed_status": false,
|
2022-11-23 18:31:31 +00:00
|
|
|
"amount": Format.satoshisToAmount(amount, coin: coin).toDouble(),
|
2022-08-26 08:11:35 +00:00
|
|
|
"recipientAmt": amount,
|
|
|
|
"worthNow": Format.localizedStringAsFixed(
|
|
|
|
value: ((Decimal.fromInt(amount) * price) /
|
2022-11-23 18:31:31 +00:00
|
|
|
Decimal.fromInt(Constants.satsPerCoin(coin)))
|
2022-08-26 08:11:35 +00:00
|
|
|
.toDecimal(scaleOnInfinitePrecision: 2),
|
|
|
|
decimalPlaces: 2,
|
|
|
|
locale: locale),
|
|
|
|
"address": address,
|
|
|
|
"timestamp": DateTime.now().millisecondsSinceEpoch ~/ 1000,
|
|
|
|
"subType": "join",
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<int> getBlockHead(ElectrumX client) async {
|
|
|
|
try {
|
|
|
|
final tip = await client.getBlockHeadTip();
|
|
|
|
return tip["height"] as int;
|
|
|
|
} catch (e) {
|
|
|
|
Logging.instance
|
|
|
|
.log("Exception rethrown in getBlockHead(): $e", level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// end of isolates
|
|
|
|
|
|
|
|
bip32.BIP32 getBip32Node(
|
|
|
|
int chain, int index, String mnemonic, NetworkType network) {
|
|
|
|
final root = getBip32Root(mnemonic, network);
|
|
|
|
|
|
|
|
final node = getBip32NodeFromRoot(chain, index, root);
|
|
|
|
return node;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// wrapper for compute()
|
|
|
|
bip32.BIP32 getBip32NodeWrapper(
|
|
|
|
Tuple4<int, int, String, NetworkType> args,
|
|
|
|
) {
|
|
|
|
return getBip32Node(
|
|
|
|
args.item1,
|
|
|
|
args.item2,
|
|
|
|
args.item3,
|
|
|
|
args.item4,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
bip32.BIP32 getBip32NodeFromRoot(int chain, int index, bip32.BIP32 root) {
|
|
|
|
String coinType;
|
|
|
|
switch (root.network.wif) {
|
|
|
|
case 0xd2: // firo mainnet wif
|
|
|
|
coinType = "136"; // firo mainnet
|
|
|
|
break;
|
|
|
|
case 0xb9: // firo testnet wif
|
|
|
|
coinType = "1"; // firo testnet
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
throw Exception("Invalid Bitcoin network type used!");
|
|
|
|
}
|
|
|
|
|
|
|
|
final node = root.derivePath("m/44'/$coinType'/0'/$chain/$index");
|
|
|
|
return node;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// wrapper for compute()
|
|
|
|
bip32.BIP32 getBip32NodeFromRootWrapper(
|
|
|
|
Tuple3<int, int, bip32.BIP32> args,
|
|
|
|
) {
|
|
|
|
return getBip32NodeFromRoot(
|
|
|
|
args.item1,
|
|
|
|
args.item2,
|
|
|
|
args.item3,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
bip32.BIP32 getBip32Root(String mnemonic, NetworkType network) {
|
|
|
|
final seed = bip39.mnemonicToSeed(mnemonic);
|
|
|
|
final firoNetworkType = bip32.NetworkType(
|
|
|
|
wif: network.wif,
|
|
|
|
bip32: bip32.Bip32Type(
|
|
|
|
public: network.bip32.public,
|
|
|
|
private: network.bip32.private,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
|
|
|
|
final root = bip32.BIP32.fromSeed(seed, firoNetworkType);
|
|
|
|
return root;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// wrapper for compute()
|
|
|
|
bip32.BIP32 getBip32RootWrapper(Tuple2<String, NetworkType> args) {
|
|
|
|
return getBip32Root(args.item1, args.item2);
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<String> _getMintScriptWrapper(
|
|
|
|
Tuple5<int, String, int, String, bool> data) async {
|
|
|
|
String mintHex = getMintScript(data.item1, data.item2, data.item3, data.item4,
|
|
|
|
isTestnet: data.item5);
|
|
|
|
return mintHex;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<void> _setTestnetWrapper(bool isTestnet) async {
|
2022-08-29 13:50:02 +00:00
|
|
|
// setTestnet(isTestnet);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/// Handles a single instance of a firo wallet
|
2023-01-12 21:20:57 +00:00
|
|
|
class FiroWallet extends CoinServiceAPI with WalletCache, WalletDB, FiroHive {
|
2022-08-26 08:11:35 +00:00
|
|
|
static const integrationTestFlag =
|
|
|
|
bool.fromEnvironment("IS_INTEGRATION_TEST");
|
|
|
|
|
|
|
|
final _prefs = Prefs.instance;
|
|
|
|
|
|
|
|
Timer? timer;
|
2023-01-12 21:20:57 +00:00
|
|
|
late final Coin _coin;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
bool _shouldAutoSync = false;
|
|
|
|
|
|
|
|
@override
|
|
|
|
bool get shouldAutoSync => _shouldAutoSync;
|
|
|
|
|
|
|
|
@override
|
|
|
|
set shouldAutoSync(bool shouldAutoSync) {
|
|
|
|
if (_shouldAutoSync != shouldAutoSync) {
|
|
|
|
_shouldAutoSync = shouldAutoSync;
|
|
|
|
if (!shouldAutoSync) {
|
|
|
|
timer?.cancel();
|
|
|
|
timer = null;
|
|
|
|
stopNetworkAlivePinging();
|
|
|
|
} else {
|
|
|
|
startNetworkAlivePinging();
|
|
|
|
refresh();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
NetworkType get _network {
|
|
|
|
switch (coin) {
|
|
|
|
case Coin.firo:
|
|
|
|
return firoNetwork;
|
|
|
|
case Coin.firoTestNet:
|
|
|
|
return firoTestNetwork;
|
|
|
|
default:
|
|
|
|
throw Exception("Invalid network type!");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
@override
|
|
|
|
set isFavorite(bool markFavorite) {
|
2023-01-12 21:20:57 +00:00
|
|
|
_isFavorite = markFavorite;
|
|
|
|
updateCachedIsFavorite(markFavorite);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
@override
|
2023-01-12 21:20:57 +00:00
|
|
|
bool get isFavorite => _isFavorite ??= getCachedIsFavorite();
|
|
|
|
|
|
|
|
bool? _isFavorite;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
@override
|
|
|
|
Coin get coin => _coin;
|
|
|
|
|
|
|
|
@override
|
|
|
|
Future<List<String>> get mnemonic => _getMnemonicList();
|
|
|
|
|
|
|
|
@override
|
|
|
|
bool validateAddress(String address) {
|
|
|
|
return Address.validateAddress(address, _network);
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Holds wallet transaction data
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<List<isar_models.Transaction>> get _txnData => isar.transactions
|
|
|
|
.filter()
|
|
|
|
.isLelantusIsNull()
|
|
|
|
.or()
|
|
|
|
.isLelantusEqualTo(false)
|
|
|
|
.findAll();
|
|
|
|
// _transactionData ??= _refreshTransactions();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// models.TransactionData? cachedTxData;
|
2022-11-07 16:24:08 +00:00
|
|
|
|
|
|
|
// hack to add tx to txData before refresh completes
|
|
|
|
// required based on current app architecture where we don't properly store
|
|
|
|
// transactions locally in a good way
|
|
|
|
@override
|
|
|
|
Future<void> updateSentCachedTxData(Map<String, dynamic> txData) async {
|
2023-01-12 00:16:17 +00:00
|
|
|
// final currentPrice = await firoPrice;
|
|
|
|
// final locale =
|
|
|
|
// Platform.isWindows ? "en_US" : await Devicelocale.currentLocale;
|
|
|
|
// final String worthNow = Format.localizedStringAsFixed(
|
|
|
|
// value:
|
|
|
|
// ((currentPrice * Decimal.fromInt(txData["recipientAmt"] as int)) /
|
|
|
|
// Decimal.fromInt(Constants.satsPerCoin(coin)))
|
|
|
|
// .toDecimal(scaleOnInfinitePrecision: 2),
|
|
|
|
// decimalPlaces: 2,
|
|
|
|
// locale: locale!);
|
|
|
|
//
|
|
|
|
// final tx = models.Transaction(
|
|
|
|
// txid: txData["txid"] as String,
|
|
|
|
// confirmedStatus: false,
|
|
|
|
// timestamp: DateTime.now().millisecondsSinceEpoch ~/ 1000,
|
|
|
|
// txType: "Sent",
|
|
|
|
// amount: txData["recipientAmt"] as int,
|
|
|
|
// worthNow: worthNow,
|
|
|
|
// worthAtBlockTimestamp: worthNow,
|
|
|
|
// fees: txData["fee"] as int,
|
|
|
|
// inputSize: 0,
|
|
|
|
// outputSize: 0,
|
|
|
|
// inputs: [],
|
|
|
|
// outputs: [],
|
|
|
|
// address: txData["address"] as String,
|
|
|
|
// height: -1,
|
|
|
|
// confirmations: 0,
|
|
|
|
// );
|
|
|
|
//
|
|
|
|
// if (cachedTxData == null) {
|
|
|
|
// final data = await _fetchTransactionData();
|
|
|
|
// _transactionData = Future(() => data);
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// final transactions = cachedTxData!.getAllTransactions();
|
|
|
|
// transactions[tx.txid] = tx;
|
|
|
|
// cachedTxData = models.TransactionData.fromMap(transactions);
|
|
|
|
// _transactionData = Future(() => cachedTxData!);
|
2022-11-07 16:24:08 +00:00
|
|
|
}
|
|
|
|
|
2022-08-26 08:11:35 +00:00
|
|
|
/// Holds wallet lelantus transaction data
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<List<isar_models.Transaction>> get lelantusTransactionData =>
|
|
|
|
isar.transactions.filter().isLelantusEqualTo(true).findAll();
|
|
|
|
// _lelantusTransactionData ??= _getLelantusTransactionData();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
/// Holds the max fee that can be sent
|
|
|
|
Future<int>? _maxFee;
|
|
|
|
@override
|
|
|
|
Future<int> get maxFee => _maxFee ??= _fetchMaxFee();
|
|
|
|
|
|
|
|
Future<FeeObject>? _feeObject;
|
|
|
|
@override
|
|
|
|
Future<FeeObject> get fees => _feeObject ??= _getFees();
|
|
|
|
|
|
|
|
@override
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<String> get currentReceivingAddress async =>
|
|
|
|
(await _currentReceivingAddress).value;
|
|
|
|
|
|
|
|
Future<isar_models.Address> get _currentReceivingAddress async =>
|
|
|
|
(await isar.addresses
|
|
|
|
.filter()
|
|
|
|
.typeEqualTo(isar_models.AddressType.p2pkh)
|
|
|
|
.subTypeEqualTo(isar_models.AddressSubType.receiving)
|
|
|
|
.sortByDerivationIndexDesc()
|
|
|
|
.findFirst())!;
|
|
|
|
|
|
|
|
Future<String> get currentChangeAddress async =>
|
|
|
|
(await _currentChangeAddress).value;
|
|
|
|
|
|
|
|
Future<isar_models.Address> get _currentChangeAddress async =>
|
|
|
|
(await isar.addresses
|
|
|
|
.filter()
|
|
|
|
.typeEqualTo(isar_models.AddressType.p2pkh)
|
|
|
|
.subTypeEqualTo(isar_models.AddressSubType.change)
|
|
|
|
.sortByDerivationIndexDesc()
|
|
|
|
.findFirst())!;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
late String _walletName;
|
|
|
|
@override
|
|
|
|
String get walletName => _walletName;
|
|
|
|
|
|
|
|
// setter for updating on rename
|
|
|
|
@override
|
|
|
|
set walletName(String newName) => _walletName = newName;
|
|
|
|
|
|
|
|
/// unique wallet id
|
2023-01-12 21:20:57 +00:00
|
|
|
late final String _walletId;
|
2022-08-26 08:11:35 +00:00
|
|
|
@override
|
|
|
|
String get walletId => _walletId;
|
|
|
|
|
|
|
|
@override
|
|
|
|
Future<bool> testNetworkConnection() async {
|
|
|
|
try {
|
|
|
|
final result = await _electrumXClient.ping();
|
|
|
|
return result;
|
|
|
|
} catch (_) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Timer? _networkAliveTimer;
|
|
|
|
|
|
|
|
void startNetworkAlivePinging() {
|
|
|
|
// call once on start right away
|
|
|
|
_periodicPingCheck();
|
|
|
|
|
|
|
|
// then periodically check
|
|
|
|
_networkAliveTimer = Timer.periodic(
|
|
|
|
Constants.networkAliveTimerDuration,
|
|
|
|
(_) async {
|
|
|
|
_periodicPingCheck();
|
|
|
|
},
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
void _periodicPingCheck() async {
|
|
|
|
bool hasNetwork = await testNetworkConnection();
|
|
|
|
_isConnected = hasNetwork;
|
|
|
|
if (_isConnected != hasNetwork) {
|
|
|
|
NodeConnectionStatus status = hasNetwork
|
|
|
|
? NodeConnectionStatus.connected
|
|
|
|
: NodeConnectionStatus.disconnected;
|
|
|
|
GlobalEventBus.instance
|
|
|
|
.fire(NodeConnectionStatusChangedEvent(status, walletId, coin));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void stopNetworkAlivePinging() {
|
|
|
|
_networkAliveTimer?.cancel();
|
|
|
|
_networkAliveTimer = null;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool _isConnected = false;
|
|
|
|
|
|
|
|
@override
|
|
|
|
bool get isConnected => _isConnected;
|
|
|
|
|
2022-09-07 15:56:10 +00:00
|
|
|
Future<Map<String, dynamic>> prepareSendPublic({
|
|
|
|
required String address,
|
|
|
|
required int satoshiAmount,
|
|
|
|
Map<String, dynamic>? args,
|
|
|
|
}) async {
|
|
|
|
try {
|
|
|
|
final feeRateType = args?["feeRate"];
|
|
|
|
final feeRateAmount = args?["feeRateAmount"];
|
|
|
|
if (feeRateType is FeeRateType || feeRateAmount is int) {
|
|
|
|
late final int rate;
|
|
|
|
if (feeRateType is FeeRateType) {
|
|
|
|
int fee = 0;
|
|
|
|
final feeObject = await fees;
|
|
|
|
switch (feeRateType) {
|
|
|
|
case FeeRateType.fast:
|
|
|
|
fee = feeObject.fast;
|
|
|
|
break;
|
|
|
|
case FeeRateType.average:
|
|
|
|
fee = feeObject.medium;
|
|
|
|
break;
|
|
|
|
case FeeRateType.slow:
|
|
|
|
fee = feeObject.slow;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
rate = fee;
|
|
|
|
} else {
|
|
|
|
rate = feeRateAmount as int;
|
|
|
|
}
|
|
|
|
|
|
|
|
// check for send all
|
|
|
|
bool isSendAll = false;
|
2023-01-12 00:16:17 +00:00
|
|
|
final balance =
|
|
|
|
Format.decimalAmountToSatoshis(availablePublicBalance(), coin);
|
2022-09-07 15:56:10 +00:00
|
|
|
if (satoshiAmount == balance) {
|
|
|
|
isSendAll = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
final txData =
|
|
|
|
await coinSelection(satoshiAmount, rate, address, isSendAll);
|
|
|
|
|
|
|
|
Logging.instance.log("prepare send: $txData", level: LogLevel.Info);
|
|
|
|
try {
|
|
|
|
if (txData is int) {
|
|
|
|
switch (txData) {
|
|
|
|
case 1:
|
|
|
|
throw Exception("Insufficient balance!");
|
|
|
|
case 2:
|
|
|
|
throw Exception(
|
|
|
|
"Insufficient funds to pay for transaction fee!");
|
|
|
|
default:
|
|
|
|
throw Exception("Transaction failed with error code $txData");
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
final hex = txData["hex"];
|
|
|
|
|
|
|
|
if (hex is String) {
|
|
|
|
final fee = txData["fee"] as int;
|
|
|
|
final vSize = txData["vSize"] as int;
|
|
|
|
|
|
|
|
Logging.instance
|
|
|
|
.log("prepared txHex: $hex", level: LogLevel.Info);
|
|
|
|
Logging.instance.log("prepared fee: $fee", level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log("prepared vSize: $vSize", level: LogLevel.Info);
|
|
|
|
|
|
|
|
// fee should never be less than vSize sanity check
|
|
|
|
if (fee < vSize) {
|
|
|
|
throw Exception(
|
|
|
|
"Error in fee calculation: Transaction fee cannot be less than vSize");
|
|
|
|
}
|
|
|
|
|
|
|
|
return txData as Map<String, dynamic>;
|
|
|
|
} else {
|
|
|
|
throw Exception("prepared hex is not a String!!!");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown from prepareSend(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
throw ArgumentError("Invalid fee rate argument provided!");
|
|
|
|
}
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown from prepareSend(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<String> confirmSendPublic({dynamic txData}) async {
|
|
|
|
try {
|
|
|
|
Logging.instance.log("confirmSend txData: $txData", level: LogLevel.Info);
|
|
|
|
final txHash = await _electrumXClient.broadcastTransaction(
|
|
|
|
rawTx: txData["hex"] as String);
|
|
|
|
Logging.instance.log("Sent txHash: $txHash", level: LogLevel.Info);
|
2022-11-07 16:24:08 +00:00
|
|
|
txData["txid"] = txHash;
|
|
|
|
// dirty ui update hack
|
|
|
|
await updateSentCachedTxData(txData as Map<String, dynamic>);
|
2022-09-07 15:56:10 +00:00
|
|
|
return txHash;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown from confirmSend(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-08-26 08:11:35 +00:00
|
|
|
@override
|
2022-09-07 15:56:10 +00:00
|
|
|
Future<Map<String, dynamic>> prepareSend({
|
|
|
|
required String address,
|
|
|
|
required int satoshiAmount,
|
|
|
|
Map<String, dynamic>? args,
|
|
|
|
}) async {
|
2022-08-26 08:11:35 +00:00
|
|
|
try {
|
2022-09-08 12:45:38 +00:00
|
|
|
// check for send all
|
|
|
|
bool isSendAll = false;
|
|
|
|
final balance =
|
2023-01-12 00:16:17 +00:00
|
|
|
Format.decimalAmountToSatoshis(availablePrivateBalance(), coin);
|
2022-09-08 12:45:38 +00:00
|
|
|
if (satoshiAmount == balance) {
|
2022-09-09 17:14:38 +00:00
|
|
|
// print("is send all");
|
2022-09-08 12:45:38 +00:00
|
|
|
isSendAll = true;
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
dynamic txHexOrError =
|
2022-09-08 12:45:38 +00:00
|
|
|
await _createJoinSplitTransaction(satoshiAmount, address, isSendAll);
|
2022-08-26 08:11:35 +00:00
|
|
|
Logging.instance.log("txHexOrError $txHexOrError", level: LogLevel.Error);
|
|
|
|
if (txHexOrError is int) {
|
|
|
|
// Here, we assume that transaction crafting returned an error
|
|
|
|
switch (txHexOrError) {
|
|
|
|
case 1:
|
|
|
|
throw Exception("Insufficient balance!");
|
|
|
|
default:
|
|
|
|
throw Exception("Error Creating Transaction!");
|
|
|
|
}
|
|
|
|
} else {
|
2022-09-24 16:16:56 +00:00
|
|
|
final fee = txHexOrError["fee"] as int;
|
|
|
|
final vSize = txHexOrError["vSize"] as int;
|
|
|
|
|
|
|
|
Logging.instance.log("prepared fee: $fee", level: LogLevel.Info);
|
|
|
|
Logging.instance.log("prepared vSize: $vSize", level: LogLevel.Info);
|
|
|
|
|
|
|
|
// fee should never be less than vSize sanity check
|
|
|
|
if (fee < vSize) {
|
|
|
|
throw Exception(
|
|
|
|
"Error in fee calculation: Transaction fee cannot be less than vSize");
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
return txHexOrError as Map<String, dynamic>;
|
|
|
|
}
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown in firo prepareSend(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
@override
|
|
|
|
Future<String> confirmSend({required Map<String, dynamic> txData}) async {
|
|
|
|
if (await _submitLelantusToNetwork(txData)) {
|
|
|
|
try {
|
|
|
|
final txid = txData["txid"] as String;
|
|
|
|
|
|
|
|
// temporarily update apdate available balance until a full refresh is done
|
|
|
|
|
|
|
|
// TODO: something here causes an exception to be thrown giving user false info that the tx failed
|
2023-01-12 00:16:17 +00:00
|
|
|
// Decimal sendTotal =
|
|
|
|
// Format.satoshisToAmount(txData["value"] as int, coin: coin);
|
|
|
|
// sendTotal += Decimal.parse(txData["fees"].toString());
|
|
|
|
|
|
|
|
// TODO: is this needed?
|
|
|
|
// final bals = await balances;
|
|
|
|
// bals[0] -= sendTotal;
|
|
|
|
// _balances = Future(() => bals);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
return txid;
|
|
|
|
} catch (e, s) {
|
2022-12-13 00:17:02 +00:00
|
|
|
//todo: come back to this
|
2022-08-26 08:11:35 +00:00
|
|
|
debugPrint("$e $s");
|
|
|
|
return txData["txid"] as String;
|
|
|
|
// don't throw anything here or it will tell the user that th tx
|
|
|
|
// failed even though it was successfully broadcast to network
|
|
|
|
// throw Exception("Transaction failed.");
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
//TODO provide more info
|
|
|
|
throw Exception("Transaction failed.");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// /// returns txid on successful send
|
|
|
|
// ///
|
|
|
|
// /// can throw
|
|
|
|
// @override
|
|
|
|
// Future<String> send({
|
|
|
|
// required String toAddress,
|
|
|
|
// required int amount,
|
|
|
|
// Map<String, String> args = const {},
|
|
|
|
// }) async {
|
|
|
|
// try {
|
|
|
|
// dynamic txHexOrError =
|
|
|
|
// await _createJoinSplitTransaction(amount, toAddress, false);
|
|
|
|
// Logging.instance.log("txHexOrError $txHexOrError", level: LogLevel.Error);
|
|
|
|
// if (txHexOrError is int) {
|
|
|
|
// // Here, we assume that transaction crafting returned an error
|
|
|
|
// switch (txHexOrError) {
|
|
|
|
// case 1:
|
|
|
|
// throw Exception("Insufficient balance!");
|
|
|
|
// default:
|
|
|
|
// throw Exception("Error Creating Transaction!");
|
|
|
|
// }
|
|
|
|
// } else {
|
|
|
|
// if (await _submitLelantusToNetwork(
|
|
|
|
// txHexOrError as Map<String, dynamic>)) {
|
|
|
|
// final txid = txHexOrError["txid"] as String;
|
|
|
|
//
|
|
|
|
// // temporarily update apdate available balance until a full refresh is done
|
|
|
|
// Decimal sendTotal =
|
|
|
|
// Format.satoshisToAmount(txHexOrError["value"] as int, coin: coin);
|
|
|
|
// sendTotal += Decimal.parse(txHexOrError["fees"].toString());
|
|
|
|
// final bals = await balances;
|
|
|
|
// bals[0] -= sendTotal;
|
|
|
|
// _balances = Future(() => bals);
|
|
|
|
//
|
|
|
|
// return txid;
|
|
|
|
// } else {
|
|
|
|
// //TODO provide more info
|
|
|
|
// throw Exception("Transaction failed.");
|
|
|
|
// }
|
|
|
|
// }
|
|
|
|
// } catch (e, s) {
|
|
|
|
// Logging.instance.log("Exception rethrown in firo send(): $e\n$s",
|
|
|
|
// level: LogLevel.Error);
|
|
|
|
// rethrow;
|
|
|
|
// }
|
|
|
|
// }
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
Future<List<String>> _getMnemonicList() async {
|
|
|
|
final mnemonicString =
|
|
|
|
await _secureStore.read(key: '${_walletId}_mnemonic');
|
|
|
|
if (mnemonicString == null) {
|
|
|
|
return [];
|
|
|
|
}
|
|
|
|
final List<String> data = mnemonicString.split(' ');
|
|
|
|
return data;
|
|
|
|
}
|
|
|
|
|
|
|
|
late ElectrumX _electrumXClient;
|
|
|
|
ElectrumX get electrumXClient => _electrumXClient;
|
|
|
|
|
|
|
|
late CachedElectrumX _cachedElectrumXClient;
|
|
|
|
CachedElectrumX get cachedElectrumXClient => _cachedElectrumXClient;
|
|
|
|
|
2022-11-09 23:48:43 +00:00
|
|
|
late SecureStorageInterface _secureStore;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
late TransactionNotificationTracker txTracker;
|
|
|
|
|
|
|
|
// Constructor
|
|
|
|
FiroWallet({
|
|
|
|
required String walletId,
|
|
|
|
required String walletName,
|
|
|
|
required Coin coin,
|
|
|
|
required ElectrumX client,
|
|
|
|
required CachedElectrumX cachedClient,
|
|
|
|
required TransactionNotificationTracker tracker,
|
2022-11-09 23:48:43 +00:00
|
|
|
required SecureStorageInterface secureStore,
|
2022-08-26 08:11:35 +00:00
|
|
|
}) {
|
|
|
|
txTracker = tracker;
|
|
|
|
_walletId = walletId;
|
|
|
|
_walletName = walletName;
|
|
|
|
_coin = coin;
|
|
|
|
_electrumXClient = client;
|
|
|
|
_cachedElectrumXClient = cachedClient;
|
2022-11-09 22:43:26 +00:00
|
|
|
_secureStore = secureStore;
|
2023-01-12 21:20:57 +00:00
|
|
|
initCache(walletId, coin);
|
|
|
|
initFiroHive(walletId);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
Logging.instance.log("$walletName isolates length: ${isolates.length}",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
// investigate possible issues killing shared isolates between multiple firo instances
|
|
|
|
for (final isolate in isolates.values) {
|
|
|
|
isolate.kill(priority: Isolate.immediate);
|
|
|
|
}
|
|
|
|
isolates.clear();
|
|
|
|
}
|
|
|
|
|
2022-09-07 15:56:10 +00:00
|
|
|
int estimateTxFee({required int vSize, required int feeRatePerKB}) {
|
|
|
|
return vSize * (feeRatePerKB / 1000).ceil();
|
|
|
|
}
|
|
|
|
|
|
|
|
/// The coinselection algorithm decides whether or not the user is eligible to make the transaction
|
|
|
|
/// with [satoshiAmountToSend] and [selectedTxFeeRate]. If so, it will call buildTrasaction() and return
|
|
|
|
/// a map containing the tx hex along with other important information. If not, then it will return
|
|
|
|
/// an integer (1 or 2)
|
|
|
|
dynamic coinSelection(
|
|
|
|
int satoshiAmountToSend,
|
|
|
|
int selectedTxFeeRate,
|
|
|
|
String _recipientAddress,
|
|
|
|
bool isSendAll, {
|
|
|
|
int additionalOutputs = 0,
|
2023-01-12 00:16:17 +00:00
|
|
|
List<isar_models.UTXO>? utxos,
|
2022-09-07 15:56:10 +00:00
|
|
|
}) async {
|
|
|
|
Logging.instance
|
|
|
|
.log("Starting coinSelection ----------", level: LogLevel.Info);
|
2023-01-12 00:16:17 +00:00
|
|
|
final List<isar_models.UTXO> availableOutputs = utxos ?? await this.utxos;
|
|
|
|
final currentChainHeight = await chainHeight;
|
|
|
|
final List<isar_models.UTXO> spendableOutputs = [];
|
2022-09-07 15:56:10 +00:00
|
|
|
int spendableSatoshiValue = 0;
|
|
|
|
|
|
|
|
// Build list of spendable outputs and totaling their satoshi amount
|
|
|
|
for (var i = 0; i < availableOutputs.length; i++) {
|
2023-01-12 00:16:17 +00:00
|
|
|
if (availableOutputs[i].isBlocked == false &&
|
|
|
|
availableOutputs[i]
|
|
|
|
.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) ==
|
|
|
|
true) {
|
2022-09-07 15:56:10 +00:00
|
|
|
spendableOutputs.add(availableOutputs[i]);
|
|
|
|
spendableSatoshiValue += availableOutputs[i].value;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// sort spendable by age (oldest first)
|
2023-01-12 00:16:17 +00:00
|
|
|
spendableOutputs.sort((a, b) => b.blockTime!.compareTo(a.blockTime!));
|
2022-09-07 15:56:10 +00:00
|
|
|
|
|
|
|
Logging.instance.log("spendableOutputs.length: ${spendableOutputs.length}",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log("spendableOutputs: $spendableOutputs", level: LogLevel.Info);
|
|
|
|
Logging.instance.log("spendableSatoshiValue: $spendableSatoshiValue",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log("satoshiAmountToSend: $satoshiAmountToSend", level: LogLevel.Info);
|
|
|
|
// If the amount the user is trying to send is smaller than the amount that they have spendable,
|
|
|
|
// then return 1, which indicates that they have an insufficient balance.
|
|
|
|
if (spendableSatoshiValue < satoshiAmountToSend) {
|
|
|
|
return 1;
|
|
|
|
// If the amount the user wants to send is exactly equal to the amount they can spend, then return
|
|
|
|
// 2, which indicates that they are not leaving enough over to pay the transaction fee
|
|
|
|
} else if (spendableSatoshiValue == satoshiAmountToSend && !isSendAll) {
|
|
|
|
return 2;
|
|
|
|
}
|
|
|
|
// If neither of these statements pass, we assume that the user has a spendable balance greater
|
|
|
|
// than the amount they're attempting to send. Note that this value still does not account for
|
|
|
|
// the added transaction fee, which may require an extra input and will need to be checked for
|
|
|
|
// later on.
|
|
|
|
|
|
|
|
// Possible situation right here
|
|
|
|
int satoshisBeingUsed = 0;
|
|
|
|
int inputsBeingConsumed = 0;
|
2023-01-12 00:16:17 +00:00
|
|
|
List<isar_models.UTXO> utxoObjectsToUse = [];
|
2022-09-07 15:56:10 +00:00
|
|
|
|
|
|
|
for (var i = 0;
|
2022-10-18 19:15:57 +00:00
|
|
|
satoshisBeingUsed <= satoshiAmountToSend && i < spendableOutputs.length;
|
2022-09-07 15:56:10 +00:00
|
|
|
i++) {
|
|
|
|
utxoObjectsToUse.add(spendableOutputs[i]);
|
|
|
|
satoshisBeingUsed += spendableOutputs[i].value;
|
|
|
|
inputsBeingConsumed += 1;
|
|
|
|
}
|
|
|
|
for (int i = 0;
|
|
|
|
i < additionalOutputs && inputsBeingConsumed < spendableOutputs.length;
|
|
|
|
i++) {
|
|
|
|
utxoObjectsToUse.add(spendableOutputs[inputsBeingConsumed]);
|
|
|
|
satoshisBeingUsed += spendableOutputs[inputsBeingConsumed].value;
|
|
|
|
inputsBeingConsumed += 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
Logging.instance
|
|
|
|
.log("satoshisBeingUsed: $satoshisBeingUsed", level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log("inputsBeingConsumed: $inputsBeingConsumed", level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('utxoObjectsToUse: $utxoObjectsToUse', level: LogLevel.Info);
|
|
|
|
|
|
|
|
// numberOfOutputs' length must always be equal to that of recipientsArray and recipientsAmtArray
|
|
|
|
List<String> recipientsArray = [_recipientAddress];
|
|
|
|
List<int> recipientsAmtArray = [satoshiAmountToSend];
|
|
|
|
|
|
|
|
// gather required signing data
|
|
|
|
final utxoSigningData = await fetchBuildTxData(utxoObjectsToUse);
|
|
|
|
|
|
|
|
if (isSendAll) {
|
|
|
|
Logging.instance
|
|
|
|
.log("Attempting to send all $coin", level: LogLevel.Info);
|
|
|
|
|
|
|
|
final int vSizeForOneOutput = (await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: [_recipientAddress],
|
|
|
|
satoshiAmounts: [satoshisBeingUsed - 1],
|
|
|
|
))["vSize"] as int;
|
|
|
|
int feeForOneOutput = estimateTxFee(
|
|
|
|
vSize: vSizeForOneOutput,
|
|
|
|
feeRatePerKB: selectedTxFeeRate,
|
|
|
|
);
|
|
|
|
|
|
|
|
if (feeForOneOutput < vSizeForOneOutput + 1) {
|
|
|
|
feeForOneOutput = vSizeForOneOutput + 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
final int amount = satoshiAmountToSend - feeForOneOutput;
|
|
|
|
dynamic txn = await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: recipientsArray,
|
|
|
|
satoshiAmounts: [amount],
|
|
|
|
);
|
|
|
|
Map<String, dynamic> transactionObject = {
|
|
|
|
"hex": txn["hex"],
|
|
|
|
"recipient": recipientsArray[0],
|
|
|
|
"recipientAmt": amount,
|
|
|
|
"fee": feeForOneOutput,
|
|
|
|
"vSize": txn["vSize"],
|
|
|
|
};
|
|
|
|
return transactionObject;
|
|
|
|
}
|
|
|
|
|
|
|
|
final int vSizeForOneOutput = (await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: [_recipientAddress],
|
|
|
|
satoshiAmounts: [satoshisBeingUsed - 1],
|
|
|
|
))["vSize"] as int;
|
|
|
|
final int vSizeForTwoOutPuts = (await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: [
|
|
|
|
_recipientAddress,
|
|
|
|
await _getCurrentAddressForChain(1),
|
|
|
|
],
|
|
|
|
satoshiAmounts: [
|
|
|
|
satoshiAmountToSend,
|
|
|
|
satoshisBeingUsed - satoshiAmountToSend - 1,
|
|
|
|
], // dust limit is the minimum amount a change output should be
|
|
|
|
))["vSize"] as int;
|
2022-12-13 00:17:02 +00:00
|
|
|
//todo: check if print needed
|
2022-09-07 15:56:10 +00:00
|
|
|
debugPrint("vSizeForOneOutput $vSizeForOneOutput");
|
|
|
|
debugPrint("vSizeForTwoOutPuts $vSizeForTwoOutPuts");
|
|
|
|
|
|
|
|
// Assume 1 output, only for recipient and no change
|
|
|
|
var feeForOneOutput = estimateTxFee(
|
|
|
|
vSize: vSizeForOneOutput,
|
|
|
|
feeRatePerKB: selectedTxFeeRate,
|
|
|
|
);
|
|
|
|
// Assume 2 outputs, one for recipient and one for change
|
|
|
|
var feeForTwoOutputs = estimateTxFee(
|
|
|
|
vSize: vSizeForTwoOutPuts,
|
|
|
|
feeRatePerKB: selectedTxFeeRate,
|
|
|
|
);
|
|
|
|
|
|
|
|
Logging.instance
|
|
|
|
.log("feeForTwoOutputs: $feeForTwoOutputs", level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log("feeForOneOutput: $feeForOneOutput", level: LogLevel.Info);
|
|
|
|
if (feeForOneOutput < (vSizeForOneOutput + 1)) {
|
|
|
|
feeForOneOutput = (vSizeForOneOutput + 1);
|
|
|
|
}
|
|
|
|
if (feeForTwoOutputs < ((vSizeForTwoOutPuts + 1))) {
|
|
|
|
feeForTwoOutputs = ((vSizeForTwoOutPuts + 1));
|
|
|
|
}
|
|
|
|
|
|
|
|
Logging.instance
|
|
|
|
.log("feeForTwoOutputs: $feeForTwoOutputs", level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log("feeForOneOutput: $feeForOneOutput", level: LogLevel.Info);
|
|
|
|
|
|
|
|
if (satoshisBeingUsed - satoshiAmountToSend > feeForOneOutput) {
|
|
|
|
if (satoshisBeingUsed - satoshiAmountToSend >
|
|
|
|
feeForOneOutput + DUST_LIMIT) {
|
|
|
|
// Here, we know that theoretically, we may be able to include another output(change) but we first need to
|
|
|
|
// factor in the value of this output in satoshis.
|
|
|
|
int changeOutputSize =
|
|
|
|
satoshisBeingUsed - satoshiAmountToSend - feeForTwoOutputs;
|
|
|
|
// We check to see if the user can pay for the new transaction with 2 outputs instead of one. If they can and
|
|
|
|
// the second output's size > DUST_LIMIT satoshis, we perform the mechanics required to properly generate and use a new
|
|
|
|
// change address.
|
|
|
|
if (changeOutputSize > DUST_LIMIT &&
|
|
|
|
satoshisBeingUsed - satoshiAmountToSend - changeOutputSize ==
|
|
|
|
feeForTwoOutputs) {
|
|
|
|
// generate new change address if current change address has been used
|
|
|
|
await checkChangeAddressForTransactions();
|
|
|
|
final String newChangeAddress = await _getCurrentAddressForChain(1);
|
|
|
|
|
|
|
|
int feeBeingPaid =
|
|
|
|
satoshisBeingUsed - satoshiAmountToSend - changeOutputSize;
|
|
|
|
|
|
|
|
recipientsArray.add(newChangeAddress);
|
|
|
|
recipientsAmtArray.add(changeOutputSize);
|
|
|
|
// At this point, we have the outputs we're going to use, the amounts to send along with which addresses
|
|
|
|
// we intend to send these amounts to. We have enough to send instructions to build the transaction.
|
|
|
|
Logging.instance.log('2 outputs in tx', level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
|
|
Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log('Change Output Size: $changeOutputSize',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log(
|
|
|
|
'Difference (fee being paid): $feeBeingPaid sats',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('Estimated fee: $feeForTwoOutputs', level: LogLevel.Info);
|
|
|
|
dynamic txn = await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: recipientsArray,
|
|
|
|
satoshiAmounts: recipientsAmtArray,
|
|
|
|
);
|
|
|
|
|
|
|
|
// make sure minimum fee is accurate if that is being used
|
|
|
|
if (txn["vSize"] - feeBeingPaid == 1) {
|
|
|
|
int changeOutputSize =
|
|
|
|
satoshisBeingUsed - satoshiAmountToSend - (txn["vSize"] as int);
|
|
|
|
feeBeingPaid =
|
|
|
|
satoshisBeingUsed - satoshiAmountToSend - changeOutputSize;
|
|
|
|
recipientsAmtArray.removeLast();
|
|
|
|
recipientsAmtArray.add(changeOutputSize);
|
|
|
|
Logging.instance.log('Adjusted Input size: $satoshisBeingUsed',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log(
|
|
|
|
'Adjusted Recipient output size: $satoshiAmountToSend',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log(
|
|
|
|
'Adjusted Change Output Size: $changeOutputSize',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log(
|
|
|
|
'Adjusted Difference (fee being paid): $feeBeingPaid sats',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log('Adjusted Estimated fee: $feeForTwoOutputs',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
txn = await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: recipientsArray,
|
|
|
|
satoshiAmounts: recipientsAmtArray,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
Map<String, dynamic> transactionObject = {
|
|
|
|
"hex": txn["hex"],
|
|
|
|
"recipient": recipientsArray[0],
|
|
|
|
"recipientAmt": recipientsAmtArray[0],
|
|
|
|
"fee": feeBeingPaid,
|
|
|
|
"vSize": txn["vSize"],
|
|
|
|
};
|
|
|
|
return transactionObject;
|
|
|
|
} else {
|
|
|
|
// Something went wrong here. It either overshot or undershot the estimated fee amount or the changeOutputSize
|
|
|
|
// is smaller than or equal to [DUST_LIMIT]. Revert to single output transaction.
|
|
|
|
Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
|
|
Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log(
|
|
|
|
'Difference (fee being paid): ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
|
|
dynamic txn = await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: recipientsArray,
|
|
|
|
satoshiAmounts: recipientsAmtArray,
|
|
|
|
);
|
|
|
|
Map<String, dynamic> transactionObject = {
|
|
|
|
"hex": txn["hex"],
|
|
|
|
"recipient": recipientsArray[0],
|
|
|
|
"recipientAmt": recipientsAmtArray[0],
|
|
|
|
"fee": satoshisBeingUsed - satoshiAmountToSend,
|
|
|
|
"vSize": txn["vSize"],
|
|
|
|
};
|
|
|
|
return transactionObject;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// No additional outputs needed since adding one would mean that it'd be smaller than 546 sats
|
|
|
|
// which makes it uneconomical to add to the transaction. Here, we pass data directly to instruct
|
|
|
|
// the wallet to begin crafting the transaction that the user requested.
|
|
|
|
Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
|
|
Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log(
|
|
|
|
'Difference (fee being paid): ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
|
|
dynamic txn = await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: recipientsArray,
|
|
|
|
satoshiAmounts: recipientsAmtArray,
|
|
|
|
);
|
|
|
|
Map<String, dynamic> transactionObject = {
|
|
|
|
"hex": txn["hex"],
|
|
|
|
"recipient": recipientsArray[0],
|
|
|
|
"recipientAmt": recipientsAmtArray[0],
|
|
|
|
"fee": satoshisBeingUsed - satoshiAmountToSend,
|
|
|
|
"vSize": txn["vSize"],
|
|
|
|
};
|
|
|
|
return transactionObject;
|
|
|
|
}
|
|
|
|
} else if (satoshisBeingUsed - satoshiAmountToSend == feeForOneOutput) {
|
|
|
|
// In this scenario, no additional change output is needed since inputs - outputs equal exactly
|
|
|
|
// what we need to pay for fees. Here, we pass data directly to instruct the wallet to begin
|
|
|
|
// crafting the transaction that the user requested.
|
|
|
|
Logging.instance.log('1 output in tx', level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('Input size: $satoshisBeingUsed', level: LogLevel.Info);
|
|
|
|
Logging.instance.log('Recipient output size: $satoshiAmountToSend',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log(
|
|
|
|
'Fee being paid: ${satoshisBeingUsed - satoshiAmountToSend} sats',
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance
|
|
|
|
.log('Estimated fee: $feeForOneOutput', level: LogLevel.Info);
|
|
|
|
dynamic txn = await buildTransaction(
|
|
|
|
utxosToUse: utxoObjectsToUse,
|
|
|
|
utxoSigningData: utxoSigningData,
|
|
|
|
recipients: recipientsArray,
|
|
|
|
satoshiAmounts: recipientsAmtArray,
|
|
|
|
);
|
|
|
|
Map<String, dynamic> transactionObject = {
|
|
|
|
"hex": txn["hex"],
|
|
|
|
"recipient": recipientsArray[0],
|
|
|
|
"recipientAmt": recipientsAmtArray[0],
|
|
|
|
"fee": feeForOneOutput,
|
|
|
|
"vSize": txn["vSize"],
|
|
|
|
};
|
|
|
|
return transactionObject;
|
|
|
|
} else {
|
|
|
|
// Remember that returning 2 indicates that the user does not have a sufficient balance to
|
|
|
|
// pay for the transaction fee. Ideally, at this stage, we should check if the user has any
|
|
|
|
// additional outputs they're able to spend and then recalculate fees.
|
|
|
|
Logging.instance.log(
|
|
|
|
'Cannot pay tx fee - checking for more outputs and trying again',
|
|
|
|
level: LogLevel.Warning);
|
|
|
|
// try adding more outputs
|
|
|
|
if (spendableOutputs.length > inputsBeingConsumed) {
|
|
|
|
return coinSelection(satoshiAmountToSend, selectedTxFeeRate,
|
|
|
|
_recipientAddress, isSendAll,
|
|
|
|
additionalOutputs: additionalOutputs + 1, utxos: utxos);
|
|
|
|
}
|
|
|
|
return 2;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<Map<String, dynamic>> fetchBuildTxData(
|
2023-01-12 00:16:17 +00:00
|
|
|
List<isar_models.UTXO> utxosToUse,
|
2022-09-07 15:56:10 +00:00
|
|
|
) async {
|
|
|
|
// return data
|
|
|
|
Map<String, dynamic> results = {};
|
|
|
|
Map<String, List<String>> addressTxid = {};
|
|
|
|
|
|
|
|
// addresses to check
|
|
|
|
List<String> addresses = [];
|
|
|
|
|
|
|
|
try {
|
|
|
|
// Populating the addresses to check
|
|
|
|
for (var i = 0; i < utxosToUse.length; i++) {
|
|
|
|
final txid = utxosToUse[i].txid;
|
|
|
|
final tx = await _cachedElectrumXClient.getTransaction(
|
|
|
|
txHash: txid,
|
|
|
|
coin: coin,
|
|
|
|
);
|
|
|
|
|
|
|
|
for (final output in tx["vout"] as List) {
|
|
|
|
final n = output["n"];
|
|
|
|
if (n != null && n == utxosToUse[i].vout) {
|
|
|
|
final address = output["scriptPubKey"]["addresses"][0] as String;
|
|
|
|
|
|
|
|
if (!addressTxid.containsKey(address)) {
|
|
|
|
addressTxid[address] = <String>[];
|
|
|
|
}
|
|
|
|
(addressTxid[address] as List).add(txid);
|
|
|
|
|
|
|
|
addresses.add(address);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// p2pkh / bip44
|
|
|
|
final addressesLength = addresses.length;
|
|
|
|
if (addressesLength > 0) {
|
|
|
|
final receiveDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_receiveDerivations");
|
|
|
|
final receiveDerivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(receiveDerivationsString ?? "{}") as Map);
|
|
|
|
|
|
|
|
final changeDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_changeDerivations");
|
|
|
|
final changeDerivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(changeDerivationsString ?? "{}") as Map);
|
|
|
|
|
|
|
|
for (int i = 0; i < addressesLength; i++) {
|
|
|
|
// receives
|
|
|
|
|
|
|
|
dynamic receiveDerivation;
|
|
|
|
|
|
|
|
for (int j = 0; j < receiveDerivations.length; j++) {
|
|
|
|
if (receiveDerivations["$j"]["address"] == addresses[i]) {
|
|
|
|
receiveDerivation = receiveDerivations["$j"];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// receiveDerivation = receiveDerivations[addresses[i]];
|
|
|
|
// if a match exists it will not be null
|
|
|
|
if (receiveDerivation != null) {
|
|
|
|
final data = P2PKH(
|
|
|
|
data: PaymentData(
|
|
|
|
pubkey: Format.stringToUint8List(
|
|
|
|
receiveDerivation["publicKey"] as String)),
|
|
|
|
network: _network,
|
|
|
|
).data;
|
|
|
|
|
|
|
|
for (String tx in addressTxid[addresses[i]]!) {
|
|
|
|
results[tx] = {
|
|
|
|
"output": data.output,
|
|
|
|
"keyPair": ECPair.fromWIF(
|
|
|
|
receiveDerivation["wif"] as String,
|
|
|
|
network: _network,
|
|
|
|
),
|
|
|
|
};
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// if its not a receive, check change
|
|
|
|
|
|
|
|
dynamic changeDerivation;
|
|
|
|
|
|
|
|
for (int j = 0; j < changeDerivations.length; j++) {
|
|
|
|
if (changeDerivations["$j"]["address"] == addresses[i]) {
|
|
|
|
changeDerivation = changeDerivations["$j"];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// final changeDerivation = changeDerivations[addresses[i]];
|
|
|
|
// if a match exists it will not be null
|
|
|
|
if (changeDerivation != null) {
|
|
|
|
final data = P2PKH(
|
|
|
|
data: PaymentData(
|
|
|
|
pubkey: Format.stringToUint8List(
|
|
|
|
changeDerivation["publicKey"] as String)),
|
|
|
|
network: _network,
|
|
|
|
).data;
|
|
|
|
|
|
|
|
for (String tx in addressTxid[addresses[i]]!) {
|
|
|
|
results[tx] = {
|
|
|
|
"output": data.output,
|
|
|
|
"keyPair": ECPair.fromWIF(
|
|
|
|
changeDerivation["wif"] as String,
|
|
|
|
network: _network,
|
|
|
|
),
|
|
|
|
};
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return results;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance
|
|
|
|
.log("fetchBuildTxData() threw: $e,\n$s", level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Builds and signs a transaction
|
|
|
|
Future<Map<String, dynamic>> buildTransaction({
|
2023-01-12 00:16:17 +00:00
|
|
|
required List<isar_models.UTXO> utxosToUse,
|
2022-09-07 15:56:10 +00:00
|
|
|
required Map<String, dynamic> utxoSigningData,
|
|
|
|
required List<String> recipients,
|
|
|
|
required List<int> satoshiAmounts,
|
|
|
|
}) async {
|
|
|
|
Logging.instance
|
|
|
|
.log("Starting buildTransaction ----------", level: LogLevel.Info);
|
|
|
|
|
|
|
|
final txb = TransactionBuilder(network: _network);
|
|
|
|
txb.setVersion(1);
|
|
|
|
|
|
|
|
// Add transaction inputs
|
|
|
|
for (var i = 0; i < utxosToUse.length; i++) {
|
|
|
|
final txid = utxosToUse[i].txid;
|
|
|
|
txb.addInput(txid, utxosToUse[i].vout, null,
|
|
|
|
utxoSigningData[txid]["output"] as Uint8List);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add transaction output
|
|
|
|
for (var i = 0; i < recipients.length; i++) {
|
|
|
|
txb.addOutput(recipients[i], satoshiAmounts[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
try {
|
|
|
|
// Sign the transaction accordingly
|
|
|
|
for (var i = 0; i < utxosToUse.length; i++) {
|
|
|
|
final txid = utxosToUse[i].txid;
|
|
|
|
txb.sign(
|
|
|
|
vin: i,
|
|
|
|
keyPair: utxoSigningData[txid]["keyPair"] as ECPair,
|
|
|
|
witnessValue: utxosToUse[i].value,
|
|
|
|
redeemScript: utxoSigningData[txid]["redeemScript"] as Uint8List?,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Caught exception while signing transaction: $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
|
|
|
|
final builtTx = txb.build();
|
|
|
|
final vSize = builtTx.virtualSize();
|
|
|
|
|
|
|
|
return {"hex": builtTx.toHex(), "vSize": vSize};
|
|
|
|
}
|
|
|
|
|
2022-08-26 08:11:35 +00:00
|
|
|
@override
|
|
|
|
Future<void> updateNode(bool shouldRefresh) async {
|
2022-11-09 22:43:26 +00:00
|
|
|
final failovers = NodeService(secureStorageInterface: _secureStore)
|
2022-08-26 08:11:35 +00:00
|
|
|
.failoverNodesFor(coin: coin)
|
|
|
|
.map(
|
|
|
|
(e) => ElectrumXNode(
|
|
|
|
address: e.host,
|
|
|
|
port: e.port,
|
|
|
|
name: e.name,
|
|
|
|
id: e.id,
|
|
|
|
useSSL: e.useSSL,
|
|
|
|
),
|
|
|
|
)
|
|
|
|
.toList();
|
|
|
|
final newNode = await _getCurrentNode();
|
|
|
|
_cachedElectrumXClient = CachedElectrumX.from(
|
|
|
|
node: newNode,
|
|
|
|
prefs: _prefs,
|
|
|
|
failovers: failovers,
|
|
|
|
);
|
|
|
|
_electrumXClient = ElectrumX.from(
|
|
|
|
node: newNode,
|
|
|
|
prefs: _prefs,
|
|
|
|
failovers: failovers,
|
|
|
|
);
|
|
|
|
|
|
|
|
if (shouldRefresh) {
|
2022-09-06 15:34:39 +00:00
|
|
|
unawaited(refresh());
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
@override
|
|
|
|
Future<void> initializeNew() async {
|
|
|
|
Logging.instance
|
|
|
|
.log("Generating new ${coin.prettyName} wallet.", level: LogLevel.Info);
|
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
if (getCachedId() != null) {
|
2022-08-26 08:11:35 +00:00
|
|
|
throw Exception(
|
|
|
|
"Attempted to initialize a new wallet using an existing wallet ID!");
|
|
|
|
}
|
|
|
|
|
|
|
|
await _prefs.init();
|
|
|
|
try {
|
|
|
|
await _generateNewWallet();
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown from initializeNew(): $e\n$s",
|
|
|
|
level: LogLevel.Fatal);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
|
|
|
|
await Future.wait([
|
2023-01-12 21:20:57 +00:00
|
|
|
updateCachedId(walletId),
|
|
|
|
updateCachedIsFavorite(false),
|
2022-08-26 08:11:35 +00:00
|
|
|
]);
|
|
|
|
}
|
|
|
|
|
|
|
|
@override
|
|
|
|
Future<void> initializeExisting() async {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Opening existing $_walletId ${coin.prettyName} wallet.",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
if (getCachedId() == null) {
|
2022-08-26 08:11:35 +00:00
|
|
|
throw Exception(
|
|
|
|
"Attempted to initialize an existing wallet using an unknown wallet ID!");
|
|
|
|
}
|
|
|
|
await _prefs.init();
|
2023-01-12 18:46:01 +00:00
|
|
|
await isarInit(walletId);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
Future<bool> refreshIfThereIsNewData() async {
|
|
|
|
if (longMutex) return false;
|
|
|
|
if (_hasCalledExit) return false;
|
|
|
|
Logging.instance
|
|
|
|
.log("$walletName refreshIfThereIsNewData", level: LogLevel.Info);
|
|
|
|
|
|
|
|
try {
|
|
|
|
bool needsRefresh = false;
|
|
|
|
Set<String> txnsToCheck = {};
|
|
|
|
|
|
|
|
for (final String txid in txTracker.pendings) {
|
|
|
|
if (!txTracker.wasNotifiedConfirmed(txid)) {
|
|
|
|
txnsToCheck.add(txid);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (String txid in txnsToCheck) {
|
|
|
|
final txn = await electrumXClient.getTransaction(txHash: txid);
|
|
|
|
int confirmations = txn["confirmations"] as int? ?? 0;
|
|
|
|
bool isUnconfirmed = confirmations < MINIMUM_CONFIRMATIONS;
|
|
|
|
if (!isUnconfirmed) {
|
|
|
|
needsRefresh = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (!needsRefresh) {
|
2023-01-12 00:16:17 +00:00
|
|
|
final allOwnAddresses = await _fetchAllOwnAddresses();
|
|
|
|
List<Map<String, dynamic>> allTxs = await _fetchHistory(
|
|
|
|
allOwnAddresses.map((e) => e.value).toList(growable: false));
|
2022-08-26 08:11:35 +00:00
|
|
|
for (Map<String, dynamic> transaction in allTxs) {
|
2023-01-12 00:16:17 +00:00
|
|
|
final txid = transaction['tx_hash'] as String;
|
|
|
|
if ((await isar.transactions
|
|
|
|
.filter()
|
|
|
|
.txidMatches(txid)
|
|
|
|
.findFirst()) ==
|
2022-08-26 08:11:35 +00:00
|
|
|
null) {
|
|
|
|
Logging.instance.log(
|
|
|
|
" txid not found in address history already ${transaction['tx_hash']}",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
needsRefresh = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return needsRefresh;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception caught in refreshIfThereIsNewData: $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<void> getAllTxsToWatch() async {
|
2022-08-26 08:11:35 +00:00
|
|
|
if (_hasCalledExit) return;
|
|
|
|
Logging.instance.log("$walletName periodic", level: LogLevel.Info);
|
2023-01-12 00:16:17 +00:00
|
|
|
List<isar_models.Transaction> unconfirmedTxnsToNotifyPending = [];
|
|
|
|
List<isar_models.Transaction> unconfirmedTxnsToNotifyConfirmed = [];
|
|
|
|
|
|
|
|
final currentChainHeight = await chainHeight;
|
|
|
|
|
|
|
|
final txTxns = await isar.transactions
|
|
|
|
.filter()
|
|
|
|
.isLelantusIsNull()
|
|
|
|
.or()
|
|
|
|
.isLelantusEqualTo(false)
|
|
|
|
.findAll();
|
|
|
|
final ltxTxns =
|
|
|
|
await isar.transactions.filter().isLelantusEqualTo(true).findAll();
|
|
|
|
|
|
|
|
for (isar_models.Transaction tx in txTxns) {
|
|
|
|
isar_models.Transaction? lTx;
|
|
|
|
try {
|
|
|
|
lTx = ltxTxns.firstWhere((e) => e.txid == tx.txid);
|
|
|
|
} catch (_) {
|
|
|
|
lTx = null;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) {
|
|
|
|
if (txTracker.wasNotifiedPending(tx.txid) &&
|
|
|
|
!txTracker.wasNotifiedConfirmed(tx.txid)) {
|
|
|
|
// get all transactions that were notified as pending but not as confirmed
|
|
|
|
unconfirmedTxnsToNotifyConfirmed.add(tx);
|
|
|
|
}
|
|
|
|
if (lTx != null &&
|
|
|
|
(lTx.inputs.isEmpty || lTx.inputs.first.txid.isEmpty) &&
|
|
|
|
lTx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) ==
|
|
|
|
false &&
|
|
|
|
tx.type == isar_models.TransactionType.incoming) {
|
|
|
|
// If this is a received that is past 1 or more confirmations and has not been minted,
|
2022-08-26 08:11:35 +00:00
|
|
|
if (!txTracker.wasNotifiedPending(tx.txid)) {
|
|
|
|
unconfirmedTxnsToNotifyPending.add(tx);
|
|
|
|
}
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
} else {
|
|
|
|
if (!txTracker.wasNotifiedPending(tx.txid)) {
|
|
|
|
// get all transactions that were not notified as pending yet
|
|
|
|
unconfirmedTxnsToNotifyPending.add(tx);
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
for (isar_models.Transaction tx in txTxns) {
|
|
|
|
if (!tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) &&
|
|
|
|
tx.inputs.first.txid.isNotEmpty) {
|
|
|
|
// Get all normal txs that are at 0 confirmations
|
|
|
|
unconfirmedTxnsToNotifyPending
|
|
|
|
.removeWhere((e) => e.txid == tx.inputs.first.txid);
|
|
|
|
Logging.instance.log("removed tx: ${tx.txid}", level: LogLevel.Info);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
|
|
|
|
for (isar_models.Transaction lTX in ltxTxns) {
|
|
|
|
isar_models.Transaction? tx;
|
|
|
|
try {
|
|
|
|
tx = ltxTxns.firstWhere((e) => e.txid == lTX.txid);
|
|
|
|
} catch (_) {
|
|
|
|
tx = null;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (tx == null) {
|
|
|
|
// if this is a ltx transaction that is unconfirmed and not represented in the normal transaction set.
|
|
|
|
if (!lTX.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) {
|
|
|
|
if (!txTracker.wasNotifiedPending(lTX.txid)) {
|
|
|
|
unconfirmedTxnsToNotifyPending.add(lTX);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (txTracker.wasNotifiedPending(lTX.txid) &&
|
|
|
|
!txTracker.wasNotifiedConfirmed(lTX.txid)) {
|
|
|
|
unconfirmedTxnsToNotifyConfirmed.add(lTX);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
|
2022-08-26 08:11:35 +00:00
|
|
|
Logging.instance.log(
|
|
|
|
"unconfirmedTxnsToNotifyPending $unconfirmedTxnsToNotifyPending",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
Logging.instance.log(
|
|
|
|
"unconfirmedTxnsToNotifyConfirmed $unconfirmedTxnsToNotifyConfirmed",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
|
|
|
|
for (final tx in unconfirmedTxnsToNotifyPending) {
|
2023-01-12 00:16:17 +00:00
|
|
|
final confirmations = tx.getConfirmations(currentChainHeight);
|
|
|
|
|
|
|
|
switch (tx.type) {
|
|
|
|
case isar_models.TransactionType.incoming:
|
2022-09-06 15:34:39 +00:00
|
|
|
unawaited(
|
|
|
|
NotificationApi.showNotification(
|
|
|
|
title: "Incoming transaction",
|
|
|
|
body: walletName,
|
|
|
|
walletId: walletId,
|
|
|
|
iconAssetName: Assets.svg.iconFor(coin: coin),
|
|
|
|
date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
2023-01-12 00:16:17 +00:00
|
|
|
shouldWatchForUpdates: confirmations < MINIMUM_CONFIRMATIONS,
|
2022-09-06 15:34:39 +00:00
|
|
|
coinName: coin.name,
|
|
|
|
txid: tx.txid,
|
2023-01-12 00:16:17 +00:00
|
|
|
confirmations: confirmations,
|
2022-09-06 15:34:39 +00:00
|
|
|
requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
|
|
),
|
2022-08-26 08:11:35 +00:00
|
|
|
);
|
|
|
|
await txTracker.addNotifiedPending(tx.txid);
|
|
|
|
break;
|
2023-01-12 00:16:17 +00:00
|
|
|
case isar_models.TransactionType.outgoing:
|
2022-09-06 15:34:39 +00:00
|
|
|
unawaited(
|
|
|
|
NotificationApi.showNotification(
|
2023-01-12 00:16:17 +00:00
|
|
|
title: tx.subType == isar_models.TransactionSubType.mint
|
|
|
|
? "Anonymizing"
|
|
|
|
: "Outgoing transaction",
|
2022-09-06 15:34:39 +00:00
|
|
|
body: walletName,
|
|
|
|
walletId: walletId,
|
|
|
|
iconAssetName: Assets.svg.iconFor(coin: coin),
|
|
|
|
date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
2023-01-12 00:16:17 +00:00
|
|
|
shouldWatchForUpdates: confirmations < MINIMUM_CONFIRMATIONS,
|
2022-09-06 15:34:39 +00:00
|
|
|
coinName: coin.name,
|
|
|
|
txid: tx.txid,
|
2023-01-12 00:16:17 +00:00
|
|
|
confirmations: confirmations,
|
2022-09-06 15:34:39 +00:00
|
|
|
requiredConfirmations: MINIMUM_CONFIRMATIONS,
|
|
|
|
),
|
2022-08-26 08:11:35 +00:00
|
|
|
);
|
|
|
|
await txTracker.addNotifiedPending(tx.txid);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (final tx in unconfirmedTxnsToNotifyConfirmed) {
|
2023-01-12 00:16:17 +00:00
|
|
|
if (tx.type == isar_models.TransactionType.incoming) {
|
2022-09-06 15:34:39 +00:00
|
|
|
unawaited(
|
|
|
|
NotificationApi.showNotification(
|
|
|
|
title: "Incoming transaction confirmed",
|
|
|
|
body: walletName,
|
|
|
|
walletId: walletId,
|
|
|
|
iconAssetName: Assets.svg.iconFor(coin: coin),
|
|
|
|
date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
|
|
|
shouldWatchForUpdates: false,
|
|
|
|
coinName: coin.name,
|
|
|
|
),
|
2022-08-26 08:11:35 +00:00
|
|
|
);
|
|
|
|
await txTracker.addNotifiedConfirmed(tx.txid);
|
2023-01-12 00:16:17 +00:00
|
|
|
} else if (tx.type == isar_models.TransactionType.outgoing &&
|
|
|
|
tx.subType == isar_models.TransactionSubType.join) {
|
2022-09-06 15:34:39 +00:00
|
|
|
unawaited(
|
|
|
|
NotificationApi.showNotification(
|
2023-01-12 00:16:17 +00:00
|
|
|
title: tx.subType ==
|
|
|
|
isar_models.TransactionSubType.mint // redundant check?
|
2022-10-15 17:57:00 +00:00
|
|
|
? "Anonymized"
|
|
|
|
: "Outgoing transaction confirmed",
|
2022-09-06 15:34:39 +00:00
|
|
|
body: walletName,
|
|
|
|
walletId: walletId,
|
|
|
|
iconAssetName: Assets.svg.iconFor(coin: coin),
|
|
|
|
date: DateTime.fromMillisecondsSinceEpoch(tx.timestamp * 1000),
|
|
|
|
shouldWatchForUpdates: false,
|
|
|
|
coinName: coin.name,
|
|
|
|
),
|
2022-08-26 08:11:35 +00:00
|
|
|
);
|
|
|
|
await txTracker.addNotifiedConfirmed(tx.txid);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Generates initial wallet values such as mnemonic, chain (receive/change) arrays and indexes.
|
|
|
|
Future<void> _generateNewWallet() async {
|
2022-11-30 15:46:28 +00:00
|
|
|
Logging.instance
|
|
|
|
.log("IS_INTEGRATION_TEST: $integrationTestFlag", level: LogLevel.Info);
|
|
|
|
if (!integrationTestFlag) {
|
2022-11-30 15:54:46 +00:00
|
|
|
try {
|
|
|
|
final features = await electrumXClient
|
|
|
|
.getServerFeatures()
|
|
|
|
.timeout(const Duration(seconds: 3));
|
|
|
|
Logging.instance.log("features: $features", level: LogLevel.Info);
|
|
|
|
switch (coin) {
|
|
|
|
case Coin.firo:
|
|
|
|
if (features['genesis_hash'] != GENESIS_HASH_MAINNET) {
|
|
|
|
throw Exception("genesis hash does not match main net!");
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case Coin.firoTestNet:
|
|
|
|
if (features['genesis_hash'] != GENESIS_HASH_TESTNET) {
|
|
|
|
throw Exception("genesis hash does not match test net!");
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
throw Exception(
|
|
|
|
"Attempted to generate a FiroWallet using a non firo coin type: ${coin.name}");
|
|
|
|
}
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("$e/n$s", level: LogLevel.Info);
|
2022-11-30 15:46:28 +00:00
|
|
|
}
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
// this should never fail as overwriting a mnemonic is big bad
|
|
|
|
assert((await _secureStore.read(key: '${_walletId}_mnemonic')) == null);
|
|
|
|
await _secureStore.write(
|
|
|
|
key: '${_walletId}_mnemonic',
|
|
|
|
value: bip39.generateMnemonic(strength: 256));
|
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
await firoUpdateJIndex(<dynamic>[]);
|
2022-08-26 08:11:35 +00:00
|
|
|
// Generate and add addresses to relevant arrays
|
|
|
|
final initialReceivingAddress = await _generateAddressForChain(0, 0);
|
|
|
|
final initialChangeAddress = await _generateAddressForChain(1, 0);
|
2023-01-12 00:16:17 +00:00
|
|
|
|
2023-01-12 18:46:01 +00:00
|
|
|
await isarInit(walletId);
|
2023-01-12 00:16:17 +00:00
|
|
|
|
|
|
|
await isar.writeTxn(() async {
|
|
|
|
await isar.addresses.putAll([
|
|
|
|
initialReceivingAddress,
|
|
|
|
initialChangeAddress,
|
|
|
|
]);
|
|
|
|
});
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
bool refreshMutex = false;
|
|
|
|
@override
|
|
|
|
bool get isRefreshing => refreshMutex;
|
|
|
|
|
|
|
|
/// Refreshes display data for the wallet
|
|
|
|
@override
|
|
|
|
Future<void> refresh() async {
|
|
|
|
if (refreshMutex) {
|
2022-08-31 17:49:19 +00:00
|
|
|
Logging.instance.log("$walletId $walletName refreshMutex denied",
|
|
|
|
level: LogLevel.Info);
|
2022-08-26 08:11:35 +00:00
|
|
|
return;
|
|
|
|
} else {
|
|
|
|
refreshMutex = true;
|
|
|
|
}
|
|
|
|
Logging.instance
|
|
|
|
.log("PROCESSORS ${Platform.numberOfProcessors}", level: LogLevel.Info);
|
|
|
|
try {
|
|
|
|
GlobalEventBus.instance.fire(
|
|
|
|
WalletSyncStatusChangedEvent(
|
|
|
|
WalletSyncStatus.syncing,
|
|
|
|
walletId,
|
|
|
|
coin,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.0, walletId));
|
|
|
|
|
|
|
|
final receiveDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_receiveDerivations");
|
|
|
|
if (receiveDerivationsString == null ||
|
|
|
|
receiveDerivationsString == "{}") {
|
|
|
|
GlobalEventBus.instance
|
|
|
|
.fire(RefreshPercentChangedEvent(0.05, walletId));
|
|
|
|
final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic');
|
|
|
|
await fillAddresses(mnemonic!,
|
|
|
|
numberOfThreads: Platform.numberOfProcessors - isolates.length - 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
await checkReceivingAddressForTransactions();
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.1, walletId));
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
await _refreshUTXOs();
|
2022-08-26 08:11:35 +00:00
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.2, walletId));
|
|
|
|
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.25, walletId));
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
await _refreshTransactions();
|
2022-08-26 08:11:35 +00:00
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.35, walletId));
|
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
final feeObj = _getFees();
|
2022-08-26 08:11:35 +00:00
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.50, walletId));
|
|
|
|
|
|
|
|
_feeObject = Future(() => feeObj);
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.60, walletId));
|
|
|
|
|
|
|
|
final lelantusCoins = getLelantusCoinMap();
|
2022-09-06 15:34:39 +00:00
|
|
|
Logging.instance.log("_lelantus_coins at refresh: $lelantusCoins",
|
2022-08-26 08:11:35 +00:00
|
|
|
level: LogLevel.Warning, printFullLength: true);
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.70, walletId));
|
|
|
|
|
|
|
|
await _refreshLelantusData();
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.80, walletId));
|
|
|
|
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.90, walletId));
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
await _refreshBalance();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(0.95, walletId));
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
await getAllTxsToWatch();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
GlobalEventBus.instance.fire(RefreshPercentChangedEvent(1.0, walletId));
|
|
|
|
|
|
|
|
GlobalEventBus.instance.fire(
|
|
|
|
WalletSyncStatusChangedEvent(
|
|
|
|
WalletSyncStatus.synced,
|
|
|
|
walletId,
|
|
|
|
coin,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
refreshMutex = false;
|
|
|
|
|
|
|
|
if (isActive || shouldAutoSync) {
|
2022-09-09 15:07:27 +00:00
|
|
|
timer ??= Timer.periodic(const Duration(seconds: 30), (timer) async {
|
2022-08-26 08:11:35 +00:00
|
|
|
bool shouldNotify = await refreshIfThereIsNewData();
|
|
|
|
if (shouldNotify) {
|
|
|
|
await refresh();
|
|
|
|
GlobalEventBus.instance.fire(UpdatedInBackgroundEvent(
|
2022-08-31 17:49:19 +00:00
|
|
|
"New data found in $walletId $walletName in background!",
|
|
|
|
walletId));
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
});
|
|
|
|
}
|
|
|
|
} catch (error, strace) {
|
|
|
|
refreshMutex = false;
|
|
|
|
GlobalEventBus.instance.fire(
|
|
|
|
NodeConnectionStatusChangedEvent(
|
|
|
|
NodeConnectionStatus.disconnected,
|
|
|
|
walletId,
|
|
|
|
coin,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
GlobalEventBus.instance.fire(
|
|
|
|
WalletSyncStatusChangedEvent(
|
|
|
|
WalletSyncStatus.unableToSync,
|
|
|
|
walletId,
|
|
|
|
coin,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
Logging.instance.log(
|
|
|
|
"Caught exception in refreshWalletData(): $error\n$strace",
|
|
|
|
level: LogLevel.Warning);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<int> _fetchMaxFee() async {
|
2023-01-12 00:16:17 +00:00
|
|
|
final balance = availablePrivateBalance();
|
2022-11-23 18:31:31 +00:00
|
|
|
int spendAmount = (balance * Decimal.fromInt(Constants.satsPerCoin(coin)))
|
|
|
|
.toBigInt()
|
|
|
|
.toInt();
|
2022-09-06 15:34:39 +00:00
|
|
|
int fee = await estimateJoinSplitFee(spendAmount);
|
2022-08-26 08:11:35 +00:00
|
|
|
return fee;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<List<DartLelantusEntry>> _getLelantusEntry() async {
|
|
|
|
final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic');
|
|
|
|
final List<LelantusCoin> lelantusCoins = await _getUnspentCoins();
|
|
|
|
final root = await compute(
|
|
|
|
getBip32RootWrapper,
|
|
|
|
Tuple2(
|
|
|
|
mnemonic!,
|
|
|
|
_network,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
final waitLelantusEntries = lelantusCoins.map((coin) async {
|
|
|
|
final keyPair = await compute(
|
|
|
|
getBip32NodeFromRootWrapper,
|
|
|
|
Tuple3(
|
|
|
|
MINT_INDEX,
|
|
|
|
coin.index,
|
|
|
|
root,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
if (keyPair.privateKey == null) {
|
|
|
|
Logging.instance.log("error bad key", level: LogLevel.Error);
|
|
|
|
return DartLelantusEntry(1, 0, 0, 0, 0, '');
|
|
|
|
}
|
|
|
|
final String privateKey = Format.uint8listToString(keyPair.privateKey!);
|
|
|
|
return DartLelantusEntry(coin.isUsed ? 1 : 0, 0, coin.anonymitySetId,
|
|
|
|
coin.value, coin.index, privateKey);
|
|
|
|
}).toList();
|
|
|
|
|
|
|
|
final lelantusEntries = await Future.wait(waitLelantusEntries);
|
|
|
|
|
|
|
|
if (lelantusEntries.isNotEmpty) {
|
|
|
|
lelantusEntries.removeWhere((element) => element.amount == 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
return lelantusEntries;
|
|
|
|
}
|
|
|
|
|
|
|
|
List<Map<dynamic, LelantusCoin>> getLelantusCoinMap() {
|
2023-01-12 21:20:57 +00:00
|
|
|
final _l = firoGetLelantusCoins();
|
2022-08-26 08:11:35 +00:00
|
|
|
final List<Map<dynamic, LelantusCoin>> lelantusCoins = [];
|
|
|
|
for (var el in _l ?? []) {
|
|
|
|
lelantusCoins.add({el.keys.first: el.values.first as LelantusCoin});
|
|
|
|
}
|
|
|
|
return lelantusCoins;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<List<LelantusCoin>> _getUnspentCoins() async {
|
|
|
|
final List<Map<dynamic, LelantusCoin>> lelantusCoins = getLelantusCoinMap();
|
2022-09-06 15:34:39 +00:00
|
|
|
if (lelantusCoins.isNotEmpty) {
|
2022-08-26 08:11:35 +00:00
|
|
|
lelantusCoins.removeWhere((element) =>
|
|
|
|
element.values.any((elementCoin) => elementCoin.value == 0));
|
|
|
|
}
|
2023-01-12 21:20:57 +00:00
|
|
|
final jindexes = firoGetJIndex();
|
2023-01-12 00:16:17 +00:00
|
|
|
final transactions = await _txnData;
|
|
|
|
final lelantusTransactionsd = await lelantusTransactionData;
|
|
|
|
|
2022-08-26 08:11:35 +00:00
|
|
|
List<LelantusCoin> coins = [];
|
|
|
|
|
|
|
|
List<LelantusCoin> lelantusCoinsList =
|
|
|
|
lelantusCoins.fold(<LelantusCoin>[], (previousValue, element) {
|
|
|
|
previousValue.add(element.values.first);
|
|
|
|
return previousValue;
|
|
|
|
});
|
2023-01-12 00:16:17 +00:00
|
|
|
|
|
|
|
final currentChainHeight = await chainHeight;
|
|
|
|
|
2022-08-26 08:11:35 +00:00
|
|
|
for (int i = 0; i < lelantusCoinsList.length; i++) {
|
|
|
|
// Logging.instance.log("lelantusCoinsList[$i]: ${lelantusCoinsList[i]}");
|
|
|
|
final txn = await cachedElectrumXClient.getTransaction(
|
|
|
|
txHash: lelantusCoinsList[i].txId,
|
|
|
|
verbose: true,
|
|
|
|
coin: coin,
|
|
|
|
);
|
|
|
|
final confirmations = txn["confirmations"];
|
|
|
|
bool isUnconfirmed = confirmations is int && confirmations < 1;
|
|
|
|
if (!jindexes!.contains(lelantusCoinsList[i].index) &&
|
2023-01-12 00:16:17 +00:00
|
|
|
transactions
|
|
|
|
.where((e) => e.txid == lelantusCoinsList[i].txId)
|
|
|
|
.isEmpty) {
|
2022-08-26 08:11:35 +00:00
|
|
|
isUnconfirmed = true;
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
|
|
|
|
// TODO: optimize the following
|
|
|
|
if ((transactions
|
|
|
|
.where((e) => e.txid == lelantusCoinsList[i].txId)
|
|
|
|
.isNotEmpty &&
|
|
|
|
!transactions
|
|
|
|
.where((e) => e.txid == lelantusCoinsList[i].txId)
|
|
|
|
.first
|
|
|
|
.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) ||
|
|
|
|
(lelantusTransactionsd
|
|
|
|
.where((e) => e.txid == lelantusCoinsList[i].txId)
|
|
|
|
.isNotEmpty &&
|
|
|
|
!lelantusTransactionsd
|
|
|
|
.where((e) => e.txid == lelantusCoinsList[i].txId)
|
|
|
|
.first
|
|
|
|
.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS))) {
|
2022-08-26 08:11:35 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (!lelantusCoinsList[i].isUsed &&
|
|
|
|
lelantusCoinsList[i].anonymitySetId != ANONYMITY_SET_EMPTY_ID &&
|
|
|
|
!isUnconfirmed) {
|
|
|
|
coins.add(lelantusCoinsList[i]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return coins;
|
|
|
|
}
|
|
|
|
|
|
|
|
// index 0 and 1 for the funds available to spend.
|
|
|
|
// index 2 and 3 for all the funds in the wallet (including the undependable ones)
|
2023-01-12 00:16:17 +00:00
|
|
|
// Future<List<Decimal>> _refreshBalance() async {
|
|
|
|
Future<void> _refreshBalance() async {
|
2022-08-26 08:11:35 +00:00
|
|
|
try {
|
2023-01-12 00:16:17 +00:00
|
|
|
final utxosUpdateFuture = _refreshUTXOs();
|
2022-08-26 08:11:35 +00:00
|
|
|
final List<Map<dynamic, LelantusCoin>> lelantusCoins =
|
|
|
|
getLelantusCoinMap();
|
2022-09-06 15:34:39 +00:00
|
|
|
if (lelantusCoins.isNotEmpty) {
|
2022-08-26 08:11:35 +00:00
|
|
|
lelantusCoins.removeWhere((element) =>
|
|
|
|
element.values.any((elementCoin) => elementCoin.value == 0));
|
|
|
|
}
|
|
|
|
final data = await _txnData;
|
|
|
|
final lData = await lelantusTransactionData;
|
2023-01-12 00:16:17 +00:00
|
|
|
final currentChainHeight = await chainHeight;
|
2023-01-12 21:20:57 +00:00
|
|
|
final jindexes = firoGetJIndex();
|
2022-08-26 08:11:35 +00:00
|
|
|
int intLelantusBalance = 0;
|
|
|
|
int unconfirmedLelantusBalance = 0;
|
2022-09-06 15:34:39 +00:00
|
|
|
|
|
|
|
for (var element in lelantusCoins) {
|
|
|
|
element.forEach((key, value) {
|
2023-01-12 00:16:17 +00:00
|
|
|
isar_models.Transaction? tx;
|
|
|
|
try {
|
|
|
|
tx == data.firstWhere((e) => e.txid == value.txId);
|
|
|
|
} catch (_) {
|
|
|
|
tx = null;
|
|
|
|
}
|
|
|
|
|
|
|
|
isar_models.Transaction? ltx;
|
|
|
|
try {
|
|
|
|
ltx = lData.firstWhere((e) => e.txid == value.txId);
|
|
|
|
} catch (_) {
|
|
|
|
ltx = null;
|
|
|
|
}
|
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
// Logging.instance.log("$value $tx $ltx");
|
|
|
|
if (!jindexes!.contains(value.index) && tx == null) {
|
|
|
|
// This coin is not confirmed and may be replaced
|
|
|
|
} else if (jindexes.contains(value.index) &&
|
|
|
|
tx == null &&
|
|
|
|
!value.isUsed &&
|
|
|
|
ltx != null &&
|
2023-01-12 00:16:17 +00:00
|
|
|
!ltx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) {
|
2022-09-06 15:34:39 +00:00
|
|
|
unconfirmedLelantusBalance += value.value;
|
|
|
|
} else if (jindexes.contains(value.index) && !value.isUsed) {
|
|
|
|
intLelantusBalance += value.value;
|
|
|
|
} else if (!value.isUsed &&
|
2023-01-12 00:16:17 +00:00
|
|
|
(tx == null
|
|
|
|
? true
|
|
|
|
: tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) !=
|
|
|
|
false)) {
|
2022-09-06 15:34:39 +00:00
|
|
|
intLelantusBalance += value.value;
|
2023-01-12 00:16:17 +00:00
|
|
|
} else if (tx != null &&
|
|
|
|
tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) ==
|
|
|
|
false) {
|
2022-09-06 15:34:39 +00:00
|
|
|
unconfirmedLelantusBalance += value.value;
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
});
|
|
|
|
}
|
2022-09-06 15:34:39 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// final int utxosIntValue = utxos.satoshiBalance;
|
|
|
|
// final Decimal utxosValue =
|
|
|
|
// Format.satoshisToAmount(utxosIntValue, coin: coin);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// List<Decimal> balances = List.empty(growable: true);
|
|
|
|
//
|
|
|
|
// Decimal lelantusBalance =
|
|
|
|
// Format.satoshisToAmount(intLelantusBalance, coin: coin);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// balances.add(lelantusBalance); 0
|
|
|
|
//
|
|
|
|
// balances.add(lelantusBalance * price); 1
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// Decimal _unconfirmedLelantusBalance =
|
|
|
|
// Format.satoshisToAmount(unconfirmedLelantusBalance, coin: coin);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// balances.add(lelantusBalance + utxosValue + _unconfirmedLelantusBalance); 2
|
|
|
|
//
|
|
|
|
// balances.add(
|
|
|
|
// (lelantusBalance + utxosValue + _unconfirmedLelantusBalance) * price); 3
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// int availableSats =
|
|
|
|
// utxos.satoshiBalance - utxos.satoshiBalanceUnconfirmed;
|
|
|
|
// if (availableSats < 0) {
|
|
|
|
// availableSats = 0;
|
|
|
|
// }
|
|
|
|
// balances.add(Format.satoshisToAmount(availableSats, coin: coin)); 4
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// wait for updated uxtos to get updated public balance
|
|
|
|
await utxosUpdateFuture;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// todo: shared total between private and public balances?
|
|
|
|
_balancePrivate = Balance(
|
|
|
|
coin: coin,
|
|
|
|
total: intLelantusBalance + unconfirmedLelantusBalance + balance.total,
|
|
|
|
spendable: intLelantusBalance,
|
|
|
|
blockedTotal: 0,
|
|
|
|
pendingSpendable: unconfirmedLelantusBalance + balance.total,
|
|
|
|
);
|
2023-01-12 21:20:57 +00:00
|
|
|
await updateCachedBalanceSecondary(_balancePrivate!);
|
2023-01-12 00:16:17 +00:00
|
|
|
// _balance = Balance(
|
|
|
|
// coin: coin,
|
|
|
|
// total: utxos.satoshiBalance,
|
|
|
|
// spendable: availableSats,
|
|
|
|
// blockedTotal: 0,
|
|
|
|
// pendingSpendable: utxos.satoshiBalanceUnconfirmed,
|
|
|
|
// );
|
|
|
|
|
|
|
|
// Logging.instance.log("balances $balances", level: LogLevel.Info);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'totalBalance',
|
|
|
|
// value: balances[2].toString());
|
|
|
|
// return balances;
|
2022-08-26 08:11:35 +00:00
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown in getFullBalance(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-09-06 21:55:01 +00:00
|
|
|
Future<void> anonymizeAllPublicFunds() async {
|
|
|
|
try {
|
|
|
|
var mintResult = await _mintSelection();
|
|
|
|
if (mintResult.isEmpty) {
|
|
|
|
Logging.instance.log("nothing to mint", level: LogLevel.Info);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
await _submitLelantusToNetwork(mintResult);
|
|
|
|
unawaited(refresh());
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception caught in anonymizeAllPublicFunds(): $e\n$s",
|
|
|
|
level: LogLevel.Warning);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Returns the mint transaction hex to mint all of the available funds.
|
|
|
|
Future<Map<String, dynamic>> _mintSelection() async {
|
2023-01-12 00:16:17 +00:00
|
|
|
final currentChainHeight = await chainHeight;
|
|
|
|
final List<isar_models.UTXO> availableOutputs = await utxos;
|
|
|
|
final List<isar_models.UTXO?> spendableOutputs = [];
|
2022-09-06 21:55:01 +00:00
|
|
|
|
|
|
|
// Build list of spendable outputs and totaling their satoshi amount
|
|
|
|
for (var i = 0; i < availableOutputs.length; i++) {
|
2023-01-12 00:16:17 +00:00
|
|
|
if (availableOutputs[i].isBlocked == false &&
|
|
|
|
availableOutputs[i]
|
|
|
|
.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) ==
|
|
|
|
true &&
|
2022-09-06 21:55:01 +00:00
|
|
|
!(availableOutputs[i].isCoinbase &&
|
2023-01-12 00:16:17 +00:00
|
|
|
availableOutputs[i].getConfirmations(currentChainHeight) <=
|
|
|
|
101)) {
|
2022-09-06 21:55:01 +00:00
|
|
|
spendableOutputs.add(availableOutputs[i]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
final List<Map<dynamic, LelantusCoin>> lelantusCoins = getLelantusCoinMap();
|
|
|
|
if (lelantusCoins.isNotEmpty) {
|
|
|
|
lelantusCoins.removeWhere((element) =>
|
|
|
|
element.values.any((elementCoin) => elementCoin.value == 0));
|
|
|
|
}
|
|
|
|
final data = await _txnData;
|
2023-01-12 00:16:17 +00:00
|
|
|
for (final value in data) {
|
2022-09-06 21:55:01 +00:00
|
|
|
if (value.inputs.isNotEmpty) {
|
|
|
|
for (var element in value.inputs) {
|
|
|
|
if (lelantusCoins
|
|
|
|
.any((element) => element.keys.contains(value.txid)) &&
|
|
|
|
spendableOutputs.firstWhere(
|
|
|
|
(output) => output?.txid == element.txid,
|
|
|
|
orElse: () => null) !=
|
|
|
|
null) {
|
|
|
|
spendableOutputs
|
|
|
|
.removeWhere((output) => output!.txid == element.txid);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
}
|
2022-09-06 21:55:01 +00:00
|
|
|
|
|
|
|
// If there is no Utxos to mint then stop the function.
|
|
|
|
if (spendableOutputs.isEmpty) {
|
|
|
|
Logging.instance.log("_mintSelection(): No spendable outputs found",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
return {};
|
|
|
|
}
|
|
|
|
|
|
|
|
int satoshisBeingUsed = 0;
|
2023-01-12 00:16:17 +00:00
|
|
|
List<isar_models.UTXO> utxoObjectsToUse = [];
|
2022-09-06 21:55:01 +00:00
|
|
|
|
|
|
|
for (var i = 0; i < spendableOutputs.length; i++) {
|
|
|
|
final spendable = spendableOutputs[i];
|
|
|
|
if (spendable != null) {
|
|
|
|
utxoObjectsToUse.add(spendable);
|
|
|
|
satoshisBeingUsed += spendable.value;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var mintsWithoutFee = await createMintsFromAmount(satoshisBeingUsed);
|
|
|
|
|
|
|
|
var tmpTx = await buildMintTransaction(
|
|
|
|
utxoObjectsToUse, satoshisBeingUsed, mintsWithoutFee);
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
int vSize = (tmpTx['transaction'] as Transaction).virtualSize();
|
|
|
|
final Decimal dvSize = Decimal.fromInt(vSize);
|
2022-09-06 21:55:01 +00:00
|
|
|
|
|
|
|
final feesObject = await fees;
|
|
|
|
|
2022-11-23 18:31:31 +00:00
|
|
|
final Decimal fastFee =
|
|
|
|
Format.satoshisToAmount(feesObject.fast, coin: coin);
|
2022-09-06 21:55:01 +00:00
|
|
|
int firoFee =
|
2023-01-12 00:16:17 +00:00
|
|
|
(dvSize * fastFee * Decimal.fromInt(100000)).toDouble().ceil();
|
|
|
|
// int firoFee = (vSize * feesObject.fast * (1 / 1000.0) * 100000000).ceil();
|
2022-09-06 21:55:01 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
if (firoFee < vSize) {
|
|
|
|
firoFee = vSize + 1;
|
2022-09-06 21:55:01 +00:00
|
|
|
}
|
|
|
|
firoFee = firoFee + 10;
|
|
|
|
int satoshiAmountToSend = satoshisBeingUsed - firoFee;
|
|
|
|
|
|
|
|
var mintsWithFee = await createMintsFromAmount(satoshiAmountToSend);
|
|
|
|
|
|
|
|
Map<String, dynamic> transaction = await buildMintTransaction(
|
|
|
|
utxoObjectsToUse, satoshiAmountToSend, mintsWithFee);
|
|
|
|
transaction['transaction'] = "";
|
|
|
|
Logging.instance.log(transaction.toString(), level: LogLevel.Info);
|
|
|
|
Logging.instance.log(transaction['txHex'], level: LogLevel.Info);
|
|
|
|
return transaction;
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
Future<List<Map<String, dynamic>>> createMintsFromAmount(int total) async {
|
|
|
|
var tmpTotal = total;
|
|
|
|
var index = 0;
|
|
|
|
var mints = <Map<String, dynamic>>[];
|
2023-01-12 21:20:57 +00:00
|
|
|
final nextFreeMintIndex = firoGetMintIndex()!;
|
2022-08-26 08:11:35 +00:00
|
|
|
while (tmpTotal > 0) {
|
|
|
|
final mintValue = min(tmpTotal, MINT_LIMIT);
|
|
|
|
final mint = await _getMintHex(
|
|
|
|
mintValue,
|
2022-09-06 15:34:39 +00:00
|
|
|
nextFreeMintIndex + index,
|
2022-08-26 08:11:35 +00:00
|
|
|
);
|
|
|
|
mints.add({
|
|
|
|
"value": mintValue,
|
|
|
|
"script": mint,
|
2022-09-06 15:34:39 +00:00
|
|
|
"index": nextFreeMintIndex + index,
|
2022-08-26 08:11:35 +00:00
|
|
|
"publicCoin": "",
|
|
|
|
});
|
|
|
|
tmpTotal = tmpTotal - MINT_LIMIT;
|
|
|
|
index++;
|
|
|
|
}
|
|
|
|
return mints;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// returns a valid txid if successful
|
|
|
|
Future<String> submitHexToNetwork(String hex) async {
|
|
|
|
try {
|
|
|
|
final txid = await electrumXClient.broadcastTransaction(rawTx: hex);
|
|
|
|
return txid;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Caught exception in submitHexToNetwork(\"$hex\"): $e $s",
|
|
|
|
printFullLength: true,
|
|
|
|
level: LogLevel.Info);
|
|
|
|
// return an invalid tx
|
|
|
|
return "transaction submission failed";
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Builds and signs a transaction
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<Map<String, dynamic>> buildMintTransaction(
|
|
|
|
List<isar_models.UTXO> utxosToUse,
|
|
|
|
int satoshisPerRecipient,
|
|
|
|
List<Map<String, dynamic>> mintsMap) async {
|
2022-12-13 00:17:02 +00:00
|
|
|
//todo: check if print needed
|
|
|
|
// debugPrint(utxosToUse.toString());
|
2022-08-26 08:11:35 +00:00
|
|
|
List<String> addressesToDerive = [];
|
|
|
|
|
|
|
|
// Populating the addresses to derive
|
|
|
|
for (var i = 0; i < utxosToUse.length; i++) {
|
|
|
|
final txid = utxosToUse[i].txid;
|
|
|
|
final outputIndex = utxosToUse[i].vout;
|
|
|
|
|
|
|
|
// txid may not work for this as txid may not always be the same as tx_hash?
|
|
|
|
final tx = await cachedElectrumXClient.getTransaction(
|
|
|
|
txHash: txid,
|
|
|
|
verbose: true,
|
|
|
|
coin: coin,
|
|
|
|
);
|
|
|
|
|
|
|
|
final vouts = tx["vout"] as List?;
|
|
|
|
if (vouts != null && outputIndex < vouts.length) {
|
|
|
|
final address =
|
|
|
|
vouts[outputIndex]["scriptPubKey"]["addresses"][0] as String?;
|
|
|
|
if (address != null) {
|
|
|
|
addressesToDerive.add(address);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
List<ECPair> elipticCurvePairArray = [];
|
|
|
|
List<Uint8List> outputDataArray = [];
|
|
|
|
|
|
|
|
final receiveDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_receiveDerivations");
|
|
|
|
final changeDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_changeDerivations");
|
|
|
|
|
|
|
|
final receiveDerivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(receiveDerivationsString ?? "{}") as Map);
|
|
|
|
final changeDerivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(changeDerivationsString ?? "{}") as Map);
|
|
|
|
|
|
|
|
for (var i = 0; i < addressesToDerive.length; i++) {
|
|
|
|
final addressToCheckFor = addressesToDerive[i];
|
|
|
|
|
|
|
|
for (var i = 0; i < receiveDerivations.length; i++) {
|
|
|
|
final receive = receiveDerivations["$i"];
|
|
|
|
final change = changeDerivations["$i"];
|
|
|
|
|
|
|
|
if (receive['address'] == addressToCheckFor) {
|
|
|
|
Logging.instance
|
|
|
|
.log('Receiving found on loop $i', level: LogLevel.Info);
|
|
|
|
// Logging.instance.log(
|
|
|
|
// 'decoded receive[\'wif\'] version: ${wif.decode(receive['wif'] as String)}, _network: $_network');
|
|
|
|
elipticCurvePairArray
|
|
|
|
.add(ECPair.fromWIF(receive['wif'] as String, network: _network));
|
|
|
|
outputDataArray.add(P2PKH(
|
|
|
|
network: _network,
|
|
|
|
data: PaymentData(
|
|
|
|
pubkey: Format.stringToUint8List(
|
|
|
|
receive['publicKey'] as String)))
|
|
|
|
.data
|
|
|
|
.output!);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (change['address'] == addressToCheckFor) {
|
|
|
|
Logging.instance.log('Change found on loop $i', level: LogLevel.Info);
|
|
|
|
// Logging.instance.log(
|
|
|
|
// 'decoded change[\'wif\'] version: ${wif.decode(change['wif'] as String)}, _network: $_network');
|
|
|
|
elipticCurvePairArray
|
|
|
|
.add(ECPair.fromWIF(change['wif'] as String, network: _network));
|
|
|
|
|
|
|
|
outputDataArray.add(P2PKH(
|
|
|
|
network: _network,
|
|
|
|
data: PaymentData(
|
|
|
|
pubkey: Format.stringToUint8List(
|
|
|
|
change['publicKey'] as String)))
|
|
|
|
.data
|
|
|
|
.output!);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
final txb = TransactionBuilder(network: _network);
|
|
|
|
txb.setVersion(2);
|
|
|
|
|
|
|
|
int height = await getBlockHead(electrumXClient);
|
|
|
|
txb.setLockTime(height);
|
|
|
|
int amount = 0;
|
|
|
|
// Add transaction inputs
|
|
|
|
for (var i = 0; i < utxosToUse.length; i++) {
|
|
|
|
txb.addInput(
|
|
|
|
utxosToUse[i].txid, utxosToUse[i].vout, null, outputDataArray[i]);
|
|
|
|
amount += utxosToUse[i].value;
|
|
|
|
}
|
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
final index = firoGetMintIndex()!;
|
2022-08-26 08:11:35 +00:00
|
|
|
Logging.instance.log("index of mint $index", level: LogLevel.Info);
|
|
|
|
|
|
|
|
for (var mintsElement in mintsMap) {
|
|
|
|
Logging.instance.log("using $mintsElement", level: LogLevel.Info);
|
|
|
|
Uint8List mintu8 =
|
|
|
|
Format.stringToUint8List(mintsElement['script'] as String);
|
|
|
|
txb.addOutput(mintu8, mintsElement['value'] as int);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (var i = 0; i < utxosToUse.length; i++) {
|
|
|
|
txb.sign(
|
|
|
|
vin: i,
|
|
|
|
keyPair: elipticCurvePairArray[i],
|
|
|
|
witnessValue: utxosToUse[i].value,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
var incomplete = txb.buildIncomplete();
|
|
|
|
var txId = incomplete.getId();
|
|
|
|
var txHex = incomplete.toHex();
|
|
|
|
int fee = amount - incomplete.outs[0].value!;
|
|
|
|
|
|
|
|
var builtHex = txb.build();
|
|
|
|
// return builtHex;
|
2023-01-12 00:16:17 +00:00
|
|
|
// final locale =
|
|
|
|
// Platform.isWindows ? "en_US" : await Devicelocale.currentLocale;
|
2022-08-26 08:11:35 +00:00
|
|
|
return {
|
|
|
|
"transaction": builtHex,
|
|
|
|
"txid": txId,
|
|
|
|
"txHex": txHex,
|
|
|
|
"value": amount - fee,
|
2022-11-23 18:31:31 +00:00
|
|
|
"fees": Format.satoshisToAmount(fee, coin: coin).toDouble(),
|
2022-08-26 08:11:35 +00:00
|
|
|
"publicCoin": "",
|
|
|
|
"height": height,
|
|
|
|
"txType": "Sent",
|
|
|
|
"confirmed_status": false,
|
2022-11-23 18:31:31 +00:00
|
|
|
"amount": Format.satoshisToAmount(amount, coin: coin).toDouble(),
|
2022-08-26 08:11:35 +00:00
|
|
|
"timestamp": DateTime.now().millisecondsSinceEpoch ~/ 1000,
|
|
|
|
"subType": "mint",
|
|
|
|
"mintsMap": mintsMap,
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<void> _refreshLelantusData() async {
|
2022-08-26 08:11:35 +00:00
|
|
|
final List<Map<dynamic, LelantusCoin>> lelantusCoins = getLelantusCoinMap();
|
2023-01-12 21:20:57 +00:00
|
|
|
final jindexes = firoGetJIndex();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
// Get all joinsplit transaction ids
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
final listLelantusTxData = await lelantusTransactionData;
|
2022-08-26 08:11:35 +00:00
|
|
|
List<String> joinsplits = [];
|
2023-01-12 00:16:17 +00:00
|
|
|
for (final tx in listLelantusTxData) {
|
|
|
|
if (tx.subType == isar_models.TransactionSubType.join) {
|
2022-08-26 08:11:35 +00:00
|
|
|
joinsplits.add(tx.txid);
|
|
|
|
}
|
|
|
|
}
|
2022-09-06 15:34:39 +00:00
|
|
|
for (final coin
|
|
|
|
in lelantusCoins.fold(<LelantusCoin>[], (previousValue, element) {
|
|
|
|
(previousValue as List<LelantusCoin>).add(element.values.first);
|
|
|
|
return previousValue;
|
|
|
|
})) {
|
|
|
|
if (jindexes != null) {
|
|
|
|
if (jindexes.contains(coin.index) && !joinsplits.contains(coin.txId)) {
|
|
|
|
joinsplits.add(coin.txId);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Grab the most recent information on all the joinsplits
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
final updatedJSplit = await getJMintTransactions(
|
|
|
|
cachedElectrumXClient,
|
|
|
|
joinsplits,
|
|
|
|
coin,
|
|
|
|
);
|
|
|
|
|
|
|
|
final currentChainHeight = await chainHeight;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
// update all of joinsplits that are now confirmed.
|
|
|
|
for (final tx in updatedJSplit) {
|
2023-01-12 00:16:17 +00:00
|
|
|
isar_models.Transaction? currentTx;
|
|
|
|
|
|
|
|
try {
|
|
|
|
currentTx = listLelantusTxData.firstWhere((e) => e.txid == tx.txid);
|
|
|
|
} catch (_) {
|
|
|
|
currentTx = null;
|
|
|
|
}
|
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
if (currentTx == null) {
|
2022-08-26 08:11:35 +00:00
|
|
|
// this send was accidentally not included in the list
|
2023-01-12 00:16:17 +00:00
|
|
|
tx.isLelantus = true;
|
|
|
|
listLelantusTxData.add(tx);
|
2022-08-26 08:11:35 +00:00
|
|
|
continue;
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
if (currentTx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS) !=
|
|
|
|
tx.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) {
|
|
|
|
listLelantusTxData.removeWhere((e) => e.txid == tx.txid);
|
|
|
|
tx.isLelantus = true;
|
|
|
|
listLelantusTxData.add(tx);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Logging.instance.log(txData.txChunks);
|
2023-01-12 00:16:17 +00:00
|
|
|
final listTxData = await _txnData;
|
|
|
|
for (final value in listTxData) {
|
2022-08-26 08:11:35 +00:00
|
|
|
// ignore change addresses
|
2022-09-09 17:14:38 +00:00
|
|
|
// bool hasAtLeastOneReceive = false;
|
2022-09-06 21:55:01 +00:00
|
|
|
// int howManyReceiveInputs = 0;
|
2022-09-09 17:14:38 +00:00
|
|
|
// for (var element in value.inputs) {
|
|
|
|
// if (listLelantusTxData.containsKey(element.txid) &&
|
|
|
|
// listLelantusTxData[element.txid]!.txType == "Received"
|
|
|
|
// // &&
|
|
|
|
// // listLelantusTxData[element.txid].subType != "mint"
|
|
|
|
// ) {
|
|
|
|
// // hasAtLeastOneReceive = true;
|
|
|
|
// // howManyReceiveInputs++;
|
|
|
|
// }
|
|
|
|
// }
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
if (value.type == isar_models.TransactionType.incoming &&
|
|
|
|
value.subType != isar_models.TransactionSubType.mint) {
|
2022-09-07 22:14:10 +00:00
|
|
|
// Every receive other than a mint should be shown. Mints will be collected and shown from the send side
|
2023-01-12 00:16:17 +00:00
|
|
|
listLelantusTxData.removeWhere((e) => e.txid == value.txid);
|
|
|
|
value.isLelantus = true;
|
|
|
|
listLelantusTxData.add(value);
|
|
|
|
} else if (value.type == isar_models.TransactionType.outgoing) {
|
2022-09-07 22:14:10 +00:00
|
|
|
// all sends should be shown, mints will be displayed correctly in the ui
|
2023-01-12 00:16:17 +00:00
|
|
|
listLelantusTxData.removeWhere((e) => e.txid == value.txid);
|
|
|
|
value.isLelantus = true;
|
|
|
|
listLelantusTxData.add(value);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// TODO: optimize this whole lelantus process
|
|
|
|
await isar.writeTxn(
|
|
|
|
() async => isar.transactions.putAllByTxid(listLelantusTxData));
|
|
|
|
|
|
|
|
// // update the _lelantusTransactionData
|
|
|
|
// final models.TransactionData newTxData =
|
|
|
|
// models.TransactionData.fromMap(listLelantusTxData);
|
|
|
|
// // Logging.instance.log(newTxData.txChunks);
|
|
|
|
// _lelantusTransactionData = Future(() => newTxData);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: 'latest_lelantus_tx_model', value: newTxData);
|
|
|
|
// return newTxData;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
Future<String> _getMintHex(int amount, int index) async {
|
|
|
|
final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic');
|
|
|
|
final mintKeyPair = await compute(
|
|
|
|
getBip32NodeWrapper,
|
|
|
|
Tuple4(
|
|
|
|
MINT_INDEX,
|
|
|
|
index,
|
|
|
|
mnemonic!,
|
|
|
|
_network,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
String keydata = Format.uint8listToString(mintKeyPair.privateKey!);
|
|
|
|
String seedID = Format.uint8listToString(mintKeyPair.identifier);
|
|
|
|
|
|
|
|
String mintHex = await compute(
|
|
|
|
_getMintScriptWrapper,
|
|
|
|
Tuple5(
|
|
|
|
amount,
|
|
|
|
keydata,
|
|
|
|
index,
|
|
|
|
seedID,
|
|
|
|
coin == Coin.firoTestNet,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
return mintHex;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<bool> _submitLelantusToNetwork(
|
|
|
|
Map<String, dynamic> transactionInfo) async {
|
2022-09-06 11:47:57 +00:00
|
|
|
final latestSetId = await getLatestSetId();
|
2022-08-26 08:11:35 +00:00
|
|
|
final txid = await submitHexToNetwork(transactionInfo['txHex'] as String);
|
|
|
|
// success if txid matches the generated txid
|
|
|
|
Logging.instance.log(
|
|
|
|
"_submitLelantusToNetwork txid: ${transactionInfo['txid']}",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
if (txid == transactionInfo['txid']) {
|
2023-01-12 21:20:57 +00:00
|
|
|
final index = firoGetMintIndex();
|
2022-08-26 08:11:35 +00:00
|
|
|
final List<Map<dynamic, LelantusCoin>> lelantusCoins =
|
|
|
|
getLelantusCoinMap();
|
|
|
|
List<Map<dynamic, LelantusCoin>> coins;
|
2022-09-06 15:34:39 +00:00
|
|
|
if (lelantusCoins.isEmpty) {
|
2022-08-26 08:11:35 +00:00
|
|
|
coins = [];
|
|
|
|
} else {
|
|
|
|
coins = [...lelantusCoins];
|
|
|
|
}
|
|
|
|
|
|
|
|
if (transactionInfo['spendCoinIndexes'] != null) {
|
|
|
|
// This is a joinsplit
|
|
|
|
|
|
|
|
// Update all of the coins that have been spent.
|
2022-09-06 15:34:39 +00:00
|
|
|
for (final lCoinMap in coins) {
|
|
|
|
final lCoin = lCoinMap.values.first;
|
2022-08-26 08:11:35 +00:00
|
|
|
if ((transactionInfo['spendCoinIndexes'] as List<int>)
|
2022-09-06 15:34:39 +00:00
|
|
|
.contains(lCoin.index)) {
|
|
|
|
lCoinMap[lCoinMap.keys.first] = LelantusCoin(
|
|
|
|
lCoin.index,
|
|
|
|
lCoin.value,
|
|
|
|
lCoin.publicCoin,
|
|
|
|
lCoin.txId,
|
|
|
|
lCoin.anonymitySetId,
|
2022-08-26 08:11:35 +00:00
|
|
|
true);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// if a jmint was made add it to the unspent coin index
|
|
|
|
LelantusCoin jmint = LelantusCoin(
|
|
|
|
index!,
|
|
|
|
transactionInfo['jmintValue'] as int? ?? 0,
|
|
|
|
transactionInfo['publicCoin'] as String,
|
|
|
|
transactionInfo['txid'] as String,
|
2022-09-06 11:47:57 +00:00
|
|
|
latestSetId,
|
2022-08-26 08:11:35 +00:00
|
|
|
false);
|
|
|
|
if (jmint.value > 0) {
|
|
|
|
coins.add({jmint.txId: jmint});
|
2023-01-12 21:20:57 +00:00
|
|
|
final jindexes = firoGetJIndex()!;
|
|
|
|
jindexes.add(index);
|
|
|
|
await firoUpdateJIndex(jindexes);
|
|
|
|
await firoUpdateMintIndex(index + 1);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
2023-01-12 21:20:57 +00:00
|
|
|
await firoUpdateLelantusCoins(coins);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
// add the send transaction
|
2023-01-12 00:16:17 +00:00
|
|
|
final transaction = isar_models.Transaction()
|
|
|
|
..txid = transactionInfo['txid'] as String
|
|
|
|
..timestamp = transactionInfo['timestamp'] as int? ??
|
|
|
|
(DateTime.now().millisecondsSinceEpoch ~/ 1000)
|
|
|
|
..type = transactionInfo['txType'] == "Received"
|
|
|
|
? isar_models.TransactionType.incoming
|
|
|
|
: isar_models.TransactionType.outgoing
|
|
|
|
..amount = Format.decimalAmountToSatoshis(
|
|
|
|
Decimal.parse(transactionInfo["amount"].toString()), coin)
|
|
|
|
..fee = Format.decimalAmountToSatoshis(
|
|
|
|
Decimal.parse(transactionInfo["fees"].toString()), coin)
|
2023-01-13 17:11:41 +00:00
|
|
|
..address.value = await isar.addresses
|
|
|
|
.filter()
|
|
|
|
.valueEqualTo(transactionInfo["address"] as String)
|
|
|
|
.findFirst()
|
2023-01-12 00:16:17 +00:00
|
|
|
..height = transactionInfo["height"] as int?
|
|
|
|
..subType = transactionInfo["subType"] == "mint"
|
|
|
|
? isar_models.TransactionSubType.mint
|
|
|
|
: transactionInfo["subType"] == "join"
|
|
|
|
? isar_models.TransactionSubType.join
|
|
|
|
: isar_models.TransactionSubType.none
|
|
|
|
..otherData = transactionInfo["otherData"] as String?
|
|
|
|
..isLelantus = true
|
|
|
|
..isCancelled = false;
|
|
|
|
|
|
|
|
await isar.writeTxn(() async {
|
|
|
|
await isar.transactions.put(transaction);
|
|
|
|
});
|
|
|
|
|
|
|
|
// final models.TransactionData newTxData =
|
|
|
|
// models.TransactionData.fromMap(transactions);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'latest_lelantus_tx_model',
|
|
|
|
// value: newTxData);
|
|
|
|
// final ldata = DB.instance.get<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'latest_lelantus_tx_model') as models.TransactionData;
|
|
|
|
// _lelantusTransactionData = Future(() => ldata);
|
2022-08-26 08:11:35 +00:00
|
|
|
} else {
|
|
|
|
// This is a mint
|
|
|
|
Logging.instance.log("this is a mint", level: LogLevel.Info);
|
|
|
|
|
|
|
|
// TODO: transactionInfo['mintsMap']
|
|
|
|
for (final mintMap
|
|
|
|
in transactionInfo['mintsMap'] as List<Map<String, dynamic>>) {
|
|
|
|
final index = mintMap['index'] as int?;
|
|
|
|
LelantusCoin mint = LelantusCoin(
|
|
|
|
index!,
|
|
|
|
mintMap['value'] as int,
|
|
|
|
mintMap['publicCoin'] as String,
|
|
|
|
transactionInfo['txid'] as String,
|
2022-09-06 11:47:57 +00:00
|
|
|
latestSetId,
|
2022-08-26 08:11:35 +00:00
|
|
|
false,
|
|
|
|
);
|
|
|
|
if (mint.value > 0) {
|
|
|
|
coins.add({mint.txId: mint});
|
2023-01-12 21:20:57 +00:00
|
|
|
await firoUpdateMintIndex(index + 1);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
// Logging.instance.log(coins);
|
2023-01-12 21:20:57 +00:00
|
|
|
await firoUpdateLelantusCoins(coins);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
return true;
|
|
|
|
} else {
|
|
|
|
// Failed to send to network
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<FeeObject> _getFees() async {
|
|
|
|
try {
|
|
|
|
//TODO adjust numbers for different speeds?
|
|
|
|
const int f = 1, m = 5, s = 20;
|
|
|
|
|
|
|
|
final fast = await electrumXClient.estimateFee(blocks: f);
|
|
|
|
final medium = await electrumXClient.estimateFee(blocks: m);
|
|
|
|
final slow = await electrumXClient.estimateFee(blocks: s);
|
|
|
|
|
|
|
|
final feeObject = FeeObject(
|
|
|
|
numberOfBlocksFast: f,
|
|
|
|
numberOfBlocksAverage: m,
|
|
|
|
numberOfBlocksSlow: s,
|
2022-11-23 18:31:31 +00:00
|
|
|
fast: Format.decimalAmountToSatoshis(fast, coin),
|
|
|
|
medium: Format.decimalAmountToSatoshis(medium, coin),
|
|
|
|
slow: Format.decimalAmountToSatoshis(slow, coin),
|
2022-08-26 08:11:35 +00:00
|
|
|
);
|
|
|
|
|
|
|
|
Logging.instance.log("fetched fees: $feeObject", level: LogLevel.Info);
|
|
|
|
return feeObject;
|
|
|
|
|
|
|
|
// final result = await electrumXClient.getFeeRate();
|
|
|
|
//
|
|
|
|
// final locale = await Devicelocale.currentLocale;
|
|
|
|
// final String fee =
|
|
|
|
// Format.satoshiAmountToPrettyString(result["rate"] as int, locale!);
|
|
|
|
//
|
|
|
|
// final fees = {
|
|
|
|
// "fast": fee,
|
|
|
|
// "average": fee,
|
|
|
|
// "slow": fee,
|
|
|
|
// };
|
|
|
|
// final FeeObject feeObject = FeeObject.fromJson(fees);
|
|
|
|
// return feeObject;
|
|
|
|
} catch (e) {
|
|
|
|
Logging.instance
|
|
|
|
.log("Exception rethrown from _getFees(): $e", level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<ElectrumXNode> _getCurrentNode() async {
|
2022-11-09 22:43:26 +00:00
|
|
|
final node = NodeService(secureStorageInterface: _secureStore)
|
|
|
|
.getPrimaryNodeFor(coin: coin) ??
|
2022-08-26 08:11:35 +00:00
|
|
|
DefaultNodes.getNodeFor(coin);
|
|
|
|
|
|
|
|
return ElectrumXNode(
|
|
|
|
address: node.host,
|
|
|
|
port: node.port,
|
|
|
|
name: node.name,
|
|
|
|
useSSL: node.useSSL,
|
|
|
|
id: node.id,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
//TODO call get transaction and check each tx to see if it is a "received" tx?
|
|
|
|
Future<int> _getReceivedTxCount({required String address}) async {
|
|
|
|
try {
|
|
|
|
final scripthash = AddressUtils.convertToScriptHash(address, _network);
|
|
|
|
final transactions =
|
|
|
|
await electrumXClient.getHistory(scripthash: scripthash);
|
|
|
|
return transactions.length;
|
|
|
|
} catch (e) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown in _getReceivedTxCount(address: $address): $e",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<void> checkReceivingAddressForTransactions() async {
|
|
|
|
try {
|
2023-01-12 00:16:17 +00:00
|
|
|
final currentReceiving = await _currentReceivingAddress;
|
|
|
|
|
|
|
|
final int txCount =
|
|
|
|
await _getReceivedTxCount(address: currentReceiving.value);
|
2022-08-26 08:11:35 +00:00
|
|
|
Logging.instance.log(
|
2023-01-12 00:16:17 +00:00
|
|
|
'Number of txs for current receiving address $currentReceiving: $txCount',
|
2022-08-26 08:11:35 +00:00
|
|
|
level: LogLevel.Info);
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
if (txCount >= 1) {
|
|
|
|
// First increment the receiving index
|
|
|
|
final newReceivingIndex = currentReceiving.derivationIndex + 1;
|
|
|
|
|
|
|
|
// Use new index to derive a new receiving address
|
|
|
|
final newReceivingAddress = await _generateAddressForChain(
|
|
|
|
0,
|
|
|
|
newReceivingIndex,
|
|
|
|
);
|
|
|
|
|
|
|
|
// Add that new receiving address
|
|
|
|
await isar.writeTxn(() async {
|
|
|
|
await isar.addresses.put(newReceivingAddress);
|
|
|
|
});
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
} on SocketException catch (se, s) {
|
|
|
|
Logging.instance.log(
|
2022-09-07 15:56:10 +00:00
|
|
|
"SocketException caught in checkReceivingAddressForTransactions(): $se\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
return;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown from checkReceivingAddressForTransactions(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<void> checkChangeAddressForTransactions() async {
|
|
|
|
try {
|
2023-01-12 00:16:17 +00:00
|
|
|
final currentChange = await _currentChangeAddress;
|
|
|
|
final int txCount =
|
|
|
|
await _getReceivedTxCount(address: currentChange.value);
|
2022-09-07 15:56:10 +00:00
|
|
|
Logging.instance.log(
|
2023-01-12 00:16:17 +00:00
|
|
|
'Number of txs for current change address: $currentChange: $txCount',
|
2022-09-07 15:56:10 +00:00
|
|
|
level: LogLevel.Info);
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
if (txCount >= 1) {
|
|
|
|
// First increment the change index
|
|
|
|
final newChangeIndex = currentChange.derivationIndex + 1;
|
|
|
|
|
|
|
|
// Use new index to derive a new change address
|
|
|
|
final newChangeAddress = await _generateAddressForChain(
|
|
|
|
1,
|
|
|
|
newChangeIndex,
|
|
|
|
);
|
|
|
|
|
|
|
|
// Add that new change address
|
|
|
|
await isar.writeTxn(() async {
|
|
|
|
await isar.addresses.put(newChangeAddress);
|
|
|
|
});
|
2022-09-07 15:56:10 +00:00
|
|
|
}
|
|
|
|
} on SocketException catch (se, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"SocketException caught in checkChangeAddressForTransactions(): $se\n$s",
|
2022-08-26 08:11:35 +00:00
|
|
|
level: LogLevel.Error);
|
|
|
|
return;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
2022-09-07 15:56:10 +00:00
|
|
|
"Exception rethrown from checkChangeAddressForTransactions(): $e\n$s",
|
2022-08-26 08:11:35 +00:00
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<List<isar_models.Address>> _fetchAllOwnAddresses() async {
|
|
|
|
final allAddresses = await isar.addresses
|
|
|
|
.filter()
|
|
|
|
.subTypeEqualTo(isar_models.AddressSubType.receiving)
|
|
|
|
.or()
|
|
|
|
.subTypeEqualTo(isar_models.AddressSubType.change)
|
|
|
|
.findAll();
|
|
|
|
// final List<String> allAddresses = [];
|
|
|
|
// final receivingAddresses =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'receivingAddresses')
|
|
|
|
// as List<dynamic>;
|
|
|
|
// final changeAddresses =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'changeAddresses')
|
|
|
|
// as List<dynamic>;
|
|
|
|
//
|
|
|
|
// for (var i = 0; i < receivingAddresses.length; i++) {
|
|
|
|
// allAddresses.add(receivingAddresses[i] as String);
|
|
|
|
// }
|
|
|
|
// for (var i = 0; i < changeAddresses.length; i++) {
|
|
|
|
// allAddresses.add(changeAddresses[i] as String);
|
|
|
|
// }
|
2022-08-26 08:11:35 +00:00
|
|
|
return allAddresses;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<List<Map<String, dynamic>>> _fetchHistory(
|
|
|
|
List<String> allAddresses) async {
|
|
|
|
try {
|
|
|
|
List<Map<String, dynamic>> allTxHashes = [];
|
2022-09-09 15:07:27 +00:00
|
|
|
|
|
|
|
final Map<int, Map<String, List<dynamic>>> batches = {};
|
|
|
|
final Map<String, String> requestIdToAddressMap = {};
|
|
|
|
const batchSizeMax = 100;
|
|
|
|
int batchNumber = 0;
|
|
|
|
for (int i = 0; i < allAddresses.length; i++) {
|
|
|
|
if (batches[batchNumber] == null) {
|
|
|
|
batches[batchNumber] = {};
|
|
|
|
}
|
|
|
|
final scripthash =
|
|
|
|
AddressUtils.convertToScriptHash(allAddresses[i], _network);
|
2022-09-12 15:22:05 +00:00
|
|
|
final id = Logger.isTestEnv ? "$i" : const Uuid().v1();
|
2022-09-09 15:07:27 +00:00
|
|
|
requestIdToAddressMap[id] = allAddresses[i];
|
|
|
|
batches[batchNumber]!.addAll({
|
|
|
|
id: [scripthash]
|
|
|
|
});
|
|
|
|
if (i % batchSizeMax == batchSizeMax - 1) {
|
|
|
|
batchNumber++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (int i = 0; i < batches.length; i++) {
|
|
|
|
final response =
|
|
|
|
await _electrumXClient.getBatchHistory(args: batches[i]!);
|
|
|
|
for (final entry in response.entries) {
|
|
|
|
for (int j = 0; j < entry.value.length; j++) {
|
|
|
|
entry.value[j]["address"] = requestIdToAddressMap[entry.key];
|
|
|
|
if (!allTxHashes.contains(entry.value[j])) {
|
|
|
|
allTxHashes.add(entry.value[j]);
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return allTxHashes;
|
|
|
|
} catch (e, s) {
|
2022-09-09 15:07:27 +00:00
|
|
|
Logging.instance.log("_fetchHistory: $e\n$s", level: LogLevel.Error);
|
|
|
|
rethrow;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-13 21:36:19 +00:00
|
|
|
bool _duplicateTxCheck(
|
|
|
|
List<Map<String, dynamic>> allTransactions, String txid) {
|
|
|
|
for (int i = 0; i < allTransactions.length; i++) {
|
|
|
|
if (allTransactions[i]["txid"] == txid) {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<void> _refreshTransactions() async {
|
|
|
|
final List<isar_models.Address> allAddresses =
|
|
|
|
await _fetchAllOwnAddresses();
|
|
|
|
|
|
|
|
final List<Map<String, dynamic>> allTxHashes =
|
|
|
|
await _fetchHistory(allAddresses.map((e) => e.value).toList());
|
|
|
|
|
2023-01-13 21:36:19 +00:00
|
|
|
List<Map<String, dynamic>> allTransactions = [];
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-13 21:36:19 +00:00
|
|
|
final currentHeight = await chainHeight;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-13 21:36:19 +00:00
|
|
|
for (final txHash in allTxHashes) {
|
|
|
|
final storedTx = await isar.transactions
|
|
|
|
.where()
|
|
|
|
.txidEqualTo(txHash["tx_hash"] as String)
|
|
|
|
.findFirst();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-13 21:36:19 +00:00
|
|
|
if (storedTx == null ||
|
|
|
|
!storedTx.isConfirmed(currentHeight, MINIMUM_CONFIRMATIONS)) {
|
|
|
|
final tx = await cachedElectrumXClient.getTransaction(
|
|
|
|
txHash: txHash["tx_hash"] as String,
|
|
|
|
verbose: true,
|
|
|
|
coin: coin,
|
|
|
|
);
|
|
|
|
|
|
|
|
if (!_duplicateTxCheck(allTransactions, tx["txid"] as String)) {
|
|
|
|
tx["address"] = await isar.addresses
|
|
|
|
.filter()
|
|
|
|
.valueEqualTo(txHash["address"] as String)
|
|
|
|
.findFirst();
|
|
|
|
tx["height"] = txHash["height"];
|
|
|
|
allTransactions.add(tx);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
final List<
|
|
|
|
Tuple4<isar_models.Transaction, List<isar_models.Output>,
|
|
|
|
List<isar_models.Input>, isar_models.Address?>> txnsData = [];
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-13 21:36:19 +00:00
|
|
|
for (final txObject in allTransactions) {
|
|
|
|
final data = await parseTransaction(
|
2023-01-12 00:16:17 +00:00
|
|
|
txObject,
|
|
|
|
cachedElectrumXClient,
|
|
|
|
allAddresses,
|
|
|
|
coin,
|
|
|
|
MINIMUM_CONFIRMATIONS,
|
|
|
|
);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-13 21:36:19 +00:00
|
|
|
txnsData.add(data);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
2023-01-13 21:36:19 +00:00
|
|
|
|
|
|
|
await addNewTransactionData(txnsData);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<void> _refreshUTXOs() async {
|
|
|
|
final allAddresses = await _fetchAllOwnAddresses();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
try {
|
2022-09-09 15:07:27 +00:00
|
|
|
final fetchedUtxoList = <List<Map<String, dynamic>>>[];
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
final Map<int, Map<String, List<dynamic>>> batches = {};
|
|
|
|
const batchSizeMax = 100;
|
|
|
|
int batchNumber = 0;
|
2022-08-26 08:11:35 +00:00
|
|
|
for (int i = 0; i < allAddresses.length; i++) {
|
2022-09-09 15:07:27 +00:00
|
|
|
if (batches[batchNumber] == null) {
|
|
|
|
batches[batchNumber] = {};
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
final scripthash =
|
2023-01-12 00:16:17 +00:00
|
|
|
AddressUtils.convertToScriptHash(allAddresses[i].value, _network);
|
2022-09-09 15:07:27 +00:00
|
|
|
batches[batchNumber]!.addAll({
|
|
|
|
scripthash: [scripthash]
|
|
|
|
});
|
|
|
|
if (i % batchSizeMax == batchSizeMax - 1) {
|
|
|
|
batchNumber++;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
for (int i = 0; i < batches.length; i++) {
|
|
|
|
final response =
|
|
|
|
await _electrumXClient.getBatchUTXOs(args: batches[i]!);
|
|
|
|
for (final entry in response.entries) {
|
|
|
|
if (entry.value.isNotEmpty) {
|
|
|
|
fetchedUtxoList.add(entry.value);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
|
|
|
|
final currentChainHeight = await chainHeight;
|
|
|
|
|
|
|
|
final List<isar_models.UTXO> outputArray = [];
|
|
|
|
int satoshiBalanceTotal = 0;
|
2022-09-06 23:27:14 +00:00
|
|
|
int satoshiBalancePending = 0;
|
2023-01-12 00:16:17 +00:00
|
|
|
int satoshiBalanceSpendable = 0;
|
|
|
|
int satoshiBalanceBlocked = 0;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
for (int i = 0; i < fetchedUtxoList.length; i++) {
|
|
|
|
for (int j = 0; j < fetchedUtxoList[i].length; j++) {
|
2022-08-26 08:11:35 +00:00
|
|
|
final txn = await cachedElectrumXClient.getTransaction(
|
2022-09-09 15:07:27 +00:00
|
|
|
txHash: fetchedUtxoList[i][j]["tx_hash"] as String,
|
2022-08-26 08:11:35 +00:00
|
|
|
verbose: true,
|
|
|
|
coin: coin,
|
|
|
|
);
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
final utxo = isar_models.UTXO();
|
|
|
|
|
|
|
|
utxo.txid = txn["txid"] as String;
|
|
|
|
utxo.vout = fetchedUtxoList[i][j]["tx_pos"] as int;
|
|
|
|
utxo.value = fetchedUtxoList[i][j]["value"] as int;
|
|
|
|
utxo.name = "";
|
|
|
|
|
|
|
|
// todo check here if we should mark as blocked
|
|
|
|
utxo.isBlocked = false;
|
|
|
|
utxo.blockedReason = null;
|
|
|
|
|
|
|
|
utxo.isCoinbase = txn["is_coinbase"] as bool? ?? false;
|
|
|
|
utxo.blockHash = txn["blockhash"] as String?;
|
|
|
|
utxo.blockHeight = fetchedUtxoList[i][j]["height"] as int?;
|
|
|
|
utxo.blockTime = txn["blocktime"] as int?;
|
|
|
|
|
|
|
|
satoshiBalanceTotal += utxo.value;
|
|
|
|
|
|
|
|
if (utxo.isBlocked) {
|
|
|
|
satoshiBalanceBlocked += utxo.value;
|
|
|
|
} else {
|
|
|
|
if (utxo.isConfirmed(currentChainHeight, MINIMUM_CONFIRMATIONS)) {
|
|
|
|
satoshiBalanceSpendable += utxo.value;
|
|
|
|
} else {
|
|
|
|
satoshiBalancePending += utxo.value;
|
|
|
|
}
|
2022-09-06 23:27:14 +00:00
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
outputArray.add(utxo);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Logging.instance
|
|
|
|
.log('Outputs fetched: $outputArray', level: LogLevel.Info);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
await isar.writeTxn(() async {
|
|
|
|
await isar.utxos.clear();
|
|
|
|
await isar.utxos.putAll(outputArray);
|
|
|
|
});
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// finally update public balance
|
|
|
|
_balance = Balance(
|
|
|
|
coin: coin,
|
|
|
|
total: satoshiBalanceTotal,
|
|
|
|
spendable: satoshiBalanceSpendable,
|
|
|
|
blockedTotal: satoshiBalanceBlocked,
|
|
|
|
pendingSpendable: satoshiBalancePending,
|
|
|
|
);
|
2023-01-12 21:20:57 +00:00
|
|
|
await updateCachedBalance(_balance!);
|
2022-09-09 15:07:27 +00:00
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance
|
|
|
|
.log("Output fetch unsuccessful: $e\n$s", level: LogLevel.Error);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Returns the latest receiving/change (external/internal) address for the wallet depending on [chain]
|
|
|
|
/// [chain] - Use 0 for receiving (external), 1 for change (internal). Should not be any other value!
|
|
|
|
Future<String> _getCurrentAddressForChain(int chain) async {
|
2023-01-12 00:16:17 +00:00
|
|
|
final subType = chain == 0 // Here, we assume that chain == 1 if it isn't 0
|
|
|
|
? isar_models.AddressSubType.receiving
|
|
|
|
: isar_models.AddressSubType.change;
|
|
|
|
|
|
|
|
isar_models.Address? address = await isar.addresses
|
|
|
|
.filter()
|
|
|
|
.typeEqualTo(isar_models.AddressType.p2pkh)
|
|
|
|
.subTypeEqualTo(subType)
|
|
|
|
.sortByDerivationIndexDesc()
|
|
|
|
.findFirst();
|
|
|
|
|
|
|
|
return address!.value;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
Future<void> fillAddresses(String suppliedMnemonic,
|
|
|
|
{int perBatch = 50, int numberOfThreads = 4}) async {
|
|
|
|
if (numberOfThreads <= 0) {
|
|
|
|
numberOfThreads = 1;
|
|
|
|
}
|
|
|
|
if (Platform.environment["FLUTTER_TEST"] == "true" || integrationTestFlag) {
|
|
|
|
perBatch = 10;
|
2022-08-29 15:02:06 +00:00
|
|
|
numberOfThreads = 4;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
final receiveDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_receiveDerivations");
|
|
|
|
final changeDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_changeDerivations");
|
|
|
|
|
|
|
|
var receiveDerivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(receiveDerivationsString ?? "{}") as Map);
|
|
|
|
var changeDerivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(changeDerivationsString ?? "{}") as Map);
|
|
|
|
|
|
|
|
final int start = receiveDerivations.length;
|
|
|
|
|
|
|
|
List<ReceivePort> ports = List.empty(growable: true);
|
|
|
|
for (int i = 0; i < numberOfThreads; i++) {
|
|
|
|
ReceivePort receivePort = await getIsolate({
|
|
|
|
"function": "isolateDerive",
|
|
|
|
"mnemonic": suppliedMnemonic,
|
|
|
|
"from": start + i * perBatch,
|
|
|
|
"to": start + (i + 1) * perBatch,
|
|
|
|
"network": _network,
|
|
|
|
});
|
|
|
|
ports.add(receivePort);
|
|
|
|
}
|
|
|
|
for (int i = 0; i < numberOfThreads; i++) {
|
|
|
|
ReceivePort receivePort = ports.elementAt(i);
|
|
|
|
var message = await receivePort.first;
|
|
|
|
if (message is String) {
|
|
|
|
Logging.instance.log("this is a string", level: LogLevel.Error);
|
|
|
|
stop(receivePort);
|
|
|
|
throw Exception("isolateDerive isolate failed");
|
|
|
|
}
|
|
|
|
stop(receivePort);
|
|
|
|
Logging.instance.log('Closing isolateDerive!', level: LogLevel.Info);
|
|
|
|
receiveDerivations.addAll(message['receive'] as Map<String, dynamic>);
|
|
|
|
changeDerivations.addAll(message['change'] as Map<String, dynamic>);
|
|
|
|
}
|
|
|
|
Logging.instance.log("isolate derives", level: LogLevel.Info);
|
|
|
|
// Logging.instance.log(receiveDerivations);
|
|
|
|
// Logging.instance.log(changeDerivations);
|
|
|
|
|
|
|
|
final newReceiveDerivationsString = jsonEncode(receiveDerivations);
|
|
|
|
final newChangeDerivationsString = jsonEncode(changeDerivations);
|
|
|
|
|
|
|
|
await _secureStore.write(
|
|
|
|
key: "${walletId}_receiveDerivations",
|
|
|
|
value: newReceiveDerivationsString);
|
|
|
|
await _secureStore.write(
|
|
|
|
key: "${walletId}_changeDerivations",
|
|
|
|
value: newChangeDerivationsString);
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Generates a new internal or external chain address for the wallet using a BIP84 derivation path.
|
|
|
|
/// [chain] - Use 0 for receiving (external), 1 for change (internal). Should not be any other value!
|
|
|
|
/// [index] - This can be any integer >= 0
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<isar_models.Address> _generateAddressForChain(
|
|
|
|
int chain, int index) async {
|
2022-08-26 08:11:35 +00:00
|
|
|
// final wallet = await Hive.openBox(this._walletId);
|
|
|
|
final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic');
|
|
|
|
Map<String, dynamic>? derivations;
|
|
|
|
if (chain == 0) {
|
|
|
|
final receiveDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_receiveDerivations");
|
|
|
|
derivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(receiveDerivationsString ?? "{}") as Map);
|
|
|
|
} else if (chain == 1) {
|
|
|
|
final changeDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_changeDerivations");
|
|
|
|
derivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(changeDerivationsString ?? "{}") as Map);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (derivations!.isNotEmpty) {
|
|
|
|
if (derivations["$index"] == null) {
|
|
|
|
await fillAddresses(mnemonic!,
|
|
|
|
numberOfThreads: Platform.numberOfProcessors - isolates.length - 1);
|
|
|
|
Logging.instance.log("calling _generateAddressForChain recursively",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
return _generateAddressForChain(chain, index);
|
|
|
|
}
|
2023-01-12 00:16:17 +00:00
|
|
|
return isar_models.Address()
|
|
|
|
..value = derivations["$index"]['address'] as String
|
|
|
|
..publicKey = Format.stringToUint8List(
|
|
|
|
derivations["$index"]['publicKey'] as String)
|
|
|
|
..subType = chain == 0
|
|
|
|
? isar_models.AddressSubType.receiving
|
|
|
|
: isar_models.AddressSubType.change
|
|
|
|
..type = isar_models.AddressType.p2pkh
|
|
|
|
..derivationIndex = index;
|
2022-08-26 08:11:35 +00:00
|
|
|
} else {
|
|
|
|
final node = await compute(
|
|
|
|
getBip32NodeWrapper, Tuple4(chain, index, mnemonic!, _network));
|
2023-01-12 00:16:17 +00:00
|
|
|
final address =
|
|
|
|
P2PKH(network: _network, data: PaymentData(pubkey: node.publicKey))
|
|
|
|
.data
|
|
|
|
.address!;
|
|
|
|
|
|
|
|
return isar_models.Address()
|
|
|
|
..value = address
|
|
|
|
..publicKey = node.publicKey
|
|
|
|
..subType = chain == 0
|
|
|
|
? isar_models.AddressSubType.receiving
|
|
|
|
: isar_models.AddressSubType.change
|
|
|
|
..type = isar_models.AddressType.p2pkh
|
|
|
|
..derivationIndex = index;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
// /// Takes in a list of isar_models.UTXOs and adds a name (dependent on object index within list)
|
|
|
|
// /// and checks for the txid associated with the utxo being blocked and marks it accordingly.
|
|
|
|
// /// Now also checks for output labeling.
|
|
|
|
// Future<void> _sortOutputs(List<isar_models.UTXO> utxos) async {
|
|
|
|
// final blockedHashArray =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'blocked_tx_hashes')
|
|
|
|
// as List<dynamic>?;
|
|
|
|
// final List<String> lst = [];
|
|
|
|
// if (blockedHashArray != null) {
|
|
|
|
// for (var hash in blockedHashArray) {
|
|
|
|
// lst.add(hash as String);
|
|
|
|
// }
|
|
|
|
// }
|
|
|
|
// final labels =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'labels') as Map? ??
|
|
|
|
// {};
|
|
|
|
//
|
|
|
|
// _outputsList = [];
|
|
|
|
//
|
|
|
|
// for (var i = 0; i < utxos.length; i++) {
|
|
|
|
// if (labels[utxos[i].txid] != null) {
|
|
|
|
// utxos[i].txName = labels[utxos[i].txid] as String? ?? "";
|
|
|
|
// } else {
|
|
|
|
// utxos[i].txName = 'Output #$i';
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// if (utxos[i].status.confirmed == false) {
|
|
|
|
// _outputsList.add(utxos[i]);
|
|
|
|
// } else {
|
|
|
|
// if (lst.contains(utxos[i].txid)) {
|
|
|
|
// utxos[i].blocked = true;
|
|
|
|
// _outputsList.add(utxos[i]);
|
|
|
|
// } else if (!lst.contains(utxos[i].txid)) {
|
|
|
|
// _outputsList.add(utxos[i]);
|
|
|
|
// }
|
|
|
|
// }
|
|
|
|
// }
|
|
|
|
// }
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
@override
|
|
|
|
Future<void> fullRescan(
|
|
|
|
int maxUnusedAddressGap,
|
|
|
|
int maxNumberOfIndexesToCheck,
|
|
|
|
) async {
|
|
|
|
Logging.instance.log("Starting full rescan!", level: LogLevel.Info);
|
|
|
|
// timer?.cancel();
|
|
|
|
// for (final isolate in isolates.values) {
|
|
|
|
// isolate.kill(priority: Isolate.immediate);
|
|
|
|
// }
|
|
|
|
// isolates.clear();
|
|
|
|
longMutex = true;
|
|
|
|
GlobalEventBus.instance.fire(
|
|
|
|
WalletSyncStatusChangedEvent(
|
|
|
|
WalletSyncStatus.syncing,
|
|
|
|
walletId,
|
|
|
|
coin,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
|
|
|
|
// clear cache
|
2022-09-06 15:34:39 +00:00
|
|
|
await _cachedElectrumXClient.clearSharedTransactionCache(coin: coin);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
// back up data
|
2023-01-12 21:20:57 +00:00
|
|
|
// await _rescanBackup();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
try {
|
|
|
|
final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic');
|
|
|
|
await _recoverWalletFromBIP32SeedPhrase(mnemonic!, maxUnusedAddressGap);
|
|
|
|
|
|
|
|
longMutex = false;
|
|
|
|
Logging.instance.log("Full rescan complete!", level: LogLevel.Info);
|
|
|
|
GlobalEventBus.instance.fire(
|
|
|
|
WalletSyncStatusChangedEvent(
|
|
|
|
WalletSyncStatus.synced,
|
|
|
|
walletId,
|
|
|
|
coin,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
} catch (e, s) {
|
|
|
|
GlobalEventBus.instance.fire(
|
|
|
|
WalletSyncStatusChangedEvent(
|
|
|
|
WalletSyncStatus.unableToSync,
|
|
|
|
walletId,
|
|
|
|
coin,
|
|
|
|
),
|
|
|
|
);
|
|
|
|
|
|
|
|
// restore from backup
|
2023-01-12 21:20:57 +00:00
|
|
|
// await _rescanRestore();
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
longMutex = false;
|
|
|
|
Logging.instance.log("Exception rethrown from fullRescan(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
// Future<void> _rescanBackup() async {
|
|
|
|
// Logging.instance.log("starting rescan backup", level: LogLevel.Info);
|
|
|
|
//
|
|
|
|
// // backup current and clear data
|
|
|
|
// final tempReceivingAddresses =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'receivingAddresses');
|
|
|
|
// await DB.instance.delete<dynamic>(
|
|
|
|
// key: 'receivingAddresses',
|
|
|
|
// boxName: walletId,
|
|
|
|
// );
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'receivingAddresses_BACKUP',
|
|
|
|
// value: tempReceivingAddresses);
|
|
|
|
//
|
|
|
|
// final tempChangeAddresses =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'changeAddresses');
|
|
|
|
// await DB.instance.delete<dynamic>(
|
|
|
|
// key: 'changeAddresses',
|
|
|
|
// boxName: walletId,
|
|
|
|
// );
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'changeAddresses_BACKUP',
|
|
|
|
// value: tempChangeAddresses);
|
|
|
|
//
|
|
|
|
// final tempReceivingIndex =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'receivingIndex');
|
|
|
|
// await DB.instance.delete<dynamic>(
|
|
|
|
// key: 'receivingIndex',
|
|
|
|
// boxName: walletId,
|
|
|
|
// );
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'receivingIndex_BACKUP',
|
|
|
|
// value: tempReceivingIndex);
|
|
|
|
//
|
|
|
|
// final tempChangeIndex =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'changeIndex');
|
|
|
|
// await DB.instance.delete<dynamic>(
|
|
|
|
// key: 'changeIndex',
|
|
|
|
// boxName: walletId,
|
|
|
|
// );
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: 'changeIndex_BACKUP', value: tempChangeIndex);
|
|
|
|
//
|
|
|
|
// final receiveDerivationsString =
|
|
|
|
// await _secureStore.read(key: "${walletId}_receiveDerivations");
|
|
|
|
// final changeDerivationsString =
|
|
|
|
// await _secureStore.read(key: "${walletId}_changeDerivations");
|
|
|
|
//
|
|
|
|
// await _secureStore.write(
|
|
|
|
// key: "${walletId}_receiveDerivations_BACKUP",
|
|
|
|
// value: receiveDerivationsString);
|
|
|
|
// await _secureStore.write(
|
|
|
|
// key: "${walletId}_changeDerivations_BACKUP",
|
|
|
|
// value: changeDerivationsString);
|
|
|
|
//
|
|
|
|
// await _secureStore.write(
|
|
|
|
// key: "${walletId}_receiveDerivations", value: null);
|
|
|
|
// await _secureStore.write(key: "${walletId}_changeDerivations", value: null);
|
|
|
|
//
|
|
|
|
// // back up but no need to delete
|
|
|
|
// final tempMintIndex =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'mintIndex');
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: 'mintIndex_BACKUP', value: tempMintIndex);
|
|
|
|
//
|
|
|
|
// final tempLelantusCoins =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: '_lelantus_coins');
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: '_lelantus_coins_BACKUP',
|
|
|
|
// value: tempLelantusCoins);
|
|
|
|
//
|
|
|
|
// final tempJIndex =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'jindex');
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: 'jindex_BACKUP', value: tempJIndex);
|
|
|
|
//
|
|
|
|
// final tempLelantusTxModel = DB.instance
|
|
|
|
// .get<dynamic>(boxName: walletId, key: 'latest_lelantus_tx_model');
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'latest_lelantus_tx_model_BACKUP',
|
|
|
|
// value: tempLelantusTxModel);
|
|
|
|
//
|
|
|
|
// Logging.instance.log("rescan backup complete", level: LogLevel.Info);
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// Future<void> _rescanRestore() async {
|
|
|
|
// Logging.instance.log("starting rescan restore", level: LogLevel.Info);
|
|
|
|
//
|
|
|
|
// // restore from backup
|
|
|
|
// final tempReceivingAddresses = DB.instance
|
|
|
|
// .get<dynamic>(boxName: walletId, key: 'receivingAddresses_BACKUP');
|
|
|
|
// final tempChangeAddresses = DB.instance
|
|
|
|
// .get<dynamic>(boxName: walletId, key: 'changeAddresses_BACKUP');
|
|
|
|
// final tempReceivingIndex = DB.instance
|
|
|
|
// .get<dynamic>(boxName: walletId, key: 'receivingIndex_BACKUP');
|
|
|
|
// final tempChangeIndex =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'changeIndex_BACKUP');
|
|
|
|
// final tempMintIndex =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'mintIndex_BACKUP');
|
|
|
|
// final tempLelantusCoins = DB.instance
|
|
|
|
// .get<dynamic>(boxName: walletId, key: '_lelantus_coins_BACKUP');
|
|
|
|
// final tempJIndex =
|
|
|
|
// DB.instance.get<dynamic>(boxName: walletId, key: 'jindex_BACKUP');
|
|
|
|
// final tempLelantusTxModel = DB.instance.get<dynamic>(
|
|
|
|
// boxName: walletId, key: 'latest_lelantus_tx_model_BACKUP');
|
|
|
|
//
|
|
|
|
// final receiveDerivationsString =
|
|
|
|
// await _secureStore.read(key: "${walletId}_receiveDerivations_BACKUP");
|
|
|
|
// final changeDerivationsString =
|
|
|
|
// await _secureStore.read(key: "${walletId}_changeDerivations_BACKUP");
|
|
|
|
//
|
|
|
|
// await _secureStore.write(
|
|
|
|
// key: "${walletId}_receiveDerivations", value: receiveDerivationsString);
|
|
|
|
// await _secureStore.write(
|
|
|
|
// key: "${walletId}_changeDerivations", value: changeDerivationsString);
|
|
|
|
//
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'receivingAddresses',
|
|
|
|
// value: tempReceivingAddresses);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: 'changeAddresses', value: tempChangeAddresses);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: 'receivingIndex', value: tempReceivingIndex);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: 'changeIndex', value: tempChangeIndex);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: 'mintIndex', value: tempMintIndex);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId, key: '_lelantus_coins', value: tempLelantusCoins);
|
|
|
|
// await DB.instance
|
|
|
|
// .put<dynamic>(boxName: walletId, key: 'jindex', value: tempJIndex);
|
|
|
|
// await DB.instance.put<dynamic>(
|
|
|
|
// boxName: walletId,
|
|
|
|
// key: 'latest_lelantus_tx_model',
|
|
|
|
// value: tempLelantusTxModel);
|
|
|
|
//
|
|
|
|
// Logging.instance.log("rescan restore complete", level: LogLevel.Info);
|
|
|
|
// }
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
/// wrapper for _recoverWalletFromBIP32SeedPhrase()
|
|
|
|
@override
|
|
|
|
Future<void> recoverFromMnemonic({
|
|
|
|
required String mnemonic,
|
|
|
|
required int maxUnusedAddressGap,
|
|
|
|
required int maxNumberOfIndexesToCheck,
|
|
|
|
required int height,
|
|
|
|
}) async {
|
|
|
|
try {
|
|
|
|
await compute(
|
|
|
|
_setTestnetWrapper,
|
|
|
|
coin == Coin.firoTestNet,
|
|
|
|
);
|
|
|
|
Logging.instance.log("IS_INTEGRATION_TEST: $integrationTestFlag",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
if (!integrationTestFlag) {
|
|
|
|
final features = await electrumXClient.getServerFeatures();
|
|
|
|
Logging.instance.log("features: $features", level: LogLevel.Info);
|
|
|
|
switch (coin) {
|
|
|
|
case Coin.firo:
|
|
|
|
if (features['genesis_hash'] != GENESIS_HASH_MAINNET) {
|
|
|
|
throw Exception("genesis hash does not match main net!");
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case Coin.firoTestNet:
|
|
|
|
if (features['genesis_hash'] != GENESIS_HASH_TESTNET) {
|
|
|
|
throw Exception("genesis hash does not match test net!");
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
throw Exception(
|
|
|
|
"Attempted to generate a FiroWallet using a non firo coin type: ${coin.name}");
|
|
|
|
}
|
|
|
|
// if (_networkType == BasicNetworkType.main) {
|
|
|
|
// if (features['genesis_hash'] != GENESIS_HASH_MAINNET) {
|
|
|
|
// throw Exception("genesis hash does not match main net!");
|
|
|
|
// }
|
|
|
|
// } else if (_networkType == BasicNetworkType.test) {
|
|
|
|
// if (features['genesis_hash'] != GENESIS_HASH_TESTNET) {
|
|
|
|
// throw Exception("genesis hash does not match test net!");
|
|
|
|
// }
|
|
|
|
// }
|
|
|
|
}
|
|
|
|
// this should never fail
|
|
|
|
if ((await _secureStore.read(key: '${_walletId}_mnemonic')) != null) {
|
|
|
|
throw Exception("Attempted to overwrite mnemonic on restore!");
|
|
|
|
}
|
|
|
|
await _secureStore.write(
|
|
|
|
key: '${_walletId}_mnemonic', value: mnemonic.trim());
|
|
|
|
await _recoverWalletFromBIP32SeedPhrase(
|
|
|
|
mnemonic.trim(), maxUnusedAddressGap);
|
|
|
|
|
|
|
|
await compute(
|
|
|
|
_setTestnetWrapper,
|
|
|
|
false,
|
|
|
|
);
|
|
|
|
} catch (e, s) {
|
|
|
|
await compute(
|
|
|
|
_setTestnetWrapper,
|
|
|
|
false,
|
|
|
|
);
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown from recoverFromMnemonic(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool longMutex = false;
|
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
Future<Map<int, dynamic>> getSetDataMap(int latestSetId) async {
|
|
|
|
final Map<int, dynamic> setDataMap = {};
|
|
|
|
final anonymitySets = await fetchAnonymitySets();
|
|
|
|
for (int setId = 1; setId <= latestSetId; setId++) {
|
|
|
|
final setData = anonymitySets
|
|
|
|
.firstWhere((element) => element["setId"] == setId, orElse: () => {});
|
|
|
|
|
|
|
|
if (setData.isNotEmpty) {
|
|
|
|
setDataMap[setId] = setData;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
2022-09-09 15:07:27 +00:00
|
|
|
}
|
|
|
|
return setDataMap;
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
Future<void> _makeDerivations(
|
2023-01-12 00:16:17 +00:00
|
|
|
String suppliedMnemonic,
|
|
|
|
int maxUnusedAddressGap,
|
|
|
|
) async {
|
|
|
|
List<isar_models.Address> receivingAddressArray = [];
|
|
|
|
List<isar_models.Address> changeAddressArray = [];
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
int receivingIndex = -1;
|
|
|
|
int changeIndex = -1;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
// The gap limit will be capped at 20
|
|
|
|
int receivingGapCounter = 0;
|
|
|
|
int changeGapCounter = 0;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
await fillAddresses(suppliedMnemonic,
|
|
|
|
numberOfThreads: Platform.numberOfProcessors - isolates.length - 1);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
final receiveDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_receiveDerivations");
|
|
|
|
final changeDerivationsString =
|
|
|
|
await _secureStore.read(key: "${walletId}_changeDerivations");
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
final receiveDerivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(receiveDerivationsString ?? "{}") as Map);
|
|
|
|
final changeDerivations = Map<String, dynamic>.from(
|
|
|
|
jsonDecode(changeDerivationsString ?? "{}") as Map);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
// log("rcv: $receiveDerivations");
|
|
|
|
// log("chg: $changeDerivations");
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
// Deriving and checking for receiving addresses
|
|
|
|
for (var i = 0; i < receiveDerivations.length; i++) {
|
|
|
|
// Break out of loop when receivingGapCounter hits maxUnusedAddressGap
|
|
|
|
// Same gap limit for change as for receiving, breaks when it hits maxUnusedAddressGap
|
|
|
|
if (receivingGapCounter >= maxUnusedAddressGap &&
|
|
|
|
changeGapCounter >= maxUnusedAddressGap) {
|
|
|
|
break;
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
final receiveDerivation = receiveDerivations["$i"];
|
|
|
|
final address = receiveDerivation['address'] as String;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
final changeDerivation = changeDerivations["$i"];
|
|
|
|
final _address = changeDerivation['address'] as String;
|
|
|
|
Future<int>? futureNumTxs;
|
|
|
|
Future<int>? _futureNumTxs;
|
|
|
|
if (receivingGapCounter < maxUnusedAddressGap) {
|
|
|
|
futureNumTxs = _getReceivedTxCount(address: address);
|
|
|
|
}
|
|
|
|
if (changeGapCounter < maxUnusedAddressGap) {
|
|
|
|
_futureNumTxs = _getReceivedTxCount(address: _address);
|
|
|
|
}
|
|
|
|
try {
|
|
|
|
if (futureNumTxs != null) {
|
|
|
|
int numTxs = await futureNumTxs;
|
|
|
|
if (numTxs >= 1) {
|
|
|
|
receivingIndex = i;
|
2023-01-12 00:16:17 +00:00
|
|
|
final addr = isar_models.Address()
|
|
|
|
..value = address
|
|
|
|
..type = isar_models.AddressType.p2pkh
|
|
|
|
..subType = isar_models.AddressSubType.receiving
|
|
|
|
..derivationIndex = i
|
|
|
|
..publicKey = Format.stringToUint8List(
|
|
|
|
receiveDerivation['publicKey'] as String);
|
|
|
|
receivingAddressArray.add(addr);
|
2022-09-09 15:07:27 +00:00
|
|
|
} else if (numTxs == 0) {
|
|
|
|
receivingGapCounter += 1;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
2022-09-09 15:07:27 +00:00
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown from recoverWalletFromBIP32SeedPhrase(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
try {
|
|
|
|
if (_futureNumTxs != null) {
|
|
|
|
int numTxs = await _futureNumTxs;
|
|
|
|
if (numTxs >= 1) {
|
|
|
|
changeIndex = i;
|
2023-01-12 00:16:17 +00:00
|
|
|
final addr = isar_models.Address()
|
|
|
|
..value = address
|
|
|
|
..type = isar_models.AddressType.p2pkh
|
|
|
|
..subType = isar_models.AddressSubType.change
|
|
|
|
..derivationIndex = i
|
|
|
|
..publicKey = Format.stringToUint8List(
|
|
|
|
changeDerivation['publicKey'] as String);
|
|
|
|
changeAddressArray.add(addr);
|
2022-09-09 15:07:27 +00:00
|
|
|
} else if (numTxs == 0) {
|
|
|
|
changeGapCounter += 1;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
}
|
2022-09-09 15:07:27 +00:00
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown from recoverWalletFromBIP32SeedPhrase(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
2022-09-09 15:07:27 +00:00
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
// If restoring a wallet that never received any funds, then set receivingArray manually
|
|
|
|
// If we didn't do this, it'd store an empty array
|
|
|
|
if (receivingIndex == -1) {
|
2023-01-12 00:16:17 +00:00
|
|
|
final receivingAddress = await _generateAddressForChain(0, 0);
|
2022-09-09 15:07:27 +00:00
|
|
|
receivingAddressArray.add(receivingAddress);
|
|
|
|
}
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
// If restoring a wallet that never sent any funds with change, then set changeArray
|
|
|
|
// manually. If we didn't do this, it'd store an empty array.
|
|
|
|
if (changeIndex == -1) {
|
2023-01-12 00:16:17 +00:00
|
|
|
final changeAddress = await _generateAddressForChain(1, 0);
|
2022-09-09 15:07:27 +00:00
|
|
|
changeAddressArray.add(changeAddress);
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
await isar.writeTxn(() async {
|
|
|
|
await isar.addresses.putAll([
|
|
|
|
...receivingAddressArray,
|
|
|
|
...changeAddressArray,
|
|
|
|
]);
|
|
|
|
});
|
2022-09-09 15:07:27 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/// Recovers wallet from [suppliedMnemonic]. Expects a valid mnemonic.
|
|
|
|
Future<void> _recoverWalletFromBIP32SeedPhrase(
|
|
|
|
String suppliedMnemonic, int maxUnusedAddressGap) async {
|
|
|
|
longMutex = true;
|
|
|
|
Logging.instance
|
|
|
|
.log("PROCESSORS ${Platform.numberOfProcessors}", level: LogLevel.Info);
|
|
|
|
try {
|
2023-01-12 18:46:01 +00:00
|
|
|
await isarInit(walletId);
|
2023-01-12 00:16:17 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
final latestSetId = await getLatestSetId();
|
|
|
|
final setDataMap = getSetDataMap(latestSetId);
|
|
|
|
final usedSerialNumbers = getUsedCoinSerials();
|
|
|
|
final makeDerivations =
|
|
|
|
_makeDerivations(suppliedMnemonic, maxUnusedAddressGap);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
await Future.wait([
|
|
|
|
updateCachedId(walletId),
|
|
|
|
updateCachedIsFavorite(false),
|
|
|
|
]);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
await Future.wait([usedSerialNumbers, setDataMap, makeDerivations]);
|
|
|
|
|
|
|
|
await _restore(latestSetId, await setDataMap, await usedSerialNumbers);
|
2022-08-26 08:11:35 +00:00
|
|
|
longMutex = false;
|
|
|
|
} catch (e, s) {
|
|
|
|
longMutex = false;
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown from recoverWalletFromBIP32SeedPhrase(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<void> _restore(int latestSetId, Map<dynamic, dynamic> setDataMap,
|
|
|
|
dynamic usedSerialNumbers) async {
|
|
|
|
final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic');
|
2022-09-09 17:14:38 +00:00
|
|
|
final dataFuture = _txnData;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
ReceivePort receivePort = await getIsolate({
|
|
|
|
"function": "restore",
|
|
|
|
"mnemonic": mnemonic,
|
|
|
|
"coin": coin,
|
|
|
|
"latestSetId": latestSetId,
|
|
|
|
"setDataMap": setDataMap,
|
|
|
|
"usedSerialNumbers": usedSerialNumbers,
|
|
|
|
"network": _network,
|
|
|
|
});
|
|
|
|
|
2022-09-09 17:14:38 +00:00
|
|
|
await Future.wait([dataFuture]);
|
2022-09-09 15:07:27 +00:00
|
|
|
var result = await receivePort.first;
|
|
|
|
if (result is String) {
|
2022-08-26 08:11:35 +00:00
|
|
|
Logging.instance
|
|
|
|
.log("restore() ->> this is a string", level: LogLevel.Error);
|
|
|
|
stop(receivePort);
|
|
|
|
throw Exception("isolate restore failed.");
|
|
|
|
}
|
|
|
|
stop(receivePort);
|
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
final message = await staticProcessRestore(
|
2023-01-12 00:16:17 +00:00
|
|
|
(await dataFuture),
|
|
|
|
result as Map<dynamic, dynamic>,
|
|
|
|
await chainHeight,
|
|
|
|
);
|
2022-09-09 15:07:27 +00:00
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
await Future.wait([
|
|
|
|
firoUpdateMintIndex(message['mintIndex'] as int),
|
|
|
|
firoUpdateLelantusCoins(message['_lelantus_coins'] as List),
|
|
|
|
firoUpdateJIndex(message['jindex'] as List),
|
|
|
|
]);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
final transactionMap =
|
2023-01-12 00:16:17 +00:00
|
|
|
message["newTxMap"] as Map<String, isar_models.Transaction>;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
// Create the joinsplit transactions.
|
|
|
|
final spendTxs = await getJMintTransactions(
|
2023-01-12 00:16:17 +00:00
|
|
|
_cachedElectrumXClient,
|
|
|
|
message["spendTxIds"] as List<String>,
|
|
|
|
coin,
|
|
|
|
);
|
2022-08-26 08:11:35 +00:00
|
|
|
Logging.instance.log(spendTxs, level: LogLevel.Info);
|
|
|
|
for (var element in spendTxs) {
|
|
|
|
transactionMap[element.txid] = element;
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
await isar.writeTxn(() async {
|
|
|
|
await isar.transactions.putAllByTxid(transactionMap.values.toList());
|
|
|
|
});
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
Future<List<Map<String, dynamic>>> fetchAnonymitySets() async {
|
|
|
|
try {
|
|
|
|
final latestSetId = await getLatestSetId();
|
|
|
|
|
|
|
|
final List<Map<String, dynamic>> sets = [];
|
2022-09-09 17:14:38 +00:00
|
|
|
List<Future<Map<String, dynamic>>> anonFutures = [];
|
2022-08-26 08:11:35 +00:00
|
|
|
for (int i = 1; i <= latestSetId; i++) {
|
2022-09-09 15:07:27 +00:00
|
|
|
final set = cachedElectrumXClient.getAnonymitySet(
|
2022-08-26 08:11:35 +00:00
|
|
|
groupId: "$i",
|
|
|
|
coin: coin,
|
|
|
|
);
|
2022-09-09 15:07:27 +00:00
|
|
|
anonFutures.add(set);
|
|
|
|
}
|
|
|
|
await Future.wait(anonFutures);
|
|
|
|
for (int i = 1; i <= latestSetId; i++) {
|
2022-09-09 17:14:38 +00:00
|
|
|
Map<String, dynamic> set = (await anonFutures[i - 1]);
|
2022-08-26 08:11:35 +00:00
|
|
|
set["setId"] = i;
|
|
|
|
sets.add(set);
|
|
|
|
}
|
|
|
|
return sets;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown from refreshAnonymitySets: $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<dynamic> _createJoinSplitTransaction(
|
|
|
|
int spendAmount, String address, bool subtractFeeFromAmount) async {
|
2023-01-12 00:16:17 +00:00
|
|
|
// final price = await firoPrice;
|
2022-08-26 08:11:35 +00:00
|
|
|
final mnemonic = await _secureStore.read(key: '${_walletId}_mnemonic');
|
2023-01-12 21:20:57 +00:00
|
|
|
final index = firoGetMintIndex();
|
2022-08-26 08:11:35 +00:00
|
|
|
final lelantusEntry = await _getLelantusEntry();
|
|
|
|
final anonymitySets = await fetchAnonymitySets();
|
|
|
|
final locktime = await getBlockHead(electrumXClient);
|
2023-01-12 00:16:17 +00:00
|
|
|
// final locale =
|
|
|
|
// Platform.isWindows ? "en_US" : await Devicelocale.currentLocale;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
ReceivePort receivePort = await getIsolate({
|
|
|
|
"function": "createJoinSplit",
|
|
|
|
"spendAmount": spendAmount,
|
|
|
|
"address": address,
|
|
|
|
"subtractFeeFromAmount": subtractFeeFromAmount,
|
|
|
|
"mnemonic": mnemonic,
|
|
|
|
"index": index,
|
2023-01-12 00:16:17 +00:00
|
|
|
// "price": price,
|
2022-08-26 08:11:35 +00:00
|
|
|
"lelantusEntries": lelantusEntry,
|
|
|
|
"locktime": locktime,
|
|
|
|
"coin": coin,
|
|
|
|
"network": _network,
|
|
|
|
"_anonymity_sets": anonymitySets,
|
2023-01-12 00:16:17 +00:00
|
|
|
// "locale": locale,
|
2022-08-26 08:11:35 +00:00
|
|
|
});
|
|
|
|
var message = await receivePort.first;
|
|
|
|
if (message is String) {
|
|
|
|
Logging.instance
|
|
|
|
.log("Error in CreateJoinSplit: $message", level: LogLevel.Error);
|
|
|
|
stop(receivePort);
|
|
|
|
return 3;
|
|
|
|
}
|
|
|
|
if (message is int) {
|
|
|
|
stop(receivePort);
|
|
|
|
return message;
|
|
|
|
}
|
|
|
|
stop(receivePort);
|
|
|
|
Logging.instance.log('Closing createJoinSplit!', level: LogLevel.Info);
|
|
|
|
return message;
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<int> getLatestSetId() async {
|
|
|
|
try {
|
|
|
|
final id = await electrumXClient.getLatestCoinId();
|
|
|
|
return id;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown in firo_wallet.dart: $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<List<dynamic>> getUsedCoinSerials() async {
|
|
|
|
try {
|
|
|
|
final response = await cachedElectrumXClient.getUsedCoinSerials(
|
|
|
|
coin: coin,
|
|
|
|
);
|
|
|
|
return response;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception rethrown in firo_wallet.dart: $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
@override
|
|
|
|
Future<void> exit() async {
|
|
|
|
_hasCalledExit = true;
|
|
|
|
timer?.cancel();
|
|
|
|
timer = null;
|
|
|
|
stopNetworkAlivePinging();
|
|
|
|
for (final isolate in isolates.values) {
|
|
|
|
isolate.kill(priority: Isolate.immediate);
|
|
|
|
}
|
|
|
|
isolates.clear();
|
|
|
|
Logging.instance
|
|
|
|
.log("$walletName firo_wallet exit finished", level: LogLevel.Info);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool _hasCalledExit = false;
|
|
|
|
|
|
|
|
@override
|
|
|
|
bool get hasCalledExit => _hasCalledExit;
|
|
|
|
|
|
|
|
bool isActive = false;
|
|
|
|
|
|
|
|
@override
|
|
|
|
void Function(bool)? get onIsActiveWalletChanged => (isActive) async {
|
|
|
|
timer?.cancel();
|
|
|
|
timer = null;
|
|
|
|
if (isActive) {
|
|
|
|
await compute(
|
|
|
|
_setTestnetWrapper,
|
|
|
|
coin == Coin.firoTestNet,
|
|
|
|
);
|
|
|
|
} else {
|
|
|
|
await compute(
|
|
|
|
_setTestnetWrapper,
|
|
|
|
false,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
this.isActive = isActive;
|
|
|
|
};
|
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
Future<dynamic> getCoinsToJoinSplit(
|
2022-08-26 08:11:35 +00:00
|
|
|
int required,
|
|
|
|
) async {
|
|
|
|
List<DartLelantusEntry> coins = await _getLelantusEntry();
|
|
|
|
if (required > LELANTUS_VALUE_SPEND_LIMIT_PER_TRANSACTION) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
int availableBalance = coins.fold(
|
2022-09-06 15:34:39 +00:00
|
|
|
0, (previousValue, element) => previousValue + element.amount);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
if (required > availableBalance) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// sort by biggest amount. if it is same amount we will prefer the older block
|
|
|
|
coins.sort((a, b) =>
|
|
|
|
(a.amount != b.amount ? a.amount > b.amount : a.height < b.height)
|
|
|
|
? -1
|
|
|
|
: 1);
|
2022-09-06 15:34:39 +00:00
|
|
|
int spendVal = 0;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
List<DartLelantusEntry> coinsToSpend = [];
|
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
while (spendVal < required) {
|
2022-08-26 08:11:35 +00:00
|
|
|
if (coins.isEmpty) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
DartLelantusEntry? chosen;
|
|
|
|
int need = required - spendVal;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
var itr = coins.first;
|
|
|
|
if (need >= itr.amount) {
|
2022-09-06 15:34:39 +00:00
|
|
|
chosen = itr;
|
2022-08-26 08:11:35 +00:00
|
|
|
coins.remove(itr);
|
|
|
|
} else {
|
|
|
|
for (int index = coins.length - 1; index != 0; index--) {
|
|
|
|
var coinIt = coins[index];
|
|
|
|
var nextItr = coins[index - 1];
|
|
|
|
|
|
|
|
if (coinIt.amount >= need &&
|
|
|
|
(index - 1 == 0 || nextItr.amount != coinIt.amount)) {
|
2022-09-06 15:34:39 +00:00
|
|
|
chosen = coinIt;
|
|
|
|
coins.remove(chosen);
|
2022-08-26 08:11:35 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-10-13 19:01:49 +00:00
|
|
|
// TODO: investigate the bug here where chosen is null, conditions, given one mint
|
2022-09-06 15:34:39 +00:00
|
|
|
spendVal += chosen!.amount;
|
|
|
|
coinsToSpend.insert(coinsToSpend.length, chosen);
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
// sort by group id ay ascending order. it is mandatory for creating proper joinsplit
|
2022-08-26 08:11:35 +00:00
|
|
|
coinsToSpend.sort((a, b) => a.anonymitySetId < b.anonymitySetId ? 1 : -1);
|
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
int changeToMint = spendVal - required;
|
2022-08-26 08:11:35 +00:00
|
|
|
List<int> indices = [];
|
|
|
|
for (var l in coinsToSpend) {
|
|
|
|
indices.add(l.index);
|
|
|
|
}
|
2022-09-06 15:34:39 +00:00
|
|
|
List<DartLelantusEntry> coinsToBeSpentOut = [];
|
|
|
|
coinsToBeSpentOut.addAll(coinsToSpend);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
return {"changeToMint": changeToMint, "coinsToSpend": coinsToBeSpentOut};
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
|
|
|
|
2022-09-06 15:34:39 +00:00
|
|
|
Future<int> estimateJoinSplitFee(
|
2022-08-26 08:11:35 +00:00
|
|
|
int spendAmount,
|
|
|
|
) async {
|
2022-10-13 19:01:49 +00:00
|
|
|
var lelantusEntry = await _getLelantusEntry();
|
2023-01-12 00:16:17 +00:00
|
|
|
final balance = availablePrivateBalance();
|
2022-11-23 18:31:31 +00:00
|
|
|
int spendAmount = (balance * Decimal.fromInt(Constants.satsPerCoin(coin)))
|
|
|
|
.toBigInt()
|
|
|
|
.toInt();
|
2022-10-13 19:01:49 +00:00
|
|
|
if (spendAmount == 0 || lelantusEntry.isEmpty) {
|
|
|
|
return LelantusFeeData(0, 0, []).fee;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
2022-10-13 19:01:49 +00:00
|
|
|
ReceivePort receivePort = await getIsolate({
|
|
|
|
"function": "estimateJoinSplit",
|
|
|
|
"spendAmount": spendAmount,
|
|
|
|
"subtractFeeFromAmount": true,
|
|
|
|
"lelantusEntries": lelantusEntry,
|
|
|
|
"coin": coin,
|
|
|
|
});
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-10-13 19:01:49 +00:00
|
|
|
final message = await receivePort.first;
|
|
|
|
if (message is String) {
|
|
|
|
Logging.instance.log("this is a string", level: LogLevel.Error);
|
|
|
|
stop(receivePort);
|
|
|
|
throw Exception("_fetchMaxFee isolate failed");
|
|
|
|
}
|
|
|
|
stop(receivePort);
|
|
|
|
Logging.instance.log('Closing estimateJoinSplit!', level: LogLevel.Info);
|
|
|
|
return (message as LelantusFeeData).fee;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|
2022-10-13 19:01:49 +00:00
|
|
|
// int fee;
|
|
|
|
// int size;
|
|
|
|
//
|
|
|
|
// for (fee = 0;;) {
|
|
|
|
// int currentRequired = spendAmount;
|
|
|
|
//
|
|
|
|
// TODO: investigate the bug here
|
|
|
|
// var map = await getCoinsToJoinSplit(currentRequired);
|
|
|
|
// if (map is bool && !map) {
|
|
|
|
// return 0;
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// List<DartLelantusEntry> coinsToBeSpent =
|
|
|
|
// map['coinsToSpend'] as List<DartLelantusEntry>;
|
|
|
|
//
|
|
|
|
// // 1054 is constant part, mainly Schnorr and Range proofs, 2560 is for each sigma/aux data
|
|
|
|
// // 179 other parts of tx, assuming 1 utxo and 1 jmint
|
|
|
|
// size = 1054 + 2560 * coinsToBeSpent.length + 180;
|
|
|
|
// // uint64_t feeNeeded = GetMinimumFee(size, DEFAULT_TX_CONFIRM_TARGET);
|
|
|
|
// int feeNeeded =
|
|
|
|
// size; //TODO(Levon) temporary, use real estimation methods here
|
|
|
|
//
|
|
|
|
// if (fee >= feeNeeded) {
|
|
|
|
// break;
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// fee = feeNeeded;
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// return fee;
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
@override
|
|
|
|
Future<int> estimateFeeFor(int satoshiAmount, int feeRate) async {
|
2022-09-06 15:34:39 +00:00
|
|
|
int fee = await estimateJoinSplitFee(satoshiAmount);
|
2022-08-26 08:11:35 +00:00
|
|
|
return fee;
|
|
|
|
}
|
|
|
|
|
2022-09-07 16:58:54 +00:00
|
|
|
Future<int> estimateFeeForPublic(int satoshiAmount, int feeRate) async {
|
2023-01-12 00:16:17 +00:00
|
|
|
final available = balance.spendable;
|
2022-09-07 16:58:54 +00:00
|
|
|
|
|
|
|
if (available == satoshiAmount) {
|
2023-01-12 00:16:17 +00:00
|
|
|
return satoshiAmount - (await sweepAllEstimate(feeRate));
|
2022-09-07 16:58:54 +00:00
|
|
|
} else if (satoshiAmount <= 0 || satoshiAmount > available) {
|
|
|
|
return roughFeeEstimate(1, 2, feeRate);
|
|
|
|
}
|
|
|
|
|
|
|
|
int runningBalance = 0;
|
|
|
|
int inputCount = 0;
|
2023-01-12 00:16:17 +00:00
|
|
|
for (final output in (await utxos)) {
|
|
|
|
if (!output.isBlocked) {
|
|
|
|
runningBalance += output.value;
|
|
|
|
inputCount++;
|
|
|
|
if (runningBalance > satoshiAmount) {
|
|
|
|
break;
|
|
|
|
}
|
2022-09-07 16:58:54 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
final oneOutPutFee = roughFeeEstimate(inputCount, 1, feeRate);
|
|
|
|
final twoOutPutFee = roughFeeEstimate(inputCount, 2, feeRate);
|
|
|
|
|
|
|
|
if (runningBalance - satoshiAmount > oneOutPutFee) {
|
|
|
|
if (runningBalance - satoshiAmount > oneOutPutFee + DUST_LIMIT) {
|
|
|
|
final change = runningBalance - satoshiAmount - twoOutPutFee;
|
|
|
|
if (change > DUST_LIMIT &&
|
|
|
|
runningBalance - satoshiAmount - change == twoOutPutFee) {
|
|
|
|
return runningBalance - satoshiAmount - change;
|
|
|
|
} else {
|
|
|
|
return runningBalance - satoshiAmount;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
return runningBalance - satoshiAmount;
|
|
|
|
}
|
|
|
|
} else if (runningBalance - satoshiAmount == oneOutPutFee) {
|
|
|
|
return oneOutPutFee;
|
|
|
|
} else {
|
|
|
|
return twoOutPutFee;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO: correct formula for firo?
|
|
|
|
int roughFeeEstimate(int inputCount, int outputCount, int feeRatePerKB) {
|
|
|
|
return ((181 * inputCount) + (34 * outputCount) + 10) *
|
|
|
|
(feeRatePerKB / 1000).ceil();
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<int> sweepAllEstimate(int feeRate) async {
|
2022-09-07 16:58:54 +00:00
|
|
|
int available = 0;
|
|
|
|
int inputCount = 0;
|
2023-01-12 00:16:17 +00:00
|
|
|
for (final output in (await utxos)) {
|
|
|
|
if (!output.isBlocked &&
|
|
|
|
output.isConfirmed(storedChainHeight, MINIMUM_CONFIRMATIONS)) {
|
2022-09-07 16:58:54 +00:00
|
|
|
available += output.value;
|
|
|
|
inputCount++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// transaction will only have 1 output minus the fee
|
|
|
|
final estimatedFee = roughFeeEstimate(inputCount, 1, feeRate);
|
|
|
|
|
|
|
|
return available - estimatedFee;
|
|
|
|
}
|
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
Future<List<Map<String, dynamic>>> fastFetch(List<String> allTxHashes) async {
|
|
|
|
List<Map<String, dynamic>> allTransactions = [];
|
|
|
|
|
|
|
|
const futureLimit = 30;
|
|
|
|
List<Future<Map<String, dynamic>>> transactionFutures = [];
|
|
|
|
int currentFutureCount = 0;
|
|
|
|
for (final txHash in allTxHashes) {
|
|
|
|
Future<Map<String, dynamic>> transactionFuture =
|
|
|
|
cachedElectrumXClient.getTransaction(
|
|
|
|
txHash: txHash,
|
|
|
|
verbose: true,
|
|
|
|
coin: coin,
|
|
|
|
);
|
|
|
|
transactionFutures.add(transactionFuture);
|
|
|
|
currentFutureCount++;
|
|
|
|
if (currentFutureCount > futureLimit) {
|
|
|
|
currentFutureCount = 0;
|
|
|
|
await Future.wait(transactionFutures);
|
|
|
|
for (final fTx in transactionFutures) {
|
|
|
|
final tx = await fTx;
|
|
|
|
// delete unused large parts
|
|
|
|
tx.remove("hex");
|
|
|
|
tx.remove("lelantusData");
|
|
|
|
|
|
|
|
allTransactions.add(tx);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (currentFutureCount != 0) {
|
|
|
|
currentFutureCount = 0;
|
|
|
|
await Future.wait(transactionFutures);
|
|
|
|
for (final fTx in transactionFutures) {
|
|
|
|
final tx = await fTx;
|
|
|
|
// delete unused large parts
|
|
|
|
tx.remove("hex");
|
|
|
|
tx.remove("lelantusData");
|
|
|
|
|
|
|
|
allTransactions.add(tx);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return allTransactions;
|
|
|
|
}
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Future<List<isar_models.Transaction>> getJMintTransactions(
|
2022-08-26 08:11:35 +00:00
|
|
|
CachedElectrumX cachedClient,
|
|
|
|
List<String> transactions,
|
2023-01-12 00:16:17 +00:00
|
|
|
// String currency,
|
2022-08-26 08:11:35 +00:00
|
|
|
Coin coin,
|
2023-01-12 00:16:17 +00:00
|
|
|
// Decimal currentPrice,
|
|
|
|
// String locale,
|
2022-08-26 08:11:35 +00:00
|
|
|
) async {
|
|
|
|
try {
|
2023-01-12 00:16:17 +00:00
|
|
|
List<isar_models.Transaction> txs = [];
|
2022-09-09 15:07:27 +00:00
|
|
|
List<Map<String, dynamic>> allTransactions =
|
|
|
|
await fastFetch(transactions);
|
2022-08-26 08:11:35 +00:00
|
|
|
|
2022-09-09 15:07:27 +00:00
|
|
|
for (int i = 0; i < allTransactions.length; i++) {
|
2022-08-26 08:11:35 +00:00
|
|
|
try {
|
2022-09-09 15:07:27 +00:00
|
|
|
final tx = allTransactions[i];
|
2022-08-26 08:11:35 +00:00
|
|
|
|
|
|
|
var sendIndex = 1;
|
|
|
|
if (tx["vout"][0]["value"] != null &&
|
2022-08-29 23:29:38 +00:00
|
|
|
Decimal.parse(tx["vout"][0]["value"].toString()) > Decimal.zero) {
|
2022-08-26 08:11:35 +00:00
|
|
|
sendIndex = 0;
|
|
|
|
}
|
|
|
|
tx["amount"] = tx["vout"][sendIndex]["value"];
|
|
|
|
tx["address"] = tx["vout"][sendIndex]["scriptPubKey"]["addresses"][0];
|
|
|
|
tx["fees"] = tx["vin"][0]["nFees"];
|
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
final txn = isar_models.Transaction()
|
|
|
|
..isLelantus = true
|
|
|
|
..txid = tx["txid"] as String
|
|
|
|
..timestamp = tx["time"] as int? ??
|
|
|
|
(DateTime.now().millisecondsSinceEpoch ~/ 1000)
|
|
|
|
..type = isar_models.TransactionType.outgoing
|
|
|
|
..subType = isar_models.TransactionSubType.join
|
|
|
|
..fee = Format.decimalAmountToSatoshis(
|
|
|
|
Decimal.parse(tx["fees"].toString()), coin)
|
2023-01-13 17:11:41 +00:00
|
|
|
..address.value = await isar.addresses
|
|
|
|
.filter()
|
|
|
|
.valueEqualTo(tx["address"] as String)
|
|
|
|
.findFirst()
|
2023-01-12 00:16:17 +00:00
|
|
|
..amount = Format.decimalAmountToSatoshis(
|
|
|
|
Decimal.parse(tx["amount"].toString()), coin)
|
|
|
|
..isCancelled = false
|
|
|
|
..height = tx["height"] as int?;
|
|
|
|
|
|
|
|
txs.add(txn);
|
2022-08-26 08:11:35 +00:00
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception caught in getJMintTransactions(): $e\n$s",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return txs;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown in getJMintTransactions(): $e\n$s",
|
|
|
|
level: LogLevel.Info);
|
|
|
|
rethrow;
|
|
|
|
}
|
|
|
|
}
|
2022-09-06 01:18:45 +00:00
|
|
|
|
|
|
|
@override
|
|
|
|
Future<bool> generateNewAddress() async {
|
|
|
|
try {
|
2023-01-12 00:16:17 +00:00
|
|
|
final currentReceiving = await _currentReceivingAddress;
|
|
|
|
|
|
|
|
final newReceivingIndex = currentReceiving.derivationIndex + 1;
|
|
|
|
|
|
|
|
// Use new index to derive a new receiving address
|
|
|
|
final newReceivingAddress = await _generateAddressForChain(
|
|
|
|
0,
|
|
|
|
newReceivingIndex,
|
|
|
|
);
|
|
|
|
|
|
|
|
// Add that new receiving address
|
|
|
|
await isar.writeTxn(() async {
|
|
|
|
await isar.addresses.put(newReceivingAddress);
|
|
|
|
});
|
2022-09-06 01:18:45 +00:00
|
|
|
|
|
|
|
return true;
|
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log(
|
|
|
|
"Exception rethrown from generateNewAddress(): $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
2022-09-06 21:52:51 +00:00
|
|
|
|
2023-01-12 00:16:17 +00:00
|
|
|
Decimal availablePrivateBalance() {
|
|
|
|
return balancePrivate.getSpendable();
|
|
|
|
}
|
|
|
|
|
|
|
|
Decimal availablePublicBalance() {
|
|
|
|
return balance.getSpendable();
|
|
|
|
}
|
|
|
|
|
|
|
|
Future<int> get chainHeight async {
|
|
|
|
try {
|
|
|
|
final result = await _electrumXClient.getBlockHeadTip();
|
2023-01-12 21:20:57 +00:00
|
|
|
final height = result["height"] as int;
|
|
|
|
await updateCachedChainHeight(height);
|
|
|
|
return height;
|
2023-01-12 00:16:17 +00:00
|
|
|
} catch (e, s) {
|
|
|
|
Logging.instance.log("Exception caught in chainHeight: $e\n$s",
|
|
|
|
level: LogLevel.Error);
|
2023-01-12 21:20:57 +00:00
|
|
|
return storedChainHeight;
|
2023-01-12 00:16:17 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
@override
|
2023-01-12 21:20:57 +00:00
|
|
|
int get storedChainHeight => getCachedChainHeight();
|
2023-01-10 23:50:22 +00:00
|
|
|
|
|
|
|
@override
|
2023-01-12 21:20:57 +00:00
|
|
|
Balance get balance => _balance ??= getCachedBalance();
|
2023-01-12 00:16:17 +00:00
|
|
|
Balance? _balance;
|
|
|
|
|
2023-01-12 21:20:57 +00:00
|
|
|
Balance get balancePrivate => _balancePrivate ??= getCachedBalanceSecondary();
|
2023-01-12 00:16:17 +00:00
|
|
|
Balance? _balancePrivate;
|
|
|
|
|
|
|
|
@override
|
|
|
|
Future<List<isar_models.UTXO>> get utxos => isar.utxos.where().findAll();
|
|
|
|
|
|
|
|
@override
|
|
|
|
Future<List<isar_models.Transaction>> get transactions =>
|
|
|
|
isar.transactions.where().findAll();
|
2023-01-12 03:23:21 +00:00
|
|
|
|
|
|
|
@override
|
|
|
|
Isar get isarInstance => isar;
|
2022-08-26 08:11:35 +00:00
|
|
|
}
|