mirror of
https://github.com/cypherstack/stack_wallet.git
synced 2024-12-26 21:39:21 +00:00
450 lines
16 KiB
Dart
450 lines
16 KiB
Dart
import 'package:hive/hive.dart';
|
|
import 'package:isar/isar.dart';
|
|
import 'package:stackwallet/db/hive/db.dart';
|
|
import 'package:stackwallet/db/isar/main_db.dart';
|
|
import 'package:stackwallet/electrumx_rpc/electrumx.dart';
|
|
import 'package:stackwallet/models/exchange/change_now/exchange_transaction.dart';
|
|
import 'package:stackwallet/models/exchange/response_objects/trade.dart';
|
|
import 'package:stackwallet/models/isar/models/blockchain_data/address.dart';
|
|
import 'package:stackwallet/models/isar/models/isar_models.dart' as isar_models;
|
|
import 'package:stackwallet/models/models.dart';
|
|
import 'package:stackwallet/models/node_model.dart';
|
|
import 'package:stackwallet/services/mixins/wallet_db.dart';
|
|
import 'package:stackwallet/services/node_service.dart';
|
|
import 'package:stackwallet/services/wallets_service.dart';
|
|
import 'package:stackwallet/utilities/amount/amount.dart';
|
|
import 'package:stackwallet/utilities/constants.dart';
|
|
import 'package:stackwallet/utilities/default_nodes.dart';
|
|
import 'package:stackwallet/utilities/enums/coin_enum.dart';
|
|
import 'package:stackwallet/utilities/flutter_secure_storage_interface.dart';
|
|
import 'package:stackwallet/utilities/logger.dart';
|
|
import 'package:stackwallet/utilities/prefs.dart';
|
|
import 'package:tuple/tuple.dart';
|
|
|
|
class DbVersionMigrator with WalletDB {
|
|
Future<void> migrate(
|
|
int fromVersion, {
|
|
required SecureStorageInterface secureStore,
|
|
}) async {
|
|
Logging.instance.log(
|
|
"Running migrate fromVersion $fromVersion",
|
|
level: LogLevel.Warning,
|
|
);
|
|
switch (fromVersion) {
|
|
case 0:
|
|
await Hive.openBox<dynamic>(DB.boxNameAllWalletsData);
|
|
await Hive.openBox<dynamic>(DB.boxNamePrefs);
|
|
final walletsService =
|
|
WalletsService(secureStorageInterface: secureStore);
|
|
final nodeService = NodeService(secureStorageInterface: secureStore);
|
|
final prefs = Prefs.instance;
|
|
final walletInfoList = await walletsService.walletNames;
|
|
await prefs.init();
|
|
|
|
ElectrumX? client;
|
|
int? latestSetId;
|
|
|
|
// only instantiate client if there are firo wallets
|
|
if (walletInfoList.values.any((element) => element.coin == Coin.firo)) {
|
|
await Hive.openBox<NodeModel>(DB.boxNameNodeModels);
|
|
await Hive.openBox<NodeModel>(DB.boxNamePrimaryNodes);
|
|
final node = nodeService.getPrimaryNodeFor(coin: Coin.firo) ??
|
|
DefaultNodes.firo;
|
|
List<ElectrumXNode> failovers = nodeService
|
|
.failoverNodesFor(coin: Coin.firo)
|
|
.map(
|
|
(e) => ElectrumXNode(
|
|
address: e.host,
|
|
port: e.port,
|
|
name: e.name,
|
|
id: e.id,
|
|
useSSL: e.useSSL,
|
|
),
|
|
)
|
|
.toList();
|
|
|
|
client = ElectrumX.from(
|
|
node: ElectrumXNode(
|
|
address: node.host,
|
|
port: node.port,
|
|
name: node.name,
|
|
id: node.id,
|
|
useSSL: node.useSSL),
|
|
prefs: prefs,
|
|
failovers: failovers,
|
|
);
|
|
|
|
try {
|
|
latestSetId = await client.getLatestCoinId();
|
|
} catch (e) {
|
|
// default to 2 for now
|
|
latestSetId = 2;
|
|
Logging.instance.log(
|
|
"Failed to fetch latest coin id during firo db migrate: $e \nUsing a default value of 2",
|
|
level: LogLevel.Warning);
|
|
}
|
|
}
|
|
|
|
for (final walletInfo in walletInfoList.values) {
|
|
// migrate each firo wallet's lelantus coins
|
|
if (walletInfo.coin == Coin.firo) {
|
|
await Hive.openBox<dynamic>(walletInfo.walletId);
|
|
final _lelantusCoins = DB.instance.get<dynamic>(
|
|
boxName: walletInfo.walletId, key: '_lelantus_coins') as List?;
|
|
final List<Map<dynamic, LelantusCoin>> lelantusCoins = [];
|
|
for (var lCoin in _lelantusCoins ?? []) {
|
|
lelantusCoins
|
|
.add({lCoin.keys.first: lCoin.values.first as LelantusCoin});
|
|
}
|
|
|
|
List<Map<dynamic, LelantusCoin>> coins = [];
|
|
for (final element in lelantusCoins) {
|
|
LelantusCoin coin = element.values.first;
|
|
int anonSetId = coin.anonymitySetId;
|
|
if (coin.anonymitySetId == 1 &&
|
|
(coin.publicCoin == '' ||
|
|
coin.publicCoin == "jmintData.publicCoin")) {
|
|
anonSetId = latestSetId!;
|
|
}
|
|
coins.add({
|
|
element.keys.first: LelantusCoin(coin.index, coin.value,
|
|
coin.publicCoin, coin.txId, anonSetId, coin.isUsed)
|
|
});
|
|
}
|
|
Logger.print("newcoins $coins", normalLength: false);
|
|
await DB.instance.put<dynamic>(
|
|
boxName: walletInfo.walletId,
|
|
key: '_lelantus_coins',
|
|
value: coins);
|
|
}
|
|
}
|
|
|
|
// update version
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo, key: "hive_data_version", value: 1);
|
|
|
|
// try to continue migrating
|
|
return await migrate(1, secureStore: secureStore);
|
|
|
|
case 1:
|
|
await Hive.openBox<ExchangeTransaction>(DB.boxNameTrades);
|
|
await Hive.openBox<Trade>(DB.boxNameTradesV2);
|
|
final trades =
|
|
DB.instance.values<ExchangeTransaction>(boxName: DB.boxNameTrades);
|
|
|
|
for (final old in trades) {
|
|
if (old.statusObject != null) {
|
|
final trade = Trade.fromExchangeTransaction(old, false);
|
|
await DB.instance.put<Trade>(
|
|
boxName: DB.boxNameTradesV2,
|
|
key: trade.uuid,
|
|
value: trade,
|
|
);
|
|
}
|
|
}
|
|
|
|
// update version
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo, key: "hive_data_version", value: 2);
|
|
|
|
// try to continue migrating
|
|
return await migrate(2, secureStore: secureStore);
|
|
|
|
case 2:
|
|
await Hive.openBox<dynamic>(DB.boxNamePrefs);
|
|
final prefs = Prefs.instance;
|
|
await prefs.init();
|
|
if (!(await prefs.isExternalCallsSet())) {
|
|
prefs.externalCalls = true;
|
|
}
|
|
|
|
// update version
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo, key: "hive_data_version", value: 3);
|
|
return await migrate(3, secureStore: secureStore);
|
|
|
|
case 3:
|
|
// clear possible broken firo cache
|
|
await DB.instance.deleteBoxFromDisk(
|
|
boxName: DB.instance.boxNameSetCache(coin: Coin.firo));
|
|
await DB.instance.deleteBoxFromDisk(
|
|
boxName: DB.instance.boxNameUsedSerialsCache(coin: Coin.firo));
|
|
|
|
// update version
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo, key: "hive_data_version", value: 4);
|
|
|
|
// try to continue migrating
|
|
return await migrate(4, secureStore: secureStore);
|
|
|
|
case 4:
|
|
// migrate
|
|
await _v4(secureStore);
|
|
|
|
// update version
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo, key: "hive_data_version", value: 5);
|
|
|
|
// try to continue migrating
|
|
return await migrate(5, secureStore: secureStore);
|
|
|
|
case 5:
|
|
// migrate
|
|
await Hive.openBox<dynamic>("theme");
|
|
await Hive.openBox<dynamic>(DB.boxNamePrefs);
|
|
|
|
final themeName =
|
|
DB.instance.get<dynamic>(boxName: "theme", key: "colorScheme")
|
|
as String? ??
|
|
"light";
|
|
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNamePrefs, key: "theme", value: themeName);
|
|
|
|
// update version
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo, key: "hive_data_version", value: 6);
|
|
|
|
// try to continue migrating
|
|
return await migrate(6, secureStore: secureStore);
|
|
|
|
case 6:
|
|
// migrate
|
|
await MainDB.instance.initMainDB();
|
|
final count = await MainDB.instance.isar.addresses.count();
|
|
// add change/receiving tags to address labels
|
|
for (var i = 0; i < count; i += 50) {
|
|
final addresses = await MainDB.instance.isar.addresses
|
|
.where()
|
|
.offset(i)
|
|
.limit(50)
|
|
.findAll();
|
|
|
|
final List<isar_models.AddressLabel> labels = [];
|
|
for (final address in addresses) {
|
|
List<String>? tags;
|
|
switch (address.subType) {
|
|
case AddressSubType.receiving:
|
|
tags = ["receiving"];
|
|
break;
|
|
case AddressSubType.change:
|
|
tags = ["change"];
|
|
break;
|
|
case AddressSubType.paynymNotification:
|
|
tags = ["paynym notification"];
|
|
break;
|
|
case AddressSubType.paynymSend:
|
|
break;
|
|
case AddressSubType.paynymReceive:
|
|
tags = ["paynym receiving"];
|
|
break;
|
|
case AddressSubType.unknown:
|
|
break;
|
|
case AddressSubType.nonWallet:
|
|
break;
|
|
}
|
|
|
|
// update/create label if tags is not empty
|
|
if (tags != null) {
|
|
isar_models.AddressLabel? label = await MainDB
|
|
.instance.isar.addressLabels
|
|
.where()
|
|
.addressStringWalletIdEqualTo(address.value, address.walletId)
|
|
.findFirst();
|
|
if (label == null) {
|
|
label = isar_models.AddressLabel(
|
|
walletId: address.walletId,
|
|
value: "",
|
|
addressString: address.value,
|
|
tags: tags,
|
|
);
|
|
} else if (label.tags == null) {
|
|
label = label.copyWith(tags: tags);
|
|
}
|
|
labels.add(label);
|
|
}
|
|
}
|
|
|
|
if (labels.isNotEmpty) {
|
|
await MainDB.instance.isar.writeTxn(() async {
|
|
await MainDB.instance.isar.addressLabels.putAll(labels);
|
|
});
|
|
}
|
|
}
|
|
|
|
// update version
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo, key: "hive_data_version", value: 7);
|
|
|
|
// try to continue migrating
|
|
return await migrate(7, secureStore: secureStore);
|
|
|
|
case 7:
|
|
// migrate
|
|
await _v7(secureStore);
|
|
|
|
// update version
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo, key: "hive_data_version", value: 8);
|
|
|
|
// try to continue migrating
|
|
return await migrate(8, secureStore: secureStore);
|
|
|
|
default:
|
|
// finally return
|
|
return;
|
|
}
|
|
}
|
|
|
|
Future<void> _v4(SecureStorageInterface secureStore) async {
|
|
await Hive.openBox<dynamic>(DB.boxNameAllWalletsData);
|
|
await Hive.openBox<dynamic>(DB.boxNamePrefs);
|
|
final walletsService = WalletsService(secureStorageInterface: secureStore);
|
|
final prefs = Prefs.instance;
|
|
final walletInfoList = await walletsService.walletNames;
|
|
await prefs.init();
|
|
await MainDB.instance.initMainDB();
|
|
|
|
for (final walletId in walletInfoList.keys) {
|
|
final info = walletInfoList[walletId]!;
|
|
assert(info.walletId == walletId);
|
|
|
|
final walletBox = await Hive.openBox<dynamic>(info.walletId);
|
|
|
|
const receiveAddressesPrefix = "receivingAddresses";
|
|
const changeAddressesPrefix = "changeAddresses";
|
|
|
|
// we need to manually migrate epic cash transactions as they are not
|
|
// stored on the epic cash blockchain
|
|
if (info.coin == Coin.epicCash) {
|
|
final txnData = walletBox.get("latest_tx_model") as TransactionData?;
|
|
|
|
// we ever only used index 0 in the past
|
|
const rcvIndex = 0;
|
|
|
|
final List<Tuple2<isar_models.Transaction, isar_models.Address?>>
|
|
transactionsData = [];
|
|
if (txnData != null) {
|
|
final txns = txnData.getAllTransactions();
|
|
|
|
for (final tx in txns.values) {
|
|
bool isIncoming = tx.txType == "Received";
|
|
|
|
final iTx = isar_models.Transaction(
|
|
walletId: walletId,
|
|
txid: tx.txid,
|
|
timestamp: tx.timestamp,
|
|
type: isIncoming
|
|
? isar_models.TransactionType.incoming
|
|
: isar_models.TransactionType.outgoing,
|
|
subType: isar_models.TransactionSubType.none,
|
|
amount: tx.amount,
|
|
amountString: Amount(
|
|
rawValue: BigInt.from(tx.amount),
|
|
fractionDigits: info.coin.decimals,
|
|
).toJsonString(),
|
|
fee: tx.fees,
|
|
height: tx.height,
|
|
isCancelled: tx.isCancelled,
|
|
isLelantus: false,
|
|
slateId: tx.slateId,
|
|
otherData: tx.otherData,
|
|
nonce: null,
|
|
inputs: [],
|
|
outputs: [],
|
|
);
|
|
|
|
if (tx.address.isEmpty) {
|
|
transactionsData.add(Tuple2(iTx, null));
|
|
} else {
|
|
final address = isar_models.Address(
|
|
walletId: walletId,
|
|
value: tx.address,
|
|
publicKey: [],
|
|
derivationIndex: isIncoming ? rcvIndex : -1,
|
|
derivationPath: null,
|
|
type: isIncoming
|
|
? isar_models.AddressType.mimbleWimble
|
|
: isar_models.AddressType.unknown,
|
|
subType: isIncoming
|
|
? isar_models.AddressSubType.receiving
|
|
: isar_models.AddressSubType.unknown,
|
|
);
|
|
transactionsData.add(Tuple2(iTx, address));
|
|
}
|
|
}
|
|
}
|
|
await MainDB.instance.addNewTransactionData(transactionsData, walletId);
|
|
}
|
|
|
|
// delete data from hive
|
|
await walletBox.delete(receiveAddressesPrefix);
|
|
await walletBox.delete("${receiveAddressesPrefix}P2PKH");
|
|
await walletBox.delete("${receiveAddressesPrefix}P2SH");
|
|
await walletBox.delete("${receiveAddressesPrefix}P2WPKH");
|
|
await walletBox.delete(changeAddressesPrefix);
|
|
await walletBox.delete("${changeAddressesPrefix}P2PKH");
|
|
await walletBox.delete("${changeAddressesPrefix}P2SH");
|
|
await walletBox.delete("${changeAddressesPrefix}P2WPKH");
|
|
await walletBox.delete("latest_tx_model");
|
|
await walletBox.delete("latest_lelantus_tx_model");
|
|
|
|
// set empty mnemonic passphrase as we used that by default before
|
|
if ((await secureStore.read(key: '${walletId}_mnemonicPassphrase')) ==
|
|
null) {
|
|
await secureStore.write(
|
|
key: '${walletId}_mnemonicPassphrase', value: "");
|
|
}
|
|
|
|
// doing this for epic cash will delete transaction history as it is not
|
|
// stored on the epic cash blockchain
|
|
if (info.coin != Coin.epicCash) {
|
|
// set flag to initiate full rescan on opening wallet
|
|
await DB.instance.put<dynamic>(
|
|
boxName: DB.boxNameDBInfo,
|
|
key: "rescan_on_open_$walletId",
|
|
value: Constants.rescanV1,
|
|
);
|
|
}
|
|
}
|
|
}
|
|
|
|
Future<void> _v7(SecureStorageInterface secureStore) async {
|
|
await Hive.openBox<dynamic>(DB.boxNameAllWalletsData);
|
|
final walletsService = WalletsService(secureStorageInterface: secureStore);
|
|
final walletInfoList = await walletsService.walletNames;
|
|
await MainDB.instance.initMainDB();
|
|
|
|
for (final walletId in walletInfoList.keys) {
|
|
final info = walletInfoList[walletId]!;
|
|
assert(info.walletId == walletId);
|
|
|
|
final count = await MainDB.instance.getTransactions(walletId).count();
|
|
|
|
for (var i = 0; i < count; i += 50) {
|
|
final txns = await MainDB.instance
|
|
.getTransactions(walletId)
|
|
.offset(i)
|
|
.limit(50)
|
|
.findAll();
|
|
|
|
// migrate amount to serialized amount string
|
|
final txnsData = txns
|
|
.map(
|
|
(tx) => Tuple2(
|
|
tx
|
|
..amountString = Amount(
|
|
rawValue: BigInt.from(tx.amount),
|
|
fractionDigits: info.coin.decimals,
|
|
).toJsonString(),
|
|
tx.address.value,
|
|
),
|
|
)
|
|
.toList();
|
|
|
|
// update db records
|
|
await MainDB.instance.addNewTransactionData(txnsData, walletId);
|
|
}
|
|
}
|
|
}
|
|
}
|