Discussion about Bitcoin. BitcoinSV restores the original Bitcoin protocol, will keep it stable, and allow it to massively scale on-chain. BSV will maintain the vision laid out by Satoshi Nakamoto in the 2008 white paper - Bitcoin: A Peer-to-Peer Electronic Cash System.
I was frustrated at how hard it is to setup for dev around (not on) Bitcoin Core so put this together. It includes Bitcoin Core via PPA and Rein's client and server under virtualenv. A highlight is Bitcoin Core being configured use testnet, 2gb pruned, serving RPC with a randomly generated password.
https://github.com/gridcoin-community/Gridcoin-Research/releases/tag/184.108.40.206 Finally! After over ten months of development and testing, "Fern" has arrived! This is a whopper. 240 pull requests merged. Essentially a complete rewrite that was started with the scraper (the "neural net" rewrite) in "Denise" has now been completed. Practically the ENTIRE Gridcoin specific codebase resting on top of the vanilla Bitcoin/Peercoin/Blackcoin vanilla PoS code has been rewritten. This removes the team requirement at last (see below), although there are many other important improvements besides that. Fern was a monumental undertaking. We had to encode all of the old rules active for the v10 block protocol in new code and ensure that the new code was 100% compatible. This had to be done in such a way as to clear out all of the old spaghetti and ring-fence it with tightly controlled class implementations. We then wrote an entirely new, simplified ruleset for research rewards and reengineered contracts (which includes beacon management, polls, and voting) using properly classed code. The fundamentals of Gridcoin with this release are now on a very sound and maintainable footing, and the developers believe the codebase as updated here will serve as the fundamental basis for Gridcoin's future roadmap. We have been testing this for MONTHS on testnet in various stages. The v10 (legacy) compatibility code has been running on testnet continuously as it was developed to ensure compatibility with existing nodes. During the last few months, we have done two private testnet forks and then the full public testnet testing for v11 code (the new protocol which is what Fern implements). The developers have also been running non-staking "sentinel" nodes on mainnet with this code to verify that the consensus rules are problem-free for the legacy compatibility code on the broader mainnet. We believe this amount of testing is going to result in a smooth rollout. Given the amount of changes in Fern, I am presenting TWO changelogs below. One is high level, which summarizes the most significant changes in the protocol. The second changelog is the detailed one in the usual format, and gives you an inkling of the size of this release.
Note that the protocol changes will not become active until we cross the hard-fork transition height to v11, which has been set at 2053000. Given current average block spacing, this should happen around October 4, about one month from now. Note that to get all of the beacons in the network on the new protocol, we are requiring ALL beacons to be validated. A two week (14 day) grace period is provided by the code, starting at the time of the transition height, for people currently holding a beacon to validate the beacon and prevent it from expiring. That means that EVERY CRUNCHER must advertise and validate their beacon AFTER the v11 transition (around Oct 4th) and BEFORE October 18th (or more precisely, 14 days from the actual date of the v11 transition). If you do not advertise and validate your beacon by this time, your beacon will expire and you will stop earning research rewards until you advertise and validate a new beacon. This process has been made much easier by a brand new beacon "wizard" that helps manage beacon advertisements and renewals. Once a beacon has been validated and is a v11 protocol beacon, the normal 180 day expiration rules apply. Note, however, that the 180 day expiration on research rewards has been removed with the Fern update. This means that while your beacon might expire after 180 days, your earned research rewards will be retained and can be claimed by advertising a beacon with the same CPID and going through the validation process again. In other words, you do not lose any earned research rewards if you do not stake a block within 180 days and keep your beacon up-to-date. The transition height is also when the team requirement will be relaxed for the network.
Besides the beacon wizard, there are a number of improvements to the GUI, including new UI transaction types (and icons) for staking the superblock, sidestake sends, beacon advertisement, voting, poll creation, and transactions with a message. The main screen has been revamped with a better summary section, and better status icons. Several changes under the hood have improved GUI performance. And finally, the diagnostics have been revamped.
The wallet sync speed has been DRASTICALLY improved. A decent machine with a good network connection should be able to sync the entire mainnet blockchain in less than 4 hours. A fast machine with a really fast network connection and a good SSD can do it in about 2.5 hours. One of our goals was to reduce or eliminate the reliance on snapshots for mainnet, and I think we have accomplished that goal with the new sync speed. We have also streamlined the in-memory structures for the blockchain which shaves some memory use. There are so many goodies here it is hard to summarize them all. I would like to thank all of the contributors to this release, but especially thank @cyrossignol, whose incredible contributions formed the backbone of this release. I would also like to pay special thanks to @barton2526, @caraka, and @Quezacoatl1, who tirelessly helped during the testing and polishing phase on testnet with testing and repeated builds for all architectures. The developers are proud to present this release to the community and we believe this represents the starting point for a true renaissance for Gridcoin!
Most significantly, nodes calculate research rewards directly from the magnitudes in EACH superblock between stakes instead of using a two- or three- point average based on a CPID's current magnitude and the magnitude for the CPID when it last staked. For those long-timers in the community, this has been referred to as "Superblock Windows," and was first done in proof-of-concept form by @denravonska.
Network magnitude unit pinned to a static value of 0.25
Max research reward allowed per block raised to 16384 GRC (from 12750 GRC)
New CPIDs begin accruing research rewards from the first superblock that contains the CPID instead of from the time of the beacon advertisement
500 GRC research reward limit for a CPID's first stake
6-month expiration for unclaimed rewards
10-block spacing requirement between research reward claims
Rolling 5-day payment-per-day limit
Legacy tolerances for floating-point error and time drift
The need to include a valid copy of a CPID's magnitude in a claim
10-block emission adjustment interval for the magnitude unit
One-time beacon activation requires that participants temporarily change their usernames to a verification code at one whitelisted BOINC project
Verification codes of pending beacons expire after 3 days
Self-service beacon removal
Burn fee for beacon advertisement increased from 0.00001 GRC to 0.5 GRC
Rain addresses derived from beacon keys instead of a default wallet address
Beacon expiration determined as of the current block instead of the previous block
The ability for developers to remove beacons
The ability to sign research reward claims with non-current but unexpired beacons
As a reminder:
Beacons expire after 6 months pass (180 days)
Beacons can be renewed after 5 months pass (150 days)
Renewed beacons must be signed with the same key as the original beacon
Magnitudes less than 1 include two fractional places
Magnitudes greater than or equal to 1 but less than 10 include one fractional place
A valid superblock must match a scraper convergence
Superblock popularity election mechanics
Yes/no/abstain and single-choice response types (no user-facing support yet)
To create a poll, a maximum of 250 UTXOs for a single address must add up to 100000 GRC. These are selected from the largest downwards.
Burn fee for creating polls scaled by the number of UTXOs claimed
50 GRC for a poll contract
0.001 GRC per claimed UTXO
Burn fee for casting votes scaled by the number of UTXOs claimed
0.01 GRC for a vote contract
0.01 GRC to claim magnitude
0.01 GRC per claimed address
0.001 GRC per claimed UTXO
Maximum length of a poll title: 80 characters
Maximum length of a poll question: 100 characters
Maximum length of a poll discussion website URL: 100 characters
Maximum number of poll choices: 20
Maximum length of a poll choice label: 100 characters
Magnitude, CPID count, and participant count poll weight types
The ability for developers to remove polls and votes
[220.127.116.11] 2020-09-03, mandatory, "Fern"
Backport newer uint256 types from Bitcoin #1570 (@cyrossignol)
Implement project level rain for rainbymagnitude #1580 (@jamescowens)
Upgrade utilities (Update checker and snapshot downloadeapplication) #1576 (@iFoggz)
Provide fees collected in the block by the miner #1601 (@iFoggz)
Add support for generating legacy superblocks from scraper stats #1603 (@cyrossignol)
Port of the Bitcoin Logger to Gridcoin #1600 (@jamescowens)
Implement zapwallettxes #1605 (@jamescowens)
Implements a global event filter to suppress help question mark #1609 (@jamescowens)
Add next target difficulty to RPC output #1615 (@cyrossignol)
Add caching for block hashes to CBlock #1624 (@cyrossignol)
Make toolbars and tray icon red for testnet #1637 (@jamescowens)
Add an rpc call convergencereport #1643 (@jamescowens)
Implement newline filter on config file read in #1645 (@jamescowens)
Implement beacon status icon/button #1646 (@jamescowens)
Add gridcointestnet.png #1649 (@caraka)
Add precision to support magnitudes less than 1 #1651 (@cyrossignol)
Replace research accrual calculations with superblock snapshots #1657 (@cyrossignol)
Publish example gridcoinresearch.conf as a md document to the doc directory #1662 (@jamescowens)
Add options checkbox to disable transaction notifications #1666 (@jamescowens)
Add support for self-service beacon deletion #1695 (@cyrossignol)
Add support for type-specific contract fee amounts #1698 (@cyrossignol)
Add verifiedbeaconreport and pendingbeaconreport #1696 (@jamescowens)
Add preliminary testing option for block v11 height on testnet #1706 (@cyrossignol)
Add verified beacons manifest part to superblock validator #1711 (@cyrossignol)
Implement beacon, vote, and superblock display categories/icons in UI transaction model #1717 (@jamescowens)
Hi fellas, this is my first post on the darknet. I read today about the Empire Market and decided to post here about a tech that might help to end this charade of constant exit scams. My post is about a DLT-based open source decentralized marketplace that will hopefully initiate a paradigm shift in the free and private trade between individuals and ends the exit-scams and leaky databases problem once and for all. It is called the Open Marketplace V3.0 and is set to be released after 3 years of hard development work by Particl Project, including one year of public beta testing. My personal belief is that they are releasing to the world a game-changing/breaking product. At minimum, due to the open source nature of the project, this is a proof of concept that is bound to shift the global eCommerce paradigm into DeCommerce. I hope my own description below of the marketplace and its technology will fuel your interest to read more about it. My even bigger hope is that some of you will decide to become first-hand beta testers as soon as the Open Marketplace v3.0 hits the testnet (eta: a few weeks). The marketplace: The Open Marketplace is designed to be private by default and decentralized with no middlemen/intermediaries whatsoever. The trades are protected by two-way automated escrow via smart-contracts that de-incentivize and penalize dishonest behavior on both sides. The Open Marketplace takes no sales commissions and charges only a tiny listing fee (<0.01$) to prevent listing spamming. All the marketplace generated fees go to the nodes that provide the hardware infrastructure for the p2p network to operate. The nodes can be public or you can run them as Tor hidden services. The V3.0 that is set to be released will be the first version open to the wider public. It will allow anyone to create easily a decentralized personal storefront or a community market or simply buy and sell goods on the already existing markets. The user-created markets/storefronts can be visible on the marketplace if the access key is publicly announced or absolutely invisible to anyone that doesn't know the access key (held by the creator). This is an intentional privacy feature and simply put without going into many technical details, if you dont have market access key there is absolutely no way to see/detect that some market exists. The technology: The Open Marketplace is crypto-agnostic and currently supports payments in BTC, PART, ZCoin (DAI, NIX are next in the pipeline and many more to come). It uses as a settlement layer its native blockchain, which is an up-to-date Bitcoin codebase with added privacy features like CT, RingCT (up to 32 mixins), Stealth addresses, etc. These privacy features are used in combination to keep the financial data like escrows, transactions, etc private and most importantly un-linkable to the actual market buys/sells. For the actual user or markets related data exchange like posted listings, buy/sell flows, encrypted user communication, built-in cryptocurrency exchange, etc, the Open Marketplace uses a custom Bitmessage variant called SMSG, which allows metadata stripped encrypted data exchange. Last but not least the Open Marketplace client has a built-in option for using Tor network via proxy. The important people: The cypherpunks behind Particl's Open Marketplace have been passionate pioneers and OGs in the privacy DLT field. For example, they were the first ever to implement features like RingCT, Bulletproofs, PoS, cold staking, etc on a Bitcoin codebase. Their privacy features implementations have been audited successfully by several respectable academics and security R&D providers, like QuarksLab. The team has been so far focused on building without any marketing and thus have remained intentionally in the shadows but the latter is planned to change with the v3.0 release. One of the steps towards that will be the initiation of the long-planned Vendor On-boarding and Outreach Program and the release of the Particl Academy, an easy to understand and learn about the project portal. Me: I am a passionate freedom and privacy advocate that discovered the project 1.5 year ago and since then has become a member of their small but like-minded community ([email protected]/discord).
Scaling Reddit Community Points with Arbitrum Rollup: a piece of cake
https://preview.redd.it/b80c05tnb9e51.jpg?width=2550&format=pjpg&auto=webp&s=850282c1a3962466ed44f73886dae1c8872d0f31 Submitted for consideration toThe Great Reddit Scaling Bake-Off Baked by the pastry chefs atOffchain Labs Please send questions or comments to [[email protected] ](mailto:[email protected]) 1. Overview We're excited to submit Arbitrum Rollup for consideration to The Great Reddit Scaling Bake-Off. Arbitrum Rollup is the only Ethereum scaling solution that supports arbitrary smart contracts without compromising on Ethereum's security or adding points of centralization. For Reddit, this means that Arbitrum can not only scale the minting and transfer of Community Points, but it can foster a creative ecosystem built around Reddit Community Points enabling points to be used in a wide variety of third party applications. That's right -- you can have your cake and eat it too! Arbitrum Rollup isn't just Ethereum-style. Its Layer 2 transactions are byte-for-byte identical to Ethereum, which means Ethereum users can continue to use their existing addresses and wallets, and Ethereum developers can continue to use their favorite toolchains and development environments out-of-the-box with Arbitrum. Coupling Arbitrum’s tooling-compatibility with its trustless asset interoperability, Reddit not only can scale but can onboard the entire Ethereum community at no cost by giving them the same experience they already know and love (well, certainly know). To benchmark how Arbitrum can scale Reddit Community Points, we launched the Reddit contracts on an Arbitrum Rollup chain. Since Arbitrum provides full Solidity support, we didn't have to rewrite the Reddit contracts or try to mimic their functionality using an unfamiliar paradigm. Nope, none of that. We launched the Reddit contracts unmodified on Arbitrum Rollup complete with support for minting and distributing points. Like every Arbitrum Rollup chain, the chain included a bridge interface in which users can transfer Community Points or any other asset between the L1 and L2 chains. Arbitrum Rollup chains also support dynamic contract loading, which would allow third-party developers to launch custom ecosystem apps that integrate with Community Points on the very same chain that runs the Reddit contracts. 1.1 Why Ethereum Perhaps the most exciting benefit of distributing Community Points using a blockchain is the ability to seamlessly port points to other applications and use them in a wide variety of contexts. Applications may include simple transfers such as a restaurant that allows Redditors to spend points on drinks. Or it may include complex smart contracts -- such as placing Community Points as a wager for a multiparty game or as collateral in a financial contract. The common denominator between all of the fun uses of Reddit points is that it needs a thriving ecosystem of both users and developers, and the Ethereum blockchain is perhaps the only smart contract platform with significant adoption today. While many Layer 1 blockchains boast lower cost or higher throughput than the Ethereum blockchain, more often than not, these attributes mask the reality of little usage, weaker security, or both. Perhaps another platform with significant usage will rise in the future. But today, Ethereum captures the mindshare of the blockchain community, and for Community Points to provide the most utility, the Ethereum blockchain is the natural choice. 1.2 Why Arbitrum While Ethereum's ecosystem is unmatched, the reality is that fees are high and capacity is too low to support the scale of Reddit Community Points. Enter Arbitrum. Arbitrum Rollup provides all of the ecosystem benefits of Ethereum, but with orders of magnitude more capacity and at a fraction of the cost of native Ethereum smart contracts. And most of all, we don't change the experience from users. They continue to use the same wallets, addresses, languages, and tools. Arbitrum Rollup is not the only solution that can scale payments, but it is the only developed solution that can scale both payments and arbitrary smart contracts trustlessly, which means that third party users can build highly scalable add-on apps that can be used without withdrawing money from the Rollup chain. If you believe that Reddit users will want to use their Community Points in smart contracts--and we believe they will--then it makes the most sense to choose a single scaling solution that can support the entire ecosystem, eliminating friction for users. We view being able to run smart contracts in the same scaling solution as fundamentally critical since if there's significant demand in running smart contracts from Reddit's ecosystem, this would be a load on Ethereum and would itself require a scaling solution. Moreover, having different scaling solutions for the minting/distribution/spending of points and for third party apps would be burdensome for users as they'd have to constantly shuffle their Points back and forth. 2. Arbitrum at a glance Arbitrum Rollup has a unique value proposition as it offers a combination of features that no other scaling solution achieves. Here we highlight its core attributes. Decentralized. Arbitrum Rollup is as decentralized as Ethereum. Unlike some other Layer 2 scaling projects, Arbitrum Rollup doesn't have any centralized components or centralized operators who can censor users or delay transactions. Even in non-custodial systems, centralized components provide a risk as the operators are generally incentivized to increase their profit by extracting rent from users often in ways that severely degrade user experience. Even if centralized operators are altruistic, centralized components are subject to hacking, coercion, and potential liability. Massive Scaling. Arbitrum achieves order of magnitude scaling over Ethereum's L1 smart contracts. Our software currently supports 453 transactions-per-second for basic transactions (at 1616 Ethereum gas per tx). We have a lot of room left to optimize (e.g. aggregating signatures), and over the next several months capacity will increase significantly. As described in detail below, Arbitrum can easily support and surpass Reddit's anticipated initial load, and its capacity will continue to improve as Reddit's capacity needs grow. Low cost. The cost of running Arbitrum Rollup is quite low compared to L1 Ethereum and other scaling solutions such as those based on zero-knowledge proofs. Layer 2 fees are low, fixed, and predictable and should not be overly burdensome for Reddit to cover. Nobody needs to use special equipment or high-end machines. Arbitrum requires validators, which is a permissionless role that can be run on any reasonable on-line machine. Although anybody can act as a validator, in order to protect against a “tragedy of the commons” and make sure reputable validators are participating, we support a notion of “invited validators” that are compensated for their costs. In general, users pay (low) fees to cover the invited validators’ costs, but we imagine that Reddit may cover this cost for its users. See more on the costs and validator options below. Ethereum Developer Experience. Not only does Arbitrum support EVM smart contracts, but the developer experience is identical to that of L1 Ethereum contracts and fully compatible with Ethereum tooling. Developers can port existing Solidity apps or write new ones using their favorite and familiar toolchains (e.g. Truffle, Buidler). There are no new languages or coding paradigms to learn. Ethereum wallet compatibility. Just as in Ethereum, Arbitrum users need only hold keys, but do not have to store any coin history or additional data to protect or access their funds. Since Arbitrum transactions are semantically identical to Ethereum L1 transactions, existing Ethereum users can use their existing Ethereum keys with their existing wallet software such as Metamask. Token interoperability. Users can easily transfer their ETH, ERC-20 and ERC-721 tokens between Ethereum and the Arbitrum Rollup chain. As we explain in detail below, it is possible to mint tokens in L2 that can subsequently be withdrawn and recognized by the L1 token contract. Fast finality. Transactions complete with the same finality time as Ethereum L1 (and it's possible to get faster finality guarantees by trading away trust assumptions; see the Arbitrum Rollup whitepaper for details). Non-custodial. Arbitrum Rollup is a non-custodial scaling solution, so users control their funds/points and neither Reddit nor anyone else can ever access or revoke points held by users. Censorship Resistant. Since it's completely decentralized, and the Arbitrum protocol guarantees progress trustlessly, Arbitrum Rollup is just as censorship-proof as Ethereum. Block explorer. The Arbitrum Rollup block explorer allows users to view and analyze transactions on the Rollup chain. Limitations Although this is a bake-off, we're not going to sugar coat anything. Arbitrum Rollup, like any Optimistic Rollup protocol, does have one limitation, and that's the delay on withdrawals. As for the concrete length of the delay, we've done a good deal of internal modeling and have blogged about this as well. Our current modeling suggests a 3-hour delay is sufficient (but as discussed in the linked post there is a tradeoff space between the length of the challenge period and the size of the validators’ deposit). Note that this doesn't mean that the chain is delayed for three hours. Arbitrum Rollup supports pipelining of execution, which means that validators can keep building new states even while previous ones are “in the pipeline” for confirmation. As the challenge delays expire for each update, a new state will be confirmed (read more about this here). So activity and progress on the chain are not delayed by the challenge period. The only thing that's delayed is the consummation of withdrawals. Recall though that any single honest validator knows immediately (at the speed of L1 finality) which state updates are correct and can guarantee that they will eventually be confirmed, so once a valid withdrawal has been requested on-chain, every honest party knows that the withdrawal will definitely happen. There's a natural place here for a liquidity market in which a validator (or someone who trusts a validator) can provide withdrawal loans for a small interest fee. This is a no-risk business for them as they know which withdrawals will be confirmed (and can force their confirmation trustlessly no matter what anyone else does) but are just waiting for on-chain finality. 3. The recipe: How Arbitrum Rollup works For a description of the technical components of Arbitrum Rollup and how they interact to create a highly scalable protocol with a developer experience that is identical to Ethereum, please refer to the following documents: Arbitrum Rollup Whitepaper Arbitrum academic paper (describes a previous version of Arbitrum) 4. Developer docs and APIs For full details about how to set up and interact with an Arbitrum Rollup chain or validator, please refer to our developer docs, which can be found at https://developer.offchainlabs.com/. Note that the Arbitrum version described on that site is older and will soon be replaced by the version we are entering in Reddit Bake-Off, which is still undergoing internal testing before public release. 5. Who are the validators? As with any Layer 2 protocol, advancing the protocol correctly requires at least one validator (sometimes called block producers) that is honest and available. A natural question is: who are the validators? Recall that the validator set for an Arbitrum chain is open and permissionless; anyone can start or stop validating at will. (A useful analogy is to full nodes on an L1 chain.) But we understand that even though anyone can participate, Reddit may want to guarantee that highly reputable nodes are validating their chain. Reddit may choose to validate the chain themselves and/or hire third-party validators.To this end, we have begun building a marketplace for validator-for-hire services so that dapp developers can outsource validation services to reputable nodes with high up-time. We've announced a partnership in which Chainlink nodes will provide Arbitrum validation services, and we expect to announce more partnerships shortly with other blockchain infrastructure providers. Although there is no requirement that validators are paid, Arbitrum’s economic model tracks validators’ costs (e.g. amount of computation and storage) and can charge small fees on user transactions, using a gas-type system, to cover those costs. Alternatively, a single party such as Reddit can agree to cover the costs of invited validators. 6. Reddit Contract Support Since Arbitrum contracts and transactions are byte-for-byte compatible with Ethereum, supporting the Reddit contracts is as simple as launching them on an Arbitrum chain. Minting. Arbitrum Rollup supports hybrid L1/L2 tokens which can be minted in L2 and then withdrawn onto the L1. An L1 contract at address A can make a special call to the EthBridge which deploys a "buddy contract" to the same address A on an Arbitrum chain. Since it's deployed at the same address, users can know that the L2 contract is the authorized "buddy" of the L1 contract on the Arbitrum chain. For minting, the L1 contract is a standard ERC-20 contract which mints and burns tokens when requested by the L2 contract. It is paired with an ERC-20 contract in L2 which mints tokens based on whatever programmer provided minting facility is desired and burns tokens when they are withdrawn from the rollup chain. Given this base infrastructure, Arbitrum can support any smart contract based method for minting tokens in L2, and indeed we directly support Reddit's signature/claim based minting in L2. Batch minting. What's better than a mint cookie? A whole batch! In addition to supporting Reddit’s current minting/claiming scheme, we built a second minting design, which we believe outperforms the signature/claim system in many scenarios. In the current system, Reddit periodically issues signed statements to users, who then take those statements to the blockchain to claim their tokens. An alternative approach would have Reddit directly submit the list of users/amounts to the blockchain and distribute the tokens to the users without the signature/claim process. To optimize the cost efficiency of this approach, we designed an application-specific compression scheme to minimize the size of the batch distribution list. We analyzed the data from Reddit's previous distributions and found that the data is highly compressible since token amounts are small and repeated, and addresses appear multiple times. Our function groups transactions by size, and replaces previously-seen addresses with a shorter index value. We wrote client code to compress the data, wrote a Solidity decompressing function, and integrated that function into Reddit’s contract running on Arbitrum. When we ran the compression function on the previous Reddit distribution data, we found that we could compress batched minting data down to to 11.8 bytes per minting event (averaged over a 6-month trace of Reddit’s historical token grants)compared with roughly 174 bytes of on-chain data needed for the signature claim approach to minting (roughly 43 for an RLP-encoded null transaction + 65 for Reddit's signature + 65 for the user's signature + roughly 8 for the number of Points) . The relative benefit of the two approaches with respect to on-chain call data cost depends on the percentage of users that will actually claim their tokens on chain. With the above figures, batch minting will be cheaper if roughly 5% of users redeem their claims. We stress that our compression scheme is not Arbitrum-specific and would be beneficial in any general-purpose smart contract platform. 8. Benchmarks and costs In this section, we give the full costs of operating the Reddit contracts on an Arbitrum Rollup chain including the L1 gas costs for the Rollup chain, the costs of computation and storage for the L2 validators as well as the capital lockup requirements for staking. Arbitrum Rollup is still on testnet, so we did not run mainnet benchmarks. Instead, we measured the L1 gas cost and L2 workload for Reddit operations on Arbitrum and calculated the total cost assuming current Ethereum gas prices. As noted below in detail, our measurements do not assume that Arbitrum is consuming the entire capacity of Ethereum. We will present the details of our model now, but for full transparency you can also play around with it yourself and adjust the parameters, by copying the spreadsheet found here. Our cost model is based on measurements of Reddit’s contracts, running unmodified (except for the addition of a batch minting function) on Arbitrum Rollup on top of Ethereum. On the distribution of transactions and frequency of assertions. Reddit's instructions specify the following minimum parameters that submissions should support: Over a 5 day period, your scaling PoC should be able to handle:
100,000 point claims (minting & distributing points)
75,000 one-off points burning
We provide the full costs of operating an Arbitrum Rollup chain with this usage under the assumption that tokens are minted or granted to users in batches, but other transactions are uniformly distributed over the 5 day period. Unlike some other submissions, we do not make unrealistic assumptions that all operations can be submitted in enormous batches. We assume that batch minting is done in batches that use only a few percent on an L1 block’s gas, and that other operations come in evenly over time and are submitted in batches, with one batch every five minutes to keep latency reasonable. (Users are probably already waiting for L1 finality, which takes at least that long to achieve.) We note that assuming that there are only 300,000 transactions that arrive uniformly over the 5 day period will make our benchmark numbers lower, but we believe that this will reflect the true cost of running the system. To see why, say that batches are submitted every five minutes (20 L1 blocks) and there's a fixed overhead of c bytes of calldata per batch, the cost of which will get amortized over all transactions executed in that batch. Assume that each individual transaction adds a marginal cost of t. Lastly assume the capacity of the scaling system is high enough that it can support all of Reddit's 300,000 transactions within a single 20-block batch (i.e. that there is more than c + 300,000*t byes of calldata available in 20 blocks). Consider what happens if c, the per-batch overhead, is large (which it is in some systems, but not in Arbitrum). In the scenario that transactions actually arrive at the system's capacity and each batch is full, then c gets amortized over 300,000 transactions. But if we assume that the system is not running at capacity--and only receives 300,000 transactions arriving uniformly over 5 days-- then each 20-block assertion will contain about 200 transactions, and thus each transaction will pay a nontrivial cost due to c. We are aware that other proposals presented scaling numbers assuming that 300,000 transactions arrived at maximum capacity and was executed in a single mega-transaction, but according to our estimates, for at least one such report, this led to a reported gas price that was 2-3 orders of magnitude lower than it would have been assuming uniform arrival. We make more realistic batching assumptions, and we believe Arbitrum compares well when batch sizes are realistic. Our model. Our cost model includes several sources of cost:
L1 gas costs: This is the cost of posting transactions as calldata on the L1 chain, as well as the overhead associated with each batch of transactions, and the L1 cost of settling transactions in the Arbitrum protocol.
Validator’s staking costs: In normal operation, one validator will need to be staked. The stake is assumed to be 0.2% of the total value of the chain (which is assumed to be $1 per user who is eligible to claim points). The cost of staking is the interest that could be earned on the money if it were not staked.
Validator computation and storage: Every validator must do computation to track the chain’s processing of transactions, and must maintain storage to keep track of the contracts’ EVM storage. The cost of computation and storage are estimated based on measurements, with the dollar cost of resources based on Amazon Web Services pricing.
It’s clear from our modeling that the predominant cost is for L1 calldata. This will probably be true for any plausible rollup-based system. Our model also shows that Arbitrum can scale to workloads much larger than Reddit’s nominal workload, without exhausting L1 or L2 resources. The scaling bottleneck will ultimately be calldata on the L1 chain. We believe that cost could be reduced substantially if necessary by clever encoding of data. (In our design any compression / decompression of L2 transaction calldata would be done by client software and L2 programs, never by an L1 contract.) 9. Status of Arbitrum Rollup Arbitrum Rollup is live on Ethereum testnet. All of the code written to date including everything included in the Reddit demo is open source and permissively licensed under the Apache V2 license. The first testnet version of Arbitrum Rollup was released on testnet in February. Our current internal version, which we used to benchmark the Reddit contracts, will be released soon and will be a major upgrade. Both the Arbitrum design as well as the implementation are heavily audited by independent third parties. The Arbitrum academic paper was published at USENIX Security, a top-tier peer-reviewed academic venue. For the Arbitrum software, we have engaged Trail of Bits for a security audit, which is currently ongoing, and we are committed to have a clean report before launching on Ethereum mainnet. 10. Reddit Universe Arbitrum Rollup Chain The benchmarks described in this document were all measured using the latest internal build of our software. When we release the new software upgrade publicly we will launch a Reddit Universe Arbitrum Rollup chain as a public demo, which will contain the Reddit contracts as well as a Uniswap instance and a Connext Hub, demonstrating how Community Points can be integrated into third party apps. We will also allow members of the public to dynamically launch ecosystem contracts. We at Offchain Labs will cover the validating costs for the Reddit Universe public demo. If the folks at Reddit would like to evaluate our software prior to our public demo, please email us at [email protected] and we'd be more than happy to provide early access. 11. Even more scaling: Arbitrum Sidechains Rollups are an excellent approach to scaling, and we are excited about Arbitrum Rollup which far surpasses Reddit's scaling needs. But looking forward to Reddit's eventual goal of supporting hundreds of millions of users, there will likely come a time when Reddit needs more scaling than any Rollup protocol can provide. While Rollups greatly reduce costs, they don't break the linear barrier. That is, all transactions have an on-chain footprint (because all calldata must be posted on-chain), albeit a far smaller one than on native Ethereum, and the L1 limitations end up being the bottleneck for capacity and cost. Since Ethereum has limited capacity, this linear use of on-chain resources means that costs will eventually increase superlinearly with traffic. The good news is that we at Offchain Labs have a solution in our roadmap that can satisfy this extreme-scaling setting as well: Arbitrum AnyTrust Sidechains. Arbitrum Sidechains are similar to Arbitrum Rollup, but deviate in that they name a permissioned set of validators. When a chain’s validators agree off-chain, they can greatly reduce the on-chain footprint of the protocol and require almost no data to be put on-chain. When validators can't reach unanimous agreement off-chain, the protocol reverts to Arbitrum Rollup. Technically, Arbitrum Sidechains can be viewed as a hybrid between state channels and Rollup, switching back and forth as necessary, and combining the performance and cost that state channels can achieve in the optimistic case, with the robustness of Rollup in other cases. The core technical challenge is how to switch seamlessly between modes and how to guarantee that security is maintained throughout. Arbitrum Sidechains break through this linear barrier, while still maintaining a high level of security and decentralization. Arbitrum Sidechains provide the AnyTrust guarantee, which says that as long as any one validator is honest and available (even if you don't know which one will be), the L2 chain is guaranteed to execute correctly according to its code and guaranteed to make progress. Unlike in a state channel, offchain progress does not require unanimous consent, and liveness is preserved as long as there is a single honest validator. Note that the trust model for Arbitrum Sidechains is much stronger than for typical BFT-style chains which introduce a consensus "voting" protocols among a small permissioned group of validators. BFT-based protocols require a supermajority (more than 2/3) of validators to agree. In Arbitrum Sidechains, by contrast, all you need is a single honest validator to achieve guaranteed correctness and progress. Notice that in Arbitrum adding validators strictly increases security since the AnyTrust guarantee provides correctness as long as any one validator is honest and available. By contrast, in BFT-style protocols, adding nodes can be dangerous as a coalition of dishonest nodes can break the protocol. Like Arbitrum Rollup, the developer and user experiences for Arbitrum Sidechains will be identical to that of Ethereum. Reddit would be able to choose a large and diverse set of validators, and all that they would need to guarantee to break through the scaling barrier is that a single one of them will remain honest. We hope to have Arbitrum Sidechains in production in early 2021, and thus when Reddit reaches the scale that surpasses the capacity of Rollups, Arbitrum Sidechains will be waiting and ready to help. While the idea to switch between channels and Rollup to get the best of both worlds is conceptually simple, getting the details right and making sure that the switch does not introduce any attack vectors is highly non-trivial and has been the subject of years of our research (indeed, we were working on this design for years before the term Rollup was even coined). 12. How Arbitrum compares We include a comparison to several other categories as well as specific projects when appropriate. and explain why we believe that Arbitrum is best suited for Reddit's purposes. We focus our attention on other Ethereum projects. Payment only Rollups. Compared to Arbitrum Rollup, ZK-Rollups and other Rollups that only support token transfers have several disadvantages:
As outlined throughout the proposal, we believe that the entire draw of Ethereum is in its rich smart contracts support which is simply not achievable with today's zero-knowledge proof technology. Indeed, scaling with a ZK-Rollup will add friction to the deployment of smart contracts that interact with Community Points as users will have to withdraw their coins from the ZK-Rollup and transfer them to a smart contract system (like Arbitrum). The community will be best served if Reddit builds on a platform that has built-in, frictionless smart-contract support.
All other Rollup protocols of which we are aware employ a centralized operator. While it's true that users retain custody of their coins, the centralized operator can often profit from censoring, reordering, or delaying transactions. A common misconception is that since they're non-custodial protocols, a centralized sequencer does not pose a risk but this is incorrect as the sequencer can wreak havoc or shake down users for side payments without directly stealing funds.
Sidechain type protocols can eliminate some of these issues, but they are not trustless. Instead, they require trust in some quorum of a committee, often requiring two-third of the committee to be honest, compared to rollup protocols like Arbitrum that require only a single honest party. In addition, not all sidechain type protocols have committees that are diverse, or even non-centralized, in practice.
Plasma-style protocols have a centralized operator and do not support general smart contracts.
13. Concluding Remarks While it's ultimately up to the judges’ palate, we believe that Arbitrum Rollup is the bakeoff choice that Reddit kneads. We far surpass Reddit's specified workload requirement at present, have much room to optimize Arbitrum Rollup in the near term, and have a clear path to get Reddit to hundreds of millions of users. Furthermore, we are the only project that gives developers and users the identical interface as the Ethereum blockchain and is fully interoperable and tooling-compatible, and we do this all without any new trust assumptions or centralized components. But no matter how the cookie crumbles, we're glad to have participated in this bake-off and we thank you for your consideration. About Offchain Labs Offchain Labs, Inc. is a venture-funded New York company that spun out of Princeton University research, and is building the Arbitrum platform to usher in the next generation of scalable, interoperable, and compatible smart contracts. Offchain Labs is backed by Pantera Capital, Compound VC, Coinbase Ventures, and others. Leadership Team Ed Felten Ed Felten is Co-founder and Chief Scientist at Offchain Labs. He is on leave from Princeton University, where he is the Robert E. Kahn Professor of Computer Science and Public Affairs. From 2015 to 2017 he served at the White House as Deputy United States Chief Technology Officer and senior advisor to the President. He is an ACM Fellow and member of the National Academy of Engineering. Outside of work, he is an avid runner, cook, and L.A. Dodgers fan. Steven Goldfeder Steven Goldfeder is Co-founder and Chief Executive Officer at Offchain Labs. He holds a PhD from Princeton University, where he worked at the intersection of cryptography and cryptocurrencies including threshold cryptography, zero-knowledge proof systems, and post-quantum signatures. He is a co-author of Bitcoin and Cryptocurrency Technologies, the leading textbook on cryptocurrencies, and he has previously worked at Google and Microsoft Research, where he co-invented the Picnic signature algorithm. When not working, you can find Steven spending time with his family, taking a nature walk, or twisting balloons. Harry Kalodner Harry Kalodner is Co-founder and Chief Technology Officer at Offchain Labs where he leads the engineering team. Before the company he attended Princeton as a Ph.D candidate where his research explored economics, anonymity, and incentive compatibility of cryptocurrencies, and he also has worked at Apple. When not up at 3:00am writing code, Harry occasionally sleeps.
Syscoin Platform’s Great Reddit Scaling Bake-off Proposal
https://preview.redd.it/rqt2dldyg8e51.jpg?width=1044&format=pjpg&auto=webp&s=777ae9d4fbbb54c3540682b72700fc4ba3de0a44 We are excited to participate and present Syscoin Platform's ideal characteristics and capabilities towards a well-rounded Reddit Community Points solution! Our scaling solution for Reddit Community Points involves 2-way peg interoperability with Ethereum. This will provide a scalable token layer built specifically for speed and high volumes of simple value transfers at a very low cost, while providing sovereign ownership and onchain finality. Token transfers scale by taking advantage of a globally sorting mempool that provides for probabilistically secure assumptions of “as good as settled”. The opportunity here for token receivers is to have an app-layer interactivity on the speed/security tradeoff (99.9999% assurance within 10 seconds). We call this Z-DAG, and it achieves high-throughput across a mesh network topology presently composed of about 2,000 geographically dispersed full-nodes. Similar to Bitcoin, however, these nodes are incentivized to run full-nodes for the benefit of network security, through a bonded validator scheme. These nodes do not participate in the consensus of transactions or block validation any differently than other nodes and therefore do not degrade the security model of Bitcoin’s validate first then trust, across every node. Each token transfer settles on-chain. The protocol follows Bitcoin core policies so it has adequate code coverage and protocol hardening to be qualified as production quality software. It shares a significant portion of Bitcoin’s own hashpower through merged-mining. This platform as a whole can serve token microtransactions, larger settlements, and store-of-value in an ideal fashion, providing probabilistic scalability whilst remaining decentralized according to Bitcoin design. It is accessible to ERC-20 via a permissionless and trust-minimized bridge that works in both directions. The bridge and token platform are currently available on the Syscoin mainnet. This has been gaining recent attention for use by loyalty point programs and stablecoins such as Binance USD.
Syscoin Foundation identified a few paths for Reddit to leverage this infrastructure, each with trade-offs. The first provides the most cost-savings and scaling benefits at some sacrifice of token autonomy. The second offers more preservation of autonomy with a more narrow scope of cost savings than the first option, but savings even so. The third introduces more complexity than the previous two yet provides the most overall benefits. We consider the third as most viable as it enables Reddit to benefit even while retaining existing smart contract functionality. We will focus on the third option, and include the first two for good measure.
Distribution, burns and user-to-user transfers of Reddit Points are entirely carried out on the Syscoin network. This full-on approach to utilizing the Syscoin network provides the most scalability and transaction cost benefits of these scenarios. The tradeoff here is distribution and subscription handling likely migrating away from smart contracts into the application layer.
The Reddit Community Points ecosystem can continue to use existing smart contracts as they are used today on the Ethereum mainchain. Users migrate a portion of their tokens to Syscoin, the scaling network, to gain much lower fees, scalability, and a proven base layer, without sacrificing sovereign ownership. They would use Syscoin for user-to-user transfers. Tips redeemable in ten seconds or less, a high-throughput relay network, and onchain settlement at a block target of 60 seconds.
Integration between Matic Network and Syscoin Platform - similar to Syscoin’s current integration with Ethereum - will provide Reddit Community Points with EVM scalability (including the Memberships ERC777 operator) on the Matic side, and performant simple value transfers, robust decentralized security, and sovereign store-of-value on the Syscoin side. It’s “the best of both worlds”. The trade-off is more complex interoperability.
Syscoin + Matic Integration
Matic and Blockchain Foundry Inc, the public company formed by the founders of Syscoin, recently entered a partnership for joint research and business development initiatives. This is ideal for all parties as Matic Network and Syscoin Platform provide complementary utility. Syscoin offers characteristics for sovereign ownership and security based on Bitcoin’s time-tested model, and shares a significant portion of Bitcoin’s own hashpower. Syscoin’s focus is on secure and scalable simple value transfers, trust-minimized interoperability, and opt-in regulatory compliance for tokenized assets rather than scalability for smart contract execution. On the other hand, Matic Network can provide scalable EVM for smart contract execution. Reddit Community Points can benefit from both. Syscoin + Matic integration is actively being explored by both teams, as it is helpful to Reddit, Ethereum, and the industry as a whole.
Total cost for these 100k transactions: $0.63 USD See the live fee comparison for savings estimation between transactions on Ethereum and Syscoin. Below is a snapshot at time of writing: ETH price: $318.55 ETH gas price: 55.00 Gwei ($0.37) Syscoin price: $0.11 Snapshot of live fee comparison chart Z-DAG provides a more efficient fee-market. A typical Z-DAG transaction costs 0.0000582 SYS. Tokens can be safely redeemed/re-spent within seconds or allowed to settle on-chain beforehand. The costs should remain about this low for microtransactions. Syscoin will achieve further reduction of fees and even greater scalability with offchain payment channels for assets, with Z-DAG as a resilience fallback. New payment channel technology is one of the topics under research by the Syscoin development team with our academic partners at TU Delft. In line with the calculation in the Lightning Networks white paper, payment channels using assets with Syscoin Core will bring theoretical capacity for each person on Earth (7.8 billion) to have five on-chain transactions per year, per person, without requiring anyone to enter a fee market (aka “wait for a block”). This exceeds the minimum LN expectation of two transactions per person, per year; one to exist on-chain and one to settle aggregated value.
Tools to simplify using Syscoin Bridge as a service with dapps and wallets will be released some time after implementation of Syscoin Core 4.2. These will be based upon the same processes which are automated in the current live Sysethereum Dapp that is functioning with the Syscoin mainnet.
The Syscoin Ethereum Bridge is secured by Agent nodes participating in a decentralized and incentivized model that involves roles of Superblock challengers and submitters. This model is open to participation. The benefits here are trust-minimization, permissionless-ness, and potentially less legal/regulatory red-tape than interop mechanisms that involve liquidity providers and/or trading mechanisms. The trade-off is that due to the decentralized nature there are cross-chain settlement times of one hour to cross from Ethereum to Syscoin, and three hours to cross from Syscoin to Ethereum. We are exploring ways to reduce this time while maintaining decentralization via zkp. Even so, an “instant bridge” experience could be provided by means of a third-party liquidity mechanism. That option exists but is not required for bridge functionality today. Typically bridges are used with batch value, not with high frequencies of smaller values, and generally it is advantageous to keep some value on both chains for maximum availability of utility. Even so, the cross-chain settlement time is good to mention here.
Ethereum -> Syscoin: Matic or Ethereum transaction fee for bridge contract interaction, negligible Syscoin transaction fee for minting tokens Syscoin -> Ethereum: Negligible Syscoin transaction fee for burning tokens, 0.01% transaction fee paid to Bridge Agent in the form of the ERC-20, Matic or Ethereum transaction fee for contract interaction.
Zero-Confirmation Directed Acyclic Graph is an instant settlement protocol that is used as a complementary system to proof-of-work (PoW) in the confirmation of Syscoin service transactions. In essence, a Z-DAG is simply a directed acyclic graph (DAG) where validating nodes verify the sequential ordering of transactions that are received in their memory pools. Z-DAG is used by the validating nodes across the network to ensure that there is absolute consensus on the ordering of transactions and no balances are overflowed (no double-spends).
Unique fee-market that is more efficient for microtransaction redemption and settlement
Uses decentralized means to enable tokens with value transfer scalability that is comparable or exceeds that of credit card networks
Provides high throughput and secure fulfillment even if blocks are full
Probabilistic and interactive
99.9999% security assurance within 10 seconds
Can serve payment channels as a resilience fallback that is faster and lower-cost than falling-back directly to a blockchain
Each Z-DAG transaction also settles onchain through Syscoin Core at 60-second block target using SHA-256 Proof of Work consensus
Z-DAG enables the ideal speed/security tradeoff to be determined per use-case in the application layer. It minimizes the sacrifice required to accept and redeem fast transfers/payments while providing more-than-ample security for microtransactions. This is supported on the premise that a Reddit user receiving points does need security yet generally doesn’t want nor need to wait for the same level of security as a nation-state settling an international trade debt. In any case, each Z-DAG transaction settles onchain at a block target of 60 seconds.
Syscoin 3.0 White Paper (4.0 white paper is pending. For improved scalability and less blockchain bloat, some features of v3 no longer exist in current v4: Specifically Marketplace Offers, Aliases, Escrow, Certificates, Pruning, Encrypted Messaging)
16MB block bandwidth per minute assuming segwit witness carrying transactions, and transactions ~200 bytes on average
SHA256 merge mined with Bitcoin
UTXO asset layer, with base Syscoin layer sharing identical security policies as Bitcoin Core
Z-DAG on asset layer, bridge to Ethereum on asset layer
On-chain scaling with prospect of enabling enterprise grade reliable trustless payment processing with on/offchain hybrid solution
Focus only on Simple Value Transfers. MVP of blockchain consensus footprint is balances and ownership of them. Everything else can reduce data availability in exchange for scale (Ethereum 2.0 model). We leave that to other designs, we focus on transfers.
Future integrations of MAST/Taproot to get more complex value transfers without trading off trustlessness or decentralization.
Zero-knowledge Proofs are a cryptographic new frontier. We are dabbling here to generalize the concept of bridging and also verify the state of a chain efficiently. We also apply it in our Digital Identity projects at Blockchain Foundry (a publicly traded company which develops Syscoin softwares for clients). We are also looking to integrate privacy preserving payment channels for off-chain payments through zkSNARK hub & spoke design which does not suffer from the HTLC attack vectors evident on LN. Much of the issues plaguing Lightning Network can be resolved using a zkSNARK design whilst also providing the ability to do a multi-asset payment channel system. Currently we found a showstopper attack (American Call Option) on LN if we were to use multiple-assets. This would not exist in a system such as this.
Web3 and mobile wallets are under active development by Blockchain Foundry Inc as WebAssembly applications and expected for release not long after mainnet deployment of Syscoin Core 4.2. Both of these will be multi-coin wallets that support Syscoin, SPTs, Ethereum, and ERC-20 tokens. The Web3 wallet will provide functionality similar to Metamask. Syscoin Platform and tokens are already integrated with Blockbook. Custom hardware wallet support currently exists via ElectrumSys. First-class HW wallet integration through apps such as Ledger Live will exist after 4.2. Current supported wallets Syscoin Spark Desktop Syscoin-Qt
Hey everyone! Welcome to Weekly Update 127. This week we’ll be going over Stakenet and DEX’s slightly more generally. You bet that we’ll also be covering the usual GitHub development of Raiden and an exciting new release! Let’s dig in. Raiden x Stakenet Stakenet project is combining Lightning Network and Masternodes into a new blockchain for running decentralized applications. One of the first decentralized applications they are building is the Stakenet decentralized exchange. Currently, the decentralized exchange supports Bitcoin, Litecoin and Stakenet via Lightning Network integration. In the future they are planning to expand the list of the supported crypto currencies with the Ethereum ERC-20 standard. Technology of choice which they believe will best support them in this integration is the Raiden Network. The Stakenet team, after dedicating some time into researching Raiden, are getting ready for the first testing of Raiden. For the first tests users will have to run a Raiden node separately and will have to use 2 different wallets (stakenet and ethereum wallet). However, after the initial setup user experience should be the same for trading as it is currently. We will be following the progress of testing closely and cover it’s progress. It will be interesting to see Raiden enabling the DEX, especially cross-chain. Development progress The development goals remain the same as they were during the previous several weeks. The first goal is to prepare the demo for the Reddit Scaling Bake-Off while the second one is to reach the Alderaan milestone in the Raiden light client. In the Raiden light client the Juggernaut testnet release was introduced. In this release the SDK has become much more robust. It includes many small bug fixes as well as some major reworks. Furthermore, it also introduces the completed withdraw functionality. As the repository where the development team is preparing the demo for the Reddit Scaling Bake-Off is not yet open sourced we can only look at the changes made in the old repositories. It seems that the Raiden team is preparing a couple of bigger features one of which is handling something called “claims”. It will be interesting to see what exactly are claims and how they are integrated in the Raiden protocol. Conclusion To wrap up this Weekly Update, it’s exciting to cover DEX’s and Raiden Network being utilized for the name of decentralization. Don’t forget to check out the new light client release. If you’re interested to learn more about this update, DEX’s or Raiden Network in general then please leave a comment and we’ll get back to you. Cheers!
People had been speculating since the dawn of crypto when the world’s largest online marketplaces, the ones of the Amazon caliber like eBay, Etsy or AliExpress, and, well, Amazon itself, would start to accept cryptocurrencies. There were a slew of rumors, opinions, and theories thickly interspersed with false reports popping up here and there of Amazon and its little cousins being on the verge of embracing cryptocurrencies. On top of that, someone has actually posted a petition on change.org to add Ether to Amazon as a payment method. by StealthEX Long story short, that was a waste of time. High hopes fell flat, and people lost religion. But not all. As the common wisdom goes, when hope dies, action begins. This exposition describes one such effort which tries to bring to fruition the idea of a decentralized marketplace for trading goods and services. And as you might have already figured it out, with a cryptocurrency as a means of payment. So let’s welcome Particl Marketplace and see what it has to offer – and what Amazon has missed.
What is it, in simple words?
Particl Marketplace is an online marketplace where you can trade goods and services. Not a big deal, you may think. However, what distinguishes it from places like Amazon as well as cryptocurrency-enabled marketplaces is the decentralized nature of purchases on Particl. You can think of it as a variety of a decentralized cryptocurrency exchange (aka DEX) where trades are being conducted on-chain. But in case of Particl, it is goods and services that are being traded, not fiat or crypto, with deals on-chain as well, fully encrypted and decentralized. Particl is a global peer-to-peer privacy-centered marketplace that uses an automated two-party escrow system. It is crypto-agnostic and designed to work with any cryptocurrency, creating a secure, highly-scalable environment supported by a privacy-focused blockchain-based platform. The team behind the project sees its mission in developing “a new decentralized, private and democratic economy” that is governed by the network of its users, with no central authority or middleman getting in the way. In the project developers’ own words, Particl enables everyone to participate in a free, anonymous exchange of all kinds of goods, without paying any fee and regardless of geographical location. To be sure, you are already thinking about Silk Road and its dark fate, and that the government is going to crack down monumentally on Particl one day. Well, the outcome may vary as the payments on the platform are made using its own cryptocurrency PART, with its laser focus on privacy and anonymity. But more on this later.
How did it grow up?
The development of the Particl project started in early 2017 with the release of the white paper describing the team’s vision for the marketplace, which was shortly followed by a successful seed funding that brought in enough funds ($750,000) to support the development of the project for a year (it turned out sufficient to last for over two years). These donations helped to establish the Particl Foundation, a non-profit Swiss organization with the goal of providing legal protection for the project to ensure its sustained development and compliance with government regulations. It receives 10% of all the staking rewards generated on the Particl network, making the project self-sustainable and free for most uses. Unlike other such projects in the crypto arena, Particl has been using its own blockchain from day one, which happened to be July 17, 2017. It was specifically designed to be crypto agnostic by supporting and working with any cryptocurrency. Additionally, it supported the smart contract tech out of the box, giving users an ability to build all kinds of decentralized applications (dApps) that can be directly integrated into the Particl marketplace. On May 31, 2018, the Particl Marketplace, the Holy Grail of the entire endeavor, was made available for alpha testing on the testnet of the project, which later split into development and stable branches. It went live with the mainnet release of the Particl Open Marketplace on August 12, 2019, which featured Particl Desktop 2.0.0, a client-side application providing user interface and built-in wallet functions. On November 25, 2019, the Particl Desktop 2.3.0 client was released that enabled Bitcoin payments and marked the introduction of untraceable transactions. With the help of the new in-wallet exchange module, everyone can easily swap their bitcoins for the native PART coin. Moreover, the module allows seamless integration of third-party accountless exchange services right into the marketplace, with StealthEX being one of them.
How is it different from other marketplaces?
The common solution many P2P marketplaces implement to protect buyers and sellers from the other party failing to honor their end of the bargain is through third-party escrow, where the “third-party” in the majority of places and cases is the platform itself that the market participants must mutually trust. In short, it is a single point of failure. And the selling (pardon the pun) point of the entire Particl’s marketplace is its decentralized escrow, which is a thing entirely between the two parties engaged. No middlemen allowed here! And these are not empty words. Particl implements the concept best known as Mutually Assured Destruction (aptly shortened as MAD), a military doctrine you are certainly familiar with, and probably even afraid of, that consists in a mutual destruction of two belligerent parties in an all-out nuclear holocaust. If you are curious, the idea stems from the game theory and has a lot to do with the Nash Equilibrium, of John Nash’s fame. In a nutshell, Particl removes the need for a trusted escrow agent by introducing MAD escrow smart contracts. A MAD escrow contract allows to lock funds in a multi-signature address that can be released only if all the parties sign off on the transaction. So both the seller and the buyer lock in the contract an agreed amount for a specified period of time, with the buyer also depositing the payment for the items purchased. The escrowed funds are released when both parties confirm the fulfillment of the agreement. Should one party break the terms, the funds remain locked for good causing a mutual financial loss until both parties agree to sign off. Another crucial aspect of Particl Marketplace is its end-to-end privacy. The problem with conventional marketplaces acting as an escrow agent is that the communication between the parties should be open to the agent for it to serve as an arbitrator. With Particl, it is no longer required, and all messages between the buyer and the seller are encrypted. Despite being public, only their recipient can decrypt them, which effectively makes messages untraceable. This is also where the PART coin turns up quite handy. It enables three different privacy modes, and with the most secure mode, the Anon mode, PART transactions utilize the RingCT privacy protocol, which hides both the amount transferred and the identity of the parties transacting. Accordingly, every part of the entire Particl trading environment is thoroughly decentralized, and the full anonymity of market participants is maintained at all times, making the platform a completely trustless marketplace. Big Brother is no longer watching you. Aside from that, you can stake PART and generate a source of passive income for yourself. Particl uses a custom Proof-of-Stake consensus protocol, allowing you to get a piece of the pie in the form of new coins created at each block according to the scheduled inflation process. The annual inflation rate is initially set to 5% and goes down 1 percentage point every year until it finally floors at 2% indefinitely. Moreover, these rates are a bare minimum as they assume that all PART coins have been staked. Otherwise, the income will be bigger and better as the same rewards are paid to fewer coins. Additionally, your passive income through staking PART will be augmented by the fees generated through the everyday marketplace operations. Whether it is network fees collected via PART transactions or marketplace listing fees paid by the sellers, all of them contribute to the stakers’ rewards. At the end of the day, staking PART can turn into a profitable business once the Particl platform starts to attract more traffic. In simple words, the more popular the market gets, the more fees it generates, the more coins the stakers earn. As PART is a standalone cryptocurrency, it can be used outside Particl Marketplace as well. So if you plan on using it for purposes other than eCommerce, it is traded on several exchanges, for example, HitBTC and Bittrex, with more exchanges to list PART in the future. There are native Particl wallets available for storing PART such as Particl Qt with Ledger support, Particl-cli, and Particl Copay Wallet, with the latter available for both the desktop and the mobile. There is also a third-party multicurrency Flare Wallet, enabling cold staking for Particl. Running Particl is a collective effort, which means no operational costs and no company bagging profits from it. The marketplace buyers don’t pay any commissions other than tiny network confirmation fees, while the sellers are only charged a small listing fee to keep spam listings to a minimum. This creates a highly competitive environment, with the sellers making more profits and the buyers having access to cheaper goods and services as a result.
What’s in the pipeline?
The next major release of the Particl Marketplace should have been Particl Desktop 2.4.0, but it was later rebranded as Particl Desktop 3.0 to reflect its breakthrough nature. It is set for release in the second half of 2020 and will enable the addition of user-created markets and storefronts, effectively turning the Particl marketplace into a network of specialized markets. And if you think about it, that makes perfect sense. Say, you have a social network account highly merited and full of karma that you want to sell, whatever your reasons might be. Then creating a dedicated market for trading such accounts privately and securely may look extremely appealing to you. Whether it is the right thing to do is another matter, of course. Kidding aside, it is obviously not about selling or offering something that the society on the whole doesn’t approve of or frowns upon. If you are a freelancer, for example, a graphics designer or a translator, you would be certainly interested in the future freelancer markets – along with your potential employers. Put simply, birds of a feather should flock together. To keep things in perspective, popular freelancer markets that exist today charge up to 10-20% of what you would get from your client if you negotiated directly. All in all, establishing communities across the marketplace seems to be the next logical step in the natural evolution and growth of the platform. In fact, it is a little surprising that the Particl team didn’t come up with this idea earlier. Meanwhile, we wish Particl success and good luck in achieving their goals and aspirations. And remember if you need to exchange your coins StealthEX is here for you. We provide a selection of more than 250 coins and constantly updating the list so that our customers will find a suitable option. Our service does not require registration and allows you to remain anonymous. Why don’t you check it out? Just go to StealthEX and follow these easy steps: ✔ Choose the pair and the amount for your exchange. For example ETH to PART. ✔ Press the “Start exchange” button. ✔ Provide the recipient address to which the coins will be transferred. ✔ Move your cryptocurrency for the exchange. ✔ Receive your PART coins! Follow us on Medium, Twitter, Facebook, and Reddit to get StealthEX.io updates and the latest news about the crypto world. For all requests message us via [email protected]. The views and opinions expressed here are solely those of the author. Every investment and trading move involves risk. You should conduct your own research when making a decision. Original article was posted onhttps://stealthex.io/blog/2020/08/26/particl-marketplace-where-sellers-meet-buyers/
Author: Gamals Ahmed, CoinEx Business Ambassador https://preview.redd.it/5bqakdqgl3g51.jpg?width=865&format=pjpg&auto=webp&s=b709794863977eb6554e3919b9e00ca750e3e704 A decentralized storage network that transforms cloud storage into an account market. Miners obtain the integrity of the original protocol by providing data storage and / or retrieval. On the contrary, customers pay miners to store or distribute data and retrieve it. Filecoin announced, that there will be more delays before its main network is officially launched. Filecoin developers postponed the release date of their main network to late July to late August 2020. As mentioned in a recent announcement, the Filecoin team said that the initiative completed the first round of the internal protocol security audit. Platform developers claim that the results of the review showed that they need to make several changes to the protocol’s code base before performing the second stage of the software testing process. Created by Protocol Labs, Filecoin was developed using File System (IPFS), which is a peer-to-peer data storage network. Filecoin will allow users to trade storage space in an open and decentralized market. Filecoin developers implemented one of the largest cryptocurrency sales in 2017. They have privately obtained over $ 200 million from professional or accredited investors, including many institutional investors. The main network was slated to launch last month, but in February 2020, the Philly Queen development team delayed the release of the main network between July 15 and July 17, 2020. They claimed that the outbreak of the Coronavirus (COVID-19) in China was the main cause of the delay. The developers now say that they need more time to solve the problems found during a recent codecase audit. The Filecoin team noted the following: “We have drafted a number of protocol changes to ensure that building our major network launch is safe and economically sound.” The project developers will add them to two different implementations of Filecoin (Lotus and go-filecoin) in the coming weeks. Filecoin developers conducted a survey to allow platform community members to cast their votes on three different launch dates for Testnet Phase 2 and mainnet. The team reported that the community gave their votes. Based on the vote results, the Filecoin team announced a “conservative” estimate that the second phase of the network test should begin by May 11, 2020. The main Filecoin network may be launched sometime between July 20 and August 21, 2020. The updates to the project can be found on the Filecoin Road Map. Filecoin developers stated: “This option will make us get the most important protocol changes first, and then implement the rest as protocol updates during testnet.” Filecoin is back down from the final test stage. Another filecoin decentralized storage network provider launched its catalytic test network, the final stage of the storage network test that supports the blockchain. In a blog post on her website, Filecoin said she will postpone the last test round until August. The company also announced a calibration period from July 20 to August 3 to allow miners to test their mining settings and get an idea of how competition conditions affected their rewards. Filecoin had announced earlier last month that the catalytic testnet test would precede its flagship launch. The delay in the final test also means that the company has returned the main launch window between August 31 and September 21. Despite the lack of clear incentives for miners and multiple delays, Filecoin has succeeded in attracting huge interest, especially in China. Investors remained highly speculating on the network’s mining hardware and its premium price. Mining in Filecoin In most blockchain protocols, “miners” are network participants who do the work necessary to promote and maintain the blockchain. To provide these services, miners are compensated in the original cryptocurrency. Mining in Filecoin works completely differently — instead of contributing to computational power, miners contribute storage capacity to use for dealing with customers looking to store data. Filecoin will contain several types of miners: Storage miners responsible for storing files and data on the network. Miners retrieval, responsible for providing quick tubes for file recovery. Miners repair to be carried out. Storage miners are the heart of the network. They earn Filecoin by storing data for clients, and computerizing cipher directories to check storage over time. The probability of earning the reward reward and transaction fees is proportional to the amount of storage that the Miner contributes to the Filecoin network, not the hash power. Retriever miners are the veins of the network. They earn Filecoin by winning bids and mining fees for a specific file, which is determined by the market value of the said file size. Miners bandwidth and recovery / initial transaction response time will determine its ability to close recovery deals on the network. The maximum bandwidth of the recovery miners will determine the total amount of deals that it can enter into. In the current implementation, the focus is mostly on storage miners, who sell storage capacity for FIL.
The current system specifications recommended for running the miner are:
NVIDIA-manufactured GPU (to be expanded).
SSD drive designated as large buffer (512GB +).
Large amount of RAM for data replication account (128GB +)
Compared to the hardware requirements for running a validity checker, these standards are much higher — although they definitely deserve it. Since these will not increase in the presumed future, the money spent on Filecoin mining hardware will provide users with many years of reliable service, and they pay themselves many times. Think of investing as a small business for cloud storage. To launch a model on the current data hosting model, it will cost millions of dollars in infrastructure and logistics to get started. With Filecoin, you can do the same for a few thousand dollars. Proceed to mining Deals are the primary function of the Filecoin network, and it represents an agreement between a client and miners for a “storage” contract. Once the customer decides to have a miner to store based on the available capacity, duration and price required, he secures sufficient funds in a linked portfolio to cover the total cost of the deal. The deal is then published once the mine accepts the storage agreement. By default, all Filecoin miners are set to automatically accept any deal that meets their criteria, although this can be disabled for miners who prefer to organize their deals manually. After the deal is published, the customer prepares the data for storage and then transfers it to the miner. Upon receiving all the data, the miner fills in the data in a sector, closes it, and begins to provide proofs to the chain. Once the first confirmation is obtained, the customer can make sure the data is stored correctly, and the deal has officially started. Throughout the deal, the miner provides continuous proofs to the chain. Clients gradually pay with money they previously closed. If there is missing or late evidence, the miner is punished. More information about this can be found in the Runtime, Cut and Penalties section of this page. At Filecoin, miners earn two different types of rewards for their efforts: storage fees and reward prevention. Storage fees are the fees that customers pay regularly after reaching a deal, in exchange for storing data. This fee is automatically deposited into the withdrawal portfolio associated with miners while they continue to perform their duties over time, and is locked for a short period upon receipt. Block rewards are large sums given to miners calculated on a new block. Unlike storage fees, these rewards do not come from a linked customer; Instead, the new FIL “prints” the network as an inflationary and incentive measure for miners to develop the chain. All active miners on the network have a chance to get a block bonus, their chance to be directly proportional to the amount of storage space that is currently being contributed to the network. Duration of operation, cutting and penalties “Slashing” is a feature found in most blockchain protocols, and is used to punish miners who fail to provide reliable uptime or act maliciously against the network. In Filecoin, miners are susceptible to two different types of cut: storage error cut, unanimously reduce error. Storage Error Reduction is a term used to include a wider range of penalties, including error fees, sector penalties, and termination fees. Miners must pay these penalties if they fail to provide reliability of the sector or decide to leave the network voluntarily. An error fee is a penalty that a miner incurs for each non-working day. Sector punishment: A penalty incurred by a miner of a disrupted sector for which no error was reported before the WindowPoSt inspection. The sector will pay an error fee after the penalty of the sector once the error is discovered. Termination Fee: A penalty that a miner incurs when a sector is voluntary or involuntarily terminated and removed from the network. Cutting consensus error is the penalty that a miner incurs for committing consensus errors. This punishment applies to miners who have acted maliciously against the network consensus function. Filecoin miners Eight of the top 10 Felticoin miners are Chinese investors or companies, according to the blockchain explorer, while more companies are selling cloud mining contracts and distributed file sharing system hardware. CoinDesk’s Wolfe Chao wrote: “China’s craze for Filecoin may have been largely related to the long-standing popularity of crypto mining in the country overall, which is home to about 65% of the computing power on Bitcoin at discretion.” With Filecoin approaching the launch of the mainnet blocknet — after several delays since the $ 200 million increase in 2017 — Chinese investors are once again speculating strongly about network mining devices and their premium prices. Since Protocol Labs, the company behind Filecoin, released its “Test Incentives” program on June 9 that was scheduled to start in a week’s time, more than a dozen Chinese companies have started selling cloud mining contracts and hardware — despite important details such as economics Mining incentives on the main network are still endless. Sales volumes to date for each of these companies can range from half a million to tens of millions of dollars, according to self-reported data on these platforms that CoinDesk has watched and interviews with several mining hardware manufacturers. Filecoin’s goal is to build a distributed storage network with token rewards to spur storage hosting as a way to drive wider adoption. Protocol Labs launched a test network in December 2019. But the tokens mined in the testing environment so far are not representative of the true silicon coin that can be traded when the main network is turned on. Moreover, the mining incentive economics on testnet do not represent how final block rewards will be available on the main network. However, data from Blockecoin’s blocknetin testnet explorers show that eight out of 10 miners with the most effective mining force on testnet are currently Chinese miners. These eight miners have about 15 petabytes (PB) of effective storage mining power, accounting for more than 85% of the total test of 17.9 petable. For the context, 1 petabyte of hard disk storage = 1000 terabytes (terabytes) = 1 million gigabytes (GB). Filecoin craze in China may be closely related to the long-standing popularity of crypt mining in the country overall, which is home to about 65% of the computing power on Bitcoin by estimation. In addition, there has been a lot of hype in China about foreign exchange mining since 2018, as companies promote all types of devices when the network is still in development. “Encryption mining has always been popular in China,” said Andy Tien, co-founder of 1475, one of several mining hardware manufacturers in Philquin supported by prominent Chinese video indicators such as Fenbushi and Hashkey Capital. “Even though the Velikoyen mining process is more technologically sophisticated, the idea of mining using hard drives instead of specialized machines like Bitcoin ASIC may be a lot easier for retailers to understand,” he said. Meanwhile, according to Feixiaohao, a Chinese service comparable to CoinMarketCap, nearly 50 Chinese crypto exchanges are often somewhat unknown with some of the more well-known exchanges including Gate.io and Biki — have listed trading pairs for Filecoin currency contracts for USDT. In bitcoin mining, at the current difficulty level, one segment per second (TH / s) fragmentation rate is expected to generate around 0.000008 BTC within 24 hours. The higher the number of TH / s, the greater the number of bitcoins it should be able to produce proportionately. But in Filecoin, the efficient mining force of miners depends on the amount of data stamped on the hard drive, not the total size of the hard drive. To close data in the hard drive, the Filecoin miner still needs processing power, i.e. CPU or GPU as well as RAM. More powerful processors with improved software can confine data to the hard drive more quickly, so miners can combine more efficient mining energy faster on a given day. As of this stage, there appears to be no transparent way at the network level for retail investors to see how much of the purchased hard disk drive was purchased which actually represents an effective mining force. The U.S.-based Labs Protocol was behind Filecoin’s initial coin offer for 2017, which raised an astonishing $ 200 million. This was in addition to a $ 50 million increase in private investment supported by notable venture capital projects including Sequoia, Anderson Horowitz and Union Square Ventures. CoinDk’s parent company, CoinDk, has also invested in Protocol Labs. After rounds of delay, Protocol Protocols said in September 2019 that a testnet launch would be available around December 2019 and the main network would be rolled out in the first quarter of 2020. The test started as promised, but the main network has been delayed again and is now expected to launch in August 2020. What is Filecoin mining process? Filecoin mainly consists of three parts: the storage market (the chain), the blockecin Filecoin, and the search market (under the chain). Storage and research market in series and series respectively for security and efficiency. For users, the storage frequency is relatively low, and the security requirements are relatively high, so the storage process is placed on the chain. The retrieval frequency is much higher than the storage frequency when there is a certain amount of data. Given the performance problem in processing data on the chain, the retrieval process under the chain is performed. In order to solve the security issue of payment in the retrieval process, Filecoin adopts the micro-payment strategy. In simple terms, the process is to split the document into several copies, and every time the user gets a portion of the data, the corresponding fee is paid. Types of mines corresponding to Filecoin’s two major markets are miners and warehousers, among whom miners are primarily responsible for storing data and block packages, while miners are primarily responsible for data query. After the stable operation of the major Filecoin network in the future, the mining operator will be introduced, who is the main responsible for data maintenance. In the initial release of Filecoin, the request matching mechanism was not implemented in the storage market and retrieval market, but the takeover mechanism was adopted. The three main parts of Filecoin correspond to three processes, namely the stored procedure, retrieval process, packaging and reward process. The following figure shows the simplified process and the income of the miners: The Filecoin mining process is much more complicated, and the important factor in determining the previous mining profit is efficient storage. Effective storage is a key feature that distinguishes Filecoin from other decentralized storage projects. In Filecoin’s EC consensus, effective storage is similar to interest in PoS, which determines the likelihood that a miner will get the right to fill, that is, the proportion of miners effectively stored in the entire network is proportional to final mining revenue. It is also possible to obtain higher effective storage under the same hardware conditions by improving the mining algorithm. However, the current increase in the number of benefits that can be achieved by improving the algorithm is still unknown. It seeks to promote mining using Filecoin Discover Filecoin announced Filecoin Discover — a step to encourage miners to join the Filecoin network. According to the company, Filecoin Discover is “an ever-growing catalog of numerous petabytes of public data covering literature, science, art, and history.” Miners interested in sharing can choose which data sets they want to store, and receive that data on a drive at a cost. In exchange for storing this verified data, miners will earn additional Filecoin above the regular block rewards for storing data. Includes the current catalog of open source data sets; ENCODE, 1000 Genomes, Project Gutenberg, Berkley Self-driving data, more projects, and datasets are added every day. Ian Darrow, Head of Operations at Filecoin, commented on the announcement: “Over 2.5 quintillion bytes of data are created every day. This data includes 294 billion emails, 500 million tweets and 64 billion messages on social media. But it is also climatology reports, disease tracking maps, connected vehicle coordinates and much more. It is extremely important that we maintain data that will serve as the backbone for future research and discovery”. Miners who choose to participate in Filecoin Discover may receive hard drives pre-loaded with verified data, as well as setup and maintenance instructions, depending on the company. The Filecoin team will also host the Slack (fil-Discover-support) channel where miners can learn more. Filecoin got its fair share of obstacles along the way. Last month Filecoin announced a further delay before its main network was officially launched — after years of raising funds. In late July QEBR (OTC: QEBR) announced that it had ceded ownership of two subsidiaries in order to focus all of the company’s resources on building blockchain-based mining operations. The QEBR technology team previously announced that it has proven its system as a Filecoin node valid with CPU, GPU, bandwidth and storage compatibility that meets all IPFS guidelines. The QEBR test system is connected to the main Filecoin blockchain and the already mined filecoin coin has already been tested. “The disclosure of Sheen Boom and Jihye will allow our team to focus only on the upcoming global launch of Filecoin. QEBR branch, Shenzhen DZD Digital Technology Ltd. (“ DZD “), has a strong background in blockchain development, extraction Data, data acquisition, data processing, data technology research. We strongly believe Filecoin has the potential to be a leading blockchain-based cryptocurrency and will make every effort to make QEBR an important player when Mainecoin mainnet will be launched soon”. IPFS and Filecoin Filecoin and IPFS are complementary protocols for storing and sharing data in a decentralized network. While users are not required to use Filecoin and IPFS together, the two combined are working to resolve major failures in the current web infrastructure. IPFS It is an open source protocol that allows users to store and transmit verifiable data with each other. IPFS users insist on data on the network by installing it on their own device, to a third-party cloud service (known as Pinning Services), or through community-oriented systems where a group of individual IPFS users share resources to ensure the content stays live. The lack of an integrated catalytic mechanism is the challenge Filecoin hopes to solve by allowing users to catalyze long-term distributed storage at competitive prices through the storage contract market, while maintaining the efficiency and flexibility that the IPFS network provides. Using IPFS In IPFS, the data is hosted by the required data installation nodes. For data to persist while the user node is offline, users must either rely on their other peers to install their data voluntarily or use a central install service to store data. Peer-to-peer reliance caching data may be a good thing as one or multiple organizations share common files on an internal network, or where strong social contracts can be used to ensure continued hosting and preservation of content in the long run. Most users in an IPFS network use an installation service. Using Filecoin The last option is to install your data in a decentralized storage market, such as Filecoin. In Filecoin’s structure, customers make regular small payments to store data when a certain availability, while miners earn those payments by constantly checking the integrity of this data, storing it, and ensuring its quick recovery. This allows users to motivate Filecoin miners to ensure that their content will be live when it is needed, a distinct advantage of relying only on other network users as required using IPFS alone. Filecoin, powered by IPFS It is important to know that Filecoin is built on top of IPFS. Filecoin aims to be a very integrated and seamless storage market that takes advantage of the basic functions provided by IPFS, they are connected to each other, but can be implemented completely independently of each other. Users do not need to interact with Filecoin in order to use IPFS. Some advantages of sharing Filecoin with IPFS:
Filecoin and IPFS CIDs share hash specifications.
Use libp2p by Filecoin nodes to create secure connections with each other.
Messaging between nodes and cluster propagation is facilitated in Filecoin by libp2p pubsub.
IPLD use for blockchain data structures.
Use Graphsync to transfer data between nodes.
Of all the decentralized storage projects, Filecoin is undoubtedly the most interested, and IPFS has been running stably for two years, fully demonstrating the strength of its core protocol. Filecoin’s ability to obtain market share from traditional central storage depends on end-user experience and storage price. Currently, most Filecoin nodes are posted in the IDC room. Actual deployment and operation costs are not reduced compared to traditional central cloud storage, and the storage process is more complicated. PoRep and PoSt, which has a large number of proofs of unknown operation, are required to cause the actual storage cost to be so, in the early days of the release of Filecoin. The actual cost of storing data may be higher than the cost of central cloud storage, but the initial storage node may reduce the storage price in order to obtain block rewards, which may result in the actual storage price lower than traditional central cloud storage. In the long term, Filecoin still needs to take full advantage of its P2P storage, convert storage devices from specialization to civil use, and improve its algorithms to reduce storage costs without affecting user experience. The storage problem is an important problem to be solved in the blockchain field, so a large number of storage projects were presented at the 19th Web3 Summit. IPFS is an important part of Web3 visibility. Its development will affect the development of Web3 to some extent. Likewise, Web3 development somewhat determines the future of IPFS. Filecoin is an IPFS-based storage class project initiated by IPFS. There is no doubt that he is highly expected. Resources :
Hi there, I'm currently running a masternode within testnet and would love to get some blockchain data to consolidate. e.g. number of masternodes, transaction per seconds,... Of course this can be expanded over time, but first I have to start by getting at least some data. Since the DeFi-Chain is a Bitcoin Fork I quickly found the following API calls list. https://en.bitcoin.it/wiki/Original_Bitcoin_client/API_calls_list However I would like to know if there is a separate list for dfi specific commands which have been added (e.g. listmasternodes).
We know that constant interaction with our users is key to developing a personal connection which is why we also devote our time “partnering” with the community, if you will. Ignite community launch Ankr is heating things up with a move to Ignite, an uncensorable and decentralized microblogging social network with features that’ll be familiar for Twitter users. The service is built upon our partner Binance’s Smart Chain, and users can check us out here. https://preview.redd.it/2t1a1bzvree51.png?width=700&format=png&auto=webp&s=700e81616206e8b07dc932d1db999f1bd727914e Celo Vote dApp launch Governance is an essential part of many prominent cryptocurrencies but can be hard for the user just looking for passive income. Our new CeloVote automates this and allows users to earn the maximum potential reward in a matter of seconds. The dApp is integrated with our platform and is just as easy as deploying any other node. https://preview.redd.it/9kgqa12yree51.png?width=700&format=png&auto=webp&s=c19123b71abb4e6a44151b8469c3e4fce951b7c0 AMA with DeFi project Meter Meter is one of the many DeFi projects we are collaborating with this month. What separates Meter from the rest is pegging its purchasing power to electricity, which results in far more stable value than any fiat currency. Our AMA (following our partnership) transcript can be read here.
DeFi Integrations Besides Meter, we were also able to offer our services to upcoming projects Plutus DeFi and Keysians. Both are nascent platforms in the sector, and we look forward to seeing how our relationship with each deepens in the following months. Polkadot, the next big thingPolkadot’s validator node integration was completed this month, allowing users to interact with the protocol looking to connect all blockchains. Additionally, Polkadot parachains Bifrost and ChainX completed integration this month as well with more on the way. Binance Smart Chain integration Binance is working on a chain supporting smart contracts, optimized for developers and Ankr is thrilled to help. Late last month we had launched developer tools to help get development going, and this month we followed up by enabling BNB staking for the Testnet! Keep an eye on this one, as we will release more news in the coming days! Bitcoin Full node integration The several hours (or days, depending on your system) that it takes to initialize the 300GB blockchain for Bitcoin core is one of the main reasons why there are so few nodes. We’re excited to announce that Bitcoin full nodes are now part of our platform — and we will suggest our community on some things you can do with all the time you’ll be saving. Keep your eyes on our blog! Additional full node integrations Following on the heels of Bitcoin’s node integration, users can also now deploy full nodes to perform data analytics for the Decred, Elastos, ETH Classic, NEO, and Nuls native blockchains. https://preview.redd.it/hfupzm40see51.png?width=700&format=png&auto=webp&s=62069d748b5249b7911fbdae8b6aabc406d21219
With the final ETH 2.0 testnet version “Medalla” being announced for an August 4th launch, we should start to see further developments for DeFi and Ethereum related dApps. Regardless, Ankr is ready for when ETH 2.0 launches on the mainnet!
Just like the first Protocol (Core 1.0), the TKEY asset used in the TkeyNet network — there are no changes in this plan. After launching TkeyNet, you can transfer TKEY to any user on the TKEY network without any restrictions.
Transactions in TkeyNet will be much faster than it was before. You can check it in practice.
The Protocol has a built-in TkeyConnect module, which allows you to connect various blockchains to our network to conduct transactions directly in the TkeyNet blockchain. Besides, TkeyConnect meets the international ISO and ISIN standards, which also allows you to conduct transactions with Fiat currencies and shares in the TkeyNet blockchain. TkeyConnect creates a flexible system, giving users the ability to store and conduct transactions in any assets, be it Bitcoin, Ethereum, Litecoin or dollars, euros, etc. https://preview.redd.it/rkroh5ulbjg51.png?width=700&format=png&auto=webp&s=c47a094c9ae3d22131bebea80badafd0dd1f8dd6
If the previous software based on the Core 1.0 Protocol, after switching to TkeyNet, wallets, a blockchain search engine, and other software related to the project will be adapted to the new Protocol.
Contact information and support service
Email addresses of the support service and other departments will be transferred to other service providers and will structure in the following areas: B2B and B2C. The list of email addresses will publish after updates are complete.
B2B (business-to-business)a term that means that a company or a division of a company sells its goods/services to corporate clients, that is, to other companies.B2C (business-to-consumer)is a term that refers to the commercial relationship between an organization (Business) and a private, so-called “end” consumer.
Information about the project, the company, and its products will subdivide into two websites: tkeycoin.com and tkey.org. The purpose of this division is to simplify product navigation, improve the appearance of pages for each product and solution, and update content. The solution is modular. The information will be structured according to sections and sites, dividing the corporate and user segments. As the products develop, the information will be updated. Technical specifications, documentation, and a description of the Protocol and its features will appear on the official website: tkey.org. Sections will fill in gradually.
Testing and launching TkeyNet
Between July 22 and July 24, Telnet was successfully launched in testnet mode. Our team is actively testing the entire TkeyNet network and its functions. The system is tested with different scenarios, its effectiveness is checked when working with high loads, and the security of the entire system is audited. Testing of TkeyNet is an important stage of production aimed at detailed research of the program code and identification of errors in the system operation. Comprehensive testing, which is carried out by our team, is necessary to determine the level of readiness of the system for subsequent operation. Testing is based on a set of test scenarios that cover the main business operations.
The testing process contains all the life cycle activities: dynamic and static indicators. The testing process involves planning, preparing, and evaluating a software product. The purpose of testing is to determine that all meet the requirements described, as well as to show that they are suitable for the stated purposes and for detecting errors.
You can buy or sell TKEY only on the exchange and not in any other source. Private transactions are subject to high risks. Once again, we remind you that at the time of updates, any transactions with TKEY will be invalid.
Safety of funds
If you use a local wallet on your computer, make a backup copy of the wallet.dat file. If you use a TkeySpace mobile wallet, make a backup copy of the private key (backup phrase).
Transactions before updates are completed
During updates-no transactions can be made on the network, which means that any private dealing made at the time of updates will be invalid. Additionally, we ask you to refrain from any actions related to TKEY until the end of updates protocol, including starting mining, local wallets, and reinstalling them.
Testing of the system and its functions takes place in a stable mode without days off. The test results that will receive at the end of this week will reflect the current state of Affairs. We will get up-to-date information about the end date of updates and the planned release date of TkeyNet. An announcement of interim test results, as well as future updates, will be published at the end of this week or early next.
Diamond in the rough forks to Blockcore tech. Solid 10x short term.
Meet x42 Protocol, the feeless blockchain that will allow anyone to host decentralized applications the easy way. X42 Protocol Price $0.01275764 Market Cap $247,173 (May 27, 2020) Explorer: https://explorer.x42.tech/ Website: https://www.x42.tech/ x42 is a decentralized cryptocurrency based on the Blockcore technology designed to be a multi-chain solution for DApps (decentralized applications) that allows for a range that goes anywhere from small indie developed games to large entrepreneurial projects that span dozens of facilities. The main idea behind x42 is to be a scalable, on-chain solution for any developer that wants to launch games and applications in general with minimal initial investment, zero transaction fees and near infinite scalability. The protocol works around a main blockchain which hosts all the x42 coins and three types of nodes. Side blockchains can be created at will by the developers that decide to use the x42 protocol to launch their projects, side blockchains are very flexible and allow for a great deal of customization. The main blockchain of the x42 protocol will have a maximum total of forty-two million coins mined into existence by the year 2030. The coins follow the same rules as most cryptocurrencies, every transaction is final, timestamped and will be registered to the blockchain ledger, blockchain explorers can be used to browse any and all transactions Every project can have its own side blockchain, in which the development team can fully test and experiment on before going live to the store, because of that need all side blockchains have access to individual testnets. FM Interview with developer: The workpaper talks about using Stratis technology - Many say this is soon to be defunct? It does, the current WP is the 1st itteration, we have a 2.0 that will be released when the servers go online. However in Dec 2019, we have switched to Blockcore as Stratis had too many limitations that didn't meet our needs without heavy modifications. We did start with stratis as the basic framework, but we modified all of the code thus far to meet the needs we have. x42 is not stratis, but did start with some of their code. The switch to blockcore is really more of a code thing and not otherwise noticed on the user end. There has been some but not a significant amount of discussion about it. Anyone that does access the new beta will be able to notice the difference as it nolonger will say stratis anywhere. On the user end, some file structures may change with mention of blockcore. Blockcore is a fork of Stratis, however the cool thing is that it isnt its own project. More details: https://www.blockcore.net/ is an open source project started by a group of blockchain developers and engineers to realize a fully integrated platform for building custom blockchains. Blockcore is the foundation for realizing blockchains and includes core functionality to create your own custom blockchain with a lot of tooling supporting your blockchain. So its been 18 months of development? When will we say a mainnet? We are currently in testnet and close to 90% ready. Xservers should be released this year. However we are not working on a timeline for mainnet with dates at the moment. Deadlines produce inferior products. The new wallet (xCore) has passed and is ready, it will work alongside the xServer. The switch to blockcore did require some code changes and we are still working with blockcore to fine tune some issues in their code as well. Security and ease of use for the end user are required. What about use of dapps in beta? We have a few in progress, but there are not many as the parameters can be changing. Although, anyone that can code a dapp in c# could drop it into their app folder and have one right now. There just won't be the user interface on the current wallet or the server support yet. Why did you call it x42? X has always been the universal unknown in mathematics, which can also mean ‘anything.’ X is believed to have come from the arab letter shin (ش), that was initially used by spanish scholars because of their inability to translate certain arabic sounds into the spanish language, so it became a synonym for an unknown thing. From there on it was noticed that the spanish language didn’t have an appropriate sound for the arabic ‘sh.’ As time passed it was adapted into the ‘ck’ sound, which in classic greek is written down with a symbol known as chi (X). 42 is known to Douglas Adams’ The Hitchhiker’s Guide to the Galaxy readers as the ‘answer to the ultimate question of life, the universe, and everything’. We believe that x42 can offer a great answer to most questions plaguing the cryptocurrency space at the moment and in the future. That mix of an unknown and an answer is where our name got its origin. The techy Details: The two main levels of the x42 architecture are: xCore – This is responsible for interacting with the infrastructure and the interface layers in addition to the node policy layers. An x42 xCore full node handles all the APIs and user interfaces. NBitcoin – This is responsible for handling messages between nodes in the Network Layer. It also bridges the gap between various Consensus Layer functions. Higher Transaction Speed: Transaction on the x42 blockchain platform happens almost instantly under 60 seconds. Also, the main x42 blockchain supports around 70 transactions per second. This is nearly 10 times the transactions supported by the Bitcoin blockchain network. Additionally, the good thing is the x42 side blockchains can be edited to hold as many transactions-per-second (TPS) as the developers want. Understanding xCore, xServer, and Wallets xCore – An xCore node is basically a device with software that has the entire x42 blockchain saved and connects the main blockchain to the side blockchains. This node seeds the main and side blockchain to all users and also has all the functionalities of a client node. Client node is any client running a wallet connected to the x42 main blockchain network. As explained in the white-paper, “The xCore can stake coins on the main blockchain, it can also run decentralised applications hosted on xServers all across the network after having it installed on the local machine”. xServer – The x42 blockchain supports several different types of servers. xServer is a special sort of wallet that runs on dedicated hardware facilitates decentralisation of applications (DApps), processing and data storage. xServers are also tasked with other activities on the network like propagating smart contracts and signing transactions. It also allows the server owner (the Gatekeeper) to get paid for hosting diverse applications. Besides, xServers can choose to host the projects launched on any of the non-private side blockchains. Projects that share a considerable part of their revenues with the xServer owners are likely to get hosted quickly and maintained by the same servers for a longer period of time. Before we understand xServer wallets, let’s brief about the Hot Wallet and Cold Wallets. A hot wallet is an online wallet always connected to the internet. This is an instant point-of-contact for receiving and sending payments. The hot wallets receive the staking rewards along with payouts from DApps, side blockchain, smart contracts, and private transactions. A cold wallet is an offline wallet and more secure from the hot wallet as it is less susceptible to online attacks. This wallet holds the collateral for the server. All the server rewards are paid out to the cold storage wallet. Coming to the xServer wallets, these are basically software holding information for both hot and cold storage wallets. The xServer wallets can run on almost any operating system. They can also host and launch decentralised applications (DApps) and generate passive income by DApp hosting. Nice article.. https://medium.com/the-consensus/x42-protocol-not-just-another-pos-mn-project-d8dd73e8846d
A Bitcoin and Ethereum client is a piece of software that runs a node and enforces the ruleset of the protocol. ... With Eth 2 Testnet this number is far better with Prysmatic Labs client only ... The standard Bitcoin client, Bitcoin Core has three networks it can run on: mainnet, testnet, and regtest. This is the real Bitcoin network, with the block chain that everyone uses. You can of course experimentwith this network, but there are a few downsides: Its generally frowned upon to experiment on a production network. The Bitcoin Testnet Faucet is quick! The Bitcoin Testnet Faucet is bech32 ready, for modern clients! No ads! No BS! No tracking stuff! Best dev experience EVER! ... - Bitcoin - 30m interval mempool snapshots - KHS9NE - spruned bitcoin client - #spruned @ freenode Both the external port and the internal port should be 8333 for Bitcoin. (You may also want to map port 18333 for Bitcoin’s testnet, although this guide does not cover using testnet.) Make sure the IP address you enter is the same one you configured in the previous subsection. After filling in the details for the mapping, save the entry. The testnet is an alternative Bitcoin block chain, to be used for testing.Testnet coins are separate and distinct from actual bitcoins, and are never supposed to have any value. This allows application developers or bitcoin testers to experiment, without having to use real bitcoins or worrying about breaking the main bitcoin chain.
Cardano Testnet Stake Pools; Bitcoin at $1 Trillion; Vitalik says Fiat Won't End ... Litecoin Performance Pattern; Canada Mining Rigs for US Clients ... FINANCIALFOX : New Shelley Testnet ... Dash: QT Wallet TestNet Guide This tutorial will help you run the latest Dash-qt client on Test-net. This video is mean't for Linux but can be used for Windows/Mac as well. Bitcoin traders were waiting for the price to drop to the support at $9,500 but this was not to be. Instead of dropping further, the price stabilized at the $9,800 level and continued its upward ... Creating a Local Bitcoin Testnet / Regtest - Programming Bitcoin - Duration: ... How to Setup Testnet for the Ethereum blockchain client - Duration: 5:44. Abhishek Singh 1,096 views. 5:44. Ties.Network founders big interview about the project. Platform's CEO Alexander Neymark, CTO Dmitry Kochin and Business Developer Natalia Tokar talk about Ties.Network functions in details and why ...