How To Find Bitcoin Mining Cost? - Bitcoinik

HydroMiner

Bitcoin mining with hydro power. Using hydro power stations in the Alps region, we reach one of the lowest prices per kilo Watt in Europe. In fact, our cost of electricity is currently 85% lower than the average in Europe.
[link]

Bitcoin mentioned around Reddit: The cost of producing solar power is rapidly declining: it now costs $50 to produce one megawatt-hour of solar power. Coal, on the other hand, costs $102 per megawatt-hour to produce. This recent chan /r/Futurology

Bitcoin mentioned around Reddit: The cost of producing solar power is rapidly declining: it now costs $50 to produce one megawatt-hour of solar power. Coal, on the other hand, costs $102 per megawatt-hour to produce. This recent chan /Futurology submitted by HiIAMCaptainObvious to BitcoinAll [link] [comments]

Cost of blocking SegWit. At 3 ExaHash, 5% hash power, 15 MW. At $0.1 per kWh that's $1500 an hour. $13M, a year and to block forever /r/Bitcoin

Cost of blocking SegWit. At 3 ExaHash, 5% hash power, 15 MW. At $0.1 per kWh that's $1500 an hour. $13M, a year and to block forever /Bitcoin submitted by BitcoinAllBot to BitcoinAll [link] [comments]

[Guide] How to make money in EFT

EDIT : Thanks to everybody for pointing out the few mistakes/improvements that can be made in this new-player level guide.
For the sake of summarizing here :
- Intel documents are NOT worth 250k. I didn't check them on the flea before writing this and for some reason I always remembered them at 250k. Game is in maintenance so I can't check the real price. That being said, it's still profitable to craft USB into Intel, it's just not x2 profitable.
- Scav case : moonshine / intel docs, some people seem to say they've never been profitable. I personally *did not* measure those, I eyeballed it. I'm working on so much shit that I didn't bother. On average I think that I'm in a net positive, but it's as believable as people saying they're not : without proof we can't really say for sure. That bein said, it's certainly more profitable to run lower-tier scav runs that are *faster* when you're online, and to run a moonshine or intel when you log off. It's more efficient to get a lot of runs while you can re-start them every time.
- Crafting moonshine : It's not profitable to spam it ; I was under the assumption that the average player who will read this will usually not play for 4-5 hours straight and will end up collecting yesterday's moonshine, craft a new one, and that's it. If that's you're rythm then yes, spam it. If you intend to play more than one craft worth's of time, then you will craft moonshine faster than you can spend it, and it's not really worth to sell it on the flea except to up your market reputation for a small loss (about 10k). So in short : craft moonshine to be able to start a moonshine run for when you log off, but you don't *need* more than that.

Check this out

Here is some actual data on the lavatory !!

Hey everybody !

I know it can be a struggle to get a stable economy in this game, especially when you die a lot. Today I'm gonna try and give a few guidelines on how to make money safely, efficiently, fast, or in any other way we can think of.
If you're struggling to stay above the 15-20 million rouble treshold, this guide is definitely for you.
Very often I'll hear newer players say "Damn I can't seem to make money, I keep loosing. Every time I take gear I die instantly". There is some truth in that. Today I'll help you improve your survival rate, but most importantly I'll unbalance the other side of the equation. When you complain about losing a lot of money, I will help you spend less by a significant margin, as well as earn more. You'll also get rid of gear fera naturally.
Remember this throughout this very, very long read : It all depends on how you want to play, and how much. Some of these tips will not fit how you want to play the game, and like Nikita always says : this game is supposed to be fun before anything else.

1. Hideout

Safety Score : 100%
Reward : Moderate but very stable.
Maxing your hideout should be one of your top priorities, probably before telling your mom how much you love her every now and then. If you're not doing either of those, the big gamer in you knows what to do.
Early wipe, save your fuel for when you're online and playing. If you're playing, your generator should definitely be running and all your stations should be crafting something.
Once you have Medstation 1, Workbench 1 and Lavatory 2, you really have no reason to turn your generator off when you're playing.
Once you have the bitcoin farm, you should never turn off the generator.
Medstation :
Craft salewas and/or IFAKs permanently. They cost 8k and sell for 15k. That's a net profit of about 25k / hour for salewas, as well as never having to buy any.
Lavatory :
Always be crafting Bleach. If you have 2 empty blue fuel, use those empty cans to craft a Magazine case.
You can then keep the magazine cases until you've enough for your liking and sell those for a good profit.
The bleach you will use to buy the 6B47 helmets which are better than the SSh-68 helmets. Buying from 2x bleach barter at ragman level 1 means you get the helmet for 18k (instead of 33k on the market). This helmet has better head coverage, less slow/negative effects, less weight, has a slot for a mount, has +11 ergonomics AND is cheaper than the 22k SSh-68. That being said, it has a slight noise reduction that the Ssh does not have. If you wear headphones I'd say this is negligible but debatable. I prefer to have the extra protection and ergonomics for sure, considering it's slightly cheaper.

You can also barter for that helmet and instantly sell it back for a profit (five times) and level up ragman money requirements.
Bleach can also be traded for the Blackjack backpack at level 4, as well as the TTV rig at level 2. You should definitely do it.
Sell excess bleach on the flea market when the prices are around 10.5k or more. (around midnight Central European Time).
Workbench :
You can buy Power Cords and craft Wires forever and always make a profit. Buy in the morning and sell in the evening for better profits (CET timezone). For even more profit, you can craft gunpowders and ammo which tend to also be ridiculously pricy at night.
Buying grenades from Peacekeeper and crafting green (Eagle) gunpowder is a good way to make a lot of money and level up Peacekeeper.
Intel Center :
You main objective is to get this one to level 3 for reduced fees and better quest rewards, but also access to the bitcoin farm at level 2.
If you need FiR for quests, craft that. When you're done craft Intel Documents at all times (buy the USB), and use it for scav case or sell for a x2 profit. ( 3x40 for USB = 120, documents sell for 250)
Bitcoin Farm :
Once you have it, spend all your money on GPU until its maxxed, then level it up even more. The BTC farm is definitely worth it. At 50GPU you need to connect every 15 hours to clic. If you can't, keep it level 2 and connect every 24 hours to clic. Even at level 1 its worth. But its much, much faster at higher levels.
From 0 to 50 GPUs it takes about 30 days to pay for itself. GPUs should not be sold until you maxxed it.
Water Collector :
Must be running at all times. Buy the components if you don't have them.
Booze Generator :
Must be running at all times. Buy the components if you don't have them.
Scav Case :
Always have it running on moonshine, and use intel documents once you're done crafting one.
Nutrition Unit :
It's not really worth crafting sugar to put in the Booze gen, as the price for chocolate is pretty much = the price of sugar. So buy the sugar instead and craft something else. I tend to craft Hot Rods when the prices are good (morning) and then use them to barter 5.45 BS Ammo with Prapor or sell for a profit.

If you do all that, you should have about 150k an hour fairly easily. Don't forget to check it between every raid.

2. Traders

Safety Score : 100%
Reward : Quite good.
Once your mom has received all the love she deserves and your hideout is taken care of, you should have max traders (traders are a requirement for most of the hideout anyway).
Traders level 4 will net you much better prices on most mods and open very good barter trades.
Buy as much as you can from barter trades. You can buy almost everything from it, and it's usually at least 25% cheaper to buy the requirements and then do the barter. Ragman4 has the CPC Armored Rig which is level 5 armor, you'll get it for about 200k instead of 250k on the flea. The Slick is also much cheaper. The Blackjack backpack is literally half priced.
You can also NOT use what you barter and just sell it back to a dealer (sometimes the same from which you bartered) for a profit as well as having 2 times the loyalty money increase (from bartering then from selling).
Another good example is buying a Recbat 14k from the market, getting an ADAR for skier, selling it to Mechanic and winning 8k just like that. You can find every single barter that nets a profit yourself and just buy-resell and you'll probably make another 100k every reset, if you really are struggling and have the patience. I personally advise to just use the equipment for yourself unless you're levelling traders, but I wouldn't go as far as buying all profitable items every reset.
Every trader at every level has good barters. You can make a full decent kit at level 1 traders for about 40k roubles on barter, instead of 90 if you buy it all. (Paca for masks, helmet for bleach, ADAR for recbatt, salewa from craft, backpack, etc. all barters)

Bleach is beautiful and is coveted in the real world for its ability to cure diseases.

3. Modding

Safety Score : 100%
Reward : Very profitable.
Don't mod out of your reach. Don't mod Meta. If money is an issue for you, having +1 ergo won't change your life.
For example,
Priced at 10k roubles
Priced at 45k Roubles

See where I'm going with this?
If you have money, sure, go for the Shift. If you wanna have fun and try, sure, go for it as well. But if you're struggling, buy 4 cobras and mod 4 guns for the price of 1% recoil which will not make you a gamer god anyway.
Also, do NOT buy mods from the flea market when you see you can buy them from traders. Look at the top of the market, if the mod is greyed out, look at the price. It means you don't have access (yet). If the price is too inflated for you, find another mod. There are always other mods. You can make 2 AKMs that have a difference of 2% recoil and 4 Ergonomics and have a 150k price difference. It's up to you. When money is the issue, this was the answer.

Note : Some guns are inherently much more expensive. Guns shooting 5.56 or 5.45 tend to be more expensive than 7.62. AKMs are VERY good budget guns. They're a bit harder to handle, but you can get a fully modded AK for 150-200k, where as you will have an entry level M4 for that price. 7.62 PS ammo is also incredibly cheap while being decent. Play 7.62 if you're struggling with money. It's not meta, but it's far more than enough, trust me. You'll rarely lose fights exclusively because you had PS ammo in an AKM. Rarely.

4. Statistical loadout balance

This is fairly simple yet overlooked a LOT. To be accurate, you need data. Personally I kept it in an excel spreadsheet, if you're hardcore you should do something similar.

A somewhat relevant spreadsheet I used a wipe ago to measure some of my stats
What you need to know about yourself for this :
These will help us measure how much you fuck up or not.
Lets make it simple.
If you have a 500k loadout and you usually extract with 100k, at 10% survival rate, that means you will spend 500k x 10 = 5.000.000 roubles over 10 raids on average, die 9 times, and earn 100k once. This very obvious example shows the loss.
Basically we're gonna try and balance that equation so that you never lose money on average. You'll have ups and downs obviously, but over a week or two, it'll smooth things out for you, like math always does in a pleasant conversation with a girl.

So what can you do to improve that equation ?

4.1 Improve survival rate

Seems simple enough, DIE LESS. You do not need to be good, smart, or special to die less. If you die a lot, do something different. If you die less, try more of that. Explore statistical advantages through different gameplay.
What can you do to die less practically? Here is a list of checkboxes you can tick depending on your money, skill, mood, or any other factor like the map and sheer luck:
Do all that, it'll give you a LOT of data to actually improve by just doing something different without really being fastestronger, just smarter.
And I repeat : you can do some of it, all of it, it depends on what you like, what you're comfortable with, and the time/investment you're putting in the game. It's okay to play at your own pace.

4.2 Reduce gear cost

The second part of our "profit equation" above is how much gear you take with you. Using previous tips, reduce that cost. Barters, cheaper mods, etc.

4.3 Increase extracted value

This one is not as tricky as it sounds. Basically there are two ways to extract with more money in the backpack :
The goal is to pay for the gear you will loose when you die while making a profit on top. That one time you extract if you have a MBSS backpack, you'll need items worth like 50k per slot to break even. If you take a tri-zip, suddenly it's only 30k per slot. If you take a blackjack and blackrock from good old ragman, suddenly it's 10k per slot. So you can break even by looting crickents and DVD players almost.
See where I'm going ? Always take a tri-zip or bigger unless you're doing something special. That way you can afford to loot shitty areas, take less risk, and survive more while having a little less value.
We'll cover that in a minute, but there are ways to loot high value items, moderate value and low value. Those have also different risk/reward.

All of those are also map specific. In woods I'll often go with a 6B3TM armored rig for 40k, no helmet, 20k headphones and a sniper rifle. Rest is pouched so does not count. That's less than 100k investment. All players tend to have low value gear so I never extract with a lot either so it balances out. But on Woods, my survival rate is 20% instead of my overall 40%. So I know it's not a map I can reliably make money on, because I measured that accurately over time. This example is very common and should make sense to you.
Same goes for interchange where I have more about 50% survival but will tend to go in with 600k worth of gear, but will also often extract with over 500k quite regularly. Different ratios, different values, different purposes.
You can measure your own data if you're willing to do so, or you can eyeball it. Eyeballing it is much faster but very inaccurate because you will tend to include emotions in the mix when you die. You'll remember losses ~2x more than your wins (that's somewhat scientifically proven), and if you're eyeballing your loadout you might think you have 600k but really you might have only 450k. I would advise to go hardcore and measure it all for price, initial loadout, losses and earnings, for each map.

5. Money runs

Now money runs are vast and numerous. All include different levels of risk and reward. It's up to you once again to find what you're willing to do for the time it takes, the fun it will give you and how much it will actually help you. You can always try them all for ~50 raids the sake of trying something different and see how your data is impacted. it doesn't have to be 50 in a row if you don't want to. As long as you keep track of it it can be over a whole wipe. You'd have your data ready for the next wipe :) Faster is better though.

5.1 Hatchling runs

Safety Score : 100%
Reward : Very Variable. Mentally exhausting.
Those are incredibly money efficient. You're investing a gear of 0 value, so whatever you extract with is 100% win, so you cannot possibly lose money that way. Is it fun? Is it rewarding? I don't care, to each is own. Statistcally speaking, hatchling runs are an efficient way to make money.
They do however require a little bit of knowledge, but not skill. You'll be much more efficient at doing these kind of runs if you know where to go, what to look for, and how to get there depending on your spawn. That being said, such knoweldge is easily found ; it's nothing complex, it just takes time to learn. Once again, depends on how much you're willing to invest (if not roubles, time).

5.2 Scav runs

Safety Score : 100%
Reward : Low-ish
Scav runs are also incredibly efficient for the same reason as hatchlings. Except those have a cooldown. Statisticall speaking I have noticed you should always run your scavs as fast as possible on the map where you extract both the fastest and most frequently.
The explanation is simple, lets make it simpler :
The scav is a button that makes you earn free money. When you press it the button becomes unpressable for some time, when you release the button you earn money (sometimes).
That means you want to release the button as often as possible. And for that, you need to release it as fast as possible. It's that simple. So make scavs incredibly fast. I'm talking "Run through" fast.
Unless you're looking for FiR items or doing something specific like annoying a streamer, you should literally run straight to the extract every single time, and loot what you have that doesn't make you go out of your way too much. Usually I suggest factory, go in, kill a random scav, loot it, get out.
Two weapons is at LEAST 50k, 100 if they have a scope. There you go. That's 100k every 20 minutes (or less with intel center). That's MUCH BETTER than going up to 150-200k but taking 30 minutes to extract, and taking more risk by spending more time in the map. Every second you're in someone can shoot. Nobody can shoot you in the hideout.
The exception to that rule is Scavs with a pilgrim which you can take on your favourite loot-run map, probably interchange or reserve. There you should just fill everything you can and extract once you're full, no matter what you have. 30 crickents and an extra gun is fine.

5.3 Stash runs

Safety Score : Very
Reward : Okay
Those are very very safe and can be done with a pistol and a backpack only. Very cheap, quite unchalleneged, for a moderate reward. Just go on a map that you like and run around and loot all stashes until you're full, then get out. You can vary the map/route depending on the traffic of players. Interchange and shoreline are good contenders for that.
It'll net you easy money. Not great money, but definitely safe.

5.4 Loot Runs

Safety Score : Moderate
Reward : Quite alright
Once you have better knowledge/skill you can start having a specific route in a specific map, depending on a specific spawn. So it'll take time to learn. Usually very similar than a hatchling run except this time you bring moderate gear and go for moderate loots. For example, instead of going for fast techlight, in-and-out interchange, you can decide "alright I'll loot 100% of Oli and the computers in the back", it'll take time, but it'll make good loot. More money than stashes, definitely will see scavs to kill, and most probably some more pvp. More risk. If you win that PvP you have even more loot as well. But overall good reward.
Loot runs need to be "scheduled" and thought of after several tries, so you know how much you can take per person depending on backpack size. For example you can't say "lets loot oli" if you have a 5-man with blackjacks, you'll all be empty. Adapt.

5.4 PvP

Safety Score : Insane
Reward : Unreliably moderate
This one is pretty obvious. Very risky, unpredictable rewards. Usually better than loot runs when you survive. I won't elaborate on this, because if you're reading this far you're probably struggling in PvP. And the rest of this guide already covers a fair bit.

6. Insurance

Safety Score : "Meh"
Reward : Very profitable.
Now this is very, very important. Always insure your gear. Always.
If you die you will get stuff back, pretty much for free. If you're really struggling people won't loot your "trash", so you WILL get it back.
If you play in a group it's very likely that people will hide your stuff too.
And most importantly : you can insurance fraud. This is the best way to balance the equation we talked about earlier. If you find a decent-ish gun, replace yours. You drop your initial investment by a significant margin, you will definitely get it back, and if you extract it's a flat profit. Weapons don't take inventory slot, so if you have two weapons that are not yours initially they will usually pay for your whole gear. I have quite often left my super-mega-modded HK just for an average M4 or other weapon that I can fight with, just so I can reduce my investment by 350k and up my reward by like 200k instantly. Replace your headphones all the time too, that's an easy -30+30k, same with helmets. even if it's a bit broken or slightly worse.
If you're struggling with money, try to leave every raid with at least 3-4 pars of your equipment that aren't yours initially.
But value the risk behind this. I won't leave my slick for a Paca at the third minute of a raid just to have that extra 28k. I won't leave my meta-modded HK for a naked mosin. But if it seems decent/doable, do it. It will pay off. Because even if you die, you still get your shit back, and gun is usually the most expensive part of the gear.

7. Final notes

It's all about balance. Find what works *for you* and try shit out. Really, try. You'll die, you'll learn, you'll adapt with data to back that up. I find it crazy that people will die and not try to learn from it. That's how you will improve as a player.
First you gotta get smarter, then you'll get better. And with time, skill, mechanics, gamesense, all that will improve on the side. Earning more will snowball in your favour. And if you know you're statistically okay, you will have a much smaller gear fear and enjoy the game more.

Sorry for the wall of text, you guys should be used to it with me by now :D I made these guides in video but not in english, so here I am typing it all for you guys.
Enjoy :)
submitted by SixOneZil to EscapefromTarkov [link] [comments]

Proposal: The Sia Foundation

Vision Statement

A common sentiment is brewing online; a shared desire for the internet that might have been. After decades of corporate encroachment, you don't need to be a power user to realize that something has gone very wrong.
In the early days of the internet, the future was bright. In that future, when you sent an instant message, it traveled directly to the recipient. When you needed to pay a friend, you announced a transfer of value to their public key. When an app was missing a feature you wanted, you opened up the source code and implemented it. When you took a picture on your phone, it was immediately encrypted and backed up to storage that you controlled. In that future, people would laugh at the idea of having to authenticate themselves to some corporation before doing these things.
What did we get instead? Rather than a network of human-sized communities, we have a handful of enormous commons, each controlled by a faceless corporate entity. Hey user, want to send a message? You can, but we'll store a copy of it indefinitely, unencrypted, for our preference-learning algorithms to pore over; how else could we slap targeted ads on every piece of content you see? Want to pay a friend? You can—in our Monopoly money. Want a new feature? Submit a request to our Support Center and we'll totally maybe think about it. Want to backup a photo? You can—inside our walled garden, which only we (and the NSA, of course) can access. Just be careful what you share, because merely locking you out of your account and deleting all your data is far from the worst thing we could do.
You rationalize this: "MEGACORP would never do such a thing; it would be bad for business." But we all know, at some level, that this state of affairs, this inversion of power, is not merely "unfortunate" or "suboptimal" – No. It is degrading. Even if MEGACORP were purely benevolent, it is degrading that we must ask its permission to talk to our friends; that we must rely on it to safeguard our treasured memories; that our digital lives are completely beholden to those who seek only to extract value from us.
At the root of this issue is the centralization of data. MEGACORP can surveil you—because your emails and video chats flow through their servers. And MEGACORP can control you—because they hold your data hostage. But centralization is a solution to a technical problem: How can we make the user's data accessible from anywhere in the world, on any device? For a long time, no alternative solution to this problem was forthcoming.
Today, thanks to a confluence of established techniques and recent innovations, we have solved the accessibility problem without resorting to centralization. Hashing, encryption, and erasure encoding got us most of the way, but one barrier remained: incentives. How do you incentivize an anonymous stranger to store your data? Earlier protocols like BitTorrent worked around this limitation by relying on altruism, tit-for-tat requirements, or "points" – in other words, nothing you could pay your electric bill with. Finally, in 2009, a solution appeared: Bitcoin. Not long after, Sia was born.
Cryptography has unleashed the latent power of the internet by enabling interactions between mutually-distrustful parties. Sia harnesses this power to turn the cloud storage market into a proper marketplace, where buyers and sellers can transact directly, with no intermediaries, anywhere in the world. No more silos or walled gardens: your data is encrypted, so it can't be spied on, and it's stored on many servers, so no single entity can hold it hostage. Thanks to projects like Sia, the internet is being re-decentralized.
Sia began its life as a startup, which means it has always been subjected to two competing forces: the ideals of its founders, and the profit motive inherent to all businesses. Its founders have taken great pains to never compromise on the former, but this often threatened the company's financial viability. With the establishment of the Sia Foundation, this tension is resolved. The Foundation, freed of the obligation to generate profit, is a pure embodiment of the ideals from which Sia originally sprung.
The goals and responsibilities of the Foundation are numerous: to maintain core Sia protocols and consensus code; to support developers building on top of Sia and its protocols; to promote Sia and facilitate partnerships in other spheres and communities; to ensure that users can easily acquire and safely store siacoins; to develop network scalability solutions; to implement hardforks and lead the community through them; and much more. In a broader sense, its mission is to commoditize data storage, making it cheap, ubiquitous, and accessible to all, without compromising privacy or performance.
Sia is a perfect example of how we can achieve better living through cryptography. We now begin a new chapter in Sia's history. May our stewardship lead it into a bright future.
 

Overview

Today, we are proposing the creation of the Sia Foundation: a new non-profit entity that builds and supports distributed cloud storage infrastructure, with a specific focus on the Sia storage platform. What follows is an informal overview of the Sia Foundation, covering two major topics: how the Foundation will be funded, and what its funds will be used for.

Organizational Structure

The Sia Foundation will be structured as a non-profit entity incorporated in the United States, likely a 501(c)(3) organization or similar. The actions of the Foundation will be constrained by its charter, which formalizes the specific obligations and overall mission outlined in this document. The charter will be updated on an annual basis to reflect the current goals of the Sia community.
The organization will be operated by a board of directors, initially comprising Luke Champine as President and Eddie Wang as Chairman. Luke Champine will be leaving his position at Nebulous to work at the Foundation full-time, and will seek to divest his shares of Nebulous stock along with other potential conflicts of interest. Neither Luke nor Eddie personally own any siafunds or significant quantities of siacoin.

Funding

The primary source of funding for the Foundation will come from a new block subsidy. Following a hardfork, 30 KS per block will be allocated to the "Foundation Fund," continuing in perpetuity. The existing 30 KS per block miner reward is not affected. Additionally, one year's worth of block subsidies (approximately 1.57 GS) will be allocated to the Fund immediately upon activation of the hardfork.
As detailed below, the Foundation will provably burn any coins that it cannot meaningfully spend. As such, the 30 KS subsidy should be viewed as a maximum. This allows the Foundation to grow alongside Sia without requiring additional hardforks.
The Foundation will not be funded to any degree by the possession or sale of siafunds. Siafunds were originally introduced as a means of incentivizing growth, and we still believe in their effectiveness: a siafund holder wants to increase the amount of storage on Sia as much as possible. While the Foundation obviously wants Sia to succeed, its driving force should be its charter. Deriving significant revenue from siafunds would jeopardize the Foundation's impartiality and focus. Ultimately, we want the Foundation to act in the best interests of Sia, not in growing its own budget.

Responsibilities

The Foundation inherits a great number of responsibilities from Nebulous. Each quarter, the Foundation will publish the progress it has made over the past quarter, and list the responsibilities it intends to prioritize over the coming quarter. This will be accompanied by a financial report, detailing each area of expenditure over the past quarter, and forecasting expenditures for the coming quarter. Below, we summarize some of the myriad responsibilities towards which the Foundation is expected to allocate its resources.

Maintain and enhance core Sia software

Arguably, this is the most important responsibility of the Foundation. At the heart of Sia is its consensus algorithm: regardless of other differences, all Sia software must agree upon the content and rules of the blockchain. It is therefore crucial that the algorithm be stewarded by an entity that is accountable to the community, transparent in its decision-making, and has no profit motive or other conflicts of interest.
Accordingly, Sia’s consensus functionality will no longer be directly maintained by Nebulous. Instead, the Foundation will release and maintain an implementation of a "minimal Sia full node," comprising the Sia consensus algorithm and P2P networking code. The source code will be available in a public repository, and signed binaries will be published for each release.
Other parties may use this code to provide alternative full node software. For example, Nebulous may extend the minimal full node with wallet, renter, and host functionality. The source code of any such implementation may be submitted to the Foundation for review. If the code passes review, the Foundation will provide "endorsement signatures" for the commit hash used and for binaries compiled internally by the Foundation. Specifically, these signatures assert that the Foundation believes the software contains no consensus-breaking changes or other modifications to imported Foundation code. Endorsement signatures and Foundation-compiled binaries may be displayed and distributed by the receiving party, along with an appropriate disclaimer.
A minimal full node is not terribly useful on its own; the wallet, renter, host, and other extensions are what make Sia a proper developer platform. Currently, the only implementations of these extensions are maintained by Nebulous. The Foundation will contract Nebulous to ensure that these extensions continue to receive updates and enhancements. Later on, the Foundation intends to develop its own implementations of these extensions and others. As with the minimal node software, these extensions will be open source and available in public repositories for use by any Sia node software.
With the consensus code now managed by the Foundation, the task of implementing and orchestrating hardforks becomes its responsibility as well. When the Foundation determines that a hardfork is necessary (whether through internal discussion or via community petition), a formal proposal will be drafted and submitted for public review, during which arguments for and against the proposal may be submitted to a public repository. During this time, the hardfork code will be implemented, either by Foundation employees or by external contributors working closely with the Foundation. Once the implementation is finished, final arguments will be heard. The Foundation board will then vote whether to accept or reject the proposal, and announce their decision along with appropriate justification. Assuming the proposal was accepted, the Foundation will announce the block height at which the hardfork will activate, and will subsequently release source code and signed binaries that incorporate the hardfork code.
Regardless of the Foundation's decision, it is the community that ultimately determines whether a fork is accepted or rejected – nothing can change that. Foundation node software will never automatically update, so all forks must be explicitly adopted by users. Furthermore, the Foundation will provide replay and wipeout protection for its hard forks, protecting other chains from unintended or malicious reorgs. Similarly, the Foundation will ensure that any file contracts formed prior to a fork activation will continue to be honored on both chains until they expire.
Finally, the Foundation also intends to pursue scalability solutions for the Sia blockchain. In particular, work has already begun on an implementation of Utreexo, which will greatly reduce the space requirements of fully-validating nodes (allowing a full node to be run on a smartphone) while increasing throughput and decreasing initial sync time. A hardfork implementing Utreexo will be submitted to the community as per the process detailed above.
As this is the most important responsibility of the Foundation, it will receive a significant portion of the Foundation’s budget, primarily in the form of developer salaries and contracting agreements.

Support community services

We intend to allocate 25% of the Foundation Fund towards the community. This allocation will be held and disbursed in the form of siacoins, and will pay for grants, bounties, hackathons, and other community-driven endeavours.
Any community-run service, such as a Skynet portal, explorer or web wallet, may apply to have its costs covered by the Foundation. Upon approval, the Foundation will reimburse expenses incurred by the service, subject to the exact terms agreed to. The intent of these grants is not to provide a source of income, but rather to make such services "break even" for their operators, so that members of the community can enrich the Sia ecosystem without worrying about the impact on their own finances.

Ensure easy acquisition and storage of siacoins

Most users will acquire their siacoins via an exchange. The Foundation will provide support to Sia-compatible exchanges, and pursue relevant integrations at its discretion, such as Coinbase's new Rosetta standard. The Foundation may also release DEX software that enables trading cryptocurrencies without the need for a third party. (The Foundation itself will never operate as a money transmitter.)
Increasingly, users are storing their cryptocurrency on hardware wallets. The Foundation will maintain the existing Ledger Nano S integration, and pursue further integrations at its discretion.
Of course, all hardware wallets must be paired with software running on a computer or smartphone, so the Foundation will also develop and/or maintain client-side wallet software, including both full-node wallets and "lite" wallets. Community-operated wallet services, i.e. web wallets, may be funded via grants.
Like core software maintenance, this responsibility will be funded in the form of developer salaries and contracting agreements.

Protect the ecosystem

When it comes to cryptocurrency security, patching software vulnerabilities is table stakes; there are significant legal and social threats that we must be mindful of as well. As such, the Foundation will earmark a portion of its fund to defend the community from legal action. The Foundation will also safeguard the network from 51% attacks and other threats to network security by implementing softforks and/or hardforks where necessary.
The Foundation also intends to assist in the development of a new FOSS software license, and to solicit legal memos on various Sia-related matters, such as hosting in the United States and the EU.
In a broader sense, the establishment of the Foundation makes the ecosystem more robust by transferring core development to a more neutral entity. Thanks to its funding structure, the Foundation will be immune to various forms of pressure that for-profit companies are susceptible to.

Drive adoption of Sia

Although the overriding goal of the Foundation is to make Sia the best platform it can be, all that work will be in vain if no one uses the platform. There are a number of ways the Foundation can promote Sia and get it into the hands of potential users and developers.
In-person conferences are understandably far less popular now, but the Foundation can sponsor and/or participate in virtual conferences. (In-person conferences may be held in the future, permitting circumstances.) Similarly, the Foundation will provide prizes for hackathons, which may be organized by community members, Nebulous, or the Foundation itself. Lastly, partnerships with other companies in the cryptocurrency space—or the cloud storage space—are a great way to increase awareness of Sia. To handle these responsibilities, one of the early priorities of the Foundation will be to hire a marketing director.

Fund Management

The Foundation Fund will be controlled by a multisig address. Each member of the Foundation's board will control one of the signing keys, with the signature threshold to be determined once the final composition of the board is known. (This threshold may also be increased or decreased if the number of board members changes.) Additionally, one timelocked signing key will be controlled by David Vorick. This key will act as a “dead man’s switch,” to be used in the event of an emergency that prevents Foundation board members from reaching the signature threshold. The timelock ensures that this key cannot be used unless the Foundation fails to sign a transaction for several months.
On the 1st of each month, the Foundation will use its keys to transfer all siacoins in the Fund to two new addresses. The first address will be controlled by a high-security hot wallet, and will receive approximately one month's worth of Foundation expenditures. The second address, receiving the remaining siacoins, will be a modified version of the source address: specifically, it will increase the timelock on David Vorick's signing key by one month. Any other changes to the set of signing keys, such as the arrival or departure of board members, will be incorporated into this address as well.
The Foundation Fund is allocated in SC, but many of the Foundation's expenditures must be paid in USD or other fiat currency. Accordingly, the Foundation will convert, at its discretion, a portion of its monthly withdrawals to fiat currency. We expect this conversion to be primarily facilitated by private "OTC" sales to accredited investors. The Foundation currently has no plans to speculate in cryptocurrency or other assets.
Finally, it is important that the Foundation adds value to the Sia platform well in excess of the inflation introduced by the block subsidy. For this reason, the Foundation intends to provably burn, on a quarterly basis, any coins that it cannot allocate towards any justifiable expense. In other words, coins will be burned whenever doing so provides greater value to the platform than any other use. Furthermore, the Foundation will cap its SC treasury at 5% of the total supply, and will cap its USD treasury at 4 years’ worth of predicted expenses.
 
Addendum: Hardfork Timeline
We would like to see this proposal finalized and accepted by the community no later than September 30th. A new version of siad, implementing the hardfork, will be released no later than October 15th. The hardfork will activate at block 293220, which is expected to occur around 12pm EST on January 1st, 2021.
 
Addendum: Inflation specifics
The total supply of siacoins as of January 1st, 2021 will be approximately 45.243 GS. The initial subsidy of 1.57 GS thus increases the supply by 3.47%, and the total annual inflation in 2021 will be at most 10.4% (if zero coins are burned). In 2022, total annual inflation will be at most 6.28%, and will steadily decrease in subsequent years.
 

Conclusion

We see the establishment of the Foundation as an important step in the maturation of the Sia project. It provides the ecosystem with a sustainable source of funding that can be exclusively directed towards achieving Sia's ambitious goals. Compared to other projects with far deeper pockets, Sia has always punched above its weight; once we're on equal footing, there's no telling what we'll be able to achieve.
Nevertheless, we do not propose this change lightly, and have taken pains to ensure that the Foundation will act in accordance with the ideals that this community shares. It will operate transparently, keep inflation to a minimum, and respect the user's fundamental role in decentralized systems. We hope that everyone in the community will consider this proposal carefully, and look forward to a productive discussion.
submitted by lukechampine to siacoin [link] [comments]

Comparison between Avalanche, Cosmos and Polkadot

Comparison between Avalanche, Cosmos and Polkadot
Reposting after was mistakenly removed by mods (since resolved - Thanks)
A frequent question I see being asked is how Cosmos, Polkadot and Avalanche compare? Whilst there are similarities there are also a lot of differences. This article is not intended to be an extensive in-depth list, but rather an overview based on some of the criteria that I feel are most important.
For better formatting see https://medium.com/ava-hub/comparison-between-avalanche-cosmos-and-polkadot-a2a98f46c03b
https://preview.redd.it/e8s7dj3ivpq51.png?width=428&format=png&auto=webp&s=5d0463462702637118c7527ebf96e91f4a80b290

Overview

Cosmos

Cosmos is a heterogeneous network of many independent parallel blockchains, each powered by classical BFT consensus algorithms like Tendermint. Developers can easily build custom application specific blockchains, called Zones, through the Cosmos SDK framework. These Zones connect to Hubs, which are specifically designed to connect zones together.
The vision of Cosmos is to have thousands of Zones and Hubs that are Interoperable through the Inter-Blockchain Communication Protocol (IBC). Cosmos can also connect to other systems through peg zones, which are specifically designed zones that each are custom made to interact with another ecosystem such as Ethereum and Bitcoin. Cosmos does not use Sharding with each Zone and Hub being sovereign with their own validator set.
For a more in-depth look at Cosmos and provide more reference to points made in this article, please see my three part series — Part One, Part Two, Part Three
(There's a youtube video with a quick video overview of Cosmos on the medium article - https://medium.com/ava-hub/comparison-between-avalanche-cosmos-and-polkadot-a2a98f46c03b)

Polkadot

Polkadot is a heterogeneous blockchain protocol that connects multiple specialised blockchains into one unified network. It achieves scalability through a sharding infrastructure with multiple blockchains running in parallel, called parachains, that connect to a central chain called the Relay Chain. Developers can easily build custom application specific parachains through the Substrate development framework.
The relay chain validates the state transition of connected parachains, providing shared state across the entire ecosystem. If the Relay Chain must revert for any reason, then all of the parachains would also revert. This is to ensure that the validity of the entire system can persist, and no individual part is corruptible. The shared state makes it so that the trust assumptions when using parachains are only those of the Relay Chain validator set, and no other. Interoperability is enabled between parachains through Cross-Chain Message Passing (XCMP) protocol and is also possible to connect to other systems through bridges, which are specifically designed parachains or parathreads that each are custom made to interact with another ecosystem such as Ethereum and Bitcoin. The hope is to have 100 parachains connect to the relay chain.
For a more in-depth look at Polkadot and provide more reference to points made in this article, please see my three part series — Part One, Part Two, Part Three
(There's a youtube video with a quick video overview of Polkadot on the medium article - https://medium.com/ava-hub/comparison-between-avalanche-cosmos-and-polkadot-a2a98f46c03b)

Avalanche

Avalanche is a platform of platforms, ultimately consisting of thousands of subnets to form a heterogeneous interoperable network of many blockchains, that takes advantage of the revolutionary Avalanche Consensus protocols to provide a secure, globally distributed, interoperable and trustless framework offering unprecedented decentralisation whilst being able to comply with regulatory requirements.
Avalanche allows anyone to create their own tailor-made application specific blockchains, supporting multiple custom virtual machines such as EVM and WASM and written in popular languages like Go (with others coming in the future) rather than lightly used, poorly-understood languages like Solidity. This virtual machine can then be deployed on a custom blockchain network, called a subnet, which consist of a dynamic set of validators working together to achieve consensus on the state of a set of many blockchains where complex rulesets can be configured to meet regulatory compliance.
Avalanche was built with serving financial markets in mind. It has native support for easily creating and trading digital smart assets with complex custom rule sets that define how the asset is handled and traded to ensure regulatory compliance can be met. Interoperability is enabled between blockchains within a subnet as well as between subnets. Like Cosmos and Polkadot, Avalanche is also able to connect to other systems through bridges, through custom virtual machines made to interact with another ecosystem such as Ethereum and Bitcoin.
For a more in-depth look at Avalanche and provide more reference to points made in this article, please see here and here
(There's a youtube video with a quick video overview of Avalanche on the medium article - https://medium.com/ava-hub/comparison-between-avalanche-cosmos-and-polkadot-a2a98f46c03b)

Comparison between Cosmos, Polkadot and Avalanche

A frequent question I see being asked is how Cosmos, Polkadot and Avalanche compare? Whilst there are similarities there are also a lot of differences. This article is not intended to be an extensive in-depth list, but rather an overview based on some of the criteria that I feel are most important. For a more in-depth view I recommend reading the articles for each of the projects linked above and coming to your own conclusions. I want to stress that it’s not a case of one platform being the killer of all other platforms, far from it. There won’t be one platform to rule them all, and too often the tribalism has plagued this space. Blockchains are going to completely revolutionise most industries and have a profound effect on the world we know today. It’s still very early in this space with most adoption limited to speculation and trading mainly due to the limitations of Blockchain and current iteration of Ethereum, which all three of these platforms hope to address. For those who just want a quick summary see the image at the bottom of the article. With that said let’s have a look

Scalability

Cosmos

Each Zone and Hub in Cosmos is capable of up to around 1000 transactions per second with bandwidth being the bottleneck in consensus. Cosmos aims to have thousands of Zones and Hubs all connected through IBC. There is no limit on the number of Zones / Hubs that can be created

Polkadot

Parachains in Polkadot are also capable of up to around 1500 transactions per second. A portion of the parachain slots on the Relay Chain will be designated as part of the parathread pool, the performance of a parachain is split between many parathreads offering lower performance and compete amongst themselves in a per-block auction to have their transactions included in the next relay chain block. The number of parachains is limited by the number of validators on the relay chain, they hope to be able to achieve 100 parachains.

Avalanche

Avalanche is capable of around 4500 transactions per second per subnet, this is based on modest hardware requirements to ensure maximum decentralisation of just 2 CPU cores and 4 GB of Memory and with a validator size of over 2,000 nodes. Performance is CPU-bound and if higher performance is required then more specialised subnets can be created with higher minimum requirements to be able to achieve 10,000 tps+ in a subnet. Avalanche aims to have thousands of subnets (each with multiple virtual machines / blockchains) all interoperable with each other. There is no limit on the number of Subnets that can be created.

Results

All three platforms offer vastly superior performance to the likes of Bitcoin and Ethereum 1.0. Avalanche with its higher transactions per second, no limit on the number of subnets / blockchains that can be created and the consensus can scale to potentially millions of validators all participating in consensus scores ✅✅✅. Polkadot claims to offer more tps than cosmos, but is limited to the number of parachains (around 100) whereas with Cosmos there is no limit on the number of hubs / zones that can be created. Cosmos is limited to a fairly small validator size of around 200 before performance degrades whereas Polkadot hopes to be able to reach 1000 validators in the relay chain (albeit only a small number of validators are assigned to each parachain). Thus Cosmos and Polkadot scores ✅✅
https://preview.redd.it/2o0brllyvpq51.png?width=1000&format=png&auto=webp&s=8f62bb696ecaafcf6184da005d5fe0129d504518

Decentralisation

Cosmos

Tendermint consensus is limited to around 200 validators before performance starts to degrade. Whilst there is the Cosmos Hub it is one of many hubs in the network and there is no central hub or limit on the number of zones / hubs that can be created.

Polkadot

Polkadot has 1000 validators in the relay chain and these are split up into a small number that validate each parachain (minimum of 14). The relay chain is a central point of failure as all parachains connect to it and the number of parachains is limited depending on the number of validators (they hope to achieve 100 parachains). Due to the limited number of parachain slots available, significant sums of DOT will need to be purchased to win an auction to lease the slot for up to 24 months at a time. Thus likely to lead to only those with enough funds to secure a parachain slot. Parathreads are however an alternative for those that require less and more varied performance for those that can’t secure a parachain slot.

Avalanche

Avalanche consensus scan scale to tens of thousands of validators, even potentially millions of validators all participating in consensus through repeated sub-sampling. The more validators, the faster the network becomes as the load is split between them. There are modest hardware requirements so anyone can run a node and there is no limit on the number of subnets / virtual machines that can be created.

Results

Avalanche offers unparalleled decentralisation using its revolutionary consensus protocols that can scale to millions of validators all participating in consensus at the same time. There is no limit to the number of subnets and virtual machines that can be created, and they can be created by anyone for a small fee, it scores ✅✅✅. Cosmos is limited to 200 validators but no limit on the number of zones / hubs that can be created, which anyone can create and scores ✅✅. Polkadot hopes to accommodate 1000 validators in the relay chain (albeit these are split amongst each of the parachains). The number of parachains is limited and maybe cost prohibitive for many and the relay chain is a ultimately a single point of failure. Whilst definitely not saying it’s centralised and it is more decentralised than many others, just in comparison between the three, it scores ✅
https://preview.redd.it/ckfamee0wpq51.png?width=1000&format=png&auto=webp&s=c4355f145d821fabf7785e238dbc96a5f5ce2846

Latency

Cosmos

Tendermint consensus used in Cosmos reaches finality within 6 seconds. Cosmos consists of many Zones and Hubs that connect to each other. Communication between 2 zones could pass through many hubs along the way, thus also can contribute to latency times depending on the path taken as explained in part two of the articles on Cosmos. It doesn’t need to wait for an extended period of time with risk of rollbacks.

Polkadot

Polkadot provides a Hybrid consensus protocol consisting of Block producing protocol, BABE, and then a finality gadget called GRANDPA that works to agree on a chain, out of many possible forks, by following some simpler fork choice rule. Rather than voting on every block, instead it reaches agreements on chains. As soon as more than 2/3 of validators attest to a chain containing a certain block, all blocks leading up to that one are finalized at once.
If an invalid block is detected after it has been finalised then the relay chain would need to be reverted along with every parachain. This is particularly important when connecting to external blockchains as those don’t share the state of the relay chain and thus can’t be rolled back. The longer the time period, the more secure the network is, as there is more time for additional checks to be performed and reported but at the expense of finality. Finality is reached within 60 seconds between parachains but for external ecosystems like Ethereum their state obviously can’t be rolled back like a parachain and so finality will need to be much longer (60 minutes was suggested in the whitepaper) and discussed in more detail in part three

Avalanche

Avalanche consensus achieves finality within 3 seconds, with most happening sub 1 second, immutable and completely irreversible. Any subnet can connect directly to another without having to go through multiple hops and any VM can talk to another VM within the same subnet as well as external subnets. It doesn’t need to wait for an extended period of time with risk of rollbacks.

Results

With regards to performance far too much emphasis is just put on tps as a metric, the other equally important metric, if not more important with regards to finance is latency. Throughput measures the amount of data at any given time that it can handle whereas latency is the amount of time it takes to perform an action. It’s pointless saying you can process more transactions per second than VISA when it takes 60 seconds for a transaction to complete. Low latency also greatly increases general usability and customer satisfaction, nowadays everyone expects card payments, online payments to happen instantly. Avalanche achieves the best results scoring ✅✅✅, Cosmos with comes in second with 6 second finality ✅✅ and Polkadot with 60 second finality (which may be 60 minutes for external blockchains) scores ✅
https://preview.redd.it/kzup5x42wpq51.png?width=1000&format=png&auto=webp&s=320eb4c25dc4fc0f443a7a2f7ff09567871648cd

Shared Security

Cosmos

Every Zone and Hub in Cosmos has their own validator set and different trust assumptions. Cosmos are researching a shared security model where a Hub can validate the state of connected zones for a fee but not released yet. Once available this will make shared security optional rather than mandatory.

Polkadot

Shared Security is mandatory with Polkadot which uses a Shared State infrastructure between the Relay Chain and all of the connected parachains. If the Relay Chain must revert for any reason, then all of the parachains would also revert. Every parachain makes the same trust assumptions, and as such the relay chain validates state transition and enables seamless interoperability between them. In return for this benefit, they have to purchase DOT and win an auction for one of the available parachain slots.
However, parachains can’t just rely on the relay chain for their security, they will also need to implement censorship resistance measures and utilise proof of work / proof of stake for each parachain as well as discussed in part three, thus parachains can’t just rely on the security of the relay chain, they need to ensure sybil resistance mechanisms using POW and POS are implemented on the parachain as well.

Avalanche

A subnet in Avalanche consists of a dynamic set of validators working together to achieve consensus on the state of a set of many blockchains where complex rulesets can be configured to meet regulatory compliance. So unlike in Cosmos where each zone / hub has their own validators, A subnet can validate a single or many virtual machines / blockchains with a single validator set. Shared security is optional

Results

Shared security is mandatory in polkadot and a key design decision in its infrastructure. The relay chain validates the state transition of all connected parachains and thus scores ✅✅✅. Subnets in Avalanche can validate state of either a single or many virtual machines. Each subnet can have their own token and shares a validator set, where complex rulesets can be configured to meet regulatory compliance. It scores ✅ ✅. Every Zone and Hub in cosmos has their own validator set / token but research is underway to have the hub validate the state transition of connected zones, but as this is still early in the research phase scores ✅ for now.
https://preview.redd.it/pbgyk3o3wpq51.png?width=1000&format=png&auto=webp&s=61c18e12932a250f5633c40633810d0f64520575

Current Adoption

Cosmos

The Cosmos project started in 2016 with an ICO held in April 2017. There are currently around 50 projects building on the Cosmos SDK with a full list can be seen here and filtering for Cosmos SDK . Not all of the projects will necessarily connect using native cosmos sdk and IBC and some have forked parts of the Cosmos SDK and utilise the tendermint consensus such as Binance Chain but have said they will connect in the future.

Polkadot

The Polkadot project started in 2016 with an ICO held in October 2017. There are currently around 70 projects building on Substrate and a full list can be seen here and filtering for Substrate Based. Like with Cosmos not all projects built using substrate will necessarily connect to Polkadot and parachains or parathreads aren’t currently implemented in either the Live or Test network (Kusama) as of the time of this writing.

Avalanche

Avalanche in comparison started much later with Ava Labs being founded in 2018. Avalanche held it’s ICO in July 2020. Due to lot shorter time it has been in development, the number of projects confirmed are smaller with around 14 projects currently building on Avalanche. Due to the customisability of the platform though, many virtual machines can be used within a subnet making the process incredibly easy to port projects over. As an example, it will launch with the Ethereum Virtual Machine which enables byte for byte compatibility and all the tooling like Metamask, Truffle etc. will work, so projects can easily move over to benefit from the performance, decentralisation and low gas fees offered. In the future Cosmos and Substrate virtual machines could be implemented on Avalanche.

Results

Whilst it’s still early for all 3 projects (and the entire blockchain space as a whole), there is currently more projects confirmed to be building on Cosmos and Polkadot, mostly due to their longer time in development. Whilst Cosmos has fewer projects, zones are implemented compared to Polkadot which doesn’t currently have parachains. IBC to connect zones and hubs together is due to launch Q2 2021, thus both score ✅✅✅. Avalanche has been in development for a lot shorter time period, but is launching with an impressive feature set right from the start with ability to create subnets, VMs, assets, NFTs, permissioned and permissionless blockchains, cross chain atomic swaps within a subnet, smart contracts, bridge to Ethereum etc. Applications can easily port over from other platforms and use all the existing tooling such as Metamask / Truffle etc but benefit from the performance, decentralisation and low gas fees offered. Currently though just based on the number of projects in comparison it scores ✅.
https://preview.redd.it/4zpi6s85wpq51.png?width=1000&format=png&auto=webp&s=e91ade1a86a5d50f4976f3b23a46e9287b08e373

Enterprise Adoption

Cosmos

Cosmos enables permissioned and permissionless zones which can connect to each other with the ability to have full control over who validates the blockchain. For permissionless zones each zone / hub can have their own token and they are in control who validates.

Polkadot

With polkadot the state transition is performed by a small randomly selected assigned group of validators from the relay chain plus with the possibility that state is rolled back if an invalid transaction of any of the other parachains is found. This may pose a problem for enterprises that need complete control over who performs validation for regulatory reasons. In addition due to the limited number of parachain slots available Enterprises would have to acquire and lock up large amounts of a highly volatile asset (DOT) and have the possibility that they are outbid in future auctions and find they no longer can have their parachain validated and parathreads don’t provide the guaranteed performance requirements for the application to function.

Avalanche

Avalanche enables permissioned and permissionless subnets and complex rulesets can be configured to meet regulatory compliance. For example a subnet can be created where its mandatory that all validators are from a certain legal jurisdiction, or they hold a specific license and regulated by the SEC etc. Subnets are also able to scale to tens of thousands of validators, and even potentially millions of nodes, all participating in consensus so every enterprise can run their own node rather than only a small amount. Enterprises don’t have to hold large amounts of a highly volatile asset, but instead pay a fee in AVAX for the creation of the subnets and blockchains which is burnt.

Results

Avalanche provides the customisability to run private permissioned blockchains as well as permissionless where the enterprise is in control over who validates the blockchain, with the ability to use complex rulesets to meet regulatory compliance, thus scores ✅✅✅. Cosmos is also able to run permissioned and permissionless zones / hubs so enterprises have full control over who validates a blockchain and scores ✅✅. Polkadot requires locking up large amounts of a highly volatile asset with the possibility of being outbid by competitors and being unable to run the application if the guaranteed performance is required and having to migrate away. The relay chain validates the state transition and can roll back the parachain should an invalid block be detected on another parachain, thus scores ✅.
https://preview.redd.it/li5jy6u6wpq51.png?width=1000&format=png&auto=webp&s=e2a95f1f88e5efbcf9e23c789ae0f002c8eb73fc

Interoperability

Cosmos

Cosmos will connect Hubs and Zones together through its IBC protocol (due to release in Q1 2020). Connecting to blockchains outside of the Cosmos ecosystem would either require the connected blockchain to fork their code to implement IBC or more likely a custom “Peg Zone” will be created specific to work with a particular blockchain it’s trying to bridge to such as Ethereum etc. Each Zone and Hub has different trust levels and connectivity between 2 zones can have different trust depending on which path it takes (this is discussed more in this article). Finality time is low at 6 seconds, but depending on the number of hops, this can increase significantly.

Polkadot

Polkadot’s shared state means each parachain that connects shares the same trust assumptions, of the relay chain validators and that if one blockchain needs to be reverted, all of them will need to be reverted. Interoperability is enabled between parachains through Cross-Chain Message Passing (XCMP) protocol and is also possible to connect to other systems through bridges, which are specifically designed parachains or parathreads that each are custom made to interact with another ecosystem such as Ethereum and Bitcoin. Finality time between parachains is around 60 seconds, but longer will be needed (initial figures of 60 minutes in the whitepaper) for connecting to external blockchains. Thus limiting the appeal of connecting two external ecosystems together through Polkadot. Polkadot is also limited in the number of Parachain slots available, thus limiting the amount of blockchains that can be bridged. Parathreads could be used for lower performance bridges, but the speed of future blockchains is only going to increase.

Avalanche

A subnet can validate multiple virtual machines / blockchains and all blockchains within a subnet share the same trust assumptions / validator set, enabling cross chain interoperability. Interoperability is also possible between any other subnet, with the hope Avalanche will consist of thousands of subnets. Each subnet may have a different trust level, but as the primary network consists of all validators then this can be used as a source of trust if required. As Avalanche supports many virtual machines, bridges to other ecosystems are created by running the connected virtual machine. There will be an Ethereum bridge using the EVM shortly after mainnet. Finality time is much faster at sub 3 seconds (with most happening under 1 second) with no chance of rolling back so more appealing when connecting to external blockchains.

Results

All 3 systems are able to perform interoperability within their ecosystem and transfer assets as well as data, as well as use bridges to connect to external blockchains. Cosmos has different trust levels between its zones and hubs and can create issues depending on which path it takes and additional latency added. Polkadot provides the same trust assumptions for all connected parachains but has long finality and limited number of parachain slots available. Avalanche provides the same trust assumptions for all blockchains within a subnet, and different trust levels between subnets. However due to the primary network consisting of all validators it can be used for trust. Avalanche also has a much faster finality time with no limitation on the number of blockchains / subnets / bridges that can be created. Overall all three blockchains excel with interoperability within their ecosystem and each score ✅✅.
https://preview.redd.it/ai0bkbq8wpq51.png?width=1000&format=png&auto=webp&s=3e85ee6a3c4670f388ccea00b0c906c3fb51e415

Tokenomics

Cosmos

The ATOM token is the native token for the Cosmos Hub. It is commonly mistaken by people that think it’s the token used throughout the cosmos ecosystem, whereas it’s just used for one of many hubs in Cosmos, each with their own token. Currently ATOM has little utility as IBC isn’t released and has no connections to other zones / hubs. Once IBC is released zones may prefer to connect to a different hub instead and so ATOM is not used. ATOM isn’t a fixed capped supply token and supply will continuously increase with a yearly inflation of around 10% depending on the % staked. The current market cap for ATOM as of the time of this writing is $1 Billion with 203 million circulating supply. Rewards can be earnt through staking to offset the dilution caused by inflation. Delegators can also get slashed and lose a portion of their ATOM should the validator misbehave.

Polkadot

Polkadot’s native token is DOT and it’s used to secure the Relay Chain. Each parachain needs to acquire sufficient DOT to win an auction on an available parachain lease period of up to 24 months at a time. Parathreads have a fixed fee for registration that would realistically be much lower than the cost of acquiring a parachain slot and compete with other parathreads in a per-block auction to have their transactions included in the next relay chain block. DOT isn’t a fixed capped supply token and supply will continuously increase with a yearly inflation of around 10% depending on the % staked. The current market cap for DOT as of the time of this writing is $4.4 Billion with 852 million circulating supply. Delegators can also get slashed and lose their DOT (potentially 100% of their DOT for serious attacks) should the validator misbehave.

Avalanche

AVAX is the native token for the primary network in Avalanche. Every validator of any subnet also has to validate the primary network and stake a minimum of 2000 AVAX. There is no limit to the number of validators like other consensus methods then this can cater for tens of thousands even potentially millions of validators. As every validator validates the primary network, this can be a source of trust for interoperability between subnets as well as connecting to other ecosystems, thus increasing amount of transaction fees of AVAX. There is no slashing in Avalanche, so there is no risk to lose your AVAX when selecting a validator, instead rewards earnt for staking can be slashed should the validator misbehave. Because Avalanche doesn’t have direct slashing, it is technically possible for someone to both stake AND deliver tokens for something like a flash loan, under the invariant that all tokens that are staked are returned, thus being able to make profit with staked tokens outside of staking itself.
There will also be a separate subnet for Athereum which is a ‘spoon,’ or friendly fork, of Ethereum, which benefits from the Avalanche consensus protocol and applications in the Ethereum ecosystem. It’s native token ATH will be airdropped to ETH holders as well as potentially AVAX holders as well. This can be done for other blockchains as well.
Transaction fees on the primary network for all 3 of the blockchains as well as subscription fees for creating a subnet and blockchain are paid in AVAX and are burnt, creating deflationary pressure. AVAX is a fixed capped supply of 720 million tokens, creating scarcity rather than an unlimited supply which continuously increase of tokens at a compounded rate each year like others. Initially there will be 360 tokens minted at Mainnet with vesting periods between 1 and 10 years, with tokens gradually unlocking each quarter. The Circulating supply is 24.5 million AVAX with tokens gradually released each quater. The current market cap of AVAX is around $100 million.

Results

Avalanche’s AVAX with its fixed capped supply, deflationary pressure, very strong utility, potential to receive air drops and low market cap, means it scores ✅✅✅. Polkadot’s DOT also has very strong utility with the need for auctions to acquire parachain slots, but has no deflationary mechanisms, no fixed capped supply and already valued at $3.8 billion, therefore scores ✅✅. Cosmos’s ATOM token is only for the Cosmos Hub, of which there will be many hubs in the ecosystem and has very little utility currently. (this may improve once IBC is released and if Cosmos hub actually becomes the hub that people want to connect to and not something like Binance instead. There is no fixed capped supply and currently valued at $1.1 Billion, so scores ✅.
https://preview.redd.it/mels7myawpq51.png?width=1000&format=png&auto=webp&s=df9782e2c0a4c26b61e462746256bdf83b1fb906
All three are excellent projects and have similarities as well as many differences. Just to reiterate this article is not intended to be an extensive in-depth list, but rather an overview based on some of the criteria that I feel are most important. For a more in-depth view I recommend reading the articles for each of the projects linked above and coming to your own conclusions, you may have different criteria which is important to you, and score them differently. There won’t be one platform to rule them all however, with some uses cases better suited to one platform over another, and it’s not a zero-sum game. Blockchain is going to completely revolutionize industries and the Internet itself. The more projects researching and delivering breakthrough technology the better, each learning from each other and pushing each other to reach that goal earlier. The current market is a tiny speck of what’s in store in terms of value and adoption and it’s going to be exciting to watch it unfold.
https://preview.redd.it/dbb99egcwpq51.png?width=1388&format=png&auto=webp&s=aeb03127dc0dc74d0507328e899db1c7d7fc2879
For more information see the articles below (each with additional sources at the bottom of their articles)
Avalanche, a Revolutionary Consensus Engine and Platform. A Game Changer for Blockchain
Avalanche Consensus, The Biggest Breakthrough since Nakamoto
Cosmos — An Early In-Depth Analysis — Part One
Cosmos — An Early In-Depth Analysis — Part Two
Cosmos Hub ATOM Token and the commonly misunderstood staking tokens — Part Three
Polkadot — An Early In-Depth Analysis — Part One — Overview and Benefits
Polkadot — An Early In-Depth Analysis — Part Two — How Consensus Works
Polkadot — An Early In-Depth Analysis — Part Three — Limitations and Issues
submitted by xSeq22x to CryptoCurrency [link] [comments]

Bitcoin Farming, Basic Arithmetic, and you

I have written this guide to dispel a common misconception I hear from this community - that putting more than one Graphics Card in your Bitcoin Farm is a great idea.
TLDR: The FIRST graphics card you put in your bitcoin farm generates a bitcoin every 20 hours. Every additional graphics card you put in your bitcoin farm generates a bitcoin every 333.33 hours. This information is misstated on the wiki and in many videos I've seen.
More Complicated Maths TLDR from u/Mekhazzio :
TLDR: The bitcoin farm has a base production rate that's much higher than the rate added by each additional graphics cards. So when investing, you shouldn't be looking at how fast the whole farm pays itself off, but how much time it takes your N>1 graphics cards to each pay for themselves, because otherwise you could have just been pocketing the pure profit from the base production rate the whole time.
At current therapist/flea-FiR values:
That is to say, adding a GPU to an already-running farm takes three weeks before you've stopped losing money on that GPU.

A pretty simple formula is utilized to determine Bitcoin Farming output. The payback period for your first graphics card is around 3 days. For each additional graphics card that you put in the payback period is over 20 days. The reason that this has confused so many people is that they credit the production from Graphics Card 1 to the payback period for the rest of the Graphics Cards.
Caveat 1: Escape from Tarkov is a video game and, at least for us players, not a business. Many video game players are completionists, and I will not begrudge anyone who wants to max out every single part of their hideout because it will feel like an achievement. This guide discusses the impact of bitcoin farming on your PMC's wallet. If you find utility in maxing out the bitcoin farm for the feeling of completion then you should do it and probably just close this guide and not worry about it.
Caveat 2: This guide will not address people who hatchet run or pistol run to put graphics cards in their secure container that will usually end up being non-FIR. There are too many variables (spawn rate, survival rate, replacement value of just doing normal Tarkov raids instead of hatchet runs) to do a decent analysis. If you end up with non-FIR graphics cards you should put them in your Bitcoin Farm.
Analysis:
The formula for bitcoin generation is as follows:
Let's simplify some unnecessary constants and make this look more like a normal mathematical function. All we have to do is multiply (1/49) * (0.15) to get this, which is equivalent and much easier to understand:
Now, let's get some ground rules for investment:
Caveat 3: Prices may change, blah blah blah, unless the IRL bitcoin market crashes the conclusions from this guide will still be accurate for the most part.
I will also note that I'm not going to include the cost for fuel needed for production. Because you can craft expeditionary fuel into mag boxes, as well as do other crafts on your workbench and med station while you have the power on, this cost is negligible. Furthermore, since my thesis is that putting more graphics cards in is not worth it, the fact is that I can prove this mathematically without even accounting for the entire cost category of fuel only strengthens my argument.
Using these assumed prices, let's take a look at some different cases.
Case 1:
Building a Bitcoin Generator and putting a single graphics card in.
To calculate cost, we add the cost of building the empty generator (300k) to the single graphics card (250k) to get 550k rouble investment.
Lets calculate revenue using our formula before:BTC Generated per Hour = 0.05 + 0.003 * (Graphics Cards - 1)BTC Generated per Hour = 0.05 + 0.003 * (1 Graphics Cards - 1)BTC Generated per Hour = 0.05 + 0.003 * (0)BTC Generated per Hour = 0.05
So we're generating 5% of a bitcoin every hour which means we'll get a bitcoin from our farm every 20 hours.
So, every 20 hours we are generating a product worth ~150k. Since we invested ~550k we need to sell:
550k investment / 150k roubles per bitcoin = 3.66 physical bitcoins in order to recoup our investment
Since we can't harvest bitcoins until they are full, we actually need to wait until we get 4 bitcoins at which point we'll be making a slight profit. Generating 4 bitcoins will take 4 bitcoins * 20 hours per bitcoin = 80 hours or a little more than 3 days.
Case 2:
Adding a second graphics card to our bitcoin farm.
Now, as discussed above I'm not worried about non-FIR graphics cards that you hatchet ran to find. If you have an FIR graphics card then you can sell it on the flea market for the 250k price that I'm using as an assumption above.
This concept is called opportunity cost and if you don't understand it I will troll you in the comments: Putting an FIR graphics card into your bitcoin farm is the same as purchasing one off of the flea market and putting it in your bitcoin farm because you had the opportunity to just sell your FIR graphics card for the same price that you can buy it.
With that out of the way, let's do some math on our 2 graphics card bitcoin farm:
BTC Generated per Hour = 0.05 + 0.003 * (Graphics Cards - 1)BTC Generated per Hour = 0.05 + 0.003 * (2 Graphics Cards - 1)BTC Generated per Hour = 0.05 + 0.003 * 1BTC Generated per Hour = 0.053
So, for the cost of 250k roubles we have increased our bitcoin per hour generation by 0.003.
The first graphics card that we added to our bitcoin farm generates us one bitcoin every 20 hours, as discussed above.
The second graphics card that we added to our bitcoin farm generates 0.003 bitcoins per hour. To calculate how many hours this takes to get 1 bitcoin we do the math of 1 / 0.003 = 333.33 hours. 333.33 hours / 24 hours per day is 13.88 or roughly 14 days.
In order to recoup our investment from the 250k roubles we used to get our second graphics card we divide 250k roubles invested by 150k roubles per bitcoin = 1.66 bitcoins. We generate one bitcoin every 14 days, so we can multiply 14 days * 1.66 bitcoins = 23 days.
This math will hold true for every additional graphics card because the function is linear.
Thus, the payback period for your 250k investment in adding a graphics card past the first one to your bitcoin farm is 23 days.
To reiterate: The FIRST graphics card you put in your bitcoin farm generates a bitcoin every 20 hours. Every additional graphics card you put in your bitcoin farm generates a bitcoin every 333.33 hours.
submitted by Death4Chairman20x70 to EscapefromTarkov [link] [comments]

Minimum Viable Issuance - Why Ethereum’s lack of a hard cap on ETH issuance is a good thing.

This post will explain how the argument used by the average Bitcoin maximalist, thinking that they have found Ethereum’s achilles heel when talking about issuance is actually highlighting one of Ethereum’s strong points and one of the main threats to the longevity of the Bitcoin network.
So first let’s answer the question which I know many people have about Ethereum:

What is Ethereum’s ETH issuance schedule?

Ethereum has an issuance policy of Minimum Viable Issuance. So what does this mean exactly? It means that the issuance of ETH will be as low as possible while also maintaining a sufficient budget to pay miners (and soon to be stakers) to keep the network secure. For example, if ETH issuance was halved, miners would drop off the network and stop mining as it is no longer profitable for them to mine. As a result, the network would be less secure as it would cost less money for an attacker to control 51% of the hash power and attack the network. This means that the Ethereum community plans to change ETH issuance as time goes on to maintain a reasonable security budget which will keep the network secure but will also keep inflation in check. We have done this twice in the past with EIP-649 and EIP-1234 which reduced block rewards from 5 ETH per block to 3 ETH and from 3 ETH to 2 ETH respectively. I previously made a graph of ETH issuance over time here: https://redd.it/it8ce7
So while Ethereum doesn’t have a strictly defined issuance schedule, the community will reject any proposals which either put the security of the network at risk such as the recent EIP-2878, or we will reject proposals which will lead to excessive network security and therefore an unnecessarily high inflation rate (or we will accept proposals which reduce issuance after price rises and therefore the security budget rises). This means that when Bitcoiners accuse the Ethereum Foundation of being no better than a central bank because they can “print more Ether”, this is completely untrue. Any proposals made by the EF which would increase issuance unnecessarily would be rejected by the community in the same way that a proposal to increase the supply of Bitcoin from 21 million to 22 million would be rejected. There is a social contract around both Bitcoin’s and Ethereum’s issuance schedules. Any networks or proposals which break the social contracts of 21 million Bitcoins and minimal viable issuance of Ether would be a breach of these contracts and the new proposed network would be labeled by the community as illegitimate and the original network would live on.

So why is minimum viable issuance better than a hard cap?

Minimum viable issuance is better than a hard cap because it puts the most important part of the network first - the security. MVI ensures that the Ethereum network will always have a security budget which keeps the cost of a 51% attack impractically high. Bitcoin on the other hand, halves its security budget every 4 years until eventually only the transaction fees pay for network security. This means that every 4 years, the amount of money paying for network security halves until eventually, the value of attacking the network becomes greater than the security budget and someone performs a 51% attack (technically the security budget only halves if terms of BTC not in dollars. However, even if the price of Bitcoin more than doubles in the time that the security budget halves, the ratio of security budget to value secured on the network still halves, doubling the financial viability of performing a network attack). The strategy to pay for the security budget once Bitcoin issuance stops is for transaction fees to secure the network since transaction fees are paid to miners. Not only does this have its own security problems which I won’t detail here, but unless Bitcoin scales on layer 1 (layer 2 scaling solutions have their own security mechanisms separate from L1), then fees would have to cost well in the thousands of dollars to secure a trillion dollar market cap Bitcoin that is secured by nothing but fees. If Bitcoin maximalists want a 10 trillion or 100 trillion dollar market cap then expect fees to go up another 10 or 100 times from there.
Ethereum on the other hand, will be able to keep its network secure with approximately 1-2% annual issuance being paid to stakers under ETH 2.0. This is because not all of the network will be staking, so if 33 million of the approximately 110 million Ether in existence stakes under ETH 2.0, then paying this 33 million Ether 6% a year (a very decent yield!) would cost just under 2 million ETH per year which would equate to less than 2% annual ETH inflation. This is also before considering EIP-1559 which will burn a portion of transaction fees which will counter the effect of this inflation and potentially even make ETH deflationary if the sum of all burned transaction fees are greater than the annual inflation. Also, under ETH 2.0, an attacker performing a 51% attack would get his funds slashed (they would lose their funds) if they attack the network, meaning that they can only perform a 51% attack once. However, in Bitcoin, anyone who controls 51% of the mining hash power could perform multiple 51% attacks without losing everything like they could in ETH 2.0.
So in conclusion, while Ethereum doesn’t have the guaranteed anti-inflation security of a hard cap, it does have the guarantee of always paying it’s miners (or stakers under ETH 2.0) enough to keep the network secure. In contrast, while Bitcoin’s social contract may guarantee a hard cap of 21 million, it cannot simultaneously guarantee network security in the long run. Eventually, its users will have to decide if they want a secure network with more than 21 million coins or a tax to pay for security or an insecure network with super high fees and a hard cap of 21 million Bitcoin.
Disclaimer: The details I covered around 51% attacks and network security are simplified. I am not an expert in this field and things are a lot more nuanced than I laid out in my simplifications above.
submitted by Tricky_Troll to ethfinance [link] [comments]

Dragonchain Great Reddit Scaling Bake-Off Public Proposal

Dragonchain Great Reddit Scaling Bake-Off Public Proposal

Dragonchain Public Proposal TL;DR:

Dragonchain has demonstrated twice Reddit’s entire total daily volume (votes, comments, and posts per Reddit 2019 Year in Review) in a 24-hour demo on an operational network. Every single transaction on Dragonchain is decentralized immediately through 5 levels of Dragon Net, and then secured with combined proof on Bitcoin, Ethereum, Ethereum Classic, and Binance Chain, via Interchain. At the time, in January 2020, the entire cost of the demo was approximately $25K on a single system (transaction fees locked at $0.0001/txn). With current fees (lowest fee $0.0000025/txn), this would cost as little as $625.
Watch Joe walk through the entire proposal and answer questions on YouTube.
This proposal is also available on the Dragonchain blog.

Hello Reddit and Ethereum community!

I’m Joe Roets, Founder & CEO of Dragonchain. When the team and I first heard about The Great Reddit Scaling Bake-Off we were intrigued. We believe we have the solutions Reddit seeks for its community points system and we have them at scale.
For your consideration, we have submitted our proposal below. The team at Dragonchain and I welcome and look forward to your technical questions, philosophical feedback, and fair criticism, to build a scaling solution for Reddit that will empower its users. Because our architecture is unlike other blockchain platforms out there today, we expect to receive many questions while people try to grasp our project. I will answer all questions here in this thread on Reddit, and I've answered some questions in the stream on YouTube.
We have seen good discussions so far in the competition. We hope that Reddit’s scaling solution will emerge from The Great Reddit Scaling Bake-Off and that Reddit will have great success with the implementation.

Executive summary

Dragonchain is a robust open source hybrid blockchain platform that has proven to withstand the passing of time since our inception in 2014. We have continued to evolve to harness the scalability of private nodes, yet take full advantage of the security of public decentralized networks, like Ethereum. We have a live, operational, and fully functional Interchain network integrating Bitcoin, Ethereum, Ethereum Classic, and ~700 independent Dragonchain nodes. Every transaction is secured to Ethereum, Bitcoin, and Ethereum Classic. Transactions are immediately usable on chain, and the first decentralization is seen within 20 seconds on Dragon Net. Security increases further to public networks ETH, BTC, and ETC within 10 minutes to 2 hours. Smart contracts can be written in any executable language, offering full freedom to existing developers. We invite any developer to watch the demo, play with our SDK’s, review open source code, and to help us move forward. Dragonchain specializes in scalable loyalty & rewards solutions and has built a decentralized social network on chain, with very affordable transaction costs. This experience can be combined with the insights Reddit and the Ethereum community have gained in the past couple of months to roll out the solution at a rapid pace.

Response and PoC

In The Great Reddit Scaling Bake-Off post, Reddit has asked for a series of demonstrations, requirements, and other considerations. In this section, we will attempt to answer all of these requests.

Live Demo

A live proof of concept showing hundreds of thousands of transactions
On Jan 7, 2020, Dragonchain hosted a 24-hour live demonstration during which a quarter of a billion (250 million+) transactions executed fully on an operational network. Every single transaction on Dragonchain is decentralized immediately through 5 levels of Dragon Net, and then secured with combined proof on Bitcoin, Ethereum, Ethereum Classic, and Binance Chain, via Interchain. This means that every single transaction is secured by, and traceable to these networks. An attack on this system would require a simultaneous attack on all of the Interchained networks.
24 hours in 4 minutes (YouTube):
24 hours in 4 minutes
The demonstration was of a single business system, and any user is able to scale this further, by running multiple systems simultaneously. Our goals for the event were to demonstrate a consistent capacity greater than that of Visa over an extended time period.
Tooling to reproduce our demo is available here:
https://github.com/dragonchain/spirit-bomb

Source Code

Source code (for on & off-chain components as well tooling used for the PoC). The source code does not have to be shared publicly, but if Reddit decides to use a particular solution it will need to be shared with Reddit at some point.

Scaling

How it works & scales

Architectural Scaling

Dragonchain’s architecture attacks the scalability issue from multiple angles. Dragonchain is a hybrid blockchain platform, wherein every transaction is protected on a business node to the requirements of that business or purpose. A business node may be held completely private or may be exposed or replicated to any level of exposure desired.
Every node has its own blockchain and is independently scalable. Dragonchain established Context Based Verification as its consensus model. Every transaction is immediately usable on a trust basis, and in time is provable to an increasing level of decentralized consensus. A transaction will have a level of decentralization to independently owned and deployed Dragonchain nodes (~700 nodes) within seconds, and full decentralization to BTC and ETH within minutes or hours. Level 5 nodes (Interchain nodes) function to secure all transactions to public or otherwise external chains such as Bitcoin and Ethereum. These nodes scale the system by aggregating multiple blocks into a single Interchain transaction on a cadence. This timing is configurable based upon average fees for each respective chain. For detailed information about Dragonchain’s architecture, and Context Based Verification, please refer to the Dragonchain Architecture Document.

Economic Scaling

An interesting feature of Dragonchain’s network consensus is its economics and scarcity model. Since Dragon Net nodes (L2-L4) are independent staking nodes, deployment to cloud platforms would allow any of these nodes to scale to take on a large percentage of the verification work. This is great for scalability, but not good for the economy, because there is no scarcity, and pricing would develop a downward spiral and result in fewer verification nodes. For this reason, Dragonchain uses TIME as scarcity.
TIME is calculated as the number of Dragons held, multiplied by the number of days held. TIME influences the user’s access to features within the Dragonchain ecosystem. It takes into account both the Dragon balance and length of time each Dragon is held. TIME is staked by users against every verification node and dictates how much of the transaction fees are awarded to each participating node for every block.
TIME also dictates the transaction fee itself for the business node. TIME is staked against a business node to set a deterministic transaction fee level (see transaction fee table below in Cost section). This is very interesting in a discussion about scaling because it guarantees independence for business implementation. No matter how much traffic appears on the entire network, a business is guaranteed to not see an increased transaction fee rate.

Scaled Deployment

Dragonchain uses Docker and Kubernetes to allow the use of best practices traditional system scaling. Dragonchain offers managed nodes with an easy to use web based console interface. The user may also deploy a Dragonchain node within their own datacenter or favorite cloud platform. Users have deployed Dragonchain nodes on-prem on Amazon AWS, Google Cloud, MS Azure, and other hosting platforms around the world. Any executable code, anything you can write, can be written into a smart contract. This flexibility is what allows us to say that developers with no blockchain experience can use any code language to access the benefits of blockchain. Customers have used NodeJS, Python, Java, and even BASH shell script to write smart contracts on Dragonchain.
With Docker containers, we achieve better separation of concerns, faster deployment, higher reliability, and lower response times.
We chose Kubernetes for its self-healing features, ability to run multiple services on one server, and its large and thriving development community. It is resilient, scalable, and automated. OpenFaaS allows us to package smart contracts as Docker images for easy deployment.
Contract deployment time is now bounded only by the size of the Docker image being deployed but remains fast even for reasonably large images. We also take advantage of Docker’s flexibility and its ability to support any language that can run on x86 architecture. Any image, public or private, can be run as a smart contract using Dragonchain.

Flexibility in Scaling

Dragonchain’s architecture considers interoperability and integration as key features. From inception, we had a goal to increase adoption via integration with real business use cases and traditional systems.
We envision the ability for Reddit, in the future, to be able to integrate alternate content storage platforms or other financial services along with the token.
  • LBRY - To allow users to deploy content natively to LBRY
  • MakerDAO to allow users to lend small amounts backed by their Reddit community points.
  • STORJ/SIA to allow decentralized on chain storage of portions of content. These integrations or any other are relatively easy to integrate on Dragonchain with an Interchain implementation.

Cost

Cost estimates (on-chain and off-chain) For the purpose of this proposal, we assume that all transactions are on chain (posts, replies, and votes).
On the Dragonchain network, transaction costs are deterministic/predictable. By staking TIME on the business node (as described above) Reddit can reduce transaction costs to as low as $0.0000025 per transaction.
Dragonchain Fees Table

Getting Started

How to run it
Building on Dragonchain is simple and requires no blockchain experience. Spin up a business node (L1) in our managed environment (AWS), run it in your own cloud environment, or on-prem in your own datacenter. Clear documentation will walk you through the steps of spinning up your first Dragonchain Level 1 Business node.
Getting started is easy...
  1. Download Dragonchain’s dctl
  2. Input three commands into a terminal
  3. Build an image
  4. Run it
More information can be found in our Get started documents.

Architecture
Dragonchain is an open source hybrid platform. Through Dragon Net, each chain combines the power of a public blockchain (like Ethereum) with the privacy of a private blockchain.
Dragonchain organizes its network into five separate levels. A Level 1, or business node, is a totally private blockchain only accessible through the use of public/private keypairs. All business logic, including smart contracts, can be executed on this node directly and added to the chain.
After creating a block, the Level 1 business node broadcasts a version stripped of sensitive private data to Dragon Net. Three Level 2 Validating nodes validate the transaction based on guidelines determined from the business. A Level 3 Diversity node checks that the level 2 nodes are from a diverse array of locations. A Level 4 Notary node, hosted by a KYC partner, then signs the validation record received from the Level 3 node. The transaction hash is ledgered to the Level 5 public chain to take advantage of the hash power of massive public networks.
Dragon Net can be thought of as a “blockchain of blockchains”, where every level is a complete private blockchain. Because an L1 can send to multiple nodes on a single level, proof of existence is distributed among many places in the network. Eventually, proof of existence reaches level 5 and is published on a public network.

API Documentation

APIs (on chain & off)

SDK Source

Nobody’s Perfect

Known issues or tradeoffs
  • Dragonchain is open source and even though the platform is easy enough for developers to code in any language they are comfortable with, we do not have so large a developer community as Ethereum. We would like to see the Ethereum developer community (and any other communities) become familiar with our SDK’s, our solutions, and our platform, to unlock the full potential of our Ethereum Interchain. Long ago we decided to prioritize both Bitcoin and Ethereum Interchains. We envision an ecosystem that encompasses different projects to give developers the ability to take full advantage of all the opportunities blockchain offers to create decentralized solutions not only for Reddit but for all of our current platforms and systems. We believe that together we will take the adoption of blockchain further. We currently have additional Interchain with Ethereum Classic. We look forward to Interchain with other blockchains in the future. We invite all blockchains projects who believe in decentralization and security to Interchain with Dragonchain.
  • While we only have 700 nodes compared to 8,000 Ethereum and 10,000 Bitcoin nodes. We harness those 18,000 nodes to scale to extremely high levels of security. See Dragonchain metrics.
  • Some may consider the centralization of Dragonchain’s business nodes as an issue at first glance, however, the model is by design to protect business data. We do not consider this a drawback as these nodes can make any, none, or all data public. Depending upon the implementation, every subreddit could have control of its own business node, for potential business and enterprise offerings, bringing new alternative revenue streams to Reddit.

Costs and resources

Summary of cost & resource information for both on-chain & off-chain components used in the PoC, as well as cost & resource estimates for further scaling. If your PoC is not on mainnet, make note of any mainnet caveats (such as congestion issues).
Every transaction on the PoC system had a transaction fee of $0.0001 (one-hundredth of a cent USD). At 256MM transactions, the demo cost $25,600. With current operational fees, the same demonstration would cost $640 USD.
For the demonstration, to achieve throughput to mimic a worldwide payments network, we modeled several clients in AWS and 4-5 business nodes to handle the traffic. The business nodes were tuned to handle higher throughput by adjusting memory and machine footprint on AWS. This flexibility is valuable to implementing a system such as envisioned by Reddit. Given that Reddit’s daily traffic (posts, replies, and votes) is less than half that of our demo, we would expect that the entire Reddit system could be handled on 2-5 business nodes using right-sized containers on AWS or similar environments.
Verification was accomplished on the operational Dragon Net network with over 700 independently owned verification nodes running around the world at no cost to the business other than paid transaction fees.

Requirements

Scaling

This PoC should scale to the numbers below with minimal costs (both on & off-chain). There should also be a clear path to supporting hundreds of millions of users.
Over a 5 day period, your scaling PoC should be able to handle:
*100,000 point claims (minting & distributing points) *25,000 subscriptions *75,000 one-off points burning *100,000 transfers
During Dragonchain’s 24 hour demo, the above required numbers were reached within the first few minutes.
Reddit’s total activity is 9000% more than Ethereum’s total transaction level. Even if you do not include votes, it is still 700% more than Ethereum’s current volume. Dragonchain has demonstrated that it can handle 250 million transactions a day, and it’s architecture allows for multiple systems to work at that level simultaneously. In our PoC, we demonstrate double the full capacity of Reddit, and every transaction was proven all the way to Bitcoin and Ethereum.
Reddit Scaling on Ethereum

Decentralization

Solutions should not depend on any single third-party provider. We prefer solutions that do not depend on specific entities such as Reddit or another provider, and solutions with no single point of control or failure in off-chain components but recognize there are numerous trade-offs to consider
Dragonchain’s architecture calls for a hybrid approach. Private business nodes hold the sensitive data while the validation and verification of transactions for the business are decentralized within seconds and secured to public blockchains within 10 minutes to 2 hours. Nodes could potentially be controlled by owners of individual subreddits for more organic decentralization.
  • Billing is currently centralized - there is a path to federation and decentralization of a scaled billing solution.
  • Operational multi-cloud
  • Operational on-premises capabilities
  • Operational deployment to any datacenter
  • Over 700 independent Community Verification Nodes with proof of ownership
  • Operational Interchain (Interoperable to Bitcoin, Ethereum, and Ethereum Classic, open to more)

Usability Scaling solutions should have a simple end user experience.

Users shouldn't have to maintain any extra state/proofs, regularly monitor activity, keep track of extra keys, or sign anything other than their normal transactions
Dragonchain and its customers have demonstrated extraordinary usability as a feature in many applications, where users do not need to know that the system is backed by a live blockchain. Lyceum is one of these examples, where the progress of academy courses is being tracked, and successful completion of courses is rewarded with certificates on chain. Our @Save_The_Tweet bot is popular on Twitter. When used with one of the following hashtags - #please, #blockchain, #ThankYou, or #eternalize the tweet is saved through Eternal to multiple blockchains. A proof report is available for future reference. Other examples in use are DEN, our decentralized social media platform, and our console, where users can track their node rewards, view their TIME, and operate a business node.
Examples:

Transactions complete in a reasonable amount of time (seconds or minutes, not hours or days)
All transactions are immediately usable on chain by the system. A transaction begins the path to decentralization at the conclusion of a 5-second block when it gets distributed across 5 separate community run nodes. Full decentralization occurs within 10 minutes to 2 hours depending on which interchain (Bitcoin, Ethereum, or Ethereum Classic) the transaction hits first. Within approximately 2 hours, the combined hash power of all interchained blockchains secures the transaction.

Free to use for end users (no gas fees, or fixed/minimal fees that Reddit can pay on their behalf)
With transaction pricing as low as $0.0000025 per transaction, it may be considered reasonable for Reddit to cover transaction fees for users.
All of Reddit's Transactions on Blockchain (month)
Community points can be earned by users and distributed directly to their Reddit account in batch (as per Reddit minting plan), and allow users to withdraw rewards to their Ethereum wallet whenever they wish. Withdrawal fees can be paid by either user or Reddit. This model has been operating inside the Dragonchain system since 2018, and many security and financial compliance features can be optionally added. We feel that this capability greatly enhances user experience because it is seamless to a regular user without cryptocurrency experience, yet flexible to a tech savvy user. With regard to currency or token transactions, these would occur on the Reddit network, verified to BTC and ETH. These transactions would incur the $0.0000025 transaction fee. To estimate this fee we use the monthly active Reddit users statista with a 60% adoption rate and an estimated 10 transactions per month average resulting in an approximate $720 cost across the system. Reddit could feasibly incur all associated internal network charges (mining/minting, transfer, burn) as these are very low and controllable fees.
Reddit Internal Token Transaction Fees

Reddit Ethereum Token Transaction Fees
When we consider further the Ethereum fees that might be incurred, we have a few choices for a solution.
  1. Offload all Ethereum transaction fees (user withdrawals) to interested users as they wish to withdraw tokens for external use or sale.
  2. Cover Ethereum transaction fees by aggregating them on a timed schedule. Users would request withdrawal (from Reddit or individual subreddits), and they would be transacted on the Ethereum network every hour (or some other schedule).
  3. In a combination of the above, customers could cover aggregated fees.
  4. Integrate with alternate Ethereum roll up solutions or other proposals to aggregate minting and distribution transactions onto Ethereum.

Bonus Points

Users should be able to view their balances & transactions via a blockchain explorer-style interface
From interfaces for users who have no knowledge of blockchain technology to users who are well versed in blockchain terms such as those present in a typical block explorer, a system powered by Dragonchain has flexibility on how to provide balances and transaction data to users. Transactions can be made viewable in an Eternal Proof Report, which displays raw data along with TIME staking information and traceability all the way to Bitcoin, Ethereum, and every other Interchained network. The report shows fields such as transaction ID, timestamp, block ID, multiple verifications, and Interchain proof. See example here.
Node payouts within the Dragonchain console are listed in chronological order and can be further seen in either Dragons or USD. See example here.
In our social media platform, Dragon Den, users can see, in real-time, their NRG and MTR balances. See example here.
A new influencer app powered by Dragonchain, Raiinmaker, breaks down data into a user friendly interface that shows coin portfolio, redeemed rewards, and social scores per campaign. See example here.

Exiting is fast & simple
Withdrawing funds on Dragonchain’s console requires three clicks, however, withdrawal scenarios with more enhanced security features per Reddit’s discretion are obtainable.

Interoperability Compatibility with third party apps (wallets/contracts/etc) is necessary.
Proven interoperability at scale that surpasses the required specifications. Our entire platform consists of interoperable blockchains connected to each other and traditional systems. APIs are well documented. Third party permissions are possible with a simple smart contract without the end user being aware. No need to learn any specialized proprietary language. Any code base (not subsets) is usable within a Docker container. Interoperable with any blockchain or traditional APIs. We’ve witnessed relatively complex systems built by engineers with no blockchain or cryptocurrency experience. We’ve also demonstrated the creation of smart contracts within minutes built with BASH shell and Node.js. Please see our source code and API documentation.

Scaling solutions should be extensible and allow third parties to build on top of it Open source and extensible
APIs should be well documented and stable

Documentation should be clear and complete
For full documentation, explore our docs, SDK’s, Github repo’s, architecture documents, original Disney documentation, and other links or resources provided in this proposal.

Third-party permissionless integrations should be possible & straightforward Smart contracts are Docker based, can be written in any language, use full language (not subsets), and can therefore be integrated with any system including traditional system APIs. Simple is better. Learning an uncommon or proprietary language should not be necessary.
Advanced knowledge of mathematics, cryptography, or L2 scaling should not be required. Compatibility with common utilities & toolchains is expected.
Dragonchain business nodes and smart contracts leverage Docker to allow the use of literally any language or executable code. No proprietary language is necessary. We’ve witnessed relatively complex systems built by engineers with no blockchain or cryptocurrency experience. We’ve also demonstrated the creation of smart contracts within minutes built with BASH shell and Node.js.

Bonus

Bonus Points: Show us how it works. Do you have an idea for a cool new use case for Community Points? Build it!

TIME

Community points could be awarded to Reddit users based upon TIME too, whereas the longer someone is part of a subreddit, the more community points someone naturally gained, even if not actively commenting or sharing new posts. A daily login could be required for these community points to be credited. This grants awards to readers too and incentivizes readers to create an account on Reddit if they browse the website often. This concept could also be leveraged to provide some level of reputation based upon duration and consistency of contribution to a community subreddit.

Dragon Den

Dragonchain has already built a social media platform that harnesses community involvement. Dragon Den is a decentralized community built on the Dragonchain blockchain platform. Dragon Den is Dragonchain’s answer to fake news, trolling, and censorship. It incentivizes the creation and evaluation of quality content within communities. It could be described as being a shareholder of a subreddit or Reddit in its entirety. The more your subreddit is thriving, the more rewarding it will be. Den is currently in a public beta and in active development, though the real token economy is not live yet. There are different tokens for various purposes. Two tokens are Lair Ownership Rights (LOR) and Lair Ownership Tokens (LOT). LOT is a non-fungible token for ownership of a specific Lair. LOT will only be created and converted from LOR.
Energy (NRG) and Matter (MTR) work jointly. Your MTR determines how much NRG you receive in a 24-hour period. Providing quality content, or evaluating content will earn MTR.

Security. Users have full ownership & control of their points.
All community points awarded based upon any type of activity or gift, are secured and provable to all Interchain networks (currently BTC, ETH, ETC). Users are free to spend and withdraw their points as they please, depending on the features Reddit wants to bring into production.

Balances and transactions cannot be forged, manipulated, or blocked by Reddit or anyone else
Users can withdraw their balance to their ERC20 wallet, directly through Reddit. Reddit can cover the fees on their behalf, or the user covers this with a portion of their balance.

Users should own their points and be able to get on-chain ERC20 tokens without permission from anyone else
Through our console users can withdraw their ERC20 rewards. This can be achieved on Reddit too. Here is a walkthrough of our console, though this does not show the quick withdrawal functionality, a user can withdraw at any time. https://www.youtube.com/watch?v=aNlTMxnfVHw

Points should be recoverable to on-chain ERC20 tokens even if all third-parties involved go offline
If necessary, signed transactions from the Reddit system (e.g. Reddit + Subreddit) can be sent to the Ethereum smart contract for minting.

A public, third-party review attesting to the soundness of the design should be available
To our knowledge, at least two large corporations, including a top 3 accounting firm, have conducted positive reviews. These reviews have never been made public, as Dragonchain did not pay or contract for these studies to be released.

Bonus points
Public, third-party implementation review available or in progress
See above

Compatibility with HSMs & hardware wallets
For the purpose of this proposal, all tokenization would be on the Ethereum network using standard token contracts and as such, would be able to leverage all hardware wallet and Ethereum ecosystem services.

Other Considerations

Minting/distributing tokens is not performed by Reddit directly
This operation can be automated by smart contract on Ethereum. Subreddits can if desired have a role to play.

One off point burning, as well as recurring, non-interactive point burning (for subreddit memberships) should be possible and scalable
This is possible and scalable with interaction between Dragonchain Reddit system and Ethereum token contract(s).

Fully open-source solutions are strongly preferred
Dragonchain is fully open source (see section on Disney release after conclusion).

Conclusion

Whether it is today, or in the future, we would like to work together to bring secure flexibility to the highest standards. It is our hope to be considered by Ethereum, Reddit, and other integrative solutions so we may further discuss the possibilities of implementation. In our public demonstration, 256 million transactions were handled in our operational network on chain in 24 hours, for the low cost of $25K, which if run today would cost $625. Dragonchain’s interoperable foundation provides the atmosphere necessary to implement a frictionless community points system. Thank you for your consideration of our proposal. We look forward to working with the community to make something great!

Disney Releases Blockchain Platform as Open Source

The team at Disney created the Disney Private Blockchain Platform. The system was a hybrid interoperable blockchain platform for ledgering and smart contract development geared toward solving problems with blockchain adoption and usability. All objective evaluation would consider the team’s output a success. We released a list of use cases that we explored in some capacity at Disney, and our input on blockchain standardization as part of our participation in the W3C Blockchain Community Group.
https://lists.w3.org/Archives/Public/public-blockchain/2016May/0052.html

Open Source

In 2016, Roets proposed to release the platform as open source to spread the technology outside of Disney, as others within the W3C group were interested in the solutions that had been created inside of Disney.
Following a long process, step by step, the team met requirements for release. Among the requirements, the team had to:
  • Obtain VP support and approval for the release
  • Verify ownership of the software to be released
  • Verify that no proprietary content would be released
  • Convince the organization that there was a value to the open source community
  • Convince the organization that there was a value to Disney
  • Offer the plan for ongoing maintenance of the project outside of Disney
  • Itemize competing projects
  • Verify no conflict of interest
  • Preferred license
  • Change the project name to not use the name Disney, any Disney character, or any other associated IP - proposed Dragonchain - approved
  • Obtain legal approval
  • Approval from corporate, parks, and other business units
  • Approval from multiple Disney patent groups Copyright holder defined by Disney (Disney Connected and Advanced Technologies)
  • Trademark searches conducted for the selected name Dragonchain
  • Obtain IT security approval
  • Manual review of OSS components conducted
  • OWASP Dependency and Vulnerability Check Conducted
  • Obtain technical (software) approval
  • Offer management, process, and financial plans for the maintenance of the project.
  • Meet list of items to be addressed before release
  • Remove all Disney project references and scripts
  • Create a public distribution list for email communications
  • Remove Roets’ direct and internal contact information
  • Create public Slack channel and move from Disney slack channels
  • Create proper labels for issue tracking
  • Rename internal private Github repository
  • Add informative description to Github page
  • Expand README.md with more specific information
  • Add information beyond current “Blockchains are Magic”
  • Add getting started sections and info on cloning/forking the project
  • Add installation details
  • Add uninstall process
  • Add unit, functional, and integration test information
  • Detail how to contribute and get involved
  • Describe the git workflow that the project will use
  • Move to public, non-Disney git repository (Github or Bitbucket)
  • Obtain Disney Open Source Committee approval for release
On top of meeting the above criteria, as part of the process, the maintainer of the project had to receive the codebase on their own personal email and create accounts for maintenance (e.g. Github) with non-Disney accounts. Given the fact that the project spanned multiple business units, Roets was individually responsible for its ongoing maintenance. Because of this, he proposed in the open source application to create a non-profit organization to hold the IP and maintain the project. This was approved by Disney.
The Disney Open Source Committee approved the application known as OSSRELEASE-10, and the code was released on October 2, 2016. Disney decided to not issue a press release.
Original OSSRELASE-10 document

Dragonchain Foundation

The Dragonchain Foundation was created on January 17, 2017. https://den.social/l/Dragonchain/24130078352e485d96d2125082151cf0/dragonchain-and-disney/
submitted by j0j0r0 to ethereum [link] [comments]

Comparing Nano's Nakamoto Coefficient

Inspired by the discussion on the cc subreddit (which I won't link to), I have some questions.
These Nakamoto coefficients aren't very comparable. Miners can reassign their hashrate at any time. Hashrate also has an ongoing, real expense. Nano votes can't be reassigned if the network is controlled, and there's no out of band "real" cost to acquire or maintain control. Thus, it's extremely misleading to try and compare these.
I would say that indeed hashrate has an ongoing, real expense so indeed, performing a 51% attack on Bitcoin will cost you on a per hour basis. On the other hand, get a 51% majority of Nano and you essentially block the network for eternity from what I understand. Bitcoin would most likely also collapse in value if a 51% attack was successfully performed, because even if it were to go offline for an hour and just a few doublespends were performed, it would undermine the store of value mantra quite strongly.
Some sides notes here are, of course, that getting a 51% majority delegates for Nano is extremely difficult or expensive, as you need to buy up a large percentage of the supply yourself or you need to convince a lot of people to delegate to you, which hopefully only works if you build services that use Nano and therefore, in both cases, you have a vested interest in ensuring the Nano network remains valuable.
On the other hand, Bitcoin miners have made large investments in ASICs which means they are strongly incentivized in the same sense, they want the Bitcoin network to remain valuable. Convincing either enough large Nano holders, or large swathes of Bitcoin hash power, would therefore be difficult.
However, wouldn't it, generally, not be easier to find hash power outside the large miners currently mining than it would be to find Nano to give yourself a majority? I'm thinking that to get a 51% majority in Nano as said earlier you need to buy up enough of the outstanding Nano, or convince holders with a vested interest in the value of the Nano network. For Bitcoin however, I could rent out a chunk of Amazon's computing power and set up my own temporary mining operation to compete with the mining pools currently available. It would still be expensive, but, I am assuming, less so than taking the Nano option (even with current market caps).
Is this a fair comparison? Or am I misrepresenting how easy it would be to get a Nano majority, or misrepresenting how difficult it would be to find alternative hash power to mine Bitcoin?
Edit: Comparing to Bitcoin because it has the most hash power, this goes for all PoW crypto.
submitted by SenatusSPQR to nanocurrency [link] [comments]

What is the cost of BTC mining in China and Russia?

I know that the vast majority of Bitcoin mining is done in China due to the abundance of cheap electricity from hydroelectric dams. Out of curiosity, does anyone actually know what their effective cost per kWh is?
I’ve also read that a lot of that hash power could move to Russia (Siberia) due to some green initiatives by the Chinese government. Whether that’s true or not remains to be seen. Does anyone have any electricity cost data for Russia?
Thanks.
submitted by asdfredditusername to Bitcoin [link] [comments]

Bitcoin's next 15 years : Year 2020–2035

2020 4Q

~ More companies follow in Microstrategy’s footsteps. Rumors of more corporate treasurers investing in BTC in boardrooms globally. A few listed large corporates announce accumulation of BTC after their buddies have all bought in (Board members, C-suite executives, family, and friends, etc.)
~ Money printing does not stop as the deflationary force of technology is too severe; the new US government formed after Biden’s win begins to adopt MMT as its primary guidance of future economic theory, led by Steph Kelton.
~ The holiday season and strong seasonality pump BTC back to $20k for the first time. Hard rejection and price fall back to $14k.

2021

~ BTC finally breaks $20k after multiple retests of overhead resistance sometime in spring
~ Almost weekly we see another corporation announcing vested interest in BTC
~ No longer in doubt that the asset class is in a bull market. Macro funds pile in. By year-end, we’re at $55k. Newspaper reports Bitcoin has now broken the $1 trillion mark. Most institutions begin scrambling to understand the asset class and set up “Digital Asset Investment teams”
~ Retail money flows to altcoins; Bitcoin is becoming too expensive for “retail” investors. The bitcoin community discusses possibly denoting BTC as sats, but majority of exchanges not interested as they derive most income from alt flows. However, most Bitcoin-only platforms switch to sats as the primary display format led by bitcoiners who now have considerable wealth and influence
~ Increasing talk that some smaller nations are now discussing the prospect of including Bitcoin on their central bank balance sheet
~ The first BTC-denominated corporate bond is launched

2022

~ Those in power have established full BTC positions, and we begin to see subtle clues that some countries are possibly accumulating BTC
~ Private banks selling BTC structured products now out in full force; custody solutions are now institutional-grade. 50% of the world’s banks have some product/solution tailored around bitcoin. The other 50% scramble.
~ Marks the top as BTC momentarily exceeds the most valuable company by market cap (~$2.5 trillion in 2022 @ $130K price). The final days of the frenzy are filled with rumors that central banks have accumulated 10% of global supply, and that it may even form part of the IMF’s global recognized reserve currencies. Crypto Twitter reaches peak “I told you so”

2023

~ The next bear market isn’t as severe as the last few; as the digital asset teams of various institutions are accumulating up to 2-5% of their AuM. It’s now commonly accepted that this asset class is here to stay and that even deploying $10 billion is no longer an issue in an asset class worth an aggregate $5 trillion.
~ BTC finds a floor 60% lower at $50K as smart money accumulates. CT screams for a 80% correction because mUh bItCoIn cYcLeS aNd fRaCtAls
~ Investment banks now have full-fledged research teams dedicated to digital assets. Calls for 80% correction too, so the smart money front-runs.
~ The middle class latches on to the wholecoining meme. “1 Bitcoin to secure a retirement; stack those sats”
~ The wealthy who are now increasingly composed of inherited wealth begin selling real estate/equities/bonds for Bitcoin but holds their BTC with their private bank. Realizing that Bitcoin supply is truly limited and sensing the “1 bitcoin to retire” meme; and that not every millionaire can own 1 bitcoin, many of the rich/ultra-rich scramble to buy 5–100 BTC each if only to cement their status as rich. 5–100 BTC costs $500K-10M (at $100k per BTC)
~ The winning product of the year is an automatic savings plan in bitcoin.

2025

~ Bitcoin is back to trading near its all-time highs of $130K after the 2024 halving cycle, however, the effect is marginal but the markets wrongly attribute it to the halving supply squeeze, building a false narrative for the next cycle in 2028.
~ Institutional money now in full-play; on hindsight we’ll realize the 10-year steady bull-run has actually begun since last year in 2023, similar to the gold bull run from 2000 to 2011
~ More exchanges finally denominate BTC in sats. $100K BTC = 0.1 cent per sat. Logging into platform displays your stack as:
“11.7m satoshis ≈ $17,500”

2028

~ Retail attempts to trade around the 2028 halving cycle. The halving cycle no longer have much of an impact, as demand now far outstrips supply changes
~ Many earlycoiners now sell between $200–400K, only to see it continue its relentless climb at a 30% annual rate
~ The first central bank announces the official addition to their balance sheets; all other central bank begins to FOMO. Cements BTC as a global reserve asset.
~ Governments ask that private ownership of bitcoin be transferred to regulated financial institutions such as their local bank where it will be held under custody. 70% of people do so.

2033

~ Many of the early-coiners now buyback at near to $1M ($20 trillion market cap), finally equaling gold’s market cap at a price of $4000+
~ Bitcoin peaks and meanders under $1M for the next decade
~ Volatility is now <10% per year, merchants begin adopting it en-masse as a medium of exchange

2035

~ 5 years of price stability leads to some merchants re-pricing certain goods in sat-terms
~ The lightning network crosses a billion channels created
~ Fiat does not go away, but most G20 countries decide to ban bitcoin as a medium of exchange for economic transactions. Ownership of bitcoin as an asset is encouraged as a store of wealth; private ownership is frowned upon and in some cases made illegal.
submitted by laobuggier to CryptoCurrency [link] [comments]

Why Osana takes so long? (Programmer's point of view on current situation)

I decided to write a comment about «Why Osana takes so long?» somewhere and what can be done to shorten this time. It turned into a long essay. Here's TL;DR of it:
The cost of never paying down this technical debt is clear; eventually the cost to deliver functionality will become so slow that it is easy for a well-designed competitive software product to overtake the badly-designed software in terms of features. In my experience, badly designed software can also lead to a more stressed engineering workforce, in turn leading higher staff churn (which in turn affects costs and productivity when delivering features). Additionally, due to the complexity in a given codebase, the ability to accurately estimate work will also disappear.
Junade Ali, Mastering PHP Design Patterns (2016)
Longer version: I am not sure if people here wanted an explanation from a real developer who works with C and with relatively large projects, but I am going to do it nonetheless. I am not much interested in Yandere Simulator nor in this genre in general, but this particular development has a lot to learn from for any fellow programmers and software engineers to ensure that they'll never end up in Alex's situation, especially considering that he is definitely not the first one to got himself knee-deep in the development hell (do you remember Star Citizen?) and he is definitely not the last one.
On the one hand, people see that Alex works incredibly slowly, equivalent of, like, one hour per day, comparing it with, say, Papers, Please, the game that was developed in nine months from start to finish by one guy. On the other hand, Alex himself most likely thinks that he works until complete exhaustion each day. In fact, I highly suspect that both those sentences are correct! Because of the mistakes made during early development stages, which are highly unlikely to be fixed due to the pressure put on the developer right now and due to his overall approach to coding, cost to add any relatively large feature (e.g. Osana) can be pretty much comparable to the cost of creating a fan game from start to finish. Trust me, I've seen his leaked source code (don't tell anybody about that) and I know what I am talking about. The largest problem in Yandere Simulator right now is its super slow development. So, without further ado, let's talk about how «implementing the low hanging fruit» crippled the development and, more importantly, what would have been an ideal course of action from my point of view to get out. I'll try to explain things in the easiest terms possible.
  1. else if's and lack any sort of refactoring in general
The most «memey» one. I won't talk about the performance though (switch statement is not better in terms of performance, it is a myth. If compiler detects some code that can be turned into a jump table, for example, it will do it, no matter if it is a chain of if's or a switch statement. Compilers nowadays are way smarter than one might think). Just take a look here. I know that it's his older JavaScript code, but, believe it or not, this piece is still present in C# version relatively untouched.
I refactored this code for you using C language (mixed with C++ since there's no this pointer in pure C). Take a note that else if's are still there, else if's are not the problem by itself.
The refactored code is just objectively better for one simple reason: it is shorter, while not being obscure, and now it should be able to handle, say, Trespassing and Blood case without any input from the developer due to the usage of flags. Basically, the shorter your code, the more you can see on screen without spreading your attention too much. As a rule of thumb, the less lines there are, the easier it is for you to work with the code. Just don't overkill that, unless you are going to participate in International Obfuscated C Code Contest. Let me reiterate:
Perfection is achieved, not when there is nothing more to add, but when there is nothing left to take away.
Antoine de Saint-Exupéry
This is why refactoring — activity of rewriting your old code so it does the same thing, but does it quicker, in a more generic way, in less lines or simpler — is so powerful. In my experience, you can only keep one module/class/whatever in your brain if it does not exceed ~1000 lines, maybe ~1500. Splitting 17000-line-long class into smaller classes probably won't improve performance at all, but it will make working with parts of this class way easier.
Is it too late now to start refactoring? Of course NO: better late than never.
  1. Comments
If you think that you wrote this code, so you'll always easily remember it, I have some bad news for you: you won't. In my experience, one week and that's it. That's why comments are so crucial. It is not necessary to put a ton of comments everywhere, but just a general idea will help you out in the future. Even if you think that It Just Works™ and you'll never ever need to fix it. Time spent to write and debug one line of code almost always exceeds time to write one comment in large-scale projects. Moreover, the best code is the code that is self-evident. In the example above, what the hell does (float) 6 mean? Why not wrap it around into the constant with a good, self-descriptive name? Again, it won't affect performance, since C# compiler is smart enough to silently remove this constant from the real code and place its value into the method invocation directly. Such constants are here for you.
I rewrote my code above a little bit to illustrate this. With those comments, you don't have to remember your code at all, since its functionality is outlined in two tiny lines of comments above it. Moreover, even a person with zero knowledge in programming will figure out the purpose of this code. It took me less than half a minute to write those comments, but it'll probably save me quite a lot of time of figuring out «what was I thinking back then» one day.
Is it too late now to start adding comments? Again, of course NO. Don't be lazy and redirect all your typing from «debunk» page (which pretty much does the opposite of debunking, but who am I to judge you here?) into some useful comments.
  1. Unit testing
This is often neglected, but consider the following. You wrote some code, you ran your game, you saw a new bug. Was it introduced right now? Is it a problem in your older code which has shown up just because you have never actually used it until now? Where should you search for it? You have no idea, and you have one painful debugging session ahead. Just imagine how easier it would be if you've had some routines which automatically execute after each build and check that environment is still sane and nothing broke on a fundamental level. This is called unit testing, and yes, unit tests won't be able to catch all your bugs, but even getting 20% of bugs identified at the earlier stage is a huge boon to development speed.
Is it too late now to start adding unit tests? Kinda YES and NO at the same time. Unit testing works best if it covers the majority of project's code. On the other side, a journey of a thousand miles begins with a single step. If you decide to start refactoring your code, writing a unit test before refactoring will help you to prove to yourself that you have not broken anything without the need of running the game at all.
  1. Static code analysis
This is basically pretty self-explanatory. You set this thing once, you forget about it. Static code analyzer is another «free estate» to speed up the development process by finding tiny little errors, mostly silly typos (do you think that you are good enough in finding them? Well, good luck catching x << 4; in place of x <<= 4; buried deep in C code by eye!). Again, this is not a silver bullet, it is another tool which will help you out with debugging a little bit along with the debugger, unit tests and other things. You need every little bit of help here.
Is it too late now to hook up static code analyzer? Obviously NO.
  1. Code architecture
Say, you want to build Osana, but then you decided to implement some feature, e.g. Snap Mode. By doing this you have maybe made your game a little bit better, but what you have just essentially done is complicated your life, because now you should also write Osana code for Snap Mode. The way game architecture is done right now, easter eggs code is deeply interleaved with game logic, which leads to code «spaghettifying», which in turn slows down the addition of new features, because one has to consider how this feature would work alongside each and every old feature and easter egg. Even if it is just gazing over one line per easter egg, it adds up to the mess, slowly but surely.
A lot of people mention that developer should have been doing it in object-oritented way. However, there is no silver bullet in programming. It does not matter that much if you are doing it object-oriented way or usual procedural way; you can theoretically write, say, AI routines on functional (e.g. LISP)) or even logical language if you are brave enough (e.g. Prolog). You can even invent your own tiny programming language! The only thing that matters is code quality and avoiding the so-called shotgun surgery situation, which plagues Yandere Simulator from top to bottom right now. Is there a way of adding a new feature without interfering with your older code (e.g. by creating a child class which will encapsulate all the things you need, for example)? Go for it, this feature is basically «free» for you. Otherwise you'd better think twice before doing this, because you are going into the «technical debt» territory, borrowing your time from the future by saying «I'll maybe optimize it later» and «a thousand more lines probably won't slow me down in the future that much, right?». Technical debt will incur interest on its own that you'll have to pay. Basically, the entire situation around Osana right now is just a huge tale about how just «interest» incurred by technical debt can control the entire project, like the tail wiggling the dog.
I won't elaborate here further, since it'll take me an even larger post to fully describe what's wrong about Yandere Simulator's code architecture.
Is it too late to rebuild code architecture? Sadly, YES, although it should be possible to split Student class into descendants by using hooks for individual students. However, code architecture can be improved by a vast margin if you start removing easter eggs and features like Snap Mode that currently bloat Yandere Simulator. I know it is going to be painful, but it is the only way to improve code quality here and now. This will simplify the code, and this will make it easier for you to add the «real» features, like Osana or whatever you'd like to accomplish. If you'll ever want them back, you can track them down in Git history and re-implement them one by one, hopefully without performing the shotgun surgery this time.
  1. Loading times
Again, I won't be talking about the performance, since you can debug your game on 20 FPS as well as on 60 FPS, but this is a very different story. Yandere Simulator is huge. Once you fixed a bug, you want to test it, right? And your workflow right now probably looks like this:
  1. Fix the code (unavoidable time loss)
  2. Rebuild the project (can take a loooong time)
  3. Load your game (can take a loooong time)
  4. Test it (unavoidable time loss, unless another bug has popped up via unit testing, code analyzer etc.)
And you can fix it. For instance, I know that Yandere Simulator makes all the students' photos during loading. Why should that be done there? Why not either move it to project building stage by adding build hook so Unity does that for you during full project rebuild, or, even better, why not disable it completely or replace with «PLACEHOLDER» text for debug builds? Each second spent watching the loading screen will be rightfully interpreted as «son is not coding» by the community.
Is it too late to reduce loading times? Hell NO.
  1. Jenkins
Or any other continuous integration tool. «Rebuild a project» can take a long time too, and what can we do about that? Let me give you an idea. Buy a new PC. Get a 32-core Threadripper, 32 GB of fastest RAM you can afford and a cool motherboard which would support all of that (of course, Ryzen/i5/Celeron/i386/Raspberry Pi is fine too, but the faster, the better). The rest is not necessary, e.g. a barely functional second hand video card burned out by bitcoin mining is fine. You set up another PC in your room. You connect it to your network. You set up ramdisk to speed things up even more. You properly set up Jenkins) on this PC. From now on, Jenkins cares about the rest: tracking your Git repository, (re)building process, large and time-consuming unit tests, invoking static code analyzer, profiling, generating reports and whatever else you can and want to hook up. More importantly, you can fix another bug while Jenkins is rebuilding the project for the previous one et cetera.
In general, continuous integration is a great technology to quickly track down errors that were introduced in previous versions, attempting to avoid those kinds of bug hunting sessions. I am highly unsure if continuous integration is needed for 10000-20000 source lines long projects, but things can be different as soon as we step into the 100k+ territory, and Yandere Simulator by now has approximately 150k+ source lines of code. I think that probably continuous integration might be well worth it for Yandere Simulator.
Is it too late to add continuous integration? NO, albeit it is going to take some time and skills to set up.
  1. Stop caring about the criticism
Stop comparing Alex to Scott Cawton. IMO Alex is very similar to the person known as SgtMarkIV, the developer of Brutal Doom, who is also a notorious edgelord who, for example, also once told somebody to kill himself, just like… However, being a horrible person, SgtMarkIV does his job. He simply does not care much about public opinion. That's the difference.
  1. Go outside
Enough said. Your brain works slower if you only think about games and if you can't provide it with enough oxygen supply. I know that this one is probably the hardest to implement, but…
That's all, folks.
Bonus: Do you think how short this list would have been if someone just simply listened to Mike Zaimont instead of breaking down in tears?
submitted by Dezhitse to Osana [link] [comments]

Make Money 10000$ Per Day With Bitcoin  Without ... How much can you earn bitcoin mining on nicehash Tim Draper Predicts $10,000 Per Bitcoin by 2018 IS BITCOIN (BTC) MINING WORTH IT 2020?? -PROFITABLE? ✅ BITCOIN HALVING 2020 DONE ✅ Bitcoins Next Move?

The middle chart below shows that it costs ~$37 in cash expenses to run 1PH/s per day on the efficient S19 class rigs at a 4c/kWh power cost, but running S9 class rigs at 4c/kWh power would cost ... (Price per KWh * Electricity consumed per transaction * number of transactions per block) / Bitcoins per Block . This gave us a price in cents so we divided by further 100 to get a price in US dollars. The map was created using MapChart.net. List of Electricity Cost Of Mining One Bitcoin By Country A site that will show the cost for electrical power to generate a 50 BTC block given a hash rate, power usage and cost of electricity as inputs. Difficulty field: BlockExplorer provides the current difficulty. Hash rate field: the value can determined from your miner client, or from the mining hardware comparison. Power usage field: the value for a GPU card itself can be determined from the ... Hash Rate: It is the total hash power of the bitcoin mining devices used for bitcoin mining and it is represented in TH/s which means Trillions Hash generated per second.For example, Bitmain Antminer S19 Pro mining hardware has a Hash Rate of 110TH/s. Power Consumption: Power Consumption refers to the total power consumption of the mining devices used for bitcoin mining. Median BTC mining power cost is 3c / kWh and cost to mine 1 BTC is ~$5000 • Hashrate could reach 260 & 360 EH/s in 12 & 24 months with upgrade cycle but needs $6.3B Capex; funding gap is $4.1B vs. industry cash generation • Price needs to appreciate by $1000 for every 10EH/s increase to stay revenue neutral per MWh • Cheap power, BTC ...

[index] [43632] [14237] [20117] [5720] [47759] [24539] [47671] [23946] [22466] [16298]

Make Money 10000$ Per Day With Bitcoin Without ...

The Bitcoin Halving 2020 happened excactly as intended at blockheight 630'000. The Bitcoin block reward got reduced by 50% to 6.25 BTC per block. What does this mean for Bitcoin miners and what is ... https://www.youtube.com/watch?v=-cWkoBJuu-k Track: CØDE - Duck Face [NCS Release] Music provided by NoCopyrightSounds. Watch: https://youtu.be/-cWkoBJuu-k Fr... Venture capitalist Tim Draper predicts $10K per Bitcoin by 2018 - Duration: 6:23. ... Tommy World Power 57,315 views. 2:26. Bitcoin Price Prediction For 2017 - Duration: 5:37. David MacGregor ... http://bitfun.co/?ref=0BE902754D91 PLAY GAMES, HAVE FUN AND EARN FREE BITCOIN Hello guys just a quick video on my mining rigs,How much I have spent and how m... Bitcoin is still trading slightly about $8,000 and yesterday I made an entrance into the market and bought some Bitcoin and Ethereum. In today's video I will...

#