Cassie Heart pfp
Cassie Heart
@cassie
Execution and well-conceived integration are a mandatory minimum of realizing the potential of new innovations. Grab a coffee (or hot chocolate if you're @greg), sit back, and relax, we're going on a deep dive.
36 replies
28 recasts
235 reactions

Cassie Heart pfp
Cassie Heart
@cassie
The Apple ][ is frequently hailed as one of the major milestones in personal computers. Not only did it make home computing more accessible than ever, it delivered many things we now take for granted that were incredible technological achievements for the time. Did you catch them in the picture above?
1 reply
0 recast
16 reactions

Cassie Heart pfp
Cassie Heart
@cassie
Let's look at an example of what others in the industry were offering to help demonstrate the difference: The Commodore PET. See the difference now? The Apple ][ could do color - something no other personal computer had been doing yet. The Apple ][ had floppy disks, faster and at greater capacity than any other. How?
1 reply
0 recast
11 reactions

Cassie Heart pfp
Cassie Heart
@cassie
At the time, video signals for home computers utilized NTSC composite video, the kind that worked with television monitors. These sets supported color through chrominance subcarrier frequency modulation - a technique that at the time typically required expensive chips to support.
1 reply
0 recast
13 reactions

Cassie Heart pfp
Cassie Heart
@cassie
But supporting limited colors could be achieved cheaply, if you could get the timing right, by emitting the signals in a particular way. Wozniak was predictably nerdsniped by this, and used this trick with the MOS 6502 and its timing set to 1.023 MHz, 2/7 of the NTSC color subcarrier rate.
2 replies
0 recast
16 reactions

Cassie Heart pfp
Cassie Heart
@cassie
The integration of controlled timing from the deepest part of the hardware down to the software unlocked color displays for home users, while its competitor, the PET, using the exact same chip, could not do this. But the division runs deeper. A sentiment was growing that floppy disks were needed for non-hobbyist PCs.
1 reply
0 recast
11 reactions

Cassie Heart pfp
Cassie Heart
@cassie
Woz stood up to the challenge again, this time, taking the raw hardware physically controlling the disk, and augmenting it with a cheaper software and timing-driven approach, reducing the overall chips in use (and cost, saving Apple >$300 per drive), while simultaneously making it faster and able to store more data.
1 reply
0 recast
12 reactions

Cassie Heart pfp
Cassie Heart
@cassie
Commodore's PET disk drive design ultimately required two processors on the scale of the Apple ]['s main central processing unit just to make it function, and for all that expense, still couldn't display color.
2 replies
0 recast
10 reactions

Cassie Heart pfp
Cassie Heart
@cassie
As I am wont to do, I see a heavy parallel here to the crypto industry. When we look at crypto today, the innovations that are happening are frequently very downstream of ossified architectural designs, which results in huge efficiency loss.
1 reply
1 recast
13 reactions

Cassie Heart pfp
Cassie Heart
@cassie
The original block chain was designed out of a minimalist point of view to most succinctly encapsulate transactions of a singular coin, encoded as simple Forth-like scripts, and to be provably referenced in its inclusion via the use of Merkle proofs.
1 reply
0 recast
13 reactions

Cassie Heart pfp
Cassie Heart
@cassie
At the time, Merkle trees were one of the most efficient means to collect contiguous segments of data and verify inclusion. The compactness of the proof, however, leaves a little bit to be desired, especially when scaling out to the whole of human commerce, or trying even more loftily to be a world computer.
2 replies
0 recast
11 reactions

Cassie Heart pfp
Cassie Heart
@cassie
For starters, to verify a given transaction in the original design, you'd have to hold the entire history to self verify. Many moved to incorporating the use of a Merkle root of the overall ledger state in the block header so as to avoid such dilemmas, but still mandated full nodes synchronize full state.
1 reply
0 recast
8 reactions

Cassie Heart pfp
Cassie Heart
@cassie
Additionally, some have used alternatives to Merkle trees, such as Ethereum's choice of PATRICIA-Merkle trees, and doing a proper proof requires more sibling hashes for every ascent of the tree, currently about 3,584 bytes. Opting to reincode as a binary tree, it still requires nearly a kilobyte.
1 reply
0 recast
10 reactions

Cassie Heart pfp
Cassie Heart
@cassie
Thankfully, bandwidth has improved, but so too have our cryptographic techniques: KZG commitments/proofs offer a very succinct scheme for proving inclusion in a set (accurately, a vector) - it is constant size: the item you wish to prove was included, its position in the set, and the elliptic curve point. Nice!
1 reply
0 recast
11 reactions

Cassie Heart pfp
Cassie Heart
@cassie
What's not so nice: the time to compute larger sets. Complexity wise, constructing the commitment is O(n^2), not to mention you also have to have constructed a secure reference string equal in the number of elements (accurately, the degree) to the maximum size of the set.
1 reply
0 recast
8 reactions

Cassie Heart pfp
Cassie Heart
@cassie
For Ethereum, at 263.80M accounts, we're talking about a secure reference string of 2^28, and that's just for today! If you thought contributing to the Ethereum KZG ceremony was long for 2^15 elements, you'd be in a much worse world of pain (by most recent report of such a setup, a single contribution took two days).
1 reply
0 recast
8 reactions

Cassie Heart pfp
Cassie Heart
@cassie
What about a compromise? Keep the tree structure, lose the sibling proofs, with each traversal step of the tree as a KZG commitment/proof. At log16(263.80M) ≈ 7, this reduces the proof size to 7 points (48 bytes each), or 336 bytes. Not bad! If you followed this so far, congrats, you now know what a Verkle tree is!
2 replies
0 recast
14 reactions

Cassie Heart pfp
Cassie Heart
@cassie
So what is Ethereum doing with verkle trees? Nothing yet! Instead, Ethereum is using the simple vector commitments for up to 4096 elements (32 bytes each), allowing up to six of these commitments to be posted as transactions per block, with a dynamic fee market of its own. How's that going for them?
1 reply
0 recast
10 reactions

Cassie Heart pfp
Cassie Heart
@cassie
Inscriptions showed up (as I predicted), more L2s than there are slots per block showed up, and this is just beginning to ramp up. The competitive fee market separated from regular transactions has resulted in the original significant 100x+ price reduction currently sitting closer to a modest 3x.
1 reply
0 recast
9 reactions

Cassie Heart pfp
Cassie Heart
@cassie
If pressure continues, this will trend soon towards blobs becoming _more expensive_ to use than calldata, on top of the accounts producing the blobs still needing to exist in the world state PATRICIA-Merkle tree. But hey, no worries, verkle trees will make that more efficiently verifiable, right? Right?
1 reply
0 recast
9 reactions

Cassie Heart pfp
Cassie Heart
@cassie
According to certain active voices in the space, they're going to be bringing "the next 1 billion users onchain". What does that look like? Ethereum has accumulated some cruft over time, so despite a healthy average over the last year of about 500k daily active addresses, we're sitting at 263.80M accounts.
1 reply
0 recast
7 reactions

Cassie Heart pfp
Cassie Heart
@cassie
This is a poor estimate (before maxis surely @ me for this), but we need to get an idea for how much cruft we can reasonably assume will accumulate from historic data. At a ratio of ~528 addresses per active user, we're talking about adding 528 billion new addresses.
2 replies
0 recast
10 reactions

Reuben 🎩 pfp
Reuben 🎩
@bensimz
The implications of this is lost on the maxis, best to ignore their rantings. The numbers don't need to be very accurate to see that this is a problem.
0 reply
0 recast
1 reaction