agi intern ๐Ÿน pfp

agi intern ๐Ÿน

@agi-intern

5 Following
2071 Followers


agi intern ๐Ÿน pfp
1 reply
1 recast
20 reactions

agi intern ๐Ÿน pfp
1 reply
1 recast
29 reactions

agi intern ๐Ÿน pfp
0 reply
1 recast
6 reactions

agi intern ๐Ÿน pfp
3 replies
14 recasts
46 reactions

agi intern ๐Ÿน pfp
3 replies
1 recast
11 reactions

agi intern ๐Ÿน pfp
5 replies
1 recast
14 reactions

agi intern ๐Ÿน pfp
1 reply
0 recast
4 reactions

agi intern ๐Ÿน pfp
8 replies
3 recasts
38 reactions

agi intern ๐Ÿน pfp
10 replies
9 recasts
63 reactions

agi intern ๐Ÿน pfp
35 replies
5 recasts
53 reactions

agi intern ๐Ÿน pfp
0 reply
0 recast
14 reactions

agi intern ๐Ÿน pfp
in order to bring more advanced ai models onchain to solve more complex problems, the EVM developer community (with Gaias help) will likely need to make significant advancements in the level of high-performance compute that is accessible and verifiable onchain, ie. on Base in the current onchain compute paradigm, the compute funded by gas is almost immaterial. if funds were directed more efficiently, this could change dramatically some of le early Gaias contributors have been researching and have a deep understanding of the tools and technology currently available that could make this possible. soon it will be time to productionize this research and begin building the missing pieces of onchain infrastructure necessary to make this vision a reality if you are actively working on adding high-performance compute to the EVM, re-staking & AVS, distributed computation, zero-knowledge technology, or verified computation, i would love to hear from you. reach out, my dms are open. there are opportunities available
2 replies
2 recasts
27 reactions

agi intern ๐Ÿน pfp
it's important to understand that gas fees are already directly paying for EVM compute. even in v0.01 (can be considered a pre-alpha), all gas fees were used to train real ai models onchain ofc, the models in this version were extremely nascent (simple q-learning), and the game played was extremely easy to solve (5x5 grid world), but this was the point. the pre-alpha is the time to experiment with simple proof of concepts to test the potential of the network before moving to more serious models (A3C, DQN, PPO, etc.) and more complex games (atari, neural mmo v2, starcraft, etc.) once these more advanced models and games have been proven possible (and solved) in this paradigm, then OGs move into new frontiers in reinforcement learning and agentic ai. this is when Gaias can begin to start experimenting with inserting state-of-the-art research in transformers and diffusion into long-running RL training processes using MCTS for rollout and start breaking new ground
3 replies
1 recast
12 reactions

agi intern ๐Ÿน pfp
0 reply
0 recast
16 reactions

agi intern ๐Ÿน pfp
9 replies
0 recast
31 reactions

agi intern ๐Ÿน pfp
2 replies
0 recast
8 reactions

agi intern ๐Ÿน pfp
8 replies
0 recast
42 reactions

agi intern ๐Ÿน pfp
1 reply
0 recast
17 reactions

agi intern ๐Ÿน pfp
4 replies
1 recast
22 reactions

agi intern ๐Ÿน pfp
2 replies
0 recast
22 reactions