๐๐ช๐พ๐ก๐ก๐พ
@gm8xx8
161 Following
131518 Followers
0 reply
3 recasts
18 reactions
1 reply
2 recasts
21 reactions
6 replies
0 recast
12 reactions
3 replies
0 recast
29 reactions
2 replies
2 recasts
150 reactions
0 reply
1 recast
24 reactions
1 reply
1 recast
7 reactions
Oasis is a 500M parameter AI model that generates real-time, interactive gameplay directly from user input, eliminating the need for a traditional game engine. Rendering immersive experiences at 20 fps with zero latency, Oasis introduces โGenerative Interactive Experiences,โ where users can interact with AI-created video or game content in real time, responding instantly to text or voice commands. Built on Decartโs optimized ViT + DiT architecture, this promises more dynamic, personalized digital experiences. Oasis is available open-source, complete with code, weights, and a live demo for exploration.
demo: https://oasis.decart.ai/welcome 1 reply
2 recasts
17 reactions
1 reply
0 recast
13 reactions
0 reply
1 recast
13 reactions
0 reply
0 recast
17 reactions
1 reply
2 recasts
22 reactions
0 reply
0 recast
3 reactions
4 replies
1 recast
6 reactions
6 replies
5 recasts
34 reactions
1 reply
1 recast
4 reactions
0 reply
1 recast
15 reactions
0 reply
0 recast
4 reactions
Zyphra introduced Zamba2-7B, a new SLM that surpasses competitors like Mistral-7B, Gemma-7B, and Llama3-8B in both performance & efficiency. Designed for on-device use, consumer GPUs, and enterprise applications.
Zamba2-7B:
> Achieves top benchmark results with faster inferenceโ25% quicker first token generation and 20% more tokens per second compared to similar models.
> Utilizes Mamba2 blocks w/ two shared attention blocks interleaved in an ABAB pattern, enhancing cross-sequence dependencies.
> Trained on 3T tokens, combining proprietary Zyda data w/ high-quality, deduplicated datasets, followed by a specialized annealing phase to improve model precision.
> Zamba2-7B achieves high throughput and low memory usage, eliminating the need for KV-caches & leveraging efficient Mamba blocks optimized for parallel hardware.
> Model was trained on 128 H100 GPUs over 50 days and will be available open-source under Apache 2.0. w/ integrations on Huggingface and PyTorch. 1 reply
9 recasts
50 reactions
0 reply
1 recast
3 reactions