𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp

𝚐π”ͺ𝟾𝚑𝚑𝟾

@gm8xx8

323 Following
129161 Followers


𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
β›½οΈŽ
0 reply
0 recast
3 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
☺︎
0 reply
0 recast
1 reaction

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
βœ”οΈ https://github.com/Mozilla-Ocho/llamafile/releases/tag/0.8.10
0 reply
0 recast
1 reaction

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
you know what impresses me? making a small model work well.
2 replies
1 recast
5 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
undertrained behemoth…but why? i’ll have to dive into this one later ☺︎ Tele-FLM-1T an open LLM released by BAAI and TeleAI with 1 trillion parameters. - supports chinese and english - apache 2.0 license - cost-effective progressive pre-training - enhanced with input/output scalers, RoPE, RMSNorm, and SwiGLU https://huggingface.co/CofeAI/Tele-FLM-1T
0 reply
0 recast
4 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
food for thought https://warpcast.com/gm8xx8/0x13fafc1d
2 replies
0 recast
6 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
it’s no surprise that OpenAI is getting into search. the rumors were spot on, and the publisher partnerships gave it away. 😈 let the search wars begin!
1 reply
0 recast
8 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
this is a significant breakthrough, marking the first successful use of MCTS with LLMs for theorem proving. 😈 https://deepmind.google/discover/blog/ai-solves-imo-problems-at-silver-medal-level/
1 reply
0 recast
8 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
SearchGPT https://openai.com/index/searchgpt-prototype/
1 reply
0 recast
3 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
deepseek πŸ‘Œ
0 reply
0 recast
2 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
1st time we see an open source model in the top spot on EQ-Bench https://eqbench.com
0 reply
0 recast
2 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
0 reply
0 recast
10 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
sqlite-lembed: A SQLite extension for generating text embeddings locally i’ll leave this here… https://alexgarcia.xyz/blog/2024/sqlite-lembed-init/index.html
0 reply
0 recast
8 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
(more links) Mistral Large 2: Quality, Performance & Price Analysis: https://artificialanalysis.ai/models/mistral-large-2 Mistral Large 2: API Provider Benchmarking & Analysis: https://artificialanalysis.ai/models/mistral-large-2/providers Mistral Large 2: - 123B parameters - 128k context window - natively multilingual - strong code & reasoning - SOTA function calling - open weights for non-commercial use its also smaller and cheaper to host than models like Llama 405B. looks like this continues Mistral’s trend of smaller, specialized models.
0 reply
0 recast
5 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
0 reply
0 recast
1 reaction

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
who wants to see Mistral-123B evaluated on SEAL against Llama 3.1? πŸ™‹β€β™‚οΈ
0 reply
0 recast
2 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
mistral large 2 model checkpoints https://huggingface.co/mistralai/Mistral-Large-Instruct-2407
0 reply
1 recast
2 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
Mistral Large 2 features: - 128k context window - supports 80+ coding languages - 123 billion dense parameters - matches Llama 3 405B in coding with 1/3 the parameters - superior function calling compared to GPT-4o and Claude - released under the Mistral Research License and it’s only wednesday πŸ”₯ https://mistral.ai/news/mistral-large-2407/
1 reply
0 recast
2 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
i’ll leave this here
1 reply
0 recast
5 reactions

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
☺︎ https://warpcast.com/swabbie.eth/0x7c10c023
0 reply
0 recast
2 reactions