Posts
481
Following
7
Followers
15
the burr's test account, sysadmin since never, forced to sysadmin since forever
there is no good cop there is no good militaryperson.

KILL 'EM ALL 1997

(art by @/garbanzzobean on Instagram)
@sierra bongwater turned freakywater
0
0
0
@sierra all ts effort for 2 likes and 1 boost vro 😭✌️
0
0
0
@sierra is this two balls and a chud
1
1
2
i asked grok to rewrite all of akkoma in rust tmrw kill-corp will be 67x faster 🙏 trust
0
3
7
@sierra idk how but ur finetuned model is somehow constantly using the chain-of-thought data even when it isn't enabled which is fucking funny
0
0
1
fine-tuning a 1B was such a bad idea
0
0
0
@sierra "grok can u make a readme.md out of what u just told me" ahh
0
0
0
who is this diddy blud on the LLM fine-tuning script
0
0
0
re: I made Bongwater-1.1 invent 20 new strings of text, and this is the result
Show content
@sierra she's so smart
0
0
1
@sierra you're making it think?? are you also fine-tuning the thinking part????
1
0
0
@sierra all LLMs are just transformers so dw abt that! :3
0
0
0

@sierra i lied on accident there is no Ministral 1B and even 3B is commercial-only. fuck. anyways

i don’t think you will be able to avoid Python sadly as the two biggest open-source frameworks to train AI models are PyTorch and TensorFlow, which are both Python…

as for how, I’m looking at it right now, as I haven’t done it since GPT-2, but i’m very curious! :)

0
0
0
@sierra there's ministral (Mistral, european), gemma (da goog), maybe da Qwen's (Alibaba)? idk how good Qwen is anymore within the world of SLMs... but anyways yea these are the 3 big ones :3

Ministral and Gemma is available in 1B so u shouldnt have much problem, idk abt Qwen tho. hope that helps! drgn_happy
1
0
1
@sierra why is there a small PINGAS for the next dialog arrow
1
0
0
@sierra the normal evolution of burrs.s3
0
0
0
outbound federation is broken on oomfie city :3
0
0
0
this is yet another federation test // ignore this thanku :)
1
0
1
@sierra green goop nitro badge
0
0
1
Show older