Damus
librekitty profile picture
librekitty
@librekitty

FOSS zealot, game developer, and lifelong learner πŸ‘ΉπŸ‘ΎπŸ“–

freedom maxi, solarpunk, cypherpunk, and anarcho-communist (anti-statist/anti-stalinist) πŸŒžπŸ’» πŸŸ₯⬛
centralization is the root of all evil

looking to connect through this protocol πŸ’œπŸ«‚
zaps/XMR to me is optional, feel free to save your sats πŸ’œβš‘

πŸ‡ΊπŸ‡ΈπŸ³οΈβ€πŸŒˆπŸ΄β€β˜ οΈ

contact πŸ’¬:

- white noise: npub1avxs6urpqyflcrua027e53kure5emfwhhqf5khf2np7rcc0gkvdqcwsa2t
- simplex: https://smp18.simplex.im/a#I1NyhGz6I1w-MTn-hS8IabZJoCNL3mZ7hkrqh1WAN2M
- matrix: @mewmeow:matrix.org
- XMPP: [email protected]

BTC:
bc1q464gm5xf7fhj4pjww8ppymksea0wm4kpd9ype3

monero/XMR:
8BooBQNxWx7YE5DqqtSeAE9a6vukxM82cHggqHeoNWs8WVucewTxhRdAf5B7Euuh35igNRYqG2b5qgo9kPbWU51eN1AZ5y6

Relays (5)
  • wss://nos.lol/ – read & write
  • wss://relay.pleb.one/ – read & write
  • wss://basspistol.org/ – write
  • wss://nostr.mom/ – read
  • wss://basspistol.org/inbox – read

Recent Notes

note1mrpv5...
librekitty profile picture
i love it!
the difference between 26b (moe) and 31b for writing is pretty big though
but i think i'll use 26b for coding and stuff, faster and more VRAM for context

i'm yet to try the smaller models on my phone, excited to try that too
eardiod · 9h
makes two of us, mine is also too old :)
eardiod · 17h
Did you use the compressed model from nvidia? I so far tried only 4bit og, but this new nvidia compression seems to be even better
Uno · 1d
I'm testing my new client! Specifically DMs, I'm gonna send you a message all you gotta do is reply if you get it.
πŸ‡΅πŸ‡Έ whoever loves Digit · 1d
But let us see pre edited versions of course
GrapheneOS · 1d
nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqwp2lcskf5cu2k44dq4cpdgk7lt8f0d0wecc68fayvun6mprcn8jsdf6gu6 nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpq0r8xl2njyepcw2zwv3a6dyufj4e4ajx86hz6v4ehu4gnpupxxp7spmxr3j nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqjfn4ghffz7uq7urll...
blackcat · 1d
πŸˆβ€β¬›πŸ–€
librekitty profile picture
anyone else still need to use tricks to remember how to spell? πŸ”‘

to this day i still spell restaurant 🍽️ as "rest😴 + aura✨ + nt🧠"
101❀️3πŸ’œ2πŸ€™1
librekitty · 2d
btw, on an unrelated spelling note, homeowner is secretly hoMEOWner πŸ πŸˆβ€β¬›
titus · 1d
Jim Carrey taught me how to spell beautiful when I saw Bruce Almighty as a youngster https://www.youtube.com/watch?v=6K3UpktQH9w&t=25
Kayne · 1d
I pronounce words in my mind how they are spelt, silent letters aren't silent, and anything that's spelt with re instead of er for instance, like centre or litre it's re instead of er when I hear the word in my mind.
skepticles · 1d
β€œThe way I remember my name is-β€œ -Luanne Platter
πŸ‡΅πŸ‡Έ whoever loves Digit · 1d
The most giga based people have some misspellings on the record because they use nostr where you can't edit podts
nicodemus · 1d
This is true, GPUs are faster for inference. But you'll also be consuming 1500 watts, have to deal with those thermal issues, and still struggle to fit a model larger than 32B with decent quantization. Alternatively, the 395 chips and their NPU are doing pretty good. Combine 2 of them and you're lo...
Gigi · 2d
considering buying hardware to run everything locally. Would should I buy? #asknostr
librekitty profile picture
intel is seriously competitive for price-to-VRAM, but i don't know about compatibility

NVIDIA is usually the clear winner for performance, 5xxx series/blackwell has support for NVFP4 quantized models
https://developer.nvidia.com/blog/introducing-nvfp4-for-efficient-and-accurate-low-precision-inference/

but you could also do like, multiple 3090s or something

hope this helps
2
librekitty · 2d
you can also go the CPU route with tons of RAM, but inference speed will be terrible compared to GPU accellerated
zaytun · 1d
I think dual 3090s would be preferable to fx a dgx spark with regards to inference speed, no? vRAM speed is higher I believe. Downside is model size limit is obviously lower on 48 gb vRAM than 128gb unified of the dgx spark.