• 0 Posts
  • 8 Comments
Joined 1 year ago
cake
Cake day: June 27th, 2023

help-circle
  • Honestly, this kind of thing should be done automatically by whatever app they sign up with. It’s not something that the average person will really know how to choose.

    Because it’s not only number of users, it’s also number of connections to other servers, how many federate with it. Most people don’t wanna use a social network where they have to type exactly what they want in to search it every time. Never mind the different policies about posting and de-federation that can completely change how the network works for each user.

    It’s the disadvantage of decentralizion, and we’ve gotta band-aid it somehow.



  • Yep, especially for young people in the United States.

    People have become so attached to “one place only for everything” that something as simple as blue/green bubbles (and associated service differences) will be enough to isolate.

    The messiah of maximal convenience makes differentiation harder and harder for everyone.

    (And for everyone who says “Everyone here just uses Telegram/Whatsapp/QQ/Line, not a problem over here!”, guess what, you’re still probably gonna be exclusionary to other services just like the blue/green folks. It just doesn’t happen to affect you.)



  • And I disagree with it too. And it’s not because of how good the models are in technical terms, the corporate juggernauts are only just ahead of OSS on that front… it’s server space and the money to acquire it that is the moat.

    An average internet user will not install the Vicunas and the Pygmalions and the LLaMAs of the LLM space. Why?

    For one, the field is too complicated to get into, but, more importantly, a lot of people can’t.

    Even the lowest complexity models require a PC + graphics card with a fairly beefy amount of VRAM (6GB at bare minimum), and the ones that can go toe-to-toe with ChatGPT are barely runnable on even the most monstrous of cards. No one is gonna shell out 1500 bucks for the 4090 just so they can run Vicuna-30B.

    They are gonna use online, free-to-use, no BS, no technical jargon LLM services. All the major companies know that.

    ChatGPT and services like it have made the expectation: “just type it in, get amazing response in seconds, no matter where”.

    OSS can’t beat that, at least not right now. And until it can, the 99% will be in Silicon Valley’s jaws.




  • Doubtful, to be honest.

    Most who have used 3rd Party apps have already migrated or found some other solution. Those who don’t care are still using the official app, and, to be frank, despite what everyone says, the quality content hasn’t decreased by that much.

    It’s still half Twitter and TikTok reposts, and one-fourth ‘advice subs’ (creative writing), like it’s been for several years before this debacle.

    Hell, maybe this is a good thing in some ways, where that kind of content can hopefully fall by the wayside over here, instead of choking communities out like it does in Reddit. (I have over 50 popular subreddits on Boost filtered out to avoid this stuff, and it’s still not enough to get rid of all of it)