A software developer and Linux nerd, living in Germany. I’m usually a chill dude but my online persona doesn’t always reflect my true personality. Take what I say with a grain of salt, I usually try to be nice and give good advice, though.

I’m into Free Software, selfhosting, microcontrollers and electronics, freedom, privacy and the usual stuff. And a few select other random things, too.

  • 2 Posts
  • 203 Comments
Joined 4 months ago
cake
Cake day: June 25th, 2024

help-circle

  • Very good idea. I mean there are frameworks for programmers to do exaclty that, like LangChain. But I also end up doing this manually. I use Kobold.cpp and most of the times I just switch it to Story mode and I get one lage notebook / text area. I’ll put in the questions, prompts, special tokens if it’s an instruct-tuned variant and start the bullet point list for it. Or click on generate after I’ve already typed in the chapter names or a table of contents. Or opened the code block with the proper markdown. So pretty much like what you lined out. It’s super useful to guide the LLM into the proper direction. Or steer it back on track with a small edit in its output, and a subsequent call to generate from there.


  • I read about some PCBs that let you repurpose old laptop screens but I’m pretty sure that’s more complicated with phones and tablets. These have tiny and very specialized electronics. Oftentimes not built in a modular way. And people tend to break them, give them away or dispose of them. I don’t see people repurposing these devices.

    And it’d be hard to bypass the boot time and Android experience. Sometimes you can flash a custom ROM like LineageOS. Though, that’s still Android. Other operating systems aren’t really a thing within that ecosystem.

    I like to flash LineageOS and then use extra phones/tablets as a kitchen radio or TV or as a TAN generator.

    You could also install Termux and install Linux software. Like a webserver or something like that.










  • Mmhm, I’m not sure if I’m entirely on the same page. Admins have complained. Users would like to run their own instances, but they can’t as the media cache is quite demanding and requires a bigger and costly virtual server. And we’re at the brink of DDoSing ourselves with the way ActivityPub syncs (popular) new posts throughout the network. We still have some room to grow, but it’s limited due to the protocol design choices. And it’s chatty as pointed out. Additionally we’ve already had legal concerns, due to media caching…

    Up until now everything turned out mostly alright in the end. But I’m not sure if it’s good as is. We could just have been lucky. And we’re forced to implement some minimum standards of handling harassment, online law, copyright and illegal content. Just saying we’re amateurs doesn’t really help. And it shifts burden towards instance admins. Same for protocol inefficiencies.

    I agree - however - with the general promise. We’re not a big company. And that’s a good thing. We’re not doing business and not doing economy of scale here. And it’s our garden which we foster and have fun at.




  • Correct answer. There is no general purpose AI model that can fit ino 1GB. These small models exist, but they do very specific small tasks. Sentiment analysis, object detection, word embeddings for vector databases…

    For coding, answering questions and generating text, you’d need like 6-8GB minimum. For maths way more than that and they’ll still be throwing dice instead of giving correct answers.