A software developer and Linux nerd, living in Germany. I’m usually a chill dude but my online persona doesn’t always reflect my true personality. Take what I say with a grain of salt, I usually try to be nice and give good advice, though.

I’m into Free Software, selfhosting, microcontrollers and electronics, freedom, privacy and the usual stuff. And a few select other random things as well.

  • 3 Posts
  • 478 Comments
Joined 4 years ago
cake
Cake day: August 21st, 2021

help-circle



  • I think after initial installation, you open a browser with the post-installation step and configure a username and password there. I’m not entirely sure, it’s been some time since I did it. But depending on installation method, I don’t think it has a provided password.

    General password advice: Check caps lock, and if you use like a German keyboard if ‘z’ and ‘y’ are swapped.



  • Yes. Steam is available on Linux, pretty easy to install and it comes with a compatibility layer (Proton) which works quite well.

    Linux is a bit different than Windows. But I’d say just using it is about as complicated as using Windows. You’ll just have to try and see whether you like it. And if it’s hard or easy for you to relearn a few things. I mean if you’re in the Browser and Steam all day, those will be the same applications and also look and work the same way. Other than that you could face some issues with gaming hardware and you have to fiddle with things, or everything works out of the box. You can’t tell beforehand.



  • Most frontends should have you covered and scale down the image appropriately (and automatically). I’m not entirely sure about that. I think working on resolutions higher than supported should either not work due to the image encoder, or lead to degraded performance. Usually they’re scaled to what the model needs, somewhere within the pipeline. You can crop them if you like, that sheds off some pixels. Or split them up and feed them in one part after the other, if that somehow makes sense. But I bet with most software you can just upload a random image and it’ll do whatever scaling is required for you.


  • Uh sorry, no. Since I don’t use vllm, I don’t know. It certainly depends on the method you choose. The activation-aware ones will use a lot of resources. Just truncating the numbers to 8bit (or whatever) uses very little resources, I did that on my laptop. Also depends on model architecture and the size of the model you feed in. Since you gave an 32b parameter model, I’d expect it to take about 64GB loaded fully into memory using 16bit floating point numbers. (32 million billion times 2 bytes.) But I don’t really know whether quantization methods load the full thing, or if they do it in chunks. You’d have to google that or try it.



  • I think MiniCPM advertises with strong OCR capabilities. Maybe try that one. But is that 384 number correct? Seems to me even the common Llama 3.2 models do images up to 1120x1120px for a while and that should be enough. MiniCPM can do even more (1.8 million pixels).

    Setup shouldn’t be too hard. Lots of interfaces and inference frameworks have vision capabilities these days. It should work pretty much out of the box with something like Ollama, AnythingLLM or llama.cpp based stuff.

    Other approaches include doing OCR by more traditional means and then feed the result into an LLM. Just make sure to check that whatever you use can read Chinese.

    I’m not entirely sure what your prompt should look like. Obviously attach the image. With ChatGPT I believe I had a bit better results when I told it to first copy and write down the text, and then translate it in a second step. And then do whatever I want after that. I’m not sure what it does if you make it do everything in one go.


  • Maybe Discover isn’t the best choice. I believe that’s made for the KDE desktop and Gnome should come with “Gnome Software” per default?! I’m not entirely sure what kind of concept Fedora has. I usually use the command line or some of the older package managers with more options and settings, so I can’t really tell what’s best here. These modern and shiny ones also regularly confuse me and I install Flatpaks by accident or whatever. Maybe try something else, maybe the Fedora community has some recommendations for a better one.







  • I think that’s a size where it’s a bit more than a good autocomplete. Could be part of a chain for retrieval augmented generation. Maybe some specific tasks. And there are small machine learning models that can do translation or sentiment analysis, though I don’t think those are your regular LLM chatbots… And well, you can ask basic questions and write dialogue. Something like “What is an Alpaca?” will work. But they don’t have much knowledge under 8B parameters and they regularly struggle to apply their knowledge to a given task at smaller sizes. At least that’s my experience. They’ve become way better at smaller sizes during the last year or so. But they’re very limited.

    I’m not sure what you intend to do. If you have some specific thing you’d like an LLM to do, you need to pick the correct one. If you don’t have any use-case… just run an arbitrary one and tinker around?


  • Thanks! I’ve updated the link. I always just use Batocera or something like that, which has Emulationstation and Kodi set up for me. So I don’t pay a lot of attention to the included projects and their development state…

    I didn’t include this, since OP wasn’t mentioning retro-gaming. But Batocera, Recalbox, Lakka, RetroPie are quite nice. I picked one which includes both Kodi and Emulationstation and I can switch between the interfaces with the gamecontroller. I get all the TV and streaming stuff in Kodi, and Emulationstaation launches the games. And I believe it can do Flatpaks and other applications as well.