I did nothing and I’m all out of ideas!

  • 0 Posts
  • 37 Comments
Joined 2 years ago
cake
Cake day: June 11th, 2023

help-circle
  • I’ve never used oobabooga but if you use llama.cpp directly you can specify the number of layers that you want to run on the GPU with the -ngl flag, followed by the number.

    So, as an example, a command (on linux) from the directory you have the binary, to run its server would look something like: ./llama-server -m "/path/to/model.gguf" -ngl 10

    Another important flag that could interest you is -c for the context size.

    This will put 10 layers of the model on the GPU, the rest will be on RAM for the CPU.

    I would be surprised if you can’t just connect to the llama.cpp server or just set text-generation-webui to do the same with some setting.

    At worst you can consider using ollama, which is a llama.cpp wrapper.

    But probably you would want to invest the time to understand how to use llama.cpp directly and put a UI in front of it, Sillytavern is a good one for many usecases, OpenWebUI can be another but - in my experience - it tends to have more half baked features and the development jumps around a lot.

    As a more general answer, no, the safetensor format doesn’t directly support quantization, as far as I know


  • That’s the bad thing about social media. If no one was doing it before, someone is now!

    Jokes aside it’s possible, but with the current LLMs I don’t think there’s really a need for something like that.

    Malicious actors usually try to spend the least amount of effort possibile for generalized attacks, because you end up having to often restart when found out.

    So they probably just feed an LLM with some examples to get the tone right and prompt it in a way that suits their uses.

    You can generate thousands of posts while Lemmy hasn’t even started to reply to one.

    If you instead want to know if anyone is taking all the comments on lemmy to feed to some model training… Yeah, of course they are. Federation makes it incredibly easy to do.


  • Hangs, reboots or does it turn off? Depending on that it could be a plethora of things.

    Do you get any error prompts? Any red alerts when restarting?

    Did you check that the monitor works steadily on another system/OS?

    Did you try another DE, even on a live usb?

    Is the HDD/SSD healthy?

    Are all the fans working? Is it the thermal protection?

    Is the PSU healthy and or the power connection damaged?

    Does your system have a centralized logging like journalctl or can you reach the single log files to check and add more information?

    It could literally be anything, even aliens.


  • AFAIK it is still a tuning of llama 3[.1], the new Base models will come with the release of 4 and the “Training Data” section of both the model cards is basically a copy paste.

    Honestly I didn’t even consider the fact they would not be giving Base models anymore before reading this post and, even now, I don’t think this is the case. I went to search the announcements posts to see if there was something that could make me think about it being a possibility, but nothing came out.

    It is true that they released Base models with 3.2, but there they had added a new projection layer on top of that, so the starting point was actually different. And 3.1 did supersede 3…

    So I went and checked the 3.3 hardware section and compare it with the 3 one, the 3.1 one and the 3.2 one.

    3 3.1 3.2 3.3
    7.7M GPU hours 39.3M GPU hours 2.02M GPU hours 39.3M GPU hours

    So yeah, I’m pretty sure the base of 3.3 is just 3.1 and they just renamed the model in the card and added the functional differences. The instruct and base versions of the models have the same numbers in the HW section, I’ll link them at the end just because.

    All these words to say: I’ve no real proof, but I will be quite surprised if they will not release the Base version of 4.

    Mark Zuckerberg on threads

    Link to post on threads
    zuck a day ago
    Last big AI update of the year:
    •⁠ ⁠Meta AI now has nearly 600M monthly actives
    •⁠ ⁠Releasing Llama 3.3 70B text model that performs similarly to our 405B
    •⁠ ⁠Building 2GW+ data center to train future Llama models
    Next stop: Llama 4. Let’s go! 🚀

    Meta for Developers

    Link to post on facebook
    Today we’re releasing Llama 3.3 70B which delivers similar performance to Llama 3.1 405B allowing developers to achieve greater quality and performance on text-based applications at a lower price point.
    Download from Meta: –

    Small note: I did delete my previous post because I had messed up the links, so I had to recheck them, whoops




  • I honestly don’t really remember the main quest progression (I last did it a really long time ago), but I think there were a couple of steps that were time or exploration/research gated…

    I can say that the story of Artemis has an ending, you should probably just try go on and research some Archive related things on the Base computer or jump around to new systems

    As a general rule, IMHO, to really enjoy it you should treat NMS as a sandbox game that happens to have a story that sometimes pops up






  • So, I can’t install aur packages via pacman?

    Nope, you have to do it manually or using an helper that abstracts the manual work away.

    AUR packages, or to be more precise the PKGBUILD files, are recipes to compile or download stuff outside from the official repositories, manage their deps and installing them on the system.

    You should always only run PKGBUILD files that you trust, they can do basically anything on your system. Checking the comments of the package in the aur repo is a good practice too.

    Also Are you quoting certain nExT gEn gAmE guy?

    …maybe


  • Also in wiki they didn’t mention anything about OpenSSL?

    Sorry, that was my bad, I wrote OpenSSL instead of openvpn. That one is probably needed too, but you should not have to pull it manually.

    Generally speaking the ArchWiki is one of the best, more structured and well maintained source of information about Linux things even for other distros, but it can too be outdated, so you should always check if the info is valid. In this case it seems so.

    In theory you should be able to just install proton-vpn-gtk-app using one of the many AUR helpers and it should Just Work™. Paru and yay are the most commonly used ones - as far as I know - and they wrap around pacman too, so you can use them to do everything packages related. Usually Arch related distro use one of them, for example EndeavourOS have yay already installed.

    At worst when you try to start protonvpn the GUI will not appear or immediately crash: if that happens, usually, you can try and run the program from the Shell and see what kind of error it returns and work your way from there. Checking if the deps listed in the wiki are installed is always a great first step.


  • Reading rorschac’s comment I assume both OpenSSL and wireguard are already installed on CachyOS, or anyway pulled by the aur package.

    If you want to make sure you can install them explicitly before protonvpn:

    paru openvpn wireguard-tools
    

    or using yay or the vanilla pacman -Syu --needed openvpn wireguard-tools (it will sync and update the system too) or how it is suggested for CachyOS to install packages. I repeat I’ve no direct experience with that one.

    If you are scared to mess things up you can always spin up a VM with CachyOS and try to install it inside that. If it all works you can then do the same on your main OS.

    As a general advice, only run in your shell commands that you are sure about.




  • The Heroic Games Launcher is (IMHO) by far the best interface to gog you can have on linux.

    You can find it on the AUR if you use arch, which makes it pretty straightforward to install.

    The next version will integrate with the Galaxy API using the comet project, which should make it even better.

    The only problem I had with it is that, once upon a time, there was a bug with downloading some games (Cyberpunk 2077, in my case) and I had to compile the git version of Gog-dl and target that in the settings… but the fact I could even do that is great by itself.