Formerly @russjr08@outpost.zeuslink.net

  • 0 Posts
  • 8 Comments
Joined 10 months ago
cake
Cake day: December 7th, 2023

help-circle

  • I personally use Sleep as Android which comes with a bunch of options to help ensure you’ve actually woken up. I utilize the “captcha” option in which when I go to turn off the alarm, it displays a screen full of sheep and all of them but one are sleeping - you have to click the one that is “awake” in order to dismiss the alarm. I guess the process wakes up my brain just enough so that I don’t go back to sleep, whereas with a regular alarm that has just a simple dismiss button I’ll absolutely either hit dismiss or one of the volume buttons to turn off the alarm before I’ve fully woken up.

    I also have it set to buzz on my watch for 90 seconds before playing a sound on my phone (which escalates in volume) - I’ve not had a problem waking up with this in the years that I’ve been using it.

    There are other options too, such as answering math questions, scanning a QR code, pressing your phone to an NFC tag, heavily shaking the phone, one called “Say cheese!” that makes you smile as hard as you can and uses the camera to detect it, and one that you have to “laugh out loud”.


  • Hmm, gotcha. I just tried out a fresh copy of text-gen-webui and it seems like the latest version is borked with ROCM (I get the CUDA error: invalid device function error).

    My next recommendation then would be LM Studio which to my knowledge can still output an OpenAI compatible API endpoint to be used in SillyTavern - I’ve used it in the past before and I didn’t even need to run it within Distrobox (I have all of the ROCM stuff installed locally, but I generally run most of the AI stuff in distrobox since it tends to require an older version of Python than Arch is currently using) - it seems they’ve recently started supporting running GGUF models via Vulkan, which I assume probably doesn’t require the ROCM stuff to be installed perhaps?

    Might be worth a shot, I just downloaded the latest version (the UI has definitely changed a bit since I last used it) and just grabbed a copy of the Gemma model and ran it, and it seemed to work without an issue for me directly on the host.

    The advanced configuration settings no longer seem to directly mention GPU acceleration like it used to, however I can see it utilizing GPU resources in nvtop currently, and the speed it was generating at (the one in my screenshot was 83 tokens a second) couldn’t have possibly been done on the CPU so it seems to be fine on my side.