• 0 Posts
  • 42 Comments
Joined 3 years ago
cake
Cake day: June 14th, 2023

help-circle

  • I had a few ideas, I’m suspicious that handbrake is falling back to CPU, maybe check the logs of the container to make sure it isn’t falling back to CPU decoding. Otherwise here are a few things I would check next:

    • If you are not using docker locally so you are already doing this, you will need to configure the docker container to pass through the GPU for quicksync to work inside the container.
    • If you are already doing that then I would make sure the device is the same name on the synology, it probably is but just to be sure.
    • you will likely need to add your user to the video and/or render group on the synology if you haven’t, especially if you are running the container as your user instead of root
    • make sure you are reading and writing to volumes that use bind mounts and not docker volumes, overlayfs is not what I would call fast and writing especially.

  • So I had a few thoughts. I’m not sure that you can use the docker device flag with a directory as you have there, I think it expects a device node, you can pass that directory as a volume (-v) though.

    If that doesn’t work you might also try running the VM with host-passthrough mode set on the CPU as well if it isn’t set that way already, sometimes that is also required for pass through to work from my experience. Also, make sure you passed through the whole device node, sometimes there are audio devices you have to pass through with the GPU device or you will get odd errors like those initialization ones you had. I’m not sure if this is the case for Intel iGPU though offhand though. Are you able to use intel_gpu_top on the VM to access the GPU? None of that is necessarily specific to proxmox though (but probably applies to anything libvirt powered) so YMMV.

    Edit: I realized you may not know what a “device node” is, that is the full path to the device, like /dev/dri/renderD128 vs /dev/dri which is actually a directory.



  • Oh it is certainly not just you, I am sometimes confused reading them even for commands I have used for years and I know what flag I am looking for but don’t remember the exact syntax or something hah! I am glad they are there but they are definitely not a complete guide to any command, especially built-ins.

    Interestingly, this is something AI has been very useful for to me, less searching because I can describe the outcome I want and it figures out what I am talking about generally.



  • Okay so when you say “unplug the power” do you mean shut it down first or just pull the plug? The latter is a great way to corrupt your storage pools as ZFS uses memory for read and write cache etc by default. You definitely need to do a graceful shutdown especially if there is data that was recently written to disk, that’s why a UPS is so recommended. That said you can usually import an existing pool when that happens, I think there is a UI menu for it now.


  • As others have said, be careful with fans if they are large, many of the plugs don’t have a very high wattage rating and are all definitely rated for 15a at max usually, you might consider a smart relay instead (like a Shelly or something).

    That said I have switches and plug-in and in-wall relays from Aqara (zigbee) and TP-Link (WiFi) and zooz (zwave) and all are fine and do the job. Not all support power monitoring if that is something that matters to you, it’s not a universal feature.





  • I mean sure but that’s a lot of words to say “I didn’t read the directions and no one caught it in a merge request review because no one else read the directions either.”

    Their documentation and examples are pretty easy to read and the site parameter is explained in the getting started guide and even linked from the readme for the JavaScript sdk, and in lots of sample configurations so I’m not sure how this made it into a release and then no one noticed the missing metrics for eleven days, sounds like lots of issues in that shop.

    The behavior of the sdk isn’t great but the proposed solution wouldn’t work because you can use custom endpoints for all of the components using endpoints on domains you own anyway.


  • Well 5ghz requires more power, has less range, and needs its own antenna so for microcontrollers this makes it pretty pointless for devices that need range and low bandwidth for sending sensor updates, especially those that are battery powered. 5ghz can also have its own issues in cities if you have a lot of use of the DFS bands as well as being worse at traversing reinforced concrete.

    Also, a 2.4ghz radio can also sometimes support other things like zigbee, BT, and BLE which can be used for other functions.

    For what it’s worth, I have probably 50 WiFi devices and the majority of them are 2.4ghz sensors or switches and other low bandwidth tasks and I don’t have any issues, even when living in an apartment complex. If you are having issues you might need different hardware or more access points or something.

    Anyway, all that to say that 2.4ghz definitely still has a lot of utility today.




  • Yeah I’m with you, I read most of it but I just don’t know where the disdain comes from. At most scales of infrastructure anymore you can use them interchangeably because the difference is immaterial in practical applications.

    Like if I am going to provision 2TB I don’t really care if it’s 2000 or 2048GB, I’ll be resizing it when it gets to 1800 either way, and if I needed to actually store 2TB I would create a 3TB volume, storage is cheap and my time calculating the difference is not.

    Wait until you learn about how different fields use different precision levels of pi.




  • I think what this person is saying is that systems and services have been monitored for metrics and logs for a long time, I know I have been doing it for more than 20 years across many OS, hardware platform, and software stack. The tools and depth of the integrations have definitely changed and gotten way better and more sophisticated but I definitely made systems that monitored and healed themselves to varying levels of efficiency since at least using Nagios in 2003 (I’m getting Perl PTSD flashbacks now hah).

    One thing that has definitely gotten better in the last 5 or so years though is code level instrumentation and tracing as well as the higher level correlation tools. I have also seen more developers and vendors way more willing to implement monitoring features in their code from the beginning leading to more data and less duct tape and guessing which has been FANTASTIC.

    Anyway, great post though, the monitoring arena is definitely way more awesome than ever before these days that is for sure.