• 0 Posts
  • 36 Comments
Joined 1 year ago
cake
Cake day: July 4th, 2023

help-circle





  • I agree that UI should always take priority. I shouldn’t have to do anything to guarantee this.

    I have HZ_1000, tickless kernel with nohz_full set up. This all has a throughput/bandwidth cost (about 2%) in exchange for better responsiveness by default.

    But this is not enough, because the short burst UI tasks need near-zero wake-up latency… By the time the task scheduler has done its re-balancing the UI task is already sleeping/halted again, and this cycle repeats. So the nice/priorities don’t work very well for UI tasks. Only way a UI task can run immediately is if it can preempt something or if the system has a somewhat idle CPU to put it on.

    The kernel doesn’t know any better which tasks are like this. The on-going EEVDF, sched_ext scheduler projects attempt to improve the situation. (EEVDF should allow specifying the desired latency, while sched_ext will likely allow tuning the latency automatically)


  • No, I definitely want it to use as many resources it can get.

    taskset -c 0 nice -n+5 bash -c 'while :; do :; done' &
    taskset -c 0 nice -n+0 bash -c 'while :; do :; done'
    

    Observe the cpu usage of nice +5 job: it’s ~1/10 of the nice +0 job. End one of the tasks and the remaining jumps back to 100%.

    Nice’ing doesn’t limit the max allowed cpu bandwidth of a task; it only matters when there is contention for that bandwidth, like running two tasks on the same CPU thread. To me, this sounds exactly what you want: run at full tilt when there is no contention.


  • The kernel runs out of time to solve the NP-complete scheduling problem in time.

    More responsiveness requires more context-switching, which then subtracts from the available total CPU bandwidth. There is a point where the task scheduler and CPUs get so overloaded that a non-RT kernel can no longer guarantee timed events.

    So, web browsing is basically poison for the task scheduler under high load. Unless you reserve some CPU bandwidth (with cgroups, etc.) beforehand for the foreground task.

    Since SMT threads also aren’t real cores (about ~0.4 - 0.7 of an actual core), putting 16 tasks on a 16/8 machine is only going to slow down the execution of all other tasks on the shared cores. I usually leave one CPU thread for “housekeeping” if I need to do something else. If I don’t, some random task is going to be very pleased by not having to share a core. That “spare” CPU thread will be running literally everything else, so it may get saturated by the kernel tasks alone.

    nice +5 is more of a suggestion to “please run this task with a worse latency on a contended CPU.”.

    (I think I should benchmark make -j15 vs. make -j16 to see what the difference is)








  • JATth@lemmy.worldtoTechnology@lemmy.world*Permanently Deleted*
    link
    fedilink
    English
    arrow-up
    16
    ·
    edit-2
    2 months ago

    So, now they are slowly (or immediately and forever, I don’t know the time span) injecting propaganda into their clone of wikipedia and they are simultaneously thus admitting they are doing it. (to further brainwash the russian citizens)

    So lettme repeat: FUCK PUTIN, and stuff your rubber clones in your ass. (which there are many of)



  • JATth@lemmy.worldtoTechnology@lemmy.worldUnsmart a smart TV
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    1
    ·
    3 months ago

    The in-rush of endorphins when the modded LCD thing actually worked probably knocked you out?

    A bit of a side story: I disassemble probably 1–5 panels in a week. (For recycling, it takes about 20-45min per panel.) The flat cables alone are so flimsy, I would say just assembling a display again from known, matched and functional parts would take days. I would triple or quadruple that to assemble a display from random parts. Considering this, that modded panel is quite amazing to me.

    In standalone panels, the PSU has a chance of killing you: the main capacitors can hold multiple joules worth of nope, charged to about 400v. So, if the main caps are not allowed to discharge (if they discharge at all that is), there is a possibility of death when disassembling a panel with an integrated PSU. Waiting “5min” is bad; a PSU needs to be unplugged for a day or two at least before the charge drains out.


  • JATth@lemmy.worldtolinuxmemes@lemmy.worldArch with XZ
    link
    fedilink
    arrow-up
    5
    arrow-down
    1
    ·
    edit-2
    3 months ago

    I just did: “rm -rf xz

    pacman -Syu
    find / -name "*xz*"  | sort | grep -e '\.xz$' | xargs -o -n1 rm -i 
    pacman -Qqn | pacman -S -
    

    (and please, absolutely don’t run above as root. Just don’t.) I carefully answered to retain any root owned files and my backups, despite knowing the backdoor wasn’t included in the culprit package. This system has now “un-trusted” status, meaning I’ll clean re-install the OS, once the full analysis of the backdoor payload is available.

    Edit: I also booted the “untrusted” system without physical access to the web, no gui, and installed the fixed package transferred to it locally. (that system is also going to be dd if=/dev/zero'd)


  • I would highly recommend Curve25519, etc., just because such keys are faster and less common than RSA public-private keys in today’s world. RSA 2048-bit keys are considered weak today, while the Curve25519 256-bit keys remain stronger. Also, the ChaCha20-Poly1305 cipher has an interesting backstory and doesn’t necessarily need hardware acceleration (which, in theory, could be borked by the HW-vendor) to obtain good performance.

    Unfortunately, some SSH front-ends don’t play nice with Curve25519 public-private keys yet… (I’m pointing at the putty SSH client, but that may have improved from the last time I had to use it)