I have an HP g3 mini and a Dell Optiplex flying around, both similarly specced. The HP has an i5 6500t and 16gb DDR4 RAM, the Dell has 8gb DDR3l, so nothing too different.

However, the Dell draws around 15W while idle, the HP one 5W.

The only difference I could think of (and that is in my power to change) is the PSU. The Dell has one of those SFF PSU for up to 180W while the HP has an external 65W power brick with a barrel jack.

So my question is: Does anyone have experience with one of those Pico PSUs? I guess they should be more efficient? I’m not planning to put anything power hungry into the optiplex.

  • 486@kbin.social
    link
    fedilink
    arrow-up
    4
    ·
    1 year ago

    Particularly in low-load scenarios there can be quite a big difference when it comes to PSU efficiency. While newer ATX PSUs have become better with regards to efficiency at low load, a Pico PSU can still be quite a bit better. Older ATX PSU often don’t even reach 60 % efficiency at 5 % load (which would be a typical load for such a system at idle), sometimes considerably less than that. At the same load a Pico PSU can easily be at 85 % efficiency.
    Of course, at higher loads the difference is way smaller.

    • cstine@lemmy.uncomfortable.business
      link
      fedilink
      arrow-up
      0
      ·
      1 year ago

      That’s fair; I wasn’t really considering how poorly performing PSUs were at extremely low loads, despite knowing that they are.

      Odd that a random brick would be substantially better than a same-era actual PSU, but I suppose it’s hard to say without more specifics.

      • shadeless@discuss.tchncs.de
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        Switching power supplies (“bricks”) are generally more efficient than linear power supplies because they lose less energy as heat. that’s were the difference comes from. (Of course they have drawbacks as well, like increased noise)

    • MangoPenguin@lemmy.blahaj.zone
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      1 year ago

      True, but at that point 60% vs 85% on a load of say 10W is a difference of about 5W draw from the wall. If you live somewhere with high electricity rates it might be worth it, but otherwise at the usual $0.10/kwh that’s about $4 per year.

      The other situation where it makes sense is off-grid setups, where wasted power is a big deal.

      • 486@kbin.social
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        OP mentioned $0.40/kWh, so that would be about $17 per year with a 5 W difference.