• lemmydividebyzero@reddthat.com
    link
    fedilink
    English
    arrow-up
    86
    arrow-down
    1
    ·
    6 days ago

    4k is enough, 60fps is enough, no smart or AI stuff is perfectly fine…

    What about reducing the energy consumption? That’s an innovation I want.

    • HereIAm@lemmy.world
      link
      fedilink
      English
      arrow-up
      67
      arrow-down
      1
      ·
      6 days ago

      I hope you mean 60hz is enough for TVs. Because I certainly don’t want that regression on my monitor 😄

      • 1984@lemmy.today
        link
        fedilink
        English
        arrow-up
        31
        ·
        6 days ago

        Totally agree. Huge difference when moving windows or gaming on a 120 Hz or higher monitor. So smooth.

            • NotMyOldRedditName@lemmy.world
              link
              fedilink
              English
              arrow-up
              9
              ·
              edit-2
              6 days ago

              When I tried a game that had a 120hz mode, I had to lower the quality a little to get it work well, but didn’t notice a big difference, but there was a difference.

              A few days later I went back to 60hz so i could increase the graphics quality , and the difference was crazy huge. I had to go back to 120hz.

              • Ibuthyr@lemmy.wtf
                link
                fedilink
                English
                arrow-up
                1
                ·
                2 days ago

                Just give it an hour or two and your brain won’t give a shit about 60 hz anymore.

          • Echo Dot@feddit.uk
            link
            fedilink
            English
            arrow-up
            5
            ·
            6 days ago

            Honestly you will. You probably never compared the two side by side.

            It’s incredibly apparent, you think it’s smooth but then when you go over to 120 Hz and then go back the difference is very apparent.

            • escapeVelocity@lemmy.ca
              link
              fedilink
              English
              arrow-up
              5
              arrow-down
              1
              ·
              6 days ago

              The forbidden comparaison.

              People live happily untill they do it. Than … Than they can’t ever go back . It’s a curse.

              Don’t taste the forbidden fruit

            • mirisgaiss@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              6 days ago

              I had a 980ti for AGES that could do 120hz at 1080. got a big 4k screen, immediately had to upgrade the gpu because it wouldn’t do 4k at anything higher than 60hz. even moving the mouse across the screen felt sluggish.

          • lukaro@lemmy.zip
            link
            fedilink
            English
            arrow-up
            2
            ·
            5 days ago

            I notice the difference between 60 an 72hz but only after a few hours when i either get or don’t have a headache from using 60hz. Visually I see no difference.

        • scala@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 days ago

          120hz on my 4k 75in TV. Didn’t bother connecting it to WiFi. It’s a dumb TV now

    • 1984@lemmy.today
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      23
      ·
      6 days ago

      Are you aware that reducing the energy consumption on monitors is competely irrelevant compared to the giant data centers coming up now, taking as much as power as a full city?

      One typical AI data center ≈ 1 TWh/year ≈ the electricity used by 100 000 average homes annually.

      A very large AI data center under construction ≈ 20 TWh/year ≈ electricity for ~2 million homes annually.

      Fun times isnt it.

      • lemmydividebyzero@reddthat.com
        link
        fedilink
        English
        arrow-up
        15
        ·
        6 days ago

        I personally do reject that kind of thinking.

        “But others are worse” reminds me of the kindergarten…

        EU has introduced energy efficiency levels for monitors based on the potential and we are far from the goal.

        I can’t tear down AI data centers, but I can choose to buy a monitor that does not heat up my living room and leads to a nicer electricity bill for me.

      • A_Random_Idiot@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        1
        ·
        6 days ago

        By that logic, I should go out and just start a murder spree, because people die anyway so whats the point.

  • AA5B@lemmy.world
    link
    fedilink
    English
    arrow-up
    95
    ·
    7 days ago

    I want a dumb tv with the price and and specs of a smart tv. Less is more

    • thatKamGuy@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      31
      arrow-down
      1
      ·
      7 days ago

      What you’re looking for are commercial screens. They’re a bit more expensive for a comparable panel, as they are intended for 24/7 use- but are about as dumb as they get nowadays.

      • NocturnalMorning@lemmy.world
        link
        fedilink
        English
        arrow-up
        20
        ·
        edit-2
        6 days ago

        A bit more expensive? I was able to get a smart tv for like 800 bucks. The same equivalent dumb tv would have been a few thousand dollars and Best Buy said they would only sell it to business accounts which was infuriating to read.

        • Doomsider@lemmy.world
          link
          fedilink
          English
          arrow-up
          9
          ·
          6 days ago

          This isn’t a peasant TV. And it doesn’t even have any tracking, I am not sure a pleb can even legally own these. Sorry, but you have to be a wealthy person who watches CP to have it in your home.

    • avg@lemmy.zip
      link
      fedilink
      English
      arrow-up
      21
      ·
      7 days ago

      You would likely have to pay more since they aren’t getting to sell your information.

      • Alcoholicorn@mander.xyz
        link
        fedilink
        English
        arrow-up
        12
        ·
        7 days ago

        *you would have to pay more because major companies know they can charge more. There isn’t a limited amount of profit a company wants to mae, and then they pick a price from that, they price it as high as the market will bear.

      • AA5B@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        6 days ago

        Realistically, not only do I not want an 8k tv, but I might not get a tv at all, if I had to do it today. We rarely watch tv anymore. The brainrot is the same but we’re much more likely to use individual screens

    • keyez@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      6 days ago

      My LG C3 not connected to internet and using an HTPC and Nvidia shield is working great so far.

      • webhead@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        6 days ago

        I don’t know why more people don’t do this. Don’t connect the TV to the Internet. Ever. Do updates using a flash drive if you must and connect whatever flavor streaming box you like. The TVs and their dumb os get slow and shitty anyway so why fuck around with it? Lol.

        • BurgerBaron@piefed.social
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 days ago

          Ppl spend so much time fucking around with Android slop boxes and I’m just like…how is using a full fat linux desktop PC under the TV with a wireless trackpad keyboard clunky? Set display scale to 200%, install Stremio, and now you can game and shitpost on your couch too as a bonus.

          • webhead@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            5 days ago

            It’s mostly because all the streaming services won’t give you higher res and HDR unless you’re using one of the boxes. You don’t need the slop tho. Nvidia shield imo has been solid for over a decade now. I’m over fire TV and those cheap shitty no name boxes are even worse. I have no idea why people fuck with those. Even a fire TV is better than that lol.

    • drgeppo@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      6 days ago

      I just recently found out about Thomson’s “Easy TVs” LINK

      Actual dumb TVs, with a good range on inputs and at a cheap price, they seem to be aimed at hotels and such.

      The downside is that they only do 1080p at 40" or 43" but the appeal of getting one before they disappear from the market is strong (also I don’t have any 4k media in my library so my worry is more about future proofing than any current necessity)

  • arthurpizza@lemmy.world
    link
    fedilink
    English
    arrow-up
    25
    ·
    6 days ago

    For the majority of people a 1080p60 with a high bitrate and 10+ bit color space will look absolutely perfect. Some can pixel peep and tell, but more people still struggle seeing when the aspect ratio is wrong on their TV.

  • magic_smoke@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    19
    ·
    6 days ago

    That’s because the answer isn’t higher resolutions, it was legally enforcing h.265 to be open source. Now the solution is AV1, but video codecs shouldn’t be locked down like that.

    To act like that was ever in favor of “protecting the sciences” is a fucking joke.

  • A_Random_Idiot@lemmy.world
    link
    fedilink
    English
    arrow-up
    24
    ·
    6 days ago

    TV manufacturers salivated at the idea of TV resolution, hoping desperately to turn the TV market into something like the PC market, in that you have to upgrade every 5ish years to stay on top of technology and use the latest stuff to artificially increase sales beyond what their already abysmal build qualities provide them.

    I’m glad the plan is failing spectacularly.

    Hopefully this forces them to think more about quality and start focusing on TVs that actually last now… You know, like we used to have 30 years ago.

  • Pyr@lemmy.ca
    link
    fedilink
    English
    arrow-up
    46
    arrow-down
    1
    ·
    7 days ago

    Fuck.

    Now instead of each new generation of TVs being slightly higher in resolution some god damn business tech executive is going to focus on some god damn bullshit to try and change or add which is going be absolutely fucking ridiculous or pointless and annoying like AI television shit or TV gaming or fuck my life. Smart TVs are bad enough but they can’t help themselves they need to change or add shit all the fucking time to “INNOVATE!!!” and show stockholder value.

    Resolution was something easy and time consuming but we can’t rely on that keeping them from fucking TV up any more.

  • garretble@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    ·
    5 days ago

    Yeahhhhh 8K is going to be pretty far off considering we still get 1080p “enhanced” trash with YoutubeTV for sports games. It looks like ass on my good, 4K TV. I can’t imagine that on an 8K display.

    Though some sports - like the Unrivaled games on HBO - are of a higher quality, you just don’t get that everywhere.

    And that’s just sports. Couple that with the fact that some people still have data caps, and I just don’t see widespread adoption any time soon.

  • ohulancutash@feddit.uk
    link
    fedilink
    English
    arrow-up
    42
    arrow-down
    1
    ·
    7 days ago

    They showed Skyfall on 70ft IMAX screens, and that film was shot 2880 x 1200. Its not all about the pixel count.

      • ohulancutash@feddit.uk
        link
        fedilink
        English
        arrow-up
        5
        ·
        edit-2
        7 days ago

        DP Roger Deakins has a forum on his site where he answers these sorts of questions. The Arri Alexa Studio was essentially developed for him, as he prefers an optical VF. It has an academy gate, and Deakins favours spherical lenses rather than anamorphic so we simply take the width of the sensor and crop down to 2.40:1 aspect ratio to get the quoted dimensions. Your link quotes the capture resolution, which will have some excess compared to the finished material.

        This was put through the DI process and at a later stage IMAX DNR’d a blow-up to 15/70mm.

  • Omega_Jimes@lemmy.ca
    link
    fedilink
    English
    arrow-up
    37
    ·
    7 days ago

    Theres a ton of other things I want my TV to do before more pixels.

    Actual functional software would be nice, better tracking on high speed shots (in particular sweeping landscapes or reticles in video games) higher frame rates and variable frame rate content, make the actual using of the tv, things like changing inputs or channels faster, oh man so much more.

    Anything but more pixels.

      • uniquethrowagay@feddit.org
        link
        fedilink
        English
        arrow-up
        13
        ·
        7 days ago

        Yes I do. I want an actual smart TV with a practical, open source, TV-optimitzed Linux OS. It’s not that software on a TV is a bad idea in itself. It’s how it’s ruined by for-profit companies.

        • balsoft@lemmy.ml
          link
          fedilink
          English
          arrow-up
          19
          ·
          6 days ago

          Nah, honestly, I think stuffing an entire computer inside a monitor and relying on it to generate/show content is a bad idea no matter what software it runs. A dumb TV + a small computing dongle requires only a tiny fraction more labor to produce than a smart TV, but it’s so much easier to upgrade in the future if you decide you need faster boot times or wanna game on the TV, etc. And if the TV breaks before the dongle does, you can also buy a new TV and keep all your settings/media without transferring anything.

          • Aceticon@lemmy.dbzer0.com
            link
            fedilink
            English
            arrow-up
            7
            ·
            6 days ago

            Also to add to this, the life-cycle of a TV display is mismatched from the live-cycle of media playing hardware or just hardware for general computing: one needs to update the latter more often in order to keep up with things like new video codecs (as for performance those things are actually implemented in hardware) as well as more in general to be capable of running newer software with decent performance.

            I’ve actually had a separate media box for my TV for over a decade and in my experience you go through 3 or 4 media boxes for every time you change TVs, partly because of new video codes coming out and partly because the computing hardware for those things is usually on the low-end so newer software won’t run as well. In fact I eventually settled down on having a generic Mini-PC with Linux and Kodi as my media box (which is pretty much the same to use in your living room as a dedicated media box since you can get a wireless remote for it, so no need for a keyboard or mouse to use it as media player) and it doubles down as a server on the background (remotely managed via ssh), something which wouldn’t at all be possible with computing hardware integrated in the TV.

            In summary, having the computing stuff separate from the TV is cheaper and less frustrating (you don’t need to endure slow software after a few years because the hardware is part of an expensive TV that you don’t want to throw out), as well as giving you far more options to do whatever you want (lets just say that if your network connected media box is enshittified, it’s pretty cheap to replace it or even go the way I went and replace it with a system you fully control)

      • Omega_Jimes@lemmy.ca
        link
        fedilink
        English
        arrow-up
        9
        ·
        7 days ago

        I mean, yes and no. I like e-arc, and I like being able to adjust settings other than v-hold. But I don’t want this slow crud fest that keeps telling me when my neighbour turns on Bluetooth on their iphone.

          • Omega_Jimes@lemmy.ca
            link
            fedilink
            English
            arrow-up
            1
            ·
            6 days ago

            Yeah, all my inputs go to the tv, then i run a wire to the receiver. This makes it so my ps5 and PC are plugged directly to the tv so i can get different resolutions are variable refresh rate and the tv can control the receiver. So when I turn something on, the tv/receiver turn on and set themselves to matching settings, Dolby, stereo, whatever. Its not huge but its a nice convinienice over the older optical connection.

      • Zozano@aussie.zone
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        7 days ago

        I want software on my TV.

        Steam Link specifically. I like streaming to my TV via Ethernet.

    • yeehaw@lemmy.ca
      link
      fedilink
      English
      arrow-up
      18
      ·
      7 days ago

      I still probably watch 90% 1080p and 720p stuff lol. As long as the bitrate is good it still looks really good.

    • Lenggo@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      7 days ago

      I have a Samsung frame because I wanted a TV that didn’t look so much like I had one, but the software is so goddam bad. The only way to switch sources quickly is to set them as a favorite which isn’t always that straight forward if you didn’t do it right away. Regardless you have to let the the home page fully render before you can even worry about that. Even the Samsung TV app which you would think would be perfectly optimized for the hardware since the same compare makes the software is barely functional and loads like a web page on AOL in 1998

      • lobut@lemmy.ca
        link
        fedilink
        English
        arrow-up
        2
        ·
        7 days ago

        I like my frame because it faces a set of windows and with all my other tvs … I would have to close the blinds to see the tv in the day time.

        However, the software is straight garbage. I didn’t even know about the favourite thing … every time I change source it would spend like a minute or two trying to figure out if it could connect to it for no reason.

  • SocialMediaRefugee@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    ·
    6 days ago

    At a certain point yours eyes can’t tell much difference. It is like music, people would obsess over tweaking their stereo systems to the point where I doubt you could physically tell the difference, it was mostly imagined.

    Huge tvs also require big rooms to make the viewing angle work. Not everyone has a room they work in. Apartments are especially too small for huge tvs.

  • Dirty AnCom@discuss.online
    link
    fedilink
    English
    arrow-up
    11
    ·
    6 days ago

    What’s interesting to me is that film is roughly, perceptually around 8K. However, very very few people have cinema-sized screens in their home, so what’s the point if it’s “only” even 80 inches?

    I think giant 8K monitors are still useful for productivity, but only for a small number of people. I personally like having multiple monitors over one big one.

  • towerful@programming.dev
    link
    fedilink
    English
    arrow-up
    26
    ·
    7 days ago

    Yeh, do: 60fps, 30 bit color… and I guess HDR?
    Do things that people can actually appreciate.
    And do them in the way that utilises the new tech. 60fps looks completely different from 24fps… Work with that, it’s a new media format. Express your talent

    • theparadox@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      7 days ago

      Sorry, the best I can do is install a camera and microphone on our next model, to spy on you and force interaction with advertisements.

      I mean video conferencing from your living room. How neat is that?