1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

HDR vs 4K for Gaming: The Breakdown

Discussion in 'Monitors' started by Kalifornication, Dec 2, 2016.

  1. Kalifornication

    Associate

    Joined: Sep 28, 2016

    Posts: 24

    Location: Insta: @KalifornicationX

    Triggered by the launch of Microsoft's Xbox One S -- a revision of the prior Xbox One console which features a slimmer form-factor, 4K Blu Ray playback, 4K game upscaling, HDR support and an overclocked processor – High Dynamic Range – also known as HDR, is set to be the next step forward in display technology.

    In this article I will provide an explanation as to what HDR is, how it works, whether or not it's an important aspect to gaming and if it's a superior visual improvement than going to 4K. Think of this article as a shortcut explanation that does away with the unneeded and vastly detailed information. While the technology is largely limited to TVs as opposed to gaming monitors which are largely preferred due to the high response times and choices for panel technology, HDR will be arriving on PC displays – with graphics cards already providing the software and hardware to support the technology. The decision to move to 4K or HDR is primarily due to the availability of 4K displays on the market, where HDR sees almost none. This leaves gamers in a difficult decision of having to wait for HDR displays and 4K HDR displays to arrive, with the latter being more expensive as it incorporates both technologies.

    What Is HDR And What Does It Do?
    HDR (high dynamic range) allows for a wider range of colours and a greater level of contrast, meaning the levels of brightness are higher and the lowest levels of black are far more deeper. This is the Dynamic Rangethe largest and smallest measurements on each end of the scale. In order to view HDR images and video, a HDR compatible display is required. To put things in to perspective let's take a look at how SDR (standard dynamic range) compares to HDR.

    An SDR display will produce an 8-bit colour range measured at 16.8 million colours, with 6 stops of dynamic range at a maximum brightness level measured at 400 nits. 16.8 million colours means each pixel can present 256 shades of red, green or blue.

    An HDR display will produce a 10-bit colour range measured over 1.07 billion colours, up to 20 stops of dynamic range, with a maximum brightness level well over 1000 nits. 1.07 billion colours means each pixel can present over 1000 shades of red, green or blue.

    [​IMG]
    Playground Games And Microsoft Studios' Forza Horizon 3. SDR.
    Image Courtesy Of NeoGaf.
    http://www.neogaf.com/forum/showthread.php?t=1283285&page=8

    [​IMG]
    HDR

    What's the benefits of an increased colour range?
    Given the additional tones of colour and the increased levels of brightness along with the deeper levels of black, the amount of details within a picture or video can be greatly increased and resolved within each individual pixel for an increased duration of time, allowing for finer details in brighter scenes, more clarity in darker scenes and a more accurate representation of the quantity of colours that reside within the image.

    [​IMG]
    An Image Comparison Of SDR Vs HDR. Image has been adjusted In Order To Demonstrate Differences In Colour On A Non-HDR Display. A HDR Compatible Display Is Required In Order To View Authentic HDR

    What Is 4K And What Does It Do?
    Most modern TVs contain a maximum resolution of 1920 X 1080p (HD). This is the number of pixels within the display on a horizontal and vertical alignment, that changes colours to present an image. Making the move to 4K, this will see TVs with an increase in resolution that's four times the number of pixels within a HD display. This resolution increase measures to 3840 X 2160p.

    [​IMG]
    WB Games Montréal's Batman: Arkham Origins. Zoomed In At 1080p And 4K Resolution

    With in an increase in resolution, more details can be resolved from the image as it is provided with more pixels to divide its information across. This means images will be sharper and crisper, and the amount of colours that's available within the dynamic range will be far more realised and clearer for the viewer to see. Within games this will increase the quality of textures, lighting and shadows. This also allows for less blur in fast moving scenes, especially when it's combined with a display featuring low response times.

    How Much Does Resolution Improve The Gaming Experience?
    Naturally, displays have always seen an increase in the number of pixels they contain. This will continue to happen as the benefits are clear. The human eye doesn't see things in pixels as they do not consist of digital technology interpreting visual information. They take in the visual information that's available to them with no pixels having to divide them in to a set number of pixels for clarity.

    While an increase in resolution is good, what's so good about these extra pixels if the range of colour remains the same? It's essentially the same amount of colours, just more of them attempting to realise additional information without displaying more colours on the screen. For instance, a game which attempts to deliver a realistic aesthetic. Those such as Ryse: Son of Rome, Rise of the Tomb Raider, Battlefield 1, Assassins Creed: Unity and Quantum Break all demonstrate this.

    They employ visual techniques often utilised in film in order to convey an image that's highly detailed, emphasises lighting and softer is as a whole, moving away from the traditional rendering techniques used in video games – in order to eliminate harsh textures and sharp geometry. While the PC versions of these games see an increase in quality over their console variants – due to the increase in resolution and graphical details, neither of these games look terrible, nor do they look more "gamey" as a result of their resolution and detail reduction when being played on a console.

    [​IMG]
    Crytek's Ryse: Son of Rome

    Masters in rendering technology and aiming to push the boundaries since their first well-known title – Remedy Games have always strived for a realistic aesthetic, one that prioritises pixel detail and techniques used in film – as opposed to increasing the resolution with no quality to back it up. When Remedy Games released their Psychological-Thriller title on the Xbox 360Alan Wake, they made the decision to focus on the technique in which the game renders detail and lighting, rather than following the route of a clearer image because of an increase in pixels.

    Displaying the game at a resolution of 960 X 540p seems rather low, especially when you consider that the vast majority of developers where targeting 1280 X 720p and 1920 X 1080p for the Xbox 360. Due to the rendering techniques that the studio employed, which largely emphasised lighting, bloom effects, and enhancing the detail within the many dark and foggy areas within the game's world, Alan Wake proved to be visually stunning – ahead of most games released at the current time. Later down the line the game was released on the PC.

    [​IMG]
    Remedy Games' Alan Wake

    With this decision one would expect the increase in resolution to deliver superior results and a far more detailed game, exceeding the original. Increasing this resolution on the PC while keeping the detail and texture levels to that of the original game didn't produce a better looking game. Only when enabling the higher levels of graphical details and textures did the game look better. The reason for this is simple. Increasing resolution alone without the necessary visual assets to back it up will only result in a crisper and sharper of image of the already available assets.

    As the original game incorporated filmic techniques in order to strike a balance between texture, detail, lighting, resolution and the like – the image that was presented proved to be adequate at the low resolution. Many games since have taken the same visual styling as Remedy's Alan Wake, and titles such as Crytek's Ryse: Son of Rome and Remedy's own Quantum Break all standout as visual showpieces. Looking at all three games in question which all saw releases on both the Xbox platform and the PC, the improvement of these games from console to PC isn't as spectacular as one would expect.

    [​IMG]
    Remedy Games' Quantum Break

    All of these games deliver an image that's startlingly low in resolution on Microsoft's platforms. Neither one of these games look bad as a result even if they do present themselves as superior on the PC which allows for resolutions ranging all the way up to 4K. Reason being, they opt for a filmic aesthetic reminiscent of Alan Wake. With that being said, this brings us back to the question of HDR and 4K, and which of these two technologies could possibly deliver an improvement to gaming, given the aforementioned games gained minimal. Keeping to the three games in question, where would they seek the most benefit?

    Extra pixels or a wide range of colour that can realise a greater amount of detail in both the darkest and brightest scenes of an image?

    To put it another way – is the quantity of pixels more important than the quality of the pixels?


    HDR For Exaggerated Realism
    HDR for titles that attempt to emulate realism through their visuals do have their benefits. And as noted in the aforementioned titles, resolution alone can never be the deciding factor as how far visual details can be pushed. This is especially true when you consider the fact that a video game rendered at a high resolution doesn't look any more detailed, natural, authentic or "real" than a DVD quality movie.

    When taking in to account that a video game is totally artificial within its characters, environments and settings, where as a movie is recorded footage of those which are real – which has then been digitally translated into a set number of colours which is then displayed through pixels, then the case stands to reason. But how about those which do not attempt realism?

    [​IMG]
    Nintendo's Pikmin 3

    Those more in line with Cartoon-like visuals, exaggerated realism, a visual quality akin to movies such as The Incredibles, Ratatouille, or Monsters, Inc. Games such as Crash Bandicoot, Jak and Daxter, Pikmin, Street Fighter, and Donkey Kong do without a doubt gain a visual benefit from rendering at a higher resolution. But since these games place a larger emphasis on colour and the amount of detail they contain which in itself isn't all too deep, but is more dependant on the lighting, geometry depth and as stated previously – colour, these games seek greater benefit from an expanded colour palette.

    [​IMG]
    Disney and Pixar Studios' The Incredibles Movie

    In terms of where HDR and 4K find more benefits, there's no doubting that games will find advantages in both. However, with all that said, it would seem that higher resolutions hold more benefits in the clarity and sharpness of movie viewing, regardless if the source material is capturing the real world or CGI. – Games are greatly dependant on an artificial level of detail in order to make the most of their rendering resolution. That detail level is currently limited in geometry, lighting and general methods for rendering, as well as the range of available colours due to SDR standards.

    So long as games continue to adopt higher resolutions as opposed to prioritising they technique in how the game is rendered and the amount of available colours which in turn provide higher levels of detail, they will always appear dated in the years that follow. This is especially true for games which focus on a realistic quality that depend on resolution to present their aesthetics. This is also why games such as Mario Kart 8, Prince of Persia (2008), Transistor, Ori and the Blind Forest, and Okami hold up far better in the modern day and the years to come, since they're visual style comes from the inspiration of Disney Pixar and cartoons. While these few number of games do look better at higher resolutions they also look great without, gaining more improvement from HDR along with those emulating realism.

    What Are The Requirements For HDR And 4K Gaming?
    While the Xbox One S may have HDR and 4K upscaling capabilities, there are some caveats to the methods implemented. HDR requires no additional processing power in order to make use of. The graphics chip of the console or PC simply needs to have the functionality implemented from a hardware standpoint. It can be implemented through software modifications but it will be severely limited. The necessary display connections for HDR are DisplayPort 1.4 or HDMI 2.0a and a display capable of displaying the HDR media. The HDR technology must also be built in to the game in order to display it.

    4K upscaling is also a function that requires no additional processing power. So long as the aspect ratio of the screen remains the same to the resolution it's displaying, either the display itself, the console or the PC will scale the pixels to fit the display. 4K or 3840 X 2160p is the four times the identical amount of pixels of HD, or 1920 X 1080P. This means the scaling will be precise when attempting to output 4K.

    [​IMG]
    Ryse: Son of Rome Upscaling Comparison. Finer Details Resolve On The Native 1080p Image, But The Image Sees No Significant Improvement Due To The Game's Rendering Technique. The 1600 X 900p Image Upscale Isn't Common Due To The Increase Being Only Around 30%, As Well As The Lack Of Displays Adopting The Resolution. Image courtesy of WCCFTech

    As 4K upscaling is simply a technique of extrapolating extra pixels which do not exist on a HD display, results will not always be satisfying. Most upscaling techniques work by stretching or blurring the colours of each pixel over a larger number of pixels in order to fill the display, which are then applied with a sharpening filter to clean up the image. Others work by having the device or system attempt to calculate which colours around each pixel should be displayed then filling the surrounding areas so that it can scale to the display. Some methods work better than others but none compare to a native 4K display with 3840 X 2160 true pixels.

    Because of the amount of pixels is greater than that of HD, additional processing power is required to display it. In the case that a user wishes to run the game at 4K resolution while keeping the same frame rate the same which it was in HD, the required processing power in order to render the game will be four times the same if not more.

    [​IMG]
    Rise Of The Tomb Raider. 4K Image Resized For The Purposes Of Viewing.

    While a 4K capable display is required to render the game in native 4K there are solutions that will allow users to gain an improvement in game resolution while keeping to a standard HD display. For owners of Nvidia and AMD graphics card, the software and driver control panels for each respective cards provide rendering options for higher resolutions being displayed on a HD display. Nvidia uses a technology called DSR – also known as dynamic super resolution, and AMD uses a technology called VSR – also known as virtual super resolution.

    These technologies allow the display to render the internal image of a game at a higher resolution than what the display is able to provide, which it then downsamples to fit the display. This works by applying an anti-aliasing solution to the image through the form of super-sampling then shrinking the image back down to the aspect ratio of the display. While there are drastic visual improvements to this solution, those which provide a larger benefit over upscaling, it doesn't look as supreme as a native 4K resolution and it require almost as much processing power as rendering the image in native 4K.

    HDR and 4K are both significant leaps for display technology and while the latter has existed for quite a number of years in PC gaming, it hasn't been widely adopted due to the extreme amount of the performance required by graphics cards, most of which have always been on the high-end as opposed to be relatively affordable by the majority of most gamers, as well as the lack of 4K capable displays in the few years prior.

    HD displays have become long-in-the-tooth, modern day graphics cards even on the low-end are more than capable of displaying games at the resolution, at 60 frames-per-second and at detail levels that far eclipse those in game consoles. Video games have always dictated and pioneered display technology and 3D rendering. With TV manufacturers refusing to acknowledge the audience, producing displays which cater to slow moving scenes focusing on image quality, as opposed to image performance which demands low input times and fast refresh rates, gaming monitors have become the number one choice for gamers.

    HDR may not be limited on gaming monitors at this current time, but for gamers who are tolerable to the specifications offered by a TV, they do have larger availability should they choose to use them. Moving to 4K and HDR will see benefits, but should the selection for display availability be one or the other, with gaming monitors that integrate both seeing substantially higher prices, which technology will you decide to adopt?
     
    Last edited: Dec 2, 2016
  2. VEGACORTEZ

    Gangster

    Joined: Jun 3, 2014

    Posts: 375

    I'd be happy with 24" 1440p 144hz or 165hz with HDR on a quality TN panel if such a thing is possible or perhaps OLED. Although if gpu's come out that can run 4k at high fps i'm happy to go back to 4k.
    I have a certified ultra HD tv by Samsung & the response times are bloody terrible with blur & smearing & at 60hz only. I use PS4 pro on it though so 30fps is not an issue on a panel like that as the input lag is pretty decent for a tv but for PC gaming at 60fps the blur & smear is just terrible and ruins gaming, 60hz monitors look much better especially TN, neither of course are a patch on high hz TN.
     
  3. VincentHanna

    Capodecina

    Joined: Jul 30, 2013

    Posts: 21,001

    FYI - Shadow warrior 2 on PC already supports HDR.

    If you wanted to test it.
     
  4. VEGACORTEZ

    Gangster

    Joined: Jun 3, 2014

    Posts: 375

    Thanks, I've actually already tried it quickly on my tv but I would rather not use HDR & play on my monitor at 120-160fps with very little motion blur. TV's are kind of unsuitable for PC gaming to me personally, consoles obviously don't really benefit from a monitor & so are fine. I really like HDR but not enough to game at 4k lower fps & crappy motion on a tv. I really wish that gaming monitors would just get HDR already & some at 1440p if possible then you get the best of everything.
     
  5. bakes0310

    Capodecina

    Joined: Oct 22, 2004

    Posts: 12,434

    Thanks for a very good thread. Haven't been following about the latest technology very well for monitors/TVs so thanks for a good explanation.
    I currently have a lg 34" super wide monitor which I love and don't think I could honestly go away from super wide monitors now.
    My question is there going to be super wide HDR monoitors. My dream monitor would be an lg 34" 3440x 1440 HDR, could I actually expect something like that in 2017?
     
  6. Greebo

    Caporegime

    Joined: Jan 20, 2005

    Posts: 34,997

    Location: Co Durham

    No chance bakes0310. The monitors we see for sale in 2017 are the ones which have already been shown in shows now and no such monitor exists. It's normally 6 months o a year after you first see a monitor before it becomes available.

    Sadly I don't think we will start and see any new decent HDR screens until 2018.
     
  7. bakes0310

    Capodecina

    Joined: Oct 22, 2004

    Posts: 12,434

    Thats too bad, but tbh i am very happy with my current monitor to which ive only had for two years. And going my past record i seem to change monitors every 4 years so i still got another 2 to go yet, so roll on 2019.
     
  8. avatar12m

    Hitman

    Joined: Aug 2, 2006

    Posts: 929

    Location: Elloughton, UK

    so how does a pc game display hdr ? is it only by using a tv style monitor as there are no ( from what I can see ) a pc hdr monitors ?

    would buying a 10bit non hdr oled pc panel be near to hdr as you can possibly get atm
     
    Last edited: Jan 13, 2017
  9. John7up

    Gangster

    Joined: Jul 7, 2013

    Posts: 109

    Location: Newcastle

    Oled yeah, definitely. In july-august maybe we'll get a chance to see in stores the 27 inch 4k hdr monitor... till then 34inch curved ultrawide 100hz gsync 3440 monitor is still the best allrounder.
     
  10. Daniel - LG

    LG Rep

    Joined: Apr 8, 2015

    Posts: 561

    Location: Normally in the car

    Hi All,

    Incase anyone wasn't aware, LG are expecting the LG 32UD99 to launch in the next 8 weeks or so - this is the first PC monitor with HDR

    This will be 32", 4K UHD IPS, with HDR 10, will include USB-C connectivity as well as usual HDMI and DP, should be Freesync supporting also

    And too top it off, looks like OCUK will have the UK Exclusive on launch stock to boot

    Dan
     
  11. spot35

    Gangster

    Joined: Apr 1, 2014

    Posts: 294

    Nice one. Will there be a gsynch variant?
     
  12. Wulfster

    Gangster

    Joined: Feb 12, 2016

    Posts: 399

    Location: 2 metres away

  13. nox_uk

    Hitman

    Joined: Sep 12, 2006

    Posts: 626

    Any ideas if will it have local dimming?

    Nox
     
  14. jeff83nz

    Gangster

    Joined: Aug 26, 2013

    Posts: 428

    Daniel is there an expected UK price and date yet?
     
  15. Daniel - LG

    LG Rep

    Joined: Apr 8, 2015

    Posts: 561

    Location: Normally in the car

    date - unsure, price I am discussing :)
     
  16. jeff83nz

    Gangster

    Joined: Aug 26, 2013

    Posts: 428

    Thanks
     
  17. Daniel - LG

    LG Rep

    Joined: Apr 8, 2015

    Posts: 561

    Location: Normally in the car

    apologies for the lack of Info

    due to panel shortage, looks like May for stock JFYI
     
  18. Astravagrant

    Gangster

    Joined: Jul 31, 2014

    Posts: 121

    ^^ any update on this?
     
  19. Flake87

    Gangster

    Joined: Jul 27, 2015

    Posts: 251

    Quite an in depth piece there, but I do wonder how many people will be able to see the difference between 16 million and 1 billion colours in a moving image ? Far more important to me at least is the difference in shades of grey where 8 bit can only manage 256 shades against the 1024 of 10 bit. This can lead to banding effects which can be seen. Whether or not this table can be correctly displayed on an 8 bit monitor is of course another point !


    [​IMG]
     
  20. Flake87

    Gangster

    Joined: Jul 27, 2015

    Posts: 251

    Another qualification to add here is that the usual GTX series of NVidia graphics cards cannot display 10 bit colour across all formats. OpenGL 10 bit which photoshop uses being a case in point, and the only cards which will render in 10 bit are the Quadro / Firepro. It might well be another case of manufacturers hobbling their cards to make sure sales of their 'pro' range are maintained.