NVIDIA GeForce RTX 2080 Ti review, SFF style.

Discussion in 'Hardware' started by 3lfk1ng, Nov 12, 2018.

  1. 3lfk1ng

    3lfk1ng King of Cable Management
    Thread Starter Site Staff Gold Supporter

    Another long but thorough read, we review the NVIDIA GeForce RTX 2080 Ti.

    Note: This is going to be an ongoing review, a first for SFFn, and it will see several updates throughout the following months. Updates will be appended to the article and also in this thread.

    Read the review here.
    annasoh323, Phuncz, K888D and 4 others like this.
  2. Windfall

    Windfall SFF Guru

    In @Josh | NFC's stream today, he said the LED's were RGB, configurable thru software.

    Please clarify, thanks!
  3. 3lfk1ng

    3lfk1ng King of Cable Management
    Thread Starter Site Staff Gold Supporter

    #3 3lfk1ng, Nov 12, 2018
    Last edited: Nov 12, 2018
    They are RGB indeed but the internal controller has the color set to green and green ONLY. I've not found a way to change them yet. I think JayzTwoCents discovered that as well.

    Corsair, ASUS, and EVGA tools do not recognize them either and I didn't find any visible way to access any header that could connect to an RGB controller.
    Windfall likes this.
  4. el01

    el01 King of Cable Management

    I think the 001 in the fan name refers to...erm... revision 001??

    the card looks subjectively ugly to me though. more of a hard-edged-low-poly person if you know what I mean.
  5. VegetableStu

    VegetableStu Shrink Way Wielder

    jay brought the FE shroud to an EVGA reference PGB and the RGB works o_o
  6. rfarmer

    rfarmer King of Cable Management
    Silver Supporter

    Nice review, wish the price was more reasonable.

    You state that the Intel CPUs only have 16 PCI lanes so when used with a M.2 PCI 3.0x4 SSD you only have x8 available for the GPU. While it is true the CPU only has 16 lanes the z370 and z390 motherboards have an additional 24 PCI lanes.

    I have a M.2 PCI 3.0x4 SSD and still have full x16 for my GPU.

  7. TeutonJon78

    TeutonJon78 Average Stuffer

    How can Nvidia make it's top of the line GPU in a perfectly good sized frame, but all the OEMS need to make these giant monstrosities?
  8. K888D

    K888D SFF Guru
    Bronze Supporter Lazer3D

    Excellent review, thank you.
  9. 3lfk1ng

    3lfk1ng King of Cable Management
    Thread Starter Site Staff Gold Supporter

    #9 3lfk1ng, Nov 13, 2018
    Last edited: Nov 13, 2018
    We removed that bit to look further into the matter. It was more in-depth when the review went live.

    However, @||| has informed us that PCIe lanes are not as clear cut and dry as their circuitry can vary from one motherboard manufacturer to another -even regardless of the chipset. Our intention was to exclude the screwiness that comes with PCIe switches on some ATX boards but it turns out that the issue was far more complicating than I could have imagined. While the issue hasn't been as prevalent since the z97 and z170 era, the board that I am on is from 2018 has a 2nd m.2 that is shared with the only PCIe x16 lane -thus dropping it to x8 when in use.

    We're discussing the possibility to crowd-source and see what other motherboards might have this 'issue' and would like to cover it in future motherboard reviews alongside PCIe bifurcation.

    Either way, there is enough hardware to get down to x8 and thankfully it wouldonly impact those with a 2080 Ti.

    I wasn't willing to take mine apart but the FE is much smaller due to the use of a vapor chamber HSF. The technology is nothing new but it's still more costly and difficult/expensive to manufacture. It's not exclusive to the FE either, some of the other OEMs like GALAX have chosen to employ it as well.

    Thank you, we will be improving on it as we go so stay tuned :)
    Phuncz and VegetableStu like this.
  10. rfarmer

    rfarmer King of Cable Management
    Silver Supporter

    Thanks for the response @3lfk1ng I can well believe different boards behave differently. My Gigabyte z170 was only giving me x8 for the GPU even though the chipset has additional lanes, I am just glad it is working properly on my z370.
  11. Phuncz

    Phuncz Lord of the Boards
    Moderator Gold Supporter SFF Purist

    Excellent review, it was a fun read with clear pictures !
  12. annasoh323

    annasoh323 Airflow Optimizer

    Excellently thorough review that captures (IMO) the right balance of data and interpretation/analysis. It seems like everyone is in the same boat as far as DLSS and RT go, so not holding my breath for seeing those parts of the article filled in anytime soon. Sounds like it's a good time for Club Pascal to lay low and see what the future may hold.

    <mumbling> well, if I shouldn't upgrade, maybe I can find a different excuse to build a new machine altogether... SFF is a terrifying drug.
  13. EdZ

    EdZ Virtual Realist
    Gold Supporter

    A few nit-picks:
    The adhesive-attached front plate only needs to be removed to disassemble the cooler into its individual components. To remove the cooler entirely form the card (e.g. for watercooling) you only need to remove the screws in the backplate and the whole thing comes off, same as with the previous FE and reference cards.
    From Derbauer's testing, the RTX cards respond better to OCed memory and stock GPU core clocks than a GPU core overclock and stock memory. Slightly better performance than a max power target 100% fan OC, but with not noticeable increase in power consumption, so likely a good choice for SFF.
    On Intel:
    As far as I am aware, no consumer socket (Hx series, AKA LGA 115x) boards feed the m.2 slots with CPU lanes. ALL feed them with lanes from the PCH, which has its own dedicated DMI 3.0 link (similar to PCIe 3.0 x4) to the CPU that is not shared by the CPU PCIe lanes, laving the x16 PCIe slot unencumbered. This is because Intel's PCIe RAID works with chipset lanes, but not CPU lanes.
    Things are a bit different on the 'enthusiast'/HEDT Xx99 platforms (which practically means the two ASRock ITX boards). There, there are plenty of CPU PCIe lanes to feed multiple m.2 slots along with an x16 slot for a GPU, and the CPU supports VROC (RAID on CPU PCIe lanes). However, the ASRock x299 ITX/AC has only two of it's m.2 ports connected to CPU lanes. The third is connected to the PCH, so it can be used for an Optane transparent cache.
    On AMD:
    With Ryzen, a single m.2 slot can be fed by an extra x4 PCIe 3.0 link from the CPU in addition to the x16 PCIe 3,0 link used for a GPU (or other card). Any additional m.2 slots are instead PCIe 2.0 (not 3.0!) and fed from the chipset. The single theoretical exception are the A300/X300 'un-chipsets', where there is no chipsets and everything is connected to the CPU directly (as it has a handful of SATA and USB links on board). This in theory leaves the PCIe 3.0 x4 link usually occupied by the chipset free for an m.2 slot, but I have not seen a board do this in practice.
    With Threadripper/Epyc, there are plenty of CPU PCIe 3.0 lanes available, but these are a pretty poor choice for gaming i nthe first place, so rather a moot point.
    The chance of 7nm being able to produce large dies (like the Pascal and Turing cards) anytime soon seems pretty low. The physics issues that TSMC face are identical to those Intel are facing with 10nm, and have also only been able to produce similarly small dies at acceptable volumes. 7nm may produce efficient small-die GPUs, but high-end GPUs are going to be sticking to larger processes or commanding very high prices (e.g. the Radeon Instinct MI60, at 330mm^2 around the size of the GTX1070's GP104 but with a pricing target in the GV100 realm). On top of process scale issues, demand will also provide upward pricing pressure: Apple are eating all TSMC's limited output, and next in line are AMD for low-volume-high-margin parts (and with a departure from the shared-die approach used on Zeppelin to Epyc getting a separate die from 'Zen 3' likely meaning those parts will remain on 14nm or move to Samsung's 10nm/8nm which are of the same feature size as 14nm).
    tl;dr a 7nm Ampere may pick up the 'lower end' (2060 on down) but are unlikely to occupy the higher end range the 2070/2080 do. Nvidia are likely making thin margins (remember their sales are mostly die-to-AIB rather than FE cards) on TU102 unless yeilds are truly exceptional*, a similar performance class GPU (even an unchanged die-shrink) on 7nm would cost more, not less; cost/transistor has been rising since 28nm.
    How it works is (relatively) simple: Nvidia uses their big SaturnV Turing-powered supercomputer to render demo runs of a game to two resolutions simultaneously: a render target resolution (e.g. 2560x1440, AKA 'low res') and a 'final' resolution (e.g. 3840x2160 AKA 'high res') with 64x SSAA (the SSAA here is to produce nice alias-free training images for the NN, because aliasing is high frequency noise that the NN could mistakenly view as a desired output). A NN is then trained on that mountain of frames to go "if you see feature X in 'low res', it should look like the same location in 'high res'". With sufficiently variation in training frames (i.e. the demo loop traverses all levels, views every entity in every angle in as many lighting conditions as possible, etc) you have a NN that takes a locally rendered 'low res' frame, and will spit out a 'high res' frame. In short, it's a 'smart' upscaled tuned to a specific game, but tuned by brute force rather than manual tweaking. The massive amount of work done to train the NN (and produce its training datasets) is all done offline only once, to produce a relatively lightweight 'inference' NN that can be run by everyone in real-time.
    It;s entirely down to there being n system with an acceptable combination of high bandwith and low end-to-end latency. Safety has nothing whatsoever to do with it as a: RF is non-ionising, so causes only surface heating (as would holding a warm mug to your head) and b: the transmitter is across the room, not on your head.
    Personally this would be a 'pro' too: it's rigid as a board, no 'GPU sag' possible unless your rear PCIe bracket itself is flimsy enough to bend!


    * This is into unrelated wild-ass-guess territory, but my suspicion is that abnormally high Turing yeilds early in the development cycle may well be the reason for the rushed release. Nvidia may have been expecting to have only a handful of viable dies to feed the Quadro RTX cards and to later sell off the lower binned remnants as a 'Titan RTX', but ended up with a glut of quality dies and popped out the GeForce RTX series at short notice. This would explain why the PCBs are Quadro-grade overspecced (not enough time to design newer consumer-grade PCBs, and why the RTX 2070 came out with a delay), the coolers are overbuilt (not enough time to pare the design down for efficient manufacture), why the lower end of the range is absent (Ampere would have lacked RT and Tensor cores entirely), and why RT and DLSS features are only starting to be implemented in engines (intention was for the Quadro RTX cards to seed development before a GeForce RTX launch in later generations). As for why they would rush them to market? The combination of high GPU pricing that could support an experimental card launch, and complete lack of competition in the high end, lets them jump the gun on hybrid rendering. It's also a bet that's worked out well for them in the past after all (Hardware T&L, unified shaders).
  14. 3lfk1ng

    3lfk1ng King of Cable Management
    Thread Starter Site Staff Gold Supporter

    #14 3lfk1ng, Nov 13, 2018
    Last edited: Nov 13, 2018
    You'll dig deep into anything, this is why I love you @EdZ. Thank you for the detailed and elaborate reply, much appreciated!

    Yes, an additional 600MHz on the memory is as easy as an additional 500MHz was on Pascal. However, I'm not sure if all OEMs are using the same memory and/or if they will continue to once the price drops so I chose to just perform a basic overclock the GPU as that SKU isn't likely to change. To make matters worse, the Voltage slider in EVGA's Precision X1 doesn't currently work, it doesn't actually adjust the voltage, in fact, it's been so buggy that it cannot save any settings/presets.

    According to W1zzard: "installing the RTX 2080 Ti in the topmost x16 slot of your motherboard while sharing half its PCIe bandwidth with another device in the second slot, such as an M.2 PCIe SSD, will come with performance penalties, even if they're small."

    So while it's not likely to affect ITX it might still be worth mentioning(??) TBPH, I'm not really sure how to illustrate this point other than, it might be a thing now that surely won't be a thing when PCIe 4.0 drops.

    On AMD, it get's overly confusing and my original attempt to simplify it wasn't as clear and in depth as needed.

    I could have researched more into the issue but I think it would warrant an entire article by itself.

    Yea, that's why I mentioned Ampere. It supposed to be their 7nm flagship from the rumor mill. I just hope it provides a greater jump than the small Volta to Turing hop. I'm guessing that we will learn more in the 1st half of next year but I don't think we will see a consumer launch until early 2020.

    Ehh, I've used Meraki wireless units that are strong enough to cause strong headaches so I would like that little wireless quip to remain. I definitely wouldn't want anything stronger mounted(or pointed) towards my head.

    The article has been updated. Thanks again EdZ.
    EdZ likes this.
  15. Kandirma

    Kandirma Trash Compacter

    I think this is an important note - it will cut down on dev time in the far off future where everyone has it. Until then it is an additional feature to be implemented.

    And, at least in competitive games, 'more realistic' lighting is going to be a CON not a PRO in many situations as it will make shadows darker instead of just applying a base level of light everywhere and then darkening select areas. It Might help for seeing shadows of players, but I suspect in most competitive [fps] games people will be disabling RTX for as long as they can to prevent shadows/dark areas from being truly dark.
  16. Elerek

    Elerek Cable-Tie Ninja

    So what are the odds of GPUs moving to a "glued together" format like Ryzen CPUs and Intel's new HEDT CPUs?
  17. EdZ

    EdZ Virtual Realist
    Gold Supporter

    The big limitation of segmented chips is that it puts a massive constraint on bandwidth and latency between the 'chiplets' (even with EMIB, the problem is far worse with through-substrate links). Thinking of a multi-die package like a multi-socket system crammed into a small space gets you into the right state of mind for thinking about how to optimise for that sort of device, and likewise with GPUs you would need to think about how to optimise for multiple discrete GPUs to get decent results out of a segmented GPU. GPUs are also much more sensitive to bandwidth than CPUs, as most graphical operations revolve around loading a big chunk of data, applying a relatively simple transform to all of it, then storing that big chunk of data ready for the next rendering step.
    For HPC, this seems a pretty decent way to go. Dangling a bunch of discrete GPUs off of the PCIe bus, or strung across an NVLink mesh, is the standard way to add GPGPU to a HPC cluster, and it works pretty well. Adapting existing techniques for job distribution to account for segmented GPUs is not a large leap for HPC, so it seems a good route forwards for reticle-bustingly large GPUs targeted at HPC like GV100.
    For consumer GPUs, I think it would be a pretty bad choice. There, you are basically only ever running one single job at any one time on any number of GPUs you may have (i.e. you are not going to be playing two games at once) which means all the problems that multi-GPU has today would apply to a segmented GPU. Over a decade of multi-GPU effort has pretty much petered down to "we give up" at this point, so developing a segmented GPU for consuemr use has two uphill battles to fight: designing an architecture that doesn't get hobbled by internal bandwidth limitations, and convincing the entire development community to reject the last decade and a half of direct experience with multi-GPU not paying dividends with "this time it's different! (for our new architecture only)".
  18. VegetableStu

    VegetableStu Shrink Way Wielder

    could the engine do a filter pass on the raytracing pass? o_o i mean in-card colour filters are a thing...
  19. Kandirma

    Kandirma Trash Compacter

    I see no reason why it couldn't, but it'd kinda-defeat the point of having ray tracing.

    The only thing I could see being a 'reasonable' solution is just bringing the black level up to a point where it's more of a dark grey
  20. Elerek

    Elerek Cable-Tie Ninja

    On the flip side, nvidia and amd are still pushing multi-gpu with renewed efforts. If multi-die becomes necessary in the gpu space, it could finally push us to the place where multi-gpu is actually worth it and scales in every game because every gpu is effectively already multi-gpu by design.