Today we find the GPU Flashback Archive delving into the not so distant past to focus on the NVIDIA 900 series of graphics cards, the first to use NVIDIA’s new Maxwell architecture which had already seen the light day in mobile GPU solutions, an indication of the direction that the company were taking at the time. Let’s take a look at the cards that were launched as part of the 900 Series, the improvements and changes that Maxwell brought and some of the more memorable scores that have been posted on HWBOT.
The first question one may well have regarding the NVIDIA 900 series is simple - what happened to the 800 series? To answer the question fully, you must first look at the direction that NVIDIA was moving at the time. A movement to expand its product offerings in order to compete in the quickly expanding mobile SoC market. The suddenly ubiquity of Android-based smartphones around the globe was fuelled in part by the development of mobile SoCs from Qualcomm, Samsung, Mediatek, Marvell, Allwinner and others. The traditional feature phone was quickly being replaced by smartphones that now required improved multi-core CPU performance, HD display support and, importantly from NVIDIA’s perspective, decent enough graphics processing to actually play 3D games. Intel and NVIDIA were two companies with plenty of R&D and marketing budget who sought to enter a new market to help bolster revenues during an inevitable slow down of desktop PC sales, a traditional cash cow for both.
The GPU Flashback Archive series continues today with a recap of the NVIDIA GeForce 700 series, a series refresh which heralds part two of the Kepler family of GPUs. We can also remember it as a time when NVIDIA launched their first ever GTX Titan card and with it, a new pricing and retail strategy for truly high-end graphics card products. Let’s take a look at the new Kepler architecture GPUs, the cards that were popular with HWBOT members and some of the more memorable scores that have been posted since launch.
The 2011-2013 period of history saw NVIDIA implement a more regular cadence to their high-end product launches and refreshes. One that saw the company launch a new GPU architecture every two years, with new product lines arriving each year. This means deriving two product lines per architecture with an improved version offered the second time out. This is what we saw with Fermi, an architecture whose potential was full realized at the second attempt. With the GeForce 700 series, which arrived proper in May 2013 with the arrival of both the GeForce GTX 780 and GTX 770, we have something different. The new cards arrived using a much bigger version of the Kepler architecture compared to what we saw on the NVIDIA 600 series.
The GPU Flashback Archive arrives today at the NVIDIA 600 series that debuted in Spring of 2012. The new range of cards showcased a new graphics architecture design and the beginning of what we might describe as the Kepler era. Let’s take a peek at the changes that the new design heralded as well as a close up view of on the GeForce GTX 680 card, the most popular 6-series card with HWBOT members historically speaking. Before we look at some notable scores that were made with the GeForce 680, let’s first kick off with an overview of what innovations arrived with the new Kepler architecture.
If we cast our minds back to 2012 we can recall a era when NVIDIA and AMD were virtually neck and neck, with successive graphic card launches from each company swinging the performance crown from side to side. The arrival of Kepler in many ways represents the beginning of the end of the competitive duopoly that is clearly absent today. Kepler helped NVIDIA push ahead of AMD in terms of graphics processor design, creating a performance lead which AMD still finds insurmountable, despite the arrival of their latest Vega-based cards. Let’s take a look at Kepler in a little detail.
This week the GPU Flashback Archive sets its sights on the GeForce 500 series from NVIDIA. Arriving in late 2010, the 500 Series was the second round of graphics cards based on the Fermi architecture which had limped over the line in the previous generation, ostensibly due to fabrication and yield issues. The new flagship GTX 580 arrived with a more polished take on the Fermi design that help NVIDIA combat the threat from AMD and their popular Radeon 5000 and 6000 series cards. As ever, let’s take a look at the new GPU, the new flagship card and a few of the outstanding scores that have been submitted to HWBOT.
To say that the NVIDIA 400 series graphics cards launch was less than smooth, would be a total understatement. The GF100 Fermi architecture GPU in fact arrived six months late with a significant number of cores hacked off. Blame was laid at the door of fabricators TSMC and a 40nm manufacturing process that clearly hadn’t been optimally adapted for NVIDIA’s Fermi, a monster chip boasting 3 billion transistors and a 529mm² die. While cards such as the GTX 480 had actually done well to make NVIDIA competitive in performance terms, the GTX 580 and its GF110 GPU was rather quickly shoved out the door just eight months later as a revised and improved version of the original.
This week in our GPU Flashback Archive series we cast our minds back to a very popular and well loved graphics card series, the GeForce 400 series. NVIDIA launched the GeForce 400 series in March 2010 armed with a new Fermi architecture that it hoped would help it compete with the successful AMD Radeon 5000 series. Let’s look at the new features that Fermi offered, the cards that were popular and the scores that were submitted to HWBOT in this era.
Compared to previous product launches from NVIDIA, the GeForce 400 series launch did not go as smoothly as hoped. September 2009 saw AMD come out with their Radeon 5000 series which made a solid case against NVIDIA 200 series offerings. It would be January before NVIDIA really started wooing tech media with tales of its forthcoming Fermi architecture lineup. It would be March 2010 before tech media actually got their hands on the new cards and several weeks after that before enthusiasts would be able to actually buy one. This was not the typical NVIDIA launch. Reasons for the delay certainly seemed to lie with issues with actual fabrication at TSMC who were not providing the yields expected on their new 40nm process. This was a problem that particularly hurt NVIDIA due to the fact that the new Fermi GPU, the GF100, was actually very large. When the GeForce 400 series finally arrived in the form of the GeForce GTX 480 and GTX 470, by most calculations they were six months late.
We are treated this week to a look at the NVIDIA 200 series of graphics cards. As well as rejigged product nomenclature, the 200 series represents a new and improved architectural approach to the GPU design from NVIDIA who managed to come up with their largest graphics chip ever. The 200 series was the latest weapon in the fight against ATI and one that proved to be fairly potent in terms of raw frame-rates. Let’s take a look at the new architecture, the graphics cards that were popular at the time with overclockers on HWBOT and of course, some of the more notable scores that have been made its introduction.
We mentioned in the previous GeForce 9 series article how this period of history shows plenty of overlap in terms of GPU series. In April 2008 NVIDIA launched the 9 series and the G92 GPU (read all about the 9 series here ) which was based on an improved but largely identical Tesla design. The 9 series served a purpose by bringing to market cheaper high-end enthusiast cards that could compete with ATI. It also eventually gave NVIDIA a chance to test out the 55nm manufacturing process from TSMC using a more familiar architecture. The GeForce 200 series initially launched on 65nm silicon with later revisions taking advantage the 55nm process.
This week the GPU Flashback Archive series turns its attention to the NVIDIA 9 Series of graphics cards that replaced the successful and much loved 8 series. Arriving in April 2008, the new series featured an updated GPU design that eventually found itself built on a new 55nm manufacturing process. The period also marks a time when ATI and NVIDIA were trading blows as equals. An era when taking the performance crown was all that mattered, creating a situation that proved how healthy competition in an industry could indeed be very beneficial for consumers. Let’s go back in time and revisit the NVIDIA 9 Series, the cards that were popular on HWBOT and some of the more notable scores that have been submitted to the database.
The era of the NVIDIA GeForce 9 series is actually one of considerable overlap. When the the 9 series became available in stores at launch on April fool’s day 2008, a full array of 8 series cards were still available in the retail channel. There’s nothing too odd about that, as the previous generation typically gets a price cut to help clear inventory. It is a little odd however when the next generation GTX 200 series arrived on shelves just three months later. Today we’ll try and keep things simple and just focus on exactly what the GeForce 9 series offered. The 9 series may always be compared to the second revision Tesla chips that followed it, but for now we’ll leave the GTX 200 series for next week’s edition.
This week’s trip down GPU memory lane is all about the NVIDIA 8 series of graphics cards, a series that marks the arrival of DirectX 10 and a wholly new GPU architecture. Arriving in late 2006, the NVIDIA 8 series remains a fondly remembered era for many enthusiasts and of course overclockers, especially the GeForce 8800 GTX a card that is still a topic of conversation with some retro-minded HWBOT members today. Let’s take a look at the hardware associated with the GeForce 8 series era, the technology and features that arrived at that time, and some of the scores and submissions that were made using popular GeForce 8800 GTX card.
The NVIDIA 8 series was officially launched on November 8th 2006 with the arrival of a new flagship card, the GeForce 8800 GTX. The card presented a new GPU to the world’s media, the NVIDIA G80, an entirely new design based on the Tesla architecture. The GPU itself was manufactured using a 90nm process, packed a groundbreaking 681 million transistors into a die measuring 484mm². The G80 was designed specifically with DirectX 10 in mind, taking advantage of many of the specific technologies and ideas introduced by Microsoft. One such feature is the implementation of unified shaders.
Today’s trip down GPU memory lane is all about the NVIDIA 7 series that arrived on the scene in June 2005. Where previous GPU designs had heralded major innovations and the introduction of entirely new technologies, the 7 series was more of an update by comparison. The new GPU arrived with a change in nomenclature and notably a change in the way that NVIDIA graphics cards were actually launched - NVIDIA and AIB partners had products shelves on the very same day that the press embargo was lifted. Let’s look at the GPUs and cards that arrived as part of the new 7 series launch, the cards that have since proved to be popular with overclockers on HWBOT and of course, the notable scores that grace our database to this day.
The NVIDIA GeForce 7800 GTX was launched on June 22nd 2005 as the company’s brand new flagship card offering. At launch the card was immediately available in the retail channel, literally the same day, which at the time was largely unheard of. This was seen as NVIDIA more or less giving ATI the proverbial finger, as previous ATI launches had tended to be prefaced with vague ‘coming soon... we hope’ messaging. The 7800 GTX was based on the G70, the successor to the NV4x series that had powered the GeForce 6 series. The change in naming scheme was apparently a marketing decision with GeForce 7 being better represented by G70 than NV47. The NV70 was largely based on the same architecture as the previous generation and the NV30 generation that preceded it. The G70 again used Shader Model 3.0 with support for the DX9.0c and OpenGL 2.1. Nothing new there. The real interest is when you consider the rendering configuration.
Welcome back to another episode in our GPU Flashback Archive series. Following on from last week’s look at the GeForce FX series, we turn our attention to its successor, the NVIDIA GeForce 6 series. After rising to a position of relative dominance in the early years of GPU design, the GeForce 4 and subsequent FX series had seen NVIDIA lose ground to ATI who had stolen a march with their highly popular Radeon 9000 series. The stage was set for a return with the launch of a new GPU design and a series of cards that required more space in your rig and additional power to deliver a truly next generation gaming experience. Let turn our minds back to 2004 and check out the technologies and features that debuted with the GeForce 6 series, plus the most popular cards of the era and the most notable scores that have been submitted here on HWBOT.
The NVIDIA GeForce 6 series arrived in tech reviewers hands in April of 2004, debuting with a new NV40 GPU and two graphics card models, the GeForce 6 Ultra which commanded a price of $499 USD, and the GeForce 6800 (often referred to as the non-Ultra) for $299 USD. Let’s first consider the GPU itself, the NV40.