Hold off those video card purchases (details on GeForce 4 family/new ATI cards)
Here's the goods folks.
NV25 (GeForce 4 Ti4400 & 4600):
-63 Million Transistors
-0.15 micron.
-4 rendering pipelines, 2 TMUs each.
-300 MHz core frequency.
- NVidia declares 4.9 Gpixels/s fillrate for the upcoming GPU. We already know that nowadays developers declare so-called
-effective fillrate for their processors, so, take these enormous numbers as "peak numbers" that do not have a lot to do with the real games in real life.
-The graphics processor will be very complex and will heat pretty well, so it will be equipped with an integrated heat sink as well as a special hardware monitoring function called NVidia Thermal control system.
-Frequency: 650 MHz DDR SDRAM.
-Four BGA memory chips on the front side of the board, just like on the FireGL 8700. However, NVidia solution also has memory heatsinks.
-Peak memory bandwidth: 10.4 GB/s.
-128MB Frame Buffer.
-Lightspeed Memory Architecture II, the new version now includes the following features:
Lossless 4:1 Z-compression: four times decreases the memory bandwidth consumed by Z-buffer
Second generation Occlusion Culling. As we know from various reviews, Hidden Surface Removal technique is more efficient in RADEON 8500 and KYRO II than in GeForce 3; in GeForce 4, NVidia introduces a new algorithm that will be more effective than its predecessor.
Quadcache Architecture.
The speed of the GeForce 4 Ti4600 based videocards is going to be extremely high compared to GeForce 3 and RADEON 8500. Additionally, there is also a couple of new features so that those interested in capabilities also had a reason to buy a new videocard
The new patent pending type of AntiAliasing will be introduced with the NV25. Its brand name is going to be Accuview AA and represent the new AA mask and technology, that saves one frame buffer write cycle. Accuview AA is another multisampling type of antialiasing. Originally, multisampling needs to write subpixel locations into the memory and only the calculate the pixel color values. Thanks to the new subpixel mask [and the improved algorithm and caching] the entire memory write cycle will be saved causing smaller perfomance decrease compared to the GeForce3. Another interesting feature of the new GPU is the nFiniteFX II engine. Basically, it is the same as on the GeForce 3, but with an additional vertex shader pipeline. Due to the increased frequency, the software that can use the power of the second vertex-shader pipeline will run up to three times faster on the GeForce 4 compared to the predecessor. We can also expect a 50% perfomance gain from the newest solutions when calculation pixel shaders due to the same reason. NVidia declares support for the so-called Z-Correct bump mapping. We couldn't get any more or less precise information about this, and we don't want to speculate on this, so we'll just say nothing this time. The final most important thing about GeForce4 is the support for the nView technology. The new, more feature-rich an easier to use version of TwinView.
That's just a little teaser. I will post more including NV17 (GeForce 4 MX) details, pictures of the new GeForce 4 cards, and info on Hercules ATI cards, the Fire GL cards, 7500DV, and new Kyro. Stay tuned.
[ 01-25-2002: Message edited by: TigerWoods99 ]</p>
NV25 (GeForce 4 Ti4400 & 4600):
-63 Million Transistors
-0.15 micron.
-4 rendering pipelines, 2 TMUs each.
-300 MHz core frequency.
- NVidia declares 4.9 Gpixels/s fillrate for the upcoming GPU. We already know that nowadays developers declare so-called
-effective fillrate for their processors, so, take these enormous numbers as "peak numbers" that do not have a lot to do with the real games in real life.
-The graphics processor will be very complex and will heat pretty well, so it will be equipped with an integrated heat sink as well as a special hardware monitoring function called NVidia Thermal control system.
-Frequency: 650 MHz DDR SDRAM.
-Four BGA memory chips on the front side of the board, just like on the FireGL 8700. However, NVidia solution also has memory heatsinks.
-Peak memory bandwidth: 10.4 GB/s.
-128MB Frame Buffer.
-Lightspeed Memory Architecture II, the new version now includes the following features:
Lossless 4:1 Z-compression: four times decreases the memory bandwidth consumed by Z-buffer
Second generation Occlusion Culling. As we know from various reviews, Hidden Surface Removal technique is more efficient in RADEON 8500 and KYRO II than in GeForce 3; in GeForce 4, NVidia introduces a new algorithm that will be more effective than its predecessor.
Quadcache Architecture.
The speed of the GeForce 4 Ti4600 based videocards is going to be extremely high compared to GeForce 3 and RADEON 8500. Additionally, there is also a couple of new features so that those interested in capabilities also had a reason to buy a new videocard
The new patent pending type of AntiAliasing will be introduced with the NV25. Its brand name is going to be Accuview AA and represent the new AA mask and technology, that saves one frame buffer write cycle. Accuview AA is another multisampling type of antialiasing. Originally, multisampling needs to write subpixel locations into the memory and only the calculate the pixel color values. Thanks to the new subpixel mask [and the improved algorithm and caching] the entire memory write cycle will be saved causing smaller perfomance decrease compared to the GeForce3. Another interesting feature of the new GPU is the nFiniteFX II engine. Basically, it is the same as on the GeForce 3, but with an additional vertex shader pipeline. Due to the increased frequency, the software that can use the power of the second vertex-shader pipeline will run up to three times faster on the GeForce 4 compared to the predecessor. We can also expect a 50% perfomance gain from the newest solutions when calculation pixel shaders due to the same reason. NVidia declares support for the so-called Z-Correct bump mapping. We couldn't get any more or less precise information about this, and we don't want to speculate on this, so we'll just say nothing this time. The final most important thing about GeForce4 is the support for the nView technology. The new, more feature-rich an easier to use version of TwinView.
That's just a little teaser. I will post more including NV17 (GeForce 4 MX) details, pictures of the new GeForce 4 cards, and info on Hercules ATI cards, the Fire GL cards, 7500DV, and new Kyro. Stay tuned.
[ 01-25-2002: Message edited by: TigerWoods99 ]</p>
Comments
<strong>Here's the goods folks.
NV25 (GeForce 4 Ti4400 & 4600):
*snip really impressive specs I don't understand*
[ 01-25-2002: Message edited by: TigerWoods99 ]</strong><hr></blockquote>
So, does that mean it's 'holy crap!' fast, or just 'smokin' fast? Those are really benchmark terms by the way.
tsukurite
? 0.15 micron.
? 2 rendering pipelines, 2 TMUs each.
? GeForce 4 MX420 - 250 MHz.
? GeForce 4 MX440 - 275 MHz.
? GeForce 4 MX460 - 300 MHz.
? 128, 64 or 32-bit memory interface.
? GeForce 4 MX420 - 166 MHz SDR SDRAM.
? GeForce 4 MX440 - 400 MHz DDR SDRAM.
? GeForce 4 MX460 - 550 MHz DDR SDRAM.
? The most powerful version of the GeForce 4 MX will be equipped with 64 MB of 550 MHz DDR SDRAM memory. Such memory is packed in BGA package.
? The GeForce 4 MX also supports Lightspeed Memory Architecture II, the new version that includes the following features:
? Lossless 4:1 Z-compression: four times decreases the memory bandwidth consumed by Z-buffer.
? Second generation Occlusion Culling
? Quadcache Architecture
? Auto Pre-Charge: The ability to tell the memory device proactively to pre-charge areas of the memory that are not in use but are likely to be used in the very near future.
? Fast Z-Clear: Faster clearing the Z-buffer of if old data.
? Crossbar-Memory Controller: two independent memory controllers, instead of four on the GeForce3 and GeForce4 Ti.
? Accuview AA
The GeForce 4 MX will have integrated TV encoder (maximum resolution 1024x768, support nView technology (dual DVI version is also to be available) and will be HDTV ready. The upcoming graphics processor will also have a video processing engine, MPEG-2 decoder MPEG2 decode algorithm which includes Inverse Quantization (IQ), Inverse Discrete Cosine Transform (IDCT), Motion Compensation, and Color Space Conversion (CSC) functions.
MORE TO COME!
<strong>NV17 aka GeForce4 MX:
• 0.15 micron.
• 2 rendering pipelines, 2 TMUs each.
• GeForce 4 MX420 - 250 MHz.
• GeForce 4 MX440 - 275 MHz.
• GeForce 4 MX460 - 300 MHz.
• 128, 64 or 32-bit memory interface.
• GeForce 4 MX420 - 166 MHz SDR SDRAM.
• GeForce 4 MX440 - 400 MHz DDR SDRAM.
• GeForce 4 MX460 - 550 MHz DDR SDRAM.
• The most powerful version of the GeForce 4 MX will be equipped with 64 MB of 550 MHz DDR SDRAM memory. Such memory is packed in BGA package.
• The GeForce 4 MX also supports Lightspeed Memory Architecture II, the new version that includes the following features:
• Lossless 4:1 Z-compression: four times decreases the memory bandwidth consumed by Z-buffer.
• Second generation Occlusion Culling
• Quadcache Architecture
• Auto Pre-Charge: The ability to tell the memory device proactively to pre-charge areas of the memory that are not in use but are likely to be used in the very near future.
• Fast Z-Clear: Faster clearing the Z-buffer of if old data.
• Crossbar-Memory Controller: two independent memory controllers, instead of four on the GeForce3 and GeForce4 Ti.
• Accuview AA
The GeForce 4 MX will have integrated TV encoder (maximum resolution 1024x768, support nView technology (dual DVI version is also to be available) and will be HDTV ready. The upcoming graphics processor will also have a video processing engine, MPEG-2 decoder MPEG2 decode algorithm which includes Inverse Quantization (IQ), Inverse Discrete Cosine Transform (IDCT), Motion Compensation, and Color Space Conversion (CSC) functions.
MORE TO COME!</strong><hr></blockquote>
All sarcasm aside. Jeezus! This thing's gonna rock! (That beats 'holy crap' as stated in my previous post ).
tsukurite
<strong>NV17 aka GeForce4 MX:
[snip]
The GeForce 4 MX will have integrated TV encoder (maximum resolution 1024x768, support nView technology (dual DVI version is also to be available) and will be HDTV ready. The upcoming graphics processor will also have a video processing engine, MPEG-2 decoder MPEG2 decode algorithm which includes Inverse Quantization (IQ), Inverse Discrete Cosine Transform (IDCT), Motion Compensation, and Color Space Conversion (CSC) functions.
MORE TO COME!</strong><hr></blockquote>
Let the countdown to Kormac rearing his head and saying "See!? I told you."
Of course, John Carmack is right now designing games that already exceed these specs!
Screed
[ 01-25-2002: Message edited by: sCreeD ]</p>
How o I post jpg pictures that I have on my computer? I want to post pics of some the GeForce 4 products and also Fire GL cards. Thanks.
If you got info on these forthcoming nVidia cards, and you say there is also going to be a Quadro version...
Is this going to show up on the Mac platform?
Would only make sense to have a nVidia GeForce4 QuadroDCC to go with Maya on Mac OS X...!
Let us know...
NVIDIA GeForce4 MX 460
NVIDIA GeForce4 MX 440
NVIDIA GeForce4 MX 420
NVIDIA Quadro4 500XGL
NVIDIA Quadro4 200/400NVS
NVIDIA Quadro4 550XGL
NVIDIA GeForce4 Ti 4600
NVIDIA GeForce4 Ti 4400
NVIDIA GeForce4 Ti 4200
NVIDIA Quadro4 900XGL
NVIDIA Quadro4 750XGL
NVIDIA Quadro4 700XGL
NV17
NV17-1
NV17M
GeForce4 420 Go (32M)
GeForce4 440 Go (64M)
Looks like nVidia's next mobile GPU will be based on the GeForce4 MX.
I had an 867 with a GeForce3. Loaded up Giants and half of the graphics options are shaded. Why? Cause even though that game was made to play like a dream on the GeForce3 Apple has yet to implement most of the advatnages of that card into OSX. We never saw FSAA on the Radeon, never saw it on the GeForce3 and I doubt we will see half of the cool technologies that the PC users will enjoy on these cards cause Apple just is not interested in implementing the technology.
<img src="graemlins/hmmm.gif" border="0" alt="[Hmmm]" />
G-News
The high-end GeForce 4 is said to be $399.
<strong>There will also be a Quadro4 apparently.
How o I post jpg pictures that I have on my computer? I want to post pics of some the GeForce 4 products and also Fire GL cards. Thanks.</strong><hr></blockquote>
Put them up on a webpage somewhere and link to them with a URL. If your 'net provider gives you web space, you can use that, or a free hosting service. Or ou can ask someone else with web space to host them for you.
If you use GeoCities, wrap them in a simple HTML page first.
<strong>
Put them up on a webpage somewhere and link to them with a URL. If your 'net provider gives you web space, you can use that, or a free hosting service. Or ou can ask someone else with web space to host them for you.
If you use GeoCities, wrap them in a simple HTML page first.</strong><hr></blockquote>
whats wrong with, ummm ANYONE's iDisk? (you can u/l stuff to the public folder of i'm pretty sure anyone's iDisk) you should be able to link from there right?
<strong>
-63 Million Transistors
</strong><hr></blockquote>
Whoa! More transistors than the G5!
<strong>Apple has yet to implement most of the advatnages of that card into OSX. We never saw FSAA on the Radeon, never saw it on the GeForce3 and I doubt we will see half of the cool technologies that the PC users will enjoy on these cards cause Apple just is not interested in implementing the technology.
</strong><hr></blockquote>
You are correct that these technologies haven't yet been implemented on the Mac, but they are coming. The latest OpenGL ARB minutes includes a lengthy discussion of Apple-initiated extensions to OpenGL to support several of these key features. We ought to see these technologies available by mid-year (earlier to developers so they can start taking advantage of it). The reason for the delay is that Apple wants it to work across all hardware, rather than each developer having to code for older hardware, ATI hardware, and nVidia hardware seperately. This is similar to Microsoft's DirectX8, but it needs buy in from the OpenGL architecture review board (ARB).
In fact Graphics card chips have taken over the CPUs in terms of instructions per second and complexity/transistor count several years ago, roughly with the second generation of graphic GPUs (Voodoo2)
G-News
Also, there is a FireGL 8800 card with 128 MB RAM.
Apparently Guillemot and ATI are coming out with a Radeon 8500XT card, with a 300/300 clock and 128 MB DDR RAM.
Another rumor is that the GeForce4 MX cards will be PCI & AGP. Hmmm.
I'm working on putting the pics of the boxes up. They are tak8ing pre-orders of these cards already apparently on the web.