QUOTE(jcliew @ Mar 1 2007, 03:08 PM)
tera flops (FLOPS-Floating Point Operations Per Sec)tera bits (a unit of binary data, either bandwidth or storage space )
err...
Definitive R6xx thread, Some more solid info
|
|
Mar 1 2007, 03:36 PM
Return to original view | Post
#1
|
|
Elite
6,139 posts Joined: Jan 2003 |
|
|
|
|
|
|
Apr 19 2007, 11:58 PM
Return to original view | Post
#2
|
|
Elite
6,139 posts Joined: Jan 2003 |
|
|
|
May 7 2007, 04:55 PM
Return to original view | Post
#3
|
|
Elite
6,139 posts Joined: Jan 2003 |
i think i saw a R600 demo unit in compuzone yesterday...
8pin + 6pin PCIE Aux power.. rite?? and and red flamy plastic shroud covering the heatsink This post has been edited by jinaun: May 7 2007, 04:56 PM |
|
|
May 7 2007, 05:06 PM
Return to original view | Post
#4
|
|
Elite
6,139 posts Joined: Jan 2003 |
QUOTE(empire23 @ May 7 2007, 04:58 PM) Saw those, unless specifically run, most of them are just show models. But here's a tip, MSI and Powercolor have awesome coolers, silent looking ones too. Usus, entah, they only sent a box to look at, a giant box. yeah.. they are running 3dmark06.. Hardware isn't here, not by a long shot. at the scene where one heavy armored guy/w heavy firepower shooting at those pathetic people behind those boxes.. averaging abt 55~fps at that point if its not R600.. then its something else disguised to look like R600.. LOL This post has been edited by jinaun: May 7 2007, 05:11 PM |
|
|
May 7 2007, 10:18 PM
Return to original view | Post
#5
|
|
Elite
6,139 posts Joined: Jan 2003 |
QUOTE(almostthere @ May 7 2007, 09:11 PM) for the LAST time, repeat after me, "R600 WILL BE 80nm", expect no less then that. And I can bet 3 pieces of ikan bawal masak sambal it won't occur during R600's lifetime probably it will face the same fate at R520... quickly replaced with R580i'll wait for R680+ then... LOL.. |
|
|
May 8 2007, 07:21 PM
Return to original view | Post
#6
|
|
Elite
6,139 posts Joined: Jan 2003 |
|
|
|
|
|
|
May 9 2007, 07:18 AM
Return to original view | Post
#7
|
|
Elite
6,139 posts Joined: Jan 2003 |
fudzilla mentioned that :
QUOTE R650 is the Radeon HD 2950XTX ATI hasn't launched the R600 yet and it is already planning its next launch of a die shrinked version. It won't be launching the 1024 MB graphics card until Q3. It plans to launch the R650 chip and brand it as Radeon HD 2950XTX, 1024 MB GDDR4. The chip is called R650 and it is a 65 nanometre core, how convenient. With this new chip ATI might have a fighting chance against the Geforce 8800 Ultra, but the real question is what will Nvidia have at that time. The cards are scheduled for late Q3, so you should expect them to arrive in August or September, unless they get postponed. http://www.fudzilla.com/index.php?option=c...id=871&Itemid=1 |
|
|
May 26 2007, 04:12 PM
Return to original view | Post
#8
|
|
Elite
6,139 posts Joined: Jan 2003 |
QUOTE Radeon HD 2900 XT lacks UVD video acceleration by Scott Wasson - 01:38 pm, May 25, 2007 I've just learned something that compels me to publish a major correction to our review of the Radeon HD 2900 XT GPU. I got the clear, distinct impression from AMD's presentations, statements, and conversations with me at its Radeon HD press event that its new UVD video decode acceleration logic was present in its entire lineup of Radeon HD graphics chips, and I relayed that information to you in our review of the product, promising to follow up with tests of this feature at a later date. True to my word, I set out yesterday to test HD video decode acceleration on a Radeon HD 2900 XT using an HD DVD drive and a version of PowerDVD supplied by AMD for such purposes. To my surprise, CPU utilization during playback on our Core 2 Extreme X6800 test system ran between 40 and 50%, well above what one would expect from a solution with full HD video decode acceleration. Naturally, I contacted AMD to inquire about the problem. I received a reply from AMD's David Baumann discussing the issue that ended with this revelation: Be aware, though, that HD 2900 XT itself does not feature UVD, this is specific to 2600 and 2400, so the levels of CPU utilization you see should be somewhat similar to the previous generation. The UVD logic handles the CPU-intensive bitstream processing and entropy decode acceleration portions of the HD video playback pipeline. These are the most notable video decode acceleration capabilities that would separate the Radeon HD 2900 series from its direct competition, the GeForce 8800 series, if the HD 2900 XT actually had them. Turns out it does not. As the email states, the video playback capabilities of the Radeon HD 2900 XT are essentially similar to those of the previous-gen Radeon X1950. So the essence of our correction is that the Radeon HD 2900 XT doesn't offer robust acceleration of HD video playback and will not likely reduce CPU utilization or power consumption substantially during high-definition video playback versus a GeForce 8800. We still intend to follow up with testing, but the lack of UVD logic on the GPU resets our expectations dramatically. With that out of the way, I believe I ought to take a moment to explain how we came to believe the Radeon HD 2900 XT had full HD video playback acceleration, an impression formed by many layers of talk from AMD, starting with the Radeon HD name. Let me share a slide with you from a presentation titled "ATI Radeon HD 2000 Series and the Ultimate HD Experience," given by AMD's David Cummings, Director of Mobile GPU Marketing. The slide looks like so: You can, of course, read for yourself that it says "Avivo HD technology makes full spec HD DVD / Blu-Ray (HD Disc) playback accessible at all price points," but I just like repeating it. That gives one a certain idea, does it not? Now, let's have a look at another slide showing what Avivo HD brings to video decode acceleration: The bit labeled "Avivo HD" shows GPU acceleration of bitstream processing and entropy decode, and makes clear it's distinct from the Radeon X1000's Avivo video processing, which lacks acceleration of those stages. Now, look at any specs list for the Radeon HD 2900 XT-say, this one from AMD's website, and you will find listed among its specs "ATI Avivo(tm) HD Video and Display Platform" and a bullet point under that saying "HD decode acceleration for H.264/AVC, VC-1, DivX and MPEG-2 video formats." At the end of the day, one gets the impression that this GPU has Avivo HD, with all that entails. Of course, AMD has left itself some wiggle room in its technical statements. The specs list above isn't technically untrue-just imprecise. The dodge built into the Cummings presentation, with its talk of making HD video playback "accessible at all price points" seems to be that high-end CPUs can handle HD video playback without as much assistance from the GPU. But that's a paper thin excuse, in my view. To make sure this wasn't simply a matter of me missing the boat-it has been known to happen, and I've got a few gray hairs promising more of the same in the future-I checked with a couple of other journalists who attended a separate Radeon HD press event the week after the one I attended. Both Marco Chiappetta from HotHardware and Ryan Shrout of PC Perspective came away from their meetings with AMD convinced the Radeon HD 2900 XT had full HD playback acceleration via UVD logic, as well. I was not alone in gathering this impression from AMD. To their credit, some reviewers did sort through the fog and identify the Radeon HD 2900 XT's lack of UVD, but they were swimming against the tide of statements from AMD itself. Nor could any of us have uncovered this fact prior to the publication of our reviews via testing, because AMD hasn't yet delivered a driver that includes the support for the Radeon HD's "full" multimedia capabilities. They initially targeted May 9 for that driver's release. AMD now says the driver is due next week. http://www.techreport.com/onearticle.x/12552 will it be an issue or not...? This post has been edited by jinaun: May 26 2007, 04:13 PM |
|
|
Dec 16 2007, 03:55 PM
Return to original view | Post
#9
|
|
Elite
6,139 posts Joined: Jan 2003 |
|
|
|
Dec 16 2007, 04:31 PM
Return to original view | Post
#10
|
|
Elite
6,139 posts Joined: Jan 2003 |
QUOTE(ben_panced @ Dec 16 2007, 04:09 PM) its embedded into the GPU with total bus width of 2560-bit (composed of three independent buses: 1024-bit write, 1024-bit read, 512-bit read/write) working at 8 GHz» Click to show Spoiler - click again to hide... « This post has been edited by jinaun: Dec 16 2007, 04:40 PM |
|
|
Dec 17 2007, 11:09 PM
Return to original view | Post
#11
|
|
Elite
6,139 posts Joined: Jan 2003 |
QUOTE(clayclws @ Dec 17 2007, 08:16 PM) I'm worried about the GDDR3 usage...it's not as energy efficient or OC-able as GDDR4. They've used GDDR4...why are they taking a step back for their flagship product? Hope it's not accurately reported. perhaps GDDR3 is much cheaper.. and clock for clock.. is faster than GDDR4 due to additional latencies associated with GDDR4GDDR4 will onli be faster than GDDR3 if its clocked high enough.. eg.. take DDR400 vs DDR2-400 |
| Change to: | 0.0254sec
0.66
7 queries
GZIP Disabled
Time is now: 21st December 2025 - 07:36 AM |