Another thing that was discusses at B3D. That sort of stuff is in all likelyhood also be done on the SPE. Sure, AI routines are easier to make using heavy integer computations, but there is nothing that makes floating point based AI routines impossible (or so they said).
Whatever BS player is doing, that's incredibly impressive if all those videos are in fact 720x480 resolution DVD vob files. I don't think I can run four MPEG1 videos smoothly on my Cel633 using regular media player.I just did a test with my 350 mhz P2. I ran three mpeg2 videos and one avi video in multiple instances of BSplayer. Here's what I concluded from the test:
What thing?Sony said the same thing about PS2
Marconelly said:Open GL used for PSP and PS3 development is very standardized, don't worry about that![]()
sonycowboy said:So, that would mean a 5.3GH system could do 20 MPEG2's vs the 48 shown here? Of course, I would assume more efficient decoders have been written since then, but then again, we don't know how efficiently the 266 actually handled the MPEG2 decoding, but that was a loooong time ago and the with progressive scan added to MPEG2 since then we're still screwed as to what kind of processing power is needed.
koam said:Is cell going to ever hit the pc/mac market?
Tellaerin said:I'd rather see IBM and Sony challenge the Wintel monopoly with a line of Cell-based personal computers, myself. It'd be nice to see some truly new computers come to market as we move into the 21st century, ones that aren't saddled with decades of legacy hardware and software architecture.
seismologist said:I have an athlon 2ghz (eqivalent to 2.8ghz p4),
I could get 8 avi's running before it started looking bad
but the avi I was playing is divx so I guess that's more
than SDTV resolution
rastex said:And when that general purpose Cell-based computer grinds to a halt because it can't do Out of Order operations I'm sure you'll be smiling then.
HokieJoe said:So to clarify, they're running 48 tiled MPEG2 streams at roughly 240 X 180 resolution/tile.
1920 x 1080 = 2,073,600
2,073,600/48= 43,200/tile
43,200= 240 x 180 resolution/tile
240 x 180= 1.333333333333 aspect ratio, which equates to a 4:3 aspect ratio (the tiles appear to be in a 4:3 format)
Give or take a few pixels, does that estimate sound about right?
gofreak said:You're thinking about things on way too high a level. Forget about concurrency between two different programs for a second, and think about concurrency within a program. There's a lot of concurrency in games waiting to be unlocked. Just because most games currently are single-threaded (and for good reason - hardware threading is a very recent thing, and multiple cores even more recent), doesn't mean they can't be concurrent going forward. They'll have to be, at least if they want to take advantage of the performance on offer...as Dr. Dobb's said, the free lunch is over as far as computing performance is concerned.
Did I say unusable? And a CGI workstation is VERY different compared to a general purpose computer.Tellaerin said:Yes, I'm sure such machines would 'grind to a halt' frequently and be practically unusable, much like the proposed Cell-based CGI workstations will. :lol
rastex said:Did I say unusable? And a CGI workstation is VERY different compared to a general purpose computer.
rastex said:And when that general purpose Cell-based computer grinds to a halt because it can't do Out of Order operations I'm sure you'll be smiling then.
seismologist said:no, they said the video is scaled down so the native resolution is higher than that.
That's the future for limited-function devices like a media server, sure. It makes sense for a device only doing media loads & decodes to have a crippled primary core scheduling tasks and secondary vector cores running them. CELL taking over in these scenarios is the consumer electronics equivalent of outsourcing: Your job is simple and doesn't require brains, just a lot of easily repetitive tasks that go parallel nicely.Vince said:And your argument itself is so very valid as the future of media and consumer electronics will see workloads bounded by how fast you can run MS Word and run SPEC benchmarks. Because, lets face it, the future is in static processing and making sure I can open Thunderbird faster than I can physically realize there is a difference.
At least Intel has large cores which keep the PC user experience functioning as expected by customers.Vince said:And the ironic thing is that even Intel is moving in this direction as their R&D roadmaps all show Cell-esque devices due around 2010 which use one or two large cores surrounded by reductionist|specialized cores reminescent of the SPU.
PizzaFarmer said:That's the future for limited-function devices like a media server, sure. It makes sense for a device only doing media loads & decodes to have a crippled primary core scheduling tasks and secondary vector cores running them. CELL taking over in these scenarios is the consumer electronics equivalent of outsourcing: Your job is simple and doesn't require brains, just a lot of easily repetitive tasks that go parallel nicely.
I don't believe that's the future for PCs (any kind of PC with direct user interaction: home PCs, workstations, etc).
At least Intel has large cores which keep the PC user experience functioning as expected by customers.
Do you seriously believe that an in-order CPU with strict limitations on latency and instruction dispatch can function well in an environment where the machine's owner expects to browse random websites and compose documents, all while their favorite IM client and mp3 player are running in the background?
Having general-purpose C/C++ code perform poorly and requiring SPE use for effective utilization is not a recipe for knocking Intel and AMD off the PC map. You can't just make up for that with a high FLOPS rating reached in a lab environment using a pathological application.
But the PS3 has OpenGL/ES and Cg which are familiar tools too, or more familiar to some than the Windows-only APIs...Srider said:The cell will be good at doing what it does best, people like MS is obviously not going to put out something not competitive. From a practicality viewpoint, xbox2 will likely have an advantage in terms of development than PS3 will at launch, since it's a familiar API to the developers. It'd be interesting to see if the PS3 gets delayed due to not enough games available at launch. We'll see at E3.
So what? This mpeg2 decoding demo shows the realworld performance of the Cell, not on paper.Srider said:Remember when the Emotion Engine came out and Sony was boasting how it can do crazy calculations comparing it to super computers? Look at PS2 now. Most of these things are simply marketing to make it appealing to potential clients. I'll believe it when a game developer gets his/her hands on the Cell and put out some good looking stuff.
Srider said:I'm simply stating that the performance gain is not necessarily as much as people claim it is.
Srider said:Remember when the Emotion Engine came out and Sony was boasting how it can do crazy calculations comparing it to super computers? Look at PS2 now
Ryudo said:Very interesting, if not useless. Does anyone honestly think the same chip will be in the PS3 for < 300 dollars ?
"PS3 Linux Kit" is the way to gogofreak said:Invoking DP performance as a reason why performance gain "isn't as much as people claim" isn't very sound. Again, compare Cell's DP performance with that of a desktop CPU.
As for Cell in PCs, I don't see it happening in the short/medium term. Apple is probably the only short/medium term avenue for getting Cell into desktop PCs, and even then it'd be as an add-on processor rather than a CPU replacement. Technicalities aside, there's a huge software mountain to overcome. Besides, STI has barely discussed its chances as a desktop replacement themselves..it's not a market they're targetting for now.
ThirdEye said:"PS3 Linux Kit" is the way to go![]()
gofreak said:As for Cell in PCs, I don't see it happening in the short/medium term. Apple is probably the only short/medium term avenue for getting Cell into desktop PCs, and even then it'd be as an add-on processor rather than a CPU replacement. Technicalities aside, there's a huge software mountain to overcome. Besides, STI has barely discussed its chances as a desktop replacement themselves..it's not a market they're targetting for now
Yes, static processing tasks performed in arbitrary order and with an arbitrary number of them running at any given time.Vince said:So, you're saying that if you were to do an abstract asymptotic analysis of future consumer electronics computational demands, you'll find that they're bounded by static processing tasks?
Well, you did respond, because clearly you're above the rest of us discussing the application of common sense and user scenarios to theory and expectations.Vince said:I don't even need to respond to this, it's that asinine.
Mind sharing how much experience you have trying to run today's general-purpose code on an in-order processor with tight cache restrictions?Vince said:First off, running all those apps on current processors utilized under 10% of the aggregate resource (eg. They aren't a limit on preformance); Secondly they aren't inheriently incapable of being utilized effecienctly in the BPA.
I used Word and web-surfing as a few popular examples of "modern general-purpose applications written in C/C++". More conjecture about CELL being powerful enough and your future of a media processing-based future.Vince said:How many people are upgrading their Pentium4 or even their P2 because it doesn't browse the web fast enough or they can't use MS Word enough? Is there anyone who believes these apps are driving demand? Of course not, you're insane if you believe this and the proof is in the talk of Longhorn becoming a demand driver and forcing turn-over -- and why is this... perhaps they're focus on digital media and demands on dynamic processing requirements? The GPU will pick-up alot of these demands, but Cell|BPA is more than capable.
In theory, this matters how? x86 has userbase. x64 has a great migration story. Intel and AMD are supposed to be shaking in their boots because of a different architecture slanted towards running software scenarios people aren't willing to pay for yet?Vince said:Intel is also 5 years behind, lets compare the 3rd Generation Cell processors being fabbed at 32nm and then see how they relate. The difference is that STI has had Cell ICs in developers and DCC hands since 2004; Intel has a few powerpoint presentations that have cool looking pictures next to the year 2010.
If only you could randomly pull a constant number for perf loss. The problem is that you've introduced bottlenecks throughout the design that didn't exist there before with a potentially huge existing number of inputs to the system that affect the resulting numbers.Vince said:Uh, yes. Lets just pull a random number and say you'll lose a net 50% preformance by getting rid of OoO when running your tasks, the 1st generation BPA's PPE (forgetting about SMT, etc) is clocked at over 4GHz. You're saying it won't be able to run MS Word, AIM and your shitty Virus Scanner to the point that the average consumer will notice the difference? This being the same consumer who is likley running a Celeron. And from an architectural PoV, OoO is a net loss of effeciency when compared with the area it utilizes - it's a losing proposition.
That's comedy gold right there.Vince said:Totally fallicious comment. I can only offer anecdotal evidence at this point, but JFYI, the recent Cool Chips demonstration by Toshiba which showed off the 6 SPUs decoding 48 MPEG2 streams concurrently and resizing them to fit in a composite 1920x1080 screen was written without the programmer doing any explicit thread scheduling. Quite the nightmare development enviroment, huh?
In your rush to show us how much you know about the history of the P4, you read my post too quickly. I didn't say the P4 was crippled for favoring SSE. I described a possible foolish path Intel could take by crippling the rest of the P4 at the expense of more SSE functionality.Vince said:PS. And BTW, the Pentium4 (Northwood) wasn't crippled because of favoring SSE. It was due to the engineering going out of control in terms of transistor budget and what was becoming an IC that was so far gone in terms of diminishing returns per area cost that Intel's Dan Boggs capped the area at that of the PPro and cut the shit out of it.
At least without OOE I don't have to worry about my FPU calculations being unpredictable.And when that general purpose Cell-based computer grinds to a halt because it can't do Out of Order operations I'm sure you'll be smiling then.
PizzaFarmer said:Yes, static processing tasks performed in arbitrary order and with an arbitrary number of them running at any given time.
I used Word and web-surfing as a few popular examples of "modern general-purpose applications written in C/C++". More conjecture about CELL being powerful enough and your future of a media processing-based future.
Then again, maybe all people will care about in five years is running Mathematica and watching ten DVDs at the same time, which would make you right and the rest of us bored out of our minds. That, and BeOS would rise from the dead to be the OS of choice for parallel mpeg-decode lovers everywhere![]()
If only you could randomly pull a constant number for perf loss. The problem is that you've introduced bottlenecks throughout the design that didn't exist there before with a potentially huge existing number of inputs to the system that affect the resulting numbers.
I pointed out that CELL is slanted towards pathological lab scenarios and your Exhibit A is a pathological lab scenario. You sure proved me wrong!
In your rush to show us how much you know about the history of the P4, you read my post too quickly. I didn't say the P4 was crippled for favoring SSE. I described a possible foolish path Intel could take by crippling the rest of the P4 at the expense of more SSE functionality.
I actually don't think our views on the user scenarios differ that much.Vince said:Which is seen in what task that will bring the PPEs in Cell or the XCPU to it's knees? We seem to have fundimentally differing views of the future of programming. I see the future based around digital media and interactivity, removing the static pages and UI of today with dynamic ones. You, obviously, don't see this as the future because if you did the bounds on what limits preformance would be more aligned with what I'm talking about, or what was stated in the links I provided..
And here's an edit of my own now that I noticed you editing old postsVince said:Outiside of stating that you program the SPUs in C/C++ and that the SPU architecture was designed by guys like Gschwind who had the compilers in mind as they wrote most of them
That's for how GCC works with SIMD aspects of SPEs - doesn't mean every other compiler does/will do the same. Some compilers DO come with native support for SIMD types(not talking specifically about Cell here) you know.In a sense, you can program the SPEs in C/C++, but intrinsics don't count.
There's nothing stopping SPEs from running general purpose code. Whether that's efficient use of the processor is up for debate, but then most windows software we run today doesn't utilize hw in most efficient ways either.I'm talking about absolute general-purpose code being offloaded through the magic of copy/paste in a text editor.
Definitely a fair point considering that there will probably be other compilers available for PS3 that expose the SPEs in different ways. Just amusing to see Sony trumpeting around the "higher-level languages this time around!" banner despite the actual implementation described so far.Fafalada said:That's for how GCC works with SIMD aspects of SPEs - doesn't mean every other compiler does/will do the same. Some compilers DO come with native support for SIMD types(not talking specifically about Cell here) you know.
How good is integer/branch support on the SPEs? I haven't seen much coverage of that besides commentary on the lack ...Fafalada said:There's nothing stopping SPEs from running general purpose code. Whether that's efficient use of the processor is up for debate, but then most windows software we run today doesn't utilize hw in most efficient ways either.
sly said:Sorry for interrupting your discussion but judging by the lack of "ooohs" and "aaahs" in this thread I'm sensing that the demo isn't that impressive. Or is it? Hypothetically, what proccessor available today will carry out the same task with ease?
We'll see as they reveal more. Anyway GCC as a starting solution obviously isn't a top choice - but I gather Sony chose it over paying royalties for IBM solutionsDefinitely a fair point considering that there will probably be other compilers available for PS3 that expose the SPEs in different ways. Just amusing to see Sony trumpeting around the "higher-level languages this time around!" banner despite the actual implementation described so far.
It's pretty complete from what I've heard so far. Admitedly I am hesitant to call any SoA SIMD solution "complete", but that's a FPU specific complaint.How good is integer/branch support on the SPEs? I haven't seen much coverage of that besides commentary on the lack ...
Well the analogy is running code not particularly designed for the target CPUAnd of course, there's a difference between general-purpose code PC utilizing an x86 CPU poorly because the code wasn't designed for it and general-purpose code trying to run well on a CPU designed around vector operations.
I see what you mean, but games are not what I'd call 'software people are not willing to pay for', and assuming that Cell indeed performs really well at that task, at a lower price - that alone justifies existence of that architecture IMO. Not just games, but things like multiple realtime thumbnail previews of HDTV streems seem just handy for future TVs. I agree in that Intel has nothing to worry about anytime soon, but there is obviously something to gaming with multicore CPUs, considering that even MS abandonesd Intel as an Xbox client to IBM's another multicore chip.In theory, this matters how? x86 has userbase. x64 has a great migration story. Intel and AMD are supposed to be shaking in their boots because of a different architecture slanted towards running software scenarios people aren't willing to pay for yet?
gofreak said:I thought I saw at one point a longhorn demo that showed multiple movie streams on a desktop, all rotated and skewed in different ways (just to show the new GUI)..it had more than 2 on screen, though not 48Anyone else remember that? I can't remember if they were all the same stream, just replicated, or not though..