Metaprogramming GPUs with Sh 139
Martin Ecker writes "With the advent of powerful, programmable GPUs in consumer graphics hardware, an increasing number of shading languages to program these GPUs has become available. One quite interesting language that - in many ways - has a very different approach than other mainstream shading languages (such as Cg or the OpenGL Shading Language) is Sh. The recently released book "Metaprogramming GPUs with Sh" by Michael McCool and Stefanus Du Toit, both major contributors to the Sh project, explains the basics of the Sh high-level shading language and the corresponding API and also goes into some of the details of the implementation. The book is intended for an audience that is already familiar with traditional shader development for programmable GPUs. Also, a firm background in 3D graphics programming and C++ is a must for the interested reader." Read on for the rest.
Metaprogramming GPUs with Sh | |
author | Michael McCool, Stefanus Du Toit |
pages | 308 |
publisher | A K Peters |
rating | 7/10 |
reviewer | Martin Ecker |
ISBN | 0321197895 |
summary | A book that describes an interesting shading language and accompanying API to program GPUs. |
Before discussing the book in more detail, I will try to give a basic overview of Sh, since most readers will not be familiar with it. For a more in-depth look at Sh, I recommend taking a look at a recently posted Gamasutra article by Michael McCool (http://www.gamasutra.com/features/20040716/mccool _01.shtml), the paper on Sh from the authors presented at the recently held SIGGRAPH 2004 conference (http://www.cgl.uwaterloo.ca/Projects/rendering/Pa pers/#algebra), and of course the Sh homepage at http://www.libsh.org.
Sh started out as a research project at the University of Waterloo (http://www.cgl.uwaterloo.ca), and it is both a shading language and a runtime API to use the Sh shaders. As a shading language Sh is embedded into C++ as a domain-specific language, which is made possible by using C++ operator overloading and by defining special tuple and matrix types that are used extensively in shader code. So instead of defining its own language that requires a full compiler, like other shading languages do, Sh uses regular C++ syntax to describe shader code, which is then dynamically (at runtime) compiled to a specific backend, such as a GPU or possibly even the CPU. In addition to compiling to a specific GPU or CPU target, Sh can also be used in a special stream mode where a shader is applied to a stream of input tuples. This is very useful for general purpose GPU programming where the GPU is basically used as an additional processor to the host CPU (see http://www.gpgpu.org for more information on the subject). Finally, Sh code can also be executed in an immediate mode where every Sh statement is directly executed on the host CPU (without being compiled into a shader program), which makes it very easy to debug shaders with any host debugger running on the CPU.
Due to the way Sh is embedded into C++, the full range of abstraction mechanisms offered by C++ can be used to structure and modularize shader code. Abstract base classes, regular functions, templates, and any other features offered by C++ can be used to develop shaders. This is an interesting consequence of the metaprogramming approach of Sh that also allows the use of software engineering principles in shader development, such as object orientation, that other shading languages currently cannot offer.
This kind of metaprogramming in C++ is used by an increasing number of libraries. For example, the Spirit parser framework (see http://spirit.sourceforge.net) uses a similar approach to describe and generate parsers directly in C++ instead of using traditional external tools, such as yacc or bison.
One of the most fascinating features of the Sh toolkit is the possibility to combine and connect shader programs to form new shader programs, which allows one to easily build complex shaders out of simple shader fragments. In a more general sense, Sh provides what can be called a shader algebra (see also the aforementioned SIGGRAPH 2004 paper), where shader programs are the objects on which special operators to combine and connect them are defined. An interesting application of this shader algebra is to specifically bind certain varying shader inputs to uniform variables and the other way around (this is what functional programming languages usually call currying). Also combining a matrix palette skinning shader with any light model shader (or any shaders that perform specific tasks, for that matter) is easily possible.
After this short introduction to the Sh toolkit, we shall now take a closer look at the book "Metaprogramming GPUs with Sh".
The book is split into three parts, an introduction, a reference, and an engineering overview.
The introduction consists of the first five chapters and discusses the basics of the Sh shading language and the API. In particular, the tuple and matrix types and the operators defined on them are presented. The way shader programs are defined and how parameters and attributes are handled is discussed, followed by the way textures are represented. Finally, the stream and channel concept used to feed data into shader programs is discussed. These introductory chapters contain a number of examples that demonstrate the presented concepts. Chapter three contains a quite interesting sample shader that uses constructive solid geometry techniques and metaprogramming in Sh to render text. While not the most useful use case, the shader shows some interesting capabilities of Sh, in particular the shader algebra operators. Chapter four on textures has some more nice sample shaders for doing shiny bump mapping, rendering wood and marble, and using Worley noise.
The second part of the book is a reference on Sh. Unlike references in many other computer books, this is not just a technical listing of the available features of Sh but is written in regular prose (with the occasional reference-like table here and there). The six chapters of the reference section describe how to setup and use the Sh library, and then discuss the available types, operators, and standard library functions more thoroughly than in the introduction. Additionally, the available backends are mentioned in the last chapter of this part of the book. A draft of the reference manual can also be found online at http://www.libsh.org/ref/online.
The final part of the book deals with engineering aspects of Sh. These final five chapters of the book discuss the details of the current implementation. The intermediate representation for shaders that is used by Sh is presented as well as how streams and textures are managed and stored internally. The interface between the Sh frontend and the various specific backends is discussed, as well as the current state of the optimizer including some further improvements that are planned in the future.
The images in the book are all in black and white except for 14 color plates in the middle of the book. The color plates and other images usually show teapots or animals, so they aren't all that exciting, but do demonstrate what the sample shaders presented in the book look like.
The book does not come with a CD-ROM, but with such a young library that is still under heavy development, putting a snapshot of the library's source code base on a CD-ROM would be a waste of resources. Sh itself as well as all sample shaders presented in the book can be downloaded from the Sh homepage at http://www.libsh.org. This website also has additional documentation, including some papers and the API reference documentation generated with Doxygen from the sources. Sh is distributed under a very liberal open source license (based on the zlib/libpng license) that also allows commercial use.
For the reader with enough expertise in 3D and shader programming, this book provides a concise and well-written introduction to Sh. The book will definitely contribute to enlarging the currently relative small user base of Sh and hopefully help the library grow and get more refined in the near future. Everyone familiar with "regular" high-level shading languages, such as Cg or the OpenGL Shading Language, should take a look at this book to see a new and interesting way of programming GPUs that the aforementioned languages do not offer.
About the review author:
The author has been involved in real-time graphics programming for more than 9 years and works as a games developer for arcade games. In his rare spare time he works on a graphics-related open source project called XEngine http://xengine.sourceforge.net.
You can purchase Metaprogramming GPUs with Sh from bn.com. Slashdot welcomes readers' book reviews -- to see your own review here, read the book review guidelines, then visit the submission page.
Wow (Score:4, Funny)
Re:Wow (Score:1, Offtopic)
Dude,
Playstation Super-computers and Clusters (Score:2)
There have been systems like this using DSPs for years; I'v
Anyone Else Notice... (Score:1, Offtopic)
Template broken? (Score:2, Offtopic)
Front Page Story (Score:4, Funny)
That is sooo... awesome.... (Score:5, Funny)
Re: (Score:2)
Re:That is sooo... awesome.... (Score:3, Informative)
Re:That is sooo... awesome.... (Score:1)
Call it "shlang" if it makes you happy. Also easier to pronounce ;-). But it's not really a language; it's really a library. The website is http://libsh.org/ [libsh.org], so calling it "libsh" is also acceptable if
Largest tease post evar! (Score:1, Offtopic)
CB
/bin/sh (Score:5, Funny)
Re:/bin/sh (Score:3, Interesting)
Yes I know-- sh in this case is shading language...
But on a more serious note, is there any reason why once cannot write to the video card (as root) with a bash script? Is there any technological reason why one cannot send in the OpenGL information this way?
If not, maybe we need command-line programs to connect to irrlicht then we can do all this
Re:/bin/sh (Score:1)
bash -> Bourne Again Shell
Re:/bin/sh (Score:2)
Re:/bin/sh (Score:2)
Why? From the bash man page:
Bottom line: bash when invoked as #!/bin/sh will behave pretty much the same as sh. Bash invoked as #!/bin/bash behaves much more sanely :-).
Try writing a shell script under linux invoked as #!/bin/sh, then write one invoked as #!/bin/b
Re:/bin/sh (Score:2)
Re:/bin/sh (Score:4, Interesting)
Re:/bin/sh (Score:1)
Re:/bin/sh (Score:2)
Re:/bin/sh (Score:2)
Don't feel bad, I read the title the same way. Maybe they should change the title to not be so misleading, maybe something with fire in the name.
There is no reason why you couldn't modify bash to add internals to support opengl calls, but bash is completely interpreted, a
Re:/bin/sh (Score:2)
Re:/bin/sh (Score:2)
SDL-perl (Score:4, Insightful)
This is a shame, because perl's a nice quick way of working through the edit-try-edit-try-edit development process when working with OpenGL and 3D geometrical stuff in general.
I don't want to learn Yet Another Scripting Language just to try out some OpenGL routines, and I don't want to slog through a ton of C code makefile junk either. Script languages are a great help there.
Re:SDL-perl (Score:1)
lisp for games (Score:1)
In college, my prof made us write a game in LISP just to prove that some languages are bad at certain things.
Just a data point: Naughty Dog is well-known for the heavy use of Lisp (an in-house dialect called GOAL) in their games, including the Crash Bandicoot and Jak and Daxter series.
Amazon linkage (Score:4, Informative)
Re:Hilarioius (Score:2)
Re:Hilarioius (Score:1)
Re:Hilarioius (Score:2)
Jason
ProfQuotes [profquotes.com]
Re:Hilarioius (Score:2)
/bin/sh (Score:4, Funny)
sed | awk | sed | grep | awk | sed | awk
sh? (Score:2, Funny)
lol (Score:3, Insightful)
lol, how many languages do we need? back in the day, knowing 2 or 3 languages very well (like c and some db language) would have been enough to land a job. add to that knowing another 2 or 3 languages well enough to feel comfortable doing small coding, and you have an excellent programmer. now it seems like there are so many languages, from python to cobol to fortran. there must easily be over 50. it is like if i wanted to work at the united nations and said i knew english and spanish and german, and they asked "well what about sanskrit and tagalog". how much is enough?
Re:lol (Score:1)
In the real world, I would expect a professional programmer with a background in OO to be able to perform competently in any OO language with very little learning curve. The only thing that really changes from language to language is the syntax.
Big paradigm shifts have been few (Sequential to functional to object oriented). There are of co
Re:lol (Score:2)
In the real world, I would expect a professional programmer with a background in OO to be able to perform competently in any OO language with very little learning curve. The only thing that really changes from language to language is the syntax.
Have to disagree with you there. OO languages are different in implementation, shorcomings, and strengths depending on the language in question. Java can do things that are difficult in Ruby, and vice versa (functions as objects), and Smalltalk can do things that
Re:lol (Score:2)
In college, I watched entire classes go through the same intro topics (how hard is it to switch between C++ and Java when you're only talking about freakin' conditionals?!) for weeks... I th
Re:lol (Score:1)
I guess it depends on how similar the languages are, but from personal experience:
a good Java programmer != a good C++ programmer, and vice versa.
Re:lol (Score:2)
How many do we want? More.
Imagine I want to do some quick parsing. Both C++ and Java suck at this. So I'll use Perl instead.
Imagine I want to write code for a GPU. Sure, I could do it in bytecode. I could do it with some library for C++. Or I could write a custom language - say, Cg or Sh - that works well for that purpose.
Imagine I want
Re:lol (Score:2)
This is exactly what I thought, except I never knew anyone who works doing assembly. It might be nice for some theory, but in practice it is never applied. I would think someone who is expert in assembly has a small niche market for themseleves and they probably make good money. My thinking is either C or C++ is a must (or maybe Java, it
Re:lol (Score:2)
It is a language that it is simple to get started with, sure, but it is, for a professional programmer, a language full of tedium and details that need to be remembered.
EIvind.
Re:lol (Score:1)
First off, I totally agree with this, people always make the analogy that languages are like tools and each job should be implemented with the most appropriate tool. Who would argue against more tools in the workshop?
But there's a huge missing link here: Languages are horribly inoperable. Even C to C++ requires some tomfoolery like
#ifdef CPLUSPLUS ...
extern "C" {
#endif
#include
#ifdef CPLUSPLUS
}
#endif
Is this a joke? The problem is that once you've used that nice band
Re:lol (Score:2)
You're right, with some exceptions, languages aren't very interoperable. Python is an exception, and the entire
COM and CORBA and so on and so forth do a reasonably good job of bridging most gaps, but unfortunately t
Re:lol (Score:1)
Re:lol (Score:4, Insightful)
Excellent plan.
If nothing else, it serves to confuse outsiders (Score:3, Funny)
What about other kinds of GPU programming? (Score:5, Interesting)
Re:What about other kinds of GPU programming? (Score:2)
As far as I know, one problem is the AGP interface, which gets data from main memory to the graphics card fast but not the other way. This will be fixed with PCI Express, so maybe when more computers have that we will also see some non-display applications of GPUs.
Re:What about other kinds of GPU programming? (Score:2)
Using GPU for searching in text and stuff like that is not useful though. The point with a GPU is that it is massively parallel and has very deep pipelines. As such it is perfect for doing the same computation on a large image but pretty us
Re:Not really. (Score:4, Interesting)
It is one place that an OS architecture which makes a clean separation of GUI and kernel would win over the mess of sundry .dlls from the Criminal Monopoly. The other place would have been the ill-fated tablet computer, had it been correctly designed, with the X serevr on the portable bit.
Because simple single-processor relatively high-performance silicon is now very affordable, the time is right to go back to parallelism, dedicated I/O and graphics processors, etc, to get even better performance.
It would be nice to have a GPU compiler to run ordinary code, such as X, but the compiler would be different for each architecture, not like AMD vs Intel, where you can use the same compiler with just the optimisations tweaked. But my thinking is to replicate a full CPU in the GPU chip or at least on the graphics card. You could put a Pentium/Athlon there, feeding the gPU as usual, fed itself from the AGP bus or its successor via a big FIFO.
But whatever way things go, I do think we are in for some big steps forward in graphics processing, at least. I am not into shaders and so on, I don't need fast 3-D or anything like that, but I know that many people do, and not just for games. But if it offloads things from my main CPU, I would find immediate benefit.
Re:Not really. (Score:1)
The other place would have been the ill-fated tablet computer, had it been correctly designed, with the X serevr on the portable bit.
I thought about this a couple of months ago - how great it'd be to have a tablet PC that was literally just an X/VNC/RDC/Citrix server. With a GPU that most of the hard work could be dedicated to, and possibly something like a Transmeta/C3 etc type CPU for networking, device management etc - which wouldn't require much
Re:Not really. (Score:2)
Putting all of X on a GPU would probably not be a good idea however. (For reasons to be mentioned later.) And while you can free up your CPU from the specifics of GU
Slight digression (Score:2)
The only problem with that is that when there is a lot of work, they are often both working full-out already. I agree it would be nice if GPUs could also do less specialized (i.e. completely different) parallel processes, but I am sure someone will be able to hack even the current generation to do say weather modelling if they really want to.
Anyway, to dig
Sh? maybe. Brook? Definitely. (Score:5, Informative)
Brook [stanford.edu], on the other hand, was written from a more C-like perspective, and approaches the GPU as a massively data-parallel stream processor (well, Sh does as well, but IMHO Brook achieves that aim more directly as is evidenced by things like iterator streams and similar kernels).
Not there yet for "real" interactive framerates (Score:5, Informative)
I think it will be a while before sh / GPU metaprogramming will be commonly used for "real" games programming. For example, their paper on Worley shaders claimed interactive rates of 14 FPS on a very simple single model for stone shading on the fastest video hardware (6800GT) currently available.
The benefit of HLSL and Cg are that they achieve performance close to or better than the PS/VS-Asembler implementations and are orders of magnitude easier to program. This allows them to be used on hundreds of objects per frame at video game "interactive" framerates (which usually starts at 30 FPS, with 60 FPS as a gold standard, and 300 FPS or other crazy numbers as the Doom/Quake standard for some reason).
Still it's good to see that there are languages evolving that will handle more complex shading algorithms and as the hardware becomes faster in 2 or 3 years from now, this may be practical for real-time use in video games.
Re:Not there yet for "real" interactive framerates (Score:5, Informative)
(Disclaimer: I'm Stefanus Du Toit, one of the authors of this book and implementer of most of the Sh compiler/library)
I think it will be a while before sh / GPU metaprogramming will be commonly used for "real" games programming. For example, their paper on Worley shaders claimed interactive rates of 14 FPS on a very simple single model for stone shading on the fastest video hardware (6800GT) currently available.
I should point out that the Worley shaders are rather large shaders. It's not that they're written in Sh that makes them so huge, it's simply the way the algorithm works. The shaders would be approximately the same size if written in Cg (or HLSL, etc.).
The benefit of HLSL and Cg are that they achieve performance close to or better than the PS/VS-Asembler implementations and are orders of magnitude easier to program.
The same holds true for Sh. Both Sh and Cg target "assembly" interfaces such as RB_fragment_program. Sh provides a superset of Cg's functionality. Because in both the case of Sh and Cg the actual program is executed by the GPU, not in any way the system itself, the only difference in performance that might be caused between the two languages would be due to the optimizer. I've actually just rewritten the Sh optimizer, and the only major optimization left to do is common subexpression elimination, and some arithmetic simplifications.
In fact, we already provide an optimization that no other shading language provides (this will be in the next release, and is working in the subversion repository) called uniform lifting. Sh will automatically discover computations which do not change from one fragment to the next, and move them to the host where they can be evaluated more efficiently. This is something that would be difficult to do in other shading languages.
So, to summarize, there is nothing intrinsic about Sh that makes it any slower than Cg. In fact, we're thinking of adding a Cg backend, so that your Sh code can automatically be converted to equivalent Cg code. Then, if you really want, you can take advantage of the backends and optimizations provided by Cg while using the modularity facilities of Sh.
Re:Not there yet for "real" interactive framerates (Score:5, Interesting)
And here's a question that's sure to anger the general Slashdot crowd -- however, the answer will be important in whether or not your language makes it into common usage in the game industry. How do you feel about the use of your technology in closed source proprietary products (like video games)?
Re:Not there yet for "real" interactive framerates (Score:4, Informative)
The zlib/libpng License
Copyright (c)
This software is provided 'as-is', without any express or implied warranty. In no event will the authors be held liable for any damages arising from the use of this software.
Permission is granted to anyone to use this software for any purpose, including commercial applications, and to alter it and redistribute it freely, subject to the following restrictions:
1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.
2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.
3. This notice may not be removed or altered from any source distribution.
Re:Not there yet for "real" interactive framerates (Score:1)
How is this different from the preshaders introduced in DirectX 9.0c? Not tha
Re:Not there yet for "real" interactive framerates (Score:2, Informative)
In Sh, though, this optimization is internal and automatic, and so invisible to the user. It's important to do it that way, because the shade
Re:Not there yet for "real" interactive framerates (Score:1)
Re:Not there yet for "real" interactive framerates (Score:4, Informative)
I happen to know a bit about that specific procedural model, and on a modern CPU, I'd expect perhaps about 0.5s to render the same pixels.
So that example is already an order of magnitude faster than CPU evaluations. And GPU speeds scale faster than CPU so this gulf will widen, quickly, with time.
But the poster is correct, GPU metaprogramming is NOT going to be common in games quite yet. But it's going to happen, absolutely, and the time frame is measured in just a year or two, not some vague future date. sh is fascinating because it is a solid attempt to give developers tools to start using the GPU for different kinds of computations, not just polygons/pixels/shading.
And the important applications are likely NOT to be games. Think 3D movie renderering, think fluid simulators, think finite element analysis, think fracture simulation.. all of those massively parallel, highly numeric applications. Graphics is just one of them.
Re:Not there yet for "real" interactive framerates (Score:2)
I agree with your time frame of a couple years. The stuff in the games industry has been moving crazy fast. Every time we hear about something at SIGGRAPH that can't be done "yet" in realtime, it seems to be about 2-3 years before the hardware is catching up and someone figures out a clever technique to do it.
Re:Not there yet for "real" interactive framerates (Score:2)
BTW, that's a minor understatement considering it's named after you - LOL
Re:Not there yet for "real" interactive framerates (Score:1)
Re:Not there yet for "real" interactive framerates (Score:1)
- - -
Phisics affected by framerate bugs:
[netquake] you can jump higher and cross water faster with faster framerate
[quakeworld] relation with mouse lag, or something similar, etc..
Re:I saw this a long time ago (Score:2, Informative)
However, people DO like to think of shaders as "data" to be managed like "assets", just like textures. Also, there are contexts like web browsers where downloading and running an ar
Re:I saw this a long time ago (Score:1)
sh is awesome! (Score:3, Funny)
Great! (Score:2)
too bad this joke doesn't even work because the blurb was basically the whole book article anywyas
/bin/sh (Score:3, Funny)
Look out for the second book, "Metaprogramming GPUs with bash", and a third and final "Metaprogramming GPUs with X11", comming soon to all fine book retailers.
haha reqs (Score:1)
in other words, you must be a slashdot regular
Re:haha reqs (Score:2)
Maybe 5 years ago, not now.
Re:haha reqs (Score:2, Insightful)
Re:haha reqs (Score:1)
Mr. McCool and Mr. Du Toit (Score:4, Funny)
Re:Mr. McCool and Mr. Du Toit (Score:1)
UNFAIR!
Du Toit's first name is Steph-ANUS!
I've seen plenty of Stephans and Stephanies, but never a Steph-ANUS!
I've heard of Staph-Anus, you get that from yo momma.
triangle drawing example in sh (Score:3, Funny)
echo #
echo ###
echo #####
That was a preemptive Sh (Score:3, Funny)
Wow! (Score:2)
"Metaprogramming"? (Score:2)
Is this "metaprogramming"? Googling turns up a variety of definitions of the term, but the term sounds like it should mean writing a program which generates a program. Which is different from a library and syntactic sugar.
Re:"Metaprogramming"? (Score:5, Informative)
From the review, it sounds like SH is basically a library, and that library invocations are dressed up through the use of operator overloading. Is this "metaprogramming"?
Yes, because rather than just executing the the operations you specify, Sh has a special "retained mode", which instead collects these operations, builds an intermediate representation of them, runs an optimizer and passes them on to a backend compiler. For example:
So, the bits that are written inside the declaration of "prg" (the above is valid C++/Sh), are actually collected into prg, instead of being executed. prg can then be compiled to a GPU (or CPU) backend and sent to the GPU to run. Thus, your C++ program is used to write Sh programs at runtime. This allows all sorts of metaprogramming techniques.
We also provide further levels of metaprogramming with the shader algebra operations, but that's at an even higher level. These let you take previously written Sh programs and combine them in various ways, at run time. See our SIGGRAPH paper for details.
The About Page [libsh.org] on our homepage [libsh.org] tries to explain this in more detail.
Re:"Metaprogramming"? (Score:3, Informative)
Another thing to emphasize: yes, Sh does compile to the *CPU* as well as the GPU. So you can use it fo
Re:"Metaprogramming"? (Score:1)
Now the program object p uses d as a constant over each data-parallel invocation of p to data. To support GPGPU, Sh supports a stream data abstraction so i
Re:"Metaprogramming"? (Score:2, Informative)
SMP GPU's (Score:1)
Is anyone working on a method of using the GPU on some of the faster video cards to offload processing power to from the cpu?
Is it even feasible?
Don't Believe Everything You Read (Score:1)
Unfortunately, the developers of Sh chose to publish the book which details features which are not yet supported in the library. Since the software is currently in the alpha stage of development and constantly changing, many of the examples described in their book do not work as expected (the minimal GLUT example doesn't even compile).
In short, a lot of work remains before they meet their full project goals. The name, however, which is an abbreviation of Serious Hack [serioushack.com], seems quite appropriate.
Re:Don't Believe Everything You Read (Score:5, Informative)
Unfortunately, the developers of Sh chose to publish the book which details features which are not yet supported in the library.
Yes, this is true. If we had written the book based entirely on what we had implemented at the time, it would however have become out of date completely very quickly. Instead we chose to write the book as a specification. Too many programming language books, and books describing systems become out of date as soon as they are published, so it's a tough decision to make.
I should point out that we are getting closer to full support of everything in the book every day. Most of the missing features are very simple things like missing library functions. Sh works, right now, and can be used to write real shaders.
With the next release I will post a fixed version of the glut example. As unfortunately happens so often, the example got broken during book writing stage...
You're welcome to browse our Issue Tracker [libsh.org]. Progress has been slow of late because I've been working hard on the optimizer, and we've been busy with non-Sh-related things. However, note that most of the issues in the track are either "easy" or "bitesized", and once development gets on track again (in a week or so) I expect most of them to be resolved quite quickly.
Re:Don't Believe Everything You Read (Score:2, Informative)
We *were* p
And math, too! (Score:2, Interesting)
This idea is not mine, but back when the first Netra came out, a couple of guys here in Houston who used to optimize Crays figured out how to use the GPU to do floating point arithmetic a whole magnitude faster than Sun knew it could be done.
My optimization skills are a little rusty, but the same architectural elements are there in the ATI graphics cards.
If some one does this, drop me a line and tell me how you did it.
Re:And math, too! (Score:1)
Floating point operations aren't much use unless you have some idea how accurate they are.
Re:And math, too! (Score:1)
But anyways, it's true that the FP on GPUs takes a bunch of shortcuts, doesn't handle various things the way it's supposed to according to the IEEE spec, drops low-order bits for some cards, etc. etc. The work at Stanford is certainly valuable in characterizing these issues. However, I suspect the next gene
Hello world (Score:1)
Mods on crack. (Score:2)