Please create an account to participate in the Slashdot moderation system

 



Forgot your password?
typodupeerror
×
Graphics Software Hardware

Renderfarm Setup Tips? 253

CarlosOlivaG4 asks: "We're in the process of acquiring and setting up a renderfarm, and I'm hoping the Slashdot community might light us up a little here. We'll use 6 to 8 nodes first, but would like to be able to expand it in the future." There was an earlier version of this question, but it dealt more with the hardware of the farm's nodes, rather than the network and software infrastructure on which these nodes would be based.

"In the hardware side, we still haven't made a choice between using AMD's Opteron or Apple's Xserve G5 (they have some very nice and price convenient cluster nodes which seem to be ideal for this kind of job), with Linux. As for the networking between them, is Gigaethernet enough or should we be going for Fiber? The software used to manage the render queues is another important point as well: I've been looking into Rush, and even though it's a commercial package, it works on all of the platforms we currently use (W2k/XP, Irix, OS X and Linux). But then there is also Dr. Queue, which is open source and is supported on at least the *NIX members of the aforementioned OS's. Other options include RenderPal and Pixar's RenderMan, but I would prefer an F/OSS alternative. Finally, it's worth noting that we'll be using the renderfarm for Maya and Adobe AfterEffects."

This discussion has been archived. No new comments can be posted.

Renderfarm Setup Tips?

Comments Filter:
  • Cinelerra (Score:5, Informative)

    by selfabuse ( 681350 ) on Tuesday June 15, 2004 @01:16PM (#9431750)
    You might want to check out Cinelerra. It has pretty good support for renderfarms. I built one out of scrap 300mhz machines, and it only took a weekend.
    • Re:Cinelerra (Score:3, Informative)

      by selfabuse ( 681350 )
      And here [heroinewarrior.com] is the link I forgot.
    • Re:Cinelerra (Score:2, Insightful)

      by dhartman ( 635124 )
      I've done a bit of work with Cinelerra and have been very pleased with the results. With three Athlon XP 1900 ish machines with no more than 512MB ram I was able to render at 30fps with full DVD resolution (720x480) output.

      The render nodes only run cli tools and do not require local storage. A bootable Knoppix cd could be made to create a temporary render node. Imagine using 10 computers in your office to render video in the off hours and reboot into Windoze in the morning with the user not knowing.

  • Simple (Score:2, Funny)

    by hanssprudel ( 323035 )

    (a) Wait for the next new processor technology to hit Slashdot.

    (b) Build a Beowulf cluster of those.
  • by hubertf ( 124995 ) on Tuesday June 15, 2004 @01:22PM (#9431824) Homepage Journal
    Check out g4u [feyrer.de] for deploying your render machines - it's a image based disk cloning tool that uses DHCP and FTP and which doesn't care what you run on your clients. (g4u itself is based on NetBSD, but that doesn't matter for the application).

    I've used g4u to setup a ~50 node video rendering cluster, see my webpage on the Regensburg Marathon Cluster [feyrer.de].

    Enjoy!

    - Hubert
  • by Jonas the Bold ( 701271 ) on Tuesday June 15, 2004 @01:22PM (#9431826)
    Don't use macs. First off, you have less choice in renderers, and second, the hardware is more expensive. Rendering is grunt work. Buy cheap systems that you can upgrade more often, and run linux or something.

    Macs are very nice hardware, but you really don't need that for rendering. For workstations they make sense, but for rendering you really want to have a lot of fast computers rather than nice computers.
    • Hehe, but to me, fast IS nice. =)

      I'm not a Machead, I'm an x86head. Always will be.
    • Um, why? Others have used Macs and this is a ideal thing for a Xserve and it obvously was not a problem for the person in this question. Both platforms he has presented are both new and are both goign to run him a lot of money. Plus Apple has a Cluster Node config that will work well for him (only one hard disk....in this case you'd use some other storage or a Xserve Raid. In fact, I bet Pixar may use G5's soon if they don't have them soon. Xserves are fast and if you make yuor cluster out of them, you


    • Yeah because when you're building something like a renderfarm, you wanna bottom-dollar it.
      • Well, I know you're trying to be a smartass, but yeah, you do.

        Rendering is rediculously parallel, so if one computer messes up, goes down, explodes, whatever, whatever it was working on can be picked up by another computer. Chances are you won't lose more than one frame on one pass of your render. Some other node can redo whatever was lost, it's not a big deal at all.

        So having lots of unreliable but fast computers is more effective than having fewer reliable high quality ones. Two cheap PCs can do more wo
    • But if you are using Shake the OS X version is $2999.00 or $2000 dollars cheaper than the Linux version. The OS X version also comes with unlimited render only nodes for free. Each Linux render node costs $1499.

      So for say 10 computers:
      Cluster node version of Xserver 10 @ $2,999.00 = $29,990*
      Shake 1 @ $2,999 = $2,999
      Total = $ 32 989

      Now the Linux version will cost:
      Shake 1 @ $4,999 = $4,999
      Render nodes 9 @ = $13 491
      Total costs software = $18 490

      This leaves you with $14 499 to buy 10 x86 boxes or

    • Depends on which renderer you're using. Except for LightWave, RenderMan, Mental Ray, Animation Master, Cinema 4D, and even Blender, to name a few, it's true, you'll have problems getting support for OSX...oh, wait, that's quite a few right there, isn't it. Nevermind.

      The question with building a headless Intel render farm is going to be licensing. Great, you can build cheap machines, but are you going to have to buy extra licenses just because you chose to go cross-platform? It's not just the main rende
  • Why run Linux? (Score:3, Insightful)

    by puppetluva ( 46903 ) on Tuesday June 15, 2004 @01:23PM (#9431830)
    Why would you run Linux on the Apple boxes? Wouldn't OSX be just as good?
    • Re:Why run Linux? (Score:4, Informative)

      by Anonymous Coward on Tuesday June 15, 2004 @01:32PM (#9431958)
      Why would you run Linux on the Apple boxes?

      Because of its unnecessary flashiness? OS X is notoriously bloated. For the command line junkies among us, Linux fits the bill.

      • Re:Why run Linux? (Score:2, Informative)

        by GFLPraxis ( 745118 )
        Actually, I find Mac OS X to run faster on my 1 GHz G4 than Mandrake Linux 10.0 on my 2.6 ghz Pentium 4. That flashiness doesn't slow it down that much, you see...the Mac makes use of the 3d graphics card (which usually does almost nothing when you're not playing a 3d game) to control all the window effects. As a result, the processor is hardly taxed.
      • Darwin (Score:4, Insightful)

        by Dav3K ( 618318 ) on Tuesday June 15, 2004 @02:19PM (#9432553)
        But more to the parent's point, OSX could be run Darwin-only, AKA BSD-style command-line interface. The bloat can be stripped out of OSX just as easily as it can for linux. Additionally, the user would have the comfort in knowing that his render farm is using the same OS as the workstations used to control it.

        As other respondants have suggested, I guess it would come down to which OS supported the entire collection of desired applications for the job.
      • Hello, anonymous! Where the hell do you get 'bloated' from?
      • I am sorry....OS X has less bloat WITH it's interface then other Operating Systems. That said, it's easy to NOT run the UI on these if you are really going for every CPU cycle you can but consider this...OS X can still run acceptably on many OLDER Macs. The 500 MHz Sawtooth G4 is still in use at work running Pather (and Jaguar before that). You CAN deactivate some of the glitz, but I have never seen anyone who has done that. These also ran Rage 128's. Very old machines running a very good OS. That sai
    • Apple has not yet released a true 64-bit version of OS X, while Gentoo released a PPC64 version a few weeks ago. [gentoo.org] If you're going to buy 64 bits of CPU, you might as well get 64 bits of OS too.
  • Software? (Score:5, Insightful)

    by m0rph3us0 ( 549631 ) on Tuesday June 15, 2004 @01:24PM (#9431848)
    What render queue products are supported by your pieces of software? Why don't you try a few of them?
    I'm sure a demo can be arranged.
    I wouldn't go blindly marching in the direction of FOSS especially in something that is valuable enough to setup a renderfarm for.

    Most importantly, find out what the people who will be using the software like and dislike about each package. And what works for them. If it saves you $30 per hour times 5 people software and hardware cost become insignifigant after one work week.

    The biggest renderfarm in the world is useless if your people can't use it. Always remember that software is only good in its ability to meet the goals of the organization it supports.
  • Our experiences (Score:5, Informative)

    by Thagg ( 9904 ) <thadbeier@gmail.com> on Tuesday June 15, 2004 @01:25PM (#9431858) Journal
    We had a renderfarm for "The Chronicles of Riddick" of 40 boxes. Each box was a dual-proc Opteron.

    We evaluated several render-queue management systems, and decided on Rush. The most persuasive arguments for using Rush were the very good experience we have heard from other users, and the simplicity of extending it to manage a variety of different tasks. I have to add Hammerhead to the list of happy customers. It did everything we could have hoped for. In particular, it was able to handle the inevitable crashing of machines pretty well.

    While it's true that Rush is a proprietary, gotta-pay-for-it system; a robust render queue management system pays for itself very quickly in the ability to make your renferfarm productive. Perhaps a render queue manager is overkill when you have just 6 or 8 systems, but once you get up to 30 or 40 it is essential.

    Our experience is all under Linux, but if you're going to be running After Effects that means that you're not going to be running Linux -- so there's not too much more I can help you with there. We did find that the dual Opterons worked much more efficiently than dual Xeons in multiprocessor rendering -- don't know about the Xserves, though. We were running mostly Maya, RenderMan, Shake, and our own in-house tools on the farm.

    This farm is unfortunately powered down now that Riddick is done -- if you need some dual opterons, let me know at thad@hammerhead.com
    • but I do work in a 3D department at a television production studio. I can't say Rush is the best, but we use it and the animators are generally happy with it. We have about 10 render only boxes and another 9 or 10 workstations that Rush let's you "online" when you are not using them (lunch, nights, weekends, long meetings...).
    • Re:Our experiences (Score:4, Interesting)

      by NanoGator ( 522640 ) on Tuesday June 15, 2004 @02:35PM (#9432740) Homepage Journal
      Welp, your post sparked some debate here at the office. I'm at a small studio making a full length animated movie. One aspect of it we've been chewing on is what to get in terms of render farm down the road. I just had a few questions for you, if you don't mind:

      1.) The words 'dual' and 'Opteron' both surprised us. We were kind of under the impression that maybe single proc machines would be better for a render farm. We were really curious why dual was chosen over single. Did the extra cost end up being worth it?

      2.) You mentioned that Opteron was more efficient than Xeons. I just had to ask: Was the particular software you were using particularly tuned to Opteron (i.e. 64-bit?) or was the 32-bit side of it just pleasant to work with? Any more insight you can share with me about the use of Opteron would be most helpful.

      3.) Did you guys end up buying a bunch of machines from a place like IBM or something, or was it more like "we bought the components and assembled ourselves..?" If it's the former, how'd you like the service?

      4.) Any regrets or things you'd do differently next time around?

      5.) Why are you getting rid of the machines used for Riddick? Or did I read that wrong?

      Appreciated,

      NanoG
      • Re:Our experiences (Score:5, Informative)

        by Thagg ( 9904 ) <thadbeier@gmail.com> on Tuesday June 15, 2004 @03:06PM (#9433147) Journal
        Good questions

        1.) The words 'dual' and 'Opteron' both surprised us. We were kind of under the impression that maybe single proc machines would be better for a render farm. We were really curious why dual was chosen over single. Did the extra cost end up being worth it?

        The computers are relatively cheap -- it's render licenses (especially for RenderMan) that are expensive. With the newest version of RenderMan, Pixar has deigned to let us use the two processors of a dual-processor machine with a single license. This lowers the cost of rendering by about 60%, if the machine rendered twice as fast with dual processors. In fact, for RenderMan, the Opterons were indeed almost twice as fast, where the Xeons were only about 50% faster.

        Our other big rendering application was Shake, and it also allowed the use of two processors with one license.


        2.) You mentioned that Opteron was more efficient than Xeons. I just had to ask: Was the particular software you were using particularly tuned to Opteron (i.e. 64-bit?) or was the 32-bit side of it just pleasant to work with? Any more insight you can share with me about the use of Opteron would be most helpful.


        Yes and no. The Opterons are the first AMD machines to implement the SSE2 instructions, which are heavily used by RenderMan. Also, the HyperChannel communication between processors on the Opteron is light-years beyond the communication between Athlons and Xeons. On the other hand, there is absolutely no advantage in the 64-bitness of the Opterons -- we were running a 32-bit Linux (RedHat 9), and we weren't using more than 4 GB memory on any of the boxes.

        3.) Did you guys end up buying a bunch of machines from a place like IBM or something, or was it more like "we bought the components and assembled ourselves..?" If it's the former, how'd you like the service?

        We hired a beige-box manufacturer. We specced it out to various places, and PC Mall built them for the best price. If I had to do it over again, I'm not sure that I wouldn't go with IBM -- while they cost a lot more, I expect that they'd build more solid systems.

        4.) Any regrets or things you'd do differently next time around?

        We bought minitower machines instead of the more trendy, space- and power-efficient 1U or blade machines. We did that so that we could potentially use the new Gelato renderer from NVidia -- that software uses the current NVidia high-performance graphics cards as an external array processor, giving significantly better render performance.

        As we didn't end up using Gelato, that was perhaps a mistake. We ended up power and HVAC constrained in the end -- as happens with almost every renderfarm I've heard of.

        5.) Why are you getting rid of the machines used for Riddick? Or did I read that wrong?

        No, you read it right. Hammerhead is a small company, typically working on just one show at a time. We don't see a use for the machines for another nine months or so, as we begin development of the next project -- and it just isn't right to leave all that compute horespower idle.

        Thad Beier
        Hammerhead Productions
        • Thad,

          I just wanted to thank you for both your insight, and in taking the time to respond. You gave us some stuff to chew on here.

          Have a good day!

          NG
        • I've seen a lot of videos that HammerHead puts out, and (as most people are) I'm very impressed and usually completely overwhelmed. Are you able to tell us which productions were these machines involved in rendering?

          Also, in the interest of understanding how much it costs to set up a significant render farm, how much does this sort of thing cost? Is it all in the PCs, or would the backplane infrastructure cost surprise us a lot?

          Really I guess what I'm asking is if you could do a cost-per-node breakdown to
          • Re:Our experiences (Score:5, Interesting)

            by Thagg ( 9904 ) <thadbeier@gmail.com> on Tuesday June 15, 2004 @04:36PM (#9434366) Journal
            Are you able to tell us which productions were these machines involved in rendering?

            These particular machines were just used for The Chronicles of Riddick. Computer technology advances so fast, has lowered in price so quickly, and movie post-production schedules are so long (six to nine months, typically) that we typically don't use any particular machines for more than a couple of films.

            Also, in the interest of understanding how much it costs to set up a significant render farm, how much does this sort of thing cost? Is it all in the PCs, or would the backplane infrastructure cost surprise us a lot?

            In fact the dominant cost, at least for us, is not the render boxes themselves. The network is a significant expense, as is the data server system. An even larger expense, though, are the licenses for the rendering software. Top-of-the-line rendering systems like RenderMan (for 3D) and Shake (for 2D) cost thousands of dollars per node. And then there are significant infrastructure costs in just electrical wiring and cooling.

            At least in the 10-to-50 server range, I would say that the costs are pretty linear. As you get bigger than that, you can start to see some economies of scale.

            At some point, it becomes profitable to start developing in-house software tools instead of buying licenses. Digital Domain's Nuke system was originally developed as a renderer for Flame, for example, so that the expensive Flame machines could be used for the interactive work and the batch rendering could happen on commodity hardware. For Riddick, we developed our own smoke-rendering system rather than use RenderMan, to free up render licenses for other parts of the movie.

            I'm afraid that an explict cost-per-node breakdown would get into details that we keep confidential, but this should give you an overview of our situation.

            Thad Beier
            Hammerhead Productions

            p.s we don't do Videos, we make Films.
            • p.s we don't do Videos, we make Films.
              My apoogies. I interchange the two words entirely too frequently... old habits die pretty hard.
            • Re:Our experiences (Score:5, Interesting)

              by dcmeserve ( 615081 ) on Wednesday June 16, 2004 @04:01AM (#9439633) Homepage Journal
              An even larger expense, though, are the licenses for the rendering software.

              What did you think of the freeware options, e.g. Aqsis [aqsis.com]?

              • Re:Our experiences (Score:3, Interesting)

                by Thagg ( 9904 )
                Well, I should have mentioned that are largest line item, by far, is the animators that drive all of this software. At least on a project of the scale of Riddick, it was best to use software with which our animators would be most productive. For rendering, that means RenderMan. There are many people who have strong experience in writing shaders and doing lighting in RenderMan -- and RenderMan is practically bullet-proof after decades now of work.

                We did try a couple of other rendering alternatives. We h
      • Re:Our experiences (Score:3, Informative)

        by forkazoo ( 138186 )
        From previous posts, I seem to recall that you use Lightwave, right? Regarding 2), head over to blanos.com, and check out his benchmarks. I just looked at 2p P4 Xeons vs 2p Opterons. Looks like Opterons wipe the floor from the specific scenes I was looking at, but I didn't see any P4 Xeon 2p results for the radiosity test scene. I'd like to see that, as LW is supposed to use SSE pretty heavily in radiosity, so the clock speed of the P4 would be a potential bonus.

        The big difference between Pentia and Op
        • Re:Our experiences (Score:3, Informative)

          by NanoGator ( 522640 )
          "From previous posts, I seem to recall that you use Lightwave, right?"

          Yes, that is correct.

          " 2), head over to blanos.com, and check out his benchmarks."

          Good idea!

          "but I didn't see any P4 Xeon 2p results for the radiosity test scene"

          Heh I just did that this morning. I ran the skull radiosity test with 8 threads. It was on a Dual P4 Xeon 2.4ghz 533 bus and Hyperthreading enabled. 119s. (I'll try to remember to log that at Blanos, time permitting...) That was running LW8, not sure if that make a
          • Indeed, figuring out exactly what Lightwave wants is annoyingly not-straightforward. Has anybody ever written something like a PC-emulator where you can tweak various properties of the emulated CPU, and run batch jobs, and time them? Or some sort of debugger you can plug into any binary, and see how much data it is loading, what sort of cache usage is going on, etc? Obviously, the app would run much slower under an emulator. That's fine, it could still be really useful to have an easy way to see why you
    • Re:Our experiences (Score:2, Informative)

      by Anonymous Coward
      The Maya renderer is not SSE optimized, so the extra fpu in the Opteron and Athlons gives them an edge. The Opterons are about equal to P4s in SSE enabled apps, the Athlons lag behind. Mac hardware is overpriced, especially if you are going to buy more in the future. And find some benchmarks - I don't think many apps use Altivec, but many do use SSE/SSE2.

      www.zoorender.com has a limited/single benchmark for Maya and Mental Ray on a variety of hardware. Pay attention to their comments about how the May
    • We evaluated several render-queue management systems, and decided on Rush. The most persuasive arguments for using Rush were the very good experience we have heard from other users, and the simplicity of extending it to manage a variety of different tasks.

      I'd add that Greg Ercolano is a very smart guy with a lot of experience in render queue management - he wrote (in the '90s) the Race render queue management system used at Digital Domain. Productive use of computing resources was perhaps more important

    • A different view:

      I work at . We used to use Rush. I was a bit clunky, but it got the job done. Unfortunately, once the # of employees got up in the hundreds and the # of render slaves got >500, Rush falls over. It just doesn't scale and it's central database can't keep up with the dispatch rate and query rate from the clients. Eventually it can't keep the farm full.

      The end result was we wrote our own distributor. It's a pretty sophisticated package that can distribute pretty much any batch processing
  • by Selecter ( 677480 ) on Tuesday June 15, 2004 @01:26PM (#9431871)
    Personally I'd go with the G5 Xserve with a few diskless cluster nodes tossed in.

    Thats only if you desire maximum ease of use with minimum setup and running hassles. The same ease of use the regular G5's have is built into all their server stuff too. I'm sure the linux dudes will have something to say about that.....

    I would take a really hard look at the ready made bio-information cluster they have all setup, and just load yer software as needed and off you go. But that's me. Some people seem to like futzing with computers.....After 20+ years doing that at work, I just wanna do what I wanna do when I wanna do it. Apple makes that easy.

    I get paid to deal with headaches, I'm not gonna deal with them at home too.

  • Oops. (Score:5, Funny)

    by Mr. Bad Example ( 31092 ) on Tuesday June 15, 2004 @01:27PM (#9431883) Homepage
    My subconscious desire to get out of the computer field as a career must be surfacing--I read this as "reindeer farm". Then reality set back in and I almost made a lame "LapLANder" joke before tasering myself.
  • XGRID (Score:4, Informative)

    by artlu ( 265391 ) <artlu@3.14artlu.net minus pi> on Tuesday June 15, 2004 @01:28PM (#9431898) Homepage Journal
    Recently, I have been working a lot with Apple's xgrid [apple.com]. We have been linking about 4 G5s/G4s together and getting impressive results. I don't understand your hardware situation, but if you are a Mac guy try it out.

    GroupShares.com [groupshares.com]
    • Re:XGRID (Score:3, Informative)


      If you're considering Apple G5s, either in workstations or in Xserves, take a look at Apple's mailing list [apple.com] for help and resources. Folks there have been working about clustering and Xgrid on the Mac for a while now.
  • Do it properly (Score:2, Informative)

    by smharr4 ( 709389 )
    If you're going to do this, do it properly. Get systems with massive amounts of I/O that will cope with all the data you're trying to throw at them. For this kind of work, you need only buy from one vendor: SGI.

    Don't bother with Intel/Linux, with dodgy hardware and the frequently-changing Linux code. Pay the money, get decent hardware with a support contract and a steady, stable, tried and trusted OS.

    Apple *may* be an appropriate choice, now that Pixar have ported RenderMan to OS X, but I don't like the i
    • Re:Do it properly (Score:2, Informative)

      by Donny Smith ( 567043 )
      >Get systems with massive amounts of I/O that will cope with all the data you're trying to throw at them.

      Render nodes get input via simple render scripts, output frames get written to the file server one by one every X seconds as they get rendred. Textures are shared but it's never "massive" and never "thrown at them" (compute nodes).
      The I/O loading is concentrated on the file server.

      > Don't bother with Intel/Linux, with dodgy hardware and the frequently-changing Linux code.

      So HP, IBM and other Int
    • One would think if Linux was so dodgy, ILM wouldn't have started a migration from SGI boxes to Linux boxes. Here are a few articles detailing how they love their new Linux setup. Not that that they don't have a few complaints (NFS performance.)

      ILM Linux Switch [linuxjournal.com]

      ILM Computers [linuxjournal.com]

      More on ILM Linux switch [linuxjournal.com]
  • pvm is the way (Score:3, Informative)

    by Goeland86 ( 741690 ) <`goeland86' `at' `gmail.com'> on Tuesday June 15, 2004 @01:31PM (#9431946) Homepage
    I think that what you're looking for is a renderfarm for computer graphics rendering, right? in that case, you should be looking at PVM or OpenMOSIX or even MPI. In either case, since you're going to have more data crunching than actual data transfer, I think that even T100 would be enough. gigabit will be nice, but fiber is not worth it. Drqueue is nice... if you can get it to work, and I didn't. We used pvmpovray for many things, and I think that might be worth a look. pvmpovray exists for gentoo with an ebuild script, which would make the installation and configuration the minimum pain for it. But that option requires a conversion from maya to povray files. Also, I don't know what the pricing is going to be like, but if it were up to me, I'd take the Opterons, because I believe they are faster, although I'm not positive on that, and because I know they're well supported under linux, and again, I think that's a more personal choice to make, but the impression I got from AMD is that you always get the most for your buck. The Opterons also let you find replacement parts or upgrades a little easier than the G5 if you burn a CPU or motherboard. That's just my $.02 worth of advice.
    • This is what support contracts are for (AKA Apple Care). I am not sure if they have on site support yet, but you can bring yout Xserve into any Apple Store and they can either fix it or take care of it for you. With Apple, you have more support then buiding your own machines. If you really need 24/7 support for your renderfarm, then you can order some Xeon servers from IBM with appropriate support contract.
  • My $0.02.. (Score:3, Interesting)

    by midifarm ( 666278 ) on Tuesday June 15, 2004 @01:32PM (#9431948) Homepage
    I just want to let you know, I'm not a F500 company with jillions of dollars to spend, hence my two cents! I've been running After Effects and FCP using a host of various Macs for offline rendering. It's like a museum of Macs that are able to run OSX and I've been able to do it rather efficiently. I'd love to delve into Maya and perhaps Renderman eventually. I must tell ya that Pixar is completely converting to G5's and OSX for Renderman. I imagine they're getting the family discount, but it says a lot when the world's leading animation studio is going in that direction. I find that I have no problems with the computers. Granted I'd love a fridge rack of XServes for offline rendering, but for the time being my motley cluster is working just fine.

    Peace

    • Damn, Motley Cluster is a hell of a name for a band....

    • Re:My $0.02.. (Score:3, Interesting)

      by Thagg ( 9904 )
      I could be wrong, but I thought that Pixar was going to stick with Intel-type machines for the renderfarm, although they are apparently moving to OS X boxes for their animator workstations.

      I'm porting all of our animation code from Linux to OS X as well -- more or less as an exercise in code portability -- and it's going pretty well. OS X 10.3 is dramatically more standards- (or at least Linux-) compliant than the earlier OS X versions were. Almost every program I have compiles with virtually no changes.
    • I do my video editing work on a old G4 using FCP 3.0 (I know, I'm behind.) Occasionally I get stuck with 8 hour renders that generate 6 gig or better render files.

      Is there any cheap solution that will let me spread this work out? I don't use AE at all. My shop is incredibly cheap so it would have to be a free or nearly free solution. I know that's asking a lot. It just isn't worth it to get more macs and put FCP express on them. That's the only thing I can think of.

      Any tips would be greatly appreciated.
      • I don't know about the older versions of FCP, but I know you can spread out renders with the new version. I'm not sure if Xgrid is something that you could use to parse that amongst machines as well. You could inquire with the ACG at Apple. I know it was originally made for scientific crap, but it might coincide with FCP. Save your money on AE and get Motion when it comes out. Good luck. If you've got any more questions feel free to ask.

        Peace

  • Light us up (Score:5, Funny)

    by Reckless Visionary ( 323969 ) * on Tuesday June 15, 2004 @01:34PM (#9431979)
    I'm hoping the Slashdot community might light us up a little here.

    Light us up the bomb!

  • by RelliK ( 4466 ) on Tuesday June 15, 2004 @01:35PM (#9431981)
    ... but it was rejected. How do you deal with terabytes of data (50+ TB), all in a single directory tree, all must be accessible to every node? This is larger than what you can store on a single filer. Also, for performance reasons, the data must be separated across multiple filers. Currently we use lots of symlinks to tie it all together into a single logical directory tree, but that's a really ugly solution. There's got to be a better way. Right? Anyone?
    • I'm sure I'm missing some obvious part of your question, but... nullfs? unionfs?
    • Currently we use lots of symlinks to tie it all together into a single logical directory tree, but that's a really ugly solution.

      I don't think a bunch of symlinks is ugly at all. If it works well - who cares?

      Are you haveing trouble with any of your simlinked directory structures - in my little FreeBSD world, I've never noticed any problem at all except for a few utilities that are aware of simlinks and will allow you to chose to traverse them or not - like rsync.

    • by Anonymous Coward
      Take a look at Veritas Storage Foundation Cluster File System. You can have your 50+ Terabytes on a SAN fabric, and each server on the fabric can have the 50+ terabyte LUN mapped to them. The Cluster File System manages all of the concurrent access to the filesystem from each node so things don't get clobbered.

      You'll get your performance through the SAN by utilizing high performance FCAL disks and multiple HBAs to your servers. You can have the load balanced across the HBAs to give you the bandwidth tha
      • I agree that a SAN could be the way to go. Moving your storage to fibre instead of trying to throw it on the same network link as your render management and all make sense.

        All the SAN solutions out there are rather a bit wonky. Not saying they won't work, they're just all wonky. If Apple's xSan actually performs in the real world the way the claim (and the way it was working at NAB) then it could be the be-all-end-all. It'll be crossplatform and all, with file-level locking. Whether it really works or
    • Try a cluster file system [polyserve.com].

      "Filers" create "hot spots" whereby often-accessed directories/files create IO bottlenecks.

      I think you can use this CFS to create a directory tree with over 200TB of data (/home/lun1, /home/lun2, .../home/lun255). You can't "tie" them together like with LVM but you do get huge throughput as opposed to a single-host bottleneck with a volume manager and/or clustered NAS filers with the hot spot problem.

    • That's a difficult question to answer without knowing something of your setup. How are the spindles organized--SAN, individual file servers NFS cross-mounting, or what? Which OSes are you running? Also, how much money are you able to spend to resolve this problem?

      If you could rebuild everything from the ground up (and had tons of money to throw at it), you'd most likely want to build a system based on a very expensive vendor solution [ibm.com].

      Assuming that you can't do that, your best bet would be to go with

      • That's a difficult question to answer without knowing something of your setup. How are the spindles organized--SAN, individual file servers NFS cross-mounting, or what? Which OSes are you running? Also, how much money are you able to spend to resolve this problem?

        We have a bunch of netapp/bluearc filers and a single symlink tree that distributes data across them, so it's NFS. Render nodes run Linux. I have no experience with SAN. Can you point me to more info?

    • You call a company like EMC that specializes in storage area networks. Chip H.
    • "How do you deal with terabytes of data (50+ TB), all in a single directory tree, all must be accessible to every node?"

      Easy! Just use Gnome 2.6 - it has super-duper spatial browser behavior [osnews.com]. All your troubles will be solved.

  • by FueledByRamen ( 581784 ) <sabretooth@gmail.com> on Tuesday June 15, 2004 @01:37PM (#9432022)
    I don't know what the setup for After Effects and such looks like, but I managed to build a pretty good Sun Grid Engine system for distributing Maya batch renders. SGE is free, works on Linux or Solaris, x86 or SPARC (I obviously used the x86 linux binaries), and seems to be very well designed. I set up a pretty solid system for netbooting the clients, running them diskless (or with local swap drives), adding new clients on the fly (all scripted), and it all worked flawlessly. You could submit a batch job and it would distribute it per-frame as an array job to all the different nodes. Or you could just run SETI on all of them...

    It's since been taken down in favor of running Alfred (because I no longer use Maya's builtin renderer, we've moved on to MTOR and PRMan), but I still have all of the files and scripts for it. If anyone's interested, I'd be happy to share: sabretooth@gmail.com [mailto]
    • ok I only have experance with a custom T&L rendering platform but the grid engine worked very well I have used it at a couple of hardware design places and it's CPU blancing was very nice indeed
      it supports

      Apple Mac OS/X
      Compaq Tru64 Unix 5.0, 5.1
      Hewlett Packard HP-UX 11.x
      IBM AIX 4.3, 5.1
      Linux x86, kernel 2.4, glibc >= 2.2
      Linux AMD64 (Opteron), kernel 2.4, glibc >= 2.2
      Silicon Graphics IRIX 6.5
      Sun Microsystems Solaris (Sparc) 7 and higher 32-bit
      Sun Microsystems Solaris (Sparc) 7 and higher 64-bit
      Sun
    • We're replacing our a**-slow "Platform LSF" commercial cpu farm management software with GridWare / Sun Grid Engine. Its great! Grid Engine performs operations on jobs in an instant just as if you were running them locally (*unlike LSF which is too slow at common tasks like job submission or job killing to be useful*). its also free, with support from sun available if you feel like wasting money.
    • Yeah, I'm using SGE with a machine room full of solaris boxes to distribute interactive jobs, it's one of those pieces of software you look at afterwards and think, "Yeah, that's nice". Nice, easy to use and flexible system which does what it says on the tin, if not the cutting edge of distributed computing.

  • by Anonymous Coward on Tuesday June 15, 2004 @01:38PM (#9432030)
    I manage a large Linux cluster that is used for a wide variety of tasks, including running Renderman. Our nodes are dual Xeon 2.4 GHz with 2 GB RAM and they are more than sufficient for the design group's rendering needs (primarily Maya). Our scheduler (OpenPBS) lets them request up to 64 processors at a time, which, I'm told by their IT guy, works well with their Renderman maitre de.

    There is a fiber interface (Myrinet) to each node used by the MPI crowd, but our rendering group doesn't use it; they seem content with the performance of Ethernet over copper. Your needs may be different, of course, but latency isn't really an issue for rendering, and copper should provide all the bandwidth you need.

    I'm not knowledgable regarding all the software packages you list there, but I'm wondering if any of them would really take advantage of a 64-bit kernel (either on Opteron or G5-PPC970). Of course you can put a PPC version of Linux on the Xserve, but not without sacrificing nearly all Apple-provided management. If you expand the cluster to a large number of nodes, or even if you keep a small number of nodes but place it in a remote location, Xserve running Linux would be painful to manage (no remote power-off/-on, remote console problems). Xserve is shiny and has the requisite blue LED's, but and AMD or Intel box (from the right vendor) would be much easier to manage remotely.

  • For one, consider creating a standard image for all of your machines and have them install using pxeboot or etherboot. If any of the machines get hosed, you can reboot them and force them to reinstall their OS in 5 minutes.

    It seems to me that fiber is a waste of money. You implement gigE using copper. I would think that most of the data transfer is going to be the scene data and then an image transfered back.

    The company i used to work for was developing a global illumination raytracer and we create
  • PCI-Express (Score:3, Interesting)

    by Viking Coder ( 102287 ) on Tuesday June 15, 2004 @01:41PM (#9432061)
    My question is - when is someone going to make a blade with PCI-Express and enough room for the latest batch of cards from NVidia and ATI?

    Seriously - this is going to become a huge issue, as more rendering is pushed out to the stream processor that is the GPU.
  • Render managers (Score:3, Interesting)

    by tikoloshe ( 515755 ) on Tuesday June 15, 2004 @01:41PM (#9432065)
    We have a 50 node dual 3Ghz render farm, 20 on w2k and 30 on linux 2.4.20 (ok Redhat9) We are currently using Muster www.vvertex.com [vvertex.com], but are not that happy with it. It was reasonably priced, and the developers are very helpful, but its not quite there yet. I have been seriously considering building my own using a SQL database (currently Postgresql, but may swich to MySql) and perl. A render manager is really just a database with a bit of network sockets and scripts to run occasionally. A simple concept that is probably going to come back and bite me :)
  • dynebolic (Score:2, Informative)

    by Anonymous Coward
    There is a distro called dynebolic which is specifically for multimedia. It features Veejay, Cinelera, MJPEG tools etc. It is supposed to run off the CD. It is supposed to use every computer hooked to the network that is running dynebolic in the manner of a render farm. That is a very attractive idea. Go around the house, put in dynebolic CDs and go nuts.

    I can't say that it is great because I haven't been able to do much of the above. Maybe you will have better luck.
  • The cheapest and easiest way to set up a renderfarm is to use someone else's computers!

    We used to run Lightwave for work-related projects, but we only had a few machines allocated to the media lab. So after hours, we'd sneak around the cube farms and pop a client boot disk in every machine we could find.

    • Only acceptable in a Uni environment although I like the way you think! :) If I could do that with our system I work on. After hours processing could jump up! Here's an idea...with a all mac shop, could you have the main cluster (using Xgrid) add the developer's nodes when they go home? I imagine adding 10-15 nodes to the cluster at night only could help clean out your render queues! Also, a good scheduler can help too (cron is too weak....make it something else).
  • by BBStriker ( 264222 ) on Tuesday June 15, 2004 @01:56PM (#9432254)
    Hey

    I've set up and administrated a number of farms over the years (doing it as I type. its.. what I do). One thing you really want to do, certainly with Maya's renderer, is to try to use the same OS and platform on your farm as you use on your user workstations. There can be subtle or even obvious differences in the render output between OS's, and since you'll have enough issues to deal with you'll want to keep cross-platform incompatabilities out of the mix. Please, trust me on this. Had to deal with Maya Irix/Win2k/Linux differences in the past.

    As for queueing software, give Condor [wisc.edu] a look-see. Free and functional. I reverse-engineered a Perl version of it before they made their source available, and my version has been run quite successfully at several animation studios and an effects house over the years. It's a well architected system for distributed computing.

    Feel free to contact me if you've got any other render system or management questions. I'm always interested in seeing how other studios approach the challenge.
  • Everything you have mentioned is dependent on what render engine you are using.

    First, figure that out, then go to the forums for that engine and ask your question there.

    I just built an 8-node renderfarm for using Vray, which means 3D Studio, which means Win2k and it was a piece of cake.

    6-8 nodes is tiny. Figure your use, then ask on the right forum. Many, many people have done this already.
  • by beegle ( 9689 ) * on Tuesday June 15, 2004 @02:01PM (#9432318) Homepage
    This is a fun idea in the abstract, but if you're looking for concrete advice, you need to give us some concrete data. Most importantly, -what- are you rendering?

    If you're at a university and you're doing some sort of bioinformatics visualization, use whatever the researchers are most comfortable with. The odds are good that this is whatever the CS department was teaching on 5 years ago. Probably Suns or Windows machines. Slave... errr, grad student labor is cheap, so use an OSS scheduling and job management system if you can.

    At most other places, a similar rule applies: use whatever the users are most comfortable with. If you're using Mac workstations and software, then it may make sense to go with a G5 rendering farm. If you're using Windows... well, okay. Windows render farms still suck, but at least buy PCs to leave your options open. Unless you're a really large organization (that is, the sort that doesn't have to resort to Ask Slashdot for research), you probably want to use products that come with support contracts. $20k/year is a pretty good deal when compared to keeping a full-time support person for the same task.
  • by bhouston ( 788429 ) on Tuesday June 15, 2004 @02:02PM (#9432326)
    At Frantic Films [franticfilms.com] we have over the past year developed our own network rendering solution: Deadline [franticfilms.com]. Our solution has just recently entered a beta testing period thus if people are so inclined one can have a look at the current product (screenshots) [franticfilms.com] and possibly download a trial version (download page) [franticfilms.com]. We used Deadline on a number of recent feature films including Scooby Doo 2 [warnerbros.com] and Paycheck [paycheckmovie.com].

    We did this because we primarily use Discreet's 3dsmax [discreet.com] (with Brazil [splutterfish.com] and V-Ray [choasgroup.com]) and Eyeon's Digital Fusion [eyeon.com]. We have found that most existing render farm solutions do not support these two packages very well -- thus we decided to develop our own custom solution. We also support After Effects [adobe.com], Alias|Maya [alias.com], AIR [sitexgraphics.com] and other RenderMan [pixar.com] compliant rendering packages.

    Of interest to the general Slashdot crowd may be that this Deadline Render Management Solution is based on the open source (BSD License) Exocortex C# library originally released with this C# 3D Engine [exocortex.org]. Deadline is built with C# in the hopes that using Mono [go-mono.com] we will be able to start supporting Linux with minimal extra effort.

    I'll be reading all the posts on this Slashdot thread but I would also appreciate any direct feedback on our current beta product. We also found solutions such as Rush and Smedge to be less than user friendly in many respects. Thus we have tried as best as we could to increase a 3D package that is not well supported by most render farm management solutions -- except for Discreet's Backburner (which we found not that that scalable.)

    • We have almost requirements here at Blur Studio. I tried the Deadline beta, and it seems like a definate improvement over Backburner =) AS you mentioned, it's not very scalable. How many machines do you guys have on your farm? I wasn't very confident in Deadline being able to handle a large number of hosts, with it's XML polling architecture.

      We have our own in-house solution as well: Assburner. Assburner is GPLed and we hope to do a public release in the next month or two. Along with our production trackin
      • Cool stuff. We will be posting a new beta on the site later today -- with a couple enhancements. I just loaded up Deadline Monitor here and it shows that we have 98 machines current rendering, 3 idle, 41 offline (user workstations) and 5 machines disabled. Overnight during really heavy production loads (i.e. back in February with Scooby Doo 2) we would have all possibly machines rendering except for the 5 or so disabled ones (which are file servers or underpowered machines.) Thus from a pragmatic standp
  • First off, the G5 XServe's are very fast machines and the cluster node is, of course, designed for just this sort of thing in the minimal amount of space. You should be aware that they are LOUD as hell though... louder than the G4 XServes, which sound like a plane taking off. That said, plan on hiding them away from your workspace because even one of them in the same room with you will drive you nuts. This is no different than a 1U server from Dell/Whoever though.

    You'll only really need to buy a single n

  • Some Tips (Score:5, Insightful)

    by mnemonic_ ( 164550 ) <jamecNO@SPAMumich.edu> on Tuesday June 15, 2004 @02:28PM (#9432655) Homepage Journal
    These are just some tips I've heard in my 4 years of experience.

    Whichever processors you go with, make sure the entire farm uses the same type. Otherwise peculiar rendering differences might occur, in things like particles, hair/fur and fluids.

    I suggest going with the Opterons just for the PC compatibility. While the CG industry is becoming more diverse hardware-wise, it is still dominated by PC's and to a much lesser extent SGI boxes (5 years ago it was all SGI). Using PC's keeps your options open. Perhaps someday you will find 3ds max and its included distributed rendreing software more suitable for a task, and that can only be used with PC's. Same goes with the Mental Ray and Brazil renderers and the Combustion compositing software. Macs just have not been widely used in the 3d graphics industry, and so the vast majority of 3d content creation software is PC and SGI only (Maya Unlimited is only available on PC and SGI, while a lower end wersion is on Mac). And VirtualPC cannot be used to emulate 3d hardware acceleration (and it shouldn't be used for anything processor intensive anyways), though this only applies to the hardware rendered viewports in the apps. Having only Macs would be risky, and could limit your capabilities significantly.

    Pixar's PRMan (Photorealistic RenderMan) is a full blown renderer, not just something to help distribute render jobs. It is generally considered the best in the industry, though MentalRay and Brazil have gained significant followings. For a cheap but effective render queueing system, check out Smegde [uberware.net]. Smedge was used by Manex Visual Effects for handling some of the effects shots in the Matrix trilogy. If you're running the Linux version of Maya (x86 only) it is not too difficult to distribute the render tasks yourself using shell scripts and the command-line renderer.

    GB Ethernet should be fine, the bottleneck will be in the actual image processing not data transfer rates. 100Mb ethenet might even get the job done, thught I'd use GB for the added speed when copying large files. YMMV of course.

    Overall I'd try to create a very flexible system, one that will definitely support the newest CG software down the road and one that ensures compatibility with everything, for those always short deadlines. Goodl luck with your rendering.
  • Gigaethernet enough or should we be going for Fiber

    Well quit mixing your media here. What do you mean GbE is enough, it runs over both copper and fiber. Now there are other layer 2 protocols that run over both copper and fiber (all though different cables) for example - what SAN are you going to use ? iSCSI, FC, SRP, NFS ?

    What networking are you going to use ? GbE, TokenRing, FDDI, ???

    What HPC/Interconnect are you going to use Infiniband, Myrinet, VI, ???

    Some of these are the same networks, some of

    • Presumeably the poster was asking about fibre channel. For an 8-node cluster, FC can support pretty high bandwidth, but so can GbE using SMB or even iSCSI. Also, GbE switches are cheap compared to fibre SANs.

  • nVidia Gelato (Score:3, Informative)

    by Guspaz ( 556486 ) on Tuesday June 15, 2004 @02:56PM (#9433017)
    You might want to look into nVidia Gelato [nvidia.com]. It's a 3D renderer that uses Quadro FX cards as secondary FPUs, supposedly doubling or more the speed of rendering. They claim it's two to six times faster than the leading renderer. There's a demo, so you can verify those claims for your uses.

    It runs under Linux, and "will function with whatever [render farm] management system you currently use.".

    To reiterate, it's a SOFTWARE renderer, that is hardware accelerated by using the video card as a co-processor.
  • imp [imp.org] is a POV-RAY based distibuted renderfarm for Win or Lin.

  • For a "Render Farm" manager Here [highend3d.com].

    Just thought this was an appropriate discussion. Also, maybe a "slashjobs" would be an interesting addition to this site.
  • by gorodish ( 788476 ) on Tuesday June 15, 2004 @05:44PM (#9435257)
    Having built two generations of renderfarms, and now working on a third, I'd suggest building it as cheap as possible. You will want to upgrade every 2 years or so, so make sure that you won't feel bad disposing of the old farm when it's time for the new.

    Regarding networking: you have to look carefully at the way the farm will be used. If you are doing any kind of compositing (which requires high I/O rates), you'll benefit from gigabit ethernet. You'll also benefit from gigabit if you have exceptionally short render times (less than 30 minutes per frame), since in this case I/O is a significant fraction of each frame's render cycle. But the longer your per-frame render time, the less necessary gigabit is. We've always used 100base and it still serves us well. Fiber is expensive and provides nothing you'll need that copper can't provide.

    The individual machines should have identical configurations and be interchangable. Your goal is to not care when an individual machine dies. In light of this, there should be no local storage of data. You can save money on support if you buy spares instead of service contracts. Warranties also work, but the big manufacturers give their worst service to warranty-only customers.

    Don't wire anything but ethernet to the machines. KVM wiring is expensive and unnecessary. Each machine should run unattended until it dies; when it does, you can wheel over a monitor and keyboard to diagnose it.

    Opterons are fast, compatible, cool, lower-power and cheap. Xserves are nice, but we've found that Darwin doesn't integrate well into a pure Unix environment. You'd also be locking yourself into a single manufacturer.

    Linux is cheap and effective, and easier to configure correctly as a server OS than as a desktop OS. There is so much commercial software available for it now that there is little reason to consider Windows or a commercial Unix. We haven't found Linux support from the big manufacturers to be all that great; if you use Linux, assume that you will have to solve most problems on your own.

On the eighth day, God created FORTRAN.

Working...