Great, I didn't know about the shared-memory option, I have to look into that. Even so, that cache is local to one machine. I think the reason why facebook, digg, wikipedia all use memcached is that it can be distributed over many machines. That gives it great scalability. From what I understand, it's basically a distributed hashmap that can be installed on all your webserver nodes or wherever else you have spare memory in effect giving you a giant virtual memory pool.
You're right, that is a good design choice. It makes for a more flexible solution.
That's right - IPB only supports one cachetype at once. I found out that today by rummaging through the registry source code. Too bad, I think the combo would be best. If I was planning on running this on a single webserver, I would go for eaccelerator. That would give the opcode caching performance boost along with the shared memory data caching. Since our client is willing to pay for the hardware, I think we'll go with memcached. It gives us more scaleability and reduces waste.