redsnake9111's forum posts
Gaming should be fun, not stressful or rushed. Some people enjoy taking their time on a game, some people enjoy burning a swath right through the campaign. Whatever is fun is what you should do. Plus those games unforrtunatly aren't worth that much now anyways.
On a second note I like to use cheats after I finish the game and just want to screw around. They aren't bad but if you use them before you beat the game you rely on them and then it makes that game too easy and boring.
I voted "fail" for now just based on the benchmarks.
But remember that you really can't judge until the drivers get more mature. It is a new architecture, after all. the 9s and 8s all have more mature driver software.
Firefox 3 is more secure than the previous firefox 2 and WAY more secure than the current version of internet explorer. You can check out this link (scroll down to vulnerabilities at the bottom of the page):
http://en.wikipedia.org/wiki/Comparison_of_web_browsers
this site will let you compare cards and answer your questions...it will also bring you milk and cookies before bed....maybe
gpureview.com
ct1615
LOL!:lol:
Thanks, but it still doesn't answer my question fully. Especially what is the difference between processor cores, processor clock, and core clock. My own inference is that the processor cores number is the number of shader cores. But is the processor clock the clock speed of each of these shader cores, or is that the core clock? Which is the clock speed for each of the shader cores and what does the other clock speed signify? I was wondering if the actual graphics processing unit is just one processor (at one of those clock speeds) that distributeds work to all the shader cores, or if the gpu is only those shader cores and they figure out what they need to do by themselves. Is there some processor that controls the whole card and what the shader cores do?
Any other responses would be greatly appreciated. Anything that anyone has to contribute is welcomed!
The best you'll get is Medium and average 30FPS, the reason for that is the crappy processor. You need atleast Dual Core processor at 3.0 GHz or Higher. I myself have the following specs:
Intel Dual Core Processor 3.0GHz
4 GB of Ram running at 800 MHz
Micro Atx Intel Motherboard (I don't know the exacts on it)
550 Watt Sli Powersupply
8800GTS SSC Edition Overclocked
300GB Hard Drive
Apevia medium black tower case with Blue LEDs
________________________________________________________
I run Crysis at full detail on 1920 x 1200 and getting average of 37 FPS, worst I've gotten so far is 16FPS when their is simultaneous action in Crysis and to much of it, I suspect it is due to my processor. I'm hoping to upgrade to AMD quad core soon!
I hope my post helped you with what you need!
P.S. I would upgrade the processor and keep the card and ram and you'll be fine. If you can't spend that much get atleast 2.8GHz at the most, or switch over to cheaper AMD and get a dual core from them, but you'll need to change your motherboard for that.
Have fun playing Crysis once you get the upgrades!!!
playabythebook1
:shock: Are you sure that you are processor limited? Most rigs (and I am guessing yours) are GPU limited by Crysis since (at least my understanding) it does most fo the visual effects computations on the GPU. Physics is obviously a different story however.
Word out of the homesite says that those with nVidia 8's and up will finally be able to help the cause like never before. Work on a CUDA-compatible client to complement the client for ATI GPUs is being tested and will likely be released in a few weeks. I currently fold anonymously using the SMP client (have a C2Q 9450) but am really looking forward to turning my 9800GTX loose on this. I'll probably sign up for the GS team once the CUDA client is released.HuusAsking
I think the beta client already has been. Check that out. Have a link:
http://folding.stanford.edu/English/FAQ-NVIDIA
Log in to comment