redsnake9111's forum posts

Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#1 redsnake9111
Member since 2006 • 615 Posts
Hey PS3 has had some big events this year. Blu-Ray won. Resistance 2, MGS4, Bioshock ported, Home, Trophies, and hardware sales rapidly increasing. The fat lady hasen't sung on these format wars. What determines the winner anyhow? Profits? Number of games that are exclusive? There isn't a winner. Just a bunch of competators, and that is what keeps the market healthy.
Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#2 redsnake9111
Member since 2006 • 615 Posts

Gaming should be fun, not stressful or rushed. Some people enjoy taking their time on a game, some people enjoy burning a swath right through the campaign. Whatever is fun is what you should do. Plus those games unforrtunatly aren't worth that much now anyways.

On a second note I like to use cheats after I finish the game and just want to screw around. They aren't bad but if you use them before you beat the game you rely on them and then it makes that game too easy and boring.

Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#3 redsnake9111
Member since 2006 • 615 Posts
Yeah well did you guys know that Activision/Blizzard is making 60% gross margins on all you suckers that are playing WoW? I am not denying that it is a good game (from what I have heard anyway) but still, anyone who plays WoW is really getting ripped off.
Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#4 redsnake9111
Member since 2006 • 615 Posts
I am in the same exact situation with the same exact hardware with the same native resolution to deal with. I suggest just grin and bear it until something that is at least 50% or more better in performance comes out at a cheaper price. Or you could upgrade your mother board to support SLI, but that also costs money.
Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#5 redsnake9111
Member since 2006 • 615 Posts

Here you go chief.

luamhtrad

OUCH:o

Talk about pricey.

Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#6 redsnake9111
Member since 2006 • 615 Posts

I voted "fail" for now just based on the benchmarks.

But remember that you really can't judge until the drivers get more mature. It is a new architecture, after all. the 9s and 8s all have more mature driver software.

Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#7 redsnake9111
Member since 2006 • 615 Posts

Firefox 3 is more secure than the previous firefox 2 and WAY more secure than the current version of internet explorer. You can check out this link (scroll down to vulnerabilities at the bottom of the page):

http://en.wikipedia.org/wiki/Comparison_of_web_browsers

Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#8 redsnake9111
Member since 2006 • 615 Posts

this site will let you compare cards and answer your questions...it will also bring you milk and cookies before bed....maybe

gpureview.com

ct1615

LOL!:lol:

Thanks, but it still doesn't answer my question fully. Especially what is the difference between processor cores, processor clock, and core clock. My own inference is that the processor cores number is the number of shader cores. But is the processor clock the clock speed of each of these shader cores, or is that the core clock? Which is the clock speed for each of the shader cores and what does the other clock speed signify? I was wondering if the actual graphics processing unit is just one processor (at one of those clock speeds) that distributeds work to all the shader cores, or if the gpu is only those shader cores and they figure out what they need to do by themselves. Is there some processor that controls the whole card and what the shader cores do?

Any other responses would be greatly appreciated. Anything that anyone has to contribute is welcomed!

Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#9 redsnake9111
Member since 2006 • 615 Posts

The best you'll get is Medium and average 30FPS, the reason for that is the crappy processor. You need atleast Dual Core processor at 3.0 GHz or Higher. I myself have the following specs:

Intel Dual Core Processor 3.0GHz

4 GB of Ram running at 800 MHz

Micro Atx Intel Motherboard (I don't know the exacts on it)

550 Watt Sli Powersupply

8800GTS SSC Edition Overclocked

300GB Hard Drive

Apevia medium black tower case with Blue LEDs

________________________________________________________

I run Crysis at full detail on 1920 x 1200 and getting average of 37 FPS, worst I've gotten so far is 16FPS when their is simultaneous action in Crysis and to much of it, I suspect it is due to my processor. I'm hoping to upgrade to AMD quad core soon!

I hope my post helped you with what you need!

P.S. I would upgrade the processor and keep the card and ram and you'll be fine. If you can't spend that much get atleast 2.8GHz at the most, or switch over to cheaper AMD and get a dual core from them, but you'll need to change your motherboard for that.

Have fun playing Crysis once you get the upgrades!!!

playabythebook1

:shock: Are you sure that you are processor limited? Most rigs (and I am guessing yours) are GPU limited by Crysis since (at least my understanding) it does most fo the visual effects computations on the GPU. Physics is obviously a different story however.

Avatar image for redsnake9111
redsnake9111

615

Forum Posts

0

Wiki Points

1

Followers

Reviews: 0

User Lists: 0

#10 redsnake9111
Member since 2006 • 615 Posts

Word out of the homesite says that those with nVidia 8's and up will finally be able to help the cause like never before. Work on a CUDA-compatible client to complement the client for ATI GPUs is being tested and will likely be released in a few weeks. I currently fold anonymously using the SMP client (have a C2Q 9450) but am really looking forward to turning my 9800GTX loose on this. I'll probably sign up for the GS team once the CUDA client is released.HuusAsking

I think the beta client already has been. Check that out. Have a link:

http://folding.stanford.edu/English/FAQ-NVIDIA