We are independent & ad-supported. We may earn a commission for purchases made through our links.

Advertiser Disclosure

Our website is an independent, advertising-supported platform. We provide our content free of charge to our readers, and to keep it that way, we rely on revenue generated through advertisements and affiliate partnerships. This means that when you click on certain links on our site and make a purchase, we may earn a commission. Learn more.

How We Make Money

We sustain our operations through affiliate commissions and advertising. If you click on an affiliate link and make a purchase, we may receive a commission from the merchant at no additional cost to you. We also display advertisements on our website, which help generate revenue to support our work and keep our content free for readers. Our editorial team operates independently from our advertising and affiliate partnerships to ensure that our content remains unbiased and focused on providing you with the best information and recommendations based on thorough research and honest evaluations. To remain transparent, we’ve provided a list of our current affiliate partners here.

How Much Electricity does a Computer Use?

Dana Hinders
By
Updated May 16, 2024
Our promise to you
EasyTechJunkie is dedicated to creating trustworthy, high-quality content that always prioritizes transparency, integrity, and inclusivity above all else. Our ensure that our content creation and review process includes rigorous fact-checking, evidence-based, and continual updates to ensure accuracy and reliability.

Our Promise to you

Founded in 2002, our company has been a trusted resource for readers seeking informative and engaging content. Our dedication to quality remains unwavering—and will never change. We follow a strict editorial policy, ensuring that our content is authored by highly qualified professionals and edited by subject matter experts. This guarantees that everything we publish is objective, accurate, and trustworthy.

Over the years, we've refined our approach to cover a wide range of topics, providing readers with reliable and practical advice to enhance their knowledge and skills. That's why millions of readers turn to us each year. Join us in celebrating the joy of learning, guided by standards you can trust.

Editorial Standards

At EasyTechJunkie, we are committed to creating content that you can trust. Our editorial process is designed to ensure that every piece of content we publish is accurate, reliable, and informative.

Our team of experienced writers and editors follows a strict set of guidelines to ensure the highest quality content. We conduct thorough research, fact-check all information, and rely on credible sources to back up our claims. Our content is reviewed by subject matter experts to ensure accuracy and clarity.

We believe in transparency and maintain editorial independence from our advertisers. Our team does not receive direct compensation from advertisers, allowing us to create unbiased content that prioritizes your interests.

Determining the amount of electricity a computer uses can be a somewhat complicated task, since it depends on what type of equipment a person has and what applications he is running. Typically, the amount of electricity a computer uses is between 65 watts and 250 watts. The monitor often needs between 35 watts and 80 watts as well. Most desktop computers have a label that lists how much power they need, but this is generally the theoretical maximum and not an average representation.

Desktop computers with faster processors use more electricity than computers with slower ones. LCD monitors only use about half of similarly sized CRT monitors, however. Accessories and peripherals, such as cable modems, routers, or webcams, contribute to a slight increase in how much power a computer uses as well.

Regardless of what type of computer a person owns, the type of work he does on that computer makes a difference in electrical consumption. Using a computer to edit digital pictures, design a website, or play a video game uses more power than reading email or completing simple word processing tasks. In addition, the amount of electricity a computer uses significantly increases when it is connected to the Internet.

One common misconception is that using a screensaver saves power, but this is not true. Its series of moving images is aimed at protecting the screen from having a static image ingrained into it.

People who are worried about high utility bills will find that a better option is to leave the computer in standby mode when it's not being used. In standby mode, a computer uses approximately 6 watts of electricity and the monitor’s electrical consumption drops to almost nothing. Of course, it's even cheaper to turn the computer completely off when it won't be used for several hours at a time.

Although many people prefer laptops because of their added convenience, it is interesting to note that a laptop computer can also result in a significant energy savings. Most laptops use between 15 watts and 45 watts. Switching to a laptop may be a smart decision for someone who is concerned about how much energy a computer uses.

EasyTechJunkie is dedicated to providing accurate and trustworthy information. We carefully select reputable sources and employ a rigorous fact-checking process to maintain the highest standards. To learn more about our commitment to accuracy, read our editorial process.
Dana Hinders
By Dana Hinders , Writer
With a B.A. in Journalism and Mass Communication from the University of Iowa, Dana Hinders brings a strong foundation to her work as a freelance writer. After discovering her passion for freelance writing following the birth of her son, Dana has been a vital part of the EasyTechJunkie team. She also showcases her versatility by creating sales copy and content for e-courses and blogs.

Discussion Comments

By anon991203 — On Jun 03, 2015

I'm doing an energy project for my science class, and I want to find out how much energy the project took. Thanks!

By anon330068 — On Apr 14, 2013

I am a software consultant. Are there any ways to know how much power is needed for intensive use of such a software?

By anon329755 — On Apr 11, 2013

I was thinking of using a solar panel for my computer since I live in the tropics and work almost 24/7 from the PC. Con you tell me how to choose the right solar panel my computer?

By anon272255 — On May 31, 2012

Just let the powerhouse gaming CPU stay on 24/7 and do not sleep or hibernate. It doesn't save much money and you be able to tell. Printers drain a lot of wattage -- not LED monitors!

By anon267027 — On May 08, 2012

Post 5 is pretty accurate.

I have a power monitor for the main electrical box by my Alienware and that uses 250 with no SLI idle, 300-350 with dual SLI idle and 350-450 SLI in a game depending on what it is doing.

By BeastKiller — On Jan 23, 2012

I have an i7 2600k processor with nvidia 2gb card and a 1500watt power supply and a 2.2kv UPS. I use the computer for about two to four hours on the internet for gaming. How much power will the system use for an hour?

By anon198415 — On Jul 19, 2011

a laptop depends on the graphics card and its processor. I guess that in these times the newest computers that cost around 1000 euros consume about 300 watts, at least!

By applenaina — On Jun 27, 2011

i want to know how much power does a cpu nd monitor consume in terms of amperage: 6A or 10A. please let me know soon.

By anon160812 — On Mar 17, 2011

I bought a 2500W(5000​W) 12V-240V power inverter so I can attach it to a 12volt 24aH battery (Not a car battery). Can anyone tell me if this will power my laptop power supply? It states 19v 3.42 amps. I just want to know if it will do the job to use a laptop from this inverter.

can anyone help answer this question as I am an electronics virgin!

By anon151920 — On Feb 11, 2011

@anon74056: You are just wrong. a quad SLI system can easily pull over 1500 watts. in fact, if you run quad SLI you will probably need two power supplies to feed it.

By anon133593 — On Dec 11, 2010

Seriously don't underestimate extremely high end gaming PCs. Look up "antec 1200 pulling 2000w" and look at the second result.

my rig: i7 930 @ 4.4ghz; 12Gb 2000mhz ddr3; 240Gb vertex 2; 2x 2tb hitatchi 7200rpm; Gigabyte X58A-UD9; Corsair AX1200w; 2 HD5870 2GBs; 3 Acer V233H monitors in eyefinity.

By anon74056 — On Mar 30, 2010

Even using brand new, highest-end power hogging components for a gaming rig (and let's include 3 or 4 way SLI all on bleeding edge cards, 2-3 large widescreen monitors, raid array(s) with all 10k rpm sata drives, etc), your system power draw will never sit at "up to and over 800 watts."

Yes, you almost certainly need a power supply rated that high (and for the setup in question, possibly higher depending on rails and amps), but even taking such a system and running the most intensive gaming benchmarks 24/7 would not pull anywhere near 800W.

Absolute peak draw on such a system, if you are using fairly inefficient hardware, can get up to right around 800W. Current games (yes, even including crysis), however, don't stress your computer to anywhere near peak load - virtually nothing will, except for benchmarking, stress testing, or folding - none of which are something a normal user is going to do.

A very high-end system will probably sit around 250W idle, 500W typical load. More typical high-end gaming systems today will be around 200W idle, and 400-450W load.

More mainstream (but still fairly expensive) gaming machines will be down to 150W idle, 300-350W load.

A 750W psu can easily handle a souped-up quad sli system, at least in terms of raw power - the issue these days is just making sure there is enough output on the +12v rails.

By anon42918 — On Aug 24, 2009

I borrowed a "Kill A Watt" energy meter from the library, and found that when my computer is in the sleep mode, it uses 105 watts, but when in used it uses 235 watts of power. That means that when my computer is in the "sleep" or hibernate mode, it is the same as leaving a 100 watt light bulb on.

My computer is an el-cheapo desk top model about three years old.

By Faliarin — On Oct 25, 2007

Beware of Gaming computers if you are saving money on electricity. A Gaming computer may use up to and over 800 Watts. An SLi or Crossfire computer with two video cards can suck up wattage.

More hard drives means more motors in your computer running and this will also eat up wattage. If you can consolidate your data from 4x40GB drives to one 200GB drive. However one drive also means that if that drive fails you may lose everything.

SLi is NVidia's Dual Video card system when you get better game proformance. Crossfire is ATI's(AMD's) dual card system. Popular brands like Dell and Gateway will have gaming machines with these dual card setups.

Check your laptop's power cable to see what wattage it is using. For desktops it is harder to find as you may need to open your case to see how many watts your desktop can handle/use. Just because it says the watts there doesn't mean your computer is using all of it.

My laptop can use up to 180watts, but I had to buy something I could game on. My Desktop is rated up to 600 watts, but I am not using all of it.

Dana Hinders

Dana Hinders

Writer

With a B.A. in Journalism and Mass Communication from the University of Iowa, Dana Hinders brings a strong foundation to...
Learn more
EasyTechJunkie, in your inbox

Our latest articles, guides, and more, delivered daily.

EasyTechJunkie, in your inbox

Our latest articles, guides, and more, delivered daily.