answersLogoWhite

0

Ohm's Law states Volts = Amps x Resistance.

You would need to apply 600 volts across 3 ohm load to have 200 Amps flow in circuit.

Not sure what you are really asking and why you mentioned 2 gauge.

User Avatar

Wiki User

14y ago

Still curious? Ask our experts.

Chat with our AI personalities

DevinDevin
I've poured enough drinks to know that people don't always want advice—they just want to talk.
Chat with Devin
MaxineMaxine
I respect you enough to keep it real.
Chat with Maxine
BeauBeau
You're doing better than you think!
Chat with Beau
More answers

To calculate the voltage drop in a circuit, you can use Ohm's Law (V = I * R). In this case, V = 200 amps * 3 ohms = 600 volts. This means that with a 200 amp current flowing through a 3 ohm 2-gauge wire, there will be a voltage drop of 600 volts.

User Avatar

AnswerBot

9mo ago
User Avatar

Add your answer:

Earn +20 pts
Q: How many volts is 2 gauge 200 amp 3 ohms?
Write your answer...
Submit
Still have questions?
magnify glass
imp
Continue Learning about Natural Sciences

How many amps does a 200 watt light bulb draw at 120 volts?

I'm assuming this is standard residential single phase. Simple calculation as noted below: Watts / Volts = Amps So: 200 Watts / 120 Volts = 1.666~ Amps If you needed to calculate for a 220 volt run with the same 200 Watts 200 Watts / 220 Volts = 0.909~ Amps Remember 80% load per circuit breaker so a 15 amp breaker you should only load to 12 amps or less. Using Watt / Volts = Amps is the same as Amps x Volts = Watts. 15 Amps X .8 (80%) = 12 Amps max per circuit (for a 15 amp breaker/fuse) So 12 Amps x 120 Volts = 1440 Watts max for a 15 amp circuit (typical 14 gauge wire) *************************************************************** If a 20 amp circuit and 12 gauge wire (smaller gauge = larger dia wire). 20 Amps X .8 (80%) = 16 Amps max per circuit So 16 Amps x 120 Volts = 1920 Watts max for a 20 amp circuit (typical 12 gauge wire)


How many volts it takes to cause a current of 1.95 amps to exist in a 200 watt lamp?

The formula to calculate the relationship between amps, volts and watts is Volts X Amps = Watts or Volts = Watts / Amps or Amps = Watts / Volts therefore; 200 Watts divided by 1.95 Amps is 102.5641 Volts.


200 watts to amps 12v?

To convert watts to amps, you can use the formula: Amps = Watts / Volts. In this case, to convert 200 watts at 12 volts to amps, it would be: 200 watts / 12 volts = 16.67 amps. So, 200 watts at 12 volts is approximately 16.67 amps.


Can I use 240 volt heater elements in a 120 volt hot water heater?

No. For an example using a 3500 watt element. The amperage through the element will be I = W/E, 3500/120 = 29 amps. The resistance of this 3500 watt element will be R = E/I, 120/29 = 4.13 ohms. Now using the 200 volts and finding the amperage I = E/R, 200/4.13 = 48 amps. The new wattage of the element has become W = A x V, 48 x 200 = 9600 watts. The element will not stand the increase in amperage and will burn open. To operate on 200 volts and still have a 3500 element in the tank you will have to find an element with a resistance of 12 ohms. I = W/E, 3500/200 = 17 amps. R = E/I = 200/17 = 12 ohms. This is a resistance three times more than the 120 volt element to achieve the 3500 watt rating.


2400w equals how many amps in 12v?

To calculate the amperage, you can use the formula: Amps = Watts / Volts. In this case, 2400 watts divided by 12 volts equals 200 amps. So, 2400 watts at 12 volts would draw 200 amps of current.