Chronology | Current Month | Current Thread | Current Date |
[Year List] [Month List (current year)] | [Date Index] [Thread Index] | [Thread Prev] [Thread Next] | [Date Prev] [Date Next] |
When activating a RL circuit with battery and switch loop
theorem gives:
emf - iR - L(di/dt) = 0 form this we get i = (emf/R)(1-e^t/timeconstant)
when allowing to decay, I want to write the loop theorem as:
-iR + L(di/dt) = 0 as there is a potential drop across the
resistor and gain in the inductor. This makes it hard to get a
negative exponent in the current equation. The text says we
can use the original equation and set emf to zero. This works
algebraically but does not set well with my concept of energy
conservation. I know that di/dt is negative which will make
the term positive but .....I also know that the self induced
emf is -Ldi/dt which could make a problem in the first
equation.
I need help reconciling the algebraic signs here with my concept of
incresing and decreasing potentials as I traverse a loop.