Could some electronics guru here please clarify this a bit ?
I have been experimenting SPICE RF amplifier designs in my spare time,and I am getting confused by the differences between textbook expressions for amplifier efficiency etc., and my experimental findings.
Consider the textbook definition for efficiency as: efficiency = Power delivered to load/DC power supplied
First, the power delivered to the load(e.g., standard
50 Ohm) can be measured from the RMS values of current through the load and the RMS value of voltage across load. On the other hand, the DC power value by definition is a steady fixed value.However SPICE experiments give different answers. Take a single ended class B RF amplifier with BFR92A and a parallel LC tuner for 500 MHz. Since the input, output impedances need to be matched, the S parameters for Ic=10mA, Vce=10V are selected. A value for the collector resistor is calculated.
If the collector current is measured fro a trial transient analysis run, the collector current does not remain at 10 mA, In fact the Ic starts at around 8.5 mA and decreases to steady state value of 1 mA. So the DC power supplied by the DC source decreases from the design value(10 mA) to 0.1x design value, and consequently only a RMS value for the supplied DC power can be used in the calculations for rfficierncy. My observations are based on real SPICE simulations, and differs from the standard definition of DC power.
Are there any flawws in this reasoning. All hints, suggestions are welcome. Thanks in advance.