Hello,
I'd like to understand the "guts" of the deflection circuit inside a fixed-frequency display. I know the most monitors today are either tri/quad sync, or truly multi-synchronous, but for learning purposes, I'd rather start with a fixed-frequency display. I'm just learning the basics of VGA timing (doing a video project on an FPGA), and I feel that these answers would give me a more rock-solid grasp on what I can and can't do with VGA.
At first, I thought that the horizontal deflection circuit would be even simpler than a normal sawtooth relaxation oscillator. My thought was that rather than using a "threshold" type device to trigger the draining of the capacitor in the oscillator, we could just use the sync pulse directly to turn a drain transistor on/off. But all the information I've read on the net seems to suggest that the sync pulses in some way add/subtract only slightly to the "normal" frequency. Can someone clarify what's actually going on here? Perhaps there isn't one correct answer (different companies, different method), but I'm guessing that there's a common method used by most.
As for the vertical deflection, it should look like a "staircase", right? Of course, there would be a sharp downward line after each "staircase" for the vertical retrace. How is this waveform generated? My initial guess, again, would be a simple integrator whose input was the horizontal sync pulses. It would reset itself on each vertical sync pulse.
However, this type of a setup would mean that the *length* of the horizontal sync pulses would directly affect how much the picture was vertically stretched. I'm not sure whether this is true or not. Question: So long as the horizontal sync pulses are at the correct frequency and are close to the correct length, does their *exact* length in fact have a great effect on the picture? I'm a bit more into the digital circuit world, so excuse the terminology, but another way of asking the question might be: Is there some part of the deflection circuit which is "edge-triggered" by the sync signal, or does the actual pulse length of the sync signal matter? (I know that's kind of a crude analogy since not all CRT's use digital circuitry - sorry)
If anyone can answer those questions, the next will be:
Tri/quad frequency monitors seem like a fairly straightforward extension of fixed-frequency monitors, but how does the deflection circuitry in *true* multi-synchronous CRT's work? I know a PLL is involved, but what is the supporting circuitry? I've only just started to become familiar with PLL's in general, so please excuse me if this is fairly obvious. We were using PLL's in the context of sinusoids, please let me know how they would be used to generate the sawtooth signal we desire.
Of course, the threshold voltage (after which we want each sawtooth pulse to end) would still be the same as with the fixed-frequency case. After all, this is directly related to the size of the CRT screen. So determining when to "reset" isn't hard. It's determining how fast the sawtooth should rise (ie, how fast the beam sweeps horizontally) that I don't understand (circuit-wise).
Thank you for your help !
Sean
PS - If anyone thinks that it would be better to a different group, let me know.