A 14-bit A to D conversion has to be accomplished for a specific design. In our
ID: 1927888 • Letter: A
Question
A 14-bit A to D conversion has to be accomplished for a specific design. In our design kit, we have two 8-bit microprocessors that implement Successive Approximation for A to D Conversions. In our case. we will have to connect the two processors in parallel to perform the 14-bit conversion. If the processors are driven by a 1 Mhz oscillator, how long will it take to do the conversion? Assume that the second processor uses it's 6 low order bits and Successive Approximation of the bits starts with the MSB. Also assume that the signals have been conditioned before they are inputted into the A to D microprocessors.Explanation / Answer
17.6 hours to perform the 14-bit conversion
Related Questions
Hire Me For All Your Tutoring Needs
Integrity-first tutoring: clear explanations, guidance, and feedback.
Drop an Email at
drjack9650@gmail.com
drjack9650@gmail.com
Navigate
Integrity-first tutoring: explanations and feedback only — we do not complete graded work. Learn more.