Electronic calibrators are fast becoming the benchmark for measurement and are replacing mechanical types of instruments for testing and calibration checks. Techniques, usage, traceability requirements, and problems are changing quickly as technology advances in the development of these instruments. Information concerning these issues is often outdated by the time the technician receives it. Electronic calibrators use a microprocessor with digital measurement. What is the difference between analog and digital? Analog is a continuous signal whereas digital is an analog signal converted into numerical data or bits that computers can understand change and store. The numerical values are converted from the analog signal at intervals of time. As the amount of time becomes shorter and shorter between the intervals of conversion, it becomes close to impossible to distinguish it from the original analog signal. Stephan Schuster chairman of the Rainier Corp., summed it up as follows: Analog is the real world and digital is a numerical representation of the real world.” Computers are the driving force behind the digital revolution because information must be digital to be used on a computer. Just like any other computer, technology is often obsolete by the time an instrument is designed, manufactured, marketed, sold, and shipped to the end user