Assuming a standard 3-15 psig range, the sensitivity of a 50 - 100 degrees Fahrenheit transmitter would be:

Prepare for the BOMA Instrumentation and Controls Test with our interactive quiz. Boost knowledge with multiple-choice questions and expert hints. Ace your exam with confidence!

To determine the sensitivity of a transmitter with a standard pressure range of 3-15 psig over a temperature range of 50 to 100 degrees Fahrenheit, we first need to calculate the total span of both the pressure and temperature ranges.

The pressure span can be calculated by subtracting the lower range value from the upper range value:

15 psig - 3 psig = 12 psig.

The temperature span is found similarly:

100°F - 50°F = 50°F.

Sensitivity is defined as the change in output (in this case, psig) per unit change in input (temperature in degrees Fahrenheit). This is calculated by dividing the pressure span by the temperature span:

Sensitivity = Pressure Span / Temperature Span = 12 psig / 50°F.

Calculating this gives:

Sensitivity = 0.24 psig/degree Fahrenheit.

This indicates that for every degree Fahrenheit change in temperature, the output pressure changes by 0.24 psig, making this sensitivity measurement essential for understanding how the transmitter will respond to changes in temperature within the specified ranges. This level of sensitivity ensures appropriate calibrations and adjustments in systems reliant on such measurements for accuracy in control and instrumentation.

Subscribe

Get the latest from Examzify

You can unsubscribe at any time. Read our privacy policy