EPM Manual Model 42i
EPM Manual Model 42i
EPM Manual Model 42i
Instruction Manual
Chemiluminescence NO-NO2-NOx Analyzer
Part Number 101350-00
25Jul2015
2007 Thermo Fisher Scientific Inc. All rights reserved.
Specifications, terms and pricing are subject to change. Not all products are available in all countries. Please
consult your local sales representative for details.
Safety Review the following safety information carefully before using the analyzer.
This manual provides specific information on how to operate the analyzer,
however, if the analyzer is used in a manner not specified by the
manufacturer, the protection provided by the equipment may be impaired.
Safety and Equipment This manual contains important information to alert you to potential safety
hazards and risks of equipment damage. Refer to the following types of
Damage Alerts alerts you may see in this manual.
FCC Compliance Changes or modifications to this unit not expressly approved by the party
responsible for compliance could void the users authority to operate the
equipment.
Note This equipment has been tested and found to comply with the limits
for a Class A digital device, pursuant to Part 15 of the FCC Rules. These
limits are designed to provide reasonable protection against harmful
interference when the equipment is operated in a commercial environment.
This equipment generates, uses, and can radiate radio frequency energy
and, if not installed and used in accordance with the instruction manual,
may cause harmful interference to radio communications. Operation of this
equipment in a residential area is likely to cause harmful interference in
which case the user will be required to correct the interference at his own
expense.
WEEE Symbol The following symbol and description identify the WEEE marking used on
the instrument and in the associated documentation.
Symbol Description
Marking of electrical and electronic equipment which applies to waste
electrical and electronic equipment falling under the Directive 2002/96/EC
(WEEE) and the equipment that has been put on the market after 13 August
2005.
Where to Get Help Service is available from exclusive distributors worldwide. Contact one of
the phone numbers below for product support and technical information
or visit us on the web at www.thermo.com/aqi.
Figure 715. Rear Panel Analog Input and Output Pins ................................... 7-27
Figure 716. Replacing the Pressure Transducer ............................................. 7-32
Figure 717. Replacing the Scrubber................................................................ 7-37
Figure 718. Replacing the I/O Expansion Board (Optional) ............................ 7-38
Figure 719. Rear Panel Board Connectors ...................................................... 7-38
Figure 720. Replacing the Measurement Interface Board ............................. 7-41
Figure 721. Replacing the Flow Transducer ................................................... 7-42
Figure 722. Replacing the Front Panel Board and the LCD Module............... 7-45
Figure 81. Hardware Components .................................................................... 8-2
Figure 91. Flow Diagram, Zero/Span Option .................................................... 9-2
Figure 92. Flow Diagram, Internal Permeation with Zero/Span Valve ............ 9-3
Figure 93. Flow Diagram, Lag Volume ............................................................ 9-12
Figure 91. Bench Mounting ............................................................................. 9-16
Figure 92. EIA Rack Mounting ........................................................................ 9-17
Figure 93. Retrofit Rack Mounting.................................................................. 9-18
Figure 94. Rack Mount Option Assembly ....................................................... 9-19
Figure B1. Flags .............................................................................................. B-14
Principle of The Model 42i operates on the principle that nitric oxide (NO) and ozone
(O3) react to produce a characteristic luminescence with an intensity
Operation linearly proportional to the NO concentration. Infrared light emission
results when electronically excited NO2 molecules decay to lower energy
states. Specifically:
NO + O 3 NO 2 + O 2 + h
Lifting When lifting the instrument, use procedure appropriate to lifting a heavy
object, such as, bending at the knees while keeping your back straight and
upright. Grasp the instrument at the bottom in the front and at the rear of
the unit. Although one person can lift the unit, it is desirable to have two
persons lifting, one by grasping the bottom in the front and the other by
grasping the bottom in the rear.
Unpacking and The Model 42i is shipped complete in one container. If there is obvious
damage to the shipping container when you receive the instrument, notify
Inspection the carrier immediately and hold for inspection. The carrier is responsible
for any damage incurred during shipment.
Use the following procedure to unpack and inspect the instrument.
1. Remove the instrument from its shipping container and set it on a table
or bench that allows easy access to both the front and rear.
Remove Packing
(2 pieces)
Remove Packing
(2 pieces)
Units without Optional I/O Board Units with Optional I/O Board
6. Check that all connectors and circuit boards are firmly attached.
1. Connect the sample line to the SAMPLE bulkhead on the rear panel
(Figure 23). Ensure that the sample line is not contaminated by dirty,
wet, or incompatible materials. All tubing should be constructed of
FEP Teflon, 316 stainless steel, borosilicate glass, or similar tubing
with an OD of 1/4-inch and a minimum ID of 1/8-inch. The length of
the tubing should be less than 10 feet.
SAMPLE
Connecting Several components are available for connecting external devices to iSeries
instruments.
External Devices
These connection options include:
Individual terminal board PCB assemblies
Terminal block and cable kits (optional)
Individual cables (optional)
For detailed information on the optional connection components, refer to
the Optional Equipment chapter. For associated part numbers, refer to
External Device Connection Components on page 7-6.
Terminal Board PCB The terminal board PCB assemblies are circuit boards with a D-Sub
Assemblies connector on one side and a series of screw terminals on the other side.
This assembly provide a convenient mechanism for connecting wires from
a data system to the analyzers I/O connectors.
The following terminal board PCB assemblies are available for iSeries
instruments:
I/O terminal board PCB assembly, 37 pin (standard)
D/O terminal board PCB assembly, 37 pin (standard)
25-pin terminal board PCB assembly, (included with optional I/O
expansion board)
I/O Terminal Board Figure 25 shows the recommended method for attaching the cable to the
terminal board using the included tie-down and spacer. Table 21
identifies the connector pins and associated signals.
Note Not all of the I/O available in the instrument is brought out on the
supplied terminal board. If more I/O is desired, an alternative means of
connection is required. See optional Terminal Block and Cable Kits.
See Detail B
See Detail A
Detail A Detail B
Assembled Connector
D/O Terminal Board Figure 26 shows the recommended method for attaching the cable to the
terminal board using the included tie-down and spacer. Table 22
identifies the connector pins and associated signals.
See Detail A
See Detail B
Detail A Detail B
Assembled Connector
25-Pin Terminal Board The 25-pin terminal board is included with the optional I/O Expansion
Board.
See Detail A
See Detail B
Detail A Detail B
Assembled Connector
Note it is best to turn the ozonator on and let the instrument run
overnight before calibration in order to obtain the most accurate
information.
Display The 320 x 240 graphics liquid-crystal display (LCD) shows the sample
concentrations, instrument parameters, instrument controls, help, and error
messages. Some menus contain more items than can be displayed at one
time. For these menus, use and to move the cursor up and
down to each item.
CAUTION If the LCD panel breaks, do not let the liquid crystal contact
your skin or clothes. If the liquid crystal contacts your skin or clothes, wash
it off immediately using soap and water.
Pushbuttons The Pushbuttons allow the user to traverse the various screens/menus.
Soft Keys The soft keys are multi-functional keys that use part of the display to
identify their function. The function of the soft keys is to provide a
shortcut to the most often used menus and screens. They are located
directly underneath the display, and user-defined labels in the lower part of
the display indicate the function of each key at that time.
To change a soft key, place the menu cursor > on the item of the selected
menu or screen you wish to set. Press followed by the selected soft
key within 1 second of pressing the right-arrow key. The edit soft key
prompt will be displayed for configuration of the new label.
Note Not all menu items may be assigned to soft keys. If a particular menu
or screen item cannot be assigned, the key assignment screen will not come
up upon entering right-arrow-soft key combinations. All items under the
Service menu (including the menu itself) cannot be assigned soft keys.
Alphanumeric Entry The alphanumeric entry screen is used to enter strings consisting of letters,
Screen numbers, and other characters. The cursor may be positioned within the
entry line using the and keys. If a character is entered over an
existing character, that character will be overwritten. Use the and
keys to switch between the entry line and the keyboard as well as to
move within the keyboard. To select a character to add to the string, use
the cursor keys to position the cursor over the desired character, and then
press the key to add that character to the entry line.
On the right side of the keyboard are special functions. BKSP is used to
move the cursor in the entry line one place to the left, deleting the
character that was to the left of the cursor and moving any character at or
to the right of the cursor one place to the left. PAGE is used to change the
keyboard character page. For the English language, this switches between
upper and lower-case alphabetic characters. SAVE stores the string from the
entry line into the parameter. Alternately, if the active cursor is moved to
the entry line, may be pressed to store the string from the entry line
into the parameter.
Firmware The Model 42i utilizes the menu-driven firmware as illustrated by the
flowchart in Figure 33. The Power-Up screen, shown at the top of the
Overview flowchart, is displayed each time the instrument is turned on. This screen is
displayed while the instrument is warming up and performing self-checks.
After the warm-up period, the Run screen is automatically displayed. The
Run screen is the normal operating screen. It displays the NO, NO2, and
NOx concentrations, depending on operating mode. From the Run screen,
the Main Menu can be displayed by pressing . The Main Menu
contains a list of submenus. Each submenu contains related instrument
settings. This chapter describes each submenu and screen in detail. Refer to
the appropriate sections for more information.
Self Test
Screen
Run Screen
Main Menu
Power-Up Screen The Power-Up screen is displayed when power is applied to the Model 42i.
This screen is displayed while the internal components are warming up and
diagnostic checks are being performed.
Run Screen The Run screen displays the NO, NO2, and NOx concentrations. The
status bar displays the time (24-hour format), the password (lock) icon,
service (wrench) icon, alarm (bell) icon, and optional zero/span sample
solenoid valve status, if installed.
The word SAMPLE on the left of the status bar indicates the analyzer has
the span/zero valve option and is in SAMPLE mode. Other modes
appear in the same area of the display as ZERO or SPAN. For more
information about the optional solenoid valves, see Chapter 9, Optional
Equipment.
When operating in dual or auto range mode two sets of coefficients are
used to calculate the NO-NO2-NOx High and Low concentrations.
Also, two averaging times are usedone for each range. The title bar
indicates which range concentrations are displayed. The words LOW
RANGE CONCENTRATION on the top of the display indicates that
the low concentration is displayed. In dual range mode, pressing the
and arrows will toggle between high and low concentrations. The
example below shows the Run screen in single range mode.
Main Menu The Main Menu contains a number of submenus. Instrument parameters
and settings can be read and modified within the submenus according to
their function. The concentration appears above the Main Menu and
submenus in every screen. The Service menu is visible only when the
instrument is in service mode. For more information on the service mode,
see Service Mode later in this chapter.
Use and to move the cursor up and down.
Press to make a selection.
Press to return to the Main Menu or to return to the Run
screen.
XXLO WCONCENTRATIONXXLO W
NO 62.7 PPB
NO2 25.5 PPB
NOx 88.2 PPB
xSAMPLE 12:34 x
MAIN MENU:
>RANGE
AVERAGING TIME
CALIBRATION FACTORS
CALIBRATION
INSTRUMENT CONTROLS
DIAGNOSTICS
ALARMS
SERVICE
PASSWORD
Range Menu The Range menu allows the operator to select the gas units, NO-NO2-NOx
ranges, and to set the custom ranges. The screens below show the range
menu in single range mode and dual/auto range modes. The only
difference between the screens are the words HI and LO to indicate
which range is displayed. For more information about the single, dual and
auto range modes, see Single Range Mode, Dual Range Mode, and
Auto Range Mode below.
In the Main Menu, choose Range
RANGE: RANGE:
>GAS UNITS PPB >GAS UNITS PPB
NO RANGE 50 HI NO RANGE 500
NO2 RANGE 50 LO NO RANGE 50
NOx RANGE 50 HI NO2 RANGE 500
SETCUSTOM RANGES LO NO2 RANGE 50
RANGE AVG DIAGS ALARM RANGE AVG DIAGS ALARM
Single Range Mode In the single range mode, the NO, NO2, and NOx channels each have one
range, one averaging time, and one span coefficient.
By default, the three analog outputs are arranged on the rear panel
connector as shown in Figure 34. See Table 32 for channels and pin
connections. Single range mode may be selected from the Range Mode
Select on page 3-76.
Note All channels are user definable. If any customization has been made
to the analog output configuration, the default selections may not apply.
Dual Range Mode In the dual range mode, there are two independent analog outputs. These
are labeled simply as the High Range and the Low Range. Each
channel has its own analog output range, averaging time, and span
coefficient.
This enables the sample concentration reading to be sent to the analog
outputs at two different ranges. For example, the low NO analog output
Note All channels are user definable. If any customization has been made
to the analog output configuration, the default selections may not apply.
Auto Range Mode The auto range mode switches the NO, NO2, and NOx analog outputs
between high and low ranges, depending on the NOx concentration level.
The high and low ranges are defined in the Range menu.
For example, suppose the low range is set to 20 ppb and the high range is
set to 100 ppb (Figure 36). Sample concentrations below 50 ppb are
presented to the low ranges analog outputs and sample concentrations
above 50 ppb are presented to the high ranges analog outputs. When the
low range is active, the status output is at 0 volts. When the high range is
active, the status output is at half of full-scale.
When the high ranges are active, the NOx concentration must drop to 95%
of the low NOx range for the low ranges to become active.
In addition to each channel having two ranges, each channel has two span
coefficients. There are two span coefficients so that each range can be
calibrated separately. This is necessary if the two ranges are not close to one
another. For example, the low NO range is set to 050 ppb and the high
NO range is set to 020,000 ppb.
By default, in the auto range mode, the analog outputs are arranged on the
rear panel connector as shown in Figure 37. See Table 34 for channels
and pin connections. Auto range mode may be selected from the Range
Mode Select on page 3-76.
Note All channels are user definable. If any customization has been made
to the analog output configuration, the default selections may not apply.
Gas Units The Gas Units screen defines how the NO, NO2, and NOx concentration
readings are expressed. Gas units of parts per billion (ppb), parts per
million (ppm), micrograms per cubic meter (g/m3), or milligrams per
cubic meter (mg/m3) are available. The g/m3 and mg/m3 gas
concentration modes are calculated using a standard pressure of 760
mmHg and a standard temperature of 20 C.
When switching the selected units from ppb or ppm to g/m3 or mg/m3,
the analog ranges all default to the highest range in that mode. For
example, when switching from mg/m3 to ppm, all the ranges default to 20
ppm. Therefore, whenever you change units, you should also check the
range settings.
In the Main Menu, choose Range > Gas Units.
Note If the units change from ppb/ppm to g/m3/mg/m3 or vice versa, the
instrument should be re-calibrated, particularly if the users standard
temperature is different from 20 C. A display warning will appear that
ranges will be defaulted and calibration parameters reset.
GAS UNITS:
CURRENTLY: PPB
SET TO: UG/M3 ?
AND SET RANGES AND CAL TOX
DEFAULT SETTINGSXXXXXXXXXX
CHANGE VALUE SAVE
NO, NO2, and NOx The NO, NO2, and NOx Ranges screen defines the concentration range of
Ranges the analog outputs. For example, a NO2 range of 050 ppb restricts the
NO2 analog output to concentrations between 0 and 50 ppb.
The display shows the current NO, NO2, or NOx range. The next line of
the display is used to change the range. The range screen is similar for the
single, dual, and auto range modes. The only difference between the
screens are the words High and Low to indicate which range is
displayed. The example below shows the NO range screen in single mode.
For more information about the dual and auto range modes, see Single
Range Mode, Dual Range Mode, and Auto Range Mode earlier in
this chapter.
Table 35 lists the available operating ranges. Table 36 lists the extended
ranges. When switching from standard to extended ranges, the PMT
voltage must be readjusted. For more information about readjusting the
PMT voltage, see Chapter 7, Servicing.
In the Main Menu, choose Range > NO, NO2, or NOx Range.
NO RANGE:
CURRENTLY: 50
SET TO: 100 ?
CHANGE VALUE
SAVE VALUE
C1, C2, and C3 are custom ranges. For more information about custom
ranges, see Set Custom Ranges below.
Set Custom Ranges The Set Custom Ranges menu lists three custom ranges, which are user-
defined. In the standard range mode, any value between 50 ppb (0.05
ppm) and 20,000 ppb (20 ppm) can be specified as a range. In the g/m3
(mg/m3) mode, any value between 100 g/m3 (0.1 mg/m3) and 30,000
g/m3 (30 mg/m3) can be specified as a range. In the extended range mode,
any value between 200 ppb (0.02 ppm) and 100,000 ppb (100 ppm) can
be specified as a range. In the g/m3 (mg/m3) mode, any value between 500
g/m3 (0.5 mg/m3) and 150,000 g/m3 (150 mg/m3) can be specified as a
range.
In the Main Menu, choose Range > Set Custom Ranges.
CUSTOM RANGES:
>CUSTOM RANGE 1 55.6
CUSTOM RANGE 2 75.0
CUSTOM RANGE 3 125.0
Custom Ranges The Custom Ranges screen is used to define the custom ranges.
The display shows the current custom range. The next line of the display is
used to set the range. To use the custom full-scale range, be sure to select it
(Custom range 1, 2, or 3) in the NO, NO2, or NOx Ranges screen. For
more information about selecting ranges, see NO, NO2, and NOx
Ranges above.
In the Main Menu, choose Range > Set Custom Ranges > Custom
Range 1, 2, or 3.
CUSTOM RANGE 1:
CURRENTLY: 55.6
SET TO: 000055.7 ?
MOVE CURSOR
CHANGE VALUE SAVE
Averaging Time The Averaging Time defines a time period (10 to 300 seconds) over which
NO, NO2, and NOx measurements are taken. The average concentration of
the NO, NO2, and NOx readings are calculated for that time period. The
front panel display and analog outputs are updated every 10 seconds for
averaging times between 10 and 300 seconds. An averaging time of 10
seconds, for example, means that the average concentration of the last 10
seconds will be output at each update. An averaging time of 300 seconds
means that the moving average concentration of the last 300 seconds will
be output at each update. Therefore, the lower the averaging time the faster
the front panel display and analog outputs respond to concentration
changes. Longer averaging times are typically used to smooth output data.
The Averaging Time screen for the single range mode is shown below. In
the dual and auto range modes, an Averaging Time Menu is displayed
before the averaging time screens. This additional menu is needed because
the dual and auto range modes have two averaging times (high and low).
The Averaging Time screen functions the same way in the single, dual, and
auto range modes. The following averaging times are available: 10, 20, 30,
60, 90, 120, 180, 240, and 300 seconds. Additional averaging times are
available in NO and NOx modes: 1, 2, and 5 seconds. For more
information about the manual mode, see Auto/Manual Mode later in this
chapter.
In the Main Menu, choose Averaging Time.
AVERAGING TIME:
CURRENTLY: 30 SEC
SET TO: 10 SEC ?
CHANGE VALUE
SAVE VALUE
Calibration Factors Calibration factors are used to correct the NO, NO2, and NOx
concentration readings that the instrument generates using its own internal
Menu calibration data. The Calibration Factors menu displays the calibration
factors. The screens below show the calibration factors menu in single
mode and dual/auto range modes. The only difference between the screens
are the words HI and LO to indicate which range is displayed.
Normally, the calibration factors are calculated automatically using the
functions described in Calibration menu later in this chapter. However,
the calibration factors can also be set manually using the functions in this
menu.
In the Main Menu, choose Calibration Factors.
NO and NOx Backgrounds The NO and NOx background corrections are determined during zero
calibration. The NO background is the amount of signal read by the
analyzer in the NO channel while sampling zero air. The NOx background
is the amount of signal read by the analyzer in the NOx channel while
sampling zero air. Although the background is expressed in terms of
concentration, the background signal is actually the combination of
electrical offsets, PMT dark currents, and trace substances undergoing
chemiluminescence. Before the analyzer sets the NO and NOx readings to
zero, it stores these values as the NO and NOx background corrections,
respectively. The NO2 background correction is determined from the NO
NO BACKGROUND:
NO: 4.4
SET BKG TO: 0.00 ?
INC/DEC
SAVE VALUE
NO, NO2, and NOx The NO, NO2, and NOx span coefficients are usually calculated by the
Coefficients instrument processor during calibration. The span coefficients are used to
correct the NO, NO2, and NOx readings. The NO and NOx span
coefficients normally has a value near 1.000. The NO2 span coefficient
normally has a value between 0.95 and 1.050.
The NO, NO2, and NOx Coefficient screens allow the NO, NO2, and
NOx span coefficients to be manually changed while sampling span gas of
known concentration. The NO, NO2, and NOx Coefficient screens operate
the same way. Therefore, the following description of the NO coefficient
screen applies to the NO2, and NOx coefficient screens as well.
The display shows the current NO concentration reading. The next line of
the display shows the NO span coefficient that is stored in memory and is
being used to correct the NO concentration. Notice that as the span
coefficient value is changed, the current NO concentration reading on the
above line also changes. However, no actual changes are made to the value
stored in memory until is pressed. Only proposed changes, as
indicated by a question mark prompt, are displayed until is pressed.
In dual or auto range modes, HIGH or LOW is displayed to indicate
the calibration of the high or low coefficient. The example below shows the
coefficient screen in dual/auto range mode.
Note The screen will show SPAN CONC NOT VALID if the
coefficient change causes the concentration to be either higher than the
selected range or less than or equal to zero.
In the Main Menu, choose Calibration Factors > NO, NO2, or NOx
Coef.
NO COEFFICIENT:
NO: 51.4
SET COEF TO: 1.000
INC/DEC
SAVE VALUE
Reset User Calibration The Reset User Calibration Defaults screen allows the user to reset the
Defaults calibration factors to factory defaults.
In the Main Menu, choose Calibration Factors > Reset User Cal
Defaults.
RESTORE RESTORE
ARE YOU SURE YOU WANT TO?x
PRESS TO CONFIRM RESTORE
Calibration Menu The Calibration menu is used to automatically set the zero background of
NO and NOx, and the NO, NO2, and NOx span coefficients. The screens
below show the calibration menu in single mode and dual/auto range
modes. The zero/span check is visible only if the zero/span option is
installed.
The calibration procedure is the same in dual, auto, or single range,
however, there are two sets of gas coefficients in dual or auto range (i.e. low
and high coefficients). This enables each range to be calibrated separately.
When calibrating the instrument in dual or auto range, be sure to use a low
span gas to calibrate the low range and a high span gas to calibrate the high
range.
In the Main Menu, choose Calibration.
CALIBRATION: CALIBRATION:
>CAL NO BACKGROUND >CAL NO BACKGROUND
CAL NOx BACKGROUND CAL NOx BACKGROUND
CAL NO COEFFICIENT CAL HI NO COEFFICIENT
CAL NO2 COEFFICIENT CAL HI NO2 COEFFICENT
CAL NOx COEFFICIENT CAL HI NOx COEFFICIENT
Calibrate NO and NOx The Calibrate NO and NOx Background screens are used to adjust the
Backgrounds instrument zero, or background. Before making an adjustment, be sure the
analyzer samples zero air until the readings stabilize. The display shows the
current NO or NOx reading.
It is important to note the averaging time when calibrating. The longer the
averaging time, the more accurate the calibration will be. To be most
accurate, use the 300-second averaging time. For more information about
calibration, see Chapter 4, Calibration.
In the Main Menu, choose Calibration > Calibrate NO or NOx
Background.
NO BACKGROUND:
NO: 1.2
CURRENTLY: 0.00 ?
SET CO TO ZERO
Calibrate NO, NO2, and The Calibrate NO Coefficient screen is used to adjust the NO span
NOx Coefficients concentration while sampling span gas of known concentration. All
calibration screens operate the same way. Therefore, the following
description of the NO calibration screen applies to the NO2 and NOx
calibration screens as well.
The display shows the current NO concentration reading and the current
NO range. The next line of the display is where the NO calibration gas
concentration is entered.
It is important to note the averaging time when calibrating. The longer the
averaging time, the more accurate the calibration will be. To be most
accurate, use the 300-second averaging time. For more information about
calibration, see Chapter 4, Calibration.
In the Main Menu, choose Calibration > Cal NO, NO2 or NOx
Coefficient.
CALIBRATE NO:
NO: 25.5
SPAN CONC: 0001.000 ?
MOVE CURSOR
CHANGE VALUE SAVE
Zero/Span Check The Zero/Span Check menu is available only if the zero/span valve option
is installed. It is used to program the instrument to perform fully
automated zero and span checks or adjustments.
In the Main Menu, choose Calibration > Zero/Span Check.
ZERO/SPAN CHECK:
>NEXT TIME 01Jan06 12:00
PERIOD HR 24
TOTAL DURATION HR 1.5
ZERO DURATION MIN 30
SPAN DURATION MIN 30
Next Time The Next Time screen is used to view and set the initial date and time (24-
hour format) of the zero/span check. Once the initial zero/span check is
performed, the date and time of the next zero/span check is calculated and
displayed.
In the Main Menu, choose Calibration > Zero/Span Check > Next
Time.
Period Hours The Period Hours screen defines the period or interval between zero/span
checks. Periods between 0 and 999 hours are acceptable. To turn the
zero/span check off, set the period to 0.
In the Main Menu, choose Calibration > Zero/Span Check > Period
HR.
ZERO/SPAN PERIOD:
CURRENTLY: 024 HRS
SET TO: 025 HRS ?
MOVE CURSOR
CHANGE VALUE SAVE
Total Duration Hour The Total Duration Hour line is the sum of the zero, span, and purge
duration minutes, and is for viewing only.
Zero/Span/Purge The Zero Duration Minutes screen defines how long zero air is sampled by
Duration Minutes the instrument. The Span and Purge Duration Minutes screens look and
function the same way as the zero duration screen. The span duration
screen is used to set how long the span gas is sampled by the instrument.
The purge duration screen is used to set how long the purge period will be
after doing a zero or span check. This gives the instrument time to flush
out the zero and span gas before any meaningful data is taken. Logged data
is flagged as taken during a purge to show that the data is suspect.
Durations between 0 and 99 minutes are acceptable. Each time a zero/span
check occurs the zero check is done first, followed by the span check. To
perform just a zero check, set the span duration to 0 (off). The same applies
to perform just a span check.
In the Main Menu, choose Calibration > Zero/Span Check > Zero,
Span or Purge Duration Min.
ZERO DURATION:
CURRENTLY: 30 MIN
SET TO: 31 MIN ?
MOVE CURSOR
CHANGE VALUE SAVE
Zero/Span Averaging The Zero/Span Averaging Time screen allows the user to set the zero/span
Time averaging time. The zero/span averaging time is used by the analyzer only
when performing an automatic zero or span check or adjustment. The
analyzers averaging time is used for all other functions. The following
averaging times are available: 1, 2, 5, 10, 20, 30, 60, 90, 120, 180, 240,
and 300 seconds.
In the Main Menu, choose Calibration > Zero/Span Check >
Zero/Span Avg Sec.
CHANGE VALUE
SAVE VALUE
Zero/Span Calibration Zero and Span Calibration Reset are toggle items that change between yes
Reset or no when selected, and are displayed only if Auto Calibration is set to
YES in the Instrument Configuration screen in the Diagnostics menu.
If the zero calibration reset line is set to yes, then not only is a zero check
done, but a zero adjustment is made. If the span calibration reset line is set
to yes, then not only is a span check done, but a span adjustment is made.
(This is how to set up a scheduled, recurring auto calibration.)
Zero/Span Ratio The Zero/Span Ratio screen is used to set the ratio of zero checks to span
checks. For example, if this value is set to 1, a span check will follow every
zero check. If this value is set to 3, there will be three zero checks between
each span check. This value may be set from 1 to 10, with 1 as default.
In the Main Menu, choose Calibration > Zero/Span Check >
Zero/Span Ratio.
ZERO/SPAN RATIO:
CURRENTLY: 1:1
SET TO: 2:1 ?
CHANGE VALUE
SAVE VALUE
Instrument Controls The Instrument Controls menu contains a number of items that may be
selected to control various instrument operational parameters. The software
Menu controls listed in this menu enable control of the listed instrument
functions.
In the Main Menu, choose Instrument Controls.
INSTRUMENT CONTROLS:
>OZONATOR
PMT SUPPLY
AUTO/MANUAL MODE
DATALOGGING SETTINGS
COMMUNICATION SETTINGS
I/O CONFIGURATION
TEMPERATURE COMPENSATION
PRESSURE COMPENSATION
SCREEN CONTRAST
SERVICE MODE
DATE/TIME
TIMEZONE
Ozonator The Ozonator screen is used to turn the internal ozonator on or off. The
display shows the status of the control line that turns the ozonator on or
off. The next line of the display shows the user-specified ozonator setting.
Under most conditions, the control line status and ozonator set status are
the same. However, as a safety precaution, the microprocessor can override
the user-specified ozonator setting. This occurs only if the ozonator flow
doesnt indicate any flow or if the NO2 converter temperature is below the
minimum alarm limit. In this case, an alarm is activated and the ozonator is
turned off. This is done to prevent the ozonator from overheating, which
will result in permanent damage to the ozonator, or if the converter
temperature drops below the minimum limit, which reduces the
effectiveness of the ozone destruct.
It is possible, however, to override the ozonator shut-off due to converter
temperature being out of range, by setting the ozonator safety OFF in the
Service Mode.
OZONATOR:
CURRENTLY: OFF
SET TO: ON ?
TOGGLE VALUE
PMT Supply The PMT Supply screen is used to turn the PMT power supply on or off.
This is useful in a troubleshooting situation.
In the Main Menu, choose Instrument Controls > PMT Supply.
PMT SUPPLY:
CURRENTLY: OFF
SET TO: ON ?
TOGGLE VALUE
Auto/Manual Mode The Auto/Manual Mode screen allows selection of the automatic mode
(NO/NOx), NO mode (manual NO), or NOx mode (manual NOx). The
auto cycle mode switches the mode solenoid valve automatically on a 10
second cycle so that NO, NO2, and NOx concentrations are determined.
The manual NO mode puts the mode solenoid valve into the open position
so that the sample gas bypasses the NO2-to-NO converter. Therefore, only
the NO concentration is determined. The manual NOx mode puts the
mode solenoid valve into the closed position so that the sample gas passes
through the NO2-to-NO converter. Therefore, only the NOx concentration
is determined. In the manual modes, additional averaging times of 1, 2,
and 5 seconds are available from the Averaging Times screen.
In the Main Menu, choose Instrument Controls > Auto/Manual
Mode.
MODE:
CURRENTLY: MANUAL NO
SET TO: NO/NOX ?
CHANGE VALUE
SAVE VALUE
DATALOGGING SETTINGS:
>SELECT SREC/LREC SREC
VIEW LOGGED DATA
ERASE LOG
SELECT CONTENT
COMMIT CONTENT
Select Srec/Lrec The Select Srec/Lrec screen is used to select the log record type for other
operations in this menu.
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Select Srec/Lrec.
TOGGLE VALUE
View Logged Data The View Logged Data screen is used to select the starting point to view
the logged data by the number of records or by date and time. Note that
both types of records cannot be viewed at the same time, only the selected
record type.
In the Main Menu, choose Instrument Controls > Datalogging Settings
> View Logged Data.
CHANGE ACCEPT
Number of Records The Number of Records screen is used to select the number of records to
view, ending with the most recent. It also shows the total number of
records that have been logged for the selected record type.
MOVE CURSOR
CHANGE VALUE SAVE
The Record Display screen (read only) displays the selected records.
Date and Time The Date and Time screen is used to set a start date and time for which to
view logged data. For example, if 20 Jan 2007 10:00 is entered, then the
first logged data record that is displayed is the first record after this time. If
set to one minute logging, this would be at 20 Jan 2007 10:01.
The Record Display screen (read only) displays the selected records.
Erase Log The Erase Log screen is used to erase all saved data for the selected record
type only (not both srecs and lrecs).
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Erase Log.
ERASE LREC LOG FILE DATA? ERASE LREC LOG FILE DATA?
ERASE ERASE
ARE YOU SURE YOU WANT TO?x
PRESS TO CONFIRM ERASURE
Select Content The Select Content submenu displays a list of 32 record fields to use to
configure the data to be logged. When a field is selected, a submenu list
appears of the items to choose from to assign to those fields. Item types are
Concentrations, Other Measurements, and Analog Inputs (if the I/O
expansion board is installed). These record fields comprise a temporary list
of items for the selected record type that must be committed via the
datalogging menu before the changes will apply. Note that committing any
changes to this list will erase all currently logged data for the selected record
type only, as the format of the stored data is changed.
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Select Content.
LREC FIELDS:
>FIELD 1 NO
FIELD 2 NOX
FIELD 3 PRES
FIELD 4 PMTT
FIELD 5 INTT
Choose Field Data The Choose Field Data submenu displays a list of the types of data that can
be logged for the current field. Choices are Concentrations, Other
Measurements, and Analog Inputs (if the I/O expansion board is installed).
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Select Content > Field 132.
Note The ANALOG INPUTS item is only displayed if the I/O expansion
board option is intalled.
Concentrations The Concentrations screen allows the user to assign one of the
concentrations to the selected record field. The selected item is shown by
<-- after it. Note that at this point, pressing indicates that these are
proposed changes as opposed to implemented changes. To change the
selected record format and erase record log file data, see Commit
Content below. Range (NOx) is only visible in auto range mode.
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Select Content > select Field > Concentrations.
CONCENTRATIONS:
>NONE
NO <--
NO2
NOx
LO NO
LO NO2
LO NOx
HI NO
HI NO2
HI NOx
RANGE (NOX)
Other Measurements The Other Measurements screen allows the user to assign one of the other
available measurement types to the selected record field. The selected item
is shown by <-- after it. Note that at this point, pressing indicates
that these are proposed changes as opposed to implemented changes. To
change the selected record format and erase record log file data, see
Commit Content below. BKG and COEF items are displayed only if
Ouput Conc Cal option is set to YES in the Instrument Configuration
screen.
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Select Content > select Field > Other Measurements.
OTHER MEASUREMENTS:
>NONE
INT TEMP
CHAMBER TEMP
COOLER TEMP
NO2 CNV TEMP
Analog Inputs The Analog Inputs screen allows the user to select the parameter (none or
analog inputs 18) to the selected record field. The selected item is shown
by <-- after it. Note that at this point, pressing indicates that these
are proposed changes as opposed to implemented changes. To change the
selected record format and erase record log file data, see Commit
Content below. The analog inputs 18 will not show if the Ouput Conc
Cal option is set to YES in the Instrument Configuration screen.
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Select Content > select Field > Analog Inputs.
ANALOG INPUTS:
>NONE
ANALOG IN 1
ANALOG IN 2
ANALOG IN 3
ANALOG IN 4
Commit Content The Commit Content screen is used to save any changes that have been
made to any of the record fields for the selected record type. Saving changes
will erase record log file data for that record type. If no changes have been
made NO CHANGES TO RECORD LIST! will appear. For more
information about selecting the content of logged data fields, see Select
Content above.
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Commit Content.
Reset to Default Content The Reset to Default Content screen is used to reset all of the datalogging
field items to default values for the selected record type. For more
information about selecting the content of logged data fields, see Select
Content above.
Configure Datalogging The Configure Datalogging menu deals with datalogging configuration for
the currently selected record type.
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Configure Datalogging.
DATALOGGING SETTINGS:
>LOGGING PERIOD MIN 60
MEMORY ALLOCATION % 50
DATA TREATMENT AVG
Logging Period Min The Logging Period Min screen is used to select the logging period in
minutes for the selected record format (srec or lrec). List of choices include:
off, 1, 5, 15, 30, and 60 minutes (default).
In the Main Menu, choose Instrument Controls > Datalogging Settings
> Configure Datalogging > Logging Period Min.
Memory Allocation The Memory Allocation Percent screen is used to select the percentage of
Percent total memory that may be used by the selected record type (lrecs or srecs).
Data Treatment The Data Treatment screen is used to select the data type for the selected
record type: whether the data should be averaged over the interval, the
minimum or maximum measured during the interval, or the current value
(last value measured). Data treatment doesnt apply to all data, just to the
concentration measurement. All other data points log the current value at
the end of the interval.
COMMUNICATION SETTINGS:
>SERIAL SETTINGS
INSTRUMENT ID
COMMUNICATION PROTOCOL
STREAMING DATA CONFIG
TCP/IP SETTINGS
Serial Settings The Serial Setting submenu is used for serial communications control and
configuration.
In the Main Menu, choose Instrument Controls > Communication
Settings > Serial Settings.
SERIAL SETTINGS:
>BAUD RATE 9600
DATA BITS 8
PARITY NONE
STOP BITS 1
RS-232/485 SEL RS-232
Baud Rate The Baud Rate screen is used to set the RS-232/RS-485 interface baud rate.
Baud rates of 1200, 2400, 4800, 9600, 19200, 38400, 57600, and 115200
are available. The analyzers default baud rate is set to 9600 to provide
backwards compatibility with the older C-series analyzers.
In the Main Menu, choose Instrument Controls > Communication
Settings > Serial Settings > Baud Rate.
BAUD RATE:
CURRENTLY: 9600
SET TO: 19200 ?
CHANGE VALUE
SAVE VALUE
Data Bits The Data Bits screen is used to set the number of serial data bits to either 7
or 8 (default).
DATA BITS:
CURRENTLY: 8
SET TO: 7 ?
CHANGE VALUE
SAVE VALUE
Parity The Parity screen is used to select the parity bit for the serial port to None
(default), Even, or Odd.
In the Main Menu, choose Instrument Controls > Communication
Settings > Serial Settings > Parity.
PARITY:
CURRENTLY: NONE
SET TO: ODD ?
CHANGE VALUE
SAVE VALUE
Stop Bits The Stop Bits screen is used to set the number of stop bits for the serial
port to 1 (default) or 2.
In the Main Menu, choose Instrument Controls > Communication
Settings > Serial Settings > Stop Bits.
STOP BITS:
CURRENTLY: 1
SET TO: 2 ?
CHANGE VALUE
SAVE VALUE
RS-232/RS-485 Selection The RS-232/RS-485 Selection screen allows the user to choose between the
RS-232 or RS-485 specification for serial communication.
Equipment Damage Disconnect the serial cable before changing the RS-
232 and RS-485 selection to prevent damage to any equipment currently
connected to the analyzer.
Instrument ID The Instrument ID screen allows the operator to edit the instrument ID
number. The ID is used to identify the instrument when using the C-Link
or MODBUS protocols to control the instrument or collect data. It may be
necessary to edit the ID number if two or more instruments of the same
model are connected to one computer. Valid Instrument ID numbers are
from 0 to 127. The Model 42i has a default Instrument ID of 42. For
more information about the Instrument ID, see Appendix B C-Link
Protocol Commands or Appendix C MODBUS Protocol.
In the Main Menu, choose Instrument Controls > Communication
Settings > Instrument ID.
INSTRUMENT ID:
CURRENTLY: 42
SET TO: 50 ?
CHANGE VALUE
SAVE VALUE
Communication Protocol The Communication Protocol screen is used to change the protocol for
serial communications. Possible choices include: C-Link, MODBUS,
Geysitech, and Streaming Data.
In the Main Menu, choose Instrument Controls > Communication
Settings > Communication Protocol.
COMMUNICATION PROTOCOL:
CURRENTLY: CLINK
SET TO: STREAMING ?
CHANGE VALUE
SAVE VALUE
Streaming Data Configuration The Streaming Data Configuration menu is used to allow for configuration
of the 8 streaming data output items, streaming interval, current data
format, and current timestamp setting. The Choose Stream Data submenu
displays a list of the analog output signal group choices to choose from.
Choices are Concentrations, Other Measurements, and Analog Inputs (if
the I/O expansion board option is installed).
Streaming Data Interval The Streaming Data Interval screen is used to adjust how frequently a new
record will be generated. The following interval times are available: 1, 2, 5,
10, 20, 30, 60, 90, 120, 180, 240, and 300 seconds.
In the Main Menu, choose Instrument Controls > Communication
Settings > Streaming Data Config > Interval.
CHANGE VALUE
SAVE VALUE
Note Add Labels, Prepend Timestamp, and Add Flags are toggle items that
change between yes or no when selected.
Choose Stream Data The Choose Stream Data screen displays a list of the types of data that can
be sent via streaming data. Choices are Concentrations, Other
Measurements, and Analog Inputs (if the I/O expansion board is installed).
In the Main Menu, choose Instrument Controls > Communication
Settings > Streaming Data Config > Item 18.
Concentrations The Concentrations screen allows the user to assign one of the
concentrations to the selected streaming data item. The currently selected
item is shown by <-- after it. Once an item is selected, pressing
will save the selected streaming data item. In dual or auto range mode,
HI or LO is displayed to indicate high or low range concentrations.
Rang (NOx) is visible only in auto range mode.
In the Main Menu, choose Instrument Controls > Communication
Settings > Streaming Data Config > select Item > Concentrations.
CONCENTRATIONS:
>NONE
NO <--
NO2
NOx
RANGE (NOX)
Other Measurements The Other Measurements screen allows the user to assign one of the other
available measurement types to the selected streaming data item. The
currently selected item is shown by <-- after it. Once an item is selected,
pressing will save the selected streaming data item. BKG and COEF
items are displayed only if Ouput Conc Cal option is set to YES in the
Instrument Configuration screen.
In the Main Menu, choose Instrument Controls > Communication
Settings > Streaming Data Config > Select Item > Other
Measurements.
OTHER MEASUREMENTS:
>NONE
INT TEMP
CHAMBER TEMP
COOLER TEMP
NO2 CNV TEMP
CHAMBER PRES
FLOW
PMT VOLTS
NO BKG
NOx BKG
LO NO COEF
LO NO2 COEF
LO NOx COEF
HI NO COEF
HI NO2 COEF
HI NOx COEF
OZONATOR FLOW
EXT ALARMS
Analog Inputs The Analog Inputs screen allows the user to assign an analog input signal
(none or analog inputs 18) to the selected streaming data item. The
currently selected item is shown by <-- after it. Once an item is selected,
pressing will save the selected streaming data item. The analog
inputs 18 will not show if the Ouput Conc Cal option is set to YES in the
Instrument Configuration screen.
In the Main Menu, choose Instrument Controls > Communication
Settings > Streaming Data Config > Select Item > Analog Inputs.
ANALOG INPUTS:
>NONE
ANALOG IN 1
ANALOG IN 2
ANALOG IN 3
ANALOG IN 4
TCP/IP Settings The TCP/IP Settings menu is used for defining parameters that are
required for Ethernet communications.
Note The instrument power must be cycled after any of these parameters
have been changed for the change to take effect.
TCP/IP SETTINGS:
>USE DHCP OFF
IP ADDR 10.209.43.237
NETMASK 255.255.254.0
GATEWAY 10.209.42.1
HOST NAME iSeries
Use DHCP The Use DHCP screen is used to specify whether to use Dynamic Host
Configuration Protocol (DHCP) or not. When DHCP is enabled, the
network dynamically provides an IP address for the instrument.
In the Main Menu, choose Instrument Controls > Communication
Settings > TCP/IP Settings > Use DCHP.
DHCP:
CURRENTLY: OFF
SET TO: ON ?
TOGGLE VALUE
CYCLE POWER TO CHANGE DHCP
IP Address The IP Address screen is used to edit the IP address. The IP address can
only be changed when DHCP is off. If DHCP is on, the instrument will
respond with NOT SETTABLE IF DHCP IS ON. For more
information on DHCP, see Use DHCP above.
In the Main Menu, choose Instrument Controls > Communication
Settings > TCP/IP Settings > IP Address.
IP ADDRESS:
CURRENT: 10.209.43.237
SET TO: 10.209.43.237 1
MOVE CURSOR
CHANGE VALUE
SAVE VALUE
Netmask The Netmask screen is used to edit the netmask. The netmask is used to
determine the subnet on which the instrument can directly communicate
to other devices. The netmask can only be changed when DHCP is off. If
DHCP is on, the instrument will respond with NOT SETTABLE IF
DHCP IS ON. For more information on DHCP, see Use DHCP
above.
In the Main Menu, choose Instrument Controls > Communication
Settings > TCP/IP Settings > Netmask.
NETMASK:
CURRENT: 255.255.254.0
SET TO: 255.255.254.0 0
MOVE CURSOR
CHANGE VALUE
SAVE VALUE
Default Gateway The Default Gateway screen is used to edit the gateway address. The
default gateway can only be changed when DHCP is off. If DHCP is on,
the instrument will respond with NOT SETTABLE IF DHCP IS ON.
For more information on DHCP, see Use DHCP above. Any traffic to
addresses that are not on the local subnet will be routed through this
address.
In the Main Menu, choose Instrument Controls > Communication
Settings > TCP/IP Settings > Gateway.
DEFAULT GATEWAY:
CURRENT: 10.209.42.1
SET TO: 10.209.42.1 1
MOVE CURSOR
CHANGE VALUE
SAVE VALUE
Host Name The Host Name screen is used to edit the host name. When DHCP is
enabled, this name is reported to the DHCP server.
In the Main Menu, choose Instrument Controls > Communication
Settings > TCP/IP Settings > Host Name.
HOST NAME:
CURRENT: ISERIES
ISERIES SSSSS ?
ABCDEFGHIJKLMN BKSP
OPQRSTUVWXYZ PAGE
0123456789 ./- SAVE
Network Time Protocol The Network Time Protocol (NTP) Server screen is used to edit the IP
Server address of the NTP server. An NTP server may be used to periodically
synchronize the instruments real-time clock with a standard. More
information about the NTP servers and a list of public servers may be
found at http://www.ntp.org.
In the Main Menu, choose Instrument Controls > Communication
Settings > TCP/IP Settings > NTP Svr.
I/O Configuration The I/O Configuration menu deals with configuration of the analyzers
I/O system. The analog input configuration is displayed only if the I/O
expansion board option is installed.
I/O CONFIGURATION:
>OUTPUT RELAY SETTINGS
DIGITAL INPUT SETTINGS
ANALOG OUTPUT CONFIG
ANALOG INPUT CONFIG
Output Relay Settings The Output Relay Settings menu displays a list of the 10 digital output
relays available, and allows the user to select the logic state or instrument
parameter for the relay selected.
Note The digital outputs may take up to one second after the assigned
state occurs to show up on the outputs.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Output Relay Settings.
Logic State The Logic State menu item is used to change the selected I/O relay to
either normally open or normally closed. The default state is open, which
indicates that a relay connected between the digital output pin and ground
is normally open and closes to trigger the digital output action.
Press to toggle and set the logic state open or closed.
Instrument State The Instrument State submenu allows the user to select the instrument
state that is assigned to the selected relay output. A submenu lists signal
types of either alarm or non-alarm to choose from.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Output Relay Settings > Select Relay > Instrument State.
Alarms The Alarms screen allows the user to select an alarm status for the selected
relay output. The currently selected item is shown by <-- after it. The
I/O board status alarm is only present if the I/O expansion board is
installed. Zero and Span Check Cal are only present if the automated
zero/span check is enabled. See the Alarms Menu section later in this
chapter for more information on alarm items.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Output Relay Settings > Select Relay > Instrument State > Alarms.
Non-Alarm The Non-Alarm screen allows the user to select a non-alarm status for the
selected relay output. The currently selected item is shown by <-- after it.
The zero, span, and sample modes are only present if the zero/span valve
option is installed.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Output Relay Settings > select Relay > Instrument State > Non-Alarm.
ZERO MODE
SPAN MODE
NO MODE
NOX MODE
SAMPLE MODE
PURGE MODE
Digital Input Settings The Digital Input Settings menu displays a list of the 16 digital inputs
available, and allows the user to select the logic state and instrument
parameter for the relay selected.
Note The digital inputs must be asserted for at least one second for the
action to be activated.
Not all of the I/O available in the instrument is brought out on the
supplied terminal board. If more I/O is desired, an alternative means of
connection is required. (See optional Terminal Block and Cable Kits.)
In the Main Menu, choose Instrument Controls > I/O Configuration >
Digital Input Settings.
Logic State The Logic State menu item is used to change the selected I/O relay to
either normally open or normally closed. The default state is open, which
indicates that a relay connected between the digital input pin and ground is
normally open and closes to trigger the digital input action.
Press to toggle and set the logic state to normally open or
normally closed.
Instrument Action The Instrument Action submenu allows the user to choose the instrument
action that is assigned to the selected digital input. The zero and span
modes are only present if the zero/span valve option is installed.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Digital Input Settings > select Relay > Instrument Action.
CHOOSE ACTION:
>NONE
ZERO MODE
SPAN MODE
NO MODE <--
NOX MODE
SET BACKGROUND
CAL TO LO SPAN
AOUTS TO ZERO
AOUTS TO FS
CAL TO HI SPAN
EXT ALARM 1
EXT ALARM 2
EXT ALARM 3
Analog Output Configuration The Analog Output Configuration menu displays a list of the analog
(Select Channel) output channels available for configuration. Channel choices include all
voltage channels, all current channels, individual voltage channels 16, and
individual current channels 16 (if the I/O expansion board option is
installed).
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Output Config.
OUTPUT CHANNELS:
>ALLOW OVER/UNDER RANGE
ALL VOLTAGE CHANNELS
ALL CURRENT CHANNELS
VOLTAGE CHANNEL 1
VOLTAGE CHANNEL 2
Allow Over/Under Range The Allow Over/Under Range screen is used to select whether or not the
analog outputs are allowed to exceed the maximum selected value of 100
mV, 1 V, 5 V, 10 V, or 20 mA or the minimum selected value of 0 V, 0
TOGGLE VALUE
Analog Output Configuration The Analog Output Configuration menu displays a list of the analog
(Select Action) output configuration choices, from which the user selects the parameter to
adjust for the selected output channel. Configuration choices include
selecting range, setting minimum/maximum values, and choosing the
signal to output.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Output Config > All Voltage Channels, All Current
Channels, Voltage Channel 16 or Current Channel 16.
Select Range The Select Range screen is used to select the hardware range for the selected
analog output channel. Possible ranges for the voltage outputs are: 0-100
mV, 0-1, 0-5, and 0-10 V. Possible ranges for the current outputs are: 0-20
mA and 4-20 mA.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Output Config > select Channel > Select Range.
Minimum and Maximum The Minimum and Maximum Value screens are used to edit the zero (0)
Value and full-scale (100) values, respectively, in percentages for the selected
analog output channel. See Table 37 for a list of choices. The minimum
and maximum output value screens function the same way. The following
example shows the set minimum value screen.
In the Main Menu, choose Instrument Controls > IO Configuration >
Analog Output Config > select Channel > Set Minimum or Maximum
Value.
Choose Signal to Output The Choose Signal to Output screen displays a list of the analog output
signal choices. Choices are grouped into three categories: Concentrations,
Other Measurements, and Analog Inputs (if the I/O expansion board
option is installed). This allows the user to select the output signal that will
be assigned to each analog channel. The Concentrations screen is shown
below. See Table 38 for a list of items for each signal group choice. The
analog inputs 18 will not show if the Ouput Conc Cal option is set to
YES in the Instrument Configuration screen.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Output Config > select Channel > Choose Signal to Output.
Analog Input The Analog Input Configuration menu displays a list of the 8 analog input
Configuration channels available for configuration. This screen is only displayed if the I/O
expansion board option is installed. Configuration includes entering the
Descriptor, Units, Decimal Places, choice of 110 points in the table, and
corresponding number of points selected. This menu is not displayed if the
Output Conc Cal option is selected.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Input Config.
POINT 2
Descriptor The Descriptor screen allows the user to enter the descriptor, or name, for
the selected analog input channel. The descriptor is used in datalogging
and streaming data to report what data is being sent out. The descriptor
may be from 1 to 3 characters in length, and defaults to IN1 to IN8 (user
input channel number).
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Input Config > select Channel > Descriptor.
Units The Units screen allows the user to enter the units for the selected analog
input channel. The units are displayed on the diagnostic screen and in
datalogging and streaming data. The units may be from 1 to 3 characters in
length, and defaults to V (volts).
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Input Config > select Channel > Units.
Decimal Places The Decimal Places screen allows the user to select how many digits are
displayed to the right of the decimal, from 0 to 6, with a default of 2.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Input Config > Select Channel > Decimal Places.
DECIMAL PLACES:
CURRENTLY: 2
SET TO: 3 ?
CHANGE VALUE
SAVE VALUE
Number of Table Points The Number of Table Points screen allows the user to select how many
points are used in the analog input conversion table for the selected
channel. The instrument uses linear interpolation between the points in
this table to determine what the reading value is based on the analog input
voltage. Each point in the table consists of an analog input voltage value
(010.5 V) and a corresponding reading value. Only two points are
necessary for linear inputs, however a larger number of points may be used
to approximate non-linear inputs. The points range from 2 to 10, with a
default of 2.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Input Config > select Channel > Table Points.
CHANGE VALUE
SAVE VALUE
Table Point The Table Point submenu allows the user to set up an individual table
point.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Input Config > Select Channel > Point 1-10.
Volts The Volts screen allows the user to set the input voltage for the selected
table point in the conversion table, from 0.00 to 10.50. The default table is
a two-point table with point 1: 0.00 V = 000.0 U and point 2: 10.00 V =
10.0 U, where U is the previously entered unit of measure.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Input Config > select Channel > select Point > Volts.
MOVE CURSOR
CHANGE VALUE SAVE
User Value The User Value screen allows the user to set the output value for the
corresponding input voltage for the selected table point in the conversion
table, from -9999999 to 99999999. The default table is a two-point table
with point 1: 0.00 V = 000.0 U and point 2: 10.00 V = 10.0 U, where U is
the previously entered unit of measure.
In the Main Menu, choose Instrument Controls > I/O Configuration >
Analog Input Config > select Channel > select Point > User Value.
MOVE CURSOR
CHANGE VALUE SAVE
Temperature The Temperature Compensation screen allows the user to turn temperature
Compensation compensation on or off. Temperature compensation changes to the
instrument's output signal that occur due to internal instrument
temperature variations. The effects of temperature changes on the analyzer's
subsystems and output have been empirically determined. This empirical
data is used to compensate for any changes in temperature. This
compensation can be used for special applications, or when operating the
instrument outside the recommended temperature range, even though the
Model 42i does not require temperature compensation for EPA
equivalency.
When temperature compensation is on, the display shows the current
internal instrument temperature (measured by a thermistor on the Interface
board). When temperature compensation is off, the display shows the
factory standard temperature of 30 C.
In the Main Menu, choose Instrument Controls > Temperature
Compensation.
TEMPERATURE COMPENSATION:
COMP TEMP: 30.0 oC
CURRENTLY: OFF
SET TO: ON ?
TOGGLE VALUE
PRESSURE COMPENSATION:
COMP PRES: 200.0 mmHg
CURRENTLY: OFF
SET TO: ON ?
TOGGLE VALUE
Screen Contrast The Screen Contrast screen is used to change the contrast of the display.
Intensities between 0 and 100% in increments of 5 are available. Changing
the screen contrast may be necessary if the instrument is operated at
extreme temperatures.
Note The optimal contrast will change from one LCD to another. If the
LCD is replaced, the contrast may need to be reset.
Note If the display contrast is not optimal, but the content on the screen is
visible, select Instrument Controls > Screen Contrast and adjust the screen
contrast. If the content on the screen is not visible, use the set contrast 10
C-Link command to set the screen contrast to mid-range, then optimize
the contrast. See contrast levels in Appendix B, C-Link Protocol
Commands for more information on this command.
SCREEN CONTRAST:
CURRENTLY: 50 %
SET TO: 60 % ?
CHANGE VALUE
SAVE VALUE
Service Mode The Service Mode screen is used to turn the service mode on or off.
Turning service mode ON locks out any remote actions and allows access
to parameters and functions that are useful when making adjustments or
diagnosing the Model 15i. The service (wrench) icon on the status bar is
shown when service mode is on. For more information about the service
mode, see Service Menu later in this chapter.
Note The service mode should be turned off when finished, as it prevents
remote operation.
SERVICE MODE:
CURRENTLY: OFF
SET TO: ON ?
TOGGLE VALUE
Date/Time The Date/Time screen allows the user to view and change the system date
and time (24-hour format). The internal clock is powered by its own
battery when instrument power is off.
In the Main Menu, choose Instrument Controls > Date/Time.
Timezone The Timezone screen is used to set the timezone for the Network Time
Protocol (NTP) server. This should be set to the timezone that the
instrument is located in. If the exact timezone is not shown in the list, it
may be entered via the C-LINK tz command (see Appendix B). The
selections are: UTC (GMT), EST (GMT+5), CST (GMT+6), MST
(GMT+7), PST (GMT+8), YST (GMT+9), HST (GMT+10), NST
(GMT+11), DLW (GMT+12), CET (GMT-1), EET (GMT-2), BST
(GMT-3), DLT (GMT-4), ECH (GMT-5), FOX (GMT-6), GLF (GMT-
7), CCT (GMT-8), JST (GMT-9), GST (GMT-10), LMA (GMT-11),
DLE (GMT-12), EDT (GMT+5/4), CDT (GMT+6/5), MDT
(GMT+7/6), and PDT (GMT+8/7).
Note The current timezone may say NULL before the timezone is set for
the first time, or if the timezone was cleared with a C-Link command.
CHANGE VALUE
SAVE VALUE
Diagnostics Menu The Diagnostics menu provides access to diagnostic information and
functions. This menu is useful when troubleshooting the instrument. The
analog input readings and analog input voltages are only displayed if the
I/O expansion board option is installed.
In the Main Menu, choose Diagnostics.
DIAGNOSTICS:
>PROGRAM VERSION
VOLTAGES
TEMPERATURES
PRESSURE
FLOW
Program Version The Program Version screen (read only) shows the version number of the
program installed. Prior to contacting the factory with any questions
regarding the instrument, please note the product model name and
program version number.
In the Main Menu, choose Diagnostics > Program Version.
PROGRAM VERSIONS:
PRODUCT: MODEL 42i
VERSION: 01.00.01.074
FIRMWARE: 09.06.19
Voltages The Voltages menu displays the current diagnostic voltage readings. This
screen enables the power supply to be quickly read for low or fluctuating
voltages without having to use a voltage meter. The I/O board item is only
displayed if the I/O expansion board option is installed.
In the Main Menu, choose Diagnostics > Voltages.
VOLTAGES:
>MOTHERBOARD
INTERFACE BOARD
I/O BOARD
Motherboard Voltages The Motherboard screen (read only) is used to display the current voltage
readings on the motherboard.
In the Main Menu, choose Diagnostics > Voltages > Motherboard
Voltages.
MOTHERBOARD VOLTAGES:
3.3 SUPPLY 3.3 V
5.0 SUPPLY 5.0 V
24.0 SUPPLY 24.1 V
-3.3 SUPPLY -3.3 V
Interface Board Voltages The Interface Board screen (read only) is used to display the current voltage
readings on the interface board.
In the Main Menu, choose Diagnostics > Voltages > Interface Board
Voltages.
I/O Board Voltages The I/O Board screen (read only) is used to display the current voltage
readings on the I/O expansion board. This menu is only displayed if the
I/O expansion board option is installed.
In the Main Menu, choose Diagnostics > Voltages > I/O Board
Voltages.
Temperatures The Temperatures screen (read only) displays the internal temperature,
reaction chamber temperature, cooler temperature, converter temperatures,
and perm oven temperatures (if the permeation oven option is installed).
The internal temperature is the air temperature measured by a sensor
located on the interface board.
In the Main Menu, choose Diagnostics > Temperatures.
TEMPERATURES:
o
INTERNAL 28.6 C
o
CHAMBER 49.0 C
o
COOLER -10.0 C
o
NO2 CONVERTER 320.7 C
o
NO2 CONV SET 325.0 C
Pressure The Pressure screen (read only) displays the reaction chamber pressure. The
pressure is measured by a pressure transducer at the reaction chamber.
In the Main Menu, choose Diagnostics > Pressure.
Flow The Flow screen (read only) displays the sample and ozonator flow rate.
The flows are measured by internal flow sensors. For more information, see
Figure 11.
In the Main Menu, choose Diagnostics > Flow.
OZONATOR FLOW OK
Analog Input The Analog Input Readings screen (read only) displays the 8 user-scaled
Readings analog readings (if the I/O expansion board option is installed).
In the Main Menu, choose Diagnostics > Analog Input Readings.
Analog Input The Analog Input Voltages screen (read only) displays the 8 raw analog
Voltages voltage readings (if the I/O expansion board option is installed).
In the Main Menu, choose Diagnostics > Analog Input Voltages.
Digital Inputs The Digital Inputs screen (read only) displays the state of the 16 digital
inputs. Pull-ups are provided on all of the inputs, so if nothing is connected
they will read (1), if an input is brought to ground, it will read (0).
In the Main Menu, choose Diagnostics > Digital Inputs.
DIGITAL INPUTS:
>INPUT 1 1
INPUT 2 1
INPUT 3 1
INPUT 4 1
INPUT 5 1
Relay States The Relay States screen displays the state of the 10 digital outputs and
allows toggling of the state to either on (1) or off (0). The relays are
restored to their original states upon exiting this screen.
In the Main Menu, choose Diagnostics > Relay States.
Press to toggle and set the relay state to open or closed for the
selected digital output.
RELAY STATE:
>OUTPUT 1 0
OUTPUT 2 0
OUTPUT 3 0
OUTPUT 4 1
OUTPUT 5 1
Test Analog Outputs The Test Analog Outputs menu is used to set each of the analog output
channels to zero or full-scale. Channel choices include all analog outputs, 6
voltage channels, and 6 current channels (if the I/O expansion board
option is installed).
In the Main Menu, choose Diagnostics > Test Analog Outputs.
Set Analog Outputs The Set Analog Outputs screen contains three choices: Set to full-scale, set
to zero, or reset to normal. Full-scale sets the analog outputs to the full-
scale voltage, zero sets the analog outputs to 0 volts, and reset returns the
output to normal operation. The analog outputs are returned to normal
operation upon exiting this screen. The following example shows the screen
when all analog outputs are set to normal operating mode.
In the Main Menu, choose Diagnostics > Test Analog Outputs > ALL,
Voltage Channel 16, or Current Channel 16.
INSTRUMENT CONFIGURATION:
>I/O EXPANSION BOARD YES
ZERO/SPAN VALVES YES
PERM OVEN YES
LAG VOLUME NO
PERM DRYER NO
CONVERTER MOLY
DILUTION NO
AUTO CALIBRATION NO
OUTPUT CONC CAL DATA NO
Contact Information The Contact Information screen displays the customer service information.
In the Main Menu, choose Diagnostics > Contact Information.
CONTACT INFORMATION:
WEB: WWW.THERMO.COM
Alarms Menu The Alarms menu displays a list of items that are monitored by the
analyzer. If the item being monitored goes outside the lower or upper limit,
the status of that item will go from OK to either LOW or HIGH,
respectively. If the alarm is not a level alarm, the status will go from OK
to FAIL. The total number of alarms detected is displayed to indicate
how many alarms have occurred.
To see the actual reading of an item and its minimum and maximum
limits, move the cursor to the item and press .
ALARMS:
ALARMS DETECTED 0
>INTERNAL TEMP OK
CHAMBER TEMP OK
COOLER TEMP OK
CONVERTER TEMP OK
Internal The Internal Temperature screen displays the current internal temperature
Temperature and sets the minimum and maximum alarm limits. Acceptable alarm limits
range from 8 to 47 C. If the internal temperature reading goes beyond
either the minimum or maximum limit, an alarm is activated and the alarm
(bell) icon appears in the status bar on the Run screen and in the Main
Menu.
In the Main Menu, choose Alarms > Internal Temp.
INTERNAL TEMPERATURE:
o
ACTUAL 28.6 C
o
>MIN 15.0 C
o
MAX 45.0 C
Min and Max Internal The Minimum Internal Temperature alarm limit screen is used to change
Temperature Limits the minimum internal temperature alarm limit. The minimum and
maximum internal temperature screens function the same way.
In the Main Menu, choose Alarms > Internal Temp > Min or Max.
INTERNAL TEMPERATURE:
o
ACTUAL MIN: 15.0 C
o
SET MIN TO: 16.0 C ?
INC/DEC
SAVE VALUE
CHAMBER TEMPERATURE:
o
ACTUAL 49.0 C
o
>MIN 48.0 C
o
MAX 52.0 C
Min and Max Chamber The Minimum Chamber Temperature alarm limit screen is used to change
Temperature Limits the minimum chamber temperature alarm limit. The minimum and
maximum chamber temperature screens function the same way.
In the Main Menu, choose Alarms > Chamber Temp > Min or Max.
CHAMBER TEMPERATURE:
o
ACTUAL MIN: 48.0 C
o
SET MIN TO: 49.0 C ?
INC/DEC
SAVE VALUE
Cooler Temperature The Cooler Temperature screen displays the current cooler temperature
and sets the minimum and maximum alarm limits. Acceptable alarm limits
range from -40 to 10 C. If the cooler temperature reading goes beyond
either the minimum or maximum limit, an alarm is activated and the alarm
(bell) icon appears in the status bar on the Run screen and in the Main
Menu.
In the Main Menu, choose Alarms > Cooler Temp.
COOLER TEMPERATURE:
o
ACTUAL -9.8 C
o
>MIN -20.0 C
o
MAX -1.0 C
Min and Max Cooler The Minimum Cooler Temperature alarm limit screen is used to change
Temperature Limits the minimum cooler temperature alarm limit. The minimum and
maximum cooler temperature screens function the same way.
In the Main Menu, choose Alarms > Cooler Temp > Min or Max.
COOLER TEMPERATURE:
o
ACTUAL MIN -20.0 C
o
SET MIN TO: -10.0 C ?
INC/DEC
SAVE VALUE
CONVERTER TEMPERATURE:
ACTUAL 320.7 oC
>MIN 300.0 oC
MAX 350.0 oC
Min and Max Converter The Minimum Converter Temperature alarm limit screen is used to change
Temperature Limits the minimum converter temperature alarm limit. The minimum and
maximum converter temperature screens function the same way.
In the Main Menu, choose Alarms > Converter Temp > Min or Max.
CONVERTER TEMPERATURE:
ACTUAL MIN 300.0 oC
SET MIN TO: 310.0 oC ?
INC/DEC
SAVE VALUE
Permeation Oven The Permeation Oven Gas Temperature screen displays the current
Gas Temperature permeation oven gas temperature and sets the minimum and maximum
alarm limits. Acceptable alarm limits range from 25.00 to 50.00 C. If the
Min and Max Permeation Oven The Minimum Permeation Oven Gas Temperature alarm limit screen is
Temperature Limits used to change the minimum temperature alarm limit. The minimum and
maximum permeation gas temperature screens function the same way.
In the Main Menu, choose Alarms > Perm Gas Temp > Min or Max.
INC/DEC
SAVE VALUE
Pressure The Pressure screen displays the current reaction chamber pressure reading
and sets the minimum and maximum alarm limits. Acceptable alarm limits
range from 150 to 300 mmHg. If the pressure reading goes beyond either
the minimum or maximum limit, an alarm is activated and the alarm (bell)
icon appears in the status bar on the Run screen and in the Main Menu.
In the Main Menu, choose Alarms > Pressure.
PRESSURE:
ACTUAL 232.5 mmHg
>MIN 150.0 mmHg
MAX 300.0 mmHg
Min and Max Pressure The Minimum Pressure alarm limit screen is used to change the minimum
Limits temperature alarm limit. The minimum and maximum pressure screens
function the same way.
In the Main Menu, choose Alarms > Pressure > Min or Max.
PRESSURE:
ACTUAL MIN: 200.0 mmHg
SET MIN TO: 210.0 mmHg?
INC/DEC
SAVE VALUE
Flow The Flow screen displays the current sample flow reading and sets the
minimum and maximum alarm limits. Acceptable alarm limits range from
0 to 1 LPM. If the sample flow reading goes beyond either the minimum
or maximum limit, an alarm is activated and the alarm (bell) icon appears
in the status bar on the Run screen and in the Main Menu.
In the Main Menu, choose Alarms > Flow.
FLOW:
ACTUAL 0.531 LPM
>MIN 0.750 LPM
MAX 1.000 LPM
Min and Max Flow Limits The Minimum Flow alarm limit screen is used to change the minimum
sample flow alarm limit. The minimum and maximum flow screens
function the same way.
In the Main Menu, choose Alarms > Flow > Min or Max.
FLOW:
ACTUAL MIN: 0.750 LPM
SET MIN TO: 0.760 LPM ?
INC/DEC
SAVE VALUE
Ozonator Flow The Ozonator Flow screen (read only) is used to display the ozonator flow
readings. If the ozonator flow reading is 0.050 LPM (50 cc) or below, an
alarm is activated, and an alarm condition screen appears as LOW. If the
ozonator flow is above 0.050, the no alarm condition screen is displayed,
indicating that the flow is acceptable. Inadequate ozonator flow will cause
the ozonator to overheat, resulting in permanent damage to the ozonator.
In the Main Menu, choose Alarms > Ozonator Flow.
OZONATOR FLOW:
ACTUAL > 0.050 LPM
Zero and Span The Zero and Span Check screens allow the user to view the status of the
Check most recent zero/span checks and set the maximum check offsets. An alarm
will be triggered if a zero or span check indicates drift that exceeds the
offset value. The zero and span check screens are visible only if the
zero/span check option is enabled and function the same way.
In the Main Menu, choose Alarms > Zero or Span Check.
ZERO CHECK:
ALARM: OK
RESPONSE: 0.000
>MAX OFFSET: 10.000
Max Zero and Span The Max Zero and Span Check Offset screens are used to change the
Check Offset maximum check offsets. The maximum zero and span offset screens
function the same way.
In the Main Menu, choose Alarms > Zero or Span Check > Max
Offset.
MOVE CURSOR
CHANGE VALUE SAVE
Zero and Span Auto The Zero and Span Auto Calibration screens (read only) allow the user to
Calibration view the status of the most recent auto background calibration or span
calibrations. The zero and span auto calibration screens are visible only if
the auto calibration option is enabled and the zero or span cal reset
function is enabled.
In the Main Menu, choose Alarms > Zero or Span Autocal.
NO, NO2, and NOx The NO Concentration screen displays the current NO concentration and
Concentration sets the minimum and maximum alarm limits. Acceptable alarm limits
range from 0 to 100,000 ppb (0 to 100 ppm) and 0 to 150,000 g/m3 (0 to
150 mg/m3). The minimum alarm may be programmed as a floor trigger
(alarm is triggered when the concentration falls below the minimum value)
or a ceiling trigger (alarm is triggered when the concentration goes above
the minimum value). If the NO concentration goes beyond either the
minimum or maximum limit, an alarm is activated and the alarm (bell)
icon appears in the status bar on the Run screen and in the Main Menu.
The NO2 and NOx concentration screens function the same way.
In the Main Menu, choose Alarms > NO, NO2, or NOx
Concentration.
NO CONCENTRATION:
ACTUAL 62.7
>MIN 0.0
MAX 100000.0
MIN TRIGGER CEILING
Min and Max NO, NO2, and The Minimum NO Concentration alarm limit screen is used to change the
NOx Concentration Limits minimum NO concentration alarm limits. The minimum and maximum
NO, NO2, and NOx concentration alarm limit screens function the same
way.
In the Main Menu, choose Alarms > Select Concentration > Min or
Max.
NO CONCENTRATION:
ACTUAL MIN: 0.0
SET MIN TO: 00020.00 ?
MOVE CURSOR
CHANGE VALUE SAVE
Min Trigger The Minimum Trigger screen allows the user to keep set the minimum
NO, NO2, and NOx concentration alarm trigger type to either floor or
ceiling. The minimum alarm may be programmed as a floor trigger (alarm
is triggered when the concentration falls below the minimum value) or a
ceiling trigger (alarm is triggered when the concentration goes above the
minimum value).
In the Main Menu, choose Alarms > Select Concentration > Min
Trigger.
MIN TRIG(CEILING/FLOOR):
ACTUAL TRIGGER: CEILING
SET TRIGGER TO: FLOOR ?
External Alarms The External Alarms menu is used to display the external alarm status for
alarm 1, 2, and 3 only if the external alarm is assigne as a digital input.
External alarms can be assigned to digital inputs using the Instruments
Controls > I/O Configuration > Digital Input Settings menu. External
alarms can also be logged, streamed, or output as a digital output by
choosing EXT ALARMS from the Other Measurements list.
In the Main Menu, choose Alarms > External Alarms.
ALARMS:
EXTERNAL ALARM 1 OK
EXTERNAL ALARM 2 OK
EXTERNAL ALARM 3 OK
Service Menu The Service menu appears only when the instrument is in the service mode.
When the service mode is active, the service (wrench) icon is displayed on
the right side of the status bar. To put the instrument into the service
mode:
In the Main Menu, choose Instrument Controls > Service Mode.
Advanced diagnostic functions are included in the service mode.
Meaningful data should not be collected when the instrument is in the
service mode.
In the Main Menu, choose Service.
SERVICE:
>PMT VOLTAGE ADJUSTMENT
RANGE MODE SELECT
CONVERTER SET TEMP
PRESSURE CALIBRATION
FLOW CALIBRATION
PMT Voltage Adjustment The PMT Voltage Adjustment screen is used to manually adjust the PMT
supply voltage. The PMT voltage adjustment screen is visible only when
the instrument is in service mode. For more information on the service
mode, see Service Mode earlier in this chapter.
CHANGE VALUE
SAVE VALUE
Range Mode Select The Range Mode Select screen is used to switch between the various range
modes: Single, Dual, and Auto Range. The range mode select screen is
visible only when the instrument is in service mode. For more information
on the service mode, see Service Mode earlier in this chapter.
In the Main Menu, choose Service > Range Mode Select.
CHANGE VALUE
SAVE VALUE
Converter Set Temperature The Converter Set Temperature screen is used to change the converter set
temperature. The converter set temperature reading is updated every
second. The converter set temperature screen is visible only when the
instrument is in service mode. For more information on the service mode,
see Service Mode earlier in the chapter.
INC/DEC
SAVE VALUE
Pressure Calibration The Pressure Calibration submenu is used to calibrate the pressure sensor
to zero, span, or restore factory default values. The pressure calibration is
visible only when the instrument is in service mode. For more information
on the service mode, see Service Mode earlier in this chapter.
The pressure sensors zero counts and span slope are displayed on the
menu.
Calibrate Pressure Zero The Calibrate Pressure Zero screen calibrates the pressure sensor at zero
pressure.
Note Wait at least 30 seconds for the reading to stabilize before saving the
value.
In the Main Menu, choose Service > Pressure Calibration > Zero.
Calibrate Pressure Span The Calibrate Pressure Span screen allows the user to view and set the
pressure sensor calibration span point.
Note Wait at least 30 seconds for the reading to stabilize before saving the
value.
In the Main Menu, choose Service > Pressure Calibration > Span.
MOVE CURSOR
CHANGE VALUE SAVE
Restore Default Pressure The Restore Default Pressure Calibration screen allows the user to reset the
Calibration pressure calibration configuration values to factory defaults.
In the Main Menu, choose Service > Pressure Calibration > Set
Defaults.
RESTORE RESTORE
ARE YOU SURE YOU WANT TO?x
PRESS TO CONFIRM RESTORE
Flow Calibration The Flow Calibration submenu is used to calibrate the flow sensor to zero,
span, or restore factory default values. The flow calibration screen is visible
only when the instrument is in service mode. For more information on the
service mode, see Service Mode earlier in the chapter.
Calibrate Flow Zero The Calibrate Flow Zero screen calibrates the flow sensor at zero flow.
Note Wait at least 30 seconds for the reading to stabilize before saving the
value.
In the Main Menu, choose Service > Flow Calibration > Zero.
Calibrate Flow Span The Calibrate Flow Span screen allows the user to view and set the flow
sensor calibration span point.
Note An independent flow sensor is required to read the flow, then the
operator enters the flow value on this screen to perform the calibration.
Note Wait at least 30 seconds for the reading to stabilize before saving the
value.
In the Main Menu, choose Service > Flow Calibration > Span.
MOVE CURSOR
CHANGE VALUE SAVE
Restore Default Flow The Restore Default Flow Calibration screen allows the user to reset the
Calibration flow calibration configuration values to factory defaults.
In the Main Menu, choose Service > Flow Calibration > Set Defaults.
RESTORE RESTORE
ARE YOU SURE YOU WANT TO?x
PRESS TO CONFIRM RESTORE
Input Board Calibration The Input Board Calibration menu is used to initiate a calibration of the
input board A/D stages. The input board calibration menu is visible only
when the instrument is in service mode. For more information on the
service mode, see Service Mode earlier in the chapter.
Manual Input Calibration The Manual Input Calibration screen is used to do a manual calibration of
the input board A/D stages per the following procedure:
Note The measurement system and the PMT are both shut off inside this
screen.
1. In the Main Menu, choose Service > Input Board Calibration >
Manual Input Cal.
Automatic Input Calibration The Automatic Input Calibration screen is used to do an automatic
calibration of the input board A/D stages. A message will be displayed after
the optimum setting has been determined.
Note The measurement system and the PMT are both shut off inside this
screen.
In the Main Menu, choose Service > Input Board Calibration >
Automatic Input Cal.
Press to leave warning screen and begin automatic calibration.
Input Frequency Display The Input Frequency Display screen is used to manually adjust the input
board gain. This may be used as a troubleshooting tool to ensure that the
input board is working properly. The gain setting and test mode are reset
upon exiting this screen. (That is, no changes are saved.)
In the Main Menu, choose Service > Input Board Calibration > Input
Frequency Disp.
Press to leave warning screen.
Use and to toggle the test signal (between ON and OFF).
Use and to change the gain between 1, 10 and 100.
Temperature Calibration The Temperature Calibration screen allows the user to view and set the
ambient temperature sensor offset, which basically adjusts the temperature
sensor so that it reads correctly. The temperature calibration is visible only
when the instrument is in service mode. For more information on the
service mode, see Service Mode earlier in the chapter.
Note Wait at least 30 seconds for the reading to stabilize before saving the
value.
MOVE CURSOR
CHANGE VALUE SAVE
Analog Output Calibration The Analog Output Calibration menu provides access to the 6 voltage
channels and 6 current channels for calibration. Current channels are
visible only if the I/O expansion board is installed. The analog output
calibration menu is visible only when the instrument is in service mode.
For more information on the service mode, see Service Mode earlier in
the chapter.
Analog Output Calibrate The Analog Output Calibrate Zero screen allows the user to calibrate the
Zero zero state of the selected analog output. The operator must connect a volt
meter to the output and adjust the output until it reads 0.0 V on the meter
for a voltage channel, or either 0 or 4 mA for a current channel (depending
on the selected range). See the set output to: field on the display.
In the Main Menu, choose Service > Analog Out Cal > select Channel
> Calibrate Zero.
Analog Output Calibrate The Analog Output Calibrate Full-Scale screen allows the user to calibrate
Full-Scale the full-scale state of the selected analog output. The operator must connect
a volt meter to the output and adjust the output until the meter reads the
value shown in the set output to: field, in either V or mA, depending on
the selected output channel.
In the Main Menu, choose Service > Analog Out Cal > select Channel
> Calibrate Full Scale.
Analog Input Calibration The Analog Input Calibration menu is used to calibrate the 8 analog input
channels at both zero and full-scale. The Analog Input Calibration menu is
visible only when the I/O expansion board is installed. The analog input
calibration menu is visible only when the instrument is in service mode.
For more information on the service mode, see Service Mode earlier in
the chapter.
Analog Input Calibrate The Analog Input Calibrate Zero screen allows the user to calibrate the
Zero zero state of the selected analog input.
In the Main Menu, choose Service > Analog Input Cal > select Channel
> Calibrate Zero.
ANALOG INPUT CAL: ZERO
DISCONNECT SELECTED INPUT!
SELECTED INPUT: INPUT 1
CURRENTLY: 0.04 V
Analog Input Calibrate The Analog Input Calibrate Full-Scale screen allows the user to calibrate
Full-Scale the full-scale state of the selected analog input.
In the Main Menu, choose Service > Analog Input Cal > select Channel
> Calibrate Full Scale. (Hook up a voltage source of 10 V to the
analog input channel.)
Permeation Oven Settings The Permeation Oven Settings menu is used for setting up and calibrating
the permeation oven option. The permeation oven settings are visible only
if the permeation oven option installed and when the calibrator is in service
mode. For more information on the service mode, see Service Mode
earlier in this chapter.
In the Main Menu, choose Service > Perm Oven Settings.
Calibrate Gas Thermistor The Calibrate Gas Thermistor menu is used for calibrating the permeation
oven gas thermistor using a water bath or known resistance.
In the Main Menu, choose Service > Perm Oven Settings > Cal Gas
Thermistor.
Water Bath The Calibrate Gas Thermistor Water Bath screen is used to view and set
the permeation oven gas temperature to a known temperature value.
In the Main Menu, choose Service > Perm Oven Settings > Cal Gas
Thermistor > Water Bath.
MOVE CURSOR
CHANGE VALUE SAVE
Resistor The Calibrate Gas Thermistor Resistor screen is used to view and set the
permeation oven gas thermistor resistance to a known resistor value.
In the Main Menu, choose Service > Perm Oven Settings > Cal Gas
Thermistor > Known Resistor.
MOVE CURSOR
CHANGE VALUE SAVE
Calibrate Oven Thermistor The Calibrate Oven Thermistor screen is used to view and set the
permeation oven thermistor resistance to a known resistor value.
In the Main Menu, choose Service > Perm Oven Settings > Cal Oven
Thermistor.
MOVE CURSOR
CHANGE VALUE SAVE
Permeation Oven Selection The Permeation Oven Setpoint screen is used to select the permeation oven
state to not present, or the desired temperature choice of 30, 35, and 45 C.
In the Main Menu, choose Service > Perm Oven Settings > Perm Oven
Selection.
MOVE CURSOR
CHANGE VALUE SAVE
Factory Calibrate Gas Thermistor The Factory Calibrate Gas Thermistor menu is used for calibrating the
permeation oven gas thermistor to either low point, high point, or set
defaults.
In the Main Menu, choose Service > Perm Oven Settings > Factory
Cal Gas Therm.
Low and High Points The Calibrate Gas Thermistor Low Point screen is used to view and set the
permeation oven thermistor resistance. The low and high point screens
function the same way.
In the Main Menu, choose Service > Perm Oven Settings > Factory Cal
Gas Therm > Low or High Point.
MOVE CURSOR
CHANGE VALUE SAVE
Set Defaults The Set Defaults screen is used to reset the configuration values to factory
defaults.
In the Main Menu, choose Service > Perm Oven Settings > Factory Cal
Gas Therm > Set Defaults.
Factory Calibrate Oven The Factory Calibrate Oven Thermistor menu is used for calibrating the
Thermistor permeation oven heater thermistor to either low point, high point, or set
defaults.
In the Main Menu, choose Service > Perm Oven Settings > Factory
Cal Oven Therm.
Low and High Points The Calibrate Oven Thermistor Low Point screen is used to view and set
the permeation oven thermistor resistance. The low and high point screens
function the same way.
In the Main Menu, choose Service > Perm Oven Settings > Factory Cal
Oven Therm > Low or High Point.
MOVE CURSOR
CHANGE VALUE SAVE
Set Defaults The Set Defaults screen is used to reset the configuration values to factory
defaults.
In the Main Menu, choose Service > Perm Oven Settings > Factory Cal
Oven Therm > Set Defaults.
Ozonator Safety The Ozonator Safety screen is used to turn the ozonator safety feature on
or off. If the ozonator safety is turned off, the ozonator will always be on,
even if the converter is not up to temperature. The ozonator shuts of if
there is a converter temperature alarm or an ozone flow alarm. If the safety
override is on, the instrument generates an ozone flow alarm.
The ozonator safety screen is visible only when the instrument is in service
mode. For more information on the service mode, see Service Mode
earlier in the chapter.
In the Main Menu, choose Service > Ozonator Safety.
OZONATOR SAFETY:
CURRENTLY: OFF
SET TO: ON ?
TOGGLE VALUE
Extended Ranges The Extended Ranges screen is used to turn the extended ranges feature on
and off. The extended ranges screen is visible only when the instrument is
in service mode. For more information on the service mode, see Service
Mode earlier in the chapter.
In the Main Menu, choose Service > Extended Ranges.
Press to toggle and set the extended ranges on or off.
EXTENDED RANGES:
CURRENTLY: OFF
SET TO: ON ?
TOGGLE VALUE
Dilution Ratio The Dilution Ratio screen allows the user to view and set the dilution ratio.
Acceptable values are 1500: 1. The default is 1:1. When this value is set,
the dilution ratio is applied to all concentration measurements. This screen
is only accessible if the dilution ratio option is installed.
In the Main Menu, choose Service > Dilution Ratio.
DILUTION RATIO:
CURRENTLY: 001.0 :1
SET TO: 002.0 :1 ?
MOVE CURSOR
CHANGE VALUE SAVE
Display Pixel Test The Display Pixel Test is used to test the LCD display. The display pixel
test is visible only when the instrument is in service mode. For more
information on the service mode, see Service Mode earlier in the chapter.
In the Main Menu, choose Service > Display Pixel Test.
BEGIN TEST
GO BACK TO MENU
Restore User Defaults The Restore User Defaults screen is used to reset the user calibration and
configuration values to factory defaults. The restore user defaults screen is
visible only when the instrument is in service mode. For more information
on the service mode, see Service Mode earlier in the chapter.
In the Main Menu, choose Service > Restore User Defaults.
RESTORE RESTORE
ARE YOU SURE YOU WANT TO?x
PRESS TO CONFIRM RESTORE
Password Menu The Password menu allows the user to configure password protection. If
the instrument is locked, none of the settings may be changed via the
front panel user interface, but they can still be changed via remote
operation. When the front panel is locked, the user can still navigate the
menu and view data, instrument parameters, and settings, but nothing can
be changed. The password (lock) icon on the right side of the status bar
indicates that the password lock is active. The items visible under the
Password menu are determined by the instruments password status.
In the Main Menu, choose Password.
PASSWORD MENU:
>SET PASSWORD
LOCK INSTRUMENT
CHANGE PASSWORD
REMOVE PASSWORD
UNLOCK INSTRUMENT
Set Password The Set Password screen is used to set the password that is used to unlock
the front panel, and is shown if the instrument is unlocked and the
password is not set.
In the Main Menu, choose Password > Set Password
ABCDEFGHIJKLMN BKSP
OPQRSTUVWXYZ PAGE
0123456789 ./- SAVE
Lock Instrument The Lock Instrument screen is used to lock the instruments front panel so
users cannot change any settings from the front panel. The lock instrument
screen is shown if the instrument is unlocked and the password is set.
In the Main Menu, choose Password > Lock Instrument
Lock/Unlock and If the instrument is locked via the front panel using Password > Lock
Local/Remote Operation Instrument, the instrument reports being in Remote mode. In this mode,
the front panel is locked, where data can be viewed, settings can be
viewed but not changed using the front panel interface, and the remote
Set commands are active.
If the instrument keyboard is unlocked via the front panel using Password
> Unlock Instrument, the instrument reports being in Local mode, the
front panel interface is unlocked, and settings can be changed from the
front panel.
Refer to C-Link Protocol Commands appendix for detailed information
about mode, allow mode, and power up mode commands.
Change Password The Change Password screen is used to change the password used to
unlock the instruments front panel. The change password screen is shown
if the instrument is unlocked.
In the Main Menu, choose Password > Change Password
ABCDEFGHIJKLMN BKSP
OPQRSTUVWXYZ PAGE
0123456789 ./- SAVE
Remove Password The Remove Password screen is used to erase the current password and
disable password protection. The remove password screen is shown if the
instrument is unlocked and the password is set.
In the Main Menu, choose Password > Remove Password
REMOVE PASSWORD:
XXXPRESSING ENTER WILLXXXX
XREMOVE CURRENT PASSWORDXX
XXXAND DISABLE LOCKINGXXXX
REMOVE PASSWORD
Unlock Instrument The Unlock Instrument screen is used to enter the password to unlock the
front panel. The unlock instrument screen is shown if the instrument is
locked.
In the Main Menu, choose Password > Unlock Instrument
ABCDEFGHIJKLMN BKSP
OPQRSTUVWXYZ PAGE
0123456789 ./- SAVE
NO + O 3 NO2 + O2
Zero Gas Generator A zero air source, such as a Thermo Scientific Model 111 Zero Air Supply or
Model 1160 Zero Air Supply, free of contaminants such as NO, NO2, and
O3 is required for dilution, calibration, and gas phase titration.
Compression The zero air source should be at an elevated pressure to allow accurate and
reproducible flow control and to aid in subsequent operations such as
drying, oxidation, and scrubbing. An air compressor that gives an output of
10 psig is usually sufficient for most applications.
Drying Several drying methods are available. Passing the compressed air through a
bed of silica gel, using a heatless air dryer, or removing water vapor with a
permeation dryer are three possible approaches.
Scrubbing Fixed bed reactors are commonly used in the last step of zero air generation
to remove the remaining contaminants by either further reaction or
absorption. Table 41 lists materials that can be effective in removing
contaminants.
Table 41. Scrubbing Materials
To Remove Use
NO2 Soda-Lime (6-12 mesh), Purafil
Hydrocarbons Molecular Sieve (4A), Activated Charcoal
O3 and SO2 Activated Charcoal
Gas Phase Titrator A gas phase titrator (GPT), such as is included in the Thermo Scientific
Model 146 Series Multigas Calibration System is used to generate NO2
concentrations from NO concentrations. Figure 41 shows the suggested
placement of the component parts of a gas phase titration apparatus.
Flow Controllers The airflow controllers should be devices capable of maintaining constant
airflows within 2% of the required flow rate. The NO flow controller
should be capable of maintaining constant NO flows within 2% of the
required flow rate.
Pressure Regulator The pressure regulator for the standard NO cylinder must have a non-
reactive diaphragm and internal parts, and a suitable delivery pressure.
Ozone Generator The ozone generator must be capable of generating sufficient and stable
levels of ozone for reaction with NO to generate NO2 concentrations in the
range required.
Note Ozone generators of the electric discharge type may produce NO and
NO2 and are not recommended.
Diverter Valve A valve can be used to divert the NO flow when zero air is required at the
manifold.
Reaction Chamber The reaction chamber used for the reaction of ozone with excess NO
should have sufficient volume so that the residence time meets the
requirements specified in this chapter.
Mixing Chamber The mixing chamber is used to provide thorough mixing of the reaction
products and diluent air.
NO Concentration Standard A cylinder containing 10 to 50 ppm NO in N2 with less than 1 ppm NO2
is usually used as the concentration standard. The cylinder must be
traceable to a National Institute of Standards and Technology (NIST) NO
in N2 Standard Reference Material or NO2 Standard Reference Material.
Procedures for certifying the NO cylinder (working standard) against an
NIST traceable NO or NO2 standard and for determining the amount of
NO2 impurity are given in EPA Publication No. EPA-600/4-75-003,
Technical Assistance Document for the Chemiluminescence Measurement
of Nitrogen Dioxide.
Assaying a Working NO Standard Use the following procedure to calculate the NO concentration standard
Against a NIST-traceable NO and NO2 impurity.
Standard
1. Generate several NO concentrations by dilution of the NO working
standard.
3. Plot the analyzer NO response (in ppm) versus the nominal diluted
NO concentration and determine the slope, SNOM.
7. Plot the analyzer NO response versus FNO/FT and determine the slope
which gives [NO]STD directly. The analyzer NOx responses to the
generated NO concentrations reflect any NO2 impurity in the NO
working standard.
8. Plot the difference between the analyzer NOx and NO responses versus
FNO/FT. The slope of this plot is [NO2]IMP.
Zero Air A source of zero air free of contaminants should be used as described earlier
in this chapter. Contaminants can cause a detectable response on the
instrument and may also react with the NO, O3, or NO2 during the gas
phase titration.
Dynamic Parameter Use the following definitions for the remainder of this chapter.
Specifications for PR = Dynamic parameter specification to ensure complete reaction of the
Gas Titrator available O3, ppm-min
[NO]RC = NO concentration in the reaction chamber, ppm
tR = residence time of the reactant gases in the reaction chamber, min
[NO]STD = Concentration of the undiluted NO standard, ppm
FNO = NO flow rate, sccm
FO = O3 generator air flow rate, sccm
VRC = Volume of the reaction chamber, cc
FT = Analyzer demand plus 10 to 50% excess
The O3 generator (ozonator) airflow rate and the NO flow rate must be
adjusted such that the following relationships hold:
Determining GPT System Use the following procedure to determine the flow conditions to be used in
Flow Conditions the GPT system.
1. Determine FT, the total flow required at the output manifold, which
should be equal to the analyzer demand plus 10 to 50 percent excess.
[NO ] OUT x FT
FNO =
[NO ] STD
5. Compute FO as:
6. Compute tR as:
VRC
tR =
FO + FNO
FD = FT - FO - FNO
Note The calibration and calibration check duration times should be long
enough to account for the transition (purge) process when switching from
sample to zero and from zero to span. This transition time is the time
required to purge the existing air.
3. Be sure the instrument is in the auto mode, that is, NO, NO2, and
NOx measurements are being displayed on the front panel display. If
the instrument is not in auto mode:
a. Press to display the Main Menu, then choose Instrument
Controls > Auto/Manual Mode.
b. Select NO/NOx, and press .
c. Press to return to the Run screen.
4. Select NO, NO2, and NOx ranges, and push the AVG soft key to
display the Averaging Time screen. It is recommended that a higher
averaging time be used for best results. For more information about the
ranges or averaging time, see the Operation chapter.
Note The averaging time should be less than the zero duration and less
than the span duration.
5. Verify that any filters used during normal monitoring are also used
during calibration.
Calibration The following procedure calibrates the analyzer using the gas phase titrator
and zero gas generator described previously in this manual. It is suggested
that a calibration curve have at least seven points between the zero and full
scale NO concentrations. Although the seven-point curve is optional, two
of whatever number of points is chosen should be located at the zero and
90% levels and the remaining points equally spaced between these values.
Note When the instrument is equipped with internal zero/span and sample
valves, the ZERO and SPAN ports should give identical responses to the
SAMPLE port when test gases are introduced. The user should calibrate the
instrument using the SAMPLE port to introduce the zero and span gas
sources.
After calibration, the zero and span sources should be plumbed to the
appropriate ports on the rear panel of the instrument, and then
reintroduced to the instrument. The instrument should give identical
responses to the test gases whether they are introduced via the SAMPLE
port or the ZERO or SPAN ports. If not, the plumbing and/or valves
should be serviced.
Connect GPT Apparatus Use the following procedure to connect the GPT apparatus to the analyzer.
to the Analyzer
1. Assemble a dynamic calibration system such as the one shown in Figure
41.
2. Ensure that all flow meters are calibrated under the conditions of use
against a reliable standard, such as a soap-bubble meter or wet-test
meter. All volumetric flow rates should be corrected to 25 C and 760
mmHg.
4. Connect the analyzer sample bulkhead input to the output of the GPT
system.
Adjust Instrument Gain Use the following procedure to adjust the instrument gain. This includes:
Setting the NO and NOx background to zero
Calibrating the NO channel to the NO calibration gas
Calibrating the NOx channel to the NOx calibration gas
Set NO and NOx Background The NO and NOx background corrections are determined during zero
to Zero calibration. The background signal is the combination of electrical offsets,
PMT dark current, and trace substances undergoing chemiluminescence.
For more detailed information, see Calibrate NO and NOx Backgrounds
in the Operation chapter.
Use the following procedure to set the NO background. Both the NO and
NOx background screens operate the same way, therefore, the following
procedure also applies to the NOx background screen.
Note The NO channel should be calibrated first and then calibrate the
NOx channel.
For detailed information about the menu parameters and the icons used in
these procedures, see the Operation chapter.
2. Adjust the GPT diluent air and O3 generator air flows to obtain the
flows determined in Dynamic Parameter Specifications for Gas Phase
Titrator earlier in this chapter. The total GPT airflow must exceed the
total demand of the analyzer. The Model 42i requires approximately
700 cc/min of sample flow, and a total GPT airflow of at least 1.5
liters/min is recommended.
a. Allow the analyzer to sample zero air until the NO, NOx, and NO2
responses stabilize.
b. After the responses have stabilized, from the Main Menu, choose
Calibration > Cal NO Background.
c. Press to set the NO reading to zero.
d. Press to return to the Calibration menu and repeat this
procedure to set the NOx background to zero.
e. Record the stable zero air responses as ZNO, ZNOX, and ZNO2
(recorder response, percent scale).
x
[NO ] OUT = FNO NOSTD
FNO + FO + FD
Where:
[NO]OUT = Diluted NO concentration at the output manifold, ppm
FNO = No flow
FO = Ozone flow
FD = Dilution flow
Calibrate the NO Channel to the Use the following procedure to calibrate the NO channel to the NO
NO Calibration Gas calibration gas.
1. Allow the analyzer to sample the NO calibration gas until the NO,
NO2, and NOx readings have stabilized.
2. When the responses stabilize, from the Main Menu, choose Calibration
> Cal NO Coefficient.
[NO ] OUT
Recorder Response (% scale) = x 100 + ZNO
URL
Where:
URL = Nominal upper range limit of the NO channel, ppm
Calibrate the NOx Channel to the Use the following procedure to calibrate the NOx channel to the NOx
NOx Calibration Gas calibration gas.
2. Verify that the NOx calibration gas concentration is the same as the
NO calibration gas concentration plus any known NO2 impurity.
The NOx line of the Calibrate NOx screen displays the current NOx
concentration. The SPAN CONC line of the display is where you enter
the NOx calibration gas concentration.
Use to move the cursor left and right and use
to increment and decrement the numeric character at the cursor.
3. Press to calculate and save the new NOx coefficient based on the
entered span concentration.
[ NO x ]OUT
Recorder Response (% scale) = x 100 + Z NOx
URL
Where:
URL = Nominal upper range limit of the NOx channel, ppm
Preparing NO, NOx, and Use the following procedures to prepare the NO, NOx, and NO2
NO2 Calibration Curves calibration curves.
9. From the Main Menu choose Calibration > Cal NO2 Coefficient.
The NO2 line of the Calibrate NO2 screen displays the current NO2
concentration. The SPAN CONC line of the display is where you enter
the NO2 calibration gas concentration.
10. Set the NO2 calibration gas concentration to reflect the sum of the
following: the NO2 concentration generated by GPT, ([NO]ORIG -
[NO]REM), and any NO2 impurity.
Use to move the cursor left and right and use
to increment and decrement the numeric character at the cursor.
Where:
[NO2]OUT = diluted NO2 concentration at the output manifold, ppm
11. Press to calculate and save the new NO2 coefficient based on the
entered span concentration.
The analyzer does a one point NO2 span coefficient calculation,
corrects the NO2 reading for converter inefficiency, and then adds the
corrected NO2 to the NO signal to give a corrected NOx signal.
If the analyzer calculates a NO2 span coefficient of less than 0.96, either
the entered NO2 concentration is incorrect, the converter is not being
heated to the proper temperature, the instrument needs servicing (leak
or imbalance), or the converter needs replacement or servicing. The
[ NO2 ]OUT
Recorder Response (% scale) = x 100 + ZNO2
URL
Where:
URL = Nominal upper range limit of the NO2 channel, ppm
12. Record the NO2 concentration and the analyzer's NO2 response.
13. Maintaining the same FNO, FO, and FD, adjust the ozone generator to
obtain several other concentrations of NO2 over the NO2 range (at least
five evenly spaced points across the remaining scale are suggested).
14. Record the stable responses and plot the analyzer's NO2 responses
versus the corresponding calculated (using the above equation for
[NO2]OUT) concentrations and draw or calculate the NO2 calibration
curve.
Note It is important that the curve be linear within 1% FS over the NO2
range. If the curve is nonlinear, the analyzer is not operating correctly,
(possible leak, or converter failure, etc.), and should be serviced. Assuming
the curve is linear, subsequent data should be reduced using this NO2
calibration curve response.
Using the Calibration Factors menu can change the calibration factors.
This is often useful in a troubleshooting situation. However, after the above
calibration procedure is completed, all subsequent data reduction depends
on the calibration parameters, remaining the same as during the initial
calibration.
Therefore, never change any calibration factor without first recording the
value so that after any troubleshooting procedure is completed, the initial
value can be re-entered thereby not altering the multipoint calibration.
Alternative Calibration Although it is recommended that a GPT system be used to calibrate the
Procedure Using NO2 analyzer, the procedure described in the Code of Federal Regulations, Title
40, Part 50, Appendix F using a NO2 permeation tube may be used as an
Permeation Tube alternative procedure for calibrating the instrument.
Calibration in Dual The dual/auto range calibration feature is used to calibrate the analyzer at
two different span levels (as opposed to a single span level in the standard
Range and Auto mode) generating a tailored multi-point calibration curve stored in the
Range Mode analyzer's memory. This feature may be used:
When widely different gas levels are being monitored, such as a factor
of 10 or greater apart
If precision and span levels are being introduced using separate tanks
If more than one multi-component cylinder is being used to calibrate
the instrument
Properly designed chemiluminescence analyzers are inherently linear over a
wide dynamic range; and under normal USEPA compliance situations this
feature is not required. Dual calibration may be used for span levels less
than a factor of 10 apart, however if this is done to correct for a significant
non-linearity, it may mask the problems causing the effect, such as, bad
calibration cylinder, leaks in sampling lines, or low ozonator output.
Use the following procedures to calibrate the analyzer in dual or auto range
mode.
Set NO and NOx Use the following procedure to set the NO and NOx background readings
Background Readings to to zero. Both the Set NO background and NOx background screens
operate the same way, and the following procedure also applies to the NO
Zero and NOx background screen.
For detailed information about the menu parameters and the icons used in
these procedures, see the Operation chapter.
2. Introduce zero air to the SAMPLE bulkhead and allow the analyzer to
sample zero air until the NO, NOx, and NO2 responses stabilize.
3. When the responses stabilize, from the Main Menu choose Calibration
> Calibrate NO Background.
1. Disconnect the source of zero air from the SAMPLE bulkhead. In its
place, connect a source of NO calibration gas of about 80% of the low
NO full-scale range.
2. Allow the analyzer to sample the low NO calibration gas until the NO,
NO2, and NOx readings stabilize.
3. When the responses are stabile, from the Main Menu choose
Calibration > Cal Lo NO Coefficient.
Calibrate Low NOx Use the following procedure to calibrate the NOx channel to the NOx
calibration gas.
2. Verify that the low NOx calibration gas concentration is the same as the
low NO calibration gas concentration plus any known NO2 impurity.
The Lo NOx field displays the current NOx concentration. The Lo
NOx Span Conc field is where you enter the low NOx calibration gas
concentration.
Use to move the cursor left and right and use
to increment and decrement the numeric character at the cursor.
3. Press to calculate and save the new low NOx coefficient based on
the entered span concentration.
Calibrate Low NO2 Use the following procedure to calibrate the NO2 channel to the NO2
calibration gas.
2. From the Main Menu choose Calibration > Cal Lo NO2 Coefficient.
The Lo NO2 field displays the current NO2 concentration. The Lo
NO2 Span Conc field is where you enter the Lo NO2 calibration gas
concentration.
3. Set the low NO2 calibration gas concentration to reflect the sum of the
NO2 concentration generated by GPT and any NO2 impurity.
Use to move the cursor left and right and use
to increment and decrement the numeric character at the cursor.
4. Press to calculate and save the new low NO2 coefficient based on
the entered span concentration.
Calibrate High NO Use the following procedure to calibrate the NO channel to the NO
calibration gas.
2. After the responses have stabilized, from the Main Menu, choose
Calibration > Cal Hi NO Coefficient.
The Hi NO field displays the current NO concentration. The Hi NO
Span Conc field is where you enter the high NO calibration gas
concentration.
Use to move the cursor left and right and use
to increment and decrement the numeric character at the cursor.
Calibrate High NOx Use the following procedure to calibrate the NOx channel to the NOx
calibration gas.
2. Verify that the high NOx calibration gas concentration is the same as
the high NO calibration gas concentration plus any known NO2
impurity.
Use to move the cursor left and right and use
to increment and decrement the numeric character at the cursor.
3. Press to calculate and save the new high NOx coefficient based
on the entered span concentration.
Calibrate High NO2 Use the following procedure to calibrate the NO2 channel to the NO2
calibration gas.
2. From the Main Menu, choose Calibration > Cal Hi NO2 Coefficient.
The Hi NO2 field displays the current NO2 concentration. The Hi
NO2 Span Conc field is where you enter the high NO2 calibration gas
concentration.
3. Set the high NO2 calibration gas concentration to reflect the sum of the
NO2 concentration generated by GPT and any NO2 impurity.
Use to move the cursor left and right and use
to increment and decrement the numeric character at the cursor.
4. Press to calculate and save the new high NO2 coefficient based
on the entered span concentration.
You can change the calibration factors by using the Calibration Factors
menu. This is often useful in a troubleshooting situation. However, after
the above calibration procedure is completed, all subsequent data reduction
depends on the calibration parameters remaining the same as during the
initial calibration.
Therefore never change any calibration factor without first recording the
value so that after any troubleshooting procedure is completed, the initial
value can be re-entered thereby not altering the multipoint calibration.
Zero and Span Check The analyzer requires initial and periodic calibration according to the
procedures outlined in this manual. Initially, the frequency of the
calibration procedure should be determined by the stability of the zero and
span checks, which may be run daily. You should generate a new
calibration curve when zero and span checks indicate a shift in instrument
gain of more than 10 percent from that determined during the most recent
multipoint calibration. You can adjust the frequency of calibration and
even zero and span checks appropriately as you gain confidence with the
instrument.
You should have a quality control plan where the frequency and the
number of points required for calibration can be modified on the basis of
calibration and zero and span check data collected over a period of time.
Note however, that the EPA requires a minimum of one multipoint
calibration per calendar quarter. Such a quality control program is essential
to ascertain the accuracy and reliability of the air quality data collected and
to alert the user if the accuracy or reliability of the data should become
unacceptable. A compilation of this kind might include items such as dates
of calibration, atmospheric conditions, calibration factors, and other
pertinent data.
Use the following procedure to perform a zero and span check.
5. When the calibration check has been completed, record the NO, NO2,
and NOx values.
Safety Precautions Read the safety precautions before beginning any procedures in this
chapter.
Replacement Parts See the Servicing chapter for a list of replacement parts.
Cleaning the Outside Clean the outside case using a damp cloth being careful not to damage the
labels on the case.
Case
Equipment Damage Do not use solvents or other cleaning products to
clean the outside case.
Ozonator Air Feed Use the following procedure to replace the ozonator air feed drying
column.
Drying Column
Replacement 1. Remove the drying column from the connector DRY AIR bulkhead on
the rear panel of the instrument.
Capillaries Inspection The capillaries normally only require inspection when instrument
performance indicates that there may be a flow problem.
and Replacement
Equipment Damage Some internal components can be damaged by small
amounts of static electricity. A properly ground antistatic wrist strap must
be worn while handling any internal component. For more information
about appropriate safety precautions, see the Servicing chapter.
Use the following procedure to inspect and replace the capillaries. This
procedure can be used to check any or all of the capillaries.
Cajon Fitting
Capillary
Capillary O-Ring
4. Remove the Cajon fitting(s) from the reaction chamber body using a
5/8-inch wrench being careful not to lose the ferrule or O-ring.
5. Remove the glass capillaries, ferrule, and O-ring. Inspect O-ring for
cuts or abrasion, and replace as necessary.
11. Connect the power cord and turn the instrument ON.
Thermoelectric Cooler Use the following procedure to inspect and clean the thermoelectric cooler
fins.
Fins Inspection and
Cleaning
Equipment Damage Some internal components can be damaged by small
amounts of static electricity. A properly ground antistatic wrist strap must
be worn while handling any internal component. For more information
about appropriate safety precautions, see the Servicing chapter.
4. Blow off the cooler fins using clean pressurized air. It may be more
convenient to vacuum the cooler fins. In either case, make sure that any
particulate accumulation between the fins has been removed.
Fan Filters Inspection Use the following procedure to inspect and clean the fan filters.
and Cleaning
1. Remove the two fan guards from the fans and remove the filters.
2. Flush the filters with warm water and let dry (a clean, oil-free purge will
help the drying process) or blow the filters clean with compressed air.
Filter
Fan Guard
Pump Rebuilding Use the following procedure to rebuild the pump Figure 53. To replace
the pump, see Pump Replacement in the Servicing chapter.
Equipment Required:
Pump Repair Kit
Allen Wrench, 3 mm
Wrench, 9/16-inch
Needlenose Pliers
1. Turn instrument off, unplug the power cord, and remove the cover.
4. Remove four screws from top plate using the 3 mm Allen wrench.
Flapper Valve
Bottom Plate
Clamping Disk
Diaphragm
8. Re-install the cover, reconnect the power cord, and turn the instrument
ON.
9. Check that the reaction chamber pressure reads between 150 and 250
mmHg.
Safety Precautions Read the safety precautions in the Preface and Servicing chapter before
performing any actions listed in this chapter.
Troubleshooting The troubleshooting guides presented in this chapter are designed to help
isolate and identify instrument problems.
Guides
Table 61 provides general troubleshooting information and indicates the
checks that you should perform if you experience an instrument problem.
Table 62 lists all the alarm messages you may see on the graphics display
and provides recommendations about how to resolve the alarm condition.
See Alarms Menu in the Operation chapter for detailed information.
Alarm NO, NO2, NOx Concentration has exceeded Check to insure range
Conc. range limit corresponds with expected
value. If not select proper
range.
Concentration low Check user-defined low set
point, set to zero.
Alarm - Motherboard Internal cables not connected Check that all internal cables
Status properly are connected properly.
Alarm - Interface Board is defective Recycle AC power to
Status instrument. If still alarming,
change board.
Alarm - I/O Exp Status
Board-Level Figure 61 and Figure 62 are board-level connection diagrams for the
common electronics and measurement system. These illustrations can be
Connection used along with the connector pin descriptions in Table 63 through Table
Diagrams 69 to troubleshoot board-level faults.
I/OEXPANSION
BOARD
EXPANSION I/O J1
25 Pin
TOMEASUREMENT SYSTEM
DIGITAL OUTPUT
BOARD
DIGITAL J2
37 Pin
OUTPUTS
11 Pin 8 Pin
J14 J3
J2
RS232/RS485 P1:A FRONT 14 Pin LCD
9 Pi n PANEL DISPLAY
J1 J4
RS232/RS485 J18 BOARD
9 Pin P1:B 34 Pin 3 Pin
J10
2 Pin
POWER_GOOD
J17
2 Pin
PJ1 PJ3 PJ2
3 Pin 3 Pin 3 Pin
24VDC
AC POWER SUPPLY
REAR PANEL AC 24VDC
AC INPUT 3 Cond P1
2 Pin
AC
FRON T PANEL
POWER SW PJ6
TRANSFORMER 3 Pin
(90VACor 240VAC
OPTIONS)
PMT HVPS
MJ10 PMT
7 Pin 2 Cond
MJ15 SAMPLEFLOW
5 Pin
MJ14 OZ FLOWSW
3 Pin
OJ1 OZONATOR
2 Pin
PJ7 COOLER
MEASUREMENT 4 Pin
INTERFACEBOARD
MJ9 AMBIENT TEMP. THERM.
2 Pin
Connector Pin The connector pin descriptions in Table 63 through Table 69 can be
used along with the board-level connection diagrams to troubleshoot
Descriptions board-level faults.
Motherboard Connector Pin Descriptions on page 6-8
Measurement Interface Board Connector Pin Descriptions on page 6-13
Front Panel Board Connector Pin Diagram on page 6-15
I/O Expansion Board (Optional) Connector Pin Descriptions on page 6-
17
Digital Output Board Connector Pin Descriptions on page 6-18
Input Board Connector Pin Descriptions on page 6-20
Temperature Control Board Connector Pin Descriptions on page 6-20
Safety Precautions Read the safety precautions before beginning any procedures in this
chapter.
Do not remove the LCD panel or frame from the LCD module.
Do not wipe the LCD polarizing plate with a dry cloth, it may easily
scratch the plate.
Do not place the LCD module near organic solvents or corrosive gases.
Firmware Updates The firmware can be updated by the user in the field via the serial port or
over the Ethernet. This includes both the main processor firmware and the
firmware in all low-level processors. Refer to the iPort manual for the
firmware update procedure.
Accessing the If the Service Menu is not displayed on the Main Menu, use the following
procedure to display it.
Service Mode
1. From the Main Menu, choose Instrument Controls > Service Mode.
The Service Mode screen appears.
Replacement Parts Table 71 lists the replacement parts for the Model 42i major
subassemblies. Refer to Figure 72 to identify the component location.
List
Table 71. Model 42i Replacement Parts
Part Number Description
100480-00 Front Panel Pushbutton Board
101491-01 Processor Board
100533-00 Motherboard
100539-00 Digital Output Board
100542-00 I/O Expansion Board (Optional)
102340-00 Front Panel Connector Board
Cable List Table 72 describes the Model 42i spare cables. See the Troubleshooting
chapter for associated connection diagrams and board connector pin
descriptions.
Table 72. Model 42i Cables
Part Number Description
101036-00 DC Power Supply 24V Output
101037-00 115VAC Supply to Interface Board
101048-00 RS-485/Data
101038-00 AC Power Switch to Motherboard
101364-00 DC Power Supply Status Monitor
101054-00 Motherboard to Front Panel Board
101035-00 DC Power Supply AC Input
101033-00 AC from Receptacle
101377-00 AC to Power Switch
101267-00 Fan Power Cable
101346-00 Temperature Control
101355-02 Signal Output Ribbon
101695-00 Permeation Oven
101050-00 Heater Power
101055-00 Main AC Receptacle Assembly
102057-00 AC to External Pump
External Device Table 73 lists the standard and optional cables and components used for
connecting external devices such as PCs and data loggers to an iSeries
Connection instrument.
Components Table 73. External Device Connection Components
Part Number Description
102562-00 Terminal Block and Cable Kit (DB25) (optional)
102556-00 Terminal Block and Cable Kit (DB37) (optional)
102645-00 Cable, DB37M to Open End Cable, Six Feet (optional)
Capillaries (2)
Pressure Transducer
Input Board
HVPS
Pump
Thermistor
Measurement
Interface Board
I/O Expansion
Board (Opt)
Power Supply
Removing the The measurement bench can be removed and the partition panel can be
lowered to improve access to connectors and components. Refer to the
Measurement Bench following steps when a procedure requires lowering the partition panel (see
and Lowering the Figure 73).
Partition Panel
SLIDE BENCH OUT
REMOVE SCREWS
Pass Through
Connectors
Remove Screw
Loosen
Captive Partition Panel
Screw
Remove Screws
Remove Screw
Figure 73. Removing the Measurement Bench and Lowering the Partition
Panel
Equipment Required:
Philips screwdriver
5. Disconnect the three connectors that pass through the center of the
partition panel.
6. Remove two screws from the left side of the case (viewed from front).
8. Remove one screw from the top front of the partition panel.
9. While holding the case securely, loosen the captive screw at the rear of
the measurement bench, and pull the measurement bench from the rear
of the case.
10. Remove the screw at the top rear of the partition panel securing the top
of partition panel to the measurement bench, and lower the panel being
careful not to put excessive tension on the cables.
Pump Replacement Use the following procedure to replace the pump (see Figure 74).To
rebuild the pump, see Pump Rebuilding in the Preventive Maintenance
chapter.
Equipment Required:
110V pump
Screwdriver
1. Turn instrument OFF, unplug the power cord, and remove the cover.
4. Loosen the four captive screws on the mounting plate and remove the
pump.
Fan/Filter Use the following procedure to replace the rear panel fan and the filter
(Figure 75).
Replacement
Equipment Required:
Fan
Fan Filter
Philips screwdriver
1. Turn the instrument OFF, unplug the power cord, and remove the
cover.
2. Remove the fan guard and filter from the rear of the instrument by
unsnapping it.
3. If the fan is not being replaced, install the new filter, snap it into place,
and skip the remaining steps.
5. Remove the four fan mounting screws along with nuts and washers and
remove the fan.
Filter
Fan Guard
PMT Cooler and Use the following procedure to replace the PMT cooler and reaction
chamber assembly (see Figure 76).
Reaction Chamber
Equipment Required:
Assembly
PMT cooler
Replacement
Wrench, 7/16-inch
Wrench, 9/16-inch
3. Snap off the temperature control board from the board mounts.
4. Remove the four screws securing the cooler shroud to the rear panel
and remove the shroud.
Cooler Shroud
Reaction Chamber
To COOLER on Measurement
Interface Board
To BENCH on
Temp. Control Board
6. Disconnect the cables from the PMT high voltage power supply, the
input board, and the measurement interface board. Remove all tie-
wraps securing the cables.
7. Loosen four captive screws holding cooler to floor plate and remove the
cooler assembly with the reaction chamber.
Note If only the cooler is being replaced, remove the PMT and reaction
chamber from the old cooler and install them on the new cooler.
Make sure that the heat shrink covered tubing between the reaction
chamber and the converter is light tight at the connections.
Photomultiplier Tube Use the following procedure to replace the PMT tube.
Replacement Equipment Required:
Photomultiplier tube and PMT base
Nut driver, 5/16-inch
Flat blade screwdriver
Philips screwdriver, small
1. Turn instrument OFF, unplug the power cord, and remove the cover.
2. Disconnect the high voltage cable from the PMT power supply and
unplug the BNC cable from the Input Board.
3. Remove six external screws holding PMT cover plate and the four
screws holding the PMT shroud to the panel and remove the PMT
cover plate (Figure 77). If the cooler fan is attached, unplug the fan
power cord if necessary.
4. Remove the three retaining screws holding PMT base assembly to the
cooler using a 5/16-inch nut driver.
5. Pull the PMT and PMT base from cooler assembly by twisting it
slightly back and forth.
PMT High Voltage Use the following procedure to replace the PMT high voltage power supply
(Figure 78).
Power Supply
Equipment Required:
Replacement
PMT high voltage power supply
1. Turn instrument OFF, unplug the power cord, and remove the cover.
3. Loosen the two retaining screws securing the assembly bracket to the
floor plate and slide the assembly towards the rear slightly and lift it off
the base screws.
4. Loosen two screws on the input box assembly and lift the input box
assembly off the power supply.
5. Remove the four screws securing the power supply to the bracket and
remove the power supply.
Assembly Bracket
HVPS
PMT Voltage Use the following procedure to adjust the PMT voltage after switching
from standard to extended ranges or vice versa.
Adjustment
WARNING The service procedures in this manual are restricted to qualified
service representatives.
1. Select the NO, NO2, and NOx ranges. Refer to Range Menu in the
Operation chapter.
2. Set the NO BKG and NOx BKG calibration factors to 0.0. Refer to
Calibration Factors Menu in the Operation chapter.
6. From the Main Menu, press to scroll to Service > press >
to scroll to PMT Voltage Adjustment > and press .
The Set PMT Voltage screen appears.
Reaction Chamber Use the following procedure to clean or remove the reaction chamber (see
Figure 79).
Cleaning or Removal
Equipment Required:
Allen Wrench, 9/64-inch
Wrench, 7/16-inch
Wrench, 9/16-inch
PMT Cooler
O-Ring (2-034)
Viton Spacer
Red Filter
Quartz Window
O-Ring (2-127)
NO2-to-NO Converter Use the following procedure to replace the converter (Figure 710).
Replacement Equipment Required:
NO2-to-NO Converter
Wrench, 7/16-inch
Wrench, 9/16-inch
Wrench, 1/2-inch
Wrench, 5/8-inch
Screwdriver
Nut driver, 1/4-inch
Nut driver, 5/16-inch
1. Turn instrument OFF, unplug the power cord, and remove the cover.
5. Loosen the four captive screws holding converter housing to floor plate.
Heater Assembly
Moly Converter Assembly
Thermocouple Leads
and Heater Connector
O3 Converter
6. Remove the six screws holding the top housing assembly to the bottom
half.
8. Loosen the heater clamp, pry heater apart no wider than necessary and
remove the converter cartridge noting the proper orientation of heater
wires and thermocouple probe.
Solenoid Valve Use the following procedure to replace the solenoid valve (Figure 711).
Replacement Equipment Required:
Solenoid valve
Wrench, 5/16-inch
Philips screwdriver
Solenoid Valve
Mounting Clip
Ozonator Assembly Use the following procedure to replace the ozonator assembly (Figure 7
12).
Replacement
Equipment Required:
Ozonator assembly
Wrench, 5/8-inch
Philips screwdriver
2. Carefully disconnect the plumbing at the glass inlet and outlet of the
ozonator.
4. Loosen the four captive screws securing the ozonator bracket to the
floor plate.
Flow Transducer
Ozonator
Ozonator Transformer
Screws (4)
5. Remove the two screws securing the ozonator to the ozonator bracket.
Ozonator Transformer Use the following procedure to replace the ozonator transformer (Figure 7
12).
Replacement
Equipment Required:
Ozonator transformer
Philips screwdriver
1. Turn instrument OFF, unplug the power cord, and remove the cover.
Input Board Use the following procedure to replace the input board (Figure 713).
Replacement Equipment Required:
Input board
Philips screwdriver
2. Disconnect the coaxial cable with BNC connector and the ribbon
cable.
3. Loosen the two screws holding the assembly bracket to the floor plate,
move the assembly towards the rear, and lift the assembly off the
screws.
4. Loosen the two screws holding the input box to the assembly bracket
and lift the input box off the screws.
Assembly Bracket
Input Board
Input Box
Assembly Bracket Screws (2)
5. Remove the four screws holding the input board to the input box and
remove the input board.
Input Board After replacing the input board, use the following procedure to calibrate the
input board.
Calibration
WARNING The service procedures in this manual are restricted to qualified
service representatives.
1. From the Main Menu, press to scroll to Service > press >
to scroll to Input Board Calibration > and press .
The Input Board Calibration screen appears.
DC Power Supply Use the following procedure to replace the DC power supply (Figure 7
14).
Replacement
Equipment Required:
DC power supply
Philips screwdriver
1. Turn instrument OFF, unplug the power cord, and remove the cover.
3. Loosen the captive screw securing the power supply to the chassis plate
and lift out the power supply.
BOTTOM VIEW
Captive Screw
4. Turn the power supply upside down and remove the four retaining
screws securing the power supply to the power supply plate and remove
the power supply.
Analog Output The analog outputs should be tested if the concentration value on the front
panel display disagrees with the analog outputs. To check the analog
Testing outputs, you connect a meter to an analog output channel (voltage or
current) and compare the meter reading with the output value set on the
Test Analog Outputs screen.
Equipment Required:
Multimeter
Use the following procedure to test the analog outputs.
Current Outputs
5. Check that the meter is displaying the zero value. If the meter reading
differs by more than one percent, the analog outputs should be
adjusted. Refer to the Analog Output Calibration procedure that
follows.
Table 74. Analog Output Channels and Rear Panel Pin Connections
Voltage Pin Current Pin
Channel Channel
1 14 1 15
2 33 2 17
3 15 3 19
4 34 4 21
5 17 5 23
6 36 6 25
Ground 16, 18, 19, 35, 37 Current Output Return 13, 16, 18, 20, 22, 24
Table 75. Analog Input Channels and Rear Panel Pin Connections
Input Channel Pin
1 1
2 2
3 3
4 5
5 6
6 7
7 9
8 10
Ground 4, 8, 11, 14
Analog Output Use the following procedure to calibrate the analog outputs if a meter
reading in the Analog Output Testing procedure differed by more than
Calibration one percent or after replacing the optional I/O expansion board.
Equipment Required:
Multimeter
2. From the Main Menu, press to scroll to Service > press >
to scroll to Analog Output Calibration > and press .
The Analog Output Cal screen appears.
Note When calibrating the analog output, always calibrate zero first and
then calibrate full-scale.
5. Use until the meter reads 0.0V (or 0.0 or 4.0 mA for a
current channel), then press to save the value.
8. Use until the meter reads the value shown in the Set
Output To line, then press to save the value.
Analog Input Use the following procedures to calibrate the analog inputs after replacing
the optional I/O expansion board. These procedures include selecting
Calibration analog input channels, calibrating them to zero volts, and then calibrating
them to full-scale using a known voltage source.
Calibrating the Input Use the following procedure to calibrate the input channels to zero volts.
Channels to Zero Volts
1. From the Main Menu, press to scroll to Service, press >
to scroll to Analog Input Calibration, and press .
5. Make sure that nothing is connected to the channel input pins and
press to calibrate the input voltage on the selected channel to
zero volts.
The screen displays 0.00 V as the voltage setting.
Calibrating the Input Use the following procedure to calibrate the input channels to full scale by
Channels to Full-Scale applying a known voltage to the channels.
Equipment Required:
Pressure Transducer Use the following procedure to replace the pressure transducer assembly
(Figure 716).
Assembly Replacement
Equipment Required:
Pressure transducer assembly
Philips screwdriver
1. Turn instrument OFF, unplug the power cord, and remove the cover.
Pressure Transducer Use the following procedure to calibrate the pressure transducer.
Calibration Equipment Required:
Vacuum pump
Note An error in the zero setting of the pressure transducer does not
introduce a measurable error in the output concentration reading.
Therefore, if only a barometer is available and not a vacuum pump, only
adjust the span setting.
Do not try to calibrate the pressure transducer unless the pressure is known
accurately.
3. From the Main Menu, press to scroll to Service > press >
to scroll to Pressure Calibration > and press .
The Pressure Sensor Cal menu appears.
5. Wait at least 10 seconds for the zero reading to stabilize, then press
to save the zero pressure value.
Temperature Control Use the following procedure to replace the temperature control board
(Figure 76).
Board Replacement
Equipment Required:
Temperature control board
Small flat-blade screwdriver
1. Turn instrument OFF, unplug the power cord, and remove the cover.
Ambient Temperature Use the following procedure to calibrate the ambient internal temperature
for the instrument.
Calibration
Equipment Required:
Calibrated thermometer or 10K 1% Resistor
3. From the Main Menu, press to scroll to Service > press >
to scroll to Temperature Calibration > and press .
The Calibrate Ambient Temperature screen appears.
Scrubber Use the following procedure to replace the ammonia scrubber (Figure 7
17).
Replacement
Equipment Required:
Ammonia scrubber
Nut driver, 3/16-inch
1. Turn instrument OFF, unplug the power cord, and remove the cover.
Silica Scrubber
Mounting Clips
Ammonia Scrubber
I/O Expansion Board Use the following procedure to replace the optional I/O expansion board
(Figure 718).
(Optional)
Equipment Required:
Replacement
I/O expansion board
Nut driver, 3/16-inch
1. Turn instrument OFF, unplug the power cord, and remove the cover.
3. Remove the two standoffs holding the I/O expansion board connector
to the rear panel (Figure 719).
4. Pop off the board from the mounting studs and remove the board.
Mounting Studs
Motherboard
Motherboard
Digital Output Board
Digital Output Board Use the following procedure to replace the digital output board (Figure 7
18).
Replacement
Equipment Required:
Digital output board
Nut driver, 3/16-inch
1. Turn instrument OFF, unplug the power cord, and remove the cover.
2. Remove the I/O expansion board (optional), if used. See the I/O
Expansion Board (Optional) Replacement procedure in this chapter.
4. Using the nut driver, remove the two standoffs securing the board to
the rear panel (Figure 719).
5. Pop off the digital output board from the mounting studs and remove
the board.
Motherboard Use the following procedure to replace the motherboard (Figure 7-18).
Replacement Equipment Required:
Motherboard
Philips screwdriver
Nut driver, 3/16-inch
1. Turn instrument OFF, unplug the power cord, and remove the cover.
2. Remove the I/O expansion board (optional), if used. See the I/O
Expansion Board Replacement procedure in this chapter.
3. Remove the digital output board. See the Digital Output Board
Replacement procedure in this chapter.
5. Using the nut driver, remove the eight standoffs securing the board to
the rear panel.
Measurement Use the following procedure to replace the measurement interface board
(Figure 720).
Interface Board
Equipment Required:
Replacement
Measurement interface board
Philips screwdriver
1. Lower the partition panel, then proceed to the next step below. Refer to
Removing the Measurement Bench and Lowering the Partition Panel
on page 7-8 in this chapter.
3. Unscrew the two screws at the top of the measurement interface board.
Pop off the measurement interface board from the two bottom
mounting studs and remove the board (Figure 720).
Flow Transducer Use the following procedure to replace the flow transducer (Figure 721).
Replacement Equipment Required:
Flow transducer
1. Turn instrument OFF, unplug the power cord, and remove the cover.
Flow Transducer
Retaining Nuts
4. Loosen the two retaining nuts securing the flow transducer to the
ozonator bracket and remove the flow transducer.
Flow Transducer Use the following procedure to calibrate the flow transducer.
Calibration Equipment Required:
Calibrated flow sensor
3. From the Main Menu, press to scroll to Service > press >
to scroll to Flow Calibration > and press .
The Flow Sensor Cal menu appears.
5. Wait at least 10 seconds for the zero reading to stabilize, then press
to save the zero flow value.
Front Panel Board Use the following procedure to replace the front panel board (Figure 7
22).
Replacement
Equipment Required:
Front panel board
1. Turn instrument OFF, unplug the power cord, and remove the cover.
2. Remove the three ribbon cables and the two-wire connector from the
front panel board.
3. Pop off the board from the two top mounting studs and remove the
board by lifting it up and off the slotted bottom support.
LCD Module
LCD Module Screws (2)
Retaining Screw
Front Panel Board
Two-Wire Connector
Bottom Support
Figure 722. Replacing the Front Panel Board and the LCD Module
LCD Module Use the following procedure to replace the LCD module (Figure 722).
Replacement Equipment Required:
LCD module
Philips screwdriver
CAUTION If the LCD panel breaks, do not to let the liquid crystal contact
your skin or clothes. If the liquid crystal contacts your skin or clothes, wash
it off immediately using soap and water.
Do not remove the LCD panel or frame from the LCD module.
Do not wipe the LCD polarizing plate with a dry cloth, it may easily
scratch the plate.
Do not place the LCD module near organic solvents or corrosive gases.
1. Turn instrument OFF, unplug the power cord, and remove the cover.
2. Disconnect the ribbon cable and the two-wire connector from the front
panel board.
4. Slide the LCD module out towards the center of the instrument.
Note The optimal contrast will change from one LCD screen to another.
After replacing the LCD screen, the contrast may need to be reset. If the
content on the screen is visible, select Instrument Controls > Screen
Contrast and adjust the screen contrast. If the content on the screen is not
visible, use the set contrast 10 C-Link command to set screen constrast to
mid range, then optimize the contrast. See the C-Link Protocol
Commands appendix for more information on this command.
Service Locations For additional assistance, service is available from exclusive distributors
worldwide. Contact one of the phone numbers below for product support
and technical information or visit us on the web at www.thermo.com/aqi.
Capillaries (2)
Pressure Transducer
Input Board
HVPS
Pump
Thermistor
Measurement
Interface Board
I/O Expansion
Board (Opt)
Power Supply
Mode Solenoid The mode solenoid valve switches analyzer operation between the NO
mode and NOx mode. It routes the ambient air sample either through the
reaction chamber (NO mode) or through the NO2-to-NO converter and
then to the reaction chamber (NOx mode).
Reaction Chamber The reaction chamber is where the sample reacts with ozone and produces
excited NO2 that gives off a photon of energy when it decays.
The reaction chamber is heated and controlled to approximately 50 C in
order to ensure the greatest instrument stability. The sample and ozone
flow capillaries and a thermistor sensor are also housed in/on the reaction
chamber assembly.
Optical Filter The optical filter housed in the reaction chamber limits the spectral region
viewed by the detector and eliminates possible interferences due to other
chemiluminescent reactions.
Pressure Transducer The pressure transducer measures the reaction chamber pressure.
Sample Flow Sensor The sample flow sensor located at the reaction chamber inlet measures the
sample flow into the reaction chamber.
Ozonator The Ozonator generates the necessary ozone concentration required for the
chemiluminescent reaction. The ozone reacts with the NO in the ambient
air sample to produce the electronically excited NO2 molecules.
Ozonator Flow The ozonator flow switch located at the ozonator inlet completes an
Switch electrical safety circuit when air flows through the sensor to the ozonator. If
airflow stops, the flow sensor breaks the electrical circuit to the ozonator
and shuts it off to prevent the ozonator from overheating.
Photomultiplier Tube The Photomultiplier tube (PMT) provides the infrared sensitivity required
to detect the NO2 luminescence resulting from the reaction of the ozone
with the ambient air sample.
Optical energy from the reaction is converted to an electrical signal by the
PMT and sent to the input board that transmits it to the processor.
Photomultiplier Tube The thermoelectric PMT cooler reduces the PMT temperature to
approximately -3 C to minimize dark current and increase instrument
Cooler sensitivity. The cooler helps to increase zero and span stability over a wide
ambient temperature range. The cooler housing also shields the PMT from
external electrical and optical interferences.
Pump The vacuum pump draws the reacted gasses out of the reaction chamber.
Sample Capillary The sample capillary along with the pump is used to control flow in the
sample line.
Dry Air Capillary The dry air capillary along with the pump is used to control flow in the dry
air line.
Software The processor software tasks are organized into four areas:
Instrument Control
Monitoring Signals
Measurement Calculations
Output Communication
Instrument Control Low-level embedded processors are used to control the various functions on
the boards, such as analog and digital I/O and heater control. These
processors are controlled over a serial interface with a single high-level
processor that also controls the front-panel user interface. The low-level
processors all run a common piece of firmware that is bundled with the
high-level firmware and loaded on power-up if a different version is
detected.
Each board has a specific address that is used to identify to the firmware
what functions are supported on that board. This address is also used for
the communications between the low-level processors and the high-level
processor.
Every tenth of a second the frequency counters, analog I/O, and digital I/O
are read and written to by the low-level processor. The counters are
accumulated over the past second and the analog inputs are averaged over
that second. The high-level processor polls the low-level processors once
per second to exchange the measurement and control data.
Monitoring Signals Signals are gathered from the low-level processors once per second, and
then processed by the high-level processor to produce the final
measurement values. The one-second accumulated counts representing the
NO/NOx are accumulated and reported for the user-specified averaging
time. If this averaging time is greater than ten seconds, the measurement is
reported every 10 seconds. The one-second average of the other analog
inputs are reported directly (no additional signal conditioning is performed
by the high-level processor).
In auto mode, every ten seconds the NO/NOx solenoid switches and the
processor waits three seconds for the reaction chamber to flush and
stabilize. After those three seconds, it accumulates the signal counts for
seven seconds before again switching the solenoid.
Measurement The calculation of the NO and NOx concentrations is lengthy and uses the
Calculations high-level processor to provide the most accurate readings. The calculation
begins by subtracting the appropriate electronic offset from the seven-
second count accumulation. Following this correction, the raw
accumulated counts are scaled according to the gain setting of the input
board.
Next, the uncorrected NO and NOx values are determined according to a
unique averaging algorithm which minimizes errors resulting from rapidly
changing gas concentrations. This algorithm results in NO and NOx values
which are stored in RAM in a circular buffer that holds all the ten second
data from the previous five minutes. This data is averaged over the selected
time interval, which can be any multiple of ten between 10 and 300 (the
manual modes have additional intervals of 1, 2, and 5 seconds).
The background values for NO and NOx, which are corrected for
temperature, are subtracted from their respective averages. The NO reading
is corrected by the stored span factor and by the temperature factor. The
NOx reading is partially corrected by the span factor, temperature factor,
and balance factor. The corrected NO value is subtracted from the partially
corrected NOx value to yield an uncorrected NO2 value. The NO2 value is
then corrected for converter efficiency to give a corrected NO2 reading.
Finally, the corrected NO2 reading is added to the corrected NO reading to
yield a fully corrected NOx value.
Output Communication The front panel display, serial and Ethernet data ports, and analog outputs
are the means of communicating the results of the above calculations. The
front panel display presents the NO, NO2, and NOx concentrations
simultaneously. The display is updated every 1-10 seconds, depending on
the averaging time.
The analog output ranges are user selectable via software. The analog
outputs are defaulted based on the measurement range. The defaults are
calculated by dividing the data values by the full-scale range for each of the
three parameters and then multiplying each result by the user-selected
output range. Negative concentrations can be represented as long as they
are within -5% of full-scale. The zero and span values may be set by the
user to any desired value.
Electronics All electronics operate from a universal switching supply, which is capable
of auto-sensing the input voltage and working over the entire operating
range.
Internal pumps and heaters all operate on 110VAC. An optional
transformer is required if operating on the 210-250VAC or 90-110VAC
ranges.
An on/off switch controls all power to the instrument, and is accessible on
the front panel.
Motherboard The motherboard contains the main processor, power supplies, a sub-
processor and serves as the communication hub for the instrument. The
motherboard receives operator inputs from the front panel mounted
function key panel and/or over I/O connections on the rear panel and
sends commands to the other boards to control the functions of the
instrument and to collect measurement and diagnostic information. The
motherboard outputs instrument status and measurement data to the front-
panel mounted graphics display and to the rear-panel I/O. The
motherboard also contains I/O circuitry and the associated connector to
monitor external digital status lines and to output analog voltages that
represent the measurement data. Connectors located on the motherboard
include:
Measurement The measurement interface board serves as a central connection area for all
Interface Board measurement electronics in the instrument. It contains power supplies and
interface circuitry for sensors and control devices in the measurement
system. It sends status data to the motherboard and receives control signals
from the motherboard.
Flow Sensor The flow sensor assembly consists of a board containing an instrumentation
Assembly amplifier and a flow transducer with input and output gas fittings. The
flow transducer output is produced by measuring the pressure difference
across a precision orifice. This unit is used for measuring the flow of sample
gas in the measurement system.
Temperature Control The temperature control board regulates and sets the temperature of the
Board reaction chamber and converter.
PMT Power Supply The PMT power supply produces high voltage to operate the photo
Assembly multiplier tube used in the measurement system. The output voltage is
adjustable from approximately 600 to 1200 volts under software control.
Input Board The input board accepts the current signal from the PMT and converts it
Assembly to a voltage, which is scaled by a factor of approximately 1, 10, or 100
depending on the full-scale range of the NO channel. The scaled voltage
signal is converted to a frequency and sent to the microprocessor.
The input board includes a test signal that can be activated under software
control. The test signal is injected at the first stage of the input board in
parallel with the PMT input. This allows the input board and the
connection to the processor system to be tested and calibrated without
using the PMT.
Digital Output Board The digital output board connects to the motherboard and provides
solenoid driver outputs and relay contact outputs to a connector located on
the rear panel of the instrument. Ten relay contacts normally open (with
power off) are provided which are electrically isolated from each other.
Eight solenoid driver outputs (open collector) are provided along with a
corresponding +24VDC supply pin on the connector.
I/O Expansion Board The I/O expansion board connects to the motherboard and adds the
(Optional) capability to input external analog voltage inputs and to output analog
currents via a connector located on the rear panel of the instrument. It
contains local power supplies, a DC/DC isolator supply, a sub-processor
and analog circuits. Eight analog voltage inputs are provided with an input
Front Panel Connector The front panel connector board interfaces between the motherboard and
Board the front panel mounted function key panel and Graphics display. It serves
as central location to tie the three connectors required for the function key
panel, the graphics display control lines, and the graphics display backlight
to a single ribbon cable extending back to the motherboard. This board
also includes signal buffers for the graphics display control signals and a
high voltage power supply for the graphics display backlight.
I/O Components External I/O is driven from a generic bus that is capable of controlling the
following devices:
Analog output (voltage and current)
Analog input (voltage)
Digital output (TTL levels)
Digital input (TTL levels)
Note The instrument has spare solenoid valve drivers and I/O support for
future expansion.
Analog Voltage The instrument provides six analog voltage outputs. Each may be software
Outputs configured for any one of the following ranges, while maintaining a
minimum resolution of 12 bits:
0-100mV
0-1V
0-5V
0-10V
The user can calibrate each analog output zero and span point through the
firmware. At least 5% of full-scale over and under range are also supported,
but may be overridden in the firmware, if required.
The analog outputs may be assigned to any measurement or diagnostic
channel with a user-defined range in the units of the selected parameter.
The voltage outputs are independent of the current outputs.
Analog Current The optional I/O Expansion board includes six isolated current outputs.
Outputs (Optional) These are software configured for any one of the following ranges, while
maintaining a minimum resolution of 11 bits:
0-20 mA
4-20 mA
The user can calibrate each analog output zero and span point through
firmware. At least 5% of full-scale over and under range are also supported,
but may be overridden in the firmware, if required.
The analog outputs may be assigned to any measurement or diagnostic
channel with a user-defined range in the units of the selected parameter.
The current outputs are independent of the voltage outputs. The current
outputs are isolated from the instrument power and ground, but they share
a common return line (Isolated GND).
Analog Voltage Eight analog voltage inputs are used to gather measurement data from
Inputs (Optional) third-party devices. The user may assign a label, unit, and a conversion
table (2 to 10 points). Each point in the conversion table consists of an
analog input voltage value (0-10.5 V) and a corresponding user-defined
reading value. Only two points are necessary for linear inputs, however a
larger number of points may be used to approximate non-linear inputs. All
voltage inputs have a resolution of 12 bits over the range of 0 to 10.5 volts.
Digital Relay The instrument includes one power fail relay on motherboard and ten
Outputs digital output relays on the digital output board. These are reed relays rated
for at least 500 mA @ 200VDC.
The power fail relay is Form C (both normally opened and normally closed
contacts). All other relays are Form A (normally opened contacts) and are
used to provide alarm status and mode information from the analyzer, as
well as remote control to other devices, such as for controlling valves during
calibration. The user may select what information is sent out each relay and
whether the active state is opened or closed.
Digital Inputs Sixteen digital inputs are available which may be programmed to signal
instrument modes and special conditions including:
NO Measure Mode
NOx Measure Mode
Zero Gas Mode
Serial Ports Two serial ports allow daisy chaining so that multiple analyzers may be
linked using one PC serial port.
The standard bi-directional serial interface can be configured for either RS-
232 or RS-485. The serial baud rate is user selectable in firmware for
standard speeds from 1200 to 19,200 baud. The user can also set the data
bits, parity, and stop bits. The following protocols are supported:
C-Link
Modbus Slave
Geysitech (Bayern-Hessen)
Streaming Data
The Streaming Data protocol transmits user-selected measurement data via
the serial port in real-time for capture by a serial printer, data logger, or PC.
RS-232 Connection A null modem (crossed) cable is required when connecting the analyzer to
an IBM Compatible PC. However, a straight cable (one to one) may be
required when connecting the analyzer to other remote devices. As a
general rule, when the connector of the host remote device is female, a
straight cable is required and when the connector is male, a null modem
cable is required.
Data Format:
1200, 2400, 4800, 9600, 19200, 38400, 57600, or 115200 BAUD
7 or 8 data bits
1 or 2 stop bit
No, odd, or even parity
RS-485 Connection The instrument uses a four wire RS-485 configuration with automatic flow
control (SD). Refer to Table 82 for the DB9 connector pin configuration.
Table 82. RS-485 DB9 Connector Pin Configuration
DB9 Pin Function
2 + receive
8 - receive
7 + transmit
3 - transmit
5 ground
Ethernet Connection An RJ45 connector is used for the 10Mbs Ethernet connection supporting
TCP/IP communications via standard IPV4 addressing. The IP address
may be configured for static addressing or dynamic addressing (set using a
DHCP server).
Any serial port protocols may be accessed over Ethernet in addition to the
serial port. Up to three simultaneous connections are allowed per protocol.
External Accessory The external accessory connector is not used in the Model 42i analyzer.
Connector This port is used in other models to communicate with smart external
devices that may be mounted hundreds of feet from the analyzer using an
RS-485 electrical interface.
Internal Zero/Span With the zero/span assembly option, a source of span gas is connected to
the SPAN port and a source of zero air is connected to the ZERO port as
and Sample Valves shown in Figure 91. Zero and span gas should be supplied at atmospheric
pressure. It may be necessary to use an atmospheric dump bypass plumbing
arrangement to accomplish this.
For more information, refer to the Installation chapter and the
Operation chapter.
Internal Permeation The Internal Permeation Span Source option is designed to provide a
simple source of span gas. It is intended as a quick, convenient check to be
Span Source used between zero and span calibrations for determining instrument
malfunction or drift. Because this option does not precisely control dilution
gas flow, it should not be used as a basis for instrument zero and span
adjustments, calibration updates, or adjustment of ambient data.
Whenever there is an indication of instrument drift or malfunction, a full
zero and multipoint calibration (Level 1) should be performed prior to
corrective action. For further information on zero, span and calibration of
air pollution monitors, refer to Section 2.0.9 of the US EPA's Quality
Assurance Handbook for Air Pollution Measurement Systems (Volume II).
Figure 92 shows how this option is integrated with the Model 42i
components. Energizing the sample valve V1 shuts off the sample flow and
permits the flow of zero air for analysis. When valves V1 and V2 are
energized, the flow of zero air mixes with air containing NO2 from the
permeation oven. This mode of operation provides a single point span
check.
Permeation Tube Use the following procedure to install the optional permeation tube.
Installation
1. Remove the oven cover.
3. Separate the glass chamber from the top assembly by twisting and
gently pulling the glass away from the top.
5. Attach the glass chamber to the top assembly by gently pushing the two
together with a slight twisting motion.
6. Replace the glass chamber assembly into the oven until the top of the
assembly is flush or slightly below the top of the oven.
Equipment Damage Do not use tools to tighten the knurled screw in the
following step.
7. Tighten the knurled screw finger tight. Do not use tools to tighten.
9. Replace the oven cover being careful to place the tubing and wiring
into the cover slot.
Where:
R = permeation rate in ng/min
MW = molecular weight
K(NO2) = 0.532
Oven Installation and Use the following procedure to install and configure the permeation oven.
Configuration
1. Physically install the permeation oven and valves into the instrument
and connect the cables and plumbing.
2. From the Perm Oven Settings menu in the Service menu, select Perm
Oven Selection, then select 45 C.
5. Go to the Factory Cal Gas Therm menu from the Perm Oven Settings
menu in the Service menu. Select Low Point, enter the exact value of
the resistor and press to calibrate the low resistance point.
8. Go to the High Point screen, enter the exact value of the resistor and
press to calibrate the high resistance point.
10. Remove resistor from POJ1 and re-attach the gas thermistor.
12. Go to the Factory Cal Oven Therm menu from the Perm Oven
Settings menu in the Service menu. Select Low Point, enter the exact
value of the resistor and press to calibrate the low resistance
point.
15. Go to the High Point screen, enter the exact value of the resistor and
press to calibrate the high resistance point.
17. Remove resistor from POJ3 and re-attach the permeation oven.
Permeation Tube Oven There are two general approaches that can be used to calibrate the
Calibration permeation tube oven. The first is to calibrate the temperature indicator
very accurately (to better than 0.02 C) and to use a permeation tube whose
weight loss has been previously determined at that temperature.
Setting Perm Oven You can use either of the two calibration methods presented here. One
Temperature method involves performing the Setting Perm Oven Temperature
procedure and then continuing with the Setting Temperature with Water
Bath procedure.
Alternatively, you can perform the Setting Perm Oven Temperature
procedure and then continue with the Setting Temperature with Known
Resistance procedure.
In either case, use the following procedure for setting the perm oven
temperature.
2. In the Main Menu, choose Service > Perm Oven Settings > Cal Oven
Thermistor.
MOVE CURSOR
CHANGE VALUE SAVE
3. Enter the exact value of the attached resistor and press to save
the offset.
Setting Temperature with Use the following procedure for setting the measure temperature with water
Water Bath bath.
1. Remove the thermistor from the permeation tube oven, but leave the
thermistor connected to the measurement interface board. Insert the
thermistor into the water bath next to an NIST traceable thermometer
(if necessary, use an extension cable to reach).
3. In the Main Menu, choose Service > Perm Oven Settings > Cal Gas
Thermistor > Water Bath.
MOVE CURSOR
CHANGE VALUE SAVE
5. Remove the thermistor from the water bath, dry, and replace into the
permeation tube oven.
6. Make sure the source of zero air is connected to the ZERO bulkhead
on the rear panel.
Setting Temperature with Use the following procedure to set the gas temperature with an accurate
Known Resistance known resistance.
3. In the Main Menu, choose Service > Perm Oven Settings > Cal Gas
Thermistor > Known Resistor.
MOVE CURSOR
CHANGE VALUE SAVE
4. Enter the exact value of the attached resistor and press to save
the offset.
6. Make sure the source of zero air is connected to the ZERO bulkhead
on the rear panel.
Determining Permeation Use the following procedure to determine the permeation rate by weight
Rate by Weight Loss loss.
1. Make sure the oven has been calibrated as described in the Permeation
Tube Oven Calibration procedure described previously.
5. Carefully remove the permeation tube from the oven and weigh to an
accuracy of 0.1mg. Perform this measurement as quickly as possible.
8. Compute the weight loss of the permeation tube from the values
determined in Steps 5 through 7.
9. Repeat Steps 5 through 8 until the weight loss has been determined to a
precision of 1-2%.
10. For the most accurate work, use the permeation tube in the same oven
that was used to determine the weight loss of permeation tube.
Determining Release Rate Use the following procedure to determine the release rate by transfer
by Transfer Standard standard.
1. Make sure the oven has been calibrated as described in the Permeation
Tube Oven Calibration procedure described previously. Also make
sure that the Transfer Standard has been properly calibrated.
2. Determine the permeation rate for the permeation tube in the Transfer
Standard, or install a certified permeation tube.
3. Allow the permeation tubes in both the 42i and the Transfer Standard
to stabilize for at least 24-48 hours.
4. Carefully calibrate the 42i using the Transfer Standard. The output of
the Transfer Standard should be connected to the SAMPLE bulkhead
on the rear panel of the 42i.
6. Measure the flow rate into the ZERO bulkhead on the rear panel of the
42i. Be sure that the source of zero air is connected. Note the flow and
measured NO2 concentration.
Sample Permeation The sample permeation dryer option is used when there may be widely
varying levels of ambient water vapor present. The dryer stabilizes the
Dryer moisture content of the sample stream providing a constant dew-point at
the outlet over a wide range of inlet ambient moisture levels.
A secondary benefit of the sample dryer option is its ability to remove
ambient levels of ammonia (NH3) from the sample stream.
Lag Volume The Model 42i is available with the following Lag Volume option.
Principle of Operation The Model 42i is based on the principle that nitric oxide (NO) and ozone
react to produce a characteristic luminescence with an intensity linearly
proportional to the NO concentration. Infrared light emission results when
electronically excited NO2 molecules decay to lower energy states.
Specifically,
NO O 3 NO 2 O 2 h
3 NO 2 Mo 3 NO MoO3
The ambient air sample enters the Model 42i through an inline Teflon
particulate filter, through a flow control capillary, and then splits between
the NO and the NOx channel. In the NO channel, the spilt sample is
directed to the common port of the three-way solenoid valve. The sample is
then routed either to the inlet tee of the reaction chamber or joins the
exhaust of the reaction chamber. In the NOx channel, the split sample is
directed to the common port of a second three-way solenoid valve after
having passed through the NO2 converter and a lag volume, as shown in
Figure 93. The "delayed" sample is then routed identically to the NO
channel.
The two channels operate 180 out of phase, that is, when the instrument is
monitoring NO, the NOx sample is being bypassed, and when the
instrument is monitoring NOx, the NO sample is being bypassed. The
solenoids switch every 5 seconds and the size of the lag volume has been
chosen so the same original sample is being monitored by both the NO and
the NOx channels. In this way, any positive or negative errors in the NO2
signal (determined by the difference between the NOx and NO readings) is
minimized - especially in a situation where the sample is changing rapidly,
e.g., an urban traffic environment.
The Model 42i stores the NO signal obtained during the first half of the
solenoid cycle, determines the NOx signal during the second half of the
solenoid cycle, and then calculates and updates a NO, NO2, and NOx
signal every 10 seconds. Averages are available then from 10 to 300
seconds.
Lag Volume Test Use the following procedure to test the Model 42i with lag volume option.
1. Set the NO range to 1,000 ppb and the averaging time to 10 seconds.
2. Introduce approximately 800 ppb NO into the Model 42i. Wait for a
stable reading and then record the NO2 reading. Next, introduce zero
air into the Model 42i. Wait until the instrument reads close to zero
(<3 ppb). Repeat the above procedure two more times. Take the
average of the three NO2 readings. The average NO2 reading should be
less than 50 ppb.
Ammonia Scrubber The ammonia scrubber is mounted internally and removes ammonia from
the sample air.
Teflon Particulate A 5-10 micron pore size, two-inch diameter Teflon element is available for
the Model 42i. This filter should be installed just prior to the SAMPLE
Filter bulkhead. When using a filter, all calibrations and span checks must be
performed through the filter.
Ozone Particulate The ozone particulate filter minimizes the potential for contamination of
the capillary and reaction chamber by trapping any particulate matter
Filter before passing through the capillary and reaction chamber.
NO2-to-NO Converter The Model 42i includes a Molybdenum NO2-to-NO converter as standard
equipment. A stainless steel converter is available as an option.
I/O Expansion Board The I/O expansion board provides six analog current output channels (0-
20 mA or 4-20 mA) and eight analog voltage inputs (0-10V). The DB25
Assembly connector on the rear panel provides the interface for these inputs and
outputs.
25 Pin Terminal The 25-pin terminal board assembly is included with the I/O expansion
board. Refer Terminal Board PCB Assemblies in the Installation
Board Assembly chapter for information on attaching the cable to the connector board. For
associated part numbers, refer to External Device Connection
Components on page 7-6.
Terminal Block and The optional terminal block and cable kits provide a convenient way to
connect devices to the instrument. These kits break out the signals on the
Cable Kits rear panel connector to individual numbered terminals.
Two types of terminal block and cable kits are available. One kit is for the
DB37 connectors and can be used for either the analog output connector
or the relay output connector. The other kit is for the DB25 connector and
can be used for the optional I/O expansion board. For associated part
numbers, refer to External Device Connection Components on page 7-6.
Each kit consists of:
one six-foot cable
one terminal block
one snap track
Note Supporting all of the connections on units with the optional I/O
expansion board requires:
two DB37 kits
one DB25 kit
Cables Table 91 identifies the optional individual cables that are available for the
instrument and Table 92 provides the cable color codes. For associated
part numbers, refer to External Device Connection Components on page
7-6.
Note Table 92 provides the color coding for both 25-pin cables and 37-
pin cables. Color codes for pins 1-25 are for 25-pin cables; color codes for
pins 1-37 are for 37-pin cables.
Mounting Options The analyzer can be installed in the configuration described in Table 93
and shown in Figure 91 through Figure 94.
Table 93. Mounting Options
Mounting Type Description
Bench Positioned on bench, includes mounting feet and front panel side-
trim handles.
EIA rack Mounted in an EIA-style rack, includes mounting slides and front
panel EIA-rack mounting handles.
Retrofit rack Mounted in an EIA-style rack, includes mounting slides and front
panel EIA-rack mounting handles. This configuration is intended
for direct replacement of a C-series instrument in an existing
rack. The rail mounting location is lower on the case and the
front mounting screw slots are in non-standard EIA locations.
they were not designed, (v) causes external to the Products such as, but not
limited to, power failure or electrical power surges, (vi) improper storage
and handling of the Products or (vii) use of the Products in combination
with equipment or software not supplied by Seller. If Seller determines
that Products for which Buyer has requested warranty services are not
covered by the warranty hereunder, Buyer shall pay or reimburse Seller for
all costs of investigating and responding to such request at Seller's then
prevailing time and materials rates. If Seller provides repair services or
replacement parts that are not covered by the warranty provided in this
warranty, Buyer shall pay Seller therefor at Seller's then prevailing time and
materials rates. ANY INSTALLATION, MAINTENANCE, REPAIR,
SERVICE, RELOCATION OR ALTERATION TO OR OF, OR
OTHER TAMPERING WITH, THE PRODUCTS PERFORMED BY
ANY PERSON OR ENTITY OTHER THAN SELLER WITHOUT
SELLER'S PRIOR WRITTEN APPROVAL, OR ANY USE OF
REPLACEMENT PARTS NOT SUPPLIED BY SELLER, SHALL
IMMEDIATELY VOID AND CANCEL ALL WARRANTIES WITH
RESPECT TO THE AFFECTED PRODUCTS.
THE OBLIGATIONS CREATED BY THIS WARRANTY
STATEMENT TO REPAIR OR REPLACE A DEFECTIVE PRODUCT
SHALL BE THE SOLE REMEDY OF BUYER IN THE EVENT OF A
DEFECTIVE PRODUCT. EXCEPT AS EXPRESSLY PROVIDED IN
THIS WARRANTY STATEMENT, SELLER DISCLAIMS ALL
OTHER WARRANTIES, WHETHER EXPRESS OR IMPLIED, ORAL
OR WRITTEN, WITH RESPECT TO THE PRODUCTS,
INCLUDING WITHOUT LIMITATION ALL IMPLIED
WARRANTIES OF MERCHANTABILITY OR FITNESS FOR ANY
PARTICULAR PURPOSE. SELLER DOES NOT WARRANT THAT
THE PRODUCTS ARE ERROR-FREE OR WILL ACCOMPLISH
ANY PARTICULAR RESULT.
Instrument Each command sent to the analyzer over the serial port must begin with the
American Standard Code for Information Interchange (ASCII) symbol or
Identification byte value equivalent of the instrument's identification number plus 128.
Number
Thermo Fisher Scientific Model 42i Instruction Manual B-1
C-Link Protocol Commands
Commands
For example, if the instrument ID is set to 25, then each command must
begin with the ACSII character code 153 decimal. The analyzer ignores any
command that does not begin with its instrument identification number. If
the instrument ID is set to 0, then this byte is not required. For more
information on changing Instrument ID, see Chapter 3, Operation.
Commands The analyzer must be in the remote mode in order to change instrument
parameters via remote. However, the command set mode remote can be
sent to the analyzer to put it in the remote mode. Report commands
(commands that dont begin with set) can be issued either in the remote
or local mode. For information on changing modes, see Chapter 3,
Operation.
The commands can be sent in either uppercase or lowercase characters.
Each command must begin with the proper instrument identification
number (ASCII) character. The command in the example below begins
with the ASCII character code 170 decimal, which directs the command to
the Model 42i, and is terminated by a carriage return CR (ASCII
character code 13 decimal).
Many of the commands have two forms. One form reads parameter from
the instruments memory, and the other writes, or updates, a parameter.
The syntax for a write command adds the word set in front of the
command and provides an argument. Command responses are generally
echoed with a data element appended.
Note If the Service Mode is active, C-Link set commands are not
allowed. This is to prevent parameters from being changed remotely while
the unit is being serviced locally.
The save and set save params commands (duplicated for backward
compatibility) store parameters in FLASH memory. It is important that
this command be sent each time instrument parameters are changed. If
changes are not saved, they will be lost in the event of a power failure.
Convert Concentration Convert concentration formats from xxxxE+yy to x.xxxE+yy. The iSeries
Formats uses more standard script than the C-Series by placing only one digit to the
left of the decimal.
Commands List Table B1 lists the 42i C-Link protocol commands. The interface will
respond to the command strings outlined below.
Table B1. C-Link Protocol Commands
Command Description Page
1 Simulates pressing soft key 1 pushbutton B-33
2 Simulates pressing soft key 2 pushbutton B-33
3 Simulates pressing soft key 3 pushbutton B-33
4 Simulates pressing soft key 4 pushbutton B-33
addr dns Reports/sets domain name server address for Ethernet port B-43
addr gw Reports/sets default gateway address for Ethernet port B-43
addr ip Reports/sets IP address for Ethernet port B-43
addr nm Reports/sets netmask address for Ethernet port B-44
addr ntp Reports the IP address for network time protocol server B-44
alarm chamber Reports/sets chamber temperature alarm maximum value B-14
temp max
no
no2
nox
high no
high no2
high nox
low no
low no2
low nox
These commands report the measured NO, NO2, and NOx concentrations
when operating in single range, or high and low NO, NO2, and NOx when
operating in dual or auto range mode. The example below shows that the
NO concentration is 67.2 ppb.
Send: no
Receive: no 6.720E+01 ppb
conv temp
This command reports the current NO2 converter temperature. The
example below reports that the current converter temperature is 320.7 C.
cooler temp
This command reports the current PMT cooler temperature. The example
below reports that the current PMT cooler temperature is -2.8 C.
flow
sample flow
These commands report the current measured flow. The example below
reports that the flow measurement is 0.700 liters/minute.
Send: flow
Receive: flow 0.7 1/m
internal temp
This command reports the current internal instrument temperature. The
first reading is the temperature being used in instrument calculations. The
second temperature is the actual temperature being measured. If
temperature compensation is on, then both temperature readings are the
same. If temperature compensation is off, a temperature of 30 C is used as
the default temperature even though the actual internal temperature is 27.2
C. The example below shows that temperature compensation is on and
that the internal temperature is 27.2 C.
pmt temp
This command reports the PMT cooler temperature. The example below
reports that the PMT cooler temperature is -2.8 C.
pmt voltage
This command reports the PMT voltage. The example below reports that
the current PMT voltage is -750 volts.
pres
This command reports the current reaction chamber pressure. The first
pressure reading is the pressure reading being used in instrument
calculations. The second pressure is the actual pressure reading being
measured. If pressure compensation is on, then both pressure readings are
the same. If pressure compensation is off, a pressure of 300 mmHg is used
as default pressure even though the actual pressure is 306.3 mmHg. The
example below shows that the actual reaction chamber pressure is 306.3
mmHg.
Send: pres
Receive: pres 753.4 mm Hg, actual 306.6
react temp
This command reports the current reaction chamber temperature. The
example below reports that the current reaction temperature is 49.0 C.
flags
This reports 8 hexadecimal digits (or flags) that represent the status of the
ozonator, PMT, pressure and temperature compensation status, gas units,
gas mode, and alarms. To decode the flags, each hexadecimal digit is
converted to binary as shown in the Figure B1. It is the binary digits that
define the status of each parameter. In the example below, the instrument is
reporting that the ozonator and PMT are both on, and that the instrument
is in the span gas mode.
Send: flags
Receive: flags 80028000
version
This command reports the version of all the firmware components. The
following example shows a list of the firmware components that were
displayed by issuing the version command.
Send: version
Receive: version
Program = 01.05.79.225
Library = 01.01.60.167
Kernal = 2.4.24-uc0-003-Thermo
Board = 81, File = /usr/application.hex
Board App = 11.3.100 BI 4.0.97
File App = 11.3.100 BI 4.0.97
list din
list dout
These commands report the current selection for the digital outputs in the
format. Output no Index number variable name active state. The active
state for digital outputs is open or closed. The active state for digital inputs
is high or low.
list lrec
list srec
list stream
list sp
These commands report the list of current selections for lrec logging data,
srec logging data, streaming data output, or the scratch pad (sp) list.
The scratch pad is a temporary memory area which is used to set up lists of
selections for lrec, srec, or streaming data items. The user can copy any of
these lists to the scratch pad, modify individual elements in the list, then
save the scratch pad back to the original list. Refer to the sp field
command for information on how to edit the scratch pad.
The following example shows the list for streaming data output.
er xy
lr xy
sr xy
x=|0|1| : Reply termination format (see set format format
command)
y = | 0 | 1 | 2 | : Output format (see set erec/lrec/srec format format
command)
These commands report the last long and short records stored or the
dynamic data record. In the example below, the command requests a long
record with no checksum, in ASCII format with text. For details on how to
decode the flag fields within these records, see the flags command.
Send: lr01
Receive: lr01
erec
This command returns a snapshot of the main operating conditions
(measurements and status) at the time the command is issued. The
following example shows a typical response.
The format is defined by the current settings of the format and erec
format commands. For details on erec formatting, see the Record Layout
Definition section at the end of this appendix. For details on how to
decode the flag fields within these records, see the flags command.
Send: erec
Receive: erec
10:11 04-06-05 flags DD008000 no 0.000 nox 0.000 no2
0.000 1 lono 147.500 lonox 0.000 lono2 0.000 1 pmtv
805.491 tempal 1 pres 172.278 pcal 150.000 smplf 0.000
ozonf 0.050 hiavgtime 10 loavgtime 10 nobkg 0.000
noxbkg 0.000 nocoef 1.000 noxcoef 1.000 no2coef 1.000
lonocoef 1.000 lonoxcoef 1.000 lono2coef 1.000 norange
100000.000 noxrange 100000.000 no2range 100000.000
lonorange 100000.000 lonoxrange 100000.000 lono2range
100000.000
lrec
srec
lrec xxxx yy
srec xxxx yy
lrec aa:bb oo-pp-qq yy
srec aa:bb oo-pp-qq yy
xxxx = the number of past records
yy = the number of records to return (1 to 10)
aa = hours (01 to 24)
bb = minutes (01 to 59)
oo = month (01 to 12)
pp = day (01 to 31)
qq = year
These commands output long or short records and dynamic data. The
output format is determined by the set lrec format, and set srec format
commands. The logging time is determined by the set lrec per and set
srec per commands. In dual range, the long records and short records
contain the high and low NO and NOx concentrations. In single range the
low NO and low NOx values are set to 0 and the high NO and high NOx
are used. In NO or NOx only mode, the pertinent high value used, other
Send: lrec 5
Receive: lrec 100 5
11:03 02-22-03 flags 54089100 no 8416E-1 nox 8458E-1
lono 6474E-1 lonox 6506E-1 pres 131.4 pmtt 53.1 intt
80.0 rctt 80.0 convt 61 smplf 0.500 ozonf 0.000 pmtv -
116
11:04 02-22-03 flags 54089100 no 8421E-1 nox 8457E-1
lono 6477E-1 lonox 6505E-1 pres 131.5 pmtt 53.1 intt
80.0 rctt 80.0 convt 61 smplf 0.500 ozonf 0.000 pmtv -
116
11:05 02-22-03 flags 54089100 no 8440E-1 nox 8456E-1
lono 6492E-1 lonox 6505E-1 pres 131.5 pmtt 53.2 intt
80.0 rctt 80.0 convt 61 smplf 0.500 ozonf 0.000 pmtv -
116
11:06 02-22-03 flags 54089100 no 8432E-1 nox 8483E-1
lono 6486E-1 lonox 6525E-1 pres 133.0 pmtt 53.0 intt
80.0 rctt 80.0 convt 61 smplf 0.500 ozonf 0.000 pmtv -
116
11:07 02-22-03 flags 54089100 no 8442E-1 nox 8383E-1
lono 6494E-1 lonox 6449E-1 pres 131.5 pmtt 53.1 intt
80.0 rctt 80.0 convt 61 smplf 0.500 ozonf 0.000 pmtv -
116
where:
erec format
lrec format
srec format
These commands report the output format for long and short records, and
dynamic data in various formats such as ASCII without text, ASCII with
text, or binary. The example below shows the output format for long
records is ASCII with text, according to Table B4.
erec layout
lrec layout
srec layout
These commands report the layout (string indicating the data formats) for
data that is sent out in response to the erec, lrec, srec, and related
commands. For details on how to interpret the strings, see Record Layout
Definition later in this appendix.
lrec per
srec per
These commands report the long and short records logging period. The
example below shows that the short record logging period is 5 minutes.
no of lrec
no of srec
These commands report the number of long and short records stored in the
long and short records memory. The example below shows that 50 long
records have been stored in the memory.
Send: no of lrec
Receive: no of lrec 50 recs
malloc lrec
malloc srec
These commands report the currently set memory allocation for long and
short records in percent of total memory.
Note Issuing these commands will clear all the logging data memory. All
the existing records should be retrieved using appropriate commands, if
required.
sp field number
This command reports the variable number and name stored at the index in
the scratch pad list.
The scratch pad is a temporary memory area which is used to set up lists of
selections for lrec, srec, or streaming data items. The user can copy any of
these lists to the scratch pad, modify individual elements in the list, then
save the scratch pad back to the original list.
The following example shows that field 5 in the scratch pad is set to index
number 13, which is for the variable pressure.
Send: sp field 5
Receive: sp field 5 13 pres
stream per
This command reports the currently set time interval in seconds for
streaming data.
stream time
This command reports if the streaming data string will have a time stamp
attached to it or not, according to Table B5.
no coef
no2 coef
nox coef
high no coef
high no2 coef
high nox coef
low no coef
low no2 coef
low nox coef
These commands report NO, NO2, and NOx coefficients in single range
mode, or the high and low range coefficients in dual or auto range mode. If
the mode is incorrect, the instrument responds with cant, wrong
settings. The example below reports that the NO coefficient is 1.000.
Send: no coef
Receive: no coef 1.000
no gas
no2 gas
nox gas
high no gas
high no2 gas
high nox gas
low no gas
low no2 gas
low nox gas
These commands report NO, NO2, and NOx span gas concentrations used
to auto-calibrate NO, NO2, and NOx coefficients. The high and low
commands are only available in dual and auto range mode. If the mode is
incorrect, the instrument responds with cant, wrong settings. The
example below reports that the NO low span gas concentration is 240.0
ppb.
no bkg
nox bkg
bkg no
bkg nox
These commands report the current NO and NOx backgrounds. The
example below reports that the NO background is 5.5 ppb.
Send: no bkg
Receive: no bkg 5.5 ppb
span dev
This command reports the span deviation (span check offset). The
following example reports that the span deviation is 1 ppb.
zero dev
This command reports the zero deviation (maximum zero check offset).
The following example reports that the zero deviation is 10 ppb.
span dur
This command reports the span duration. The following example reports
that the span duration is 10 minutes.
zero dur
This command reports the zero duration. The following example reports
that the zero duration is 10 minutes.
zs period
This command reports the zero/span (z/s) period. The following example
reports that the zero/span period is 24 hours.
Send: zs period
Receive: zs period 24 hr
zs avg time
This command reports the zero/span (z/s) averaging time in seconds. The
following example reports that the zero/span averaging time is 30 seconds,
according to Table B2.
isc
iscreen
This command retrieves the framebuffer data used for the display on the
iSeries instrument. It is 19200 bytes in size, 2-bits per pixel, 4 pixels per
byte arranged as 320 by 240 characters. The data is sent in RLE encoded
form to save time in transmission. It is sent as a type '5' binary C-Link
response with no checksum.
The RLE encoding consists of a 0 followed by an 8-bit count of
consecutive 0xFF bytes. The following 'c' code will expand the incoming
data.
while (rlecount)
{
*(ptr++) = 0;
rlecount--;
}
}
else if (*(rlescreen + i) == 0xff)
{
unsigned char rlecount = *(unsigned char *)(rlescreen + ++i);
while (rlecount)
{
*(ptr++) = 0xff;
rlecount--;
}
}
}
}
To convert this data into a BMP for use with windows, it needs to be
turned into a 4BPP as that is the smallest windows can display. Also note
that BMP files are upside down relative to this data, i.e. the top display line
is the last line in the BMP.
menutext
This command displays the text of the menu item where the cursor is
currently positioned. The following example shows that the cursor is
positioned at the instrument controls menu item.
Send: menutext
Receive: menu text main menu instrument controls
sc
screen
This command is meant for backward compatibility on the C series. Screen
information is reported using the iscreen command above.
Send: screen
Receive: screen
This is an I series
Instrument. Screen
Information not
available
Measurement range no
range no2
Configuration range nox
high range no
high range no2
high range nox
low range no
low range no2
low range nox
These commands report NO, NO2, and NOx range in single range mode,
or the high and low ranges in dual or auto range mode. If the mode is
incorrect, the instrument responds with cant, wrong settings. The
example below reports that the NO full-scale range is set to 50 ppb,
according to Table B6 and Table B7.
Send: range no
Receive: range no 0: 5000E-2 ppb
custom range
range = | 1 | 2 | 3 |
This command reports the user-defined value of custom range 1, 2, or 3.
The example below reports that custom range 1 is defined to 5.50 ppb.
Send: custom 1
Receive: custom 1 550E-2 ppb
range mode
This command reports the current range mode.
gas mode
This command reports the current mode of sample, zero, or span. The
example below reports that the gas mode is sample.
set sample
set sample gas
These commands set the zero/span valves to the sample mode. The
example below sets the instrument to sample mode, that is, the instrument
is reading the sample gas.
set zero
set zero gas
These commands set the zero/span valves to the zero mode. The example
below sets the instrument to zero mode that is, the instrument is reading
the sample gas.
set span
set span gas
These commands set the zero/span valves to the span mode. The example
below sets the instrument to span mode that is, the instrument is sampling
span gas.
gas unit
This command reports the current gas units (ppb, ppm, g/m3, or mg/m3).
The example reports that the gas unit is set to ppb.
meas mode
This command reports which measurement mode (NO/NOx, NO, or
NOx) is active. The example below reports that the measurement mode is
set to NO.
pres comp
This command reports whether pressure compensation is on or off. The
example below shows that pressure compensation is on.
temp comp
This command reports whether temperature compensation is on or off.
The example below shows the temperature compensation is off.
Hardware contrast
This command reports the screens level of contrast. The example below
Configuration shows the screen contrast is 50%, according to Table B8.
Send: contrast
Receive: contrast 10:50%
date
This command reports the current date. The example below reports the
date as December 1, 2004.
Send: date
Receive: date 12-01-04
ozonator
This command reports the ozonator is on or off. The example below
reports that the ozonator is on.
Send: ozonator
Receive: ozonator on
ozonator flow
This command reports the current ozonator flow. The example below
reports that the current ozonator flow is 0.050 LPM.
ozonator safety
This command reports the status of the ozonator safety on or off. The
example below reports that the ozonator safety is on.
ozonator status
This command reports the status of the ozonator and safety. The example
below reports that the ozonator is off.
pmt status
This command reports the status of the PMT on or off. The example
below reports that the PMT is on.
save
set save params
This command stores all current parameters in FLASH memory. It is
important that each time instrument parameters are changed, that this
command be sent. If changes are not saved, they will be lost in the event of
a power failure. The example below saves the parameters to FLASH
memory.
time
This command reports the current time (24-hour time). The example
below reports that the internal time is 2:15:30 pm.
Send: time
Receive: time 14:15:30
addr gw
This command reports the default TCP/IP gateway address.
Send: addr gw
Receive: addr gw 192.168.1.1
Note This command cannot be used when DHCP is on. Refer to the
DHCP command that follows for additional information.
addr ip
This command reports the IP address of the analyzer.
Send: addr ip
Receive: addr ip 192.168.1.15
Note This command cannot be used when DHCP is on. Refer to the
DHCP command that follows for additional information.
addr nm
This command reports the IP netmask.
Send: addr nm
Receive: addr nm 255.255.255.0
Note This command cannot be used when DHCP is on. Refer to the
DHCP command that follows for additional information.
addr ntp
This command reports the IP address for the network time protocol server.
See Network Time Protocol Servier in the Communications Settings
section of the Operation chapter for more information.
Send: addr ip
Receive: addr ip 192.168.1.15
baud
This command reports the current baud rate for the serial port
(RS232/RS485). The example below reports that the current baud rate is
9600 baud.
Send: baud
Receive: baud 9600
Note After the command is sent, the baud rate of the sending device must
be changed to agree with the instrument.
dhcp
This command reports the current state of use of the Dynamic Host
Configuration Protocol (DHCP) as on or off. DHCP is used to assign an
IP address to the instrument automatically. The following example shows
that DHCP is on.
Send: dhcp
Receive: dhcp on
Note If DHCP is changed from on to off and then the IP address, the
netmask address, or the gateway address is changed, you must cycle power
to the instrument before the change takes effect. Until you cycle the power,
the address assigned by the DHCP server will still be used and reported as
the current address.
format
This command reports the current reply termination format. The example
below shows that the reply format is 00, which means reply with no
checksum, according to Table B9.
Send: format
Receive: format 00
where xxxx = 4 hexadecimal digits that represent the sum of all the
characters (bytes) in the message
host name
This command reports the host name string. The following example
reports the host name is set to iSeries. This command returns bad cmd if
no host name has been set.
instr name
This command reports the instrument name.
instrument id
This command reports the instrument id.
Send: instrument id
Receive: instrument id 42
Note Sending this command via RS-232 or RS-485 will require the host to
use the new id for subsequent commands.
mode
This command reports what operating mode the instrument is in: local,
service, or remote. The example below shows that the instrument is in the
remote mode.
Send: mode
Receive: mode remote
Note The instrument will always respond to the command mode with
the status of the password lock as mode local or mode remote
regardless of the above setting.
The following example sets the instrument to accept the set mode local
or set mode remote commands.
Send: set allow mode cmd 1
Receive: set allow mode cmd 1 ok
power up mode
This command reports the current power up mode setting either 0 =
local/unlocked or 1 = remote/locked according to Table B11. The default
value is 0; power up in local/unlocked mode. The following example shows
that the instrument is configured to power up in the remote/locked mode.
Send: power up mode
Receive: power up mode 1
program no
This command reports the analyzers model information and program
version number, which will be dependant on the current version.
Send: program no
Receive: program no iSeries 42i 01.00.01.074
tz
This command reports the tz timezone string for the NTP server. See
Network Time Protocol Server in the Communications Settings
section of the Operation chapter for more information.
Send: tz
Receive: tz EST+5EDT
set tz string
This command sets the timezone string for the instrument for use with the
NTP server, where string is a standard timezone string. Common strings are
listed in the Timezone screen description in Chapter 3.
B13. This command responds with feature not enabled if the I/O
expansion board is not detected.
dig in
This command reports the status of the digital inputs as a 4-digit
hexadecimal string with the most significant bit (MSB) being input 16.
Send: dig in
Receive: dig in 0xff7f
din channel
This command reports the action assigned to the digital input channel and
index number of the corresponding active state. The following example
reports input 5 to be assigned an index number 9 corresponding to action
of analog outputs to zero with the active state being high.
Send: din 5
Receive: din 5 9 AOUTS TO ZERO high
dout channel
This command reports the index number, output variable and the active
state assigned to digital output channel. The following example reports
output 4 to be assigned an index number 11 corresponding to action of
general alarm.
Send: dout 4
Receive: dout 4 11 GEN ALARM open
dtoa channel
This command reports the outputs of the 6 or 12 Digital to Analog
converters, according to Table B15. The following example shows that the
D/A #1 is 97.7% full-scale.
Send: dtoa 1
Receive: dtoa 1 97.7%
All channel ranges are user definable. If any customization has been made
to the analog output configuration, the default selections may not apply.
This command reports the current state of the MODBUS registers (analog
outputs). Output is in floating point format with the pair of registers start
and start+1 appearing as the left-most bit.
relay stat
This command reports the current relay logic normally open or normally
closed, if all the relays are set to same state, that is all open or all closed.
The example below shows that the status when all the relays logic is set to
normally open.
Note If individual relays have been assigned different logic then the
response would be a 4-digit hexadecimal string with the least significant
byte (LSB) being relay no 1.
For example:
Note If the command is sent without an appended relay number then all
the relays are assigned the set logic of normally open/closed.
Record Layout The Erec, Lrec Srec layouts contain the following:
Definition A format specifier for parsing ASCII responses
A format specifier for parsing binary responses
In addition to these the Erec Layout contains:
A format specifier for producing the front-panel displays
Values are read in using either the ASCII or binary format specifiers and
converted to uniform internal representations (32-bit floats or 32-bit
integers). These values are converted into text for display on the screen
using the format specifier for the front-panel display. Normally, the
specifier used to parse a particular datum from the input stream will be
strongly related to the specifier used to display it (such as, all of the floating
point inputs will be displayed with an 'f' output specifier, and all of the
integer inputs will be displayed with a 'd' specifier).
Format Specifier for The first line of the Layout response is the scanf-like parameter list for
ASCII Responses parsing the fields from an ASCII ERec response. Parameters are separated
by spaces and the line is terminated by a \n (the normal line separator
character). Valid fields are:
%s - parse a string
%d - parse a decimal number
%ld - parse a long (32-bit) decimal number
%f - parse a floating point number
%x - parse a hexadecimal number
%lx - parse a long (32-bit) hex number
%* - ignore the field
Note Signed versus unsigned for the integer values does not matter; it is
handled automatically.
Format Specifier for The second line of the Layout response is the binary parameter list for
Binary Responses parsing the fields from a binary response. Parameters MUST be separated
by spaces, and the line is terminated by a '\n'. Valid fields are:
t - parse a time specifier (2 bytes)
D - parse a date specifier (3 bytes)
i - ignore one 8-bit character (1 byte)
e - parse a 24-bit floating point number (3 bytes: n/x)
E - parse a 24-bit floating point number (3 bytes: N/x)
f - parse a 32-bit floating point number (4 bytes)
Format Specifier for The subsequent lines in the ERec Layout response describe the appearance
EREC Layout of the full panel. The full instrument panel as it appears on the screen has
two columns of lines. Each line is composed of three major components:
(1) a text field, (2) a value field, and (3) a button. None of these three
components is required. The text field contains statically displayed text.
The value field displays values which are parsed out of the response to a
DATA/ERec command. It also displays, though background changes,
alarm status. The button, when pressed, triggers input from either a dialog
box or a selection list. There are five kinds of buttons, B, I, L, T, and N.
Each line in the layout string corresponds to one line on the display. The
layout string describes each of the three major fields as well as translation
mechanisms and corresponding commands.
Text The first field in the layout string is the text. It is delimited by a ':'. The
string up to the first ':' will be read and inserted in the text field of the line.
Value String This is followed by a possible string enclosed in quotes that is used to place
a string into the value field.
Value Source The value source, which is the item (or word) number in the DATA/ERec
response, appears next. This is followed by an optional bitfield designator.
The datum identified by the value source can be printed as a string 's',
hexadecimal 'x', decimal 'd', or floating point 'f', or binary 'b' number.
Typically, bitfield extractions are only done for decimal or hexadecimal
numbers.
Floating-point numbers can be followed with an optional precision
specifier which will be used as an argument to printf's %f format (e.g., a
field of '4' would be translated into the printf command of '%.3f').
Alternately, the special character '*' can precede the precision specifier; this
causes an indirection on the precision specifier (which now becomes a field
number).
This is useful when formatting, for example, numbers which have varying
precision depending on the mode of the instrument.
Binary numbers can also have an optional precision specifier which is used
to determine how many bits to print. For example, the specifier 'b4' will
print the lowest four bits of the parsed number.
There are serious restrictions on where an 's' field may appear: currently
sources 1 and 2 must be 's', and no others may be 's'.
Alarm Information The value source is followed by optional alarm information, indicated by a
commercial at sign '@' with a source indicator and a starting bit indicator.
All alarm information is presumed to be two bits long (low and high). The
bitfield extraction is performed on the integer part of the source. Typical
alarm information would appear as '@6.4'.
Translation Table Then, there appears an optional translation table within braces '{}'. This is
a string of words separated by spaces. An example translation table would
be '{Code_0 Code_1 Code_2 Code_3}'. The value, once extracted is used
as a zero-based index into the translation table to determine the string to
display.
Selection Table Then there appears an optional selection table within parentheses '(...)'.
This is a string of numbers separated by spaces '(0 1)'. The selection table
lists the translation table entries which the user may select from when
setting the parameter. This is not necessarily the same as the entries which
may be displayed.
Button Designator Then there appears an optional button designator. This will be one of 'B',
'I', 'L', 'T', or 'N'.
B- Indicates a button which pops up an input dialog prompting the
user for a new value using the designated input format. The input
format is specified from the 'B' through the subsequent semicolon.
IIndicates a button which pops up a selection list with input
translation. That is, the values read are translated before they are
compared to the selection list options.
LIndicates a button which pops up a selection list without any
translation. The output value is number of the selected option.
TIndicates a button which pops up a selection list with output
translation. The number of the option selected is used as an index into
the translation table to generate an output string.
NIndicates a button which only sends the subsequent command to
the instrument. No user-prompting happens.
The following string through an optional | or the end of the line is the
command which is to be sent to the instrument upon the completion of the
button selection. The command string should normally contain print-style
formatting to include the user input. If a | is present, it indicates a
command which is sent to the instrument upon successful completion of
the button command to update the value field.
This is not currently used.
'Concentrations\n'
'\n'
' NO:3s\n'
This is a line which appears slightly indented. The text field is 'NO', the
value is taken from the third element of the data response, and interpreted
as a string.
This is a line which also appears slightly indented. The next field is also
'NO', but the value is taken from the eighteenth element of the data
response, again interpreted as a string. A button appears on this line which,
when pressed, pops up an input dialog which will state "Please enter a new
value for NO using a d.ddd format." The string entered by the user is used
to construct the output command. If the user enters, for example, '1.234',
the constructed command will be 'set no coef 1.234'.
This is a line which appears slightly indented, the title is again 'NO', and
the value the twenty-first element of the data response, interpreted as a
floating-point number. There is a no-translation button which creates a
selection list of twelve "Code nn" options. The number of the user
selection is used to create the output command.
This is a line which has a title of 'Mode', and value taken from the sixth
field of the data response. There is a bitfield extraction of bits 12 through
13 from the source (the value type is not important here because the value
is being translated to an output string). Once the bits have been extracted,
they are shifted down to the bit-zero position. Thus, the possible values of
this example will be 0 through 3. The translation list shows the words
which correspond to each input value, the zeroth value appearing first (0 ->
local, 1 -> remote, etc.). The selection list shows that only the first two
values, in this case, are to be shown to the user when the button is pressed.
The 'T' button indicates full translation, input code to string, and user
selection number to output string.
'\xC'
This shows that the bitfield end (the second part of a bitfield specification)
is optional. The bitfield will be one bit long, starting in this case at the
eleventh bit.
This shows the use of indirect precision specifiers for floating point
displays. The background value is taken from the 7th element, and the
precision specifier is taken from the 8th. If the asterisk were not present, it
would indicate instead that 8 digits after the decimal point should be
displayed.
Serial Communication The following are the communication parameters that are used to configure
the serial port of the iSeries to support MODBUS RTU protocol.
Parameters
Number of Data bits : 7 or 8
Number of Stop bits : 1 or 2
Parity : None, Odd, or Even
Data rate : 1200 to 115200 Baud (9600 is default)
TCP Communication iSeries Instruments support the MODBUS/TCP protocol. The register
definition is the same as for the serial interface. Up to three simultaneous
Parameters connections are supported over Ethernet.
TCP connection port for MODBUS : 502
Application Data Here are the MODBUS ADU (Application Data Unit) formats over serial
and TCP/IP:
Unit Definition
Slave Address The MODBUS save address is a single byte in length. This is the same as
the instrument ID used for C-Link commands and can be between 1 and
127 decimal (i.e. 0x01 hex to 0x7F hex). This address is only used for
MODBUS RTU over serial connections.
Function Code The function code is a single byte in length. The following function codes
are supported by the instrument:
Data The data field varies depending on the function. For more description of
these data fields, see Function Codes below.
Error Check In MODBUS over Serial an error check is included in the message. This is
not necessary in MODBUS over TCP/IP because the higher-level protocols
ensure error-free transmission. The error check is a two-byte (16 bit) CRC
value.
Function Codes This section describes the various function codes that are supported by the
Model 42i.
(0x01/0x02) Read Coils / Read Coils / Inputs read the status of the digital outputs (relays) in the
Read Inputs instrument. Issuing either of these function codes will generate the same
response.
These requests specify the starting address, i.e. the address of the first
output specified, and the number of outputs. The outputs are addressed
starting at zero. Therefore, outputs numbered 116 are addressed as 015.
The outputs in the response message are packed as one per bit of the data
field. Status is indicated as 1 = Active (on) and 0 Inactive (off). The LSB
of the first data byte contains the output addressed in the query. The other
outputs follow toward the high end of this byte, and from low order to
high order in subsequent bytes. If the returned output quantity is not a
multiple of eight, the remaining bits in the final data byte will be padded
with zeros (toward the high order end of the byte). The Byte Count field
specifies the quantity of complete bytes of data.
Note The values reported may not reflect the state of the actual relays in
the instrument, as the user may program these outputs for either active
closed or open.
Request
Function code 1 Byte 0x01 or 0x02
Starting Address 2 Bytes 0x0000 to maximum allowed by instrument
Quantity of outputs 2 Bytes 1 to maximum allowed by instrument
Unit Identifier 1 Byte 0x00 to 0xFF (Passed back in response)
Response
Function code 1 Byte 0x01 or 0x02
Byte count 1 Byte N*
Output Status n Byte N = N or N+1
Error Response
Function code 1 Byte 0x01 or 0x02
Exception code 1 Byte 01=Illegal Function, 02=Illegal Address,
03=Illegal Data, 04=Slave Device Failure
Request
Field Name (Hex)
Function 0x01
Starting Address Hi 0x00
Starting Address Lo 0x02
Quantity of Outputs Hi 0x00
Quantity of Outputs Lo 0x0D
Response
Field Name (Hex)
Function 0x01
Byte Count 0x03
Output Status 210 0xCD
Output Status 1115 0x0A
The status of outputs 210 is shown as the byte value 0xCD, or binary
1100 1101. Output 10 is the MSB of this byte, and output 2 is the LSB.
By convention, bits within a byte are shown with the MSB to the left, and
the LSB to the right. Thus, the outputs in the first byte are 10 through 2,
from left to right. In the last data byte, the status of outputs 15-11 is shown
as the byte value 0x0A, or binary 0000 1010. Output 15 is in the fifth bit
position from the left, and output 11 is the LSB of this byte. The four
remaining high order bits are zero filled.
(0x03/0x04) Read Holding Read holding / input registers reads the measurement data from the
Registers / Read Input instrument. Issuing either of these function codes will generate the same
response. These functions read the contents of one or more contiguous
Registers registers.
These registers are 16 bits each and are organized as shown below. All of
the values are reported as 32-bit IEEE standard 754 floating point format.
This uses 2 sequential registers, least significant 16 bits first.
The request specifies the starting register address and the number of
registers. Registers are addressed starting at zero. Therefore, registers
numbered 116 are addressed as 015. The register data in the response
message are packed as two bytes per register, with the binary contents right
justified within each byte. For each register, the first byte contains the high
order bits and the second contains the low order bits.
Request
Function code 1 Byte 0x03 or 0x04
Starting Address 2 Bytes 0x0000 to maximum allowed by instrument
Quantity of Registers 2 Bytes 1 to maximum allowed by instrument
Response
Function code 1 Byte 0x03 or 0x04
Byte count 1 Byte 2 x N*
Register value N* x 2 Bytes N = N or N+1
*N = Quantity of Registers
Error Response
Function code 1 Byte Function code + 0x80
Exception code 1 Byte 01=Illegal Function, 02=Illegal Address,
03=Illegal Data, 04=Slave Device Failure
Request
Field Name (Hex)
Function 0x03
Starting Address Hi 0x00
Starting Address Lo 0x09
No. of Registers Hi 0x00
No. of Registers Lo 0x04
Response
Field Name (Hex)
Function 0x03
Byte Count 0x06
Register value Hi (10) 0x02
Register value Lo (10) 0x2B
Register value Hi (11) 0x00
The contents of register 10 are shown as the two byte values of 0x02 0x2B.
Then contents of registers 1113 are 0x00 0x00, 0x00 0x64 and 0x00 0x64
respectively.
(0x05) Force (Write) The force (write) single coil function simulates the activation of the digital
Single Coil inputs in the instrument, which triggers the respective action.
This function code is used to set a single action to either ON or OFF. The
request specifies the address of the action to be forced. Actions are
addressed starting at zero. Therefore, action number 1 is addressed as 0.
The requested ON/OFF state is specified by a constant in the request data
field. A value of 0xFF00 requests the action to be ON. A value of 0x0000
requests it to be OFF. All other values are illegal and will not affect the
output. The normal response is an echo of the request, returned after the
state has been written.
Note This function will not work if the instrument is in service mode.
Request
Function code 1 Byte 0x05
Starting Address 2 Bytes 0x0000 to maximum allowed by instrument
Output Value 2 Bytes 0x0000 or 0xFF00
Response
Function code 1 Byte 0x05
Starting Address 2 Bytes 0x0000 to maximum allowed by instrument
Output Value 2 Bytes 0x0000 or 0xFF00
Error Response
Function code 1 Byte Function code + 0x80
Request
Field Name (Hex)
Function 05
Output Address Hi 00
Output Address Lo 05
Output Value Hi FF
Output Value Lo 00
Response
Field Name (Hex)
Function 05
Output Address Hi 00
Output Address Lo 05
Output Value Hi FF
Output Value Lo 00
MODBUS Parameters Table C1 through Table C3 lists the MODBUS addresses supported for
the Model 51i.
Supported
IMPORTANT NOTE The addresses in the following tables are Protocol
Data Unit (PDU) addresses. Verify the coil number on your MODBUS
master to ensure that it matches the coil number on the instrument.
In addition to the coils listed in the Read Coils table, coils in the Write
Coils table can also be read.
Note For additional information on how to read registers and interpret the
data, refer to the (0x03/0x04) Read Holding Registers / Read Input
Registers section in this appendix.
Register Variable
Number
55 & 56 ANALOG IN 1/NO BKG** I/O Expansion Board Option*
57 & 58 ANALOG IN 2/NOx BKG** I/O Expansion Board Option*
59 & 60 ANALOG IN 3/LO NO COEF** I/O Expansion Board Option*
61 & 62 ANALOG IN 4/LO NO2 COEF** I/O Expansion Board Option*
63 & 64 ANALOG IN 5/LO NOx COEF** I/O Expansion Board Option*
65 & 66 ANALOG IN 6/HI NO COEF** I/O Expansion Board Option*
67 & 68 ANALOG IN 7/HI NO2 COEF** I/O Expansion Board Option*
69 & 70 ANALOG IN 8/HI NOx COEF** I/O Expansion Board Option*
71 & 72 OZONATOR FLOW
73 & 74 NOT USED
75 & 76 NOT USED
77 & 78 NOT USED
79 & 80 NO CORRECTION CONC O2 Sensor Option
81 & 82 NO2 CORRECTION CONC O2 Sensor Option
83 & 84 NOx CORRECTION CONC O2 Sensor Option
85 & 86 NOT USED
87 & 88 NOT USED
89 & 90 LOW NO CORRECTION CONC O2 Sensor Option
91 & 92 LOW NO2 CORRECTION CONC O2 Sensor Option
93 & 94 LOW NOx CORRECTION CONC O2 Sensor Option
95 & 96 NOT USED
97 & 98 NOT USED
99 & 100 HIGH NO CORRECTION CONC O2 Sensor Option
101 & 102 HIGH NO2 CORRECTION CONC O2 Sensor Option
103 & 104 HIGH NOx CORRECTION CONC O2 Sensor Option
105 & 106 NOT USED
107 & 108 NOT USED
109 & 110 NOT USED
111 & 112 O2 % Internal O2 Sensor Option
113 & 114 O2 SENS TEMP Internal O2 Sensor Option
115 & 116 NOT USED
117 & 118 NOT USED
119 & 120 NOT USED
Register Variable
Number
121 & 122 EXT ALARMS
*
The analog inputs 18 will not show if the Ouput Conc Cal option is set to YES in the Instrument
Configuration screen.
**
BKG and COEF items are displayed only if Ouput Conc Cal option is set to YES in the Instrument
Configuration screen.
Note Writing 1 to the coil number shown in the following table will
initiate the action triggered listed in the table. This state must be held for
at least 1 second to ensure the instrument detects the change and triggers
the appropriate action.
Note The coils within each coil group in the following table are mutually
exclusive and will not be triggered if there is conflict. Before you assert (1)
one coil in a group, make sure the other coils in the group are de-asserted
(0).
Reading a Write Coil To read a write coil, issue a read coil command for that coil. For example,
to view the state of write coil 101, issue a read coil 101.
Serial Communication The following are the communication parameters that are used to configure
the serial port of the iSeries to support Geysitech protocol.
Parameters
Number of Data bits : 7 or 8
Number of Stop bits : 1 or 2
Parity : None, Odd, or Even
Data rate : 1200 to 115200 Baud (9600 is default)
TCP Communication iSeries Instruments support the Geysitech/TCP protocol over TCP/IP. The
register definition is the same as for the serial interface. Up to three
Parameters simultaneous connections are supported over Ethernet.
TCP connection port for Geysitech: 9882
Instrument Address The Geysitech instrument address has a value between 0 and 127 and is
represented by 3 digit ASCII number with leading zeros or leading spaces if
required (e.g. Instrument address of 1 is represented as 001 or <SP><SP>1)
The instrument Address is the same as the Instrument ID used for C-Link
and MODBUS commands. This can be set via the front panel.
The Instrument Address is represented by <address> in the examples
throughout this document.
Note Device IDs 128 through 247 are not supported because of
limitations imposed by the C-Link protocol.
Block Checksum The Block Checksum Characters are calculated beginning with a seed value
of 00000000, binary (0x00), and bitwise exclusive ORing with each of the
Characters <BCC> characters of the command string (or response) including the framing
characters <STX> and <ETX>. The checksum works as an error check. The
command terminator determines the presence or absence of <BCC>.
If a command is terminated by <ETX> then the next two characters are the
checksum, if the command is terminated with <CR> no checksum is
attached.
The block checksum is represented by two characters, which represent a 2
digit hex number (1byte) (e.g. 1 byte 0xAB hex checksum will be
represented by the two characters A & B).
The checksum is referred to as <BCC> throughout this document.
Geysitech Commands The following commands are supported by the Geysitech protocol:
Instrument Control Command (ST)
Data Sampling/Data Query Command (DA)
Instrument Control Command There are three control commands supported by the Geysitech protocol.
(ST) This <control command> is a single letter, which triggers an action in the
instrument. These commands are active only when service mode is inactive
and the zero/span option is present.
Command N switches the instrument gas mode to Zero mode.
Command K switches the instrument gas mode to Span mode.
Command M switches the instrument gas mode to Sample mode.
The following are the different acceptable formats of the ST command:
<STX>ST<address><control command><ETX><BCC>
OR
<STX>ST<address><control command><CR>
OR
<STX>ST<address><SP><control command><CR>
OR
<STX>ST<address><SP><control command><ETX><BCC>
The <address> is optional, which means it can be left out completely. The
<address> if present must match the Instrument Address. Additional space
can be present after the <address>.
If the received command does not satisfy the above formats or if the
<address> does not match the Instrument Address the command is ignored.
This is a sample command to switch the instrument to zero mode,
instrument id 5:
<STX>ST005<SP>N<CR>
Data Sampling/Data This command DA initiates a data transfer from the instrument. The
Query Command instrument responds with measurement data, which depends on the range
mode and is listed in Measurements reported in response to DA
(DA) command below.
The command structure for a data query command is as follows:
<STX>DA<address><ETX><BCC>
The <address> is optional, which means it can be left out completely. The
<address> if present must match the Instrument Address. Additional space
can be present after the <address>.
If the <address> is left out then no space is allowed in the query string.
A command with no address is also a valid command.
The following are the different acceptable formats of the DA command
with Instrument Address 5:
<STX>DA<CR>
<STX>DA005<CR>
<STX>DA<SP><SP>5<ETX><BCC>
<STX>DA<ETX><BCC>
The data query string is valid and will be answered with data transmission
only if the command starts with <STX> which is followed by the characters
DA, and the <address> (if present) matches the Instrument Address, and
the command is terminated with either <CR> with no checksum or <ETX>
followed by the correct checksum <BCC>.
Example:
Geysitech Protocol with transmission of three concentrations (Instrument
ID is 1, Operation Status is 03, Error Status is 04):
Data Query String: <STX>DA<CR>
Reply String:
<STX>MD03<SP>001<SP>+2578+01<SP>03 <SP>04<SP>0000000000 <SP>002 <SP>
Address First Concentration(E-format)=25.78 Address+1
+5681+00<SP>03<SP>04<SP>0000000000<SP>003<SP>+1175+01<SP>03<SP>04<SP
Second Concentration = 5.681 Address+2 Third Concentration=11.75
0000000000<SP><CR>
Measurements reported in The following measurements reported in response to DA command are for
response to DA command the Model 42i.
Single Range Mode The 3 measurements reported in single range mode include:
NO
NO2
NOx
Dual/Auto Range Mode The 6 measurements reported in dual or auto range modes include:
low NO
low NO2
low NOx
high NO
high NO2
high NOx
Operating and Error See Table D1 for operating status and Table D2 for error status for the
Status Model 42i.