Using your analyser

Operation Manual
Running Order
How it works
Your DXL parts and Accessories
DXL radiation Safety
Using your DXL Analyser
Using your Analyser with your PC
Advanced Settings
Supervisor Examination
How XRF Works
How does XRF work?
Each of the elements present in a sample produces
a unique set of characteristic x-rays that is a
"fingerprint" for that specific element. XRF analysers
determine the chemistry of a sample by measuring the
spectrum of the characteristic x-ray emitted by the
different elements in the sample when it is illuminated
by x-rays. These x-rays are emitted either from a
miniaturized x-ray tube, or from a small, sealed
capsule of radioactive material.
How does XRF work?
o A fluorescent x-ray is created when an x-ray of
sufficient energy strikes an atom in the sample,
dislodging an electron from one of the atom's
inner orbital shells.
o The atom regains stability, filling the vacancy
left in the inner orbital shell with an electron
from one of the atom's higher energy orbital
o The electron drops to the lower energy state by
releasing a fluorescent x-ray, and the energy of
this x-ray is equal to the specific difference in
energy between two quantum states of the
XRF at work
Atomic Level Process of Fluorescence Production
XRF at work
When a sample is measured using XRF, each
element present in the sample emits its own unique
fluorescent x-ray energy spectrum.
By simultaneously measuring the fluorescent x-rays
emitted by the different elements in the sample, the
XL2 can rapidly determine those elements present
in the sample and their relative concentrations - in
other words, the elemental chemistry of the sample.
Excitation Source
Macro Level X-Ray Production
Technical Illustration of a NITON Analyzer
DSP converts analog pulses to digital;
sends to CPU.
Digitized value represents original
energy of characteristic X-Ray.
Over measuring time, each element’s
energy accumulated into a series of
element/energy channels.
Spectrum contains qualitative and
quantitative information from the
Si PIN Detector
Your DXL parts and
DXL Battery and charger
DXL Battery
DXL Battery Charger
DXL Accessories Cont
Prolene Windows
USB Cable
Small Spot Locator
Rear Dust Cover
Battery Cover
DXL Accessories Cont
Software CD
Quick Start Guide
DXL radiation Safety
The Thermo Scientific Model Niton DXL series analyser contains an x-ray
tube which emits radiation into a shielded sample chamber. The x-ray
tube emits radiation only when a sample is being measured. During this
time, indicator lights surrounding the analyser light up to alert personnel
that the x-ray tube is on. The shielded sample chamber is designed to
reduce radiation dose rates to less than 2.5 micro Sieverts (or
equivalently 0.25 millirem) per hour at a 5 centimetre distance from any
point along the surface of the analyser. The sample chamber is also
designed with an interlock system so that the x-ray tube cannot be
energized unless the sample chamber door is closed. The purpose of
this interlock system is to prevent a person from placing any part of their
body into the primary, unshielded x-ray beam.
With the shielded and interlocked sample chamber, the Thermo Scientific
Model Niton DXL series analyser can be operated safely with minimal
programmatic and administrative controls. The following list summarizes
the key safe use guidelines that all operators should be made aware of:
Overview Cont
 • Operators should be provided basic radiation safety information. Consult with your local
radiation Protection Advisor (RPA) to determine the specific operator training requirements
for your needs. At a minimum, operators should read and understand the safety information
in this manual.
 • Where required by RPA, obtain authorization to use this analytical device. Authorization is
typically granted in the form of a registration, license, certificate, or permit. Your local
authority can be a valuable resource for safe use guidelines.
 • Do not attempt to override interlocks or in any way interfere with their proper operation.
 • The DXL analyser incorporates engineering controls that are designed to prevent access
to the sample chamber during a measurement. Never place a part of your body in the
sample chamber if the "X-RAY ON" indicator lights are on.
 • Maintain the analyser in accordance with the instructions in this User's Guide.
 • High voltage and high intensity x-ray beam hazards exist inside the instrument. Do not
attempt to disassemble or service your analyser beyond the maintenance instructions
described in this Resource Guide. Never remove parts or components except as described
in this Resource Guide. Service must be performed at an authorized service centre.
 • Use caution when lifting or moving the analyser to prevent strains or back injuries.
Time, Distance and Shielding
 Radiation Protection Basics
 Reasonable effort should always be made to maintain exposure to radiation as far below dose limits as
is practical. This is known as the ALARA (As Low as Reasonably Achievable) principle. For any given
source of radiation, three factors will help minimize your radiation exposure: Time, Distance, and
 Time
 The longer you are exposed to a source of radiation the longer the radiation is able to interact in your
body and the greater the dose you receive. Dose increases in direct proportion to length of exposure.
 Distance
 Intensity of radiation becomes weaker as it spreads out from a source since the same amount of
radiation becomes spread over a larger area. Based on geometry alone, dose increases and
decreases with an inverse-squared relation to your distance from the source of radiation (additional
dose rate reduction comes from air attenuation). For example, the radiation dose one foot from a
source is nine times greater than the dose three feet from the source.
 Shielding
 Shielding is any material that is placed between you and the radiation source. The more material
between you and the source, or the denser the material, the less you will be exposed to that radiation.
The sample chamber of the Thermo Scientific Model Niton DXL series analyser incorporates various
forms of shielding.
Radiation Dose Rates
 The Niton DXL Series analyser is designed to limit the radiation dose rate to no
more than 2.5 micro Sieverts per hour at a 5 cm distance from any point along
the surface of the instrument under worst case operating conditions. Worstcase operating conditions are as follows:
 • X-ray tube voltage at its maximum of 45 tube kilovolts;
 • X-ray tube current at its maximum of 0.1 tube milliamps;
 • An attenuated primary x-ray beam - i.e. no sample present, or;
 • A solid plastic sample present to maximize scatter radiation.
 The dose rates shown in the following table were measured while the analyser
was operating under these worst-case conditions. Each measured value
represents the maximum dose rate measured over the general area described.
A Thermo Scientific Model MicroRem LE dose rate meter (SN:9049 calibrated
3/29/2012) was used to perform these measurements. The measured values
listed are net above background. The background at the time and location of
measurement was 0.1 μSv per hour.
 The dose rate meter has an estimated minimum detection limit of 0.1 μSv per
hour. The measurements were performed on August 17, 2012.
Storage and Transportation
 Storage
 Regulations in some jurisdictions may require that you store your analyser in a secured
area to prevent access, use, and/or removal by unauthorized individuals. Storage
requirements may vary by location, particularly with regard to storage at temporary job
sites or away from your primary storage location such as hotels and motels and in
vehicles. You should contact your local Radiation Control Authority to identify the specific
storage requirements in your jurisdiction.
 Transportation
 Transport of lithium ion batteries is regulated by most transport authorities. End-users
should obtain additional information and training regarding the local requirements for
transport of lithium ion batteries, as appropriate for the specific transport modes that may
be used. In particular, for air transport, most jurisdictions have adopted the regulatory
guidance published by the International Air Transport Association (IATA). These IATA
regulations provide instructions for the safe transport of lithium ion batteries by air in
Packing Instructions 965 (for batteries packed alone) and Packing Instruction 966 (for
batteries packed with equipment). You will find additional information about lithium ion
battery safety and transportation in the Start up Operations and Standard Maintenance
Sections of this Resource guide.
 It is recommended that you ship the Niton DXL in its original shipping container and foam
to protect the sensitive measuring equipment inside the analyser.
Using your DXL
Inspecting your analyser
Battery Installation and Charging
Unscrew the captive screw on
the battery door and remove the
Grasp the cloth tag on
the bottom of your
analyzer’s battery and
firmly pull the battery
out into your hand.
Battery Installation and Charging Cont
Place the old battery aside and slide the
new battery into the cavity in the rear of
the analyser. The battery is keyed, and
will only insert fully one way. If it bottoms
before it is fully inserted, remove the
battery, turn it over top to bottom, and
reinsert it.
Press in until the battery clicks into place.
Replace the battery door and screw in the captive screw.
Recharging The Battery Pack
The battery pack is normally trickle charged in the analyser when the analyser is plugged in,
but it is faster to use the battery recharger - especially if you normally use battery power a lot.
Fully recharging a depleted battery pack in the charger takes approximately 2 hours.
Remove the battery pack from the analyser.
2. Place the battery pack upside down into the charger. The battery pack is keyed, and will
only fit into the charger fully one way. If your battery pack is resting on the back of the charger
rather than sliding all the way to the bottom, remove the battery pack, turn it around, and reinsert it into the charger.
3. The red light blinks when the charger needs a calibration. The red
light is steadily on when the charger has a fault.
4. The blue light blinks when the charger is running a calibration. The
blue light is steadily on when the calibration is complete.
5. The green light blinks when the charger is charging a battery. The
green light is steadily on when the charging is complete.
Inspecting your analyser
The LED Power and Safety Indicators
The LED Power Indicators are located surrounding the Power and Trigger Buttons on the
Control Panel. These LED Lights are on whenever your analyser is powered on.
The LED Safety Indicators are located on the sides of your analyser, just above the Carrying
Handles; on the rear of your analyser, just above the Port Cover; and surrounding the Cover
Release Button near the top edge of the Control Panel where it meets the Cover. These LED
lights are on whenever a sample is being analysed or whenever a System Check is in
The Control Panel
The control panel is located on the analyser’s front housing, directly below the
Touch Screen. The control panel consists of a Power Button and a larger
Trigger Button. Using the touch screen you may navigate through all of the
analyser’s screens and menus. The Trigger Button to the right of the Power
Button is used to initiate readings. The Power Button both controls the power to
the analyser and serves as an "escape" button. When the Power Button is
pushed and immediately released, it functions as an "escape", and brings you
back to the Main Menu from the current screen in the menu system.
Replacing the Measurement Window
CAUTION: Take every precaution to prevent damage to the solid beryllium surface of the tube and detector behind the analysis
window. Both the x-ray tube and detector are located directly behind the analysis window and each has a small surface of solid
beryllium or beryllium oxide. Beryllium-containing materials, in solid form and as finished parts, present no particular health
hazard. However, exposure to the dust or fumes from beryllium metal or metal oxides has the potential to cause serious health
WARNING: Before you begin, cut off power to your analyser!
WARNING: In the event that there is known or suspected damage to the solid beryllium surface of the tube or detector, the
following precautions are recommended. Use latex or other disposable gloves for any handling or clean up of visible beryllium
fragments or contamination. Collect fragments into a thick plastic bag, seal the bag tightly with adhesive tape, and affix a label
clearly indicating “Danger Beryllium”. If there has been any inadvertent contact with skin, thoroughly wash affected skin area with
soap and water before eating, drinking, or smoking. Contact your health and safety staff and/or Thermo Fisher Scientific customer
support for further instruction if needed. Dispose of beryllium waste in accordance with all federal, state or local regulations.
Remove the old Measurement
Window from the bracket.
Clean the Window area thoroughly, using a clean, guaranteed
lint-free cloth and isopropyl alcohol.
Measurement Window is
Kapton - P/N 187-4280
CAUTION! Do not use fingers to
press window into place! Use a
smooth, hard surface such as back of
Start up procedure
To turn on the analyzer, depress the On/off/escape button on the control panel until
the Touch Screen comes on.
On startup, the screen will show by a Start Screen which will automatically count
down from 4 to 0 in increments of one second.
When the start up is complete, the Start Screen will be replaced by the Logon
Screen. Tap anywhere on this screen to continue.
The Logon Screen will be
replaced by a Warning
Screen, advising you that
this analyser produces
radiation when the lights
are flashing. You must
acknowledge this warning
by selecting the Yes
button before logging on.
Selecting the No button
will return you to the
Logon Screen.
Select your 4 digit security
code, followed by the Enter
button. The default
password is 1-2-3-4
Start up procedure
After you have completed the log on procedure, the word
"USER" will appear on the bottom of the screen, then the
Main Menu will appear. Note that security codes are
editable. This will be covered in ‘advanced settings’. There
we can change passwords and set User Privileges.
Please Note - Your analyser will need to set the temperature of
the detector to -25ºC before it can be used. If you attempt to use
the analyser before this procedure the instrument will display the
warning ‘Please wait, cooling detector’. This procedure will take
no longer than 60 seconds once logged in and will remain stable
until the analyser is powered down.
System check
Every DXL analyser will have the ‘system check’ function.
This function is an internal check to maintain the instruments
fundamental parameters calibration, check tube output and
detector resolution.
This is an important operation to perform as over time the
instruments detector can suffer from electronic drift slightly
shifting the calibration curve.
Running the ‘system check’ can keep the instruments
calibration curve in line which will continue to provide the
best possible results from the analyser.
System check
Select the System Check
Icon on the Main Menu to
perform a system check. We
recommend that you perform
a system check once every
working day, as part of your
normal start up procedure.
If you attempt to initiate
a system check while the
cover is open, a Safety
Message will be
displayed informing you
the "Lid is open", and
you will return to the
Main Menu.
While performing the
system check, your screen
will show a progress bar
indicating the progress of
the check. When it is done,
the screen will show a
100% completion.
Click yes to continue the
system check.
Please make sure at this time
there is nothing on the front
of the instrument that could
affect the reading.
If however there is an error
message we advice you to call
Niton UK Service on 01256
397860 or email
[email protected]
Calibrating the Touch Screen
Select the Calibrate Touch Screen icon to
re-calibrate the analyser's touch screen
display. This procedure establishes the
display boundaries for the touch screen
Analyze mode and tools
The Tools Menu can be accessed in two ways,
and the options are different depending on the
way you access it. The main Tools Menu is
accessed by selecting the Analyze Icon from the
Main Menu, then selecting the Tools button from
the slide down window on the Ready to Analyze
From the main menu select
analyze to go in to the testing
mode for the instrument. From
this mode you will be able to
perform various tasks including
taking an analysis, adding data
and set averaging.
The Tools Menu enables you to
perform common data-related
tasks such as printing and
averaging readings. Select a
task from the menu to initiate
that task. The options available
can vary depending on the
selected mode.
The AuDIT algorithm determines whether or not a surface is plated. AuDIT can detect plating as thick as 8mm. Since most plating
is in the 2-3mm range, this can usually detect plated objects. Heavily plated objects with a plating greater than 8mm thick will read
as Gold Plate Not Detected.
AuDIT can be toggled on or off from your Tools Menu. This toggle is only available in Precious Metals Mode. Selecting the
AuDIT:Off button will turn AuDIT on, and change the button to "AuDIT:On". Selecting the AuDIT:On button will turn AuDIT off, and
change the button to "AuDIT:Off".
AuDIT uses four separate tests run automatically to determine whether or not a
sample is plated.
1. The first test is an iterative comparison of X-ray intensity signatures. This
finds most examples of plating.
2. Nickel is often used as a pre-plate, and high proportions of Ni in a reading
are a good indicator of plating.
3. Plating's often have a low Karat value when averaged with the substrate, so
Karat values of less than 9 are flags indicating plating.
4. A Karat rating that is not one of the standard Karat percentages - within 0.5
karat of 9kt, 10kt, 14kt, 18kt, 22kt, or 24kt (referred to as Out of Plumb) - also
strongly indicate that this is a plating.
Only if the sample passes all four
tests is it labelled Gold Plate Not
Detected". This does not mean
that there is no plating, but that
the presence or absence of
plating cannot be determined by
the analyser.
X-ray Intensity
Hi Ni
Low Karat
Non-Std. Karat Rating
AuDIT Additional Methods of Plating Detection
Analyze the item in several different areas
Variance of more than 1-2% in Au content can be a positive indication that an item is plated.
Look for identifying marks (hallmarks)
Compare to your results. Discrepancies may indicate that an item is plated.
Note: You may come across some Italian jewellery that has 18k gold plating over 14k gold. This is hallmarked as 14k or 585,
but will likely show 16-17k on XRF.
“Smell” the item
A metallic, copper-like smell (similar to copper-based coins such as USA pennies) indicates the possible presence of a copper
substrate under gold plating.
Use a strong magnet
A magnetic draw on the item may indicate a magnetic substrate under gold plating (gold alloys are not magnetic).
As a final and last resort…
analyze a spot, perform a deep file or grind, and then analyze the same spot again. A reduced gold content (more than 1-2%)
indicates a thinning of the gold plating layer.
AuDIT Messages
If AuDIT is not
enabled, a white on
black message
stating "AuDIT
Disabled" will display
on the Results
Screen while in
Precious Metals
If AuDIT detects
what looks like gold
plating on the
material, a black on
red message
stating "Gold Plate
Probable" will
display on the
Results Screen.
If AuDIT finds too
much Nickel in the
sample, a black on
yellow message
stating "High Ni
Content" will
display on the
Results Screen.
If AuDIT detects
what is either
unplated gold or
very thickly plated
gold, a black on
white message
stating "Gold Plate
Not Detected" will
display on the
Results Screen.
If AuDIT detects
what may be
gold plate, but
isn't sure, a black
on yellow
message stating
"Gold Plate
Suspect" will
display on the
Results Screen.
When AuDIT
finds a Karat
rating less than
8.5 in the
sample, a black
on yellow
message stating
"Low Karat" will
display on the
results Screen.
When AuDIT finds a Karat rating other than the standard Karats,
a black on yellow message stating "Non-Standard Karat" will
display on the results Screen.
Using your analyser
1.Turn the analyser on. Follow the screen instructions and “Log On” as the operator using either the default password or a
custom one as designated by the user in an NDU file. See advanced settings for details.
2. Wait five (5) minutes before using the analyser, allowing the instrument electronics to stabilize.
3. Verify that the date is set properly for data tracking purposes. See advanced settings for details.
4. (Optional) Connect the analyser to a computer via the included USB cable.
5. During analysis and detector calibrations, it is important to ensure that the analyser is not exposed to strong
electromagnetic fields, including those produced by computer monitors, hard drives, mobile telephones, walkie talkies, etc.
Keep a minimum two (2) feet (0.7 meters) distance between the analyser and electronic devices.
6. From the Main Menu, select System Check icon then the "Yes" button.
6.1. System Check calibrates the detector and verifies it is operating to specifications. After starting the process, no
further user interaction is required during this operation. When the instrument is finished performing the check, the unit
will show “System OK”. If the analyser shows any other message displaying a fault contact Niton UK.
Using your analyser
1 Open the lid using the push button on the front of the lid.
2 Place the item to be tested over the measurement window and close the lid. Use the camera to line up the item.
3 Press the start/stop button to start the test and again to finish the test once you are happy with the result.
4 Once the test is finished you can open the lid and take out the item. Opening the lid whilst the test is ongoing will end
the test.
The analyser will often display a correct alloy identification and accurate chemistry result before the above specified time
interval. If the accuracy meets the user’s requirements, it is not necessary to measure for the full time.
Longer measurements might be necessary if low concentrations of alloy elements must be determined.
Measure the supplied calibration check sample AT LEAST once a shift. If correct, continue work. If incorrect, redo System
Check and re-take the past 2 hours of results.
UNDERSIZED OR NON-CONTACT SAMPLES (samples that do not make contact with or that do not fully cover the
measurement aperture)
For samples that do not fully cover the measurement aperture, increase the testing time by increasing the time in inverse
proportion to the decrease in percentage of aperture covered. For example: a ring only covers ½ of the aperture, so increase
the measurement time by two (e.g., from 10 to 20 seconds per filter for chemistry).
Averaging Forward
Enables you to average different readings together from this analysis forward. Select the Avg Forward button to initiate future
sample averaging. Avg Forward will set up an automatic personal averaging protocol to be followed until your analyser is shut
down, or this feature is disabled. To begin, select the number of readings you want to average from the virtual numeric keypad.
Your analyser will calculate an average reading after that number of tests, and continue this pattern until stopped. For
example, if you select 3 on the virtual keypad, the analyser will automatically calculate, average, and store a reading for every
three tests you take, storing the individual readings along the way.
The range number is selected using a virtual numeric keypad on your analyser similar to the keypad used for login. Select the
digits in the range number from the keypad, then select the E button to enter the number. The C button will clear all, and the
“<“ button will clear the last digit entered. The average will automatically be displayed.
Avg Back (Alt)
Enables you to average different readings together from this
analysis backward. Select the Avg Back option to initiate
backwards sample averaging. Avg Back will take the number of
readings you select and average their analytical results. The
range is counted from the last reading backward by the number
of readings selected. If your last reading was #15, selecting 3
would average readings #13, 14, and 15. The average is
calculated, displayed, and stored into memory as the next
sequential reading number, in this case, #16.
The range number is selected using a virtual numeric keypad on
your analyser similar to the keypad used for login. Select the
digits in the range number from the keypad, then select the E
button to enter the number. The C button will clear all, and the
“<“ button will clear the last digit entered. The average will
automatically be displayed.
Data management
Viewing Data
Use the Data Screen to view previously taken test
result readings. When the View Data icon is
selected, the Results screen of your most recent test
is shown on the Touch Screen.
Using the buttons on the control panel, you may view different readings or additional data for individual readings. Your analyser
will display the standard screen analysis. Selecting the Complete List icon will display a complete scrolling elemental chemistry
listing. Each press of the Down Button scrolls the screen down to the next element. You can also use the scroll bar along the
right side to scroll or page through the elements.
Scrolling Down Through the Complete Listing of Elements
Selecting the Left Arrow icon (<--) on your analyser's touch
screen will display the previous reading, or if the first reading is
currently displayed, the last reading. Selecting the Right Arrow
icon (-->) on your analyser's touch screen will display the next
reading, or if the last reading is currently displayed, the first
reading in memory. Your analyser can store up to 10,000
readings. You can also look at the complete x-ray spectra for
each reading stored in the analyser's memory.
Erasing Data
Select the Erase Readings icon to erase all accumulated test readings from your analyser. Selecting the Erase Readings
icon will bring up a confirmation screen asking you “Are you sure?” with options to select “YES” or “NO”. Selecting the Yes
Button will erase all test reading data from your analyser. Selecting the No Button will return you to the Erase Menu.
Note - We recommend that you download all your readings
into an NDT file for recording purposes before erasing all data.
Using Your
Analyser With
Your PC
PC Requirements
We recommend the
following system
configuration for
Niton Data Transfer
The installation
program installs NDT
and NDTr. You may
install the files in any
directory on your PC.
Minimum Configuration
Recommended Configuration
133 MHz
Operating System
Windows 98, Millennium Edition,
2000, XP, Vista, Windows 7
Windows XP
64 MB
128 MB
Free Disk Space
10 MB
Installing NDT
Close all the programs that you have open on your PC.
Insert the CD into the CD-ROM drive.
If the installation program starts automatically, follow the instructions on the
screen. If the installation program does not start, go to the next step.
Double-click My Computer.
Double-click the Compact Disc icon:
Double-click the Setup Application icon.
Follow the instructions on the screen.
Starting NDT Software
Click the Start button.
Select Programs.
Select Thermo Niton.
Click NDT.
Connecting the XRF Analyser to Your PC
Insert the Standard USB connector on the USB cable into a USB port on your computer.
Open the Port Cover on the XRF Analyser.
Turn on the analyser and insert the mini USB connector on the USB cable into the USB port.
4. Upon initial installation, insert the NDT disk located behind the foam in your case. Follow the prompts and install
the USB driver located on the disk.
Connecting the XRF Analyser to Your PC (cont)
The USB cable
1. Insert the NDT CD and close out any dialogue box that
pops up. The driver is located on this disk.
2. Click on
“Control Panel”
and locate the
Manager”. If it is
not available
direct under
“Control Panel”,
look under
“System and
Security” then
3. Open “Device Manager”
4. Plug in instrument using the USB cable
5. Message will appear “Device Driver
Software Not Successfully Installed”
Connecting the XRF Analyser to Your PC (cont)
6. In “Device Manager”,
“Unknown Device” will appear
under “Other Devices”
7. Right click on “Unknown
Devices”; select “Update Driver
8. Click on “Browse My
Computer for Driver Software”
Connecting the XRF Analyser to Your PC (cont)
9. Click “Browse” button; select CD drive or
the location of the driver if you are not
installing from the NDT CD (recommended).
10. Click “OK”
11. Click “Next”
12. A Security Dialog Box will appear. Select
“Install This Driver Software Anyway?”
13. Driver will install; close out.
14. Insert the Standard USB connector on the
USB cable into a USB port on your computer.
Downloading Data
To download data you have collected offline:
Make sure that the XRF Analyser is connected to your computer.
Turn on the XRF Analyser.
Note - Wait at least 30 seconds after turning on the XRF Analyser
to begin downloading files. The System Start screens do not allow
Start the Niton Data Transfer program (NDT) from your PC.
Click the Download button.
The Download dialog box will open.
Downloading Data (cont)
In the Download dialog box, Select the Test button to test the serial
connection to the Analyser.
6. You should get a
pop-up window
informing you that the
connection tested
If the test fails, there
is a problem with
your serial port
setup. Press
‘settings’ to set the
correct com port and
try again
Downloading Data (cont)
In the Download dialog box, click the Connect button.
8. Click the
Query Readings
button. This will
return a list of all
current readings
on your analyser.
The list appears
in the large white
box in the
Download dialog
Downloading Data (cont)
Select the readings that you want to download. There are two ways to do this.
Downloading Data (cont)
a. Click the boxes next to each of the
reading numbers to select or de-select
individual readings. You can select a range
of readings by pressing the shift key, then
selecting the first and last reading in the
range. All readings from the first reading
selected to the last will then be selected.
b. When the readings are downloaded,
the available modes are also downloaded.
Click the boxes on the left to select or deselect all the readings of a specific
available type. You can also use the ShiftClick method of selecting a range of
readings as described above.
Downloading Data (cont)
The download generates a data file containing the selected readings. To
save the file for later use:
Enter the path for the file in the
Destination Folder field. You can use the ...
button to browse.
Enter a name for the file in the File
Name field.
Caution: - Some characters are not allowed in the file name. Characters
such as the "#" sign will cause an error. Niton recommends using only
alphanumeric characters "-", "_" and the space character when naming a
Downloading Data (cont)
Click the Download button.
When the progress bar shows that all the
readings are downloaded, click the Done
You should now see the readings you selected for download displayed, one reading
per horizontal line. The data has been saved to the folder and filename you
indicated prior to downloading.
Downloading Data (cont)
Error Messages while Downloading
Error Message
Error Message
Couldn't open \\.\COM7
Error Code: 2
Select another COM port.
WARNING: 38400 baud
rate not supported.
The port \\.\COM2 is in use
Select another COM port.
This indicates a potential problem. Test
the serial port. If there is a problem
connecting, switch baud rate on both the
analyser and the NDT software to 115200.
Please Open the Port
Click the Connect button.
Hardware Not Responding
Hardware Not Ready
Turn on the XRF Analyser.
Incorrect Data in reading #
Reading will be skipped.
Error code:
This indicates a version mismatch
between your instrument code and the
NDT code running on your computer. Use
a version of NDT that matches the version
number of the software on your analyser.
Incorrect Data in reading #
Reading will be skipped.
Error code:
This indicates a version mismatch
between your instrument code and the
NDT code running on your computer. Use
a version of NDT that matches the version
number of the software on your analyser.
Incorrect Data in reading #
Reading will be skipped.
Error code:
This indicates a version mismatch
between your instrument code and the
NDT code running on your computer. Use
a version of NDT that matches the version
number of the software on your analyser.
WARNING: 115200 baud
rate not supported.
This indicates a potential problem. Test
the serial port. If there is a problem
connecting, switch baud rate on both the
analyser and the NDT software to 38400.
If you are using a serial cable, check that
the cable is inserted snugly.
If you are using a serial cable, select the
other COM port.
If you are using the wireless USB adapter,
connect the serial port. See the "Installing
and Using Bluetooth" manual for complete
instructions on setting up the Bluetooth
adapter to work with your analyser.
Check that the spare battery is fully
The Serial Port connection
failed: RFCOMM
connection failed
Check that the battery is fully charged.
SH4 Successfully Communicating Result: SUCCESS;
Indicates normal working connection
Creating and Using Reports
Niton Data Transfer creates a new report whenever you download data.
To open an existing report
Click the Open button to open the
‘Open Existing Report’ dialog box.
Click the name of the report to
highlight it.
3. Click
the Open
Creating and Using Reports (cont)
4. The report is presented with one line per measurement. The report doesn't all
fit onto one screen, but you can scroll right if the fields you are interested in are not
visible. If they are still not visible read the following section on changing visible
You can copy readings from
one report file and paste them
into another file, or into a new
Select the readings you
want to copy, using the shift
or control keys as
Creating and Using Reports (cont)
Right click on the selected files, and select Copy Readings (Binary) from
the Pop-up Menu.
Open a destination report file to transfer
the readings into.
Creating and Using Reports (cont)
Right-click on any of the readings, and select Paste Readings (Binary) from the
Pop-up Menu.
Creating and Using Reports (cont)
The new readings will be highlighted in colour to distinguish them from the
readings already in place.
Creating and Using Reports (cont)
If you paste readings from yet another report file, they will be displayed in another
colour. Each group of readings copied in will have a different colour.
Creating and Using Reports (cont)
The colour distinctions will disappear when you save the file.
Creating and Using Reports (cont)
The original information is still there, though, and can be retrieved by right
clicking on a reading and selecting Report Properties.
Creating and Using Reports (cont)
Each analyser whose readings have been
pasted into the Compound Report will have
a separate tab in properties. The tab for
the analyser which was used to take the
reading you highlighted will be displayed
on top.
You cannot upload a compound file into an
analyser. Compound files are for data
correlation only.
Configuring Reports for Printing
Select ‘Customize’ from the ‘Tools’ Menu.
2. Select the
‘Report Print’ tab.
This brings up the Report Customization
window. Select the ‘certificate print’ tab
to change the certificate layout.
Customizing the Report
Like the Certificate Print option, the Report Print option adds company information,
company logo, headers, and footers to a standard report printout. The data input
into the various report fields is arbitrary - you can put anything you want to into any
of the fields. For example, you can type test location into the company address
field. The names given to the fields do not limit the data placed into the fields.
The Company Fields Font
Size Selector controls the
size of the font used in the
Company Name, Address,
and Location fields. The
Headers & Footers Font
Size Selector controls the
size of the font used in the
Header and Footer.
The report will print every
optional field selected in the
Print Options Checklist.
Customizing the Report
To use the Customize Report dialog box to add and remove columns:
Click the View Menu.
Click Customize Report...
3. Click the boxes next to
each of the column
headings to add or remove
each column from the
Customizing the Report
When a checked box is
selected, the check is
removed and the box
becomes empty.
If you want, you can
click the boxes in the
Selection area to
select or de-select a
group of headings.
Customizing the Report
If you want the changes to apply to
all the reports, click the box next to
Apply Global. If this box does not
have a check mark, the changes
only apply to the report that is
currently open.
Click the OK Button.
The columns appear in the order in which
they are listed in the Customize Report
Dialog Box. If you change the position of
the columns, and then use the Customize
Report Dialog Box, the columns return to
their original positions.
To change the position of a column, click and
drag the column heading. A shadow of the
heading will appear, indicating where the
heading will be placed.
Report Properties
Viewing the Report Properties
The report properties are:
• Analyzer Serial Number
• Report File Name
• File Version
• Report Download Date
• Total Readings in the Report
To view the report properties, click the
Report Properties button,
or select Properties from the View Menu.
Creating Passwords
Select the File menu
2. Select “New” then
“New Password File”.
Your screen should
look like this:
Creating Passwords (cont)
3. Create a unique security key, then
select the OK Button
This step is optional.
You can tick the
‘Disable User File
Security’ box to skip
this part.
Creating Passwords (cont)
Your screen should look like this:
Creating Passwords (cont)
Right click, then select “New User”
Creating Passwords (cont)
6. Enter a user name and password, then select the privileges assigned to this
user. Selecting the Check All check box will result in enabling all features.
Creating Passwords (cont)
Warning: it is recommended that only users at the highest level have access to the
“Safety” feature. This should be unchecked for all other operators.
Creating Passwords (cont)
Select the Update Button
You are now
ready to upload
your password
file to the
analyser. If you
wish to add more
user passwords
simply follow
steps 5 to 7.
Creating Passwords (cont)
8. Be sure the analyser is switched on; connect the analyser using USB
or serial connection.
9. Select the
Upload icon.
Creating Passwords (cont)
Your screen should look like this:
11. Select the Settings Button and
choose the comm port that your analyser
is connect to.
12. Select the Connect Button, then the
Upload Button.
13. Upon completion, you will receive a “File
Upload Successfully Completed” message.
14. Click the OK Button; save your password
file at this time by selecting the File icon then
“Save As.
15. Restart your analyser; your password
file should be successfully installed.
Changing Precision (Sigma Value)
Sigma is the symbol used for Standard Deviation, a measure of how much a set
of numbers deviates from the mean. For example, each of the three data sets
{0, 0, 14, and 14}, {0, 6, 8, and 14} and {6, 6, 8, 8} has a mean of 7. Their
standard deviations are 7, 5, and 1, respectively. The third set has a much
smaller standard deviation than the other two because its values are all close to
7. In a loose sense, the standard deviation tells us how far from the mean the
data points tend to be.
The number of standard deviations between the process mean and the nearest
specification limit is given in sigmas. As process standard deviation goes up, or
the mean of the process moves away from the center of the tolerance, the sigma
number goes down, because fewer standard deviations will then fit between the
mean and the nearest specification limit.
Changing Precision (Sigma Value) (cont)
Confidence Intervals
Confidence intervals assume that the data are from an approximately normally
distributed population - generally, sums of many independent, identically distributed
random variables tend towards the normal distribution as a limit. Using this
assumption, about 68 % of the values must be within 1 standard deviation of the
mean, about 95 % of the values must be within two standard deviations, about 99.7
% must lie within 3 standard deviations, and about 99.99% of the values must lie
within 4 standard deviations.
The greater the sigma value of the test, the more confident you can be that the
sample is as it appears, but the more difficult and time consuming the testing must
be to verify this. That's why it's important to use the most appropriate sigma value
for the test. By adjusting the sigma value for each type of test, you can optimize the
process for your needs.
Changing Precision (Sigma Value) (cont)
Adjusting the Sigma Values
The sigma values
are listed in the
column headed s.
The default value is
2 sigma. You can
change this value
by selecting the
down arrow next to
the value, which
opens up a dropdown menu from
which you can
select the desired
sigma value by
clicking on it.
When you have
changed the
sigma values to
the appropriate
number, select
the Save button
to save these
settings for use.
Select the Close
button to return
to the previous
screen without
saving any
Sorting the Element Display
Select the Sort
Element Display
icon to configure
sorting criteria
used for
analysis display.
Selecting the
Sort Element
Display icon
opens up the
Sort Criteria
Select the mode you wish to change, and the Sorting Options Screen will
appear (next slide).
Sorting the Element Display (cont)
On the left of the
display are elements,
each with its currently
selected display option
beside it to the right.
The element list is
ranked by importance,
with the most important
element on top, and
each one lower down of
less importance than
the one above it.
By selecting an element and using the arrow buttons to the right of the list, you can
change its ranking. Use the Top Arrow Button to move an element one rank closer to
the top with each click. Use the Bottom Arrow Button to move an element one rank
closer to the bottom with each click.
Sorting the Element Display (cont)
Display Options
The Display Options Drop Down Menu, allows you to change the display status
of any element to one of three states:
• Normal - The standard state. Element displays only when the elemental value
is greater than the limit of detection.
• Always - Always display the results for this element. Use this state for
elements critical to all of your analyses.
• Never - Never display the results for this element. Use this state for
elements which are unimportant to your work. This makes your instrument
display less complex.
Display Units
Select the Display Settings icon to choose between Karats and mg/g displays when taking readings, to set the max measurement
time, and to change the number of digits in the Karat rating for the reading.
Select the downward
pointing triangle next
to the Units field to
choose either the
Karat or the mg/g
display for the
The Set Display Units Menu Path
The Set Display Units Screen
Select the central Max Reading
Time button to set the maximum
reading time before the reading is
terminated. This defaults to 1000
seconds. The Max Reading Time
virtual keypad works similarly to
all virtual keypads, with the "C"
button clearing the field, the "<"
button backspacing over the last
digit, and the "E" button entering
the new time.
Select the downward pointing triangle next to the
Karat Display Digits field to choose either the 2
digit Karat Display (X.Y) or the single digit Karat
Display (X).
When you have changed the display units to the
appropriate values, select the Save button to save
these settings for use. Select the Close button to
return to the previous screen without saving any
Pseudo-Elements are constructs you can create which will be treated like elements in analysis, showing up in analysis as if they
were actual elements. In setting up a Pseudo Element, you must name it with at least 3 characters and at most 6 characters except you cannot use * (asterisk) and space. You may create up to 15 pseudo-elements, but you may not have more than 64
displayed values. You can set up equations with the following operands:
You can Add detected levels of elements or Pseudo Elements together or to a constant. For example: sum applicable elements
for FAC Analysis - (FAC = Cr+Cu+Mo), or sum elements for Residual Element Application - (REA = Cr+Ni+Cu).
You can Subtract detected levels of elements or Pseudo Elements from each other or a constant. For example: find the level of
clinker material C2S in cement - (C2S = (2.87*SiO2)-(0.754*Ca3SiO5))
You can Multiply detected levels of elements or Pseudo Elements by a constant. For example: Convert elemental data to oxide
form - (CaO = Ca*1.4).
You can Divide detected levels of elements or Pseudo Elements by a constant. For example: Convert Au concentration to Karat%
- (Karat = ((Au/100)*24))
You can use parentheses to stack order of calculation, as in the Karat and C2S examples above.
The Pseudo Elements Screen enables you to Create, Edit, Save,
and Delete Pseudo Elements. Below is a Pseudo Element for
correction of the gold readings due to Rh flashing which causes
attenuation of the gold signal. The correction is based on
assumptions that Rh plating will lower the gold signal by a
percentage of its displayed value, which will only work for low Rh
concentrations (<5%). We will call it Au-Rh.
To set up a new Pseudo
Element, select the New
Button. To Edit an existing
Pseudo Element, select the
Edit Button. To access the
library of saved Pseudo
Elements, select the Library
Button. To Delete an existing
Pseudo Element, select the
Delete Button. To Close the
window and return to the
previous screen, select the
Close Button.
1. Select the Keyboard Button to bring up the Virtual Keyboard.
2. Type in the
name you
chose to call
your Pseudo
5. Select the element you want to
add to the equation. It will appear
in the Current Element window.
4. Select the
triangle button
to open the
6. Select the Add Button to add
the Current Element to the
3. Select Enter.
7. Using the Virtual
Numeric Keypad integral
to the editor, type in the
other parts of the
equation. If other
elements are required,
they can be added as
needed from the dropdown list.
8. When the equation is
complete, select the Close
9. The Pseudo Element will
show up in the Mode Editor
The Finished Pseudo
Element in the Mode
Editor Window
Select the Pseudo
Element you want to
edit from the Mode
Editor window, then
select the Edit
This will open up the
Pseudo Element
Equation Editor with
the selected Pseudo
Element pre-loaded.
1. Select the
Pseudo Element
you want to
delete from the
Mode Editor
2. Select
the Delete
3. Select the
Close Button to
return to the
The Library
allows you to
easily access
your stored
Elements. To
add a Pseudo
Element to your
library, select
the Library
This will
display those
which are not
currently in
your library.
Select the
Element you
wish to add.
Select the Add Button.
Your new Pseudo Element
will display in your results

similar documents