Dental and Medical Problems

Dent Med Probl
Impact Factor (IF 2024) – 3.9
Journal Citation Indicator (JCI 2024) - 1.36
Scopus CiteScore (2024) – 5.0
Index Copernicus Value (ICV 2023) – 181.00
MNiSW – 70 pts
ISSN 1644-387X (print)
ISSN 2300-9020 (online)
Periodicity – bimonthly


 

Download original text (EN)

Dental and Medical Problems

2023, vol. 60, nr 4, October-December, p. 559–567

doi: 10.17219/dmp/152923

Publication type: original article

Language: English

License: Creative Commons Attribution 3.0 Unported (CC BY 3.0)

Download citation:

  • BIBTEX (JabRef, Mendeley)
  • RIS (Papers, Reference Manager, RefWorks, Zotero)

Cite as:


Arıcan Alpay B, Büyük C, Atav Ateş A. Comparison of gray-level detectability on computer monitors among several dental specialties: A web-based study. Dent Med Probl. 2023;60(4):559–567. doi:10.17219/dmp/152923

Comparison of gray-level detectability on computer monitors among several dental specialties: A web-based study

Burçin Arıcan Alpay1,A,B,C,D,E,F, Cansu Büyük2,A,B,C,D,E,F, Ayfer Atav Ateş3,A,B,C,D,E,F

1 Department of Endodontics, School of Dental Medicine, Bahçeşehir University, Istanbul, Turkey

2 Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Istanbul Okan University, Turkey

3 Department of Endodontics, Faculty of Dentistry, Istinye University, Istanbul, Turkey

Abstract

Background. Diagnosis in dentistry begins with the correct reading and interpreting of the dental radiograph.

Objectives. The aim of the present study was to examine the effects of the imaging technique used, the dentistry specialty and the years of experience on the gray-level perception among dentists.

Material and methods. A custom web application was developed. Dentomaxillofacial radiologists (DentRads), endodontists (Ends) and general dental practitioners (GDPs) were invited via e-mail to parti­cipate in the study. A total of 46 participants met the requirements of the test. The test comprised 2 webpages. On the 1st page, the participants were asked for information such as gender, specialty, the years of experience, and the imaging techniques they used. Then, on the 2nd page, they were welcomed with instructions and directions, and asked to rearrange 85 gray color tones represented by square bars of equal dimensions. These mixed gray bars were placed in 4 rows according to the principles of the Farnsworth–Munsell 100-hue test (FM). Each clinician’s test results were recorded in a database. The individual’s level of recognition of gray tones was evaluated through the total error score (TES), which was calculated using a web-based independent scoring software program. Lower TES values were a desirable result, indicating fewer misplacement, while higher scores indicated more misplacements of gray tones. The testing time (TT) was recorded automatically.

Results. The years of the participants’ experience as dentists or specialists did not affect TES or TT. The dentists who used the charge-coupled device-complementary metal oxide semiconductor (CCD-CMOS) had lower TES values than those who used analog radiographs (p < 0.05).

Conclusions. While the specialty and the years of experience did not affect the clinicians’ ability to recognize gray tones, the digital imaging techniques (photostimulable phosphor (PSP) and CCD/CMOS) could improve the clinicians’ gray-level perception.

Keywords: grayscale, dental radiology, contrast resolution

Introduction

Dental radiology procedures are among the most neces­sary and common ones in modern dentistry; they are used for the diagnosis of apical and periapical abnormalities, treatment plans, supporting treatment, the evaluation of care, and forensic situations.1 Therefore, radiographic image interpretation plays an important role at the forefront in all branches of dentistry before, during and after treatment. The ability to identify subtle differences in grayscale images, combined with good know­ledge of anatomy, often leads to a successful diagnosis.

Contrast resolution is defined as the number of possible shades of gray that a pixel can carry, which also expresses the bit depth. Digital images have a bit depth from 8 (256 shades of gray) to 16 (65,536 shades of gray).2 However, even if the monitor has a wide contrast range, the observer may not be able to perceive details.3 One of the most recognized methods in medical imaging used to measure the limits of the human visual system (HVS) is Barten’s model.4 It comprises a series of experiments at different luminance levels that characterize the contrast sensitivity of HVS. Based on Barten’s model, a calibration standard has been described for medical imaging monitors. The standard, called Digital Imaging and Communication in Medicine (DICOM®) Grayscale Standard Display Function (GSDF), states the precise display luminance that should be produced for a given input value.4

The GSDF provides that different monitors can be set to have the same grayscale response, resulting in the improvement of the perceptual linearity of observers.3 While GSDF calibration in medical-grade grayscale monitors is accomplished during the manufacturing process, for consumer-grade monitors, the calibration may be achieved with medical application software.5 However, considering that clinical dental practice requires relatively high room illuminance, and that medical-grade monitors and the additional software are costly, dental radiographs are most often interpreted on consumer-grade monitors.6 Furthermore, most of the affordable and practical imaging monitors support only 8 bits.4 Therefore, GSDF has not been greatly implemented.

In dentomaxillofacial radiology, contrast resolution refers to the dental image quality in terms of clear differences in gray shades with regard to enamel and dentin, root canal morphology, the pulp system, the marginal bone levels, air, and the trabecular bone pattern.7, 8 Dia­gnosing root fractures, resorptions or fractured root canal instruments, and determining the working length or accessory canals are some of the challenging radiological tasks for dentists, requiring the ability to differentiate fine details. In addition, the correct interpretation of the root canal system in terms of canal morphology and number of canals is fundamental for the successful outcome of endodontic therapies.8, 9 Thus, contrast sensiti­vity is an important competence when evaluating dental radiological images.10 Ganesan et al. concluded that the observer’s experience has a great impact on their visual search pattern.11

To evaluate the quality of the radiographic imaging system, there are some suggested methods, such as the perceptibility curve tests or the receiver operating cha­racteristic (ROC) curve tests.12 However, when it comes to evaluating the perception of the observer, there is no quantitative study in the literature that assesses the ability of dentists to differentiate gray tones. The Farnsworth–Munsell 100-hue test (FM) is widely used for color discrimination, including 85 colored caps.13 The accuracy of the observer in arranging the caps (forming a gradual transition in chroma) is gauged with the total error score (TES). The TES increases with the number of misplacements. Starting from this point, the authors of the present study claim that FM, which provides a numerical score, might be modified for the discrimination of gray tones.

Diagnosis in dentistry begins with the correct reading and interpretation of the dental radiograph.14 All dentists, whether specialists or not, use radiographs in their clini­cal routines. However, some branches of dentistry use radio­graphs not only for diagnosis, but also in all process steps. Endodontists (Ends) use radiographs pre-, during and post-treatment. Dentomaxillofacial radiologists (DentRads) routinely focus on normal and pathological differences on the scans obtained with cone-beam computed tomography (CBCT) and all other radiographic imaging systems in daily clinical practice. To the best of the authors’ knowledge, no studies have evaluated the ability of dentists to interpret contrast resolution. Therefore, the present study aimed to examine the influence of the imaging technique used, the dentistry specialty and the years of experience on the ability to perceive gray tones among dental professionals, as well as the testing time (TT). The null hypotheses were as follows:

1. There is no difference in the perception of gray tones and TT between the dentistry branches.

2. Neither the years of experience nor the imaging techniques used by clinicians have an effect on the interpretation of contrast resolution.

Material and methods

This laboratory study was conducted according to the Preferred Reporting Items for Laboratory studies in Endodontology (PRILE) 2021 guidelines.15 The study was approved by the institutional ethics committee at Istanbul Okan University, Turkey (approval No. 21.10.2020/24).

This research was a web-based study. Prior to the main study, 3 pilot studies were carried out. In the 1st and 2nd pilot studies, the participants were asked to rearrange 32 or 64 gray color bars from white to black. It was shown that they all reordered whitish and blackish colors almost faultlessly, but had problems arranging light to dark gray colors. Based on the outcomes of the pilot studies, the main study was reconstructed focusing on gray tones, which were a major problem for dentists. Then, the 3rd pilot study was applied to 20 people under the conditions of the main study, using 85 gray tones selected conse­quently (22–106) from a 128-tone (7 bits/pixel) grayscale. Afterward, the web page was blocked, and the pilot study was concluded. A reliability test was performed. Consider­ing the coefficient of correlation between the TES and TT values as a result of the 3rd pilot study (ρ = −0.581), the minimum number of participants to be included, with 95% confidence (1−α) and 95% test strength (1−β), was determined to be 27.

In light of these data, the individuals who participated in the pilot studies were not included in the main study, and other DentRads, Ends, and general dental practitioners (GDPs) were invited to participate. Invitations were sent to randomly selected members of relevant associations in Turkey via e-mail. A reminder e-mail was automatically sent to those who did not respond to the first e-mail within 10 days. Then, after 20 days, the system was closed.

The inclusion criteria for the respondents were as follows:

– the participant should be a GDP, End or DentRad;

– the participant should have had an eye examination in the last 18 months;

– the participant should not have any eye defects, or if they did, they should have been fixed with glasses or contact lenses;

– the participant should be using a MacBook Air laptop or be familiar with it.

The exclusion criteria for the respondents were as follows:

participants who did not visit the website after the 2 e-mails;

– participants who had an eye defect (glaucoma, cataract, etc.);

– participants who refused to use a MacBook Air laptop;

– participants who had been diagnosed with attention deficit hyperactivity disorder (ADHD).

A custom web application was developed by a web developer using HTML, CSS, JavaScript, and PHP. The application consisted of 2 webpages. On the 1st page, the participants were asked to provide personal information, including gender, specialty, the years of experience, and the imaging techniques (i.e., charge-coupled device-complementary metal oxide semiconductor (CCD-CMOS), photostimulable phosphor (PSP), analog) used in their daily clinical routines (Figure 1). In addition, they were directed to calibrate their monitors, using the link provided (macOS User Guide; https://support.apple.com/en-hk/guide/mac-help/mchlp1109/mac). The participants were not allowed to proceed to the 2nd page without confirming that the necessary calibration was performed. After providing informed consent to participate in the study, the participants were directed to the 2nd page.

The 2nd page included test instructions and directions (Figure 1). Initially, the participants were informed about the contents of the test. According to the instructions, the test was to be taken on a MacBook Air laptop (Apple Inc., California, USA) at a distance of approx. 60 cm from the monitor (3 spans). The participants were offered a MacBook Air laptop to use if they did not have one. The test design was inspired by FM. Eighty-five gray color tones represented with square bars of equal length and width were randomly placed in 4 rows according to the principles of FM, as follows: 22 bars in the 1st row (white-like gray) and 21 bars in the 2nd (light gray), 3rd (dark gray) and 4th (black-like gray) rows. These gray tones were selected consequently (22–106) from a 128-tone (7 bits/pixel) grayscale (Figure 1). The lightest (1–21) and darkest (107–128) gray tones were excluded. Each tone was numbered from 1 to 85, but the enumeration was not visible to the participants. On this page, the participants were asked to rearrange the bars from lightest to darkest in the correct order. They were only allowed to drag and change the bars in the same row.

The participants who got acquainted with the instructions were asked to start the test by clicking the “start” button. Then, when they completed the test, they were asked to click the “submit” button.

The test results of each clinician were recorded in a PHP database. The individual’s level of recognition of gray tones was evaluated through TES. The score was calculated using a web-based independent scoring software program for FM (the classical method), which was coded by Béla Török (https://www.torok.info/colorvision/fm100.htm) (Figure 2). Lower TES values were a desirable result, indicating fewer misplacements, while higher scores indicated more misplacements of gray tones. The testing time was recorded automatically.

Statistical analysis

Statistical analysis was performed using the IBM SPSS Statistics for Windows software, v. 23.0 (IBM Corp., Armonk, USA). The data was assessed for normality using the Shapiro–Wilk test. The one-way analysis of variance (ANOVA) was used to compare normally distributed data. The Mann–Whitney U test and the Kruskal–Wallis test compared non-normally distributed data. The results were presented as mean and standard deviation (M ±SD) for normally distributed data, and as median (minimum–maximum) (Me (min–max)) for non-normally distributed data. The significance level was set at p < 0.05.

Forty-six participants completed the study, and the power of the test (99.9%) was obtained as a result of the post-hoc power analysis.

Results

Sixty-eight dentists visited the website within 30 days. Sixteen visitors did not complete the test, 3 visitors sub­mitted their tests in less than 15 s, and the TES values of 3 other visitors were higher than 500. Therefore, the results of these 22 participants were discarded from the study. A total of 46 individuals met the requirements of the test. The numbers of respondents by dentistry branch were 16 GDPs, 15 Ends and 15 DentRads. Eighteen participants were male, and 28 were female. The TT values according to gender, branch, the years of experience, and the radio­graphic method used by the participants are shown in Table 1. The time to complete the test ranged from 280 s to 1,228 s. There was no correlation between TT and gender, the years of experience or the radiographic method used (p > 0.05). However, DentRads spent significantly more time than GDPs to complete the test (p < 0.05).

The mean TES values for the participants are shown in Table 2. Higher TES values represented more misplacements of the gray bars. The TES results ranged from 3 to 435. The TES results of the specialty groups were GDPs > Ends > DentRads, but the differences were not statistically significant (p > 0.05).

The participants used various imaging techniques in their clinical practice. The respondents were allowed to choose more than one answer to the question regarding the imaging methods. Five participants reported using more than one technique, which included PSP and CCD-CMOS (the hybrid group). While the first choice of Ends (n = 6; 40.00%) and DentRads (n = 13; 86.67%) was PSP, the first choice of GDPs (n = 7; 43.75%) was analog radio­graphy. In terms of TES, there were no significant differences between the PSP and hybrid groups as compared to the other groups. The dentists who used CCD-CMOS had significantly lower TES values than those who used analog radiographs (p < 0.05) (Table 2).

Although the lowest TES values were obtained by professionals with 0–5 years of experience, the years of the participants’ experience as dentists (Experience 1) or specialists (Experience 2) did not affect the TES results (p > 0.05) (Table 2). Also, no statistically significant differences were observed between specialists (DentRads and Ends) and GDPs regarding both the TES and TT values (p < 0.05) (Table 3).

The order of gray tones entered into the software can also be visualized by a ring chart (Figure 3). The misplaced gray bars are shown in red. Accordingly, the upper left part of the diagram represents the 1st row, the lower left part – the 2nd row, the lower right part – the 3rd row, and the upper right part – the 4th row. The participants often misplaced the bars representing the white-like gray and light gray colors. The 4th row (black-like gray) had the highest rate of correct placement.

Discussion

It is well known that radiographs are the main adjunct to the clinical examination, and dentists need them in both operative and preventive dentistry. Dentomaxillofacial radiologists and endodontists are the main groups that use radiographic imaging techniques in diagnosis, follow-up sections, in pre-treatment, treatment and post-treatment procedures. On the other hand, GDPs constitute the dominant group of dentists, dealing with all fields of dentistry, and they use all types of radiographs in their clinical routine. Identifying and improving the ability of dentists to detect gray tones, as well as determining the factors affecting this ability, would shorten the diagnostic process, reduce costs and positively affect the prognosis of treatment.16 Therefore, the present study aimed to compare the contrast resolution recognition abilities of the abovementioned 3 groups of dentists. According to the results, there was a difference between the dentistry branches with regard to TT, as DentRads spent significantly more time than GDPs to complete the test. The years of experience had no effect on TES or TT. However, the dentists who used CCD-CMOS had significantly lower TES values than those who used analog radiographs (p < 0.05). For this reason, the 1st null hypothesis was partially rejected, and the 2nd hypothesis was also partially rejected.

When the TES and TT values were analyzed, it turned out that DentRads spent the longest time completing the test, but no statistically significant differences were observed between this group and the other groups with regard to the TES values. Given that radiologists often view radiographs on larger screens in dark and private rooms, it is understandable that the focusing time will be longer on a 13-inch screen. In addition, it has been shown that the computer vision syndrome (CVS), which is defined as an eye and vision problem due to computer use,17 manifested mainly as difficulty with focusing, is quite common among oral radiologists.18, 19, 20 This situation and the prolonged focusing time may affect their work performance.

The lowest TES values were obtained in the group with 0–5 years of experience for both Experience 1 (as a dentist) and Experience 2 (as a specialist). Considering that the number of years of experience is directly proportional to the age of the person, it can be concluded that older participants had more difficulty with distinguishing and ranking gray tones. In addition, this computer-based study may have attracted more attention from the younger generation. Another possibility is that young participants may follow new developments and use digital systems more intensively. However, one of the limitations of this study may be that the participants were asked about their years of experience, not their age.

Radiographic interpretation could be influenced by both objective21, 22 and subjective factors.23, 24 Kamburoğlu et al. compared observer agreement between Ends and DentRads in the evaluation of endodontically and non-endo­dontically treated teeth for the presence of periapical lesions by using CBCT with 2 different voxel sizes.25 They reported that the voxel size and the specialty affected intra- and inter-observer reliability. The DentRads group showed better intra- and inter-observer agreement under all conditions.25 In this study, in which the contrast resolution perception ability of the participants was compared using the TES values, no significant differences were found between the dentistry branches. However, the DentRads group showed better results and lower SD values. The reason for this difference between these 2 stu­dies may be the number of specialists participating in the study. While only 2 DentRads and 2 Ends participated in the study by Kamburoğlu et al., the number of participants in our study was at least 15 for each group examined. In addition, since this study measured the dentists’ ability to recognize contrast resolution and not their radiographic diagnostic ability, the participants’ expertise may not have affected the outcomes.

There were several limitations to this study. The study was conducted amid the coronavirus disease 2019 (COVID-19) pandemic, when the world had come to a stand­still. People all around the world had to provide online services from their homes, including many professions that one may not have expected. Although this study was primarily meant as a face-to-face project, unfortunately, this design could not be implemented under COVID-19 conditions. In addition, considering a previous study that stated there were no differences between a computer-based FM (CBFM) and a manual FM (MFM) regarding TES and TT, and also suggested that CBFM was a very reliable, cost-effective and rapid method,26 our online study could also provide valuable outcomes.

The most challenging problem for these kinds of stu­dies is to design a test that can be standardizable and applicable using virtual methods. Ophthalmologists use various methods, such as the Spaeth/Richman contrast sensitivity test (SPARCS) and the Pelli–Robson (PR) chart test, in the assessment of contrast sensitivity, which are difficult to monitor and translate to the dentistry field.27 The FM test, which assesses HVS, is used for the detection of color blindness. It could be performed either via the Internet or by using physical derivatives. The data obtained from the test proposes several quantitative pieces of information. In this study, only the application protocol and design of the FM test were adapted.

To obtain standardization, several precautions were taken, including screen calibration, detailed guidance for the test environment and the use of the same brand of computer. Based on a study by Koenderink et al.,28 the distance from the monitor was determined as approx. 60 cm. The order of the gray bars in each row was established using www.random.org, and each participant saw the same random sequence. In this way, inter-participant standardization was ensured. The participants were also directed to a webpage for screen calibration. This webpage link was placed with a button for the ease of use on the 1st page of the application. Digital zoom was also auto­matically forbidden by the system. All participants were instructed to use laptops of the same model to standardize the factors related to the monitor and display resolution. MacBook Air laptops were provided to 5 participants who did not have one.

In the present study, the participants used MacBook Air laptop monitors rather than medical monitors or desktop monitors. Medical monitors, which have complex techno­logy, are excellent pathfinders for both medical and dental radiologists during diagnostic processes. However, physicians cannot access these functional monitors in their clinical practice because of their high cost. It was not preferred to use desktop monitor models, since they may have different resolution characteristics and cause calibration problems, and it was not possible to standardize monitor brands. The MacBook Air laptop with the Display Calibrator Assistant enables the control of screen calibration. It has support for millions of colors and a native resolution of 2,560 × 1,600 at a density of 227 pixels per inch. It automatically adjusts the white point of the user’s display to the color temperature of the environment for a more natural viewing experience. Due to their display features, MacBook Air laptops were used in the present study.

The main concern of the researchers was to supply suitable ambient light in the evaluation room for the participants. Although a high ambient light level (>1,000 lux) is almost always present in dental clinics, a low ambient light level (<50 lux) is recommended to enhance the dentist’s performance during the detection of caries.29, 30 It has been shown that reporting rooms should have good ambient light. It should be adjustable, as bright as the computer screen, and neither too bright nor too dark.19 Baltacıoğlu et al. reported that different ambient light levels did not affect the diagnostic ability of observers.31 In light of this information, ambient light was ignored in this study. However, the participants were instructed in the e-mail invitation to take the test under dim light conditions.

The respondents were asked about the imaging techniques they were using in their clinical routines, and the relationship between the method used and the TES result was assessed. It is broadly accepted that PSP plates have a wider dynamic range than CCD and CMOS sensors.32 Moreover, both digital systems (PSP plates and CCD/CMOS sensors) offer better contrast resolution than analog systems. On the other hand, the view box of analog films provides better luminance levels than typical digital monitors.32 Since the study was designed to assess the arrangement of digital data on the monitor, the lower TES results of the digital system users as compared to the analog system users were consistent.

Dental radiographs play a fundamental role in each step of treatment. However, it is not known whether the deve­lopment of digital radiography technology has improved clinicians’ ability to prepare a proper treatment plan.23 This is because we interpret digital radiographs rather than read them.33 Goldman et al. reported that many factors, such as the experience of the observer, the conditions in which radiographs are examined and the settings of the X-ray systems, could influence this interpretation.33 From this point of view, although the years of the participants’ experience were evaluated and the examining conditions were standardized in the current study, the results may not be generalizable when considering both observational and technical factors. Interpreting radiographic images poses challenges, particularly when discerning anatomi­cal malformations within the teeth. The fine details of dental structures may not be easily discernible from radio­graphs. Additionally, the presence of hard tissues can lead to superimpositions on the images, complicating the interpretation of the results. The designed test cannot represent these factors completely. Additionally, dentists normally do not calibrate their personal monitors every time they use them. For this reason, the present study may be a pathfinder for further studies that will focus on contrast resolution under clinical conditions.

Conlusions

Within the limitations of the present study, while the specialty and the years of experience did not affect the clinicians’ ability to recognize gray tones, the digital imag­ing techniques (PSP and CCD/CMOS) could improve the clinicians’ gray-level perception. Although the contrast sensitivity of the observer plays a fundamental role when evaluating dental radiographs, it is not the only factor that determines the outcomes. Further studies of the contrast resolution recognition ability of clinicians, supported by clinical cases, are needed.

Ethics approval and consent to participate

The study was approved by the institutional ethics committee at Istanbul Okan University, Turkey (approval No. 21.10.2020/24). All participants provided informed consent.

Data availability

The datasets generated and/or analyzed during the current study are available from the corresponding author on reasonable request.

Consent for publication

Not applicable.

Tables


Table 1. Intergroup comparisons of the testing time (TT) [s] at 95% confidence interval (CI)

Group

TT [s]

M ±SD

Me (min–max)

test statistic

p-value

Gender

male (n = 18)

571.7 ±270.1

480.5 (280–1,064)

U = 213.5

0.386

female (n = 28)

630.6 ±256.7

596.5 (281–1,228)

Branch

GDPs (n = 16)

535.4 ±260.9

435 (280–1,068)a

χ2 = 7.438

0.024

Ends (n = 15)

532.3 ±227.4

446 (286–1,074)ab

DentRads (n = 15)

759.7 ±238.5

769 (281–1,228)b

Radiographic imaging technique

analog (n = 9)

478.1 ±221.0

403 (280–983)

χ2 = 4.513

0.211

CCD-CMOS (n = 10)

595.4 ±196.7

630.5 (286–843)

PSP (n = 22)

688.1 ±292.4

701 (281–1,228)

hybrid (n = 5)

510.2 ±212.3

450 (286–831)

Work experience as a dentist
[years]

0–5 (n = 10)

647.7 ±261.9

677 (297–1,068)

F = 1.810

0.160

5–10 (n = 21)

676.2 ±290.6

750 (280–1,228)

10–15 (n = 9)

465.7 ±163.3

433 (286–831)

15–20 (n = 6)

513.0 ±193.9

476 (281–843)

Work experience as a specialist
[years]

0–5 (n = 18)

687.9 ±274.1

729 (281–1,228)

F = 0.585

0.564

5–10 (n = 8)

579.8 ±252.9

519 (286–1,074)

>10 (n = 4)

590.0 ±186.8

557 (403–843)

M – mean; SD – standard deviation; Me – median; min – minimum; max – maximum; GDP – general dental practitioner; End – endodontist; DentRad – dentomaxillofacial radiologist; CCD-CMOS – charge-coupled device-complementary metal oxide semiconductor; PSP – photostimulable phosphor; U – Mann–Whitney U test; χ2 – Kruskal–Wallis test; F – ANOVA; the values marked with different superscript letters were significantly different (p < 0.05).
Table 2. Intergroup comparisons of the total error score (TES) values at 95% confidence interval (CI)

Group

TES

M ±SD

Me (min–max)

test statistic

p-value

Gender

male (n = 18)

132.4 ±112.8

100 (6–435)

= 279.5

0.536

female (n = 28)

110.5 ±88.8

97.5 (3–309)

Branch

GDPs (n = 16)

151.2 ±125.2

143 (3–435)

F = 2.251

0.118

Ends (n = 15)

124.9 ±88.3

113 (6–309)

DentRads (n = 15)

78.9 ±59.2

53 (6–188)

Radiographic imaging technique

analog (n = 9)

196.9 ±82.7

190 (82–324)b

χ2 = 10.268

0.016

CCD-CMOS (n = 10)

91.5 ±130.8

35 (6–435)a

PSP (n = 22)

108.3 ±82.0

100 (3–309)ab

hybrid (n = 5)

81.4 ±57.0

85 (6–155)ab

Work experience as a dentist
[years]

0–5 (n = 10)

83.1 ±86.9

69 (3–280)

χ2 = 4.583

0.205

5–10 (n = 21)

124.7 ±117.4

85 (16–435)

10–15 (n = 9)

118.1 ±74.2

115 (6–251)

15–20 (n = 6)

160.8 ±67.4

171.5 (35–218)

Work experience as a specialist
[years]

0–5 (n = 18)

88.5 ±71.6

66 (6–309)

χ2 = 2.138

0.343

5–10 (n = 8)

108.6 ±88.9

112 (6–267)

>10 (n = 4)

149.0 ±80.2

171.5 (35–218)

The values marked with different superscript letters were significantly different (p < 0.05).
Table 3. Intergroup comparison of the testing time (TT) and total error score (TES) values

Group

TT [s]

TES

GDPs
(n = 16)

M ±SD

535.4 ±260.9

151.2 ±125.2

Me (min–max)

435 (280–1,068)

143 (3–435)

Specialists (n = 30)
Ends (n = 15)
DentRads (n = 15)

M ±SD

646.0 ±256.5

101.9 ±77.5

Me (min–max)

606 (281–1,228)

91.5 (6–309)

Test statistic

U = 172.0

U = 195.0

p-value

0.117

0.299

Figures


Fig. 1. First (A) and second (B) webpages of the application
Fig. 2. Data entry page of a web-based independent scoring software program coded by Béla Török (https://www.torok.info/colorvision/fm100.htm)
Fig. 3. Ring charts of the participants whose TES values were 3 (A), 53 (B), 151 (C), 280 (D), 324 (E), and 1,035 (F) (discarded from the study)
The red bars show the incorrect placements and their accumulations. For example, in Fig. 3A, gray tones numbered 40, 41 and 42 in the 2nd row were misplaced, and this area was shown as a single red row.

References (33)

  1. Supriyanti R, Setiadi AS, Ramadhani Y, Widodo HB. Point processing method for improving dental radiology image quality. Int J Electr Comput Eng. 2016;6(4):1587–1594. doi:10.11591/ijece.v6i4.9986
  2. Wenzel A, Møystad A. Work flow with digital intraoral radiography: A systematic review. Acta Odontol Scand. 2010;68(2):106–114. doi:10.3109/00016350903514426
  3. Kagadis GC, Walz-Flannigan A, Krupinski EA, et al. Medical imaging displays and their use in image interpretation. Radiographics. 2013;33(1):275–290. doi:10.1148/rg.331125096
  4. Kimpe T, Tuytschaever T. Increasing the number of gray shades in medical display systems – how much is enough? J Digit Imaging. 2007;20(4):422–432. doi:10.1007/s10278-006-1052-3
  5. Salazar AJ, Aguirre DA, Ocampo J, Camacho JC, Díaz XA. DICOM gray-scale standard display function: Clinical diagnostic accuracy of chest radiography in medical-grade gray-scale and consumer-grade color displays. AJR Am J Roentgenol. 2014;202(6):1272–1280. doi:10.2214/AJR.13.11509
  6. Kallio-Pulkkinen S, Haapea M, Liukkonen E, Huumonen S, Tervonen O, Nieminen MT. Comparison of consumer grade, tablet and 6MP-displays: Observer performance in detection of anatomical and pathological structures in panoramic radiographs. Oral Surg Oral Med Oral Pathol Oral Radiol. 2014;118(1):135–141. doi:10.1016/j.oooo.2014.04.005
  7. Aziman C, Hellén-Halme K, Shi XQ. A comparative study on image quality of two digital intraoral sensors. Dentomaxillofac Radiol. 2019;48(7):20190063. doi:10.1259/dmfr.20190063
  8. Reda R, Zanza A, Bhandi S, De Biase A, Testarelli L, Miccoli G. Surgical-anatomical evaluation of mandibular premolars by CBCT among the Italian population. Dent Med Probl. 2022;59(2):209–216. doi:10.17219/dmp/143546
  9. Nascimento EH, Nascimento MC, Gaêta-Araujo H, Fontenele RC, Freitas DQ. Root canal configuration and its relation with endodontic technical errors in premolar teeth: A CBCT analysis. Int Endod J. 2019;52(10):1410–1416. doi:10.1111/iej.13158
  10. Bruno MA. 256 shades of gray: Uncertainty and diagnostic error in radiology. Diagnosis (Berl). 2017;4(3):149–157. doi:10.1515/dx-2017-0006
  11. Ganesan A, Alakhras M, Brennan PC, Mello‐Thoms C. A review of factors influencing radiologists’ visual search behaviour. J Med Imaging Radiat Oncol. 2018;62(6):747–757. doi:10.1111/1754-9485.12798
  12. Yoshiura K. Image quality assessment of digital intraoral radiography – perception to caries diagnosis. Jpn Dent Sci Rev. 2012;48(1):42–47. doi:10.1016/j.jdsr.2011.09.001
  13. Farnsworth D. The Farnsworth–Munsell 100-Hue Test for the Examination of Color Discrimination. Baltimore, USA: Munsell Color Company; 1957:2–5.
  14. Kim IH, Singer SR, Hong DJ, Mupparapu M. Fundamentals of radiographic interpretation for the dentist. Dent Clin North Am. 2021;65(3):409–425. doi:10.1016/j.cden.2021.02.001
  15. Nagendrababu V, Murray PE, Ordinola‐Zapata R, et al. PRILE 2021 guidelines for reporting laboratory studies in endodontology: A consensus‐based development. Int Endod J. 2021;54(9):1482–1490. doi:10.1111/iej.13542
  16. Olchowy C, Olchowy A, Hadzik J, Dąbrowski P, Mierzwa D. Dentists can provide reliable shear wave elastography measurements of the stiffness of masseter muscles: A possible scenario for a faster diagnostic process. Adv Clin Exp Med. 2021;30(6):575–580. doi:10.17219/acem/134875
  17. Londoño-Candonaza FE, Fiori-Chincaro GA, Agudelo-Botero AM, Llaguno-Rubio J, Arriola-Guillén LE. Occupational health in oral radiologists: A review. Dent Med Probl. 2021;58(3):405–410. doi:10.17219/dmp/134789
  18. Teo C, Giffard P, Johnston V, Treleaven J. Computer vision symptoms in people with and without neck pain. Appl Ergon. 2019;80:50–56. doi:10.1016/j.apergo.2019.04.010
  19. Atwal SS, Prasad A, Deepak D, Agarwal K. Health issues among radiologists: Toll they pay to their profession. J Clin Diagn Res. 2017;11(4):TM01–TM02. doi:10.7860/JCDR/2017/17023.9537
  20. Ranasinghe P, Wathurapatha WS, Perera YS, et al. Computer vision syndrome among computer office workers in a developing country: An evaluation of prevalence and risk factors. BMC Res Notes. 2016;9:150. doi:10.1186/s13104-016-1962-1
  21. Neves FS, Freitas DQ, Flores Campos PS, Ekestubbe A, Lofthag-Hansen S. Evaluation of cone-beam computed tomography in the diagnosis of vertical root fractures: The influence of imaging modes and root canal materials. J Endod. 2014;40(10):1530–1536. doi:10.1016/j.joen.2014.06.012
  22. Brito-Júnior M, Santos LA, Faria-e-Silva AL, Pereira RD, Sousa-Neto MD. Ex vivo evaluation of artifacts mimicking fracture lines on cone-beam computed tomography produced by different root canal sealers. Int Endod J. 2014;47(1):26–31. doi:10.1111/iej.12121
  23. Tewary S, Luzzo J, Hartwell G. Endodontic radiography: Who is reading the digital radiograph? J Endod. 2011;37(7):919–921. doi:10.1016/j.joen.2011.02.027
  24. Patel S, Dawood A, Mannocci F, Wilson R, Ford TP. Detection of periapical bone defects in human jaws using cone beam computed tomography and intraoral radiography. Int Endod J. 2009;42(6):507–515. doi:10.1111/j.1365-2591.2008.01538.x
  25. Kamburoğlu K, Koç C, Sönmez G, Elbahary S, Rosen E, Tsesis I. Effect of cone beam CT voxel size and dental specialty status on the agreement of observers in the detection and measurement of periapical lesions. Oral Surg Oral Med Oral Pathol Oral Radiol. 2021;132(3):346–351 doi:10.1016/j.oooo.2021.04.053
  26. Ghose S, Parmar T, Dada T, Vanathi M, Sharma S. A new computer-based Farnsworth Munsell 100-hue test for evaluation of color vision. Int Ophthalmol. 2014;34(4):747–751. doi:10.1007/s10792-013-9865-9
  27. Habtamu E, Bastawrous A, Bolster NM, et al. Development and validation of a smartphone-based contrast sensitivity test. Trans Vis Sci Technol. 2019;8(5):13. doi:10.1167/tvst.8.5.13
  28. Koenderink J, Van Doorn A, Gegenfurtner K. Colour order. I-Perception. 2019;10(5):2041669519872516. doi:10.1177/2041669519872516
  29. Haak R, Wicht M, Hellmich M, Nowak G, Noack M. Influence of room lighting on grey-scale perception with a CRT and a TFT monitor display. Dentomaxillofac Radiol. 2002;31(3):193–197. doi:10.1038/sj/dmfr/4600668
  30. Hellén-Halme K, Lith A. Carious lesions: Diagnostic accuracy using pre-calibrated monitor in various ambient light levels: An in vitro study. Dentomaxillofac Radiol. 2013;42(8):20130071. doi:10.1259/dmfr.20130071
  31. Baltacıoĝlu İH, Eren H, Yavuz Y, Kamburoğlu K. Diagnostic accuracy of different display types in detection of recurrent caries under restorations by using CBCT. Dentomaxillofac Radiol. 2016;45(6):20160099. doi:10.1259/dmfr.20160099
  32. White SC and Pharoah MJ. Oral Radiology: Principles and Interpretation. 7th ed. St. Louis, MO: Elsevier; 2014:78–99.
  33. Goldman M, Pearson AH, Darzenta N. Endodontic success – who’s reading the radiograph? Oral Surg Oral Med Oral Pathol. 1972;33(3):432–437. doi:10.1016/0030-4220(72)90473-2