eISSN: 1897-4309
ISSN: 1428-2526
Contemporary Oncology/Współczesna Onkologia
Current issue Archive Manuscripts accepted About the journal Supplements Addendum Special Issues Editorial board Reviewers Abstracting and indexing Subscription Contact Instructions for authors Publication charge Ethical standards and procedures
Editorial System
Submit your Manuscript
SCImago Journal & Country Rank
1/2024
vol. 28
 
Share:
Share:
Original paper

Deep feature extraction and fine κ-nearest neighbour for enhanced human papillomavirus detection in cervical cancer – a comprehensive analysis of colposcopy images

Lipsarani Jena
1, 2
,
Santi Kumari Behera
1
,
Srikanta Dash
3
,
Prabira Kumar Sethy
3, 4

  1. Veer Surendra Sai University of Technology, Burla, India
  2. GITA Autonomous College, Bhubaneswar, India
  3. Sambalpur University, India
  4. Guru Ghasidas Vishwavidyalaya, Bilaspur, C.G., India
Contemp Oncol (Pozn) 2024; 28 (1): 37–44
Online publish date: 2024/04/26
Article file
Get citation
 
PlumX metrics:
 

Introduction

Cervical cancer, predominantly initiated by persistent human papillomavirus (HPV) infection acquired through sexual contact, is a formidable global health challenge. Human papillomavirus is the primary aetiological factor, responsible for 12% of all malignancies and ranking as the second leading cause of death among women worldwide [1]. This association between HPV and cervical cancer is particularly pronounced, with certain types of HPV being categorised as high-risk due to their potential to trigger cervical cancer development. Among these high-risk types, HPV-16 and HPV-18 have emerged as prominent culprits, leading to persistent infections that significantly contribute to the onset of cervical cancer [2].

The impact of HPV infection on cervical cancer risk is further exacerbated in women living with HIV, who face a six-fold increase in susceptibility to cervical cancer compared with their HIV-negative counterparts. This underscores the intricate interplay between HPV and immunocompromised states, emphasising the urgent need to address cervical cancer risk in vulnerable populations [1].

Human papillomavirus, a diverse group of viruses with the capacity to infect not only the genital area, but also the mouth and throat, primarily spreads through sexual contact. Despite the prevalence of HPV exposure in sexually active populations, the immune system typically clears the infection in most cases. However, persistent infections may occur when the immune system fails to eliminate the virus, thus paving the way for the development of abnormal cervical cells. Progression from persistent infection to cervical cancer is a well-established continuum that underscores the critical importance of timely detection and intervention [2].

This research article delves into the crucial nexus between HPV infection and cervical cancer, exploring avenues to enhance understanding, detection, and, ultimately, the prevention of this devastating disease. By investigating novel methodologies for HPV classification in colposcopy images, this study aimed to provide valuable insights into the existing body of knowledge in medical imaging and cancer detection, addressing a notable gap in current research endeavours.

Cervical cancer is strongly associated with certain types of HPV. Human papillomavirus is a group of related viruses that can infect the genital area as well as the mouth and throat. Among the many types of HPV, some are considered high risk because they can lead to the development of cervical cancer. The link between HPV and cervical cancer is well established. Persistent infection with high-risk HPV types, particularly HPV-16 and HPV-18, is the leading cause of cervical cancer. Human papillomavirus is primarily transmitted through sexual contact, and most sexually active individuals are exposed to the virus at some point in their life. However, in most cases, the immune system clears the infection. Persistent infection can occur when the immune system fails to eliminate the virus. Over time, this can lead to the development of abnormal cells in the cervix, which can progress to cervical cancer if not detected and treated [3].

Preventive measures, including the administration of HPV vaccines, are advised to reduce the risk of cervical cancer associated with HPV in both males and females. Vaccines targeting the prevalent high-risk genotypes of HPV, namely HPV-16 and HPV-18, confer protection against the disease and substantially reduce the risk of developing cervical cancer. Cervical cancer screening, which encompasses HPV testing and Pap examinations, improves the efficacy of treatment by facilitating early detection of cervical abnormalities. Consistent gynaecological screening and prompt medical attention in response to unusual symptoms or test results can aid in the detection and treatment of cervical abnormalities and cancer in their early stages. The public must understand the connection between HPV and cervical cancer if they are motivated to participate in screening programs and take preventative measures. To mitigate the potential for HPV-associated cervical cancer, it is imperative to maintain regular communication with healthcare professionals and rigorously adhere to prescribed vaccination and screening protocols. Some samples depicted in Figure 1 tested positive or negative for HPV.

Fig. 1

Colposcopy image sample of human papillomavirus positive (A) and negative (B)

/f/fulltexts/WO/54021/WO-28-54021-g001_min.jpg

Various techniques are employed to detect HPV, including biopsies, RNA and DNA testing, visual examinations using acetic acid or Lugol’s iodine, and Pap smears. A healthcare professional uses a spatula or brush to collect cells from the cervical region during a Pap smear. The collected cells are examined for irregularities under a microscope. Standard screening may potentially incorporate an HPV DNA test, which examines cell samples concurrently with cervical cancer risk assessment for high-risk HPV DNA. As an indicator of viral replication, the presence of active HPV infection in the cells was ascertained by RNA detection. Visual investigation using acetic acid or Lugol’s iodine is feasible in contexts where resources are limited. The application of a solution to the cervix may induce discernible alterations in colour or texture, which may indicate the existence of abnormalities. Biopsies are performed when abnormalities are detected during the screening process. Collecting a tissue sample from the site of interest (e.g. the cervix) for microscopic examination is the initial step. The primary objective of this test is to assess the existence of HPV-associated abnormalities and provide comprehensive data regarding their severity. Early detection and prevention of cervical cancer are dependent on screening procedures.

Numerous variables have the potential to impact cervical cancer screening methods, including the Pap smear and HPV testing. An absence of understanding regarding the importance of screening and the correlation between HPV and cervical cancer may contribute to hesitancy. Fear, anxiety, distress, unease, and potential for cancer diagnosis are typical responses to cancer screening. Avoidance behaviours can be impacted by prevailing cultural and community norms, in addition to systemic taboos concerning reproductive health. Sexually transmitted infections may cause individuals to be reluctant to seek treatment due to concerns such as stigma, humiliation, and invasion of privacy. Obstacles to the timely delivery of screening services may include financial limitations, geographical restrictions, and scarcity of healthcare services. Frequent misunderstandings regarding the screening procedure and its objectives can cause some to forsake it. Additionally, individuals may become complacent if they believe that they are not at risk. Education, public health campaigns, and community outreach programs should be implemented to address these issues. This is essential for increasing awareness, refuting misconceptions, enhancing early detection and prevention programs, and increasing the cervical cancer screening rates.

Because of significant limitations in infrastructure, resources, and financial means, the installation of cervical cancer screening apparatus in every healthcare facility situated in remote regions is not viable. Thus, colposcopy represents the most pragmatic and effective approach for establishing economical screening initiatives for HPV detection. To accurately diagnose cervical cancer in remote rural areas, paramedical professionals must be able to identify HPV using simplified, practical, and readily comprehensible methods. Consequently, an automated method is necessary for HPV detection in cervical cancer patients.

The major contributions of this study are as follows:

  • colposcopy images were utilised in a novel manner to classify HPV, as described in this study. Detection of cervical cancer, the second most prevalent malignancy among women worldwide, is the primary objective of this approach. This is an innovative method for classifying HPVs,

  • by investigating the uncharted domain of HPV-based colposcopy image identification in the context of cervical cancer, this study fills a substantial void in the existing body of knowledge. This novel perspective contributes to our understanding of the potential applications of medical imaging in the context of cervical cancer screening,

  • low-income and underdeveloped regions can benefit from colposcopy screening, according to the study, because it does not require biopsy specimens and only requires a compact, cost-effective setup. Within healthcare systems characterised by resource constraints, this attribute underscores the potential for extensive integration,

  • within the methodological framework, feature extraction and robust dataset augmentation were performed using the EfficientNetB0 architecture. An exceptionally precise and effective classification system was generated through the experimental integration of 19 distinct convolutional neural network (CNN) architectures, followed by fine-tuning using the fine κ-nearest neighbour (KNN algorithm),

  • the proposed technique shows remarkable performance in classification, as shown by its validation accuracy of 99.9% and area under the curve (AUC) of 99.86%. The system maintained strong performance when evaluated using test data, achieving an accuracy of 91.4% and an AUC of 91.76%. The efficacy of the integrated approach is substantiated by these remarkable outcomes, which provide a reliable methodology for HPV classification in colposcopy images; this has the potential to enhance cervical cancer diagnosis on a global scale.

Material and methods

Related work

Numerous investigations have significantly advanced the domain of computer-aided diagnosis (CAD) in cervical pathology by employing distinct methodologies to enhance the precision of lesion identification and classification. Zhang et al. [4] introduced an innovative CAD technique targeted at automatically identifying cervical precancerous lesions, with a particular emphasis on distinguishing cervical intraepithelial neoplasia 2 (CIN2) or higher-level lesions in cervical images. Their methodology incorporated a comprehensive pre-processing stage encompassing data augmentation and extraction of regions of interest. A remarkable accuracy of 73.08% and an AUC of 0.75% were attained on a test set comprising 600 images through the utilisation of pre-trained DenseNet convolutional neural networks to fine-tune the parameters of each layer. To enhance the diagnostic efficiency of positive low-grade squamous intraepithelial conditions such as cervical cancer and CIN, Li et al. [5] examined the application of time-lapse colposcopic images. Utilising feature fusion and key-frame feature-encoding networks, they constructed their system. The aforementioned networks encoded the characteristics extracted from colposcopic images captured prior to and at various points during the acetic acid test. The researchers achieved a noteworthy classification accuracy of 78.33% using a graph convolutional network with edge features to process a massive dataset comprising 7668 images. Yu et al. [6] introduced an innovative approach to the analysis of colposcopy images by incorporating time-series parameters. The researchers achieved a sensitivity of 95.68%, specificity of 987.22%, and accuracy of 96.72% with their gated recurrent convolutional neural network (C-GCNN) when integrating multistate cervical images for CIN grading. To enhance the DenseNet model’s performance in classifying cervical cancer through the utilisation of colposcopy images, Saini et al. [7] developed ColpoNet. Their framework achieved an accuracy of 81.353% when applied to a dataset consisting of 400 CIN1 and 400 CIN2/CIN3/CIN4 images. Xu et al. [8] resolved the issue of image preprocessing for the classification of cervical neoplasia through the implementation of a two-way augmentation technique. The methodology successfully categorised cervical lesions with an impressive 82.31% accuracy using an optimised support vector machine and CNN models. Cho et al. [9] classified neoplasms in 5 ways and pre-processed images of considerable complexity on a heterogeneous dataset. By implementing ResNet-15, they achieved notable advancement in the discipline, attaining an accuracy rate of 51.7%.

Although these studies have significantly advanced the state of the art in colposcopy image analysis, it is noteworthy that the existing literature predominantly focuses on the classification of various cancer types. Surprisingly, there is a discernible gap in research on the detection of HPV using colposcopy images. The present study sought to fill this gap by introducing a novel methodology explicitly designed for HPV classification, thereby expanding the scope of research in medical imaging and cancer detection. Table 1 provides a comparison of existing work in which only colposcopy and neoplasia images are taken for classification.

Table 1

Comparison of existing work

AuthorMethodologyResultsConclusions
Zhang et al. [4]CAD approach, DenseNet CNN, ROI extraction, data augmentationAccuracy: 73.08%, AUC: 0.75Successful automatic identification of cervical pre-cancerous lesions, distinguishing CIN2 or higher
Li et al. [5]Deep learning framework, key-frame feature encoding, feature fusion, E-GCNClassification accuracy: 78.33%Efficient diagnosis of LSIL+ using time-lapsed colposcopic images. E-GCN proved to be suitable for feature fusion
Yu et al. [6]Gated recurrent CNN, time series considerations, multistate cervical images mergingAccuracy: 96.87%, sensitivity: 95.68%, specificity: 98.72%High accuracy in CIN grading using colposcopy images and time series considerations
Saini et al. [7]ColpoNet architecture (enhanced DenseNet model)Accuracy: 81.353%Successful cervical cancer classification using the ColpoNet model
Xu et al. [8]Two-way augmentation, CNN, SVMAccuracy: 82.31%Effective two-way augmentation for preprocessing, achieving high accuracy in cervical neoplasia classification
Cho et al. [9]Image pre-processing, ResNet-15, five-way neoplasia classificationAccuracy: 51.7%Neoplasia classification using ResNet-15, with a moderate accuracy of 51.7%

[i] AUC – area under the curve, CAD – computer-aided diagnosis, CIN – cervical intraepithelial neoplasia, CNN – convolutional neural network, E-GCN – graph convolutional network with edge features, ROI –regions of interest, LSIL+ – low-grade squamous intraepithelial, SVM – support vector machine

About dataset

This dataset was sourced from the International Agency for Cancer Research (IACR) upon request. It contains samples of normal saline, acetic acid, acetic acid with a green filter, and Lugol’s iodine. In this study, samples after acetic acid treatment were considered.

Methodology

Image pre-processing

The dataset utilised in this research, obtained from IACR, comprises 2760 images categorised as either HPV positive or HPV negative. To effectively train the deep learning-based image classification model, 80% of the dataset was allocated for training, and the remaining 20% was reserved for testing. Acquiring a substantial volume of images, particularly within the specialised domain of colposcopy, presents a challenge. Consequently, the dataset underwent augmentation procedures, including flipping, rotation, contrast variation, brightness variation, random rotation, and translation, as listed in Table 2. These operations expanded the dataset five-fold by incorporating horizontal and vertical flips and left and right rotations. To address balance, adjustments were made to reach 690 samples for both HPV-positive and HPV-negative categories. The images were resized to 224 × 224 pixels in accordance with the preprocessing requirements of the model. The final dataset comprised 2760 samples, of which the proportion of HPV-positive and HPV-negative samples was balanced. The dataset provides an extensive and varied assortment that is well-suited for rigorous training and evaluation of the image classification model.

Table 2

Details of dataset

HPVOriginalBalanceAugmentation (left, right, horizontal, vertical)Augmentation (brightness, contrast, intensity)
Positive1381386902760
Negative1511386902760
Total      5520

[i] HPV – human papillomavirus

Feature extraction using EfficientNetB0

Feature extraction, a critical stage in data analysis, involves the conversion of unprocessed inputs into a format that is more comprehensible to users [10]. The extraction of HC features – image qualities that are precisely defined according to the particular attributes of the target region – is an inherent component of this procedure. Because of their straightforward extraction process, HC features have been widely employed by researchers, particularly in situations involving limited datasets. By employing these attributes, which were formulated in conjunction with domain specialists, a distinct training dataset was obtained, resulting in a visual representation that was more lucid than that of alternative attributes. An additional way to increase the desirability of particular attributes is to assign them varying degrees of importance in relation to the attainment of the goal. However, these characteristics may be difficult to discern when accompanied by complex visuals. This necessitates the implementation of an alternative approach to feature extraction, i.e. the application of deep learning models (DLMs). In particular, the significance of DLMs and CNNs has increased substantially over the past decade owing to advancements in processing capacity. This development has facilitated the training of more intricate networks, leading to an increase in the output while reducing the required time. Convolutional neural networks, which are employed as feature extractors, can extract features independently from input images. The versatility of the feature extraction methods is demonstrated by the utilisation of both handcrafted features and DLMs to meet the demands of various datasets and analytical purposes. To obtain optimal attributes using this method necessitates extensive training and considerable variation; however, the utilisation of DLMs offers the benefit of autonomously executing crucial preparatory tasks [11]. This image depicts the intricate design of EfficientNetB0. EfficiencyNet topologies optimise performance [12] by balancing scalability in accordance with available resources. In contrast to constructing EfficientNet topologies within a predetermined budget and subsequently adjusting them with additional resources, scalability is maintained for depth, breadth, and resolution. By adopting this methodology, the design could optimise the utilisation of its resources. Consequently, they erect succession of more intricate structures. The most fundamental form of architecture, EfficientNet-B0, incorporates 16 MobileNet [13] inverted bottleneck MBConv layers in 3 or 5 dimensions, following a convolution layer for feature extraction. Subsequently, fully linked, convolution, and pooling layers were implemented. EfficientNet-B0 streamlines and improves the training procedure by utilising its 5.3 million trainable parameters.

Classification using K-nearest neighbour

A basic neural network classifier assigns each pixel to the same class as the training data with the nearest intensity. However, this method may lead to errors if the nearest neighbour is an outlier in another class. To enhance the robustness of this approach, the KNN classifier was employed, where K patterns were considered. The κ-nearest neighbour classifier is categorised as a non-parametric classifier because it does not assume any specific statistical structure for the data [14]. The steps of the KNN algorithm comprise including classes in the training set, examining K items near the item to be classified, and placing the new item in the class with the highest number of nearby items. This process has a time complexity of O(q) for each tuple to be classified, where q represents the training set size. The κ-nearest neighbour classification utilises a training dataset with both the input and target variables. By comparing the test data, which only contain input variables, to this reference set, the distance of the unknown to its K-nearest neighbours determines its class assignment. This assignment can be achieved either by averaging the class numbers of the K-nearest reference points or by obtaining a majority vote for them [15].

Feature vectors are analogous to data points when the KNN algorithm is used to classify images. X represents the feature vector of the image requiring categorisation, whereas Xi denotes the feature vector of each individual image in the training set.

The Euclidean distance (d (X, Xi) is computed between X, and each Xi, using the formula:

dX,Xi=j=1nXjXij2 1

In this instance, n represents the number of features in the image. The algorithm identifies K-nearest companions by selecting images with minimum distances. In the context of image classification, the class label for the image is determined by a majority vote among its nearest neighbours. The mathematical representation is as follows:

y=argmaxi=1KIyi=c2

where the indicator function is denoted as I(.), yi is the class label of the ith neighbouring node, and c is a class label. An effective and versatile technique for picture classification that does not impose strict assumptions about the underlying data distribution is provided by this method, which enables the KNN classifier to classify an image according to the predominant class within its K-closest neighbours. Fine KNN is a kind of nearest-neighbour classifier that uses a single neighbour for classification, allowing it to make subtle differences between classes.

Colposcopy models were employed in a methodical fashion to categorise HPV images. An extensive dataset was initially compiled and subsequently augmented to ensure a robust model training and to enhance diversity. Following this, we trained the EfficientNetB0 architecture, which is renowned for its efficient data collection of pertinent information, to extract features from colposcopy images using the improved dataset. A comparative analysis of 19 distinct CNN configurations was conducted to ascertain the optimal CNN for the given task. The selection criteria for identifying a model that performs exceptionally well in HPV categorisation were determined by employing accuracy metrics. Following this, the chosen model was modified by utilising the KNN classification algorithm. Precise predictions were generated by identifying patterns within the feature space using this approach. A robust and precise HPV classification system was ensured through the implementation of this all-encompassing strategy, which incorporated data augmentation, effective feature extraction, model selection, and KNN algorithm fine- tuning. Figure 3 illustrates a simplified description of the proposed methodology.

Fig. 2

Architecture of EfficientNet-B0 as feature extractor

/f/fulltexts/WO/54021/WO-28-54021-g002_min.jpg
Fig. 3

Block diagram for proposed classification model

/f/fulltexts/WO/54021/WO-28-54021-g003_min.jpg

Data availability

The data are available on genuine request. The code generated and/or analysed during the current study is available from the corresponding author upon reasonable request.

Results

The described approach was executed on an HP Victus laptop equipped with a 12th generation Intel Core i7 CPU operating on a Windows 11 platform. The laptop featured an integrated NVIDIA graphics processing unit, and the analysis was performed using the MATLAB 2022a software. To evaluate various CNN classifiers, we examined their confusion matrix metrics, including accuracy, sensitivity, specificity, precision, false-positive rate, F1 score, Matthew’s correlation coefficient (MCC), and κ-values. The augmented dataset was randomly split into training (80%) and testing (20%) sets to assess model performance. To ensure the classifier’s ability to generalise to unseen patients, dataset splits were randomised. The results reported in this study represent the average outcomes of 10 independent runs. ROC curves were generated using the results of these 10 runs. Table 3 displays the performance metrics of 13 CNN models alongside the proposed model.

Table 3

Performance comparison of convolutional neural network models for classification of human papillomavirus in percentage

Model nameAccuracySensitivitySpecificityPrecisionFPRF1 scoreMCCκ-value
AlexNet76.9968.6885.5182.5314.4974.85547953.99
Vgg1680.8978.0883.7082.7316.3080.34618761.78
Vgg1979.3564.1394.5792.1905.4375.64616258.70
ResNet1865.4932.6198.3795.2416.3048.58411230.98
ResNet5086.1491.6780.6282.5419.3886.8772.7372.28
ResNet10186.6892.0381.3483.1418.6687.3673.7973.37
DenseNet20184.1572.9395.4794.154.5381.1270.1268.30
Squeeze Net65.1332.7997.4692.822.5448.4639.6730.25
Darknet1983.4294.3872.4677.417.5485.0668.5166.85
Darknet5378.9992.0365.9472.9904.0681.4160.0560.05
Xception87.3285.6988.9588.5811.5087.1174.6874.64
InceptionResnetV283.3375.0091.6790.0008.3381.8267.6166.60
InceptionV382.4368.3066.5695.2003.7479.5467.6164.86
NasnetLarge87.0577.1796.9296.1603.7385.6375.5874.09
NasnetMobile74.5595.4753.6267.3104.6378.9554.0649.09
MobileNetV286.2385.3387.1486.9012.8686.1172.4872.46
Google Net72.8392.2153.4466.4506.5677.2445.6545.65
EfficientNetB (0)92.7590.9496.2095.8203.4392.6285.5185.51

[i] FPR – false-positive rate, MCC – Matthew’s correlation coefficient

The confusion matrix and AUC for the validation dataset are presented in Figure 4. The results revealed an impressive accuracy of 99.9%, indicating exceptional classification performance of the model. Additionally, the AUC for validation was 99.86%, emphasising the robustness of the model in assessing the ROC curve. These metrics, displayed graphically, affirm the efficacy and reliability of the proposed methodology for HPV classification using colposcopy images.

Fig. 4

Performance of validation data confusion matrix (A), area under the curve (B)

/f/fulltexts/WO/54021/WO-28-54021-g004_min.jpg

Discussion

A concise overview of the performance of the model on the test dataset is presented in Figure 5, which incorporates the confusion matrix and AUC. The 91.4% accuracy rate of the model demonstrates its utility in the classification of assignments. Furthermore, the efficacy of the model in analysing the ROC curve was underscored by a test data AUC of 91.76%. The results are shown in Figure 5. These results demonstrated that the proposed method for classifying HPVs using colposcopy images is dependable and efficient, particularly when evaluated using fresh data.

Fig. 5

Performance of test data confusion matrix (A), area under the curve (B)

/f/fulltexts/WO/54021/WO-28-54021-g005_min.jpg

Conclusions

To date, there have been no reports of cervical cancer diagnosed using HPV extracted from colposcopy images. However, this study addressed the feasibility of cervical cancer screening in rural areas of developing and low-income countries through the introduction of an innovative analytical framework that utilises colposcopy images to classify HPV. Colposcopy, which does not require biopsy specimens for processing and has a compact and cost- effective setup, is a feasible screening alternative, particularly in resource-constrained regions. The methodology of this research is characterised by a systematic approach, commencing with comprehensive dataset augmentation to enhance the robustness and breadth of the model. Following this, features were extracted using the efficient and well-known EfficientNetB0 architecture. This architectural design excels at gathering critical image attributes. Following the evaluation of 19 unique CNN architectures, we successfully identified a model that exhibited exceptional accuracy during the model selection procedure. The fine KNN method can be employed to refine the classification capabilities of the model, enabling indistinct distinctions between classes with only one neighbouring class. The proposed methodology yielded exceptional results, as shown by its 99.9% validation accuracy and 99.86% AUC. An AUC of 91.76% and an accuracy of 91.4% on the test data further demonstrated the excellent performance of the model. These compelling indicators prove that the integrated approach is effective and provides a reliable structure for HPV classification. The findings of this study offer promising evidence supporting the potential clinical application of this revolutionary methodology, thereby establishing a foundation for subsequent advancements in medical imaging. To enhance the practicality and impact of this ground-breaking approach across various therapeutic domains, subsequent endeavours should concentrate on empirical verification and further improvement.

Disclosures

Institutional review board statement

Not applicable.

We express our gratitude to Eric Lucas from the Early Detection, Prevention & Infections Branch at the International Agency for Research on Cancer, WHO, for generously supplying us with the colposcopy images.

Financial support and sponsorship

None.

Conflicts of interest

None.

References

1 

Mahmoodi P, Fani M, Rezayi M, et al. Early detection of cervical cancer based on high-risk HPV DNA-based genosensors: a systematic review. Biofactors 2019; 45: 101-117.

2 

Singini MG, Singh E, Bradshaw D, et al. Usefulness of high-risk HPV early oncoprotein (E6 and E7) serological markers in the detection of cervical cancer: a systematic review and meta-analysis. J Med Virol 2023; 95: e27900.

3 

Burd EM. Human papillomavirus and cervical cancer. Clin Microbiol Rev 2003; 16: 1-17.

4 

Zhang T, Luo YM, Li P, et al. Cervical pre-cancerous lesions classification using pre-trained densely connected convolutional networks with colposcopy images. Biomed Signal Process Control 2020; 55: 101566.

5 

Li Y, Chen J, Xue P, et al. Computer-aided cervical cancer diagnosis using time-lapsed colposcopic images. IEEE Trans Med Imaging 2020; 39: 3403-3415.

6 

Yu Y, Ma J, Zhao W, Li Z, Ding S. MSCI: a multistate dataset for colposcopy image classification of cervical cancer screening. Int J Med Inform 2021; 146: 104352.

7 

Saini SK, Bansal V, Kaur R, Juneja M. ColpoNet for automated cervical cancer screening using colposcopy images. Mach Vision Appl 2020; 31: 1-15.

8 

Xu T, Zhang H, Xin C, et al. Multi-feature-based benchmark for cervical dysplasia classification evaluation. Pattern Recognit 2017; 63: 468-475.

9 

Cho BJ, Choi YJ, Lee MJ, et al. Classification of cervical neoplasms on colposcopic photography using deep learning. Sci Rep 2020; 10: 13652.

10 

Samet H. Foundations of multidimensional and metric data structures. Morgan Kaufmann, Burlington 2006.

11 

Bansal P, Gehlot K, Singhal A, Gupta A. Automatic detection of osteosarcoma based on integrated features and feature selection using binary arithmetic optimization algorithm. Multimed Tools Appl 2022; 81: 8807-8834.

12 

Tan M, Le Q. Efficientnet: rethinking model scaling for convolutional neural networks. In International Conference on Machine Learning 2019, 6105-6114.

13 

Howard AG, Zhu M, Chen B, et al. Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv 2017; 1704.04861.

14 

Selvaraj H, Selvi ST, Selvathi D, Gewali L. Brain MRI slices classification using least squares support vector machine. Int J Intelligent Computing Med Sci Image Process 2007; 1: 21-33.

15 

Ramteke RJ, Monali KY. Automatic medical image classification and abnormality detection using k-nearest neighbour. Int J Adv Computer Res 2012; 2: 190.

Copyright: © 2024 Termedia Sp. z o. o. This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0) License (http://creativecommons.org/licenses/by-nc-sa/4.0/), allowing third parties to copy and redistribute the material in any medium or format and to remix, transform, and build upon the material, provided the original work is properly cited and states its license.
 
Quick links
© 2024 Termedia Sp. z o.o.
Developed by Bentus.