Category: غير مصنف
Materials Engineering and Nanotechnology for Enhancing Structure Resistance to Seismic Activity
Book Review: “Arbitration” by Judge Dr. Muhammad Walid Mansour
Multiple Health and Environmental Impacts of An Innovative Livestock Feed Intake Diet
Introduction
Humanity is struggling to create chemical-free, GMO-free, hormone-free, antibiotic-free, safely clean, high-protein nutritional food with low-to-neutral carbon and water footprints at a reasonable price (1). Therefore, improving the sustainability of dairy operations is a current key goal in the dairy sector, and one critical task to increase sustainability is to reduce environmental consequences from dairy production (2). Hence, when evaluating feedstuffs to establish their nutritional contents and inclusion rates in dairy rations, the environmental impact, as well as production responses, should be taken into account. In fact, tremendous effort was put into finding new livestock dietary formulas to tackle the above-enlisted challenges around the globe (3,4). Cattles, and particularly dairy production systems, significantly contribute to green house gas (GHG) emissions and global warming mostly through the creation of methane (CH4) (5). In fact, methane is the largest contributor to global warming from the dairy sector, with a 28 times higher impact compared to carbon dioxide (CO2) over a hundred-year period (6). Therefore, the transformation of our production systems with a particular focus on lowering GHG emissions has gained priority (7). In this context, lessening environmental impacts from dairy production is one critical task to improve sustainability of dairy operations. Thus, the environmental impact, as well as production responses, should be considered when evaluating feedstuffs and determining their nutritional values and inclusion rates in dairy rations. Yan et al. (8), demonstrated that one attribute of energy-efficient cows is that less methane is produced relative to the amount of energy consumed or milk produced. Other studies have shown that breeding for cattle with high feed efficiency may also result in decreased daily enteric methane generation, due to the strong genetic and phenotypic association between daily methane output and residual feed intake (9,10). On another side, profitability will rise due to improved feed efficiency because feed expenses are the main expense on dairy farms. Feed efficiency are expressed in various ways, including feed conversion efficiency (milk output over feed intake). In theory, improved feed efficiency decreases daily methane production due to a lower methane per kilogram of dry matter intake (DMI) at a given production level (11), while decreased methane production (e.g., due to nutritional strategies) does not necessarily improve feed efficiency. However, experimental data are inconsistent on the link between residual feed intake (RFI) and methane emission, while research has primarily focused on beef cattle (12,13) rather than on lactating dairy cows (14) .Recent developments in livestock nutrition have primarily concentrated on three areas: improving our understanding of the nutritional needs of livestock, identifying the supply and availability of nutrients in feed ingredients, and developing the least expensive diets that effectively combine nutrient requirements and nutrient supply (15-17). In line with this strategy, the main objectives of our applied research are to reverse the livestock conventional common practices from high multi-dimensional polluter into a low polluter sector with a low water footprint; improving the quality of protein and fatty acids profiles for animal and human better health and well-being, and reducing the overall cost compared to organic practices. To achieve the objectives of this project, nearly a decade worth of applied research resulted in the development of a new balanced feed intake diet composed of a clean fresh sprouted highly nutritional mix, namely Mahjoub Feedstock Diet (MFD), produced in soil-less vertical farming in a controlled environment. Neutral carbon footprint (NCF) and low water footprint (LWF) resulted into chemicals-free, hormones and genetically modified (GMO) free husbandry practice at a local facility in Damascus, Syria. This study is a continuation of our applied research, and it is worth mentioning that it is the first to evaluate whether manure nutrients, NH3 emissions and milk quality were affected by feeding cows with MFD.
Material and Methods
Mahjoub’s feedstock diet
Mahjoub feedstock (MFD) diet is an innovation primed for livestock feed. This innovation in feed ingredients covers efficiency, profitability, environmental footprint, animal health, and welfare. All chemical and physical analyses were conducted in Cumberland Valley Analytical Services (USA) according to standard and accredited protocols. MFD, with Mahjoob’s Intellectual Properties, is a clean fresh sprouted diet produced in a controlled-environment vertical farming powered entirely by clean renewable energy resulting in a neutral carbon footprint and a very low water footprint at a local facility in Damascus, Syria.
Animals and Treatments
We conducted our experiments in a randomized complete block design. We fed four “Holstein” cows (average 506 ± 100 kg) on MFD diet over a period of two years. For manure comparison only, we collected fresh manure samples from cows fed a local common basal diet-containing soybean as a control (CON) and compared its chemical and physical composition to fresh manure collected from cows fed on MFD. On the other hand, we compared the composition of milk-fat produced by cows fed on MFD to a world-renowned brand butter fat sample. MFD was prepared once a day in the morning and fed to cows four times/24h, namely at 7am, 13:00, 17:00 and 20:00. Notably, cows were free-stall most of the day with access to outdoor and fed through designated feeding box. Finally, cow bedding was made of dried odour-less dried manure. Sample Collection and Measurement – milk production and composition. Cows were milked 2x daily with milk yields average around the year approximately 25 liter/day. Milk samples were obtained by automated milking machine and collected into clean and steamed containers, with measurements performed within one hour at the laboratories of the National Commission for Biotechnology (NCBT), Damascus, Syria. GC-MS standard protocols were used for fatty acid analysis (Thermo Scientific, USA) while amino acid analysis was performed by amino acid analyser (Agilent, USA). All chemical and colorometric assays for total protein and manure analysis were performed using standard protocols at NCBT.
Results
Feed Diet Analysis
Chemical analyses were performed on Mahjoub’s feedstock diet (MFD) (Table 1).
Effect of Mahjoub’s feedstock diet on Manure Nutrient Content. The characteristics of the manure samples from cows fed MFD and meal local common diet are shown in Table 2.
Table 2. Characteristics of the manure samples from cows fed Mahjoub’s feedstock diet and meal local common diet
Properties | Unit | Basis | Manure from Cows fed MFD | Manure from Cows fed CON |
Dry matter | wt% | wet | 17.96 | 17.86 |
Volatile solids | wt% | dry | 86.64 | 80.70 |
Ash | wt% | dry | 13.36 | 19.26 |
Carbon | wt% | dry | 50.22 | 46.80 |
Nitrogen | wt% | dry | 1.65 | 2.20 |
C/N | wt% | dry | 30.40 | 21.27 |
Ammonia | wt% | wet | 0.07 | 0.21 |
Fiber | wt% | dry | 53.11 | 49.58 |
Calcium | wt% | dry | 2.34 | 2.07 |
Phosphorus | wt% | dry | 0.17 | 0.26 |
Potassium | wt% | dry | 0.67 | 0.72 |
Sodium | wt% | dry | 0.17 | 0.14 |
Chloride | wt% | dry | 0.21 | 0.83 |
Iron | wt% | dry | 0.011 | 0.012 |
Electrical conductivity | μS/cm | – | 180 | 470 |
PH | – | – | 7.7 | 7.25 |
MFD = Mahjoub’s feedstock diet; CONT = fed basal diet containing soybean (meal local common diet)
Analysis of Protein and Fat in Milk Products
We compared the amino acid (AA) profile between MFD and cow milk to explore the possible cow/rumen conversion of AA in vivo (Fig 1). To study the effect of MFD on fat profile, we compared the fat contents in milk-fat from MFD-fed cows to a globally well-known butter brand (Fig 2).
Fig. 1. Comparison of amino acid concentrations between the original MFD (g/day of feed intake) and cow milk (g/day produced) fed on MFD.
Fig. 2. Fatty acids composition in average milk fat from cows fed on MFD in comparison with a renowned fat brand. USF: unsaturated fat, SF: saturated fat, TF: trans-fat, w9: omega 9, w6: omega 6, w3: omega 3, PUSF: polyunsaturated fat, MUSF: monounsaturated fat, CLA: conjugated linoleic acid.
Discussion
MFD composition
Results showed several privileged characteristics of MFD when compared to conventional cow feed diets (18) (Table 1) explicitly; high protein, low fat, low volatile fatty acids (VFA), high soluble protein SP/crude protein CP, neutral dietary cation-anion difference (DCAD), low starch, high acid detergent fibers (ADF), and near neutral pH. This composition may reflect the sprouted non-stiff format of MFD and makes it a unique high protein diet, which might positively reflect on cow health.
Manure Composition
Despite the fact that manure is a valuable fertilizer, it has the potential to harm the environment in terms of odor, air, soil, and water quality (19). As various types of gases (e.g., NH3, greenhouse gases, and H2S) are created from manure via microbial fermentation or chemical changes, farm odor and a reduction in air quality at stalls and during manure storage before application to the field may occur (20). It is worth knowing that the amount of gas generated by manure is determined by both internal and external factors. External influences include chemical forms of nutrients and nutrient concentrations, temperature, humidity, wind, bedding, manure storage system, and so on. Internal factors may include the cow genetic makeup and the microflora residing in their intestines.
In our study, we assessed changes in manure characteristics as well as potential gas emissions from manure. In fact, feeding the cows on MFD diet tended to increase manure pH compared to controls (7.7 vs.7.26) (Table 1).
The content of organic matters was greater for MFD versus CONT, without a difference in dry matter (DM). It is worth noting that manure nitrogen content was lower for MFD versus CONT (1.65 vs. 2.2 %), and this could be a factor that potentially lowers NH3 emissions from manure because manure N, in the form of urea, contributes to NH3 emitted from manure (21,22). Our results showed that the cumulative ammonia production for MFD was lower than its production from CONT by a factor of three (0.07 vs. 0.21 %). The degree of the decrease in NH3 emissions by MFD in this study is similar to the decrease observed when feeding cows on a
low-protein diet (23). Thus, our data demonstrate that the NH3 -emitting potential of manure can be reduced using MFD without decreasing dietary protein content, as the high protein content of MFD was not associated with high manure nitrogen, as one would expect.
Table 1. Chemical analysis of Mahjoub’s feedstock diet
Properties | Unit | Basis | MFD |
Dry matter | wt% | wet | 24.9 |
Ash | wt% | dry | 5.56 |
Crude protein | wt% | dry | 28.2 |
Soluble Protein | wt% | dry | 17.6 |
Crude fat (fat ether extract) | wt% | dry | 2.83 |
Starch | wt% | dry | 11.8 |
Starch | wt% | NFC | 44.0 |
Soluble Fiber | wt% | dry | 9.51 |
Soluble Fiber | wt% | NFC | 35.4 |
Volatile fatty acids | wt% | dry | 5.58 |
Lactic Acid | wt% | dry | 2.20 |
Lactic Acid | wt% | VFA | 39.4 |
Acetic Acid | wt% | dry | 3.38 |
Propionic Acid | wt% | dry | 0.19 |
Ammonia | wt% | dry | 17.3 |
Lignin | wt% | dry | 4.24 |
Soluble Fiber | wt% | dry | 9.51 |
Non Fiber Carbohydrates | wt% | dry | 26.8 |
Non-Structural Carbohydrates | wt% | dry | 13.4 |
Acid detergent fibres | wt% | dry | 25.1 |
Total digestible nutrients. | wt% | dry | 68.4 |
PH | 6.68 |
MFD= Mahjoub’s feedstock diet; NFC= non-fiber carbohydrate; VFA= volatile fatty acids
The C/N ration in manure from MFD-fed cows was profoundly higher compared to manure from locally fed cows (30.4 vs. 21.27), while total nitrogen was lower (1.65 vs 2.2) and fiber content was close (53.11% vs. 48.58%), compared to manure from locally common fed cow. The profoundly lower electrical conductivity (180 vs. 470 μS/cm) and higher pH (7.7 vs 7.25) in MFD-fed compared to locally fed cows may enhance the applicability of the fresh manure from the former as a proposed soil substrate replacement. Worth to mention, the low ammonia concentration may have resulted in a near-no odour of manure.
Amino Acid Profile
Our results show a major increase in several amino acids upon feeding on MFD, specifically two essential AA (proline and glutamate), and leucine, a non-essential AA. Interestingly, these three previous AAs were proposed to play a main role in regulating and enhancing the immune response in both cows and humans (24,25). In fact, it is well known that amino acids regulate the activation of many immune cells including T and B lymphocytes, natural killer cells and macrophages, in addition to controlling gene expression and the production of antibodies and cytokines (24). Nevertheless, one major finding about MFD-fed cows was the antibiotic-free wellbeing of the four cows in study over the last two years. This wellbeing is supported by physical in addition to biochemical analyses of several cow plasma biomarkers, all of which were continuously within reference ranges throughout the study (data not shown).
Fat Profile
The results showed several excellent features of the MFD on human health and wellbeing (26), including: slightly higher unsaturated fat and lower 6/3 ratio compared to brand fat, a favourable profile in many health compromised situations including heart disease (27,28). More importantly, fat from MFD contained substantially favourable lower trans-fat (TF) in MFD-milk fat compared to the brand fat (0.79 vs 3.13, respectively). In fact, previous research proved a direct link between TF and many diseases including cardiovascular, breast cancer and disorders of nervous system, etc (29). Additionally, MFD-fat contained two fold levels of conjugated linoleic acid (CLA) in comparison to the brand fat (0.62 vs. 0.31, respectively). CLA has several beneficial health effects as it reduces body fat and consequently alleviates the risk for cardiovascular diseases and cancer. In addition, CLA modulates immune and inflammatory responses as well as improves bone mass (30). Finally, both saturated C15 and C17 were markedly higher in MFD-fed cow fat compared to brand (C15; 0.86 vs. 0.40 g/100g fat) and (C17; 0.87 vs. 0.42 g/100g fat), respectively. C15 odd saturated fatty acids are linked to supporting metabolic and heart health, while both C15 and C17 fatty acids are associated with lower risks for cardiovascular diseases and mortality (31,32). Taken together, the MFD-milk fat profile suggest an enhanced human wellbeing.
Conclusion
This study was the first to evaluate whether manure nutrients, NH3 emissions and milk quality were affected by feeding cows with Mahjoub’s feedstock diet. Our results show a major increase in several amino acids in the milk of cows fed with MFD, which we propose to play a main role in regulating and enhancing the immune response in cows. Indeed, this could be supported by the fact that cows fed on MFD were antibiotic-free well-being for many years.
On another hand, our results indicate that the NH3-emitting potential of cow manure were reduced by MFD without decreasing dietary protein content. Hence, a beneficial goal was achieved without jeopardizing the cow immune response relying on adequate protein concentration in the diet .Finally, the low ammonia values in MFD-fed cow manure, low total nitrogen, high fiber compared to local common-fed cow manure, low electrical conductivity and alkaline pH, will enhance the applicability of the fresh manure from MFD-fed cows as a proposed soil substrate replacement and may have resulted in a near-no odour of manur. More studies on the long-term incubation of manure will be necessary to understand H2S emissions during manure storage. In this context, further research is planned and ongoing; our preliminary results show predictable privileged characteristics of MFD on both environment and cow/human wellbeing.
The Frequency of Rs8134378 and Rs12329760 Gene Polymorphism in A Cohort of Healthy and Compromised Syrians
The Priorities for UN Climate Change Conference COP27
The conference that was held in November 2022, at Sharm el-Sheikh, Egypt, defined five priorities, which can be briefed as following; First, COP27 closed with a breakthrough agreement to provide loss and damage funding for vulnerable countries hit hard by floods, droughts and other climate disasters. This was widely lauded as a historic decision, because for the first time countries did recognize the need for finance to respond to loss and damage associated with the catastrophic effects of climate change, and ultimately agreed to establish a fund. Second, it intended to keep 1.5°C within reach, which requires global greenhouse gas emissions to peak before 2025 and to be reduced by 43% by 2030. This means the global economy must “mitigate” climate change – in other words, countries must reduce or prevent the emission of greenhouse gases to get where they need to be by 2030. Third, hold businesses and institutions to account; this new phase of implementation also means a new focus on accountability when it comes to the commitments made by sectors, businesses and institutions, so the transparency of their commitments will be a priority of UN Climate Change in 2023. Fourth, providing financial support for developing countries; finance is at the heart of all what the world is doing to combat climate change. As a result, mitigation, adaptation, loss and damage, climate technology – all of it requires sufficient funds to function properly and to yield the desired results. Finally, making the pivot toward implementation, which is an important step, because climate pledges aren’t worth the paper they’re written on if they aren’t taken off the page and turned into concrete action.
The Effectiveness of CNN in Evaluating Ultrasound Image Datasets: Diagnosing Polycystic Ovary Syndrome (PCOS) as an Example
Introduction
Polycystic ovary syndrome (PCOS) is one of the most prevalent conditions affecting women of reproductive age. It affects 6%–20% of premenopausal women globally (1). Ovarian dysfunction and an excess of androgen are the two main symptoms of PCOS. Menstrual abnormalities, hirsutism, obesity, insulin resistance, cardiovascular disease, in addition to emotional symptoms like depression (2) are all common among PCOS patients (3). As a result, it is crucial for the accurate diagnosis and treatment of PCOS. In 1985, Adams et al. (4) discovered that polycystic ovaries have an abnormally high number of follicles, also termed multifollicularity. It was suggested that PCOS be diagnosed when at least two of the three following features were present at an expert meeting in Rotterdam in 2003: (i) oligo- or anovulation, (ii) clinical and/or biochemical hyperandrogenism, or (iii) polycystic ovaries (5). The latter of which can be detected using ultrasonography which offers the highest contribution to the diagnosis of PCOS (6). Early detection of PCOS is important because it can help to manage the symptoms and reduce the risk of long-term health issues. Although ultrasound images have some disadvantages of strong artifacts, noise and high dependence on the experience of doctors, they are still considered as one of the most widely used modalities in medical diagnosis. Many artificial intelligence systems have been developed to help doctors. Convolutional Neural Networks (CNNs) and deep learning have achieved great success in computer vision with its unique advantages (7). Many diseases are diagnosed using different Deep Learning Models (7). Some examples include the detection of COVID-19 using lung ultrasound imagery achieving 89.1% accuracy using InceptionV3 network (8), the use of deep learning architectures for the segmentation of the left ventricle of the heart (9), and the classification of breast lesions in ultrasound images obtaining an accuracy of 90%, sensitivity of 86%, and specificity of 96% by utilizing the GoogLeNet CNN (10). As we can see, deep learning has proved its potential and the vital role it can provide in benefitting and assisting practitioners that use ultrasonography as a tool for diagnosis. This paper is discussing the robustness of deep learning in diagnosing PCOS. Since arificial intelligence (AI) and deep learning algorithms can quickly and reliably assess vast volumes of data, they can be utilized to diagnose PCOS in ultrasound scans. AI and deep learning algorithms are expected to examine ultrasound images to find patterns and traits that are indicative of PCOS in the case of PCOS detection. This can help to increase the speed and accuracy of diagnosis as it can be done more accurately and efficiently than by manual analysis. Furthermore, the application of AI and deep learning in the diagnosis of PCOS can decrease the workload for medical professionals and free them up to concentrate on other responsibilities. Overall, the use of AI and deep learning in the detection of PCOS in ultrasound images has the potential to improve the accuracy, efficiency, and accessibility of healthcare. This was the motive to tackle such an important health issue that affects millions of women worldwide and apply the potential of deep learning to help nullify this crucial problem. Obtaining a viable and correct ultrasound dataset for this task is difficult as the annotation of medical images requires significant professional medical knowledge, which makes the annotation very expensive and rare as well as the ethical issues and sensitivity of such dataset which can pose another problem. Therefore, resorting to a publicly available dataset was chosen to accelerate the work on this project. After observing some of the related work, one publicly available PCOS dataset was utilized in the training of PCONet, a CNN developed by Hosain AK et al. that detects PCOS from ovarian ultrasound images with accuracy of 98.12% on test images as well as fine-tuning InceptionV3 model achieving 96.56% accuracy (11). The PCOS dataset is publicly available on Kaggle (12). Other related work includes Wenqi Lv et al. who utilized image segmentation technique using U-Net on scleral images then a Resnet model was adapted for feature extraction of PCOS achieving classification accuracy of 0.929, and AUC of 0.979 (13). Subrato Bharati et al. used clinical attributes of 541 women in which 177 are infected with PCOS to be utilized in a machine learning model that uses random forest and logistic regression to predict PCOS disease for which the testing accuracy achieved is 91.01% (14). Sakshi Srivastava et al. employed a fine-tuned VGG-16 model to train on their dataset that consists of ultrasound images of the ovaries to detect the presence ovarian cyst with 92.11% accuracy obtained (15). In this paper, the Kaggle dataset highlighted previously is used to conduct training by enhancing the power of transfer learning to train an existing model architecture that is pre-trained on thousands of images in advance. The training on this dataset achieves excellent results. However, after achieving great results and after further inspection on the dataset, it turned out that the publicly available PCOS dataset that was utilized in training the fine-tuned model and that other authors (11) have been using in their research is extremely erroneous and full of misleading information. This will be discussed in details in this paper.
Methodology
Dataset Description
An appropriate dataset is vital for the proper functioning of any deep learning framework.
Thus, the publicly available PCOS ultrasound images available on Kaggle (12) is used. This same dataset is referred to as dataset A in the highlighted paper of Hosain AK et al. (11) mentioned previously and will be referred to as dataset A in this paper as well. A screenshot of the website providing this data is shown in Figure 1 and Figure 2.
Figure 1: Screenshot of the publicly available PCOS dataset on Kaggle consisting of `infected` and `notinfected` ovarian ultrasound images referred to as Dataset A
Figure 2: Statistics of Dataset A show that it is downloaded 301 times out of 2394 views. This means that almost 1 out of every 10 viewers downloads this dataset for utilization in research/projects
The dataset consists of 3856 ultrasound images divided into 2 classes which are labeled as: `infected` and `notinfected`. The latter depicts healthy ovaries and the first indicates the presence of PCOS. These images are partitioned into train and test sets in which 1932 images belong to the test set and the rest belong to the train set. However, the same images seem to be repeated in the test and train directories. Therefore, one of the directories was neglected. Figure 3 depicts a sample of the ultrasound images present in this dataset.
Figure 3: Samples of infected and healthy ovaries in Dataset A
Data Pre-Processing
There are several steps that may be taken to preprocess medical ultrasound images before they are input into a deep learning model for training. It’s important to note that the specific preprocessing steps will depend on the characteristics of the images and the specific requirements of the deep learning model being used. Since the model chosen for this project is the DenseNet201 architecture which is a pre-trained model on the ImageNet dataset, the ovarian ultrasound images that will be the input to this model need to be pre-processed the same way as the images this model was originally trained on. These images vary in size and dimensions therefore they need to be preprocessed so that they are uniform in this aspect. The preprocessing pipeline include:
- Resizing all the images into the same size of (224, 224), which is the image size that DenseNet201 was trained on.
- Transforming the images into tensors. In Pytorch which is the deep learning of choice for this project, a tensor is a multi-dimensional array that is similar to a Numpy array, but can be operated on by the GPU, which makes it more efficient for certain types of computations. Tensors can be used to store a wide variety of data, including images, videos, and audio, and are an important building block in PyTorch. The resulting tensor has the same number of dimensions as the original image, with the size of each dimension being the size of the corresponding image dimension. The pixel values of the image are also normalized to the range [0,1] when converted to tensors.
- Normalize the images: It is common to normalize the images by subtracting the mean and dividing by the standard deviation of the image pixels. This helps to center the data and improve the model’s ability to learn from it. The mean and standard deviation values for each color channel are:
mean = [0.485, 0.456, 0.406] std = [0.229, 0.224, 0.225]
This is done due to it being mentioned as a necessary pre-process in the documentation (16) of DenseNet201.
- Transforming all ultrasound images to grayscale as this will eliminate the unnecessary color channels if they exist and the resultant image will have only 1 color channel due to it being grayscale.
- The images and their corresponding labels are turned into batches where each batch consists of 32 images. This is very beneficial for a variety of reasons:
- Memory constraints: Training a deep learning model on a large dataset can require a lot of memory. Grouping the images into batches allows the model to train on a subset of the data at a time, which can be more memory efficient.
- Computational efficiency: Processing the images in batches can be more computationally efficient, as it allows the model to make better use of the parallel processing capabilities of modern hardware such as GPUs.
- Stochastic gradient descent: When training a model using stochastic gradient descent, it is common to update the model’s weights using the gradients computed on a small batch of data rather than the entire dataset. This is because computing the gradients on the full dataset can be computationally expensive, and using a smaller batch of data can provide a good approximation.
- Regularization: Using small batches of data can also introduce noise into the learning process, which can act as a form of regularization and help the model generalize better to new data.
Model Description
After pre-processing the data, a classification model is needed. Reviewing related work, most of research done in the medical field that uses deep learning utilizes transfer learning to train on the data through a pre-trained model. In fact, the conclusion of paper (17) mentions that transfer learning is the best available option for training even if the data model is pre-trained on, it is weakly related to data in hand.
Thus, in order to train on this dataset, the DenseNet architecture (18) was chosen. DenseNet is one of the leading architectures used in ultrasonography analysis when conducting transfer learning as table 3 in this paper (19) shows. DenseNet is a variation of the traditional CNN architecture that uses dense blocks, where each layer in a block is connected to every other layer in the block. This allows the network to pass information from earlier layers to later layers more efficiently, alleviate the problem of vanishing gradients in deeper networks, and can improve performance on tasks such as image classification. There are several variations of the DenseNet architecture, such as DenseNet121, DenseNet201, and DenseNet264. Each variation refers to the number of layers it consists of. DenseNet201 is the variation chosen for this project as it has a moderate number of layers and it is not too computationally demanding. DenseNet and many other pre-trained architectures are pre-trained on the ImageNet dataset which is long-standing landmark in computer vision (20). It consists of 1,281,167 training images, 50,000 validation images, and 100,000 test images belonging to 1000 classes. These images consist of a wide variety of scenes, covering a diverse range of categories such as animals, vehicles, household objects, food, clothing, musical instruments, body parts, and much more. A question might arise from this information, why would a model trained on a such unrelated dataset might be used for training on and predicting medical images? In fact, the paper titled “A scoping review of transfer learning research on medical image analysis using ImageNet” (19) discusses this very topic. After inspecting tens of research papers and studies that utilize ImageNet models to train on medical datasets, Author proves that transfer learning of ImageNet models is a viable option to train on medical datasets. The idea behind transfer learning is that although medical datasets are different from non-medical datasets, the low-level features (e.g., straight and curved lines that construct images) are universal to most of the image analysis tasks (21). Therefore, transferred parameters (i.e., weights) may serve as a powerful set of features, which reduce the need for a large dataset as well as the training time and memory cost (21). The structure of DenseNet is shown in figure 4:
Figure 4: An ultrasound image with size (224, 224, 1) as an input to the DenseNet model using its weights and architecture to make a prediction
Model Fine-Tuning
Transfer learning can be utilized to import the DenseNet201 model and fine-tune it to adjust it to according to the dataset. This fine-tuning in this study involves 2 stages:
- Adjusting the very first layer to make it accept grayscale images that are composed of 1 color channel as opposed to 3 color channels that the ImageNet dataset consists of which is used to train the DenseNet201 model on. This is a better method than expanding a single channel to channels because it requires additional resources to store and process the additional channels that don’t provide any new information.
- ImageNet dataset consists of 1000 classes. Therefore, the DenseNet201 model also has 1000 corresponding outputs, 1 output probability for each class. This dataset consists of 2 classes only. Thus, the very last layer is adjusted to output 1 probability only which will be rounded to either 0 if the value is below 0.5 which indicates that the image is `infected` or 1 if the value is above 0.5 which signifies that the image is `not infected`. The raw output of the model i.e., the logit which is the final unnormalized score of the model (unbounded real number) is converted into a probability (range 0à1) using the sigmoid function: σ(x)=1/(1+e^(-x) ) where x is the model logit. After the logit is converted into a probability, it is converted into label (0 or 1) by rounding the probability into either 0 or 1 to indicate the prediction of the model.
After fine-tuning the model, the total number of parameters in the model is 18,088,577 in which 1,921 is trainable, and 18,086,656 is non-trainable (frozen).
Picking Loss Function and Optimizer
When working on binary classification in PyTorch, the most common loss function to use is binary cross-entropy loss, also known as log loss. This loss function is appropriate for binary classification problems where the output of the model is a probability, and the goal is to minimize the difference between the predicted probabilities and the true labels. The loss is calculated as:
loss = -(y * log(p) + (1 – y) * log(1 – p))
As for the optimizer, the two most common optimizers used are Adam and Stochastic Gradient Descent (SGD). The latter was the choice for this project as this paper (22) mentions that SGD generally performs better on image classification tasks. The optimizer and the learning rate are closely related, as the optimizer uses the learning rate to determine the step size when making updates to the model parameters. The learning rate opted for is 0.01.
Results:
After conducting the above steps and setting the learning rate to 0.01, the optimizer is set to SGD and training the model is initiated for 15 epochs. Moreover, the model had its pre-trained weights frozen except for its final Linear layer. This method is recommended in the paper (19) for models trained on ImageNet architectures. Exceptional results were achieved with 100% accuracy on the train set and 99.83% on the test set. The following two figures show the loss and accuracy during training in addition to the confusion matrix obtained by the results of the trained model predicting on the test set. Since the results are suspiciously perfect, we decided to investigate more using another dataset.
Figure 5: Results of loss and accuracy during training epochs
Figure 6: Confusion Matrix for Dataset A
Working with a New Dataset
Same experiment using the same parameters was repeated on another dataset: Dataset B. Another publicly available dataset is published by Telkom University Dataverse (23). Figure 7 shows a screenshot of the website providing this data. Dataset B consists of 54 ultrasound images 14 of which are classified as PCOS and the rest are normal. Figure 8 below shows a sample from this dataset.
Figure 7: Screenshot of the publicly available PCOS dataset by Telkom University Dataverse referred to as Dataset B in this paper which is an ovarian ultrasound dataset that is annotated by specialist doctors
Figure 8: Samples of infected and healthy ovaries in Dataset B
However, to keep the two experiments identical and due to the problem of dataset B being relatively small in data number, data augmentation techniques such as random horizontal flipping, random vertical flipping, random brightness alteration, and random rotation are applied using the Python library `imgaug`. Data augmentation increases the number of ultrasound images which improves the training process massively.
Dataset B Results
After training the same model again for 15 epochs on this new dataset and keeping all the hyperparameters fixed, the following results and confusion matrix are obtained:
Figure 9: Results of accuracy and loss during training epochs with the new data for Dataset B
Figure 10: Confusion Matrix for Dataset B
As figure 9 shows, the train accuracy increases steadily until it reaches 83.33% while the test accuracy remains relatively unchanged at 62.92%. The same can be observed for train and test loss where the train loss steadily decreases but not for the test loss. This is a clear indication of overfitting and the inability of the model to generalize well on unseen data. The poor model performance is also confirmed in the confusion matrix where the number of true positives and true negatives is unsatisfactory.
Table 1 exhibits the precision, recall and F1-score for the infected data points in both datasets:
Table 1: Precision, recall, and F1-score for the infected data points in both datasets
Further inspection on the dataset A and after consulting a professional specialist in this medical field, it turned out that the dataset is highly erroneous and misleading. The `notinfected` class which is supposed to represent the healthy ovaries having no sign of PCOS are in fact not images of ovaries at all. Rather, they are ultrasound images of uterus which completely falsify this dataset.
Conclusion:
Two experiments were conducted on two different datasets as called in this paper: Dataset A and Dataset B. Dataset A gave much better results but, it turned out that the dataset is highly erroneous and misleading. Therefore, data quality is of the utmost importance when training deep learning models, especially in the medical and health fields. The results of this study, show the ability of CNN and deep learning models in detecting the suspicious findings from the datasets. The DenseNet201 model not performing well on dataset B could be due to a variety of reasons such as the complexity of the ultrasound images or the relatively small number of data points available in the original dataset. To address this, the entire DenseNet201 model can be trained rather than freezing the feature extractor to perhaps produce better accuracy results on the test data of dataset B as training only the classifier seems to be not sufficient for this task. Also, experimenting with different learning rates and optimizers could yield more satisfactory results. However, the accuracy and reliability of the model’s predictions depend heavily on the quality of the data used to train it. If the data is flawed or biased, the model will likely produce inaccurate or unreliable results even if the results appear to be satisfactory. In the medical and health field, this can have serious consequences as it can lead to incorrect diagnoses or treatment recommendations, potentially causing harm to patients. Therefore, it is essential to ensure that the data used to train these models is of the highest quality and represents accurately the population intended to serve. This includes ensuring that the data is free from errors, and represents the target population, and has been collected using appropriate methods. Ensuring data quality is an ongoing process that requires continuous monitoring and improvement.