<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-05-05T18:33:28Z</responseDate><request verb="GetRecord" identifier="oai:uvadoc.uva.es:10324/80444" metadataPrefix="didl">https://uvadoc.uva.es/oai/request</request><GetRecord><record><header><identifier>oai:uvadoc.uva.es:10324/80444</identifier><datestamp>2026-04-22T07:32:21Z</datestamp><setSpec>com_10324_1134</setSpec><setSpec>com_10324_931</setSpec><setSpec>com_10324_894</setSpec><setSpec>col_10324_1213</setSpec></header><metadata><d:DIDL xmlns:d="urn:mpeg:mpeg21:2002:02-DIDL-NS" xmlns:doc="http://www.lyncode.com/xoai" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="urn:mpeg:mpeg21:2002:02-DIDL-NS http://standards.iso.org/ittf/PubliclyAvailableStandards/MPEG-21_schema_files/did/didl.xsd">
<d:DIDLInfo>
<dcterms:created xmlns:dcterms="http://purl.org/dc/terms/" xsi:schemaLocation="http://purl.org/dc/terms/ http://dublincore.org/schemas/xmls/qdc/dcterms.xsd">2025-12-10T13:11:34Z</dcterms:created>
</d:DIDLInfo>
<d:Item id="hdl_10324_80444">
<d:Descriptor>
<d:Statement mimeType="application/xml; charset=utf-8">
<dii:Identifier xmlns:dii="urn:mpeg:mpeg21:2002:01-DII-NS" xsi:schemaLocation="urn:mpeg:mpeg21:2002:01-DII-NS http://standards.iso.org/ittf/PubliclyAvailableStandards/MPEG-21_schema_files/dii/dii.xsd">urn:hdl:10324/80444</dii:Identifier>
</d:Statement>
</d:Descriptor>
<d:Descriptor>
<d:Statement mimeType="application/xml; charset=utf-8">
<oai_dc:dc xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns:dc="http://purl.org/dc/elements/1.1/" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
<dc:title>Machine Learning Representation of Loss of Eye Regularity in a Drosophila Neurodegenerative Model</dc:title>
<dc:creator>Diez Hermano, Sergio</dc:creator>
<dc:creator>Ganfornina Álvarez, María Dolores</dc:creator>
<dc:creator>Vega-Lozano, Esteban</dc:creator>
<dc:creator>Sánchez Romero, Diego</dc:creator>
<dc:description>Producción Científica</dc:description>
<dc:description>The fruit fly compound eye is a premier experimental system for modeling human&#xd;
neurodegenerative diseases. The disruption of the retinal geometry has been historically&#xd;
assessed using time-consuming and poorly reliable techniques such as histology or&#xd;
pseudopupil manual counting. Recent semiautomated quantification approaches rely&#xd;
either on manual region-of-interest delimitation or engineered features to estimate the&#xd;
extent of degeneration. This work presents a fully automated classification pipeline&#xd;
of bright-field images based on orientated gradient descriptors and machine learning&#xd;
techniques. An initial region-of-interest extraction is performed, applying morphological&#xd;
kernels and Euclidean distance-to-centroid thresholding. Image classification algorithms&#xd;
are trained on these regions (support vector machine, decision trees, random forest,&#xd;
and convolutional neural network), and their performance is evaluated on independent,&#xd;
unseen datasets. The combinations of oriented gradient C gaussian kernel Support&#xd;
Vector Machine [0.97 accuracy and 0.98 area under the curve (AUC)] and fine-tuned&#xd;
pre-trained convolutional neural network (0.98 accuracy and 0.99 AUC) yielded the best&#xd;
results overall. The proposed method provides a robust quantification framework that&#xd;
can be generalized to address the loss of regularity in biological patterns similar to the&#xd;
Drosophila eye surface and speeds up the processing of large sample batches.</dc:description>
<dc:date>2025-12-10T13:11:34Z</dc:date>
<dc:date>2025-12-10T13:11:34Z</dc:date>
<dc:date>2020</dc:date>
<dc:type>info:eu-repo/semantics/article</dc:type>
<dc:identifier>Front Neurosci. 2020 Jun 4;14:516</dc:identifier>
<dc:identifier>https://uvadoc.uva.es/handle/10324/80444</dc:identifier>
<dc:identifier>10.3389/fnins.2020.00516</dc:identifier>
<dc:identifier>Frontiers in Neuroscience</dc:identifier>
<dc:identifier>14</dc:identifier>
<dc:identifier>1662-453X</dc:identifier>
<dc:language>spa</dc:language>
<dc:rights>info:eu-repo/semantics/openAccess</dc:rights>
<dc:rights>http://creativecommons.org/licenses/by-nc-nd/4.0/</dc:rights>
<dc:rights>Attribution-NonCommercial-NoDerivatives 4.0 Internacional</dc:rights>
</oai_dc:dc>
</d:Statement>
</d:Descriptor>
<d:Component id="10324_80444_1">
<d:Resource ref="https://uvadoc.uva.es/bitstream/10324/80444/1/2020%20Frontiers%20Method.pdf" mimeType="application/pdf"/>
</d:Component>
</d:Item>
</d:DIDL></metadata></record></GetRecord></OAI-PMH>