{"id":3835,"date":"2023-01-24T11:35:37","date_gmt":"2023-01-24T09:35:37","guid":{"rendered":"https:\/\/gaz-temporal.i3a.es\/?p=3835"},"modified":"2025-12-23T01:20:28","modified_gmt":"2025-12-22T23:20:28","slug":"alejandro-valero","status":"publish","type":"post","link":"https:\/\/gaz.i3a.es\/es\/alejandro-valero\/","title":{"rendered":"Alejandro Valero"},"content":{"rendered":"<div id=\"pl-gb3835-69f225f4f06b1\"  class=\"panel-layout wp-block-siteorigin-panels-layout-block\" ><div id=\"pg-gb3835-69f225f4f06b1-0\"  class=\"panel-grid panel-has-style\" ><div class=\"siteorigin-panels-stretch panel-row-style panel-row-style-for-gb3835-69f225f4f06b1-0\" data-stretch-type=\"full-width-stretch\" ><div id=\"pgc-gb3835-69f225f4f06b1-0-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb3835-69f225f4f06b1-0-0-0\" class=\"so-panel widget widget_sow-hero panel-first-child panel-last-child\" data-index=\"0\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-hero so-widget-sow-hero-default-93415d0e2dbf-3835 so-widget-fittext-wrapper\"\n\t\t\t data-fit-text-compressor=\"0.85\"\n\t\t>\t\t\t\t<div class=\"sow-slider-base\" style=\"display: none\" tabindex=\"0\">\n\t\t\t\t\t<ul\n\t\t\t\t\tclass=\"sow-slider-images\"\n\t\t\t\t\tdata-settings=\"{&quot;pagination&quot;:true,&quot;speed&quot;:800,&quot;timeout&quot;:8000,&quot;paused&quot;:false,&quot;pause_on_hover&quot;:false,&quot;swipe&quot;:true,&quot;nav_always_show_desktop&quot;:&quot;&quot;,&quot;nav_always_show_mobile&quot;:&quot;&quot;,&quot;breakpoint&quot;:&quot;780px&quot;,&quot;unmute&quot;:false,&quot;anchor&quot;:null}\"\n\t\t\t\t\t\t\t\t\t\tdata-anchor-id=\"\"\n\t\t\t\t>\t\t<li class=\"sow-slider-image\" style=\"visibility: visible;;background-color: #1e73be\" >\n\t\t\t\t\t<div class=\"sow-slider-image-container\">\n\t\t\t<div class=\"sow-slider-image-wrapper\">\n\t\t\t\t<h3 style=\"text-align: center\"><a href=\"..\/team\/\">Investigadores<\/a><\/h3>\n<h1 style=\"text-align: center\"><strong>Alejandro Valero<\/strong><\/h1>\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t<\/li>\n\t\t<\/ul>\t\t\t\t<ol class=\"sow-slider-pagination\">\n\t\t\t\t\t\t\t\t\t\t\t<li><a href=\"#\" data-goto=\"0\" aria-label=\"mostrar diapositiva 1\"><\/a><\/li>\n\t\t\t\t\t\t\t\t\t<\/ol>\n\n\t\t\t\t<div class=\"sow-slide-nav sow-slide-nav-next\">\n\t\t\t\t\t<a href=\"#\" data-goto=\"next\" aria-label=\"diapositiva siguiente\" data-action=\"next\">\n\t\t\t\t\t\t<em class=\"sow-sld-icon-thin-right\"><\/em>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"sow-slide-nav sow-slide-nav-prev\">\n\t\t\t\t\t<a href=\"#\" data-goto=\"previous\" aria-label=\"diapositiva anterior\" data-action=\"prev\">\n\t\t\t\t\t\t<em class=\"sow-sld-icon-thin-left\"><\/em>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\t\t\t\t<\/div><\/div><\/div><\/div><\/div><\/div><\/div>\n\n<div id=\"pl-gb3835-69f225f4f1151\"  class=\"panel-layout wp-block-siteorigin-panels-layout-block\" ><div id=\"pg-gb3835-69f225f4f1151-0\"  class=\"panel-grid panel-no-style\" ><div id=\"pgc-gb3835-69f225f4f1151-0-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb3835-69f225f4f1151-0-0-0\" class=\"so-panel widget widget_sow-image panel-first-child panel-last-child\" data-index=\"0\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-image so-widget-sow-image-default-8b5b6f678277-3835\"\n\t\t\t\n\t\t>\n<div class=\"sow-image-container\">\n\t\t<img \n\tsrc=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/Alejandro_Valero_360x270-1-300x225.png\" width=\"300\" height=\"225\" srcset=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/Alejandro_Valero_360x270-1-300x225.png 300w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/Alejandro_Valero_360x270-1-16x12.png 16w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/Alejandro_Valero_360x270-1.png 360w\" sizes=\"(max-width: 300px) 100vw, 300px\" alt=\"\" \t\tclass=\"so-widget-image\"\/>\n\t<\/div>\n\n<\/div><\/div><\/div><div id=\"pgc-gb3835-69f225f4f1151-0-1\"  class=\"panel-grid-cell\" ><div id=\"panel-gb3835-69f225f4f1151-0-1-0\" class=\"so-panel widget widget_sow-image-grid panel-first-child\" data-index=\"1\" ><div class=\"panel-widget-style panel-widget-style-for-gb3835-69f225f4f1151-0-1-0\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-image-grid so-widget-sow-image-grid-default-5ff4073610f5-3835\"\n\t\t\t\n\t\t>\t<div\n\t\tclass=\"sow-image-grid-wrapper\"\n\t\t\t\t\t>\n\t\t\t\t\t<div class=\"sow-image-grid-image\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<a href=\"https:\/\/scholar.google.es\/citations?user=oUnoffMAAAAJ&#038;hl=en&#038;oi=ao\"\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\ttarget=\"_blank\" \t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\trel=\"noopener noreferrer\" \t\t\t\t\t\t\t\t\t\t\t>\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"37\" height=\"37\" src=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2020\/10\/google-scholar.png\" class=\"sow-image-grid-image_html\" alt=\"\" title=\"\" srcset=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2020\/10\/google-scholar.png 37w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2020\/10\/google-scholar-12x12.png 12w\" sizes=\"auto, (max-width: 37px) 100vw, 37px\" \/>\t\t\t\t\t\t\t\t\t<\/a>\n\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t<div class=\"sow-image-grid-image\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<a href=\"https:\/\/www.researchgate.net\/profile\/Alejandro-Valero\"\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\ttarget=\"_blank\" \t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\trel=\"noopener noreferrer\" \t\t\t\t\t\t\t\t\t\t\t>\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"37\" height=\"37\" src=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/researchgate_icon.png\" class=\"sow-image-grid-image_html\" alt=\"\" title=\"\" srcset=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/researchgate_icon.png 37w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/researchgate_icon-12x12.png 12w\" sizes=\"auto, (max-width: 37px) 100vw, 37px\" \/>\t\t\t\t\t\t\t\t\t<\/a>\n\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t<div class=\"sow-image-grid-image\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<a href=\"https:\/\/dblp.org\/pid\/02\/7848.html\"\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\ttarget=\"_blank\" \t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\trel=\"noopener noreferrer\" \t\t\t\t\t\t\t\t\t\t\t>\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"37\" height=\"37\" src=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/dblp_icon.png\" class=\"sow-image-grid-image_html\" alt=\"\" title=\"\" srcset=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/dblp_icon.png 37w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/dblp_icon-12x12.png 12w\" sizes=\"auto, (max-width: 37px) 100vw, 37px\" \/>\t\t\t\t\t\t\t\t\t<\/a>\n\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t<div class=\"sow-image-grid-image\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<a href=\"https:\/\/orcid.org\/0000-0002-0824-5833\"\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\ttarget=\"_blank\" \t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\trel=\"noopener noreferrer\" \t\t\t\t\t\t\t\t\t\t\t>\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"37\" height=\"37\" src=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/orcid_icon.png\" class=\"sow-image-grid-image_html\" alt=\"\" title=\"\" srcset=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/orcid_icon.png 37w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/orcid_icon-12x12.png 12w\" sizes=\"auto, (max-width: 37px) 100vw, 37px\" \/>\t\t\t\t\t\t\t\t\t<\/a>\n\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t<div class=\"sow-image-grid-image\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<a href=\"https:\/\/www.scopus.com\/authid\/detail.uri?authorId=57202633862\"\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\ttarget=\"_blank\" \t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\trel=\"noopener noreferrer\" \t\t\t\t\t\t\t\t\t\t\t>\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"37\" height=\"37\" src=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/scopus_icon.png\" class=\"sow-image-grid-image_html\" alt=\"\" title=\"\" srcset=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/scopus_icon.png 37w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/scopus_icon-12x12.png 12w\" sizes=\"auto, (max-width: 37px) 100vw, 37px\" \/>\t\t\t\t\t\t\t\t\t<\/a>\n\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t<div class=\"sow-image-grid-image\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<a href=\"https:\/\/publons.com\/researcher\/3171409\/alejandro-valero\/\"\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\ttarget=\"_blank\" \t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\trel=\"noopener noreferrer\" \t\t\t\t\t\t\t\t\t\t\t>\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"37\" height=\"37\" src=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/publons_icon.png\" class=\"sow-image-grid-image_html\" alt=\"\" title=\"\" srcset=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/publons_icon.png 37w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/publons_icon-12x12.png 12w\" sizes=\"auto, (max-width: 37px) 100vw, 37px\" \/>\t\t\t\t\t\t\t\t\t<\/a>\n\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t<div class=\"sow-image-grid-image\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<a href=\"https:\/\/www.lens.org\/lens\/profile\/249529457\/scholar\"\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\ttarget=\"_blank\" \t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\trel=\"noopener noreferrer\" \t\t\t\t\t\t\t\t\t\t\t>\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"37\" height=\"37\" src=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/lens_icon.png\" class=\"sow-image-grid-image_html\" alt=\"\" title=\"\" srcset=\"https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/lens_icon.png 37w, https:\/\/gaz.i3a.es\/wp-content\/uploads\/2023\/01\/lens_icon-12x12.png 12w\" sizes=\"auto, (max-width: 37px) 100vw, 37px\" \/>\t\t\t\t\t\t\t\t\t<\/a>\n\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t<\/div>\n<\/div><\/div><\/div><div id=\"panel-gb3835-69f225f4f1151-0-1-1\" class=\"so-panel widget widget_sow-editor panel-last-child\" data-index=\"2\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-editor so-widget-sow-editor-base\"\n\t\t\t\n\t\t>\n<div class=\"siteorigin-widget-tinymce textwidget\">\n\t<p>&nbsp;<\/p>\n<p><strong>Associate Professor<\/strong><\/p>\n<p><strong>Email:<\/strong> <a href=\"mailto:alvabre@unizar.es\">alvabre@unizar.es<\/a><\/p>\n<p><strong>Address:<\/strong><br \/>\nDepartment of Computer Science and Systems Engineering<br \/>\nUniversidad de Zaragoza<br \/>\nCalle Mar\u00eda de Luna, 1<br \/>\nAda Byron Building<br \/>\n50018 Zaragoza, Spain<\/p>\n<\/div>\n<\/div><\/div><\/div><\/div><\/div>\n\n<div id=\"pl-gb3835-69f225f4f4077\"  class=\"panel-layout wp-block-siteorigin-panels-layout-block\" ><div id=\"pg-gb3835-69f225f4f4077-0\"  class=\"panel-grid panel-has-style\" ><div class=\"panel-row-style panel-row-style-for-gb3835-69f225f4f4077-0\" ><div id=\"pgc-gb3835-69f225f4f4077-0-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb3835-69f225f4f4077-0-0-0\" class=\"so-panel widget widget_sow-headline panel-first-child\" data-index=\"0\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-headline so-widget-sow-headline-default-71f7c5d8712d-3835\"\n\t\t\t\n\t\t><div class=\"sow-headline-container\">\n\t\t\t\t\t\t\t<h5 class=\"sow-headline\">\n\t\t\t\t\t\tBIOGRAPHY\t\t\t\t\t\t<\/h5>\n\t\t\t\t\t\t\t\t\t\t\t<div class=\"decoration\">\n\t\t\t\t\t\t<div class=\"decoration-inside\"><\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n<\/div><\/div><div id=\"panel-gb3835-69f225f4f4077-0-0-1\" class=\"so-panel widget widget_sow-editor panel-last-child\" data-index=\"1\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-editor so-widget-sow-editor-base\"\n\t\t\t\n\t\t>\n<div class=\"siteorigin-widget-tinymce textwidget\">\n\t<p>Alejandro Valero received the BS, MS, and PhD degrees in Computer Engineering from the <em>Universitat Polit\u00e8cnica de Val\u00e8ncia<\/em>, Spain, in 2009, 2011, and 2013, respectively. From 2013 to 2015 he was a Visiting Researcher with Northeastern University, Boston (MA), USA, and the University of Cambridge, UK. From 2016 to 2021 he was an Assistant Professor with the Department of Computer Science and Systems Engineering, <em>Universidad de Zaragoza<\/em>, Spain. Since 2021 he is an Associate Professor with the same department and institution. Prof. Valero has taught several courses on computer organization, including digital design, computer organization and design, heterogeneous systems programming and design, data center design, and operating systems. His PhD research contributions to the design of high-performance, energy-efficient CPU memory subsystems were recognized by multiple entities. He received the Intel Doctoral Student Honor Program Award in 2012 and the Gold Medal in the ACM Student Research Competition (SRC) held in the 27th International Conference on Supercomputing (ICS 2013). His research interests mainly focus on the design of memory hierarchies in terms of performance, energy efficiency, and reliability for different microprocessors: CPU systems, general-purpose GPUs, and accelerators for computer vision algorithms. Prof. Valero has participated in more than 20 national and local funded research projects and has published more than 30 papers in the main venues of the computer architecture area, such as the IEEE\/ACM International Symposium on Microarchitecture (MICRO), the International Conference on Parallel Architectures and Compilation Techniques (PACT), IEEE Transactions on Computers, and IEEE Transactions on Very Large Scale Integration (VLSI) Systems. He has served as Technical Program Committee member in a significant number of conferences, workshops, and research competitions, like the Design Automation and Test in Europe (DATE) conference, the IEEE International Conference on Computer Design (ICCD), the Performance Modeling, Benchmarking, and Simulation of High Performance Computer Systems (PMBS) workshop, and the ACM SRC Grand Finals. He is also a frequent reviewer in top journals of his area, such as IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, IEEE Transactions on Dependable and Secure Computing, and ACM Transactions on Design Automation of Electronic Systems. He was a recipient of the Outstanding Reviewer Award in the Design Methods and Tools track at the DATE 2024 conference. Prof. Valero is a member of the ACM, the <em>Sociedad de Arquitectura y Tecnolog\u00eda de Computadores<\/em> (SARTECO), the Aragon Institute of Engineering Research (I3A), and an affiliated member of the High Performance, Edge, And Cloud Computing (HiPEAC) European Network of Excellence.<\/p>\n<\/div>\n<\/div><\/div><\/div><\/div><\/div><div id=\"pg-gb3835-69f225f4f4077-1\"  class=\"panel-grid panel-has-style\" ><div class=\"panel-row-style panel-row-style-for-gb3835-69f225f4f4077-1\" ><div id=\"pgc-gb3835-69f225f4f4077-1-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb3835-69f225f4f4077-1-0-0\" class=\"so-panel widget widget_sow-headline panel-first-child\" data-index=\"2\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-headline so-widget-sow-headline-default-244eb6bef45a-3835\"\n\t\t\t\n\t\t><div class=\"sow-headline-container\">\n\t\t\t\t\t\t\t<h5 class=\"sow-headline\">\n\t\t\t\t\t\tPUBLICATIONS\t\t\t\t\t\t<\/h5>\n\t\t\t\t\t\t\t\t\t\t\t<div class=\"decoration\">\n\t\t\t\t\t\t<div class=\"decoration-inside\"><\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n<\/div><\/div><div id=\"panel-gb3835-69f225f4f4077-1-0-1\" class=\"so-panel widget widget_sow-editor panel-last-child\" data-index=\"3\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-editor so-widget-sow-editor-base\"\n\t\t\t\n\t\t>\n<div class=\"siteorigin-widget-tinymce textwidget\">\n\t<div class=\"teachpress_pub_list\"><form name=\"tppublistform\" method=\"get\" action=\"\"><a name=\"tppubs\" id=\"tppubs\"><\/a><div class=\"teachpress_filter\"><select class=\"default\" name=\"yr\" id=\"yr\" tabindex=\"2\" onchange=\"teachpress_jumpMenu('parent',this, 'https:\/\/gaz.i3a.es\/es\/alejandro-valero\/?')\">\r\n                   <option value=\"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=#tppubs\">Todos los a\u00f1os<\/option>\r\n                   <option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2026#tppubs\" >2026<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2025#tppubs\" >2025<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2024#tppubs\" >2024<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2023#tppubs\" >2023<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2022#tppubs\" >2022<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2021#tppubs\" >2021<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2020#tppubs\" >2020<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2019#tppubs\" >2019<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2018#tppubs\" >2018<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2017#tppubs\" >2017<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2016#tppubs\" >2016<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2015#tppubs\" >2015<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2014#tppubs\" >2014<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2013#tppubs\" >2013<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2012#tppubs\" >2012<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2011#tppubs\" >2011<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2010#tppubs\" >2010<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2009#tppubs\" >2009<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2008#tppubs\" >2008<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2007#tppubs\" >2007<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2006#tppubs\" >2006<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2005#tppubs\" >2005<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2004#tppubs\" >2004<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2003#tppubs\" >2003<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2002#tppubs\" >2002<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2001#tppubs\" >2001<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2000#tppubs\" >2000<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1999#tppubs\" >1999<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1998#tppubs\" >1998<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1997#tppubs\" >1997<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1996#tppubs\" >1996<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1995#tppubs\" >1995<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1994#tppubs\" >1994<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1989#tppubs\" >1989<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1987#tppubs\" >1987<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1985#tppubs\" >1985<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=1791#tppubs\" >1791<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=0000#tppubs\" >0000<\/option>\r\n                <\/select><select class=\"default\" name=\"type\" id=\"type\" tabindex=\"3\" onchange=\"teachpress_jumpMenu('parent',this, 'https:\/\/gaz.i3a.es\/es\/alejandro-valero\/?')\">\r\n                   <option value=\"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=#tppubs\">Todas las tipolog\u00edas<\/option>\r\n                   <option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=article#tppubs\" >Art\u00edculos de revista<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=book#tppubs\" >Libros<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=incollection#tppubs\" >Book Sections<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=inproceedings#tppubs\" >Proceedings Articles<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=mastersthesis#tppubs\" >Tesis de m\u00e1ster o tesina<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=misc#tppubs\" >Miscel\u00e1nea<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=phdthesis#tppubs\" >Tesis doctorales<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=proceedings#tppubs\" >Actas de congresos<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=techreport#tppubs\" >Informes t\u00e9cnicos<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=workshop#tppubs\" >Workshops<\/option>\r\n                <\/select><\/div><input type=\"hidden\" name=\"trp-form-language\" value=\"es\"\/><\/form><div class=\"tablenav\"><div class=\"tablenav-pages\"><span class=\"displaying-num\">41 registros<\/span> <a class=\"page-numbers button disabled\">&laquo;<\/a> <a class=\"page-numbers button disabled\">&lsaquo;<\/a> 1 de 9 <a href=\"https:\/\/gaz.i3a.es\/es\/alejandro-valero\/?limit=2&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"p\u00e1gina siguiente\" class=\"page-numbers button\">&rsaquo;<\/a> <a href=\"https:\/\/gaz.i3a.es\/es\/alejandro-valero\/?limit=9&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"\u00faltima p\u00e1gina\" class=\"page-numbers button\">&raquo;<\/a> <\/div><\/div><div class=\"teachpress_publication_list\"><h3 class=\"tp_h3\" id=\"tp_h3_2025\">2025<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_article\">Art\u00edculos de revista<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Valero, Alejandro;  Lorente, Vicente;  Petit, Salvador;  Sahuquillo, Julio<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('849','tp_links')\" style=\"cursor:pointer;\">Dual Fast-Track Cache: Organizing Ring-Shaped Racetracks to Work as L1 Caches<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">IEEE Transactions on Computers, <\/span><span class=\"tp_pub_additional_volume\">vol. 74, <\/span><span class=\"tp_pub_additional_number\">no 8, <\/span><span class=\"tp_pub_additional_pages\">pp. 2812-2826, <\/span><span class=\"tp_pub_additional_year\">2025<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 0018-9340<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_849\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('849','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_849\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('849','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_849\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('849','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_849\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Valero2025,<br \/>\r\ntitle = {Dual Fast-Track Cache: Organizing Ring-Shaped Racetracks to Work as L1 Caches},<br \/>\r\nauthor = {Alejandro Valero and Vicente Lorente and Salvador Petit and Julio Sahuquillo},<br \/>\r\nurl = {https:\/\/www.computer.org\/csdl\/journal\/tc\/2025\/08\/11022726\/27fzlt4rw88},<br \/>\r\ndoi = {10.1109\/TC.2025.3575909},<br \/>\r\nissn = {0018-9340},<br \/>\r\nyear  = {2025},<br \/>\r\ndate = {2025-08-01},<br \/>\r\nurldate = {2025-08-01},<br \/>\r\njournal = {IEEE Transactions on Computers},<br \/>\r\nvolume = {74},<br \/>\r\nnumber = {8},<br \/>\r\npages = {2812-2826},<br \/>\r\nabstract = {Static Random-Access Memory (SRAM) is the fastest memory technology and has been the common design choice for implementing first-level (L1) caches in the processor pipeline, where speed is a key design issue that must be fulfilled. On the contrary, this technology offers much lower density compared to other technologies like Dynamic RAM, limiting L1 cache sizes of modern processors to a few tens of KB. This paper explores the use of slower but denser Domain Wall Memory (DWM) technology for L1 caches. This technology provides slow access times since it arranges multiple bits sequentially in a magnetic racetrack. To access these bits, they need to be shifted in order to place them under a header. A 1-bit shift usually takes one processor cycle, which can significantly hurt the application performance, making this working behavior inappropriate for L1 caches. Based on the locality (temporal and spatial) principles exploited by caches, this work proposes the Dual Fast-Track Cache (Dual FTC) design, a new approach to organizing a set of racetracks to build set-associative caches. Compared to a conventional SRAM cache, Dual FTC enhances storage capacity by 5\u00d7 while incurring minimal shifting overhead, thereby rendering it a practical and appealing solution for L1 cache implementations. Experimental results show that the devised cache organization is as fast as an SRAM cache for 78% and 86% of the L1 data cache hits and L1 instruction cache hits, respectively (i.e., no shift is required). Consequently, due to the larger L1 cache capacities, significant system performance gains (by 22% on average) are obtained under the same silicon area.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('849','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_849\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Static Random-Access Memory (SRAM) is the fastest memory technology and has been the common design choice for implementing first-level (L1) caches in the processor pipeline, where speed is a key design issue that must be fulfilled. On the contrary, this technology offers much lower density compared to other technologies like Dynamic RAM, limiting L1 cache sizes of modern processors to a few tens of KB. This paper explores the use of slower but denser Domain Wall Memory (DWM) technology for L1 caches. This technology provides slow access times since it arranges multiple bits sequentially in a magnetic racetrack. To access these bits, they need to be shifted in order to place them under a header. A 1-bit shift usually takes one processor cycle, which can significantly hurt the application performance, making this working behavior inappropriate for L1 caches. Based on the locality (temporal and spatial) principles exploited by caches, this work proposes the Dual Fast-Track Cache (Dual FTC) design, a new approach to organizing a set of racetracks to build set-associative caches. Compared to a conventional SRAM cache, Dual FTC enhances storage capacity by 5\u00d7 while incurring minimal shifting overhead, thereby rendering it a practical and appealing solution for L1 cache implementations. Experimental results show that the devised cache organization is as fast as an SRAM cache for 78% and 86% of the L1 data cache hits and L1 instruction cache hits, respectively (i.e., no shift is required). Consequently, due to the larger L1 cache capacities, significant system performance gains (by 22% on average) are obtained under the same silicon area.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('849','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_849\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/www.computer.org\/csdl\/journal\/tc\/2025\/08\/11022726\/27fzlt4rw88\" title=\"https:\/\/www.computer.org\/csdl\/journal\/tc\/2025\/08\/11022726\/27fzlt4rw88\" target=\"_blank\">https:\/\/www.computer.org\/csdl\/journal\/tc\/2025\/08\/11022726\/27fzlt4rw88<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/TC.2025.3575909\" title=\"DOI de seguimiento:10.1109\/TC.2025.3575909\" target=\"_blank\">doi:10.1109\/TC.2025.3575909<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('849','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2024\">2024<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_article\">Art\u00edculos de revista<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Toca-D\u00edaz, Yamilka;  Tejero, Rub\u00e9n Gran;  Valero, Alejandro<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('850','tp_links')\" style=\"cursor:pointer;\">Shift-and-Safe: Addressing permanent faults in aggressively undervolted CNN accelerators<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">Journal of Systems Architecture, <\/span><span class=\"tp_pub_additional_volume\">vol. 157, <\/span><span class=\"tp_pub_additional_pages\">pp. 1-13, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 1383-7621<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_850\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('850','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_850\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('850','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_850\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('850','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_850\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Toca-D\u00edaz2024,<br \/>\r\ntitle = {Shift-and-Safe: Addressing permanent faults in aggressively undervolted CNN accelerators},<br \/>\r\nauthor = {Yamilka Toca-D\u00edaz and Rub\u00e9n Gran Tejero and Alejandro Valero},<br \/>\r\nurl = {https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1383762124002297},<br \/>\r\ndoi = {https:\/\/doi.org\/10.1016\/j.sysarc.2024.103292},<br \/>\r\nissn = {1383-7621},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-12-01},<br \/>\r\nurldate = {2024-12-01},<br \/>\r\njournal = {Journal of Systems Architecture},<br \/>\r\nvolume = {157},<br \/>\r\npages = {1-13},<br \/>\r\nabstract = {Underscaling the supply voltage (Vdd) to ultra-low levels below the safe-operation threshold voltage (Vmin) holds promise for substantial power savings in digital CMOS circuits. However, these benefits come with pronounced challenges due to the heightened risk of bitcell permanent faults stemming from process variations in current technology node sizes. This work delves into the repercussions of such faults on the accuracy of a 16-bit fixed-point Convolutional Neural Network (CNN) inference accelerator powering on-chip activation memories at ultra-low Vdd voltages. Through an in-depth examination of fault patterns, memory usage, and statistical analysis of activation values, this paper introduces Shift-and-Safe: two novel and cost-effective microarchitectural techniques exploiting the presence of outlier activation values and the underutilization of activation memories. Particularly, activation outliers enable a shift-based data representation that reduces the impact of faults on the activation values, whereas the memory underutilization is exploited to maintain a safe replica of affected activations in idle memory regions. Remarkably, these mechanisms do not add any burden to the programmer and are independent of application characteristics, rendering them easily deployable across real-world CNN accelerators. Experimental results show that Shift-and-Safe maintains the CNN accuracy even in the presence of almost a quarter of the total activations with faults. In addition, average energy savings are by 5% and 11% compared to the state-of-the-art approach and a conventional accelerator supplied at Vmin, respectively.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('850','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_850\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Underscaling the supply voltage (Vdd) to ultra-low levels below the safe-operation threshold voltage (Vmin) holds promise for substantial power savings in digital CMOS circuits. However, these benefits come with pronounced challenges due to the heightened risk of bitcell permanent faults stemming from process variations in current technology node sizes. This work delves into the repercussions of such faults on the accuracy of a 16-bit fixed-point Convolutional Neural Network (CNN) inference accelerator powering on-chip activation memories at ultra-low Vdd voltages. Through an in-depth examination of fault patterns, memory usage, and statistical analysis of activation values, this paper introduces Shift-and-Safe: two novel and cost-effective microarchitectural techniques exploiting the presence of outlier activation values and the underutilization of activation memories. Particularly, activation outliers enable a shift-based data representation that reduces the impact of faults on the activation values, whereas the memory underutilization is exploited to maintain a safe replica of affected activations in idle memory regions. Remarkably, these mechanisms do not add any burden to the programmer and are independent of application characteristics, rendering them easily deployable across real-world CNN accelerators. Experimental results show that Shift-and-Safe maintains the CNN accuracy even in the presence of almost a quarter of the total activations with faults. In addition, average energy savings are by 5% and 11% compared to the state-of-the-art approach and a conventional accelerator supplied at Vmin, respectively.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('850','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_850\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1383762124002297\" title=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1383762124002297\" target=\"_blank\">https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1383762124002297<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/https:\/\/doi.org\/10.1016\/j.sysarc.2024.103292\" title=\"DOI de seguimiento:https:\/\/doi.org\/10.1016\/j.sysarc.2024.103292\" target=\"_blank\">doi:https:\/\/doi.org\/10.1016\/j.sysarc.2024.103292<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('850','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Toca-D\u00edaz, Yamilka;  Palacios, Reynier Hern\u00e1ndez;  Tejero, Ruben Gran;  Valero, Alejandro<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('851','tp_links')\" style=\"cursor:pointer;\">Flip-and-Patch: A fault-tolerant technique for on-chip memories of CNN accelerators at low supply voltage<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">Microprocessors and Microsystems, <\/span><span class=\"tp_pub_additional_volume\">vol. 106, <\/span><span class=\"tp_pub_additional_pages\">pp. 1-13, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 0141-9331<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_851\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('851','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_851\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('851','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_851\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('851','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_851\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Toca-D\u00edaz2024b,<br \/>\r\ntitle = {Flip-and-Patch: A fault-tolerant technique for on-chip memories of CNN accelerators at low supply voltage},<br \/>\r\nauthor = {Yamilka Toca-D\u00edaz and Reynier Hern\u00e1ndez Palacios and Ruben Gran Tejero and Alejandro Valero},<br \/>\r\nurl = {https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0141933124000188},<br \/>\r\ndoi = {https:\/\/doi.org\/10.1016\/j.micpro.2024.105023},<br \/>\r\nissn = {0141-9331},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-04-01},<br \/>\r\nurldate = {2024-04-01},<br \/>\r\njournal = {Microprocessors and Microsystems},<br \/>\r\nvolume = {106},<br \/>\r\npages = {1-13},<br \/>\r\nabstract = {Aggressively reducing the supply voltage (Vdd) below the safe threshold voltage (Vmin) can effectively lead to significant energy savings in digital circuits. However, operating at such low supply voltages poses challenges due to a high occurrence of permanent faults resulting from manufacturing process variations in current technology nodes. This work addresses the impact of permanent faults on the accuracy of a Convolutional Neural Network (CNN) inference accelerator using on-chip activation memories supplied at low Vdd below Vmin. Based on a characterization study of fault patterns, this paper proposes two low-cost microarchitectural techniques, namely Flip-and-Patch, which maintain the original accuracy of CNN applications even in the presence of a high number of faults caused by operating at Vdd &lt; Vmin. Unlike existing techniques, Flip-and-Patch remains transparent to the programmer and does not rely on application characteristics, making it easily applicable to real CNN accelerators.<br \/>\r\nExperimental results show that Flip-and-Patch ensures the original CNN accuracy with a minimal impact on system performance (less than 0.05% for every application), while achieving average energy savings of 10.5% and 46.6% in activation memories compared to a conventional accelerator operating at safe and nominal supply voltages, respectively. Compared to the state-of-the-art ThUnderVolt technique, which dynamically adjusts the supply voltage at run time and discarding any energy overhead for such an approach, the average energy savings are by 3.2%.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('851','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_851\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Aggressively reducing the supply voltage (Vdd) below the safe threshold voltage (Vmin) can effectively lead to significant energy savings in digital circuits. However, operating at such low supply voltages poses challenges due to a high occurrence of permanent faults resulting from manufacturing process variations in current technology nodes. This work addresses the impact of permanent faults on the accuracy of a Convolutional Neural Network (CNN) inference accelerator using on-chip activation memories supplied at low Vdd below Vmin. Based on a characterization study of fault patterns, this paper proposes two low-cost microarchitectural techniques, namely Flip-and-Patch, which maintain the original accuracy of CNN applications even in the presence of a high number of faults caused by operating at Vdd &lt; Vmin. Unlike existing techniques, Flip-and-Patch remains transparent to the programmer and does not rely on application characteristics, making it easily applicable to real CNN accelerators.<br \/>\r\nExperimental results show that Flip-and-Patch ensures the original CNN accuracy with a minimal impact on system performance (less than 0.05% for every application), while achieving average energy savings of 10.5% and 46.6% in activation memories compared to a conventional accelerator operating at safe and nominal supply voltages, respectively. Compared to the state-of-the-art ThUnderVolt technique, which dynamically adjusts the supply voltage at run time and discarding any energy overhead for such an approach, the average energy savings are by 3.2%.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('851','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_851\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0141933124000188\" title=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0141933124000188\" target=\"_blank\">https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0141933124000188<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/https:\/\/doi.org\/10.1016\/j.micpro.2024.105023\" title=\"DOI de seguimiento:https:\/\/doi.org\/10.1016\/j.micpro.2024.105023\" target=\"_blank\">doi:https:\/\/doi.org\/10.1016\/j.micpro.2024.105023<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('851','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_inproceedings\">Proceedings Articles<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Toca-D\u00edaz, Yamilka;  Tejero, Rub\u00e9n Gran;  Valero, Alejandro<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('852','tp_links')\" style=\"cursor:pointer;\">Ensuring the Accuracy of CNN Accelerators Supplied at Ultra-Low Voltage<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_pages\">pp. 92-95, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>, <span class=\"tp_pub_additional_isbn\">ISBN: 979-8-3503-8040-8<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_852\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('852','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_852\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('852','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_852\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('852','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_852\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{Toca-D\u00edaz2024c,<br \/>\r\ntitle = {Ensuring the Accuracy of CNN Accelerators Supplied at Ultra-Low Voltage},<br \/>\r\nauthor = {Yamilka Toca-D\u00edaz and Rub\u00e9n Gran Tejero and Alejandro Valero},<br \/>\r\nurl = {https:\/\/ieeexplore.ieee.org\/document\/10817950},<br \/>\r\ndoi = {https:\/\/doi.org\/10.1109\/ICCD63220.2024.00024},<br \/>\r\nisbn = {979-8-3503-8040-8},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-11-18},<br \/>\r\nurldate = {2024-11-18},<br \/>\r\njournal = {Proceedings of the 42nd IEEE International Conference on Computer Design (ICCD 2024)},<br \/>\r\npages = {92-95},<br \/>\r\nabstract = {Underscaling the supply voltage (Vdd) to ultra-low levels below the safe-operation threshold voltage (Vmin) brings significant energy savings in digital CMOS circuits but introduces reliability challenges due to increased risk of bitcell permanent faults. This work explores the impact of such faults on the accuracy of a CNN inference accelerator supplying on-chip activation memories at ultra-low Vdd. By examining fault pat-terns, activation values, and memory usage, this paper proposes two microarchitectural techniques exploiting activation outliers and activation memory underutilization. These approaches are cost-effective, do not require programmer intervention, and are application-independent. Experimental results show that the proposed approaches maintain the original CNN accuracy and achieve energy savings by 2.1 % and 8.2 % compared to the state-of-the-art technique and a conventional accelerator supplied at Vmin, respectively, with a negligible impact on the system performance (less than 0.25 %).},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('852','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_852\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Underscaling the supply voltage (Vdd) to ultra-low levels below the safe-operation threshold voltage (Vmin) brings significant energy savings in digital CMOS circuits but introduces reliability challenges due to increased risk of bitcell permanent faults. This work explores the impact of such faults on the accuracy of a CNN inference accelerator supplying on-chip activation memories at ultra-low Vdd. By examining fault pat-terns, activation values, and memory usage, this paper proposes two microarchitectural techniques exploiting activation outliers and activation memory underutilization. These approaches are cost-effective, do not require programmer intervention, and are application-independent. Experimental results show that the proposed approaches maintain the original CNN accuracy and achieve energy savings by 2.1 % and 8.2 % compared to the state-of-the-art technique and a conventional accelerator supplied at Vmin, respectively, with a negligible impact on the system performance (less than 0.25 %).<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('852','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_852\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/ieeexplore.ieee.org\/document\/10817950\" title=\"https:\/\/ieeexplore.ieee.org\/document\/10817950\" target=\"_blank\">https:\/\/ieeexplore.ieee.org\/document\/10817950<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/https:\/\/doi.org\/10.1109\/ICCD63220.2024.00024\" title=\"DOI de seguimiento:https:\/\/doi.org\/10.1109\/ICCD63220.2024.00024\" target=\"_blank\">doi:https:\/\/doi.org\/10.1109\/ICCD63220.2024.00024<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('852','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2023\">2023<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_inproceedings\">Proceedings Articles<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Toca-D\u00edaz, Yamilka;  Mu\u00f1oz, Nicol\u00e1s Landeros;  Tejero, Ruben Gran;  Valero, Alejandro<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('853','tp_links')\" style=\"cursor:pointer;\">On Fault-Tolerant Microarchitectural Techniques for Voltage Underscaling in On-Chip Memories of CNN Accelerators<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_pages\">pp. 138-145, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>, <span class=\"tp_pub_additional_isbn\">ISBN: 979-8-3503-4419-6<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_853\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('853','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_853\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('853','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_853\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('853','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_853\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{Toca-D\u00edaz2023,<br \/>\r\ntitle = {On Fault-Tolerant Microarchitectural Techniques for Voltage Underscaling in On-Chip Memories of CNN Accelerators},<br \/>\r\nauthor = {Yamilka Toca-D\u00edaz and Nicol\u00e1s Landeros Mu\u00f1oz and Ruben Gran Tejero and Alejandro Valero},<br \/>\r\nurl = {https:\/\/ieeexplore.ieee.org\/document\/10456839},<br \/>\r\ndoi = {https:\/\/doi.org\/10.1109\/DSD60849.2023.00029},<br \/>\r\nisbn = {979-8-3503-4419-6},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-09-06},<br \/>\r\nurldate = {2023-09-06},<br \/>\r\njournal = {Proceedings of the 26th Euromicro Conference on Digital System Design (DSD 2023)},<br \/>\r\npages = {138-145},<br \/>\r\nabstract = {Aggressively underscaling the supply voltage (Vdd) below the safe voltage (Vmin) margin is an effective solution to attain substantial energy savings. Unfortunately, operating at such low voltages is challenging due to the high number of permanent faults as a result of variations in the manufacturing process of current technology nodes. This work characterizes the impact of permanent faults on the accuracy of a Convolutional Neural Network (CNN) inference accelerator with on-chip activation memories supplied at low Vdd below Vmin. Based on these observations, this paper proposes a couple of low-cost microarchitectural techniques, referred to as flipping and patching, that ensure the accuracy of CNN applications despite the presence of permanent faults. Contrary to prior work, the proposed techniques are transparent to the programmer and do not depend on application characteristics. Experimental results show that the proposed techniques maintain the original CNN accuracy with a minimal impact on system performance (less than 0.05%), while reducing the energy consumption of activation memories by 11.2% and 46.7% compared to those of a conventional accelerator operating at safe and nominal supply voltages, respectively.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('853','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_853\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Aggressively underscaling the supply voltage (Vdd) below the safe voltage (Vmin) margin is an effective solution to attain substantial energy savings. Unfortunately, operating at such low voltages is challenging due to the high number of permanent faults as a result of variations in the manufacturing process of current technology nodes. This work characterizes the impact of permanent faults on the accuracy of a Convolutional Neural Network (CNN) inference accelerator with on-chip activation memories supplied at low Vdd below Vmin. Based on these observations, this paper proposes a couple of low-cost microarchitectural techniques, referred to as flipping and patching, that ensure the accuracy of CNN applications despite the presence of permanent faults. Contrary to prior work, the proposed techniques are transparent to the programmer and do not depend on application characteristics. Experimental results show that the proposed techniques maintain the original CNN accuracy with a minimal impact on system performance (less than 0.05%), while reducing the energy consumption of activation memories by 11.2% and 46.7% compared to those of a conventional accelerator operating at safe and nominal supply voltages, respectively.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('853','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_853\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/ieeexplore.ieee.org\/document\/10456839\" title=\"https:\/\/ieeexplore.ieee.org\/document\/10456839\" target=\"_blank\">https:\/\/ieeexplore.ieee.org\/document\/10456839<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/https:\/\/doi.org\/10.1109\/DSD60849.2023.00029\" title=\"DOI de seguimiento:https:\/\/doi.org\/10.1109\/DSD60849.2023.00029\" target=\"_blank\">doi:https:\/\/doi.org\/10.1109\/DSD60849.2023.00029<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('853','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><\/div><div class=\"tablenav\"><div class=\"tablenav-pages\"><span class=\"displaying-num\">41 registros<\/span> <a class=\"page-numbers button disabled\">&laquo;<\/a> <a class=\"page-numbers button disabled\">&lsaquo;<\/a> 1 de 9 <a href=\"https:\/\/gaz.i3a.es\/es\/alejandro-valero\/?limit=2&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"p\u00e1gina siguiente\" class=\"page-numbers button\">&rsaquo;<\/a> <a href=\"https:\/\/gaz.i3a.es\/es\/alejandro-valero\/?limit=9&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"\u00faltima p\u00e1gina\" class=\"page-numbers button\">&raquo;<\/a> <\/div><\/div><\/div>\n<\/div>\n<\/div><\/div><\/div><\/div><\/div><\/div>\n\n\n<p><\/p>","protected":false},"excerpt":{"rendered":"","protected":false},"author":1,"featured_media":3836,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[238,239],"tags":[],"class_list":["post-3835","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-reseacher","category-team"],"_links":{"self":[{"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/posts\/3835","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/comments?post=3835"}],"version-history":[{"count":18,"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/posts\/3835\/revisions"}],"predecessor-version":[{"id":4150,"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/posts\/3835\/revisions\/4150"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/media\/3836"}],"wp:attachment":[{"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/media?parent=3835"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/categories?post=3835"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/gaz.i3a.es\/es\/wp-json\/wp\/v2\/tags?post=3835"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}