Keynote Talks
Sparse Matrix Algorithms for Evolving Neural Networks
Carlos Ordonez
Department of Computer Science, University of Houston, USA
Abstract
There is significant research into sparse and dense matrix computations, with high-performance computing techniques, reducing time complexity and improving parallel speedup mostly with ample main memory, considering I/O on secondary storage as a less important aspect. On the other hand, there has been important work in the database, data mining and big data communities accelerating the computation of machine learning models on large data sets. However, massive neural networks and constantly changing data sets are pushing matrix computation demands further. We first present a survey on three key problems identifying research issues: maintaining a large data set updated under frequent matrix entry insertions and deletions, sparse matrix addition/multiplication and recomputing a deep neural network when a sparse data set changes frequently. We then propose a research agenda focusing on those three major problems solved with parallel I/O efficient algorithms storing and processing matrices with coordinate tuples (like a database relational table): matrix entry insertion/deletion, matrix addition/multiplication and assembling these algorithms into state-of-the-art neural networks. We argue coordinate tuples complement and can potentially replace established main memory storage mechanisms, like dense arrays and compressed row/column formats. In summary, we believe database-inspired, parallel I/O efficient, algorithms tailored for sparse matrices can help updating, explaining and monitoring evolving neural networks on large dynamic data sets.
![]() |
BiographyCarlos Ordonez studied at UNAM (top research university in Mexico), getting a B.Sc. in applied math and an M.S. in computer science. He continued PhD studies at the Georgia Institute of Technology, focusing on optimizing machine learning and data mining analysis with parallel algorithms, removing main memory limitations and enhancing their accuracy. During his PhD, Carlos joined NCR Corporation collaborating in the optimization of machine learning, statistics and cubes on the Teradata parallel DBMS, under an SMP distributed architecture. After working almost 10 years years at NCR, Carlos joined the Department of Computer Science at the University of Houston, where he currently leads the Data-Intensive Parallel Algorithms for AI group, focusing on making theory practical. During 2012 and 2013 Carlos regularly visited MIT, collaborating with Turing award winner, Michael Stonebraker, working on new-generation parallel DBMSs (columnar, array, lockfree transactions) to solve large-scale linear algebra and graph algorithms. Carlos worked as a visiting researcher with ATT Labs-Research (formerly the famous ATT Bell Labs where C, C++ and Unix were invented), where he conducted research on analyzing streams with statistical methods on massive data sets. His research projects have been funded by NSF and NIH grants. |
From Data Silos to Data Mesh: A Case Study in Financial Data Architecture
Mariusz Sienkiewicz
Director, Supervisory Data Analysis Center, Polish Financial Supervision Authority, Poland
Abstract
Successful data analytics implementation requires seamless access to both data and related metadata. In many organizations, analytics challenges arise from Data Silos, which impede cross-functional access to data and knowledge sharing across the organization. This talk presents practical insights from a data architecture transformation project conducted at a large institution with over 1,400 employees and overseeing over 2,000 market entities. The organization faced significant analytical and operational challenges due to the presence of Data Silos—isolated repositories associated with specific business areas. To address these limitations, the institution initiated a transition to a Data Mesh architecture to improve data availability and enhance analytical capabilities. This talk explains the rationale behind the persistence of silos, evaluates alternative architectural models, and justifies the choice of Data Mesh based on organizational context. Key elements of the transformation include developing a data management framework, implementing a data catalog, creating a data lake to provide data input flexibility, and establishing a common analytics platform based on Data Domains. While the project is still ongoing, the talk describes the methods being implemented and shares early results, key learnings, and practical recommendations for institutions undertaking similar architectural transitions.
![]() |
BiographyMariusz Sienkiewicz is a data architecture and governance expert with extensive experience in transforming large-scale data environments within highly regulated financial institutions. He currently serves as Director of the Supervisory Data Analysis Center at the Polish Financial Supervision Authority, where he leads strategic initiatives focused on data management, data integration, and analytics infrastructure supporting financial market oversight. |