Julio Faracco
At least, I have more than 10 years dedicated to Linux Operating Systems. During this time, I have worked with embedded systems, robotics, virtualization, software development and kernel (today). During my M.Sc., I also dedicated my time to work on HPC problems including the ones specific to machine learning techniques. Finally, no matter which subject I'm involved with I always try to contribute to open source communities.
Red Hat
Job title –Software Engineer
Session
Have you ever had to deal with training machine learning models where the data is very large? If the data does not fit in main memory, then how can you use GPUs if their memories are even smaller? Many of these cases require strategies for handling data sets. In this presentation, we will introduce DASF, a framework that brings together lazy data loading techniques using Dask, acceleration techniques using RAPIDS AI, and other techniques that facilitate the use of large data in ML pipelines locally or in HPC environments. We will also present a show case carried out with a company in the oil and gas sector.