Sr. AI Engineer @
IBM
MSc. Data Science & Robotics @
ETHZ •
EPFL
Background articulated around AI, DataOps & MLOps.
Expertise in GenAI (Text, Image, Audio), Deep L and Computer Vision.
Focus on Sofware Engineer practices and Business Value
I am passionate about the foundation of AI (GenAI, ML, DL, CV) and its deployment and management in industrial settings.
Over the past three years, I have been working in both academic and industrial environments, advancing the frontiers of research
and integrating the latest findings in Generative AI, Deep Learning, and Machine Learning into impactful softwares and applications.
I see the modern practice of a Machine Learning software engineer as follow;
Technically I am deeply interested into
While(True) : Sensors -> Data Wrangling -> Machine Learning -> Real Word
[PSEUDO CODE 1.] Synergy of my bachelor's in Microengineering and my Master's in Data Science and Robotics
I am open to connect for Machine Learning / Data Science / Software Engineer related topics
Programation |
Proficiency : Python, Pyspark Good : C/C++, SQL, HTML, CSS |
Machine Learning Library | PyTorch, PyTorch Lightning, TensorFlow, Keras, Scikit-Learn |
Development Tools | Git & Github, bash/UNIX, Jira |
Cloud and Deployment | AWS, Azure, Dockers |
Operating Systems | Linux, Windows |
CPU/GPU Computing | SLURM, Kubernetes, DDP and Multiprocessing |
Hackathon Microsoft x IBM | 1st /250 teams - May 2024
Hackathon UBS | Finalist Top 10/230 teams - Nov 2023
Hackathon Databricks | 2nd/130 teams - Sep 2023
Nominated project | 50 years event @ EPFL - Sep 2019
Nowadays, it is an amazing time to work in AI. LLMs and vLLMs serve as interfaces for computers to understand human language and the human world. By connecting these interfaces to other components of an operating system, we enable interactions with tools (software), functions (Python, C++), internal databases (RAG), and memory management.
This vision is revolutionary and will radically change how we interact with computers, making our daily lives more productive. All the giant tech companies are racing toward this vision by launching products, conducting research, and initiating open-source projects.
I am excited to be part of this journey and look forward to contributing to this transformative path.
François Chollet Software Engineer @ Google |
Pr. Yann LeCun Chief AI Scientist @ Facebook |
Pr. Andrex Ng Co-founder of Google Brain |
Pr. Michael Bronstein Head of Graph Learning @ Twitter |
![]() |
![]() |
![]() |
![]() |
“The future of Ai will be discrete as well as continuous” | “[...] self-supervised learning (SSL) is one of the most promising ways to build [...] a form of common sense in AI systems.” | “The consistency of the data is paramount” | “Geometric Deep Learning aims to bring geometric unification to deep learning [...]” |
Deep learning is limited; we can use deep learning for a continuous problem where the data is interpolative and has a learnable manifold with a dense sampling across the entire surface of the manifold between which we need to make predictions. For Francois Chollet, generalization itself is by far the most crucial feature of artificial intelligence. The early history of AI focused on defining priors in data structures and algorithms and tended to leave out experience and learning. The field of AI post the deep-learning revolution seems to have the opposite obsession. In the last few years, there has been much emphasis on learning as much as possible from scratch. The connection to Chollet’s ideas is that the deep learning era focuses on maximally exploiting experience in the form of training data. In Chollet’s view, task-specific skills cannot tell us anything about an agent’s intelligence because it is possible to “buy arbitrary performance” by simply defining better priors or utilizing more experience. “Defining better priors or training from more experience reduces uncertainty and novelty and thus reduces the need for generalization.” |
In recent years, the AI field has made massive strides in developing AI systems that learn from vast amounts of carefully labeled data. However, moving forward, it seems impossible to annotate the vast amounts of data with everything that we care about. Supervised learning is a bottleneck for allowing more intelligent generalist models to do various jobs and gain new abilities without extensive amounts of labeled data. For Yann LeCun a promising way to approximate such common sense is self-supervised learning (SSL). Self-supervised tasks act as a proxy strategy to learn representations of the data using pseudo labels. These pseudo labels are created automatically based on the attributes found in the data. Nevertheless, the outcome of this created task is habitually dismissed. Instead, we focus on the learned intermediate representation with the hypothesis that this representation can offer excellent semantic and benefit a diversity of useful downstream tasks. |
Data is everything in modern-day machine learning but is often neglected and not handled properly in AI projects. As a result, hundreds of hours are wasted on tuning a model built on low-quality data. That’s the main reason why the model's accuracy is significantly lower than expected - it has nothing to do with model tuning. The Data-Centric Architecture treats data as a valuable and versatile asset instead of an expensive afterthought. Data-centricity significantly simplifies security, integration, portability, and analysis while delivering faster insights across the entire data value chain. Andrew Ng’s idea is relatively simple; let us hold the model architecture fix assuming it is good enough and instead iteratively improve the data quality. Instead of asking, “What data do I need to train a useful model?”, the question should be: “What data do I need to measure and maintain the success of my ML application?” |
Modern machine learning operates with large high-quality dataset, which together with the appropriate computational resources, motivate the design of rich function space with the capacity to interpolate over the data points. "Symmetry, as wide or narrow as you may define its meaning, is one idea by which man through the ages has tried to comprehend and create order, beauty, and perfection. (Hermann Weyl) Since the early days, researchers have adapted neural networks to exploit the low dimensional geometric arising from physical measurements, such as grids in images, sequences in time series or position and momentum in the molecule, and their associated symmetries such as translation or translation rotation. Remarkably, the essence of deep learning is built from two simple algorithmic principles: first, the notion of representation or feature learning, and second, learning by local gradient-descent type methods, typically implemented as backpropagation. Geometric Deep Learning unifies a broad class of ML problems from the perspectives of symmetry and invariance. These principles not only underlie the breakthrough performance of convolutional neural networks and the recent success of graph neural networks but also provide a principled way to construct new types of problem-specific inductive biases |
Take Away for Ai in Industry :
|
Take Away for Ai in Industry :
|
Take Away for Ai in Industry :
|
Take Away for Ai in Industry :
|