I am a Data Scientist with expertise in AI, machine learning, and data analysis. With a strong
background in predictive modeling and deep learning, I enjoy turning
data into actionable insights to help businesses thrive.
Thomas More De Nayer - Sint-Katelijne-Waver, Belgium | 09/2021 -
01/2024
Developed supervised anomaly detection algorithms using K-Nearest Neighbors and Isolation
Forests on synthetic dataset for energy distribution systems.
Deployed regression and deep neural networks (DNNs) for time series prediction of
single household energy consumption, directly
enhancing resource allocation efficiency and minimizing downtime.
Performed system and network maintenance with a focus on troubleshooting issues that
involved data collection and analysis, ensuring optimal network performance.
Developed basic Python scripts to automate repetitive tasks such as system monitoring and
log
analysis, improving efficiency by identifying anomalies in system performance.
Performed system and network maintenance with a focus on troubleshooting issues that
involved data collection and analysis, ensuring optimal network performance.
Developed basic Python scripts to automate repetitive tasks such as system monitoring and
log
analysis, improving efficiency by identifying anomalies in system performance.
Worked closely with mentors/colleagues and teams to gather system performance data,
contributing to ongoing
optimization projects. Analyzed data patterns to detect system failures before they
occurred.
Assisted in the configuration and deployment of hardware and software, integrating systems
in a
scalable and data-driven manner, which supported broader IT strategy.
Projects
A showcase of my recent work in AI and Machine Learning.
This project demonstrates the development of a powerful retrieval-based question-answering
system using LangChain and Hugging Face. The system leverages state-of-the-art language
models to provide accurate and contextually relevant answers to user queries.
The Automated Image Captioning with AI project is designed to automatically generate
descriptive captions for images extracted from a given website URL or a local folder.
Leveraging state-of-the-art machine learning models, this project provides an interactive
web interface using Gradio, making it easy for users to input a URL or select images from a
folder, generate captions, and save the results.
An advanced text-based adventure game powered by Large Language Models (LLMs) that
demonstrates the practical application of AI/ML in interactive entertainment. The game
features dynamic quest generation, intelligent NPC interactions, and content safety
validation.
This project aims to predict earthquake magnitudes and locations by analyzing seismic data to
identify patterns and trends. The goal is to provide valuable insights for earthquake
preparedness and mitigation.
This project aims to classify images of an industrial product from a QA dataset provided by
the AI4IM project. The
dataset consists of QA images of injection-moulded products taken by four different cameras
from four different directions.
This is a AI classification project using PyTorch and 2 classes. Used trained model in a
Flask application and showed the AI classification prediction results on web page. Also
using two Docker containers with base images from Nvidia's nvidia/cuda from Docker hub.
In this project, I share the results and implementations of building a powerful
retrieval-based question-answering system using LangChain and Hugging Face. I walk you
through the process step-by-step, leveraging the capabilities of state-of-the-art language
models. For this project, I used markdown files from the Weights & Biases documentation as
the data source.
Key Features:
Loading and Preparing Data
Splitting Documents into Sections
Generating Embeddings
Storing Embeddings in Chroma
Alternative Solution: Prompt-Based Approach
Creating a RetrievalQA Chain
Running the Query
Technologies: LangChain, Hugging Face, Python
Practical Applications: Enhanced information retrieval, improved user
experience, and efficient knowledge management.
The Automated Image Captioning with AI project is designed to automatically generate
descriptive captions for images extracted from a given website URL or a local folder.
Leveraging state-of-the-art machine learning models, this project provides an interactive
web interface using Gradio, making it easy for users to input a URL or select images from a
folder, generate captions, and save the results in any format to enable modifying later as
well.
Development Process:
Situation: A news agency publishes hundreds of articles daily, each
containing several images relevant to the story. Writing appropriate and descriptive
captions for each image manually is a tedious task and might slow down the publication
process. The agency needed a solution to expedite this process while ensuring the
captions were accurate and contextually relevant.
Task: Develop an automated image captioning tool that could generate
descriptive captions for images extracted from a given website URL or a folder. The tool
needed to be user-friendly, efficient, and capable of producing high-quality captions
that enhance accessibility and improve SEO.
Action: Implemented an automated image captioning program that works
directly from a URL or a folder. The user provides the URL or selects images from a
folder, and the code generates captions for the images found. The output is a text file
that includes all the image URLs or image file names along with their respective
captions.
Result: By integrating this automated image captioning tool, the news
agency is able to expedite its publication process significantly. The tool ensures that
all images come with appropriate descriptions, enhancing accessibility for visually
impaired readers and improving the website's SEO. This broadens the agency's reach and
engagement with a more diverse audience base.
Key Features:
Image Extraction: Automatically extracts image URLs from the
provided website URL.
Caption Generation: Utilizes the Salesforce/blip-image-captioning-large model
to generate descriptive captions for each image.
Interactive Interface: Provides a user-friendly interface using Gradio for
easy interaction.
Modify Captions: Allows users to edit the generated captions.
Save Captions: Allows users to save the generated captions to a text file.
Clear Interface: Includes a "Clear" button to reset the interface and clear
all data.
This project demonstrates transfer learning using ResNet18 for image classification. The
model was trained on a dataset of hymenoptera images and deployed using a Flask application.
Deployment in a Flask web application, leveraging Docker containerization
Practical Applications: This project showcases the use of transfer learning
to create a scalable image classification model. By leveraging a pre-trained ResNet18 model
and fine-tuning it on a custom dataset, the project demonstrates how to achieve high
accuracy in image classification tasks. The deployment of the model in a Flask application
provides a user-friendly interface for image classification, making it accessible to a wide
range of users.
Problem we are trying to solve: To build a Machine Learning model, that can
evaluate images
as good or bad, enabling the company to use the model for quality control
purposes. This is a project to classify images from QA dataset from the AI4IM
project.
Using AI principles:
Feature engineering
Hyperparameter tuning
Detect & avoid overfitting
Method selection
Appropriate performance criteria
Test the result on independent test set
Challenge => To build a single class classifier that has trained on
enough good images so that it can detect bad images by only looking at them as seeing they are not good
images, so that classify them as 'not good'='bad' images
Methods: Random Forest Regression, Gradient Boosting Regression, Support Vector
Regression (SVR) (regression version of Support Vector Machine (SVM)), Neural Networks using
TensorFlow/Keras, XGBoost Regression
Using AI principles:
Data preprocessing: Data cleaning, Pre-processing,
Feature engineering and Splitting the Data
Model Training: Trained the machine learning models using the preprocessed data.
Hyperparameter Tuning: Random Search, Bayesian Optimization
Inference/Predictions: Used the trained models to make earthquake predictions.
Data Visualization: Visualized earthquake patterns, trends, and predictions using the provided
visualization tools.
Data: The project uses earthquake data from Turkey, including features like date,
location, latitude, longitude, magnitude, depth, and more. The dataset is available in the data
directory and in dataset location: Turkey 20 Years Earthquakes CSV. Please unzip the dataset and place it in the
root directory.
Education
Formal training, hands-on experience, and educational-theoretical background.
Bachelor of Computer Science
Specialization in Artificial Intelligence
Thomas More University of Applied Sciences - Sint-Katelijne-Waver, Belgium |
2021
- 2024
Graduated Cum Laude with a thesis on Medical Image Processing with
AI:
Scale Detection.
Solid fundamental knowledge on how ML systems work and statistics.
Skills Acquired: Python, TensorFlow, Keras, YOLO, EasyOCR, OpenCV,
PyTorch,
Scikit-learn, Pandas, Matplotlib, NumPy, Feature engineering, Data preprocessing and
cleaning and more.
Key Achievements: Developed AI-powered
medical
imaging solutions that
improved diagnostic accuracy by 71.43%.
Soft Skills: Collaboration
with cross-functional teams,
problem-solving,
innovative thinking and research. Effective communication skills and teamwork, empathy,
attention to detail
Practical Experience, Industry Exposure:
Gained deep knowledge of AI frameworks, machine learning, and neural networks through
real-world projects.
Engaged in guest lectures and workshops by industry leaders from top AI, healthcare,
and technology companies.
Hands-on experience with real-world datasets and practical problem-solving
scenarios.
Gained practical problem-solving experience through real-world projects involving:
Computer vision, deep learning, and medical imaging.
NLP, AI applications in healthcare, finance, and traffic
management.
Large Language Models (LLMs) and AI pipelines.
Workshops on AI ethics, business applications, adversarial learning, MLOps, and
deployment strategies in real-world environments.
Theoretical Knowledge and Credit Courses on AI and Machine Learning
Participated in Kaggle competition of facial recognition.
Developed ML models applying the knowledge about different A.I. Frameworks into
practice.
Understanding of the theoretical underpinnings of these frameworks and their
application in solving real-world problems.
A.I. Applications
Covered topics include common AI applications such as computer vision, NLP,
health, Industry, Consultancy, MLOps, time-series forecasting, insurance &
finance. Gained understanding of AI ethics & dangers and explored on-edge AI.
Exposed to real-world AI use cases presented by industry experts. Acquired
skills to identify and understand different aspects of AI use cases and apply
knowledge to practical scenarios.
Engaged in an AI project:
Worked on a visual QA dataset from the AI4IM project, implementing
models to classify images captured by different cameras.
Utilized Python frameworks, tools, and models of choice to build a
working AI solution, with a focus on achieving high performance on an
independent test set.
Project evaluation focused on application of AI principles such as
feature engineering, hyperparameter tuning, avoidance of overfitting,
and method selection.
Explored the theoretical aspects of AI applications and their impact on various
industries.
A.I. Fundamentals
Covered fundamental techniques of AI and Machine Learning, including
classification, regression, decision trees, neural networks, deep learning,
ensemble methods, and reinforcement learning.
Developed skills to select appropriate AI algorithms based on problem settings
and complexity requirements.
Evaluated the relevance of AI techniques for specific applications and
understood the connections between different techniques.
Learned to explain the relationship between given data and the quality of output
in AI applications.
Gained a deep understanding of the theoretical foundations of AI and ML
algorithms.
Python for A.I.
Developed proficiency in using Python and associated libraries such as NumPy,
Pandas, Matplotlib, OpenCV, Scikit-learn, TensorFlow, and Keras for implementing
AI solutions.
Developed an AI project in Python. Successfully implemented various AI models,
including Neural Networks, Decision Trees, SVM, and k-NN, and effectively
visualized input data, labels, and results using appropriate plots.
Explored AI frameworks in Python and developed a project using them.
Gained theoretical knowledge on the implementation and optimization of AI models
using Python.
Cloud for A.I.
Explored cloud deployment strategies for AI models, including using platforms
like AWS.
Learned about cloud-based tools and services for AI, such as AWS SageMaker.
Developed skills in deploying, scaling, and managing AI models in the cloud.
Gained theoretical knowledge on the benefits and challenges of cloud-based AI
solutions.
Full-Stack Web Development Course
HYF - Brussels, Belgium | 2019 - 2020
Hands-on experience with HTML, CSS, JavaScript, React, Node.js, and
MongoDB.
Key Project: Built dynamic web applications using modern JavaScript
frameworks, focusing on user interactivity and backend integration.
Soft Skills: Teamwork, communication, and project management under
tight deadlines. Presented projects to non-technical audiences, gaining insights into industry
practices.
Worked on projects using Agile methodology.
Developed skills in HTML, CSS, JavaScript, React, Node.js, and databases.
Collaborated with team members to build full-stack web applications.
Gained hands-on experience with version control systems like Git and GitHub.
Bachelor of Systems Engineering
Specialisation in Mechanical Engineering
Turkish State University - Ankara, Turkey | 2010 - 2014
Graduated with distinction in Systems Engineering.
Gained expertise in system analysis, design, and engineering processes.
Key Skills: Systems analysis, problem-solving, and teamwork.
×
Volunteer Work
Contributing to the community through various initiatives.
React Website Development - Red Cross CBS Codeathon 6.0
The EGG Brussels | 17/10/2019 – 19/10/2019
Improved and developed various functionalities, including:
A custom-developed SMS gateway.
Responsive versions of the report/line list.
First versions of two analytics graphs.
A mobile reporting application.
Improvements to the CBS website and a light version of analytics for limited bandwidth
connections.