Hello, I am

Fatima Yousif

MSc in Intelligent Field Robotic Systems (Currently, Spain)

ROS
OpenCV
ROS2
TensorFlow
PyTorch
Python
C++
MATLAB
Gazebo
Stonefish
Git
Gazebo
ROS
ROS2
C++
OpenCV
MATLAB
TensorFlow
Stonefish
Git
Python
PyTorch
ROS
PyTorch
Git
ROS2
C++
Stonefish
OpenCV
Gazebo
TensorFlow
MATLAB
Python
ROS
OpenCV
Python
ROS2
Git
Stonefish
Gazebo
TensorFlow
MATLAB
PyTorch
C++

About Me

I am an Erasmus Mundus Joint Masters Scholar in Intelligent Field Robotic Systems at University of Girona, Spain and the University of Zagreb, Croatia. My research interests span a diverse range of cutting-edge disciplines, including robotics, computer vision, and machine/deep learning.

I am passionate about exploring the intersection of these fields to push the boundaries of technology and create innovative solutions that address complex real world challenges.

Additionally, I have proven leadership qualities demonstrated in global communities including Google Developer Student Clubs, TEDx, Google Developer Groups Live, U.S. Embassy programs, and 10Pearls.

Education

2023 - 2025

Erasmus Mundus Joint Masters in Intelligent Field Robotic Systems | University of Girona & University of Zagreb

Semester I & II in Girona: Autonomous Systems, Machine Learning, Multiview Geometry, Probabilistic Robotics (Kalman Filtering), and Robot Manipulation, Localization (SLAM), Planning, Perception (Computer Vision), and Intervention.

Semester III in Zagreb: Aerial Robotics, Multi-Robot Systems, Human-Robot Interaction, Robotic Sensing, Perception, & Actuation, Deep Learning, and Ethics & Technology.

2018 - 2022

B.E. in Software Engineering | Mehran University of Engineering and Technology

Agent Based Intelligent Systems, Data Science & Analytics, Simulation & Modeling, Cloud Computing, Statistics and Probablity

CGPA 3.96 / 4.00 - Silver Medal Distinction & First Position

Experiences

03/2025 - 05/2025

Master Thesis/Intern | Saxion Smart Mechatronics and Robotics Research Group

My research work on “Vision-based Tracking and Following of a Moving Target Using a UAV”, focused on challenges mainly in RGB modality for robust person detection and tracking, dealing with occlusions and identity switches, and real-time prediction in dynamic environments. The work integrates a hybrid system combining filtering-based tracking (Kalman Filter), vision-based deep learning models (YOLO), and adaptive control strategies. Overall, this research seeks to leverage situational-aware techniques to enable intelligent UAV deployment in security-critical scenarios detecting potential suspicious targets and track them in real-time with minimal response time.

10/2024 - 10/2024

ROSCon 2024 Diversity Scholar | Open Robotics, Denmark

I secured a diversity scholarship to attend the ROSCon 2024 in Denmark where I had the opportunity to network with companies and ROS contributors globally. I specifically got extensive hands-on experience by attending the workshops named “Open source, open hardware hand-held mobile mapping system for large scale surveys” which gave exposure to essential processes such as LIDAR odometry and multi-session refinement for large-scale mapping and “ros2_control” where we learned about controller chaining, fallback controllers, and async controllers.

06/2024 - 08/2024

Robotics Intern | Paltech Robotics GmbH

Worked on testing and comparing two new ultrasonic sensors i.e. Bosch and Valeo for the collision avoidance task to include the safety braking feature (setting thresholds to slow down or stop the robot with ROS2) which involved performing multiple field tests of different high grass.

My Projects

Sim2Real: Controlling a Swarm of Crazyflies using Reynolds Rules and Consensus Protocol

This project implements swarm control for Crazyflies UAVs using Reynolds Rules for flocking and a Consensus Protocol for coordinated movement. It integrates rendezvous and formation control in ROS2 and Gazebo, enabling agents to converge and maintain geometric formations. Tested in both simulation and real-world environments, the system demonstrates adaptability and scalability.

Stereo Visual-Odometry (VO) on the KITTI Dataset

Implementation of Stereo VO pipeline in Python on the KITTI dataset. It processes stereo image data using SIFT, feature matching using BFMatcher, triangulation of points, to estimate the motion of a camera (w.r.t its starting position) in 3D space using the approach of minimizing the 3D to 2D reprojection error with PnP and RANSAC.

Aerial Robotics

In this Aerial Robotics course, lab work included design and implementations of attitude control of a quadrotor, cascade control of a single quadrotor axis in MATLAB, cascade horizontal control of quadrotor in the Gazebo simulator and on the real DJI Tello quadrotor.

Deep Learning

In this Deep Learning course lab work, PyTorch implementations included working on logistic regression and gradient descent, implementing fully connected models on the MNIST dataset, Convolutional models for image classification tasks on MNIST and CIFAR, Recurrent models for analysis of sentiment classification with the Stanford Sentiment Treebank (SST) dataset followed by detailed implementations on metric embeddings.

Human Detection and Tracking

This project focuses on human detection and tracking using the state-of-the-art YOLOv9 object detection model and the DeepSORT multi-object tracking algorithm. The methodology integrates Kalman filtering for motion prediction and deep learning-based appearance matching. The system is tested under various conditions, addressing challenges such as occlusions, identity switches, and tracking interruptions.

Frontier Based Exploration Using Kobuki Turtlebot

Using RGB-D camera mounted on a Kobuki Turtlebot, the project integrates advanced path planning techniques, combining the RRT* algorithm with Dubins path to map unknown environments with the primary objective of enabling the Turtlebot to autonomously explore unknown environments by identifying and navigating to frontiers. A hybrid control system, which merges PID control with principles from the Pure Pursuit Controller is used. Validated both in simulation and real world.

Monocular Visual Odometry (M-VO) for an Autonomous Underwater Vehicle (AUV)

M-VO for an AUV through an integrated approach combining extended Kalman filter (EKF) based navigation. The methodology employs SIFT feature detection and FLANN matching to (offline / post) process images from a ROSBag. A key contribution of this work is the incorporation of EKF to provide a refined estimation of the vehicle´s motion and trajectory.

Pose Based SLAM using the Extended Kalman Filter (EKF) on a Kobuki Turtlebot

Implemented a Pose-Based P- EKF SLAM system integrating IMU and 2D LiDAR data on a Kobuki Turtlebot. The algorithm maintains robot pose history for map building and localization, using ICP for scan matching and robust state updates. Validated in both simulated (Stonefish) and real-world environments, PEKFSLAM demonstrated superior accuracy and stability compared to conventional EKF-based SLAM approaches.

Kinematic Control System for a Mobile Manipulator, based on the Task-Priority Redundancy Resolution Algorithm

Designed and implemented a kinematic control system for a mobile manipulator (Kobuki Turtlebot 2 with a 4-DOF uArm Swift Pro), using a task-priority redundancy resolution algorithm. Developed in ROS and tested in the Stonefish simulator, the system performed complex pick-and-place tasks, including ArUco marker-based navigation.

SLAM - Differential Drive Mobile Robot

Simultaneous Localization and Mapping (SLAM) algorithms for a differential drive mobile robot with python simulations and plotting.

Behaviour trees for pick-place of objects

Implemented behavior trees using the py_trees library and validated the approach through TurtleBot simulations in RViz, demonstrating path planning and obstacle avoidance across various complex environments.

Pick and Place Application with the Staubli TS60 and TX60 Robot

Worked on industrial manipulators - TS60 and TX60 robots for classification, assembling pieces and performing pick-place tasks on simulation alongside real-robot implementation.

Palletizing Application with UR3e Collaborative Robot (CoBot)

Worked with the 6 DOF CoBot to develop a pick-and-place program with pallets to achieve simple palletizing application. Testing was conducted collaboratively in the laboratory.

Stereo Visual Odometry (VO) for Grizzly Robotic Utility Vehicle

Developed VO pipeline from stereo camera calibration, feature extraction, and matching using SURF features and utilizing bucketing strategies and circular matching for accurate apparent motion computation and effective noise/outlier rejection, Structure from motion (2D-to-2D, 3D-to-2D, and 3D-to-3D) for triangulation and refinement using bundle adjustment. The final VO trajectory was also extensively compared with GPS-generated ground truth data.

Event Based Cameras (EBC)

Analyzed EBC data by comparing it with ground truth using DAVIS. Employed a frame-based approach to convert raw event data into frames suitable for CNNs and RNNs, and applied motion compensation techniques.

Machine Vision Projects

Worked on tasks such as Augmented Reality, Camera Calibration, Detecting Aruco markers, and Generating Fiducial Makers with computer vision, and image processing in CPP.

Reinforcement Learning-Based Path Planning for Autonomous Robots in Static Environments

Implemented the Q-learning algorithm on a point (omnidirectional) robot for path planning and navigation purposes.

Image Captioning Deep Learning Model

In this undergraduate research project on image captioning, developed using an end-to-end deep learning system that combines computer vision and natural language processing to automatically generate descriptive captions for visual content.

Contact Me