• Designing Cultural Verification Frameworks for Large Language Models (LLMs) (Mandeep Rathee)
    Current alignment techniques, such as RLHF (Reinforcement Learning from Human Feedback), typically rely on a centralized set of "safety guidelines" that reflect the cultural norms of the model's developers. This often leads to cultural misalignment, where the LLM may inadvertently erase local nuances, apply inappropriate social taboos, or fail to recognize regional linguistic etiquette.
    This thesis aims to move beyond "cultural awareness" toward Cultural Verification. The student will design a systematic framework to evaluate and verify whether an LLM’s outputs adhere to the specific socio-cultural norms, value systems, and historical contexts of a target demographic.

    Key Responsibilities & Tasks:

    The research will be structured around four primary pillars, beginning with the development of a comprehensive Taxonomy of Cultural Correctness to define the specific dimensions—such as social hierarchy, religious sensitivities, and gender roles—necessary for nuanced LLM evaluation. Building upon this foundation, the study will propose a Verification Methodology centered on a "Cultural Guardrail" architecture, utilizing both programmatic checks and human-in-the-loop oversight to validate model responses. To test the efficacy of this framework, a specialized Benchmark will be curated, featuring a localized "red-teaming" suite designed to expose cultural hallucinations and latent insensitivities. Finally, a Comparative Analysis will be conducted to measure the performance of leading models, including GPT-4, Llama 3, and Claude, against these proposed designs across two or more distinct cultural contexts.



    Anforderungen:
    - Experience with LLMs, HuggingFace
    - LLM fine-tuning
    - Python
    - Knowledge of German culture and language is a plus.

    Verwandte Arbeiten:
    - Atari et al. (2023): "Which Humans? Probabilistic Expectations and Social Values in LLMs."
    - Wang, Jiahao, Songkai Xue, Jinghui Li, and Xiaozhen Wang. "Diverse Human Value Alignment for Large Language Models via Ethical Reasoning." In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, vol. 8, no. 3, pp. 2637-2648. 2025.
    - Venktesh, V., Mandeep Rathee, and Avishek Anand. "Trust but verify! a survey on verification design for test-time scaling." arXiv preprint arXiv:2508.16665 (2025).


    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Mandeep Rathee wenn Sie an diesem Thema interessiert sind.
  • Object and Texture Recognition via Acoustic Sensing in a Humanoid Robotic Hand (Wadhah Zai El Amri)
    In this project, you will contribute to the cutting edge field of robotic tactile perception. Working with a humanoid robotic hand equipped with highly sensitive vibration microphones, your goal is to give the robot the ability to "feel" and identify what it is touching. As the robotic hand manipulates a diverse set of everyday objects, the embedded sensors will capture complex acoustic and vibration signals. You will use this raw sensory data to develop an intelligent system capable of accurately recognizing both the specific object and its surface texture.

    Key Responsibilities & Tasks:

    - Experimental Design and Data Collection: Design and execute robust data collection protocols.
    - Signal Processing: Preprocess the raw audio and vibration signals. This involves filtering noise, segmenting the data, and extracting meaningful temporal and frequency domain features (e.g., spectrograms) that characterize different textures and materials.

    - Machine Learning Pipeline Development: Design, implement, and train artificial neural networks. The pipeline will take the processed acoustic signals as input and output the correct object and texture classifications.

    - Benchmarking and Evaluation: Rigorously test your models. You will benchmark different neural network architectures and classical machine learning algorithms against one another, analyzing key performance metrics like accuracy, robustness, and computational efficiency.

    Anforderungen:
    - Mathematics: A strong foundation in Linear Algebra and the mathematical principles underlying machine learning.

    - Programming: Proficiency in Python (essential for deep learning frameworks like PyTorch) and C++ (essential for hardware communication and robotics integration).

    - Signal Processing: Solid understanding of digital signal processing techniques, including Fourier transforms, filtering, and time frequency analysis.

    Verwandte Arbeiten:
    https://arxiv.org/abs/2601.20555

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Wadhah Zai El Amri wenn Sie an diesem Thema interessiert sind.
  • Multi-Layer Traversability Map for an Autonomous Off-Road Mobile Robot (Wadhah Zai El Amri)
    In this thesis/project, you will create a multi-layer traversability map for an autonomous off-road mobile robot. You will use an RGB-D camera to build a robot-centric map with derived geometric features, including slope, step height, holes, roughness, etc. The map should be generic so that it can be used with multiple robots; also, the robot size and capabilities will be parametrized. You will use datasets for outdoor environments, including roadside, grasslands, and forest edges.

    The tasks include:

    - Generate 2.5D elevation maps.
    - Derive Traversability Features (Geometry Layer), i.e., slope, step height, holes, roughness, etc.
    - Create a Semantic Layer (from RGB) to distinguish between grass, dirt, asphalt, rock, vegetation, obstacle, etc.
    - Define a robot-aware traversability scoring, considering the volume of the robot and its capabilities, to create a cost map for path planning.

    Anforderungen:
    - Linear Algebra
    - C++, Python, ROS/ROS2

    Verwandte Arbeiten:
    https://nav2.org
    https://github.com/ANYbotics/grid_map

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Wadhah Zai El Amri wenn Sie an diesem Thema interessiert sind.
  • Body-Centric Teleoperation via Pose Estimation (Malte Kuhlmann)
    Traditional teleoperation often fails when the operator moves out of a fixed sensor’s "sweet spot." This thesis aims to improve this by estimating the operator's arm pose relative to their own torso, enabling a more flexible and intuitive workspace.

    Key Objectives:
    - Research State-of-the-Art human pose estimation models for real-time 3D skeletal tracking.
    - Develop a robust pipeline to extract 3D landmarks from RGB-D or monocular feeds.
    - Modify the existing ROS control architecture to transform arm coordinates into a body-relative coordinate system

    Anforderungen:
    - Python, PyTorch, Git, Linux
    - ROS
    - Machine Learning, Computer Vision

    Verwandte Arbeiten:
    - Ce Zheng, Wenhan Wu, Chen Chen, Taojiannan Yang, Sijie Zhu, Ju Shen, Nasser Kehtarnavaz, and Mubarak Shah. 2023. Deep Learning-based Human Pose Estimation: A Survey. ACM Comput. Surv. 56, 1, Article 11 (January 2024), 37 pages. https://doi.org/10.1145/3603618

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Malte Kuhlmann wenn Sie an diesem Thema interessiert sind.
  • Personalized Hand-to-Robot Mapping for Teleoperation (Malte Kuhlmann)
    This thesis aims to improve teleoperation precision by automatically adjusting the hand-to-robot mapping based on the operator's unique anatomy. By using markerless tracking, we can estimate finger lengths in real-time to create a calibrated, intuitive control loop.

    Key Objectives:
    - Keypoint Tracking: Implement state-of-the-art markerless 3D hand tracking.
    - Biometric Estimation: Calculate individual finger lengths from extracted keypoints.
    - Control Integration: Automate personalized adjustments within a ROS-based teleoperation workflow.

    Anforderungen:
    - Python, PyTorch, Linux, Git
    - ROS (Robot Operating System)
    - Machine Learning, Computer Vision

    Verwandte Arbeiten:
    - Javier Romero, Dimitrios Tzionas, and Michael J. Black. 2017. Embodied hands: modeling and capturing hands and bodies together. ACM Trans. Graph. 36, 6, Article 245 (December 2017), 17 pages. https://doi.org/10.1145/3130800.3130883
    - Daanish M. Mulla, Nigel Majoni, Paul M. Tilley, Peter J. Keir, Two cameras can be as good as four for markerless hand tracking during simple finger movements, Journal of Biomechanics, Volume 181, 2025, 112534, ISSN 0021-9290, https://doi.org/10.1016/j.jbiomech.2025.112534.

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Malte Kuhlmann wenn Sie an diesem Thema interessiert sind.
  • Road-Shoulder Traversability from Stereo/LiDAR for Mowing Robots (Dr. Nicolás Navarro)
    In this thesis, you will derive road-shoulder traversability from stereo/LiDAR, capturing slope, curb, ditch, and surface roughness to enable safe planning near road edges, and validate predictions against real field logs.

    The tasks include:
    * Build a pipeline to fuse stereo depth and/or LiDAR into terrain primitives: slope, curb/edge step, ditch depth, roughness (e.g., surface variance, ESF/planarity).
    * Convert primitives to a traversability costmap/risk layer with clear thresholds and uncertainty.
    * Evaluate on off-road/roadside datasets (e.g., RELLIS-3D, RUGD) and replay field logs for end-to-end validation (near-miss/stop rates, path feasibility).
    * Report metrics: ROC/PR for traversable vs. non-traversable, curb F1/AP, slope/roughness error vs. ground truth or survey labels, and planning success on replays.

    Anforderungen:
    * Robotics/Perception basics; 3D geometry & point-cloud processing
    * Python/C++ (ROS 2), Git, Linux

    Verwandte Arbeiten:
    Jiang, P., Osteen, P., Wigness, M., & Saripalli, S. (2021). RELLIS-3D Dataset: Data, Benchmarks and Analysis. IEEE ICRA, 1110–1116. https://doi.org/10.1109/ICRA48506.2021.9561251

    Wigness, M., Eum, S., Rogers, J., & Kwon, H. (2019). A RUGD Dataset for Autonomous Navigation and Visual Perception in Unstructured Outdoor Environments. IEEE/RSJ IROS. https://doi.org/10.1109/IROS40897.2019.8968283

    Gasparino, M. V., Sivakumar, A. N., Liu, Y., Velasquez, A. E. B., Higuti, V. A. H., Tran, H., & Chowdhary, G. (2022). WayFAST: Navigation with Predictive Traversability in the Field. IEEE Robotics and Automation Letters, 7(4), 10651–10658. https://arxiv.org/abs/2203.12071

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Dr. Nicolás Navarro wenn Sie an diesem Thema interessiert sind.
  • Autonomous Off-road driving using 3D Voxel Maps (Dr. Nicolás Navarro)
    In this thesis, you will build a multi-modal 3D voxel mapping stack (elevation + occupancy/ESDF with semantics) for lane-less roadside terrain, producing uncertainty-aware traversability that outperforms 2D costmaps.

    The tasks include:
    * Integrate stereo/LiDAR (+ optional IMU/GNSS) into a real-time voxel map (TSDF/ESDF); maintain map updates under motion.
    * Fuse semantic cues (grass/curb/obstacle masks) into elevation/occupancy; model uncertainty for risk-aware planning.
    * Validate planner-ready layers (ESDF, traversability) and benchmark vs. 2D baselines on roadside datasets.
    * Run ablations on depth sparsity/noise, voxel resolution, and update rates.

    Anforderungen:
    * Robotics/SLAM basics; probability & linear algebra
    * C++ and/or Python; ROS 2; Git, Linux

    Verwandte Arbeiten:
    Oleynikova, H., Taylor, Z., Fehr, M., Nieto, J., & Siegwart, R. (2017). Voxblox: Incremental 3D Euclidean Signed Distance Fields for On-Board MAV Planning. IEEE/RSJ IROS, 1366–1373. https://doi.org/10.1109/IROS.2017.8202315

    Millane, A., Oleynikova, H., Wirbel, E., Steiner, R., Ramasamy, V., Tingdahl, D., & Siegwart, R. (2024). nvblox: GPU-Accelerated Incremental Signed Distance Field Mappingnvblox: GPU-Accelerated Incremental Signed Distance Field Mapping. IEEE RA-L / ICRA. https://ieeexplore.ieee.org/document/10611532

    Miki, T., et al. (2022). Elevation Mapping for Locomotion and Navigation using GPU. IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). https://ieeexplore.ieee.org/document/9981507

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Dr. Nicolás Navarro wenn Sie an diesem Thema interessiert sind.
  • Continual Learning for Invasive Species Detection in Field Robotics (Dr. Nicolás Navarro)
    In this thesis, you will design a continual learning pipeline to handle distribution shift (new invasive grass types, seasonal/region changes) without catastrophic forgetting. Training occurs offboard (in the cloud/lab) with periodic rollout to robots; the system must propose new classes and preserve prior performance across the entire fleet.

    The tasks include:

    * Build a class-incremental pipeline (rehearsal/coresets, regularization, knowledge distillation) with reproducible splits (old vs. new classes).
    * Add novelty/OOD detection to flag potential new species; integrate weak labels and active learning loops.
    * Define metrics and dashboards for pre/post deployment (mAP on old classes, mAP on new classes, average forgetting, calibration).

    Anforderungen:
    * Machine Learning, Deep Learning (classification/detection)
    * Computer Vision (PyTorch/TensorFlow), Python, Git, Linux

    Verwandte Arbeiten:
    Pagé-Fortin, M. (2023). Class-Incremental Learning of Plant and Disease Detection: Growing Branches with Knowledge Distillation. Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (ICCVW 2023). https://doi.org/10.1109/ICCVW60793.2023.00066

    Li, D., Yin, Z., Zhao, Y., & Zhang, H. (2024). Rehearsal-based Class-Incremental Learning Approaches for Plant Disease Classification. Computers and Electronics in Agriculture, 224, 109211. https://doi.org/10.1016/j.compag.2024.109211

    Shmelkov, K., Schmid, C., & Alahari, K. (2017). Incremental Learning of Object Detectors without Catastrophic Forgetting. Proceedings of the IEEE International Conference on Computer Vision (ICCV 2017), 3420–3429. https://doi.org/10.1109/ICCV.2017.368

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Dr. Nicolás Navarro wenn Sie an diesem Thema interessiert sind.
  • Robust Grass Recognition under Variable Conditions (lighting/weather/season/motion) (Dr. Nicolás Navarro)
    In this thesis, you will develop a robust vision method for recognizing invasive grass patches under challenging lighting conditions, weather conditions, seasonal changes, and motion blur—mitigating domain shift via domain generalization and test-time adaptation.
    The tasks include:

    * Curate a multi-condition dataset (day/night, sun/cloud, rain/fog, seasons, motion blur) and define train/val/test splits for robustness.
    * Implement strong baselines (e.g., SegFormer/FCN) with targeted augmentations/invariances (exposure, color, blur, weather).
    * Add domain generalization (e.g., style/appearance diversification) and test-time adaptation (entropy- or consistency-based).
    * Evaluate robustness across conditions (mIoU/mAP by domain; calibration), and validate in field tests on roadside scenes.

    Anforderungen:
    * Machine Learning, Computer Vision
    * Python, PyTorch/TensorFlow; Git, Linux

    Verwandte Arbeiten:
    Lottes, P., Behley, J., Milioto, A., & Stachniss, C. (2018). Fully Convolutional Networks with Sequential Information for Robust Crop and Weed Detection in Precision Farming. IEEE Robotics and Automation Letters, 3(4), 3097–3104. https://doi.org/10.1109/LRA.2018.2846289

    Lottes, P., Behley, J., Chebrolu, N., Milioto, A., & Stachniss, C. (2020). Robust Joint Stem Detection and Crop–Weed Classification Using Image Sequences for Plant-Specific Treatment in Precision Farming. Journal of Field Robotics, 37(1), 20–34. https://doi.org/10.1002/rob.21901

    Wang, P., Tang, Y., Luo, F., Wang, L., Li, C., Niu, Q., & Li, H. (2022). Weed25: A Deep Learning Dataset for Weed Identification. Frontiers in Plant Science, 13, 1053329. https://doi.org/10.3389/fpls.2022.1053329

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Dr. Nicolás Navarro wenn Sie an diesem Thema interessiert sind.
  • Replication and Comparative Evaluation of Graph-Based Multi-Target Multi-Camera Tracking Methods (Dr. Marco Fisichella)
    This thesis focuses on the replication and evaluation of two state-of-the-art approaches for multi-target multi-camera tracking (MTMCT), both of which integrate graph-based representations for associating tracked entities across disjoint views. The selected methods—LaMMOn (Language Model + GNN-based tracking) and graph-based tracklet features—demonstrate promising results in online and offline MTMCT scenarios, respectively, using advanced deep learning and graph similarity techniques.

    The student will be expected to reproduce the results presented in the following two papers, carefully re-implementing their pipelines and evaluating their performance under consistent experimental settings:

    LaMMOn: Language Model Combined Graph Neural Network for MTMCT in Online Scenarios
    Link: https://link.springer.com/article/10.1007/s10994-024-06592-1

    Graph-based Tracklet Features for MTMCT
    Link: https://ieeexplore.ieee.org/abstract/document/10121650

    Both methods target real-world intelligent transportation applications and operate on benchmark datasets such as CityFlow. A critical aspect of the thesis will be the comparative evaluation of these methods under controlled conditions to analyze their strengths, limitations, and practical deployment feasibility.

    ## Central research questions: ##
    1. What are the computational and design trade-offs between language model-based and graph-based approaches to MTMCT?
    2. How do the two systems compare in terms of tracking accuracy (e.g., IDF1, HOTA), real-time performance (FPS), and modularity?

    ## Key components of the thesis: ##
    1. Re-implementation of both pipelines, respecting the architecture and methods described in the papers.
    2. Setup of evaluation experiments on datasets such as CityFlow, TrackCUIP.
    3. Quantitative comparison using tracking metrics (e.g., HOTA, IDF1, MOTA, FPS).
    4. Qualitative analysis of system behavior in edge cases or failure scenarios.
    5. Optional: Visualization tools for tracking results and error patterns.

    Anforderungen:
    - Strong programming skills in Python.
    - Experience with PyTorch (or TensorFlow).
    - Basic knowledge of computer vision, deep learning, and graph neural networks.
    - Familiarity with tracking evaluation protocols and benchmark datasets is a plus.

    Verwandte Arbeiten:
    - LaMMOn: language model combined graph neural network for multi-target multi-camera tracking in online scenarios: https://link.springer.com/article/10.1007/s10994-024-06592-1
    - Multi-vehicle multi-camera tracking with graph-based tracklet features: https://ieeexplore.ieee.org/abstract/document/10121650
    - AIC21-MTMC baseline: https://github.com/LCFractal/AIC21-MTMC
    - CityFlow Dataset: https://www.aicitychallenge.org/
    - MGMN: Multilevel Graph Matching Networks for Deep Graph Similarity Learning
    Paper: https://arxiv.org/abs/2007.04395
    Source: https://github.com/ryderling/MGMN
    - SimGNN: A Neural Network Approach to Fast Graph Similarity Computation
    Paper: https://arxiv.org/abs/1808.05689
    Source: https://github.com/benedekrozemberczki/SimGNN

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Dr. Marco Fisichella wenn Sie an diesem Thema interessiert sind.
  • Federated Fine-Tuning of Large Language Models with Heterogeneous Low-Rank Adaptations (Dr. Marco Fisichella)
    This thesis investigates the intersection of federated learning and parameter-efficient fine-tuning of Large Language Models (LLMs), with a focus on Low-Rank Adaptation (LoRA) techniques. The goal is to develop and evaluate strategies for fine-tuning LLMs across decentralized clients, each using their own local LoRA modules tailored to their data distributions.

    ## The central research questions are: ##
    1. How can clients collaboratively fine-tune a shared LLM backbone using local, heterogeneous LoRA modules without sharing raw data?
    2. How can we effectively aggregate model updates in the presence of statistical heterogeneity across clients?
    3. What are the privacy, communication, and performance trade-offs of federated LoRA-based fine-tuning?

    The thesis will build upon recent work that shows it is possible to train LLMs in federated environments by exchanging only adapter weights. This project will extend that idea by exploring heterogeneous LoRA configurations—i.e., each client may choose its own rank, structure, or optimization setup—and study how such diversity affects convergence, model utility, and communication efficiency.

    ## Key components of the thesis: ##
    1. Implementation of federated fine-tuning protocols using LoRA and related adapter methods (e.g., QLoRA, AdaLoRA).
    2. Support for heterogeneous client-side LoRA configurations (e.g., different ranks, tasks, or target layers).
    3. Evaluation on benchmark NLP tasks (e.g., sentiment analysis, question answering) using common federated learning setups (e.g., LEAF, FedNLP).
    4. Investigation of aggregation strategies that accommodate non-iid client updates (e.g., FedAvg with normalization, FedProto, FedAdam).
    5. Optional exploration of privacy-preserving extensions (e.g., differential privacy, secure aggregation).

    ## Expected outcomes: ##
    1. A practical system that enables federated, privacy-conscious, and communication-efficient fine-tuning of LLMs.
    2. Empirical insights into the effects of heterogeneity in LoRA configurations across clients.
    3. A set of recommendations for scalable and modular federated LLM training pipelines.

    Anforderungen:
    Solid programming skills in Python.

    Experience with PyTorch and Hugging Face Transformers.

    Familiarity with basic concepts in machine learning, NLP, and distributed systems.

    Verwandte Arbeiten:
    FLoRA: Federated Fine-Tuning Large Language Models with Heterogeneous Low-Rank Adaptations.
    Link: https://arxiv.org/pdf/2409.05976

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Dr. Marco Fisichella wenn Sie an diesem Thema interessiert sind.
  • Development of an Artificial Intelligence system for converting images into tactile representations (Prof. Dr. techn. Wolfgang Nejdl)
    The tactile interpretation of images requires the transformation of two-dimensional visual information into three-dimensional representations accessible through touch. This conversion presents significant challenges, particularly in the preservation of crucial expressive elements that in visual reality are highlighted through light and shadow.

    The project focuses on case studies of pictorial art; for example, works by Caravaggio, where chiaroscuro creates a true 'tactile map' through:
    - Anatomies highlighted by shadows (muscles, ribs, veins)
    - Movement-defining fabric folds
    - Depth of environments
    - Textures of objects (fruit, fabrics, metals)

    The objective of this thesis is to develop an AI system that overcomes current limitations in the automatic conversion of images into tactile representations, focusing on:
    - Recognition of elements that in visual reality are highlighted by light and shadow
    - Translation of these elements into appropriate depth maps
    - Preservation of key narrative elements
    - Optimisation of high relief rendering
    The project is being developed in parallel with tactile image production experts who will test the effectiveness of the developed solutions.

    The project offers the opportunity to:
    - Working on a concrete problem of cultural accessibility
    - Developing an innovative image interpretation system
    - Collaborating with experts in the field of tactile representations
    - Contributing to innovation in art accessibility

    Anforderungen:
    - Image processing
    - Interest in accessibility and art
    - Ability to interpret visual elements from a tactile perspective

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Prof. Dr. techn. Wolfgang Nejdl wenn Sie an diesem Thema interessiert sind.
  • Optimizing Robot Interaction using Image-based Tactile Sensors (Wadhah Zai El Amri)
    Tactile sensing presents a promising opportunity for enhancing the interaction capabilities of today’s robots. DIGIT [1] is a commonly used tactile sensor that enables robots to perceive and respond to physical tactile stimuli.

    In situations where the sensor is unavailable or experiment repetitions are costly, the value of a reliable, real-time simulation becomes evident. Such a simulation can effectively estimate sensor outputs for various touch scenarios. Such simulation would offer a good alternative to gathering data in different setups and environments.

    While several studies have introduced simulations for the DIGIT sensor, such as TACTO [2] and Taxim [3], they predominantly rely on rigid-body simulations and overlook the crucial soft-body aspect of the sensor's gel tip. This oversight diminishes the accuracy of the simulations and their ability to faithfully replicate real-world tactile interactions.


    Goals of the thesis

    Collect real image outputs and simulated sensor surface deformation.
    Train a machine learning algorithm to generate output images using DIGIT surface deformation mesh.
    Perform manipulation/grasping tasks with a real UR5 robot to assess and evaluate the performance of the trained algorithm.

    Anforderungen:
    Prior Knowledge or interest

    Machine Learning
    Robotic Operating System (ROS)
    Python
    Linux

    Verwandte Arbeiten:
    [1]: M. Lambeta, P.-W. Chou, S. Tian, B. Yang, B. Maloon, V. R. Most, D. Stroud, R. Santos, A. Byagowi, G. Kammerer, D. Jayaraman, and R. Calandra, “DIGIT: A Novel Design for a Low-Cost Compact High-Resolution Tactile Sensor With Application to In-Hand Manipulation,” IEEE Robotics and Automation Letters, vol. 5, no. 3, pp. 3838–3845, 2020.
    [2]: S. Wang, M. Lambeta, P.-W. Chou, and R. Calandra, “TACTO: A Fast, Flexible, and Open-Source Simulator for High-Resolution Vision-Based Tactile Sensors,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 3930–3937, 2022.
    [3]: Z. Si and W. Yuan, “Taxim: An example-based simulation model for gelsight tactile sensors,” IEEE Robotics and Automation Letters, 2022.

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Wadhah Zai El Amri wenn Sie an diesem Thema interessiert sind.
  • Graph Neural Networks for Semantic Table Interpretation (Dr. Simon Gottschalk)
    Semantic Table Interpretation (STI) is the task of understanding the concepts in a table. This includes (i) cell-entity linking where objects mentioned in table cells are linked to resources such as Wikipedia, (ii) column-type annotation where the type of objects in a column is identified and (iii) column-property annotation which is about finding the relation between two columns [1]. One approach to perform these three tasks is the use of graph neural networks (GNNs) where the table, its columns, rows and cells are represented as a graph [2]. By training on a large corpus, the GNN performs the three tasks through node and edge classification [3].

    The goal in this thesis is to develop a GNN that performs the three tasks jointly and to demonstrate that this joint training leads to more accurate annotations than when performing these tasks in isolation.

    As a basis of this master thesis, we already have a method for table-to-graph conversion, a basic GNN for STI and an evaluation pipeline setup which will need to be extended to demonstrate the effect of joint training and its superiority over existing approaches.

    Your goals are as follows:
    - Understand STI and GNNs
    - Extend an GNN (implemented in PyTorch Geometric) for STI
    - Conduct experiments to evaluate the effectiveness of the GNN

    Anforderungen:
    Ideally, you have experience in:
    - Python (mandatory)
    -- PyTorch and PyTorch Geometric
    - Machine learning
    -- Graph Neural Networks
    - Linux
    - Knowledge Graphs

    Verwandte Arbeiten:
    [1] Jiménez-Ruiz, E., Hassanzadeh, O., Efthymiou, V., Chen, J., Srinivas, K.: SemTab 2019: Resources to Benchmark Tabular Data to Knowledge Graph Matching Systems. In: The Semantic Web: 17th International Conference, ESWC 2020, Heraklion, Crete, Greece, May 31–June 4, 2020, Proceedings 17. pp. 514–530. Springer (2020)
    [2] Pramanick, A., Bhattacharya, I.: Joint Learning of Representations for Web-tables, Entities and Types using Graph Convolutional Network. In: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume. pp. 1197–1206 (2021)
    [3] Wang, D., Shiralkar, P., Lockard, C., Huang, B., Dong, X.L., Jiang, M.: TCN:Table Convolutional Network for Web Table Interpretation. In: Proceedings of the Web Conference 2021. pp. 4020–4032 (2021)

    Sprache der Projekt-/Abschlussarbeit:
    Englisch

    Kontakt:
    Bitte kontaktieren Sie Dr. Simon Gottschalk wenn Sie an diesem Thema interessiert sind.