PostDoc Position – Physical AI Research

 

As a PostDoc researcher and co-founder of the Physical AI group, you will develop methods that enable robots to learn from
demonstrations, corrections, and autonomous experience, and deploy them in real-world settings.

  • Vision-Language-Action (VLA) Models: Design and implement VLA architectures integrating vision, language, and action for dexterous manipulation, building on large pre-trained vision-language backbones (e.g., 5B-parameter VLMs).
  • Reinforcement Learning from Experience: Develop RL pipelines – offline RL, advantage-conditioned policies – enabling robots to grow beyond pure imitation, achieving human-level and superhuman robustness through autonomous experience.
  • Long-Horizon Task Mastery: Investigate credit assignment across extended tasks via learned value functions, enabling robots to detect and correct compounding errors in complex real-world scenarios.
  • Sim-to-Real Transfer & World Models: Bridge simulation and deployment using world models, self-supervised representations (JEPA, DINOv3), and transfer techniques for robust generalization.
  • Medical & Clinical Robotics: Partner with imaging and clinical groups to apply Physical AI in healthcare robotics, combining LFB’s sensor expertise with embodied intelligence.

Candidate Profile:

  • PhD in Computer Science, Electrical Engineering, Robotics, Physics, or a related field
  • Strong background in deep learning, with experience in reinforcement learning, imitation learning
  • Hands-on experience with PyTorch and large-scale model training; familiarity with VLA or foundation model architectures is a strong advantage
  • Publication record in top-tier venues (NeurIPS, ICML, ICLR, CoRL, ICRA, or equivalent)
  • Drive to work at the intersection of Physical AI, embodied intelligence, and real-world deployment
  • Excellent communication skills in English; German is advantageous but not required

 


PhD Position: AI for Automated Surgical Planning

 

Your Research Impact
As a PhD researcher, you will develop AI methods that automate surgical planning.

  • 3D Anatomical Intelligence
    Develop representation learning approaches based on DINOv2 for high-resolution CT and CBCT/DVT data to
    capture complex anatomical structures.
  • Self-Supervised Learning
    Design learning strategies that leverage large volumes of unlabeled medical imaging data.
  • Multimodal Surgical Reasoning
    Combine visual models with multimodal language models to translate clinical descriptions into surgical
    planning instructions.
  • Automated Planning
    Integrate AI representations with geometric algorithms to propose osteotomy planes and surgical plans.

Candidate Profile:

  • Very good Master’s degree in Computer Science, Physics, Biomedical Engineering, or a related field
  • Strong background in AI/ML and experience with Python/PyTorch
  • Interest in self-supervised learning or vision transformer models
  • Motivation to work on interdisciplinary medical AI challenges

HiWi Offers

 

We continuously hire HiWis and would be happy to hear from you—feel free to come and talk to us anytime!.

 Studentische Hilfskraft (m/w/d) im Bereich Videosignalverarbeitung
Ansprechpartner: Mathias Wien