Biomechanical- and Image-Guided Surgical Systems (BIGSS) Laboratory – Dr. Mehran Armand
The Biomechanical- and Image-Guided Surgical Systems (BIGSS) laboratory is a collaboration between researchers at the Johns Hopkins University and the Johns Hopkins University Applied Physics Laboratory. This laboratory focuses on developing innovative computer-aided surgical guidance systems involving novel robots, advanced imaging, and real-time biomechanical assessments to improve surgical outcomes.
Photoacoustic & Ultrasonic Systems Engineering (PULSE) Lab – Dr. Muyinatu Bell
The PULSE Lab, directed by Dr. Muyinatu A. Lediju Bell, integrates light, sound, and robots to develop innovative biomedical imaging systems that simultaneously address unmet clinical needs and improve patient care. Our emphasis is diagnostic and surgical ultrasound and photoacoustic technologies, with applications in neurosurgery, cancer detection and treatment, and women’s health. We maintain a constant eye toward interfacing our technologies with real patients to facilitate clinical translation. The PULSE Lab is affiliated with the Laboratory for Computational Sensing and Robotics, the Malone Center for Engineering in Healthcare, and the Carnegie Center for Surgical Innovation, with dedicated laboratory space at both the Johns Hopkins University Homewood Campus and the Johns Hopkins Hospital School of Medicine.
Medical UltraSound Imaging and Intervention Collaboration (MUSiiC) – Dr. Emad Boctor
The MUSiiC research lab, headed by Dr. Emad Boctor, develops innovative ultrasound technologies for medical applications ranging from prostate and breast cancer treatment to liver ablation and brachytherapy, among others. The group is based on a collaboration among researchers from Johns Hopkins Medical School, Johns Hopkins Whiting School of Engineering, and partners from other academic institutions and industry.
Robot and Protein Kinematics Lab (RPK) – Dr. Gregory Chirikjian
Dr. Gregory Chirikjian directs the Robot and Protein Kinematics Lab in LCSR. This lab is involved in research in computational structural biology (in particular, computational mechanics of large proteins), conformational statistics of biological macromolecules, developed theory for ‘hyper- redundant’ (snakelike) robot motion planning, hyper-redundant robotic manipulator arms, modular self-reconfigurable robots, applied mathematics (applications of group theory in engineering), self-replicating robotic systems.
Locomotion In Mechanical and Biological Systems (LIMBS) – Dr. Noah Cowan
The LIMBS laboratory, directed by Noah J. Cowan, strives to uncover principles of animal and robot sensory guidance. For animals this is an analysis problem: we reverse engineer the biomechanical and neural control principles underlying animal movement. For robotics, this is a design problem: we incorporate biological inspiration and engineering insights to synthesize new approaches to robot control. This research program includes several projects. Learn more by visiting the following website.
Computational Interaction and Robotics Lab (CIRL) – Dr. Gregory Hager
The Computational Interaction and Robotics Laboratory, directed by Dr. Gregory Hager, is devoted to the study of problems that involve dynamic, spatial interaction at the intersection of imaging, robotics, and human-computer interaction. The laboratory has a number of ongoing projects in this area. The Language of Motion project is seeking to develop new methods to recognize and evaluate skilled human manipulation, with a particular emphasis on surgery. Data is collected using a da Vinci Surgical robot, and processed into gesture-based models that support skill evaluation, training, and human-robot collaborative task execution. The Manipulating and Perceiving Simultaneously (MAPS) project seeks to apply principles of computer vision to tactile sensing, with the goal of developing new methods for haptic object recognition. The lab’s most recent work aims to develop Generic Perception to support general-purposes manipulation of objects in the physical world. The laboratory also works in the area of medical imaging. Interactive computer-aided diagnostic systems based on images are also an area of interest.
Advanced Medical Instrumentation, and Robotics (AMIRo) – Dr. Iulian Iordachita
The Advanced Medical Instrumentation and Robotics Research Laboratory (AMIRo), directed by Dr. Iulian Iordachita, conducts research to aid and support the robotic assisted medical technology encompassing medical diagnosis and therapy, and clinical research. The main goal is to create the future medical robots and devices that will help clinicians to deliver earlier diagnosis and less invasive treatments at lower cost and in shorter time.
Sensing, Manipulation, and Real-Time Systems Laboratory (SMARTS Lab) – Dr. Peter Kazanzides
Dr. Peter Kazanzides heads the SMARTS lab, which works on components and integrated systems for computer-assisted surgery. This includes the integration of real-time imaging, such as video and ultrasound, to enable robotic assistance in more challenging environments, such as minimally invasive surgery and microsurgery. Research in component technologies includes high-performance motor control, electromagnetic and inertial sensing, and sensor fusion. The lab also performs research in system architectures, applying component-based software engineering concepts to provide a uniform programming model for multi-threaded, multi-process, and multi-processor systems.
Autonomous Systems, Control, and Optimization Laboratory (ASCO) – Dr. Marin Kobilarov
The Autonomous Systems, Control and Optimization Laboratory (ASCO), directed by Dr. Marin Kobilarov, aims to develop intelligent robotic vehicles that can perceive, navigate, and accomplish challenging tasks in uncertain, dynamic, and highly constrained environments. The lab performs research in analytical and computational methods for mechanics, control, motion planning, and reasoning under uncertainty, and in the design and integration of novel mechanisms and embedded systems. Application areas include mobile robots, aerial vehicles, and nano satellites.
Terradynamics Lab – Dr. Chen Li
Aero- and hydrodynamics have helped us understand how animals fly and swim and develop aerial and aquatic vehicles that work well. By contrast, we know little about how animals move so well through almost any terrain, and even the best robots struggle in terrain like building rubble or loose Martian soil. Analogous to aero- and hydrodynamics, we are creating terradynamics, new physics models of locomotor-terrain interactions, to understand animal locomotion and improve robotic mobility in complex terrain common in the real world.
Computer Aided Medical Procedures (CAMP) – Dr. Nassir Navab
The CAMP laboratory aims at developing the next generation solutions for computer assisted interventions. The complexity of surgical environments requires us to study, model and monitor surgical workflow enabling the development of novel patient and process specific imaging and visualization methods. Due to the requirements of flexibility and reliability we work on novel robotized multi-modal imaging solutions and to satisfy the challenging usability requirements we focus on data fusion and its interactive representation within augmented reality environments. The lab creates a bridge across the Atlantic ocean by hosting researchers working at both of Prof. Navab’s groups at JHU in Baltimore and TU in Munich.
Computer Integrated Interventional Systems (CIIS) Laboratory – Dr. Russell Taylor
Professor Russell Taylor directs the Computer Integrated Interventional Systems (CIIS) laboratory. This lab exists to develop surgical systems that integrate novel computer and human/machine interface technologies that will revolutionize surgical procedures, extending the surgeon’s abilities to achieve better outcomes at lower costs. Some of the recent research projects include robot assisted microsurgery (steady hand eye robot), surgical control and planning, snake robot, deformable human anatomical models, smart surgical instruments, treatment plan optimization in radiation oncology, image overlay, laparoscopic-assisted robot system, robot assisted ultrasound and MRI compatible robotics.
Dynamical Systems and Control Laboratory (DSCL) – Dr. Louis Whitcomb
Prof. Louis Whitcomb directs the DSCL lab and research focusing on problems in the navigation, dynamics, and control of linear and nonlinear dynamical systems, observers, nonlinear systems analysis, modeling, and sensing relevant to robots that interact dynamically in extreme environments. The principal focus is on problems motivated by two application areas that share a common underlying mathematical framework – underwater robot vehicles and robot manipulators.
Computational Sensory-Motor Systems Lab (CSMS) – Dr. Ralph Etienne-Cummings
Dr. Ralph Etienne-Cummings directs the CSMS lab. The lab’s current research includes various experiments to understand neurophysiology of spinal neural circuits, to interface with them, to decode their sensory-motor relationships, and to use these relationships to control biomorphic robots. The lab is developing brain-like computational systems to mimic the object detection, recognition, and tracking found in humans and primates. The plan is to continue to expand this area of research, while leveraging the laboratory’s expertise in VLSI circuits and systems, visual and acoustical information processing, neuromorphic computation systems and biomorphic robotics.
Networked and Spatially Distributed Systems (NSDS) – Dr. Dennice Gayme
The Networked and Spatially Distributed Systems (NSDS) group directed by Dr. Dennice Gayme, studies large-scale networked and interconnected systems using computational and theoretical methods from applied mathematics, dynamics, controls, optimization and fluid mechanics. The group focuses on two main application areas (1) the development of control oriented models for turbulent shear flows and (2) creating algorithms and tools to help facilitate the integration of renewable power sources into a “smarter” electric power system. Recent work on shear flow turbulence aims to leverage knowledge of wind farm flow physics to design wind farm and grid level algorithms that can facilitate larger scale grid-integration of wind energy.
Photonics and Optoelectronics Laboratory – Dr. Jin U. Kang
The Photonics and Optoelectronics Laboratory, directed by Jin U. Kang, conducts experimental and theoretical investigations in the area of photonics and optoelectronics with an emphasis on developing novel fiber optic imaging and sensor systems for medical applications. Specifically the lab develops high-speed real-time optical coherence tomography system that can guide surgical procedures and allow doctors to make accurate prognosis of the surgical outcome. In addition we develop an array of “smart surgical tools” that uses fiber optic OCT distal sensors to ensure safe and precise surgical maneuvers. Also a large effort is devoted to developing an array of sub-millimber endoscopic imaging systems that allows imaging of brain activities in awake freely moving mice.
Image Analysis and Communications Laboratory(IACL) – Jerry Prince
The Image Analysis and Communications Lab (IACL) in the Department of Electrical and Computer Engineering at Johns Hopkins University is led by Professor Jerry Prince. Research focuses on image and signal processing in medical imaging and video processing. Specific areas of technical interest include filter banks, wavelets, multivariate systems, signal decomposition, time-frequency and time-scale analysis, active contours and deformable geometry, computed tomography, magnetic resonance imaging, and optical flow.
Urology Robotics (URobotics) – Dr. Dan Stoianovici
Urology Robotics is a research and education program dedicated to advance the technology used in Urology. The main focus of the lab is in the development of robots for real-time Image-Guided Interventions. The application range of the lab technologies extends to other medical specialties and industry. The program is based on a multidisciplinary integrated team of students, engineers, and clinicians working in partnership from the bench to the bedside. The lab is specialized in the development of surgical robotic systems and especially in robotics for image-guided intervention (IGI). Besides Urology, the instruments and systems created in the lab apply to a larger area of medical fields, such as Interventional Radiology. The lab is part of the Brady Urological Institute (Urology Department at the Johns Hopkins Medicine) and is located at the Johns Hopkins Bayview Medical Center.
Vision, Dynamics and Learning Lab (VDL) – Dr. Rene Vidal
Our research spans a wide range of areas in biomedical imaging, computer vision, dynamics and controls, machine learning and robotics. In particular, we are interested in inference problems involving geometry, dynamics, photometry and statistics, such as (1) inferring models from images (image/video segmentation and structure from motion), static data (generalized PCA) or dynamic data (identification of hybrid systems), and (2) using such models to accomplish a complex mission (land a helicopter, pursue a team of evaders, follow a formation).