Ai co-pilot bronchoscope robot
Ai co-pilot bronchoscope robot"
- Select a language for the TTS:
- UK English Female
- UK English Male
- US English Female
- US English Male
- Australian Female
- Australian Male
- Language selected: (auto detect) - EN
Play all audios:
ABSTRACT The unequal distribution of medical resources and scarcity of experienced practitioners confine access to bronchoscopy primarily to well-equipped hospitals in developed regions,
contributing to the unavailability of bronchoscopic services in underdeveloped areas. Here, we present an artificial intelligence (AI) co-pilot bronchoscope robot that empowers novice
doctors to conduct lung examinations as safely and adeptly as experienced colleagues. The system features a user-friendly, plug-and-play catheter, devised for robot-assisted steering,
facilitating access to bronchi beyond the fifth generation in average adult patients. Drawing upon historical bronchoscopic videos and expert imitation, our AI–human shared control algorithm
enables novice doctors to achieve safe steering in the lung, mitigating misoperations. Both in vitro and in vivo results underscore that our system equips novice doctors with the skills to
perform lung examinations as expertly as seasoned practitioners. This study offers innovative strategies to address the pressing issue of medical resource disparities through AI assistance.
SIMILAR CONTENT BEING VIEWED BY OTHERS REAL-TIME MULTIFACETED ARTIFICIAL INTELLIGENCE VS IN-PERSON INSTRUCTION IN TEACHING SURGICAL TECHNICAL SKILLS: A RANDOMIZED CONTROLLED TRIAL Article
Open access 02 July 2024 HETEROGENEITY AND PREDICTORS OF THE EFFECTS OF AI ASSISTANCE ON RADIOLOGISTS Article Open access 19 March 2024 CIVIL LIABILITY FOR THE ACTIONS OF AUTONOMOUS AI IN
HEALTHCARE: AN INVITATION TO FURTHER CONTEMPLATION Article Open access 23 February 2024 INTRODUCTION Lung diseases, such as lung cancer, chronic obstructive pulmonary disease and pneumonia,
represent a significant global health burden, with millions of individuals affected yearly1,2. Early detection and intervention are crucial to mitigate the impact of these diseases, reduce
morbidity and improve patient outcomes3. Bronchoscopy, a minimally invasive diagnostic and therapeutic procedure, has emerged as an essential tool in detecting, treating and managing various
lung diseases4,5,6. During bronchoscopy, the bronchoscope should be inserted and manipulated gently to avoid abrupt or forceful movements, which can cause discomfort or injury to airway
structures. Additionally, maintaining a central position during bronchoscopy allows better visualisation of the airway anatomy and helps prevent injury to the airway mucosa or other
structures. This is especially important when navigating through tight or tortuous airways, strictures, or areas with masses. However, this procedure requires a high level of skill and
experience, resulting in a significant disparity in the quality of care provided by expert and novice doctors7,8. The availability of bronchoscopic services is predominantly limited by the
need for more experienced doctors in underdeveloped regions, resulting in a critical barrier to health care access for vulnerable populations9. The need for specialised training and
expertise in bronchoscopy further exacerbates this issue, as many health care systems in these areas need help to support the development and maintenance of such skills. Consequently, this
knowledge gap hinders the establishment and expansion of bronchoscopic services in underprivileged settings, perpetuating the cycle of inadequate health care provision for lung disease
patients10,11. Innovations combining the precision and dexterity of robotic systems with the guidance of expert doctors could help to resolve these clinical and technical challenges12,13,14.
The development of robotic platforms and devices for bronchoscopy has seen significant progress in recent years, with systems such as the Monarch Platform15 and the Ion Endoluminal System16
leading the way. The Monarch Platform is equipped with an internal bronchoscope catheter with a 4.2 mm diameter and an external sheath of 6 mm. Its subtle steering control and flexibility,
allowing deeper access into the peripheral regions of the lungs, surpass those of conventional bronchoscopes17 (9th vs. 6th airway generations). The Ion Endoluminal System boasts a fully
articulated 3.5 mm catheter with a 2 mm working channel, enhanced stability, superior flexibility and the added advantage of shape perception18. Notably, studies indicate that these
platforms exhibit a favourable diagnostic yield, ranging from 81.7% to 92%, for lung nodules with sizes between 14.8 mm and 21.9 mm. Moreover, the complication rates reported are
minimal19,20,21. These findings suggest that these platforms can play a transformative role in the future management of pulmonary conditions. In addition to the Monarch and Ion platforms,
several other bronchoscope robotic systems are under development by academic institutions or have entered early-stage research to address sensing and control issues for
doctors22,23,24,25,26,27. Nevertheless, despite its advantages, current telerobotic bronchoscopy faces several challenges, including a steep learning curve and lack of autonomy. At present
the integration of artificial intelligence (AI) techniques into bronchoscopy is further expanding the horizons of this burgeoning field28. By leveraging advanced algorithms, such as machine
learning and computer vision technologies29, researchers are developing image-guided navigation systems that process and interpret bronchoscopic imagery30, facilitating real-time
localisation31, tracking32 and interventional path planning33 for endoscopy and enabling precise navigation within the bronchial tree. These software systems enhance the accuracy and
efficiency of bronchoscopic procedures. Furthermore, by providing automated, continuous guidance throughout such a procedure34, an image-guided system can help reduce the cognitive load on
the operating doctor, allowing the doctor to focus on other critical aspects of the procedure35. However, these systems present safety concerns during bronchoscopic procedures because they
rely on bronchoscope localisation in preoperative CT36,37,38, which may result in misregistration and unsafe steering of the robot due to the limited field of view and body–CT visual
discrepancies. Concerns have been raised about the risk of complications, such as pneumothorax and bleeding, underlining the need for ongoing research on and optimisation of these platforms.
We report an AI co-pilot bronchoscope robot for safe pulmonary steering. At the hardware level, we have designed a bronchoscope robot with a quick catheter replacement feature (utilising
thin catheters for deep lung examination and thick catheters for examination and biopsy) based on magnetic adsorption, offering advanced performance that meets clinical requirements. At the
software level, we have developed an AI–human shared control algorithm based on learning from experts, which can accommodate discrete human interventions while reducing the reliance on
doctor expertise. Overall, the presented robotic system enhances safety while maintaining efficiency in bronchoscopy by providing novice doctors with increased autonomy. Through tests of
this bronchoscope system on realistic human bronchial phantoms with simulated respiratory behaviour performed by a novice doctor with the AI co-pilot and by an expert, we demonstrate that
our system enables novice doctors to access different bronchi proficiently. Subsequently, to validate the safety and effectiveness of our system under physiologically relevant conditions, we
assess the system’s steering capability in vivo using a live porcine lung to mimic the human bronchus. Our system carries the potential to improve the diagnosis and management of pulmonary
disorders. It is anticipated that the cost and logistical barriers associated with the adoption of such platforms will decrease in the future, helping to overcome the challenge of medical
resource disparities and contributing to the improvement of global health outcomes. RESULTS DESIGN OF THE AI CO-PILOT BRONCHOSCOPE ROBOT Figure 1 provides an overview of our AI co-pilot
bronchoscope robot (Supplementary Notes 1–3 and Supplementary Figs. 1–4), which is designed to be deployed in clinical settings for bronchoscopic procedures (Supplementary Movie 1). The
bronchoscope robot is integrated with a robotic arm and placed alongside the operating table, and it is teleoperated by a doctor using a remote-control console to steer the bronchoscopic
catheter. The robotic arm is used to adjust the intubation posture of the catheter for lung steering. The steering control system is composed of four linear motors for tendon actuation and
four force sensors to measure the actuation force. To facilitate user-friendly replacement of the catheter, the steering control system and catheter are designed to connect by magnetic
adsorption. The bronchoscopic catheter is composed of a proximal section with high stiffness and a distal section with low stiffness. A braided mesh is used in the proximal section for
increased stiffness, while the distal section is composed of many small hinge joints for steering control. Both sections are covered with a thin thermoplastic urethane (TPU) layer for
waterproofing. The tip of the bronchoscopic catheter consists of two LED lights for lighting and a microcamera for observation. Two types of catheters are designed, with diameters of 3.3 mm
(with a 1.2 mm working channel) and 2.1 mm (without a working channel), both of which are smaller than those of the Monarch Platform and the Ion Endoluminal System, enabling access to
bronchi of the ninth generation or deeper in average adult patients. To improve the quality and consistency of bronchoscopic procedures and enable novice doctors to perform bronchoscopy as
safely and proficiently as experienced specialists, we have designed an AI–human shared control algorithm to minimise the risk of damaging surrounding tissues while maintaining efficiency.
Figure 2a illustrates an overview of the algorithm workflow. The algorithm’s core is a policy network that takes a bronchoscopic image and a discrete human command (up, down, left, right, or
forward) as input to predict a steering action (pitch and yaw angle rates) for the robot’s orientation, which can be converted into tendon actuation through inverse kinematics and a
low-level controller. As shown in Fig. 2b, the policy network training process consists of three steps: (a) establishment of a virtual bronchoscopy environment, (b) data preparation, and (c)
Sim2Real adaptation. In the first step, an airway model is segmented from the preoperative CT volume to establish a virtual bronchoscopy environment. The airway centrelines are extracted by
means of the Vascular Modeling Toolkit (VMTK) to serve as reference paths. By simulating the bronchoscope robot in this virtual environment, we can render its observed image and depth.
Supplementary Note 4 presents the virtual environment and simulated robot configurations. In the second step, human commands and actions for each image are automatically generated for model
supervision by an artificial expert agent (AEA) guided by privileged robot pose information and the reference airway centrelines, resulting in training samples consisting of images, depths,
human commands and steering actions. For the third step, we propose a Sim2Real adaptation module to enhance the diversity and photorealism of the training samples. The domain adaptation part
of this module translates rendered images into a more realistic style while preserving the bronchial structure by means of depth supervision, ensuring that the corresponding action
supervision remains invariant. The domain randomisation part randomly alters the image appearances or adds noise to the human commands. Based on the dataset prepared as described above, a
data aggregation algorithm (DAgger)39 is employed for on-policy artificial expert imitation to eliminate distribution mismatch. Because every training sample is generated automatically, the
entire training process is intervention-free. In practice, the input rendered images and AEA-annotated commands are replaced with real images and a novice doctor’s commands in the real
bronchoscope robot, driving the policy network to navigate the robot through the airway safely and smoothly. The discrete human commands are mapped to five regions of the teleoperator
(Supplementary Fig. 3b), reducing the level of human intervention and the cognitive load compared to conventional teleoperated robots with continuous human guidance. SIMULATION RESULTS AND
IN VITRO EVALUATION We quantitatively assessed the performance of the proposed method through simulation experiments, in which airway models containing up to 5th-generation bronchi were
utilised for training and evaluation (Fig. 3a). Two kinds of bronchoscopic images were rendered from airway models with pink and realistic textures, referred to as Sim-style and Real-style
images, respectively (Fig. 3b). The realistic textures were generated by extracting actual bronchial wall textures from real historical clinical bronchoscopic videos. The training
environments established based on the two airway models contained 74 and 84 reference paths, respectively. The test environment was built on the basis of another airway model with a
realistic texture that contained 60 paths (Supplementary Fig. 5). The policy network trained using Sim-style images with domain adaptation and randomisation (Sim+A + R) achieved the highest
success rate (calculated as the ratio of the numbers of successful paths to all paths, detailed definitions of which can be found in Supplementary Note 9 and Supplementary Fig. 17) of ~93.3%
on the test paths (Fig. 3c). This performance surpassed that of networks trained without domain randomisation (Sim+A; ~80.0%), without any domain adaptation or randomisation (Sim; ~31.8%),
using only Real-style images (Real; ~81.8%), and employing the baseline domain adaptation approach40 (Sim+A(b); ~71.8%). Similarly, our Sim+A + R method exhibited the highest successful path
ratio (calculated as the ratio of the completed path length over the total path length) of ~98.9 ± 4.7%, surpassing those of Real (~96.5 ± 8.0%), Sim+A (~96.4 ± 7.4%), Sim+A(b) (~92.9 ±
12.4%) and Sim (~75.2 ± 22.6%) (Fig. 3d). In terms of the trajectory error—the Euclidean distance between the predicted and reference paths—Sim+A + R also demonstrated the lowest error of
~1.04 ± 0.21 mm, compared to ~1.23 ± 0.28 mm for Real, ~1.37 ± 0.26 mm for Sim+A, ~2.57 ± 0.54 mm for Sim+A(b), and ~3.36 ± 0.66 mm for Sim (Fig. 3e). Specific results on each path are shown
in Supplementary Fig. 6b. Notably, the Sim method underperformed, indicating overfitting of the policy network in the Sim-style image domain. Sim+A and Real showed similar performances,
emphasising the effectiveness of our domain adaptation module. Sim+A outperformed Sim+A(b), validating the advantage of preserving the bronchial structure during domain adaptation. Sim+A + R
achieved the best results, demonstrating that our proposed domain adaptation and randomisation procedures enable successful knowledge transfer between simulation and reality. The
generalisation ability of Sim+A + R has also been tested on airway models from 10 different patients (Supplementary Fig. 7). The domain adaptation module is necessary for translating
Sim-style images (the source domain) into realistic images (the target domain). Thus, we qualitatively and quantitatively evaluated the image translation performance of our method. We
selected three types of unpaired realistic images to represent target domains for training and evaluation, namely Real-, Phantom- and Clinical-style images (Supplementary Note 5 and
Supplementary Fig. 8a). These images were collected from virtual airway models rendered with realistic texture mapping, real bronchoscopy in silicone phantoms, and real bronchoscopy in live
pigs, respectively. As shown in Fig. 3f and Supplementary Fig. 8b, our method successfully performs image translation without changing the bronchial structure. Our method outperforms
AttentionGAN32 in terms of the structural similarity index measure (SSIM) across Real-style (0.91 vs. 0.80), Phantom-style (0.95 vs. 0.70) and Clinical-style (0.96 vs. 0.78) translations,
demonstrating the superior structure-preserving property of our method. Additionally, our approach achieves a higher peak signal-to-noise ratio (PSNR) than AttentionGAN (25.16 dB vs. 12.25
dB) between translated and target images in the Real style (Fig. 3g). We conclude that our method can successfully generate images with the target domain style while preserving the bronchial
structure, facilitating Sim2Real adaptation. To assess the proposed AI co-pilot bronchoscope robot, experiments were conducted on a bronchial phantom made of silica gel replicating
structured derived from human CT lung data (Fig. 4a). A crank-rocker mechanism-based breathing simulation system (Supplementary Note 7 and Supplementary Fig. 10) was designed to emulate
human respiration (15 cycles per minute). Two bronchial phantoms with distinct bronchial structures were employed for in vitro evaluation (Fig. 4b). An expert (chief doctor) and a novice
doctor (medical intern) were invited to perform bronchoscopic procedures using the robot without the AI co-pilot as a benchmark, while another novice doctor (attending doctor) also
participated using the robot with the AI co-pilot. All procedures were performed on the same path using the teleoperator. The medical intern had no experience with bronchoscopy, while the
attending doctor had a little experience (<5 years and <100 cases per year, compared to the chief doctor, who had >20 years of experience and >200 cases per year), as indicated
in Supplementary Table 1. They were both presented with two demonstrations of robotic intubation, with and without the AI co-pilot, to learn how to operate the system. During the evaluation,
the medical intern performed teleoperated bronchoscopy without the AI co-pilot along the bronchus path for three trials, with image errors decreasing from 62.39 ± 1.91 pixels to 43.54 ±
2.01 pixels (Fig. 4c, Supplementary Note 8 and Supplementary Figs. 11–13). Under the same conditions, the expert achieved more precise and stable bronchoscopic operation (with an image error
of 31.45 ± 1.19 pixels). Assisted by our AI–human shared control algorithm, the attending doctor achieved even better bronchoscopic operation performance (17.63 ± 0.46 pixels) than the
expert in the first trial. Next, the expert and the attending doctor carried out experiments on both sides of two phantoms, with the attending doctor using the robot with the AI co-pilot
(Supplementary Figs. 14–16 and Supplementary Movie 2 and 3). Detailed operation error comparisons are depicted in Fig. 4d, e. In Phantom 1, AI-assisted operation (19.14 ± 0.50 pixels)
exhibited a significantly lower operation error than expert operation (38.84 ± 0.84 pixels). In Phantom 2, despite a small initial error, as shown in the Path 66 results, expert operation
failed to maintain a low error during insertion into deeper bronchial airways, while AI-assisted operation consistently maintained a low image error, keeping the bronchoscope centred in the
image. In addition, a specific analysis of AI control performance and mode switching between AI and teleoperation is described in Supplementary Notes 11 and 12 and Supplementary Figs. 20–23.
IN VIVO DEMONSTRATION WITH A LIVE PORCINE LUNG MODEL We further evaluated the performance of the AI co-pilot bronchoscope robot in a minipig, whose bronchial structure closely resembles
that of the human bronchus. The pig was purchased from Zhuhai BestTest Bio-Tech Co., Ltd., solely based on the pig’s health condition (Supplementary Note 13). The Wuzhishan pig was female
and three months old. The protocols for animal experiments were approved by the Institutional Animal Care and Ethics Committee of Zhuhai BestTest (IAC(S)2201003-1). During the experiment,
the bronchoscope catheter was inserted through the oropharynx into the pig’s bronchial airways, while the doctor, seated at the console, controlled the robot to accomplish teleoperation and
steering. Prior to clinical trials, a physical examination and a whole-body CT scan were conducted to reconstruct the bronchial structure (Fig. 5a) and ensure the pig’s good health. We
selected two porcine bronchus paths for in vivo demonstration and conducted clinical trials with the expert and the attending doctor, the latter of whom was assisted by the AI co-pilot
(Supplementary Fig. 24). As seen in the endoscopic images presented in Fig. 5b and Supplementary Movie 4, both doctors achieved steering through the porcine bronchus to beyond the
5th-generation bronchi (diameter ~2.5 mm), yielding almost identical visualisation results. As seen from the actuation displacement and actuation force (Fig. 5c) measurements during
bronchoscopy, the attending doctor achieved smoother steering with the AI co-pilot than the expert did. To quantitatively analyse the control effects of the two volunteers, the mean values
and fluctuation ranges of the actuation displacement and actuation force are illustrated in Fig. 5d. It is evident that AI-assisted steering resulted in a lower mean value and fluctuation
range than the expert’s operation overall. The operation error comparison further indicates that our proposed AI co-pilot bronchoscope robot maintains better bronchus centring than expert
teleoperation. Based on an analysis of eight repeated trials on the live porcine lung, the attending doctor could perform bronchoscopy with a 11.38 ± 0.16 pixel operation error in
collaboration with the AI co-pilot, achieving performance as good as or even better than that of the expert (16.26 ± 0.27 pixels). Considering the pixel-to-millimetre calibration results
reported in Supplementary Note 9 and Supplementary Fig. 18, the AI co-pilot group could achieve a mean 3D positioning error of less than 0.73 mm in all procedures. To further characterise
the autonomy properties of the proposed system, we compared the number of doctor interventions between the attending doctor with the AI co-pilot and the expert. The statistical results in
Fig. 5g demonstrate that the number of interventions, as defined in Supplementary Note 9, with the use of the AI co-pilot was significantly lower than that during expert teleoperation,
greatly reducing the doctor’s physical exertion and cognitive load during the bronchoscopic operation and further illustrating the autonomy of our proposed robotic bronchoscopy system.
Furthermore, the NASA Task Load Index survey (Supplementary Note 10 and Supplementary Fig. 19) was also completed for a comprehensive assessment of human workload, demonstrating a
significant reduction in both physical and mental burden when using our AI-assisted system. DISCUSSION Bronchoscopic intervention is preferred for sampling suspected pulmonary lesions owing
to its lower complications. Recently, robot-assisted technologies, such as the Monarch Platform and the Ion Endoluminal System, have been introduced into bronchoscopic procedures to enhance
manoeuvrability and stability during lesion sampling. However, due to the high cost of robotic bronchoscope systems and the expertise needed by doctors, the proliferation of this technology
in underdeveloped regions is limited. Our study presents a low-cost comprehensive AI co-pilot bronchoscope robot (Supplementary Table 2) to improve the safety, accuracy, and efficiency of
bronchoscopic procedures. The proposed system, which incorporates a shared control algorithm and state-of-the-art domain adaptation and randomisation approaches, bridges the gap between
simulated and real environments, ensuring generalisability across various clinical settings. Moreover, this AI co-pilot bronchoscope robot enables novice doctors to perform bronchoscopy as
competently and safely as experienced specialists, reducing the learning curve for bronchoscopic procedures and ensuring a consistent quality of care. Our in vitro and in vivo evaluation
results demonstrate the efficacy of the proposed AI co-pilot bronchoscope robot in achieving insertion into deep bronchial airways with high precision and reduced intervention from the
operator. Notably, the AI–human shared control algorithm maintains better bronchus centring and exhibits lower operation errors than an expert operator, validating the robustness and
clinical potential of our approach. Additionally, the domain adaptation and randomisation techniques effectively mitigate overfitting and facilitate seamless knowledge transfer between the
simulated and real image domains, ultimately contributing to the overall success of the system. Despite these promising results, there are several areas for future research and development.
Our bronchoscope robot relies upon tendon actuation for precise steering control and is fed into the deep lungs by means of “follow-the-tip” motion41,42. In alignment with this methodology,
the proximal section of the catheter is engineered to exhibit a substantially increased stiffness in comparison to the distal section. However, a large bending angle (approaching 180°) of
the distal section presents great challenges in effecting smooth feed movement of the catheter, particularly when negotiating the upper pulmonary regions. A soft untethered magnetic catheter
design has the potential to improve the capabilities of bronchoscopy for deep lung examination and is worthy of study. In addition, it is essential to assess the robustness of the proposed
method in a broader range of clinical scenarios, including patients with varying bronchial anatomies, pathologies, or respiratory conditions. Extensive testing on a diverse patient
population will be necessary to validate the applicability of the intelligent bronchoscope robot in real-world settings. Considering the difference in teleoperators between our AI co-pilot
system and existing robotic or hand-held bronchoscopy systems, the relevance between the previous experience of doctors in current teleoperators and the proficiency of operating our system
is worth further studying. The integration of additional sensing modalities, such as ultrasound or optical coherence tomography, can also be considered to provide complementary information
to guide the bronchoscope robot. Fusing multiple data sources could improve the accuracy and safety of AI-assisted steering, offering more comprehensive diagnostic and therapeutic support.
In addition, the explainability of our AI co-pilot system was investigated by analysing the reasons for the decision-making of the AI during bronchoscopic procedures. We conducted an
experiment on the interpretability of the proposed policy network using three styles of image pairs. We generated gradient-weighted class activation maps (GradCAM) from the last
convolutional layer of the policy network to represent the network’s attention and visualised the fused images by overlaying the GradCAM results onto the original images. In the resulting
images, highlighted regions indicated the key clues that our policy network paid attention to when making decisions. As shown in Supplementary Fig. 25, our network has learned to focus on
the bronchial lumens, and as the distance between the robot and the bronchial wall increases, the attention value becomes larger. This indicates that our network concentrates on the
structural information of the airway and utilises it to predict safe actions, keeping the centre of the bronchial lumen at the centre of the image. As a result, our AI co-pilot robot is able
to remain centred in the airway and stay as far away as possible from the bronchial wall during bronchoscopic procedures. In conclusion, our AI co-pilot bronchoscope robot offers a
promising avenue for enhancing the quality and consistency of bronchoscopic procedures. The system’s robust performance in both simulations and in vivo experiments demonstrates its potential
to revolutionise bronchoscopy and empower novice doctors to perform these procedures confidently. Looking forward, we anticipate that our approach can be adapted and applied to other
medical procedures requiring precise navigation and manipulation, ultimately improving patient outcomes and reducing health care disparities. METHODS OVERVIEW OF THE AI CO-PILOT SYSTEM Our
AI co-pilot bronchoscope robot is divided into two main parts: a hardware system and an AI co-pilot algorithm. At the hardware level, the bronchoscope robot employs tendon-driven mechanics,
leveraging four linear motors to precisely steer the bronchoscope catheter and an electric slide for feed movement. Additionally, our robotic system boasts an innovative magnetic adsorption
method for rapid replacement of the catheter. At the software level, an AI–human shared control algorithm is designed to steer the robot safely. The core of the algorithm is a policy
network, which takes both bronchoscopic images and human commands as inputs to predict steering actions that will control the tip of the bronchoscope robot to remain at the centre of the
airway, helping prevent injury to the airway mucosa. To train the policy network, a virtual environment is created to simulate bronchoscopic procedures and collect training data, and then
domain adaptation and randomisation techniques are used to enhance the training samples. The training process involves a novel artificial expert agent for automatic data annotation and does
not require human intervention. The generator for domain adaptation is pretrained by using virtual bronchoscopic images and unpaired historical bronchoscopy videos, which are easy to access
at hospitals, enabling an annotation-free training stage. With the aid of our AI co-pilot bronchoscope robot, the level of human intervention and the cognitive load imposed on doctors can be
significantly reduced compared to traditional teleoperated robots. AI–HUMAN SHARED CONTROL WORKFLOW The working pipeline of our AI–human shared control algorithm in practical use is
described as follows. The bronchoscope robot acquires bronchoscopic images during the procedure, and a doctor gives discrete human commands (e.g., left, down, right, up or forward) to
determine the high-level direction of the robot. The bronchoscopic images and human commands are input into a trained policy network to predict continuous steering actions (i.e., rotation
angle rates Δ_θ_ and Δ_φ_) to control the robot’s head such that it remains centred at the bronchial lumen for safety. The predicted steering actions are converted into continuous tendon
displacements of the linear motors through inverse kinematics and a low-level controller, forming a closed-loop control system. POLICY NETWORK ARCHITECTURE The policy network is designed
with a multi-task structure, where the main task is steering action prediction and the side task is depth estimation. The learning of the depth estimation task alongside the main task can
encourage the network to recognise the 3D bronchial structure and learn a more generalised scene representation for decision-making. The policy network takes a bronchoscopic image (I) and a
human command (_c_) as inputs, and its outputs include the predicted steering action and estimated depth. Its architecture features an image feature extractor ΦE, a depth decoder ΦD and five
branched action heads \({\{{{{\Phi }}}_{{{{{{\rm{A}}}}}}}^{{{{{{\rm{i}}}}}}}\}}_{i=1}^{5}\), each of which is responsible for predicting steering actions in response to one of five human
commands (left, right, up, down and forward). ΦE is based on ResNet-34, and ΦD is built on a transposed convolutional network, which has skip connections with ΦE. The action heads are based
on multilayer perceptrons (MLPs) and are optionally activated by the human command _c_ through a five-way switch. The depth decoder and action heads share the same representation extracted
by the feature extractor. For alignment with the input channels of the MLPs, the features extracted from ΦE are flattened to a 512-d vector before being input into the chosen action head.
The specific architecture of the policy network is summarised in Supplementary Table 3. TRAINING STRATEGY For training the policy network, a virtual bronchoscopy environment is established
based on the segmented airway from preoperative thorax CT scans, as introduced in detail in Supplementary Note 4. In this study, we employ an imitation learning framework to train the policy
network. Given an expert policy Π*, a dataset _D_ of state–command–action triples (S,_c_,A*) can be created by executing Π* in the virtual bronchoscopy environment. S represents the state
of the environment, which corresponds to the image observed through the bronchoscope robot’s camera. _c_ denotes the human command, and
\({{{{{{\bf{a}}}}}}}^{{{{{{\boldsymbol{*}}}}}}}={{{{{{\mathbf{\pi }}}}}}}^{{{{{{\boldsymbol{*}}}}}}}({{{{{\bf{s}}}}}},c)\) represents the expert steering action. The objective of imitation
learning is to train a policy network Π, parameterised by Θ, that maps any given S and _c_ to a steering action A that is similar to the corresponding expert action A*. By minimising a loss
function \({{{{{{\mathcal{L}}}}}}}_{a}\), the optimal parameters Θ* can be obtained as follows: $${{{{{{\mathbf{\theta
}}}}}}}^{{{{{{\boldsymbol{*}}}}}}}=\mathop{{{{{{\rm{arg}}}}}}\min}\limits_{{{{{{\mathbf{\theta }}}}}}} \mathop{\sum }\limits_{i}^{N}{{{{{{\mathcal{L}}}}}}}_{a}\left({{{{{{\boldsymbol{\pi
}}}}}}}^{{{{{{\boldsymbol{*}}}}}}}\left({{{{{{\bf{s}}}}}}}_{{{{{{\bf{i}}}}}}},{c}_{i}\right),{{{{{\boldsymbol{\pi
}}}}}}\left({{{{{{\bf{s}}}}}}}_{{{{{{\bf{i}}}}}}},{c}_{i}{{{{{\rm{;}}}}}}{{{{{{\mathbf{\theta }}}}}}}\right)\right)$$ (1) where _N_ is the size of dataset _D_. In the conventional imitation
learning framework, the expert policy Π* is executed by human experts in the environment to collect expert data for training; however, this process is excessively time consuming in practice.
In addition, when a behaviour cloning strategy is used to train the policy Π, cascading error and distribution mismatch problems may occur in the inference stage. In our work, an artificial
expert agent (AEA) is designed to simulate a human expert and automatically execute the expert policy in the virtual bronchoscopy environment, thereby providing the human command _c_ and
annotating the ground-truth expert action A* for state S. Thus, the demonstration burden on human experts can be eliminated. We choose the dataset aggregation algorithm DAgger as the
imitation learning strategy. The initial dataset is constructed by placing the camera sequentially at waypoints along the centreline and labelling the ground-truth actions and commands
obtained from the AEA. A supplementary dataset is then obtained by running the policy network Π in the virtual environment and generating frame-by-frame labels with the AEA, namely, the
on-policy training process. In the training stage, we choose the L2 loss to implement the action loss, as follows:
$${{{{{{\mathcal{L}}}}}}}_{a}\left({{{{{{\bf{a}}}}}}}_{{{{{{\bf{i}}}}}}},{{{{{{\bf{a}}}}}}}_{{{{{{\bf{i}}}}}}}^{{{{{{\boldsymbol{*}}}}}}}\right)=\frac{1}{N}\mathop{\sum
}\limits_{i=1}^{N}{{{{{{\rm{||}}}}}}{{{{{{\bf{a}}}}}}}_{{{{{{\bf{i}}}}}}}-{{{{{{\bf{a}}}}}}}_{{{{{{\bf{i}}}}}}}^{{{{{{\boldsymbol{*}}}}}}}{{{{{\rm{||}}}}}}}_{2}^{2}$$ (2) For depth
estimation, a ground-truth depth D* can be rendered corresponding to the current observation S, accordingly, the depth loss can be computed as
$${{{{{{\mathcal{L}}}}}}}_{{depth}}\left({{{{{\bf{d}}}}}},{{{{{{\bf{d}}}}}}}^{{{{{{\boldsymbol{*}}}}}}}\right)=\frac{1}{{NM}}\mathop{\sum }\limits_{i=1}^{N}\mathop{\sum
}\limits_{j=1}^{M}{{{{{{\rm{||}}}}}}{{{{{{\bf{d}}}}}}}_{{{{{{\bf{ij}}}}}}}-{{{{{{\bf{d}}}}}}}_{{{{{{\bf{ij}}}}}}}^{{{{{{\boldsymbol{*}}}}}}}{{{{{\rm{||}}}}}}}_{2}^{2}$$ (3) where _N_ is the
size of the whole dataset, _M_ is the number of pixels of each depth, and D is the estimated depth of the policy network. In the training process, each rollout of a bronchoscopy procedure is
terminated by a series of ending conditions, which are described in Supplementary Note 6. ARTIFICIAL EXPERT AGENT This section introduces the process of human command generation and
ground-truth expert action annotation by the artificial expert agent (AEA). During the training phase, a substantial number of rollouts of virtual bronchoscopy procedures should be performed
with human commands, and numerous steering actions must be labelled to ensure adequate samples for training the policy network. This task is labour-intensive and time-consuming for doctors,
and the consistency of the resulting human annotations cannot be guaranteed. To address this challenge, we introduce the AEA to automatically provide human commands and annotate
ground-truth steering actions based on privileged robot pose information and reference airway centrelines. As shown in Supplementary Fig. 5b, the ground-truth steering action \([\Delta
{\theta }^{*},\Delta {\varphi }^{*}]\) is calculated as follows: $$\Delta {\theta }^{\ast }=\arccos
\left(\frac{\mathop{{{{{\bf{O}}}}}_{{{{\bf{c}}}}}{{{{\bf{P}}}}}_{{{{\bf{a}}}}}}\limits^{\rightharpoonup }\cdot
{{{\bf{z}}}}}{\overline{{{{{\bf{O}}}}}_{{{{\bf{c}}}}}{{{{\bf{P}}}}}_{{{{\bf{a}}}}}^{{\prime} }}}\right)$$ (4) $$\Delta {\varphi }^{\ast }=\arcsin
\left(\frac{\overline{{{{{{{\bf{P}}}}}}}_{{{{{{\bf{a}}}}}}}{{{{{{\bf{P}}}}}}}_{{{{{{\bf{a}}}}}}}^{{\prime}
}}}{\overline{{{{{{{\bf{O}}}}}}}_{{{{{{\bf{c}}}}}}}{{{{{{\bf{P}}}}}}}_{{{{{{\bf{a}}}}}}}}}\right)$$ (5) where PA is the target waypoint on the centreline that the robot should be directed
towards in the next step, OC is the origin of the camera coordinate system, and \({{{{{{\bf{P}}}}}}}_{{{{{{\bf{a}}}}}}}^{{{{\prime} }}}\) is the projection point of Pa on the
\({{{{{\bf{x}}}}}}{{{{{{\bf{O}}}}}}}_{{{{{{\bf{c}}}}}}}{{{{{\bf{y}}}}}}\) plane. PA can be determined from the current position of the robot and a fixed distance _d__a_ along the centerline.
First, the nearest waypoint \({{{{{{\rm{P}}}}}}}_{{{{{{\rm{n}}}}}}}\) on the centreline from the robot’s head is selected. Then, \({{idx}}_{{Pa}}\), i.e., the index of
\({{{{{{\bf{P}}}}}}}_{{{{{{\bf{a}}}}}}}\) among all waypoints on the centreline, can be calculated as $${{idx}}_{{Pa}}=\mathop{{{{{{\rm{arg}}}}}}\min }\limits_{m}\left|\mathop{\sum
}\limits_{k=n}^{m}\overline{{{{{{{\bf{P}}}}}}}_{{{{{{\bf{k}}}}}}}{{{{{{\bf{P}}}}}}}_{{{{{{\bf{k}}}}}}{{{{{\boldsymbol{+}}}}}}{{{{{\bf{1}}}}}}}}-{d}_{a}\right|$$ (6) where
\({{{{{{\bf{P}}}}}}}_{{{{{{\bf{k}}}}}}}\) para_denotes a certain waypoint that lies on the centerline and
\(\overline{{{{{{{\bf{P}}}}}}}_{{{{{{\bf{k}}}}}}}{{{{{{\bf{P}}}}}}}_{{{{{{\bf{k}}}}}}{{{{{\boldsymbol{+}}}}}}{{{{{\bf{1}}}}}}}}\) is the distance between
\({{{{{{\bf{P}}}}}}}_{{{{{{\bf{k}}}}}}}\) and its neighbour \({{{{{{\bf{P}}}}}}}_{{{{{{\bf{k}}}}}}{{{{{\boldsymbol{+}}}}}}{{{{{\bf{1}}}}}}}\). Thus, the ground-truth steering action
\([\Delta {\theta }^{*},\Delta {\varphi }^{*}]\) can be annotated for training the policy network. The principle of human command generation is based on the fact that doctors consistently
consider both a far navigation target and a near steering target during bronchoscopy procedures. The far navigation target allows the doctor to assess the risks of the upcoming operation and
decide where needs to be examined. The near steering target ensures that the bronchoscope remains at the centre of the airway as much as possible for local safety. The far navigation target
may be approximate yet correct, signifying the desired location the bronchoscope should reach in the near future, similar to the human command in our policy network. For instance, at the
junction of the primary and secondary bronchi, the doctor should decide where to examine in the near future. The policy network receives an approximate human command (left or right) as input
and generates precise safe steering actions for controlling the robot. Thus, in the AEA, the human command is determined based on a far target waypoint
\({{{{{{\bf{P}}}}}}}_{{{{{{\bf{f}}}}}}}\) and the robot’s current position. The index of \({{{{{{\bf{P}}}}}}}_{{{{{{\bf{f}}}}}}}\) can be computed as follows:
$${idx}_{Pf}=\mathop{{{{{{\rm{arg}}}}}}\min }\limits_{m}\left|\mathop{\sum
}\limits_{k=n}^{m}{\overline{{{{{{\bf{P}}}}}}_{{{{{\bf{k}}}}}}{{{{{\bf{P}}}}}}_{{{{{{\bf{k}}}}}}{+}{{{{{\bf{1}}}}}}}}}-{d}_{f}\right|$$ (7) where _d__f_ is the length of the centreline
between PN and PF, satisfying _d__f_ > _d__a_. After that, PF is projected into the image coordinate system with the known intrinsic parameters of the camera to generate the 2D projected
point PF. The discrete human command _c_ can be computed as $$c=\left\{\begin{array}{c}{forward},0^\circ \le \angle
{{{{{{\bf{P}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{{\bf{O}}}}}}}_{{{{{{\bf{c}}}}}}}{{{{{\bf{z}}}}}}\le \tau\hfill \\ {up},\,\angle
{{{{{{\bf{P}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{{\bf{O}}}}}}}_{{{{{{\bf{c}}}}}}}{{{{{\bf{z}}}}}} > \tau \cap 45^\circ < \angle {{{{{{\bf{p}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{\bf{Ox}}}}}}\le
135^\circ\hfill \\ {down},\,\angle {{{{{{\bf{P}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{{\bf{O}}}}}}}_{{{{{{\bf{c}}}}}}}{{{{{\bf{z}}}}}} > \tau \cap 135^\circ < \angle
{{{{{{\bf{p}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{\bf{Ox}}}}}}\le 225^\circ\hfill \\ {left},\, \angle {{{{{{\bf{P}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{{\bf{O}}}}}}}_{{{{{{\bf{c}}}}}}}{{{{{\bf{z}}}}}} >
\tau \cap 225^\circ < \angle {{{{{{\bf{p}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{\bf{Ox}}}}}}\le 315^\circ\hfill \\ {right},\, \angle
{{{{{{\bf{P}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{{\bf{O}}}}}}}_{{{{{{\bf{c}}}}}}}{{{{{\bf{z}}}}}} > \tau \cap (0^\circ < \angle {{{{{{\bf{p}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{\bf{Ox}}}}}}\le
45^\circ \cup 315^\circ < \angle {{{{{{\bf{p}}}}}}}_{{{{{{\bf{f}}}}}}}{{{{{\bf{Ox}}}}}}\le 360^\circ )\end{array}\right.$$ (8) where O is the origin of the image coordinate system and τ
is the threshold angle of the forward cone for deciding whether to continue forward in the current airway. The five discrete human commands that can be generated by the AEA are encoded as
one-hot vectors for input into the policy network. In practice, the input AEA-annotated human commands are replaced with the doctor’s commands in the real bronchoscope robot, driving the
policy network to safely and smoothly pass through the airway. The human commands are mapped to five regions of the teleoperator (Supplementary Fig. 3b), reducing the doctor’s cognitive load
compared to that imposed by conventional teleoperated robots with continuous human intervention. SIM2REAL ADAPTATION DOMAIN ADAPTATION AND TRAINING STRATEGY To improve the performance of
the policy network in clinical scenarios, domain adaptation is necessary to reduce the gap between the simulated and real environments. Generative adversarial networks (GANs), which are
often used in computer vision for image domain adaptation, can serve our purpose. The generator G of such a GAN attempt to generate realistic-style images from simulated images, while the
discriminator D attempts to distinguish between generated and real samples. Notably, in clinical scenarios, it is still challenging to pair every bronchoscopic video frame with simulated
images rendered from CT airway models due to limited available manpower and the significant visual divergence between body and CT images. When only unpaired data are used for training,
existing unpaired image translation methods, such as CycleGAN, often misinterpret crucial structural information of the bronchus as part of the style to be translated, leading to inaccurate
structures in the generated images. To address these issues, we propose a structure-preserving unpaired image translation method leveraging a GAN and a depth constraint for domain
adaptation. As shown in Supplementary Fig. 9, the network consists of a generator, a discriminator and a depth estimator. Sim-style images rendered from airway models with pink textures are
collected to represent the source domain, and their corresponding depths are rendered to provide depth supervision. Unpaired clinical images from historical bronchoscopic videos, which are
easy to access at hospitals, serve to represent the target domain. In the training stage, Sim-style images X are fed into the generator to translate them into paired realistic-style images
G(X). Then, the discriminator takes both the translated realistic-style images and the unpaired clinical images Y as input. The adversarial loss is formulated as $${ {\mathcal L}
}_{GAN}({{{{{\bf{G}}}}}},{{{{{\bf{D}}}}}},{{{{{\bf{x}}}}}},{{{{{\bf{y}}}}}})={{\mathbb{E}}}_{{{{{{\bf{y}}}}}} \sim {p}_{data}({{{{{\bf{y}}}}}})}[\log
{{{{{\bf{D}}}}}}({{{{{\bf{y}}}}}})]+{{\mathbb{E}}}_{{{{{{\bf{x}}}}}} \sim {p}_{data}({{{{{\bf{x}}}}}})}[\log (1-{{{{{\bf{D}}}}}}({{{{{\bf{G}}}}}}({{{{{\bf{x}}}}}})))]$$ (9) Following image
translation, the realistic-style images are fed into the depth estimator for the generation of estimated depths. The depth estimation task can be supervised by the rendered depths
corresponding to the input rendered images, ensuring that the 3D structure information of each generated image remains consistent with that of the original rendered image. The depth
constraint is provided by the depth loss, which is expressed as $${ {\mathcal L} }_{{{{{{\rm{depth}}}}}}}({{{{{\bf{d}}}}}},{{{{{{\bf{d}}}}}}}^{\ast })=\frac{1}{N}\mathop{\sum
}\limits_{i=1}^{N}{\left\| {{{{{{\bf{d}}}}}}}_{{{{{{\bf{i}}}}}}}-{{{{{{\bf{d}}}}}}}_{{{{{{\bf{i}}}}}}}^{\ast }\right\| }_{2}^{2}$$ (10) where _N_ is the number of pixels in the depth image,
D is the predicted depth and D* is the corresponding rendered depth of the input rendered image. As shown in Supplementary Table 4, the backbone of our generator is based on the architecture
of AttentionGAN, which explicitly decouples the foreground and background of an image through the introduction of a self-attention mechanism and has shown state-of-the-art performance in
recent image translation tasks. The generator G is composed of a parameter-sharing feature extractor GE, an attention mask generator GA and a content mask generator GC. The discriminator is
based on the architecture of CycleGAN. The depth estimator comprises a ResNet-34 network for feature extraction and a transposed convolutional network for feature decoding. Specific network
architectures of our generator and discriminator can be found in Supplementary Table 4. DOMAIN RANDOMISATION To improve the generalisation ability of the policy network, several domain
randomisation techniques are designed to randomly alter image appearances or add noise to human commands. (i) Roll rotation: Since the bronchoscope robot’s roll angle is set to 0 in the
virtual environment (Supplementary Note 4), we randomly rotate the airway model’s roll angle for each rollout. This prevents overfitting of the policy network on limited pulmonary postures
and encourages the learning of a more general safe steering policy, regardless of environmental changes. For each rollout, the starting point is randomly placed within the first third of the
reference path to collect more challenging data from the deep, thin bronchus, as the trachea and main bronchus account for a large proportion of the entire path. (ii) Light intensity: To
improve the policy network’s robustness against variations in light intensity between simulated and clinical scenarios, we randomly adjust the light intensity of the virtual bronchoscopy
environment for each observation. (iii) Command disturbance: Humans may make mistakes when operating robots in practice, such as continuing to control the robot’s heading towards the
bronchial wall even once a collision has occurred, potentially causing damage or even perforation. To address this, we randomly add disturbances to the human commands before they are input
into the policy network when the robot is less than 1 mm away from the bronchial wall during training. This ensures that the policy network will prioritise safe steering over erroneous human
control, enhancing safety and reducing the doctor’s cognitive load. (iv) Image attributes: To further improve the generalisation ability of the policy network, we employ four data
augmentation methods during the training process to randomly change various attributes of the input images, including brightness, contrast, saturation and hue. IMPLEMENTATION DETAILS The
AI-human shared control algorithm is implemented using Python (v3.7.11). To establish the virtual bronchoscopy environment, the airway models are segmented from CT scans using 3D Slicer
(v4.10.2). The airway centrelines are then extracted using VMTK (v1.4.0). For robot simulation, data acquisition and human command generation, the implementation tools are Pyrender (v0.1.45)
and PyBullet (v3.2.2). The policy network and Sim2Real adaptation module are implemented using the PyTorch platform (v1.9.1) and trained on an NVIDIA GTX 2080Ti GPU. The learning rate is
set to 10-4 for training the policy network and 2×10-4 for training the Sim2Real adaptation module. The batch size for training is set to 64, and both networks are trained for 300 epochs.
The acquired images from the simulated and real camera have a size of 400×400, and they are resized to 200×200 before being inputted into the network. The data processing tools include NumPy
(v1.19.5), OpenCV (v4.5.5.64) and VTK (v8.2.0). The data analysis and visualisation are implemented using Matplotlib (v3.3.4) and MATLAB R2022a. REPORTING SUMMARY Further information on
research design is available in the Nature Portfolio Reporting Summary linked to this article. DATA AVAILABILITY The data used for establishing virtual bronchoscopy environment and training
networks are available at https://zenodo.org/records/10077275, whereas the trained network model data used in simulated, in-vitro and in-vivo experiments is available at
https://zenodo.org/records/10077290. Other data needed to evaluate the conclusions are provided in the main text, and Supplementary files. Source data are provided with this paper. CODE
AVAILABILITY The code is available on GitHub (https://github.com/LiuLiluZJU/AI-Co-Pilot-Bronchoscope-Robot)43. The DOI for the code is https://doi.org/10.5281/zenodo.10077315. The repository
includes virtual environment establishment, data acquisition, image processing, visualisation, network training and testing code. REFERENCES * Siegel, R. L., Miller, K. D., Fuchs, H. E.
& Jemal, A. Cancer statistics. _CA: Cancer J. Clin._ 72, 7–33 (2022). PubMed Google Scholar * Minna, J. D., Roth, J. A. & Gazdar, A. F. Focus on lung cancer. _Cancer Cell_ 1, 49–52
(2002). Article PubMed CAS Google Scholar * Richeldi, L. et al. Digital Lung auscultation: will early diagnosis of fibrotic interstitial lung disease become a reality? _Am. J. Respir.
Crit. Care Med._ 200, 261–263 (2019). Article PubMed Google Scholar * Fulkerson, W. J. Fiberoptic bronchoscopy. _N. Engl. J. Med._ 311, 511–515 (1984). Article PubMed CAS Google
Scholar * Reck, M., Heigener, D. F., Mok, T., Soria, J.-C. & Rabe, K. F. Management of non-small-cell lung cancer: recent developments. _Lancet_ 382, 709–719 (2013). Article PubMed
CAS Google Scholar * Criner, G. J. et al. Interventional bronchoscopy. _Am. J. Respir. Crit. Care Med._ 202, 29–50 (2020). Article PubMed Google Scholar * Haponik, E. F. et al.
Bronchoscopy training: current fellows’ experiences and some concerns for the future. _Chest_ 118, 625–630 (2000). Article PubMed CAS Google Scholar * Obaseki, D., Adeniyi, B., Kolawole,
T., Onyedum, C. & Erhabor, G. Gaps in capacity for respiratory care in developing countries. Nigeria as a case study. _Ann. Am. Thorac. Soc._ 12, 591–598 (2015). Article PubMed Google
Scholar * Singh, N. et al. Lung cancer in India. _J. Thorac. Oncol_. 16, 1250–1266 (2021). * Rankin, N. M., McWilliams, A. & Marshall, H. M. Lung cancer screening implementation:
complexities and priorities. _Respirology_ 25, 5–23 (2020). Article PubMed Google Scholar * Froudarakis, M. E., Bouros, D. & Siafakas, N. M. Endoluminal metastases of the
tracheobronchial tree: is there any way out? _Chest_ 119, 679–681 (2001). Article PubMed CAS Google Scholar * Shademan, A. et al. Supervised autonomous robotic soft tissue surgery. _Sci.
Transl. Med._ 8, 337ra364–337ra364 (2016). Article Google Scholar * Chen, A. I., Balter, M. L., Maguire, T. J. & Yarmush, M. L. Deep learning robotic guidance for autonomous vascular
access. _Nat. Mach. Intell._ 2, 104–115 (2020). Article Google Scholar * Saeidi, H. et al. Autonomous robotic laparoscopic surgery for intestinal anastomosis. _Sci. Robot._ 7, eabj2908
(2022). Article PubMed PubMed Central CAS Google Scholar * Ho, E., Cho, R. J., Keenan, J. C. & Murgu, S. The feasibility of using the “artery sign” for pre-procedural planning in
navigational bronchoscopy for parenchymal pulmonary lesion sampling. _Diagnostics_ 12, 3059 (2022). Article PubMed PubMed Central Google Scholar * Reisenauer, J. et al. Ion: technology
and techniques for shape-sensing robotic-assisted bronchoscopy. _Ann. Thorac. Surg._ 113, 308–315 (2022). Article PubMed Google Scholar * Murgu, S. D. Robotic assisted-bronchoscopy:
technical tips and lessons learned from the initial experience with sampling peripheral lung lesions. _BMC Pulm. Med._ 19, 1–8 (2019). Article Google Scholar * Simoff, M. J. et al.
Shape-sensing robotic-assisted bronchoscopy for pulmonary nodules: initial multicenter experience using the Ion™ Endoluminal System. _BMC Pulm. Med._ 21, 1–13 (2021). Article Google Scholar
* Benn, B. S., Romero, A. O., Lum, M. & Krishna, G. Robotic-assisted navigation bronchoscopy as a paradigm shift in peripheral lung access. _Lung_ 199, 177–186 (2021). Article PubMed
Google Scholar * Kalchiem-Dekel, O. et al. Shape-sensing robotic-assisted bronchoscopy in the diagnosis of pulmonary parenchymal lesions. _Chest_ 161, 572–582 (2022). Article PubMed
Google Scholar * Ost, D. et al. Prospective multicenter analysis of shape-sensing robotic-assisted bronchoscopy for the biopsy of pulmonary nodules: results from the PRECIsE study. _Chest_
160, A2531–A2533 (2021). Article Google Scholar * Van Lewen, D. et al. A millimeter-scale soft robot for tissue biopsy procedures. _Adv. Intell. Syst._ 5, 2200326 (2023). Article PubMed
Google Scholar * Pittiglio, G. et al. Patient-specific magnetic catheters for atraumatic autonomous endoscopy. _Soft Robot_ 9, 1120–1133 (2022). Article PubMed PubMed Central Google
Scholar * McCandless, M. et al. A soft robot for peripheral lung cancer diagnosis and therapy. _Soft Robot_. 9, 754–766 (2022). Article PubMed Google Scholar * Pittiglio, G. et al.
Personalized magnetic tentacles for targeted photothermal cancer therapy in peripheral lungs. _Commun. Eng._ 2, 50 (2023). Article Google Scholar * Kato, T., King, F., Takagi, K. &
Hata, N. Robotized catheter with enhanced distal targeting for peripheral pulmonary biopsy. _IEEE ASME Trans. Mechatron_ 26, 2451–2461 (2020). Article Google Scholar * Masaki, F. et al.
Technical validation of multi-section robotic bronchoscope with first person view control for transbronchial biopsies of peripheral lung. _IEEE. Trans. Biomed. Eng._ 68, 3534–3542 (2021).
Article PubMed Google Scholar * Dupont, P. E. et al. A decade retrospective of medical robotics research from 2010 to 2020. _Sci. Robot._ 6, eabi8017 (2021). Article PubMed PubMed
Central Google Scholar * Esteva, A. et al. Deep learning-enabled medical computer vision. _NPJ Digit. Med._ 4, 5 (2021). Article PubMed PubMed Central Google Scholar * Yoo, J. Y. et
al. Deep learning for anatomical interpretation of video bronchoscopy images. _Sci. Rep._ 11, 23765 (2021). Article ADS PubMed PubMed Central CAS Google Scholar * Quan, S. Y. et al.
Clinical evaluation of a real-time artificial intelligence-based polyp detection system: a US multi-center pilot study. _Sci. Rep._ 12, 6598 (2022). Article ADS PubMed PubMed Central
Google Scholar * Wang, P. et al. Development and validation of a deep-learning algorithm for the detection of polyps during colonoscopy. _Nat. Biomed. Eng._ 2, 741–748 (2018). Article
PubMed Google Scholar * Bao, Y. et al. Study on the interventional path planning method of bronchoscope catheter. _J. Mech. Sci. Technol._ 36, 2643–2652 (2022). Article Google Scholar *
Merritt, S. A., Khare, R., Bascom, R. & Higgins, W. E. Interactive CT-video registration for the continuous guidance of bronchoscopy. _IEEE Trans. Med. Imaging_ 32, 1376–1396 (2013).
Article PubMed PubMed Central Google Scholar * Ehrmann, D. E. et al. Evaluating and reducing cognitive load should be a priority for machine learning in healthcare. _Nat. Med._ 28,
1331–1333 (2022). Article PubMed CAS Google Scholar * Banach, A., King, F., Masaki, F., Tsukada, H. & Hata, N. Visually navigated bronchoscopy using three cycle-consistent generative
adversarial network for depth estimation. _Med. Image Anal._ 73, 102164 (2021). Article PubMed PubMed Central Google Scholar * Shen, M., Giannarou, S. & Yang, G.-Z. Robust camera
localisation with depth reconstruction for bronchoscopic navigation. _Int. J. Comput. Assist. Radiol. Surg._ 10, 801–813 (2015). Article PubMed Google Scholar * Gu, Y., Gu, C., Yang, J.,
Sun, J. & Yang, G.-Z. Vision–kinematics interaction for robotic-assisted bronchoscopy navigation. _IEEE Trans. Med. Imaging_ 41, 3600–3610 (2022). Article PubMed Google Scholar *
Ross, S. & Bagnell, D. in Proceedings of the thirteenth international conference on artificial intelligence and statistics. _JMLR Workshop Conf Proc_, 661–668 (2010). * Tang, H., Liu,
H., Xu, D., Torr, P. H. & Sebe, N. Attentiongan: unpaired image-to-image translation using attention-guided generative adversarial networks. _IEEE Trans. Neural Netw. Learn. Syst._ 34,
1972–1987 (2021). Article Google Scholar * Russo, M. et al. Continuum robots: an overview. _Adv. Intell. Syst._ 5, 2200367 (2023). Article Google Scholar * Dupourqué, L. et al.
Transbronchial biopsy catheter enhanced by a multisection continuum robot with follow-the-leader motion. _Int. J. Comput. Assist. Radiol. Surg._ 14, 2021–2029 (2019). Article PubMed Google
Scholar * J. Zhang, et al. _AI Co-pilot Bronchoscope Robot v1.0.0_. https://doi.org/10.5281/zenodo.10077315 (2023). Download references ACKNOWLEDGEMENTS This work was supported by the
National Key R&D Program of China under Grant no. 2021ZD0114500 (R.X.); the National Natural Science Foundation of China under Grants no. 62373322 (Y.W.), no. T2293724 (H.L.) and no.
62303407 (H.L.); the Key R&D Program of Zhejiang under Grants no. 2023C01176 (Y.W.) and no. 2022C01022 (H.L.); and the Zhejiang Provincial Natural Science Foundation of China under Grant
no. LD22E050007 (H.L.). AUTHOR INFORMATION Author notes * These authors contributed equally: Jingyu Zhang, Lilu Liu. AUTHORS AND AFFILIATIONS * State Key Laboratory of Industrial Control
and Technology, Zhejiang University, 310027, Hangzhou, China Jingyu Zhang, Lilu Liu, Pingyu Xiang, Qin Fang, Xiuping Nie, Rong Xiong, Yue Wang & Haojian Lu * Institute of Cyber-Systems
and Control, Department of Control Science and Engineering, Zhejiang University, 310027, Hangzhou, China Jingyu Zhang, Lilu Liu, Pingyu Xiang, Qin Fang, Xiuping Nie, Rong Xiong, Yue Wang
& Haojian Lu * Department of Thoracic Surgery, First Affiliated Hospital, School of Medicine, Zhejiang University, 310009, Hangzhou, China Honghai Ma & Jian Hu Authors * Jingyu Zhang
View author publications You can also search for this author inPubMed Google Scholar * Lilu Liu View author publications You can also search for this author inPubMed Google Scholar * Pingyu
Xiang View author publications You can also search for this author inPubMed Google Scholar * Qin Fang View author publications You can also search for this author inPubMed Google Scholar *
Xiuping Nie View author publications You can also search for this author inPubMed Google Scholar * Honghai Ma View author publications You can also search for this author inPubMed Google
Scholar * Jian Hu View author publications You can also search for this author inPubMed Google Scholar * Rong Xiong View author publications You can also search for this author inPubMed
Google Scholar * Yue Wang View author publications You can also search for this author inPubMed Google Scholar * Haojian Lu View author publications You can also search for this author
inPubMed Google Scholar CONTRIBUTIONS J.Z. designed and fabricated the whole bronchoscope robot. L.L. designed the AI–human shared control algorithm. J.Z., L.L. and P.X. developed the
software and carried out all the experiments. Q.F. and X.N. validated the work. H.M. and J.H. provided technical support for the clinical experiments. R.X., Y.W. and H.L. provided technical
support for the algorithm design and funding support. J.Z. and L.L. wrote the paper. CORRESPONDING AUTHORS Correspondence to Rong Xiong, Yue Wang or Haojian Lu. ETHICS DECLARATIONS COMPETING
INTERESTS TThe authors declare no competing interests. PEER REVIEW PEER REVIEW INFORMATION _Nature Communications_ thanks Andrea Moglia and the other, anonymous, reviewer(s) for their
contribution to the peer review of this work. A peer review file is available. ADDITIONAL INFORMATION PUBLISHER’S NOTE Springer Nature remains neutral with regard to jurisdictional claims in
published maps and institutional affiliations. SUPPLEMENTARY INFORMATION SUPPLEMENTARY INFORMATION PEER REVIEW FILE DESCRIPTION OF ADDITIONAL SUPPLEMENTARY FILES SUPPLEMENTARY MOVIE 1
SUPPLEMENTARY MOVIE 2 SUPPLEMENTARY MOVIE 3 SUPPLEMENTARY MOVIE 4 REPORTING SUMMARY SOURCE DATA SOURCE DATA RIGHTS AND PERMISSIONS OPEN ACCESS This article is licensed under a Creative
Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the
original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in
the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended
use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit
http://creativecommons.org/licenses/by/4.0/. Reprints and permissions ABOUT THIS ARTICLE CITE THIS ARTICLE Zhang, J., Liu, L., Xiang, P. _et al._ AI co-pilot bronchoscope robot. _Nat Commun_
15, 241 (2024). https://doi.org/10.1038/s41467-023-44385-7 Download citation * Received: 13 June 2023 * Accepted: 12 December 2023 * Published: 04 January 2024 * DOI:
https://doi.org/10.1038/s41467-023-44385-7 SHARE THIS ARTICLE Anyone you share the following link with will be able to read this content: Get shareable link Sorry, a shareable link is not
currently available for this article. Copy to clipboard Provided by the Springer Nature SharedIt content-sharing initiative
Trending News
Rural crime wave hits scotland, wales and northern ireland - farmers weeklyFARMERS IN Northern Ireland, Wales and Scotland need to tighten up their security because the regions are fast becoming ...
Kitchen sink: basin of information👮♂️ THE SPRINGDALE POLICE chief plans to retire next year. (_Northwest Arkansas Democrat-Gazette)_ 🎥 A NEW MOVIE — a We...
The aarp minute: december 6, 2019Memorial Day Sale! Join AARP for just $11 per year with a 5-year membership Join now and get a FREE gift. Expires 6/4 G...
Kids now spend nearly as much time watching tiktok as youtube in us, uk and spain | techcrunchA new study on kids’ app usage and habits indicates a major threat to YouTube’s dominance, as kids now split their time ...
Co-operation in Bibliography | NatureABSTRACT THE movement for co-operation in bibliography has received an important impetus from the Bataafsche Petroleum M...
Latests News
Ai co-pilot bronchoscope robotABSTRACT The unequal distribution of medical resources and scarcity of experienced practitioners confine access to bronc...
Clinical puzzle answer: painful red gumsYou have full access to this article via your institution. Download PDF This is the answer to the clinical puzzle in the...
Bigg boss 14: just who is nikki tamboli?She’s beautiful, she’s fierce and has already got everyone rooting for her to win _Bigg Boss 14_. NIKKI TAMBOLI is makin...
Manchester united face psg and leipzig after tough champions league drawManchester United face a challenging Champions League group stage after being placed alongside Paris Saint-Germain, RB L...
The research suggests there are more benefits to consuming foods high in calcium, such as milk and…The research suggests there are more benefits to consuming foods high in calcium, such as milk and cheese, than there ar...