The Previous, Current and Eventual Fate of Automated Medical Procedure

Advances in Robotics & Automation

ISSN: 2168-9695

Open Access

Commentary - (2022) Volume 11, Issue 10

The Previous, Current and Eventual Fate of Automated Medical Procedure

Antonello Rizzi*
*Correspondence: Antonello Rizzi, Department of Information Engineering, Sapienza University of Rome, Rome, Italy, Email:
Department of Information Engineering, Sapienza University of Rome, Rome, Italy

Received: 05-Oct-2022, Manuscript No. ara-22-78288; Editor assigned: 07-Oct-2022, Pre QC No. P-78288; Reviewed: 10-Oct-2022, QC No. Q-78288; Revised: 15-Oct-2022, Manuscript No. R-78288; Published: 20-Oct-2022 , DOI: 10.37421/2168-9695.2022.11.234
Citation: Rizzi, Antonello. “The Previous, Current and Eventual Fate of Automated Medical Procedure.” Adv Robot Autom 11 (2022): 234.
Copyright: © 2022 Rizzi A. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.


In 2004, the US's Safeguard Progressed Exploration Tasks Organization (DARPA) hung a $1 million award for any gathering that could plan an independent vehicle that could travel itself through 142 miles of harsh territory from Barstow, California, to Primm, Nevada. After thirteen years, the Branch of Guard declared another honor not so much for a robot vehicle this time, but rather for independent, mechanical specialists [1]. Robots have been tracked down in the working suite since the 1980s for things like holding a patient's appendages set up, and later for laparoscopic medical procedure, in which specialists can utilize remote-controlled robot arms to work on the human body through minuscule openings rather than gigantic cuts. In any case, generally these robots have been, fundamentally, super extravagant variants of the surgical blades and forceps specialists have been utilizing for a really long time, extraordinarily complex, in truth, and equipped for working with unimaginable accuracy, yet devices in the specialist's hands.

In spite of many difficulties, that is evolving. Today, five years after that grant declaration, engineers are moving toward building autonomous machines that not exclusively can cut or stitch, yet in addition plan those cuts, make do and adjust. Specialists are working on the machines' capacity to explore the intricacies of the human body and direction with human specialists [2]. Yet, the genuinely independent mechanical specialist that the military might imagine, very much like really driverless vehicles, May in any case is quite far off. What's more, their greatest test may not be mechanical, however persuading individuals utilizing them is alright. Like drivers, specialists should figure out how to explore their particular surroundings, something that sounds simple on a basic level yet is perpetually muddled in reality. Genuine streets have traffic, development hardware, walkers, everything that don't be guaranteed to appear on Google Guides and which the vehicle should figure out how to keep away from.

The way that bodies move represents a further intricacy. A couple of robots as of now show some measure of independence, with one of the exemplary models being a gadget with the (perhaps a-piece on-the-button) name ROBODOC, which can be utilized in hip medical procedure to shave down bone around the hip attachment. However, bone's generally simple to work with and, once got into place, doesn't move around a lot [3].

One of the most encouraging choices for such powerful circumstances couples the utilization of cameras and complex following programming. In mid-2022, for instance, scientists at Johns Hopkins College utilized a gadget called the Savvy Tissue Independent Robot to sew two finishes of cut off digestive tract back together in an anesthetized pig a possibly jiggly task because of this visual framework. A human administrator labels the finishes of the digestive system with drops of fluorescent paste, making markers the robot can follow [4]. Simultaneously, a camera framework makes a three dimensional model of the tissue utilizing a matrix of light focuses projected onto the area. Together, these advances permit the robot to see what is before it.

Assuming that something moves during the medical procedure, you can recognize and follow it. The robot can then utilize this visual data to foresee the best game-plan, giving the human administrator various designs to browse or checking in with in the middle between stitches. In tests, STAR functioned admirably all alone, however not impeccably. Altogether, 83% of the stitches should be possible independently, yet the human actually needed to step in the other 17% of an opportunity to address things. A large portion of the issue was that the robot experienced a little difficulty tracking down the right point at specific corners and required a human to prod it into the perfect place, he says. More up to date, yet-to-be-distributed preliminaries presently have achievement rates in the high 90s [5]. Later on, the human may just have to support the arrangement, and then watch it go, no intercession required.

Conflict of Interest



  1. Zhang, Jiwang, Ning Zhang, Mintang Zhang and Liantao Lu, et al. “Microstructure and mechanical properties of austempered ductile iron with different strength grades.” Mater Lett 119 (2014): 47–50.
  2. Google Scholar, Crossref

  3. Peng, Xiao, Kai Dong, Cuiying Ye and Yang Jiang, et al. “A Breathable, Biodegradable, Antibacterial, and Self-Powered Electronic Skin Based on All-Nanofiber Triboelectric Nanogenerators.” Sci Adv 6 (2020): 1–10.
  4. Google Scholar, Crossref, Indexed at

  5. Ballantyne, Garth H. “Robotic surgery, telerobotic surgery, telepresence, and telementoring. Review of early clinical results.” Surg Endosc 16 (2002): 1389–1402.
  6. Google Scholar, Crossref, Indexed at

  7. Camarillo, David B., Thomas M. Krummel and J. Kenneth Salisbury. “Robotic technology in surgery: past, present, and future.” Am J Surg 188 (2004): 2S–15S.
  8. Google Scholar, Crossref, Indexed at

  9. LeCun, Yann, Yoshua Bengio and Geoffrey Hinton. “Deep learning.” Nature 521 (2015): 436–444.
  10. Google Scholar, Crossref, Indexed at

Google Scholar citation report
Citations: 1004

Advances in Robotics & Automation received 1004 citations as per Google Scholar report

Advances in Robotics & Automation peer review process verified at publons

Indexed In

arrow_upward arrow_upward