Open Access Article
Edy Mariano
*a,
Yannis Codereya,
Yasmine El Goumia,
Jasper Tan
a,
Tanguy Cavagna
a,
Jean-Charles Cousty
a,
Vincenzo Scamarcio
bc,
Josie Hughes
b and
Pascal Miéville
a
aSwiss Cat+ West Hub, Ecole Polytechnique Fédérale de Lausanne EPFL, 1015 Lausanne, Switzerland. E-mail: edy.mariano@epfl.ch
bCREATE Lab, Ecole Polytechnique Fédérale de Lausanne EPFL, 1015 Lausanne, Switzerland
cSUNMIL, Ecole Polytechnique Fédérale de Lausanne EPFL, 1015 Lausanne, Switzerland
First published on 17th October 2025
Laboratory automation is an active field in biology, drug discovery, and more recently in synthetic chemistry and materials science. Local automation has existed in the field for quite some time, but long-range or total laboratory automation is much less developed. In this article, we present a complete, open and decentralized global automation system called the 2D drone swarm system. It is based on a simple approach of small mobile robots moving autonomously in a dedicated track suspended above the scientific equipment for the long-distance sample and closely connected to localized robotic arms dedicated to short-distance transfers, interaction with scientific equipment and direct sample processing. This approach is inspired by the Kiva/Amazon model, where isolated autonomous mobile robots automatically deliver goods to external operators. It is also inspired by the modern automotive industry, such as Tesla's Gigafactories, to provide an evolutionary and flexible system that can adapt to numerous types of tasks with a minimum amount of resources and easily adapt to different types of workstations. This global automation system is controlled directly from the Laboratory Scheduler by a Robot Subscheduler, coded in an open-source environment, which takes care of all mobile and local robot operations. The result is an operator and scientific equipment safe, cost and energy-efficient, easily extensible and open-source global laboratory automation system that can be adapted to many different applications and laboratories.
![]() | ||
| Fig. 1 Adapted from Thurow and Fleischer.3 The four possible combinations of the open vs. closed and centralized vs. decentralized criteria allow the different laboratory automation systems to be described. (a) Representation of a closed and centralized automation system (CCAS), where a fixed number of workstations (WS#) are directly interconnected through a localized connection element (CE), (b) representation of an open and centralized automation system (OCAS), where a variable number of workstations are interconnected through a localized connection element. (c) Representation of a closed and decentralized automation system (CDAS) where a mobile link is used to connect a defined number of workstations. (d) Representation of an open and distributed automation system (ODAS) where a mobile link is used to interconnect an open number of workstations. | ||
The first combination is the closed and centralized automation system (CCAS, Fig. 1a). An example would be two workstations, a liquid handling station (WS1) used to prepare calibration solutions directly connected to an analyzer (WS2, e.g. HPLC, UV reader…) with a stationary arm serving as a connecting element (CE, e.g. SCARA arm) to transfer the plates. All commercially available automated analytical tools with internal autosamplers and analyzers (e.g. DAD, MS, FID, UV/VIS, NMR,…) should also be considered as the CCAS. The second combination is the open and centralized automation system (OCAS, Fig. 1b). There are several remarkable examples of such automation systems in the literature, such as the robotic platform for flow synthesis of organic compounds4 or the Chemputer.5 Commercial preparative stations (e.g. Chemspeed, Unchained Lab,…) or pipetting stations (e.g. Agilent Bravo, Hamilton, Opentrons,…) can be assigned to one of these first two categories, depending on their configuration and ease of workflow configuration. As previously described, these first two automation systems are localized and cannot be extended out of their defined space to expand their chemical capabilities, for example to include an autonomous solid sampling station connected to a large library of molecules, or to introduce complex analytical tools such as high-field NMR spectrometers or X-ray. To increase the overall ability to automate different types of workstations in the laboratory, it is necessary to turn to a distributed automation system.6,7 The first decentralized system is the closed and decentralized system (CDAS). This corresponds to several localized automated systems, either CCAS or OCAS, interconnected with a rigid conveyor (e.g. Chemspeed Flexshuttle) or autonomous guided vehicles (AGV). As explained by Thurow and Fleischer,3 in such a CDAS: “… the process to be performed is predetermined and can only be changed by reprogramming the entire system. Thus, the system is considered a closed system. The integration of additional components is only possible with a considerable amount of time and money”.
In order to extend the spatial domain of operation, an ODAS must certainly consist of a combination of mobile robots with stationary ones, possibly connected to a linear axis. Thurow and Junginger8 describe three possible scenarios for the use of mobile robots in laboratory automation:
Scenario 1 – mobile robot as a transport system, where simple mobile platforms without on-board manipulation are used to move samples between robotized stations. In this scenario, the transfer between the mobile platforms and the scientific equipment is performed by stationary robot arms fixed on the stations, where stationary indicates the arm's base remains fixed relative to its immediate platform or environment.
Scenario 2 – mobile robot as a transport system with a transfer function, where, in a full automation vision, mobile platforms are combined with an on-board stationary robot arm (6-axis, SCARA) and appropriate sensors to transfer the samples to and from the scientific equipment or storage stations. (e.g. Biosero-Omron,9 KMR iiwa Kuka, Augsburg, Germany)
Scenario 3 – mobile robot as a transport system with a transfer and manipulation function, where mobile platforms are combined with one or more on-board stationary robot arms (6-axis, SCARA) and adequate sensors. These components allow the samples to be transferred to and from the scientific equipment or storage stations, but also to take over some processing of the samples (e.g. Wolf et al. mobile manipulator MOMA,6,10 Cooper et al. in 2020 as the Robot Chemist11 which is a KMR iiwa Kuka robot with extended processing function).
The authors mention a number of limitations for the different strategies mentioned above. For scenarios 2 & 3, one of the problems is related to the limited universality of the gripping tools, which ultimately forces the use of gripper changers, which are not necessarily easy to set up on a mobile station and also create complex programming of the robot. The use of multiple arms could also be considered to improve the versatility of such a MOMA.12 The authors also explain that the alignment of the mobile platforms and the robotic arms with respect to the scientific equipment stations can be complex, requiring more complex sensors and programming. They question the feasibility of such solutions in automated life science laboratories. We could also question several other points of the above strategies. For all of them, we should consider the significant footprint of these mobile platforms in the laboratory, especially when we consider that at least some level of redundancy is required for continuity of operations (battery charging time, troubleshooting) and task parallelization. In addition, even if we aim for full automation of the laboratory, human operators will certainly be needed for maintenance and recovery from automation crashes. The mobile platform will therefore have to deal with unpredictable humans with a significant risk of collision. Finally, for scenarios 2 & 3, we can question the ability to parallelize, because the more time each mobile station spends on local transfer tasks and sample processing, the less time it will have available for long-distance transfer operations. In scenario 3, the issue is even more critical, and task parallelization could quickly be limited by the available number of MOMA robots, and possibly by the available area on which to move them. With respect to the above limitation list, it appears that scenario 1 is the one presenting the smallest number of limitations. The question, then, is whether we can envision an ODAS based on scenario 1, that is highly parallelizable while maintaining the flexibility to perform complex tasks, that does not constrain the laboratory design by its footprint, that maximizes the safety of people, the ease of alignment with equipment, and that can operate robustly around the clock at a limited cost.
If we look at the logistics warehouse industry, it handles a broad range of load units and packaging types, from large pallets to small individual items. These differences often dictate which handling and transport systems are most suitable. We also see an evolution towards the ODAS approach, which can be adapted in different ways depending on the material format. In the beginning, it was mostly humans walking through long aisles of goods to pick and collect the desired items. Then came an approach using conveyors with static humans to finally move to the more recent strategy used in companies like Amazon with the Kiva robotic system.19,20 The Kiva system is based on an ensemble or swarm of Autonomous Semi-Guided Vehicles (ASGVs)21 that transport the required samples (shopping carts) to human operators or localized static robots that assemble the customer packages. Semi-guided vehicles follow predefined routes using simple infrastructure cues (e.g., floor QR codes or magnetic tape), offering more flexibility than traditional fixed-path AGVs but less autonomy than fully fledged AMRs. In practice, the KIVA ASGVs use QR codes printed on the floor and an embedded camera combined with a collision detector to crisscross the floor in a dedicated, human-free space which enables flow separation (a clear division between humans and mobile robots operating in different areas). A similar system at Walmart allowed the number of picks by operators to be increased from 200–400 h−1, for the conveyor strategy, to up to 600–700 h−1.
To summarize the last generation of ODAS observed in the automotive and logistic-warehouse industries, we can identify three interesting strategies that will help us to achieve our objectives as defined above: (1) stationary robots able to perform different complex tasks and also take over transfers, (2) a swarm of simple autonomous mobile robots dedicated only to transfer adequate samples over long distances and (3) a spatial separation between humans and mobile robots.
It is composed of an ensemble (swarm) of small mobile robots controlled by a central intelligence and placed in a track on top of the scientific equipment (Fig. 2a), thus providing a flow separation between the ground floor dedicated to humans and scientific equipment, and the first floor dedicated to sample transfer. This system forms the Long-Range part of the combined Connection Element (LR-CE, dark gray in Fig. 2c). The flow separation is represented by the dotted line in Fig. 2c and is similar to that in the Amazon/Kiva system. The scientific equipment is grouped into clusters to form stations (Fig. 2b) where the drones can stop to drop or load samples using a localized multi-tasking connecting arm robot. The arm robots are interconnected to the drone swarm system and together form the global connecting element of the automated laboratory (CE, in gray in Fig. 2c). The arms are the Short-Range part of this global Connecting Element (SR-CE, light gray in Fig. 2c). To ensure smooth operation and perfect mutual alignment, the LR-CE and SR-CE are controlled by the same central intelligence called the Robot Subscheduler (Fig. 4, in gray). The connecting robot arms are used for local and medium-range transfers with the addition of a linear axis and, thanks to gripper changers, they are also used to perform processing tasks (PT#) on samples (e.g. pipetting, dilution) or on scientific equipment (e.g. loading samples). This is similar to what we described earlier from the evolution of the automotive industry. Fig. 2c shows the organization of this new ODAS, using the same schematic structure derived from Thurow and Fleischer3 in Fig. 1.
![]() | ||
| Fig. 2 Schematic description of the ODAS 2D drone swarm for laboratory automation, global connectivity element for the automated laboratory (CE, in gray in 2c). In (a) we present a CAD view of the actual Swiss Cat+ laboratory with the 2D drone swarm track (long-range link LR-CE, in dark gray in 2c) visible above the equipment and with the 6-axis collaborative robots (Cobots) used as short-range links (SR-CE#, in light gray in 2c) to the different instruments (e.g. LC, NMR). In (b) an image of a station with a mobile robot on the plexi/aluminum track. The red glare comes from the light source of the industrial camera used for fine alignment of the 6-axis cobot (Universal Robot, UR5e) with respect to the mobile robot (locally developed). The 6-axis arm is used to connect both the visible plate labeler (Agilent PlateLoc, bottom right) and our self-developed robotic workstation for automatic sample distribution for optical spectroscopy and NMR. In (c), the schematic description of this ODAS based on Fig. 1, in the center we find the 2D drones dedicated to long-range sample transfer (LR-CE, in dark gray), the dashed line around them symbolizes the flow separation between the mobile robots (first floor) on the one hand and the scientific equipment level on the other hand, operated either by the short-range collaborative connection elements (SR-CE, in light gray) or potentially by operators. The localized collaborative robotic arms (Cobot), coupled with industrial cameras for precise localization of the mobile robots and equipped with gripper changers, form the base of the stations. These robotic arms are also capable, thanks to gripper changers, of performing a wide variety of processing tasks (PT#) as well as loading/unloading samples directly into scientific instruments. This is facilitated by the fixed relative positioning between the arm and the instrument. They can also be connected to additional logistics tools, such as SCARA Cobots and linear axes, to extend the range of operations and connect to more workstations (WS#). | ||
As illustrated in Fig. 3a through Fig. 3c, the stations may be of disparate types and may combine disparate tools for local logistics and processing task functions. However, they must at a minimum consist of a Cobot and a camera for precise alignment of the cobot gripper with the sample carried by the mobile robots. In this concept, as in the Amazon/KIVA logistics system described above, the mobile robots are of the ASGV type.23 In particular, at Swiss Cat+, we have developed a dedicated swarm of small autonomous semi-guided vehicles (ASGVs), approximately 20 cm in diameter and 2 kg in weight (Fig. 3d). These vehicles are hereafter referred to as “drones.” Long-range trajectories are controlled by distributed Aruco codes, as described by Figueiredo et al.21 (see Progressive precision model below). At the station, high precision is required for loading and unloading the samples. For this purpose, industrial cameras positioned on the station read the Datamatrix code24,25 affixed to the drones, allowing the Robot-Subscheduler (see the Scheduling section below) to confirm the identity of the drone and associate it with all operations, and to define the drone's positioning (x, y and yaw) with submillimeter precision in order to adjust the fine movements of the six-axis Cobots. A separate publication will provide more detailed technical information about the drones and their control.
![]() | ||
| Fig. 3 (a) This is an illustration of the analytical and liquid sample preparation station, which is equipped with a six-axis Universal Robot UR3 Cobot (depicted in the upper right of the image) for the transfer of samples from the drones to the vertical storage tower. From the storage tower, a SCARA KX-2 1000 mounted on a 2.5-meter linear axis serves all the scientific equipment, including an Agilent Intuvo GC-MS equipped with PAL automation, two Agilent Bravo liquid handling stations, a microcentrifuge, a PlateLoc, and a Plate Labeler. (b) A front view of the same stations as in Fig. 2b. Here, a 6-axis UR5e is used to connect the track to an Agilent PlateLoc and to a locally developed ensemble of valves used to collect the flows from an Agilent 1290 preparative LC and distribute the samples directly to a Bruker Invenio FT-IR spectrometer, an Agilent Cary 60 UV-VIS spectrometer, and fill and seal 1. 7 mm NMR rack tubes and SBS sample plates are prepared for injection into the Agilent SFC-IM-Q-TOF (6560C). This second locally developed station, derived from an Agilent RapidFire, provides internal sample transfer via microfluidics or a SCARA Precise Automation PF3400 robotic arm. (c) A development done with Techvolver.27 The apparatus is designed to be integrated with a UR5e robot, enabling the use of manual pipettes. It is currently employed for direct sample processing and to enhance the flexibility of the stations. (d and e) Depiction of a real image and a CAD representation, respectively, of the drones utilized within the track for long-distance transfers. The drones are equipped with a wide-angle camera, which is used for locating the distributed Aruco codes. Additionally, a Datamatrix code22,23 is employed in conjunction with the industrial cameras to inform the system of the drone's identity and its precise alignment with respect to the 6-axis cobot and the SBS format sample plate.28 | ||
Each drone is controlled by a dedicated Trajectory Controller (TC) implemented in the ROS2 framework26 (Fig. 5). The task assignment to each ASGV is performed by the Fleet Manager (ROS2 framework) according to the high-level tasks received in cascade from the Laboratory Scheduler and the intermediate-level tasks received from the Robot Subscheduler.
The sampling rate of the camera to capture Aruco code positions is 5 Hz. The dynamic trajectory is then controlled by a combination of absolute position information provided by the previously described Aruco system, wheel odometry, and a drone-mounted IMU through an Extended Kalman Filter (EKF).29,30 The accuracy of the dynamic positioning is contingent upon the positioning system and several mechanical aspects, including the wheel friction factor, motor and gear types and precision, and the track material and geometry. The Swiss Cat+ system is capable of achieving a dynamic positioning accuracy of 2 centimeters in the x and y dimensions when a visual location (Aruco) is available and can reach a maximum error of up to 6 centimeters in between visual locations (located every approximately 50 centimeters), as indicated by the slashed gray zone in Fig. 4. The average x, y, and yaw positioning uncertainties are summarized below. For clarity, both the corresponding standard deviations (in meters or radians) and their variances (in m2 or rad2) are reported. The maximum values are primarily caused by an initial peak during the Kalman filter initialization; after convergence, the uncertainty stabilizes at around 6 cm in the x and y directions and approximately 0.06 rad in yaw.
| Std. Dev. x (m) | Variance x (m2) | Std. Dev. y (m) | Variance y (m2) | Std. Dev. yaw (rad) | Variance yaw (rad2) | |
|---|---|---|---|---|---|---|
| Minimum | 0.019 | 0.00036 | 0.019 | 0.00036 | 0.019 | 0.00036 |
| Maximum | 0.088 | 0.00775 | 0.078 | 0.00608 | 0.078 | 0.00613 |
| Average | 0.044 | 0.00195 | 0.043 | 0.00181 | 0.043 | 0.00182 |
This value ensures a smooth and uninterrupted trajectory, thereby facilitating secure and reliable drone navigation within the designated track. It also allows for safe and efficient crossings between different sections of the track and enables precise positioning within the industrial camera zone at the target.
Upon reaching the target station with the previously described centimeter accuracy (as indicated by the circle in Fig. 4b), the drone enters the field of view (FOV) of an industrial camera (Basler ACE 2 R a2A1920-51gcPRO with Fujinon HF8XA-5M) mounted above the track at an average height of 46 cm, corresponding to 34 cm above the Datamatrix codes. The camera mounting height is defined to provide a sufficient field of view (FOV) to encompass the entire long-range trajectory uncertainty domain (FOV = 14 cm in the case of Swiss Cat+), thus ensuring systematic capture of the Datamatrix mounted on the drone at the station. The industrial cameras are equipped with commercial image recognition software (http://www.cognex.com), which is used to identify and read the Datamatrix code printed on the top of the drones on the side of the sample plate (Fig. 3d and e). Upon confirmation by the Robot-Subscheduler that the drone has reached the target station, a single image is captured, significantly reducing the amount of data transmitted. The combination of the industrial camera and the image recognition software provides the sample position (as an offset text file transmitted via TCP-IP) to the Universal Robot 6-axis cobot controller with an average precision of 3.6 px in x and y.25 Under the Swiss Cat+ experimental conditions (Fig. 4b), the x accuracy was 0.3 mm, the y accuracy was 0.4 mm, and the yaw rotation angle was 1.63°. This level of accuracy is consistent with the nominal accuracy of the UR 6-axis Cobots (pose repeatability according to ISO 9283: ±0.03 mm31).Most of the instruments have at least 2 mm of compliance, the cumulated 0.7 mm of errors is in range to allow the SBS sample plates to be loaded and unloaded on the drones with minimal risk of failure. Once the drone loading and unloading operations have been completed as previously described, the precision of operations on the scientific instrument is directly defined by the 6-axis Cobot nominal accuracy due to the arm's location in a fixed reference with preprogrammed and optimized motions. This approach ensures the safest level of operation for the scientific equipment, thereby eliminating any risk of misalignment. Moreover, this approach is advantageous for human operators as it is straightforward to secure the Cobot with light curtains or other zone-defining devices. This strategy allows for the data transfer rate to be limited, as the Cobot motions are locally encoded into the Cobot controller and only need to be called by the Robot-Subscheduler.
![]() | ||
| Fig. 5 Adapted from Cousty et al.,22 a general schematic description of the GLAS-based global orchestration and scheduling architecture of the Swiss Cat+ West Hub laboratory. The Level 2 Robot Sub-Scheduler, which is responsible for controlling the 2D drone swarm system on a larger scale, is represented by a red, dotted, and rounded rectangle. The Level 2 Robot Sub-Scheduler receives orders from the Level 1 Main Laboratory Scheduler (REST API) and oversees the operations of the Fleet Manager. The latter is responsible for identifying and transferring tasks to the most appropriate drones based on proximity and availability. Subsequently, the trajectory control is dynamically regulated by the ROS2 trajectory controller, which is based on the adaptive precision positioning model described above. The ensemble of 6-axis cobots (Universal Robot, piloted through RTDE API) and Scara PAA KX2 cobots (dedicated API) as well as the 400 MHz NMR spectrometer (Bruker NEO 400, API co-developed with Bruker Switzerland) are also controlled directly by the level 2 Robot Sub-Scheduler to ensure a seamless interaction between the drones and the cobot used for operations. | ||
The Main Laboratory Scheduler (REST API in the Swiss Cat+ West Hub infrastructure) receives a high-level operational workflow in the format of a JSON file from a dedicated Human Computer Interface (HCI) that will be presented in a subsequent publication or from an automated generator such as ChemCrow.32 The workflow is transformed into an intermediate-level set of tasks, which are distributed to the appropriate ensemble of equipment. Each workflow instance is assigned to the relevant equipment upon arrival and executed sequentially according to a first-in–first-out (FIFO) principle, while multiple workflow instances can proceed in parallel provided they do not interfere with one another. In the context of robotic operations, all tasks pertaining to robots, including those involving drones, 6-axis robots, and SCARA robots, are forwarded to the level 2 Robot Sub-Scheduler (Fig. 5). The level 2 Robot Sub-Scheduler divides the intermediate-level tasks into detailed, pre-encoded sequences, which are then routed to the appropriate robots. It is important to note, however, that while the present work relies on such pre-encoded sequences for reliability, the architecture has been conceived with modularity in mind and could be extended in future to interface with intermediate task description languages (e.g., XML/JSON or domain-specific workflows), thereby supporting scalability and adaptation to new equipment. The tasks pertaining to the 6-axis arms and SCARA cobots, in addition to those associated with the 400 MHz spectrometer (Bruker NEO 400), are conveyed directly to dedicated application programming interfaces (APIs) for operational control. The tasks dedicated to the drones are conveyed to the level 3 Fleet Manager, which automatically assigns each task to the nearest available drone and transmits the detailed task to the dedicated Trajectory Controller (ROS2 framework). Subsequently, the Trajectory Controller directs the drone to the designated point, taking into account the avoidance of collisions with other drones, and notifies the Robot Sub-Scheduler via the Fleet Manager upon the completion of the task. The aforementioned cascade of tasks is illustrated in Fig. 6 through the example of a request for a high-performance liquid chromatography (HPLC) analysis of samples prepared in synthesis station A.
Fig. 8 depicts the various stages of the workflow described in Fig. 7, presented as images. The complete movie of the same sequence is available in the SI.
![]() | ||
| Fig. 8 Depiction of the various stages of the high-level workflow of the 2D drone swarm system operations. These stages pertain to the transfer of NMR tube racks from the filling station to the high-resolution NMR spectrometer autosampler (Bruker SampleJet), as illustrated in Fig. 7. In Steps 2 and 5.1, the operation of industrial cameras with red-light LEDs is observed. These cameras are used to capture the precise position of the drone and to inform the 6-axis UR5e of the necessary loading (Step 2) or unloading (Step 5.1) of the NMR rack for transfer from and to the drone. In Step 4, the top-mounted Aruco code utilized by the drone-mounted camera is observed, which enables the camera to orient itself during long-range transfers. The complete video of the same sequence is available as the SI. | ||
The first is to be cost-effective and easily scalable. In fact, the system is mostly open sourced (Arduino, Raspberry Pi, Open CV for video processing, ROS2 for drone control). The drones are locally 3D printed, and the track is made only of easily accessible materials (aluminum profiles and Plexiglas© plates). The cost of each drone is about 500 EUR and the whole track, for our 27 m long x 7 m large lab, is about 20 kEUR. If we consider the price of the robotic arms (about 15 to 25 kEUR, depending on the size and gripper used), the total cost for our system of 6 stations is in the range of 150 kEUR (similar to the price of a MoMA single mobile station with a robotic arm on it). Each new station would only cost the price of an additional robot arm and the time needed to reconfigure the mapping of the stations within the Robot Subscheduler. For comparison, in terms of scalability, a MoMA and our system have different trade-offs. MoMA scalability is constrained by the system's large footprint, which makes deploying multiple units impractical in smaller laboratories. However, MoMAs can scale more effectively in large facilities or when multiple distant laboratories need to be connected, where their mobility provides a clear advantage. By contrast, our system scales through modularity: increasing the number of small drones or adding extra arms at selected stations allows throughput to grow almost linearly with the number of added units, making the approach particularly well suited for medium-sized laboratories where space and cost efficiency are critical. In addition, if one of the drones fails, it can be easily removed from the circuit, replaced, and repaired by local personnel, as compared to a MoMA system which would need a specialized technician, and block the system in a small lab with only a few systems. UR robots are also easily programmed locally thanks to their user-friendly development environment and large availability of adaptable tools, grippers and open-source libraries.
Beyond cost and scalability, a second advantage of our system is its ability to be optimized. Since the number of drones can easily be increased, the long-distance transfers can easily be parallelized and continuously feed all stations without bottlenecks. Thus, the stations are optimally fed and can all run in parallel, depending essentially on the processing time at each station. This last point could also be optimized by using dynamic scheduling, as is common in industry (e.g. SyntQ by Optimal, https://optimal-tech.co.uk/), to improve the overall uptime in the lab. A third advantage, related to scenario 1 as described above, is the ease and safety of six-axis robotic operations with respect to complex and expensive scientific equipment. The fact that these robotic arms are static makes them easy to align with an accuracy of less than a millimeter, without the risk of relative position changes. In addition, the static positioning of the arm robots allows them to be surrounded and equipped with a wide variety of tools, greatly expanding their operational capabilities (e.g. at the Swiss Cat+ West Hub, one of the 6-axis UR5e robots is equipped with two different grippers, an automated screw tool, a Techvolver single pipette, and a vacuum pump).
A fourth advantage that we can mention here is the robustness of the operation, which is achieved thanks to the massive, low-cost redundancy of the drones. This makes the system resilient in terms of battery charging time of the drones and also in case of failure of a single drone. This is possible because of the open-source concept of the drones, but also because the drones are kept as simple as possible. They mainly consist of sensors (wide-angle camera, IMU), motors, microcontrollers and batteries. All the intelligence is delocalized in a central static lab server (the Robot Subscheduler), thus reducing the financial cost and energy consumption of each drone. This, in turn, helps to minimize the size of the required battery and keep it lightweight and energy efficient. Finally, it makes the system easily expendable by decoupling the management of long-distance transfers by the Fleet Manager and the Trajectory Controller from the management of local transfers thanks to cobots. Each station is a point on the map of the trajectory controller and in principle does not require any physical adaptation of the track. The local station can be fully adapted to the specific needs of the scientific equipment involved.
As briefly mentioned in Section 3, it is worth discussing the existing commercial solutions for automated laboratory logistics. The most relevant in this stage is the Formulatrix Rover (https://formulatrix.com/lab-automation/rover-autonomous-plate-handler), which is an industrial ODAS system composed of mobile robots, a centralized fleet-management system, and an overhead, ceiling-mounted track with lifts that allows mobile robots to move between stations equipped with plate storage. Each robot also includes an onboard transfer function based on a small linear actuator, enabling plate transfer to compatible storages and instruments. While this solution presents similarities to the herein described system, it is effective for laboratories that already operate with highly standardized, automation-ready or at least automation-adapted instruments. Our approach introduces several unique advantages. First, our system is open-source, lowering barriers for adoption and enabling customization by academic and industrial users alike. Second, it is designed to be modular and easily deployable in diverse laboratory environments without requiring specialized infrastructure such as static lifts, thereby reducing the overall system complexity. Third and most important, rather than focusing solely on plate transport, our platform integrates both mobile robots and robotic arms within a unified Robot Scheduler. This broader robotics ecosystem enables not only sample transfer but also direct manipulation of instruments and handling of complex tasks, including those involving instruments that are not inherently automation-ready. We have also noted that, although the Formulatrix Rover includes an on-board transfer function that is suitable for simple plate movements, it still requires integration with additional robotic arms to operate more complex instruments. Finally, we emphasize that our software architecture allows for flexible scheduling across heterogeneous non-necessarily automation-ready systems, which can even incorporate the Formulatrix Rover itself if desired.
On the side of limitations, we must first mention that a problem can occur if a drone has a failure, and even though such events are relatively rare, it can limit access to a station. To mitigate this, our system already incorporates basic failure-detection mechanisms, such as monitoring the continuity of ROS 2 communication and checking consistency between commanded motions and the drone's reported localization. Detected failures are also reported to the Robot Sub-Scheduler and the Fleet Manager so that the fleet can adapt. Additionally, we are working on a rescue robot that will automatically remove the blocking drone and plan to extend failure detection with onboard health monitoring in future work. In addition, an operator can manually resolve such issues via the Robot Scheduler interface when the rescue robot is not available or not implemented. A second point that needs to be clarified is the high level of robotics, IT and programming skills that are not necessarily available in standard labs to develop the complete planning chain for such a fully automated lab. But this is a general problem for all automated labs, and younger scientists are largely trained in coding. Also, more and more useful pieces of code are available either through GitHubs or can be generated through chatbots such as ChatGPT. We will soon make all our scheduling codes available via a free GitHub that will be announced with the next more technical release. Finally, developing such a completely integrated automation requires ideally a significant planification work. It can be added to an existing laboratory but will definitely benefit from an initial analysis and conception.
The reason for using a fixed industrial camera mounted on the track for precise positioning during sample loading and unloading, instead of reusing the wide-angle camera mounted on the drone, which would limit equipment and development needs, is the ability to establish a stable transform matrix between the industrial camera and the robot arm base coordinate system. With our software, this calibration can even be automated, making it repeatable and less error-prone. If we had used the onboard camera, we would have needed to localize fixed ArUco markers placed in the workspace and then precisely measure their distance and orientation relative to the robot base. This procedure is more cumbersome, reducing calibration stability. In addition, the fixed camera is directly connected via Ethernet and RJ45 cabling to the local network, which ensures industrial-grade communication with the robot arms. In contrast, the onboard camera communicates over Wi-Fi, which is more prone to packet loss and disconnections, reducing reliability for this crucial step. Finally, the fixed camera benefits from controlled lighting conditions, ensuring consistent and reliable vision detection. The onboard camera, however, is exposed to varying ambient lighting and reflections from the environment, which can compromise detection robustness and increase false measurements, especially critical in the final positioning phase where millimetric precision is expected.
No human participant data were collected or used in this study. Supplementary information: SI videos referenced in the article and all references listed above in Data availability. See DOI: https://doi.org/10.1039/d5dd00342c.
| This journal is © The Royal Society of Chemistry 2025 |