DOI:
10.1039/D6DD00177G
(Paper)
Digital Discovery, 2026, Advance Article
AI-designed and AI-implemented control systems for bespoke scientific instrumentation: application to scanning microscopy
Received
7th April 2026
, Accepted 7th May 2026
First published on 8th May 2026
Abstract
The pace of innovation in custom scientific instrumentation is frequently bottlenecked by the complexity of software engineering. While hardware designs evolve rapidly, developing robust and integrated control systems remains resource-intensive and often exceeds the software expertise available in experimental laboratories. Here, we present an AI-assisted workflow for constructing and validating an integrated control system for a bespoke scientific instrument, demonstrated on the Scanning Helium Microscope (SHeM). We emphasise that this work does not develop or train new AI models; instead, we use publicly accessible, general-purpose Large Language Models (LLMs) as practical engineering tools. Using these models, we co-develop a modular software stack spanning hardware interfaces, communication middleware, scan control, and user interfaces. To reduce the risk of deploying AI-generated code to fragile hardware, we introduced a digital sandbox that emulates instrument behaviour and supports pre-deployment verification, together with cross-validation using a second LLM and mandatory human review. We demonstrate successful deployment on a physical SHeM for 2D imaging and diffraction measurements, with behaviour consistent with a manually developed control system. This work provides a reproducible and safety-oriented template for AI-assisted software engineering in bespoke scientific instrumentation.
1 Introduction
Progress in experimental physics is often driven by new instruments. From the scanning tunnelling microscope1 to modern quantum sensors,2–4 custom hardware enables measurements beyond the reach of commercial tools.5 However, a persistent gap has emerged between rapid advances in hardware design and the creation of integrated control systems that operate the whole instrument with an efficient user experience. Modern instruments depend on tightly synchronised control of multiple subsystems,6,7 but building such infrastructure demands software engineering skills that are uncommon in university science laboratories.8–10 As a result, many promising instruments remain limited by fragile or incomplete control stacks.
Large Language Models (LLMs)11–13 offer a potential route to reduce this software bottleneck. They can generate functional code from natural language,13–15 lowering the barrier to instrument programming and enabling demonstrations of self-driving laboratory workflows.16–18 Recent studies have already demonstrated the power of LLMs in controlling diverse materials-science instrumentation14 and achieving joint control of commercial hardware alongside non-commercial, bespoke components.15 Beyond simple script generation, these systems have enabled natural–language interaction with microscopy and user-facility tools,14,19,20 and more recently, autonomous microscopy platforms have integrated language models with computer vision to enable content-driven measurement strategies.21 These developments show substantial promise, but practical challenges remain for bespoke instruments: most reports focus on a single device, isolated scripts, or operation through mature APIs, whereas customised systems often require a control system spanning hardware interfaces, communication, scan logic, and user-facing tools. Safety concerns further arise when deploying stochastic AI outputs directly to hardware control,22,23 particularly when synchronisation or scan-logic errors can damage hardware or silently corrupt datasets.
In this work, we present an AI-assisted workflow for developing and deploying an integrated control stack for a real multi-subsystem scientific instrument, the Scanning Helium Microscope (SHeM).24–28 Our contribution is not in demonstrating new capabilities of LLMs themselves, but in showing how existing, general-purpose models can be used as practical engineering tools to rapidly construct complete control systems for bespoke instruments. We do not claim the first use of LLMs in instrument control; rather, our contribution is a safety-oriented and reproducible engineering approach for translating experimental intent into a working control system for bespoke hardware. For clarity, we define the “integrated control system” in this context as the combination of: hardware control for motion and detection subsystems, a communication layer for commands and data streaming, scan control logic (trajectory generation, position verification, and dwell handling), and user interfaces for operation and monitoring. To mitigate risk, we develop a digital sandbox that emulates instrument-like responses and streams synthetic data through the same interface, enabling validation of AI-generated code prior to deployment. We further employ cross-validation using an independent LLM as a review channel with final human-verified acceptance, and we validate the resulting system on a physical SHeM with experimental data showing performance comparable to a manually engineered control system.
2 Methods
2.1 Instrumentation
The SHeM is a novel type of microscope that utilises charge-neutral thermal helium atoms as probing particles. A typical SHeM is composed of three main parts: beam generation, sample manipulation, and signal detection. Beam generation is based on a free jet expansion27,29 followed by micro-collimation and in current SHeM implementations, no software control is required. The helium microprobe is incident on the sample, which is scanned in order to create images. Sample manipulation is performed by stacking commercially available piezoelectric motors to achieve motions in X, Y, Z, and azimuthal rotation. Signal detection, used to give the intensity in each pixel, is achieved by collecting the helium beam scattered in a given direction, then ionising it in a home-made detector and reading the current using a picoammeter. SHeM is capable of generating images over a large lateral scale with sub-micrometre resolution30,31 and performing spot-profile diffraction;32,33 it is designed for performing high precision imaging without any beam damage or invasive sample preparation, while collecting local surface information through spot-profile diffraction. These functions require a complex, tightly synchronised control system.
2.2 AI-assisited development route
To facilitate the development of the SHeM control system, we utilised two frontier Large Language Models (LLMs): ChatGPT-5 (OpenAI) and Claude Opus 4.1 (Anthropic). Both models were accessed via their respective official web interfaces. While a formal quantitative evaluation of LLM performance is beyond the scope of this study, our development experience revealed complementary strengths: ChatGPT-5 was found to be particularly effective during the initial conceptual brainstorming and architectural design phases, providing structured and insightful high-level frameworks. In contrast, Claude Opus 4.1 demonstrated higher proficiency in generating functional, low-level source code and implementing specific hardware communication logic. By leveraging these distinct capabilities, we established a robust workflow from conceptualisation to deployment.
2.2.1 Conceptual system architecture Co-designed with AI. The workflow for utilising AI in the development of the control system for the Scanning Helium Microscope (SHeM) is shown in Fig. 1 as a demonstration of the methodology. At the beginning of the project, the overall concept of modularising the control system and separating the development process into distinct steps was established through iterative interactions with the AI. The first prompt illustrates how we engage the AI for conceptual design. We provide the essential background information and the overall aim of the project, followed by an open request for ideas. The outcome of this conceptual design from the AI is a division of the control system into four major modules: the hardware control layer, communication layer, scan control layer, and user interface layer. These modules form a logical progression from low-level hardware interaction through to the front-end user interface.
 |
| | Fig. 1 A schematic figure illustrating the workflow of human–AI interactions in developing the control system for the scanning helium microscope. The modular structure of the system and the separation of its functional components were refined iteratively through discussions with the AI. The initial prompt demonstrates the strategy used to engage the AI for conceptual development. The full control system is separated into four main modules, where the specific prompts used to generate each module are also shown. All prompts are summarised from many iterations with the AI at that stage to demonstrate the idea. | |
2.2.2 Hardware control layer generation. In the second prompt, with the conceptual design, we ask the AI to develop code for sample manipulation (via the Attocube nanopositioners ECC100 controller) and for signal detection (with the RBD 9103 picoammeter reading currents from the home-made helium detector) by stating the function we want to achieve. The manufacturer-provided API libraries are supplied to the AI during this stage. We also specify that synchronous functions should not be implemented at this point, as AI tends to build up future functions based on the conceptual design. The outcome is two separate pieces of code for controlling the nanopositioner and the picoammeter.
2.2.3 Communication layer generation. The third prompt focuses on the development of the communication layer. We give the AI the code for the hardware control it generated and the conceptual design, which serves as a base for the communication layer. In this design, the code generated in the hardware control layer is modified so that both the picoammeter and the nanopositioner encoders publish their signals and position data with timestamps under separate topics on a shared MQTT (Message Queuing Telemetry Transport, a lightweight, publish/subscribe network protocol designed for machine-to-machine communication) server. The server also accepts commands issued from the subsequent scan control layer.
2.2.4 Scan control layer generation. The fourth prompt describes the interaction with the AI used to develop the most central module of the entire system: the scan control layer. Similar to the conceptual design stage in Prompt 1, this module required extensive iteration with the AI to create a framework suitable for future instrument development. The outcome of the conceptual design in this specific layer is having four components: MQTT command functions, which issue movement instructions to the hardware through an MQTT communication layer; scan pattern generators, which compute coordinates for each pixel in the scan; position verification and dwell control functions, which check whether the nanopositioner has reached the target position, manage dwell times, and extract a data point from the continuously published picoammeter signal while accounting for detector lag; a high level scan controller, which integrates all previous components and executes the complete scanning sequence. Fig. 1 demonstrates several scan patterns, including “snake” and “raster” trajectories in both rectangular and polygonal geometries. It also shows the spot-profile diffraction mode, in which the sample is moved in X and Z to maintain a constant beam incidence, which allows the SHeM to vary the detection angle.32 The outcome of this stage is a group of codes that fulfil our requirement for controlling the scan. However, since this layer contains complex geometric transforms and synchronisation logic, where even small computational inaccuracies risk hardware collisions or measurement artefacts, which in our experience are not capability strengths for LLM AI models.13,34,35 Thus, further verification is required, which is presented below.
2.2.5 User interface layer generation. Finally, the fifth prompt concerns cross-platform system compatibility. Since the underlying modules are written in C++ and Python, which are widely supported across operating systems, both a graphical user interface (GUI) and a command line interface (CLI) can be implemented as frontend clients that access the same underlying functionality. Most laboratory scientists do not have the ability to develop a good user interface10 as it is usually not part of their expertise. However, AI is good at building such interfaces based on its knowledge from the front-end coding community.13,36,37 For the SHeM system, the AI helped transition our operational workflow from a series of fragmented and command-line-based scripts to a single dashboard, as shown in Fig. 2. This integrated interface organises the complex control stack into four logical regions: connectivity and system metadata, motion control gallery, live display, and dynamic data visualisation.
 |
| | Fig. 2 A screenshot of the GUI. Region 1: connectivity and metadata settings. manages MQTT client protocols, file storage paths, and detector lag compensation parameters. Region 2: integrated scan control. Configures manual axis positioning, 1D/2D scan geometries, and automated multi-dimensional series logic, including a global emergency stop mechanism. Region 3: live telemetry display. Provides real-time instrument status, including axis coordinates and signal intensities. Region 4: live display. Features live rendering of 2D images, 1D line plots, and scan path previews to monitor experiment progress. Details of the GUI can be found in the data and materials availability section. | |
2.3 Sandbox simulation environment
Given the high cost and fragility of the nanopositioners and custom detector, direct deployment of unverified AI-generated code onto physical hardware is risky. As a mitigation, we employed a “sandbox” approach. We prompted the AI to develop a digital sandbox, as shown in Fig. 3, a simulation environment designed to mimic the physical instrument's behaviour. The simulator connects to the same MQTT server as the real hardware, subscribing to movement commands and publishing positions and signals based on synthetic data. It is important to clarify that the sandbox was not intended to function as a high-fidelity physical simulator. Instead, its primary purpose is to isolate and verify key aspects of the control system, including MQTT communication integrity, scan sequencing, coordinate handling, and user interface responsiveness. As such, the scope of the sandbox is intentionally limited to logical validation under nominal operating conditions. This method allowed us to validate the integrity of the MQTT communication, the accuracy of the scan control, and the responsiveness of the user interface in a virtual environment before physical implementation. To reduce the risk of correlated errors between the control code and the simulation environment, the fundamental rules governing the sandbox were explicitly defined by the authors based on the known geometry and operation of the SHeM instrument. These include coordinate transformations within a global reference frame, the definition of a centre of rotation (COR) for azimuthal motion, and the geometric coupling between Z translation and lateral displacement. The role of the AI was restricted to implementing these predefined constraints, rather than generating them autonomously. In addition, the implemented transformations were cross-checked against analytical expectations and prior experimental observations.
 |
| | Fig. 3 A schematic figure of the sandbox used to simulate the behaviour of hardware in the real SHeM. A synthetic digital image is used as the sample, which can be positioned at any location, translated, rotated, and interpolated by settings to mimic real sample behaviour. The output is in the same format as the real hardware to test the robustness of other parts of the control system. | |
As illustrated in Fig. 3, the sandbox operates on a digital image that serves as the virtual sample surface. The image is assigned an initial position within a global reference frame, and all pixels outside this region are treated as background, producing zero intensity for simplicity. The sample can then be translated, rotated, and interpolated in three dimensions, where each axis has a speed setting for testing the on-position check logic in the scan control layer. After constructing the sample plane, the sandbox defines a probe point that reads both the positions and the corresponding intensity value at that location. The intensity will be multiplied by a gain value with an offset, which can be randomly generated for each reading to mimic the poisson background noise in a real detector. These data are streaming using MQTT in the same format used by the physical hardware. The current implementation does not include higher-order physical effects such as mechanical hysteresis, thermal drift, or real-time communication latency. While these factors are important for quantitative performance prediction and robustness testing, they are not required for the present objective of verifying logical correctness.
Human supervision remains essential throughout development. Logical inconsistencies in coordinate transformations, timing control, and data handling are identified through iterative inspection and refinement. Within these constraints, the sandbox provides a controlled and safe environment for validating the control architecture before deployment on physical hardware.
2.4 Cross-validation with multiple AI models
Alongside the “sandbox” approach, we employed a cross-validation workflow using two independent LLMs mentioned above (ChatGPT-5 and Claude Opus 4.1) since different training datasets reduce correlated hallucination errors.38,39 This process is illustrated in Fig. 4. For each module of the control system, initial code or design suggestions were generated by one model and subsequently reviewed by the second model. This comparison was useful, as these two models frequently produced different interpretations of the same prompt, enabling the detection of hallucinations, errors, or bugs generated by a single model. Discrepancies between models were resolved by a human reviewer: we adopted the smallest modification that satisfied the stated interface constraints and passed sandbox validation. In practice, many cross-model disagreements were traceable to missing assumptions in the prompt; therefore, resolution frequently involved tightening the prompt with explicit constraints and rerunning a scoped generation step. No change affecting scan logic was permitted to proceed to supervised hardware operation unless it passed sandbox verification.
 |
| | Fig. 4 The schematic illustrates the workflow for cross-validating outputs from two AI models. Examples of unwanted behaviours encountered during the development of the SHeM control system are shown, along with a summary of the prompts used throughout the cross-validation process. | |
During development, in addition to the syntax errors that are easily detected by the coding environment, many other issues were identified through a combination of cross-validation between AI models and human review of sandbox behaviour. The sandbox played a critical role in revealing errors introduced by AI during code generation, including hallucinations, logic errors, and mathematical errors, as illustrated in Fig. 4. Hallucinations such as generating additional entities not requested or referencing nonexistent functions were typically detected through standard syntax checks, whereas most logical and mathematical errors in the scan control layer emerged only when the code was exercised in the sandbox and evaluated by a human reviewer.
Although cross-validation and sandbox testing mitigated the majority of risks, AI-generated code remains unvalidated under certain conditions. In particular, direct interaction with hardware without an intermediate verification layer is hazardous, and human inspection of every script remains essential. Silent logical inconsistencies, such as misaligned device timestamps or incorrect synchronisation between independent hardware components, may still occur and cannot be reliably detected by AI cross-validation. For these reasons, all AI-generated code was subjected to manual inspection and offline testing. No model was permitted to issue direct hardware commands without first passing sandbox verification.
2.5 LLM interaction protocol and prompt engineering
We first summarise the practical protocol used during AI-assisted development. The control system was developed in an iterative prompt–review–test cycle. Each prompt included: the objective and required behaviours, the relevant API constraints and available commands, the required input/output signatures, and explicit exclusions (e.g. not generate placeholder code for future development). Generated code was first subjected to static checks (syntax, imports, functions consistency), then executed in the digital sandbox where applicable, and finally reviewed by a human for logic and safety before any supervised hardware deployment.
During extended interactions, we observed recurring failure modes: hallucinated functions or entities not present in the provided API, logical errors in coordinate transforms and scan sequencing, mathematical inconsistencies in scan geometry or dwell timing, and regression in code quality, such as duplicated implementations and overcomplicated abstractions. As an estimate, these failure modes were encountered routinely during early iterations (e.g., frequently in initial generations for complex scan logic), but were effectively eliminated through the structured workflow described. We mitigated these issues by enforcing a stable module interface, keeping prompts and code changes scoped to a single module per iteration, and using sandbox execution as a verification for scan-logic changes. To concretely illustrate these limitations and provide a transition to our proposed workflow, Table 1 demonstrates how a common logic error was identified and resolved. Additional representative prompt–response excerpts and correction cycles are provided in the SI.
Table 1 Representative correction cycle: improving position verification logic using sandbox validation
| Interaction stage |
Content/code snippet |
| Initial user prompt |
“I need a position check function to ensure the stage has reached the target before acquiring data.” |
| LLM response |
While current_position ! = target_position |
| Pass |
| Analysis: the equality check assumes exact numerical matching. In practice, due to noise and finite resolution, the stage position rarely matches the target exactly, causing the loop to stall indefinitely |
| Sandbox check and human feedback |
In the sandbox (with artificial noise added to the position readings), the stage approaches the target but never reaches an exact match, causing the scan to stall. The AI is asked to fix this issue by adding tolerance to the position verification |
| LLM response |
Tolerance = 5 × 10−9 # 5 nm |
| While abs (current_position – target_position) > tolerance |
| Pass |
| Outcome: the revised condition allows reliable position verification under realistic noise and resolution limits. Sandbox validation confirms stable scan progression without stalling |
2.6 Maintainability and extendability
The control system may be required to change over time in order to accommodate future developments of the instrument. Beyond operational reliability, it is therefore important that the control system remains easy for a human or another AI model to understand and modify. It was found that the high-level design led to a clear conceptual separation of concerns between low-level communication, control logic, and the user interface. This logical modularity meant that new changes often only required modification of a single component's logic, and individual code segments were generally very clear and accompanied by comprehensive comments. However, to prioritise immediate operational capability and rapidly produce scientific results, the final deployment was consolidated into a single Python script.
While this single-file approach successfully drove the experimental hardware and facilitated data acquisition, it significantly increased the cost and difficulty of long-term maintenance. Housing the entire codebase in one monolithic file makes navigation, version control, and isolated testing cumbersome for a human developer. Furthermore, the model had a tendency to rely on unstructured data types to pass information between the internal layers of the program. This meant future development would rely heavily on comments to interpret data flow, making it almost impossible for automated tooling, such as type checkers, to identify simple bugs. When tasked with implementing additional features within this single-file structure, the model frequently prioritised the shortest path to functional correctness, often at the expense of engineering best practices. This approach led to the gradual accumulation of technical debt, specifically through code duplication and the creation of overly complex functions with excessive arguments. Future maintenance, such as fixing a bug in the scanning logic, would require tracking changes across multiple repeated sections of the large file. Consequently, while the current architecture was sufficient for proof-of-concept experimental work, adopting a strictly multi-file architecture is highly recommended for future iterations to physically enforce modularity and reduce maintenance costs.
Despite these architectural trade-offs, we found that many of the AI's structural issues were possible to fix with additional prompting, and they were often detectable through the application of standard code linters. All AI-generated code operates entirely offline in the instrument environment, ensuring that no experimental data or hardware access is transmitted to external servers. Manual inspection and sandbox-based verification mitigate the risk of unintended vulnerabilities or hidden dependencies.
3 Results
With the full control system validated in the sandbox, we deployed the AI-generated system on the physical SHeM to assess its performance during real measurements. In Fig. 5, to provide a clear benchmark, we compared datasets across three scenarios: in Fig. 5(a), a reference 2D image and diffraction measurement acquired from a synthetic digital sample in the sandbox, followed by experimental MoS2 and biofilm datasets obtained using the AI-generated control system; in Fig. 5(b), corresponding experimental datasets acquired using the original manually written control software. These comparisons allow us to evaluate the control system made by the AI in a real experimental environment. In the left panel of Fig. 5(a), the sandbox results confirm the expected behaviour of the scan controller. The image reproduces the input digital sample, and the diffraction measurement yields the expected straight line, demonstrating that the system successfully keeps the probe fixed on the same pixel throughout the full Z range.
 |
| | Fig. 5 A schematic comparison of results obtained using the AI-generated control system and a manually developed control system is shown. (a) From left to right, a digital image is used as the input sample in the sandbox, where both the 2D scan and the Z-dependent diffraction scan exhibit the expected behaviour. The straight diffraction line confirms that the scan controller maintains the probe on the same pixel across the full Z range, as required for diffraction measurements. In the middle panel, a real MoS2 sample is imaged using the AI-generated control system, producing both a spatial image and a diffraction profile that closely match those obtained using the manually developed control code. The right panel shows an image of a biofilm sample acquired using the SHeM at a resolution of 250 nm, representing operation under optimal instrument conditions. (b) For comparison, the left panel shows MoS2 data from ref. 40, which exhibits similar contrast and diffraction characteristics to those obtained using the AI-generated control system. The right panel shows biofilm data from ref. 41 (reproduced with permission), also acquired at a resolution of 250 nm. There is no degradation in performance relative to manual control that was observed within experimental uncertainty. | |
When the AI-generated control system is deployed on the physical SHeM, the resulting spatial image and diffraction measurement from a MoS2 flake on a SiO2 substrate (middle panel of Fig. 5(a)) exhibit the same behaviour observed in data acquired using the manually written control system on a separate sample of MoS2 flakes on hBN/SiO2 (left panel of Fig. 5(b).40 This agreement demonstrates that the AI-generated scan logic functions correctly on real hardware. Both MoS2 spatial images were acquired with identical settings (2.5 µm per pixel resolution and a 1 s dwell time), while both diffraction measurements used a 3 s dwell per point and a 50 µm step size.
To further evaluate the system under optimal instrument conditions, a biofilm sample was imaged at a higher resolution of 250 nm using the AI-generated control system (right panel of Fig. 5(a)). Comparing this to biofilm data previously acquired with the manual control system at the same resolution (right panel of Fig. 5(b)),41 no degradation in performance relative to the manual control was observed within experimental uncertainty. Across all tested samples, the overall image quality and diffraction signatures remain consistent between the two control systems. Minor variations in noise levels or fine-scale contrast are attributed to detector fluctuations and sample differences rather than to the control logic itself.
The successful deployment of this system highlights a shift in experimental engineering: this exemplifies a transition from manual coding to AI-assisted development and validation. While the image quality is equivalent to the manual implementation, the disparity in development velocity is profound. The complete development cycle, from conceptual design to experimental deployment, was compressed from a timeline of many months for the manually written control system to approximately three weeks. Crucially, this acceleration did not come at the cost of reliability. The sandbox ensured that errors, AI hallucinations, and bugs were resolved within the sandbox environment, resulting in fault-free operation during physical deployment. This confirms that the combination of generative AI for code production and sandbox verification constitutes a pipeline for overcoming the software bottleneck in custom instrumentation.
4 Conclusion and outlook
We have presented an end-to-end, AI-generated control system for the bespoke instrument, using the Scanning Helium Microscope as a demonstration of the methodology. Through careful direction of Large Language Models, we successfully integrated different hardware subsystems into a single control system. The sandbox simulation environment bridged the gap between the stochastic nature of generative AI and the operational concerns of using AI-generated code in real hardware. Because the methodology is not tied to any specific hardware or probe physics, this work provides a pathway for rapidly transforming bespoke laboratory concepts into operational instruments across many scientific domains.
Looking forward, this modular control system enables extension toward autonomous instrument operation, as the control stack is software-defined and API-driven, which is suitable for machine learning. Future work will focus on integrating reinforcement learning algorithms into the scan control layer to achieve autonomous calibration and image recognition to select regions for diffraction measurements. Overall, the AI-assisted software engineering significantly reduces the bottleneck in experimental science, freeing researchers to focus on exploration and innovation in their scientific field rather than concentrating on software development and maintenance.
Author contributions
Ke Wang: writing – original draft, writing – review & editing, conceptualisation, methodology, software, validation, formal analysis, investigation, resources, data curation, visualisation, project administration. David J. Ward: writing – review & editing, conceptualisation, methodology, software, project administration. Matthew Ord: writing – review & editing. Boyao Liu: writing – review & editing. Andrew P. Jardine: writing – review & editing, conceptualization, project administration.
Conflicts of interest
The authors declare no competing interests.
Data availability
The “sandbox” simulator and the scanning control system can be found on: https://github.com/Okarl96/shem_gui and https://doi.org/10.17863/CAM.129966 (Apollo—University of Cambridge Repository). A minimal working example, including example input data and configuration files, is provided in the repository to enable reproduction of the sandbox validation workflow.
Supplementary information (SI): summarised prompt examples. See DOI: https://doi.org/10.1039/d6dd00177g.
Acknowledgements
Ke Wang acknowledges financial support from the Chinese Scholarship Council (CSC) and the Cambridge Trust. This work was supported by EPSRC (Grant No. EP/R008272/1) and EPSRC (Award No. EP/T00634X/1). The work was performed in part at CORDE, the Collaborative R&D Environment established to provide access to physics-related facilities at the Cavendish Laboratory, University of Cambridge.
Notes and references
- G. Binnig, H. Rohrer, C. Gerber and E. Weibel, Surface Studies by Scanning Tunneling Microscopy, Phys. Rev. Lett., 1982, 49, 57–61, DOI:10.1103/PhysRevLett.49.57.
- M. J. Molaei, A review on nanostructured carbon quantum dots and their applications in biotechnology, sensors, and chemiluminescence, TALANTA, 2019, 196, 456–478, DOI:10.1016/j.talanta.2018.12.042.
- C. T. Matea, et al., Quantum dots in imaging, drug delivery and sensor applications, Int. J. Nanomed., 2017, 12, 5421–5431, DOI:10.2147/IJN.S138624.
- N. P. de Leon, et al., Materials challenges and opportunities for quantum computing hardware, SCIENCE, 2021, 372(6539), 253+, DOI:10.1126/science.abb2823.
- J. M. Pearce, Building Research Equipment with Free, Open-Source Hardware, SCIENCE, 2012, 337(6100), 1303–1304, DOI:10.1126/science.1228183.
- P. T. Starkey, et al., A scripted control system for autonomous hardware-timed experiments, Rev. Sci. Instrum., 2013, 84(8), 085111, DOI:10.1063/1.4817213.
- V. Chandrasekhar and M. M. Mehta, RTSPM: Real-time Linux control software for scanning probe microscopy, Rev. Sci. Instrum., 2013, 84(1), 013705, DOI:10.1063/1.4775717.
- N. A. Ernst, J. Klein, M. Bartolini, J. Coles and N. Rees, Architecting complex, long-lived scientific software, J. Syst. Software, 2023, 204(C) DOI:10.1016/j.jss.2023.111732.
- D. S. Katz, et al., Software Sustainability & High Energy Physics, Tech. Rep., 2020 DOI:10.5281/ZENODO.4082137.
- G. Wilson, et al., Best Practices for Scientific Computing, PLoS Biol., 2014, 12(1), 1–7, DOI:10.1371/journal.pbio.1001745.
- K. G. Barman, et al., Large physics models: towards a collaborative approach with large language models and foundation models, Eur. Phys. J. C, 2025, 85(9), 1066, DOI:10.1140/epjc/s10052-025-14707-8.
- H. Pan, et al., Quantum many-body physics calculations with large language models, Commun. Phys., 2025, 8(1), 49, DOI:10.1038/s42005-025-01956-y.
- I. M. M. M. Stefano Bistarelli, Marco Fiore, Usage of Large Language Model for Code Generation Tasks: A Review, SN Comput. Sci. c, 2025, 6(6), 673, DOI:10.1007/s42979-025-04241-5.
- D. Febba, K. Egbo, W. A. Callahan and A. Zakutayev, From text to test: AI-generated control software for materials science instruments, Digit. Discov., 2025, 4(1), 35–45, 10.1039/d4dd00143e.
- Y. Xie, K. He and A. Castellanos-Gomez, Toward Full Autonomous Laboratory Instrumentation Control with Large Language Models, Small Struct., 2025, 6(8), 2500173, DOI:10.1002/sstr.202500173.
- F. Hase, L. M. Roch and A. Aspuru-Guzik, Next-Generation Experimentation with Self-Driving Laboratories, Trends Chem., 2019, 1(3), 282–291, DOI:10.1016/j.trechm.2019.02.007.
- J. R. Kitchin, The evolving role of programming and LLMs in the development of self-driving laboratories, APL Mach. Learn., 2025, 3(2), 026111, DOI:10.1063/5.0266757.
- Y. Liu, et al., Building an affordable self-driving lab: Practical machine learning experiments for physics education using Internet-of-Things, APL Mach. Learn., 2025, 3(4), 046105, DOI:10.1063/5.0283529.
- S. Mathur, N. van der Vleuten, K. G. Yager and E. H. R. Tsai, VISION: a modular AI assistant for natural human-instrument interaction at scientific user facilities, Mach. Learn.: Sci. Technol., 2025, 6(2), 025051, DOI:10.1088/2632-2153/add9e4.
- Y. Liu, M. Checa and R. K. Vasudevan, Synergizing human expertise and AI efficiency with language model for microscopy operation and automated experiment design, Mach. Learn.: Sci. Technol., 2024, 5(2), 02LT01, DOI:10.1088/2632-2153/ad52e9.
- J. Yang, et al., Zero-Shot Autonomous Microscopy for Scalable and Intelligent Characterization of 2D Materials, ACS Nano, 2025, 19(40), 35493–35502, DOI:10.1021/acsnano.5c09057.
- E. Ginzburg-Ganz, et al., Statistical Foundations of Generative AI for Optimal Control Problems in Power Systems: Comprehensive Review and Future Directions, Energies, 2025, 18(10), 2461, DOI:10.3390/en18102461.
- G. Recupito, et al., Technical debt in AI-enabled systems: On the prevalence, severity, impact, and management strategies for code and architecture, J. Syst. Software, 2024, 216 DOI:10.1016/j.jss.2024.112151.
-
(a) D. Maclaren, B. Holst, D. Riley and W. Allison, Focusing elements and design considerations for a scanning helium microscope (SHeM), Surf. Rev. Lett., 2003, 10(2–3), 249–255 CrossRef CAS;
(b) 7th International Conference on the Structure of Surfaces, Newcastle, Australia, 2002, 21–26, DOI:10.1142/S0218625X03005062.
-
(a) M. Barr, et al., A design for a pinhole scanning helium microscope, Nucl. Instrum. Methods Phys. Res. Sect. B Beam Interact. Mater. Atoms, 2014, 340, 76–80 CrossRef CAS;
(b) 20th International Workshop on Inelastic Ion-Surface Collisions (IISC), Australia, 2014, 16–21, DOI:10.1016/j.nimb.2014.06.028.
- M. Bergin, D. J. Ward, J. Ellis and A. P. Jardine, A method for constrained optimisation of the design of a scanning helium microscope, Ultramicroscopy, 2019, 207, 112833, DOI:10.1016/j.ultramic.2019.112833.
- C. Zhao, et al., A multi-detector neutral helium atom microscope, Vacuum, 2025, 234, 114006, DOI:10.1016/j.vacuum.2024.114006.
- M. Barr, et al., Unlocking new contrast in a scanning helium microscope, Nat. Commun., 2016, 7, 10189, DOI:10.1038/ncomms10189.
- M. Barr, K. M. O’Donnell, A. Fahy, W. Allison and P. C. Dastoor, A desktop supersonic free-jet beam source for a scanning helium microscope (SHeM), Meas. Sci. Technol., 2012, 23(10), 105901, DOI:10.1088/0957-0233/23/10/105901.
- S. M. Lambrick, et al., True-to-size surface mapping with neutral helium atoms, Phys. Rev. A, 2021, 103(5), 053315, DOI:10.1103/PhysRevA.103.053315.
- M. Bergin, et al., Standardizing resolution definition in scanning helium microscopy, Ultramicroscopy, 2022, 233, 113453, DOI:10.1016/j.ultramic.2021.113453.
- N. A. von Jeinsen, et al., 2D Helium Atom Diffraction from a Microscopic Spot, Phys. Rev. Lett., 2023, 131(23), 236202, DOI:10.1103/PhysRevLett.131.236202.
- M. Bergin, et al., Observation of diffraction contrast in scanning helium microscopy, Sci. Rep., 2020, 10(1), 2053, DOI:10.1038/s41598-020-58704-1.
- S. Frieder and et al., Mathematical Capabilities of ChatGPT Advances in Neural Information Processing Systems 37th Conference on Neural Information Processing Systems (NeurIPS), in Advances In Neural Information Processing Systems 36 (Neurips 2023), ed. A. Oh and et al., New Orleans, LA, 2023 Search PubMed.
- J. C. Dunlap, R. Sissons and R. Widenhorn, Descending an inclined plane with a large language model, Phys. Rev. ST Phys. Educ. Res., 2025, 21(1), 010153, DOI:10.1103/PhysRevPhysEducRes.21.010153.
- K. Moran, C. Bernal-Cárdenas, M. Curcio, R. Bonett and D. Poshyvanyk, Machine Learning-Based Prototyping of Graphical User Interfaces for Mobile Apps, IEEE Trans. Software Eng., 2020, 46(2), 196–221, DOI:10.1109/TSE.2018.2844788.
- P. Duan, J. Warner, Y. Li and B. Hartmann, Generating Automatic Feedback on UI Mockups with Large Language Models, in Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems, CHI ’24 Association for Computing Machinery, New York, NY, USA 2024, DOI:10.1145/3613904.3642782.
- M. Chelli, et al., Hallucination Rates and Reference Accuracy of ChatGPT and Bard for Systematic Reviews: Comparative Analysis, J. Med. Internet Res., 2024, 26, e53164, DOI:10.2196/53164.
- I. D. Mienye and T. G. Swart, Ensemble Large Language Models: A Survey, Information, 2025, 16(8), 688, DOI:10.3390/info16080688.
- N. von Jeinsen, et al., Helium atom micro-diffraction as a characterisation tool for 2D materials, arXiv, 2024, preprint arXiv:2409.20461, DOI:10.48550/arXiv.2409.20461.
- N. A. von Jeinsen, et al., Surface visualisation of bacterial biofilms using neutral atom microscopy, J. Microsc., 2026, 301(1), 107–115, DOI:10.1111/jmi.70038.
|
| This journal is © The Royal Society of Chemistry 2026 |
Click here to see how this site uses Cookies. View our privacy policy here.