DOI:
10.1039/D5RA09726F
(Paper)
RSC Adv., 2026,
16, 11306-11313
Large language models in materials science: assessing RAG evaluation frameworks through graphene synthesis
Received
16th December 2025
, Accepted 4th February 2026
First published on 27th February 2026
Abstract
Retrieval-Augmented Generation (RAG) systems increasingly support scientific research, yet evaluating their performance in specialized domains remains challenging due to the technical complexity and precision requirements of scientific knowledge. This study presents the first systematic analysis of automated evaluation frameworks for scientific RAG systems, using graphene synthesis in materials science as a representative case study. We develop a comprehensive evaluation protocol comparing four assessment approaches: RAGAS (an automated RAG evaluation framework), BERTScore, LLM-as-a-Judge, and expert human evaluation across 20 domain-specific questions. Our analysis of automated evaluators reveals that BERTScore lacks the interpretability and score sensitivity required to distinguish meaningful performance difference, while LLM-as-a-Judge failed to capture retrieval augmentation benefits. In contrast, RAGAS successfully captured relative performance improvements from retrieval augmentation, identifying performance gains in RAG-augmented systems (0.52-point improvement for Gemini, 1.03-point for Qwen on a 10-point scale), and demonstrating particular sensitivity to retrieval benefits in smaller, open-source models. However, it still exhibits fundamental limitations in absolute score interpretation for scientific content. These findings establish methodological guidelines for scientific RAG evaluation and highlight critical considerations for researchers deploying AI systems in specialized domains.
Introduction
The integration of Large Language Models (LLMs) into scientific research workflows has accelerated rapidly, yet their evaluation in specialized domains remains methodologically underdeveloped. While general-purpose LLMs like GPT-3.5 demonstrate broad knowledge synthesis capabilities,1,2 their performance in technical fields like materials science presents unique evaluation challenges that existing frameworks inadequately address.
Retrieval-Augmented Generation (RAG)3 systems offer a promising solution to enhance LLM performance in scientific domains by incorporating domain-specific literature in real-time. Applications in fields such as education4,5 and medicine6 have demonstrated their potential for domain-specific tasks, while specialized models such as LLaMP,7 PaperQA8 and PaperQA2 (ref. 9) highlight their promise in scientific applications. Despite this progress, evaluation remains a major challenge, requiring frameworks capable of assessing not only factual accuracy but also the quality of scientific reasoning, context utilization, and domain-specific knowledge integration. Current evaluation approaches vary widely,10–15 with no established methodology for scientific applications where accuracy and precision are critical.
This gap is especially critical in materials science, where accurate knowledge synthesis can directly influence experimental design and research outcomes.16 Benchmarks such as GPQA17 and MaScQA18 have been developed to evaluate LLMs in the scientific domain, but they primarily rely on structured question formats. For example, the MaScQA dataset includes multiple-choice, matching, predefined numerical, and open numerical questions. While this structure ensures coverage across subfields, it falls short in evaluating open-ended questions. The significance of this shortcoming has been demonstrated in prior work,19 where an LLM chose the correct multiple-choice answer but produced flawed reasoning when asked to justify it, highlighting the risk of hallucinated reasoning in open-ended use.
Recent evaluation frameworks, such as RAGAS,20 propose comprehensive assessment of RAG systems through multiple metrics including factual correctness, context recall, and faithfulness. In contrast to MaScQA's structured formats, these approaches typically evaluate open-ended responses using an evaluator LLM.21 While promising,22,23 these frameworks have not been systematically validated in scientific contexts, where the nature of knowledge and the consequences of errors differ significantly from general applications.
This study addresses this gap by providing the first systematic evaluation of automated RAG assessment frameworks in a scientific domain. Using graphene synthesis as a representative case study, we investigate how well automated evaluation methods capture the performance characteristics essential for scientific applications. We evaluate four response modes, consisting of two RAG-LLM systems (Qwen and Gemini) and their baseline LLMs, using four evaluators: RAGAS, BERTScore, an LLM judge, and a panel of subject matter experts (Fig. 1). Our analysis reveals both the capabilities and fundamental limitations of current evaluation approaches, providing methodological guidance for researchers deploying RAG systems in specialized domains such as self-driving labs.24,25 The implications might extend beyond materials science to any technical field requiring precise knowledge synthesis, offering insights into the broader challenge of evaluating AI systems in specialized domains where accuracy and reliability are paramount.
 |
| | Fig. 1 Overview of the evaluation workflow for assessing the performance of RAG-augmented and baseline LLMs. The figure shows the generation of answers from four LLM configurations (RAG-Qwen, RAG-Gemini, standard Qwen, and standard Gemini), and their evaluation using four methods: RAGAS, BERTScore, LLM judge, and a panel of subject matter experts. Human annotations serve as the benchmark for comparing the reliability and alignment of the automated evaluation metrics, with a focus on assessing the efficacy of RAGAS as a robust evaluation tool for RAG-LLMs. | |
Experimental procedures
Database preparation and RAG pipeline construction
We developed a domain-specific RAG pipeline by curating 300 peer-reviewed papers on graphene synthesis from leading scientific publishers including ScienceDirect, American Chemical Society Publications, and Springer Nature. Papers were selected based on detailed methodological descriptions of graphene production processes. For each paper, we extracted complete synthesis methodology sections, expanding technical abbreviations (GO: graphene oxide, RGO: reduced graphene oxide, CVD: chemical vapor deposition, PMMA: poly(methyl methacrylate)) to ensure contextual clarity and including complete bibliographic metadata.
Vector embeddings were generated using OpenAI's state-of-the-art26 text-embedding-3-large model via the LangChain framework, producing 3072-dimensional vectors optimized for semantic similarity retrieval. The embeddings were integrated into a Pinecone vector database configured for cosine similarity-based retrieval, with systematic batch processing to ensure efficient resource utilization.
Question and ground truth establishment
To enable systematic evaluation across different question types, we developed a structured question taxonomy based on scientific query characteristics commonly encountered in materials science research. A materials science expert with extensive graphene research experience created 20 questions spanning four categories (Table S1, SI): major fabrication methods27 (8 questions), synthesis of graphene derivatives (6 questions), application-specific synthesis strategies (2 questions), and mechanistic understanding of materials and processes (4 questions).
Ground truth answers were established through expert consensus involving three materials science researchers with specialized knowledge in graphene synthesis and characterization.28–30 A lead expert first drafted the initial ground truth answers based solely on professional knowledge, without access to the RAG corpus. These answers were then independently reviewed by two additional experts under a double-blind protocol. Revisions were iteratively discussed until all three experts reached full consensus, at which point the ground truth answers were finalized. This procedure minimizes corpus-induced bias and ensures that the ground truth reflects genuine domain knowledge rather than database-specific information.
We note, however, that the relatively limited size and diversity of the dataset may not capture the full range of scientific reasoning challenges, potentially underrepresenting differences between models of varying scales.
Obtaining responses from modes of questioning
To obtain a diverse range of responses for the Q-GT dataset, we employed four response modes: RAG-augmented and standard versions of two LLMs, Gemini-2.5-Flash and Qwen2.5-7B-Instruct. Gemini, released by Google on June 17, 2025, and freely accessible via a public API, was run with a default temperature of 1.0 and top p of 0.95 to balance determinism and variability. Qwen, an open-source model developed by Alibaba Cloud and accessed through the Hugging Face Transformers library, was prompted in a chat format with a system message followed by the user query. Tokenization was handled by “AutoTokenizer”, and responses were generated with a maximum of 1028 tokens.
To generate responses with the RAG-LLM, queries are embedded using the same process described in database preparation, and the five most similar contexts are retrieved from the Pinecone database via cosine similarity. These contexts are then combined with the query in a prompt template (Section S2, SI) to produce contextually informed responses (Fig. 2). The baseline modes used the same LLMs without retrieval, providing a clear reference point for assessing the impact of augmentation on response quality.
 |
| | Fig. 2 RAG-LLM workflow for response generation. User queries are transformed into vector embeddings and compared against the embedded research papers in the Pinecone database. The top five most similar papers are retrieved, and their contexts are combined with the original query in a structured prompt, which is subsequently passed to the LLM to generate the final RAG-LLM answer. | |
Comprehensive evaluation framework
We implemented four distinct evaluation approaches to assess both absolute performance and relative sensitivity to retrieval augmentation. An overview of the inputs and evaluation setup across all four methods, BERTScore, LLM judge, expert human evaluation, and RAGAS, is illustrated in Fig. 3.
 |
| | Fig. 3 Workflow of the four evaluation approaches. All evaluators require both the model generated answer and the ground truth as inputs. BERTScore uses only these to compute semantic similarity. RAGAS, in addition to the model answer and ground truth, also incorporates the retrieved context to assess context recall and faithfulness. LLM judge and human evaluations depend on clearly specified instructions or rubrics, as they do not follow a standardized scoring framework like BERTScore or RAGAS. | |
(1) RAGAS metrics: we evaluate our RAG pipeline using RAGAS, an open-source framework that implements multiple quantitative metrics for comprehensive RAG system assessment we employed three core RAGAS metrics using GPT-4o as the evaluator with temperature set to 0 for consistency: Factual Correctness (FC), Context Recall (CR), and Faithfulness (FF).
FC assesses overlap between generated responses and ground truth using claim-level decomposition. RAGAS offers three evaluation modes for this metric: precision, recall, and F1 score. In this study, we adopt the recall mode, as it offers the most intuitive interpretation and is more easily transferable when replicating the evaluation process with a panel of human experts.
CR measures alignment between retrieved context and ground truth. This metric is calculated as:
CR effectively captures the retriever's ability to surface relevant documents.
FF quantifies how well generated answers remain grounded in provided context, measuring the proportion of answer claims inferable from retrieved context, scoring from 0 to 1 (higher values indicate better alignment):
(2) BERTScore metric: to compare against the FC metric, we also employed BERTScore31 as an alternative method for evaluating semantic alignment between generated answers and ground truths. We used the facebook/bart-large-mnli model via Hugging Face transformers, employing recall mode for consistency with RAGAS factual correctness evaluation. All inputs were verified to remain within the 1022 token limit.
(3) LLM-as-a-judge (LLM Judge) metric: we implemented a custom evaluation approach using GPT-4o (temperature 0) with carefully engineered prompts (Section S3, SI) instructing the model to assess response quality on a 0–10 scale, similar to academic grading rubrics. The evaluation criteria emphasized content similarity and factual alignment with ground truth. Similar evaluation schemes have also been proposed in other domains, such as PaperBench,32 which evaluates the ability of LLM agents to reproduce machine learning papers from scratch using hierarchical rubrics and LLM judges.
(4) Expert human evaluator metric: nine subject matter experts independently evaluated responses using the same instructions and 0–10 scoring criteria as the LLM judge approach. Each response was evaluated by exactly three experts, with averaged scores used for analysis. Experts were blinded to response sources, and response positions were randomized across evaluation sheets to minimize bias.
Results
Performance patterns across evaluation methods
The evaluation revealed distinct patterns in how different assessment approaches capture RAG system performance. Human evaluators assigned the highest average scores (mean = 6.41, σ = 1.89), followed by LLM judge (6.00, σ = 2.24), BERTScore (5.89, σ = 0.70), and RAGAS (3.76, σ = 2.38). The substantial variation in both central tendency and dispersion across evaluators highlights fundamental differences in how each approach interprets and scores scientific content.
Retrieval augmentation impact analysis
Human evaluation revealed clear performance hierarchies demonstrating the value of retrieval augmentation for scientific applications. RAG-Gemini achieved the highest average score (6.92), followed by RAG-Qwen (6.68), standard Gemini (6.37), and standard Qwen (5.68). The performance improvements from retrieval augmentation were substantial: 0.55 points for Gemini and 1.00 points for Qwen, representing 9% and 17% relative improvements respectively.
Notably, within the evaluated dataset, the impact of retrieval was more pronounced for smaller open-source models: RAG-Qwen not only exceeded the performance of standard Gemini despite its smaller size (7B vs. Gemini's larger scale) but also showed nearly twice the relative gain in FC compared to Gemini-2.5-Flash. This demonstrates that retrieval enhances smaller open-source models to the point where they can compete effectively with larger proprietary alternatives in domain-specific applications.
Automated evaluator alignment analysis
The relationship between automated evaluators and human judgment reveals critical insights about evaluation framework reliability in scientific contexts. RAGAS exhibited the largest absolute deviation from human scores (73.5% average difference) yet demonstrated the highest sensitivity to retrieval-augmented performance improvements (Fig. 4a). RAGAS successfully captured the relative performance gains observed by human evaluators: 0.52-point improvement for Gemini (vs. 0.55 human-observed) and 1.03-point improvement for Qwen (vs. 1.00 human-observed).
 |
| | Fig. 4 Comparison of factual correctness scores across LLMs and evaluators. (a) RAGAS vs. human scores for RAG and non-RAG variants of Gemini and Qwen. RAGAS underestimates factual correctness scores and shows poor alignment with human evaluation. (b) BERTScore vs. human scores across same LLMs. BERTScore show low variance in scores. (c) LLM Judge vs. human scores, showing the closest alignment in both distribution and median values. (d) Mean factual correctness scores across RAG and non-RAG LLMs by evaluator. Human and RAGAS reflect consistent factual gains from RAG augmentation, unlike LLM judge. | |
BERTScore showed minimal absolute deviation (8.78%) but suffered from restricted score distribution (σ = 0.70), clustering most outputs between 5.19–6.59 (Fig. 4b). When applied to human evaluation patterns, only 24% of human scores fell within BERTScore's expected range, indicating poor alignment with human scoring patterns despite superficial agreement in average scores. Consequently, BERTScore lacks the interpretability and responsiveness needed for evaluating factual correctness.
LLM judge demonstrated both low absolute deviation (7.53%) and appropriate score distribution (σ = 2.24), providing the closest overall alignment with human evaluation patterns (Fig. 4c). However, LLM judge failed to capture retrieval augmentation benefits consistently, incorrectly favouring standard Gemini over RAG-Gemini and showing minimal differentiation between Qwen variants (Fig. 4d).
To capture RAG system behaviour beyond factual correctness, RAGAS reports two additional component metrics. These metrics cannot be used for comparative assessment, as they are only defined for RAG-augmented models.
(1) Context recall (CR): inter-model CR scores were nearly identical (Table 1), as expected under fixed retrieval conditions, with minor variation reflecting the LLM-based nature of RAGAS. At the question level, however, CR scores varied substantially (mean = 3.92, σ = 3.26), with several questions receiving zeros. Two main factors accounted for this. First, the metric struggled to recognize domain-specific terminology; for example, Question 16 retrieved factually correct and relevant contexts yet still scored zero. Second, genuine retrieval failures occurred in cases requiring deeper expertise. Question 20, which probed mechanistic understanding, consistently showed poor retrieval, underscoring the limitations of similarity-based retrieval approaches for complex scientific reasoning (Section S4, Question 20, SI).
Table 1 Context recall and faithfulness scores (mean ± standard deviation) for LLMsa
| Large language model |
Context recall |
Faithfulness |
| RAGAS component metric analysis. |
| RAG-Gemini |
3.92 ± 3.29 |
9.20 ± 1.54 |
| RAG-Qwen |
3.91 ± 3.23 |
7.32 ± 2.21 |
(2) Faithfulness: FF scores demonstrated model-specific patterns, with RAG-Gemini showing higher consistency (9.20 ± 1.54) compared to RAG-Qwen (7.32 ± 2.21). The higher variability in RAG-Qwen's scores indicates fluctuations in maintaining faithfulness to retrieved information, reflecting differences in contextual grounding during generation. For instance, in Question 19, both models retrieved identical relevant contexts, but RAG-Qwen produced unfaithful reasoning (FF = 4.0, FC = 1.7) as shown in Section S6 (SI), whereas RAG-Gemini remained faithful (FF = 7.1) and achieved a much higher factual correctness score (FC = 8.0).
Discussion
Methodological implications for scientific RAG evaluation
Our findings reveal that automated evaluation frameworks face fundamental challenges when applied to scientific content, reflecting the specialized nature of scientific knowledge and reasoning. For instance, BERTScore and LLM judge exhibit evaluator-specific limitations, namely restricted score distributions and failure to consistently capture retrieval augmentation benefits, respectively. In the case of RAGAS, Question 16 highlighted a structural weakness of the CR metric: despite retrieval of factually correct and relevant context, the score was zero, revealing its inability to map domain-specific terminology and limiting its utility in scientific applications.
By contrast, the FC metric was sensitive to retrieval augmentation benefits, but its poor absolute alignment with human judgment restricts its value for standalone assessment. The FF metric added complementary insights by capturing nuanced differences in how LLMs maintained contextual grounding, with these patterns further validated by corresponding FC scores.
Taken together, these observations indicate that RAGAS is most effective for comparative studies of RAG-based question and answering systems, where relative performance differences are the primary focus, rather than for applications requiring precise absolute performance levels. In such contexts, particularly those involving complex reasoning, expert human evaluation remains essential for determining deployment readiness.
Framework selection guidelines for scientific applications
Based on our systematic comparison, we propose the following guidelines for evaluation framework selection in scientific RAG applications: For comparative studies evaluating factual correctness of multiple RAG systems under identical conditions, RAGAS provides quick and reliable relative performance assessment despite absolute score limitations. Among its component metrics, FF remains useful for capturing additional dimensions of evaluation, whereas CR requires further refinement before it can be reliably applied in scientific contexts.
For standalone system evaluation where absolute performance interpretation is critical, expert human evaluation remains the gold standard. However, resource constraints may call for hybrid strategies in which LLM judges provide preliminary screening, with human evaluation applied only once outputs surpass a quality threshold warranting closer assessment. For rapid development cycles where frequent evaluation is needed, LLM judge approaches offer the best balance of human alignment and practical accessibility, though their limitations in capturing retrieval benefits must be considered.
BERTScore, despite its superficial alignment with human averages, provides insufficient discrimination and should be avoided for scientific evaluation where subtle performance differences are important.
Limitations and future directions
Our study focuses specifically on graphene synthesis, limiting direct generalizability to other scientific domains. However, the methodological insights about evaluation framework behavior likely extend to other technical fields with similar requirements for precision and domain expertise.
Future research should explore domain-adapted evaluation frameworks that incorporate scientific reasoning patterns and field-specific accuracy requirements. Developing evaluation approaches that can reliably assess conceptual understanding and reasoning quality, rather than primarily factual recall, represents a critical need for advancing AI applications in scientific research.
Conclusion
This study provides a systematic analysis of automated evaluation frameworks for scientific RAG systems, revealing both capabilities and fundamental limitations of current approaches. While RAGAS's factual correctness metric can effectively capture relative performance improvements from retrieval augmentation, it exhibits significant challenges in absolute score interpretation for scientific content.
Our findings suggest that RAG systems can provide meaningful performance improvements for scientific applications, particularly for smaller open-source models, narrowing the performance gap with larger proprietary alternatives in our evaluation setting. However, the evaluation challenges identified highlight the need for more sophisticated assessment approaches tailored to scientific domains.33–36
The methodological guidelines developed through this analysis provide practical guidance for researchers deploying RAG systems in scientific applications. By understanding when different evaluation approaches succeed or fail, researchers can make informed decisions about evaluation strategies appropriate for their specific needs and constraints.
Beyond the immediate findings, this work establishes a foundation for developing evaluation methodologies appropriate for AI systems in specialized domains where accuracy and reliability are paramount. The insights gained might extend beyond materials science to any technical field requiring precise knowledge synthesis and reasoning capabilities.
Author contributions
Z. H. C. was responsible for the conceptualization, data curation, investigation, design of methodology, implementation and validation of the computer code, and writing of the manuscript. M. O. was responsible for the data collection, design of methodology and reviewing of the manuscript. S. D., M. D. and X. X. were responsible for data collection and reviewing of the manuscript. J. C. G. was responsible for data curation and data collection. B. S. T., J. W., Y. L. and A. J. were responsible for data collection. L. W. T. N. was responsible for conceptualization, data collection, design of methodology, provision of computing resources, review and editing of the manuscript.
Conflicts of interest
There are no conflicts to declare.
Data availability
Code used to carry out the project and data for RAG training and RAGAS evaluation can be found at https://doi.org/10.17632/ry7phxn4js.3.
Supplementary information (SI): (i) a 20-question graphene-synthesis question–ground truth dataset spanning multiple synthesis and processing routes, (ii) the exact prompt templates used for RAG and standard models to enable answer generation, (iii) the LLM-as-a-judge evaluation prompt , and (iv) the complete answer sets for all questions from each evaluated configuration (RAG-Gemini, Gemini, RAG-Qwen, and Qwen), to provide clear documentation of the dataset, prompts, and model outputs used in the analysis. See DOI: https://doi.org/10.1039/d5ra09726f.
References
- H. Touvron, T. Lavril, G. Izacard, X. Martinet, M. A. Lachaux, T. Lacroix, et al., LLaMA: Open and Efficient Foundation Language Models, 2023, Available from: http://arxiv.org/abs/2302.13971.
- OpenAI, J. Achiam, S. Adler, S. Agarwal, L. Ahmad, I. Akkaya, et al., GPT-4 Technical Report, 2023, Available from: http://arxiv.org/abs/2303.08774 Search PubMed.
- P. Zhao, H. Zhang, Q. Yu, Z. Wang, Y. Geng, F. Fu, et al., Retrieval-Augmented Generation for AI-Generated Content: A Survey, 2024, Available from: http://arxiv.org/abs/2402.19473.
- E. Kasneci, K. Sessler, S. Küchemann, M. Bannert, D. Dementieva and F. Fischer, et al., ChatGPT for good? On opportunities and challenges of large language models for education, Learn. Individ. Differ., 2023, 103, 102274, DOI:10.1016/j.lindif.2023.102274
. - M. Thway, J. Recatala-Gomez, F. S. Lim, K. Hippalgaonkar and L. W. T. Ng, Harnessing GenAI for Higher Education: A Study of a Retrieval Augmented Generation Chatbot's Impact on Learning, J. Chem. Educ., 2025, 102(9), 3849–3857, DOI:10.1021/acs.jchemed.5c00113
. - A. J. Thirunavukarasu, D. S. J. Ting, K. Elangovan, L. Gutierrez, T. F. Tan and D. S. W. Ting, Large language models in medicine, Nat. Med., 2023, 29(8), 1930–1940, DOI:10.1038/s41591-023-02448-8
. - Y. Chiang, E. Hsieh, C. H. Chou and J. Riebesell, LLaMP: Large Language Model Made Powerful for High-fidelity Materials Knowledge Retrieval and Distillation, 2024, Available from: http://arxiv.org/abs/2401.17244.
- J. Lála, O. O'Donoghue, A. Shtedritski, S. Cox, S. G. Rodriques and A. D. White, PaperQA: Retrieval-Augmented Generative Agent for Scientific Research, 2023, Available from: http://arxiv.org/abs/2312.07559.
- M. D. Skarlinski, S. Cox, J. M. Laurent, J. D. Braza, M. Hinks, M. J. Hammerling, et al., Language agents achieve superhuman synthesis of scientific knowledge, 2024, Available from: http://arxiv.org/abs/2409.13740.
- O. Tippins, T. Alvarez, J. Novak, R. Martinez, E. Thompson and V. Williams, Domain-Specific Retrieval-Augmented Generation Through Token Factorization: An Experimental Study, 2024, Available from: https://www.techrxiv.org/users/841132/articles/1231256.
- S. Siriwardhana, R. Weerasekera, E. Wen, T. Kaluarachchi, R. Rana and S. Nanayakkara, Improving the Domain Adaptation of Retrieval Augmented Generation (RAG) Models for Open Domain Question Answering, Trans. Assoc. Comput. Linguistics, 2023, 11, 1–17, DOI:10.1162/tacl_a_00530
. - S. Simon, A. Mailach, J. Dorn and N. Siegmund, A Methodology for Evaluating RAG Systems: A Case Study On Configuration Dependency Validation, 2024, Available from: http://arxiv.org/abs/2410.08801.
- K. Zhu, Y. Luo, D. Xu, R. Wang, S. Yu, S. Wang, et al., RAGEval: Scenario Specific RAG Evaluation Dataset Generation Framework, 2024, Available from: http://arxiv.org/abs/2408.01262.
- S. Roychowdhury, S. Soman, H. G. Ranjani, N. Gunda, V. Chhabra and S. K. Bala, Evaluation of RAG Metrics for Question Answering in the Telecom Domain, 2024, Available from: http://arxiv.org/abs/2407.12873.
- D. Galla, S. Hoda, M. Zhang, W. Quan, T. D. Yang and J. Voyles, CoURAGE: A Framework to Evaluate RAG Systems, in Natural Language Processing and Information Systems, ed. A. Rapp, L. Di Caro, F. Meziane, V. Sugumaran, Springer Nature Switzerland, Cham, 2024, pp. 392–407, DOI:10.1007/978-3-031-70242-6_37
. - L. W. T. Ng, G. Hu, R. C. T. Howe, X. Zhu, Z. Yang, C. G. Jones, et al., 2D Material Production Methods, in Printing of Graphene and Related 2D Materials: Technology, Formulation and Applications, ed. L. W. T. Ng, G. Hu, R. C. T. Howe, X. Zhu, Z. Yang, C. G. Jones, et al., Springer International Publishing, Cham, 2019, pp. 53–101, DOI:10.1007/978-3-319-91572-2_3
. - D. Rein, B. L. Hou, A. C. Stickland, J. Petty, R. Y. Pang, J. Dirani, et al., GPQA: A Graduate-Level Google-Proof Q&A Benchmark, 2023, Available from: http://arxiv.org/abs/2311.12022.
- M. Zaki, M. Jayadeva and N. M. A. Krishnan, MaScQA: investigating materials science knowledge of large language models, Digital Discovery, 2024, 3(2), 313–327, 10.1039/D3DD00188A
. - C. Bajan and G. Lambard, Exploring the expertise of large language models in materials science and metallurgical engineering, Digital Discovery, 2025, 4(2), 500–512, 10.1039/D4DD00319E
. - S. Es, J. James, L. Espinosa-Anke and S. Schockaert, Ragas: Automated Evaluation of Retrieval Augmented Generation, 2025, Available from: http://arxiv.org/abs/2309.15217.
- H. Wei, S. He, T. Xia, A. Wong, J. Lin and M. Han, Systematic Evaluation of LLM-as-a-Judge in LLM Alignment Tasks: Explainable Metrics and Diverse Prompt Templates, 2024, Available from: http://arxiv.org/abs/2408.13006.
- E. A. Mullins, A. Portillo, K. Ruiz-Rohena and A. Piplai, Enhancing classroom teaching with LLMs and RAG, 2024, Available from: http://arxiv.org/abs/2411.04341.
- K. Suresh, N. Kackar, L. Schleck and C. Fanelli, Towards a RAG-based summarization for the Electron Ion Collider, J. Instrum., 2024, 19(07), C07006, DOI:10.1088/1748-0221/19/07/C07006
. - L. W. T. Ng, N. G. An, L. Yang, Y. Zhou, D. W. Chang and J. E. Kim, et al., A printing-inspired digital twin for the self-driving, high-throughput, closed-loop optimization of roll-to-roll printed photovoltaics, Cell Rep. Phys. Sci., 2024, 5(6), 102038, DOI:10.1016/j.xcrp.2024.102038
. - A. K. Y. Low, J. J. W. Cheng, K. Hippalgaonkar and L. W. T. Ng, Self-Driving Laboratories: Translating Materials Science from Laboratory to Factory, ACS Omega, 2025, 10(28), 29902–29908, DOI:10.1021/acsomega.5c02197
. - OpenAI, Vector embeddings, Available from: https://platform.openai.com/docs/guides/embeddings.
- R. S. Perala, N. Chandrasekar, R. Balaji, P. S. Alexander, N. Z. N. Humaidi and M. T. Hwang, A comprehensive review on graphene-based materials: From synthesis to contemporary sensor applications, Materials Science and Engineering R: Reports, Elsevier Ltd, 2024, vol. 159, DOI:10.1016/j.mser.2024.100805
. - L. W. T. Ng, G. Hu, R. C. T. Howe, X. Zhu, Z. Yang, C. G. Jones, et al., Structures, Properties and Applications of 2D Materials, in Printing of Graphene and Related 2D Materials: Technology, Formulation and Applications, ed. L. W. T. Ng, G. Hu, R. C. T. Howe, X. Zhu, Z. Yang, C. G. Jones, et al., Springer International Publishing, Cham, 2019, pp. 19–51, DOI:10.1007/978-3-319-91572-2_2
. - N. Macadam, L. W. T. Ng, G. Hu, H. H. Shi, W. Wang and X. Zhu, et al., 100 m min−1 Industrial-Scale Flexographic Printing of Graphene-Incorporated Conductive Ink, Adv. Eng. Mater., 2022, 24(5), 2101217, DOI:10.1002/adem.202101217
. - L. W. T. Ng, G. Hu, R. C. T. Howe, X. Zhu, Z. Yang, C. G. Jones, et al., Printing Technologies, in Printing of Graphene and Related 2D Materials: Technology, Formulation and Applications, ed. L. W. T. Ng, G. Hu, R. C. T. Howe, X. Zhu, Z. Yang, C. G. Jones, et al., Springer International Publishing, Cham, 2019, pp. 135–178, DOI:10.1007/978-3-319-91572-2_5
. - T. Zhang, V. Kishore, F. Wu, K. Q. Weinberger and Y. Artzi, BERTScore: Evaluating Text Generation with BERT, 2019, Available from: http://arxiv.org/abs/1904.09675.
- G. Starace, O. Jaffe, D. Sherburn, J. Aung, J. S. Chan, L. Maksin, et al., PaperBench: Evaluating AI's Ability to Replicate AI Research, 2025, Available from: http://arxiv.org/abs/2504.01848.
- X. Xiao, M. Chalh, Z. R. Loh, E. Mbina, T. Xu and R. C. Hiorns, et al., Strategies to achieve efficiencies of over 19% for organic solar cells, Cell Rep. Phys. Sci., 2025, 6(1), 102390, DOI:10.1016/j.xcrp.2024.102390
. - M. M. B. A. Mohamed, Y. X. Ang, R. J. H. Tan, L. S. Tung, X. Xiao and M. Das, et al., Investigating the fire dynamics of mounted PV weathering effects and material changes, iScience, 2025, 28(9), 113410, DOI:10.1016/j.isci.2025.113410
. - L. Song, P. Liu, Y. Liu, J. Pei, W. Cui and S. Liu, et al., Hardware Implementation of Bayesian Decision-Making with Memristors, Adv. Electron. Mater., 2025, 11(16), e00134, DOI:10.1002/aelm.202500134
. - Z. Lin, Y. Li, M. Das, C. Liang, X. Xiao and Z. Yen, et al., Direct Integration of Biomass-Derived Furan Polymers for Enhanced Stability and Efficiency in Hybrid Perovskite Solar Cells, Adv. Funct. Mater., 2025, 35(26), 2423635, DOI:10.1002/adfm.202423635
.
|
| This journal is © The Royal Society of Chemistry 2026 |
Click here to see how this site uses Cookies. View our privacy policy here.