To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
This study presents a comparative evaluation of sentiment analysis models applied to a large corpus of expert wine reviews from Wine Spectator, with the goal of classifying reviews into binary sentiment categories based on expert ratings. We assess six models: logistic regression, XGBoost, LSTM, BERT, the interpretable Attention-based Multiple Instance Classification (AMIC) model, and the generative language model LLAMA 3.1, highlighting their differences in accuracy, interpretability, and computational efficiency. While LLAMA 3.1 achieves the highest accuracy, its marginal improvement over AMIC and BERT comes at a significantly higher computational cost. Notably, AMIC matches the performance of pretrained large language models while offering superior interpretability, making it particularly effective for domain-specific tasks such as wine sentiment analysis. Through qualitative analysis of sentiment-bearing words, we demonstrate AMIC’s ability to uncover nuanced, context-dependent language patterns unique to wine reviews. These findings challenge the assumption of generative models’ universal superiority and underscore the importance of aligning model selection with domain-specific requirements, especially in applications where transparency and linguistic nuance are critical.
The goal of this chapter is to develop a theory of Formal Features that will capture and unify many of the generalisations we have arrived at in previous chapters and, more specifically, to develop a theory of Case, agreement and movement, showing how these three notions are intertwined. The core notion is the Agree relation.
Two salient notions of sameness of theories are synonymy, aka definitional equivalence, and bi-interpretability. Of these two definitional equivalence is the strictest notion. In which cases can we infer synonymy from bi-interpretability? We study this question for the case of sequential theories. Our result is as follows. Suppose that two sequential theories are bi-interpretable and that the interpretations involved in the bi-interpretation are one-dimensional and identity preserving. Then, the theories are synonymous.
The crucial ingredient of our proof is a version of the Schröder–Bernstein theorem under very weak conditions. We think this last result has some independent interest.
We provide an example to show that this result is optimal. There are two finitely axiomatized sequential theories that are bi-interpretable but not synonymous, where precisely one of the interpretations involved in the bi-interpretation is not identity preserving.
Artificial intelligence is dramatically reshaping scientific research and is coming to play an essential role in scientific and technological development by enhancing and accelerating discovery across multiple fields. This book dives into the interplay between artificial intelligence and the quantum sciences; the outcome of a collaborative effort from world-leading experts. After presenting the key concepts and foundations of machine learning, a subfield of artificial intelligence, its applications in quantum chemistry and physics are presented in an accessible way, enabling readers to engage with emerging literature on machine learning in science. By examining its state-of-the-art applications, readers will discover how machine learning is being applied within their own field and appreciate its broader impact on science and technology. This book is accessible to undergraduates and more advanced readers from physics, chemistry, engineering, and computer science. Online resources include Jupyter notebooks to expand and develop upon key topics introduced in the book.
Distinguishing between different phases of matter and detecting phase transitions are some of the most central tasks in many-body physics. Traditionally, these tasks are accomplished by searching for a small set of low-dimensional quantities capturing the macroscopic properties of each phase of the system, so-called order parameters. Because of the large state space underlying many-body systems, success generally requires a great deal of human intuition and understanding. In particular, it can be challenging to define an appropriate order parameter if the symmetry breaking pattern is unknown or the phase is of topological nature and thus exhibits nonlocal order. In this chapter, we explore the use of machine learning to automate the task of classifying phases of matter and detecting phase transitions. We discuss the application of various machine learning techniques, ranging from clustering to supervised learning and anomaly detection, to different physical systems, including the prototypical Ising model that features a symmetry-breaking phase transition and the Ising gauge theory which hosts a topological phase of matter.
Studies conducted during the COVID-19 pandemic found high occurrence of suicidal thoughts and behaviours (STBs) among healthcare workers (HCWs). The current study aimed to (1) develop a machine learning-based prediction model for future STBs using data from a large prospective cohort of Spanish HCWs and (2) identify the most important variables in terms of contribution to the model’s predictive accuracy.
Methods
This is a prospective, multicentre cohort study of Spanish HCWs active during the COVID-19 pandemic. A total of 8,996 HCWs participated in the web-based baseline survey (May–July 2020) and 4,809 in the 4-month follow-up survey. A total of 219 predictor variables were derived from the baseline survey. The outcome variable was any STB at the 4-month follow-up. Variable selection was done using an L1 regularized linear Support Vector Classifier (SVC). A random forest model with 5-fold cross-validation was developed, in which the Synthetic Minority Oversampling Technique (SMOTE) and undersampling of the majority class balancing techniques were tested. The model was evaluated by the area under the Receiver Operating Characteristic (AUROC) curve and the area under the precision–recall curve. Shapley’s additive explanatory values (SHAP values) were used to evaluate the overall contribution of each variable to the prediction of future STBs. Results were obtained separately by gender.
Results
The prevalence of STBs in HCWs at the 4-month follow-up was 7.9% (women = 7.8%, men = 8.2%). Thirty-four variables were selected by the L1 regularized linear SVC. The best results were obtained without data balancing techniques: AUROC = 0.87 (0.86 for women and 0.87 for men) and area under the precision–recall curve = 0.50 (0.55 for women and 0.45 for men). Based on SHAP values, the most important baseline predictors for any STB at the 4-month follow-up were the presence of passive suicidal ideation, the number of days in the past 30 days with passive or active suicidal ideation, the number of days in the past 30 days with binge eating episodes, the number of panic attacks (women only) and the frequency of intrusive thoughts (men only).
Conclusions
Machine learning-based prediction models for STBs in HCWs during the COVID-19 pandemic trained on web-based survey data present high discrimination and classification capacity. Future clinical implementations of this model could enable the early detection of HCWs at the highest risk for developing adverse mental health outcomes.
This paper reports the results of a cross-sectional study investigating the acquisition of the syntactic properties associated with the null subject (meso-)parameter in English as a second language (L2) among Hebrew-speaking youngsters (18-year-olds). The two languages differ concerning these properties, with Hebrew allowing null subjects and related properties (although inconsistently) and English disallowing these properties altogether. One hundred four intermediate learners and 97 English-speaker controls provided grammaticality judgments and corrections concerning constructions involving expletive and referential null subjects, post-verbal subjects, and complementizer-trace sequences. The results reveal limited evidence for transfer from the learners’ mother tongue (first language [L1]) and indicate that learners have met the native standard concerning null and post-verbal subjects. These findings support both the meso-parametric view of cross-linguistic variation and feature reassembly on functional heads in L2 acquisition, while partially rejecting the Interpretability Hypothesis. Learners nevertheless deviate from the native standard concerning complementizer-trace sequences. This finding is unaccounted for by the meso-parametric approach, feature reassembly, or interpretability, but can instead be attributed to L1 transfer. Controls also demonstrate variability concerning complementizer-trace sequences, suggesting that the performance of all participants regarding this configuration is affected by processing difficulties, lower frequency in the input, and methodological issues with the items and/or the task.
In the applications of maximum likelihood factor analysis the occurrence of boundary minima instead of proper minima is no exception at all. In the past the causes of such improper solutions could not be detected. This was impossible because the matrices containing the parameters of the factor analysis model were kept positive definite. By dropping these constraints, it becomes possible to distinguish between the different causes of improper solutions. In this paper some of the most important causes are discussed and illustrated by means of artificial and empirical data.
Generalized structured component analysis (GSCA) is a multivariate method for examining theory-driven relationships between variables including components. GSCA can provide the deterministic component score for each individual once model parameters are estimated. As the traditional GSCA always standardizes all indicators and components, however, it could not utilize information on the indicators’ scale in parameter estimation. Consequently, its component scores could just show the relative standing of each individual for a component, rather than the individual’s absolute standing in terms of the original indicators’ measurement scales. In the paper, we propose a new version of GSCA, named convex GSCA, which can produce a new type of unstandardized components, termed convex components, which can be intuitively interpreted in terms of the original indicators’ scales. We investigate the empirical performance of the proposed method through the analyses of simulated and real data.
In this paper, we study the employment of $\Sigma _1$-sentences with certificates, i.e., $\Sigma _1$-sentences where a number of principles is added to ensure that the witness is sufficiently number-like. We develop certificates in some detail and illustrate their use by reproving some classical results and proving some new ones. An example of such a classical result is Vaught’s theorem of the strong effective inseparability of $\mathsf {R}_0$.
We also develop the new idea of a theory being $\mathsf {R}_{0\mathsf {p}}$-sourced. Using this notion, we can transfer a number of salient results from $\mathsf {R}_0$ to a variety of other theories.
Turbulent flows are chaotic and multi-scale dynamical systems, which have large numbers of degrees of freedom. Turbulent flows, however, can be modeled with a smaller number of degrees of freedom when using an appropriate coordinate system, which is the goal of dimensionality reduction via nonlinear autoencoders. Autoencoders are expressive tools, but they are difficult to interpret. This article aims to propose a method to aid the interpretability of autoencoders. First, we introduce the decoder decomposition, a post-processing method to connect the latent variables to the coherent structures of flows. Second, we apply the decoder decomposition to analyze the latent space of synthetic data of a two-dimensional unsteady wake past a cylinder. We find that the dimension of latent space has a significant impact on the interpretability of autoencoders. We identify the physical and spurious latent variables. Third, we apply the decoder decomposition to the latent space of wind-tunnel experimental data of a three-dimensional turbulent wake past a bluff body. We show that the reconstruction error is a function of both the latent space dimension and the decoder size, which are correlated. Finally, we apply the decoder decomposition to rank and select latent variables based on the coherent structures that they represent. This is useful to filter unwanted or spurious latent variables or to pinpoint specific coherent structures of interest. The ability to rank and select latent variables will help users design and interpret nonlinear autoencoders.
The purpose of this study was to explore the electroencephalogram (EEG) features sensitive to situation awareness (SA) and then classify SA levels. Forty-eight participants were recruited to complete an SA standard test based on the multi-attribute task battery (MATB) II, and the corresponding EEG data and situation awareness global assessment technology (SAGAT) scores were recorded. The population with the top 25% of SAGAT scores was selected as the high-SA level (HSL) group, and the bottom 25% was the low-SA level (LSL) group. The results showed that (1) for the relative power of $\beta$1 (16–20Hz), $\beta$2 (20–24Hz) and $\beta$3 (24–30Hz), repeated measures analysis of variance (ANOVA) in three brain regions (Central Central-Parietal, and Parietal) × three brain lateralities (left, midline, and right) × two SA groups (HSL and LSL) showed a significant main effect for SA groups; post hoc comparisons revealed that compared with LSL, the above features of HSL were higher. (2) for most ratio features associated with $\beta$1 ∼ $\beta$3, ANOVA also revealed a main effect for SA groups. (3) EEG features sensitive to SA were selected to classify SA levels with small-sample data based on the general supervised machine learning classifiers. Five-fold cross-validation results showed that among the models with easy interpretability, logistic regression (LR) and decision tree (DT) presented the highest accuracy (both 92%), while among the models with hard interpretability, the accuracy of random forest (RF) was 88.8%, followed by an artificial neural network (ANN) of 84%. The above results suggested that (1) the relative power of $\beta$1 ∼ $\beta$3 and their associated ratios were sensitive to changes in SA levels; (2) the general supervised machine learning models all exhibited good accuracy (greater than 75%); and (3) furthermore, LR and DT are recommended by combining the interpretability and accuracy of the models.
Human gait trajectory prediction is a long-standing research topic in human–machine interaction. However, there are two shortcomings in the current gait trajectory prediction technology. The first shortcoming is that the neural network model of gait prediction only predicts dozens of future time frames of gait trajectory. The second shortcoming is that the gait prediction neural network model is uninterpretable. We propose the Interpretable-Concatenation former (IC-former) model, which can predict long-term gait trajectories and explain the prediction results by quantifying the importance of data at different positions in the input sequence. Experiments prove that the IC-former model we proposed not only makes a breakthrough in prediction accuracy but also successfully explains the data basis of the prediction.
Elementary first-order theories of trees allowing at most, exactly $\mathrm{m}$, and any finite number of immediate descendants are introduced and proved mutually interpretable among themselves and with Robinson arithmetic, Adjunctive Set Theory with Extensionality and other well-known weak theories of numbers, sets, and strings.
To answer the question of what responsible AI means, the authors, Jaan Tallinn and Richard Ngo, propose a framework for the deployment of AI which focuses on two concepts: delegation and supervision. The framework aims towards building ‘delegate AIs’ which lack goals of their own but can perform any task delegated to them. However, AIs trained with hardcoded reward functions, or even human feedback, often learn to game their reward signal instead of accomplishing their intended tasks. Thus, Tallinn and Ngo argue that it will be important to develop more advanced techniques for continuous high-quality supervision – for example, by evaluating the reasons which AIs give for their choices of actions. These supervision techniques might be made scalable by training AIs to generate reward signals for more advanced AIs. Given their current limitations, however, Tallinn and Ngo call for caution when developing new AI: we must be aware of the risks and overcome self-interest and dangerous competitive incentives in order to avoid them.
Kristiansen and Murwanashyaka recently proved that Robinson arithmetic, Q, is interpretable in an elementary theory of full binary trees, T. We prove that, conversely, T is interpretable in Q by producing a formal interpretation of T in an elementary concatenation theory QT+, thereby also establishing mutual interpretability of T with several well-known weak essentially undecidable theories of numbers, strings, and sets. We also introduce a “hybrid” elementary theory of strings and trees, WQT*, and establish its mutual interpretability with Robinson’s weak arithmetic R, the weak theory of trees WT of Kristiansen and Murwanashyaka, and the weak concatenation theory WTCε of Higuchi and Horihata.
Like a hydra, fraudsters adapt and circumvent increasingly sophisticated barriers erected by public or private institutions. Among these institutions, banks must quickly take measures to avoid losses while guaranteeing the satisfaction of law-abiding customers. Facing an expanding flow of operations, effective banking relies on data analytics to support established risk control processes, but also on a better understanding of the underlying fraud mechanism. In addition, fraud being a criminal offence, the evidential aspect of the process must also be considered. These legal, operational, and strategic constraints lead to compromises on the means to be implemented for fraud management. This paper first focuses on the translation of practical questions raised in the banking industry at each step of the fraud management process into performance evaluation required to design a fraud detection model. Secondly, it considers a range of machine learning approaches that address these specificities: the imbalance between fraudulent and nonfraudulent operations, the lack of fully trusted labels, the concept-drift phenomenon, and the unavoidable trade-off between accuracy and interpretability of detection. This state-of-the-art review sheds some light on a technology race between black box machine learning models improved by post-hoc interpretation and intrinsic interpretable models boosted to gain accuracy. Finally, it discusses how concrete and promising hybrid approaches can provide pragmatic, short-term answers to banks and policy makers without swallowing up stakeholders with economical and ethical stakes in this technological race.
A Leibniz class is a class of logics closed under the formation of term-equivalent logics, compatible expansions, and non-indexed products of sets of logics. We study the complete lattice of all Leibniz classes, called the Leibniz hierarchy. In particular, it is proved that the classes of truth-equational and assertional logics are meet-prime in the Leibniz hierarchy, while the classes of protoalgebraic and equivalential logics are meet-reducible. However, the last two classes are shown to be determined by Leibniz conditions consisting of meet-prime logics only.
A notion of interpretation between arbitrary logics is introduced, and the poset $\mathsf {Log}$ of all logics ordered under interpretability is studied. It is shown that in $\mathsf {Log}$ infima of arbitrarily large sets exist, but binary suprema in general do not. On the other hand, the existence of suprema of sets of equivalential logics is established. The relations between $\mathsf {Log}$ and the lattice of interpretability types of varieties are investigated.
We investigate which part of Brouwer’s Intuitionistic Mathematics is finitistically justifiable or guaranteed in Hilbert’s Finitism, in the same way as similar investigations on Classical Mathematics (i.e., which part is equiconsistent with $\textbf {PRA}$ or consistent provably in $\textbf {PRA}$) already done quite extensively in proof theory and reverse mathematics. While we already knew a contrast from the classical situation concerning the continuity principle, more contrasts turn out: we show that several principles are finitistically justifiable or guaranteed which are classically not. Among them are: (i) fan theorem for decidable fans but arbitrary bars; (ii) continuity principle and the axiom of choice both for arbitrary formulae; and (iii)$\Sigma _2$ induction and dependent choice. We also show that Markov’s principle MP does not change this situation; that neither does lesser limited principle of omniscience LLPO (except the choice along functions); but that limited principle of omniscience LPO makes the situation completely classical.