cs.CE

44 posts

arXiv:2501.01148v1 Announce Type: cross Abstract: In this paper we address the problem of performing Bayesian inference for the parameters of a nonlinear multi-output model and the covariance matrix of the different output signals. We propose an adaptive importance sampling (AIS) scheme for multivariate Bayesian inversion problems, which is based in two main ideas: the variables of interest are split in two blocks and the inference takes advantage of known analytical optimization formulas. We estimate both the unknown parameters of the multivariate non-linear model and the covariance matrix of the noise. In the first part of the proposed inference scheme, a novel AIS technique called adaptive target adaptive importance sampling (ATAIS) is designed, which alternates iteratively between an IS technique over the parameters of the non-linear model and a frequentist approach for the covariance matrix of the noise. In the second part of the proposed inference scheme, a prior density over the covariance matrix is considered and the cloud of samples obtained by ATAIS are recycled and re-weighted to obtain a complete Bayesian study over the model parameters and covariance matrix. ATAIS is the main contribution of the work. Additionally, the inverted layered importance sampling (ILIS) is presented as a possible compelling algorithm (but based on a conceptually simpler idea). Different numerical examples show the benefits of the proposed approaches

E. Curbelo, L. Martino, F. Llorente, D. Delgado-Gomez1/3/2025

arXiv:2501.00037v1 Announce Type: cross Abstract: Several numerical simulations of a co-axial particle-laden swirling air flow in a vertical circular pipe were performed. The air flow was modeled using the unsteady Favre-averaged Navier-Stokes equations. A Lagrangian model was used for the particle motion. The gas and particles are coupled through two-way momentum exchange. The results of the simulations using three versions of the k-epsilon turbulence model (standard, re-normalization group (RNG), and realizable) are compared with experimental mean velocity profiles. The standard model achieved the best overall performance. The realizable model was unable to satisfactorily predict the radial velocity; it is also the most computationally-expensive model. The simulations using the RNG model predicted additional recirculation zones. We also compared the particle and parcel approaches in solving the particle motion. In the latter, multiple similar particles are grouped in a single parcel, thereby reducing the amount of computation.

Osama A. Marzouk, E. David Huckaby1/3/2025

arXiv:2412.12616v2 Announce Type: replace Abstract: This paper presents a rigorous derivation of equations to evaluate the macroscopic stress tensor, the couple stress tensor, and the flux vector equivalent to underlying microscopic fields in continuous and discrete heterogeneous systems with independent displacements and rotations. Contrary to the classical asymptotic expansion homogenization, finite size representative volume is considered. First, the macroscopic quantities are derived for a heterogeneous Cosserat continuum. The resulting continuum equations are discretized to provide macroscopic quantities in discrete heterogeneous systems. Finally, the expressions for discrete system are derived once again, this time considering the discrete nature directly. The formulations are presented in two variants, considering either internal or external forces, couples, and fluxes. The derivation is based on the virtual work equivalence and elucidates the fundamental significance of the couple stress tensor in the context of balance equations and admissible virtual deformation modes. Notably, an additional term in the couple stress tensor formula emerges, explaining its dependence on the reference system and position of the macroscopic point. The resulting equations are verified by comparing their predictions with known analytical solutions and results of other numerical models under both steady state and transient conditions.

Jan Eli\'a\v{s}, Gianluca Cusatis1/3/2025

arXiv:2305.11586v3 Announce Type: replace Abstract: Gaussian process regression is widely applied in computational science and engineering for surrogate modeling owning to its kernel-based and probabilistic nature. In this work, we propose a Bayesian approach that integrates the variability of input data into the Gaussian process regression for function and partial differential equation approximation. Leveraging two types of observables -- noise-corrupted outputs with certain inputs and those with prior-distribution-defined uncertain inputs, a posterior distribution of uncertain inputs is estimated via Bayesian inference. Thereafter, such quantified uncertainties of inputs are incorporated into Gaussian process predictions by means of marginalization. The setting of two types of data aligned with common scenarios of constructing surrogate models for the solutions of partial differential equations, where the data of boundary conditions and initial conditions are typically known while the data of solution may involve uncertainties due to the measurement or stochasticity. The effectiveness of the proposed method is demonstrated through several numerical examples including multiple one-dimensional functions, the heat equation and Allen-Cahn equation. A consistently good performance of generalization is observed, and a substantial reduction in the predictive uncertainties is achieved by the Bayesian inference of uncertain inputs.

Dongwei Ye, Weihao Yan, Christoph Brune, Mengwu Guo1/3/2025

arXiv:2501.00643v1 Announce Type: cross Abstract: The design space of dynamic multibody systems (MBSs), particularly those with flexible components, is considerably large. Consequently, having a means to efficiently explore this space and find the optimum solution within a feasible timeframe is crucial. It is well known that for problems with several design variables, sensitivity analysis using the adjoint variable method extensively reduces the computational costs. This paper presents the novel extension of the discrete adjoint variable method to the design optimization of dynamic flexible MBSs. The extension involves deriving the adjoint equations directly from the discrete, rather than the continuous, equations of motion. This results in a system of algebraic equations that is computationally less demanding to solve compared to the system of differential algebraic equations produced by the continuous adjoint variable method. To describe the proposed method, it is integrated with a numerical time-stepping algorithm based on geometric variational integrators. The developed technique is then applied to the optimization of MBSs composed of springs, dampers, beams and rigid bodies, considering both geometrical (e.g., positions of joints) and non-geometrical (e.g., mechanical properties of components) design variables. To validate the developed methods and show their applicability, three numerical examples are provided.

Mehran Ebrahimi, Adrian Butscher, Hyunmin Cheong, Francesco Iorio1/3/2025

arXiv:2501.00450v1 Announce Type: cross Abstract: This work presents an overview of mesh-induced errors commonly experienced by cell-centred finite volumes (CCFV), for which the face-centred finite volume (FCFV) paradigm offers competitive solutions. In particular, a robust FCFV solver for incompressible laminar flows is integrated in OpenFOAM and tested on a set of steady-state and transient benchmarks. The method outperforms standard simpleFoam and pimpleFoam algorithms in terms of optimal convergence, accuracy, stability, and robustness. Special attention is devoted to motivate and numerically demonstrate the ability of the FCFV method to treat non-orthogonal, stretched, and skewed meshes, where CCFV schemes exhibit shortcomings.

Davide Cortellessa, Matteo Giacomini, Antonio Huerta1/3/2025

arXiv:2408.09904v3 Announce Type: replace Abstract: Nowadays, harmful effects of climate change are becoming increasingly apparent. A vital issue that must be addressed is the generation of energy from non-renewable and often polluting sources. For this reason, the development of renewable energy sources is of great importance. Unfortunately, too rapid spread of renewables can disrupt stability of the power system and lead to energy blackouts. One should not simply support it, without ensuring sustainability and understanding of the diffusion process. In this research, we propose a new agent-based model of diffusion of photovoltaic panels. It is an extension of the q-voter model that utilizes a multi-layer network structure. The novelty is that both opinion dynamics and diffusion of innovation are studied simultaneously on a multidimensional structure. The model is analyzed using Monte Carlo simulations and the mean-field approximation. The impact of parameters and specifications on the basic properties of the model is discussed. Firstly, we show that for a certain range of parameters, innovation always succeeds, regardless of the initial conditions. Secondly, that the mean-field approximation gives qualitatively the same results as computer simulations, even though it does not utilize knowledge of the network structure.

Tomasz Weron1/3/2025

arXiv:2408.16709v2 Announce Type: replace Abstract: This paper establishes a data-driven modeling framework for lean Hydrogen (H2)-air reaction rates for the Large Eddy Simulation (LES) of turbulent reactive flows. This is particularly challenging since H2 molecules diffuse much faster than heat, leading to large variations in burning rates, thermodiffusive instabilities at the subfilter scale, and complex turbulence-chemistry interactions. Our data-driven approach leverages a Convolutional Neural Network (CNN), trained to approximate filtered burning rates from emulated LES data. First, five different lean premixed turbulent H2-air flame Direct Numerical Simulations (DNSs) are computed each with a unique global equivalence ratio. Second, DNS snapshots are filtered and downsampled to emulate LES data. Third, a CNN is trained to approximate the filtered burning rates as a function of LES scalar quantities: progress variable, local equivalence ratio and flame thickening due to filtering. Finally, the performances of the CNN model are assessed on test solutions never seen during training. The model retrieves burning rates with very high accuracy. It is also tested on two filter and downsampling parameters and two global equivalence ratios between those used during training. For these interpolation cases, the model approximates burning rates with low error even though the cases were not included in the training dataset. This a priori study shows that the proposed data-driven machine learning framework is able to address the challenge of modeling lean premixed H2-air burning rates. It paves the way for a new modeling paradigm for the simulation of carbon-free hydrogen combustion systems.

Quentin Mal\'e, Corentin J Lapeyre, Nicolas Noiray1/3/2025

arXiv:2501.00001v1 Announce Type: new Abstract: In this paper, a mathematical model is developed to describe the evolution of the concentration of compounds through a gas chromatography column. The model couples mass balances and kinetic equations for all components. Both single and multiple-component cases are considered with constant or variable velocity. Non-dimensionalisation indicates the small effect of diffusion. The system where diffusion is neglected is analysed using Laplace transforms. In the multiple-component case, it is demonstrated that the competition between the compounds is negligible and the equations may be decoupled. This reduces the problem to solving a single integral equation to determine the concentration profile for all components (since they are scaled versions of each other). For a given analyte, we then only two parameters need to be fitted to the data. To verify this approach, the full governing equations are also solved numerically using the finite difference method and a global adaptive quadrature method to integrate the Laplace transformation. Comparison with the Laplace solution verifies the high degree of accuracy of the simpler Laplace form. The Laplace solution is then verified against experimental data from BTEX chromatography. This novel method, which involves solving a single equation and fitting parameters in pairs for individual components, is highly efficient. It is significantly faster and simpler than the full numerical solution and avoids the computationally expensive methods that would normally be used to fit all curves at the same time.

A. Cabrera-Codony, A. Valverde, K. Born, O. A. I. Noreldin, T. G. Myers1/3/2025

arXiv:2501.00258v1 Announce Type: new Abstract: In optimizing real-world structures, due to fabrication or budgetary restraints, the design variables may be restricted to a set of standard engineering choices. Such variables, commonly called categorical variables, are discrete and unordered in essence, precluding the utilization of gradient-based optimizers for the problems containing them. In this paper, incorporating the Gumbel-Softmax (GSM) method, we propose a new gradient-based optimizer for handling such variables in the optimal design of large-scale frame structures. The GSM method provides a means to draw differentiable samples from categorical distributions, thereby enabling sensitivity analysis for the variables generated from such distributions. The sensitivity information can greatly reduce the computational cost of traversing high-dimensional and discrete design spaces in comparison to employing gradient-free optimization methods. In addition, since the developed optimizer is gradient-based, it can naturally handle the simultaneous optimization of categorical and continuous design variables. Through three numerical case studies, different aspects of the proposed optimizer are studied and its advantages over population-based optimizers, specifically a genetic algorithm, are demonstrated.

Mehran Ebrahimi, Hyunmin Cheong, Pradeep Kumar Jayaraman, Farhad Javid1/3/2025

arXiv:2501.00267v1 Announce Type: new Abstract: Heretofore, the Serret-Frenet frame has been the ubiquitous choice for analyzing the elastic deformations of beam elements. It is well known that this frame is undefined at the inflection points and straight segments of the beam where its curvature is zero, leading to singularities and errors in their numerical analysis. On the other hand, there exists a lesser-known frame called Bishop which does not have the caveats of the Serret-Frenet frame and is well-defined everywhere along the beam centerline. Leveraging the Bishop frame, in this paper, we propose a new spatial, singularity-free low-order beam element based on the absolute nodal coordinate formulation for both small and large deformation applications. This element, named ANCF14, has a constant mass matrix and can capture longitudinal, transverse (bending) and torsional deformations. It is a two-noded element with 7 degrees of freedom per node, which are global nodal coordinates, nodal slopes and their cross-sectional rotation about the centerline. The newly developed element is tested through four complex benchmarks. Comparing the ANCF14 results with theoretical and numerical results provided in other studies confirms the efficiency and accuracy of the proposed element.

Mehran Ebrahimi, Adrian Butscher, Hyunmin Cheong1/3/2025

arXiv:2501.00310v1 Announce Type: new Abstract: How to accurately quantify the uncertainty of stochastic dynamical responses affected by uncertain loads and structural parameters is an important issue in structural safety and reliability analysis. In this paper, the conditional uncertainty quantification analysis for the dynamical response of stochastic structures considering the measurement data with random error is studied in depth. A method for extracting the key measurement condition, which holds the most reference value for the uncertainty quantification of response, from the measurement data is proposed. Considering the key measurement condition and employing the principle of probability conservation and conditional probability theory, the quotient-form expressions for the conditional mean, conditional variance, and conditional probability density function of the stochastic structural dynamical response are derived and are referred to as the key conditional quotients (KCQ). A numerical method combining the non-equal weighted generalized Monte Carlo method, Dirac function smoothing technique, and online-offline coupled computational strategy is developed for calculating KCQs. Three linear/nonlinear stochastic dynamical examples are used to verify that the proposed KCQ method can efficiently and accurately quantify the uncertainty of the structural response considering measurement conditions. The examples also compare the traditional non-conditional uncertainty quantification results with the conditional uncertainty quantification results given by KCQs, indicating that considering measurement conditions can significantly reduce the uncertainty of the stochastic dynamical responses, providing a more refined statistical basis for structural safety and reliability analysis.

Feng Wu, Yuelin Zhao, Li Zhu1/3/2025

arXiv:2501.00611v1 Announce Type: new Abstract: This work proposes a gradient-based method to design bone implants using triply-periodic minimal surfaces (TPMS) of spatially varying thickness to maximize bone in-growth. Bone growth into the implant is estimated using a finite element based mechanobiological model considering the magnitude and frequency of in vivo loads, as well as the density distribution of the surrounding bone. The wall thicknesses of the implant unit cells are determined via linear interpolation of the thicknesses over a user defined grid of control points, avoiding mesh dependency and providing control over the sensitivity computation costs. The TPMS structure is modeled as a homogenized material to reduce computational cost. Local properties of the implant are determined at run-time on an element-by-element basis using a pre-constructed surrogate model of the TPMS's physical and geometric properties as a function of the local wall thickness and the density of in-grown bone. Design sensitivities of the bone growth within the implant are computed using the direct sensitivity method. The methodology is demonstrated on a cementless hip, optimizing the implant for bone growth subject to wall thickness constraints to ensure manufacturability and allow cell infiltration.

David Cohen (School of Mechanical, Aerospace, and Manufacturing Engineering, University of Connecticut), Juli\'an A. Norato (School of Mechanical, Aerospace, and Manufacturing Engineering, University of Connecticut)1/3/2025

arXiv:2501.01010v1 Announce Type: new Abstract: Predicting Bitcoin price remains a challenging problem due to the high volatility and complex non-linear dynamics of cryptocurrency markets. Traditional time-series models, such as ARIMA and GARCH, and recurrent neural networks, like LSTMs, have been widely applied to this task but struggle to capture the regime shifts and long-range dependencies inherent in the data. In this work, we propose CryptoMamba, a novel Mamba-based State Space Model (SSM) architecture designed to effectively capture long-range dependencies in financial time-series data. Our experiments show that CryptoMamba not only provides more accurate predictions but also offers enhanced generalizability across different market conditions, surpassing the limitations of previous models. Coupled with trading algorithms for real-world scenarios, CryptoMamba demonstrates its practical utility by translating accurate forecasts into financial outcomes. Our findings signal a huge advantage for SSMs in stock and cryptocurrency price forecasting tasks.

Mohammad Shahab Sepehri, Asal Mehradfar, Mahdi Soltanolkotabi, Salman Avestimehr1/3/2025

arXiv:2412.18564v1 Announce Type: new Abstract: Aircraft design optimization traditionally relies on computationally expensive simulation techniques such as Finite Element Method (FEM) and Finite Volume Method (FVM), which, while accurate, can significantly slow down the design iteration process. The challenge lies in reducing the computational complexity while maintaining high accuracy for quick evaluations of multiple design alternatives. This research explores advanced methods, including surrogate models, reduced-order models (ROM), and multi-fidelity machine learning techniques, to achieve more efficient aircraft design evaluations. Specifically, the study investigates the application of Multi-fidelity Physics-Informed Neural Networks (MPINN) and autoencoders for manifold alignment, alongside the potential of Generative Adversarial Networks (GANs) for refining design geometries. Through a proof-of-concept task, the research demonstrates the ability to predict high-fidelity results from low-fidelity simulations, offering a path toward faster and more cost effective aircraft design iterations.

Apurba Sarker12/25/2024

arXiv:2412.18441v1 Announce Type: new Abstract: This paper provides a normalized field product approach for topology optimization to achieve close-to-binary optimal designs. The method employs a parameter-free density measure that implicitly enforces a minimum length scale on the solid phase, allowing for smooth and transition-free topologies. The density evaluation does not rely on weight functions; however, the related density functions must have values between 0 and 1. The method combines the SIMP scheme and the introduced density function for material stiffness interpolation. The success and efficacy of the approach are demonstrated for designing both two- and three-dimensional designs, encompassing stiff structures and compliant mechanisms. The structure's compliance is minimized for the former, while the latter involves optimizing a multi-criteria objective. Numerical examples consider different volume fractions, length scales, and density functions. A volume-preserving smoothing and resolution scheme is implemented to achieve serrated-free boundaries. The proposed method is also seamlessly extended with advanced elements for solving 3D problems. The optimized designs obtained are close to binary without any user intervention while satisfying the desired feature size on the solid phase.

Nikhil Singh, Prabhat Kumar, Anupam Saxena12/25/2024

arXiv:2412.18174v1 Announce Type: new Abstract: Recent advancements have underscored the potential of large language model (LLM)-based agents in financial decision-making. Despite this progress, the field currently encounters two main challenges: (1) the lack of a comprehensive LLM agent framework adaptable to a variety of financial tasks, and (2) the absence of standardized benchmarks and consistent datasets for assessing agent performance. To tackle these issues, we introduce \textsc{InvestorBench}, the first benchmark specifically designed for evaluating LLM-based agents in diverse financial decision-making contexts. InvestorBench enhances the versatility of LLM-enabled agents by providing a comprehensive suite of tasks applicable to different financial products, including single equities like stocks, cryptocurrencies and exchange-traded funds (ETFs). Additionally, we assess the reasoning and decision-making capabilities of our agent framework using thirteen different LLMs as backbone models, across various market environments and tasks. Furthermore, we have curated a diverse collection of open-source, multi-modal datasets and developed a comprehensive suite of environments for financial decision-making. This establishes a highly accessible platform for evaluating financial agents' performance across various scenarios.

Haohang Li, Yupeng Cao, Yangyang Yu, Shashidhar Reddy Javaji, Zhiyang Deng, Yueru He, Yuechen Jiang, Zining Zhu, Koduvayur Subbalakshmi, Guojun Xiong, Jimin Huang, Lingfei Qian, Xueqing Peng, Qianqian Xie, Jordan W. Suchow12/25/2024

arXiv:2412.18350v1 Announce Type: new Abstract: Density Functional Theory (DFT) stands as a widely used and efficient approach for addressing the many-electron Schr\"odinger equation across various domains such as physics, chemistry, and biology. However, a core challenge that persists over the long term pertains to refining the exchange-correlation (XC) approximation. This approximation significantly influences the triumphs and shortcomings observed in DFT applications. Nonetheless, a prevalent issue among XC approximations is the presence of systematic errors, stemming from deviations from the mathematical properties of the exact XC functional. For example, although both B3LYP and DM21 (DeepMind 21) exhibit improvements over previous benchmarks, there is still potential for further refinement. In this paper, we propose a strategy for enhancing XC approximations by estimating the neural uncertainty of the XC functional, named Residual XC-Uncertain Functional. Specifically, our approach involves training a neural network to predict both the mean and variance of the XC functional, treating it as a Gaussian distribution. To ensure stability in each sampling point, we construct the mean by combining traditional XC approximations with our neural predictions, mitigating the risk of divergence or vanishing values. It is crucial to highlight that our methodology excels particularly in cases where systematic errors are pronounced. Empirical outcomes from three benchmark tests substantiate the superiority of our approach over existing state-of-the-art methods. Our approach not only surpasses related techniques but also significantly outperforms both the popular B3LYP and the recent DM21 methods, achieving average RMSE improvements of 62\% and 37\%, respectively, across the three benchmarks: W4-17, G21EA, and G21IP.

Sizhuo Jin, Shuo Chen, Jianjun Qian, Ying Tai, Jun Li12/25/2024

arXiv:2411.03474v3 Announce Type: replace Abstract: Automated analysis of high-resolution transmission electron microscopy (HRTEM) images is increasingly essential for advancing research in organic electronics, where precise characterization of nanoscale crystal structures is crucial for optimizing material properties. This paper introduces an open-source computational framework called GRATEV2.0 (GRaph-based Analysis of TEM), designed for real-time analysis of HRTEM data, with a focus on characterizing complex microstructures in conjugated polymers, illustrated using Poly[N-9'-heptadecanyl-2,7-carbazole-alt-5,5-(4',7'-di-2-thienyl-2',1',3'-benzothiadiazole)] (PCDTBT), a key material in organic photovoltaics. GRATEV2.0 employs fast, automated image processing algorithms, enabling rapid extraction of structural features like d-spacing, orientation, and crystal shape metrics. Gaussian process optimization rapidly identifies the user-defined parameters in the approach, reducing the need for manual parameter tuning and thus enhancing reproducibility and usability. Additionally, GRATEV2.0 is compatible with high-performance computing (HPC) environments, allowing for efficient, large-scale data processing at near real-time speeds. A unique feature of GRATEV2.0 is a Wasserstein distance-based stopping criterion, which optimizes data collection by determining when further sampling no longer adds statistically significant information. This capability optimizes the amount of time the TEM facility is used while ensuring data adequacy for in-depth analysis. Open-source and tested on a substantial PCDTBT dataset, this tool offers a powerful, robust, and accessible solution for high-throughput material characterization in organic electronics.

Dhruv Gamdha, Ryan Fair, Adarsh Krishnamurthy, Enrique Gomez, Baskar Ganapathysubramanian12/25/2024

arXiv:2408.07843v2 Announce Type: replace Abstract: There is a continuing interest in using standard language constructs for accelerated computing in order to avoid (sometimes vendor-specific) external APIs. For Fortran codes, the {\tt do concurrent} (DC) loop has been successfully demonstrated on the NVIDIA platform. However, support for DC on other platforms has taken longer to implement. Recently, Intel has added DC GPU offload support to its compiler, as has HPE for AMD GPUs. In this paper, we explore the current portability of using DC across GPU vendors using the in-production solar surface flux evolution code, HipFT. We discuss implementation and compilation details, including when/where using directive APIs for data movement is needed/desired compared to using a unified memory system. The performance achieved on both data center and consumer platforms is shown.

Ronald M. Caplan, Miko M. Stulajter, Jon A. Linker, Jeff Larkin, Henry A. Gabb, Shiquan Su, Ivan Rodriguez, Zachary Tschirhart, Nicholas Malaya12/25/2024