The book is interwoven according to the intrinsic logics of modern most important applications of electrospun nanofibers. It discusses such application-oriented nanofibers as self-healing vascular nanotextured materials, biopolymer nanofibers, soft robots and actuators based on nanofibers, biopolymer nanofiber-based triboelectric nanogenerators, metallized nanofibers, and heaters and sensors based on them. It also includes such topics as the injectable nanofibrous biomaterials, fibrous hemostatic agents and their interaction with blood, as well as electrospun nanofibers for face-mask applications. The book also details polyelectrolytes-based complex nanofibers and their use as actuators. It also covers drug release facilitated by polyelectrolytes-based complex nanofibers. The fundamental aspects of electrospinning of polymer nanofibers discussed in the final part of the book link them to the applications described in the preceding chapters. Such topics as polymer solution preparation and their rheological properties, e.g., viscoelasticity and the related spinnability, the electrical conductivity of polymer solutions, and the cascade of the physical phenomena resulting in formation of nanofibers encompass the experimental aspects. Also, the general quasi-1D equations used for modeling of formation of electrospun polymer nanofibers, and the numerical aspects of their solution are discussed in detail, including such modeling-driven applications as nanofiber alignment by electric focusing fields.
Early detection of prediabetes is crucial to preventing its progression to diabetes. Providing individuals with a personalized sense of their risk could improve prevention efforts. While complex mathematical models that simulate metabolic and inflammatory processes offer detailed and patient-specific insights, their computational cost usually makes them impractical for real-time prediction on mobile platforms. This work introduces a long short-term memory (LSTM) surrogate for the MT2D model, that simulates the main metabolic and inflammatory processes undergoing the transition to prediabetes. The model is developed using a dataset of 43 669 simulated subjects, each with lifestyle inputs and biomarker outputs over six months. Using 8 time series inputs, the surrogate predicts the dynamics of 11 key metabolic and inflammatory outputs, closely replicating the behaviour of the MT2D model. After training, the proposed LSTM model reduces computational time from an average of 8.4 hours to 0.1 seconds per simulation, making it suitable for mobile device deployment. The model achieves root mean squared errors on the order of 10-2 on scaled data, and shows promise for prediabetes risk assessment by capturing trends in inflammatory biomarkers. This surrogate model can provide real-time and patient-specific insights into the metabolic health, potentially improving the understanding of prediabetes risk.
Surrogate
LSTM
Prediabetes
Risk
Input to Output Prediction
Dynamical System
Monitoring surface and vegetation conditions is crucial for analyzing the impact of climate change on natural resources, especially in regions susceptible to extreme events like land and forest dryness caused by summer heatwaves. Traditional satellite indices, including NDVI, have limitations in distinguishing between barren soil and distressed vegetation. This study shows the potential of two recently validated indices, the Emissivity Contrast Index (ECI) and the Water Deficit Index (WDI), to assess vegetation stress and woodland degradation. These indices, derived from Infrared Atmospheric Sounding Interferometer (IASI) data, utilize an Optimal Interpolation scheme for upscaling and remapping. The effectiveness of ECI and WDI has been validated through a comparison with Surface Soil Moisture (SSM). The methodology allows for simultaneous assessment of surface hydric stress, identifying regions at risk of drought and forest fires. This approach has been applied to southern Italy during year 2023, an area which has been impacted by strong heatwaves in the last decade. These indices could demonstrate significant effectiveness when estimated using high-resolution sounders, such as the Surface Biology and Geology Observing Terrestrial Thermal Emission Radiometer (SBG OTTER). This would allow for more effective monitoring of small, heterogeneous areas.
Emissivity
Infrared
Satellite
Soil Water Stress
Vegetation stress
Since the Laplace transform plays a central role in the solution of differential equations, it seems natural to extend it in the field of fractional calculus, since many applications of this topic have been proposed, and are becoming more and more important. In this paper we extend the classical Laplace Transform by replacing the usual kernel with a suitable one, both in the classical and Laguerre-type case, obtained by constructing the reciprocal of some exponential-type functions with respect to an appropriate differential operator. Some examples are shown, derived using the computer algebra system Mathematica.
Electroencephalography (EEG) source imaging aims to reconstruct brain activity maps from the neuroelectric potential difference measured on the skull. To obtain the brain activity map, we need to solve an ill-posed and ill-conditioned inverse problem that requires regularization techniques to make the solution viable. When dealing with real-time applications, dimensionality reduction techniques can be used to reduce the computational load required to evaluate the numerical solution of the EEG inverse problem. To this end, in this paper we use the random dipole sampling method, in which a Monte Carlo technique is used to reduce the number of neural sources. This is equivalent to reducing the number of the unknowns in the inverse problem and can be seen as a first regularization step. Then, we solve the reduced EEG inverse problem with two popular inversion methods, the weighted Minimum Norm Estimate (wMNE) and the standardized LOw Resolution brain Electromagnetic TomogrAphy (sLORETA). The main result of this paper is the error estimates of the reconstructed activity map obtained with the randomized version of wMNE and sLORETA. Numerical experiments on synthetic EEG data demonstrate the effectiveness of the random dipole sampling method.
EEG imaging
inversion method
random sampling
sLORETA
underdetermined inverse problem
wMNE
Using a mathematical model of concrete carbonation that describes the variation in porosity as a consequence of the involved chemical reactions, we both validated and calibrated the related numerical algorithm of degradation. Once calibrated, a simulation algorithm was used as a forecasting tool for predicting the effects on the porosity of concrete exposed to increasing levels of CO2 emissions, as well as to rising temperatures. Taking into account future projections of environmental modifications deriving from climate changes, some scenarios were produced numerically by the mathematical algorithm that showed the effects of different pollution levels and global warming on the porosity of Portland cement in a time window of years. Finally, a theoretical study on the effects of pollution levels on the carbonation constant determining the advancement in the carbonation front was carried out for the analyzed scenarios.
concrete carbonation
reaction and diffusion models
climate changes
model parameter estimation
mathematical algorithms
The computation of n-point Gaussian quadrature rules for symmetric weight functions is considered in this paper. It is shown that the nodes and the weights of the Gaussian quadrature rule can be retrieved from the singular value decomposition of a bidiagonal matrix of size n/2. The proposed numerical method allows to compute the nodes with high relative accuracy and a computational complexity of O(n). We also describe an algorithm for computing the weights of a generic Gaussian quadrature rule with high relative accuracy. Numerical examples show the effectiveness of the proposed approach.
LENTO: Unpredictable Latency-based continuous authEntication for Network inTensive IoT envirOnments
Mohammed AlSadi
;
Roberto Di Pietro
;
Flavio Lombardi
;
Matteo Signorini
Started as a hyped technology a few years ago, IoT is now a reality providing sensing and computing capabilities from SCADA systems to households. At their core, IoT devices connect to the outside world to share sensed or computed data. However, the sensitivity and privacy of shared data has made access management a stringent need also for the IoT. In particular, continuous authentication could solve a few security issues, like session hijacking, via checking device legitimacy for each exchanged message and preventing attackers from pretending their actions came from authenticated devices. To date, device-to-device (D2D) continuous authentication still relies on tokens/certificates or devices' fingerprints such as battery levels or location. The cited solutions, while being not always implementable on resource constrained devices, provide low-entropy and thus sporting a non negligible probability of being guessable during impersonation attacks. In this paper, we overcome the above limitations with LENTO: unpredictable Latency-based continuous authEntication for Network inTensive IoT envirOnments. In addition to a thorough analysis, we also offer experimental validation of our proposal. We have deployed LENTO as an additional authentication module of the well-known NextCloud platform, and we have performed an extensive experimental campaign. Collected results confirm our working hypothesis. Network delays can be exploited as random seeds in continuous authentication protocols as they provide as much entropy as standard approaches. To the best of our knowledge, our approach is the first continuous authentication protocol relying purely on the network characteristics, regardless of the underneath computing base trustworthiness. Given the minimal overhead introduced by our solution, it provides continuous authentication even for those devices that cannot afford to run (defacto) standard protocols. As such, LENTO could be retrofitted, offering enhanced security to a plethora of nowadays unsecured devices.
Continuous authentication
Internet of Things (IoT)
Security
Latency
Cloud
Network
This work introduces a novel extension of the 0/1 Knapsack Problem in which we consider the existence of so-called forfeit sets. A forfeit set is a subset of items of arbitrary cardinality, such that including a number of its elements that exceeds a predefined allowance threshold implies some penalty costs to be paid in the objective function value. A global upper bound on these allowance violations is also considered. We show that the problem generalizes both the Knapsack Problem with conflicts among item pairs and the Knapsack Problem with forfeit pairs, that have been previously introduced in the literature. We present a polynomial subcase by proving the integrality of its LP relaxation polytope and, we introduce three heuristic approaches, namely a constructive greedy, an algorithm based on the recently introduced Carousel Greedy paradigm and a hybrid Memetic/Carousel Greedy algorithm. Finally, we validate the performances for the proposed algorithms on a set of benchmark instances that consider both random and correlated data.
Knapsack Problem
Conflicts
Forfeit sets
Carousel Greedy
Memetic algorithm
The IAS-MEEG Package: A Flexible Inverse Source Reconstruction Platform for Reconstruction and Visualization of Brain Activity from M/EEG Data
Calvetti Daniela
;
Pascarella Annalisa
;
Pitolli Francesca
;
Somersalo Erkki
;
Vantaggi Barbara
We present a standalone Matlab software platform complete with visualization for the reconstruction of the neural activity in the brain from MEG or EEG data. The underlying inversion combines hierarchical Bayesian models and Krylov subspace iterative least squares solvers. The Bayesian framework of the underlying inversion algorithm allows to account for anatomical information and possible a priori belief about the focality of the reconstruction. The computational efficiency makes the software suitable for the reconstruction of lengthy time series on standard computing equipment. The algorithm requires minimal user provided input parameters, although the user can express the desired focality and accuracy of the solution. The code has been designed so as to favor the parallelization performed automatically by Matlab, according to the resources of the host computer. We demonstrate the flexibility of the platform by reconstructing activity patterns with supports of different sizes from MEG and EEG data. Moreover, we show that the software reconstructs well activity patches located either in the subcortical brain structures or on the cortex. The inverse solver and visualization modules can be used either individually or in combination. We also provide a version of the inverse solver that can be used within Brainstorm toolbox. All the software is available online by Github, including the Brainstorm plugin, with accompanying documentation and test data.
Objective There is increasing interest in simultaneous endovascular delivery of more than one drug from a drug-loaded
stent into a diseased artery. There may be an opportunity to obtain a therapeutically desirable uptake profile of the two drugs
over time by appropriate design of the initial drug distribution in the stent. Due to the non-linear, coupled nature of diffusion
and reversible specific/non-specific binding of both drugs as well as competition between the drugs for a fixed binding site
density, a comprehensive numerical investigation of this problem is critically needed.
Methods This paper presents numerical computation of dual drug delivery in a stent-artery system, accounting for diffusion
as well as specific and non-specific reversible binding. The governing differential equations are discretized in space, followed
by integration over time using a stiff numerical solver. Three different cases of initial dual drug distribution are considered.
Results For the particular case of sirolimus and paclitaxel, results show that competition for a limited non-specific binding
site density and the significant difference in the forward/backward reaction coefficients play a key role in determining the
nature of drug uptake. The nature of initial distribution of the two drugs in the stent is also found to influence the binding
process, which can potentially be used to engineer a desirable dual drug uptake profile.
Conclusions These results help improve the fundamental understanding of endovascular dual drug delivery. In addition, the
numerical technique and results presented here may be helpful for designing and optimizing other drug delivery problems
as well.
dual drug delivery
mathematical modelling
binding
stent
We present a new image scaling method both for downscaling and upscaling, running with any scale factor or desired size. The resized image is achieved by sampling a bivariate polynomial which globally interpolates the data at the new scale. The method's particularities lay in both the sampling model and the interpolation polynomial we use. Rather than classical uniform grids, we consider an unusual sampling system based on Chebyshev zeros of the first kind. Such optimal distribution of nodes permits to consider near-best interpolation polynomials defined by a filter of de la Vallée-Poussin type. The action ray of this filter provides an additional parameter that can be suitably regulated to improve the approximation. The method has been tested on a significant number of different image datasets. The results are evaluated in qualitative and quantitative terms and compared with other available competitive methods. The perceived quality of the resulting scaled images is such that important details are preserved, and the appearance of artifacts is low. Competitive quality measurement values, good visual quality, limited computational effort, and moderate memory demand make the method suitable for real-world applications.
Image downscaling
Image upscaling
de la Vallée-Poussin interpolation
Chebyshev nodes
This paper presents a new software framework for solving large and sparse linear systems on current hybrid architectures, from small servers to high-end supercomputers, embedding multi-core CPUs and Nvidia GPUs at the node level. The framework has a modular structure and is composed of three main components, which separate basic functionalities for managing distributed sparse matrices and executing some sparse matrix computations involved in iterative Krylov projection methods, eventually exploiting multi-threading and CUDA-based programming models, from the functionalities for setup and application of different types of one-level and multi-level algebraic preconditioners.
In a subsoil bioremediation intervention air or oxygen is injected in the polluted region and then a model for unsaturated porous media it is required, based on the theory
of the dynamics of multiphase fluids in porous media.
In order to optmize the costs of the intervention it is useful to consider the gas as compressible and this fact introduces nonlinearity in the mathematical model.
The physical problem is described by a system of equations and the unknowns are: pollutant; bacteria concentration; oxygen saturation and oxygen pressure.
Then, by algebraic manipulations, the model is reduced a to a nonlinear system of partial differential equations describing: oxygen saturation, oxygen density and bacteria concentration. For the proposed model, the results of some simulation experiments performed using COMSOL Multiphysics will be shown.
porous media
subsoil bioremediation
mathematical models
We present a mathematical model describing the evolution of sea ice and meltwater during summer. The system is described by two coupled partial differential equations for the ice thickness h(x,t) and pond depth w(x,t) fields. The model is similar, in principle, to the one put forward by Luthije et al. (2006), but it features i) a modified melting term, ii) a non-uniform seepage rate of meltwater through the porous ice medium and a minimal coupling with the atmosphere via a surface wind shear term, ?s (Scagliarini et al. 2020). We test, in particular, the sensitivity of the model to variations of parameters controlling fluid- dynamic processes at the pond level, namely the variation of turbulent heat flux with pond depth and the lateral melting of ice enclosing a pond. We observe that different heat flux scalings determine different rates of total surface ablations, while the system is relatively robust in terms of probability distributions of pond surface areas. Finally, we study pond morphology in terms of fractal dimensions, showing that the role of lateral melting is minor, whereas there is evidence of an impact from the initial sea ice topography.
sea-ice
melt ponds
wind shear
heat flux
probability distribution
surface ablation
This booklet contains all the abstracts of the results which are going to be presented at the IMACS World Congress taking place in Rome at theEngineering Faculty of University 'La Sapienza', September 11-15, 2023.This is the 21st one in a series of World Conferences whose complete list goes back to 1955 and covered the whole continents. The subsequent World Conferences, usually, take place every three years. Unfortunately, due to the COVID pandemics, IMACS2023, initially scheduled in 2020, had to be postponed also in order to follow the spirit of the IMACS World Conference which prescribes to gather scientists in presence from all over the world. So, in the present occasion, the whole participants are expected to convene in Rome for exchanging their works, ideas and experiences.This Book of Abstracts, reflecting the Congress structure, is organized in sections: Keynote Lectures, General Session, Mini-symposia, Special Sessions and Posters. According to the IMACS philosophy, different aspects of applied mathematics are represented with a special interest towards the numerical methods and solutions.
Estimates networks of conditional dependencies (Gaussian graphical models) from multiple classes of data (similar but not exactly, i.e. measurements on different equipment, in different locations or for various sub-types). Package also allows to generate simulation data and evaluate the performance. Implementation of the method described in Angelini, De Canditiis and Plaksienko (2022) <doi:10.3390/math10213983>.
Graphical Model
Estimation Network
Group Lasso penalty
Background: Major Depressive Disorder (MDD) is a psychiatric illness that is often associated with potentially life -threatening physiological changes and increased risk for suicidal behavior. Electroencephalography (EEG) research suggests an association between depression and specific frequency imbalances in the frontal brain re-gion. Further, while recently developed technology has been proposed to simplify EEG data acquisition, more research is still needed to support its use in patients with MDD.Methods: Using the 14-channel EMOTIV EPOC cap, we recorded resting state EEG from 15 MDD patients with suicidal ideation (SI) vs. 12 healthy controls (HC) to investigate putative power spectral density (PSD) between -group differences at the F3 and F4 electrode sites. Specifically, we explored 1) between-group alpha power asymmetries (AA), 2) between-group differences in delta, theta, alpha and beta power, 3) correlations between PSD data and scores in the Beck's Depression Inventory-II (BDI-II), Beck's Anxiety Inventory (BAI), Reasons for Living Inventory (RFL), and Self-Disgust Questionnaire (SDS).Results: When compared to HC, patients had higher scores on the BAI (p = 0.0018), BDI-II (p = 0.0001) or SDS (p = 0.0142) scale and lower scores in the RFL (p = 0.0006) scale. The PSD analysis revealed no between-group difference or correlation with questionnaire scores for any of the measures considered.Conclusions: The present study could not confirm previous research suggesting frequency-specific anomalies in depressed persons with SI but might suggest that frontal EEG imbalances reflect greater anxiety and negative self -referencing. Future studies should confirm these findings in a larger population sample.