首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
The pyrolysis models undergoes a fast development due to the enhancement of both the computational power and the new test used to characterize the behaviour of materials under thermal stresses, which were widely used in the field of the chemical engineering to obtain the reaction rates. Thus, these models allow us to characterize either the transient heat of material (thermal inertia) or the complete chemical scheme of mass loss processes (kinetic triplet). The pyrolysis model needs a several number of parameters what does the optimization of a suitable set of parameters a difficult task. Two kinds of materials have been investigated; the first one was a real material which mass loss process was characterized as “one-step” reaction and the second one as a “two-steps” process. Further, it has been analyzed the influence of some algorithm features (initial population number, parameter range, crossover influence) in the optimization time and also in the accuracy of results.  相似文献   

2.
Obtaining pressure radiated by flat surfaces is not a new problem. This problem has studied by the complexity of the topic and its application to design flat speakers. These kind of speakers are the speakers that we use in televisions, ceilings, cinema screens, panels, etc. in this cases usually we have rectangular speakers.The single source model is the simplest model to convert the surface vibration to radiated pressure in a point. This is an easy model but it is very slow, especially when we want calculate at high frequencies. For rectangular surfaces there are models that use relatively complex auxiliary functions. In this case the calculation is accelerated but its implementation is more complicated and is necessary to particularize each situation.This paper presents the decomposition of a rectangular surface in several circular surfaces, by means of area associations, seeking a rapid method based on circular pistons whose behavior is known with an acceptable error in the allocation.  相似文献   

3.
We present a novel fully explicit time integration method that remains stable for large time steps, requires neither matrix inversions nor solving a system of equations and therefore allows for nearly effort-less parallelization. In this first paper the proposed approach is applied to solve conduction heat transfer problems, showing that it is stable for any time step as is the case with implicit methods but with a much lower computation time.  相似文献   

4.
This paper presents a new procedure to deal with the delamination problem found in laminated composites, based in a continuum mechanics formulation. The procedure proposed obtains the composite constitutive performance with the Serial/Parallel mixing theory, developed by F. Rastellini. This theory characterizes composite materials by coupling the constitutive behaviour of the composite components, imposing an iso–strain relation among the components in the fibre (or parallel) direction and an iso-stress relation in the remaining directions (serial directions). The proposed procedure uses a damage formulation to characterize the constitutive behaviour of matrix component in order to obtain the stress-strain performance of this material.With these two formulations, the delamination phenomenon is characterized naturally by the numerical simulation, being unnecessary the definition of special elements or computationally expensive techniques like the definition of contact elements or mesh separation. Matrix failure, as a result of the stress state found in it, leads to a reduction of the stiffness and strength capacity of the composite in its serial directions, among them, the shear component. This stiffness reduction provides a composite performance equivalent to what is found in a delaminated material.To prove the ability of the formulation proposed to solve delamination problems, the End Notch Failure test is numerically simulated and the results obtained are compared with experimental ones. The agreement found in the results with both simulations, numerical and experimental, validate the proposed methodology to solve the delamination problem.  相似文献   

5.
The possibilities of computational methods for assessing the response of cable supported bridges under wind action are considered in this work. The main objective is to study the possibilities of substituting wind tunnel campaigns by computer based analyses, particularly at the early design stage. The preliminary proposed design for a continuous cable-stayed bridge with two main spans of 650 m and a single box girder deck has been considered as a case study. The force coefficients of the deck cross-section have been computed and the unsteady response associated to vortex-shedding has been simulated using CFD commercial software. Furthermore, an in-house piece of software has been employed to obtain the response for flutter and buffeting phenomena adopting the hybrid approach; with that purpose the experimental flutter functions of a similar box girder deck were adopted. The computational results have been validated by comparison with similar experimental results published by other researchers. It has been verified that the set of adopted methods offers reliable results with moderate costs; therefore, the proposed approach is very suitable at the early design stage of long span bridges or at conceptual design works.  相似文献   

6.
This paper shows the solution to the problem of seismic wave propagation in 2-D using generalized finite difference (GFD) explicit schemes. Regular and irregular meshes can be used with this method.As we are using an explicit method, it is necessary to obtain the stability condition by using the von Neumann analysis. We also obtained the star dispersion formulas for the phase velocities for the P and S waves, as well as the ones for the group velocities.As the control over the irregularity in the mesh is very important in the application of this method, we have defined an index of irregularity for the star (IIS) and another for the cloud (IIC), analyzing its relationship with the dispersion and time step used in the calculations.  相似文献   

7.
Optimum design of structures has been traditionally focused on the analysis of shape and dimensions optimization problems. However, more recently a new discipline has emerged: the topology optimization of the structures. This discipline states innovative models that allow to obtain optimal solutions without a previous definition of the type of structure being considered. These formulations obtain the optimal topology and the optimal shape and size of the resulting elements. The most usual formulations of the topology optimization problem try to obtain the structure of maximum stiffness. These approaches maximize the stiffness for a given amount of material to be used. These formulations have been widely analyzed and applied in engineering but they present considerable drawbacks from a numerical and from a practical point of view. In this paper the author propose a different formulation, as an alternative to maximum stiffness approaches, that minimizes the weight and includes stress constraints. The advantages of this kind of formulations are crucial since the cost of the structure is minimized, which is the most frequent objective in engineering, and they guarantee the structural feasibility since stresses are constrained. In addition, this approach allows to avoid some of the drawbacks and numerical instabilities related to maximum stiffness approaches. Finally, some practical examples have been solved in order to verify the validity of the results obtained and the advantages of the proposed formulation.  相似文献   

8.
This work analyzes the influence of the discretization error contained in the Finite Element (FE) analyses of each design configuration proposed by the structural shape optimization algorithms over the behavior of the algorithm. The paper clearly shows that if FE analyses are not accurate enough, the final solution provided by the optimization algorithm will neither be optimal nor satisfy the constraints. The need for the use of adaptive FE analysis techniques in shape optimum design will be shown. The paper proposes the combination of two strategies to reduce the computational cost related to the use of mesh adaptivity in evolutionary optimization algorithms: (a) the use of the algorithm described by Bugeda et al. [1] which reduces the computational cost associated to the adaptive FE analysis of each geometrical configuration and, (b) the successive increase of the required accuracy of the FE analyses in order to obtain a considerable reduction of the computational cost in the early stages of the optimization process.  相似文献   

9.
The verification of the slender compressed members of a structure is a well formulated and codified problem. However, the non-linearity of the design process, combined with the no linearity of the physical problem requires an iterative process to define the piece that solves a given problem. This paper demonstrates an identity relationship between seemingly different problems, based on which we develop a theoretical approach and a simple procedure to determine, without iteration and with little margin for error, the piece needed for a given problem. The paper provides a statistical model to check that the provided procedure is safe and effective in the 98% of the problems in buildings —pillars and bars from trusses—. The presented approach therefore greatly facilitates decisions in this area.  相似文献   

10.
The motivation of this work is the modeling of the hardening precipitate and hardness evolutions of fully hardened heat treatable aluminium alloys during friction stir welding (FSW) and/or heat treatment processes. The model used is based on the kinetics of dissolution of precipitates model for hardened aluminium alloys given by Myhr and Grong (1991). This model contains a single independent variable, the time, and a single state variable, the volume fraction of hardening precipitates. A key point of this model is the identification of the effective activation energy for precipitates dissolution and the master curve defining the model, which was given by a look-up table. The goal of this work is to find an estimation of the effective activation energy and to model the dissolution rate of hardening precipitate in aluminium alloys using neural networks, avoiding the use of look-up tables. For this purpose a new and more convenient parametrization of the master curve is defined, a neural networks class is proposed, an objective functional is defined and a variational problem including independent parameters is solved. The novel methodology has been applied to different aluminium alloys, including the AA 6005A T6, AA 7449 T79 and AA 2198 T8. Experimental tests have been carried out for those aluminium alloys in order to get the HV1 hardness after isothermal heat treatments at different temperatures and for different treatment time durations. The effective activation energy for hardening precipitates dissolution and the master curve of the model have been obtained, using different network architectures, for the aluminium alloys considered in this work.  相似文献   

11.
A general methodology to optimize the weight of power transmission structures is presented in this article. This methodology is based on the simulated annealing algorithm defined by Kirkpatrick in the early ‘80s. This algorithm consists of a stochastic approach that allows to explore and analyze solutions that do not improve the objective function in order to develop a better exploration of the design region and to obtain the global optimum. The proposed algorithm allows to consider the discrete behavior of the sectional variables for each element and the continuous behavior of the general geometry variables. Thus, an optimization methodology that can deal with a mixed optimization problem and includes both continuum and discrete design variables is developed. In addition, it does not require to study all the possible design combinations defined by discrete design variables. The algorithm proposed usually requires to develop a large number of simulations (structural analysis in this case) in practical applications. Thus, the authors have developed first order Taylor expansions and the first order sensitivity analysis involved in order to reduce the CPU time required. Exterior penalty functions have been also included to deal with the design constraints. Thus, the general methodology proposed allows to optimize real power transmission structures in acceptable CPU time.  相似文献   

12.
This study introduces a new algorithm for the ant colony optimization (ACO) method, which has been proposed to solve global optimization problems with continuous decision variables. This algorithm, namely ACO-FRS, involves a strategy for the selection of feasible regions during optimization search and it performs the exploration of the search space using a similar approach to that used by the ants during the search of food. Four variants of this algorithm have been tested in several benchmark problems and the results of this study have been compared with those reported in literature for other ACO-type methods for continuous spaces. Overall, the results show that the incorporation of the selection of feasible regions allows the performing of a global search to explore those regions with low level of pheromone, thus increasing the feasibility of ACO for finding the global optimal solution.  相似文献   

13.
The development of more and more potent computers and the recent research in the field of hidroinformatics makes possible the free surface flow modelling in two dimensions caused by earthen dam failures. In this paper, the results obtained by uni-dimensional model (HEC-RAS) and two-dimensional model (CARPA) are compared. The use of the HEC-RAS software assumes the hypothesis of unidimensionality to be true, no infiltration and existence of a minimal initial flow. The comparison is made by analyzing the effect of these hypothesis in the downstream flow hydrographs. The used models reproduce the water discharge generated by a possible failure of dam number 1 of the 5th District of the Segarra-Garrigues Irrigation Project in the Ebro river basin in Spain.  相似文献   

14.
Currently, many structures existing in seismic areas are highly vulnerable because they have been built without the use of seismic design codes or by using outdated codes. Often, methods for assessing the vulnerability of the structures do not take into account that their seismic behavior is dynamic and highly nonlinear and, moreover, that the structural characteristics and action have large uncertainties. This article aims to assess the vulnerability of structures taking into account that the mechanical properties of materials and the seismic action are random variables, by using advanced techniques based on the Monte Carlo method and on the nonlinear stochastic dynamics. The results obtained with these techniques are compared with those corresponding to a standard vulnerability assessment, based on deterministic models, in order to highlight the differences between both approaches. The main conclusion of this work is the need to address the vulnerability assessment problem from a probabilistic perspective which, combined with advanced nonlinear static and dynamic structural analysis techniques, provides a powerful tool giving information impossible to be captured by means of deterministic models. Finally, detailed results obtained for a building with waffle slabs, which is a structural typology widely used in Spain, are included and discussed.  相似文献   

15.
The recent requirements of Spanish regulations and directives, on their turn based on European directives, have led to the development of a new two dimensional open channel flow modelling tool. The tool, named Iber, combines a hydrodynamic module, a turbulence module and a sediment transport module, and is based in the finite volume method to solve the involved equations. The simulation code has been integrated in a pre-process and post-process interface based on GiD software, developed by CIMNE. The result is a flow and sediment modelling system for rivers and estuaries that uses advanced numerical schemes, robust and stable, which are especially suitable for discontinuous flows taking place in torrential and hydrologically irregular rivers.  相似文献   

16.
17.
Uncertainties are a daily issue to deal with in aerospace engineering and applications. Robust optimization methods commonly use a random generation of the inputs and take advantage of multi-point criteria to look for robust solutions accounting with uncertainty definition. From the computational point of view, the application to coupled problems, like fluid-dynamics (CFD) or fluid-structure interaction (FSI), can be extremely expensive. This work presents a coupling between stochastic analysis techniques and evolutionary optimization algorithms for the definition of a stochastic robust optimization procedure. At first, a stochastic procedure is proposed to be applied into optimization problems. The proposed method has been applied to both CFD and FSI problems for the reduction of drag and flutter, respectively.  相似文献   

18.
19.
This work presents a numerical study on the turbulent flow of air with dispersed water droplets in separators of mechanical cooling towers. The averaged Navier-Stokes equations are discretised through a finite volume method, using the Fluent and Phoenics codes, and alternatively employing the turbulence models k ? ?, k ? ω and the Reynolds stress model, with low-Re version and wall enhanced treatment refinements. The results obtained are compared with numerical and experimental results taken from the literature. The degree of accuracy obtained with each of the considered models of turbulence is stated. The influence of considering whether or not the simulation of the turbulent dispersion of droplets is analyzed, as well as the effects of other relevant parameters on the collection efficiency and the coefficient of pressure drop. Focusing on four specific eliminators (‘Belgian wave’, ‘H1-V’, ‘L-shaped’ and ‘Zig-zag’), the following ranges of parameters are outlined: 1  Ue  5 m/s for the entrance velocity, 2  Dp  50 μm for the droplet diameter, 650  Re  8.500 for Reynolds number, and 0.05  Pi  5 for the inertial parameter. Results reached alternately with Fluent and Phoenics codes are compared. The best results correspond to the simulations performed with Fluent, using the SST k ? ω turbulence model, with values of the dimensionless scaled distance to wall y+ in the range 0.2 to 0.5. Finally, correlations are presented to predict the conditions for maximum collection efficiency (100 %), depending on the geometric parameter of removal efficiency of each of the separators, which is introduced in this work.  相似文献   

20.
This article shows an efficient implementation of a dynamic semi-recursive formulation for large and complex multibody system simulations, with interesting applications in the automotive field and especially with industrial vehicles. These systems tend to have a huge amount of kinematic constraints, becoming usual the presence of redundant but compatible systems of equations. The maths involved in the solution of these problems have a high computational cost, making very challenging to achieve real-time simulations.In this article, two implementations to increase the efficiency of these computations will be shown. The difference between them is the way they consider the Jacobian matrix of the constraint equations. The first one treats this matrix as a dense one, using the BLAS functions to solve the system of equations. The second one takes into account the sparse pattern of the Jacobian matrix, introducing the sparse function MA48 from Harwell.Both methodologies have been applied on two multibody system models with different sizes. The first model is a vehicle IVECO DAILY 35C15 with 17 degrees of freedom. The second one is a semi-trailer truck with 40 degrees of freedom. Taking as a reference the standard C/C + + implementation, the efficiency improvements that have been achieved using dense matrices (BLAS) have been of 15% and 50% respectively. The results in the first model have not improved significantly by using sparse matrices, but in the second one, the times with sparse matrices have been reduced 8% with respect to the BLAS ones.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号