US 5815394 A Abstract A computational model for optimal and automated design of structures based on neural dynamic is disclosed. The disclosed neural dynamic model for structural optimization integrates a penalty function method, Lyapunov stability theorem, Kuhn-Tucker condition and neural dynamic concepts. The model, which has global convergence, is general and may be applied to many nonlinear programming problems such as the design of automobiles, airplanes, fighters, building structures, and high-performance computer chips.
Claims(9) 1. A method of selecting structural members for an engineering design optimization for an engineering structure comprising structural members:
(a) defining a plurality of general non-linear constraints of said engineering structure; (b) determining a design parameter of said structural members of said engineering structure; (c) optimizing said design parameter in response to said plurality of general non-linear constraints through application of a learning rule through the integration of a Kuhn-Tucker necessary condition for at least one local minimum of a Lyapunov energy functional involving an exterior penalty function method and through application of a pseudo-objective function in the form of said Lyapunov energy functional; and (d) selecting structural members of said engineering structure in accordance with said optimization of said design parameter. 2. A computer for optimizing the engineering design for an engineering structure comprising structural members, said computer comprising:
a microprocessor being encoded with programming instructions, said instructions adapted to: (a) define a plurality of general non-linear constraints of said engineering structure; (b) determine a design parameter of said structural members of said engineering structure; and (c) optimize said design parameter in response to said plurality of general non-linear constraints through application of a learning rule through the integration of a Kuhn-Tucker necessary condition for at least one local minimum of a Lyapunov energy functional involving an exterior penalty function method and through application of a pseudo-objective function in the form of said Lyapunov energy functional. 3. A method of constructing an engineering structure comprising structural members and being of an optimized engineering design, said method comprising:
(a) defining a plurality of general non-linear constraints of said engineering structure; (b) determining a design parameter of said structural members of said engineering structure; (c) optimizing said design parameter in response to said plurality of general non-linear constraints and through application of a pseudo-objective function in the form of a Lyapunov energy functional involving an exterior penalty function method; (d) selecting structural members of said engineering structure in accordance with said optimization of said design parameter; and (e) assembling said structural members so as to form said engineering structure. 4. A method according to claim 3 wherein said Lyapunov energy functional has at least one local minimum, and wherein said optimization of said design parameter is done through application of a learning rule through the integration of a Kuhn-Tucker necessary condition for said at least one local minimum of said Lyapunov energy functional.
5. A method according to claim 3 wherein said plurality of general non-linear constraints comprises at least one non-linear constraint.
6. A method according to claim 3 wherein said structural members have a collective cost and weight, and wherein said design parameter is selected from the group consisting of said cost and said weight.
7. A method according to claim 3 wherein said optimization is done in accordance with a neural network topology for optimization of structures with discrete members, said neural network topology comprising:
(a) a variable layer comprising a plurality of design variables, each of said design variables corresponding to a node in said variable layer; (b) a competition layer comprising a plurality of sections, each of said sections corresponding to a node in said competition layer; (c) a constraint layer comprising a plurality of constraints, each of said constraints corresponding to said discrete members and each of said constraints corresponding to a node in said constraint layer; (d) an interpolation layer comprising a plurality of sectional properties, each of said sectional properties corresponding to said sections; (e) inhibitory connections between each of said nodes in said variable layer and each of said nodes in said constraint layer, each of said inhibitory connections assigned a respective sensitivity coefficient of constraint; (f) excitatory connections between each of said nodes in said variable layer and each of said nodes in said competition layer, each of said excitatory connections having encoded therein a design variable of each of said sections; and (g) excitatory connections between each of said nodes in said competition layer and each of said nodes in said interpolation layer, each of said excitatory connections having encoded therein a sectional property of each of said sections. 8. A method of selecting structural members for an engineering design optimization for an engineering structure comprising structural members:
(a) defining a plurality of general non-linear constraints of said engineering structure; (b) determining a design parameter of said structural members of said engineering structure; (c) optimizing said design parameter in response to said plurality of general non-linear constraints and through application of a pseudo-objective function in the form of a Lyapunov energy functional involving an exterior penalty function method, wherein said optimization is done in accordance with a neural network topology for optimization of structures with discrete members, said neural network topology comprising: (i) a variable layer comprising a plurality of design variables, each of said design variables corresponding to a node in said variable layer; (ii) a competition layer comprising a plurality of sections, each of said sections corresponding to a node in said competition layer; (iii) a constraint layer comprising a plurality of constraints, each of said constraints corresponding to said discrete members and each of said constraints corresponding to a node in said constraint layer; (iv) an interpolation layer comprising a plurality of sectional properties, each of said sectional properties corresponding to said sections; (v) inhibitory connections between each of said nodes in said variable layer and each of said nodes in said constraint layer, each of said inhibitory connections assigned a respective sensitivity coefficient of constraint; (vi) excitatory connections between each of said nodes in said variable layer and each of said nodes in said competition layer, each of said excitatory connections having encoded therein a design variable of each of said sections; and (vii) excitatory connections between each of said nodes in said competition layer and each of said nodes in said interpolation layer, each of said excitatory connections having encoded therein a sectional property of each of said sections; and (d) selecting structural members of said engineering structure in accordance with said optimization of said design parameter. 9. A computer for optimizing the engineering design for an engineering structure comprising structural members, said computer comprising:
a microprocessor being encoded with programming instructions, said instructions adapted to: (a) define a plurality of general non-linear constraints of said engineering structure; (b) determine a design parameter of said structural members of said engineering structure; and (c) optimize said design parameter in response to said plurality of general non-linear constraints and through application of a pseudo-objective function in the form of a Lyapunov energy functional involving an exterior penalty function method, wherein said optimization is done in accordance with a neural network topology for optimization of structures with discrete members, said neural network topology comprising: (i) a variable layer comprising a plurality of design variables, each of said design variables corresponding to a node in said variable layer; (ii) a competition layer comprising a plurality of sections, each of said sections corresponding to a node in said competition layer; (iii) a constraint layer comprising a plurality of constraints, each of said constraints corresponding to said discrete members and each of said constraints corresponding to a node in said constraint layer; (iv) an interpolation layer comprising a plurality of sectional properties, each of said sectional properties corresponding to said sections; (v) inhibitory connections between each of said nodes in said variable layer and each of said nodes in said constraint layer, each of said inhibitory connections assigned a respective sensitivity coefficient of constraint; (vi) excitatory connections between each of said nodes in said variable layer and each of said nodes in said competition layer, each of said excitatory connections having encoded therein a design variable of each of said sections; and (vii) excitatory connections between each of said nodes in said competition layer and each of said nodes in said interpolation layer, each of said excitatory connections having encoded therein a sectional property of each of said sections. Description The present invention is in the field of design automation and optimization, particularly as it pertains to structural engineering. In the design of structures, systems have been developed to achieve optimization through the use of algorithms. However, algorithms of the prior art often fail in terms of convergence and stability, particularly for large nonlinear engineering systems. For instance, existing Computer-Aided Design (CAD) software systems have rudimentary optimization capabilities and can hardly handle large nonlinear systems. Another problem with prior art systems is that the data models employed do not take advantage of computing resources available today. Optimization of large structures with thousands of members subjected to actual constraints of commonly-used design codes requires an inordinate amount of computer processing and can be done only on multiprocessor supercomputers. Automation of design, and/or its optimization, particularly of one-of-a-kind engineering systems, is considered a particularly challenging problem. The inventors and others have been working on creating novel design theories and computational models with two broad objectives: automation and optimization. (Adeli, H. and Alrijleh, M. M., Roof Expert, PC AI, Vol. 1, No. 2, pp. 30-34 (1987); Adeli, H. and Balasubramanyam, K. V., A Novel Approach to Expert Systems for Design of Large Structures, AI Magazine, Winter, pp. 54-63 (1988); Adeli, H. and Balasubramanyam, K. V., Expert Systems for Structural Design--A New Generation, Prentice-Hall, Englewood Cliffs, N.J. (1988); Paek, Y. and Adeli, H, STEELEX: A Coupled Expert System for Integrated Design of Steel Structures, Engineering Applications of Artificial Intelligence, Vol. 1, No. 3, pp. 170-180 (1988); Adeli, H. and Kamal, O., Parallel Processing in Structural Engineering, Elsevier, London (1993); and Adeli, H. and Hung, S. L., Machine Learning--Neural Networks, Genetic Algorithms, and Fuzzy Systems, John Wiley and Sons, New York. (1995)). It is an object of the present invention to provide an improved design process that may be advantageously applied to the design and/or optimization of structures, such as aerospace, automotive, and civil engineering structures, including those that are one-of-a-kind or that may be mass produced, especially those that involve highly nonlinear systems. Such a process involves the use of an appropriate data model and algorithms that are robust and provide rapid convergences even for such highly nonlinear systems. It is also an object of the present invention to produce structures arrived at through the improved design process of the present invention. Such structures may include any structure, such as aerospace, automotive, and civil engineering structures, including those that are one-of-a-kind or that may be mass produced. It is also an object of the present invention to design structures by optimizing structural parameters while reducing the amount of computing time required for the task. In view of the following description of the invention or through its practice, other advantages, or the solution to other problems, may become apparent. The present invention includes a design process involving the use of new computing paradigms to create computational models for structural design automation. The present invention addresses problems in prior art systems by exploiting a neural dynamic model and implementing algorithms on a distributed memory multiprocessor. The method of the present invention is robust and converges fast even for highly nonlinear systems. The method of the present invention facilitates design automation and optimization of large engineering systems. It may be applied to any type of resource optimization problem that may be described as a nonlinear programming problem including various engineering designs such as automobiles, airplanes, fighters, building structures, and high-performance computer chips. For example, the method may be applied to minimum weight design steel high-rise building structures subjected to the highly nonlinear constraints of actual design codes such as the American Institute of Steel Construction (AISC) Allowable Stress Design (ASD) and Load Resistance Factor Design (LRFD) specifications where nonlinear second order effects have to be taken into account. The computational model of the present invention finds the minimum weight design for very large structures subjected to multiple dead, live, and wind loadings automatically. A high degree of parallelism may be exploited by the neural dynamic computing model of the present invention. The robust data parallel neural dynamic model of the present invention may be used for discrete optimization of large structures such as in the problem of determining the minimum weight design of steel structures made of commercially available shapes. In this application of the present invention, the design consists of preliminary design, structural analysis, and the selection of the final members of the steel structure. As such, the final optimum solution for minimum weight design of steel structures may be obtained automatically without any intervention from the user. Parallelism may be exploited in four stages of the neural dynamic model as applied to minimum weight design of a steel structure: mapping the continuous design variables to discrete sections using a trained counter propagation neural (CPN) network, generating the element of stiffness matrices in the local coordinates, transforming them to the global coordinates, and solving the resulting simultaneous linear equations using the preconditioned conjugate gradient (PCG) method, evaluation of the constraints based on the AISC ASD or LRFD specifications, and computation of the improved design variables from the nonlinear neural dynamic model. Robustness of the structural optimization algorithm of the present invention is of considerable significance since many structural optimization algorithms show convergence or instability problems when applied to large structures such as highrise building structures subjected to the actual highly nonlinear constraints of the AISC ASD or LRFD specifications. FIG. 1 is a topological diagram of an artificial neural network; FIG. 2 is an conceptual diagram of the feedforward and feedback recall mechanism; FIG. 3 is topological diagram of a neural dynamic model in accordance with the present invention; FIG. 4 is a system organization diagram for a CM-5 computer used in accordance with a preferred embodiment of the present invention; FIG. 5 is a distributed data structure diagram for a parallel recalling process in accordance with a preferred embodiment of the present invention; FIGS. 6a and 6b are flow diagram for data parallel distributed algorithm in accordance with a preferred embodiment of the present invention; FIG. 7 is an illustration of a first building structure designed in accordance with a preferred embodiment of the present invention; FIG. 8 is a floor plan of the building structure of FIG. 7; FIG. 9 is an illustration of the relationships between the cross-sectional area and moments of inertia for commercially available steel I shapes; FIG. 10 is an illustration of the relationships between the cross-sectional area and radii of gyration for commercially available steel I shapes; FIG. 11 is the convergence history for the automated design of the building structure of FIG. 7 in accordance with a preferred embodiment of the present invention; FIG. 12 is an illustration of a second building structure designed in accordance with a preferred embodiment of the present invention; FIGS. 13a and 13b are illustrations of modules for the building structure of FIG. 12; FIG. 14 is the convergence history for the automated design of the building structure of FIG. 12 in accordance with a preferred embodiment of the present invention; FIG. 15 is an illustration of a third building structure designed in accordance with a preferred embodiment of the present invention; FIG. 16 is a floor plan of the building structure of FIG. 15; and FIG. 17 is the convergence history for the automated minimum weight design of the building structure of FIG. 15 in accordance with a preferred embodiment of the present invention. The following description focuses on a presently preferred embodiment of the present invention directed to the minimum weight design of large steel structures. The minimum weight design of large steel structures is merely illustrative of the types of engineering design problems to which the method of the present invention may be applied. The method of the present invention may be applied to any type of resource optimization problem that can be described as a nonlinear programming problem. It may be used for automation and optimization of various engineering designs such as automobiles, airplanes, fighters, building structures, and high-performance computer chips. Neural network models inspired by biological nervous systems are providing a new approach to problem solving. Neural network applications in structural engineering have been based on backpropagation and counter propagation. Optimization is an area in which the neural network has not been widely used. Two different types of optimization problems may be solved using neural networks. The first type is control of mobile robots in which a neural network is used to learn the relationship between sensory input and behavior. The second type is classical optimization in which a neural network is used to find a node configuration (an equilibrium point) that minimizes an energy or objective function. Referring now to FIG. 1, an artificial neural network (ANN) is a directed graph composed of nodes 300, 302, 304, 306, 308, 310, 312 and connections between nodes 314, 316, 318. Generally, an artificial neural network consists of three elements: an organized topology of interconnected nodes (network), a method of encoding information (learning), and a method of recalling information (retrieving). ANN topology consists of layers of nodes linked by weighted interconnections. A node has an activation function that evaluates inputs and generates an output as an input to other nodes. A layer that receives inputs from outside network is called an input layer 326 and a layer that emits computed output to the environment is an output layer 330. The layers lying between the input and output layers are called hidden layers 328. Characteristics of ANN topologies are described in terms of connection and interconnection types. (1) Connection types. There are two primary connection types: excitatory and inhibitory. An excitatory connection increases the value of input to a connected node and is usually represented by a positive sign. On the other hand, an inhibitory connection decreases the value of input to a connected node and is usually represented by a negative sign. (2) Interconnection types. The three interconnection types are intra-layer (lateral connection) 320, inter-layer 322, and recurrent connections 324. Intra-layer connections are connections between nodes in the same layer 320. Inter-layer connections are connections between nodes in different layers 322. Recurrent connections are direct loops of connections to the same node 324. Learning may be defined as any change in the weights to produce some desirable state, and a learning method is a rule that adjusts the weights to the desirable state. Learning methods may be classified into two categories: supervised learning and unsupervised learning. (1) Supervised learning. Learning is performed on the basis of direct comparison of the output of the network with the given desired output. Backpropagation is a widely used example of supervised learning algorithms. (2) Unsupervised learning. Learning goal is not defined. The network is expected to create categories from the correlations of the input data, and to produce output corresponding to the input category. Counterpropagation is an example of unsupervised learning algorithms. Referring to FIG. 2, a recalling method may be defined as the method of finding the corresponding output from the given input after the learning process. The ANN recall mechanisms may be divided into two main categories. One is the feedforward recall 340 and the other is the feedback recall mechanism 350. During the feedforward recall 340, an input 342 is passed through nodes 344 and weights 346, and the corresponding output 348 is produced in one pass. In the feedback recall mechanism 350, an input 352 is passed through the nodes 354 and weights 356, and then an output 358 is produced which is fed back into an input 352 or a specific layer 360 until there is no change in the weights. Neural Dynamics Model A dynamic system may be defined as a system of differential equations whose state changes in time. ##EQU1## where X(t)= X In a neural dynamic system, a system of differential equations reflecting the characteristics of problems to be solved describes how the node activations change, where connection weights and nodes represent the dynamic system and outputs from the nodes represent the state of the system at a given instance in time. A pseudo-objective function is formulated by an exterior penalty function method that transforms the constrained optimization problem to a corresponding unconstrained optimization problem: ##EQU2## where F(X) is the objective function (the weight of the structure), P(X) is the penalty function, and r
g
g The system of differential equations is obtained by taking the derivative of the pseudo-objective function with respect to time, and forcing the pseudo-objective function (Eqn. 2) to satisfy the Lyapunov stability theorem: ##EQU4## The resulting system of differential equations in the neural dynamics model is ##EQU5## Therefore, the dynamic system always evolves such that the value of the pseudo-objective function for the structural optimization problem does not increase (dV/dt≦0). A CounterPropagation Neural network (CPN) is a mapping neural network employing both Kohonen learning and Grossberg learning rules. Optimal design of steel structures according to the commonly-used design specifications such as the AISC Allowable Stress Design (ASD) specifications requires computation of allowable member stresses. Allowable stresses in turn are functions of cross sectional properties of the members. The allowable compressive stress of members in space axial-load structures is a function of the slenderness ratio, KL/r, where K is the effective length factor and r is the value of the radius of gyration of the cross-section. Assuming K=1, the allowable stress of a compression member is governed by r The design variables in the structural optimization problem are cross-sectional areas of the members. But, the allowable compressive stress is a function of the radius of gyration of the cross-section. To map the two variables, a CPN network is trained to learn the relationship between the cross-sectional area and the radius of gyration for the aforementioned subset of W shapes. The input is the cross-sectional areas of W shapes and the output is their radii of gyration about major (r The relationship between the cross-sectional area and the radii of gyration is recalled in each iteration of the optimization process using the trained CPN. This recalling process consists of two steps. The first step is to find the winning node among the nodes in the competition layer for a given input. When a continuous relationship is assumed between the input and output variables the number of winning nodes may be more than one (up to the number of nodes in the competition layer). In contrast, when input and output are discrete variables the number of winning nodes for each variable is set to one and the value of the connection weight for the link connecting the winning node to the variable is close, but greater than or equal to the value of the variable. Thus, a winning node is selected for each variable. After selecting the winning nodes for all the variables, the output (radii of gyration) is recalled in the second step based on the output from the winning nodes. The trained CPN maps an arbitrary given instance (design variable) to the nearest encoded output (radii of gyration). The present invention is based on a hybrid counter propagation-neural dynamics model and a new neural network topology. It may be used for structural optimization of large structures with discrete members. The nonlinear neural dynamics model is based on Eqn. 8 and acts as an optimizer to produce improved design solutions for given design constraints starting from some initial design solution and using first order sensitivity information. The sensitivity information and the magnitudes of the constrained violations, required for operations of the neural dynamics model, are evaluated based on the discrete cross sectional properties. The discrete cross sectional properties are provided by the trained CPN. The hybrid neural dynamics model consists of four components: the neural dynamics model, counter propagation neural network, structural analysis, and calculation of sensitivity coefficients. The neural dynamics model consists of two distinct layers: the variable layer and the constraint layer. Nodes in the constraint layer receive the discrete cross sectional properties from the CPN as inputs, evaluate the prescribed constraints, and generate the magnitudes of constraint violations as outputs. The functional activations at the nodes in the variable layer receive information about the search direction (encoded as the weights of the links connecting the constraint layer to the variable layer) and magnitudes of constraint violations as inputs and generate the improved design solutions as outputs. The number of stress constraints is equal to or a multiple of the number of members in the structure (for the case of multiple loadings). This number is in the thousands for a large structure with a few thousand members. As such, the number of violated stress constraints requiring computation of sensitivity coefficients tends to be very large. Thus, it requires an excessive amount of CPU time. A design linking strategy may be employed as is commonly done in actual design of structures. The members grouped as one design variable have the same cross sectional properties; but, each member in the group has different magnitude of the constraint violation and different sensitivity coefficients for a given design solution. To accelerate the optimization process and reduce the required CPU time for computation of sensitivity coefficients, only the most violated constraint in each group of members linked together as one design variable is allowed to represent the status of the stress constraint for that group. Therefore, a competition is introduced in the constraint layer to select the most critical node among the nodes belonging to one linked design variable. Let O
O where Di is the number of members grouped as the ith design variable. The counter propagation part of the model consists of two layers: competition layer and interpolation layer. Nodes in the competition layer receive the values of improved design solutions from the nodes in the variable layer, calculate the Euclidean distances between the input and the connection weights, and select the winning node. Nodes in the interpolation layer recall the corresponding cross sectional properties encoded in the connection weights associated with the winning node. The adjoint variable method is used for sensitivity calculations. By using the adjoint variable method direct inverting of the structure stiffness matrix in calculation of derivatives of displacements with respect to design variables is avoided. Sensitivity coefficients of the objective and constraints functions are also linked similar to design variables (described in steps 2 and 10 of the hybrid algorithm to be presented later). Referring to FIG. 3, the topology of the hybrid counter propagation-neural dynamics model is shown. The model is comprises four layers: a constraint layer 10, a variable layer 12, a competition layer 14, and an interpolation layer 16. The number of nodes in the variable layer 12 is equal to the number of independent design variables (K) in the structural optimization problem. The number of nodes in the constraint layer 10 is equal to total number of constraints (M+N) imposed on the structure. The number of nodes in the competition layer (T) 14 is equal to the number of available sections (e.g., 61 for 61 W shapes). The number of nodes in the interpolation layer (S) 16 is equal to the number of sectional properties required for calculation of the allowable stresses. This number is 3 for space axial-load structure (A, r Three types of connections are used to link nodes: inter-layer connection 28, intra-layer connection 30, and recurrent connection 32 . Nodes between different layers are connected by inter-layer links. In the neural dynamics parts of the model, sensitivity coefficients of constraints are encoded as weights of the inhibitory connections from the constraint layer to the variable layer (W Intra-layer connections that link nodes in the same layer are employed to select the winning nodes in the constraint and competition layers. In the variable layer, the sensitivity coefficients of the objective function are assigned to the recurrent connections. The hybrid counter propagation-neural dynamics algorithm for structural optimization with discrete member sizes subject to displacement and stress (including buckling) constraints is presented in the following steps. Step 1. Initialize the design variables X, and select an initial value for the penalty parameter r and the tolerance for the optimization convergence, ε. Step 2. Compute the linked sensitivity coefficients for the original objective function (weight of the structure). ##EQU6## where X, is the ith independent design variable, Y Step 3. Set the iteration counter, n=0. Step 4. Feed forward the values of design solutions from the variable layer to the nodes in the competition layer of the CPN, and calculate the Euclidean distance between the ith design variable (i=1, K) and the connection weights U Step 8. Calculate the outputs of nodes in the constraint layer using the following activation functions. For stress constraint:
O For displacement constraint:
O By using Eqns.(16) and (17), the output will be zero when constraints are satisfied and equal to the exact value of the violation when a constraint is violated. Step 9. Select the winning node in the constraint layer representing the most violated stress constraint among the members grouped as the ith design variable (i=1, K).
O
O There is no competition among the nodes representing displacement constraints. Step 10. Calculate the sensitivity coefficients for the constraints associated with winning nodes (with nonzero output) and assign them as the weights of the inhibitory connections from the constraint layer to the variable layer. ##EQU10## Step 11. Send the output from the nodes in the constraint layer (O
X(n)=X(n-1)+∫X(n-1)dt Eqn. (23) The improved values of design variables at the nth iteration, X(n), are calculated using the Euler method for numerical integration of the integral term in Eqn. (23). Step 13. Check the optimization convergence of the change in the design variables by computing the following error function: ##EQU12## If ψ(n)≦ε and outputs from the winning nodes in the constraint layer are all zero (O and go to step 4. Eqn. (25) produces stable and nonoscillating convergence. Data Parallel Neural Dynamics Model As noted above, a neural dynamic model for optimal design of structures may be created by integrating penalty function method, Lyapunov stability theorem, Kuhn-Tucker conditions, and the neural dynamic concept. A pseudo-objective function in the form of a Lyapunov energy functional is defined using the exterior penalty function method. The Lyapunov stability theorem guarantees that solutions of the corresponding dynamic system (trajectories) for arbitrarily given starting points approach an equilibrium point without increasing the value of the objective function. In other words, the new neural dynamic model for structural optimization problems guarantees global convergence and robustness. But, this does not guarantee the equilibrium point is a local minimum. The Kuhn-Tucker condition may be used to verify that the equilibrium point satisfies the necessary conditions for a local minimum. In one embodiment of the present invention, the minimum weight design of large steel structures may be determined. In order to achieve automated optimum design of realistic structures subjected to actual design constraints of commonly-used design codes such as the American Institute of Steel Construction (AISC) Allowable Stress Design (ASD) and Load and Resistance Factor Design (LRFD) specifications, a hybrid counterpropagation neural (CPN) network-neural dynamic is employed for discrete optimization of such structures consisting of commercially available sections such as the wide-flange (W) shapes. Parallelism may be exploited in four stages of the nonlinear neural dynamic optimization model as applied to minimum weight design of steel structures: mapping the continuous design variables to discrete sections using a trained counter propagation neural (CPN) network), generating the element of stiffness matrices in the local coordinates, transforming them to the global coordinates, and solving the resulting simultaneous linear equations using the preconditioned conjugate gradient (PCG) method, evaluation of the constraints based on the AISC ASD or LRFD specifications, and computation of the improved design variables from the nonlinear neural dynamic model. Referring again to FIG. 3, the topology of the nonlinear neural dynamic model is shown. The model is comprised of four layers: a constraint layer 10, a variable layer 12, a competition layer 14, and an interpolation layer 16. The layers are interconnected by four types of connections: inhibitory recurrent connection 28, intra-layer connection 30, excitatory connection 32, and inhibitory connection 34. The number of nodes in the variable layer 12 corresponds to the number of independent design variables K in the structural optimization problem. The design variables are the cross-sectional areas of the members. For design of steel structures, a database of cross-sectional properties is needed for computation of element stiffness matrices and evaluation of the AISC ASD or LRFD constraints. For a general space frame structure, these properties as defined in Table 1 include S=11 entities (A, I The recalling process in the CPN network is done in two steps. In the first step, for each design variable a competition is created among the nodes in the competition layer for selection of the winning node. The weights of the links between the variable and the competition layers represent the set of cross-sectional areas of the available standard shapes. The weight of the link connecting the winning node to the variable is the one closest to the value of the variable itself but not smaller than that. In the second step, discrete cross-sectional properties encoded in the form of weights of links between the competition and the interpolation layers are recalled. The weights of the links connecting the winning node to the nodes in the interpolation layer are the cross-sectional properties corresponding to an improved design variable. In the second stage of the nonlinear neural dynamic model for optimal design of steel structures, element stiffness matrices are first generated in the elements' local coordinates systems and then transformed to the global coordinates system. For solution of the resulting linear simultaneous equations, direct methods are not appropriate for distributed memory computers because of their large memory requirements. Direct methods require the assembly of the structure stiffness matrix which may be very large for a structure with thousands of members. Consequently, iterative methods such as the conjugate gradient method are deemed more appropriate for distributed memory computers where the size of local memory is limited, for example, to 8 MB. Computational efficiency of the conjugate gradient method may be improved by adopting preconditioning techniques. As such, a data parallel preconditioned conjugate gradient (PCG) method may be used in a preferred embodiment of the present invention. The third stage consists of constraint evaluation using the nodal displacements and member stresses obtained in the previous stage. For minimum weight design of steel structures, three types of constraints may be considered: fabricational, displacement, and stress (including buckling) constraints. Referring again to FIG. 3, nodes in the constraint layer 10 receive the discrete cross-sectional properties from the CPN as inputs, evaluate the prescribed constraints, and generate the magnitudes of constraint violations as outputs. The functional activations at the nodes in the variable layer 12 receive information about the search direction (encoded as the weights of the links inhibitory connections 34! connecting the constraint layer 10 to the variable layer 12) and magnitudes of constraint violations as inputs and generate the improved design solutions as outputs. The number of nodes in the constraint layer 10 is equal to total number of constraints imposed on the structure. There are as many constraint layers as number of loading combinations acting on the structure. Referring to FIG. 3, two constraint layers 18, 20 are shown. The number of stress constraints may be in the thousands for a large structure with thousands of members. As such, the number of violated stress constraints requiring computation of sensitivity coefficients tends to be very large. To accelerate the optimization process and reduce the required CPU time for computation of sensitivity coefficients, only the most violated constraint in each group of members linked together as one design variable is allowed to represent the status of the constraints for that group. Therefore, a competition is introduced in the constraint layer 10 to select the most critical node among the nodes belonging to one linked design variable. For the LRFD code, the primary stress constraint for a general beam-column member is a highly nonlinear and implicit function of design variables in the following form: ##EQU14## where P
M
M where M In the final stage, the nonlinear neural dynamic model acts as an optimizer to produce improved design variables from initial design variables. It consists of a variable layer and a number of constraint layers equal to the number of different loading conditions. The dynamic system describing how nodes activations change are described by ##EQU18## where X To derive the dynamic system represented by Eqn. (32), the pseudo-objective function is formulated by transforming the constrained optimization problem into a corresponding unconstrained optimization problem using an exterior penalty function method. The resulting pseudo-objective function is in turn transformed to a Lyapunov function by verifying the stability of the pseudo-objective function. The dynamic system (Eqn. 32) is derived by taking the derivative of the Lyapunov function with respect to time and making the value of the derivative less than or equal to zero. Therefore, the changes of node activations (the changes in the design variables) in the neural dynamic model are controlled such that the state of design variables approaches to a desirable state (a local minimum) without increasing the value of the pseudo-objective function for the structural optimization problem. The changes in the values of design variables at the nth iteration are calculated by numerical integration of Eqn. (32) using the Euler method. The improved values of design variables at the nth iteration are obtained as follows:
X(n)=X(n-1)+∫X(n-1)dt i=1, K Eqn. (33) Connection Machine CM-5 Architecture A preferred embodiment of the present invention may be developed to operate on a massively parallel computer such as the CONNECTION MACHINE CM-5. The present invention is not, however, limited to any particular operating environment. Those skilled in the art will find that the methods of the present invention may operate in a variety of environments (e.g., Encore Multimax or Cray YMP8/864). The description of the present invention in relation to an implementation for the CM-5 is illustrative only. Referring to FIG. 4, the main components of the CM-5 system are a number of processing nodes (PN) 50, partition manager (PM) 52, and two high-speed, high-bandwidth communication networks called data and control networks 54, 56. A PN 50 has four vector units (identified as VU0 58, VU1 62, VU2 66, VU3 70) with a total of 32 MB of memory (8 MB per VU 60, 64, 68, 72) and can perform high-speed vector arithmetic computations with a theoretical peak performance of 128 MFLOPS. A VU 58, 62, 66, 60 is the smallest computational element (processing element, PE) in the system which executes vector operations on data in its own local (parallel) memory 60, 64, 68, 72. A 2 Data on the CM-5 are divided into two classes: scalar (or serial) and parallel. Scalar data is stored in a PM memory 74 and operated on serially by the PM. Operations involved in scalar or serial arrays are minimized in order to achieve high parallel processing performance. Parallel data are distributed across parallel memories of VUs 60, 64, 68, 72. The performance of a data parallel algorithm depends highly on the parallel data distribution method. Preferably, parallel data is distributed so that all PEs may work on data located in their own memory without any communication. A block distribution with detailed array layout may be used to distribute parallel data. In the block distribution, an array is divided into equal size blocks (the size of a block is specified in the array layout directive) and each block is assigned to one PE. The programming language CM FORTRAN uses array processing features of FORTRAN 90 to express data parallel algorithm. Array operations on a single array or array section (subset of an array) are processed concurrently on elements of the array allocated to the local VU memories 60, 64, 68, 72. Array operations involving more than two arrays or array sections may be done efficiently without inter-processor communications only when corresponding elements of arrays are stored in the memory of the same VU (called conformable arrays). Two arrays are conformable when their dimensions and lengths are the same. For example, given two conformable one dimensional arrays X Data Parallel Algorithms and Data Structures In accordance with a preferred embodiment of the present invention, four data parallel structures and algorithms are developed for the illustrative problem of optimal design of steel structures. The first one is for recalling the cross-sectional properties of the available standard shapes. The second one is for the PCG method used for the solution of the linear simultaneous equations. The third one is for evaluation of the AISC ASD or LRFD code constraints. The last one is for the operation of the neural dynamic network. Recalling Cross-Sectional Properties of the Available Standard Shapes For each design variable X
d where T is the number of available W shapes and 0 K competitions (the number of design variables) are required at each design iteration. To perform K competitions concurrently, K design variables are divided into the same number of groups as the number of available PEs (N For any given X Referring to FIG. 5, in order to reduce the communication cost, private copies of U step 1. Broadcast private copies of U(1:K, 1:T) and V(1:T, 1:S) to the PEs where the design variables are allocated. (Each PE executes roughly N step 2. Calculate the Euclidean distances, d(1:K, 1:T) between the design variables X(1:K) and the connection weights U(1:K, 1:T).
d(1:K,1:T)=∥U(1:K,1:T)-X(1:K)∥ Eqn. (37) step 3. Set the values of 0(1:K, 1:T) and calculate the modified Euclidean distances, d(1:K, 1:T): where
U(1:K,1:T)-X(1:K)<0
0(1:K,1:T)=10 else where
0(1:K,1:T)=0 end where
d(1K,1:T)=d(1:K,1:T)+0(1:K,1:T) Eqn. (38) step 4. Select the winning nodes based on the modified Euclidean distances, and set the outputs of the winning nodes, Z(i=1:Kj=1:T), to 1.0 and remaining nodes to 0. ##EQU21## step 5. Recall S discrete cross-sectional properties (S=11 for ASD and S=13 for LRFD) based on the encoded weights V(j=1: T,k=1,S): ##EQU22## Preconditioned Conjugate Gradient Method A data parallel PCG method is presented with an appropriate data structure for the solution of linear equilibrium equations resulting from a finite element analysis of the structure. Two levels of parallelisms are exploited: structural elements level and degrees-of-freedom level. Each structural member and each degree-of-freedom is assigned to a VU explicitly. For concurrent generation of element stiffness matrices, each element stiffness matrix of a framed structure containing 2N A diagonal preconditioner is created by assembling the diagonal elements of element stiffness matrices according to the nodal displacement connectivity matrix in order to accelerate the search process in the conjugate gradient method. The diagonal preconditioner is transformed into a vector with a length N The major computations required for the PCG algorithm consist of basic linear algebra operations such as matrix-vector multiplications, dot products, scalar-vector multiplications, and vector additions. All vectors involved in the basic operations have the same data structure, a degree-of-freedom per VU. Thus, the latter two operations are executed concurrently without any inter-processor communication. However, communications are involved in both matrix-vector multiplications and dot products. Communication overhead involved in dot product operations, performed through the fast control network in the CM-5 system, is negligible compared to that of matrix-vector multiplications. A matrix-vector multiplication is required between the unassembled global stiffness matrix K and the search direction vector p for computation of the search step size. In general, communication involved in the matrix-vector multiplications is unavoidable on distributed memory computer because the length and dimensions of the matrix and the vector are different. In accordance with a preferred embodiment of the data parallel algorithm, the matrix K is distributed at element level and the vector p is distributed at nodal degree-of-freedom level. The matrix-vector multiplication is parallelized at element level by transforming the direction vector p into an element level direction matrix in the form of two dimensional array ptemp(2N step 1. Set up the diagonal preconditioner (D) using the nodal displacement connectivity matrix IDJ(2N (In steps 2 and 3, each PE executes roughly N step 2. Calculate the initial residual (error) vector (r
r
p
z
ρ=z step 3. Check the convergence of the PCG algorithm using the norm of the residual (error) vector, ∥r∥. If ∥r∥<, then u is the nodal displacement vector, and stop. Otherwise do the following steps. (In steps 4 to 6, each PE executes N step 4. Transform the direction vector into a temporary two dimensional matrix using the gather operation and the nodal displacement connectivity matrix IDJ(2N step 5. Do the concurrent matrix-matrix multiplications at element level.
vtemp(2N 6. Transform the matrix ptemp into the vector v using the connectivity matrix IDJ(2N
scatter v(vd),ptemp(2N (Each PE executes roughly N step 7. Compute the step size for the search direction, γ=p/α, where α=p step 8. Compute the improved nodal displacement vector.
u(1:N step 9. Calculate the new residual vector.
r(1:N step 10. Compute new search direction vector.
p(1:N
z(1:N
ρ=z(N
β=ρ Go to step 3 Constraint Evaluation For integrated design of steel structures, three different kinds of constraints may be considered in the formulation of the structural optimization problems: fabricational, displacement, and stress constraints. Displacement constraints may be evaluated directly using the nodal displacement vector from the PCG solver. Stress constraints may be based on the AISC ASD or LRFD specifications. Constraints evaluation requires cross-sectional properties of W shapes used for members in a structure. They may be obtained from the trained CPN as described previously. The cross-sectional properties are distributed element-wise such that all cross-sectional properties for a member reside in the memory of a single VU. N Operation of the Neural Dynamic Network In a neural network model, the input to a node is calculated by the weighted sum of outputs from the nodes in the previous layer. In model of the present invention, the input to the ith node in the variable layer is the sum of three terms (Eqn. 32): the inhibitory recurrent connection weight of the ith node in the variable layer, C In the neural dynamic system of the present invention, the computation of the inputs to the nodes in the variable layer (Eqn. 32) are parallelized at both network node (neuron) and weight levels. For the node level parallelism, a node in each layer is assigned to a PE. In turn, each PE computes roughly N For the computations of inputs to the nodes in the variable layer, in order to distribute the loads among VUs as evenly as possible each column of the matrix representing the connection weights, wji(j=1:M+N, i=i The computations of inputs to the nodes in the variable layer are parallelized at both node (neuron) and weight level as follows: step 1. Broadcast private copies of O (Each PE executes roughly N step 2. Calculate the weighted sum of outputs from the nodes in the constraint layer (concurrent processing at node and weight levels) ##EQU23## step 3. Calculate the inputs to the nodes in the variable layer (concurrent processing at node level)
X(1:K)=X(1:K)-C(1:K) Eqn. (46) step 4. Calculate the outputs of the nodes in the variable layer (improved values of the design variables) by numerical integration (concurrent processing at node level).
X(1:K)=X(1:K)+IX(1:K) Eqn. (47) Referring to FIGS. 6a and 6b, a flow diagram of a preferred embodiment of the present invention is presented. FIGS. 6a and 6b illustrate the method of the present invention for a preferred embodiment and illustrate communications between VUs that occur during processing. As shown in FIGS. 6a and 6b, a series of operations are performed on data within each VU 90, 92, 94. The operations described in FIGS. 6a and 6b correspond to the stages described earlier. In Stage 1--Mapping Continuous Design Variables, starting values are initialized 96 and discrete cross-sectional properties are recalled 98. Following the completion of Stage 1, processing results are communicated across VUs 100 so the next processing stage may begin. In Stage 2--Generating Element Stiffness Matrices, element stiffness matrices are determined 102 and linear equations for nodal displacement are solved 104, preferably, using PCG. Processing results are then communicated across VUs 106. In Stage 3--Constraint Evaluation, member stresses are calculated 108, constraints are evaluated according to known specifications 110, and sensitivity coefficients are calculated for violated constraints 112. Processing results are then communicated across VUs 114. In Stage 4, connection weights are updated 116 and outputs of nodes in the constraint layer are determined 118 and communicated across VUs using a feedforward mechanism 120. Finally, inputs to nodes in the variable layer of the neural network are calculated 122 and values of improved design variables are computed 124. Processing continues until convergence occurs 126. As may be apparent, the present invention may be embodied in a computer program stored on magnetic media such as a hard disk drive or tape or another type of media. Neural network architectures and programming methods are well-known to those of ordinary skill in the art. Application of the present invention to the minimum weight design of large steel structures illustrate the benefits and advantages of the present invention. As will be apparent from the examples, the method of the present invention may be applied to steel structures of arbitrary size and configuration. Examples 1 and 3 are space moment resisting frames with cross bracings. Example 2 is a space axial-load (truss) structure. The modulus of elasticity of steel is E=198.91 Gpa (29000 ksi) and the specific weight is p=76.97 kN/m Referring to FIG. 7, this example is a 36-story irregular steel-moment-resisting space frame with setbacks and cross bracings. The structure has 1384 nodes and an aspect ratio of 4.7. The 3228 members of the structure are divided into 186 groups. The structure is divided into three 12-story sections 130, 132, 134 as indicated in FIG. 7. Referring to FIG. 8, in sections 1 (130) and 2 (132) of the structure, four different types of columns are used in every two stories as follows: corner columns 146, outer columns 148, inner columns in unbraced frames 150, and inner columns in braced frames 152. In section 3, there are only the first three different types of columns 146, 148, 150. In section 1 and 2, the beams of each story are divided into 3 groups: outer beams 154, inner beams in braced frames 156, and inner beams in unbraced frames 158. In section 3, there are only the first two types of beams 154, 156. Two different types of bracings are used in every three stories: one type in the longitudinal and another type in the transverse direction. For displacement constraints, the interstory drift is limited to 0.004 times the story height in x and y directions 160. The loading on the structure consists of dead load of 2.88 kPa 60 (psf), live load 2.38 kPa of (50 psf), and roof live load of 2.38 kPa (50 psf). The lateral loads due to wind are computed according to the Uniform Building Code (UBC) (1994). Lateral forces are determined by assuming a basic wind speed of 113 km/h (70 mph), exposure C (generally open area), and importance factor of 1. There are 295 W shapes in the AISC LRFD manual. Not all are considered economical shapes as beam and column members used in high-rise building structures. Broadly speaking, suitable W shapes for beam members are those with large moment of inertia I Three trends may be observed in FIGS. 9 and 10. For one group of W shapes (identified by 1 in FIGS. 9 (174) and 10 (184)), the moment of inertia with respect to the major axis, I The lower bounds for column/bracing and beam members are 35.7 cm The convergence histories for designs based on AISC ASD and LRFD codes are shown in FIG. 11. Minimum weights of 21.51 MN (4836.6 kips) and 20.34 MN (4573.5 kips) are obtained using the ASD and LRFD codes, respectively. These translate into 0.59 kPa (12.44 psf) and 0.56 kPa (11.76 psf) for ASD and LRFD codes, respectively, when the total weight of the structure is divided by the total floor area provided by the structure. The minimum weight design based on the LRFD code is 5.4% lighter than that based on the ASD code. Referring to FIG. 12, this example is a large and very tall hollow space steel structure with a height of 492.9 m (1617 ft). The structure has 1801 nodes, 8904 members, and an aspect ratio of 7.2. It has three main sections 200, 202, 204 and two smaller sections 206, 208. As shown in FIG. 13a, each section consists of 24 modules. Each module has 24 equal sides at each floor level with a height of 6.7 m (22 ft). The 8904 members of the structure are divided into 297 groups. In the three main sections (FIG. 12 200, 202, 204), each module has four different types of members indicated in FIG. 13a. Member types 1 (210) and 2 (212) are horizontal members. Member type 3 (214) is slightly inclined from the vertical line (4 degrees in section 1 (FIG. 12 200), 3 degrees in section 2 (FIG. 12 202), and 2 degrees in section 3 (FIG. 12 204)). Finally, the cross bracing members are grouped as member type 4 216. In the truncated section 4 (FIG. 12 206), the grouping of the members is similar to section 1 to 3. In the conical section 5 of the structure (FIG. 12 208), there are 24 members grouped as one type as shown in FIG. 13b. For displacement constraints, the interstory drift is limited to 0.0004 times the story height in any horizontal direction. The equivalent of a uniform vertical load of 1.92 kPa (40.0 psf) is assigned to the nodes of each floor. The structure is subjected to wind loads based on the UBC code similar to Example 1. For this superhigh-rise structure, none of the W shapes commonly used for columns (with relatively high r The convergence histories for designs based on AISC ASD and LRFD codes are shown in FIG. 14. Minimum weights of 429.5 MN (96564.3 kips) and 419.7 MN (94354.8 kips) are obtained using the ASD and LRFD codes, respectively. The minimum weight design based on the LRFD code is 2.3% lighter than that based on the ASD code. Referring to FIG. 15, this example is a 144-story steel superhigh-rise building structure with a height of 526.7 m (1728 ft). The structure, divided into three section 220, 222, 224, is a modified tube-in-tube system consisting of a space moment-resisting frame with cross bracings on the exterior of the structure. The structure has 8463 nodes and an aspect ratio of 7.2. The 20,096 members of the structure are divided into 568 groups. Referring to FIG. 16, in section 1 of the structure 220, five different types of columns are used in every two stories as follows: corner columns in the outer tube 226, non-corner columns in the outer tube 228, columns in the middle tube 230, columns in the inner tube 232, and the center column 234. Four different types of beams are used in every two stories as follows: beams in the outer tube 236, beams in the middle tube 238, beams in the inner tube 240, and beams connecting and within the tubes 242. Two types of bracings cover every six stories: one type covers the faces AB, CD, DE, and AF, an the other type covers the two corner regions BC and EF. In section 2, four different types of columns are used in every two stories as follows: corner columns in the outer tube 226, non-corner columns in the outer tube 228, columns in the inner tube 230, and the center column 234. The beams of every two stories are divided into three groups: beams in the outer tube 236, beams in the inner tube 238, and beams connecting and within the tubes 240. The same type of bracings is used in every six stories. In section 3, three different types of columns are used in every two stories as follows: corner columns 226, outer columns 228, and a center column 234. The beams of every two stories in this section are divided into two groups: inner beams 236 and outer beams 238. The same type of bracings is used in every six stories. For displacement constraints, the interstory drift is limited to 0.0004 times the story height in both x and y directions. The loading on the structure consists of dead load of 2.40 kPa (50 psf), live load 2.40 kPa (50 psf) in section 1 (FIG. 15 220) and dead load of 2.40 kPa (50 psf), live load 1.92 kPa (40 psf), and roof live load 1.92 kPa (40 psf) in sections 2 (FIG. 15 222) and 3 (FIG. 15 224). The structure is subjected to wind loads based on the UBC code similar to Example 1. In this example, wind loads are applied in three different directions x, y, and u (FIG. 14). The same set of 162 W shapes used in Example 1 are used as potentially economical beam members. The same set of 72 W and built-up shapes used in the Example 2 are used as potentially economical sections for columns and bracing members. The lower bounds for column/bracing and beam members are 35.7 cm2 (5.54 in. 2) and 33.9 cm The convergence histories for designs based on AISC ASD and LRFD codes are shown in FIG. 17. Minimum weights of 682.2 MN (153381.4 kips) and 669.3 MN (150467.2 kips) are obtained using the ASD and LRFD codes, respectively. These translate into 1.57 kPa (34.43 psf) and 1.54 kPa (33.78 psf) for ASD and LRFD codes, respectively, when the total weight of the structure is divided by the total floor area provided by the structure. The minimum weight design based on the LRFD code is about 2.0% less than that based on the ASD code. It may be noted that the amount of steel used in the currently tallest building structure in world, the 109 story Sears building in Chicago with a height of 445.0 m, is about 33 psf. The present invention, based on neurocomputing and parallel processing paradigms, is a data model and related algorithms for structural design automation. A neural dynamic model is presented for optimal design of structures by integrating penalty function method, Lyapunov stability theorem, Kuhn-Tucker conditions, and the neural dynamic concept. The present invention may be used for automation and optimization of various engineering designs such as automobiles, airplanes, fighters, building structures, and high-performance computer chips. A=cross-sectional area of a member A b C C d=depth a W shape d gj(x)=jth constraint function h=clear distance between flanges of a W shape I I J=torsional constant K=stiffness matrix k=number of design variables M=number of elements in the structure M M M M D=diagonal preconditioner N=number of constrained degrees-of-freedoms N N N N N N O P=nodal force vector p=search direction vector P P A=cross-sectional area of a member b dg d/A D F(X)=objective function g g I I J=torsional constant K=number of design variables L=length of a member M=number of members in the structure N=number of constrained degrees of freedoms P(X)=penalty function t t t t S=number of nodes in the interpolation layer S T=number of nodes in the competition layer V(X)=pseudo-objective function (Lyapunov function) X X X=vector of design variables ε=tolerance for convergence ∇F(X)=gradient vector of the objective function ∇g ∇g ρ=specific weight of the member σ ψ(n)=error function. r=residual (error) vector t t t t S=number of nodes in the interpolation layer S S T=number of nodes in the competition layer t t U u=nodal displacement vector V w X X X=vector of design variables Y Z Z Z φ φ θ Patent Citations
Non-Patent Citations
Referenced by
Classifications
Legal Events
Rotate |