question
stringlengths 19
6.88k
| answer
stringlengths 38
33.3k
|
---|---|
Problem Statement: Is it possible to model an Integrated Coal Gasification Combined-Cycle Power (IGCC) process with Aspen Plus? | Solution: Aspen Plus is shipped with a number of example files including a few examples of coal gasification. Both Aspen Plus backup (.bkp) files and PDF files describing the details are included in the install directory for most examples. The files for a model of an Integrated Coal Gasification Combined-Cycle Power (IGCC) process are located in the following location:
V9 and earlier: C:\Program Files (x86)\AspenTech\Aspen Plus Vx.x\GUI\Examples\IGCC
V10 and later: C:\Program Files (x86)\AspenTech\Aspen Plus Vx.x\GUI\Examples\Power\Coal Gasification
This model simulates an Integrated Coal Gasification Combined-Cycle Power (IGCC) process with different sections of the plant modeled as hierarchy blocks (model templates). There are also examples for an entrained flow coal gasifier (Aspen_Plus_Model_for_Entrained_Flow_Coal_Gasifier.apwz) and a moving bed coal gasifier (Aspen_Plus_Model_for_Moving_Bed_Coal_Gasifier.apwz) shipped with Aspen Plus in the Examples directory.
The model includes the following sections:
Sizing of the coal
Gasification unit
Air Separation (ASU)
Gas cleaning unit
Water-gas shift unit
Ammonia unit
Methanizer
Combined cycle power generation
Figure 1 shows the process flowsheet of the IGCC process.
The process is as follows:
The coal feed with particle size distribution is mixed with water in the Sizing section and undergoes crushing and screening.
The air separation unit (ASU) uses air to reach nearly pure Oxygen and Nitrogen. Using Radfrac-rigorous method to separate the air after pretreatment. The resulting Nitrogen product is 99.83 mole % pure, and the Oxygen product is 95 mole % pure.
The coal-water slurry is mixed with 95% O2 separated from air in the coal gasification section and converted into middle-low heating value syngas.
Corrosive components such as sulfide, nitride and dust are removed from the raw syngas in the cleaning section. The H2S-rich regeneration gas from the acid gas removal system is then fed into the Claus plant, producing elemental sulfur.
The Desulfuration section converts the hydrogen sulfide into sulfur.
To capture the carbon dioxide, a WGS reactor containing a two sections in series with intercooling converts a nominal 96% of the carbon monoxide to carbon dioxide.
The plant will operate at extremely low emissions of regulated air pollutants and will isolate carbon dioxide so that it can be captured. Ammonia is produced from Hydrogen and Nitrogen.
The carbon monoxide and Hydrogen are synthesized here into methane (by-product) in the Methanation section.
Following the cleaning section, the syngas is fed into the Combined Cycle Power Generation section, where the combustion energy is converted in electric energy at high efficiency.
The IGCC model provides a useful description of the process. The simulation takes advantage of Aspen Plus’s capabilities for modeling solid components. This includes tracking component attributes and particle size distribution, and estimating properties for coal. It also produces Methane, Sulfur and Ammonia as by-products.
The model may be used as a guide for understanding the process and the economics, and also as a starting point for more sophisticated models for plant design and specifying process equipment.
Keywords: None
References: None |
Problem Statement: How do I get Data Regression and Property Analysis Results to match? | Solution: In order to illustrate how to get Data Regression and Property Analysis (formerly called Prop-Table) results to match, an example file using Vapor Pressure vs. Temperature data to regress PLXANT parameters is attached.
By default Data Regression uses a Maximum Likelihood algorithm. This means that there are errors in the inputs (in this case Temperature) as well as in the outputs (in this case the vapor pressure).
The regression profile has values for PL-est that correspond to T-est. If the Temperature data is given a standard deviation of zero (as in Data set TNODEV) then the values of PL-est correspont to the
exact Temperatures in the data. The results of the regression (R-TNODEV) that uses this data will give exactly the same results as a Property Table. I have also included a data set where the temperature does have error (TDEV) and a regression (R-TDEV) for comparison.
Generally it is better to use the full Maximum Likelihood equation and have errors in all of your
variables even if you are fitting another correlation since even a correlation has inherent error. However, if you are presenting and comparing the results, it makes sense to set the deviation on the independent variables to zero.
Keywords: DRS, TGS, Data Regression, Table Generation, Analysis, Data , Prop-Table, standard deviation
References: None |
Problem Statement: AspenONE V10.0 SLM Server Installation Guide. | Solution: Please refer to attached pdf for more information.
Keywords: SLM, License Server, Installation
References: None |
Problem Statement: When using the ABE Explorer Simulation importer in ABE V10 to import a PRO/II V9.X file the following error appears:
"Unable to create simsci OLE automation object". | Solution: In order to import a PRO/II V9.X file using the ABE Explorer Simulation importer in ABE V10, the ProII.ini file should properly initialize. As the screenshot below shows the PROII.INI that resides in the program folder, as it existed in previous versions, is NOT the correct version to should be referenced. (Although, please note, the screenshot attached says so):
Thus, in the newer version of PRO/II, they have taken a step-change to creating a customized PROII.INI for each user as their Window’s profile are created. So, simply changing the directory location to the "User" folder will help to solve the issue.
Keywords: Error: Unable to create simsci OLE automation object, Simulation Importer, PROII.INI, PRO/II.
References: None |
Problem Statement: Is it possible to create a design specification with variable tolerance and limits? | Solution: Attached is an example of a Design Specification with variable tolerance and limits. See file - Design-spec2.bkp.
A design specification designates that the inlet and outlet entropies of a Heater block HX1 are equal. The temperature of HX1 is chosen as the manipulated variable. Temperature limits cannot be set a priori, but it is known that the isentropic temperature will be within 75oF of the inlet temperature. The tolerance for the specification is a function of the entropy.
The inlet and outlet entropy and the inlet temperature of the block HX1 are the sample variables. These are all Stream Variables. The entropy of the inlet stream HX1-IN is called SIN. The outlet entropy of the outlet stream HX1-OUT is called SOUT. The temperature of stream HX1-IN is called TIN.
The design specification sets the inlet entropy SOUT equal to the inlet entropy SIN.
The tolerance is specified as the variable TOL. TOL is specified as 0.0001 times the absolute value of the entropy of the inlet stream SIN on the Fortran sheet of the design specification
The design specification is satisfied when |SOUT - SIN| < TOL.
Fortran expressions such as TOL specified on the Fortran sheet can be calculated and used in any part of the specification expression: the spec, the target or the tolerance.
The heater temperature is the manipulated variable. The design specification convergence block will find the heater temperature that makes SOUT=SIN.
The manipulated variable is specified in the heater block just as if there were no design specification. The specified value is the initial estimate used by the design specification convergence block.
The design specification convergence block will not try a temperature less than the inlet temperature TIN - 75F or greater than TIN + 75F, even if the solution to the objective function lies outside this range. The limits become alternative specifications if the design specification cannot be achieved. The initial estimate entered in the reactor block lies within these limits.
You do not have to specify convergence of the design specification. Aspen Plus will automatically generate a convergence block to converge the specification.
For more information see the Aspen Plus Help topic Simulation and Analysis Tools -> Sequential Modular Flowsheeting Tools -> Design Specifications: Feedback Control.
Keywords: None
References: None |
Problem Statement: Is it possible to model a Pipeline Gas from Coal Model with Aspen Plus? | Solution: Aspen Plus is shipped with a number of example files including a few examples of coal gasification. Both Aspen Plus backup (.bkp) files and PDF files describing the details are included in the install directory for most examples. The files for a model of a Pipeline Gas from Coal Model are located in the following location:
V9 and earlier: C:\Program Files (x86)\AspenTech\Aspen Plus Vx.x\GUI\Examples\Pipeline Gas
V10 and later: C:\Program Files (x86)\AspenTech\Aspen Plus Vx.x\GUI\Examples\Power\Coal Gasification
This model simulates a process to manufacture pipeline gas from coal. The model includes the following features:
A set of chemical species including conventional, solid and nonconventional for this process.
Typical process areas including: coal crushing and screening, gasification, shift reaction, methanation, power generation, and the main streams connecting these units.
The specification for nonconventional solid components.
Definition of property model parameters with user data.
Figure 1 shows the process flowsheet which includes: coal sizing and screening, gasification, shift reaction, methanation, and power generation.
__________________________________________________________________________________
Figure 1Coal Gasification Process Flowsheet
The large coal particles (no less than 1mm), are delivered to the three stage coal crushing and screening area for particle reduction. After crushing, more than 93 % of the coal particles are less than 0.18mm and no particles are larger than 0.36mm. Water is mixed with coal feed to aid conveying and crushing. The water phase is separated from the coal before it enters the gasification area.
The small coal particles (containing 25 wt% moisture), are dried using hot nitrogen gas (132 ℃) to reduce the moisture to 10 wt%. The dried coal, with pure oxygen and LP steam is fed to the gasification furnace to produce crude coal gas (908℃, 2.4Mpa). A dust wiper is used to remove the ash generated in the gasification furnace from the crude coal gas. 90% of the crude coal gas is fed to the shift reaction area, and the remaining gas is fed to the power generation area.
In the shift reaction area, 65 % of the feed coal gas is cooled to 380 ℃, mixed with LP steam and passed through two shift reactors in series. To get a higher equilibrium fraction of CO, the gas from the first shift reactor is cooled to 210 ℃ before passing to the second shift reactor. After shift reaction, the gas mixed with that did not undergo shift reaction is fed to the methanation unit.
In the methanation unit the by-product H2S is removed from the coal gas. Then carbon particles, S, SO2 and SO3 are removed. After that, coal gas is fed to the methanation reactor in which almost all of the CO is converted to methane. The product gas from the methanation reactor is purified to remove all of the acid gas (HCL, CO2) and almost all of the water.
Mixed with hot air, 10 % of the coal gas from the gasification furnace is fed to a gas turbine. Waste heat from the gases exiting the turbine is used to boil water for use in a two stage steam turbine. The turbines are connected to electrical generators.
Process summary
Area Purpose
Coal Crushing and Screening Reduce coal particle size
Coal Gasification Decompose coal to produce coal gas
Shift Reaction Change H2/CO ratio for methanation
Methanation Coal gas purification and methanation to produce pipeline gas
Power Generation Burn coal gas to generate electrical power
Keywords: None
References: None |
Problem Statement: I want understand the Simulation Messages in Aspen Custom Modeler:
Is the Step Time different to the integrating time?
Why is the step size "accepted"?
What influences the number of integration steps?
Can I change the frequency of these messages?
Do these messages have anything to do with simulation speed? | Solution: The best way to explain the messages is with an example. Please see below (and attached) and marked-up screenshot of the simulation messages with an explanation of the different sections.
The Step count refers to the variable step integrator which is selected in Solver Options | Integrator | Step size. If the step size is chosen to be Fixed, then the step size is not be evaluated and the step count will not appear in the simulation messages.
Step size evaluation is done for variable step integration. The step size must be such that the error (x - x_predicted) is less than the set tolerances, and therefore error divided by tolerance (error/tol) is less than 1. If the error is larger than the integration tolerance, the step is rejected. If the error is much smaller than the tolerance, then the step size is increased.
The number of steps is influenced by the Absolute integration and Relative integration error tolerances set in Solver Options | Integrator. The tighter these tolerances, the more steps will be made. This is because the step size must be smaller in order to calculate an error less than the set tolerance (or error/tol < 1).
The messages report the integration steps as often as has been set in Run Options | Time control | Communications. This only affects the frequency communication to the messages window. This won't make any difference to the number of actual integration steps taken. For example, if the Communication is set to 0.02 then the "Integrating from timeX to timeY" will be such that timeY - timeX = 0.02. Decreasing this interval (e.g. from 0.02 to 0.2) will dramatically speed up your simulation run time and increasing this interval (e.g. from 0.02 to 0.02) will dramatically slow down your simulation run time.
Keywords: None
References: None |
Problem Statement: How to find Class View Application in the CLE (Class Library Editor). | Solution: One can find the Class View Application right-clicking under the Class View in question and selecting Properties | Application:
This information is important to know, in order to see which datasheet, symbol or label use this Class View in specific.
Keywords: Class View Application, Properties.
References: None |
Problem Statement: Is it possible to control reactor selectivity? | Solution: Attached is an example of using a design specification to control reactor selectivity with the reactor temperature. See file Design-spec.bkp.
The temperature of an RGibbs block REACT is manipulated to control the selectivity of component ESTER versus ETOH at a value of 2.50 +/- 0.01. This example assumes that temperature was specified for block REACT on the RGibbs Specification sheet. The RGibbs specification becomes the initial estimate for the design specification.
The molar flow rate of ESTER and of ETOH in stream PROD are the sample variables. These variables are called FESTER and FALC, respectively.
The design specification is FESTER/FALC = 2.50.
The design specification is satisfied when |FESTER/FALC -2.50| < 0.01.
Fortran expressions such as FESTER/FALC can be used in any part of the specification expression: the spec, the target or the tolerance.
The reactor temperature is the manipulated variable. The design specification convergence block will find the reactor temperature that makes FESTER/FALC=2.5.
The manipulated variable is specified in the reactor block just as if there were no design specification. The specified value is the initial estimate used by the design specification convergence block.
The design specification convergence block will not try a temperature less than 50F or greater than 150F, even if the solution to the objective function lies outside this range. The limits become alternative specifications if the design specification cannot be achieved. The initial estimate entered in the reactor block lies within these limits.
You do not have to specify convergence of the design specification. Aspen Plus will automatically generate a convergence block to converge the specification.
For more information see the Aspen Plus Help topic Simulation and Analysis Tools -> Sequential Modular Flowsheeting Tools -> Design Specifications: Feedback Control.
Keywords: None
References: None |
Problem Statement: We have developed some user models (property, heat transfer, pressure drop, unit operations...) in C++ and would like to use them from Aspen Plus. Is that possible? And how? | Solution: Aspen Plus offers a wealth of interfaces to user subroutines written in Fortran, as documented in the User Models
Keywords: None
References: Guide. We provide template subroutines in \Program Files (x86)\AspenTech\Aspen Plus xx.x\Engine\user. The user written Fortran code has to be compiled using the built-in interface to the compiler ASPCOMP and linked using the built-in linking utility ASPLINK using the Aspen Plus Simulation Engine window.
Thanks to the compatibility between compilers it is possible to use the Fortran templates we provide to create a stub that will access your C++ functions. Although it is certainly possible to link the C++ object files directly to Aspen Plus, the procedure of using the stub is preferable since it uses the documented interface.
Once you have a Fortran stub, you can pass whatever parameter to your C++ functions, including pointers to the Aspen Plus commons. In fact the actual code can be written in any programming language compatible with the calling and linking of the Microsoft Visual Studio family of compilers.
The C++ object files and the Fortran stub object files will have to be linked using the ASPLINK utility.
In the attachments one example of such a stub is offered, together with a dummy C++ function to demonstrate the link.
The following has to be noted:
Create one stub function only in each FORTRAN file
Compile the C++ code with a command such as: cl -c kvalue.cpp
Try the link once and then look in the LD file for an error message similar to the following: stub_kvlu.obj : error LNK2001: unresolved external symbol _KVALUE@20
This is the Fortran stub not finding the uppercase function it expects. Now in the same LD file look for something like:
EXPORTS ?KValue@@YGXQAN0NN0=?KValue@@YGXQAN0NN0@Z
EXPORTS ?KValue@@YGXQAN0NN0@Z
Then to fix the case sensitivities and the match the export name, in the corresponding FORTRAN stub add the following line:
!DEC$ ATTRIBUTES STDCALL, ALIAS:'?KValue@@YGXQAO0NN0@Z' :: KValue
(Note we are using the name given in the second EXPORT line)
A final note of warning: if you have multi-dimensional arrays you will need to review accurately how your code accesses the elements. Fortran and C++ may do that in a different way.
KeyWords: |
Problem Statement: Is it possible to accessing a mole fraction profile in a RadFrac distillation column? | Solution: Attached is an example of accessing a mole fraction profile in a RadFrac distillation column. See the file Molefrac-profile.bkp.
You can use the vector variable types to access an entire block profile, stream or substream at once. Aspen Plus interprets the variable you assign to the vector as an array variable. You do not need to dimension it.
This table shows the vector variables:
Variable type Description
Block-Vec Unit operation block vector
Stream‑Vec Stream vector
Substream‑Vec Substream variable
Compattr‑Vec Component attribute vector
PSD‑Vec Substream Particle Size Distribution (PSD) vector
Un‑Cor‑Vec Temperature‑dependent unary property parameter
Bi‑Cor‑Vec Temperature‑dependent binary property parameter vector
Aspen Plus generates a variable by adding the letter L to the beginning of the variable name which you assign to the vector. The value of this variable is the length of the vector. You can use the variable in Fortran statements, but you cannot change its value.
You can use the Block‑Vec variable type to access column profiles for the following multi‑stage separation models:
In this model Variables depend on
RadFrac Stage and composition
MultiFrac Stage, section, and composition
Extract Stage
PetroFrac Stage, composition, and stripper number
SCFrac Section and composition
You can also use Block‑Vec to access the following block result profiles:
MHeatX zone analysis
RBatch time profiles
RPlug length profiles
Aspen Plus automatically:
Interprets the variable you assign to the profile as an array variable
Dimensions the variable
Aspen Plus generates a variable by adding the letter L to the beginning of the variable name which you assigned to the block vector. The value of this variable is the length of the array. You can use the variable in Fortran statements, but you cannot change its value.
The order of values in the array depends on which variable you select. All values are in SI units, regardless of the Units specifications on the Define sheet.
Examples of vector variables dependent on component number and stage number are liquid and vapor composition profiles in RadFrac, MultiFrac, Extract, or PetroFrac. The values are stored as one‑dimensional arrays. All component values for stage or segment 1 are at the beginning, followed by all of the component values for stage or segment 2, and so on. The number of components and the component order are the same as on the Components Specifications Selection sheet.
For a column with three components and five stages, the liquid composition profile is stored as follows:
Array Index Value for
1 Component 1, stage or segment 1
2 Component 2, stage or segment 1
3 Component 3, stage or segment 1
4 Component 1, stage or segment 2
.
.
.
15 Component 3, stage or segment 5
For more information see the Aspen Plus Help topic Simulation and Analysis Tools -> Sequential Modular Flowsheeting Tools -> Accessing Flowsheeting Variables.
In the attached example, the entire liquid mole fraction profile of a RadFrac column with three components is accessed. The value for the second component on the fifth stage is written to the Control Panel, using a Fortran block. On the Define sheet of the Calculator block, variable XPROF of the type Block-Vec. On the Calculator Calculate sheet, include these Fortran statements:
WRITE(NTERM,10) C* TOTAL NUMBER OF COMPONENTS IS 3 *
NCOMP = 3 C* COMPONENT TO BE ACCESSED IS 2 *
ICOMP = 2 C* STAGE TO BE ACCESSED IS 5 *
ISTAGE = 5 C* CALCULATE INDEX INTO XPROF *
II = NCOMP*(ISTAGE-1) + ICOMP WRITE(NTERM,20) XPROF(II)
10 FORMAT(' MOLE FRACTION OF 2ND COMPONENT ON 5TH STAGE ') 20 FORMAT(10X,F10.2)
Keywords: None
References: None |
Problem Statement: How can binary parameters from one property method be used in another property method? | Solution: You can generate VLE and LLE data for a binary system, using a specified property method. Then, you can use the generated data in data regression to determine the binary parameters for another property method. This generated data can be regressed alone or with other data sets. With this feature you can convert parameters between different property models. General instructions for this process can be found in the Aspen Plus help.
The attached example file shows how to determine the NRTL binary parameters for water/isobutanol using LLE data generated from the UNIQUAC model at 10 and 40 degrees C. The UNIQUAC binary parameters are from the built-in Aspen Plus LLE-LIT databank. Note that the Property Method of the generated data can NOT be the same as the Property Method being used in the regression.
In general, the steps required are:
To specify generation of binary VLE and LLE data:
In the navigation pane, click Data. The Data object manager appears.
To create a new Data ID, click New. In the Create New ID dialog box, enter an ID or accept the default. Choose MIXTURE in the Select Type list box, and click OK.To edit an existing ID, select the Data ID from the object manager, and click Edit.
On the Setup sheet, choose the type of property data in the Data Type list box:
Select To generate this data
TXY, PXY, or TPXY VLE
TXX or TPXX LLE
Note: Do not select the GEN-TPXY or GEN-TPXX data type
Select the components from the Available Components list and click the right arrow button to move them to the Selected Components list.
In the Temperature and Pressure fields, if active, specify a constant temperature or pressure at which the data will be generated.
Click the Data tab.
On the Data sheet, click the Generate Data button.
In the Generate Binary VLE or LLE Data dialog box, select a property method, and a Henry's Components ID and Chemistry ID, if applicable.
Click the Generate button. This process can repeated for addition data sets at different temperatures or pressures. The Data sheet displays the liquid phase compositions for which data are to be generated for the regression. When you run the regression, these parameters will be generated and displayed on the Regression | Results | Profiles sheet.
Keywords: None
References: None |
Problem Statement:
How do I set up a shortcut drying model? How do I define solid stream types? How do I overcome dryer model errors and warnings? What does it mean to define moisture in the solid stream class? | Solution:
Have a look at the attached Aspen Plus V10 simulation file, Drying_start.bkp.
1. Components:
In this example, cobalt solid on a zeolite substrate is being dried in air. The cobalt and the zeolite components are of Type: Solid. The component list also contains air and water, these are of Type: Conventional.
2. Stream class:
The stream class for the simulation must be defined to take into account a solid stream class. This has been defined in Simulation Environment | Setup | Specifications | Stream class. Stream class MIXCIPSD has been chosen for this example which means that streams classes Mixed and Solid will be used, and that the Solid stream will be specified with Particle Size Distribution (PSD).
3. Moisture:
Water has been chosen to be the moisture component in the solid stream. This is specified in Setup | Solids
4. PSD Mesh:
The PSD Mesh is defined in Setup | Solids | PSD. The mesh is the list of size fractions that all solid streams' PSD will be reported on. This example has a geometric mesh type with smallest size 1 micron and largest size 60 microns.
5. Flowsheet:
The cobalt, zeolite and water are joined together in mixer model B1 before being transferred to Dryer model B2. Air is input to the dryer. The dryer model uses the Shortcut type and is specified with a pressure of 1 atm and a temperature of 100 °C. The dryer model is specified to output a solid stream with a moisture fraction of 0.0001. The exit PSD has also been defined for the dryer model.
6. Reset and Run the model without making any changes.
You will find the simulation runs with 1 error:
** ERROR
SOLID EXISTS IN THE MIXED SUBSTREAM.
SHORTCUT DRYER DOES NOT SUPPORT THIS. BLOCK BYPASSED.
This means that your solid components are not correctly assigned to the Solid stream class. The stream results show that none of the water has been removed from the solid stream through the dryer (because the dryer has been bypassed).
7. Resolve the error:
Open up input stream S1. Notice that there are two stream classes available for definition: Mixed and CI Solid. The zeolite has been defined in the Mixed stream whereas it should be defined in the CI Solid stream. Remove it from the Mixed stream and rather input the flowrate to the CI Solid stream. The Mixed stream inputs must be completely cleared for that form to be accepted.
The PSD for this inlet stream must also be defined - you can choose your own PSD, or copy the values from the Dryer.
You need to do the same for stream S2, the cobalt.
8. Reset and Run the model after correcting the allocation of the solids to the Solids stream class.
Now you will find that the simulation runs with 1 warning.
Block: B2 Model: DRYER
* WARNING
A CALCULATED EVAPORATION RATE IS NEGATIVE
This is caused by the absence of water in the Solid substream. The dryer needs moisture in the CISOLID substream to be able to run.
9. Resolve the warning:
Have a look at stream S3. This stream is defining the water present on the cobalt-zeolite. However, the water needs to be present in the Solid stream class for the dryer model to work. Therefore, the water needs to be defined as moisture in the solid. This is why we defined water as the moisture component in step 3. The water content in S3 must be defined in stream S1 or S2 (or split between the two, but it doesn't make a difference) in the Solid stream class and S3 can be deleted.
10. Reset and run with water defined in S1/S2 Solid stream class
The simulation now runs without errors or warnings. Have a look at the stream results and find that the water content has moved from the Solid stream class in S4 to the Mixed stream class of S7. The dryer now works correctly.
The final corrected flowsheet is attached as Drying_Complete.bkp.
Keywords: None
References: None |
Problem Statement: When Using HeatX block in EO mode, one cannot see results for block. In EO-Variables table there are only variables for connected streams.
Is it something wrong with block specification? Where to find variables like: Duty, PDrop, U, A for the HeatX block? | Solution: HeatX block allows phase splitting into 2 or 3 outlet streams for each side of the heat exchanger.
However, multiply product streams are not supported in EO-mode, and when using this feature HeatX block display in EO-Variables table only results for streams connected to heat exchanger.
To see all block variables, like: Duty, PDrop, U, A, user should specify single outlet stream for each side of the heat exchanger, or 2 product streams only while 1 of those is free water.
One can specify flash block located after HeatX block, to perform phase splitting for heat exchanger output stream.
This information is not available in the online help, but when trying to use this feature one can see the following information in the control panel:
INFORMATION WHILE INSTANTIATING THE EQUATION-ORIENTED UNIT OPERATIONS
BLOCK: "B1" (MODEL: "HEATX")
MULTIPLE PRODUCT STREAMS ARE NOT SUPPORTED IN EO UNLESS
THERE ARE 2 PRODUCT STREAMS AND 1 OF THOSE IS FREE
WATER
Fixed in Version
246203: There is a plan to include note about this limitation in the online help: EO usage notes for HeatX block, in version V11.
Keywords: HeatX block, EO mode, multiply product streams
References: None |
Problem Statement: Column section pressure drop is not updated when the overall section efficiency is changed in Column Hydraulics. Is overall section efficiency used? | Solution: In Aspen Plus V10, the overall section efficiency on the Column's Column Internals | Sections | Geometry | Design Parameter form is used in the pressure drop calculations if "Update pressure drop" is selected on the Column's Column Internals | Sections form. However, the overall "Section pressure drop" result on the Column's Column Internals | Section | Results | Summary and the "Total pressure drop" results on the By Tray forms are not updated when there is a change to the overall section efficiency. The workaround is to review the stage "Pressure" results reported on the Column's Profiles | TPFQ Form which are updated correctly.
Note that in Aspen Plus V9, the overall section efficiency is not available.
Please see the attached example file.
Fixed in Version
Targeted for a future release
Key Words
Column Hydraulics, Section Pressure Drop, Efficiency
VSTS 21707
Keywords: None
References: None |
Problem Statement:
No stream results reported in Aspen Plus | Solution:
There is an option in Aspen Plus on the Setup | Report Options | General sheet to skip the report generation completely.
This option is sometimes used to turn the report off when models for real-time optimization since the user would never see the report anyway. However, it has the side effect of preventing Aspen Plus from generating the stream report data into the User Interface. To fix this issue, make sure that the “Generate a report file” box is checked.
Keywords: None
References: None |
Problem Statement: How do I launch Aspen Plus from the Command Line (DOS-like window)? | Solution: To open Aspen Plus from the command line rather than the graphical user interface, please navigate to Start | AspenTech | Aspen Plus VX | Customize Aspen Plus VX
Keywords: None
References: None |
Problem Statement: Hydraulics T-Junction Mixer arm angle reference | Solution: To specify the design of the T-Junction Mixer:
On the Design tab, click the Data page:
In the Side Arm Angle to Main field, type in the angle value of the mixer side arm.
In the Side Arm Angle to Vertical field, type in the angle value of the mixer side arm.
In the Main Arm Angle to Vertical field, type in the angle value of the mixer main arm.
See the following sketch with axis references for the Side Arms angles calculation:
Keywords: Hydraulics, Side Arm Angle,
References: . |
Problem Statement: How to access batch time from Aspen Batch Modeler in Aspen Simulation Workbook.
An important result from a batch model is the batch time. This tip shows you how to access the batch time from an Aspen Batch Modeler simulation through Aspen Simulation Workbook. | Solution: The batch time is the variable CurrentTime. For batch model B1, it is B1.CurrentTime.
1. Open attached Batch Modeler V10 file BenzeneMCBDCB.bspf and run it. Find that it ends running at Batch time = 1.39 hours.
2. Close Batch Modeler file.
3. Open predefined Excel and ASW file Batch_ASW.xlsx and connect to the Batch Modeler file BenzeneMCBDCB.bspf
4. See that CurrentTime is 1.39 hours
5. Reset simulation in the Excel file, find that CurrentTime goes to zero.
6. Run simulation in the Excel file, see CurrentTime changing with the run time and it ends at 1.39 hours, which matches the Batch Modeler Batch time.
7. Find current time in the organizer by creating a query to search for variables with dimension = time and status = calculated.
Keywords: Aspen Batch Modeler, Batch Time, CurrentTime, Aspen Simulation Workbook
References: None |
Problem Statement: How do I prevent error -939519987 when installing a new license file?
Description of error message :
Server: Local license server (NO-NET)
Bucket: default
Type: Standard License
Code: -939519987
Description: This may be due to a time-restriction on the license | Solution: Note : This solution is applicable only for stand-alone license
This error is encountered only when the license file is not properly installed and initialized.
Also, it is possible that the old license file (which has already expired) is still present in the same location
In order to solve this error, the older license file has to be moved to a new location, to do this, go to the following locations:
C:\Program Files\Common Files\AspenTech Shared
If you are using a x64 Windows, please also go to the folder:
C:\Program Files (x86)\Common Files\AspenTech Shared
And search for license files like these ones:
Select all the license files from the indicated folders and move them to another location ( a new folder called old licenses could be created and move the files there).
Then, install the new license file again, for this, follow the KB 111593
If this doesn't work, then manually save the license file in the above mentioned folders.
Keywords: Error code -939519987
License installation error
References: None |
Problem Statement: In Stream | Properties, there is water dew point[Gas] property and it is indicated at Liquid phase property.
User might get confused when 'Water Dew Point' is indicated in liquid phase. | Solution: Water Dew Point - Liquid Phase : This is the temperature when aqueous phase is formed based on the current composition and pressure of liquid phase.
After separate this liquid phase and decrease the temperature, user can find that aqueous phase is formed at -25.41 ℃. Thus, water dew point means 2nd liquid formation temperature in this case.
Similarly, Water Dew Point - Vapor Phase : This is the temperature when aqueous phase is formed based on the current condition of vapor phase.
Water dew point is calculated by simple flash calculation to find when aqueous phase if formed.
Water Dew Point - Aqueous Phase : As aqueous phase is pure water, we can see 99.96℃ which is the dew point of water at specified pressure (in this case, 1 atm) .
If the stream doesn't have water contents, It will show '(EMPTY)' as a water dew point. In this case, user can use 'HC Dew Point' to analyze HC liquid formation temperature.
Keywords: Water Dew Point, Saturated, Property
References: None |
Problem Statement: Adding Server Name to SLM Configuration Wizard gets a Server Error:
"Error_while_checking_out_license_(SLM_LicenseProfile) 1000
Do you still wish to add the server?" | Solution: In this situation, the user moved the network license file, moved the USB dongle to another machine, installed the SLM Server, but did not install the dongle driver.
To resolve this issue follow these steps:
1. On the SLM Server, install the dongle driver from the aspenONE Engineering media and ensure to check the box for USB dongle driver.
or download the attached dongle driver from this KB article.
2. On the client machine, navigate to C:\Program Files (x86)\Common Files\AspenTech Shared\
Then, right click on the SLMConfigurationWizard, run as Administrator, add the SLM Server Name, and then click Apply button.
3. Run Aspen HYSYS or Aspen Plus and open a case to verify the license works.
Keywords: SLM Configuration Wizard, dongle driver, SLM License Manager, Configure, Server Error, and error 1000
References: None |
Problem Statement: How to copy values from a snapshot or a result with a script? What is the difference between ResultCount and SnapshotCount? How to trap an error if the snapshot or result does not exist? | Solution: This script can be used to copy values from a result.
' Before using any of the Automation methods and properties, you need to issue
' a Refresh command. This command ensures that automation has access to all the
' currently available snapshots and results in the simulation.
' If you do not issue a Refresh command, you may find that the snapshots and results
' to which you have access do not match the snapshots and results in the current
' simulation.
' refresh the snapshots & results
Application.Simulation.Results.Refresh
' Returns the number of results that are available for the current simulation.
' This includes all kept results and results contained in result files.
count = application.simulation.results.resultcount
application.msg "there are " & cstr(count) & " results available."
' now we want to "use" one result called "test"
dim aResult
set aResult = Application.Simulation.Results.FindResult("test")
SourceFixed = False
SourceFree = True
SourceInitial = True
DestinationFixed = False
DestinationFree = True
DestinationInitial = True
SourcePath = "B1"
DestinationPattern = "~"
Application.Simulation.Results.CopyValues aResult, _
SourceFixed, SourceFree, SourceInitial, _
DestinationFixed, DestinationFree, DestinationInitial, _
SourcePath, DestinationPattern
If you want to copy the values from a snapshot, the same principle can be used. See the attached file for further examples.
What is the difference between ResultCount and SnapshotCount?
Assume we have the following snapshots and results in the simulation.
When you run the script "counting_snapshots" from the attached example you will get the following messages:
ResultCount= 6
The results are:
test
test2
test33
test4
test3
test1234
SnapshotCount = 4
The Snapshots are:
test1234
testsnap
Initial Specification
Initial Specification
As you can see test1234 is both a result and a snapshot: this is because it is flagged as a "kept results", but it still current with the simulation (dark icon instead of a dimmed icon). This means also that when you want to delete "test1234", you need to delete it twice, first as a results, and secondly as a snapshot (or the other way around, that does not matter).
How to trap an error if the snapshot or result does not exist?
Use the "on error resume next". See VBScript documentation for details. Note that VBScript supports only on error resume next, while VBA or Visual Basic offers more options (on error goto 0 and on error goto label). There is no way in VBScript to restore to the default error handler.
' first refresh the snapshots
application.simulation.results.refreshsnapshots
on error resume next
dim asnap
set asnap = application.simulation.results.findsnapshot("test3")
' write a message if the snapshot does not exist
if err.number <> 0 then
application.msg err.description
err.clear
else
application.msg asnap.description
application.msg asnap.DateTime
end if
The scripts in the example file are not testing for errors to keep the code short.
KeyWords
script, snapshot, results, automation
Keywords: None
References: None |
Problem Statement: The Constant Superimposed Backpressure (BP) is the pressure at the valve tail pipe outlet imposed by the discharge environment. The constant superimposed backpressure is the backpressure which will not vary based on the state of the discharge system (that is, simultaneous discharge from other valves in the plant). The default for this value is the atmospheric pressure.
The default Constant Superimposed BP for Safety Analysis PSV scenarios is set in the Preferences Manager. In the Safety Analysis Environment, in the Home Ribbon, you will find the Preferences Manager. In the General Setup you can edit the default value for Constant Superimposed BP.
Changing the Constant Superimposed BP value in the Preferences Manager will not change the Constant Superimposed BP for any scenarios that already exist in the workspace. It will only apply to newly created PSVs. This means that to use the updated Constant Superimposed BP you would have to delete your existing PSVs and recreate them from scratch. This Tech Tip covers a workaround to avoid having to delete your PSV scenarios. | Solution: To change the superimposed BP for existing scenarios, rather than starting your PSV design from the beginning, you can add the pressure difference between your desired value and default value to the Variable Superimposed BP instead. This will have the same effect as updating the Constant Superimposed BP because the total Superimposed BP is the sum of Constant and Variable. Update the Variable Superimposed BP in the Scenario Setup by clicking Edit next to the Total Backpressure field.
The Total Backpressure is the pressure at the outlet flange of the relieving device. This is the pressure used to calculate the properties of the fluid at the exit of the relief device. This pressure is the sum of the built-up backpressure and the superimposed backpressure (constant and variable).
The Variable Superimposed BP is the pressure at the valve tail pipe outlet imposed by the discharge environment; usually the result of one or more PSVs discharging into a common header. The variable superimposed backpressure is the backpressure generated by the simultaneous discharge of other valves when the scenario under consideration is causing the relieving from other valves in the plant (for example, general power failure). The S.I. Variable Backpressure term is added to the S.I. Constant Backpressure term and is in gauge units.
For example, if you have created a PSV scenario using the default Constant Superimposed BP of 1.013 barA, but you want to work with a Constant Superimposed BP of 2 barA, then you can make the Variable Superimposed BP equal to (2-1.013=) 0.987 barA. This is the same as having Constant Superimposed BP = 2 and Variable Superimposed BP = 0.
Keywords: Safety Analysis, Backpressure, Constant Superimposed, Superimposed backpressure
References: None |
Problem Statement: How to remove or edit the characters next to the big arrows of feed and products streams?
A file may become corrupt, and in its recovery it may have put characters next to the arrow icons of feed and product streams.
What are these characters called? How can I remove them? How can I change them? | Solution: The characters next to the feed and product arrows are called terminators. They can be found in the Flowsheet preferences under File | Options | Flowsheet | Formatting. Change the text in the terminators box to change the characters next to the feed and product stream arrow icons.
Keywords: Flowsheet formatting, terminators
References: None |
Problem Statement: How to use the optimization tool in Aspen Plus, an example used in the the removal of CHC and BTEX from groundwater by stripping with air.
Process description:
Groundwater flowrate is 10 m3/h containing 5 mg/L tetrachloroethylene, 1 mg/L trichloroethylene, 5mg/L benzene, 3 mg/L toluene. The groundwater has a hardness of 15 °dH and a pH of 6. The organics are removed from the wastewater by stripping with air in a multistage absorption column. The target product stream must contain less than 10 µg/l CHC's and less than 10 µg/l BTEX. The air to water ratio must be less than 100 and the flowrate of air must be minimized to reduce downstream air treatment.
This tech tip takes you through the setup of such a model, including how to use design specifications to reach process targets, as well as using optimization to find the optimum air flowrate. | Solution: Properties Environment:
The components selected are WATER (ground water), C2CL4, C2HCL3, BENZENE, TOLUENE (groundwater contaminants), N2, O2 (air), HCL (for water pH), CACO3 and MGCO3 (for water hardness). The property method was set up using the Electrolyte Wizard, because acid, CaCO3, MgCO3 in water will form an ionic system and we want to report pH. The Electrolyte Wizard added these ions to the component +list: MG2+, CA2+, H3O+, CO2, HCO3-, CL-, OH-, CO3-- and the associated equilibrium and dissociation chemistry was created for these ions.
Simulation Environment:
The flowsheet consists of a mixer, to make up the specified groundwater, and a RadFrac block to model the absorption column.
The mixer input are an acid stream, a water stream with low concentrations of organics, and a Ca and Mg stream to account for the mineral content of the water. The acid flowrate is controlled by a Design Spec called pH to target a pH in the groundwater stream of 6.
A calculator block called HARDNESS calculates the parameters to control the hardness of the groundwater stream. Stream volume flowrate is found from mass flow and density. Stream hardness is calculated by the concentration of Ca and Mg. Hardness in mg/L is then converted to °dH to use the 15 °dH setpoint. A design specification called HARDNES has then been setup to target the °dH to 15 by varying the flowrate of the Ca-Mg stream to the mixer.
A calculator block called FDRATIO ratios the airflow and the groundwater flowrate. This is used in the optimization study as a constraint (C-2), that FDRATIO must be less than equal to 100. The other constraint (C-1) sums the content of CHC's in the column bottoms and it is specified that this must be less than or equal to 0.01 mg/L.
The optimization (O-1) is to minimize the content of CHC's and BTEX in the column bottoms subject to the constraints, by varying the flowrate of air.
After 3 iterations, the optimization is achieved with constraints met. The final air flowrate is 1289 kg/h and the final CHC+BTEX content in the column bottoms is 3.4 µg/l.
Keywords: Groundwater stripping, air stripping, optimization, electrolyte, water hardness
References: None |
Problem Statement: Hydraulics T-Junction Mixer arm angle reference | Solution: To specify the design of the T-Junction Mixer:
On the Design tab, click the Data page:
In the Side Arm Angle to Main field, type in the angle value of the mixer side arm.
In the Side Arm Angle to Vertical field, type in the angle value of the mixer side arm.
In the Main Arm Angle to Vertical field, type in the angle value of the mixer main arm.
See the following sketch with axis references for the Side Arms angles calculation:
Keywords: Hydraulics, Side Arm Angle,
References: . |
Problem Statement: Why is the pure component thermal conductivity (K) different to the mixture thermal conductivity (KMX) for a stream containing a single component?
In the attached Aspen Plus V10 file called Default.bkp, see the stream results for S1. This stream contains only ethylene. However, the pure component thermal conductivity reported for ethylene is different to the mixture thermal conductivity reported for the stream. | Solution: The reason why the two values are different is because the selected property method (RK-SOAVE in this case) is using different routes to calculate thermal conductivity for pure components and for mixtures, and because the stream is at high pressure, these two models give quite different results. To see which routes are being used, and to change the routes to achieve the same result for pure component and mixture thermal conductivity, follow these steps:
1. Navigate to the Properties Environment | Methods | Selected Methods | RK-SOAVE | Routes. Find property KVMX (vapor thermal conductivity of mixture) and hover over the route ID KVMX01. You will see that the route chosen uses the Wassiljewa-Mason-Saxena model to calculate KVMX.
3. Find KV (vapor thermal conductivity of pure component) and hover over the route ID KV01. This route is using the Stiel-Thodos Pressure Correction model to calculate KV. This is why different values are calculated for mixture vs. pure component thermal conductivity.
4. To get the same results for mixture and pure component thermal conductivity, both properties should be calculated using the same model. For high pressures, you can change the KVMX route to use the Stiel-Thodos pressure correction model as is used for KV. Change the Route ID to KVMX02 to do this.
5. Once the selected routes are using the same model, the results for thermal conductivity will match. The updated file is attached (Route Changed.bkp).
Keywords: Thermal conductivity, pure component, mixture, KVMX, KV
References: None |
Problem Statement: When looking for the Data and Configuration Wizard in V10, it is not available in the Start Menu. Neither can it be found through using keywords such as "data", "transfer" or "wizard". | Solution: The reason Data and Configuration Wizard cannot be found using words like “data”, “transfer” or “wizard” is due to the fact that it is not available as a shortcut in the Start Menu.
The executable, Infoplus21SettingsWizard.exe, for Data and Configuration Wizard can be found in C:\Program Files (x86)\AspenTech\InfoPlus.21\db21\code or C:\Program Files\AspenTech\InfoPlus.21\db21\code depending on whether the installed Aspen InfoPlus.21 is 32-bit or 64-bit.
In order to make is searchable, the following steps can be performed.
Browse to C:\Program Files\AspenTech\InfoPlus.21\db21\code in Windows Explorer.
Right-click on Infoplus21SettingsWizard.exe.
Select Send to | Desktop (create shortcut).
Rename the shortcut as Aspen Infoplus.21 Data and Transfer Wizard.
Cut and past into C:\ProgramData\Microsoft\Windows\Start Menu\Programs\Aspen Manufacturing Execution for example.
Once the above steps had been performed, it is accessible from the Start Menu and can be searched using words like "data", "transfer" or "wizard".
Keywords:
References: None |
Problem Statement: Why is the isentropic exponent for a mixture containing a single component not equal to that of the pure component?
How to customize the property method to result in the identical isentropic coefficients for mixture and pure component when the stream contains only a single component. This is done by editing the liquid molar volume calculation route. | Solution: Isentropic exponent is calculated using molar volume. For identical pure component and mixture isentropic coefficients for a stream containing a single component, check that the molar volume for a mixture and a pure component are being calculated with the same model.
The attached video (ISENEX.wrf) will take you through step-by-step instructions on how to check the calculation route for molar volume. The video is saved as a Windows Media Video (.wmv) and as a WebEx Recording (.wrf). For information on viewing this file, please go to https://www.webex.com/play-webex-recording.html.
Keywords: Isentropic exponent
Isentropic coefficient
Molar volume
Property method customization
Selected methods
Property method routes
Property method models
THRSWT
Rackett
VLMX
ISENEX
ISENEXMX
References: None |
Problem Statement: Trouble reading stream properties from XML file with multiple units of measure
Cause
This problem can happen when the simulation requests a specific unit of measure (UOM) in one Property Set that is not the default unit and does not request this UOM in any other Property Set. The XML file needs to provide the information in more than one unit of measure. If the particular UOM is requested for the property in ALL property sets requesting that property, the values would show in the single header. If a UOM is NOT requested for a particular phase, no value will be returned. Aspen Plus separates the tags in order to satisfy the request.
All the tags are at the same level in the XML file. When an external program gets to the first KMX__PBS, it sees there are no values and moves on. It gets to the next KMX__PBS and may skip it since it already read that tag leading to the values not being read. The values are present, but are not readable by normal programming methods used for reading an XML file. | Solution: If there are no values for a property in their XML report, either (1) evaluate the XML for secondary property tags that may contain the data just separated out, or (2) review the property sets to see if they are requesting inconsistent UOMs for the same property in different property sets.
Either add the other UOM to the property set list or removing extra UOMs by defining desired UOMs in the global units. The issue with adding units of measure is that ANY time a user adds an additional UOM to the property set, the values get split from the main property tag, leading to possible instances of no values readable by code.
Keywords: XML, property set
References: None |
Problem Statement: What do the following .bat file in the APED folder do?
DBCreate.bat
DBInfo.bat
DBRestore.bat
DBStart.bat
DeleteDBInstance.bat | Solution: These batch files can be used to perform some basic functions in the APED Database Manager. Please note that it is not necessary to use these batch files. The commands used in these batch files are already integrated into APED GUI tools, Database Manager and Database Tester. These batch files are still delivered because occasionally they may help for trouble-shootings or quick check.
Four of these files, DBInfo.bat, DBStart.bat, DBCreate.bat, and DeleteDBInstance.bat apply to SQLLocalDB only while DBRestore.bat can apply to both SQLLocalDB and SQL Server. Unlike SQL Server, SQLLocalDB runs as an application instead of a server so that it can runs only as needed and stops when not needed. In general, it consumes much smaller memory as compared to SQL Server. Each user has its own instance regardless of its windows administrative privilege.
The batch files DBInfo.bat, DBStart.bat, and DBCreate.bat can be used to manage a SQLLocalDB instance. The following is an example to illustrate how can be accomplished:
DBInfo.bat prints the status of the SQLLocalDB instance, AspenInstance36:
Running:
Stopped:
When the instance does not exist, you will see this:
When the instance is stopped, run DBStart.bat to restart the instance:
When the instance does not exist, run DBCreate.bat to create the instance; it will automatically start the instance:
It is possible to add custom databases to SQLLocalDB using the DeleteDBInstance.bat and DBRestore.bat batch files using the steps below:
Copy the backup files of their custom databases to C:\ProgramData\AspenTech\APED V10.0.
Run DeleteDBInstance.bat. This will drop any existing database to the SQLLocalDB instance and then delete the instance.
Run DBRestore.bat. This will run DBRestore.exe which is one of APED exe files delivered. It will check the status of the SQLLocalDB instance and then decide what batch files above need to be called. This will ensure the SQLLocalDB is running and ready for restoring databases from all backup files, including custom database backup files.
Keywords: None
References: None |
Problem Statement: Aspen Economic Evaluation には3つの製品がありますが、違いは何ですか? | Solution: 添付の資料にまとめましたので参照ください。
Keywords: Japanese, 日本文、Aspen Capital Cost Estimator, Aspen Process Economic Evaluator, Aspen In-Plant cost estimator, 特徴
References: None |
Problem Statement: Why do I get an error massage in Aspen EDR while converting the shortcut HeatX block to shell & tube through Aspen Plus. Three property data sets were created for stream & while running the same in EDR gives error massage – “Specific enthalpies for the stream do not increase with temperature”. | Solution: It is observed that in some cases, while converting from shortcut exchanger to shell & tube mode through Aspen Plus, the three property sets would be generated with respect to pressure & if user run the simulation - gets Input Error - 1802 error massage for property.
This is due to the temperature difference at the last data points where the temperature is getting decreased instead of increasing as compared to earlier data points. It may happen with single component flow with feed and outlet are covering two phases or near two phases as shown in below snap.
This issue could be resolved with Aspen Plus file by tightening the flash tolerance of stream of HeatX block at options to make it work. This is a single component flow and the bubble and dew supposed to be the same:
Keywords: None
References: None |
Problem Statement: Why is critical pressure Pc of Triethylene Glycol (TEG) in Aspen HYSYS completely different from reported values from other sources and literature? | Solution: TEG critical pressures reported in many literature are around 450 psig, while that value Pc in Aspen HYSYS is 191 psig. The incorrect value in HYSYS is due to historical reason when the software was developed from Hyprotech - likely that it was adjusted to match TEG vapor pressure for HYSYS PR . We have decided to not change the value to maintain consistency inside the software, as physical property models have already been re-parameterized to account for this discrepancy.
Keywords: TEG, critical pressure, Pc
References: None |
Problem Statement: Aspen Batch Modelerのバッチ蒸留塔やバッチ反応器のチュートリアルや例題ファイルはありますか? | Solution: チュートリアル
チュートリアルはAspen Batch Modelerのアプリケーションの中のHelpメニューからアクセスできます。
Aspen Batch Modelerを起動 (Start / Programs / AspenTech / Process Modeling V8.8 / Aspen Batch Modeler)
ToolbarからHelpメニューを検索 (Help / Aspen Batch Modeler Contents)
画面左から、“Getting Stated with Aspen Batch Modeler”をクリック。
サブメニューから、“Tutorial”をクリック。
例題ファイル
例題ファイルはAspen Batch Modelerのインストールフォルダーの中にあります。
例えばCドライブにインストールした場合は、下記フォルダーにあります。
C:\Program Files\AspenTech\Aspen Batch Modeler V8.8\Examples
Keywords: Japanese, 日本文, Batch Modeler, Tutorial, 自習書
References: None |
Problem Statement: Some settings can be changed when clicking in areas of dialog box that are far to the right of the actual checkbox. | Solution: When maintaining settings in the dialog boxes, it is possible to inadvertently change settings by clicking far to the right of an actual checkbox. For example, a click in the area highlighted below will change the setting “Allow Zero Purchases for VPOOL Components”. Please be careful about where clicks are made to avoid unintended changes or use Cancel to close the dialog unless you have specifically made and verified a desired change as we investigate the issue.
Keywords: None
References: None |
Problem Statement: How does EXTRA-ML on the RadFrac Convergence | Advanced work? | Solution: If a design specification in a RadFrac column does not converge, there is generally a message in the Control Panel about the Middle Loop not converging.
In later versions, this warning suggests that the user can specify EXTRA-ML:
* WARNING WHILE EXECUTING UNIT OPERATIONS BLOCK: "REFINING" IN "DISTILL"
(MODEL: "RADFRAC") (UDL3ZR.7)
Outside loop tolerance was satisfied but DESIGN SPEC iteration
(Middle loop) did not converge to requested tolerance.
Try to specify EXTRA-ML=1 in Convergence | Advanced.
EXTRA-ML is the extra number of middle loop convergence iterations if the design spec tolerance when using a nested method for convergence (TOLML) is not satisfied. EXTRA-ML can be between 0 and 5. The default is 0. EXTRA-ML will only be used when there is already a problem with the convergence.
Also note that TOLML (default 0.001) is not used in the middle loop convergence test. It is used strictly as a post solution check after the outside loop is converged.
Keywords: None
References: None |
Problem Statement: Aspen Audit & Compliance Administrator is showing database error and the account used to configured Aspen Audit and Compliance Database service had been verified to be using the correct password. | Solution: In this case, this may have been caused by a change in the database related to the account used specifically if it is Microsoft SQL Server.
In Microsoft SQL Server Management Studio, go to Security | Logins and check on the Properties of the account used which generally is AspenAE.
In General page, check that the Default database is set to AuditAndCompliance.
In Server Roles page, make sure that AspenAE is not assigned sysadmin role.
In User Mapping page, check that the Default Schema is set to AspenAE.
Specifically note that members of the sysadmin fixed server role are automatically mapped to dbo.
This can also be verified by logging into Microsoft SQL Server Management Studio using the AspenAE and issue the below query.
SELECT SCHEMA_NAME();
The SCHEMA_NAME() will return the default schema of the caller. If it returns dbo instead of AspenAE, check whether sysadmin role was given to AspenAE.
Keywords: E_FAIL
Database error
0x80040e14
Could not find stored procedure 'USP_EVENTHEADER'
References: None |
Problem Statement: There is a feature in Aspen PIMS for SQL Server Database Support. It allows you to send results to a SQL Server database or both Access and SQL server. But first, you have to configure an Aspen PIMS output database in SQL using scripts comes from the Aspen PIMS installation. This solution describes the procedure for creating an Aspen PIMS SQL database. This procedure is for Aspen PIMS versions 10 and earlier. In version 11 there is a tool that creates a SQL database for you and the steps below are not necessary. | Solution: Aspen PIMS model execution has a unique identifier, Solution ID, associated with each run. When an execution run occurs, the solution data is written to an output database. All report data is obtained from the output database that is specified from the General Model Settings | Output Database.
The following information is applicable if you plan to store Aspen PIMS model data results in a Microsoft SQL Server database.
Create the SQL Server Instance
We recommend that a qualified Database Administrator (DBA) create and configure the SQL Server database for the use with Aspen PIMS.
For best results, and in order to avoid disk contention, at a minimum, please install the following on different physical drives:
? Operating System
? Database
? Logs
For additional information or clarification on SQL Server installation recommendations, please consult Microsoft for the latest best practices.
After installing SQL Server, perform the following related to your server:
1 Enable Allow Remote Connections.
2 Enable TCP/IP.
3 Check the default SQL Server port; 1433 should be automatically added to the Windows Firewall exceptions list. This is a trouble shooting tip.
Check with your IT department for specific details related to your organization.
Create the SQL Server Database
Complete the following procedures on the computer where SQL Server is installed to create the target Aspen PIMS results database.
Create the database using the provided SQL script
1 On the machine where PIMS is installed, go to:
C:\Program Files\AspenTech\Aspen PIMS\Enterprise Configuration\Database.
2 Using SQL Server Management Studio or a text editor, open the PimsDbCreationScript.sql script. This script is used to create the database.
3 Replace all instances of %databaseName% with the name of the database that is being created, e.g. ResultsPIMS.
4 Replace all instances of %databasepath% with the path of where the database will be created, e.g., "C:\Program Files\Microsoft SQL Server\MSSQL.1\MSSQL\DATA." Please also notice that the path does not contain a final backslash.
5 Replace all instances of %logfilepath% with the path of where the log file will be created, e.g., "C:\Program Files\Microsoft SQL Server\MSSQL.1\MSSQL\DATA." Please also notice that the path does not contain the final backslash.
6 Replace all instances of %initialdatabasesize% with the initial size for the database, e.g., "2048KB."
7 Replace all instances of %initiallogfilesize% with the initial size for the log file, e.g., ?1024KB.?
In summary, these are the items that should have been replaced in the script, %databaseName%, %databasepath% (without the final backslash), %logfilepath% (without the final backslash), %initialdatabasesize%, %initiallogfilesize%.
8 Save the file.
9 Execute the script.
Create the database objects using the provided schema (i.e., script)
The referenced script creates the necessary database objects or tables.
1 In the same directory as the database creation script, for example, C:\Program Files\AspenTech\Aspen PIMS\Enterprise Configuration\Database, use SQL Server Management Studio or a text editor to open PimsTableCreationScript.sql.
2 Replace all instances of %databaseName% with the name of the database that is being created, e.g., ResultsPIMS.
3 Save the file.
4 Execute the script.
The database scripts should run without error. If there are any errors, trouble shoot. Make sure you have sufficient privileges to execute the scripts and create objects.
Create Database User Accounts
Create database login accounts and users and assign them to their appropriate database roles based on your local IT security policy.
The minimum database privileges for Aspen PIMS users to read and write is db_datawriter and db_datareader on the database.
Test SQL Server Setup and User Access
Use the following procedure to test SQL Server login credentials and to test the connection to your results database.
1 Test the login by logging onto SQL Server with SQL Server authentication mode using SQL Server Management Studio from the same machine on which SQL Server is installed.
2 Test the connection has been set up correctly by launching PIMS.
3 Go to the General Model Settings dialog box.
4 Click the Output Database tab.
5 Select SQL Server from the Database Type list dropdown.
6 Click the ??? associated with the Connection String field to open the Data Link Properties dialog box.
7 Complete the SQL Server data connection information using the SQL database name you created and using "Microsoft OLE DB Provider for SQL Server" as the provider.
8 Click Test Connection to verify connection to the database is successful. After the successful message is received, you are ready to run PIMS with output going to your new SQL database.
Keywords: Database, SQL Server, PIMS SQL Database, Model Setting, Output Database
References: None |
Problem Statement: Step by step ABE V10 installation instructions | Solution: 1. To install the Aspen Engineering Suite (AES) V10 software, including ABE V10, open the AES V10 media folder and lauch the 'Setup.exe' file.
2. The AspenOne Installer window will open. Select the 'Install aspenOne Products' option in the Welcome page.
3. Read an accept the 'aspenONE Software License Terms' and click on next.
4. On the 'Product Selection' page, select the Aspen Basic Engineering 'End-User Tools' as well as the 'Configuration Tools' and choose the correct 'ABE Server' option:
ABE Solo Server – A lightweight install on a single PC, in which any projects created in ABE may only be accessed from that PC.
ABE Server Standalone – A full install, typically on a server machine, enabling multi-user collaboration.
5. On the 'Licensing & Security' page, enter the license server name and click on the 'Add Server' button or browse to the location of a License file. You can leave this blank and configure the license later from the 'SLM Config Wizard'.
6. If you chose to install the Aspen Basic Engineering 'Standalone Server', use this screen to configure an account used to run the services associated with the server. The installed user will appear, enter the associated password for the installing users account.
7.Whether you chose to install the 'Standalone Server' or 'Solo Server', use this screen to configure an account used to run the 'Broker' services. We recommend to use the default Name (AZ191_UserName) and use 'Aspen100' for the password. Finally, click on the 'Next' button.
8. The installation will start. The estimated time for the full installation is of about 30 minutes.
9. Once the installation processs is completed, click on the 'Finish' button. The system reboot is requered, please click on the 'Reboot Now' button on the next window.
10. After restarting your PC, please download the Cumulative Patch 1 (CP1) for ABE V10 from the Aspentech Support website and follow the instructions to install it:
Cumulative Patch CP1 for Aspen Basic Engineering V10.0
11. After installing CP1, please restart your computer. Now ABE V10 is completely installed. Please try to create a new workspace from ABE - Administration tool and connect to the workspace using the end-user tools.
Keywords: ABE V10 Installation, Solo Server, Standalone Server.
References: None |
Problem Statement: This article features an example of a particle size separation process using classifier, screen and cyclone models. The example incorporates solid substreams, PSD and various solids models including a cyclone in design mode. The attached example (PSD classification.bkp) was created in Aspen Plus V10. | Solution: The flowsheet is based on a basic harvested grain cleaning process with a feed mixture of rice grain, rice husk and soil stones. The husk, stones and fines are removed from the rice which can then be polished in a mill.
Components
Rice and Husk are defined as non-database solids and stone is assumed to be silicon dioxide
Air is added to the component list as it is used in the fines separating unit and the cyclone
Simulation environment setup
In Setup | Global Settings, the stream class is chosen to be MIXCIPSD (includes conventional solids with particle size distribution)
In Setup | Stream Class, substreams for rice, husk and stones are declared so that we can see the individual PSD for each substream in the results
In Setup | Solids | PSD a mesh is declared to be used in all solids streams througout the flowsheet.
Feed stream setup
In FEED | CI Solid the flowrate of each substream is declared, together with the size distribution across the PSD mesh.
Flowsheet
The first step is to remove fines from the feed mixture. In the model, this is done by a two-stage classifier (CLASS1 and CLASS2). CLASS1 separates based on particle density and CLASS2 separates based on particle settling velocity. An air stream feeds CLASS1 and passes to CLASS2 and out with the fines.
The fines stream (FINESOUT) is then fed to a cyclone model in design mode. With a given separation efficiency, the cyclone dimensions are calculated.
The heavy stream from the air driven classifier are fed to a three stage physical screening process based on screen aperture size.
The middle cut from the screens is fed to a final classifier to separate the stones from the rice, and this separation is based on particle density.
Results
Some of the husk is removed from the rice in the fines removal.
Most of the husk in removed in the first screening stage to bag 1.
Some of the stones pass over the screens to bag 2.
Most of the stones are the same size as the rice and are removed in the destoner before the majority of the rice is sent to mill (out of scope).
The vent gas is cleaned of all fines in a cyclone. The dimensions of the cyclone are calculated for a Barth 1 - Rectangular inlet type.
Keywords: None
References: None |
Problem Statement: Sometimes you are going to notice that your model is optimized to optimal when using a set of unit of measure but infeasible using another. For example, if you start with MBbl and MLbs and change to Bbl and MT, your model will be infeasible if you do not change any other components. Does this mean that changing the text of UOM is changing the optimization path, and also are we using text to scale the objective function? | Solution: Yes, volumetric UOM in the setting dialog is used to set the scaling during optimization.
Scaling works fine in both the cases (Volume is either MBbl or Bbl) as the objective function is solved in both the cases.
However, few properties are out of spec (when UOM is MBbl), mainly due to the conflict between blend quantity ( for example 75000 MBbl), which means, system is trying to blend 7.5 million barrels per day, which is again in conflict with the available inventories (in 1000s, not in millions), causing spec calculation goes wrong.
There are two options to resolve this case:
(1) Keep the existing UOM (MBbl), but change the overall quantity in 1000s (i.e. 75.0 instead of 75000.0)
or
(2) Keep the quantity to 75000, but change the UOM to Bbl
If you open each event you will find out you have the quantity (number) entered with the UOM you inputted in the model settings.
Keywords: None
References: None |
Problem Statement: This knowledge base article describes how to resolve the following error message. This error occurs when I try to launch my Aspen HYSYS application.
Required Component Missing. Please see the ReadMe file for further information.
The application is unable to continue due to error 80040154. | Solution: This knowledge base article describes why error 80040154 occurs and different ways to resolve the error.
The error message may occur because the STRGXI2.dll is missing, not registered properly, or it is registered under a different directory. The proper location for STRGXI2.dll file is at
"C:\Program Files (x86)\Common Files\AspenTech Shared"
Therefore, Aspen HYSYS may not launch properly because SLM Tools is not installed or registered properly.
Other symptoms of this root cause besides error 80040154 may be:
SLM Tools 7.3 is installed on Windows 7. User is unable to see SLM Tools 7.3 listed on the Add/Remove programs or from the program file shortcuts.
AspenTech Uninstall Utility shows SLM Tools 7.3, but you just installed SLM Tools 8.4 or above.
The STRGXI2.dll version is for SLM Tools 8.4 or above, but SLM Tools 7.3 is listed in the AspenTech Uninstall Utility. This indicates that there is an install version conflict issue.
SLM Configuration Wizard is displaying blank system information when clicking on the Config button and no information will populate the dialog box.
There are several methods to resolve the error 80040154 depending on the scenario:
Method 1: Locating the STRGXI2.dll
In this case the STRGXI2.dll file does not exist in the Shared folder. Follow these steps to fix it:
On Windows, click Start > Search and then search for STRGXI2.dll without using quotes to search.
If a copy of the STRGXI2.dll is found on your computer, copy it to the "C:\Program Files (x86)\Common Files\AspenTech Shared" folder?
Then register STRGXI2.dll by using the DOS Window and changing the prompt to the location of the STRGXI2.dll file. Finally, type regserv32 STRGXI2.dll
Another way is to go to Start > Run and copy and paste this command into the run window.
regsvr32 "C:\Program Files (x86)\Common Files\AspenTech Shared\STRGXI2.dll"
You should get a confirmation that the dll was registered successfully.
Open HYSYS
If you don't locate STRGXI2.dll file on your computer:
Install SLM TOOLS from Installation media or go to www.support.aspentech.com and click on Knowledge base > login > enter the 6666005150, open the patch Software License Manager (SLM) V10 (Build 414) and download the aspenONE_V10_SLM_ClientTools.zip and Install it.
Method 2: Uninstall the existing SLM Tools from the computer then reinstall SLM Tools.
The proper way to uninstall is to use the Add/Remove programs or the AspenTech Uninstall Utility from Program Files\AspenTech. Manually delete the STRGXI2.dll from the shared folder and then reboot the computer. Finally, install SLM Tools and configure the SLM Configuration Wizard to point locally or to the network license.
Keywords: Required Component Missing, Error 80040154, 80040154
References: None |
Problem Statement: It is common to see after running an MPIMS/XPIMS model, the top part of full solution report has different economic values from economic summary section of the report. How to resolve this? | Solution: This is a known issue.
MPIMS local model buy and sell tables need to match global supply and demand tables. A mismatch would cause the global economic summary report objective function different than the objective function reported at the top of the full solution report.
Resolution Steps: Add the mismatched stream tags in local buy and sell models from global supply and demand tables in the emails. After the mismatch is solved, the two objective function numbers would be the same in the global full solution report.
Keywords: None
References: None |
Problem Statement: There might be active penalties set up in your PIMS model. For example, if you have PENALTY column filled out in CAPS and SCALE, or infeasibility breaker price/cost in BUY or SELL. You know that the penalty factor is activated. However, in the full solution report, you don’t see the “Model has active penalties” at the very top although it exists in the matrix. This happens in PIMS V10. | Solution: This is a defect that is fixed in V10 CP3 and beyond. The defect number is 382050 in the release notes. Please have it installed and you will see the penalty section reported.
If you are using V8.8 and already on the latest patch (CP5), you don’t have to worry about this issue. This is not seen in V8.8 at all.
Keywords: None
References: None |
Problem Statement: How do you enter an equilibrium constant (Keq) expression for a reaction? There does not seem to be a way to do this for the equilibrium reactors REquil or RGibbs. | Solution: Equilibrium reactions that have the Keq computed from a built-in expression are available using GENERAL Reactions in a RCSTR, RPLUG, RBATCH, or BatchOP. The Reaction Class needs to be specified as "Equilibrium" and then the reaction parameters can be specified as "Compute Keq from built-in expression" on the the Equilibrium tab. Additional Equilibrium, Kinetic, LHHW, and Custom reactions can also be specified in the same Reaction paragraph if desired.
Equilibrium reactions in Powerlaw reaction sets are supported in RCSTR, RPlug, RBatch, and BatchOp. However, in RPlug, RBatch, and BatchOp, the reactors solve an integration problem in which adding equilibrium reactions makes solving the integration more difficult. Alternative ways of specifying equilibrium reactions without making the system harder to solve include specifying them as reversible kinetic reactions using General reaction sets and specifying them in Chemistry, but note the limitations on using reactors with electrolyte systems.
Equilibrium constant expressions can also be specified for distillation columns using REAC-DIST reactions for reactions of type EQUIL.
Keywords: None
References: : VSTS 231505
KeyWords:
REquil
RGibbs
RCSTR
Chemistry
Keq |
Problem Statement: In PIMS V10, you might experience the following:
When generating the case comparison file, only the runs generated in the current session of PIMS are available to for case comparison generation. | Solution: s are retained between sessions if I save and close the model and re-open.
Solutions are not available for case comparison generation if I save the model as a new file name, although the files are kept in the directory.
Solutions are not available for case comparison generation if I paste in solution spreadsheets from other colleagues runs.
Furthermore, if you are a user of PIMS V8.8 and previous versions, you will see that the case comparison generation behaves differently in earlier versions. Why is that?
Solution
The reason this is happening is because in V10 Case Comparison works different than V8.8, it pulls directly from the results database that you saved your case runs. For example, if you select your database maintenance option to “keep existing”, run a set of cases in the model, and run another set in the next session, then both sets of cases should be in the case comparison dialog box.
Unlike V8.8, copying the solution files from another machine don’t work because it does not come from your results database. It used to come from solution files in V8.8, but this was changed in V10.
Keywords: None
References: None |
Problem Statement: How do you enter an equilibrium constant (Keq) expression for a reaction? There does not seem to be a way to do this for the equilibrium reactors REquil or RGibbs. | Solution: Equilibrium reactions that have the Keq computed from a built-in expression are available using GENERAL Reactions in a RCSTR, RPLUG, RBATCH, or BatchOP. The Reaction Class needs to be specified as "Equilibrium" and then the reaction parameters can be specified as "Compute Keq from built-in expression" on the the Equilibrium tab. Additional Equilibrium, Kinetic, LHHW, and Custom reactions can also be specified in the same Reaction paragraph if desired.
Equilibrium reactions in Powerlaw reaction sets are supported in RCSTR, RPlug, RBatch, and BatchOp. However, in RPlug, RBatch, and BatchOp, the reactors solve an integration problem in which adding equilibrium reactions makes solving the integration more difficult. Alternative ways of specifying equilibrium reactions without making the system harder to solve include specifying them as reversible kinetic reactions using General reaction sets and specifying them in Chemistry, but note the limitations on using reactors with electrolyte systems.
Equilibrium constant expressions can also be specified for distillation columns using REAC-DIST reactions for reactions of type EQUIL.
Keywords: None
References: : VSTS 231505
KeyWords:
REquil
RGibbs
RCSTR
Chemistry
Keq |
Problem Statement: How to setup a centralized SQL Server to host Aspen Properties Database | Solution: This solution article will help you to host the Aspen Properties Databases on a Central SQL Server Machine.
Please follow the below steps to host the Aspen Properties Databases on a SQL Server:
· Insert aspenONE Vx.x Engineering Installation Media.
· Go to .\aspenoneengdvd\core\APEDDB\ folder and run APEDD.exe file.
· All 4 databases APEOSVxx, APVxx, FACTVxx and NISTVxx will get extracted in the same location.
· Restore all the databases on a compatible SQL Server. See our platform support page for supported server versions. https://home.aspentech.com/platform-support
· Create the following user name with password on the SQL Server:
o Login Name: apeduser2
o Password: Aproperty88#
· Give apeduser2 login name full access to APEOSVxx, APVxx, FACTVxx and NISTVxx Databases.
· Configure SQL Server to allow remote Database connection, Please follow KB Solution: 125084
The centralized SQL Server is ready to be used for Remote Aspen Properties Database.
Keywords: Centralize
Central
SQL Server
APED
Aspen Properties Database
Share
Remote Connection
Citrix
Database
References: None |
Problem Statement: Upgrading a Batch Extractor db from V8.7 to V10 fails with error “An internal error occurred within Wizard.dll. Error: Error HRESULT E_FAIL has been returned from a call to a COM component.” | Solution: This error is the result of a missing line in the BatchExtractor.xml file
Missing line from BatchExtractor.XML
<Upgrade from="12.7.0" to="12.8.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1270toV1280_schema_mssql.ddl</Step></Upgrade>
Note: Adding the above line into the ..\AspenTech\DatabaseWizard\Configuration\BatchExtractor.XML file allows the upgrade to complete without error. See example below.
Attached is FixedBatchExtractor.xml that contains the missing line.
Open Windows Explorer and navigate to ..\AspenTech\DatabaseWizard\Configuration\
Rename BatchExtractor.xml to BatchExtractorOld.xml
Download FixedBatchExtractor.xml from this article and copy to ..\AspenTech\DatabaseWizard\Configuration\
Rename FixedBatchExtractor.xml to BatchExtractor.xml
Execute the Aspen Database Wizard to upgrade the db.
BatchExtractor.XML Example
<Upgrades>
<Upgrade from="1.4.10" to="1.4.11"><Step name="schema" type="2">BatchExtractor\EXTRACTORV01410toV01411_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="1.4.11" to="1.4.12"><Step name="schema" type="2">BatchExtractor\EXTRACTORV01411toV01412_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="1.4.12" to="1.4.13"><Step name="schema" type="2">BatchExtractor\EXTRACTORV01412toV01413_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="1.4.13" to="1.4.14"><Step name="schema" type="2">BatchExtractor\EXTRACTORV01413toV01414_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="1.4.14" to="1.4.15"><Step name="schema" type="2">BatchExtractor\EXTRACTORV01414toV01415_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="1.4.15" to="11.2.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV01415toV1120_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="11.2.0" to="11.2.1"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1120toV1121_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="11.2.1" to="11.3.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1121toV1130_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="11.3.0" to="12.0.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1130toV1200_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="12.0.0" to="12.4.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1200toV1240_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="12.4.0" to="12.5.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1240toV1250_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="12.5.0" to="12.7.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1250toV1270_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="12.7.0" to="12.8.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1270toV1280_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="12.8.0" to="14.0.0"><Step name="schema" type="2">BatchExtractor\EXTRACTORV1280toV1400_schema_mssql.ddl</Step></Upgrade>
<Upgrade from="14.0.0" to="15.0.0">
<Step name="schema" type="2">BatchExtractor\EXTRACTORV1400toV1500_schema_mssql.ddl</Step>
<Step name="defdata" type="7">BatchExtractor\EXTRACTORVCUR_defdata_mssql.ddl</Step>
<Step name="compile" type="8">BatchExtractor\EXTRACTORVCUR_compile_mssql.ddl</Step>
</Upgrade>
</Upgrades>
</Sqlserver>
Target version for fix:
V11
Keywords: None
References: None |
Problem Statement: How do you simulate a polyolefin purge bin model? | Solution: Attached is an example of a polyolefin purge bin represented using a User Aspen Custom Modeler block in Aspen Plus. This model can be used in V10.
Aspen Custom Modeler (ACM) is a general purpose flowsheet simulation tool which lets you develop your own steady-state and dynamic models. ACM is an equation-based simulator - even very complex models can solve in minutes or seconds. Models developed in ACM can be packaged to be used in Aspen Plus and Aspen Plus Dynamics.
Keywords: None
References: None |
Problem Statement: I am getting an error saying 'Model Open Cancelled' when I try to open a file in V10. What is the reason and is there a fix? | Solution: This issue is specific to V10 CP4 and earlier versions. The issue is likely occurring when you open the file on a network drive. AFR inadvertently puts the VstaTemplate.dll and VstaTemplate.pdb files into the model file instead of the installation drive.
Following is a fix for the issue:
Make a copy of the model file
Change the file extension to .zip and extract the zipped file
Open the unzipped folder and delete the following 2 files
VstaTemplate.dll
VstaTemplate.pdb
Rezip the folder
Change the extension back to .afr
Open the model file
Key words
Model open cancelled, network drive
Keywords: None
References: None |
Problem Statement: When you experience any issues with Baseload service not starting up | Solution: Baseload service is not starting up
1 - Click on Baseload Tokens icon - in AspenONE SLM License Manager
2 - On the left side of screen expand the "+" symbol for corresponding tokens
If you see the value of "0" - zero - Then open the services panel by typing services.msc - in the start run
3 - Scroll down the list to the baseload service - right click and stop this service
4 - Scroll to the Sentinel RMS service - right click and stop this service
5 - Wait 15 seconds
6 - Now start up the services in order:
Scroll to the Sentinel RMS service - right click and start this service
Scroll to the baseload service - right click and start this service
7 - Try again with the Baseload tokens
Keywords: None
References: None |
Problem Statement: Why does error “An error occured while attempting to save the workbook” shows up? | Solution: When a simulation is converged and the Results for Streams are available, you have the possibility of sending the results to Excel, both by Right-Clicking on the corner of the table or at the top of the window on the Report section:
When clicking the Send to Excel option, the next window comes up, allowing you to choose from new Excel spreadsheets or to already created ones, the later being available by clicking on the Add tables to existing workbook box:
Once the “Add tables to existing workbook box” is selected, you can Export those tables to Excel. After the Export is done, an option shows up for Open Excel file. This will open the file you exported the tables to.
After you repeat the same procedure and select the same workbook to add the tables to WITHOUT closing the previous file, the table is exported to the same workbook and a “new” Excel file with the same name is opened. This file isn’t seen, but the export was actually made on the same file and now there are two Excel files with the same name but one with only one table and the other with two tables on it:
Both Aspen Plus and Excel allow this to happen. When you try to do so one third time WITHOUT closing the Excel files, the following error message pops up:
To avoid this, simply close the Excel files already open and the error message should go away.
Keywords: Excel, Export, Plus, Results
References: None |
Problem Statement: In the online help there is a description about option codes which can be used. In theory one should find description for all possible models and option codes which can be used.
However, when checking ESRK2 model, one can see that 2 option codes can be selected, but documentation is not saying what is the meaning of that option codes.
What is the meaning of option codes which can be used for ESRK2 model? | Solution: ESRK2 model is described as: Redlich- Kwong equation of state model (PML implementation) with option to calculate enthalpy of water from NBS steam tables.
One can specify two option codes:
Option code Value Description
1 0 This option code determines if water is present in current mixture.
0: Do Not identify Water
1 (default) 1: Identify Water
2 0 (default) This option code Set the cubic root finder method
0: Analytical
1 1: Numerical
Fixed in version
This is a known issue with missing information in the Aspen Plus/ Aspen Properties documentation.
Reported in a defect: 265914. Documentation will be improved in version V11.
Keywords: Option Codes, ESRK2 model
References: None |
Problem Statement: In PSCP, the Cut Strategy options in the dropdown list are not mapped to the corresponding function correctly. | Solution: The Cut Strategy options in the dropdown list are actually mapped to these settings:
Default Strategy – No Cuts
No Cuts – Conservative Cut Strategy
Conservative Cut Strategy – Moderate Cut Strategy
Aggressive Cut Strategy – Aggressive Cut Strategy (correctly mapped)
The bug is targeted to be fixed in a future release.
Recommended settings:
Using a more aggressive cut strategy can result in a faster MIP solve or a better integer solution.
For the existing users, it is recommended to use ‘Aggressive Cut Strategy’ for MIP models.
Keywords: None
References: None |
Problem Statement: Aspen Plus documentation on rigorous plug flow reactor model (RPlug) states that residence time is calculated based on the phase fraction mole flowrate of each stream passing through the block (see equation below). Phases handled are vapor, liquid, liquid and solid (conventional and non-conventional).
Since RPlug also handles non-conventional solid streams, which have no molecular weight (and therefore no mole flowrate), how does RPlug calculate residence time when non-conventional solids are present? | Solution: Residence time is defined by the volume of the reactor divided by its total volume flow rate. What RPlug does is to sum all volumes from each stream phase present to estimate the total volume passing through the block. To calculate such volume, instead of using “molar fraction of phase j” as the equation above suggests, it uses “mass fraction of phase j”.
Since non-conventional solids are described in terms of mass density and mass flow, a total volume can then be estimated and accounted for in RPlug residence time calculation.
Note: for all stream types other than non-conventional solid, the calculation of residence time in terms of mole phase fraction does not pose any issues, since the units will still be consistent with residence time definition.
Fixed in Version
VSTS 294608: Aspen Plus Help Guide Documentation has been adjusted to account for non-conventional solids when explaining Residence Time Calculation. Fix should be available in next release.
Keywords: RPlug; Non-Conventional Solids; Residence Time; Mass basis; Mole basis;
References: None |
Problem Statement: Error message when adding variables to Aspen Online project and variables added don’t appear in grid. | Solution: When adding a variable to the project, Aspen Online gives an error message with regards to the unit of measure:
After selecting done, the Variable-Tag grid is empty and none of the added variables show up. This is a known issue and has to do with the variable physical type not getting correctly built in the model project database file, variables table.
To resolve this issue, exit Aspen Online and install the latest patch. Once the latest patch is installed, run the off-line to online utility to enable Aspen Online (re)build the variables table in the appropriate database file.
Then when you add variables to the workspace, they will add correctly.
Workaround
There is no workaround to this problem except to either ignore it or install the patch. It is highly recommended to install the patch.
Keywords: Physical type not recognized, variable-tag mapping, adding variables, variable grid error, variable grid blank
References: None |
Problem Statement: During a very large regression run, Aspen Plus/Properties may take a considerable amount of time to perform all calculations before finally converging. For those complex cases, an error message may appear stating that the engine has crashed.
This is a time-out issue, meaning that the engine regression is taking too long so the GUI quits and pops a misleading message. | Solution: A quick workaround is to edit your local mm.ini file located at Program Files (x86)\AspenTech\Aspen Plus V10.0\GUI\Xeq and set the first inter_timeout to blank. This will make the timeout period infinite and will allow time for the regression to fully run regardless of how long it might take.
Note: this procedure should only be performed as a workaround in case you are facing this issue, it should not be used as a preventive action.
Fixed in Version
V10 with Aspen Plus V10 CP4 Emergency Patch 9 (EP09)
Keywords: Regression; Engine; Crash; Aspen Properties; Properties Environment; Aspen Plus
References: None |
Problem Statement: There is a conflict in the Help Content for the Control Valve Option- Handling Densities for Multi-Phase Systems Rigorously. This Option can be found from within the Valve Property view, Rating Tab, Options Page
(see snapshot below):
The Help displays two conflicting explanations for the same option as follows:
Handle multi-phase flows rigorously option
Handle Densities for multi-phase flows rigorously option | Solution: This conflicting explanation is a mistake and a Known Issue. The correct explanation for this option should read:
Options Page
The Options page lets you select the method to handle multi-phase streams.
Handle Densities for Multi-Phase Systems Rigorously
For the Handle densities for multi-phase systems rigorously check box:
The Valve uses the phase densities to calculate the flow rate for the vapor and liquid. If cleared, the overall density is used.
Note: The Handle densities for multi-phase systems rigorously check box can only be selected if you selected one of the following Sizing Methods on the Rating tab | Sizing page: ANSI/ISA method or Manufacturer specific methods.
Handle Multi-Phase Flows Rigorously
For the Handle multi-phase flows rigorously check box:
The Valve uses the rigorous calculation method and obtain better accuracy of the flow rates and pressure drop for both vapor and liquid flow.
If selected, the liquid flow contribution to the total flow rate depends on a cutting point (0.1): if the outlet vapor molar fraction is greater than this cutting point (i.e. V_frac > 0.1), then the liquid flow contribution to the total flow is set to zero, that is, total flow rate = vapor flow rate.
Otherwise:
Where:
Vfeff = Effective mass vapor fraction,
Fliq = Liquid flowrate based on the ISA liquid equation
Fvap = Vapor capacity based on the vapor model
If cleared, the liquid flow rate and vapor flow rate are all counted with a weighted fraction in the total flow rate, and the total flow rate is adjusted with the correction factor Fm.
Note: The Handle multi-phase flows rigorously check box can only be selected if you selected one of the following Sizing Methods on the Rating tab | Sizing page: ANSI/ISA method or Manufacturer specific methods.
When the Handle multi-phase flows rigorously check box is selected, the following additional equations are used:
Fm = f (Vgas) is a function of Vgas, which is based on the following data:
Vgas [15]= {0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, .85, 0.9, .925, 0.95, 0.99,1}
Fm [15] = {0, 0.092, 0.184, 0.276, 0.371, 0.469, 0.576, 0.711, 0.876, 0.965, 1.0, 0.980, 0.91, 0.50, 0}
where:
Fl = Mass flowrate of liquid as if it is the only phase present. Fl is not the actual mass flowrate of liquid. It is obtained by setting the vapor fraction to 0 in, for example, the flow equation (5) (for ANSI/ISA valves) or (47) (for other valves).
Fv = Mass flowrate of vapor as if it is the only phase present. Fv is not the actual mass flowrate of vapor. It is obtained from the valve-dependent vapor flow rate equations and setting the vapor to 1, where applicable.
Cv = Standard liquid sizing coefficient
Cg = Cg for gas phase
C1 = Cg/Cv ratio for valve
Vmf = Vapor mass fraction
Vf = Molar vapor fraction
Mwv = Vapor molecular weight
Mwl = Liquid molecular weight
Vgas = Gas volume fraction
Fm = Cv correction factor, a function of the gas volume fraction
Dvap = Mass density of vapor phase
Dliq = Mass density of liquid phase
Ftotal = Overall mass flowrate
Keywords: Hysys Control Valve Option, Hysys Handle Multi-Phase flow rigorously, VSTS 413987, Hysys V12 Help, defect in Hysys Help
References: Fisher Controls International LLC. ANSI/ISA/IEC Valve Sizing. Catalog 12. Emerson Process Management, March 2012. Web.
This has been fixed in V12 to reflect the correct explanation in the documentation.
Fixed In V12 |
Problem Statement: Hysys freezes at the stage of loading Properties Databases | Solution: There are several reasons why the program may not run and freeze at the stage of loading Properties Databases:
1. If the computer has multiple versions installed, then the problem may be related to the conflict of installed versions of the property databases. To solve this, select the version that will be loaded by default: go to the list of the program / Set Version - Aspen Properties V10, make sure that version 36 is set as default.
2. User may not have permission to work with files in the Program Files / AspenTech and Documents / AspenTech folder. Allow access for users to change all folders and files in this directory.
3. Temporary Preference file HYSYS V10 (HYSYS V10 Aspen 0. PRF) may be damaged. This file is located in disk C: / Users / username / AppData / Roaming / AspenTech.
Move it temporarily to another location, and delete it in C C: / Users / username / AppData / Roaming / AspenTech. When you run HYSYS again, it should be created again.
Key words
Hysys freezes, crashed, stop running, loading Properties Databases
Keywords: None
References: None |
Problem Statement: VSTS 256618: When running Case Studies, specified units for a variable in the Vary tab are ignored. | Solution: When running Case Studies on Aspen Plus, the units specified for a Vary variable property aren't retrieved correctly.
For example, consider a Case Study that has Pressure and mmHg selected as the modified variable unit on the Vary tab:
As seen on the previous screenshot, Pressure is specified as Vary 2 (there are 2 Cases created). In this example, one case has 1 mmHg as Pressure and the other has 50 mmHg:
When the Case Study is run, the user can notice in the Case Study Results that the Pressure (Vary 2 in this case) has different units than the ones specified on the Sensitivity Analysis | Vary input.
The Case Study keeps using the units specified for the whole simulation by the Unit Set (PSIA in the selected ENG Unit Set):
Workaround
Current workaround is to convert units manually and specify them in the units defined by the simulation Unit Set.
Fixed in Version
VSTS 256618: Target for a future release.
Keywords: Case Study, Units, Unit Set, Plus
References: None |
Problem Statement: For Polymer example, case file and its manual is different. This KB is to notice this defect.
The exponent for gel effect's initial estimate is different between manual and case file (psea.bkp) .
Case file has 0.1 while the manual describe 2.0. Moreover, regressed result is slightly different.
> Case File Input
>Manual | Solution: Case file is incorrect.
2.0 is right approach for this simulation. The reaction rates slow down at high conversions because the polymer molecular mobility decreases (so the reaction becomes diffusion limited). The diffusion rates of small molecules in the polymer phase remains relatively high. Thus the gel effect term has a compound (f^2) effect of reactions involving two polymers where the mobility of each of the polymers is reduced by solution viscosity.
Fixed in Version
Will be fixed in V11
Keywords: Aspen Plus, Aspen Polymers, Get effect
References: None |
Problem Statement: Live agents may fail to process data from certain historians. The following error message will be seen in the Agent Service log file:
Could not load file or assembly 'Mtell.Sensor.Common, Version=36.0.1.17186, Culture=neutral, PublicKeyToken=e217e27c1617e401' or one of its dependencies. The system cannot find the file specified.
For a standard install, the Agent Service log file is typically found in the folder “C:\ProgramData\AspenTech\Aspen Mtell\Agent Service”. | Solution: 1. Stop the Agent Service on the machine on which it is installed. You can do this either using the Task Manager or the Service Console.
2. Copy Mtell.Sensor.Common.dll from the System Manager folder to the Agent Service folder. The System Manager folder path is typically “C:\Program Files\AspenTech\Aspen Mtell\Suite\Tools\System Manager” for a standard install. The Agent Service folder path is typically “C:\Program Files\AspenTech\Aspen Mtell\Mtell Agent Service” for a standard install.
3. Re-start the Agent Service. You can do this either using the Task Manager or the Service Console.
Keywords: Agent Service
Live Agents
References: None |
Problem Statement: When a controller that belongs to an Aspen Composite changes structure (i.e, new variables, change variables order, new sub-controllers), the Composite will produce this error and shed all controllers.
"OTF workspace required > allocated space"
In addition, the Production Control Web Server (PCWS) may not showing all the sub-controllers and/or variables of the controller. The Composite engine doesn't recognize the changes to its member controllers thus it is reading wrong data from the newly changed controller. | Solution:
Redeploy the Composite application will force it to reload all the information needed to run the application properly.
Keywords: Composite, OTF
References: None |
Problem Statement: How to fix a ENG installation that hangs in the Hysys installation at 9% | Solution: If your Engineering suite installation is stuck at 9% of the Hysys installation for more than 25 minutes and the last action recorded is:
ATRegisterFile.BA3F7B28_1307_4630_902E_55D684AF3A97. Return value 1.
You can verify this by checking the current process running (1) or by checking the last recorded action in the log file (2).
If this is your current situation, then open Task Manager and go to the running programs tab (details tab in windows 10) and search for the EOServer.exe or AspenHysys.exe and end that task.
This should allow the installation con complete on its own.
Keywords: None
References: None |
Problem Statement: How to converge Short Cut Column when making a large change to Feed? | Solution: Short Cut column needs hard reset to solve in Yield Fraction mode when making a large change to Feed.
The column uses an Equation Oriented solution to solve. It uses the last successful solution as a starting point for the next solve. Normally this makes it much faster to solve. However, if the actual solution is far from the starting point, the column may fail to converge. Changing the yield basis from Mass to Volume or vice versa forces the column to do a hard reset. It will re-initialize the column by solving the column in ECP mode. The ECP values are calculated by taking splitting the column into square cuts that match the user specified yield fractions. The TBP cut points of these square cuts is the starting ECP values.
To converge the column in the Design / Specs page change the yield basis from Mass to Volume or vice versa.
This issue is solved in V11.
Key words
Short-Cut Column, Column convergence
Keywords: None
References: None |
Problem Statement: Under certain circumstances, Aspen Utilities Planner (AUP) is unable to run a Multi-Period Optimization. The user will get the message “ERROR: 6 - Arithmetic operation resulted in an overflow”
Optimization Complete - solution found
Normal Completion
Feasible Optimisation | Solution: found
Processing Results...
Update Optimization Results...
Results generated
Calculating marginal costs
Xpress-MP licensed by Dash to AspenTech for Aspen Engineering Suite
Fixing globals...
ERROR: 6 - Arithmetic operation resulted in an overflow.
END
POSTPROCESS
ERROR: 6 -
END
Successful: Exiting Optimisation
Processing results...
Editing Simulation
Current snapshots have been saved to file snpA0001.snp
Workaround
The error can be fixed with the Aspen Utilities Planner (AUP) V8.8 Emergency Patch 3 (EP03).
This emergency patch requires Aspen Utilities Planner V8.8 (with Aspen Plus Family V8.8 cumulative patch CP2) installed in the computer.
Note: If you require an Emergency Patch for AUP V9 or V10, please contact [email protected]
Fixed in version
Targeted for a future release
Defect ID: VSTS 366661
Keywords: Multi-Period Optimization, Marginal Cost, Error 6, Arithmetic operation, Overflow.
References: None |
Problem Statement: Re-ordering of component list causes change in calculated duty for stream = VSTS 403057 | Solution: It is possible to select binary parameters for a pair of components from a databank that is NOT selected on the databanks sheet. These binary parameters are used and may result in significant differences in calculations on the flowsheet such as heat duties.
When components are re-ordered, the parameters which are retrieved from a non-default databank will be deleted. This will cause the simulation results to change.
Take for example the Water/Catechol pair. The NRTL-1 binary parameters for this pair are not retrieved automatically from the default databanks:
However, we can select the water/catechol pair and find that, in fact, the binary parameters for this pair can be found on one of the NIST Databanks. We can select NISTV100 NIST-HOC:
Now, if we go back to the component list on the Components | Specifications | Selection tab and click the Reorder button to re-order any component (water, for example), any binary parameters which were retrieved from a non-default databank will be erased:
This behavior is also seen if you select a pair from a databank that is not the default. This is a bit subtler:
Now re-order water and ethanol. The pairs go back to the default databank order.
Workaround
The workaround is to not re-order or to make sure all databanks are selected, and parameters are selected from the databanks in the default order. You can also copy and paste parameters back as USER values.
Fixed in Version
VSTS 403057: Target for a future release.
Keywords: Binary, parameters, databanks, pair
References: None |
Problem Statement: Aspen Plus V10 Exchange is still unable to communicate with the Exchange server after installing the Emergency Patch 12 (aspenONE Exchange will stop being able to communicate with the Exchange server after Feb 4th, 2019. Apply patches today to avoid disruption) | Solution: In this case, the users will need to manually extract and place the fix file (*.dll) on their machine:
Please follow the next steps:
1. Be sure Aspen Plus V10 is closed
2. Download the file AP_V10.0_EP12.exe from aspenONE Exchange V10.0 Emergency Patch
3. Unzip the *.exe file
3. Open the unzipped folder and go to ...\AspenTech\APrSystem V10.0\GUI\xeq
4. Copy the file 'AspenTech.Exchange.SecureServices.dll' and paste it in the route: C:\Program Files (x86)\AspenTech\Aspen Plus V10.0\GUI\Xeq
5. If needed, replace the file.
Finally, be sure the required Ports and URL's are open for connectivity (How do I resolve "Error 407: Proxy Authentication Required" when trying to connect to aspenONE Exchange?)
Keywords: Emergency Patch, Exchange download issue, Login error, Exchange issue.
References: None |
Problem Statement: We have scheduled biases defined in the model, but there is no daily results published in _BIASRESULTS table. How does it work? | Solution: For the versions before V11 there is a documentation error in Help file saying daily results will be stored in _BIASRESULTS.
APS has always written periodic data in _BIASRESULTS which is not a usual convention. In V10 we agreed to move periodic results to _ZBIASRESULTS to match with usual convention. But, there never was a logic put in place to write out daily results.
If user defines the PUB_BIAS_RESULTS keyword in CONFIG table with value Y then APS will write periodic results in _ZBIASRESULTS table otherwise, they won't be published at all.
To conclude, currently there is no logic in APS that collects daily bias results.
Keywords: Daily results
Biasing
Schedule bias
References: None |
Problem Statement: How do I resolve an Aspen Icarus Reporter - Database Connection Error? | Solution: This error may be resolved by un-installing and re-installing Microsoft Access Database 32-bit via the following link:
Microsoft Access Database Engine 2010 Redistributable
Note this issue is related to the old reporter (version 10 and older). It does not apply to the new reporter (V11 and newer).
Keywords: Access, Icarus, interactive, reporter, error
References: None |
Problem Statement: There is an enhancement implemented in APS V10 CP2 – VSTS 453264 “Enhancement - Publish required information for pipeline line fill”, which requires db upgrade.
If users plan to use this feature then they can upgrade their database following the usual upgrade process i.e. run DBUpdate using the new OrionDBGen.mdb that gets delivered with the new patch. But patch won't overwrite the existing OrionDBGen file which is unversioned but modified. That means you’ve updated the existing db file on machine (in gold or CP1) before applying patch. | Solution: That is a Known Issue, which will be resolved for future V10, V11 and V12 patches. Please use attached OrionDBGen.mdb in DBUpdate utility as a Baseline DB to upgrade your unversioned but modified file to use feature from VSTS 453264.
Keywords: CP2
OrionDBGen
Patch
DBUpdate
References: None |
Problem Statement: Quick Tips for using the Object-Oriented Modeling Framework (OOMF) language in the Aspen Plus Engine | Solution: Aspen Plus Equation Oriented (EO) commands are written in the OOMF language therefore these commands can be accessed at the Aspen Plus engine level and by-passing the Aspen Plus user interface.
This enables the user carry out quick operations, check EO modeling settings or print OOMF commands with or without referencing a specific EO process model
To access the OOMF layer, launch the Aspen Plus engine using the Customize-Aspen Plus VX DOS interface and then type OOMF and enter:
Next, try using the following command to view available units of measure for molar flow:
Print uom for mole-flow
Now try converting a known molar flow, say 135 kgmole/hr, to std_m3/hr using the following command:
Echo 135 {kmol/sec} #{scmh}
Using the Help command will display other useful OOMF commands...
… And using ctrl+c clears the command and resets the command prompt window
Additional tips: resize the window permanently by going to the Customize Aspen Plus windows application via the OS menu, right-click and select properties:
The Layout tab in the Properties form will allow you to resize the window and increase the buffering capacity of the display:
Keywords: OOMF, OOMF engine, EO commands, Aspen Plus OOMF
References: None |
Problem Statement: How do I create a user-defined prop-set property to calculate enthalpy of steam? | Solution: To get the enthalpy in the steam tables reference state, you need to add 15970312.47 J/kmol to the enthalpy value to account for the enthalpy of water at its triple point.
The attached example shows a user-defined prop-set property to calculate enthalpy of steam. The prop-set displays steam enthalpy based on an alternate reference state.
The name of the new property is HSTEAMT. The name of the subroutine must be declared on the Properties.Advanced.User_Property form. The name that is assigned to the object used to store this property becomes the name for the user property. This name is assigned before the User-Property form is entered. The user property must also be specified in a property set to be included in a stream report.
To use this example without a compiler, simply put the .bkp, .dll and .opt files in a working directory together. To modify the subroutine, a compiler is needed.
Note: This subroutine was created using a 32-bit compiler, which means it can only be used in 32-bit versions of Aspen Plus (V10 or earlier). For newer versions (V11 or after), the subroutine would need to be re-compiled using a 64-bit compiler.
Subroutine Code:
SUBROUTINE USRHST(T, P, FV, FL, BETA, NCPACK, IDX, FLOW, Y, X,
+ X1, X2, Z, NBOPST, KDIAG,
+ KPDIAG, XPCTLV, IPHASE, NAME, PRPVAL,
+ SF, S, ISUBS, CAT, FLOWS,
+ NSUBS, IDXSUB, ITYPE)
C
Keywords: None
References: None |
Problem Statement: In Aspen Utilities Planner Excel user interface, ‘Do you want to save the profiles and tariff in the databases?’ message (figure 1) may pop up. How to get rid of this message?
Figure 1. Do you want to save the profiles and tariff in the databases?’ message | Solution: Please follow the steps showed in figure 2 (can be downloaded in Attachment section). In Excel Developer Ribbon, click Visual Basic button. Find the ThisWorkbook code in pop-up window. After commenting out the related VBA code by using single quotation marks and saving the changes, this message will not pop up again.
Figure 2. Solution: commenting out the related VBA code by using single quotation marks
Keywords: Aspen Utilities Planner, Excel, VBA, Database
References: None |
Problem Statement: When I open Aspen Flare System Analyzer, I get the following error.
Error Message: Unexpected exception: Resolution of the dependency failed, Type = “FnetApplication.PFDScreen”, name = “(none)”.
Source: Microsoft.Practices.Unity
Target: DoBuildUp
Why does this error occur? How do I fix this error? | Solution: With Aspen Flare System Analyzer V10, in certain customer machines there may be a crash when the customer opens the process flowsheet or open an old case.
To resolve the issue, suggest installing the following patches in orderly manner.
1. Cumulative Patch V10.0 CP1 for Aspen Flare System Analyzer
https://esupport.aspentech.com/apex/S_SoftwareDeliveryDetail?id=a0e0B00000DiA2pQAF
2.Emergency Patch1 for Aspen Flare System Analyzer V10.0 CP1
https://esupport.aspentech.com/apex/S_SoftwareDeliveryDetail?id=a0e0B00000FIazYQAT
3. Emergency Patch 2 for Aspen Flare System Analyzer V10.0 CP1
https://esupport.aspentech.com/apex/S_SoftwareDeliveryDetail?id=a0e0B00000FIsRKQA1
To verify that the CP1 installation was successful, verify that the following file has the specified version number:
Note: If you unable to find above patches on our support website or it's still not resolved after installing three patches, please do feel free to contact AspenTech support team.
Keywords: Error, Resolution of the dependency failed, Microsoft.Practices.Unity
References: None |
Problem Statement: Is it possible to modify the Binary Parameters (BIP) for an Aspen HYSYS Fluid Package using VBA? | Solution: When using VBA, the user will not find a variable to call and edit the Binary Parameters through Automation.
However, the user can still call and edit these variables using backdoor monikers into their VBA code.
For more information on how to evoke backdoor variables and obtain the backdoor monikers please review article 000031380
Keywords: Parameters, Package, Automation, OLE, Backdoor
References: None |
Problem Statement: How to report the radius of gyration (RGYR)? | Solution: The Radius of gyration or gyradius of a body about an axis of rotation is defined as the radial distance of a point from the axis of rotation at which, if whole mass of the body is assumed to be concentrated, its moment of inertia about the given axis would be the same as with its actual distribution of mass.
The RGYR can be retrieved from the pure components parameters, as shown in the following article (How to retrieve pure component parameters such as molecular weight, Tc, etc in Aspen Plus?)
Or can be estimated from the Parachor method with the property estimation (PCES) tool as it is used in the Hayden-O'Connell (HOC) model; therefore, the user needs to have one of the xxxx-HOC property methods specified to be able to see RGYR on the parameter form.
To estimate the RGYR, the user can follow these steps:
1- Ensure that one of the xxxx-HOC property methods is selected.
NRTL-HOC
UNIF-HOC
UNIQ-HOC
VANL-HOC
WILS-HOC
2- In the setup tab of estimation folder select "Estimate only the selected parameters" with the "Pure component scalar parameters" option and in the Pure component tab select "PARC" as parameter for All the components with the PARACHOR method.
3- Click in retrieve parameters and select the review ribbon of the Pure components folder.
Keywords: RGYR, Gyradius, Parachor, HOC,
References: None |
Problem Statement: As mentioned in the following article Aspen Plus COM Automation using Matlab the user can call an Aspen Plus model and change the inlet parameters by using our Matlab code.
However, if the user wants to access to a parameters that has a "space" in its path (i.e. Aspen.Application.Tree.FindNode("\Data\Blocks\B1\Input\MOLE_YIELD\O2 MIXED")) Matlab does not recognize the address and won't create the link. | Solution: To solve this issue the user can follow these steps:
Use a calculator block and link the desire variable in the embedded excel file.
Create a live link (with the special paste option of excel) between the cell of the embedded an a cell of a normal xls file.
In Matlab access to the specific cell of the xls file. So, the workflow of the code will be the following:
The data in Matlab will be transfer to the xls file (excel)
The xls file will send the data to the apmbd file
The calculator will read the value of the apmbd file
The calculator will input the value to the yield in the reactor
Keywords: Matlab, Space, Path, Node, Call, Access, Calculator, Excel
References: None |
Problem Statement: If the user has different reaction sets and wants to evaluate their performance in a kinetic reactor, then ASW can be use to switch among different reaction sets available already in the Aspen Plus model. | Solution: To choose a specific reaction set, please go through the following steps:
1- Open the organizer form and click on the Browser button.
2- Select the "RXN_ID" variable from the variable path as shown in the screenshot below.
Note: Instead of "RI" the user should point to the name tag of the kinetic reactor
3- Create a table in Aspen Simulation Workbook as mentioned in the following article "How do I create a table in Aspen Simulation Workbook?" and change the current RXN_ID tag to that of the kinetic reactor.
Note: The RXN_ID must be the same as that used in Aspen Plus.
Keywords: Reaction, Set, Change, RXN_ID, ASW
References: None |
Problem Statement: Can I force a specific order to solve objects in my simulation? | Solution: The default order of calculation in HYSYS is generally recommended for most simulations. However, in some specific cases, based on detailed knowledge of your model, and understanding of the degrees of freedom involved in the solution, you may wish to have the objects in your flowsheet solve in a specific order.
The Calculation Levels in HYSYS set control over the order in which the objects in a flowsheet are calculated. Most objects have a default Calculation Level of 500; Columns and Sub-Flowsheets (templates) have a default Calculation Level of 2500 while Recycles and Adjusts are set by default to 3500.
HYSYS will attempt to solve the operation(s) with the lowest calculation levels to calculate first, for example an operation with a calculation level of 200 will solve before an operation with a level of 500.
The solvers accommodate Sub-Flowsheets nested multiple layers deep, which allowed the Calculation Levels to be local. This means that if a given object is within a Sub-Flowsheet which has a specified Calculation Level, the Calculation Level of the Sub-Flowsheet governs when the object is solved. Calculation levels of the operations within the subflowsheet are relative to the other objects in that same flowsheet.
Keywords: Order, Calculation, Object, Operation, Stream, Level.
References: None |
Problem Statement: How to be able to view the correct Trend data for a stream in the APS Event Screen ?
The screen-shot below shows the issue: | Solution: A Trend screen graphically depicts the Minimum bounds, Maximum bounds and the Points at which bounds have not been maintained.
Moving the cursor over a trend line displays a tooltip that shows the time corresponding to the mouse position as well as the Y-values of all curves. Charts that are highlighted in red indicate that these trends have exceeded trend limits.
However sometimes on opening the model for the first time, some Trend values will not display correct trend or show the correct values of the variable being trended.
In the above screenshot, we are trying to trend the Alkylate stream, on opening the model and reloading it, we were unable to see the values of the trend despite there being events on the event screen.
To solve this issue, we just need to "Simulate All" Events one time. This will simulate all the events that are scheduled for the refinery operations
This option can be found from the Simulator Menu > Simulate all in the Event Screen as per the below screenshot,
once this is done, we can see that the Trend is shown correctly for the Alkylate stream as per below screenshot:
Hence as a good practice, after opening the model , one must reload the simulator and then simulate all the events one time to have the correct Trend results shown.
Keywords: None
References: None |
Problem Statement: How to access Aspen Process Manual(APM) from Aspen Plus | Solution: It is accessible from All Contents under Resource tab as below:
1. Go to the resource tab, and click All Content
2. Search keyword “Aspen Process Manual”, click on the green arrow button.
3. Help file will appear, and you can access the following.
When opening a manual on a specific topic (i.e. Bulk Solids Handling, Drying, etc), it will consume 6 tokens for each category/topic. If you open multiple manuals under the same category/topic, it will only consume the initial 6 tokens (i.e. Bulk Solids Handling: Prime moves for Bulk Solids and Mechanical Conveyors)
Key Words
Aspen Process Manual, AspenPlus
Keywords: None
References: None |
Problem Statement: Aspen Plus allows to model dynamically vessels undergoing pressure relief due to a fire or heat input specified by the user as mentioned in the following link How to perform a Fire scenario with the Pres Relief analysis. This feature can calculate the vessel wetted area based on the vessel geometry, the initial composition and the fire standard. | Solution: For example, the wetted area of the following vessel will be calculated by the cylinder area (Pi*Diameter*Length) and the heads areas.
Note: The tooltip for the head area is "Surface area of a single head". Since the vessel has 2 heads the head area is added twice, so the total area is 339.854 sqft (cylinder area) + 2*189 sqft (heads areas) = 717.854 sqft.
So, the wetted area is 0.75*717.854 = 538.39 sqft which is what the user sees under the Parameters Tab of the Dynamic Results
Keywords: Pres Relief, Wetted area, Vessel, Fire, NFPA-30, API-520, API-2000
References: None |
Problem Statement: Previous to V8.0, when the user creates a profile plot for a column (composition, temperature, etc.), the plot is created as an independent object. Because of this, if the user resets the simulation and/or changes something in the column and re-runs, the old plot stays the same and a new plot is generated from the new run. | Solution: This is different in newer versions since by default the plots are automatically updated every time the simulation is re-run. So, if the user resets or changes anything in the column and re-runs, the plots get updated. To compare plots after different runs, please follow these steps:
1. After the first simulation is run, create the plot.
2. On the Format tab uncheck the Auto update option.
3. Make the desired changes and re-run the simulation.
4. Generate the second plot and add or merge it to the first plot.
5. Compare the different plots
Notes:
The Auto update option must be unchecked before using Merge Plot.
The Merge Plot option is not available for the Customize Plot
Keywords: Profiles, Merge, Reset, Update, Column
References: None |
Problem Statement: How do I use Aspen Online Modeling (Plant Data) feature to import plant data to my Aspen Hysys flowsheet model? | Solution: This solution provides a training guide and a pre-built example to demonstrate how Plant Data can be used in an Aspen Hysys flowsheet model. The example imports plant data tags from excel and does not require the user to have a connection to a Process Data Server
Keywords: Plant Data, Aspen Hysys, validated model, Plant data to Aspen Online
References: None |
Problem Statement: How do the different License States impact the functionality of aspenONE Process Explorer V8 and higher? | Solution: This knowledge base article describes the various License States for aspenONE Process Explorer V8.x and the impact in functionality for each state.
License Granted: occurs when a license has been granted successfully. In License Granted state:
·aspenONE Process Explorer provides complete functionality.
License Denied: occurs when either the license server denies the request checkout or the license server cannot be contacted and the license grace period, if any, has elapsed. In License Denied state:
·aspenONE Process Explorer displays a message dialog indicating that it cannot acquire a license.
·Click X button to close the dialog box and exit aspenONE Process Explorer.
License Timeout Period: occurs if the license has been granted but aspenONE Process Explorer is no longer able to contact the license server. If contact with the license server is not restored, then the License Timeout period can last for a timeout period specified in the license key. The default timeout period is 5 minutes. In License Timeout Period state:
·aspenONE Process Explorer will continue to run for a timeout grace period (default is 5 mins).
·aspenONE Process Explorer displays a warning icon in the Ubar indicating that there is a problem. Click the warning icon for a brief description such as “License in timeout period (0 00:04:00).”
Extended Grace Period: occurs if the license has been granted at least once, the license server cannot be contacted and the license timeout, if any, has elapsed. In Extended Grace Period state:
·aspenONE Process Explorer will continue to run for an extended grace period (default is 10 days).
·aspenONE Process Explorer displays a warning icon in the Ubar indicating that there is a problem. Click the warning icon for a brief description such as “License in Grace period (0 01:00:00).”
Server License Denied: occurs when aspenONE Process Explorer attempts to connect to a server that is in License Denied mode.
·aspenONE Process Explorer will not connect to a server that is in License Denied mode.
·Tags that reference a server that is in License Denied mode will not display data values.
NOTE: The license behavior may differ with other MES products. Click on the product below to review its license behavior.
InfoPlus.21
Aspen Production Record Manager
CIM-IO
Process Explorer
Keywords: A1PE, MES, License Behavior, License State, Granted, Timeout, Grace Period, Denied
References: None |
Problem Statement: What is the meaning of SG_60/60api, SG_H2O60nbs and SG_H2O4? | Solution: In Aspen HYSYS, the user can select different units of Liquid density when a hypothetical is create:
The units SG_60/60api refers to the mass density of a liquid at 60 F over the mass density of water at 60 F (based on API definition).
The units SG_60nbs refers to the mass density of a liquid at 60 F over the mass density of water at 60 F (based on the NBS definition).
The units for SG_H2O4 refers to the mass density of a liquid at 4 C over the mass density of water at 4 C
Note: NBS is the "The National Bureau of Standards" also knows as National Institute of Standards and Technology (NIST)
So, the difference is methods or standards.
Keywords: Hypothetical, Liquid density, NBS, API, H2O4
References: None |
Problem Statement: What does the ‘Load Model Once’ option mean in the Aspen Online, Model Specification Options mean? | Solution: When the Aspen Online Engine cycles through the model run, the Windows Service loads the process simulator application before attempting to run the model.
When the option is set to Load Model Once, it means that for the entire time the engine is loaded, the simulator model will be loaded for the case run and will not be closed even after the run is completed.
When running Equation Oriented (EO) models using Aspen Online, they are always loaded once. Therefore, if the model is EO, this option is greyed out as in the snapshot below:
For other model types, the user can decide if the engine should close the model after running a complete case or keep the model open. In concept, if the Load Model Once option is checked, it means the simulator license will be checked out for the duration the Aspen Online Engine is loaded. If this option is not checked, once the Aspen Online Engine completes the case run, the simulator license will be released.
Keywords: Load Model Once, Aspen Online Engine
References: None |
Problem Statement: How do I use Aspen Online Modeling (Plant Data) feature to import plant data to my Aspen Plus flowsheet model? | Solution: The document provides an overview of the Plant Data feature and how it is used in Aspen Plus. The solution includes a workflow presentation and example files to act as a guidance on deploying and validating an Aspen Plus process model using Plant Data
Keywords: Plant Data, validated model, Aspen Plus, Plant data to Aspen Online
References: None |
Problem Statement: How to create a filter view for objects? | Solution: This example shows you how to write a KB script to create a filter view in ABE (Aspen Basic Engineering).
ABE allows users to filter objects through specific rules defined in a Filter Module. This Filter Module is a KB script written in VB code. The script will produce an additional filter view in ABE Explorer. Hence, applying Filter Modules is a simple solution to customize equipment view.
As with any other KB scripts, users need to compile the module in the Rules Editor environment. Please refer to the following article if you would like to understand further how to compile a KB script in Rules Editor:
https://esupport.aspentech.com/S_Article?id=000047646
Unlike Rules and Methods Modules that require execution every time users need to use the modules, Filter Modules immediately apply to the workspace after the script is enabled.
In this example, we will look at a simple scenario in which user needs to filter objects by their class:
Take an instance when the user wants to only view ReciprocatingPump objects, the script is simply:
'Code starts
[!Modulename=ShowOnlyReciprocatingPump!]
[? Class=,ReciprocatingPump,
DisplayName="All Reciprocating Pumps" ?]
AZFilter RecipPump()
End AZFilter
'Code ends
Alternatively, if the user needs to exclude ReciprocatingPump out of its parent class and view only other sub-classes, we can exploit a conditional statement as below. The end results will only show objects belonging to DiaphragmPump, MeteringPump, and RotaryPump.
'Code starts
[!Modulename=FiltersOutReciprocatingPump!]
'Access the parent class "PositiveDisplacementPumps"
[? Class=PositiveDisplacementPump,
DisplayName="All but Reciprocating Pumps" ?]
AZFilter RecipPump2()
RecipPump2 = true
if self.isderivedfrom("ReciprocatingPump") then
RecipPump2 = false
end if
End AZFilter
'Code ends
Users are recommended to use the Class Library Editor to get a high level view of each class and its branches. More filter script examples are provided the under "KBs" folder in the WorkspaceLibraries
Keywords: None
References: None |
Problem Statement: Aspen DMCplus controller can continue to run with Calibrate/Adaptive Modeling if the key SLM_APC_AM exists in the license file. For v9.0 and later, if this key is missing and the controller has Calibrate/Adaptive feature enabled, it would exit with a "License Configuration Error". | Solution: To keep the Calibrate/Adaptive feature enabled on the controller, it would have to be converted to a DMC3 controller.
To keep the controller as a DMCplus controller, Calibrate/Adaptive needs to be turned off in the controller.
For ACO-based controllers, open up the latest ccf file from the online server. Go to Tools>Options. Under the General Tab, un-check the Enable Adaptive Modeling features check box
Save the ccf and reload the controller via APC Manage or PCWS
For RTE-based controller, use DMC3 Builder to open the project contained the latest copy of the application. Go to Deployment and click on Deploy
Un-check the Enable adaptive modeling checkbox and re-deploy the controller
The DMCplus controller now can operate normally.
Note: If the mode is changed to Calibrate accidentally, the engine would reset the mode back to Control and issue an warning message
"Requested Mode Calibrate is not permitted by licensing."
Keywords: DMCplus, DMC3, SLM_APC_AM, License configuration error
References: None |
Problem Statement: How can I achieve the actual field reactor outlet temperature without lowering the inlet temperature to the equilibrium reactor in my model? | Solution:
Engineers tend to want to accurately represent verified variables from the field on a model. This can happen with a reactor, in which the engineer has verified temperatures of both inlet and outlet of 215 °C and 250 °C respectively. Issue arises when the outlet temperature calculated by the program is larger or smaller than this value. If we want an accurate representation of both temperatures, we cannot try and re-specify the inlet temperature until we meet the desired outlet T, so another workflow is required.
This example utilizes an equilibrium reactor to show the process. The model simulates methanol formation through two different reactions:
Considerations
The example also has some other light end components on the process, such as Nitrogen and Methane. The selected fluid package is Peng-Robinson, since water composition and interactions with other components is neglected and process handles mostly light end hydrocarbons with no hypocomponents.
The flowsheet’s Unit Operations are a Simple Weighted Heat Exchanger and the equilibrium reactor. The Heat Exchanger is used to pre-heat the Feed stream with a recycle stream coming from the overhead of the reactor; in said exchanger, the process feed is fed to the Shell Side and the recycle stream on the tube side. Shell-side outlet is then fed into the equilibrium reactor.
Reactions take place on an adiabatic reactor with no dP and no associated volume. One of the most useful values we can check on the Equilibrium Reactor window is the Reaction Heat (Heat of Reaction at 25 °C). This reported value is a result of an enthalpy balance between the reactants and products.
We notice the Heat of Reaction is retrieved from:
Hreaction(25 °C) = Hform(Products) - Hform(Reactants)
Said Heat of Reaction is the reason behind the calculated outlet temperature, so the need of affecting this enthalpy balance arises.
We could try changing the Heat of Formation parameter for a particular component, but this approach is not recommended, since it would affect every enthalpy calculation for the selected component through the whole simulation.
The workflow selected for this example was adding an energy stream to the reactor. This automatically adds one degree of freedom to the unit operation and, with this, we can specify the desired outlet temperature for the retrieved vapor product.
Keywords:
reactor, heat, field, values, compare, reaction, HYSYS
References: None |
Problem Statement: Surface Jet Pump Unit Operation Extension | Solution: The Surface Jet Pump estimates the performance of liquid surface jet pump (also known as ejector) under a given set of flow conditions, This unit also also indicates the approximate size of the surface jet pump required. This extension does not model multi-phase flow due to its complex nature. This is a third party extension and this is the reason this unit is not available in the Model Palette. Users will need to download this from the attachment and register this before using this in Aspen HYSYS. Users can also download this extension from AspenONE Exchange - https://esupport.aspentech.com/S_Article?id=000047152
To register an extension in Aspen HYSYS, go to Tools | Preferences | Extensions and click the Register an Extension button. Browse to the folder location containing the HYSYS extension and double click on the file to be registered. You will require full Administrative right on your PC to register an extension.
Full details on surface jet pumps and their applications are given in Caltec's “Surface Jet Pump’s Engineer’s Handbook", available via website - www.caltec.com. For further questions about this extension, please contact CaltecThis extension was created by a third party and that’s why it’s not shown in Hysys Unit Operation Palette.
Key words
SJP, Surface Jet Pump, ejector, Caltec
Keywords: None
References: None |
Problem Statement: How to track the overall composition for the leaving phase and the liquid that stays in a KO Drum | Solution: If the user wants to track the overall composition for the leaving phase and the liquid that stays in a KO Drum, the user can:
For the overall:
Check the composition of the inlet stream in the results tab.
For the vapor phase:
Check the composition of the outlet stream of the KO drum.
For the liquid phase: Unfortunately, as the liquid won't continue in the network the composition won't be reported. So, the user can:
Manually get the composition of the liquid with a mass balance calculation.
Copy and paste in Aspen HYSYS the overall compositions and flash conditions (T and P or VF) from the inlet of the KO drum.
Note: The fluid package must be the same or similar to achieve the accurate results.
Keywords: Composition, Phases, KO Drum, Stay, Leave
References: None |
Problem Statement: As shown in the attached file, a sensitivity analysis is setup to evaluate the effect the of the packing height (from 1 m to 6 m) for the purity of the distillate, but the mass transfer is not affected. | Solution: The section packed height will not affect the equilibrium if the film resistance is being ignore for both phases, in the Rate Based Setup/Sections tab.
If the user ignores the film resistance, the column will pass from a rate based model to a equilibrium because there isn't any mass transfer resistance and it always will get the maximum purity of the theoretical stages. So the user must select one of the available options to calculate the film resistance.
Keywords: Rate Based, Equilibrium, Height, Packing, Film, Resistance, Mass transfer
References: None |
Problem Statement: I got a result warning 1876: Phase mixing effects for the two liquid phases in stream 1 give an effective viscosity over 50% above the higher of the two individual phase viscosities, in the temperature range... Alternative calculation methods are available. How to troubleshoot? | Solution: When you see this warning message, it means there are two liquid phases formed in the exchanger: oil phase and water phase. The flow is often in the form of an emulsion, which means droplets of one phase are carried along by the flow the dominant, continuous phase. The presence of such droplets can significantly increase the viscosity of the dominant phase, so effective viscosities higher than either the oil or water often occur. When one phase is present in relatively small amounts, the situation is straightforward, but when the amounts of the phases are comparable, there will be significant uncertainties in the calculated viscosity.
This warning message is produced whenever a predicted viscosity is more than 50% greater than the oil viscosity. It means the result maybe too conservative, and you may consider using an alternative method to calculate the viscosity.
In EDR, there are five methods to calculate the viscosity for two liquid phases:
The HTFS selected method (HTFS emulsion for single phase and boiling, higher viscosity for condensing). NOTE: Selections of the higher viscosity may be conservative
The HTFS emulsion method
The HYSYS emulsion method
Use higher viscosity
The old HTFS method (the only option in V7.2.1 and earlier - basic Brinkman method)
You can choose the method under Input- Program Options- Methods/Correlations- General.
For more details on these viscosity methods, please check Help menu topic: Shell&Tube: Viscosity Method for Two Liquid Phases
Keywords: Shell&Tube
Viscosity
Two liquid phases
References: None |
Problem Statement: Even when the Depressuring tool is solved, some cells of the profile results are empty (including the peak flow information) | Solution: The cells are empty (including peak flow information) as the default strip chart was modify. So, Aspen HYSYS won't refresh the data in the report as the Depressuring tool uses the information from the default strip chart.
To avoid that, you have two options. The first one, do not modify the default stripchart; instead create and configure a new one as shown in the following article How do I create a strip chart?.
The second option, in the default stripchart click on "View Strip Chart". Right click, select Graph control and in the curves tab uncheck the "Show Curve" box; that won’t affect the link between the strip chart and it will report only the variables shown in the strip chart.
Keywords: Strip Chart, Modify, Default, Curve, Results, Empty, Converge
References: None |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.