question
stringlengths 19
6.88k
| answer
stringlengths 38
33.3k
|
---|---|
Problem Statement: Sometimes a simulation model contains user updated pure component and binary interaction parameters. These parameters can be used in other process models. The steps to create an enterprise database for these parameters are discussed in this document. | Solution: In newer versions of Aspen Plus, you no longer need to make the division of files into DAT and INP, as described in How to create an enterprise database for pure component and binary interaction parameters from a simulation model (V7.3) . Now exporting the DFM file, modifying it, and saving it as an INP format is enough to create a new database.
The following steps are to be done in Aspen Plus.
1) Open the file that contains the modified parameters from which you want to create the database.
2) From Setup | Report Options, select Property tab, and check DFMS format input file (DFM file) check box.
Note: In order to prevent duplicated data in your DFM file, please uncheck all boxes except the two marked on the screenshot above.
3) Run the model in the Properties Environment. If there are any error messages in your simulation, those must be solved before proceeding with the database creation.
4) Once results are available, go to the Simulation Environment and click on File | Export.
5) Select the file type (.rep) and save the report in a place of your preference (two files will be created: a .dfm and .rep).
The following steps are to be done in Notepad:
6) Open the .dfm file generated in step 5 using Notepad or any ASCII text editor and change the following (in attachments, there is an example file of what the dfm file should look like. Refer to Original DFM.
Erase any duplicated data (if you haven’t deactivated the checkboxes in step 2.
Cut and paste the data New Prop between the fields PROP-DATA and NEW COMP. Make sure they have a blank line separating them.
Make sure that the fields FILE CSTDB XYZPWD New and WRFILE XYZPWD have the exact same name in the characters XYZPWD (regardless of what you decide to name them). This will become the name of your databank in Aspen Properties.
Note: In attachments, a modified version of the txt file is available as an example reference.
7) After editing, save the file as (.inp). You may rename it. This file will be used to create the database.
Note: Make sure you are not saving it as (.inp.txt) instead of .inp (it is necessary to modify the file type).
The following steps are to be done in Aspen Properties Dababase Manager:
8) Open the Database Manager, right click on Database Manager Aspen Properties and select Create New Database. You might need administrator’s rights to work with this tool.
9) Import legacy data files and select the proper authentication mode (SQL or Windows), Login Name and password and click Next.
10) If you want, choose to add security. If not, just click on Next.
11) Select the .inp file saved and click on Finish. Once the conversion is finished, the databank will be available for your reference within the Database Manager.
12) Reopen Aspen Plus and you’ll see that a new enterprise database has been created with your data under Components | Specifications | Enterprise Database tab.
Note: Adding the database to the APED system does not automatically add it to your simulation file. You will need to make sure that the enterprise database is added to your file and that it has preference over the other default databases by moving it to the top of the list. You can set your property databank as default by opening your simulation file and going to File > Options > Properties Basis. This needs to be done to each machine where the database has been registered.
Keywords: enterprise database; user database; databanks; .inp file; DFMS;
References: None |
Problem Statement: On using V11 MSC Token installation media, on running the silent installation package, the 64-bit path does not get installed and is missing from the environment variable. The installation is successful but only the 32-bit path gets installed, and the 64-bit path is missing. | Solution: This is a known issue and fixed in V12 defect 537559, 535148. When we use silent installation, these products will be affected.
The current workaround for V11 is to:
Copy the SetEnvironmentVariabe.exe (found in the attachment section of this article) to your local path, the same location as aspenONE Manufacturing and Supply Chain V11.bat.
2. Change aspenONE Manufacturing and Supply Chain V11.bat and add the command to bat.
(SetEnvironmentVariable.exe aspencommon64 C:\Program Files\Common Files\AspenTech Shared)
(SetEnvironmentVariable.exe aspencommon86 C:\Program Files (x86)\Common Files\AspenTech Shared)
Keywords: silent
unattended
installation
64-bit
Environment variable
missing
References: None |
Problem Statement: Back to Basics :
What is difference between T. BUY & T. SELL in PIMS. | Solution: 1) T. BUY
Table BUY is used for identifying the materials that can be purchased and to define the purchase constraints on individual materials.
Structure of T. BUY
2) T. SELL
Table SELL is used to identify the materials that can be sold and to define sales constraints on individual materials.
Structure of T. SELL
The major difference in T. BUY & T. SELL is,
While creating T. BUY it should include the COST column as it includes the materials that are to be purchased and T. SELL must include the PRICE column as it includes the materials to be sold
Key words:
Table BUY
Table SELL
Keywords: None
References: None |
Problem Statement: What is the single character nomenclature to be used for 100 models that can be attached to GLOBAL model. | Solution: As PIMS supports 100 models that can be attached to GLOBAL model which can be defined with the help of single character nomenclature. Below are the characters that can be used as single character nomenclature for 100 models.
Characters from A to Z = 26 models
Characters from a to z = 26 models
Characters from 0 to 9 = 10 models
Total characters = 62 models.
For the remaining 38 models’ characters from the extended ASCII can be used as PIMS supports the us of most of the characters in the set of extended ASCIIs.
Hence, in the list of extended ASCII characters from 128-255 can be used which are additional characters beyond the mentioned ones for 100 models that can be attached to GLOBAL model.
Below is the list of characters that can be used-
Note: It is not a good idea to use mathematical symbols as plant identifiers such as + - / . ^ * ( ).This will be a problem if the model includes nonlinear equations on the tree.
Key words:
ASCII Characters
Global model
Keywords: None
References: None |
Problem Statement: How do I specify wind velocity in an Aspen HYSYS pipe segment? | Solution: To include wind velocity in the Heat Transfer Coefficient calculation of a Pipe Segment, please go to Rating tab | Heat Transfer | Estimate HTC form:
The Ambient Medium dropdown menu contains three option for medium specification: Air, Water and Ground. Select Air.
It will display a field for Velocity specification. For wind temperature please enter the Ambient Temperature and select the Include Outer HTC checkbox.
At this point, ff all the required information is entered properly, you will see the Overall Heat Transfer result on the Heat Loss page.
Note: This method is only available in Steady State simulations.
Keywords: Wind velocity, pipe segment
References: None |
Problem Statement: How do I resolve the error message in Aspen Plus RadFrac Column Hydraulics “unable to compute curve at the specified minimum pressure drop per unit packed height”. | Solution: While RadFrac column hydraulics, after adding the column internal, results are available. But the hydraulic plots display “unable to compute curve at the specified minimum pressure drop per unit packed height”.
This issue might be observed while rating the column, due to this the curve may not be calculated properly.
This warning could be removed by modifying the pressure drop calculation methods available under column internals Section Design parameter & modify the “pressure drop calculation methods”.
There is also option available minimum pressure drop per unit packed height user may also modify this as well to remove the above warning on pressure drop.
For more details on the “pressure drop calculation methods, please refer Aspen Plus online help topic.
Keywords: Rates-based Modeling, Pressure Drop Calculation method, Curve, Hydraulic Plot
References: None |
Problem Statement: I get the following message when executing a kinetic reactor such as a RCSTR:
** ERROR
M-BAL LOOP DID NOT CONVERGE IN 200 ITERATIONS. CHECK
STOICHIOMETRY/MOLECULAR WEIGHTS.
How do I remove this error or warning message? | Solution: This warning/error is related to the mass balance. Users should first check if there area any errors in the molecular weight of a product or reactant or errors in the stoichiometry of a reaction.
This error can also occur if there are reactions with zeroth-order reactants which are not present or if there are unreasonable rate constants, especially activation energies.
Check the convergence option under the block, for example for RCSTR reactor, use below options to modify the parameters those affect the mass balance.
More information on troubleshooting RCSTR Mass Balance Convergence failureSolutions is in the help under Using the Simulation Environment -> Unit Operation Models
Keywords: M-BAL LOOP, Mass balance error
Mass Balance error
References: Manual -> Reactors -> RCSTR Reference -> Troubleshooting RCSTR -> RCSTR Mass Balance Convergence Failures:
Symptom |
Problem Statement: Sample example of Sour Water Stripping with Phenol | Solution: Sour Water Stripping with Phenol
Sour Water Stripping With Phenol.hsc demonstrates the difficulty of removing phenol using the sour water stripping process. In this case, phenol remains in the treated water because of its low volatility. The presence of phenol in the sour water may require the inclusion of a separate phenol extraction step in the water treatment process (Kohl A. L. and R. B. Nielsen, 1997).
Keywords: Sour Water, Stripping, phenol, etc.
References: None |
Problem Statement: Which non-condensable components are included within Aspen HYSYS component list? | Solution: Henry's Law cannot be selected explicitly as a property method in HYSYS. However, HYSYS uses Henry's Law when an activity model is selected and non-condensable components are included within the component list.
HYSYS considers the following components non-condensable:
Component Simulation Name
CH4 Methane
C2H6 Ethane
C2H4 Ethylene
C2H2 Acetylene
H2 Hydrogen
He Helium
Ar Argon
N2 Nitrogen
O2 Oxygen
NO NO
H2S H2S
CO2 CO2
CO CO
Keywords: Henry Law, Non-Condensable, etc
References: None |
Problem Statement: How to define Henry components in Aspen HYSYS? | Solution: In Aspen HYSYS, HYSYS Databank, there is no Henry's component as in Aspen Plus or Aspen Properties. User can use Aspen Properties databank in Aspen HYSYS and assign the component as Henry's.
Add Component List from Aspen Properties Databank.
Add required components.
Henry Comps tab is available to defining Henry Comp List.
Select Henry component in Available components by using forward arrow.
Keywords: Henry Comps, Aspen Properties Databank, etc.
References: None |
Problem Statement: What OPC servers has the Aspen CIMIO for OPC interface been tested with and what were the results? | Solution: We have connected Aspen CIM-IO for OPC interface to several dozen different OPC servers. Since the OPC server list is ever growing and the OPC standard has become fairly well known we have stopped keeping track of all the servers. We can talk to any OPC server that conforms to the 1.0 or 2.0 DA specification. Performance varies, but is almost always better than other interface methods. Generally the performance limitation is the device or API layer the OPC server must maintain. For some systems we have found the performance to be near the limits of CIM-IO at 1200 pts/sec. Some of the different OPC servers we have fielded are:
Siemens WinCC
Honeywell HCI
Yokogawa Centum CS
Fisher Rosemount Delta-V
Fisher Rosemount Pint (Chip)
Rockwell RSLINX
Intellution FIX
We have also tested against all of Matrikon's OPC servers - http://www.matrikon.com/products/drivers.htm
Keywords: CIMIO
CIM-IO
OPC
References: None |
Problem Statement: In this KB, we will discuss different kinds of A1PE plots. | Solution: 1. Event Overlay Plot:
- The event overlay plot is used to display a single process variable trended across multiple event time ranges.
Note: tag aliases are supported.
Figure 1: Event Overlay Plot
2. Multi-Event Plot:
- The Multi-Event plot is used to aggregate process variables and/or characteristics plotted event to event.
Figure 2: Multi-Event Plot
3. Golden Profile Plot:
- The Golden Profile plot is used to monitor reference tags in real-time against an ideal or golden event profile
Figure 3: Golden Profile Plot
Keywords: A1PE plots , Event Overlay , Multi-Event , Golden Profile
References: None |
Problem Statement: What is the slug flow rate unit? | Solution: The equation for Slug Flow Rate is:
(Slug Flow Rate) / (Overall Liquid Flow Rate) = LS* HS (ULS + UGS) / ((LS + LB) * ULS)
The equation gives the fraction of the total liquid flowrate that is transported as slugs (hence it is unitless value).
Where,
LS - slug length (as reported by HYSYS) - m
LB - bubble length - m
HS - slug holdup (always less than one if Gregory Aziz Mandhane is used)
The liquid holdup is defined as the ratio of the volume of a pipe segment occupied by liquid to the volume of the pipe segment at a given point in the pipe segment. It is unitless fraction which varies from 0 (for all gas flow) to 1 (for all liquid flow).
In a slug, allowance is made for gas bubbles in the slug. The liquid holdup in the slug is the fraction of the pipe cross-sectional area occupied by liquid where this liquid/gas mixture bridges the pipe with the liquid being the continuous phase. The model takes into account the shape of the interface from the back of one slug to the front of the next one. This is called the bubble region and the local film thickness is not uniform. The quoted film holdup for this region is the average liquid holdup from the back of one slug to the front of the next one.
Key Words:
Slug Flow Rate, etc
Keywords: None
References: None |
Problem Statement: How to perform dynamic modeling for acid gas cleaning process in HYSYS? | Solution: The Acid Gas Technology, also known as Acid Gas Property Package, it does not fully support dynamic simulation in previous version(Pre V11. The reason can be refer to KB 000060096-Dynamic Simulation with Acid Gas Technology in Aspen HYSYS.
Now from V12, Aspen HYSYS can support Acid gas cleaning dynamics using reduced order modeling.
And its support limited Amine (MDEA, DEA) as well as light hydrocarbons (methane, ethane). it would be expected to enlarge more amines in future version.
The model range limited as below:
- Amine < 50%,
- CO2, H2S < 25%,
- Methane, Ethane < 10%,
- H2O > 30%
User required to set the related parameter for acid gas dynamic so that can apply reduced order modeling technology:
Go to Properties Environment | Fluid Packages | Basis-1|Phase Handling | Enthalpy & Fugacity Models and select Reduced Order Model.
Then just take other regular dynamic setting for the model preparation, please refer to below KB000061994:
Getting Started with Aspen HYSYS Dynamics.
Keywords: Acid gas dynamics, Amine, MDEA, DEA, Reduced order modeling.
References: None |
Problem Statement: The customer wants to consolidate their MS SQL Servers and this means they want to run Aspen Framework Server using a database on a remote machine. Is this a supported environment? | Solution: Yes this is supported. It is assumed that for AFW the SQL Server can be installed on a either on the local machine or on another machine. However, some changes are necessary after the basic AFW installation.
1) Copy the C:\Program Files (x86)\AspenTech\AFW\DBUtility folder from the AFW server to the SQL server
2) On the SQL server, execute the CreateAFWDB.exe tool and create the AFWDB database
3) Back to the AFW server, edit the registry key
HKEY_LOCAL_MACHINE\SOFTWARE\Wow6432Node\AspenTech\AFW\PROVIDER key from MSACCESS to MSSQL
4) In the C:\Program Files (x86)\AspenTech\AFW\DBUtility folder, execute the PfwDataSource.exe tool and configure the AFW server to use the remote AFWDB created in the SQL server
Note, the CreateAFWDB.exe tool is installed with the AFW Server Components group in the installer. If this option is disabled you would first have to uninstall Aspen Local Security Server product to enable the AFW Server Components group.
Keywords: AFW
Remote SQL Server
Framework
Local Security
References: None |
Problem Statement: Do we have any examples of drying operations for Biomass/Food Applications? How do we consider drying of biomass? | Solution: Aspen Plus can be used for the processes using nonconventional components for biomass or food applications.
This article provides an example using biomass drying which will demonstrate the application for building the Food/Biomass dryer. You can simulate biomass or food material as a non-conventional solid component.
The biomass is characterized by proximate analysis, ultimate analysis, and sulfur analysis. User need to use MIXNCPSD stream type – this has both mixed and non-conventional solids sub-streams, with particle size distribution for the non-conventional sub-stream.
The dryer model considers the moisture outside the particles (in this example specified it in the MIXED sub-stream) and the moisture content of the cells (specified as moisture percentage in PROXANAL).
Keywords: Food application, biomass drying, non-conventional components drying, Dryer for food applications, biomass dryer
References: None |
Problem Statement: Why there is a different temperature of the split streams than its original stream after the split block? | Solution: In the splitter block after the split, by default Aspen engine makes the flash calculation. Due to this, there is an increase in the entropy of the stream (stream results) due to liquefaction, which resulted in the temperature change. There is some kind of flashing happening inside this closed loop.
This can be compromised by:
1. Using Vapor only in the flash calculation option.
2. Disabling Perform heat balance calculations option found in Block options -> Simulation options
Now, the temperature of the split streams are the same as its original stream after the split block.
Keywords: Temperature, Split Block, Aspen Plus
References: None |
Problem Statement: How to load an Aspen InfoPlus.21 Snapshot | Solution: STOP Aspen Infoplus.21 using the IP.21 Manager
Note the snapshot path found in the DBClock command line parameters.
Copy your backup file to the location above.
Rename the old file with a name that indicates that it is an old file. For example 'InfoPlus21_old_DATE.snp'.
Name the backup file as InfoPlus21.snp or the same name as the snapshot that is typically loaded.
Start Aspen Infoplus.21 Manager.
Keywords: Snapshot
InfoPlus.21
IP.21
TSK_DBCLOCK
References: None |
Problem Statement: When trying to manipulate an element of data array created using the Process Data Add-in, the following error will be displayed:
You cannot change part of an array. | Solution: There are quite a few rules about how arrays work. The following Microsoft website article explains these rules.
https://support.office.com/en-us/article/Rules-for-changing-array-formulas-8A80C2C1-44C1-42C3-88C1-37A3A7C2D367
Specifically, there is a bullet on this page which states:
* You can't delete cells in an array formula (you'll see a You cannot change part of an array error), but you can delete the entire formula and start over.
You can change the formula for the complete array using ctrl-shift-enter.
Keywords: Array
Excel Add-in
References: None |
Problem Statement: Can Aspen HYSYS generate isobars for a condensing stream? | Solution: The Heating blocks in HYSYS (heater, cooler, heat exchanger) can generate property tables and plots to generate these heating curves.
After any of these units has solved simply go to the Performance tab and select the Table section.
You will be able to obtain the heating curves for both the shell side and tube side for heat exchangers.
Keywords: Heating Curves, Condensing Stream, Aspen HYSYS
References: None |
Problem Statement: Why the temperature profile result for tail pipe section (upstream & downstream) are inconsistent when running a blowdown analysis? | Solution: The temperature profile result for pipe section (upstream & downstream) are inconsistent is due to the lack of a component to represent the change of diameters between the orifice plate and the pipe segment. Physically we do not have this component, so there will be a change in the temperatures due to the change of diameters. As you may already know, the orifice plate uses the same principle as a Venturi nozzle, so this change of diameters results in a change of fluid pressure and temperature as well.
Keywords: Aspen HYSYS, Blowdown, Temperature Profile
References: None |
Problem Statement: How efficiently Aspen DMC3 | Solution: can be used to reduce energy & how they can minimize waste by reducing variability in the process & improve quality of products.
Solution
Within identified challenges to achieve the sustainability goals by automatically adapting and optimizing to ever changing plant conditions. Embedded AI to build seed APC models from historical data combined with state-of-the-art adaptive process controlSolution to sustain peak plant performance.
The below figure shows the improvements with performance of process control with DMC3 model with adaptive process control .
The Aspen DMC3Solution have following features and capabilities through which the objectives are achieved:
Powerful machine learning algorithms to mine the wealth of historical process data to build seed models.
Adapt models in the background & rapidly deploy
Quickly adjust controller tunning & Optimization objectives
Easily accounts for severe process non-linearities
Remotely monitor performance with secure data access
Keeps models current while simultaneously optimizing for high energy efficiency.
Apply adaptive process control to all major energy consuming units.
Set control strategies to optimize for carbon reduction.
Aspen DMC3Solution have successfully implemented with good results in ethylene process, reduce energy consumption in aromatics separation area, drives savings on fresh H2 import & vent gas reduction, Delayed Coker in refinery (DCU) and Power plants etc.
Key words:
Sustainability , Aspen DMC3
Keywords: None
References: None |
Problem Statement: Why the APRM batch jobs scheduled on the APRM server should are not able to fetch data from IP21 server After a reboot of IP21 server?
error message in the log file [ Batch21Services_AtlServer__x64.16.0.1.156 ] :
B21BAI-60271: Could not connect to server <IP21_Server_Name>: -2147220990 | Solution: In this case the IP21 server and APRM server are different server but it is recommended to keep APRM and IP21 on the same server.
The maximum time the server will try to reconnect the IP21 server is 15 minutes after that it will give error that could not connect to IP21 server.
We need to stop the services from both end till the time IP21 server is up or else activity should be within max.15 min and IP21 & APRM application should be on same server is recommended.
For running the APRM service smoothly after IP21 server update follow to update the Patches for IP21 as below:
Recommended Patches:
For APRM server -V11 – the following Patch should be installed:
https://esupport.aspentech.com/apex/S_SoftwareDeliveryDetail?id=a0e4P00000OwGNUQA3
Download the patch then unblock it then copy it to server then unzip it then run the update agent by run as administrator.
Detail procedure to update the Patch explained in PDF.
Key words:
APRM , IP21
Keywords: None
References: None |
Problem Statement: If RTO Watch is configured with Aspen OnLine, the CIM-IO connection will fail. | Solution: Navigate to C:\Program Files (x86)\AspenTech\CIM-IO\io\Cimio_rto and open the file named cimio_rto.exe.config. Change the AspenOnLineVersionMinor string from 0 to 1, save the file, then restart CIM-IO.
Keywords: None
References: None |
Problem Statement: AFW performs automatic logging in C:\inetpub\logs\LogFiles. In some scenarios, these logs can grow too big in size and fill up the disk on the web server. This KB presents a | Solution: to turn off that logging operation.
Solution
Open IIS Manger, expand the server name and the sites folder
Go to Under default Web Site, Aspentech, Afw. Then, select logging under IIS.
Click on disable. This will prevent AFW logs to grow too big on the web server
Keywords: …AFW, Web server, AFW Logs, IIS, ineput
References: None |
Problem Statement: What is the difference in between Ignore bad quality Data and Allow Live agent processing with bad quality data in Aspen Mtell System Manager? | Solution: Ignore bad quality data – Default enabled
This determines whether bad quality historical data is allowed into the training data set. When enabled, data points with bad quality are excluded from the training data set. When disabled, all data points are saved to the training data set, regardless of quality. This option does not apply to live processing of new data for live agents.
Allow live agent processing with bad quality data – Default disabled
For machine learning live agents, this determines whether bad quality data obtained during live processing of new data is included in the training data set. Normally it is not.
Note: The machine learning live agents always run with the latest data obtained from the historian, even if the data is bad.
For other live agents (Rule, CM, PM), this determines whether the agents run when the latest data reported by the historian has bad quality. Normally, Aspen Mtell does not run non-machine-learning live agents against this data, instead returning an error indicating that the agent received bad quality data. If Allow live agent processing with bad quality data is enabled then these agents run even when the data is flagged as bad.
Keywords: Historian Quality
Mtell Data
TDS
Remove Bad Quality
References: None |
Problem Statement: How to Set up an Automatic Analysis using PID Watch? | Solution: The most useful mode for Aspen PID Watch analysis is its ability to automatically schedule the scanning and analysis of groups of PID loops.
1) Go to the Process Control Web Viewer.
2) Click on the history tab.
3) In the navigation window, open the Aspen Watch server and select the PID Loops section. Once into the PID Summary page, click the Group Configuration option (circled in figure below).
4) To define a new group, define the group name (use VirtualP_Loops), description (PID loops from Virtual Plant Simulation), the start (the next 15 min value available) and end times (15 mins into future from start time), and the recurrence interval (1 day).
5) Select the loops to comprise the group with the Available/Selected tool and click apply.
Later, to see the completed analysis, click on the Analysis tab.
Keywords: PID Watch, Schedule
References: None |
Problem Statement: In this KB, we will use the final corrected model to configure the controller tuning parameters and add constraints. | Solution: The following calculations are added as per below screenshot:
Figure 1: Calculation 1
Figure 2: Calculation 2
Figure 3: Calculation 3
Now, the constraints are added because we want the temperature across the three passes to be uniform. So, we want to force the gain between the flowrate and temperatures to be the same. The following figure illustrates how it is done.
Figure 4: Constraints
The traditional tuning method was utilized because the LP costs are already provided in the Excel Sheet
Figure 5: Tuning – inputs
Figure 6: Tuning - outputs
Move suppression values are recommended to be the following, for a CDU processes:
Where X often ranges from 2 to 5.
*Assuming operations do not like to see feed rate or pressure setpoints move quickly.
After performing these changes, we will look at the predictions in order to evaluate model’s accuracy before updating the master model. This step has been done for the same process in a previous KB I made, please check it out in below link:
How can predictions be improved using DMC3 Builder?
https://esupport.aspentech.com/S_Article?id=000094781
Keywords: Aspen DMC3 Controller , Configure
References: None |
Problem Statement: In this KB, we will examine the assembled model and fix any collinearity issues & carry out transform estimation part. | Solution: We will perform following steps in this KB:
Transforming nonlinear valve Ops
Collinearity Analysis
Transforming FC122SP Vs FC122OP
Figure 1: Piecewise Linear FC122SP Vs FC122OP
Transforming FC123SP Vs FC123OP
Figure 2: Piecewise Linear FC123SP Vs FC123OP
Transforming FC124SP Vs FC124OP
Figure 3: Piecewise Linear FC124SP Vs FC124OP
Transforming FC125SP Vs FC125OP
Figure 4: Piecewise Linear FC125SP Vs FC125OP
Collinearity analysis – variable selection
Figure 5: Collinearity analysis - step 1
In the RGA Analysis Results, the highlight option is clicked to display the collinear pairs.
Figure 6: RGA Analysis Results
In the above screenshot, we can see that the RGA values are large which means they are nearly collinear. Typically, if the RGA is greater than 10, it represents collinearity. However, if it is infinitely large that means they are perfectly collinear which is OK.
Figure 7: RGA before repair
After clicking on repair, the software will fix the collinearity issue detected earlier as shown in the below figure.
Figure 8: RGA Repair
Inf means infinity which states that the RGA is infinitely large which indicates they are fixed to be perfectly collinear.
Keywords: Aspen DMC3 Builder, Collinearity
References: None |
Problem Statement: In this KB, we will learn how to Import dataset into DMC3 Builder and use the dataset to build the controller configuration upon. | Solution: In this step, we will import dataset into Aspen DMC3 Builder on the data provided:
A new project named Crude Unit Heater Controller is created.
Figure 1: Crude Unit Heater Controller Project
Importing dataset into the project by clicking on import.
Figure 2: importing Dataset
Clicking start to interpolate the values in the dataset, then closing to add the dataset to the navigation bar.
Figure 3: Adding dataset to navigation bar
Model identification process & time to steady state is set at 60 mins as per the problem statement.
Figure 4: Model ID
Keywords: Aspen DMC3 Controller, model identification
References: None |
Problem Statement: The goal of this series of articles is to learn how to build a DMC3 controller with DMC3 Builder interface using a pre-collected step test data for Crude Unit Heater. | Solution: The various parts that need to be completed as part of this series include-
How to Build a DMC3 Controller Series. Part 1: Import Data
https://esupport.aspentech.com/S_Article?id=000097929
How to Build a DMC3 Controller Series. Part 2: Fix Gain Collinearity Problems and Transform Estimation
https://esupport.aspentech.com/S_Article?id=000097930
How to Build a DMC3 Controller Series. Part 3: Configure the Controller Application
https://esupport.aspentech.com/S_Article?id=000097937
How to Build a DMC3 Controller Series. Part 4: Configure Dynamics and Simulate Controller
https://esupport.aspentech.com/S_Article?id=000097938
How to Build a DMC3 Controller Series. Part 5: Deploy the Controller Online
https://esupport.aspentech.com/S_Article?id=000097939
The pre-collected step test data as well as an excel sheet containing some tuning information are attached to this article. Please download it them from the attachment section.
Additional Information
TTSS- 60 mins
The Charge Heater Fuel Gas Flow Valve OP and PV limits are equally important, but they are more important than the Interheater Skin Temperature limits
Reactor wait (TX115.PV) and Reactor Mole ratio (RX171PV) is equally important as PC141OP and SC151OP, however they are less important than the Interheater skin temperature limits.
The Interheater Delta Temperature are the least important limits
A new input calculation named TDX116 should be created (Description- Differential Temperature Measurements 116) and enter the formula shown below-
DT = T1 - T2
The above should be repeated. The input for TDX116PV and TDX117PV the binding details for
TDX116PV = TI111PV – TI1112PV
TDX117PV = TI113PV – TI112PV
TDX118PV = TI114PV – TI113PV
The final model should look like this.
Keywords: DMC3 Controller , Build Controller , DMC3 Builder
References: None |
Problem Statement: In this step, we will deploy the controller online. | Solution: Before deploying the controller online, the connections were tested first as shown below.
Figure 1: Connection tests
Now that the connections are successful, the controller can be deployed.
Figure 2: Deployment
After the deployment step, the controller is running now which means it can be accessed using PCWS.
Figure 3: Controller running
Before going to PCWS, Aspen Watch for the controller was started.
Figure 4: Aspen Watch
The following screenshot is the PCWS interface of the controller.
Figure 5: PCWS Interface
Keywords: DMC3 Controller , Deployment , Online
References: None |
Problem Statement: How do I disable the user registration dialog in aspenONE?
The first time a user launches any aspenONE product, he gets presented with a User Registration dialog. The purpose of this registration is to create/update a user record with AspenTech Support. Registering offers different benefits which are listed on the Support Web site under http://support.aspentech.com/supportpublic/RegistrationBenefits.html | Solution: Although we strongly recommend that User Registration dialogs would be displayed for all the users of our softwares, they are cases where System Administrators want to disable them in their environment. Here is how to proceed.
1) DISABLING USER REGISTRATION DIALOGS AT 'USER LEVEL' (Applies to all aspenONE versions)
Define the following registry entry in the user's profile before he launches any aspenONE product for the first time:
Registry Key: HKEY_CURRENT_USER\Software\AspenTech
Value Name: UserRegistered
Value Type: REG_SZ
Value Data: 1
2) DISABLING USER REGISTRATION DIALOGS AT 'MACHINE LEVEL' (Applies to aspenONE and later)
With 2004.1 and later, user registration can be turned off for the whole computer by setting the following registry entry:
Registry Key: HKEY_LOCAL_MACHINE\SOFTWARE\AspenTech or HKEY_LOCAL_MACHINE\SOFTWARE\Wow6432Node\AspenTech
Value Name: DoNotRegister
Value Type: REG_SZ
Value Data: 1
With these entries in the registry, user(s) will not be presented with the User Registration dialog.
Keywords: registration page, disable, registry section
References: None |
Problem Statement: How to model pressure rise due to the fire heat load input for delayed orifice opening to blowdown system? | Solution: For time delay on orifice opening, you need to enter the initial opening time.
In blowdown, the orifice opens at time t = 0 sec. The above scenario shows that we need to delay the activation of blowdown valve by 20 sec. After these 20 sec, the orifice will be ON.
To simulate a delay in the orifice opening in BLOWDOWN, you need to enter a small number for the orifice opening diameter (for example 0.01 in). The value can be increased later at the desired time interval.
Keywords: Blowdown, Orifice, Opening Time
References: None |
Problem Statement: How can I control the number of Engineering Drawings on my Aspen Capital Cost Estimator project? | Solution: Aspen Capital Cost Estimator (ACCE) provides a field to control the amount of Drawings and Diagrams to be included on the estimate.
In order to fully control all these Drawings, it is important to first review the results generated by ACCE’s system, this will allow you to understand which drawing types need to be adjusted. For details on how to find the System Generated results please review Article 97340
Once you have identified the Drawings that will be modified, you will use the Drawing Types and Drawing Count forms under the Engineering Workforce folder form in the Project Basis View tab.
To modify the existing Drawing count:
Open the Drawing Count form
Select either Basic or Detailed as the engineering phase, depending on the type of drawing you wish to modify.
Enter a Drawing Number – this is a value unique to each drawing type to be included in ACCE. Review the list of Numbers on the Icarus
Keywords: Engineering, Drawing, Diagram, Fees, Adjust, Overwrite
References: guide, Chapter 31, Standard Engineering Drawing Types.
Enter the Contractor number that will receive this change.
Set the Action Code. Select ADD to include additional drawings to the count or select CHANGE to manually specify the drawings to be used in the project. If the CHANGE action was used, you can specify a percent adjustment instead of the count.
Run the estimate to review the new results
To add drawings not included in the system:
Open the Drawing Types form
Select either Basic or Detailed as the engineering phase, depending on the type of drawing you wish to add.
Enter a Drawing Number between 91-99 – this is a value unique to each drawing type to be included in ACCE. Review the list of Numbers on the Icarus Reference guide, Chapter 31, Standard Engineering Drawing Types.
Enter the Drawing Title to be used on the reports
Enter the Drawing Account (i.e. Piping, Instrumentation, Civil etc.). The engineering hours per drawing will be accounted to this category.
Enter the drawing discipline (i.e. Design, Drafting, Model)
Enter the total man hours required to produce the drawing.
Use the Drawing Count form and the steps described in this article to set the amount of drawings to be included. Use the Action code ADD (CHANGE option is not allowed for Drawing numbers 91-99).
Run the estimate to review the new results |
Problem Statement: Is it possible to split a workspace into sub-projects and merge them altogether into one final project later in ABE (Aspen Basic Engineering)? | Solution: In our training, we cover the use of Project Management feature to split a workspace into projects to handle different alternative cases on the design. In this workflow, the user selects only one of the projects to commit and merge back to the master project.
The Project Management feature can extend beyond this workflow to allow the user to segregate the entire design into multiple sections. The difference with the above workflow is that multiple projects will need to be committed back to the master workspace, instead of only one project. Support for this workflow is useful when the project scope covers multiple engineering areas (such as in a gas processing facility, there are amine treatment unit, dehydration unit, sulfur recovery unit, etc.) To implement this workflow, the user can allocate one unit to one project. And the projects can then be merged back altogether one by one using Commit button. Some important details are mentioned below:
1. When you are inside the Project Management window in the master workspace, proceed to commit the first project by clicking on the Commit button.
2. After this step, you won't be able to commit the second project immediately. The Commit button will either be grayed out or the following message will pop up:
3. The workflow requires ABE to re-sync the master workspace (parent project) with the changes made from committing the first project. So you need to log in to the second workspace to perform this sync. Proceed to log in to the second project and launch the Project Management tool, you will find the window displaying a message: The parent project has changed. Click here to update this project. Click on this message, ABE will display a dialogue to show you the updates, proceed to click OK to finish reviewing all updates.
4. The workflow requires ABE to re-sync the master workspace (parent project) with the changes made from committing the first project. So you need to log in to the second workspace to perform this sync. Proceed to log in to the second project and launch the Project Management tool, you will find the window displaying a message: The parent project has changed. Click here to update this project. Click on this message, ABE will display a dialogue to show you the updates, proceed to click OK to finish reviewing all updates.
5. After this step, the sync has been done for the second project. You can now log back to the master workspace and commit the second project as in step 1. Step 2 and 3 will need to be repeated to continue committing additional projects.
Keywords: Projects, Split projects, Commit button grayed out, unable to commit projects
References: None |
Problem Statement: Which are unit operations supported for malfunction scenarios in Aspen HYSYS Dynamics? | Solution: Starting in Aspen HYSYS V11, user can now model equipment failures, malfunction, or performance deterioration quickly and easily in Aspen HYSYS Dynamics for the following unit operations:
Air-cooled heat exchangers
Heat Exchangers
Pumps
Pressure relief devices (PSVs)
Valves
User can access this feature on the Dynamics tab | Malfunction page of the desired unit operation.
Keywords: Malfunction, performance deterioration, equipment failure, etc.
References: None |
Problem Statement: Customer may not be able to run the V12 Emergency Patch installer directly due to some security restrictions. | Solution: Users can expand the emergency patch installer to a local folder and then run the EP.exe to start the emergency patching process.
Using a ZIP extract tools such as 7zip or WinZip, customers can extract the contents of emergency patch installer.
After the file extraction, run the EP.exe from the extracted folder to start the patching process.
To run the EP.exe silently provide the “/s” parameter as part of the command.
Keywords: Emergency Patch (EP)
References: None |
Problem Statement: Do Aspen Mtell Rule Agents follow specific asset offline conditions? How do they behave in terms of asset status? | Solution: Aspen Mtell Rule Agents are designed specifically to only consider the Boolean rule defined in the agent. Offline status/condition of the Asset is not taken into consideration while processing Live Rule Agents. These offline conditions must be incorporated within the Boolean function while creating the Rule agent.
Below is an Example of incorporating the offline condition to prevent Offline period Rule agent triggers:
1) The intent of the rule function is to trigger an alert whenever sensor COMPR1_MBRG1_TEMP.F_CV exceeds 100
2) The Equipment is offline when COMPR1_MTR_AMPS.F_CV dips below 5
3) The AND operator is used to trigger whenever COMPR1_MBRG1_TEMP.F_CV exceeds 100 and the equipment is online i.e. COMPR1_MTR_AMPS.F_CV exceeds 5
Keywords: Asset status rule agent
Offline rule agent
AND operator
References: None |
Problem Statement: The granularity shown in the Processed Data tab of an agent's View Trend does not match the granularity of the TDS under which it is created. For example, an hourly TDS may have processed data every minute. This often happens when the sensor data source is CSV files.
This also may cause agents to take a long time to open trends (i.e. Probability Trend, View Trend, View Trend Lines). | Solution: When importing your CSV historian data, you may have set the Interpolation Mode to None. Follow the below steps to change your Interpolation Mode to Linear or Stair Step.
Note: This procedure will delete the data currently imported in your TDS(s). This means you will need to retrain all the agents under that TDS.
1. Open Aspen Mtell System Manager.
2. Navigate to the Configuration tab. In Settings, select Sensor Data Sources.
3. Select the CSV sensor data source that is not respecting the TDS granularity.
4. Where it says Interpolation Mode, select either Linear or Stair Step. In the top ribbon, press Save.
5. Press Import Files.
6. In the window that opens, press Select Files.
7. Select the CSV historian files(s) that you want to re-import, then press Open.
8. In the Import CSV Files window, press Import.
9. Wait for the CSV files to import. Once they finish importing, press Close.
10. Open Aspen Mtell Agent Builder.
11. Navigate to the Machine Learning tab.
12. Select a TDS that uses the CSV data.
13. In the top ribbon, press Import Data.
14. Select Manually select Assets. In the dropdown, select the asset(s) that you want to use. Press OK.
15. In the newly opened prompt, press No.
16. In the next prompt, select No.
Note: The following step will delete the data you previously imported into your TDS. This means all of your agents under this TDS will need to be retrained. Please proceed with caution.
17. In the window that opens, select Delete all pre-existing data between start and end date for selected assets. Then, press Import Data.
18. Repeat steps 12 - 17 for additional TDS's that use the CSV historian data.
19. Wait for the jobs in the job queue to finish.
20. Your agents' processed data should now adhere to the TDS granularity.
Keywords: Training Data Set
References: None |
Problem Statement: This error message is very general, this KB article explains some tips and tricks to successfully import data. | Solution: Apply the following recommendations in the text file
Sort data from Oldest to newest. First sample should be the oldest
Import the txt file in excel and save it as tab delimited. Then try to import in IQ Model
In the File Format window, make sure to select the appropriate “starting in field” or tag names.
In the expert section, specify date and time format to match the txt format. This is an important check.
In the expert section, you can also change the delimiters accordingly.
Add a blank row below the last sample in the text file otherwise the last sample will not be imported.
If none of these steps seem to work, you can use the IQ excel add in to import data or contact Aspentech Support and send the txt file.
See https://esupport.aspentech.com/S_Article?id=000096115 about how to configure the IQ Excel add in
Keywords: IQModel, Inferential, txt., Specify Data
References: None |
Problem Statement: How is the the CP/CV ratio for a mixture calculated in Aspen Plus? | Solution: If the user requests the CV/CP ratio to be computed (by including CPCVMX in the stream report), Aspen Plus uses the following relations:
CV = DHMX + T*DVMX**2/DVDP
CP/CV = DHMX / (DHMX + T*DVMX**2/DVDP)
where:
P - System Pressure (N/M2)
T - System temperature (K)
DVDP - Partial derivative of mixture molar volume with respect to pressure at constant temperature (M3/(KMOL-N/M2))
DHMX - Partial derivative of mixture enthalpy with respect to temperature at constant pressure (CP) (J/KMOL)
DVMX - Partial derivative of mixture molar volume with respect to temperature at constant pressure (M3/KMOL-K)
Analytical derivatives are used to compute DVDP, DHMX, DVMX. But if the option Use analytical derivatives is disabled in the setup Simulation Options Calculation form, numerical perturbation is used instead. For example for DVDP:
PDEL = 1.001*P for Liquid phase
PDEL = 0.999*P for Vapor phase
DVDP = (VMXp - VMXdp)/ (P-PDEL)
where:
PDEL - pressure perturbation (Pa)
VMXp - Molar Volume at system pressure (M3/KMOL)
VMXdp - Molar Volume at PDEL pressure (M3/KMOL)
Liquids:
If the liquid volume model (e.g Rackett) is not a function of Pressure, i.e. DVDP=0, CV is set equal to CP and CP/CV=1.
Keywords: CP/CV
CPCVMX
CPCV
COMPR
compressor
isenthalpic
References: None |
Problem Statement: Under Input | Components, I noticed the listed Types are not correct. How do I fix these? | Solution: In order to modify the component Types after they have been added to a file, we can use the Aspen Flare System Analyzer (AFSA) Spreadsheet Export/Import Feature:
In AFSA, go to File | Export Case In
Choose .mdb | .xls | .xml format
Click Browse
Navigate to the desired location, provide a File name, change the Save as type to Excel (.xls) and click Save
Ensure Clear all existing data before export is not checked and click Next
Choose Create a new export definition file and click Next
Under Objects | Components, check the box for Export this type of data
Go to the Field Details tab and check both Name and Type in the Export column, and then click Next
Choose whether or not to save the export definition file and then export the data (options 1 or 2, not 3)
Once the export process is complete, navigate to the location decided in Step 4 above and open the Excel .xls file
In the Type column, change the values using the key provided below. Note you enter one of the twelve values (0-11), not the description (HC, Misc, etc.). Do not alter the component names.
Save the Excel file and then close it.
Return to AFSA and go to File | Import Case
Click Browse
Navigate to the location from Step 4 above, change the file type to .xls, select the Excel file and then click Open
Ensure Import As New is not checked and click Next
Choose Create a new import definition file and click Next
Repeat Steps 7 - 9 above, except this time the options will display import instead of export
Once the import process is complete, return to the Components form and confirm the types have been correctly modified
Component Types Key
0 - HC
1 - Misc
2 - Amine
3 - Alcohol
4 - Ketone
5 - Aldehyde
6 - Ester
7 - Carbacid
8 - Halogen
9 - Nitrile
10 - Phenol
11 - Ether
Note that before adding components to a file, you have the opportunity to change the default component types using the Components DataBase Editor.
Keywords: Components, type, spreadsheet, export, import
References: None |
Problem Statement: I am receiving the following errors when evaluating my project. How do I resolve these errors?
FATAL> 'W1 - 2' CWF INPUT IS NOT ALLOWED FOR PRIME CONTRACTOR RUN
FATAL> 'W1 - 2' WORKFORCE DATA ERRORS PREVENT PREPARATION OF PROJ ESTIMATE | Solution: There are two primary reasons for these errors. Both involve setting up multiple contractors with either of the two situations below.
The first reason is the option to Suppress default equipment/area/project bulks is set to M. Please see this KB for more info:
https://esupport.aspentech.com/S_Article?id=000080420
The second reason is the file was set up to Include pipeline areas. This choice is made when starting a new file. A project which includes pipeline areas does not allow for multiple contractors.
To confirm if the file includes pipeline areas, go to Project Basis View | General Project Data | Allow Pipeline Areas input. If there is a P, then that indicates the file includes pipeline areas and therefore will not allow for multiple contractors, thus giving the error when evaluating.
Keywords: Allow, pipeline, areas, error, multiple, contractors,
References: None |
Problem Statement: Can the length of my file name cause the Reporter error shown below?
Aspen Economics Reporter
Cannot open database
<users file name> requested by the login. The login failed.
Login failed for user '<user>'.
Reporter application will close. | Solution: Yes, it is possible to receive the error above if the file name is too long (too many characters). Please try shortening the file name, saving and then re-running the report again.
If the issue persists, then please contact AspenTech Support.
Keywords: Reporter, error, login, failed, file, name, length
References: None |
Problem Statement: How do I resolve the following error when I run a heat exchanger that involves the heating of a cold stream (water) to the above supercritical condition in using Aspen Plus?
I get the following errors:
** ERROR
DEW POINT CALCULATIONS CANNOT BE PERFORMED BECAUSE THE SPECIFIED PRESSURE
IS ABOVE THE CRITICAL PRESSURE. FLASH FAILED FOR COLD STREAM DURING DEW
POINT CALCULATION. SUBSEQUENT ZONE ANALYSIS AND AREA CALCULATION
ARE HIGHLY SUSPECT
** ERROR
BUBBLE POINT CALCULATIONS CANNOT BE PERFORMED BECAUSE THE SPECIFIED
PRESSURE IS ABOVE THE CRITICAL PRESSURE. FLASH FAILED FOR COLD STREAM
DURING BUBBLE POINT CALCULATION. SUBSEQUENT ZONE ANALYSIS AND AREA
CALCULATION ARE HIGHLY SUSPECT | Solution: The error messages are valid since the stream is being flashed as Vapor-Liquid.
In order to get rid off the error messages, you will need to change the valid phases of the Cold stream in the Heat Exchanger block to Vapor-only to avoid the Vapor-Liquid flash and prevent the error. This makes sense if you are above the critical pressure already.
This setting is on the Heat Exchanger | Options | Flash Options sheet.
Keywords: Heat Exchanger, Cold Stream, Aspen Plus
References: None |
Problem Statement: How to get rid off the following error message The embedded Excel object could not be located when launching a simulation file in Aspen Plus? | Solution: If the case file is saved as a backup (.BKP) file, it should have an associated .APMBD file which is the file Aspen Plus uses to store the embedded excel file.
When the following error message appears, it means that the .APMBD file is missing.
In order to get rid off the above error message, you will need to locate the .APMBD file and put it in the same directory as the .BKP.
If saving as a compound (.APWZ) file, the reference file and any associated files such as the .APMBD with the Excel Calculator will be saved as one zipped file.
Keywords: Excel, apmbd, Aspen Plus
References: None |
Problem Statement: After installing the Matrikon OPC Simulation Server the following error shows up when launching either GDOT Web Pages (GDOT-Simulation and GDOT-Online) or the Aspen Unified Web Page:
HTTP Error 500.0 – Internal Server Error
Calling LoadLibraryEx on ISAPI filter “C:\Windows\Microsoft.NET\Framework\v4.0.30319\\aspnet_filter.dll failed
Installing the Matrikon OPC Simulation Server also installed the following programs as prerequisites:
Microsoft.NET Framework 1.1 Service Pack 1
Visual C++ 2008 x86 Runtime v9.0.30729.6161
Visual C++ 2008 x64 Runtime v9.0.30729.6161
These could also be associated with problems with the Internet Information Server on the machine. | Solution: To solve this issue, it is necessary to perform the next procedure:
Go to the Internet Information Services (IIS) Manager, click on the server and double click on ISAPI Filters.
Locate the filter which is associated with the Executable C:\Windows\Microsoft.NET\Framework\v4.0.30319\\aspnet_filter.dll
You will see that there is a double backslash, right-click on Edit for the specific filter and delete one of the backslashes like in the images below:
Click on OK and then refresh Aspen Unified web page or any of GDOT web viewers to confirm theSolution worked well.
Keywords: Aspen Unified, GDOT web viewer, Matrikon OPC Simulation Server
References: None |
Problem Statement: Sites with more than one Aspen InfoPlus.21 server may wish to view tag data from these remote data sources in their local InfoPlus.21 client applications (e.g. Aspen Process Explorer, SQLplus, aspenONE Process Explorer or the InfoPlus.21 Administrator.) To do this, the system administrator or user must first configure the ADSA to recognize these remote data sources. | Solution: Users may configure remote data sources on their own PC as a user defined data source (which will only be available on that client PC.) The option to configure public data sources (accessible to all users) is only available on the ADSA server.
To search for ADSA servers on the network, go to Start-> Programs-> AspenTech-> Common Utilities-> ADSA Client Config Tool. The Directory Server field lists all the ADSA servers on the network.
To configure a new data source:
Go to Start | Programs | AspenTech | Common Utilities | Data Source Config Tool.
Click on the Add button in the Aspen Public Data Sources dialog box.
In the New Data Source dialog box, type in a name, preferably something recognizable, such as the nodename of the remote IP.21 server.
Next, click on the Add button to add the required ADSA service components.
From the list of ADSA service components, first select Aspen DA for IP.21. You will be prompted for a Host Name and Group Number. The Host Name should be the nodename of the remote IP.21 server, and the Group Number should be 200. Click OK and the Aspen DA for IP.21 service component will be added into the Configured Service list.
Similarly, add the following service components: Aspen Process Data (IP.21) and Aspen SQLplus service component. For the SQLplus service component, you will need to enter the Port Number of the SQLplus service from the NT services file (usually 10014).
If you are using Aspen IP.21 Process Browser, you must add the Aspen process Subscription (Generic) service component also. The Port Number in this case is obtained from the ipgcsi service in the Windows Services file (usually port # 10013).
When you have finished adding all the required service components, click on the OK button to exit the New Data Source dialog box.
You may continue to add other data sources (pointing to the other IP.21 servers) by repeating the instructions above. Be sure to change the Host Name fields to the nodename of the correct IP.21 server.
After adding all the IP.21 data sources, click OK to exit the Aspen Public Data Sources dialog box.
To make the changes take effect immediately, you will need to restart the Aspen Data Source Directory service.
If you open Process Explorer, you will now see a list of all the IP.21 servers (data sources) by clicking on the Data Source field of the tag in the legend.
If you have installed ADSA on the other IP.21 servers, you will have more than one ADSA server on the network. You can go ahead and configure the ADSA settings for the other remote servers by following the same procedure above. If different data sources are added to different ADSA servers, client users must take care to select the correct ADSA server in Process Explorer (from Tools menu-> Options, ADSA tab).
Important note:
You must be able to resolve the nodenames of all the remote servers on the network for this to work. A good way to check this is to do a reverse lookup (c:\> nslookup servername), or simply add the nodename and IP address into the Hosts file.
If database security has been implemented on the remote IP.21 servers, you need to be a member of the local roles to read/write to the remote database.
Keywords: data source
multiple
remote
ADSA
user data source
References: None |
Problem Statement: Should the DMCplus Controller fail to converge to a | Solution: an LP Internal Error message is issued and the controller sheds (goes open-loop)?
In many cases, prior to shedding, the controller may issue an LP ExtendedSolution Warning (without shedding). This warning should already prompt the user to investigate:
* conditioning the controller's model as shown inSolution 112044, and
* increasing the LP iteration parameters, and
* adjustment of certain LP tolerance and weighting parameters.
ThisSolution suggests LP parameters that can be adjusted in the CCF to allow for additional iterations. In addition, LP tolerance and weighting parameters that can be adjusted are also shown.
Solution
Once the LP Internal Error messages are observed, the user should investigate conditioning the model (.mdl) as shown inSolution 112044. ThisSolution addresses the most common source of this error message namely model ill-conditioning. You can use either the GAIN MATRIX ANALYSIS tool within DMCplus Model to perform the conditioning (first available in AMS6.0), or the DMCplus Desktop utility ModelSVD.exe.
In certain cases, the controller model is not ill-conditioned, however the choice of Engineering units may lead to an excessively high ratio of (biggest gain ratio)/(smallest gain ratio). This potential problem is also
clearly marked by the Gain Matrix Analysis tool. The user may change Engineering units by using an input transformation to reduce this ratio to a more acceptable value (suggested ratio say less than 10e+4).
Should the conditioned model still invoke the LP Internal Error or ExtendedSolution messages, then its possible to increase the maximum allowable LP iterations as shown below. Thereafter, if necessary, its possible to relax the LPSolution tolerance parameters as shown below.
PARAMETERS GOVERNING LP ITERATIONS
The CCF tuning parameters that govern the allowable primary and secondary iterations respectively are: MXLPIT and MXRFAC. Both are located in the CCF's Configure section.
The default values and suggested increases are shown in the table below.
CCF Parameter
(Configure Section)
definition of parameter
default
adjusted value
MXLPIT
Max Primary LP iterations
#MV * 100
times 1.5 or 2
MXRFAC
Max Secondary LP iterations
2% of MXLPIT
times 1.5 or 2
You need to save the CCF and reload the online controller for these changes to take effect.
PARAMETERS GOVERNING LPSolution TOLERANCES AND WEIGHTING
The following parameters govern the LP solver'sSolution tolerances and their default and adjusted values are shown in the table below.
CCF Parameter
(Configure Section)
definition of parameter
default
adjusted value
worst case value
EPSCST
LP matrix cost tolerance
1.00E-08
1.00E-07
1.00E-06
EPSDV0
divide by zero tolerance
1.00E-08
1.00E-07
1.00E-06
EPSFEA
steady stateSolution feasibility
1.00E-05
1.00E-04
1.00E-03
Similarly, you need to save the CCF and reload the online controller for these changes to take effect.
Keywords: DMCplus
LP Internal Error
LP Extended
References: None |
Problem Statement: AspenTech has successfully tested 500 filesets per repository and is aware of customers using even higher numbers (up to 600).
Whether it is to work around the 500 fileset limit or some other technical reason, we are aware that sometimes customers want to split an existing repository. This | Solution: addresses this issue.
Solution
Here are the steps to split a repository:
Let's suppose you have 20,000 tags in your database and you want to take 10,000 of them and point to a new repository without losing old history for any of them :
You will initially need enough disk to hold a copy of all current filesets in the repository.
a. COMPLETE A BACKUP FIRST (INCLUDING THE RECOMMENDED METHOD OF BACKING UP BOTH ACTIVE AND NON-ACTIVE FILESETS AS DESCRIBED IN ASPEN INFOPLUS.21 RELEASE NOTES)
b. Create another repository.
c. Create same number of filesets in the new repository as in the old repository.
d. Set size of each fileset in the new repository to be the same as the old repository
e. Use the Admin tool to stop the current repository.
f. Copy (normal windows copy) the contents of 'all' filesets (arc.dat, arc.byte, and arc.key files) of old repository into corresponding filesets for new repository.
g. Stop/Restart IP.21 to get it to recognize the new repository and new filesets. So at this stage we have 2 repositories covering the same tags for the same period in history
h. Modify 10,000 of the tags using 3 SQL statements.
o Turn Archiving Off
o Point records to new repository
o Turn Archiving back on.
Because the filesets were copied to the new repository, all tags can still see old history. Now because there are only 10,000 tags per repository, it should take twice as long to shift at the 1 gig maximum size.. Eventually, once enough fileset shifts have taken place, if you do not need to see so far back in history, you can reduce the number of filesets per repository, or the size of each fileset.
YOU MIGHT WANT TO TRY STEP -(H) ON ONE TAG FIRST, JUST TO DEMONSTRATE TO YOURSELF THAT YOU CAN STILL SEE ALL OLD HISTORY.
Keywords: Large
Split
Repository
Filesets
Record
References: None |
Problem Statement: Install Package to enable Hybrid Modelling in V10 and V11 of Aspen Plus and Aspen HYSYS | Solution: Hybrid Modeling, new in aspenONE Engineering V12, combines AI and First Principles to create a comprehensive, accurate model more quickly without requiring significant expertise and deployed as a sensor or unit operation model in Aspen Plus or Aspen HYSYS. Hybrid Models are built using the Aspen AI Model Builder SaaS application – contact your AspenTech rep to get a login. Hybrid Models can be deployed to V10 and V11 of Aspen Plus and Aspen HYSYS after applying a hybrid model install package. Download the package, attached below, and run the install setup.exe. The package includes documentation describing best practices on using hybrid models in V10 and V11 of Aspen Plus and Aspen HYSYS.
Keywords: Hybrid Model, Extension, USER2, AI Model Builder, Install Package
References: None |
Problem Statement: KB000063468 describes about an example script created using Aspen SQLplus to return the last known Good value of a single tag historized in Aspen Infoplus.21.
Customer may want to get the last Good value for multiple tags. This KB article provides a sample script to return the last Good value for the tags based on a tag list. | Solution: 1- Download attached Temp.zip file.
2- Un-zip Temp.zip and copy the Temp folder to C Drive on the Infoplus.21 server.
3- Update the TagList.txt to include your tag names
3- Launch Aspen SQLplus query writer on IP.21 server and open main.SQL at C:\Temp\Query
4- Execute the main.SQL from query Writer
5- Open the output file at C:\temp\Output
Note: The default period is the latest 2 hours. You may adjust as the requirement.
Keywords: None
References: None |
Problem Statement: How to place “Gap” at Cross point? | Solution: 'Gap' is typically used to show that crossing connection are unrelated(i.e. piping system cosmetic reason)
Here's a quick step to place an in-line gasp on a connection:
1. On the Gapping toolbar, ensure that ‘Standard’ is selected for the Size (this is the default).
2. Select a connection. Two new features now appear, on or alongside the selected connection:
A red square below, which appears at the nearest cross point to the cursor
Either the Gap icon or Heal icon below , by the cursor
Gap: Heal:
3. Left-click to toggle between opening and healing the gap. The Gap and Heal icons help you keep track of your progress:
With the Gap icon visible, left-clicking creates a gap
With the Heal icon visible, left-clicking heals the gap
As an example, the picture below shows an open gap. The Heal icon is visible, so when you left-click, the gap is healed.
Keywords: None
References: None |
Problem Statement: The Fire Heater will heat all the radiant zone streams equally using energy from the combustion, to distribute all energy from the combustion; set a steady state efficiency percentage. | Solution: In a Fire Heater Block Design | Parameters | Steady State Parameters, Efficiency must be specified.
If the efficiency is 0, then all the energy from the combustion will go into heating the flue gas. If the efficiency is 100%, then all the energy will go to heating the radiant zone process streams and the flue gas stream will not be heated.
Key Words:
Fire Heater Efficiency, Steady State Efficiency
Keywords: None
References: None |
Problem Statement: When setting up Reducing Gas Generator block specifications, Air Stream and Burn Stoichiometry [%] are required parameters, yet, they cannot be specified together. If both values are entered the Reducing Gas Generator block will display the message Cannot specify burn stoichiometry and air flow. | Solution: The Burn Stoichiometry [%] field is linked to the flowrate of the Air Stream connected to the Reducing Gas Generator. If the flowrate of the Air Stream is specified, a calculated value appears in this field. If you specify the Burn Stoichiometry [%], HYSYS calculates the flowrate of the Air Stream accordingly. The default value is 90%.
Key Words:
Reducing Gas Generator, Air Flowrate, Burn Stoichiometry
Keywords: None
References: None |
Problem Statement: The Aspen Exchanger Design Template (EDT) can only be crated from the Aspen Exchanger Design and Rating standalone application. Files with this saved as *.EDT, are created to store geometry and design specification for an exchanger type.
This Template file is compatible to be imported into HYSYS and Aspen Plus. | Solution: To create an *EDT file follow the listed steps:
Properly enter or calculate the exchanger's geometry, make sure results match what is wanted.
Go to File ribbon | Save As, locate where the file will be saved.
Under the Save as type, select from the options Aspen Exchanger Design Document Template (*.EDT)
Key Words:
EDT, Save EDT, EDT file, Create EDT file
Keywords: None
References: None |
Problem Statement: The buckling distance value is directly calculated from specifying stiffening rings in the Shell Cylinder geometry specifications, this can be reviewed on the Results | Input Summary | Cylinders/Covers/Flanges | Shell Cyl. | Stiffening Rings. | Solution: If the Stiffening Rings options is chosen as yes and the number is properly entered in Exchanger Geometry | Shell | Shell Cylinder, is well know that If the user enters the number of stiffening rings and doesn’t enter the shell cylinder length or the shell cylinder length for external pressure, the buckling length will be set to the number of stiffening rings in meters. Due to this bug, the user can only enter the number of stiffening rings if they input the shell cylinder length.
Alternatively, the user can enter the buckling length, for both front and rear head.
Key Words:
buckling distance, stiffening rings, EDR mechanical, shell and tube mechanical
Keywords: None
References: None |
Problem Statement: Water is an universal solvent and can dissolves almost every solute. Certain factors can influence the solubility of a substance such as temperature and pressure.
Solubility is the new bond formation between the solute molecules and solvent molecules. The objective of this example library is to use the Sensitivity Analysis in Aspen Plus varying the temperature in the SEP101 block to determine maximum concentration of propanol in water | Solution: In Aspen Plus create a Sensitivity Analysis that varies the three phase separation tank's temperature (SEP101) to review the change of concentration of propanol in both outlet liquid streams.
Key Words:
Sensitivity Analysis, Liquid Liquid Equilibrium, Liquid Concentration, Propanol in Water, Aspen Plus
Keywords: None
References: None |
Problem Statement: The objective of this example library is to set a Direct Fire Heater for a Sulfur Recovery Unit and adjust the outlet gas temperature varying the Fuel mass flow. | Solution: Use the adjust block to fix outlet temperature conditions of the product gas of a Direct Fire Heater block, varying the fuel flow into the unit operation.
Key Words:
Direct Fire Heater, Sulfur Recovery Unit, Aspen HYSYS
Keywords: None
References: None |
Problem Statement: The objective of this example library is to set a Selective Oxidation Converter for a Sulfur Recovery Unit in Aspen HYSYS and adjust the outlet gas temperature varying the air mass flow. | Solution: Use the adjust block to fix outlet temperature conditions of the product gas of a Selective Oxidant Converter, varying the air flow into the unit operation.
Key Words:
Selective Oxidant Converter, Fixed temperature, Sulfur Recovery Unit, Aspen HYSYS
Keywords: None
References: None |
Problem Statement: The objective of this example library is to use the Fire Heater unit and Recycle block in Aspen HYSYS to simulate a simple fire heater with recirculation of the product from the radiant section into the convective section. | Solution: The recycle block will help to transfer the results get from radiant calculations going to the convective feed into the fire heater block and performs an analysis for both radiant and convective section of the fire heater for the same process stream.
Key Words:
Fire Heater, Convective recirculation, HYSYS
Keywords: None
References: None |
Problem Statement: Shell and tube heat exchangers have many modifications of the basic configuration which can be used to solve special problems in the process industries, adding extra supports to baffles configuration to help prevent vibration issues, also to match specific bundle arrangement. | Solution: The objective of this Example library is to match the following support arrangement for a single segmental baffle specification (supports in red)
We will achieve the objective by defining the following:
Number of Supports between central baffles.
Number of supports at front head end space
number of supports at rear head end space.
Key words:
Baffle Supports, Shell and tube baffle supports,
Keywords: None
References: None |
Problem Statement: A contact condenser or spray condenser is used for the condensation of humid water vapor by direct contact with water. The inlet water is at a temperature less than the dew point of air in the chamber. | Solution: This example library shows how to simulate a contact condenser in aspen HYSYS, with a mixer block that will feed both vapor and cooling water; a two phase separator block, a cooler for the non-condensables leaving the separator, and a adjust block that will help us fix the amount of vapor leaving the tower and vary the cooling feed into it.
Key Words:
Contact Condenser, Spray Condenser
Keywords: None
References: None |
Problem Statement: How to restore or backup a Workspace in ABE V12? | Solution: In V11 and older versions, the tool used was AZ Backup tool. Follow the next article steps for those versions.
Since V12, the Backup & Restore Utility is the new tool for backup and restore. It can be easily accessed from the ABE Administration.
In the Administration tool go to Action menu and select Backup and Restore Tool.
In this same tool find the Help section by clicking on the question mark icon.
Steps to create a Backup
Open the Backup and Restore Tool.
Check Backup Workspace option and click on Next.
Select the Workspace to backup from the list and click on Next.
Select the location where the backup file will be saved and click on Next.
Finally, click on Backup button. Messages about the process status will be displayed in the textbox.
The extension of the Backup file is *.abebkp.
Steps to restore
Create a new Workspace. This workspace will contain the data of the restored one.
Open the Backup and Restore Tool, check Restore Workspace and click on Next.
Select the created Workspace’s folder. The Workspaces folders are stored in the path: C:\AspenZyqadServer\Basic EngineeringXX.X\Workspaces\<Name of the Workspace>
Note: XX.X refers to the version of ABE: 38.0 for V12, 37.0 for V11 and 19.1 for V10.
Specify the backup file path and click on Next.
Finally, click on Restore button. A warning message will appear telling the data of the specified Workspace will be overwritten. Click on Yes.
Messages about the process status will be displayed in the textbox.
Close the Backup and Restore tool before opening the restored Workspace.
Keywords: Workspace, Backup, Restore, ABE, V12
References: None |
Problem Statement: How do I get the composite curves related to the heat exchangers in my Aspen HYSYS simulation? | Solution: You cannot get the composite curves directly from Aspen HYSYS, you need to do it using Aspen Energy Analyzer.
1. Open Aspen Energy Analyzer (Same version as Aspen HYSYS)
2. Create a new project
3. Click on Data Extraction from Aspen HYSYS and the Extraction Wizard opens. Click next
4. Select the Simulation File to Import
5. On step 2 of 7, you can segment your streams to improve accuracy, it may take a long time to extract the information. Aspen HYSYS case will open. Click next
6. On step 3 of 7 you can select the flowsheet you want to work with. All the sub-flowsheets will appear at this moment.
7. On step 4 of 7 Aspen Energy Analyzer will choose from the utility database (based on temperature levels), on step 5 you can add more utilities to use.
8. Step 7 deals with the economic data, if you don't have any real data, you may leave the default values.
9. Click Finish
You will have one scenario in which you can select from the drop-down list to see the Composite Curve as you can see below.
Note: You can delete the streams you don't want to be added to the heat integration directly in Aspen Energy Analyzer. (Table at the bottom of the above screenshot)
Keywords: Composite curves, heat exchanger network
References: None |
Problem Statement: Is there a difference if I use the Case Study utility in Aspen HYSYS or Multi-Case with one single core? | Solution: Yes, the performance in Multi-Case is better because of the nature of the solver and the additional tools that are only available in Multi-Case such as the ability to select a combination of Nested and Discrete analysis, which is not possible in HYSYS.
Keywords: Multi-Case; Single Core; Performance; Case-Study;
References: None |
Problem Statement: Warning message when importing new price set in MBO through EIU is wrongly worded | Solution: When using EIU to load a new price set into MBO, users might get the below warnings about the cells containing Product Purchase prices.
In this case, there is no issue with the EIU validation or the default template - both perform correctly. But the warning message should be worded as:
Cell in XXX worksheet PriceSet is invalid. It should be '[Tag] [COMPONENT] [PURCHASE|SALES|BLEND] ' or '[Tag] [PRODUCT] [SALES] ' or '[Tag] [TANK] [INCR|DECR]’
Once the cell entries are corrected as per the correct warning message, the warnings would be resolved. This will be addressed in the upcoming V14 release.
Keywords: None
References: None |
Problem Statement: User calc trims the decimals and outputs only whole numbers to the event dialog box | Solution: If user has the Crude Receipt user calc sheet defined (CRDRCPT) and let's say that the only thing the user calc does is copy the inputs to the outputs. Now the Crude Receipt event has 3 tanks T009, T001, T003 defined to receive the receipt volume. Lets say the user has defined that T009 will receive 117.283, T001 will receive 113.244 and T003 will receive the balance. The user calc will trim the volumes to whole numbers.
The User Calc Sheet setup:
After transferring to the event:
The workaround currently is to use the automation in addition to the User calc to get the values into the event. This is targeted to be fixed in the upcoming V11, V12 patches and in the V14 release.
Keywords: None
References: None |
Problem Statement: What does all the locking information mean in SLM License Manager? | Solution: SLM Locking information is required when generating a license. SLM lock codes are uniquely generated hexadecimal codes that are generated from each locking type. The license will be tied to a specific lock code on your computer, and this will prevent the license from working on a different computer. The SLM Lock Tool is located in the aspenONE SLM License Manager or you can download it from KB 136832.
AspenTech will select the locking type based on the business model. Typically, AspenTech uses:
1. Sentinel Hardlock: It comes from the dongle which we usually use to access dongle-type license file.
2. Hostname & IP Address: It is your PC/server’s computer name and the computer’s static IP address,
3. Ethernet Address: It comes from computer’s network card.
4. Disk ID: It comes from computer’s hard disk.
Keywords: SLM license manager
Locking info
References: None |
Problem Statement: Why does the error “Cannot set value for field with no defined case” appear when editing a field? | Solution: This error appears when the field has not a case assigned. Therefore, a case must be selected for that field or fields.
See the screenshots below to review the defined case of a field.
In Excel Datasheet Editor:
In Legacy Datasheet Editor:
To know how to define and display multiple cases information, check this article:
How do you define a datasheet to show data from different cases?
Keywords: Field, Datasheet, Datasheet Editor
References: None |
Problem Statement: How to debug a dll KB script using Visual Studio? | Solution: The attached guide will show users the workflow to utilize Visual Studio to debug Aspen Basic Engineering (ABE) custom dll scripts. The document focuses primarily on establishing the code interaction with ABE process rather than explaining how to debug Visual Basic code. The document assumes users have some familiarity with Visual Basic coding and are comfortable with debugging code.
Key Words
Debugging, MS Visual Studio, attach process, ABEversion.info, ZyqadStrongNamingKey.snk
Keywords: None
References: None |
Problem Statement: When running the batch file InstallBatchAPC.bat, it fails with the error message: Error: Invalid Syntax. Type REG ADD /? for usage
Root Cause
The InstallBatchAPC.bat utility command line that produces this error is shown below:
REG ADD %REGKEYIP21%\%IP21Version%\group200\RegisteredTasks\TSK_DBCLOCK /v \PAR /t REG_EXPAND_SZ /d %DBCLOCKPAR% /f >> BatchAPClog.txt
This is a rare defect (ID 630068) where the batch script fails in the REG ADD command on non-default drives when there are two sub-directory levels. For example, when using these non-default directories:
Non-default 64-bit: D:\Program Files (x86)\AspenTech (x64) AND Non-default 32-bit: D:\Program Files (x86)\AspenTech
Non-default 64-bit: D:\Program Files\AspenTech AND Non-default 32-bit: D:\Program Files\AspenTech (x86)
Re | Solution: If you are experiencing this uncommon issue, please submit a Support ticket and we can provide you with a hot fix for the batch file: https://esupport.aspentech.com/s_casesubmit
The fix is to include triple-double quotes () around the InfoPlus.21 64-bit directory location.
For example: SET DBCLOCKPAR=DOUBLE 20000000 SNAPSHOT=%IP21g200Dir%\InfoPlus21.snp
The target fix for this Defect 630068 is V14.
Keywords: Batch, APC, InstallBatchAPC.bat, fail, error
References: None |
Problem Statement: When working with traditional ramps with imbalance (MXNIMB>0), the upper and lower imbalance limits are pinched at 0.0 to force the ramp to be “balanced” every cycle, forcing the SS optimizer to consume one degree of freedom. Moreover, in PCWS these values are always zero which is the designed behavior.
However, the upper and lower imbalance limits are always being calculated to determine if there is an imbalance big enough to turn the controller off. In this sense, these limits are not zero. This KB provides a procedure to visualize those values in RTE controllers. | Solution: TheSolution is to implement two input calculations in the controller for the high and lower imbalance limits. Two user defined entries will be created to store these values each cycle.
========== UpperRampImbalance=============
'Calculation to calculate the UpperRampImbalance when MXNIMB>0 and store it on a user entry
'Cycle Period in Minutes
FREQM = FREQ/60
'RHORIZ in minutes. RHORIZM
RHORIZM = RHORIZ/60
'Logical Statement. Activate only when working with traditional ramps with imbalance. Otherwise use -9999
If (MXNIMB > 0 and Current < HL)then
'Calculate the high limit when meas is within limits
Upper_limit = (HL-Current)/(RHORIZM/FREQM)
elseif (MXNIMB > 0 and Current > HL) then
'Calculate the high limit when meas is above high limit
Upper_limit = 0
else
Upper_limit = -9999
end if
============== LowerRampImbalance===============
'Calculate the LowerRampImbalance when MXNIMB>0 and store it on a user entry
'Cycle Period in Minutes
FREQM = FREQ/60
'RHORIZ in minutes. RHORIZM
RHORIZM = RHORIZ/60
'Logical Statement. Activate only when working with traditional ramps with imbalance. Otherwise use -9999
If (MXNIMB> 0 and Current > LL) then
'Calculate the lower limit when meas is within limits
Lower_limit = (LL-Current)/(RHORIZM/FREQM)
elseif (MXNIMB> 0 and Current < LL) then
'Calculate the lower limit when meas is below the low limit
Lower_limit = 0
else
Lower_limit = -9999
end if
=========PCWS Detail page for the ramp CV==========
Keywords: …Ramp, upper imbalance limit, lower imbalance limit, Maximum Number of imbalances.
References: None |
Problem Statement: Infoplus21 can be played as OPC UA server when IP21OpcUAServerHost.exe is running at backend. Are there any connections limit for Infoplus.21 OPCUA Server ? | Solution: The number of connections supported by IP21 OPCUA Server are defined in the C:\ProgramData\AspenTech\InfoPlus.21\db21\group200\tsk_opcua_server.opcua.config.xml file.
<MaxSessionCount>100</MaxSessionCount>
The default value is 100, which means that maximum 100 client connections will be supported.
Key Words:
OPCUA
OPC UA
tsk_opcua_server.opcua.config
Connection limit
Limit
Keywords: None
References: None |
Problem Statement: CIM-IO Interface Manager in v11.0.1 will allow the user to specify the OPC HDA Server and the maximum history recovery time period, in minutes, for a selected instance of a CIM-IO for OPC interface.
This KB article provides an example for using CIM-IO for OPC DA/HDA to get data from one Aspen InfoPlus.21 to another. | Solution: In this example, we will set up three boxes. Source IP.21 server as OPC DA/HDA server, CIM-IO server and Target IP.21 server.
1- Configure the source IP.21 server to simulate OPC-HDA server.
InfoPlus.21 OPC-DA layer only supports reading current data. However, it is possible, using third-partySolutions, to access historical data.
1) Install MatrikonOPCExplorer.exe using default options first.
2) Install MatrikonOPC_IP21_Web.exe using default options and set up Matrikon.OPC.IP21.1 server.
a) Launch MatrikonOPC Server for IP21
b) Right-click Server Configuration | select Define New....
c) Select IP.21 Server and Click OK.
d) Type a name such as IP21 and Click OK.
e) The server has been created successfully.
3) Verify the new created server from Matrikon OPC Explorer
2- Configure CIM-IO for OPC DA/HDA interface on the CIM-IO server
1) Launch Cim-IO Interface Manager (CIM)
2) Highlight Cim-IO Interfaces and Click on Create a new One
3) Select Cim-IO for OPC interface and Click on Next
4) Type Interface name, Interface description etc..
5) Discover the OPC DA server | select Matrikon.OPC.IP21.1 | click on Next
6) Click on OPC HDA tab | Click on Discover | Select Matrikon.OPC.IP21.1 | set History Recovery (mins) | Click on Apply Changes | Restart the interface.
3- Configure the logical device on Target IP21 server using Cim-IO IP21 connection manager(CCM).
4- Configure the transfer record
Important note: History recovery is supported for all types of IoUnsol records (IoUnsolDef, IoLongTagUnsdef, IoLLTagUnsDef) but NOT supported for any types of IoGet records (IoGetDef, IoLongTagGetdef, IoLLTagGetDef).
5- Verify the data collection is working for the tags.
6- Simulate communication issue in between OPC server and CIM-IO server. i.e. stop the interface in CIM for few mins. start the interface and verify the History recovery.
Key words:
OPC
HDA
OPC HDA
History recovery
Keywords: None
References: None |
Problem Statement: There is a need to support history recovery from OPC Server (Classic) to cover instances when the data collection mechanism loses communication with the server. | Solution: CIM-IO Interface Manager in v11.0.1 will allow the user to specify the OPC HDA Server and the maximum history recovery time period, in minutes, for a selected instance of a CIM-IO for OPC interface.
(Note: Feature is considered to be disabled if the maximum history recovery time period is zero.)
After re-connecting to the OPC server (following network glitch or startup) , CIM-IO for OPC will:
Check a local data store to retrieve the last timestamp that was recorded.
Calculate the history recovery start time.
Ask the third-party OPC server for any process values that have been saved since the calculated history recovery start time.
Forward any retrieved history to the CIM-IO client (IP.21)
Please refer to KB000097855 - Example using CIM-IO for OPC DA/HDA to get data from one Aspen InfoPlus.21 to another. for more configuration information.
Key Words:
HDA
OPC
OPC HDA
CIM-IO for OPC
History recovery
Keywords: None
References: None |
Problem Statement: This Knowledge Base article provides guidance on how to manually insert and edit tag data to the Aspen InfoPlus.21 (IP.21) database from aspenONE Process Explorer. | Solution: The aspenONE Process Explorer application allows users to manually insert new data points for tags and also edit data points that have already been written to the database.Solution 1 of this article will illustrate how to write a new data point for a selected tag to the database.Solution 2 will show how to edit a data point that has already been written to the database.
1.
To begin entering a new value for a tag to the IP.21 database you will first need to create a trend plot that includes the tags that you would like to manually enter data for. Next, you will need to click on the column settings button found at the bottom right of the webpage. With the column settings popup open you can enable manual data entry for a Value.
With the Column settings properly configured as shown above you will now have the option to enter a value for any of the tags you have within your trend plot. Once you have manually entered a value you will then want to choose the timestamp that is appropriate for this value. 'Use current time' is the default setting shown but this can be changed by unchecking this selection and entering the desired timestamp for the new data point.
2.
If a user wants to edit a datapoint for a tag from aspenONE Process Explorer this can be done by selecting the tag and clicking on the 'Show detail' button shown below.
With the Show Detail popup open select the History tab and click on manual data entry. After selecting manual data entry you will then see the most recent dataset values for this tag. Values from the dataset will then be able to be edited. If the user would like to edit a value from the dataset that is older than the one shown, then the user will need to click on the Time window button to show values for the tag during past time periods.
Note: When using the manual data entry functions within aspenONE Process explorer to either enter new datapoints or edit existing ones the Status of these data points will show either Manual or Good/M within the IP.21 database repeat area for this tag.
Keywords: aspenONE Process Explorer
Manual
Data
IP21
References: None |
Problem Statement: Scattered around the InfoPlus.21 Installation manual you will see the following statements
Whenever history file sets are moved to a new system, their MAP.DAT file must also be moved.
Backups of history file sets without the MAP.DAT file could be useless, causing complete loss of historical data if the MAP.DAT file is lost.
This MAP.DAT file is absolutely necessary to access archives from a database and is specific to a snapshot. This file is not recreated if deleted. If deleted, you have to use a virgin snapshot. This file is not reproducible or rebuildable.
The Support Group has seen a few cases of historical data either being lost, or appearing corrupt, due to a user breaking one of the above rules. However, when you look in Windows Explorer, this file is really a very small file. This document attempts to explain the contents of the MAP.DAT file which on default installations is located in C:\ProgramData\AspenTech\InfoPlus.21\c21\h21\dat, as well as explain why it is so important to the History system. | Solution: First of all let's define a 'tag' as an InfoPlus.21 data record that contains a History Repeat Area. This includes history for Analogs, Discretes, Text, Logs, etc. Whenever a tag is created, the user needs to open the Memory repeat area, then define a History Repository, and finally turn on Archiving. The act of turning on Archiving causes the history system to interact with the Map.Dat file. Similarly when inserting into, or searching for, history, there are interactions with the Map.dat.
The Map.dat file contains a series of unique Templates, and each Template has an Index Number. Each Unique Template describes the structure of a Unique History Repeat Area. This structure is defined by the Definition record. For example, for any users that use Ip_AnalogDef, there will be one, and only one, Template that describes the repeat area of the 4 fields in specific order (and size) of Ip_Trend_Time, Ip_Trend_Qlevel, Ip_Trend_Qstatus and Ip_Trend_Value. This Unique Template will be assigned a Unique Index number -- let's use the Index Number '3' for Ip_AnalogDef purely for this example.
As mentioned, when archiving is turned on for a tag, the history system will look in the Map.dat. It is looking for a Template that matches the structure of the History Repeat Area of the tag they are creating. If that Template exists then the corresponding Index Number is written in the header of the data record itself. If the Template does not exist, then a new Template is created, a new Index number is assigned to it, and that new number is written to the header of the new data record. In the example above, any Ip_AnalogDef 'tag', that is collecting history, will have the Index number '3' in its header.
Therefore, when doing a history search on a 'tag', via tools such as Process Explorer or SQLplus, the software uses the Index Number to find the Template, to enable it know how to search the Filesets for the History Values. Obviously any mis-match will cause problems.
To demonstrate potential problems, let's imagine that we moved the snapshot and history to a different IP.21 system, but we did not move the Map.dat - instead we used the Map.dat already on the new IP.21 system. Now let's imagine that the Map.Dat on the new IP.21 also has an IP_AnalogDef Template, BUT its index number is '4' instead of '3'. We now do a history search.
Because we moved the snapshot, the 'tag' header says to go and look at Template #3, but, because we are using the wrong Map.dat, the #3 Template was not used for Ip_AnalogDef - it was another unique structure. This will cause totally unpredictable results with the likelihood of not being able to see any of the history for that tag.
Another problem example could be where you have a Production System and a Test System. On both systems you create two new definition records, CustomOneDef and CustomTwoDef with associated tags. However, on one of the systems you create CustomOneDef first and on the other you created CustomTwoDef first. You should be able to see that the Index Numbers in the two Map.dat files will be interchanged, meaning again that if you move Snapshot and History from one system to the other, you MUST also bring across the Map.Dat.
The bottom line is that the Index Number and History Repeat Area structure in the tag MUST match the correct Index Number and Template in the Map.Dat, and the only way that can be achieved is to take note of all the warnings listed above, and in the Installation manuals, and take good care of the Map.dat file.
Keywords: Map.dat
History
InfoPlus.21
References: None |
Problem Statement: You can use the Redefine Wizard to add a field to the fixed and repeat area of a record.
1. When you add a field to the history repeat area, what effect does it have on the history already on disk?
2. Is history rewritten?
3. Or does the redefine wizard modify the contents of map.dat? | Solution: The Redefine Wizard modifies the internal history retrieval code to support changing the structure of a history occurrence. Previous software releases had warned against changing the structure of a history occurrence because it would result in the loss of existing history. The Redefine Wizard accommodates the current history as long as the data type or size of the fields common to the two definitions are the same and the fields defined in the obsolete definition are also defined in the newer definition record.
1. When you add a field to the history repeat area, what effect does it have on the history already on disk? Answer: Nothing happens to the history that has already been stored.
2. Is it rewritten? Answer: No.
3. Or does the Redefine Wizard do something like changing the contents of map.dat? Answer: The redefine wizard creates a new snapshot. The map.dat is updated by loaddb when the new snapshot is loaded into memory.
Keywords: Redefine
Wizard
History
Custom record
References: None |
Problem Statement: This article describes the best practice to be followed for variable type settings, while configuring and deploying Aspen GDOT online applications | Solution: Note on variable type settings for Binary or Integer MV’s
Aspen GDOT projects deployed online are most of the time, configured to solve continuous optimization rather than a mixed integer optimization problem. However, even with a continuous optimization problem, it is a practice to often legitimately include Binary or Integer variables to provide information on current configuration.
For example, Binary variables are used to specify whether a connected upstream unit is running or not, while Integer variables are used to specify different states/ modes.
By default, a Binary Manipulated Variable (MV) in GDOT is specified as Type 2 and an Integer MV as Type 1.
However, when these these Binary or Integer MV’s are included for information purposes only (e.g., specify if a unit is running, etc.) then instead of specifying these MV’s as Type 2 or Type 1, it is recommended instead to specify them as Type 0 (same as a continuous MV).
In addition we should specify that these MV’s are OFF for the optimizer. This will ensure that
i) The optimizer uses the specified value as is (doesn’t try to use it as an MV for optimization)
ii) The continuous optimizer is invoked rather than the mixed integer optimizer.
This will avoid unexpected results.
Keywords: Best practice for continuous optimization
Mixed integer problem
Binary or Integer manipulated variable
References: None |
Problem Statement: Why do I get Error 1068:The dependency service failed to start, while starting Aspen APC DMC plus data service?
This error could appear for Aspen APC Web provider data service as well. | Solution: Error 1068 is a specific windows service error. There could be two reasons for this error:
1. When one of the dependent services is not started for the service which we are trying to start.
The Aspen APC DMC plus data service actually has two dependencies services, which are as below:
- ACO Utility server &
- Aspen APC Message Log service
Thus, to start Aspen APC DMC plus data service successfully, these two services have to remain started as well.
The list of dependent service will be available in properties of service-> Dependency.
Please ensure the dependent services are running. Start the services if not running already.
If the dependent service is not at all available under list of dependencies, it indicates that the installation did not take place completely.
Hence, theSolution is to un-install & perform a re-install of the Aspen software. After re-installation, make sure the dependent services are visible in dependency list and are running.
2. Another possible reason for Error 1068 is - Windows service account ID or password has been changed
When a Windows service account ID or password is changed, it will be mandatory to manually update the password for Aspen APC DMC plus data service with the new login id or password.
This also needs to be done for the dependency services and any other Windows service that uses the service account within the Control Panel.
To update the password, please follow the below:
Start à Control Panel à Services à Aspen APC DMC plus data service
Double click this service and update the new ID and password in the Log On As area. This procedure needs to be followed for all other Windows services that use the service account.
A reboot of server or restart the affected Aspen services will be required after the logon properties are updated.
Keywords: Error 1068:The dependency service failed to start
Error 1068
Aspen APC DMC plus data service
Windows service
Aspen Web provider data service
References: None |
Problem Statement: Which AspenTech application is appropriate to do pressure drop calculation? AspenTech has multiple options to do the pressure drop calculation. | Solution: The options to use hydraulics calculation in AspenTech are
1. Pipe Segment (Available in HYSYS, from Aspen Tech, No additional license needed except for OLGAS.). SPT group OLGA?s correlation is available to check pressure drop correlation for multiphase system.
Note: OLGAS is the steady-state flow correlation based on OLGA. OLGA is used for dynamic multiphase flow. Visit SPT Group website, for more information: http://www.sptgroup.com/en/Products/olga/OLGA-Steady-State/.
The Pipe Segment is used to simulate a wide variety of piping situations ranging from single or multiphase plant piping with rigorous heat transfer estimation, to a large capacity looped pipeline problems. It is the most simple unit operation available in industry and hence very easy to setup. It takes only few minutes to setup and do calculation for both steady state and dynamic system. Pipe segment checks for sonic flow if indicated by the option in the Calculation page of the Design tab.
Limitation: It is NOT recommended
to use where there is already specialty software available to do precise calculation. For example if you try to run in dynamics with a pipeline which run few hundred meter vertically. The results will be very different for such system. Such calculations are done for upstream operation and Aspen Hydraulics or OLGA is recommended to use to do Dynamic study.
to use for piping Network where flow is splitting or merging. When the flow splits (TEE) or merges (Mixer), correlation such as miller chart are needed to a pressure drop calculation across TEE or Mixer. Such pressure drop in Hysys pipe segment is ignored. Aspen Flarenet or Aspen Hydraulics should be used for such calculation. Also for flow in loops, it is easier to solve such network in Aspen Hydraulics or Aspen Flarenet.
to use when velocity is very high (close to sonic velocity ? Mach 1.0). Pipe unit operation does not model choking or advanced effects such as shock waves, momentum balances, and so on. To model choking and other advanced effects, use Profess, OLGA pipe extensions, or Aspen Hydraulics flow sheet in Dynamics. For steady state, Aspen Flarenet is used. Please note that for multiphase system, the sonic velocity (C^2 = DP/D(rho) at constant entropy) may be very low such as 20m/s. This is due to large variation in density for multiphase system.
Solver: The Pipe Segment can solve in either direction. TheSolution procedure generally starts at the end where the temperature is known (temperature is typically not known on both ends). HYSYS then begins stepping through the pipe from that point, using either the specified pressure, or estimating a starting value. If the starting point is the pipe outlet, HYSYS steps backwards through the pipe. At the other end of the pipe, HYSYS compares the calculatedSolution to other known information and specifications, and if necessary, restarts the procedure with a new set of starting estimates. The calculated pressure drop accounts for fittings (K = A+B * fT, DP = K* Rho V^2/2), frictional (Calculated by correlation such as Begg?s and Brill), and hydrostatic effects static head. The static head is independent of velocity and flow rate. For a constant geometry (pipe diameter) and no elevation change, there is no static pressure drop.
The Pipe also solves for the length if you provide one pressure, two temperature specifications, and the duty. Both length and diameter calculations can only be done for pipes with a single segment.
Caution: When two liquid phases are present, appropriate volume based empirical mixing rules are implemented to calculate a single pseudo liquid phase. Therefore, caution should be exercised in interpreting the calculated pressure drops for three-phase systems. Actual pressure drops can vary dramatically for different flow regimes, and for emulsion systems.
2. Compressible Gas Pipe (CGP): (Available in Aspen HYSYS ? No additional license needed) The CGP unit operation is primarily designed for transient calculations with streams. Steady state calculations have been implemented primarily for initialization of the Pipe State prior to transient calculations. Tip: The gas pipe model and the PF solver can not solve simultaneously, the gas pipe has to contribute some equations to the PF solver, and the way it does that can cause conflicts. To solve the Gas pipe move the P spec to the Product stream and the F spec to the Feed.The CGP calculates the parameters for the PF equation it contributes to the integrator by perturbing the end flows slightly; it uses a base step size of 0.05 kg/s for the perturbation, but increases this if the resulting flow gives a Mach number less than 0.001. In CGP if the cell length is 25.6m, so with a time step of 0.02s the velocity would need to be at least 1280 m/s to cover one of the cells, which is actually greater than the sonic velocity. Unfortunately this sort of effect is a problem with the fixed cell size configuration that HYSYS uses. Ideally the cell grid should be adaptive so that during the perturbation phase the cells can be made much smaller near the ends of the pipe. The Compressible Gas Pipe does not support an energy stream. CGP calculates the speed of sound and Mach number that is not available in Pipe segment.
3. PIPESYS: (Available in HYSYS, it is product from Neotec. Pipesys license is needed to run it). Visit Neotec website for more information. http://www.neotec.com/products/black_oil.html.
Developed by Neotec as an integrated pipeline flow model for AspenTech HYSYS, PIPESYS relies on Notec?s technology for single and multiphase flow modeling for pipeline hydraulics. PIPESYS can be licensed as an integratedSolution in Aspen HYSYS. PipeSYS is similar to HYSYS pipe segment except that here more flexibility to choose separate correlation for vertical, inclined pipe. PIPESYS can NOT run in dynamic mode.
4. Aspen Hydraulics: (Available in HYSYS, From Aspen Tech - Needs Hysys Upstream license). In order to install Aspen Hydraulics in 2006.5 you must install the Upstream Operations feature in the AES. If you don't have HYSYS upstream license, then you would not be able to use Aspen Hydraulics.
Aspen Hydraulics is used for modeling transient multiphase hydrocarbon flows in wells, pipelines, and other hydraulic components. TheSolution engine solves mass, momentum and energy equations for each phase using a one-dimensional finite difference scheme. Appropriate flow pattern maps and constitutive relationships are provided for wall and interfacial friction as well as heat transfer, and a model for multi-component phase-change is included. Pipeline and hydraulic network simulations can be solved in Steady State mode or Dynamic mode on a single network, with the ability to switch between the two modes and also switch between solvers. In Aspen Hydraulics at steady-state, the inlet stream must have an empty pressure, with all the remaining conditions specified, and the outlet stream must just have the outlet pressure specified. The streams connected to the pipe in the Hysdraulics sub-flowsheet keep all the information and the ones inside the sub-flowsheet just recompute, sometimes following the specs in the Transfer Basis tab of the sub-flowsheet.
It is used in steady state as well as dynamics. Like OLGA, it is also used to study the liquid build up as well as see the effect of launching pig. Some sample files are added to the sample folder (example: C:\Program Files\AspenTech\Aspen HYSYS V7.0\Samples\AspenHydraulics)
4. PipeSIM and PipeSIM Net (Available in HYSYS, From Schlumberger - Needs HYSYS Upstream license (From Aspen Tech). Note: Some people needed HYSYS Pipesim_Link license as well. Developers are investigating the problem).
Visit Schlumberger website for more information about the product: http://www.slb.com/content/services/software/production/pipesim/index.asp?
PIPESIM is a steady-state multiphase flow simulator used for the design and diagnostic analysis of oil and gas production systems. PIPESIM is used for one pipe and PIPESIM Net is used for pipe network. In HYSYS, for feed stream, user?s need to define the flow rate, temperature and composition. In PIPESIM / PIPESIM NET, the inlet stream must have an empty pressure, with all the remaining conditions specified, and the outlet stream must just have the outlet pressure specified.
Limitation: It is only steady sate solver to model hydraulics for oil well.
5. OLGA: (Available in Aspen HYSYS, from SPT group, OLGA uses HYSYS_Hydraulics license if it is available. It can also use HYSYS_OLGA_link (for backwards compatibility) if HYSYS_hydraulics is not available. It also needs Olga license (From SPT group)). Visit: http://www.sptgroup.com/Products/olga/
OLGA is simulator for engineering the flow of oil, water and gas in wells, pipelines and receiving facilities. It is used to study the well dynamics for a multiphase pipeline flow. It is used in dynamics only to study the liquid buildup in pipe line as well as used to study the effect of launching pig.
6. Aspen Flarenet: It is a separate product from Aspen Tech and widely used to design flare network. It is also used to model a vent system. It can handle multiphase system as well it can solve pipe network with much ease.
Solver: The ambient pressure is used as a boundary condition to calculate the back pressure for relief valve. It calculates noise and also checks for choked flow. The calculation is suitable for sonic velocity as well. Here kinetic energy (KE) is included for enthalpy balance that is very important to calculate the outlet temperate at the relief valve. Further relief valve here follows different API to calculate rated flow. To accurate calculate the relief valve outlet temperature, an option to calculate temperature based on entropy balance is also available. Normally enthalpy balance is performed to calculate the outlet temperature.
7. Aspen Pipe: Sale of Aspen PIPE has been discontinued as of March 2006. Customers with support contracts for Aspen PIPE will continue to receive support for the product through June 30, 2008. At present we are no more supporting this product.
Keywords: Pipesegment, Pipe, Hydraulics, Aspen Hydraulics, Gas Pipe, Compressible Gas Pipe, CGP, Pipesys, PIPESIM, PIPESIM NET, OLGA
References: None |
Problem Statement: When importing a MOL file on the Components | Molecular Structure | Structure and Functional Group sheet, after the MOL file is selected and the Molecule Editor closed, the structure is not saved. | Solution: Import it from the Components | Specifications | Selection sheet using the User Defined wizard.
Fixed in Version
V12.1
Keywords: None
References: : VSTS 573046, 567999 |
Problem Statement: A simulation has an error message that the extrapolated EOS volume root was used for property calculations. What does this mean and how do you fix it?
The message is
** ERROR
FLASH RESULTS ARE WITHIN TOLERANCE, BUT MAY BE UNSATISFACTORY BECAUSE
EXTRAPOLATED EOS VAPOR VOLUME ROOT WAS USED FOR PROPERTY CALCULATIONS.
Cause
In Aspen Plus, this appears while a block is being executed; look for the most recent Block: line in the Control Panel output to determine which one. A line of the message which follows the above may indicate where in the block the error occurred, such as problem occurred while flashing outlet stream. The error may also appear in report pass calculations, in which case the first line will indicate what was being calculated.
In determining the phases for flash calculations using an equation of state, the Aspen Physical Property System calculates properties for the system in both vapor and liquid phases by solving an equation. In some cases, the equation has no | Solution: (root) for a particular phase, and a pseudo-root is calculated based on pressure and the derivative of pressure with respect to volume. This pseudo-root represents an estimate of the properties for the phase which could not be solved.
The fugacities of the phases are determined from these roots and the pseudo-root if applicable, and from these, K-values are computed, which are used to determine the phases. It is possible that the K-values indicate the presence of a phase which did not have a real root in the equation of state. When this happens, the error is issued to indicate that the results might not be valid.
One of the causes for the lack of a root is when the system is beyond the critical point, where there are not separate vapor and liquid phases but only a single fluid phase. This leads to only one root, which the model assigns to a particular phase, resulting in the other phase getting a pseudo-root.
Solution
Check parameters for the components and property models used in the system, especially any parameters which you entered or estimated, or for which there is no value and default values are being used. Also consider using a different property model better suited for the system in question.
If the system is beyond the critical point over the entire range of calculations performed in some block or analysis, consider setting its Valid Phases to a single phase. The message indicates which phase is being extrapolated (vapor in the example); choose the other phase.
If you have reviewed the results and you are comfortable with the calculated properties despite the use of a pseudo-root, you can reduce the severity of this message or disable it completely by setting the Flash extrapolated root error level on the Setup | Calculation Options | Flash Convergence sheet. But be aware that this applies to all flash calculations in the model.
Keywords: None
References: None |
Problem Statement: How does the Adjusted Total Project Cost calculated and why is it different from Total Project Cost? | Solution: Sometimes when you run a complete Project Evaluation in Aspen Process Economic Analyzer (APEA), you can display the default HTML Item Report (Excel), and in this report you will found difference between Adjusted Total Project Cost and Total Project Cost in Project Summary sheet.
The Adjusted Total Project Cost is an adjustment based on the date of the project. For example V11 version, the cost basis is for 1-April 2018 (1st Qtr 2018) and therefore if you put the current date (like 5-Nov 2020) as estimated start date then it will assume some cost escalation for the difference in the date.
You can check the estimated start date setting in Project Basis view | Basis for Capital costs | General Specs, if you set it on 1-April 2018, do not give any escalation.
Then the Adjusted Total Project Cost will be same as Total Project Cost.
Below KnowledgeBase article would be useful for you to explain the escalation.
How is Escalation calculated for Icarus products? https://esupport.aspentech.com/S_Article?id=000082840
Keywords: Adjusted Total Project Cost, Total Project Cost, Escalation
References: None |
Problem Statement: How does FCC-Cat/Oil Ratio calculated and why the reported value is different from manually checking result? | Solution: Cat/Oil Ratio is the the ratio of FCC catalyst and feed oil, while if manually calculate the result based on FCC data, you will find that it is little difference from the FCC reported.
Take an example as below:
Example file: C:\Program Files\AspenTech\Aspen HYSYS V11.0\Samples\Refining Cases- FCC_w_Hydrotreater.hsc
Manually calculation: FCC-Cat/Oil ratio=Catalyst circulation rate/Feed Rate=2.595e6/1.77013e5=14.6599,
FCC model shown Cat/Oil ratio value=14.69,
This is a little confused that has the difference. Actually its about how the FCC feed model was set up. the feed model within the FCC removes H2S, it is a function of the sulfur processed in the feed since not all feed sulfur is cracked. The first thing we do in the feed is remove that H2S part so that the model satisfies the sulfur balance, but since it is removed, that no longer shows up as feed to the FCC and so the mass going to the riser is missing the H2S. The cat/oil ratio is calculated internally to the FCC riser model, so that accounts for the small discrepancy in what it reported for the cat/oil ratio and what you would get if you calculate by hand.
Two methods can verify above conclusion:
1. Removed Sulfur in feed by considering Sulfur content%[wt] and “Fraction of Feed S Processed” within the feed, then calculate cat/oil ratio, and get the exactly the same value.
FCC-Cat/Oil ratio=2.595e6/[1.77013e5-(1.77013e5*0.003688*0.5]=14.69
2.Alternatively, you could just find the H2S yield on the Results | Product Yields page and subtract that from the feed. So Cat/Oil ratio would be:
Catalyst circulation rate / (Feed Rate – H2S product rate)=2.595e6/(1.77013e5-497.1287)=14.69
Keywords: FCC, Cat/Oil Ratio, H2S yield, Fraction of Feed S Processed
References: None |
Problem Statement: How to select correct property package for modeling Crystallizer in Aspen HYSYS? | Solution: If modelling Crystallizer in HYSYS, user need to use OLI Electryte pkg, this need OLI license from OLI system company.
But if user is capable of using Aspen Plus, then recommend use Aspen Plus to model the crystallizer rigorously with more accurate properties databank, Elec Wizard help, crystallization reaction and crystallization block setup, also solid handling knowledge is required for user.
Please refer to below two knowledge base articles with example to understand in Aspen Plus:
Crystallizer Example in Aspen Plus ( BaSO4 solubility ) https://esupport.aspentech.com/S_Article?id=000025652 )
A Crystallizer Block Example (NaCl, Sugar solubility ) https://esupport.aspentech.com/S_Article?id=000037757 )
Keywords: Crystallizer, Crystallization reaction, OLI Electryte pkg
References: None |
Problem Statement: Fire calculations in safety cannot involve vapor evolution other than from vaporizing liquid. Currently the wetted fire calculations in the Safety Environment only support vapor-liquid phase equilibrium. | Solution: Currently the wetted fire calculations in the Safety Environment only support vapor-liquid phase equilibrium.
Systems where vapor is produced in other ways (most notably amine systems where a reversible chemical reaction will take up acid gases) will not complete a wetted fire calculation because the vapor phase is not in VLE alone with the liquid phase.
The relieving fluid for this calculation is currently specified as saturated vapor, but for this kind of system, that is not the correct reference condition.
Key Words:
PSV, Safety, Fire Scenario, Vapor Evolution Fire Scenario.
Keywords: None
References: None |
Problem Statement: How to get rid off the following error code 1008 message The base load token license not yet checked out by the SLM Base Load Token Service when launching Aspen PIMS? | Solution: In order to get rid off the following error 1008 message:
Please start/restart the Aspen Baseload Service and Sentinel RMS License Manager Services on the license server and launch Aspen PIMS from the client machine.
This should force the Baseload Service to check out the baseload token and the error 1008 should get resolved.
Keywords: Aspen PMIS, Error 1008
References: None |
Problem Statement: After installing Windows 10 Version 20H2, the Aspen Properties Enterprise Database (APED) Manager middle pane is blank.
The exact version is
Edition Windows 10 Enterprise
Version 20H2
Installed on 4/5/2021
OS build 19042.867
Experience Windows Feature Experience Pack 120.2212.551.0
The Database Manager looks like this: | Solution: The missing pane can be resolved by copying the file mmc.exe.config from C:\ProgramData\AspenTech\APED V12.0 to C:\Windows\SysWOW64 (Modify path to match the version you are using).
Alternatively, back up the existing file, then add the following missing lines from the mmc.exe.config file in the APED folder into the file in the Windows folder:
<startup useLegacyV2RuntimeActivationPolicy=true>
<supportedRuntime version=v4.0 />
</startup>
Keywords: None
References: None |
Problem Statement: Is there a way to manually set the status of 'Shifted' or 'Changed' on a file set in Aspen InfoPlus.21 (IP.21) apart from performing a manual file set shift or writing to a file set? Is there a history-related tool that could be run or setting that could be selected? | Solution: At the time of this writing (November 2020) there is not a tool that can perform these actions. This has been confirmed with AspenTech Development.
Keywords: fileset
References: None |
Problem Statement: How do I use Aspen Mtell Health Check Script on a user machine? | Solution: Aspen Mtell Health Check Script will allow customers to get Machine Learning agent information report from the Mtell Server, please find attached example. The PowerShell script can be executed on any Mtell client machine to generate a CSV file.
1. Download Mtell Health Check Script.zip
2. Extract the zip file content in C:\Temp folder
3. Open Config.xml file and update the following parameters and save the config.xml file
Company Name: Enter the name of your company
Database Name: Enter the name of your database on the SQL Server
Database Server: Enter the hostname of your Mtell Server
To find Database Name and Database Server name, follow below steps
Launch Aspen Mtell Agent Builder
Click Machine Learning Tab and click on Database
4. Double click the CST_GatherMtellData_October.ps1 file to execute
Note: the script will automatically login to the database using the default password. If it fails to login using the default password, please enter ID\Password to connect to the database.
5. A CSV file will be generated in same folder.
Keywords: Health check
HC
Mtell Report
References: None |
Problem Statement: What are the units for the Langmuir-Hinshelwood-Hougen-Watson (LHHW) pre-exponential factor and equilibrium constant? | Solution: The pre-exponential factor in the LHHW rate expression has very complex units which depend on the concentration basis selected, whether a reference temperature is specified, and the degree of concentration dependency of the driving force and absorption terms. The concentration is converted to SI units before the rate is calculated. The pre-exponential factor has the units needed to make the overall rate expression have SI rate units, that is kmol/(sec-m^3) for a rate basis of reaction volume or kmol/(sec-kg catalyst) for a basis of catalyst weight (version 2004.1 and higher).
The units for the pre-exponential factor can be written as
units for k = (kmol/sec-m3t) / (K or R)TE (conc)DFCE
-OR-
units for k = (kmol/sec-kg catalyst) / (K or R)TE (conc)DFCE
where:
TE = Temperature exponent. If a reference temperature is specified, this is zero because the reference temperature cancels out the units of the temperature. If a reference temperature is not specified, this is n, the temperature exponent specified. The temperature unit can be Kelvin or Rankine based on the units specified for the reference temperature.
conc. = Units of concentration. This depends on the concentration basis selected.
Concentration Basis
Units of Measure
Molar concentration
(kmol/m3)
Molality
(mol/kg water)
Mole fraction or mass fraction
(dimensionless)
Partial pressure
(N/m2)
Mass concentration
(kg/m3)
Note: fugacity units are the same as partial pressure, that is, N/m2
DFCE = Driving force concentration exponent. The sum of all the driving force concentration exponents (vi) for all components in term 1 of the driving force expression.
The equilibrium constants in the driving force (K1 and K2) and adsorption expressions (Ki) may also have units. K1 in the driving force expression is always dimensionless, but K2 has units to make the second term of the driving force expression have the same units as the first; that is, it has concentration units from the above table to a power equal to the difference in the net concentration exponents of the two terms. In the adsorption expression, each equilibrium constant has concentration units inverse of the concentration it is multiplied by, so that each term is a dimensionless number.
Refer to the Help Contents under Using Aspen Plus | SM Flowsheeting Tools and Reactions | Specifying Reactions | Specifying LHHW Reactions for Reactors and Pressure Relief Systems | Rate-Controlled Reactions for LHHW for all of the equations for the LHHW reaction.
Keywords: reaction rate expression
References: None |
Problem Statement: Why can I not find many of my IP21 tags when searching through Map Sensors?
Users may also get the following error message when trying to map sensors:
Unable to load tags. Error reading definition record (ID=116) fixed fields. Access Denied | Solution: Aspen Mtell uses the TextViewDef definition record query in to search for all the tags in Aspen IP21. All the sensor definitions need to be added in the All tags query record. Please have your IP21 administrator follow the steps below
1. Launch Aspen InfoPlus.21 Administrator
2. Click Find and select InfoPlus.21 Record
3. Type All tags and click Find Now
4. Expand All tags and select #Query_Lines
5. Include all the Analog Tag Definition Record names with the closed bracket
Keywords: Unable to find tags
Cannot view tags
no sensors found
Unable to map sensors
References: None |
Problem Statement: When running a stored procedure record in the SQL Plus Query of Aspen Process Data add-in, If we entered a regular Excel date format into the referenced date cell as date parameter, it will display the error Error:ProcessDataFunctions::ATSQLPlusQuery: DAY field out of range when converting CHARACTER to TIMESTAMP.
How to resolve this error? | Solution: The add-in does not transform the stored procedure's parameters into a different format. What you can do is to use Excel's function to convert the Date format into the SQLplus needed Text format into another cell and use that cell to input into Excel add-in. The error will be resolved after this procedure.
Here is an example I did to convert B1 cell from DateTime to SQLplus timestamp format.
=TEXT(B1,dd)&-&TEXT(B1,mmm)&-&TEXT(B1,yy)& &TEXT(B1, hh:mm:ss.0)
Keywords: DAY field out of range when converting CHARACTER to TIMESTAMP
Aspen Process Data add-in
SQL Plus Query
References: None |
Problem Statement: How to change the Regression Type of the XY Chart's regression equation on A1PE? | Solution: Here is the procedure.
1. Select the RegEqn on the tag legend area, then click the Edit button on the fly-out.
2. Select the Trend Type tab, change the Regression type to the desired one. The regression type includes Linear, Quadratic, Inverse, Logarithmic, Exponential, and Power.
3. Click OK and the regression equation of the XY chart will be modified.
Keywords: regression equation
XY chart
a1pe
References: None |
Problem Statement: I'm getting the following connection failed message, related to SSL Security error: | Solution: If your registry settings for SSL3.0 are set to the following, you will not be able to connect. This applies only to SSL 3.0 (not 2.0). Here are the keys which will cause the connection issue for SSL 3.0:
[HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\SecurityProviders\SCHANNEL\Protocols\SSL 3.0]
[HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\SecurityProviders\SCHANNEL\Protocols\SSL 3.0\Client]
Enabled=dword:00000000
DisabledByDefault=dword:00000001
[HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\SecurityProviders\SCHANNEL\Protocols\SSL 3.0\Server]
Enabled=dword:00000000
DisabledByDefault=dword:00000001
Now, by simply modifying your SSL 3.0 setting to the following, you will be able to connect to the database:
[HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\SecurityProviders\SCHANNEL\Protocols\SSL 3.0]
[HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\SecurityProviders\SCHANNEL\Protocols\SSL 3.0\Client]
Enabled=dword:00000001
DisabledByDefault=dword:00000000
[HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\SecurityProviders\SCHANNEL\Protocols\SSL 3.0\Server]
Enabled=dword:00000001
DisabledByDefault=dword:00000000
To generalize:
If you have SSL 3.0 -> Disabled, TLS 1.2 enabled: DB connection will fail
If you have SSL 3.0 -> enabled, TLS 1.2 enabled: DB connection will be successful
If you do not have SSL 3.0 keys, TLS 1.2 enabled: DB connection will be successful
First Configuration that works:
SSL 3.0 Disabled:
TLS 1.0 Client - Disabled
TLS 1.0 Server Enabled
TLS 1.1 Client and Server Enabled
TLS 1.2 Client and Server Enabled
Second Configuration that works
SSL 3.0 Client and Server Disabled
TLS 1.0 Client and Server Enabled
TLS 1.1 Client and Server Enabled
TLS 1.2 Client and Server Enabled
Keywords: None
References: None |
Problem Statement: When installing or upgrading to V12 for the APC Web Server on a Windows Server 2012 R2, the following symptoms were observed:
Connecting to web clients like the PCWS Web Page results in HTTP Error 503 on the web browser
In IIS Manager, the application pool AspenAPCAppPool would go into Stopped state after starting it manually or performing IISRESET
As a result of the AspenAPCAppPool crashing, the Aspen Security Pool would also crash but would be able to stay running if started again manually
Event Viewer Logs also showed errors regarding DotNetCore (.NET Core) application | Solution: The root cause of this issue seems to be an incompatibility for Windows Server 2012 R2 with the Microsoft .NET Core version 3.1.2. This can be resolved by installing the more recent version (as of June 2021) of ASP.NET Core Runtime v3.1.16 Hosting Bundle, which can be downloaded from here: https://dotnet.microsoft.com/download/dotnet/3.1
It is important to note that the Hosting Bundle is required to be downloaded and installed, not the individual components:
Once this is installed on the APC Web Server, restart the AspenAPCAppPool from IIS Manager Application Pools to verify that it stays running. Then restart the Aspen APC Web Provider Data Service and connect to the PCWS web page.
Keywords: aspenapcapppool, apc, application, pool, crash, stop, error, 503, pcws, web, server, .net, dotnetcore
References: None |
Problem Statement: Is it possible to use an HTRI block from the Aspen Plus user interface? | Solution: The recommended method is to use the HTRI Cape Open-Aspen Plus CAPE-OPEN connection in Aspen Plus. HTRI-IST has been replaced by the CAPE-OPEN interface. HTRI has a CAPE-OPEN interface, which plugs nicely into the Aspen Plus CAPE-OPEN socket. There is a CAPE-OPEN HTRI 32-bit version for Aspen Plus V10 and earlier and a 64-bit version for Aspen Plus V11 and higher.
HTRI version 8.2 or higher is needed to run with Aspen Plus V11.
CAPE-OPEN doesn't require any license key on the Aspen Plus side. There is a separate CAPE-OPEN block on the Model palette. Go to Library |
Keywords: HTRI, Cape Open
References: s, then check the CAPE-OPEN entry. You will then see a CAPE-OPEN tab on the Model palette. It will automatically display any CAPE-OPEN models that you have registered on your computer (you won't see any models on this tab if you don't have any CAPE-OPEN models registered).
HTRI developed this CAPE-OPEN interface for their heat exchanger models; it should work with Aspen Plus or HYSYS. Since AspenTech does not offer support for third-party CAPE-OPEN interfaces, customers should work directly with the suppliers of the interfaces. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.