question
stringlengths 19
6.88k
| answer
stringlengths 38
33.3k
|
---|---|
Problem Statement: How do flags SERVER_MONITORS and SERVER_ERROR_LIMIT affect Aspen Production Execution Manager functionality? | Solution: Both the SERVER_MONITORS and SERVER_ERROR_LIMIT flags should be set on the server. There is no technical reason to put it in one file or another, all files are picked up by the server during Apache startup. However they are meaningless for the clients, so you would want to avoid config.m2r_cfg, which is often used by the clients. Therefore we recommend flags.m2r_cfg on the server.
These flags were created to provide a mechanism to allow designated workstation(s) to be made aware of specific errors relating to server evaluation, errors such as a problem with custom code execution (typically, if not one-line boolean evaluations, the Recipe conditions are filled with custom code that controls the Recipe progress.) So for example, in a test recipe, if you put a statement that results in an error, such as a SQL_SELECT_ONE that returned multiple rows, or SQRT(-1), then any workstation on the SERVER_MONITOR list would receive notification of that error. This provides an early warning system of Recipe errors, so they can be proactively investigated.
This is useful, because for particularSolutions, Basic Phases may remain on the screen for many hours, or even days, and the ability to receive a warning of a problem with Condition evaluation can help you respond more quickly to the problem, instead of waiting for operators to complain that Basic Phases they expect to see are not ready for execution.
These two flags require that the Watchdog mechanism be in place, since the Watchdog mechanism is used to send warning messages (KB article 121390 discusses the Watchdog.) As long as the Watchdog active, then add both these flags on the server, preferably to the flags.m2r_cfg file, as described above. One or more workstations can be designated as a Server monitor (if this flag remains undefined, then Server Monitor messages are not sent.) For example:
SERVER_MONITORS=Workstation_A,Workstation_B,Workstation,C
With this setting in a compiled file, after an Apache restart, then those three workstations will now receive SERVER_MONITOR messages.
SERVER_ERROR_LIMIT can be optionally defined. When not defined, all errors will be broadcast. By defining, for example:
SERVER_ERROR_LIMIT=5
then only the 5 most recent messages will be sent. During every evaluation cycle, the RPL is checked again for errors, so keeping the limit low is smart, since it is meant as a warning system, and not a detailed debugging window. Once an error happens, the APEM APIServer debug file on the Server is a good starting point to see what errors are being logged.
Keywords: debug
monitor
Condition evaluation
cycle
References: None |
Problem Statement: Does Aspen eBRS cache its Local Security context in case Aspen Local Security or Aspen Framework is not available? | Solution: Aspen eBRS does not cache security information. For a high degree of control over user permissions all checks are done in real time, requiring constant availability of the Security Server.
Keywords: ALS
AFW
References: None |
Problem Statement: eBRS Basic Phase code can only be executed when the compiled Basic Phase is referenced in a Procedure Function Chart PFC that is then executed via a test order. This makes development harder, since many steps have to be taken before the BP code can run. | Solution: Two eBRS Version 2006 new features are Test and Manage mode, which create a shortcut enabling code to run more directly. These two new features are enabled by checking the Manage checkbox in the Basic Phase list of a Basic Phase library:
Test Mode
Once Manage checkbox is checked, the Execute button (has a lightning bolt symbol) can be used to run the Basic Phase immediately. This is Test mode. A limitation is that no order context exists, so if the Basic Phase code depends on global variables, or system variables such as $ORDER_COD, the code will not execute correctly.
To help offset this limitation, it is possible to create a text file populated with simulated order environment information. Create a text file with this name and location:
<AeBRS>\MOC\config\BPLExecutionEnvironment.txt
Example information to add to the file would be values for global or system variables. For known $ variables (think of these as system variables) the proper datatype is given by the system. All other variables should be defined as $$ variables (i.e. available to all Basic Phases within an RPL) and they are limited to string type. An example file is attached.
Manage Mode
Manage mode means double-clicking on a Basic Phase after it has been added to the PFC Chart in a Recipe:
The advantage to Manage mode over test is that any parameter values for arguments defined in the RPL document PFC chart are available. In this case the predefined string for Thick Crust pizzas is displayed in the Basic Phase. In Test Mode, the BP would still display, but the box would be blank.
Manage mode is also available from within Order context -- review KB 126608 for more detail on that functionality, and how related environment variable $EBR_PHASE_MANAGE plays a part.
Keywords:
References: None |
Problem Statement: When creating a verification test order on a V7.3 system where the GML library has been imported at installation time, an error like the following occurs:
And errors like the following are written to the MOC debug file:
09:26:19: m2rDatabaseConnection(IP21).URL : jdbc:odbc:DRIVER={AspenTech SQLplus};HOST=ServerName;CHARTIME=No;CHARINT=No
09:26:21: Exception
java.sql.SQLException: [AspenTech][SQLplus] Host name not found
at sun.jdbc.odbc.JdbcOdbc.createSQLException(JdbcOdbc.java:6957)
at sun.jdbc.odbc.JdbcOdbc.standardError(JdbcOdbc.java:7114)
at sun.jdbc.odbc.JdbcOdbc.SQLDriverConnect(JdbcOdbc.java:3073)
at sun.jdbc.odbc.JdbcOdbcConnection.initialize(JdbcOdbcConnection.java:323)
at sun.jdbc.odbc.JdbcOdbcDriver.connect(JdbcOdbcDriver.java:174)
at java.sql.DriverManager.getConnection(DriverManager.java:582)
at java.sql.DriverManager.getConnection(DriverManager.java:185)
at m2r.DataModel.m2rDatabaseConnection.connect(m2rDatabaseConnection.java:195)
at Util.dbConnection.dbConnection$ExternalDBConnection.<init>(dbConnection.java:70)
at Util.dbConnection.dbConnection.connect(dbConnection.java:32) | Solution: This happens because the GML library import also adds a set of generic connections to GML-related datasources, such as Aspen Production Record Manager (APRM), Aspen Manufacturing Master Data Manager mMDM and InfoPlus.21.
These connections to external datasources are maintained in the Config module, via the gear/mesh icon, here:
For the APRM, mMDM and other connections, if no password is defined, then the connection is automatically bypassed without attempting to check it. However for InfoPlus.21 (considered an external connection) it is automatically checked because an entry exists for it under Additional default DB connections. So to avoid this error, either the connection must be fully defined or, if not needed, remove the DEFAULT_DB_CONNECTIONS reference.
To fully configure the InfoPlus.21 connection, (1) edit the IP21_URL line to specify the HOST value, (2) edit IP21_USER and IP21_PASSWORD. The finished connection should look like this:
To remove it, remove IP21 from DEFAULT_DB_CONNECTIONS:
Once the connection is fully configured, or the reference to it removed, save the changes made in the Editor via the Toolbar Save icon, with the disc image:
The Database connection error should now be resolved.
Keywords: No phases for this RPL {0}.
Database error:[AspenTech][SQLplus] Host name not found
Design with ERRORS
References: None |
Problem Statement: Every action performed in Aspen Production Execution Manager (formerly Aspen eBRS), whether by a user, or programatically, is not automatically recorded in the audit trail. Is it possible to extend the types of activities recorded to the audit trail? | Solution: Calls to the following functions are, by default, not sent to the audit trail:
AssignWorkstation
SetOrderState
SetOrderDetails
SetOrderParamDetails
SetOperState
CreateOrderFromMasterRecipe
CreateOrderFromProcedureLogic
CreateOrderFromControlRecipe
However before calling any of those functions, if a value is given to the System Variable REASON_OF_CHANGE, then the activity generated by the functions will be added to the audit trail.
Keywords: eBRS
References: None |
Problem Statement: An Aspen Production Execution Manager crash occurs, and the following entries are found in the debug file:
12:58:17: Exception
com.microsoft.sqlserver.jdbc.SQLServerException: The transaction log for database 'AeBRS' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases
at com.microsoft.sqlserver.jdbc.SQLServerException.makeFromDatabaseError(Unknown Source)
at com.microsoft.sqlserver.jdbc.SQLServerStatement.getNextResult(Unknown Source)
at com.microsoft.sqlserver.jdbc.SQLServerPreparedStatement.doExecutePreparedStatement(Unknown Source)
at com.microsoft.sqlserver.jdbc.SQLServerPreparedStatement$PrepStmtExecCmd.doExecute(Unknown Source)
at com.microsoft.sqlserver.jdbc.TDSCommand.execute(Unknown Source) | Solution: This error occurs if the SQL Server logging file does not allow unrestricted growth. To resolve the issue:
1. Open SQL Server Enterprise Manager, and expand to the AeBRS Database.
2. Choose Properties on the AeBRS Database.
3. From the Properties dialog, click on Files, on the left.
4. Note the properties for AeBRS_log1, and its Autogrowth parameters - it is likely restricted.
5. Click the . . . indicator for the AeBRS_log1 line. Likely under Maximum File Size you need to choose the radio button Unrestricted File Growth.
Of course when choosing Unrestricted File Growth it is important to know how much free space exists on the disk where the database is saved and have a good idea of the rate of growth of the database.
Keywords:
References: None |
Problem Statement: Why do unrelated screens and MOC itself freeze when I use the WAIT or WAIT_UNTIL functions? | Solution: Both WAIT and WAIT_UNTIL stop execution of the thread they are executed on, for the specified time. Since all GUI-related components are executing on the same thread, they are all stopped. Current phase actions (the one where the call was made, other phases, MOC itself.) Everything GUI-related is paused until the WAIT / WAIT_UNTIL time period elapses.
The intended use for WAIT-related functions is non-GUI threads, i.e. async threads, code blocks, screen pre-actions. Short WAIT statements can be helpful on GUI-related threads when for example a file is created on the drive, and a pause is needed for a second or two while the file system updates, before attempting to access that file.
A work-around would be to run a concurrent action that references your screen of interest, disables it, then re-enables it after the desired time period has elapsed, essentially controlling the behavior of one element of a thread from a separate thread.
Keywords: Block
pause
freeze
timer
counter
References: None |
Problem Statement: Aspen Production Execution Manager versions prior to V7.1 had a blue background. V7.1 was hard-coded to white. However some customers, who use MOC maximized, understandably did not like the all-white screen. Later versions still default to white, as shown here in V7.3: | Solution: V7.2 and later offer a set of configuration keys in the flags.m2r_cfg file that can be used to control the background (and in fact entire color scheme) for MOC. The default values at installation are as follows:
Even though there are multiple interactions of color schemes possible, as outlined by the flag logic above, if the goal is only to change the background color of MOC, just uncomment the second MDIAPP_DESKTOP_COLOR flag, run codify_all.cmd, and reopen MOC:
Keywords: change color
color scheme
desktop
References: None |
Problem Statement: Version 2004.2, with Patch 9, and all later versions of Aspen Production Execution Manager (APEM), add a new environment variable, $EBR_PHASE_MANAGE.
What is this variable used for? | Solution: A summary of the possible values for this environment variable is included in the APEM Design
Keywords:
References: Guide, starting with the V7.2 release. The variable itself receives its value depending on the context in which a phase is executed:
VALUE
EXPLANATION
NULL
Basic Phase is being executed as part of standard Order execution, via whatever method (Order Tracking, Workstation, Automatic Activation, etc.)
DESIGN
Basic Phase is being executed from the RPL Designer. This context is commonly used to verify that parameter values have been specified correctly, since RPL parameters are visible to the Basic Phase from inside the Designer tool.
TRACKING
Basic Phase is being executed via a right-click (Properties, choose Manage) in the Order Tracking module. This means the Basic Phase has Order context, since it is executing as part of an order.
This environment variable allows the programmer to finely tune Basic Phase behavior when it is invoked via Manage option. (For more information on the Manage feature, see KB 121637.) By making a test of the $EBR_PHASE_MANAGE value in the pre-action of a Basic Phase for example, a designer could choose to exit (essentially disallowing Manage functionality for whatever BP's had that code added.)
This feature allows execution of any phase, independent of the order flow. You can execute already completed phases using Manage, or phases not yet ready to execute. This takes advantage of phases that do work, perhaps managing parameter values of the recipe, or reading / writing data to some external source. But in Manage mode, once you hit ?OK? or enter the Password to validate the phase, there is nothing added to the screenshot report or Batch Record, and the RPL does not progress further as a result of the Manage work.
Additional Notes originally included in Patch 2004.2.9 documentation:
BPL code independent execution is enabled according to the following use-cases:
1) BPL-components (BPs, local-oper & scripts) can be execute for testing:
parameter values take defined defaults
BPL/BP environment set according to executing BPL/BP - BP environment being NULL for local-oper and scripts
- user/workstation environment set as usual
Order-environment may be set using
<AeBRS>\MOC\config\BPLExecutionEnvironment.txt ? it may - include order-global variables, restricted to strings
2) BPs can be enabled for management execution checking on new column in BPL/BP table.
3) Management enabled BP can be executed from Order tracking Manage option, independently of the state of the phase:
BP is executed with the same parameter values as regular execution would have
Order environment is same as regular phase execution
Additionally, $EBR_PHASE_MANAGE:=TRACKING (it is NULL in regular executions)
4) Management enabled BP can be executed from Designer double clicking on the phase:
BP is executed with parameter values according to RPL parameter values
Environment limited to BPL/BP, user/workstation and $STEP_ID, $STEP_NAME, $STEP_DESCRIP (arrays) and $STEP_PATH=UP\Op\Ph
Additionally, $EBR_PHASE_MANAGE:=DESIGN (it is NULL in regular executions)
Customer is supposed to design BP logic to work accordingly in case BP is run for management.
Notice that independent execution:
Can be executed many times simultaneously ? every execution window receives a unique name that appears in the active window menu/list (View option in designer): active window list allows restoring/bringing to
front/selecting executing window
Does not record to batch record and screenshots.
Cannot be interrupted, it is cancelled by closing execution window |
Problem Statement: When executing BATCH_RECORD_READ from Aspen Production Execution Manager (APEM) to get batch information from Aspen Production Record Manager, data retrieved does not include decimals from numeric data. | Solution: To resolve Modify the entry APRM_URL in the APEM configuration files with the following content:
APRM_URL = jdbc:odbc:DRIVER={AspenTech ODBC driver for Production Record Manager};HOST=<Server name>;LocaleID=1033;
Thus, the parameter LocaleID=1033 forces to the driver to return decimal numbers with dot (.) and not with comma (,). This is the way that APEM understand.
However, the Boolean is fixed for English so you can either use 0 or non-zero. Or True, On, Yes.
Example: Mixed char types in single column
DRIVER={AspenTech ODBC driver for Production Record Manager};HOST=localhost;charfloat=Y;charint=N;chartime=Y;localeid=1033
select characteristicname, value from batches(*) where batchid=15715
MyBoolean, Yes
Example: Single char type in column. Different connection options.
DRIVER={AspenTech ODBC driver for Production Record Manager};HOST=localhost;charfloat=Y;charint=N;chartime=Y;localeid=1033
select myboolean[0] from batches where batchid=15715
Yes
The locale id is defined by windows whic can be found on the Microsoft link below.
http://msdn.microsoft.com/en-us/library/0h88fahh/
Keywords: BATCH_RECORD_READ
Aspen Production Execution Manager
Decimal
References: None |
Problem Statement: No session could be identified error is displayed when visiting the Production Execution Manager Order Execution web page: | Solution: For a newly installed Production Execution Manager system to successfully display the Order Management or Order Execution pages, several additional configuration steps are needed:
1. To have an accurate audit trail, it is required to add the nodename of any PC that will be allowed to visit either the Order Management or Order Execution web pages into the Config module:
In this case the server itself, DOBEASE708 has been added. This allows control over which PCs are considered part of the Production Execution Manager system, preventing any PC not in this list from visiting the web pages.
IMPORTANT NOTE: In some network environments the server does not recognize the short form of a nodename. In that case add the fully qualified form instead:
2. Next, it is necessary to configure the Aspen Production Execution Manager ADSA Service. See KB 121129 for more details on the ADSA configuration.
3. Lastly, the user account must be added to the Web User role in Aspen Local Security, if not already a member of the Production Execution Manager Administrator role.
Keywords: resolve
URL
References: None |
Problem Statement: Is an audit trail entry created if 'REASON_OF_CHANGE' is not set for APEM functions? | Solution: No. But if there is no specific reason to be documented, and an audit trail entry is still desired, then set
REASON_OF_CHANGE:=
Keywords: SetOrderParamDetails
References: None |
Problem Statement: When running the AeBRSInstaller module on Windows Server 2008, various Access is denied messages may be displayed, such as:
These errors occur, even when logged into the Server as Administrator because of stricter security policies in Windows. The Installation manual advises to run with higher privileges, but how exactly to do that? | Solution: The AeBRSInstaller java module is started from a Command prompt. To run with elevated privileges, instead of clicking directly on the Command icon, right-click, and choose Run As Administrator:
Now start the installer from within this high-privilege context, and there should be no more Access is denied warnings:
Keywords: C:\Program Files (x86)\ApacheSoftware Foundation\Tomcat 5.5\webapps\AeBRSserver\aebrs_api.tmp (Access is denied)
C:\Program Files (x86)\AspenTech\AeBRS\cfg_source\config.m2r_cfg
References: None |
Problem Statement: Using the INSERT function to add multiple lines to a User table can result in a Duplicate Key error. | Solution: Add a WAIT(0.1) statement between each use of INSERT -- this allows time for the timestamp (which is being used by the table as the primary column) to increment at least a tenth of a second.
Keywords: None
References: None |
Problem Statement: What is mean by Minimum Temperature Approach in HeatX block of Aspen Plus? | Solution: Minimum temperature approach is defined by the minimum temperature difference of hot and cold fluid when looking at their respective heat curves for a given heat exchanger service (with known configuration and heat balance). It is generally taken as the temperature difference between the lowest temperature of the hot fluid and the highest temperature of the cold fluid.
For example, you draw the diagram of heat movement and you will see the minimum distance between the two line is the minimum temperature approach. The closer the cold fluid approaches the hot, the larger the heat transfer area.
The minimum approach temperature is specified in Aspen Plus (HeatX) is to avoid temperature crossovers during convergence. If the approach temperature is reached at either end of the exchanger or at any point of phase transition internal to the exchanger, theSolution is accepted as converged.
Keywords: HeatX, Minimum Approach Temperature, Aspen Plus
References: None |
Problem Statement: While using MOC, or working on a Basic Phase in MOC or via the web client, you may see timeout messages after a few minutes of no activity:
Inactivity message in MOC
Inactivity message in Order Tracking
(meaning a Basic Phase is on the screen)
Inactivity message seen during web-based execution
By default Aspen Production Execution Manager sets timeouts for these three activities according to these defaults:
INACTIVITY_PERIOD controls a timeout for the overall MOC session. OPERATION_INACTIVITY_PERIOD is a specific timeout for running Basic Phases within MOC. For example an Administrator may allow MOC to be open for eight hours based on a single authorization (INACTIVITY_PERIOD=28800), but only want to allow five minutes for a Basic Phase that is being actively worked on (OPERATION_INACTIVITY_PERIOD=300.) WEB_INACTIVITY_PERIOD controls the timeout of both the Order Management and Order Execution web pages.
Notice that for OPERATION_INACTIVITY_PERIOD it is provoked by trying to hit OK, thus committing the work done on the screen. The user can leave the Basic Phase up on the screen indefinitely, but once the timeout has been exceeded, a valid user id and password will have to be supplied. By design, there is no cancel option. Once a user hits OK, credentials must be given: | Solution: The above settings are kept in the path.m2r_cfg file, typically located here:
C:\Program Files\Aspentech\AeBRS\cfg_source
In a standard configuration, this file should only be edited on the server. Once it is changed, run codify_all.cmd (found in the same directory) to produce updated compiled configuration files. The next time MOC clients log in, they will launch with the updated settings.
To suppress these timeouts entirely, allowing indefinite logins, set the flags to 0.
Keywords: Session timed out
Your session has timed out. Please log in again to resume
eBRS
References: None |
Problem Statement: As a Designer it can be very useful to hide certain screen controls and only make them visible to the user when needed. How is that done? | Solution: Version 2006.5 and later have a run-time visible property for every Screen Component. The Basic Phase shown below is a demonstration, and is attached to thisSolution (it is a zip file containing an Aspen Production Execution Manager .CHK file.) To load it, create a new, empty Basic Phase, then choose File, Import, and Browse to the .CHK file):
The Visible Property of each component is toggled by the code in the Actions property of an associated button. Three different variations on toggling the property are shown. For the Label (first component, upper left), here is a traditional coding approach:
Notice that since the code is running inside the same screen, to address each component correctly there is no need to include the Basic Phase screen name (which is MAIN), but for the second component, an Expression, that addressing is included, and it works fine. This syntax can be useful to address components in another Basic Phase building block (for example manipulating a screen component from a Concurrent Action component) :
Last, and used for all other Components, is a much more compact NOT assignment. Here is the one for the Refreshable Expression:
The :=! operator can be read as whatever value this variable currently holds, make it the opposite. This handily switches the Visible property every time the code executes, all in one line.
Keywords:
References: None |
Problem Statement: How can I customize the distribution of one component in two liquid phases by defining a KLL relation based on composition? | Solution: In the following example, we have three components: a mixture of two components (C1 an dC2) that is fed to an extractor block along with a third component (C3) that is used to extract the first component. The heavy phase is made by C3, and we can assume that no C3 exists in the light liquid as well as no C2 exists in the heavy phase.
The subroutine makes the following assumptions:
The first component in the list corresponds to the component that will be distributed in the two liquid phases. In order to use the subroutine for your case make sure you add the component of interest first in the component list.
No equations exist for KLL for components two and three, if you need to add this the fortran subroutine should be modified. Also if more than three components exist in your simulation.
The equation for the distribution of the first component in the two liquids is a power relation with the form:
K=a*X^b.
K - Distribution coefficient Y/X for n components.
a - Coefficient 1, Real number 1 in the subroutine parameters.
b - Coefficient 2, Real number 2 in the subroutine parameters.
X the composition of C1 in the first liquid or C1 in the second liquid. To change the relation that will be used, the integer should be 1 for composition based on first liquid and 2 for composition based on second liquid.
In the attachments you will find:
An Excel spreadsheet with sample experimental data that is used to make the regression of the expression that related the KLL vs the composition.
Aspen Plus document that contains the subroutine and two Decanters with this subroutine: one of them solves based on the composition on L1 and the other on the composition for L2.
Keywords: Aspen Plus, KLL, Subroutine, Extractor, Decanter, Liquid Liquid
References: None |
Problem Statement: Customer had written a query using the Aspen Production Execution Manager (APEM) DB_UPDATE function to update InfoPlus.21 records. However an error similar to this was displayed:
Database error:[Aspentech][SQLplus] Invalid integer: 1286631664492
The error seemed strange since customer in fact was not sending an integer across in the update.
When running the same query syntax from SQLplus (using UPDATE instead of DB_UPDATE, the latter being an APEM-specific command) it worked fine. Query was similar to this:
update MyIP21TextTag set ip_input_value = 'c', QTIMESTAMP(ip_input_value) = '09-Oct-10 13:54:58', QSTATUS(ip_input_value) = 'Good' | Solution: The error occurred because by default the DB_UPDATE function will attempt to send a signature generated based on the data for traceability purposes. When using DB_UPDATE to insert data to an APEM table, the signature is necessary. But other databases, without an accompanying Signature column, will reject the signature data being sent. However Signature is an optional parameter for DB_UPDATE, as explained in the APEM Design
Keywords: None
References: Guide:
Signature
Optional STRING containing the name of the signature field.
Optional parameter. If omitted it is assumed that the name of this field is EBR_SIGNATURE.
If the string is passed, it indicates that there is no signature.
Once the code using DB_UPDATE specified the Signature parameter as , it worked successfully. |
Problem Statement: Some 3rd party add-ins for Microsoft Excel can cause issues if they are installed and active on the same machine as Aspen PIMS. This could involve dialogs with error messages or in some cases PIMS crashing/hanging during Excel Automation functions. This is because COM add-ins can be written in a way that interferes with Excel Automation and Excel automation is used by PIMS to read model input tables.
An example of this would be a COM add-in that is written such that mscoree.dll is its listed location. An add-in's location can be determined in Excel by going to OPTIONS | Add-Ins, as shown below:
Notice that all of the add-ins listed above have specific file locations and the files are generated by the makers of the add-in. However sometimes an add-in will list a Microsoft file as its location. Using a Microsoft file for this (such as mscoree.dll) can cause problems for other applications that use Excel Automation such as Aspen PIMS. | Solution: For COM add-ins which use the MScoree.dll file as their location, the bestSolution is for the maker of the add-in to change it so it no longer cites mscoree.dll as its location. If the maker of the add-in is unwilling to do that, then there are some workarounds that can be tried in an attempt to mitigate the interference from the COM add-in.
Below is a list of possible workarounds that can help with various symptoms of COM add-in interference. None of these are assured to resolve all interference issues, however, each of them have worked in some scenarios.
Always make backups and verify these changes are compatible with your company IT policies before making any changes.
If none of the workarounds is effective, then the add-in will have to be deactivated on the machine when PIMS is being run.
WORKAROUNDS THAT MAY HELP:
1. Update Registry settings for ActiveX:
a. Click Start, click Run, type regedit, and then click OK
b. Expand the following registry subkey: HKEY_CURRENT_USER\SOFTWARE\Microsoft\Office\Common
c. Right-click Common, point to New, and then click Key
d. Type Security, and then press ENTER to name the new subkey
e. Right-click Security, point to New, and then click DWORD Value
f. Type UFIControls, and then press ENTER to name the value
g. Double-click UFIControls
h. In the Value data box, type 1, and then click OK
i. Expand the following registry subkey: HKEY_CURRENT_USER\Software\Microsoft\VBA\
j. Right-click VBA, point to New, and then click Key
k. Type Security, and then press ENTER to name the new subkey
l. Right-click Security, point to New, and then click DWORD Value
m. Type LoadControlsInForms, and then press ENTER to name the value
n. Double-click LoadControlsInForms
o. In the Value data box, type 1, and then click OK
p. Quit Registry Editor.
2. Update registry settings for how Excel loads add-ins
a. Using similar procedure as above, set a value of 9 in: HKEY_CURRENT_USER\Software\Microsoft\Office\Excel\AddIns\ExcelCompanion (LoadBehavior)
3. Set your Excel Trust options so that all add-ins are trusted even if they are not signed (mscoree.dll cannot be signed)
a. In Excel, go to Options | Trust Center | Add-ins
b. Uncheck the option “Require Application Add-ins to be signed by a Trusted Publisher
Keywords: None
References: None |
Problem Statement: I think that two files are the same, but they are giving different results. How do I find the difference? | Solution: The easiest way to compare two simulations is to compare their input (.inp) files. The input file is a text summary of the simulation that is sent to the Aspen Plus calculation engine. You can go to File Export File and select Input Files (*.inp) or click on Input in the Summary area of the Home ribbon. These are text files that can be compared by hand or using one of the many file comparison tools such as Winmerge (http://winmerge.org/) or the Windows command line tool File Compare.
C:\windows\system32\fc.exe
file compare from the dos prompt: fc file1 file2 > diff.txt
Input language is documented in the Aspen Plus Input Language Guide available in the Documentation area of the website.
Keywords: None
References: None |
Problem Statement: In this example, you want to evaluate the effects that a change in crude will have on your CDU, mainly in the hydraulics. When issues arise, you learn how to use Aspen HYSYS to try various | Solution: s that meet your objectives using HYSYS, HYSYS Petroleum Refining, and Column Analysis.
Solution
You will learn to:
Evaluate column performance using Column Analysis in Aspen HYSYS:
Identify various hydraulic issues using visual tools
Run a case study in Aspen HYSYS to evaluate feed scenarios:
Monitor changes in yields and potential hydraulic issues
Troubleshoot potential problems arising from changes in feed:
Evaluate feasible operational changes to accommodate feed change
Consider additional ways to accommodate changes in feed at turnaround:
Adjust column temperature profiles and side operations
Note:This example uses Aspen HYSYS V9 CP1 (35.0.1.271)
Keywords: Aspen Petroleum Refining, Aspen HYSYS, Distillation Crude Unit, CDU, Aspen HYSYS Petroleum Refining, Column, Crude Column, Atmospheric Column, Distillation, Pump Around, Side Stripper, ACU, Crude Distillation Unit, Atm Crude Distillation Unit, Assay Management, Petroleum Feeder, Assays
References: None |
Problem Statement: Can I create my own sequence for a flowsheet? What happens if I specify an incorrect sequence (either full or partial)? | Solution: The Convergence / Sequence / Specifications form is designed to allow you to override theSolution sequence chosen by Aspen Plus. The calculation order for all or part of the flowsheet can be modified. Within a sequence, you can specify a previously specified sequence for a subset of the flowsheet. For large flowsheets, it is useful to build up the sequence specification in this manner. For more information on how to do this consult the help page by pressing F1 on your keyboard while you are viewing the Sequence form.
On each row of the form, you can enter one of either the beginning of a loop, the end of a loop, a block ID, or a sequence ID for part of a flowsheet. If you have no recycle loops or your simulation does not require an iterativeSolution, you can leave the Loop-return column empty. You only need to fill this column if you need to solve the flowsheet iteratively.
If a full/partial sequence is specified, Aspen Plus uses the complete full/partial sequence exctly as typed with these exceptions:
If you specify Conv-Options Defaults Check Sequence, Aspen Plus checks whether all loops in a sequence are torn. If a loop is not torn, Aspen Plus displays an error message.
If you specify Execute Before or Execute After in a Calculator block, Aspen Plus inserts the Calculator block into your sequence.
A Design-Spec automatically generates convergence blocks for design specifications and inserts them into your sequence
If you specify a partial sequence, but inadvertently leave a block out of the sequence, Aspen Plus may create an extra Tear stream to couple the blocks included by the partial sequence with the omitted block. If Tear stream convergence is required inside a user specified sequence, the user MUST specify the Convergence block and include the Convergence block within the partial sequence. Otherwise, the simulation will appear to complete normally, but may be out of mass balance.
Keywords: DESIGN-SPEC
design specification
References: None |
Problem Statement: What is the logic behind the selection of the binary databanks? | Solution: The list of Selected databanks is determined by
The list of databanks specified as searchable on the Tools/Options/Component data tab.
The property methods selected has the same property models as those supported by the databank.
Binary parameters are retrieved from the databanks specified to be searched on the Tools/Options/Component Data tab. Note that binary parameter databanks support only some models (vapor phase or liquid phase), so a given databank may not contain a parameter for a specific model.
For example, the VLE-RK databank is designed for activity-coefficient property methods that use the RK equation of state for the vapor phase and the following activity coefficient models: NRTL, Wilson, UNIQUAC. Therefore, if you selected the NRTL-RK property method, the Databank tab for the NRTL-1 binary interaction parameter form will show the databank VLE-RK as the Selected databank.
If the databank is not designed for the model, it will not be selected even if it is on the Tools/Options/Component data tab. For example, if ENRTL-RK and HENRY as the only Binary Databanks and NRTL is selected as the property Method, only the HENRY databank is listed on the Parameters/Binary Interaction/Henry/Databank sheet.
The ENRTL-RK databank is not. The databank ENRTL-RK is keyed on the property method ELECNRTL or its variations (ENRTL-HG). If one of these property methods is selected, the ENRTL-RK databank will be selected on the databank tab.
The default databank search order is the global databank search order from the Tools | Options dialog box on the Component Data tab. By default, all databanks containing parameters for the selected property methods are searched. However, this default is only set when the form is first initialized, when you first specify a property method that uses a particular parameter. If you specify one property method, and later specify another property method that uses the same parameter but which has different associated databanks, the databank search order will remain as specified for the first property method until you change it on this sheet.
For example, if you specify method NRTL-RK, the default databank for NRTL-1 is set to VLE-RK. If you later specify method NRTL-HOC, the NRTL-1 Databanks sheet does not automatically update to use databank VLE-HOC.
Keywords: binary activity coefficient model
References: None |
Problem Statement: TEGを含むストリームでWater Dew point Temperatureが実際よりも高い数値を示す。 | Solution: HYSYSではStreamにTEGが含まれている場合、Water Dew point Temperatureが高めに表示される傾向があります。HYSYSは温度を下げながら順次Flash計算を行ない、最初の液相出現温度をHC Dew point, 2番目の液相出現温度をWater Dew pointとして認識します。しかしStreamにTEGが含まれている場合、TEGが2番目の液相出現温度となるために、HYSYSはこれをWater Dew pointとして認識してしまい、実際のWater Dew pointよりも高い数値を示す結果となります。
以上の現象を回避し、より正確なWater Dew pointを確認する方法として、Component Splitterを使用する方法があります。Component Splitterは入口に1つと出口に2つのStreamと共に構成され、ユーザーによって指定された成分毎の分配率にしたがって、1つのStreamを2つのStreamに分割する機能を有しています。これによってTEGを取り除いたStreamのWater Dew pointを計算することで、TEGの影響を受けない、より正確なWater Dew pointを確認することが可能となります。
Keywords: 日本語、TEG, Water Dew point Temperature, Japanese
References: None |
Problem Statement: Why can I not see FlowsheetWide when using Derivative analysis? | Solution: To select the Flowsheewide for your derivative analysis, the user needs to be in the Main Simulation environment. If you activate the derivative analysis when you are in a subflowsheet, then you will not be able to see this Flowsheetwide option, which should be like the screenshot below:
This is also applicable for Case Studies, which has to be run in the Main Simulation environment.
Keywords: Derivative analysis, Flowsheetwide
References: None |
Problem Statement: What do the letters in the Phases column of the Property table (Stream Analysis tool) mean? | Solution: The phase column indicates the phases that have been detected at each pair of independent property values as per below:
V - vapour phase
L - Light liquid (hydrocarbon rich) phase
H - Heavy liquid (aqueous) phase
The combination of these letters (either 2 or 3) represents the number of specific phases respectively. e.g. H-V represents a 2-phase fluid comprising of a Heavy liquid (aqueous) phase and vapour; L-H-V represents a 3-phase mixture of all of the aforementioned phases.
Keywords: vapor, HC phase, Aq phase
References: None |
Problem Statement: What is the flammability limit? | Solution: This example estimates the flammability potential of liquid mixtures containing water and another component, typically an organic compound.
Aqueous mixtures are often vulnerable to flammability problems. The flammability analysis must cover a broad range of compositions and temperatures, and experimental measurements to define safe ranges are usually time consuming and expensive. The method presented in this example, which is based upon the work of Merck technologists ( Sharkey, J; Gruber, G.; Muzzio, D., Prediction of the Flammability Range for Chemical Systems Using Aspen, presented at AspenWorld 2002, 27 October - 1 November 2002, Washington, D.C.) provides an Excel model to estimate the flammability potential of an aqueous mixture.
A simple and reliable assumption is that a vapor mixture will support combustion if its adiabatic flame temperature exceeds a threshold value, typically 1,400 K. The calculation procedure in this example is to first compute the vapor composition of the aqueous mixture in physical equilibrium with air, and then to react the vapor mixture with available oxygen. The reaction first forms CO, and then CO2 if sufficient O2 is available. The reaction extent may thus be limited by the organic or by oxygen. The example reports and charts the adiabatic flame temperature as a function of liquid composition, so the user is able to obtain a visual image of the flammability range of the liquid.
How to use the example
The example has several organic solvents available for analysis. In order to run the calculation, set the following inputs (as shown in the segment of the Excel worksheet copied below):
Component ID of the organic (ETHANOL)
Temperature (60.07?C)
Pressure (1 atm)
Maximum solvent composition (0.20)
Example:
The Excel spreadsheet will compute the adiabatic flame temperature as a function of the liquid mole fraction and will produce the following chart.
The calculation predicts that the lower and upper flammability limits for an aqueous ethanol mixture at 60.07?C occur at mole fractions of 0.014 and 0.101, respectively.
The calculation procedure also works for mixtures that exhibit liquid immiscibility. The following chart presents the predicted adiabatic flame temperature for aqueous mixtures of 3-methyl-1-butanol (also known as isoamyl alcohol) at 52?C.
The calculation predicts that the lower and flammability limit for an aqueous 3-methyl-1-butanol mixture at 52A?C occurs at a mole fraction of 0.0018. Beyond a liquid mole fraction of about 0.0022, the liquid phase becomes immiscible and an organic-rich phase forms in addition to the water-rich phase. While the system has two liquid phases, the vapor phase composition is constant, hence the adiabatic flame temperature becomes constant.
The adiabatic flame temperature of aqueous 3-methyl-1-butanol mixtures at 52A?C will rise at high organic concentrations after the water phase disappears, which occurs at 3-methyl-1-butanol mole fractions greater than about 0.5. The results are shown in the figure below. Note that all aqueous mixtures of 3-methyl-1-butanol beyond a mole fraction 0.0018 are unsafe.
It is interesting and instructive to study another system that exhibits liquid-phase immiscibility: aqueous mixtures of n-hexane. The figure below presents the adiabatic flame temperature of aqueous mixtures of n-hexane at 25A?C.
The solubility of n-hexane in water is about 3 ppm and beyond this concentration two liquid phases exist and the adiabatic flame temperature remains constant until the water phase disappears. It is interesting that the adiabatic flame temperature goes through a maximum at the very low concentration of under 1 ppm. This suggests that mixtures of n-hexane in water that are below the saturation limit are actually more susceptible to flammability problems than the pure n-hexane liquid.
Calculation Details
Aspen Properties Excel applications are always based upon an Aspen Properties file, in this case a file named FlammabilityLimit.aprbkp. FlammabilityLimit.aprbkp is a standard Aspen Properties file that uses the NRTL-RK property method. In addition to the organics of interest, the Aspen properties file contains water, the components of air (N2, O2, Ar), and the products of combustion (CO and CO2). All the model parameters are retained at default values; this example demonstrates the predictive capability of Aspen Properties. In addition, it will be easy to add new compounds to the Aspen Properties file.
The calculation procedure in the Excel example is as follows:
1. Air Composition
The composition of air is assumed to be 78.06 mol% N2, 20.99 mol% O2, and 0.95 mol% Ar. The calculated results mainly depend on the O2 concentration in the air; the other components are chemically inert.
2. Chemical Formula
The chemical formulae of the organic compounds have been entered manually into the spreadsheet. As a check on the correctness of the empirical formula, the molecular weight is computed and compared against the value calculated by Aspen Properties.
3. Bubble Point Pressure and Vapor Composition of Liquid The bubble point pressure and vapor composition of the liquid mixture is based an Aspen Properties TVFlash3 function call (with vapor fraction set to zero). TVFlash3 must be used in preference to TVFlash because of the possibility that the liquid will split into two phases at the specified concentration. The concentrations of water and the organic in the vapor phase are assumed to be based upon partial pressures.
An analogous equation is used to calculate the vapor-phase concentration of water.
The above equation assumes that the vapor-phase ideality is not critical to the reliability of the results. This assumption is expected to be valid since flammability calculations are typically done at low pressures, about1 atm.
The concentrations of N2, O2 and Ar are calculated by assuming that they fill the remainder of vapor phase in proportion to their concentration in air.
4. Enthalpy
The enthalpy of the starting vapor mixture is computed using the Aspen Properties MixtureProperties function call. This is the enthalpy that is held constant to compute the adiabatic flame temperature.
5. CO Formation
The chemical formula of the organic is used to compute the extent of the first combustion reaction that forms CO. The reaction extent is limited by the availability of either the organic or oxygen. The Aspen Properties PHFlash function call is used to compute the temperature of the reaction mixture after the formation of CO.
6. CO2 Formation
CO is assumed to react with any remaining oxygen to form CO2. The reaction extent is limited by the availability of either the carbon monoxide or oxygen. The Aspen Properties PHFlash function call is again used to compute the temperature of the reaction mixture after the formation of CO2.
Evaluation and Analysis
Comparison with Experimental Data
All predictive methods must first be evaluated against experimental data before they can be used with confidence. The following chart presents a comparison between data and predictions for the lower flammability limit of ethanol-water mixtures. The calculations were done using the Goal Seek capability within Excel. Note that the guess provided to Goal Seek determines whether the upper or lower flammability limit is found.
Review of the results for ethanol-water indicates that the predicted lower flammability is biased low by about 5A?C. No data are available to test the upper flammability limit, but this quantity is usually less important for safety analysis. Similar evaluations for a range of compounds will establish the confidence with which the predictive method can be used for the compounds of interest.
Training
Non-experts can easily use the Excel model. It thus provides a cost-effective way to give a wide group of technologists a feel for how the flammability limit varies with compound, temperature and composition.
Production of Safety Charts
Groups such as NFPA (National Fire Protection Association) provide guidelines such as: The combustible concentration shall be maintained at or below 25 percent of the lower flammable limit (LFL). It is straightforward to use this Excel model to develop safety charts for a specified organic and safety rules.
Incorporation in Safety Programs
The procedures and methods shown in this example can easily be incorporated into other safety-related software.
Extensions
Additional Compounds
It is easy to add new compounds to the example. These compounds must first be added to the Aspen Properties file (FlammabilityLimit.aprbkp). The user must ensure that NRTL parameters for the water-organic are available, or the bubble point and vapor-phase composition predictions will not be accurate; if not, predictive methods like UNIFAC may be considered. New compounds in the Aspen properties file are automatically available when the Excel file is opened, but the component ID's in the Excel file must be updated carefully; for example note that the chemical formula is entered manually in the spreadsheet.
Multiple Organics
Real mixtures, of course, contain multiple organics. There is no problem with computing the lower flammability limit since it is a good assumption that all the organic material will combust. There is some ambiguity with the upper flammability limit since we have to decide which of the organics will react with the limited oxygen. But, as noted above, the lower flammability limit is of primary concern. Thus there is no problem with extending the example to handle multiple organics.
Other Elements
The present example has been set up to handle compounds containing the elements C, H and O. There is no problem with extending the model to compounds containing other elements if the products of combustion (reaction with O2) are identified
Keywords: Aspen Properties Excel template
Aspen Properties Excel example
References: None |
Problem Statement: Available phases in Aspen Plus include Liquid 1 and Liquid 2. How do I change the phase assigned as Liquid 1 in Aspen Properties and in Aspen Plus simulation environment? | Solution: The default selection done by Aspen Plus is to assign the phase with higher density as Liquid 2. It is possible however to change the basis on which these denominations are attributed, in both Aspen Properties and blocks in Aspen Plus Simulation environment. This is done based on the concept of key component of the second liquid phase. The software will consider the phase with the higher mole fraction of key components as the second liquid phase.
Consider a water-octanol stream (50 kmol/hr of each component).
In Aspen Properties, create a Property Set to display the mass density of each phase.
1. Go to Property Sets. Add PS-1. Select the mass density for the mixture in each phase: MASSRHOM.
2. Run an analysis for the Mixture. Verify that, for example using NRTL method at 25 deg. C, MASSRHOM is 0.846 gm/cc for Liquid 1 and 0.994 gm/cc for Liquid 2.
3. In the Qualifiers page in PS-1, select 1st liquid as Phase of first column, add second column with 2nd liquid as phase. Select 2nd liquid key component for 2nd liquid as Octanol.
4. Run analysis again. Check that Liquid 1 now has a density of 0.994 gm/cc.
This is also possible in blocks, such as the decanter.
1. Attaching the above defined stream to a decanter, and creating two output streams for first and second liquid, verify that second liquid is the water-rich stream.
2. Open the Decanter specification form. In the section Key components to identify 2nd liquid phase, select Octanol. Then, define a component mole fraction as the threshold value (in this case, it was set as 0.5). Run the simulation. Check that the stream connected to the decanter as Liquid 2 is now the organic phase.
Keywords: Liquid-Liquid equilibrium, Properties, Second Liquid Phase, Decanter
References: None |
Problem Statement: How do I get Design Specification's Spec variables in Custom Table? | Solution: When you have large flowsheet and too many Design Specification's then it will be very difficult to change the Target variable value. So it is better to create Custom Table and access those Spec variables from it. You can able to place this Custom Table on flowsheet. Please go through following procedures:
1) First create Custom Table
2) Go to Design Specifications | Spec
3) Then instead of copying specific variable components value you need to right click on the Spec window and select Copy Variables option
4) Then go to Custom Table right click and Paste, you can delete unwanted Spec variables
Keywords: Custom Table, Design Specification, Spec Variables, Aspen Plus
References: None |
Problem Statement: When using MOC from server or client you get the error below. | Solution: This error message refers to a SQL server deadlock. This deadlock can occur for many reasons such as Locks, Worker threads, and memory available.
To resolve this issue there can be many steps.
1. Check Other MOC clients to ensure they are not currently using the resources or tables.
2. Check that the SQL server is running. You may need to consult with your SQL server Admin.
3. If you are running an order you may need to cancel order or wait for the workstation currently accessing the order to finish. If no workstation is active then you may need to remove lock from SQL server.
4. Restart MOC or Restart Machine. Refer toSolution titled
Why do I get the error Error: Module Already running. Module is currently executed by unknown when starting MOC?
5. As this is a deadlocking issue for Sql Server you may also want to refer to the Microsoft Sql server support.
http://msdn.microsoft.com/en-us/library/ms178104.aspx
Keywords: None
References: None |
Problem Statement: How do you enter Fisher data for a control valve? | Solution: Currently, we do not have the Fisher or Fisher-Rosemount control valve libraries available in Aspen Plus.
However, the data for any valve can be specified directly in the User Interface instead of using a valve from the built-in library. This data must be entered for every valve in every flowsheet which of course can be a nuisance.
The valve unit operation uses tables that define these valves. You can customize these tables by changing the appropriate data files. In this case it would be the valve.dat file located in the AspenTech \ Aspen Plus xxxx \ GUI \ CUSTOM directory where ever Aspen Plus has been installed. This file can be used to expand the default valve library which currently does not contain the Fisher data. Data can be added to this file using the fixed format shown the Aspen Plus System Management reference manual, Chapter 6, Appendix A.
Then you would need to process the file using the Aspen plus User Interface Table Building System (MMTBS). For more information about running MMTBS and applying your changes, see Chapter 4 in the same manual.
http://support.aspentech.com/CustomerSupport/Documents/Engineering/Aspen%20Plus/2004.1/AspenPlus2004.1SystemManagementGuide.pdf
Other notes:
The Valve Input Valve Parameters sheet contains a table of valves.
The user chooses:
Valve type
Manufacturer
Series/style
Size
Aspen Plus retrieves these specifications from the table:
Valve position
Flow coefficient (Cv)
Pressure drop ratio factor (XT)
Pressure recovery factor (FI)
Keywords: valve, cv
References: None |
Problem Statement: A good modelling practice is to recurse only the properties that you will use downstream as part of a constraint, in another pool or because it drives the economics of a unit.
These are the main uses of recursed properties:
Delta-Base structure
Blending
Another pool
Limit properties in a submodel, e.g. using table PROCLIM
The error vectors (RqqqTAG) of recursed properties are distributed to the following row types:
EqqqDST, LqqqDST, GqqqDST, RqqqDST, NqqqDST, XqqqDST (qqq is a quality tag, DST is the destination tag)
In most of the cases, if a property is recursed but doesn't have such a destination, it should be removed from the recursion to simplify the model and reduce the use of resources to solve the model.
Exceptions would be SPG to report the right weight in the material balance, or properties called through PROWS for reporting purposes.
How do you identify which properties are not used downstream and are candidates to be removed from the recursion? | Solution: To identify unused recursed properties, do the following:
Run your model.
In the model tree, go to theSolution Files | MPS Files section.
Right-click on the MpsProb.Mps file (this is the initial model matrix file) or on the MpsBcd.Mps (this is the final model matrix final including the results ) and select the option Simplify MPS...
The resulting report will identify all the properties for each pool that are not used in the model. This are candidates for removal from model without impacting the results.
The report shown was created using this procedure:
In this case, for example, it indicates that the property RON is being recursed for ALK. If you look in the model, you will see that the gasolines blend by DON (Road Octane), therefore RON is not being used and the recursion structure in the submodel SALK can be removed as well as the entry in table PGUESS.
A similar analysis for AR1 (Atmospheric Resid 1) shows that in the destinations of this pool, the properties AFC, API, BNT, CON, KFC, PPI, SPV are not used, and therefore are candidates for removal.
In this case, as AR1 is a crude cut, you only have to remove the entries for these properties in table PGUESS for AR1.
When you run this report, the resulting MPS file will have the corresponding recursion structure removed. However, the best use of this report is to identify which properties should not be recursed and permanently removed from the model.
Note: This method just identifies the properties that are recursed but for which there is no error distribution to the following row types:
EqqqDST, LqqqDST, GqqqDST, RqqqDST, NqqqDST, XqqqDST (qqq is a quality tag)
If you are recursing a property to just use it in PROWS, or to use them in the reports (e.g. SPG or SUL), they will also be identified in the list. So in the end, you have to work on the list manually to see which properties are recursed for no use whatsoever and then decide whether they can be taken out from the model or not.
NOTE: The procedure above is specifically for models using Distributive Recursion. PIMS Advanced Optimization (PIMS-AO) does not create the MPSBCD file required for this procedure. If you are running your model in PIMS-AO, you can switch to DR, run the model and then follow this procedure.
Keywords: None
References: None |
Problem Statement: How do I Upgrade a Shortcut Dryer to a Convective Dryer? | Solution: The attached Aspen Plus V8 demo will show you how to upgrade a shortcut dryer to a convective dryer.
Keywords: Solids Capabilities, Unit Operations, Set Up, Screens, Substreams, Particle Types
References: None |
Problem Statement: Can a vibration analysis be performed in Shell&Tube Mechanical? | Solution: A vibration analysis cannot be performed in Shell&Tube Mechanical because the fluid flow parameters are not available in the program. The vibration analysis is always available in our thermal design program, Aspen Shell&Tube Exchanger.
The way in which a vibration analysis can be performed using the final mechanical design is transferring all the geometry from Shell&Tube Mechanical to Shell&Tube Exchanger and running again. This can be done automatically with the Change Model/Transfer Button.
Keywords: Vibration analysis, mechanical design
References: None |
Problem Statement: When using the pipe segment, under ratings, heat transfer in Aspen HYSYS, I saw that it is possible to estimate the heat loss to surrounding based on the insulation of the pipe (Pipe, Inner HTC, Outer HTC, and Insulation). What is the best way to model heat loss in the case of multiple insulation layers? | Solution: To model heat loss across pipe with multiple insulation layer, you are recommended to use Aspen Hydraulics based Pipe Segment. To use Aspen Hydraulics, user should have Hysys Upstream license. The snapshot below shows the location wherein you can add multiple layers for Outer Insulation and Inner insulation. It also provides user to handle different material (and associated properties) of insulation for a given pipe segment.
In case you do not have an Upstream license, as a workaround in Hysys Pipe Segment, you can goto Rating->Heat transfer and select Insulation Type as User Specified (see the below snapshot); manually fill in the data for Effective Thermal conductivity and Effective insulation material Thickness to approximately mimic the actual scenario.
Keywords: Pipe Segment, Multiple Insulation Layers
References: None |
Problem Statement: Mercury is critical among the flow assurance topics. Mercury is very common in natural gas, petrochemical, and refinery feeds. The presence of mercury, even in low level, can cause damage to aluminum equipment by amalgamates with aluminum and creating brittle alloy. Mercury can also poison catalysts. Its emission level is regulated so it needs to be removed before release. | Solution: Using Aspen HYSYS to:
Monitor the behavior of mercury in the fluid
Implement the Mercury Removal Unit (MRU) at appropriate location
Monitor the downstream process
You will learn to:
Leverage Free Mercury Analysis in Aspen HYSYS
Monitor the mercury phase distribution using tools within Aspen HYSYS
Use HYSYS Component Splitter as a shortcut MRU
Keywords: Aspen HYSYS Upstream, Aspen HYSYS, Mercury, Mercury Removal Unit, Mercury Partitioning
References: None |
Problem Statement: Why is the saddle position in the Outline drawing not located in the same location as in the Setting Plan? | Solution: The Outline drawing can be displayed from Results | Drawings | All Drawings | Outline
The Setting Plan can be displayed from Results | Drawings | All Drawings | Setting Plan
Sometimes, the saddle is not located in the same position for both drawings. The outline drawing is not in proportion as the setting plan is.
To match both drawings, the user needs to specify the saddle location in Input | Input | Exchanger Geometry | Horizontal Supports and specify location number 4.
Run again the file. Now the saddle location will be matching in both drawings
Keywords: Setting Plan, Outline, Saddle location
References: None |
Problem Statement: In the BCU Administrator where you bind the characteristic value to a tag or Alias name, there is a checkbox which allows user to select the option Extrapolate to current time. This | Solution: describes the purpose of this option and more importantly the effect it's going to pose on the data getting stored in Aspen Production Record Manager (APRM)Solution
The values of the characteristics are read from Aspen Infoplus.21 (IP.21) database at the time when a trigger is fired. Unless there is a value in IP.21 history with exact timestamp for the specific time the trigger got fired, the value for the characteristic will be interpolated between the two values with timestamps directly before and after the trigger time.
For example, say there are following data points in IP.21:
Timestamp
Data Value
11:05:00
102.5
11:15:00
102.7
11:25:00
102.6
11:35:00
102.4
If a trigger is executed at 11:20:00 it will record an interpolated value between the two data points which bridge the trigger time. This data value depends on the value of the characteristic tag's IP_STEPPED attribute that will be set within IP.21.
If it is set to Stepped, the tag's value is assumed to stay constant until the next data point. In this scenario considering above mentioned data points, a value of 102.7 at 11:15 will be assumed to be constant until next value comes in at 11:25. So, the characteristic value at 11:20 i.e. the time at which trigger executes will be 102.7.
If it is Interpolated, the tag's value is assumed to vary smoothly between the two data points, so the value of the characteristic at 11:20 would be halfway between 102.7 and 102.6 i.e. 102.65.
If there is no value in IP.21 after the trigger timestamp, then the BCU will wait until it sees the next timestamp in the tag history and then interpolate the characteristic value. For example, if the trigger execute at 11:40:00, but there is no data point in IP.21 after this timestamp BCU waits for the new IP.21 value. During this time the unit status in the Scheduler shows waiting for data. Until this value is read, new batches for the unit are not processed and the BCU is stuck waiting to calculate the characteristic.
In most cases, the waiting for data status is fine. The BCU works correctly even if it is behind current time and will usually catch up quickly. However, the situation can turn problematic if you are waiting on a tag that does not regularly update its historical values. For example, if you have a tag that receives its value via SQLplus and the query that sets the tag value may not update regularly. Moreover, when it updates the new value is same as the previous value. During all this the BCU might get stuck waiting for this value for hours, or potentially even days, depending on how often the query is executed.
To work around this, the checkbox Extrapolate to current time allows the user to specify that the characteristic should not be interpolated, but instead extrapolated from the most recent value in history. In this case, the characteristic value is set to equal the value in history previous to the time of the trigger. Again using the above data set as an example, if the trigger conditions are met at 11:40:00 and the Extrapolate to current time option is selected, then the characteristic value for the trigger will be set to equal 102.4, the most recent value entered in history.
This Extrapolate option does prevent holding up the BCU from processing new batches while it waits for a new IP.21 value. In certain circumstances it can potentially lead to incorrect batch data. For this reason, Administrators should be cautious when using this feature.
For example, an extrapolated characteristic value could be significantly different than the value that would have been interpolated. Consequently, you should use the Extrapolate feature when you know, or at least are very confident, that a new historical value will not differ much from the previous historical value.
Problems could also arise when buffering data with Store and Forward. If the Cim-IO / IP.21 connection fails, the tag value will no longer get updated with new values into history until the store files gets buffered completely. If Extrapolate is turned on, the BCU will continue to use the last data point logged prior to the creation of the store file as the characteristic value. If the values for the characteristic tag changed substantially during the buffering process, the data for that characteristic could be incorrect. In this situation, it is possible to obtain the correct characteristic values by rerunning the unit in the scheduler from the time before store file was created.
Keywords: extrapolate
waiting for data
BCU
References: None |
Problem Statement: Can we incorporate component fugacity into adsorption isotherm? | Solution: Aspen Adsorption V9 allows using component fugacity instead of component partial pressure. However the isotherm type “Partial Pressure” still needs to be selected and Aspen Properties need to be used to supply component fugacity values. To enable this options please follow:
Gas_Bed > Configure Block > Isotherm > Use Fugacity in Isotherm > True
Key Words
Isotherm, Fugacity, Partial pressure
Keywords: None
References: None |
Problem Statement: Sulfur recovery is a core process in both midstream and downstream oil & gas industries. In Aspen HYSYS V9, Sulsim technology (by Sulphur Experts) has been fully integrated within Aspen HYSYS. In Aspen HYSYS V9, you can configure, simulate, and optimize the Sulfur Recovery Unit (Claus Process) to recover elemental sulfur from gaseous H2S, COS, CS2, and SO2. Since these gases are harmful to the environment and can affect petroleum product quality, achieving sulfur removal targets and satisfying regulatory requirements across a range of feeds and conditions are critical for operations | Solution: The integrated Aspen HYSYS Sulsim product includes a new Sulsim (Sulfur Recovery) property package. When you add a Sulsim (Sulfur Recovery) property package to your simulation, the component list will automatically include all supported components, allowing you to easily integrate your sulfur recovery simulation with other gas plant processes.
Keywords: Sulfur Recovery, Claus Process, Sulsim, Sulphur Recovery, SRU, Modified-Claus Process, Challenged Feed, Sulfur, Sulphur, H2S, COS, CS2, Reaction Furnace, Catalytic Converter, Tail Gas, Flare, Incinerator, Hydrogenation Bed, Waste Heat Exchanger, Condenser, Titania, Alumina, Selective Oxidation Converter, Sub-Dewpoint, Amine, Adjust, Air Demand Analyzer, ADA, Case Study, Variable Navigator, Kinetic Value, Checker Wall
References: None |
Problem Statement: How is sonic velocity calculated? | Solution: The Mach number is Velocity/Sonic Velocity.
The sonic velocity in a fluid is given by
V = SQRT[g*k*(dP/dRHO)]
where g = dimensional constant
k = ratio of specific heats (CP/CV)
P = absolute pressure
RHO = fluid density
In Aspen Plus, you can control the way in which the heat capacity and the density of the fluid are calculated, thereby improving the results for the sonic velocity.
If you are using a property method for the molar volume that does not have pressure dependency, such as RACKETT, the sonic velocity will not be calculated for the liquid phase. In V7.3 and higher, there is a warning printed when calculating CV, CVMX, CPCV, CPCVMX, SONVEL, SONVELMX, ISENEX, and ISENEXMX when the liquid volume model is not pressure-dependent (as is the case with the Rackett model). With these models, the dV/dP contribution for CV is not calculated because the liquid volume model always provides a zero.
If possible, we recommend using a valid equation of state (EOS) property method to calculate the sonic velocity. For fluids available within REFPROP, REFPROP is the best. Other generalized EOS models such as BWRS should also do well in many cases. For property methods that use the RACKETT model, we would generally recommend switching to the COSTALD method for the volume. So for VLMX the route to be used would be VLMX22 and for VL the route to be used would be VL06.
Here is comparison of a few different property methods for a sonic velocity calculation for water at 25C.
BWRS
NRTL
NRTL(with COSTALD) REFPROP
PRES TOTAL TOTAL TOTAL TOTAL TOTAL TOTAL TOTAL TOTAL
RHO SONVEL RHO SONVEL RHO SONVEL RHO SONVEL
WATER WATER WATER WATER WATER WATER WATER WATER
bar gm/cc m/sec gm/cc m/sec gm/cc m/sec gm/cc m/sec
1 1.00 1501.25 0.99 1524.00 1.00 1864.14 1.00 1496.70
10 1.00 1502.83 0.99 1524.00 1.00 1866.75 1.00 1498.21
100 1.00 1518.31 0.99 1524.00 1.00 1892.74 1.00 1513.27
1000 1.04 1524.00 0.99 1524.00 1.02 2139.08 1.04 1663.07
The sonic velocity for a vapor/liquid mixture is calculated differently. It is the same method used in the PIPE model to calculate MACH number. The method is described in the following reference:
Sonic Velocity in Two-Phase Systems by D. L Nguyen, E. R. F. Winter and M. Greiner, Int. J. Multiphase Flow, Vol 7, pp. 311-320, 1981.
Note that when sonic velocity for total phase is requested, pure component sonic velocity SONVEL may different from mixture sonic velocity SONVELMX for a mixture containing only one component. This occurs when property methods (such as PENG-ROB) are used which use different liquid volume models for pure and mixture. Sonic velocity for total phase uses the Nguyen-Winter-Greiner mixing rule, which requires the liquid volume.
Keywords: property set
sonvel
sonic velocity
mach
cpcv
References: None |
Problem Statement: Operators for the C2 Splitter identified that the ethylene loss increased in the bottom ethane stream recycled back to the furnace. Since the tower has been in operation, the plant suspects the fouling in thermosiphon reboiler accumulates over time which reduces the boil up ratio of the reboiler. To reduce the fouling, the plant needs to use antifouling chemicals in the reboiler at the cost. Due to the low margin pressure in the current business environment, the plant management would like to verify their suspicion backed up by the data before deploying the de-fouling operation of the reboiler. With this initiative, the plant engineering group is asked to come up with ideas on how to estimate the fouling in reboiler. | Solution: An Excel interface for a high fidelity EO (Equation Oriented) Aspen Plus case was created using Aspen Simulation Workbook to monitor the column and exchanger fouling. This interface can connect directly to process historian data, or any other data pulled into Excel, to influence the EO calculations for the model. The interface allows for all aspects of the ethane-ethylene (C2) splitter to be monitored across different operational data points (time stamps). The monitored conditions include, but are not limited to, column flooding, reflux rate, column operating pressure, profitability, and compressor power supply. Further, process data can be used to detect fouling conditions inside of the heat exchange by leveraging plant data, an Aspen Plus model, and an EDR (Exchanger Design and Rating) model.
Keywords: C2 Splitter, Ethane, Ethylene, Column, Flooding, Compressor, Heat Exchanger, Fouling, Operations Support, Equation Oriented, Process Data, IP.21, EDR, Aspen Plus, ASW
References: None |
Problem Statement: I need to have language localization in the MOC. I have changed the Windows localization, but the MOC still appears in English. | Solution: If the MOC does not pick up the Windows localization it is possible to set an override. This is done by enabling the key OVERRIDE_LOCALE in flags.m2r_cfg using the two character iso639 code. The codes available are as follows:
· Chinese - zh
· French - fr
· German - de
· Italian - it
· Japanese - ja
· Korean - ko
· Portuguese - pt
· Russian - ru
· Spanish - es
These codes can be used even if the Windows language pack is not installed, provided you are running APEM V7.3 CP2 or higher.
For example, to set the language to Japanese define: OVERRIDE_LOCALE = ja save the file and then run codify_all.cmd. When the MOC next starts it will be in Japanese.
These changed are typically done on the APEM server and located in
C:\Program Files (x86)\AspenTech\AeBRS\cfg_source folder.
Keywords: MOC
localization
References: None |
Problem Statement: How important is the number of discretization points of adsorbent bed? | Solution: Aspen Adsorption gas bed models are spatially discretized. Discretization methods are used to solve the necessary Partial Differential Equations (PDE) across the bed. Larger number of discretization points will ensure a smaller distance between nodes and hence better prediction of adsorbate distribution on adsorbent surface.
By default, the number of nodes is set to 20. However user can increase it by following: Gas_Bed > Configure Block > Configure > General
Although increasing number of nodes improves model accuracy, however it slows the simulation run speed. So the user is advised to make balance between these two performance criteria.
Key Words
Gas_Bed, Nodes, Discretization
Keywords: None
References: None |
Problem Statement: Even after a Windows account has been added to an Aspen Production Execution Manager role in Security with MOC privileges, log-in can fail with the message:
Full name required for valid Aspen Production Execution Manager users | Solution: One of the requirements of 21 CFR 11 compliance is as follows:
11.50 Signature manifestations.
(a) Signed electronic records shall contain information associated with the signing that clearly indicates all of the following:
(1) The printed name of the signer;
Since Windows log-in identities are not always equivalent to the printed name of the signer, it is required that the Windows account Full Name field be used to help enable compliance:
The Full Name is then captured as part of the audit trail so that the printed name is easily associated with the action taken. Here for example, is the audit message generated when the above account is used to copy an RPL:
Notice in the last line how it is clear that the user Jack Johnson, user account jackt432a, part of domain dobease3 performed the specified Created action shown in the RPL audit trail.
Keywords: Full name required for valid AeBRS users
APEM
lock out
failure
validate
References: None |
Problem Statement: This | Solution: outlines new features in the Aspen Production Record Manager V8.4 administrator.Solution
Aspen Production Record Manager V8.4 introduces the following new feature:
Enhancing the Management of the Production Record Manager Services
Aspen Production Record Manager Administrator has been enhanced to handle the management of the Production Record Manager services. A new node name Local Services has been added to the console tree. The Local Services node contains the following sub-nodes:
Server: Administrates the Aspen Production Record Manager Service.
BCU: Administrates the Aspen Production Record Manager BCU Service.
ODBC: Administrates the Aspen Production Record Manager ODBC Service.
Extractor: Administrates the Aspen Production Record Manager Extractor Server.
Batch Connect: Administrates the Aspen Production Record Manager Batch Connect for OpenBatch Service.
BPD: Administrates the Aspen Production Record Manager BusinessProcessDocument Service.
By using the Local Services node, you can manage the Aspen Production Record Manager services from the Aspen Production Record Manager Administrator directly instead of from the Services control panel. In addition, BCU properties and Extractor properties can be set by using the respectively sub-node.
The Local Services node functions only when you manage the services on the local machine with administrator privileges.
Keywords: APRM
Aspen Production Record Manager
V8.4
Batch.21
Administrator
BCU
ODBC
Extractor
Batch Connect
BPD
References: None |
Problem Statement: After a successful installation, when trying to start MOC as a user, it fails to open and you may see messages like:
Access Denied
Access was refused
System could not find specified path
Java Virtual Machine Launcher : Could not find the main class. Program will exit!
This can happen even though other client machines are connecting successfully to the same AeBRS server.
The best eBRS troubleshooting tool is typically hints from the MOC debug file. However in cases like these where MOC does not open up completely, the process which writes debug has not even started yet. So these problems are more of a challenge to analyze. In this case, edits to the AeBRS.cmd file, for troubleshooting purposes, can help understand what is happening:
1. Do a copy and paste on the AeBRS.cmd file so you can revert easily to the original after troubleshooting.
2. Change this line, removing the letter W:
SET J=C:\j2sdk1.4.2_04\bin\JAVAW.EXE
becomes
SET J=C:\j2sdk1.4.2_04\bin\JAVA.EXE
3. To keep the process from opening in a separate window, also edit the START line. Your memory usage parameters may vary from this example line. The key edit here is just getting rid of the START-related information in the first part of the line:
START= /i AeBRS MOC %J% -Xmx256M -Xms64m -XX:NewSize=5m -cp
becomes
%J% -Xmx256M -Xms64m -XX:NewSize=5m -cp
4. Additionally, add a pause command as the last line of that file, so that the DOS window stays open after execution.
Now doubleclick on the MOC shortcut (or Start, Programs, etc.) which will now use your edited AeBRS.cmd file, and leave an open DOS window that can be reviewed for helpful troubleshooting information.
Once the problem is resolved, delete the AeBRS.cmd file, and rename Copy of AeBRS.cmd back to AeBRS.cmd.
Here is an example problem that can be diagnosed from this DOS window:
at java.io.WinNTFileSystem.createFileExclusively(Native Method) at java.io.File.createNewFile(File.java:825)
Could not open/create prefs root node Software\JavaSoft\Prefs at root 0x80000002 Windows RegCreateKeyEx(. . .) returned error code 5.
Could not open windows registry node Software\JavaSoft\Prefs at root 0x0000002 Windows RegOpenKey(. . .) returned error code 2. | Solution: This error condition can result when a strict user policy prohibits the user from writing to the AeBRS folder and all child sub-folders. It is necessary, for AeBRS to work correctly, to grant Read/Write/Change privileges to this folder, and all subfolders:
C:\Program Files\Aspentech\AeBRS
Keywords: freeze
fail
access denied
crash
install
References: None |
Problem Statement: When manipulating a .CCF, either with Simulate or some other advanced control process, you get error 15141. You might also get Combuf Duplicate Usage, and it might even be extended to describe a specific tag that is causing the problem. Or as in example used below in the | Solution: Solution
This error is caused by a duplicated name in the Common Buffer. In this case, the following message was seen in the pop-up window when trying to save from Build.
--- Internal Validation Finished. No errors detected. -----
Checking the CCF with DMCplus Simulate.
24-Sep-2007 11:30:51 GETCCF 15141 ComBuf Duplicate Usage DB013
Error. DMCINI Failed - 15141
This indicates that the entry in question is DB013. The issue is that the user had defined DB as a user defined entry for a number of variables in both the Independent and Dependent sections. CV 13 was the first one where there was also a DB in the MV section so you got this error... i.e. there was a DB in the 13th array index for both the MV's and the CV's. Rename the entries to include MV in their name (for the MVs) and CV in there name (for CVs). For this example: DBMV and DBCV This will correct the error.
A similar error was found specifically mentioning DMCINI 15141 when trying to simulate a ccf from Build, opening a CCF directly with Simulate, or loading a CCF with Manage. However, the specific parameter name and MV/CV number combination was not being mentioned. In this instance, it is still recommended to run through the number combinations for all USER parameters in the CCF.
When building the controller context (for instance, Build validating using Simulate, Simulate, Manage load), every parameter has to have a unique name. For Independents and Dependents, the position number is appended to the parameter name. That is why Independents and Dependents have different parameter names for the same indication, such as the critical indicator, CRIIND and CRIDEP. Internal to the controller, these become CRIIND001, CRIIND002, CRIIND003, etc., CRIDEP001, CRIDEP002, etc.
To fix a CCF experiencing the DMCINI 15141 error, change the USER parameter names. An easy way to identify the user-defined parameters is to sort on the type column in Build. The user-defined entries will say USER.
A good practice is to always append IND or MV to Independent parameters and DEP or CV to Dependent parameters. For example, rather than calling them APPLE in both MV's and CV's, it would better to call them APPLEIND and APPLEDEP.
Keywords: None
References: None |
Problem Statement: Is it possible to know whether the Aspen HYSYS Petroleum Refining license is being used or not? | Solution: In order to know if the Aspen HYSYS Petroleum Refining license is being used, go to File | About | Licensing Information. The Licensing Information window will be displayed and it will show all the active licenses being used in the simulation case.
The Aspen HYSYS Petroleum Refining license may be found as ‘REFSYS’. If the REFSYS license is being used, then it will appear listed.
Keywords: Petroleum Refining, REFSYS, Petroleum Assay, License.
References: None |
Problem Statement: How maldistribution risk is handled when twin headers are used? | Solution: Maldistribution risk can be reviewed in Results |Thermal/Hydraulic Summary | Pressure Change | Main distributors tab
Twin headed distributors are also used in some cases, they are made up of basic distributor form repeated in each half the exchanger width.
Distributors with twin headers are handled the same way as other centrally symmetric distributors. The right hand pressure change reported in the results is actually the central path pressure change.
Maldistribution risk is assessed based on the total path pressure changes, not the individual distributor pressure changes. All of the left hand changes and right hand changes are summed and compared to the mean. If either side is too far off of the mean, it is a maldistribution risk and “Yes” will be displayed.
Keywords: maldistribution risk, plate fin, twin headers
References: None |
Problem Statement: When running cases under AO/XLP, the following message is encountered on one machine and it is impossible to continue work:
 Failed to validate | Solution: ID in PrSolutionID table. Please run the model again and generate the reports
What is the meaning of this message and what needs to be done in order to successfully continue work?
Solution
There can be several underlying causes for seeing this message. The first sentence in this message actually quite precisely describes what Aspen PIMS has encountered and why it stopped. The second sentence is providing a remedy for the most common cause of this behaviour which is a user error. However we have identified over the years other situations which may lead to this message being invoked. Here we will list all known causes up to this point.
1. Interrupted run - this is the most benign and most common of the situations. User has interrupted the execution of a set of cases on one of the cases. Sometimes this may lead to database corruption. If the user attempts to run just the reports (Run/Reports...) for the cases which did solve successfully this message will be invoked and the remedy is to follow what it says: run the model again. This has been experienced occasionally under DR.
2. Running AO features Goal Programming or Feedstock Basket Reduction without setting the Access database maintenance to: KEEP EXISTING. Remedy: set Access database maintenance to: KEEP EXISTING.
3. Experiencing the error message after a PIMS upgrade. Sometimes during and Aspen PIMS version upgrade, the database changes too or it may become corrupt. Remedy:
· Go to the model folder
· Select the file Results.mdb
· Delete this file
· Run PIMS again
4. Running PIMS on a machine with an unsupported version of Office. This error comes up when the database is not populated correctly. Remedy: install a supported version of Office.
5. Using a SQL database that was not created properly. Test: Use the same machine to run our Volume Sample model with Access database to verify that everything works correctly. Remedy: recreate SQL database.
6. Caseparallel.exe crashing - this can be verified by trying to run under AO/XLP and with trace level set to medium (Model settings -> Non-linear Model (XNLP), Advanced tab: Trace Level field: Medium). If the CaseParallelRankZero.log is not created this is a clear sign that Caseparallel.exe crashed. So far we have experienced and documented three underlying causes:
A) Corrupted CaseParallel.exe. Remedy: replace the CaseParallel.exe file with a one from a machine where PIMS runs fine.
B) System file corruption. One of the system files might be corrupt. The remedy here is to use Windows system file checker utility to repair any corrupt file:
1. Open a Command Window in Administrator mode:
* click Start
* click All Programs, then Accessories
* right click on the Command Prompt option,
* on the drop down menu which appears, click on the Run as Administrator option.
* If you haven't disabled User Account Control (and you should not!) you will be asked for authorization. Click the Continue button if you are the administrator or insert the administrator password.
2. Start the System File Checker
* In the Command Prompt window, type: sfc /scannow,
* press Enter.
* You will see that the system scan will begin.
The scan may take some time and windows will repair/replace any corrupt or missing files.
C) Conflict with Windows' Customer Experience Improvement Program - there is a known issue of an exception being thrown due to a conflict with this Windows feature. Test is actually to apply the remedy and see if it helps:
1. Go to control panel and go to system and security.
2. Click Action Center > Change Action Center settings.
3. Click Customer Experience Improvement Program settings.
4. Select No, I don't want to participate in the program and click Save changes.
5. Start the control panel and click Administrative Tools > Task Scheduler.
6. In the Task Scheduler (Local) pane of the Task Scheduler dialog box, expand the Task Scheduler Library > Microsoft > Windows nodes and open the Application Experience folder.
7. Disable the AITAgent and ProgramDataUpdater tasks.
8. In the Task Scheduler Library > Microsoft > Windows node, open the Customer Experience Improvement Program folder.
9. Disable the Consolidator, KernelCEIPTask, and Use CEIP tasks.
10. After this PIMS should run without issues.
If none of the above suggestions resolve the issue and the CaseParallelRankZero.log file was not created, then please collect the following information and send it to [email protected].
What version of PIMS is installed?
Was PIMS installed on a new machine or was it an upgrade on a machine with a prior version of PIMS?
Send a copy of the XSLP_Control.log found in the model directory
Send the event viewer details from both the windows logs node and also the application. To get the event viewer details open the event and in that dialog, go to the details tab. On that tab there will be a COPY button at the bottom. Click the COPY button and paste the contents into a text file.
Keywords: CaseParallel.exe, Case stacking, Error
References: None |
Problem Statement: Why are the wind and seismic loads not applied to the supports by default in Aspen Shell&Tube Mechanical? | Solution: Wind and Seismic calculations depend on specific input parameters that govern how big of a load the equipment will experience.
If these parameters are not known for the site where the equipment is going to be installed, the calculated loads may be too large or too little.
However, if the designer needs to have these loads applied, even if not specific to the installation site, then turn the option to ‘Yes’ under Input | Program Options | Loads-Ext/Wind/Seismic/Weights/Ins-Den | Seismic Loads tab.
Keywords: Wind and Seismic loads, Program Options, Input
References: None |
Problem Statement: How do you switch the version for Aspen Properties in Excel? | Solution: When you have multiple Aspen Properties versions installed, the procedure for switching from one version to another is as follows:
1. Open an empty Excel and from Options | Add-ins, choose Excel Add-Ins in the Manage field and click Go. Then, uncheck Aspen Properties. This will release Aspen Properties Excel Add-in from the current version. Quit Excel.
2. Set the version by going to Start | All Programs | AspenTech | Process Modeling Vx.x | Aspen Properties and run the Set Version - Aspen Properties Vx.x
3. Open an empty Excel again, from Options | Add-ins, choose Excel Add-Ins in the Manage field and click Go. Then, click Browser to find the current Aspen Properties Excel Add-in file Aspen Properties.xla. This will install the current Aspen Properties Excel Add-in. Quit Excel.
Now you can open any Excel Add-in file that was created with an version earlier than the current version. For instance, the file was created with V7.3, you can reopen it with V8.0, V8.2, V8.4, V8.6, and so no.
When you reopen an Excel Add-in file on a different machine or with a different version configured, you may see some errors initially. This is because Excel tries to use the original settings in the file. If Excel cannot find the original settings, it will issue an error. But don't worry too much about these warning/error. Excel usually automatically tries to fix the problems with settings and recalculates all calculations. So you just need to be patient and let Excel finish fixing the problem and updating the calculations and it may take a couple to a few minutes.
An example is attached. You can follow the steps described to see how it works when you switch from one version to another.
1. Configure Excel with Aspen Properties Excel Add-in V8.4, including linking to C:\Program Files (x86)\AspenTech\Aspen Properties V8.4\Engine\Xeq\Aspen Properties.xla:
2. Quit Excel.
3. Copy the files to a local folder :
4. Now double click Vertical-Ranges.xlsx. Click Yes when you see this dialog:
5. You'll see the following:
6. The link to Vertical-Ranges.aprbkp still points to the original path on the user’s machine (your machine) and the vertical cells for mole-frac are shown as #VALUE!. The message says the aprbkp does not exist at the original location, WHICH IS CORRECT. Just click OK.
7. After you click OK, Excel will start updating and correcting the above problems. It may take a couple of minutes. So just be patient to wait and do not do anything else. Once it is done, you'll see the following. The path to Vertical-Ranges.aprbkp is fixed. The vertical cells for mole-frac are also fixed with the correct values. Just click Close.
8. Now I'll do another vertical cell calculation as shown below:
9. The results are the same and correct:
10. Notice that a new aprpdf file is created:
Keywords: None
References: : CQ00584540, CQ00584544, CQ00584539 |
Problem Statement: What is new in Glycol Package in Aspen HYSYS? | Solution: The glycol package in Aspen HYSYS contains several improvements from the previous versions. This knowledge baseSolution is intended to highlight these improvements.
- Estimation of interaction parameters (IPs) has been improved. The estimation is done using the UNIFAC model. However, previously the HYSYS database does not contain any UNIFAC functional group information for more than 600 compounds. This prevented the IPs to be estimated in the older versions. The necessary UNIFAC data have now been added allowing the IPs to be be estimated.
- IPs developed for the Clean Fuel package, which covers mercaptanes and sulfides have been incorporated into the Glycol package. This comprises 83 IP values.
- We found that IPs for many hydrocarbons with water resulted in very large deviation in predicted solubility of the hydrocarbons in the aqueous phase. Improvements are made for water with C6 to C20 by removing the activity coefficient IPs and using kij.
- The Glycol package is very sensitive to missing IPs. There are large gaps between light gases with hydrocarbons and hydrocarbons with hydrocarbons. Therefore, new IPs (kijs) have been added for 244 binary pairs.
Note: If a simulation was created in a older HYSYS version then it may be necessary to reset the IPs via the Binary Coeffs page under the Fluid Package (see screenshot below) in HYSYS to make the new IPs available for the simulation.
Keywords: TEG Dehydration, Interaction Parameters (IPs), Clean Fuel Package
References: None |
Problem Statement: When using the compressible gas pipe unit operation with the compressible gas method, the mass density reported on the Performance tab > View Profile is incorrect. If we change the method to perfect gas, the mass density gives a correct value. Why is that? | Solution: This happens due to the flash calculation of the inlet stream. The perfect gas method always uses the ideal gas equation for density, so the calculation is not affected by the flash result of the inlet stream. On the other hand, the compressible gas method uses the flash results of the inlet pipe to calculate the density.
The compressible gas pipe operation performs a Pressure - Temperature (P-T) flash. If the inlet stream is calculated with a Pressure - Vapour fraction (P-Vf) flash, then the P-T flash for the same composition can give different results of vapour fraction. This is mainly when the inlet stream is close to a phase boundary.
For example, when the pressure and vapour fraction of 1 is specified in the inlet stream and the stream is close to a phase boundary, the P-T flash of the same stream can give a result of vapour fraction of 0 (see below).
TheSolution is to make sure that the flash calculation for the inlet stream is P-T. Hence, the user needs to specify pressure and temperature to define the stream.
Keywords: Compressible gas pipe, density, compressible gas method
References: None |
Problem Statement: How do I find a block or stream in a large flowsheet? | Solution: If you are working in a large model, it may be difficult to locate a particular block, stream, or other object. You can use the Find Object command to help find an object:
Right-click a blank area of the flowsheet and then click Find Object. The Find Object window appears.
Initially, the search is set to Plant, the top level of hierarchy encompassing the entire project. You can limit the search by selecting a hierarchy or a folder representing a specific kind of object from the tree at the left. If you want to limit the search to only the selected hierarchy, clear the Search in Hierarchies box; with the box checked, the search also looks inside deeper hierarchy levels.
Type a name or partial name in the search box at the top. As you type, objects whose names contain the string typed so far as listed in the box below, along with their types and full hierarchical names.
Move the mouse over any search result. A tooltip will appear containing the description and status of the object.
Click the short name in any of these search results to jump to that object in the flowsheet, or open the default form for objects not displayed on the flowsheet.
The search includes all named objects in the current environment. In addition to blocks and streams, you can find Calculators, Design Specs, Reaction blocks, etc. Find Object is also available in the Properties environment where it is limited to the kinds of objects available in that environment.
Video demonstrating Find Tool in Aspen Plus
Keywords: locate
References: None |
Problem Statement: In the Naphtha reformer model for a semi-regen reactor type, the user can enter the midpoint weighting factors for the reactors in the catalyst page located in the Reactor tab . What are the midpoint weighting factors of the reactors and what values should you enter? | Solution: The midpoint weighting factor default value is 0.67, but it can be changed if the user wants. The value should always be between 0 and 1. This value should be the same in both calibration and simulation.
The reformer model considers the Start-Of-Run (SOR) and End-Of-Run (EOR) conditions as applied to the catalyst life. The start point tells the coke on catalyst at the beginning and the end point tells the coke on catalyst at the end, but it is the coke on catalyst at the midpoint that is actually affecting the yields. The default value is 0.67 because coke builds up increasingly, so it was determined that 2/3 was about the time of the midpoint for coke formation.
This can be changed by the user, but it may be best to not change it by too much. By setting the midpoint to 1, that means that the whole simulation is running at the end point conditions.
Keywords: midpoint weighting factor, reformer, coke on catalyst
References: None |
Problem Statement: What should I check if I am receiving notification of a negative pressure in the pipe segment? | Solution: Usually, this can be alleviated by decreasing the flow rate or increasing the diameter. Other option is to increase the inlet pressure. Check these numbers first and adjust according to your needs.
If the status bar of the pipe segment unit operation says Not Solved and you get the error in the trace window Negative pressure calculated on increment....., probably the flow is choked.
Go to Design | Calculations and toggle the Check Choked Flow check box. Now, if the status bar of the pipe-segment says Flow is Choked, please refer to the following Article 000029597
Keywords: Pipe Segment, Negative, Absolute Pressure
References: None |
Problem Statement: When creating a Stream Summary Template from scratch without using a pre-defined template using the Manage Template and clicking on ‘New’ button:
The properties that you have added will not be shown on the generated Stream Summary Template. | Solution: In order to show those added properties to the new Stream Summary Template, at least the ‘Show Total Stream’ option has to be select/toggle on under ‘Edit Stream Summary Template’ window in the Scope (tab), as the screenshot below shows:
Otherwise those added properties will not be produced when using this template.
The steps recommended as best practice to create an user Stream Summary Template are:
In order to start a new Stream Summary Template select a pre-defined or other user template as starting point and click on ‘Save as New’ under Stream Summary (ribbon) | Template (sub-group):
Provide with a name and description for the new created template.
As soon as you click on OK having provided with the name and description, the ‘Edit Stream Summary Template’ window will open. Otherwise, it can be open clicking on ‘General Options’ in the ‘Stream Summary Options group’.
Select and navite through the appropriate tab and select or de-select, add or remove the desire properties, calculation, label options, qualifiers, etc.
Always click on the ‘Apply’ button, in order to make changes.
Finally, click ‘OK’ to apply those changes to the new created template.
Keywords: Stream Summary Template, Add properties, Scope, Show Total Stream.
References: None |
Problem Statement: Getting Fortran compiler error when using PIPESIM in Aspen HYSYS Upstream. | Solution: When using PIPESIM link in Aspen HYSYS Upstream, user may see this dll error, as shown below:
Warning during libifport.dll initialization: Non-compatible version of Fortran libifcore DLL was loaded to process address space.
In the case, if you copy the two files libifcoremd.dll and libmmd.dll from the PIPESIM directory into C:\Windows\system32, the problem goes away.
Keywords: PIPESIM, DLL error
References: None |
Problem Statement: How do I configure a PID controller with multiple process variables and one control valve? | Solution: You can only connect one process variable (PV) to each PID controller in Aspen HYSYS. If you need two PV then this can be configured with two PID controllers and the output (OP) from each can be connected to a selector block. The output from the selector block can be linked with the control valve.
The selector block can process the output based on the mode chosen in the Parameters page as shown in the screenshot below.
Â
The user can also apply the Hand Select mode. This mode allows the user to select output (OP) from a specific controller. The signal can be manually switched or can be linked with a spreadsheet where the switch can be configured with the conditional statement.
Keywords: PID Controller, Selector Block
References: None |
Problem Statement: I am using the new Process Data COM add-in within Excel and I unexpectedly receive the error Data source name is either empty or referencing empty cell(s). This does not happen with the old (legacy) add-in when I reference the very same data source. | Solution: There is a known issue with the use of data source names (or indeed tag names) which are also valid Excel cell references. The legacy add-in did not allow cell referencing of the data source and so data source names that resembled Excel cell references were not a problem. With the introduction of the COM add-in, cell referencing for data source names is now supported to bring it into line with the existing tag name functionality. Thus, this unexpected behavior has also been introduced.
Example: A data source named IP21 would work fine with the legacy add-in but with the COM add-in you are likely to get the error:
Data source name is either empty or referencing empty cell(s)
This happens because cell address IP21 is most probably empty - ie. the data source name has been misinterpreted as an Excel cell reference. This is not what you intended but is a mistake that is easily made - especially since the drop down box currently lists the data source in a manner that leads to its misuse and confusion.
There are two approaches you could take.
Surround the data source name with double quotes, eg. IP21
Use cell referencing correctly (in the example above, set data source to Sheet1!B2 and then specify IP21 in cell B2).
Keywords: Tag name
ADSA data source
Cell reference
Excel
Add-in
References: None |
Problem Statement: I ran Ranging Analysis on my crudes, but I don't really understand the output. Can you please explain what the Flexibility Index and Utility Index numbers mean and how I use them? | Solution: When Ranging Analysis completes it provides output on the selected variables including a Flexibility Index and a Utility Index value.
Flexibility Index is defined as the fraction (in terms of total feed) of the optimum amount that can be displaced for a predefined drop in the economic objective function. For example, if my optimalSolution says I should run 100,000 B/D of WTS, the Flexibility Index would be a fraction indicating how much of that 100,000 B/D I could remove and still achieve within 1% of my original objective function. The 1% is user designated. If I could remove all of the WTS and run other crudes instead and still achieve within 1% of my original objective function, then the Flexibility Index for WTS would be 1. If I could not remove any of the WTS without losing more than 1% of the objective function, then the Flexibility Index of WTS would be 0.
The Utility Index is defined as the fraction of the remaining total feed that this particular crude could displace for a predefined drop in the economic objective function - typically 1%. For example, I have an optimalSolution that says I should run 100,000 B/D of WTS, 50,000 B/D of WTI, and 25,000 B/D of Brent crudes. If the Utility Index of Brent is 0.6, then this means that I could replace 60% of the feed that is not Brent with additional Brent. In this case, that would mean that 60% of (100,000 B/D WTS + 50,000 B/D of WTI) could be replaced with Brent crude. If I made this replacement, my economics would stay within my predefined threshold of my original objective function.
These indices give information about which crudes are critical to have in your plan. For example if a crude has a low flexibility index, then little of that crude can be removed, or your objective function will drop below your threshold. Therefore it is critical to secure the source of this crude. If a crude has a high flexibility index like 0.9, then almost all of that crude could be replaced with other crudes and it is less urgent to secure.
When replacing a high flexibility index crude, the high utility index crudes are good candidates. These are crudes whose rate can be substantially increased without dropping below your objective function threshold.
Keywords: None
References: None |
Problem Statement: Why does Shell&Tube Mechanical consider 1.0 as the value for the joint efficiency for both longitudinal and circular seams when there is no radiography used, the material is ERW (Electric Resistance Welding) pipe and the welded joint is a corner joint? | Solution: ERW pipe is equivalent to seamless (ASME BPVC VIII-1, UW-12(d)).
Please refer to table UW-12 from the ASME code. Corner joints (type no.7) shows N/A for joint efficiencies category C and D welds with note 5, repeated here for convenience.
(5) There is no joint efficiency E in the design equations of this Division for Category C and D corner joints. When needed, a value of E not greater than 1.00 may be used.
This is the reason why Shell&Tube Mechanical considers joint efficiency value as 1.0 for this construction.
Keywords: Joint efficiency, ERW pipe, longitudinal seams, circular seams
References: None |
Problem Statement: What does warning 873: “The Minimum design Metal Temperature (MDMT entered is lower than the MDMT calculated” mean? | Solution: Aspen Shell & Tube Mechanical calculates all components MDMT to qualify for the exemptions to Charpy Impact Tests. If the component material is Charpy Impact tested, the program does not take that into consideration.
Keywords: Warning 873, Charpy, MDMT, Impact Test
References: None |
Problem Statement: Is it possible to display Russian characters in Aspen SQLplus Query Writer? For some versions of Aspen SQLplus, patches are available providing localization for Russian, but these only refer to Aspen SQLplus Reports. | Solution: It is possible to display Russian characters in the Query Writer. To do this, on the Aspen SQLplus client machine go to the Windows Control Panel and open Region and Language. In the Region and Language window on the Formats tab choose Russian (Russia). On the Keyboards and Language tab make sure the Russian keyboard is available. Note it is not necessary to change the display language to Russian. Then go to the Administrative tab and click the Change system locale button and select Russian (Russia). It is necessary to restart the system for the changes to take effect.
Once this has been done, you can select the Russian keyboard from the language bar, enter Russian characters in the aspen SQLplus Query Writer and display Russian characters in the results.
Please see article 145980-2 for more information about localization with Aspen InfoPlus.21 and the additional steps required to successfully update records in the database.
Keywords: Chinese
Japanese
Korean
References: None |
Problem Statement: When activity coefficient models are used outside the range where parameters were fit, results can be completely unreliable.
Discussion:
In the attached example file, Property Analysis is used to tabulate the properties of an equimolar mixture of methanol and heptane at 1 atm and a temperature range of 25 to 300 C. The property method selected is NRTL-RK, and NRTL parameters are retrieved from the NRTL-RK databank. This file was selected because the parameters extrapolate particularly badly with temperature.
When using any of the binary databanks that has been regressed by Aspen Technology, users can determine the range of conditions used to fit the parameters using the following method:
Go to the Properties/Parameters/Binary Interaction/NRTL-1 form.
Click in a cell containing any of the NRTL parameters.
Press F1.
The online help shows that the data used to fit the parameters were isobaric at 760 mmHg, the compositions ranged from 0.05 to 0.95 mole fraction, and the temperature range was 58.8 to 98.4 C. The statistics on the fit are good, and a quick check with the literature shows that the parameters fit the temperature and composition of the azeotrope well. It should be noted, however, that it is usually a poor idea to fit temperature-dependent parameters to isobaric data such as this, and this is the root of the extrapolation problems we will see.
The Property Analysis results show some odd behavior. Everything looks good at temperatures from 20 to perhaps 140 C. Between 20 and 50 C we see only liquid phase results, and from 50 to 70 C we see vapor-liquid equilibrium results. From 70 to 140 C we see only vapor. But at 150 C and above, a liquid phase returns, and this liquid phase persists to the highest temperatures studied. | Solution: The predicted reappearance of liquid at high temperatures is the consequence of using an activity coefficient model outside its range of validity. The property table shows the calculated activity coefficients at 150 C for both methanol and heptane are approximately 0.17, and the values of the activity coefficient continue to decrease as the temperature is raised. The unreasonably low activity coefficients are causing the false prediction of liquid at high temperatures. Extrapolating these parameters to lower temperatures also causes problems.
How can a user avoid this sort of problem?
Know the conditions under which parameters were fit. The Graphical User Interfaces requires that the user view the binary parameter forms before running a simulation -- use the online help at this time to check out the range of data regressed. (Users who regress their own parameters are encouraged to use the TLOWER and TUPPER fields on the binary interaction parameter form to record the range of validity of their parameters).
Do not extrapolate activity coefficients too far in temperature. Even when parameters have been fit prudently, extrapolation of more than 50 C is likely to result in unacceptable loss of accuracy, and there can be problems with even smaller extrapolations. For examples such as this where the data have been over-fit, problems are more likely.
Parameters should be regressed at conditions as close as possible to the actual process conditions. If the built-in parameters do not cover the conditions of interest, users should obtain data at those conditions and use it to evaluate the performance of the built-in parameters. The data can also be used in the Data Regression System to determine new parameters.
Regress as few terms as possible. Activity coefficient equations are non-linear. Fitting higher order terms can make it more likely for the equation to extrapolate in unpredictable ways or to have multipleSolutions.
Keywords: data regression
gamma model
References: None |
Problem Statement: When creating a Stream Summary Template from scratch without using a pre-defined template using the Manage Template and clicking on ‘New’ button:
The properties that you have added will not be shown on the generated Stream Summary Template. | Solution: In order to show those added properties to the new Stream Summary Template, at least the ‘Show Total Stream’ option has to be select/toggle on under ‘Edit Stream Summary Template’ window in the Scope (tab), as the screenshot below shows:
Otherwise those added properties will not be produced when using this template.
The steps recommended as best practice to create an user Stream Summary Template are:
In order to start a new Stream Summary Template select a pre-defined or other user template as starting point and click on ‘Save as New’ under Stream Summary (ribbon) | Template (sub-group):
Provide with a name and description for the new created template.
As soon as you click on OK having provided with the name and description, the ‘Edit Stream Summary Template’ window will open. Otherwise, it can be open clicking on ‘General Options’ in the ‘Stream Summary Options group’.
Select and navite through the appropriate tab and select or de-select, add or remove the desire properties, calculation, label options, qualifiers, etc.
Always click on the ‘Apply’ button, in order to make changes.
Finally, click ‘OK’ to apply those changes to the new created template.
Keywords: Stream Summary Template, Add properties, Scope, Show Total Stream.
References: None |
Problem Statement: Getting Fortran compiler error when using PIPESIM in Aspen HYSYS Upstream. | Solution: When using PIPESIM link in Aspen HYSYS Upstream, user may see this dll error, as shown below:
Warning during libifport.dll initialization: Non-compatible version of Fortran libifcore DLL was loaded to process address space.
In the case, if you copy the two files libifcoremd.dll and libmmd.dll from the PIPESIM directory into C:\Windows\system32, the problem goes away.
Keywords: PIPESIM, DLL error
References: None |
Problem Statement: The tasks created in the model folder can’t be activated. | Solution: Tasks created in the model folder can only be activated in the flowsheet folder by click expanding the blocks and viewing the contents.
Keywords: Task, model folder, activation
References: None |
Problem Statement: Why do I see less sulfur conversions in the Thermal Stage result in the Performance Summary compare to the results in the Furnace, as shown below: | Solution: User may see this difference if the Allow simulation of back-reactions in the WHE is checked in the Waste Heat Exchanger in the thermal stage, as shown below:
By default, the Allow simulation of back-reactions in the WHE check box is selected; you can clear it if desired. Various reactions occur in a typical Waste Heat Exchanger. These reactions include H2 + S re-association reactions and CO + S reactions. Select this check box to use empirical models (originally created by Sulphur Experts) which predict the extent of these reactions for your simulation. We recommend that you select the Empirical furnace model for your Reaction Furnace when this check box is selected, since these models were developed directly with the Empirical Kinetic reaction furnace models.
Keywords: Sulsim, Sulfur conversion
References: None |
Problem Statement: How do I obtain critical properties values from Excel VBA? | Solution: The code for accessing critical properties utilities are given below. Please refer KBSolution 112361 for more information on VBA automation for linking Excel to Aspen HYSYS.
Option Explicit
Dim hyApp As HYSYS.Application
Dim hyCase As HYSYS.SimulationCase
Dim hyFlowSheet As HYSYS.Flowsheet
Dim hyUtility As HYSYS.CriticalPropertyUtility
Dim hyStream As ProcessStream
Dim Pc As Double
Set hyApp = GetObject(, HYSYS.Application) 'Only works if the HYSYS case is open
'Get the currently open case
Set hyCase = hyApp.ActiveDocument
Set hyFlowSheet = hyCase.Flowsheet
'Create a Critical Property Utility
Set hyUtility = hyCase.UtilityObjects.Add(Critical-1, CriticalPropertyUtility)
For Each hyStream In hyFlowSheet.Streams
'If this is a material stream
If Not hyStream.IsEnergyStream Then
'Attach the stream
hyUtility.AttachedStream = hyStream
'An example to get Pseudo Critical pressure value. Values can be displayed in the spreadsheet cells
Pc = hyUtility.PseudoPCValue
End If
Next hyStream
Keywords: VBA, Critical Properties Utility
References: None |
Problem Statement: Â How do you run the Refinery Reactor models on a 32 bit operating system? | Solution: The 32 bit machine by default only allocates up to 2 GB of memory for any given application, whereas 64 bit operating system allows up to 4 GB for a 32 bit application. The 32 bit operating systems have a flag that can be set to allow an application to use up to 3 GB for an application. This is critical for the reactor models because Aspen HYSYS uses Aspen Properties and Equation Oriented solver (EO) which both require large amounts of memory.Â
To enable the 3GB switch:
1. Right-click on the Command Prompt icon in the Accessories program group of the Start menu. Click Run as Administrator.
2. At the command prompt, enter: bcdedit /set IncreaseUserVa 307.
3. Restart the computer.
Keywords: 32 bit operating system, Windows Vista, Windows 7, or Windows 8
References: None |
Problem Statement: In the pipe segment under Design / Parameters, there are several correlations available to model vertical pipe flow. There is, however, not enough information about the correlations recommended for downward flow. | Solution: For downward flow, the recommended correlations are:
· Beggs and Brill: A study by Payne and Palmer improved the hold-up calculation for inclined flows. This is included in the Aspen HYSYS model.
· OLGAS: Has been validated with a lot of experimental/production data
· Tulsa Unified Model: Has fairly limited validation. It is recommended as an improvement to Beggs and Brill. Should be good when slugging anticipated
· ProFES: Uses the HTFS model.
Keywords: downward flow, pipe correlations
References: None |
Problem Statement: The file size keeps increasing when running the dynamic simulation. How do I reset the the file to its original size? | Solution: The dynamic simulation typically contains several strip charts to display the process conditions with time. As the Dynamic simulation runs, the data histories fill up with with time. The number of data each variable can store depends on the logger size. The default size is 300. The default option can be changed from File | Options or specified for each strip chart.
The Dynamic simulation will store the data for all the process variables and the file size will increase. To revert the file to its original size the user can set the current time in the Integrator zero. This will remove data from the strip chart historian. The file size will be smaller when you save the file with zero current time.
Keywords: Dynamic Simulation, Strip Chart, Data Logger
References: None |
Problem Statement: In Aspen Shell & Tube Exchanger version 8.4, there is a tab for opening *.EDT templates as shown below. This option does not appear in V8.8.2. Where can I find this option for version 8.8.2? | Solution: In Aspen Shell & Tube Exchanger V8.8.2, there are no new template selections because the file extension.*.EDT is added to the file type in the Open dialog as shown below.
You will need to locate the template files on your machine which have been installed in the following default directory:
C:\Program Files (x86)\AspenTech\Aspen Exchanger Design and Rating V8.8.2\Dat\Template
Keywords: EDT, template, location
References: None |
Problem Statement: How can I see the physical properties for each phase? | Solution: To view the properties for each phase, you must expand the view using the plus sign button on the Physical Properties page.
This will show the following:
Where:
F = bulk fluid (Overall)
V = vapour phase
L = hydrocarbon liquid phase
W = aqueous liquid phase
M = mixed liquid phase
Keywords: Phase, Properties
References: None |
Problem Statement: Which of the reactor temperature bias variables corresponds to in the catalytic reformer? | Solution: In catalytic reformer units, the feed temperature to the train of reactors is normally known or can be easily manipulated by means of a pre-heater (fired heater).
The inlet temperature for each reactor can be entered in the Reactor Control Page under the Operations tab.
If the inlet temperature for each reactor is not known, it can be calculated based on other parameters.
To calculate based upon the delta inlet temperature, a base temperature is used as a reference temperature for biasing the individual reactor inlet temperatures.
where
Reactor(i) Inlet Temperature = Reactor Inlet
Keywords: Temperature, reactor, catalytic reformer, bias, reference
References: Temperature + Reactor(i) Temperature Bias
The temperature bias is a tuning parameter that allows the user more flexibility to tune the product stream's temperatures. It will only affect the inlet temperatures of each reactor. However, the inlet temperature will have an effect on the performance of the reformer.
In the example below, the inlet and reference temperatures are calculated based on the temperature bias and WAIT (weighted average inlet temperature).
WAIT = Tr1 x CWr1 + Tr2 x CWr2 + Tr3 x CWr3 / CWr1 + CWr2 + CWr3
Where
Tr(i) = inlet temperature of each reactor
CWr(i) = catalyst distribution of each reactor
The catalyst distribution of the reactors is calculated on the catalyst loading page in the design section of the reformer.
Tr1 = Tr ref + Tr1 bias
Tr2 = Tr ref + Tr2 bias
Tr3 = Tr ref + Tr3 bias
Tr(i) = temperature bias of each reactor
Then,
505 = ((Tr ref + 0) x 0.1444 + (Tr ref + 1) x 0.2556 + (Tr ref + 2) x 0.6) /1
Tr ref = 503.5 C
Tr1 = 503.5 + 0 Ã Tr1 = 503.5 C
Tr2 = 503.5 + 1Ã Tr2 = 504.5 C
Tr3 = 503.5 + 2 Ã Tr3 = 505.5 C
The inlet temperature is then used to calculate the WABT (weighted average bed temperature)
WABT = AVTr1 x CWr1 + AVTr2 x CWr2 + AVTr3 x CWr3 / CWr1 + CWr2 + CWr3
Where
AVTri = the average bed temperature
AVTri =Â Inlet temperature reactor i + Outlet temperature reactor i / 2
The inlet and outlet temperature values for each reactor are reported in the Results tab / Reactors
Then,
AVTr1 = (503.5 + 370) / 2 Ã AVTr1 = 436.79 C
AVTr2 = (504.5 + 403.8) / 2 Ã AVTr2 = 454.17 C
AVTr3 = (505.5 + 426.9) / 2 Ã AVTr3 = 466.23 C
WABT = (436.79 x 0.1444 + 454.17 x 0.2556 + 466.23 x 0.6) / 1
WABT = 458.9 C |
Problem Statement: When running a case on rating mode with the specifications of heat duty, humid air, humidity ratio, inlet and outlet temperature, the results of the outlet temperature in the X-Side do not match the input values and an input warning 560 alerting that the specified outlet temperature has been adjusted based upon the specified tubeside process parameters. | Solution: When both the tubeside and airside have sufficient user-specified process parameters to calculate estimated heat duties on both sides. The input warning 560 is issued due to an incompatible estimated/initial heat duties calculated from specified process parameters on both sides. Aspen Air Cooled exchanger would adjust the airside outlet temperature based upon the calculated tubeside heat duty.
Later, once all required process parameters are determined from initial heat balance calculation, the program would perform rating mode calculation. The rating calculation is carried out by repeated Simulation calculations to iteratively find the area ratio that matches the heat duty obtained from initial heat balance calculation or user-specified value. Each Simulation run with the projected area ratio calculates the air side outlet temperature and the heat duty based on the specified air cooler geometry. Therefore, it is possible that the calculated air side outlet temperature reported in the results will be different from the input/user-specified value and/or the adjusted value reported in the input warning message.
Keywords: Input warning 560, humid air, humidity ratio
References: None |
Problem Statement: Packed Column is sized in steady state. See example screen shot below
Whilst transitioning into AspenHYSYS Dynamics, User receives this error message in the Trace Window:
Please supply appropriate packing data for tower.
The information from the Section Sizing has been transferred to the column, however in the Dynamics mode, the Packing Type data which is used in the calculations of the Section Sizing has not been transferred.
User will notice that Robbins Factor in the Packing Properties (Dynamics) has not been populated. | Solution: There are two options to resolve this issue:
1. User can type in the Robbins factor if known
Or
2. User can select the Packing Type and the corresponding default Robbins factor will be populated.
Keywords: Robbins factor, packing data, Packing Type, Packed Column
References: None |
Problem Statement: How to model a Fully Welded Plate Heat Exchanger using Aspen Exchanger Design & Rating? | Solution: Fully Welded Plate Heat Exchanger consists of four columns frame, top and bottom plates as well as four side plates.
This heat exchanger is used in oil and gas, chemical and petrochemical industries.
A Fully Welded Plate Heat Exchanger cannot be modeled in EDR yet. As an alternative, only the heat transfer calculations could potentially be modeled as multiple plate fin exchangers (LNG unit operation) using Aspen HYSYS.
Keywords: Welded plate, plate exchanger
References: None |
Problem Statement: How does #MIX treat properties while mixing of streams, when some of the member streams have it and some of them don’t? | Solution: While using #MIX (equivalent of a mixer unit), the “mixed” stream is going to take a volumetric average of the properties that are fed into it (unless otherwise specified). If two streams go into a mixture and one of the streams has a property and the other doesn’t, the mixed stream will take the existing property value and apply it to the mix.
Consider two streams (Have deliberately chosen two streams which have no relation to each other – which means they have different properties)
Stream 1 Stream 2
Common Properties:
VOL, WGT, SPG, SUL, SPPM, SPV, API, OXY
Stream 1 additional properties:
F3=
Stream 2 additional properties:
FVT, T400, ARO, N2A, CBI, LUM, KFC
Now a #MIX structure is created in the POST sheet combining the two above into a dummy stream DMSTR
Here are the simulated results for the mixed stream:
As seen above, it takes the average of the common properties like SPG, but for properties only existing in one stream like F3=, that value propagates to the mixed stream.
Keywords: None
References: None |
Problem Statement: How do I prevent the error, Stream Segment Temperatures below solver tolerance, Check Stream - xxx when I try to run a Recommended Design? | Solution: In Aspen Energy Analyzer V8.8 and older versions, the solver tolerance for segment stream is 0.1 degree C for Recommended Design. So user will receive this error if any stream segment temperature difference is below 0.1 degree C.
To solve this issue in current version, you can delete any stream segment which has temperature difference lower than 0.1 degree C. Please note the overall enthalpy or heat load for the stream has to remain the same as in the base case. ThisSolution will be effective if the total temperature difference for the stream is greater than 1
In Aspen HYSYS V9 (or Aspen Plus V9) and newer versions, user is able to modify the segmentation tolerance in the Activated Energy Analysis environment | Setup | Data Extraction:
Keywords: Recommended Design, Segment Stream, Solver Tolerance
References: None |
Problem Statement: How to display Control and Backup lines with the Surge line on Compressor curves profile? | Solution: Aspen HYSYS allows the user to display Control and Backup lines with the Surge line for Compressor.
To display both the lines, users have to click on show SurgeCurve box and Show Control/ Backup box under curves option on Compressor curves profile. This facility is available from HYSYS V9 onwards.
Please find below screenshot for reference.
Keywords: Control and Backup lines, Surge line, Compressor curves
References: None |
Problem Statement: How do I add new users to the APS Database? | Solution: To add a new user you’d have to populate one or two tables:
Table GROUPS (this defines classes of users and their privileges in the model)
Table USERS (this is the username that you would login with, and each user should belong to a group)
Recommended is updating these two tables from the SQL interface and not through the APS interface.
Starting with table GROUPS:
This sample has two different classes of users: ADMIN and Read_Only
ADDSCRN field: to allow access for a user belonging to this group to add new event screens/control variables on event screens etc (Scheduling side)
EDITMODEL field: to allow access for a user belonging to this group to add model elements (units/tanks/pipelines etc) and flowsheet elements (Modeling side)
The privilege is turned on by adding a switch “1”
So above, the ADMIN group has all possible privileges, and the Read_Only has none (he is basically just a viewer)
This is the first table that is to be populated.
If you already have the GROUPS table defined, then the next step is adding an entry to table USERS:
In the above, the ID field is what you would enter on the APS login screen, and if you have any passwords set in the PASSWORD field.
For every user you create in this table, you have to mention the group that they belong to in the GROUP_ field (In this example ADMIN or Read_Only)
The ID field can be any string, or it can also be a domain name (as you see above with the Windows2000\xxx)
Keywords: None
References: None |
Problem Statement: How do I add new users to the APS Database? | Solution: To add a new user you’d have to populate one or two tables:
Table GROUPS (this defines classes of users and their privileges in the model)
Table USERS (this is the username that you would login with, and each user should belong to a group)
Recommended is updating these two tables from the SQL interface and not through the APS interface.
Starting with table GROUPS:
This sample has two different classes of users: ADMIN and Read_Only
ADDSCRN field: to allow access for a user belonging to this group to add new event screens/control variables on event screens etc (Scheduling side)
EDITMODEL field: to allow access for a user belonging to this group to add model elements (units/tanks/pipelines etc) and flowsheet elements (Modeling side)
The privilege is turned on by adding a switch “1”
So above, the ADMIN group has all possible privileges, and the Read_Only has none (he is basically just a viewer)
This is the first table that is to be populated.
If you already have the GROUPS table defined, then the next step is adding an entry to table USERS:
In the above, the ID field is what you would enter on the APS login screen, and if you have any passwords set in the PASSWORD field.
For every user you create in this table, you have to mention the group that they belong to in the GROUP_ field (In this example ADMIN or Read_Only)
The ID field can be any string, or it can also be a domain name (as you see above with the Windows2000\xxx)
Keywords: None
References: None |
Problem Statement: I would like to model a process unit which its capacity would decrease by 0.2*X+4 when the production amount of X is more than 500 tons/day. How do I model this? | Solution: Let us use attached simple LP model as reference. Note that this is a simple LP model that is used for demonstration purpose and does not represent a full refinery configuration.
I have a process unit SABC, its feeds are ARL and ARH, its products are LN1, LPG, OFG.
The constraint is such that when LN1 production is more than 500 ton/day, the feed intake should reduce by 0.2*LN1 production+4. This would require MIP modeling.
Create a dummy submodel SXYZ, and use a G-row to store the amount of LN1 product in excess of 500 tons/day into a new column “Add”.
Create another 2 columns in T. SXYZ, set column ONE as Bivalent variable in T. BOUNDS so that it can only take value of 0 or 1, and set column FHD as a semi-continuous variable in T. MIP so that it should take a value of 0 or threshold value 0.5
Create G-row and L-row to control column ONE and FHD to be zero or active, and create a E-row to make sure column ONE and FHD are either both active or both zero.
Finally, add the capacity decrement of 0.2*LN1 production+4 into the capacity row.
Keywords: MIP, capacity, linear decrease, bivalent variable, threshold value, semi-continuous variable, process unit
References: None |
Problem Statement: User Property Model example: liquid vapor pressure subroutine with regression of the user model parameters. | Solution: The attached example illustrates how to implement a user property model in Aspen Plus, in this case for the calculation of the liquid vapor pressure (PL). Each user property model should include a principal subroutine that calculates and returns the desired physical properties. Since the principal subroutines are called directly by the appropriate physical property monitors, they have a fixed name and argument list structure. They also contain model-specific parameters that are passed to the user subroutine during calculations. Property user models are documented in the User Models Guide Chapter 6.
The user subroutine in this example returns the liquid vapor pressure (PL) and takes the values of the model-specific parameter called PL0UA to perform the calculations. In the example the coefficients of the T-dependent parameter (PL0UA) are regressed from experimental data coming from the NIST. For simplicity the model implements a version of the extended Antoine equation reduced to 5 coefficients.
In order to make the user model active in the corresponding property routes, PL0USR should be selected:
Please note also highlighted in the screen shot above the model-specific parameter for liquid vapor pressure PL0UA. This parameter should be added as follows:
Keywords: User Physical Property Models, Fortran Subroutines, Regression.
References: None |
Problem Statement: The TS Cond. Coef. (Tube Side Condensation Coefficient) and TS Film Coef. (Tube Side Coefficient) are shown on the Interval Analysis page under Results / Calculation Details / Analysis along Tubes. What is the meaning of these values? | Solution: The TS Film Coefficient is the total condensation coefficient considering the multi-component effect. It is calculated using the equation below:
TS Film Coeff. = Heat flux / (TS Fouling temperature – TS Bulk temperature)
The TS Cond. Coefficient is the pure component TS condensation coefficient.
The correlations used to calculate the condensation heat transfer model are selected on the Condensation page under Inputs / Program Options / Methods / Correlations.
The Silver method is among the most reliable and well-tested method to capture the multi-component effect when no information about the component is provided.
The heat transfer coefficients (TS Cond. Coefficients) are lower with the HTFS-Mass transfer method because it is taking into account the additional resistance of mass transfer (diffussional effects) in the coefficient.
Keywords: Condensation, film, coefficient
References: None |
Problem Statement: Is there a way to apply an elevation to fittings with the pipeline segment tool? | Solution: All fittings in Aspen HYSYS are point fittings in, that they are all considered to have zero length and elevation. As a consequence, there is no hydrostatic head associated with any library or user defined fitting.
In order to include the effect of a non-zero elbow elevation, you will need to append a corresponding section of pipe after each fitting.
Keywords: elevation, fittings, pipe
References: None |
Problem Statement: When opening Aspen Plus, the following error is reported:
Setting up Projects directory failed for restricted user
See screenshot below for reference: | Solution: The reason why this error message appears when opening Aspen Plus is that the Economic Evaluation (EE) resiliency is being executed to enable the Economics ribbon in Aspen Plus. This is by design. Trying to run any EE standalone product (such as: APEA, ACCE, or AICE) and let resiliency complete helps to solve this issue.
Keywords: Project Directory Error, Error message, Economic Evaluation.
References: None |
Problem Statement: In the input form of the Nozzles inlet and outlet pipes of Air Cooled Heat Exchangers:
For pipe sizes of 3 ½ inches and bigger, the OD/ID and wall thickness are based on Nominal Pipe Schedule 80 by default.
For pipe diameters smaller than 3 ½ inches, OD/ID and wall thickness of schedule 160 are automatically selected.
Why is this the case? | Solution: The change in the wall thickness from schedule 80 to schedule 160 for the nominal pipe sizes less than 3 ½ inches is based on the requirement on the minimum nozzle neck nominal thickness from API 661 Petroleum, Petrochemical, and Natural Gas Industries Air-Cooled Heat Exchangers for General Refinery Service.
All the default values of OD/ID/wall thickness are automatically set by the program based this requirement.
To change the pipe schedule, User should input the actual OD, ID, or wall thickness without selecting Nominal Pipe Size.
Keywords: Nozzles, Inlet, Outlet, pipes, Air Cooled Heat Exchanger, Nominal Pipe Schedule,
References: None |
Problem Statement: Why do I see the same data point appears twice in the stream data table in EDR? | Solution: EDR doesn't generate duplicate temperature points. What happens is that EDR will add the bubble and dew points to the existing temperature array that is based on the Temperature range and Number of points specified in the Properties page. If the bubble and dew points are extremely close to one of the existing temperature points, user may see same data point appears twice in the Stream Properties table.
In this situation, user can either delete one data point manually or change the Temperature range or Number of points to solve the issue.
Keywords: EDR, Property data point
References: None |
Problem Statement: What are the boundary conditions in Hydraulics subflowsheet in Aspen HYSYS Upstream? | Solution: In Aspen Hydraulics, the pressure and flowrate values of the inlet and outlet streams are considered as boundary conditions. These boundary conditions need to be specified correctly in order to make the subflowsheet solve. The total number of boundary specs should be equal to the total number of boundary streams.
For a single pipe in the flowsheet, three boundary conditions are allowed:
For a pipeline network, there are some restrictions user needs to consider:
- Flowrate specification is only allowed on inlet streams.
- For network with multiple inlets, if one of the streams has both mass flow and pressure specified, then all other inlets must have pressure specifications.
- For network with multiple outlets, all the outlets need pressure specifications.
For Mixer unit, four types of boundary conditions are supported:
For splitter unit, two types of boundary conditions are supported:
There are some built-in examples inside Aspen HYSYS that users can check/study these boundary condition. The location for these files are: Open Aspen HYSYS| Resources | Examples | AspenHydraulics.
If you need further assistance on a specific case, please contact AspenTech Support.
Keywords: Aspen Hydraulics, Boundary Conditions
References: None |
Problem Statement: What are the Nozzle entrance and exit connected to? | Solution: The Nozzle entrance and exit are for the other ends of the connected piping, not for the PRD flange connections.
An entrance would be used for the nozzle where the inlet piping connects to the vessel.
Historically as a guidance by API, an exit would be used at the end of outlet piping but API will revise this in future and recommend against counting the exit loss for a discharge to atmosphere, because it is already included in the pressure drop correlations.
Keywords: PSV, Rupture Disk, PRD, Equivalent Length, Nozzles, entrance, exit, inlet piping,
References: None |
Problem Statement: How to configure alarms in Aspen InfoPlus 21 database? | Solution: This article describes about the detailed steps to configure alarm for record belonging to IP_Analog definition family, in Aspen InfoPlus21 database
The first step to take when configuring a record for Alarm processing is to specify an Alarm Format.
The Alarm Format determines how to display the Alarm State message in the IP_ALARM_STATE field.
The default Alarm Format record supplied with the database is called IP_AlarmMessages and can be selected from the drop-down list box.
Fields which store parameters concerned with Alarm Processing:
IP_HIGH_HIGH_LIMIT: If the value of contents of the IP_INPUT_VALUE field exceeds the number specified here, the IP_ALARM_STATE field changes to indicate a severe high alarm.
IP_HIGH_LIMIT: If the value of contents of the IP_INPUT_VALUE field exceeds the number specified here, the IP_ALARM_STATE field changes to indicate a high alarm.
IP_LOW_LIMIT: If the value of contents of the IP_INPUT_VALUE field exceeds the number specified here, the IP_ALARM_STATE field changes to indicate a low alarm.
IP_LOW_LOW_LIMIT: If the value of contents of the IP_INPUT_VALUE field exceeds the number specified here, the IP_ALARM_STATE field changes to indicate a severe low alarm.
IP_LIMIT_DEADBAND: Specifies the minimum amount that data for the tag must vary before leaving an alarm state. This field is used to reduce nuisance alarms.
IP_ALARM_STATE: Determines the state message that is associated with the value in IP_INPUT_VALUE. This field does not require a value at the time that the tag is defined. It updates automatically as the tag is used to gather data. The selector record specified by the field IP_ALARM_FORMAT defines the choices for IP_ALARM_STATE. Each time this field changes, an entry is made in an Aspen InfoPlus.21 Alarm Summary.
IP_ACKNOWLEDGEMENT: When the IP_ALARM_STATE field shows that an alarm has occurred, the contents of the IP_ACKNOWLEDGEMENT field automatically changes from ACK to UNACK. Records whether or not an alarm state is acknowledged or unacknowledged.
IP_MESSAGE_SWITCH: Turns on/off writing the alarm state and alarm state acknowledgment messages to the log file. Setting IP_MESSAGE_SWITCH to ON causes all IP_ALARM_STATE and IP_ACKNOWLEDGEMENT changes to be logged to special log records. By default both of these records point to the log record IP_LOG_RECORD.
Once all the above fields are specified, the respective alarm state will appear, whenever the IP_INPUT_VALUE field crosses the limits.
Keywords: Alarm processing
IP_ALARM
InfoPlus21 database
References: None |
Problem Statement: What is the operating pressure range for the Hydrate Formation Utility? | Solution: The Hydrate Formation utility will be performing under critical region, and the model has been tested against a wide range of experimental data, with upper limits between 100 and 1000 bar (except for Methane hydrates, which have a max pressure of 10,000 bar). The model should be applicable to this pressure range .
Keywords: Hydrate Formation utility, operating pressure range
References: None |
Problem Statement: Aspen Process Explorer allows viewing statistic of Batch data using an 'Ad Hoc SPC' Chart for Aspen Production Record Manager.An Ad Hoc SPC chart shows the values of the selected statistic or characteristic for a set of selected batches that were added to the description section of the chart.
The AdHoc SPC plot may display fewer batches than what has been added to the plot. For example, suppose 20 Batches were added to the Batch Legend section of the chart, only 18 of them might be displayed across the plot with a subgroup marker | Solution: There are two possible reasons for the missing subgroup markers:
The slider bar just directly under the Plot area is not expanded all the way in both directions. The global time span must accommodate the Batch Start Time of the first batch and the Batch End Time of the last batch
The value for the Characteristic is Null or missing. From a client tool such as the Aspen Batch Query Tool, investigate the 'Value' of the Characteristic for those missing Batches. If the 'Value' is Null or missing, the whole Batch is omitted from the Plot Area even though it appears in the drop-down of the description section.
Keywords:
References: None |
Problem Statement: Why do I receive a dTmin violation when using Recommended Design? | Solution: The Recommended Design tool gives you software-generated designs. It does not follow the Pinch Design Method rules, but it tries to come up with networks that have minimum total cost index. It does not maintain specified dT minimum. It may violate deltaT minimum if it is justified by cost. Therefore, you should use engineering judgement when reviewing these design results.
Keywords: Recommended Design, dTmin violation
References: None |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.