question
stringlengths 19
6.88k
| answer
stringlengths 38
33.3k
|
---|---|
Problem Statement: On what basis are the Reaction Profiles values in Radfrac? | Solution: The reaction profiles are the amount of component moles generated or consumed per stage.
Keywords: Radfrac, reaction, profiles
References: None |
Problem Statement: When trying to generate a ternary diagram I get the message: No plot or table produced due to calculation error. Please see Control Panel for details. However, the control panel does not show any error or any useful information. | Solution: This message occurs when two liquid phases are not found. A ternary diagram needs two liquid phases.
Troubleshooting Tips:
1. Does your system contain two liquid phases? If not, ternary diagram is not applicable. Try generating a binary phase diagram instead.
If you think there should be two liquid phases, please try the following tips.
2. You need to use a property method that can predict two liquid phases. Use an activity coefficient model such as NRTL.
3. Check if binary interaction parameters are available for the components. If they are not there, include additional or even all of the available binary databanks. If this does not yield the required parameters, you can estimate them from UNIFAC by checking the box at the bottom of the binary parameter form or provide the parameters by yourself. Data regression can be used to fit the parameters.
Keywords: Ternary diagram, no plot or table produced due to calculation error. Please see Control Panel for details
References: None |
Problem Statement: How do you install the PPDS data bank into the Aspen Properties Database Manager (APED) and into the Aspen Plus User Interface? | Solution: Installing PPDS in APED
1. Start | Programs | AspenTech | Process Modeling V7.x | Aspen Properties | Aspen Properties Database Manager.
2. Right click on the Aspen Physical Properties Databases and select Create a New Database. A Create A new Properties Database Wizard window pops up.
3. Click Next and define the window as following
Database Type: SQL Server
Server: xxx\SQLEXPRESS
Refer to your SQLEXPRESS id. Start | Programs | Microsoft SQL Server 2005 | SQL Server Management Studio Express. Check the Server name.
Login Name: apeduser (V7.0 and onwards)
Password: Aprop100
Database Name: PPDS
4. Click Next and click Import Files. Browse your *.INP format file.
5. Click Finish. It will take some minutes to implement. After completing data conversion, a Report button activates. Meanwhile, a PPDS lists on the Database catalogues.
Apply PPDS from the GUI interface
1. Open a bank simulation file.
2. Go to data browser. Components | Specifications | Enterprise Database | in the Available databanks of left hand column, scroll down arrow, the PPDS is located most bottom. Select it and move to Selected databanks right hand column and move it up to the top.
Keywords: PPDS, in-house databank, APED
References: None |
Problem Statement: How can Aspen RateSep be tuned to match plant data? | Solution: There are three frequently used tuning parameters that can be used to tune Aspen RateSep models to match plant data. They are: interfacial area factor, heat transfer factor, and holdup scale factors (for liquid and vapor).
Aspen RateSep models can be tuned with built-in tuning parameters located on the Tray or Pack Rating Aspen RateSep input form:
Holdup scale factors are applicable only when kinetic reactions are involved. These parameters can be used to tune the amount and/or extent of kinetic reactions in the column.
Interfacial area factor is typically tuned to match the key composition of the key component. Take HCl absorption as an example. The key composition is the mass fraction of HCl in the vapor product from the top. Note that increasing the interfacial area factor increases the mass transfer and heat transfer directly. Therefore, if you want to increase mass transfer without increasing heat transfer, you should tune heat transfer in the opposite direction. Stated another way, the heat transfer factor and interfacial area factor are inversely proportional, i.e., heat transfer factor = 1/interfacial area factor. More specifically, if the interfacial area factor is set to 5, then the Heat Transfer Factor should be set to 0.2 so that there is no net tuning for heat transfer.
Heat transfer factor is typically used to tune temperature profiles to match plant data. Of course, as mentioned in the previous paragraph, it is also used for reverse tuning.
Keywords: rate-based distillation column, Aspen RateSep, tuning parameter, holdup scale factor, interfacial area factor, heat transfer factor
References: None |
Problem Statement: File loads incomplete due to missing column packing specifications in 2006.5. This did not happen in earlier releases. | Solution: In version 2006.5, a number of changes were made in the handling of Pack-Rating and Pack-Sizing sections. One change was omitted from the compatibility note for that release: It used to be allowed to omit any specification for which only a single option was available (for instance, for packing type GRID, vendor MTL, there is only STANDARD material and size, and these could be omitted). Since version 2006.5, it has been required to enter all specifications for Pack-Rating and Pack-Sizing even if only one option is available, to avoid ambiguities when bringing files forward to new versions where additional packing vendors, materials, and/or sizes may have been added.
Starting in V7.2, when one of these files is loaded, if there is still only one option available for an incomplete packing specification, this option will be filled in automatically so that the specification will be complete. However, for certain cases, where additional packings have been added since the version the file was created in, the specification cannot be filled in and will still load as incomplete. If you want to ensure you have the same specifications as in your original version, see the reference tables at the end of the Column Design and Rating chapter of the Input Language Guide for the version your file was created in to determine the implied values for these missing specifications.
Keywords: None
References: None |
Problem Statement: When viewing the sensitivity analysis results form in Aspen Plus, the first column says either OK, Warnings, or Errors. When this result form is pasted into Excel, these statuses are replaced by numbers. What do those numbers mean? | Solution: The numeric status codes for the sensitivity analysis, when pasted into Excel are as follows:
0 Successful run
1 Errors
2 Warnings
Keywords: None
References: None |
Problem Statement: Why is the enthalpy calculation for a stream containing coal in my model incorrect? | Solution: For a nonconventional (NC) component, attributes PROXANAL, ULTANAL and SULFANAL need to be specified consistently. If the three elements in SULFANAL (PYRITIC, SULFATE, and ORGANIC) do not sum to the ULTANAL value for sulfur, enthalpy calculation for the NC stream will be incorrect. In addition to sulfur, other consistency requirements are:
ULTANAL value for ash equals the PROXANAL value for ash.
ULTANAL values sum to 100.
PROXANAL values for FC, VM, and ASH sum to 100.
In V7.2 and higher, warning messages have been added to alert users if the attributes are not consistent. For example:
* WARNING IN THE STREAM PARAGRAPH WHICH BEGINS ON LINE 73
STREAM NAME: SIN-1 (COATTR.5)
THE SUM OF ATTRIBUTED COMPONENT VALUES IS NOT 100.0 FOR
COMPONENT: ASH1 TYPE: ULTANAL
IN SUBSTREAM NCPSD OF STREAM SIN-1 .
* WARNING IN THE STREAM PARAGRAPH WHICH BEGINS ON LINE 73
STREAM NAME: SIN-1 (COATTR.11)
THE SUM OF ATTRIBUTE SULFANAL IS NOT EQUAL TO SULFUR IN ATTRIBUTE ULTANAL
FOR COMPONENT: ASH1 IN SUBSTREAM NCPSD OF STREAM SIN-1
Keywords: Nonconventional component, attributes, consistency requirements
References: None |
Problem Statement: The BatchSep column model was added to Aspen Plus for version 2006.5 however I cannot find it in the Model Library. Where is it? | Solution: By default, the BatchSep model is not enable during the installation of Aspen Plus. To get the model to appear in the Columns tab in the Model Library, go to the Library menu in Aspen Plus and select the
Keywords: BatchSep
Model Library
Library reference
References: s option. In the Library References window, make sure the ACM Built-In library option is checked. |
Problem Statement: When you click on a Stream Table that has been place on the Process Flowsheet you cannot resize the table like other objects by dragging the corners of the table. How do you make the table larger or smaller? | Solution: The size of the Stream Table can be modified by changing the font size used for the text in the table. To change the font size, make sure the Draw toolbar is visible on the Process Flowsheet window. If it is not visible, you can select it from the View | Toolbar menu. With the Draw toolbar visible, click on the Stream Table to make it selected. Then, from the Draw toolbar, select a larger font size to increase the table size or a smaller font size to decrease the table size.
Keywords: Stream Table
Process Flowsheet
References: None |
Problem Statement: Why are the tray rating results for Koch Glitch and Flexi are so very different with Koch's KGtower, the design software from KOCH. The Aspen Plus results indicate that the tray is more loaded than what the Koch software does. For example, the flooding factor in Aspen Plus is 0.9, but in KGtower it is 0.7. | Solution: The results from KGTower and from Aspen are actually very similar to one another. First, one must make sure that one chooses Bulletin 960-1 for the Flooding calculation method in Aspen Plus on the RadFrac | Tray Rating | Setup | Design / Pdrop sheet. More importantly, however, it is standard practice to size and rate trays using the liquid from tray flowrate with the vapor to tray flowrate not the liquid to tray flowrate with the vapor from tray flowrate. The example values in the problem statement used this last convention rather than the accepted one. These flows are found on the RadFrac | Profiles | Hydraulics sheet when Include hydraulic parameters is checked on the RadFrac | Report | Property Options sheet. On this sheet, the Report type for the format of the hydaulic parameter report can be selected to be Extended report with feed properties to also get the stage feed flows.
Keywords: radfrac, tpsar
References: : CQ00217006 |
Problem Statement: How to Perform Assay Data Analysis without process flow diagram?
For a given True boiling point curve, How to generate other curve results like ASTM D86, ASTM D 1160? or Vice versa? | Solution: Yes it is possible to do Assay Data Analysis without having process flow diagram.
It is possible to generate other curves as a results like ASTM D86, ASTM D1160 for a given TBP curve.
Example:
1). Open attached .bkp file
2). Go to Data Browser/Set Up/Specification/ under Global Settings please select Run Type/ Assay Data Anaysis
3). Data Browser/Components/Specification
Define light ends, Oil-1 & Oil-2 as Assay from drop down menu and Mix Oil as Blend
4). Data Browser/Components/Assay/Blend/Edit Oil-1 and specify TBP
Keywords: Assay Data Analysis, ASTM D86, TBP, ASTM D1160 etc;
References: Curve data, API Gravity, Light
ends fraction, API Gravity Reference curve then go to optional tab then select Report Distillation curve as select
ASTM D 86 and ASTM D 1160
5). Similarly you can give input specification data for Oil-2
6). Data Browser/Components/Assay/Blend/Edit blend MixOil then define fractions for Oil-1 and Oil-2
then select Report Distillation curve as select ASTM D 86 and ASTM D 1160
7). Run the simulation by selecting Next option
8). Check Individual Assay results as well as Blend Results as per below path
Data Browser/Components/Assay/Blend/Oil-1/Results/Curves
Data Browser/Components/Assay/Blend/Oil-2/Results/Curves
Data Browser/Components/Assay/Blend/MixOil/Results/Curves |
Problem Statement: What is the purpose of the ANALYZE SPARSITY command? | Solution: This command applies only to Equation Oriented run mode. It can be typed in the command line of the control panel.
ANALYZE SPARSITY is to be used for model development to compare the analytic and numerical sparsity patterns. It will list any non-zeroes that are in conflict between the two patterns. It is normally not an issue if the analytical sparsity has a non-zero that does not exist in the numerical sparsity as this may be due to numerical effects - such as a variable with zero value. Of more concern are cases where there is a numerical non-zero that does not show up in the analytical pattern. That could indicate a model error and may cause convergence difficulties. If there are no instances like this, then there should be no problem.
Keywords: EO
References: None |
Problem Statement: What is the source of the Kij parameters for the Equation of State (EOS) models? | Solution: Kij's for the Peng-Robinson, Redlich-Kwong-Soave, Lee-Kesler Plocker models came from
Vapor-Liquid Equilibria for Mixtures of Low Boiling Substances, H. Knapp, R. Doring, L. Oellrich, U. Plocker and J. M. Prausnitz, Dechema Chemistry data Series, Vol. VI.
Kij's for the Soave-Redlich-Kwong model in the SRK-ASPEN databank were developed by AspenTech.
Pure component and Kij's for the BWR-Lee-Starling model came from
M.R. Brule, C.T. Lin, L.L. Lee, and K.E. Starling, AIChE J., Vol. 28, (1982) p. 616.
Watanasiri et al., AIChE J., Vol. 28, (1982) p. 626.
Keywords: reference
PENG-ROB
BWR-LS
LK-PLOCK
RK-SOAVE
References: None |
Problem Statement: What is the Mostafa method for property estimation? | Solution: The Mostafa method is a group contribution method for estimating solid standard Gibbs free energy of formation (DGSFRM), solid standard enthalpy of formation (DHSFRM), and parameters for the solid heat capacity correlation (CPSPO1).
This method is applied to solid inorganic salts which are divided to cations, anions and ligands.
? For DGSFRM, the reported average errors for 687 diverse solid inorganic salts, was 2.06% (Ind. Eng. Chem. Res., 34, 4577, 1995).
? For DHSFRM, the reported average errors for 938 diverse solid inorganic salts, was 2.57% (Ind. Eng. Chem. Res., 34, 4577, 1995).
? For CPSPO1, the reported average errors for 664 diverse solid inorganic salts, was 3.18% (Ind. Eng. Chem. Res., 35, 343, 1996).
To use this method, it is that it is necessary to enter the Mostafa groups directly on the Properties | Molecular Structure | Functional Groups sheet.
A list of these groups can be found in the help under Aspen Plus
Keywords: pces
References: | Physical Property Data Reference Manual | Group Contribution Method Functional Groups | Table 3.7A Mostafa Method Functional Groups. |
Problem Statement: How can the RGibbs reactor unit operation model yield products for which there are no reactions? How does RGibbs work? | Solution: RGibbs does not follow any specified reaction at all. It predicts product composition by minimizing Gibbs free energy.
RGibbs considers inlet composition and conditions and then looks for all the components specified in your component list. Then, it calculates the outlet compositions of all the components that minimize Gibbs free energy subject to atom balance constraints.
You can have RGibbs consider only a specific set of reactions. To do this you can restrict the chemical equilibrium by specifying temperature approach or molar extent for the reactions. You must specify the stoichiometric coefficients for a complete set of linearly independent chemical reactions, even if only one reaction is restricted. SeeSolution 113360 for more details.
Also, RGibbs cannot directly handle phase equilibrium between solids and fluid phases (for example, water-ice equilibrium). To work around this, you can list the same component twice on the Components | Specifications| Selection sheet, with different component IDs. Then, in order to have RGibbs use the properties for the correct phase, specify both component IDs on the RGibbs | Setup | Products sheet and specify one in each phase. SeeSolution 102346 for an example of how to accomplish this.
Keywords: RGibbs, reaction, reactor
References: None |
Problem Statement: What are the references for the equation of state binary interaction parameters defined in Aspen Plus databanks? | Solution: The binary interaction parameters used in Aspen Plus have been defined using the parameters taken from the following references:
Kij's for Peng-Robinson, Redlich-Kwong-Soave, Lee-Kesler Plocker models
Vapor-Liquid Equilibria for Mixtures of Low Boiling Substances, H. Knapp, R. Doring, L. Oellrich, U. Plocker and J. M. Prausnitz, Dechema Chemistry data Series, Vol. VI.
Pure component and Kij?s for Benedict-Webb-Rubin-Lee-Starling model
M.R. Brule, C.T. Lin, L.L. Lee, and K.E. Starling, AIChE J., Vol. 28, (1982) p. 616.
Watanasiri et al., AIChE J., Vol. 28, (1982) p. 626.
SRK-ASPEN parameters
These were regressed by AspenTech using data from the DECHEMA database, but we have not kept a record of the data used.
Keywords: bip eos
rk-soave
peng-rob
bwr-ls
References: None |
Problem Statement: How does the software search for scripts to run in the Equation-Oriented (EO) command line? | Solution: When invoking a script, one can hard code the path of the file to call. If the path is not hard coded during the invoke call, the search order for a script is:
1. Local folder where model resides
2. The folders defined in script_path
Once the script is found, then the software will not look further, but will use the file found.
Keywords: Script_path, search, order, folder
References: None |
Problem Statement: The Aspen Plus documentation reports that the pure component ETHYLENE databank is used with the SRK method. However, looking at the SRKIJ-1 form, the source for the data is listed as SRK-ASPEN. | Solution: The Ethylene data bank contains the kij parameters for the SRK model but the data bank itself resides on the Engine side. There is no equivalent data in the graphical user interface.
The data bank SRK-ASPEN in the graphical user interface contains all the kij's that are included in the Ethylene data bank. However, it also contains kij parameters for additional pairs of components i.e., it is a super-set of the Ethylene data bank.
If you want to only use the data from the Ethylene data bank, go to the SRKIJ-1 form (under Properties/Parameters/Binary Interaction). On the Databanks tab, un-select the SRK-ASPEN data bank. Make sure that Ethylene is selected in the Components/Specification/Databank tab. All the relevant kij's will be retrieved from the Ethylene databank on the engine side.
Keywords: GUI, database, databank
References: None |
Problem Statement: I am sending results from various blocks to the control panel. When I make a small change I want to see all the key results data rather than just the data that is affected by the change I made. How do I do this? | Solution: By default only the blocks that are affected by any change are calculated. You can change this behaviour from the Convergence | Conv Options | Sequencing form. Uncheck the option Use affected block logic.
Now all blocks on the flowsheet will be recalculated. Normally this will not add significantly to the simulation time as the starting point for unaffected blocks will be theSolution from the previous run.
Keywords: None
References: None |
Problem Statement: Is SLM Client 7.x supported on Windows 7? | Solution: Aspen Plus can crash in unexpected ways on Windows 7 when the SLM Client has not been updated to SLM Client 8.x. Please make sure to install the SLM Client 8.x version from the DVD. Also note that SLM 8.x clients can not talk SLM Server 7.x.
Keywords: None
References: None |
Problem Statement: Aspen did not report a warning indicating that the feed pressure is lower than the stage pressure even though the feed pressure was slightly less than the column pressure at that stage.
It seems that the feed stream pressure to RadFrac has to be lower by a certain amount in order for the Feed Pressure is lower than Stage Pressure warning message to appear. In other words, the feed stream pressure can not simply be lower but has to be significantly lower for the warning message to appear.
Is there a quantitative answer as to how much lower the feed stream pressure should get before the warning message appears? | Solution: Only if the ratio of feed pressure to column stage pressure is less than 0.99, is a warning is issued.
Keywords: column tower
References: : CQ00259060 |
Problem Statement: The below help screen capture lists the default values of efficiency used for the PUMP block in Aspen Plus. On the 3rd item on the help screen, it says for pump discharge pressure < 4.137 e7 Pa (and the solid flow rate is 0), the default efficiency is calculated using an efficiency curve based on water.
What is the reference used for this water efficiency curve? | Solution: The default efficiency curve is from Plant Design and Economics for Chemical Engineers, by Peters, M. S. and K.D. Timmerhaus, McGraw Hill, New York, 1968.
It is read from the curve in Fig. 13.36 in page 459. There is newer version of the book published in 1981.
Keywords: None
References: None |
Problem Statement: Why there is a temperature rise when pump efficiency is 100%? | Solution: Pump calculations assume constant density as inlet. In reality there is a change (which will be small) in density. This difference introduces some error in the calculations. When Aspen Plus calculates the outlet conditions from the enthalpy, it finds a higher temperature than inlet. If there were no change in density at all across the pump, at 100% efficiency there would not be a temperature rise.
Keywords: Pump, temperature, rise, 100%, efficiency
References: None |
Problem Statement: Is there an option to enable or disable display of property parameters in the Aspen Plus User Interface? | Solution: Aspen Plus V7.0 includes a new feature to allow the user to control whether or not to retrieve and display the property parameters (such as binary and pair parameters) in the Aspen Plus User Interface. By default, binary and pair parameters are retrieved and displayed automatically on the Properties | Parameters | Binary Interaction and Properties | Parameters | Electrolyte Pair forms. Pure component parameters can be displayed by using the Review button on the Components | Specifications | Global form.
Note: This option applies only when the Aspen Properties Enterprise Databases (APED) are used.
The user may prefer not to retrieve and display these parameters in the User Interface because:
? The property parameters required to model a particular process have been developed and stored in a user database. It is essential that all process models for this process use the same physical property data. Therefore, the process models have been configured to retrieve all the required parameters from the databases. If the content of the database is updated, the simulation results will change accordingly. If parameters are displayed in the User Interface, then there is a high probability that they will be changed by the users, which may lead to incorrect simulation results.
The user has his own set of property parameters that he entered on the forms. He does not want any of the built-in parameters from the system databases to be used.
The following methods can be used:
? Use the Tools/Options/Properties menu to set the Databank retrieval in GUI control setting.
This information is also attached as a PDF file.
The option specified here will be saved in the registry and used when you start a new blank simulation. Any changes you made on this form will not affect the currently opened simulation.
? Use the Components | Specifications | Enterprise Database tab. Check or uncheck the ?Retrieve parameters automatically in GUI? check box.
This setting affects the current simulation and will be saved in the backup file.
To facilitate selection of databanks other than pure component databanks to be used in the simulation, the binary or pair parameter databanks can be selected on the Components | Specifications | Enterprise Database tab.
Keywords: None
References: None |
Problem Statement: Why does the vapor fraction in Aspen Plus not match that in Aspen Hetran or Aspen TASC or Aspen Tasc+? | Solution: Aspen Hetran, Aspen TASC and Aspen Tasc+ use a mass basis for everything including vapor fraction. Aspen Plus typically uses a molar basis. You can report the vapor fraction on a mass basis in Aspen Plus by adding the property MASSVRA in a prop-set and then using it in the stream report (in all versions) or in custom stream summary (version 2006 and above). The value of MASSVFRA and the vapor fraction in Aspen Hetran, Aspen TASC or Aspen Tasc+ should match.
Keywords: vapor fraction, hetran, tasc, tasc+
References: None |
Problem Statement: What is the difference between NAME2RECID and DECODNAM?
Both access routines, NAME2RECID and DECODNAM, serve the same purpose. They both accept a record name as an argument and return a record id. DECODNAM is superior since it returns error information.
NAME2RECID preceded DECODNAM. In some sense DECODNAM has made NAME2RECID obsolete. However, they are both still available in IP.21 v6.x although NAME2RECID is deprecated. | Solution: NAME2RECID
According to the SETCIM Library
Keywords:
References: Manual, access routine NAME2RECID converts a record name to record ID [NAME2RECID(ptbuff, numchars)].
NOTE: For access routines such as the NAME2RECID routine that searches all nodes to find data, the nodes are searched in the order listed in the configuration file.
DECODNAM
According to the InfoPlus.21? 6.0 Database API Manual, access routine DECODNAM decodes a record name to its record ID [DECODNAM(ptbuff, numchars, recid, error)]. Most of the database access functions accept record IDs as arguments. A record ID is a 32-bit integer that identifies the record to be accessed. The least significant 16 bits contains the record ID. A Server Index may be embedded in some bits in the most significant bits of the record ID. The Server Index is an index into the API?s server node list.
The function DECODNAM() accepts a string containing a record name and returns a long integer containing the corresponding record ID.
When a program linked to setcim.dll calls DECODNAM(), any record ID returned will reside in the local database.
When a program linked to infoplus21_api.dll calls DECODNAM(), then DECODNAM() searches all connected databases for the record with the given name. If it finds a database containing the specified record, it will return a long record ID. A portion of the 16 most significant bits in the returned record ID will contain the node ID. The 16 least significant bits of the returned record ID will contain the ID of the record within that database.
The function DECODRAF() works in a similar fashion as DECODNAM() except that it returns a record ID and field tag after accepting a string containing a record name and field name. After DECODNAM() or DECODRAF() returns a record ID, the calling program can then pass the returned record ID as input to other database access functions.
Some database access functions accept multiple record IDs as inputs. Generally, these functions require all record IDs to have the same embedded node ID. If any Server Index does not match, the routine returns a NODE_MISMATCH error. |
Problem Statement: There is an InfoPlus.21 External task called TSK_HBAK designed for use with records defined by HistoryBackupDef. When activated on a Scheduled basis, certain History Filesets will be backed up to a user-defined Disk location by TSK_HBAK.
This KnowledgeBase | Solution: attempts to describe exactly how the task decides which Filesets will be saved each time it is activated. We also attempt to describe how it knows which ones were saved and thus will not be re-saved the next time it runs unless they have changed.Solution
When viewed from the IP.21 Administrator, each History Fileset has a 'Status'. KBSolution 111740 explains all different Status possibilities and their meanings.
Any time a HistoryBackupDef record is scheduled, the TSK_HBAK task looks at the 'Status' of each fileset in the defined Repository to see if it includes Changed or Shifted. If the 'SaveChanged' and/or 'SaveShifted' switches in the HistoryBackupDef record are set to YES it will then backup all Filesets with a Status that includes Changed or Shifted respectively.
The very first time you schedule a Repository to be backed-up it will backup ALL Filesets that contain data. That is because at that stage every Fileset is either the Active one, or has a Status that includes Shifted.
When 'SaveActive' is set to YES, it scans for T1 > T2 to determine the Active one.
Once TSK_HBAK has completed, whether it is the first time or subsequent times, for any Fileset that has been backed-up, the Fileset Status is then reset to just Mounted. This prevents TSK_HBAK from backing it up a second time without any changes.
In ALL versions, the information relevant to backups ( the Fileset Status, and the T1 \ T2 settings for each Fileset ) is stored in Shared memory while IP.21 is running. This information is updated each time anything changes in History. Shared Memory is what TSK_HBAK uses to read Fileset information and update with changes.
On a 'clean' shutdown of IP.21, this information is transferred via TSK_H21T to the file
...\AspenTech\InfoPlus.21\c21\h21\Dat\CONFIG.DAT
On 'startup' of IP.21, this information is read from CONFIG.DAT into Shared Memory via TSK_H21_INIT.
Nothing that the backup uses relevant to Filesets ( Status, Start\End Times, FilePath etc. ) is read from the Snapshot.
Keywords: None
References: None |
Problem Statement: How can you stop and start an Aspen InfoPlus.21 external task without using the InfoPlus.21 Manager? | Solution: Use the Aspen InfoPlus.21 utility TSK_CLIENT. You can find TSK_CLIENT in the default code directory for Aspen InfoPlus.21 (<drive>:\Program Files\AspenTech\InfoPlus.21\db21\code, where <drive> is the disk drive where InfoPlus.21 is installed).
As an example, to stop the external task TSK_IQ1 using TSK_CLIENT, enter the command from the DOS command window:
<drive>:\Program Files\AspenTech\InfoPlus.21\db21\code\TSK_CLIENT /stop_task TSK_IQ1
To start TSK_IQ1 using TSK_CLIENT, enter the command:
<drive>:\Program Files\AspenTech\InfoPlus.21\db21\code\TSK_CLIENT /start_task_by_name TSK_IQ1
For more details on how to use TSK_CLIENT, run it from a command window with no arguments to get a listing of all parameters.
Keywords: TSK_CLIENT
Stop external task
Start external task
References: None |
Problem Statement: How to resolve Data Binding Failed for Calculation 'name' : Record not found. | Solution: When using CalcCreate.xls (an Excel .xls file that allows you to create multiple Calculations and bind Parameters at once, usually found in :\Program Files (x86)\AspenTech\Aspen Calc\Excel), the error message Data Binding Failed for Calculation 'name' : Record not found may sometimes appear.
In this particular case, the error is triggered by the constants (b: 1,2). When they are typed with quote marks ( ), the process is successful.
Keywords: Aspen Calc, CalcCreate.xls
References: None |
Problem Statement: When using aspenOne Update, several patches may be selected and installed. User may encounter the error
Error 1704: An installation of AspenCalc is currently suspended.
after the AspenCalc patch is downloaded and installed.
For example, the following patches are selected online and installed:
- AspenCalc 2006.5 Cumulative Patch 5
- Infoplus.21 2006.5 Cumulative Patch 5
- SQLPlus 2006.5 Cumulative Patch 5
During the installation of Aspen Infoplus.21 2006.5 Cumulative Patch 5, a message is prompted 'Error 1704: An installation of AspenCalc is currently suspended...' | Solution: Launch Windows Registry Editor.
Navigate to HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\Windows\CurrentVersion\Installer\Inprogress.
If the Inprogress key exists, then delete it.
Restart the installation of the stalled product installation (e.g. Aspen InfoPlus.21 Patch 5).
Keywords: aspenOne Update
Error 1704
AspenCalc
References: None |
Problem Statement: What are the different files generated during Aspen plus run? | Solution: Description of Generated Files during each Aspen Plus RUN:
*.apprj : Aspen Plus physical property data (text)
*.his : Aspen Plus history file (text), contains input specification and run diagnostics
*.adv : Aspen + generate a file containing information of advent. Advent is an old product Aspen no longer provides.
*.appdf : Aspen plus property definition file
*.bka,*.bks : .bks, .bka files were created by the engine when input translation edit runs and sent by the GUI.
*.cmd : Command file
*.cmp : Compiler out put message file. It is reporting FORTRAN compilation output for dynamic link. It may report whether compiler is found in the usual file path names for the compiler
*.co1,*.co2 : Comment content file
*.jnl : Journal file. A record of all interactive commands and input changes
*.def : Run definition file. Each Aspen Plus run creates a run definition (.def) file containing file specifications for the run.
*.dfm : This file contains property parameter results of PCES and data regression in the Data File Management System format (DFMS).
*.dll : Dynamic link libraries
*.out : Output text file that combines both the history and report files.
*.rep : Aspen Plus simulation results (text)
*.sum : another Aspen Plus simulation result file (text), different format than *.rep
*.xml : another Aspen Plus simulation result file (text), different format than *.rep
*.prd : prediction file created during export prediction
*.rp1 :contains short form of aspen plus usage ( report file)
*.rp2 : contains a listing of each run made. Detail report file
*.sto : standard Property method table file
*.hxt : The user interface help and prompt files
*.utl : utility model file
*.fil : contains spec tables.Solution file ( e.g.; Tables.fil)
*.for : FORTRAN file
*.ld : Linker diagnostics output for dynamic linking
*.ikc : The IK-CAPE neutral file as a way to transfer/ select physical property parameters and data between different applications.
*.inm : input file generated from engine
*.srt : user interface table sort file
*.apt : application Template file
*.cpm : control panel message file
Keywords: .appdf, .bkp, .apw
References: None |
Problem Statement: What are Area Penalty Parameters on the MHeatX | Input | Zone Analysis sheet? | Solution: The Area penalty parameters become active when there is a crossover in or at either end of a zone. The LMTD is approximated to a small enough value in order to drive the UA to a large value (penalty), thus causing the solver (Design-Spec or Optimization) to move away from this region. This drives the manipulated variable in the correct direction. The Temperature approach is a threshold value for temperature difference at either end of the zone, below which the penalty is applied. The larger the Penalty value, the steeper the function for the LMTD becomes as the temperature difference approaches zero.
Keywords: MHEATX, Area Penalty
References: None |
Problem Statement: When I do a NRTL parameter estimation on Aspen Plus, why doesn't Aspen Plus estimate aij and aji (aij and aji = 0)? | Solution: When you perform a binary parameter estimation for NRTL parameters, using a UNIFAC method, the only estimated parameters are: aij, aji, bij, bji. Cij is a fxed value (depending on the type of your system). Other parameters (dij, eij, eji, fij and fji) are all zero.
The parameters aij and aji are dependent on temperature. If TLOWER = TUPPER, then aij = 0 and aji = 0. The property estimation default temperature is 298.15 K (25 C) for TLOWER and TUPPER.
If you want to change this temperature, go to Properties | Estimation | Input | Binary and include a new temperature value for the field Temperature (note that units for temperature on this field needs to be entered in the input units of measurement for the simulation as specified on the Setup | Specifications | Global sheet). If you include more than one temperature, TLOWER will be different from TUPPER and then aij and aji parameters will be different from zero.
Keyboard
aij
aji
binary parameters
NRTL
TLOWER
TUPPER
Property Estimation
aij = 0
aji = 0
Keywords: None
References: None |
Problem Statement: What is the correct procedure to run Aspen Properties with user.dll in Excel Add in | Solution: When you use a *.aprbkp file that references user model .DLLs via a linker options (DLOPT) file, the DLOPT file must be specified by its full pathname in the *.aprbkp file even if the file is located in your normal working folder. This is because when Aspen Properties is launched by Excel, its working folder is elsewhere as determined by Excel. To set this option, open the file in Aspen Properties. From the Calculate menu, click Settings, and update the Linker options field.
In addition to this, when you reload this file during a later session, you must start Excel first and open the file from the File | Open command, rather than double-clicking the file icon, to ensure that the user models load properly.
Keywords: Excel Add-in, .OPT, .dll
References: None |
Problem Statement: What is the fractional capacity in pack rating results? | Solution: There are three definitions of approach to flood that are used frequently. The first, and most common, is flooding at constant L/V. The second is flooding at constant liquid load (used often for scrubbers for which you want to process as much gas as you can with the liquid you have available). The third is flooding at constant vapor load (used very rarely for strippers where you want to process as much liquid as you can with the vapor available).
The following discussion applies to the Eckert method. The Eckert method is based on the diagram given on Perry's handbook page 18-22 (see also Kister Distillation Design p 480) which gives the flooding limit for a packing. The x-axis is the flow parameter, L/G (rhog/rhol)^0.5. The y-axis is G^2 F psi mu^0.2 / rhog / rhol.
The maximum fractional capacity is calculated in the following way:
1. Calculate the value of x = L/G (rhog/rhol)^0.5 (on the example it is about 0.02)
2. Read the value of y = G^2 F psi mu^0.2/rhog/rhol/gc from the flooding curve (on the example, it is about 0.2)
3. This gives the gas rate at flooding Gflood = (y rhog rhol gc / F / psi / mu^0.2)^0.5
4. Evaluate the superficial velocity of the gas VS = G/rhog (m/s or ft/s)
5. Evaluate the capacity factor CS = VS * (rhov / (rhol - rhov))^0.5 (m/s or ft/s)
6. Evaluate the superficial velocity of the gas at flooding VSflood = Gflood/rhog
7. Evaluate the capacity factor at flooding CSflood = VSflood * (rhov / (rhol - rhov))^0.5 (m/s or ft/s)
This gives the fractional capacity = CS / CSflood which is reported in the pack rating results. In other words, it gives the vertical distance between the operating point and the flooding line, or for the given flow parameter (L/G) by how much you need to change the diameter of column (flooding at constant L/V).
Variables
rhog: gas density (lb/ft3)
rhol: liquid density (lb/ft3)
mu: liquid viscosity (cP)
psi: density ratio (density of water / density of liquid)
Fp: packing factor (1/ft)
L: liquid rate (lb/s ft2)
G: gas rate (lb/s ft2)
gc: conversion factor (32.2)
Keywords: None
References: None |
Problem Statement: Acol+ or Tasc+ 2006 do not run from Aspen Plus 2006. There is a message that there is a floating point error in the heat exchanger program and the block is bypassed. | Solution: The most likely cause of this problem is that both HTFS+ and Aspen Plus have not been patched compatibly. To resolve this issue apply Aspen HTFS+ 2006 Cumulative Patch 6 (CP6) and Aspen Plus 2006 Cumulative Patch 3 (CP3). The two patches do not have to be installed in any specific sequence.
If the problem persist after applying patches to both programs perform following steps to try to resolve the issue (For the following you must have Administrative rights).
1. From the Start Run, go to Aspen Plus 2006 | Aspen Plus Registry Fix utility. Highlight Aspen Plus 20.0 then click OK.
2. In the directory where you have installed HTFS+ 2006, normally C:\Program Files\AspenTech\Aspen HTFS+ 2006\XEQ then
a. Double click on BjacRegClean.exe.
b. Click on the Search Registry button.
c. Click on Clean up.
d. Close.
e. In the same folder can you then click on regcontrols.bat ( to register files).
3. Confirm from Start | ... HTFS+2006 | Version Control Utility that the tick is shown against version 20.0.
4. Open any sample case to check the perfromance.
If you have multiple versions of Aspen Plus and Tasc+ on your machines then you can run the Aspen B-JAC/HTFS+ Version Control Utility that is normally in the folder C:\Program Files\AspenTech\Aspen HTFS+ 2006.5\XEQ\BJACVC.exe or similar folder. Highlight a version and then click on Set, the registries will be set so different versions can link with one another.
Always use compatible versions (for example: Aspen Plus 2006.5 and Tasc+ 2006.5) of Aspen Plus and Tasc+ to avoid any issues.
Keywords: Floating point error, Acol+, Tasc+, Aspen Plus
References: None |
Problem Statement: What are the difference between various calculations type for Heat Exchanger (HeatX) block in Aspen Plus? | Solution: In the HeatX block the following calculation types are possible:
Design mode - This mode is available for shortcut, Hetran, Aerotran, Tasc+, and Acol+ methods. Given a specified load, HeatX will determine the area required and the number of shells in series that are required. In addition, the rigorous methods will also define the optimum geometry for the exchanger.
Rating mode - This mode is available for shortcut, detailed, and all rigorous methods. Given a specified load and exchanger geometry, HeatX will predict the percent over or under surface for the unit. In addition, the rigorous methods will also estimate the level of fouling that may exist in the unit if the unit is over surfaced for those specified operating conditions.
Simulation mode - This mode is available for the shortcut, detailed, and all rigorous methods. Given an area or geometry for an exchanger, the methods will determine the outlet conditions for the hot and cold side streams.
Maximum Fouling mode - This mode is only available for the Tasc+ method. This method determines the amount of fouling resistance based on the given process condition and geometry.
Keywords: HeatX, Design Mode, Rating Mode, Simulation Mode, Maximum Fouling Mode
References: None |
Problem Statement: Is there a way to create a user convergence block using VBA and define a tear stream for the convergence block? | Solution: To create a user convergence block e.g. C1 using Wegstein:
go_Simulation.Tree.Data.Convergence.Convergence.Elements.Add (C1!WEGSTEIN)
To define a tear stream in convergence block C1:
go_Simulation.Tree.Data.Convergence.Convergence.C1.Input.TOL.Elements.Label(0,0) = tear_stream_name
Keywords: Visual Basic Application, VBA, convergence block, tear stream
References: None |
Problem Statement: The values never change for Local Parameter variables calculated by local FORTRAN in a Design-Spec. | Solution: This is by design. Only the Calculator block has the ability to change the information flow direction (import or export) on the parameter variable types. Other objects that use define forms, Optimization, Sensitivity, Design-Spec, etc. assume that parameter types are READ-ONLY (import flow direction) and hence these variables are NOT updated after the local FORTRAN calculations in the object (Design-spec, etc). If you change the value of one of these variables, that change lasts only to the end of execution of the Fortran code on the sheet, and does not change the Aspen Plus variable it is associated with. In a design-spec convergence loop, the VARY variable is the only variable that is being manipulated to meet the SPEC.
If the reason for defining these variables as local-parameter was to watch the variable's value in the control panel, consider using Fortran WRITE statements:
c 3456789
X = Var1 / Var2
Write (*,100) X
100 Format ('The value of X is:',f12.4)
Keywords: None
References: None |
Problem Statement: How do you activate the Furnace option in the PetroFrac block? | Solution: Typically, the feed to an atmospheric crude column goes through a furnace before entering the unit. To activate the Furnace option within the PetroFrac block, first go to the Streams sheet within the basic column Setup form. Select the appropriate stream (i.e., feed) and choose Furnace from the Convention dropdown list. Then navigate to the Furnace sheet to enter operating specifications for the furnace flash conditions.
Keywords: furnace, distillation, petroleum, crude, PetroFrac
References: None |
Problem Statement: Where can I find in the documentation is a listing of all functional groups? | Solution: This information is in the help.
1. Open the Aspen Plus Help topics from the Help menu.
2. In Aspen Plus 2004.1 and earlier, click on the Aspen Physical Property System Help in the right hand pane. In Aspen Plus 2006, this step is not necessary.
3. Go to Aspen Physical Property System
Keywords: functional group
References: (2004.1 and earlier) or Aspen Plus Reference (2006 and future).
4. Go to Physical Property Data Reference Manual.
5. Go to Group Contribution Method Functional Groups and choose which functional group you wish to view or print the topic.
The help files can be downloaded from the Documentation section of the support web site if you want to look at a different version than you have installed. |
Problem Statement: Is there a way to launch a specific version of Aspen Plus from Excel/VBA code? | Solution: There are two ways to open an Aspen Plus model with a specified version of Aspen Plus.Solution document #120589 discusses using the GetObject method, and this document discusses using the CreateObject method.
To have control over opening a specific version of Aspen Plus, the CreateObject method needs an extra qualifier. One of the optional qualifiers in the CreateObject argument list is the version number.
To open version 2004.1:
Set go_Simulation = CreateObject(Apwn.Document.13.2)
go_Simulation.InitFromArchive2 (lv_FilePathName)
To open version 2006:
Set go_Simulation = CreateObject(Apwn.Document.20.0)
go_Simulation.InitFromArchive2 (lv_FilePathName)
where: lv_FilePathName is a string variable containing the path and file name of the Aspen Plus model to be opened
The first line in the command opens the specific version of Aspen Plus, and the InitFromArchive method in the second line opens the specified file in the previously opened Aspen Plus session.
Keywords: VBA, visual basic, excel, version, open, create object, get object, Aspen Plus
References: None |
Problem Statement: Is it possible to enter interactive commands inside hierarchy blocks when running Aspen Plus using Input Language? | Solution: The command SCOPE needs to be used to gain access inside a hierarchy block during an interactive Aspen Plus run. In particular, the user should type SCOPE hierarchy_name in the Aspen Plus prompt in order to force Aspen Plus to activate the interactive functionality within the hierarchy block hierarchy_name. Subsequently, it is possible to use any of the interactive commands withing that specific hierarchy. For instance, the user can enter stop points before or after blocks that belong to this hierarchy in the regular fashion by typing STOP BEFORE (or AFTER) BLOCK block_name. Following this methodology, it is also possible to DISPLAY the results of a block in that hierarchy after a stop point by typing DISPLAY BLOCK block_name. Finally, the command SCOPE exits the hierarchy mode and allows the user to enter interactive commands for flowsheet entities (e.g. blocks, streams) outside the hierarchy.
Note: To run the simulation interactively, you must include the following expression in your Aspen Plus input file:
SIMULATE INTERACTIVE=YES
Otherwise Aspen Plus assumes that your simulation runs non-interactively.
Keywords: Input Language, Interactive Commands, Hierarchy, STOP, DISPLAY, IAP,
References: None |
Problem Statement: How do I get my simulation to only re-execute the stream results? My entire simulation re-executes when I simply change the units of measure in my prop-set that is reported. | Solution: Aspen Plus restarts the calculation and loses all results if a Prop-Set referenced in the flowsheet is changed in any way. This included changing the units of measure a Property. In 2006, the new stream summary will not have this problem. Users will be able to add properties for the stream report without losing results.
If the Prop-Set is already in the flowsheet or you add a new one, it is possible to only re-execute the stream results.
Follow these steps:
1. Add a new Properties Prop-Set if necessary.
2. Add the Prop-Set to the Stream Results on the Setup | Report Options | Stream sheet.
3. Select Control Panel from the View menu.
4. Right mouse click on the last block in the simulation and select Move To.
Then, when you re-open the stream to look at your results again, you will see the new properties. This will not work if you change a prop-set. When anything in the prop-sets are changed, the Aspen Plus will restart and will lose all results.
Keywords: property set
re-initialize, restart, stream results
References: None |
Problem Statement: What is the liquid-to-solid mass ratio in a Counter-Current Decanter (CCD)? The ratio does not seem to have an effect on the overflow or underflow at each stage but does affect the concentration of a solute in the final underflow. | Solution: The CCD assumes a perfect separation of the solids (i.e. all of the solids go into the cake and none go into the effluent.)
The liquid to solid ratio determines how much liquid is in the cake. With one set of ratios specified in the block, the amount liquid in the cake is determined. This amount does not depend on the feed flow of liquid (other than that you need a minimum flow). You can change the amount of wash to the CCD, and you will see that the total flow in the cake stream will be constant.
The liquid-solid ratio is the ratio in the cake. For example, if you specify that all stages have a ratio of 1, then there will be an equal amount of liquid to the solid. For example, if there is 1500 kg/hr of solid, there would be 1500 kg/hr of liquid. If you specify that all stages have a ratio of 4, then there will be 4 times the amount of liquid to solid, in this case, it would be 4 * 1500 or 6000 kg/hr of liquid.
Keywords: None
References: None |
Problem Statement: How does one know what the Equation-Oriented (EO) variables are in the RPlug, RCSTR, and Valve blocks? For example what is Z? | Solution: Scroll to the right and look at the Description column. Z is the mixture mole fraction.
If this column is not visible, right click on the column headers, select more, and check Description. The Description is blank for many unit operations, but it is available for some such as RPlug, RCSTR, and Valve that were created in ACM.
Keywords: None
References: None |
Problem Statement: What unit operation models do not support true approach electrolyte chemistry? | Solution: Certain unit operation models do not support true species electrolyte chemistry.
These models are
DSTWU (Shortcut distillation design using the Winn-Underwood- Gilliland method)
Distl (Shortcut distillation rating using the Edmister method)
SCFrac (Shortcut distillation for complex petroleum fractionation units)
MultiFrac (Rigorous fractionation for complex columns)
PetroFrac (Petroleum refining fractionation)
BatchSep (batch distillation column)
Extract (Rigorous liquid-liquid extraction)
REquil (Equilibrium reactor)
RBatch (Batch reactor)
In addition, the solids models do not support salt formation within the block from true species electrolytes (though the liquid compositions will be resolved), except for Crystallizer when using the option that salt precipitation is calculated from the chemistry.
RPlug (Plus flow reactor) and RCSTR (Continuous stirred tank reactor) can be used with true species electrolyte chemistry provided that the reactions do not contain electrolytic equilibrium reactions (such as H2O <-> H+ + OH-) and no species participates in both reactions and chemistry.
Keywords: None
References: None |
Problem Statement: Aspen RateSep does not support solids, salt and conversion reactions. However Aspen RateSep will check for salt formation after | Solution: .
What happens if the feed to the column contains precipitated salts as well as ionic electrolyte species? For example, if the feed contains the solid salt NaOH(s) as well as Na+ and OH- ions, how will the NaOH(s) be treated in the column?
Solution
All Salts in the feeds first get dissociated into their ions. Then, those ions will distribute on all stages, following the liquid profiles of the RadFrac calculations. No Salt equilibrium will be calculated for RateSep.
At the end of the simulation Aspen RateSep will check the solubility index and print out the information in the Control Panel in form of a warning.
* WARNING
SALTS MAY FORM ON SOME STAGES
NAOH(S) HAS THE HIGHEST SOLUBILITY INDEX (2.09284)
ON STAGE 3 EXCEEDING THE SATURATION LIMIT (1.00000)
If solids are present in a rate-limited column, you should use efficiencies in a equilibrium-based column to approximate the mass-transfer limitation.
Keywords: rate-based distillation
References: None |
Problem Statement: apwn.exe Application Error. The memory could not be read. | Solution: The error is commonly attributed to some registry corruption and can be fixed by running the fix registry utility in Aspen Plus.
Go to Windows-> Start-> Programs-> Aspentech-> Process Modeling-> Aspen Plus-> Aspen Plus Registry Fix Utility.
Select the version to be fixed if there is more than one and click ok.
Now aspen plus file are able to be opened.
Keywords: Apwn.exe, Aspen Plus, The memory could not be read, Registry Fix utility
References: None |
Problem Statement: In the calculator block, there is no way to access the unit of measure for any scalar property parameters. In the define variable result, there is no unit of measure reported. Is it possible to change the units of measure for property parameters. | Solution: This was a design decision. It is too difficult to do unit conversions to elements of a vector for temperature-dependent parameters. Similarly, we decided not to convert the scalar ones either. All accessed property parameters are in SI units.
Keywords: None
References: : CQ00360503 |
Problem Statement: A catastrophic hardware failure can have serious business consequences. Having a plan for how to recover effectively from this type of failure can help minimize the effect on your operation. | Solution: Catastrophic hardware failures fall into three main categories:
A. Network Card / Cabling / Equipment
B. Motherboard / Power Supply
C. Hard drive
Failures in any of these areas can bring the historicization of your data to a halt. ThisSolution presents a plan for dealing with each scenario.
A. NETWORK CARD / CABLING / EQUIPMENT
This is the easiest problem to fix. When your clients lose connection to the InfoPlus.21 server, but the InfoPlus.21 server itself is still up and responsive, then you need to diagnose and respond to a network problem. Use these steps:
1. Find out if other PC's are having connectivity problems with other servers on your network. If so, then the problem may be a network hub, router, switch or other component.
2. Try to ping the InfoPlus.21 server from another node on the network.
3. Try to ping a node from the InfoPlus.21 server.
If you have reached step 3, and know the problem is with the InfoPlus.21 box, then:
4. Check the network cable connecting your InfoPlus.21 server to the network. Try replacing it to make sure the cable has not failed.
5. Bring the InfoPlus.21 server down, and replace the network card. Bring it back up and check connectivity.
For scenario A, verifying the network card, network cable and network infrastructure should lead you to a reSolution of the problem.
B. MOTHERBOARD / POWER SUPPLY
If the box goes dead, then the likely cause is the power supply. If the screen freezes, or shows garbage, and fails to boot up again, you may have a motherboard failure.
Once you determine that the box itself has a problem that will necessitate replacement or repair, the quickest way back to a functioning system is to have an identical box on hand, and simply switch the hard drives. After switching hard drives, configure the new box to have the same name on the network as the original box before starting up InfoPlus.21 for the first time. Since the Manufacturing Suite is dependent on nodenames and not IP addresses, as long as the new box has the same name on the network, and the hardware configuration is identical, the transition to the new box should be straightforward.
If there is no identical hardware, you will not be able to do a simple swap of the disk drive from your old server. You will have to either (1) recover files off the hard drive, or (2) recover files off your network backup, and then follow the steps outlined in Section C below for a disk failure.
C. HARD DRIVE
A hard drive failure is the scenario that requires the most work to recover from. Recovery in this scenario is also dependent on the effectiveness of your network backup policy. Another factor is having accurate notes on the original InfoPlus.21 setup; did the old hard drive have logical partitions? Was InfoPlus.21 installed on C, and the history archives kept on D? Ideally, step one below should result in a PC configured as closely as possible to how your original InfoPlus.21 server was when it was originally built.
The scenario below will assume a hard drive failure for a system on a network that performs a nightly backup of the server.
1. Put a new hard drive into the existing InfoPlus.21 server. (If you don't have a spare hard drive, and need to replace both the PC and the hard drive, these instructions will work, as long as you make sure to give the new PC the same name on the network as the old PC.)
2. Make sure your NT operating system and service pack level are the same as the previous system.
3. Install Aspen's Manufacturing Suite on the new PC. Use your notes to make sure you install the correct layered products and point count authorized from your license server. You now have a system that is ready to replace your old one on the network.
ADSA. If your ADSA server was on the same box, configure the new server with the same name used previously. This will allow the existing clients to see the new server without reconfiguring their ADSA. If your ADSA is separate, as long as you are configuring your new PC to have the same nodename on the network, and the same InfoPlus.21 group, the ADSA server will successfully connect clients to the new box without any reconfiguration.
Aspen Local Security, If the Aspen Local Security server is running on the same system the rebuilding the security data base may also be necessary. See KB article How to copy the Aspen Local Security configuration to a new server
4. Recover the following files from your network back-up:
infoplus21.snp
Group configuration file (*.ipg)
map.dat
config.dat
c21_config (???)
All history files
Any custom files, such as VB apps, code, .sql queries, etc
Here are typical locations for these files:
c:\Program Files\Aspentech\InfoPlus.21\db21\Group200\infoplus21.snp
c:\Program Files\Aspentech\InfoPlus.21\db21\Group200\[mygroupconfig].ipg
c:\Program Files\Aspentech\InfoPlus.21\c21\h21\dat\map.dat
c:\Program Files\Aspentech\InfoPlus.21\c21\h21\dat\config.dat
c:\Program Files\Aspentech\InfoPlus.21\c21\h21\arcs\arc1, arc2 . . .
Find the newly installed versions of the above files, rename them with a .bkp extension, then copy the files you recovered from your network backup system to the same location.
Notes on recovered files:
Config.dat. Only copy over config.dat if you are (1) using the same PC, or (2) able to reconfigure an existing PC to use the same nodename.)
History files. For your history files, place them in an identical NT file structure and, if your old InfoPlus.21 setup used UNC file paths to specify the location, make sure to create an identically named share on the main repository folder.
The history files available on your old system may have amounted to hundreds of megabytes or more. Recovering all these files is not necessary to get your system back up and running. All that matters is recovering the active archive. If you have multiple repositories, you will need to recover the active archive for each repository, and place it in an arc# file folder for the appropriate repository. Remember that recovery of any archive consists of three files that work together:
arc.byte
arc.key
arc.dat
If your old system had a querydef configured to pause archiving and copy the active archive to a backup location on a physically separate hard drive every night it may simplify your search for the right archive(s).
If you do not have access to the file set that was the active archive at the time of the crash, you will need to use the h21arcadmin utility to create a new fileset to be used on startup.
Pages 8-1 - 8-3 of the InfoPlus.21 Administration Manual shows how to use h21arcadmin. Please refer to that section of the manual for a complete listing of arguments and syntax. If our failure happened sometime on May 23rd, here is an example of the archive we might create in response:
h21arcadmin -c -s 23 May 01 00:00:00 -e 31 May 01 12:59:59 -f /archives/may01
There are several factors to keep in mind when creating an archive.
1. To qualify as the active archive, all that matters is that (a) the start time is before current time, and (b) the end time is after current time. Please look at the additional points listed below for how to fine tune these parameters.
2. Do you have a CIM-IO system that has been buffering since the hardware failure? If so, it is imperative that you create the archive with a start time which at least slightly pre-dates the beginning of your buffer file. If you do not do this, the data will be dropped when unbuffering starts, because there will be no archive with a correct timespan to receive the data.
3. Also in relation to the start time, make sure you don't set it so far back that you run the risk of overlapping the end time of another fileset, which would create another problem to deal with.
4. Set the end time at least a couple of hours into the future. Don't set it five minutes into the future, and there is no need to set it five days. When your archive is designated as the active archive on startup, the end-time becomes irrelevant, but if you set it too soon, you may not be done with your other tasks before your first startup; if your end time is then later than current time, the archive you prepared will not be used, and your startup may fail.
h21arcadmin can be used to arbitrarily set the beginning and end time of any given archive. Therefore, if you later found a copy of the active archive that had data up to, let's say 2:45:00 that morning, the start time of the new archive could be moved to 2:45:01 a.m., so the old active archive, with its data, could be mounted. It's important to use h21arcadmin to make sure that start and end times between archives do not overlap.
5. Start your new InfoPlus.21 system.
With all configuration files in place, and an archive file ready to receive data, your InfoPlus.21 system can be started.
6. Verify that your new system is working correctly by looking at some of the following system functions:
Repositories. Check the repositories by looking at the Properties menu for each repository in the InfoPlus.21 Administrator.
CIM-IO. If you had a Store and Forward system that was buffering while you replaced your InfoPlus.21 server, you can use F5 to refresh your Administrator Window to observe the IO_%_RECOVERY parameter in your device record to monitor unbuffering.
Client Tools. Verify that client PC's can still access the InfoPlus.21 server using their existing ADSA configuration.
Keywords:
References: None |
Problem Statement: While processing input specification, Aspen Plus either crashes with TERMINAL ERROR, displaying UNABLE TO LOAD DYNAMIC LIBRARY, or severe error such as COULD NOT RESOLVE USER OR IN-LINE FORTRAN SUBROUTINE(S). | Solution: Placing system loose files in the working folder where model files reside could cause the crash with message UNABLE TO LOAD DYNAMIC LIBRARY. Similarly binary files such as *.obj or *.dll files generated by previous versions of compilerswould cause error messages such as COULD NOT RESOLVE USER OR IN-LINE FORTRAN SUBROUTINE(S).
Below is an example of the error message:
To resolve this problem, remove the system binary files from your working folder.
Similarly, after upgrading compilers, it is user's responsibility to recompile all in house Fortran routines in order to generate a set of compatible binary files. The binary .obj files are generally not compatible between compiler versions.
Keywords: UNABLE TO, UNABLE TO LOAD, UNABLE TO LOAD DYNAMIC, UNABLE TO LOAD DYNAMIC LIBRARY, LOAD, DYNAMIC, LIBRARY, smeoutils, TERMINAL ERROR, RESOLVE, USER, IN-LINE, FORTRAN, SUBROUTINE
References: None |
Problem Statement: This knowledge base article describes how Aspen InfoPlus.21 handles the situation when a 21CFR11 enabled snapshot is loaded in a non-21CFR11 enabled Aspen InfoPlus.21 server. | Solution: The 21CFR11 compliance flags are set in 2 places:
? In the system registry
In the snapshot.
Aspen InfoPlus.21 checks the 21CFR11 compliance flags in both the registry and the snapshot during each startup sequence. Therefore, if either of these flags are enabled the machine is placed in full 21CFR11 compliance mode. The consequences of this operation are:
? If you load a 21CFR11 enabled snapshot on a non-21CFR11 enabled Aspen InfoPlus.21 server, the server will become 21CFR11 enabled. This prevents users from taking a snapshot from a 21CFR11 enabled server to another server in order to make un-audited changes to the snapshot.
If you try to load a non-21CFR11 enabled snapshot on to a 21CFR11 enabled Aspen InfoPlus.21 server the snapshot will become 21CFR11 enabled. This prevents users from attempting to disable the 21CFR11 compliance feature once it has been enabled on a given server.
It is a 21CFR11 requirement that the 21CFR11 compliance flag can never be disabled once it has been enabled on a given Aspen InfoPlus.21 server. The aforementioned operations ensure that the 21CFR11 compliance feature cannot be disabled.
Keywords: audit
trail
track
log
References: None |
Problem Statement: When doing column design using RadFrac with the Pack Sizing option, Flexipac 1X or 2X packing gives me the smallest diameter column required. However, when I convert to RateSep and select Flexipac for Pack Rating, RateSep requires the input of the corrugation dimensions. Several points:
Why does Aspen even have Flexipac as an option if dimensions are required?
Koch does not publish these in their documentation. Can we get this information from AspenTech?
RateFrac does not require the user to supply this information, why does RateSep?
Under Bravo-Rocha-Fair, it wants me to input Corrugation angle, Base width, Side dimension, and Height for the Structure. If you select Billet and Schultes, it wants me to input critical surface tension and B&S CL and CV factors. I don't know input for either. Also, it is asking for Stichlmair Void fraction. | Solution: We do not provide the parameters for Flexipac because we do not have the official data from the vendor. RateFrac uses a build-in correlation for any packing and it does not require these parameters.
It is possible to search for these parameters on the Internet, or in some books and papers. For example, the book Distillation Design by Kister (page 446) has these parameters for Flexipac.
It is also possible to talk to the vendor to request these parameters. Sulzer provided all the packing parameters to us, but these data are marked as confidential and not accessible to the users.
Below are some approximate values for geometry parameters for some structured packings:
Side (mm) Crimp Height(mm) Base(mm)
Flexipac 500Y 9 6.36 12.73
Flexipac 250Y 18 12.73 25.46
Flexipac 125Y 36 25.46 50.92
Flexipac 64Y 72 50.92 101.82
These same dimensions should apply to the X versions of these same packings, except the inclination angle is 30 instead of 45 degrees.
BX gauze 8.9 6.4 12.7
These values will be added as defaults for v7.
Keywords: None
References: None |
Problem Statement: We like using the hierarchy block; however, there is often a problem with work flow. For example, if a modelling team member fine tunes the sequence of a flowsheet and imports this into another simulation as a hierarchy, the new sequence may not be the same as in the original standalone simulation. The new sequence may be more efficient, but in terms of decomposing the work this does not help. How do you make each Hierarch block have their own independent sequence? | Solution: Within the Hierarchy block, go to the EO Configuration | Solve Options | Specifications form and check Sequence separately using hierarchy. This option applies to the Sequential Modular (SM) run mode even though the setting is on the Equation-Oriented (EO) Configuration form of the Hierarchy blocks.
See the attached example .bkp file where the Sequence separately using hierarchy option is specified for both hierarchy blocks. When you click the Step button on the control panel, you can see the sequence generates two convergence blocks for the convergence of the 2 recycles inside the hierarchies and one for the overall recycle between the two hierarchies. If the Sequence separately using hierarchy option is unchecked (default), then a single convergence block is created for the simultaneous convergence of the two tear streams (which is correct, there's no need for 3 tear streams in this simulation).
Keywords: None
References: None |
Problem Statement: When I change the data set (say from 1 to 2) of a model of a property method and then try to use this method within a regression case to fit some data, the respective model does not appear in the drop-down list in the parameters tab in the regression case. Can you regress parameters for a second data set? | Solution: This is by design. Data regression can only manipulate the first data set. Users should do the regression for the first data set, then enter the obtained parameters for any data set number that they wish for the simulation run.
Keywords: DRS
References: : CQ00327851 |
Problem Statement: Have there been performance improvements with newer releases of Aspen Plus? | Solution: The simulation time has been decreasing with every newer version.
Following is a test with a large simulation file. The time information is from the history file. The elapsed time is the time it took to run the problem. If the time increases in newer releases, the machine may need more RAM and possibly a larger pagefile.
Please note that the time between Report generated and PDF updated is large in 2006, because of generating additional stream results for Custom Stream Summary, new in 2006, even for engine runts. However, when run from the GUI, if the stream summary forms are not opened yet, these calculations should not occur. In any case, it is only accounting for a relatively small fraction of the total time, and is taking place only after the simulation has converged. In 2006.5, the additional stream results are not generated until they are needed.
The machine we used was a WIndows Server 2003 with the following:
Intel 4 with 3.2 GHz CPU
1.98 GB RAM
2046 MB Paging file size
The versions were (see View History for the Engine build number):
2004.1 with CP7 (13.2.7 Build 27)
Keywords: None
References: None |
Problem Statement: What is the 'ProgID' needed to launch the Aspen Properties CAPE OPEN Thermo System? | Solution: The property system owns the property packages, and the system of interest must be registered. The Aspen Plus physical property system is registered as
AtCOProperties.COPropertySystem
This is mentioned in the reference documentation (e.g. page 285 of AspenPlusUserModelsV7_1-Ref.pdf). However the ProgID is actually version specific. You can add the internal version number to the ProgID:
? 2006.5: AtCOProperties.COPropertySystem.21
V7.0: AtCOProperties.COPropertySystem.22
V7.1: AtCOProperties.COPropertySystem.23
V7.2: AtCOProperties.COPropertySystem.24
Keywords: None
References: None |
Problem Statement: My model runs to completion in Sequential Modular (SM) mode. however, during Equation-Oriented (EO) synchronization, the following warning occurs...
* WARNING WHILE DETERMINING SPARSITY FOR CALCULATOR BLOCK: C-1
UNABLE TO CONNECT VARIABLE C-1.BLK.B1_MASS-FLOW_FLOWS_MIXED_XXX
TO ACCESSED VARIABLE B1.BLK.XXX_MASS_COMPA
DUE TO UNKNOWN ACCESSED VARIABLE EO-NAME. CONNECTION WILL BE IGNORED.
In this example, C-1 is a Calculator block, B1 is a SEP block, XXX one of the exit streams of block B1 and COMPA is a component in that stream. Therefore, the failed connection is from a Calculator block to a component flow in a SEP block. | Solution: To reduce the size of the EO problem, by default components are removed from streams if their concentration is zero. In the SMSolution, all of component COMPA goes to outlet stream XXX so COMPA flow in the remaining SEP block exit stream is dropped and the original mass flow spec for stream XXX is not put into the EO variable list. To get around this problem, open the Data Browser and navigate to Blocks | B1 | Block Options | EO Options. In the Remove components: field, choose Never. Now, when you synchronize to EO, the connection will always be made.
Keywords: Equation Oriented
EO
Remove components
EO synchronization
References: : CQ00315884 |
Problem Statement: A simulation appears to complete normally, but examination of the material balance (for either the whole simulation or one particular block) shows that the simulation is out of material balance. Why? | Solution: There are several possible causes.
1. The simulation contains a MULT block or a DUPL block. These blocks create/destroy mass in the simulation, so the simulation will inherently be out of material balance.
2. A DESIGN-SPEC (including RADFRAC SPEC/VARY) or a FORTRAN block may be varying the flow of a stream that is not a fresh feed stream. In general, this should not be done.
3. A FORTRAN block that changes the flow of a fresh feed stream is sequenced incorrectly.
4. The simulation may have a TEAR stream with a very large flow relative to all flowsheet FEEDS and PRODUCTS. The default tolerance for converging the TEAR stream (1.d-4) may need to be tightened to get the overall flowsheet in material balance. If so, the tolerance for individual blocks within the convergence loop should also be tightened.
5. The default convergence parameters may have been altered in an unsuitable manner.
A. The default convergence tolerance may have been increased. E.g. if the tolerance for a recycle stream is set to 0.1, it should be expected that the BLOCK next to the TEAR stream may be out of material balance by up to 10%.
B. The TRACE threshold may have been modified for the TEAR stream.
C. Convergence may only be required for a specific COMP-GROUP. Components not in the COMP-GROUP may be causing the block/flowsheet to be out of material balance.
If any individual blocks have an inlet flow, but no outlet flow, a SENSITIVITY block may be reinitializing all of the flowsheet streams (REINIT STREAMS=ALL). Instead, the user should supply a list of streams that should be reinitialized.
Keywords: Material balance problem,
References: None |
Problem Statement: How can I change components from one substream to another? | Solution: You can use a dummy RStoic block, as shown in the attached example file.
In stream 2, due to the chemistry defined, there is some NACL(S) in the MIXED substream.
The RStoic block B1 is set with the same temperature and no pressure drop, and with the reaction NACL(S)(Mixed) --> NACL(S)(Cisolid) with a fractional conversion of 1, so that all of it goes to the CISOLID substream.
Keywords: substream, solids
References: None |
Problem Statement: The temperatures calculated when reporting the TSOL and TFREEZ (temperature at which any component freezes out of a mixture at a given pressure) property-sets for a component in the 1st and 2nd liquid, the temperatures calculated do not seem correct. | Solution: This type of analysis cannot be applied if the phase composition changes, in this case due to splitting.
When dealing with two phases, the assumption for the property-set calculation is that the phase will remain as is. The property set is changing the temperature of one of the liquid phases and calculating the fugacity coefficient, this assumption could be incorrect as the two liquid phases could further split. If the mixture does not further split, the results are correct.
Note that in V9.0 and earlier, the TFREEZ values for the first and second liquid phases are reversed. I.e., the first liquid phase values reported are for the second liquid phase. The TSOL phases are correct.
Keywords: freezing
freeze out
References: : CQ00732690 |
Problem Statement: In a previous release of Aspen Plus or Aspen Properties, there was an option to add a button on the Toolbar that would allow access to a local version of the Detherm databank using the Detherm Server Setup Utility available from the Start menu. How do I enable this option? | Solution: Access to a local version of the Detherm databank from Aspen Plus or Aspen Properties is not supported anymore. The recommended workflow is to work within your local Detherm installation and then export the desired data into an Aspen Plus or Aspen Properties Data Regession System (DRS) input file. The INP file can then be imported into Aspen Plus or Aspen Properties.
Keywords: Detherm
References: None |
Problem Statement: How do you use NRTL-SAC parameters for a component in a simulation using another property method such as NRTL? | Solution: This document explains how you can estimate NRTL binary interaction parameters for a new component based on NRTL-SAC-derived parameters (such as NRTL-SAC parameters derived through solubility regressions). A license for Polymers Plus in addition to Aspen Plus is required to use this methodology. ThisSolution will be updated for V7.1 since the NRTL-SAC data forms have changed for that release.
NRTLSAC is the current thermo model within Aspen Properties. In V7.1, the NRTL-SAC model will be added which will have the new data forms. Both of these models are based on the underlying research done on NRTL-SAC.
Once the NRTL binary parameters are defined, the new component can be used in NRTL calculations with other components whose binary parameters are already specified. This allows you to perform activity calculations, VLE calculations, etc. using NRTL incorporating the new component.
The ability to estimate parameters for NRTL based on NRTL-SAC parameters is actually one of the capabilities within the Aspen Physical Property System. For instance, you could alternatively generate VLE data using the predictive UNIFAC property method and use the generated data to determine the binary parameters for the Wilson property method.
In the following example, however, we will focus purely on using NRTL-SAC parameters to determine NRTL binary parameters.
(Note - the following example is based on Aspen Properties 2006.)
In this example, we have started from an Aspen Properties aprbkp file which contains various solvents, as well as the NRTL-SAC regressed parameters (X, Y+, Y-, Z) for a component named DRUG. (See figure below.) The aprbkp file that we started with is called 130_Solvents_NRTLSAC.aprbkp, and it is included for your reference.
Now open the Properties folder in the tree on the left, and choose Specifications. NRTL has been chosen as the property method for this file:
Next we need to specify what type of data to regress. In the left pane of the Browser, click the Data folder. The Data object manager appears. Click the New button. In the Create New ID dialog box, enter an ID or accept the default. (We have named the new ID NRTLPRM in this example.) Then choose MIXTURE in the Select Type list box, and click OK:
On the Setup sheet, choose the type of property data in the Data Type list box. In this case, choose GEN-TPXY (near the bottom of the drop-down list). Next, choose which components you would like to regress NRTL binary parameters for. In this case, we will regress binaries for a Drug-Acetone system at 25 deg C:
Now select the Data tab. This tab will be used to specify the liquid phase compositions which will be used during the regression. On this tab, click the Generate Data button. In the Generate Binary VLE or LLE Data dialog box, select the NRTLSAC property method (at the bottom of the list of methods):
Now click the Generate button to fill in the compositions that will be used for the calculations:
Next, go to the Regression subfolder within the Properties folder in the tree on the left. Select the Binary subfolder. Within the Setup tab, choose NRTL as the property method (you may have to change it from NRTLSAC) and choose NRTLPRM (or whatever you used as the Data Object ID) as the Data set. (Delete any other Data sets that are already in this file. To do this, select the rows containing the additional data sets, then right-click and select Delete Row.) When you are done, the Setup tab should look like this:
Now select the Parameters tab. Choose Binary parameter in the Type field. For the Name/Element field, choose NRTL. Then click on the space next to the Name/Element drop-down menu (circled in red on the figure below) and enter a 1.
Continue filling out this dialogue box with the appropriate information. In this case, we have chosen Acetone and Drug. Note that we have also entered a binary parameter regression for Drug and Acetone (so that we get both parameters).
Finally, click the Start button (the triangular blue play button near the top) to run the regression. Choose the run the Binary regression (not Pure).
Click OK in the dialogue box above. When the regression is complete, you can go to the Properties/Regression/Binary/Results folder to see the regressed parameters:
To see how good the fit is, you can also click on the Residual and Profiles tabs.
The final Aspen Properties file built from this example entitled 130_Solvents_Final.aprbkp is included for your reference.
Keywords: None
References: None |
Problem Statement: Can components in a CISOLID or NC substream participate in electrolyte Chemistry? | Solution: Only components in the MIXED substream can participate in Chemistry. The CISOLID substream and the NC substream are inert with respect to phase equilibrium. Salt reactions need to involve a solid specied in the MIXED substream.
CISOLID or NC components can be moved to the MIXED substream (so they can participate in Chemistry) with an RStoic block.
Keywords: electrolytes
salt salts
References: None |
Problem Statement: When trying to populate an Aspen CIM-IO transfer record via Aspen Sqlplus, the following error appears:
Error reading memory in the repeat area. Cannot follow chain to field.
If you go to the InfoPlus.21 Administrator, enter some number in the repeat area (IO_#TAGS) and double click, there are no occurrences available. | Solution: The problem actually is no an Aspen SQLplus problem. Rather, it is an Aspen Infoplus.21 configuration issue. What is happening is that the user is Selecting a ghost field from a record in which the ghost selector field is not pointing at a correct selection from the ghost selector record.
For the specific IOGetDef record the repeat area of the transfer record has indeed links to 2 ghost selector records, io-deadbands and io-eu-conv. (The fixed area also has a link to the io-groups ghost selector). You need to check 1st_selection_value of the ghost selector, for example io-deadbands and make sure it is configured correctly (by default it has No deadband). It might be that it has a number instead.
Keywords: Repeat area
Can not follow chain
References: None |
Problem Statement: Is is possible to have an adiabatic SEP or SEP2 block? | Solution: There is no way of specifying an adiabatic SEP or SEP2 block. You need to use a design specification and vary the temperature. You also need a Calculator to equate all of the temperatures in all of the outlet streams. The Calculator block should be sequenced to execute BEFORE the SEP or SEP2 block.
An Aspen Plus example file. This file can be opened in 2004.1 and higher.
Keywords: None
References: None |
Problem Statement: This knowledge base article explains for which cases the Aspen InfoPlus.21 API routine WHIS21DAT() can return an invalid key timestamp error. | Solution: WHIS21DAT() would return an invalid key timestamp error if it could not insert an event because the specified key timestamp was:
(a) Outside the allowable range (1980-2035)
(b) Older than the oldest allowed time for the repeat area. Normally this would be set to roughly the time that the repeat area was created. See utility XOLDESTOK
(c) Older than allowed by the history repository's past time parameter
(d) Older than the T1 time of the oldest archive
WHIS21DAT() would also return an invalid key timestamp error if called in WHIS_TYPE_MODIFY mode and there was no existing history events having the specified key timestamp
Keywords: API
ROUTINES
PROCEDURES
ERROR
TIMESTAMP
WRITING TO HISTORY
References: None |
Problem Statement: Is it possible to rename a repository? (e.g. the default TSK_DHIS) | Solution: Changing the name a history repository can be done, but it requires several steps. It is generally not recommended unless you feel it is absolutely necessary.
To rename the repository you need to remove all references to it. In the case of TSK_DHIS, many records use it as the default repository.
The simplest way is to create a new repository and its filesets and then change the tags to use this repository instead.
The fileset files from the repository you are renaming (e.g. TSK_DHIS) can then be copied to this new repository. The old data will then be available for tags.
Note that filesets can be moved between repositories, as they do not contain any repository-specific information.
The steps below describe the approach that will work. It is suggested that you do this procedure on a test server first and take precise notes of all you do. Only when satisfied with the procedure should it be performed on the live server.
1. Create a new repository with the name you desire.
2. In the new folder on the hard drive you defined as the repository, define your new set of file folders for all the fileset files (identical to the existing ones).
3. In the InfoPlus.21 Manager, set any tasks that gather data to Skip During Startup (it is mainly CIM-IO tasks, but beware of other external programs you may have that gather data and write it to your history archives).
4. Shut down InfoPlus.21. At this point, if you have Store and Forward, you should be buffering data.
5. Physically copy all fileset files from their existing location to the new fileset locations of the new repository.
6. Start up InfoPlus.21. Because you have Skip During Startup for your CIM-IO and related tasks, no new data should be written to the old location.
7. Run an SQLPlus script to change the value of your IP_REPOSITORY field in all of your tags that curently use TSK_DHIS, to the new repository name.
8. Start your CIM-IO tasks manually from the InfoPlus.21 Manager list. Your Store and Forward buffer should now unbuffer to the archives in the new location, under the new name.
9. Remove all old filesets and delete the old repository, using the IP.21 Administrator. As a last step, remove the old set of folders on your hard drive.
Keywords: Renaming
Repositories
TSK_DHIS
References: None |
Problem Statement: When running the InfoPlus.21 Database Upgrade Wizard, the following error is received:
Error in setting new database size: <xxxxxxxx>. Code = -18
e.g.: Error in setting new database size: 64000000. Code = -18 | Solution: Increase the allocated memory in the command-line parameter for TSK_DBCLOCK so that the value is greater than the number in the error. For the example above, TSK_DBCLOCK's parameter should then be:
DOUBLE 65000000
or greater.
Keywords: Upgrade Wizard
References: None |
Problem Statement: When running the automated history backup record (defined by HistBackupDef) errors can result for the current active fileset. The following error message may appear in the h21arcbackup.log file.
ARCBACKUP ERROR: Could not find active fileset for repository <Repositiory_Name>. | Solution: This message occurs if the current active fileset has been marked as reserved. If the reserved status is removed from the fileset, the backup utility should function without further errors. The active fileset must have a status of Mounted, not Reserved and not Read-Only.
Keywords: automatic
archive
file
References: None |
Problem Statement: The Administrator tool may be very slow to expand, and show the InfoPlus.21 records when first opened. | Solution: If there are any invalid names in the ADSA directory server, it will cause long delays in opening the Administrator. The administrator checks each of the listed data sources and if one is invalid, or it is unable to connect (if InfoPlus.21 is shutdown on that data source), it will wait until it times out before moving on to the next ADSA data source. The time out is dependent on the dcom configuration for the entire network protocol, ranging from a 1 minute timeout to 20 minutes.
To prevent the long delay, remove any invalid data sources listed in the ADSA.
Keywords: performance
References: None |
Problem Statement: IO_Activate on the Get record(s) stays at 'Yes', does not change back to 'No'. | Solution: This problem is the result of the main and async tasks not being started. In one customer's case, these tasks were not started because of other problems. There was an issue with TSK_BCU_START, and since that was failing to start, the remaining tasks didn't start.
Once again, this was indirectly caused by using Terminal Services. The customer knew that he couldn't use the IP.12 Manager via Terminal Services, so he was stopping and restarting IP.21 from the Administrator. That doesn't give you the popup when there's a problem with a task. Going to the IP.21 server itself and using the IP.21 Manager, he was able to skip that task (then mark it as Skip during startup) and continue with the startup of the other tasks.
Keywords: cimio
cim-io
References: None |
Problem Statement: What does this error message in the error.log file mean:
ARCHIVE: h21_maybe_resync() error point <point id>
Where <point id> is some large number. | Solution: H21_maybe_resync() error point is generated by the history archiver when a call to an API function, h21_maybe_resync(), fails. The archiver just prints out the pointid. For example if the error is:
ARCHIVE: h21_maybe_resync() error point 33688603,
The actual record ID in question is 32899 (33688603/1024). Check that record ID for any unusual behavior or configuration. It is very unusual for the archiver to call h21_maybe_resync(). It is only called if a program has called WHIS21DAT() with the ADDNEW mode to queue a history event that has a timestamp that matches the key timestamp of a history event that already exists. An SQLplus INSERT could cause this.
Keywords:
References: None |
Problem Statement: What are the effects of enabling 21CFR11 Compliance in InfoPlus.21? | Solution: Activating the compliance-enabling features in IP.21 has important implications. To ensure the integrity of the system, it is required that once activated, the compliance-enabling features can never be deactivated, except by uninstalling the Aspen Framework software.
The reason for activating the compliance-enabling features needs to be clarified. The purpose for activating the compliance-enabling features is to activate special processing rules to help enforce compliance. Aspen Framework will always log security modifications even if the compliance-enabling features have not yet been activated. (Of course, this assumes that the client-side Alarm and Event components have been properly installed and configured.) In other words, Aspen Framework always logs changes. Therefore, activating the compliance-enabling features does not activate change logging, but instead activates the enforcement rules. The enforcement rules will disallow changes to security data if the Alarm and Events system is unstable. This is in accordance with regulations that require the logging of security changes. If the compliance-enabling features are not yet activated, then this rule is not enforced, and therefore changes to security data are not prohibited.
To summarize the above, the following restrictions are placed on an InfoPlus.21 database after Compliance has been enabled:
a. the inability to reload a snapshot;
b. the compliance-enabling features can never be deactivated, except by uninstalling the Aspen Framework software; and
c. if the Alarm and Events system is unstable, changes to security data will not be allowed.
Additionally, the following step should be performed on the Aspen Framework server:
a. If a site is planning to activate the compliance-enabling features for Aspen Framework, it is recommended that they limit the number of roles and securable objects that are created prior to activation. This is because there will be no official record of the changes to security data generated before the compliance features are activated;
b. Ensure that the Aspen Framework is properly installed and operational;
c. Ensure that the Alarm and Events server is properly installed and configured; and
d. Ensure that the Alarm and Event's client-side components are installed on the Aspen Framework server.
To test the system, you should first turn on Audit Trail only and use the Aspen Framework Security Manager to create a temporary role, and then delete it. If everything is properly configured, the previous changes should appear in the Alarm and Event's log. Use the Web.21 viewer to verify.
Keywords:
References: None |
Problem Statement: To users of the Aspen Manufacturing Suite products, one confusing item is the difference between activating tasks (executables) and activating database records. | Solution: ACTIVATING TASKS
When Aspen InfoPlus.21 is started, a number of External Tasks are started. Examples are TSK_SAVE, TSK_PLAN, TSK_IQ1, TSK_SQL_SERVER etc. When started, these external tasks do not immediately do any processing. They will show up as running but in fact they go to sleep and hibernate until woken up to do some processing. They are woken up when an Aspen InfoPlus.21 Database Record is activated and needs to use one of these tasks.
A very good example of this is a QueryDef record. This record typically contains a Query and a Schedule Time. When the Schedule Time is reached, the QueryDef record is Activated, the External task (such as TSK_IQ1) wakes up, and the activated record is put in the queue to be processed by TSK_IQ1.
Occassionally, you may wish to 'activate a record' from either a DOS prompt or from an SQLplus Query ...
ACTIVATING RECORDS
Despite the name, the way to Activate a 'Record' manually is via ACTTSK. This is both an Aspen InfoPlus.21 executable and an SQLPlus Command.
Suppose you have a record called Report1
From SQLplus, the ACTTSK 'command' syntax for activating the InfoPlus.21 record is as follows:
ACTTSK record [, priority]
for example
ACTTSK 'report1', 200;
ACTTSK can take the same clauses as SELECT. See the SQLplus On-Line Help for more details.
From a DOS prompt, the ACTTSK 'executable' is used to activate the record. This executable resides in the directory specified by the environment variable %SETCIMCODE%
For example:
%SETCIMCODE%\acttsk
Enter record name, or <CR> to quit : report1
Enter activation priority, or <CR> to quit : 200
Keywords: Activation
ACTTSK
References: None |
Problem Statement: How to use the Concatenation operator in Aspen Calc. | Solution: Concatenation combines two text strings, such as:
PlantArea = “FCC�
TagType = “Pressure�
Stream = “145�
ReturnValue = PlantArea & “_� & TagType & Stream
So, ReturnValue will equal “FCC_Pressure145�.
Keywords: None
References: None |
Problem Statement: A query stored in an Aspen InfoPlus.21 record has been accidently overwritten. Is it possible to recover the query? | Solution: Compressed queries defined by CompQueryDef, ProcedureDef and ViewDef are typically stored in the following path.
\Program Files\AspenTech\InfoPlus.21\db21\Group200\SQL.
To restore compressed queries, it is necessary to obtain a copy of the file(s) from a backup of the above directory. The backup must have been saved prior to the undesired changes.
Other queries--defined by QueryDef and TextViewDef--are saved as part of the database snapshot. To recover the original query, load a copy of the snapshot that was saved prior to the undesired changes. The snapshot can be loaded by right-clicking on the database in the Aspen InfoPlus.21 Administrator and selecting Load Snapshot.
Depending on site-specific procedures, queries defined by QueryDef and TextViewDef may also be saved to the server as .sql files (File > Save As...). In such a case, it is possible to recover the original query from the .sql file, allowing the problem to be corrected without loading an older snapshot. If the .sql file is opened in SQLplus (Record > Open...) and then saved (Record > Save [Save As...]), the snapshot will be updated with the query from the .sql file.
Keywords: SQL
definition
References: None |
Problem Statement: When starting InfoPlus.21 using the InfoPlus.21 Manager tool, the startup doesn't complete successfully and fails when running the TSK_DBCLOCK task.
The System event viewer may contain an error: Error creating event object, error code = 183 | Solution: This behavior is typically seen when attempting to use the InfoPlus.21 Manager tool via a Terminal Services connection.
TheSolution is to use the InfoPlus.21 Administrator utility to start InfoPlus.21 in this case, or log on directly to the InfoPlus.21 console and use the InfoPlus.21 Manager.
Keywords: terminal server
References: None |
Problem Statement: During installation of InfoPlus.21 (IP.21), you will be forced to reboot the computer. After the reboot, several installation steps are executed, among them the addgroup procedure. But, there are conditions that will cause addgroup to not complete properly. One of the symptoms is that in the IP.21 Manager, the list of defined tasks would appear empty. | Solution: #108434 lists several reasons why this might occur. Usually it has to do with permissions on the account that is being used.
However, there is another condition which may cause this.
If you attempt to run the addgroup executable manually, from a DOS prompt, you might see the following:
Creating group directories.
Updating registry and copying files...
The keyname is Software\AspenTech\Infoplus.21\5.0\group200
A security repository already exists for this database. Cannot create new one. Failed to add new group)
Attempting to execute the removegroup executable will not work either. Again, attempting to run it manually from a DOS window, it would say:
IP21 Access is denied
Solution
There is an entry in the registy that must be removed, before addgroup will run to completion. Go into regedit using Start | Run to HKEY_Local_Machine/Software/Aspentech/Security/Registry Security Server. When you highlight that, if there is an entry underneath it. (It will look like a long hex string.) Delete that entry. Do not delete the Registry Security Server, but the entry underneath it. Then try again to run addgroup. It should now run successfully.
Keywords: addgroup
security repository
References: None |
Problem Statement: DOES AspenTech Officially Support running Aspen Process Recipe and Transition Manager v2006.5 on servers that are using SQL Server 2005 SP3 as the Relational Database? | Solution: NO, with v2006.5 of Aspen Process Recipe and Transition Manager, AspenTech DOES NOT yet support running this software version with SQL Server 2005 SP3, and AspenTech does not yet have plans to support SQL Server 2005 SP3 for v2006.5.
However, even though AspenTech has not tested with SQL Server 2005 SP3 and therefore DOES NOT support running on that RDBMS version, it is a known fact that historically development and support have seen minimal impact with application and use of the various MS SQL Server updates that have occurred over the years.
It's possible that support for running Aspen Process Recipe and Transition Manager on SQL Server 2005 SP3 may be available starting with a later future version release SOMETIME AFTER V7.1 or BEYOND V8.0.
For SQL Server 2005 SP3 Support, please check back with AspenTech Support for Updates as needed regarding the Status of such Support in V7.1 or later versions.
Summary of Current Support and Plans for Future Support:
APR / ATM has NOT been officially tested on MS SQL Server 2005 SP3.
For initial v2006.5 software testing, these RDBMS Server software versions were used without any Service Packs:
MS SQL Server 2005
MS SQL Server Express 2005
By Construction phase in the v2006.5 development cycle, this software was upgraded to SP2.
MS SQL Server 2005 SP2 (Standard)
MS SQL Server Express 2005 SP2
NOTE that there were no software issues noted by these MS SQL Server SP2 updates
The same RDMS Server software [MS SQL Server 2005 SP2 and MS SQL Server Express 2005 SP2] was also used for testing in V7.1
For V8, expectations were to support the following Server RDMS software:
MS SQL Server 2005 SP2 (Standard and Enterprise)
MS SQL Server 2008 (Standard and Enterprise)
MS SQL Server Express 2005 SP2 (only if required for client RDMS)
Keywords: APR
ATM
Aspen Process Recipe
Aspen Transition Manager
SQL Server 2005 SP3
RDBMS Support
References: None |
Problem Statement: After restart AspenTech Calculator Engine, values displayed in AspenCalc calculation results do not match last value execution for bound and non-bound parameters (input or output). | Solution: ThisSolution provides a detailed explanation about the behavior of Aspen Calc for displayed calculation result.
Consider the next calculation:
The images show a simple calculation: LastRun will get the current date while the ReturnValue will receive a random number from a record called Value1 and saved it on another record called Result. Notice that LastRun is a non-bound parameter while ReturnValue is bound to an IP.21 record.
After execution of this calculation, AspenCalc show this:
The last execution values are displayed on AspenCalc and values bound to an IP.21 record are saved on the record repeat area (in this case 85.78146 is saved on record Result).
These values are displayed on AspenCalc even after AspenCalc has been closed and reopen it. Also, these values are updated after every calculation execution; but these values are lost if AspenTech Calculation Engine is restarted showing something like:
So, after closing AspenCalc, restart AspenTech Calculation Engine service and reopen AspenCalc, the last execution values displayed by AspenCalc do not correspond to the values displayed on calculation last execution.
The reason for this is that all values saved as part of the interface are volatiles, this mean; these values are not saved within the .atc file, but just kept in memory by the AspenTech Calculation Engine service. So, after a restart, since engine service do not have any values for any calculation, values displayed by AspenCalc correspond to the last value saved on last .atc update, usually made by the Calculation Wizard:
In this case, the value displayed on AspenCalc after the Engine service restart is the same value saved on last execution made by the Calculation Wizard (on 11/12/2014) which save changes in the .atc file.
The workaround to prevent confusion for displayed values on AspenCalc after AspenTech Calculator engine service restart is to execute the calculation before viewing it.Â
Calculations will execute via schedule, COS activation or, via Aspen SQLPlus. The following is an example Aspen SQLPlus script.
local calccmd, CalcName, Calculation;
calccmd = CreateObject('CalcScheduler.CalcCommands');
For each CalcName in calccmd.GetCalculationList Do
     Calculation = calccmd.GetCalculationObject(CalcName);
     Calculation.Execute;
End;
Keywords: AspenCalc
bound
non-bound
execution
References: None |
Problem Statement: When Aspen InfoPlus.21 starts up, one of the tasks that can take a long time to start is LOADDB. This task reads the snapshot file from disk and loads it into shared memory. What exactly is it doing and why can it take so long to start? | Solution: When Aspen InfoPlus.21 starts up, there are 7 tasks that have to start in this order:
TSK_C21_WIN_INIT
TSK_H21_INIT
TSK_H21_ARCCK
TSK_H21_MNTTAB
TSK_H21_PRIME
TSK_DBCLOCK
LOADDB
The H21 tasks initialize history and launches the h21archive.exe processes responsible for the processing and reading of historical data. TSK_DBCLOCK allocates the amount of shared memory necessary for LOADDB to load the snapshot into. Specifically, LOADDB reads the snapshot file referenced in the Command line parameters for the task. While it is loading the file into memory, it is synchronizing the last values collected in memory for the tags with what was last stored in the filesets on disk during the last InfoPlus.21 shutdown.
Using IP_AnalogDef records as an example, Aspentech recommends that the IP_#_OF_TREND_VALUES for each tag be set to a value of 2. This means that the last 2 values for each tag will be kept in memory. These 2 values most likely have been processed by the historian and are stored in the active fileset on disk. When Aspen InfoPlus.21shuts down, all of the in memory values are processed to disk. When InfoPlus.21 starts up, LOADDB synchronizes these 2 in memory values with the 2 most recent occurrences from history.
Suppose there are 2 values for Tag X in memory, one at 3:00 and one at 3:01 PM. At 3:01:30.0, Aspen InfoPlus.21 is shutdown. These values, if they haven't already been processed by the historian, are processed and stored in history on disk. Aspen InfoPlus.21 is restarted at 3:30 PM. LOADDB knows that the last 2 values that were in memory were at 3:00 and 3:01 - because these values and timestamps are in the snapshot. But, it also checks history to make sure that there is not something newer than 3:01. If it finds values newer than 3:01, it replaces what is in memory with what it found in history.
Loaddb is synching every tag's values in memory with what is on disk. This can take a long time if:
1. There are a lot of tags.
2. Loaddb has to look way back in history to find the values to synch.
3. Perhaps there is a virus scanner in the way.
Keywords: None
References: None |
Problem Statement: After a new installation of Aspen Calc, the error Unavailable Node:<nodename> Please delete/add nodes to your selection might appear on startup and the Aspen Calc application might hang at getting the nodelist.
This error appears because Aspen Calc is unable to locate the specific node name mentioned in the error. | Solution: Aspen Calc retrieves node names from the registry at startup. The Default nodes can be configured in the registry.
To correct this error, perform the steps mentioned below:
Open the Windows registry and check the following:
HKEY_CURRENT_USER\Software\Aspentech\Aspen Calc\Calc Sources
The node name must be mentioned against System Key as shown below:
If the node name is not present or needs to be corrected, Right Click on the key name and click on Modify.
If you want to have Aspen Calc to connect to more than one Node on startup, the Count key can be increased.
After making this change re-open Aspen Calc and see if it can connect to the nodes mentioned in the registry.
Keywords: Calc
Nodename
Unavailable
References: None |
Problem Statement: There are two types of Schedule Groups you can create in Aspen Calc:
· Interval – run at regular, evenly spaced intervals e.g. every 2 hours
· Periodic – run at a specified time e.g. every Monday or every 1st day of the month
What happens to Aspen Calc Periodic Schedule Group on the 30th or 31st day of each month, if the specified day of the month does not exist for a given month? | Solution: If the specified day of the month does not exist for a given month, then the Periodic Schedule Group configured to execute calculations on that day skips to the next month for which that day is valid. For example, if the day-of-month is 31st, then after running the January 31st scheduled calculations, it will set the next scheduled run date to March 31st.
If you want to ensure that your scheduled calculations run on the last day of every month, set your Periodic Schedule Group to execute on End of Month instead of specifying the actual day of the month.
Keywords: None
References: None |
Problem Statement: What should one do if the error,shift_archive() open failure is noticed in the error.log file for a repository? | Solution: The presence of the error shift_archive() open failure in the error.log file indicates that the archiving process was unable to successfully open an archive file (this is most likely encountered during a file set shift). If another file set is not available to shift into, the archiving process will shut down and the data will begin buffering to the event.dat file. In fact, the shutdown of the archiving process might have led one to examine the error.log where this error was discovered.
Below is list of steps that should remedy the problem where the archiving process cannot open an archive file.
Make sure that an archive folder exists for each archive file set. Pay special attention to the next file set that is due to be shifited into. At the same time, verify that the path to the archive file is correct from within the IP.21 Administrator. Do this by expanding the appropriate repository (such as TSK_DHIS) icon and double clicking on the 'File Sets' icon at the bottom of the record. In the File Set display make sure the FS_FILE_PATH and the CUR_FS_FILE_PATH fields point to the proper archive folder. If one determines the path to be incorrect at this point, the file set path can be modified by right clicking on the repository, selecting Properties, and modifying the pending repository path. The new repository path will be used once the database has been restarted.
It is also possible to obtain this error if the history repository (and file sets) are stored on a remote drive so that network problems prevent the h21archive process from successfully accessing the repository. Under NO condition should a repository ever be set up on a remote drive. Not only could the above error occur, but the system will be undoubtedly plagued with performance problems. All repositories and file sets MUST be installed on a local drive.
If the file path to the repository is deemed acceptable, one should next verify that the arc folder as well as all files in that folder (arc.dat, arc.byte, & arc.key) do not have read-only access. Read-only access will prevent the h21archive process from accessing the file set. From the IP.21 Administrator, right click on the file set number and select Properties. In the properties view make sure the Read-Only checkbox is not checked. Note: Unchecking Read-Only from this checkbox will not necessarily undo individually given read-only permissions on the individual files within the folder. For completeness, one should also verify that, within the arc folder, the arc.dat, arc.byte, and arc.key do not have read-only permissions.
Another possibility is one of the three files in the target fileset directory may be locked by another unknown process (for instance a network backup or virus-checking utility.) Attempt to delete the files in the fileset directory that is the target of the shift process. If it is impossible to delete one of these files manually, this indicates the source of the problem. It may be necessary to reboot the system to break the lock on the file. Once the file is no longer locked by another process the shift should happen successfully.
Keywords: archive
stopped
collect data
paused
historian
error message too long
shifting archive due to
open archive file failure
References: None |
Problem Statement: When LOADDB, the task that loads an InfoPlus.21 snapshot into memory, runs, the CPU in the NT Task Manager is usually running with some significant activity. There have been a couple of cases where LOADDB takes a long time to load the snapshot and, during this process, there is little or no CPU activity. LOADDB appears to hang. | Solution: There appear to be conditions where significant CPU time can be taken up by a virus scanner but NOT reported as such in the NT task manager. Also, exclusion settings in the virus scanner appear to not always be in effect, depending upon other settings in the virus scanner.
One example is V4.0.5 of Network Associates Netshield for WinNT. (This was once known as Dr. Soloman/Network Associates) The scanner was set to scan selected file types and for known macros in all files. And, in another setting, scan for unknown macro viruses.
A snapshot that had taken 5 minutes to load was now taking 50 minutes to load after an upgrade to version 4.0.5 of this virus scanner. And, during this time, the NT Task Manager showed no CPU activity by either LOADDB or the virus scanner software. When the scan selected file types and for known macros in all files was changed to scan selected file types only (.exe, .dll, .sys, .doc, etc...) and the scan for unknown macro viruses was deselected, the snapshot loading time returned to normal.
How was it determined that the virus scanner was the culprit? The number of files scanned is displayed when you click on the VShield icon next to the system clock. This was showing about 1,000,000 files per day whereas before the upgrade, it was reading about 500 per day.
So, in the event of a system displaying poor performance such as InfoPlus.21 startup times, it is a good idea to eliminate the virus scanner as a possible cause early on in the troubleshooting process.
Keywords: LOADDB
virus
cpu
References: None |
Problem Statement: What is the difference between a normal repeat area and a history repeat area? | Solution: InfoPlus.21 records can have two types of repeat areas. One type is a normal repeat area and the other type is a history repeat area. A normal repeat area is a place to store a list of values, like those found in Selector records; whereas, history repeat areas store long-term data for tags such as values and timestamps like those found in IP_AnalogDef records.
You can get to the Normal repeat area of a Selector record by expanding the #_OF_SELECTIONS field, and you can get to the history repeat area of an IP_Analog record by expanding the #_OF_TREND_VALUES field.
Often times we are asked if it is possible to delete history data out of the history repeat areas. The answer is NO, you CANNOTdelete history data out of the history repeat areas. However, you can change the value and/or quality status for the unwanted point within the history repeat areas to bad so that it won''t trend in Process Explorer, but you will not be able to completely delete or remove the history value from the database. One can use the UPDATE function within SQLplus to accomplish this task, and an example of the update function can be found starting on page 4-18 of the SQLPlus User''s Manual.
Keywords:
References: None |
Problem Statement: One or more external tasks fail with the following message in the IP.21 manager''s error file:
System error 5AF interfacing external task name of the task - process id | Solution: The WinError.h file then indicates that error (5AF) 1455 means: The paging file is too small for this operation to complete.
Increase the size of the windows page file in order to fix this.
In general, the performance of the system will not have been optimal if you receive this error message. It is advisable to increase your server''s physical memory as well.
Keywords: 5AF
1455
external task
References: None |
Problem Statement: When adding file sets using the Aspen InfoPlus.21 Administrator, the message
Failed to create directory for file set
is returned. This knowledge base article explains how to resolve this error. | Solution: The file sets are not always accessed directly as C:\Program Files\AspenTech\Infoplus.21\c21\h21\arcs, but rather they can be accessed as a shared directory such as IP21g200His. Make sure that both the directory permissions and the share permissions allow the User Account which adds the file sets to create a new directory and to add files.
To do this, open Windows Explorer. Navigate to the directory where the file sets are to be added. For example, the default location is c:\Program Files\AspenTech\Infoplus.21\c21\h21\arcs. This location may be different for your installation. Once you see the directory, move the mouse pointer over the directory name and click the right mouse button. An option menu will pop up. Select Properties.
In the Properties dialog box, choose the Security tab and make sure the user account used to add the file sets has permission to modify, read & execute, read, write, and full control. Next, choose the Sharing tab. And click on the Permissions button. Make sure the User Account has permission to change, read, and full control.
The User Account as defined here is the account which is used to start InfoPlus.21 Task Services. You can see which account is used to start this service by going to
Start | Settings | Control Panel | Administrative Tools | Services | Aspen InfoPlus.21 Task Services
Select the logon account tab.
Keywords: archive
fileset
permission
References: None |
Problem Statement: When attempting to save an InfoPlus.21 snapshot, you may be unable to save it, and get an error 1909. | Solution: This error means that the operating system is unable to (over)write the specified file. So, check the filename, write-rights, accessibility of the directory, etc.
Keywords: 1909
snapshot
References: None |
Problem Statement: This knowledge base article outlines AspenTech's current policy regarding the support of the Aspen Manufacturing Suite products with the Microsoft Windows Vista operating system.
Note: AspenTech's general operating system support policy can be found here | Solution: The releases of the Aspen Manufacturing Suite up to and including v2006 are not supported on the Microsoft Windows Vista operating system. Several problems were discovered when tests were done using v2006 on Microsoft Windows Vista.
Official support for Microsoft Windows Vista starts with v2006.5.
Keywords: OS
WinVista
References: None |
Problem Statement: Pages 3 & 4 of the Aspen InfoPlus.21 v2006 Release Notes discuss the new feature Aspen InfoPlus.21 Health Monitor. Please see Knowledge Base article 119859 for the Release Notes document.
http://support.aspentech.com/webteamasp/KB.asp?ID=119859
The work of the Health Monitor is performed by the external task defined in the Aspen InfoPlus.21 Manager called TSK_HLTH.
This article discusses three possible reasons why that task either does not run, or will not stay running.
1) License problems:
a) The document mentions the need for an IP.21 Health Monitor license key. There are several ways to check for a specific key, such as browsing the License File, or using the SLM Tool called WLMAdmin.
The required License Key needed here is SLM_IP21_HEALTH
If you don't have this license then the task TSK_HLTH will fail with the following message in the .OUT file:
No appropriate License found for this application
b) The very large majority of Aspen Manufacturing Licenses are checked during installation only.
The SLM_IP21_HEALTH is different in that it is a RUNTIME license.
In reality this license is only read once at 'runtime'. At that point a registry key is added to confirm that the key has been read from the license file:
HKEY_LOCAL_MACHINE\Software\AspenTech\InfoPlus.21\9.0\group200\InfoPlus.21_Health_Application
(where the 9.0 will vary depending on the current version).
The value of the key is encrypted and should not be changed.
c) If you believe you have the license key but it is not being read, and the registry key is not being added, it is probably because the license cannot be found at runtime.
The re | Solution: for this is to go to
Start | Programs | AspenTech | Common Utilities | SLM Configuration Wizard
From here you need to point to either your network based SLM server, or your Standalone License file. This tells the software which license file contains this specific license key.
2) Skip During Startup.
By default, during installation, TSK_HLTH is set to skip during startup in the IP.21 Manager.
When you are ready to use TSK_HLTH on a regular basis then the skip during startup box should be unchecked.
3) Database records not correctly loaded
Extra InfoPlus.21 database records are needed for the health monitor to run successfully.
During the upgrade from v2004.2 or earlier, you must 'upgrade the snapshot' so that these new records will be loaded.
Keywords: None
References: None |
Problem Statement: Is it possible to delete old data or test data from a repository and start it brand new?
Sometimes you have a repository that has old data (i.e. test data), such as a site that has a test system that contains old data. After the test phase is completed, they would like to remove the test data or old data collected in the repository. What is the best way to delete old data or test data from the repository to start brand new but still keep the old repository name?
Scenario:
Company A has a production system that just went through validation. They don't want to have validation data in their repository (i.e. test data that had been collected in the repository), and they want to start production in fileset 1 with no test data in it. Furthermore, they want the repository name to remain the same. | Solution: You only need to stop the IP.21 database and delete the arc.dat, arc.byte, and arc.key files out of each fileset folder of the Repository. Then restart the IP.21 database. Since there aren't any fileset files, the first fileset should mount and the rest should have a Status of NONE.
Keywords: fileset
repository
test data
old data
References: None |
Problem Statement: From the InfoPlus.21 Manager, you can create a custom list of Defined Tasks (TSK_xxxx) that execute when InfoPlus.21 is started. If you have multiple InfoPlus.21 databases, and desire the same list of tasks to be defined on all systems, it is possible to create the list once, and copy it over to the other computers. The list of Defined Tasks that you see in the InfoPlus.21 Manager is actually information stored in the Windows registry. It is, by default, not stored in any kind of file.
However, the registry settings for the Defined Tasks can be saved into a file, and this is the file you would copy to multiple computers. This is accomplished by using the File/Save Group Configuration menu item from the InfoPlus.21 Manager. The list is saved into a file with the extension of .ipg. This file can then be copied over to another system, and a File/Restore Group Configuration can be done. The Restore option is only available when IP.21 is not running.
However, at this point, on a computer where the .ipg file has been restored, if you tried to edit any tasks in the list, to change paths or parameters, you would see an error message stating you do not have permissions to make edits. There are some registry keys in Windows Registry related to security that will need to be modified. There is a unique system code for each computer imbedded in the registry, and the system code from the first computer is in the .ipg file that you copied to the second computer, and it needs to be modified. Also, the data source name of the first computer was in the .ipg file, and will need to be changed in the registry. | Solution: To change the registry keys, follow the steps below:
On the computer where you Restored an .ipg file:
Go to Start/Run..., and type regedit
Go to HKEY_LOCAL_MACHINE/Software/AspenTech/Security/Registry Security Server.
There will be a long hexadecimal looking number for that key. Copy this long string. This is the unique system code for this computer.
Go to HKEY_LOCAL_MACHINE/Software/AspenTech/InfoPlus.21/x.x/Group200/
On the right-hand side, find SecurityRepository. Its value will also be a long hexadecimal number, but it will be the system code from the first computer. Right mouse click on SecurityRepository, and choose Modify. Highlight the Value, and right mouse click, and select Paste. This replaces the incorrect system code with the correct one that you copied from the previous key.
On this same window, select Data Source Name. The value will be the ADSA data source name of the first computer. Right mouse click and choose Modify. Change the value to the ADSA data Source name of the computer you are now on.
Exit regedit. You should now have permission to make changes to the group configuration on the second machine.
x.x - your version of AMS software
Keywords: .ipg
group configuration
defined tasks
References: None |
Problem Statement: If the loaddb task fails, and the error file for loaddb says:
Database version is too old for InfoPlus.21.
this indicates that the snapshot Aspen InfoPlus.21 is trying to load is corrupt. | Solution: Stop Aspen InfoPlus.21
Go to the Aspen InfoPlus.21 manager and double click on loaddb in the Defined Tasks area.
Look in the Command Line Parameters field for the name of the snapshot you are trying to load:
%SETCIM%\%SETCIM_GROUP_NAME%\InfoPlus21.snp
InfoPlus21.snp is the name of the snapshot that is attempting to be loaded.
Go to C:\Program Files\AspenTech\InfoPlus.21\db21\Group200 and replace the old snapshot with a previously saved snapshot that is known to be good - ie. rename InfoPlus21.snp to InfoPlus21_corrupt.snp. Rename the last previousy good snapshot to InfoPlus21.snp.
Start up Aspen InfoPlus.21
Keywords: old
snapshot corrupt
InfoPlus.21
References: None |
Problem Statement: Not all drivers for Oracle 8 function correctly when used to access SQLA tags. | Solution: The following 8.1.7 drivers have been tested with v6 of the AMS Suite and an Oracle 8.1.7.0 view. The results are included below:
Oracle ODBC Driver
Result
8.1.7.0
Succeeded
8.1.7.4
Succeeded
8.1.7.5
Succeeded
8.1.7.6
Failed. pd_server hangs with r21_initialize: Logging In
8.1.7.7
Failed. pd_server hangs with r21_initialize: Logging In
8.1.8.10
Failed. pd_server hangs with r21_initialize: Logging In
Keywords: driver
Oracle
ODBC
References: None |
Problem Statement: How do you change IP_PLANT_AREA from a record defined by Select10Def to one defined by Select20Def? | Solution: By default, both IP_AnalogDef and IP_DiscreteDef records have the field IP_PLANT_AREA set to the record PLANT-AREAS, which is defined by Select10Def. That means the longest plant-area name can only have 10 characters. To allow for longer names (e.g. 20 characters), here are the steps to make that change:
1) In the InfoPlus.21 Administrator, go to IP_AnalogDef | Fields | #_OF_FIELDS_IN_REC
2) For IP_PLANT_AREA, change the FIELD_FORMAT_RECORD from the default PLANT-AREAS (a Select10Def record) to another selector record (defined by Select20Def).
Repeat these steps for IP_DiscreteDef if so desired.
Keywords:
References: None |
Problem Statement: Attempting to create a new record defined by FieldNameDef or FieldLongNameDef record results in the following error.
When clicking OK to acknowledge the error, the record is created in an UNUSABLE state with a field number value of zero..
Attempting to make the new record usable will result in the same error. | Solution: To resolve the problem, it is necessary to update the FIELD_NUMBER field with a non-zero value. Once a non-zero number is specified, it is possible to make the record USABLE.
A convention for field check codes that helps prevent duplicate field numbers is shown in the following list. If more than one application is being developed for the same InfoPlus.21 system, application developers should adopt additional conventions. Field Check Codes are the high six bits of Field Number
Field Check Codes
Field Numbers
Used By
0 ? 3
0001 - 0FFF
User Applications
4 - 31
1001 - 7FFF
Base InfoPlus.21, Layered Products, and future use
32 - 63
8001 - FFFF
Future Use
The Aspen InfoPlus.21.Database Developer's Guide describes the valid field numbers.The field numbers for each record do not need to be unique.
Keywords: invalid
value
field
number
fieldnamedef
fieldlongnamedef
References: None |
Problem Statement: The distribution of AspenTech's AspenOne v2004.2 is delivered as a set of NineDVDs
One is the Documentation DVD,
One is the Admin DVD.
The main Six are dedicated to the Processing Suite, such as Disc-2 AMS for Aspen Manufacturing Suite.
One is called Patches.
So what is the Patches DVD for? Should I use the Patches DVD instead of, or in addition to, one of the main six DVDs ? | Solution: The answer depends on the products you are installing or upgrading.
A 'very small number' of products, mostly Engineering, were not upgraded on the main Suite DVD for v2004.2.
In other words, the 2004.2 main DVD is actually a redistribution of the v2004.0 or v2004.1 only for those specific products.
For those products only you should use the Patches DVD to get to the latest patched v2004.0.x or v2004.1.x version.
See pages 41 and 42 in the AspenOne Update 2 Deployment Guide for a list of these few products as well as how to use the Patches DVD.
The 'very large majority' of Aspen manufacturing Suite (AMS) products, including Aspen InfoPlus.21, SQLPlus, Process Explorer, Batch.21 etc., WERE upgraded on the main Disc-2 AMS.
Therefore the ONLY way to Upgrade from ANY previous version to v2004.2 for these products is to use Disc-2 AMS
If you are upgrading from v2004.0 or v2004.1 then you will be presented with the Update method.
If you are upgrading from v6.0.1 or earlier then you will be presented with the Install method.
If this is a new installation of AspenTech software you will be presented with the Install method
Note:
The only AMS products for which the Patch DVD is needed are related to the Advanced Control Family.
e.g. AspenDMCPlus, AspenSmartStep, AspenIQ and AspenWatch
Again see pages 41 and 42 in the AspenOne Update 2 Deployment Guide for a list of these few products as well as how to use the Patches DVD.
Keywords:
References: None |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.