question
stringlengths
19
6.88k
answer
stringlengths
38
33.3k
Problem Statement: A question is sometimes asked of Support: How many tags per second can the Aspen InfoPlus.21 Historian handle? Can the Historian be overloaded? Can data be lost? The answer is yes, at some point, there is a threshold where the historian simply can't keep up anymore. But, most likely, the IP.21 historian can handle any load handed to it. If not, there is generally something in the configuration that can be modified to optimize performance.
Solution: A lot of factors affect the performance of history: how much memory, how fast disks are, number of repositories, number of tags per repository, scan rate, etc. When a user thinks that history data may be missing, most likely the data is in the history queue, but may not be viewable for a period of time making the data appear to be missing. If you wait a while, you should see the data. What do we mean by wait a while? All of the tag data is written to the history event queue (in memory) as part of the IP.21 API. The h21archive.exe process then takes over and pulls the data out of the queue and places it into the cache and then eventually to the filesets on disk. During the time that the data is in the queue, IT IS NOT VIEWABLE by any of the client tools. If you wait a while, the data should be available. Will increasing the history repeat area count field make a difference? No. It does not matter how many values are configured to be held in memory, the IP.21 API holds the data buffered in the history queue. Can the queue get backed up and make the period of time that the data is not available longer? Yes. Perhaps a fileset shift has to occur or the archive task is paused/stopped for some reason. What can you do to monitor performance? You can monitor the CHECK QUEUE (by right clicking on the repository in the IP.21 Administrator). If you are continually seeing the QUEUE STATE as OVERFLOW or DISK, but the h21archive.exe process remains running, then perhaps the history system is not able to keep up. There are a couple of configuration parameters that can be checked. What is the CACHE RATE? (This is found on the same CHECK QUEUE screen.) It should remain at or very near to 100%. There are occasions when it falls below 100%, but as a rule, it should be at 100%. If you are continually seeing this value below 100 percent, maybe the NUMBER OF POINTS configuration in the repository PROPERTIES needs to be increased. (Found by right clicking on the repository and choosing PROPERTIES.) This number needs to be greater than or equal to the number of tags collecting history data for that repository. Further to this, the CACHE SIZE (PROPERTIES\ADVANCED TAB) can be increased to hold more data in memory. The standard 256 bytes times the number of tags is the allocation in memory. If there is enough memory on the box, this 256 can be increased. It is recommended, as a precautionary measure, that Support is contacted before any changes are made to the CACHE SIZE. Keywords: cache queue 106671 References: None
Problem Statement: What is the maximum number of Aspen InfoPlus.21 points?
Solution: Any record in the database that can hold a history repeat area is considered to be a point. It is possible to have just over two million points if the machine has enough virtual memory. If history is turned on the amount of virtual memory is reduced allowing fewer points. It is recommended to not have more then a 250,000 records with history enabled in a single database. Keywords: Maximum Points Points InfoPlus.21 References: None
Problem Statement: How to determine, through an Aspen SQLplus query, the number of cache buckets in an Aspen InfoPlus.21 history repository, along with the number of points which will be historized for that repository?
Solution: Please execute the query, which is included in the attachments for thisSolution, using the Aspen SQLplus Query Writer. It is also included below as a reference (some sections have been color coded in case it is helpful)... ---------- local ats, objRepos, objReposCollection; local ActFileSet int; Local Numtags int; Local NumCacheBuckets int; Local SystemCommand char(256); -- -- The macro h21summaryPath contains the path to the utility h21summary. -- Change the macro if necessary. The query uses h21summary to get the number of points in the -- active file set of a repository. macro h21summaryPath = 'C:\Program Files\AspenTech\InfoPlus.21\c21\h21\bin\h21summary'; -- -- Connect to the IP21 History Administrator Object and -- get the list of repositories -- ats = createobject('Aspen.IP21.AtIP21HistAdmin'); objReposCollection = ats.getRepositoriesCollection; -- for each objRepos in objReposCollection do -- -- Get the active file set for the repository, the number of cache buckets, and the number of -- points being stored in the active file set. -- ActFileSet = objRepos.GetCurrentFileset; NumCacheBuckets = objRepos.NumberOfPoints; SystemCommand = '&h21summaryPath -r'||objRepos||' -a'||ActFileSet; NumTags = (select substring(7 of line) from (system SystemCommand) where linenum=1); write 'Repository '||pad(trailing objRepos to 16)||' Number of cache buckets = '||NumCacheBuckets||' Number of points historized = '||NumTags; end ---------- Please also seeSolution 000050633, How to get detailed Aspen InfoPlus.21 repository and fileset information using Aspen SQLplus script for additional, helpful details. Knowing the cache buckets and history points can assist IP.21 database administrators in sizing the cache effectively. The number of cache buckets should be slightly larger than the number of history points. Here is an example of what the output could look like: Repository TSK_DHIS Number of cache buckets = 1000 Number of points historized = 595 Repository TSK_DHIS_AGGR Number of cache buckets = 400 Number of points historized = 311 Keywords: History Administration Scripting Interface Sizing InfoPlus.21 history cache Cache slots WBB Burton Balch SQLplus References: None
Problem Statement: When trying to connect to the default Aspen Local Security Server URL, http://localhost/AspenTech/AFW/security/pfwauthz.asp, users get The page cannot be found followed by http Error 404 - File or directory not found messages.
Solution: Microsoft Internet Information Service (IIS) had Active Server Pages disabled. To correct this, open Internet Information Service Manager (IIS) and under Active Service Extensions and change Active Server Pages from Prohibited to Allowed and then reset IIS. Once this is changed to Allowed then the security page presents the blank page as expected. Keywords: http, 404, File not found References: None
Problem Statement: How to override the FIR model for Aspen Inferential Qualities that uses .mdl file (FIR)
Solution: The Aspen IQ application that was built with a FIR model with have 4 model parameters that can be changed in the Aspen Production Control Web Server (this can done only with an 'Engineering' access level): - Model Gain, MDLGAIN - Model Dead Time, MDLFILTDT - Model Filter Tau 1, MDLFILT1 - Model Filter Tau 2, MDLFILT2 By changing these model parameters, the original FIR model will be replaced by the SOPDT (Second Order Plus Dead Time) model defined by these parameters. Starting V11, DMC3 Builder includes a conversion tool for FIR to Laplace type models. The reason is that GDOT Dynamics are introduced in the Laplace form. To access this conversion tool go to Master Model Conversion. In the DMC3 Conversion tool you have the option to select “use standard form” and see the dumping and tau value for second order type models. Since Aspen IQ is asking for t1 and t2 then this indicates a second order overdamped (Dumping>1). An overdamped second order system may be the combination of two first order systems. In the above example: Tp1Tp2 = T^2 = 5.31^2 Tp1 + Tp2 = 2*5.31*1.15 So, you have to solve the system Tp1Tp2 = 28.1961 Tp1 + Tp2 = 12.213 Using solve function from excel TP1 = MDLFILT1 = 3 minutes and TP2 = MDLFILT2 = 9 minutes. As you can see, Aspen IQ dynamics are very simple. If you want to use a more complex model is better to keep your FIR curve. Keywords: IQ FIR SOPDT DMC3 Builder References: None
Problem Statement: This article describes in detail the procedure to upgrade Aspen CIMIO server from one version to higher version.
Solution: The upgrade of Aspen CIMIO could be either on the same server hosting an older version of the software or on completely new server. In both the cases, backup of below files from existing server hosting Aspen CIMIO should be taken: CIMO logical device definition file Windows services file Scenario 1 - Aspen CIMIO server is installed on a completely new server: Install the new version of CIMIO in target server (new server). Apply all the latest available patches for the new version, available on AspenTech’s support website. Copy the above backed up files from source CIMIO server and paste in the target server, in the location specified below for default installations. CIMO logical device definition file – C:\Program Files (x86)\AspenTech\CIM-IO\etc Windows services file – C:\Windows\System32\drivers\etc 4. Use Aspen CIMIO interface manager to create the CIMIO logical device. 5. Perform test API to confirm the CIMIO logical device is able to fetch data from OPC server Note: While creating the logical device, port numbers used for TCP\IP communication will be automatically selected by tool. Article #000094184 describes the detailed procedure to change port number for CIMIO logical device, created using Aspen CIMIO interface manager. Scenario 2 - Aspen CIMIO server new version will be installed on the same machine where older version of CIMIO resides. Uninstall CIMIO from the server Perform a registry clean up to remove older version registry setup Delete manually any AspenTech folders, if existing in server Perform a fresh installation of new version of Aspen CIMIO server Apply all the latest available patches for the new version, available on AspenTech’s support website. Copy the above backed up files from source CIMIO server and paste in the target server, in specified location. Use Aspen CIMIO interface manager to create the CIMIO logical device. Perform test API to confirm the CIMIO logical device is able to fetch data from OPC server In case of any errors with CIMIO communication, please check CIMIO_msg.log file located at C:\Program Files (x86)\AspenTech\CIM-IO\log. Please contact AspenTech support for assistance. Keywords: Aspen CIMIO server Server upgrade CIMIO upgrade CIMIO installation References: None
Problem Statement: How are NodeIds constructed by the InfoPlus.21 OPC UA Server for measurement nodes in the DA branch or IP_INPUT_VALUE node in the RAW branch?
Solution: We will demonstrate with an example of how the NodeIds are constructed by the InfoPlus.21 OPC UA Server for the Measurement Nodes in the DA branch or IP_INPUT_VALUE node in the RAW branch. Here we have used IP_AnalogDef::ATCAI record in InfoPlus.21 for reference. The NodeId is generated using a sequential combination of 5 different attributes of a record. Those 5 attributes in the sequence are as below: 1. Indicates if the node in InfoPlus.21 OPC UA Server is a repeat area field in the corresponding InfoPlus.21 record: 0 - Indicates it is not a repeat area field 1 - Indicates it is a repeat area field Example: 0 for IP_INPUT_VALUE/Measurement 2. Indicates the kind of node: 0 - Used for Nodes under RAW 1 - Used for Nodes under DA Example: 0 - Used for Objects->RAW->IP_AnalogDef->ATCAI->IP_INPUT_VALUE 1 - Used for Objects->DA->IP_AnalogDef->ATCAI->Measurement 3. Record ID of the tag: Example: 2078 - Record ID of ATCAI 4. Field ID: Example: 29785(0x7459) - Used for IP_INPUT_VALUE/Measurement 5. Definition ID: It is the Id of the definition record. Example: 971 - Id of IP_AnalogDef for ATCAI record. Step 1: Values of the above 5 fields are converted into a sequence of byte stream of length 20. Format of Byte Stream as per below screenshot: Step 2: This byte stream is then converted to a Base64 string. An internet search should help you find more information about converting a byte array into a Base64 string. Step 3: The Base64 string is finally prepended with 'b=' to represent an Opaque NodeId - the b= indicates that it uses a Base64 character string. Keywords: OPCUA IP21 References: None
Problem Statement: How have integer variables changed for 64-bit code?
Solution: Starting in V11, Aspen Plus is compiled as a 64-bit program and user models must also be compiled as 64-bit code. All compiled user models must be re-compiled for Aspen Plus V11. In 64-bit Aspen Plus, INTEGER variables are now 64-bit integers. We use the non-default /integer_size:64 option. For writing interfaces from C or C++, use the CPP_LONGINT type defined in APrSystem V11.0\Engine\Commons\Utilsx64.h for integer variables. The INTEGER variables used to hold Hollerith strings (used as arguments in some subroutines) are also 64-bit, though they will only use 32 of those bits to hold four characters per integer and the interface will work as it did before. (Use format string A4.) This results in the component formula no longer matching H2 and results in the message. If you have declared variables as INTEGER*4 or LOGICAL*4, you need to change these declarations to INTEGER or LOGICAL. Floating point variables are still 64-bit variables defined as REAL*8 or DOUBLE PRECISION in the user model interfaces. This does not change from past versions and should maintain maximum compatibility with results from past versions. Most names (Component IDs, Parameters, etc.) in Aspen Plus have a maximum of 8 characters and are stored and two integer words. Since INTEGER was changed from INTEGER*4 to INTEGER*8 when we converted to 64-bit, they are still two integer words, but they now have extra spaces in them. For example, ID(1)=’WATE’ ID(2)=’R ‘ became ID(1)=’WATE ‘ ID(2)=’R ‘ This is to maintain consistency and since they are printed with a format of A4 in the report file. Keywords: None References: None
Problem Statement: What do the messages ARCHIVE: Active/prime mismatch 2 a=175 p=174 t+123 store_event store_history failure 2 program shutdown that appear in the file error.log for a repository mean, and how can I correct the problem so I can start my repository?
Solution: You can see these messages when trying to insert old historical data for a point. In the message ARCHIVE: Active/prime mismatch 2 a=175 p=174 t+123, a is the active file set number for the repository, p is the file set number containing the most recent data for point, and t+ is the target file set number where the old historical data for the point is to be written. The file set containing the most recent data for the point is called the prime file set. When you insert old historical data for a point, Aspen InfoPlus.21 must update the history sequence numbers for the point beginning from where the historical data is inserted forward to the most recent recording. To do this, the historian must update the headers of all the file sets between the prime file set with the most recent data (pointed to by p) and the target file set (pointed to by t+). The message indicates one of the file set headers is incorrect, and this causes the repository to stop. Unfortunately, the message does not indicate which file set caused the problem. To correct the problem, start the Aspen InfoPlus.21 Manager, and select Actions->Repair Archive. Select the repository that is not running and repair every file set between the prime file and the target file set inclusive. This may take several hours. After repairing the file sets, restart Aspen InfoPlus.21. While the repository is stopped, Aspen InfoPlus.21 will buffer data being read from the process in the file event.dat stored in root folder of the repository, and the historian will recover the data after restarting Aspen InfoPlus..21. Keywords: repository stopped repository not running repository off no history collected References: None
Problem Statement: How can Aspen SQLplus be used to insert multiple analog tags into Aspen InfoPlus.21 (IP.21)?
Solution: 1. Download the tags.csv file in the attachments of thisSolution 2a. Copy the file to C:\temp -- OR -- 2b. Modify the SQLplus code below to accommodate the location where you have placed the tags.csv file 3. Use the following SQLplus code in the Aspen SQLplus Query Writer: INSERT INTO IP_AnalogDef (NAME, IP_DESCRIPTION, IP_REPOSITORY, IP_#_OF_TREND_VALUES, IP_VALUE_FORMAT) SELECT SUBSTRING (1 OF LINE BETWEEN ','), SUBSTRING (2 OF LINE BETWEEN ','), SUBSTRING (3 OF LINE BETWEEN ','), SUBSTRING (4 OF LINE BETWEEN ','), SUBSTRING (5 OF LINE BETWEEN ',') FROM 'C:\temp\tags.csv' WHERE LINEnum > 1; The csv file contains all the tags and related information; if modified to accommodate additional fields it must match the order of the fields in the .CSV file. NOTE: For more information on inserting into history, please see the Aspen SQLplus online help andSolutions #100949 and #103040. Additional Information: Keywords: Insert into Multiple tags References: None
Problem Statement: - How to check if Data Lookback has been turned on in an Aspen Mtell equipment set?
Solution: - There is no direct indicator of DataLookback applied to specific Equipment Set or Sensor in Mtell Agent Builder. The best way to know this is by exporting the Equipment Set profile as below:- 1) Open Aspen Mtell Agent Builder and select the Equipment set in focus (1) in the Machine Learning Tab 2) Click Export Equipment Set (2) to export the details of the equipment set in a .CSV file. 3) Specify the target Location and name (3) of the CSV file to save 4) Click Save (4) to execute the .CSV file creation 5) Open the CSV file and navigate to columns U and V, named DataLookBack and DataLookBackUnits, respectively. DataLookBack specifies the duration of lookback and DataLookBackUnits specifies the Unit in time. The naming convention starts from 0 for seconds and increases by increments of 1 for subsequent units like minutes, hours and days. Example:- U=1; V=2 is a DataLookBack of 1 Hour. Note:- DataLookBack applied to any one sensor of an equipment set is applicable to all the sensors in the equipment set while Agent processing. Keywords: Data Look Back Look-Back Equipment Set Look back CSV Sensor delay Timestamp delay DataLookBack References: How do I turn on Data Lookback in Aspen Mtell?
Problem Statement: How do I disable the default credentials for the Tomcat server? Sometimes a user may want to disable the default security options used by Tomcat server. This can either be done by setting a different password to the default “admin” account or creating a new account with its own password using the steps below.
Solution: Search Windows for AspenOne Credentials Under the Search Engine Configuration tab, check the Tomcat Basic Authentication for Search Security check box. In the User Name input box - Type in the username of your choice In the New Password input box - Type in the password of your choice In the Confirm Password input box - Type in the same password again. If the password matches, a text string shows as Passwords Match! as the screenshot below. Click the APPLY button to create the new user with the password or reset password for an existing user. A message box (like below) will pop up to inform the user on the status of configuration changes. Acknowledge the dialog by clicking the OK button. If it succeeds, please activate the configuration changes by doing the follows: restart IIS restart Tomcat through Windows Services, i.e. right clicking the Apache Tomcat 9.0 Tomcat9 and choosing Restart from the popup menu. Keywords: None References: None
Problem Statement: How to resolve the error “Number = -2147217900 Description=ORA-01735; Invalid ALTER TABLE option” received in APEM database wizard during migration for Oracle Database?
Solution: The problem is with the constraint statement in the migration script “ALTER TABLE EBR_USER DROP EBR_USER_KEY2”. This issue is approved for a fix in V12.0. For V11.0.1, the workaround would be a simple text change in the DB wizard file. The workaround would be: Locate the file in the DB wizard installation: EBRSV1600toV1601_schema_oracle.ddl Edit this file, change the fifth line as From: ALTER TABLE EBR_USER DROP EBR_USER_KEY2; to: ALTER TABLE EBR_USER DROP CONSTRAINT EBR_USER_KEY2; Keywords: Oracle, Constraint, Migration, APEM, -2147217900 References: None
Problem Statement: Is it possible to retrieve Aspen InfoPlus.21 (IP.21) history data using an OPC UA client?
Solution: Yes! When a connection is made via an OPC UA client to IP.21, browse down through the RAW section: Keep drilling down until the desired tag is found and then scroll through until the history field is located (IP_TREND_VALUE) in this case: Keywords: None References: None
Problem Statement: What are some ways to determine how long the Aspen InfoPlus.21 (IP.21) database has been running (commonly referred to as the amount of uptime)?
Solution: Here is an Aspen SQLplus query which could be used to determine how long the IP.21 database has been running: write 'The current time is: '||' '||getdbtime; write ''; write 'The InfoPlus.21 was started: '|| tsk_save.last_load_time; write ''; write 'It has been running for: '|| getdbtime - tsk_save.last_load_time || ' (hours : minutes : seconds.tenths of seconds)'; Here is an example of what the output would look like: The current time is: 16-OCT-18 18:04:54.8 The InfoPlus.21 was started: 15-OCT-18 11:02:32.5 It has been running for: +031:02:22.3 (hours : minutes : seconds.tenths of seconds) The query could be enhanced as follows (the additional line writes the result to an existing IP_TextDef record (called 'DA_Text01' in this example)): write 'The current time is: '||' '||getdbtime; write ''; write 'The InfoPlus.21 was started: '|| tsk_save.last_load_time; write ''; write 'It has been running for: '|| getdbtime - tsk_save.last_load_time || ' (hours : minutes : seconds.tenths of seconds)'; DA_Text01.ip_input_value = getdbtime - tsk_save.last_load_time; If the enhanced query is saved as a QueryDef record (called 'Dan - Uptime' in this example) and then scheduled (every ten seconds in this example): then the values could be displayed on an Aspen Process Explorer or aspenONE Process Explorer graphic: Keywords: None References: None
Problem Statement: How do I add old data to a new Aspen InfoPlus.21 record using the QTIMESTAMP function of Aspen SQLplus?
Solution: One might think that using the typical Aspen SQLplus update statements to place data in the IP_INPUT_VALUE and IP_INPUT_TIME fields within a record would work. For example: UPDATE IP_AnalogDef SET ip_input_time = '10-NOV-05 18:00:00.0', ip_input_value = 1; WHERE name = 'atcai'; When you do this, IP_INPUT_VALUE is transferred through compression and IP_VALUE_TIME is updated to the current time, which is not what we want. How do we work around the problem? TheSolution is to use the QTIMESTAMP function provided in Aspen SQLplus. Use an SQLplus statement like: UPDATE IP_AnalogDef SET ip_input_value = 1, QTIMESTAMP (ip_input_value) = '11-DEC-02 20:00:00.0' WHERE name = 'atcai'; The above statement will insert values at the correct time into history provided that the tag does not already contain history with a date that is newer (younger) than the date you're trying to enter. If neither of those conditions are true, you most likely will want to use the insert into history method described inSolution #102892. NOTE: If you encounter an error message that looks like the following: Error writing to <tagname> IP_INPUT_VALUE: Data changed but no history generated (bad key time stamp) at line <#>. You most likely need to increase the number of hours in your repository's PAST time parameter. Do this from the Aspen InfoPlus.21 Administrator tool, by right-clicking on the repository name, choosing Properties, and increasing the PAST parameter to be far enough in the past to encompass the time's effect. You may also need to run the XOLDESTOK.EXE utility, found in the code directory, so that your record can receive data older than when it was created. Instructions for using the XOLDESTOK utility are inSolution #103040. For newer versions of the Aspen software there is an XOLDESTOK function provided in Aspen SQLplus. Refer to the help files for syntax guidelines. Also, note that filesets must exist that span the time frame of the data you are trying to insert. Keywords: QTIMESTAMP history old data error inserting into history field is not changeable at line 100949-2 References: None
Problem Statement: How can the oldest value in history be found for a tag using Aspen SQLplus?
Solution: A query like the following can be used to interrogate history to locate the oldest value. This query assumes that the tag in question is defined against IP_AnalogDef: select Name as Record Name, min(ip_trend_time) as Oldest Value from IP_AnalogDef where name like 'ATCL10%' group by name Result: Record Name Oldest Value ------------------------ -------------------- ATCL101 09-MAY-16 09:05:12.0 ATCL102 09-MAY-16 09:05:12.0 ATCL103 09-MAY-16 09:05:12.0 Keywords: References: None
Problem Statement: How can you correct the message Invalid Data Source when displaying aspenOne Explorer plots and graphics?
Solution: First, close all open web browser (Microsoft Internet Explorer, Google Chrome, or Mozilla Firefox) sessions. Next, go to the web server and look for a file called: AtProcessDataREST.config located in: C:\inetpub\wwwroot\\AspenTech\ProcessData Open the file using Windows Notepad and look for the section: <!-- Default data source: If no data source passed to the service, then use this data source Specify ADSA data source name here --> <DefaultDataSource>localhost</DefaultDataSource> If 'localhost' is listed like it is above, change it from localhost to the data source name. For example, if the data source name is FlyingFortress then change it so it looks like this: <!-- Default data source: If no data source passed to the service, then use this data source Specify ADSA data source name here --> <DefaultDataSource>FlyingFortress</DefaultDataSource> Then save the file and open a Windows OS command prompt using a right-click and 'Run as Administrator' and issue the command: iisreset Then start a new web browser session, navigate to the aspenONE Process Explorer page, and see if the problem has been alleviated. Keywords: References: None
Problem Statement: What is the physical interpretation of the “Pipe Efficiency” term which is under Advanced/Methods tab?
Solution: Pipe efficiency term is like a Correction Factor. There is only a numerical/math meaning. This variable can be used to match real plant data, doing the following calculation: Dp fric corrected= Dp fric (from correlation) / pipe efficiency The pipe block uses correlation to calculate the frictional term of the pressure drop (Darcy, Beggs...,etc). These types of correlations are created from experimental data under fixed conditions (Specific compounds, range of T, P, or other physical properties). In case the simulation conditions are different from those used to create the correlations, the numerical calculations could be different from those measured in a real plant. Therefore, having this factor is useful to adjust the results. You can rely on using this factor to match the actual friction pressure drop data. Keywords: Pipe Efficiency, Factor, Aspen Plus References: None
Problem Statement: What Aspen SQLplus query may be used to disable replication for AW_MSCDEF tags whose AW_PROCESSING field is set to 'OFF' (that is, are set to 'Scan Off' in Aspen Watch)?
Solution: The following SQLplus query will do the job: -- -- Please save a snapshot before running this query -- -- Note - this query assumes that replication -- has been turned on at the database level -- for (select Name, recid as ID from AW_MSCDEF where AW_PROCESSING = 'OFF') do write Name || ' has been changed...'; setreplic(ID, 0); end; Keywords: None References: None
Problem Statement: Problems with the Aspen InfoPlus.21 Administrator tool on the Aspen InfoPlus.21 Server node - it won't expand even though it works fine on client machines. After a very long time, the Administrator on the server comes back with an error similar to: General RPC error: wait for callback: bad select!
Solution: This problem can be caused by lookup problems when there are two network cards on the server. There are two approaches that could help: Use Notepad and add the IP address and node name of the second card to the hosts file: %SystemRoot%\system32\drivers\etc\hosts View network connections - you can open this via the Control Panel. Then open the Advanced Settings dialog via the Advanced menu item in the Window menu bar - you can see this menu by clicking Organize | Folder and search options | View tab | tick the “Always show menus” checkbox. Change the order of the bindings. This is necessary even if the second network card has been disabled. A restart of TSK_ADMIN_SERVER task in Aspen InfoPlus.21 Manager may also be required. Keywords: NIC hangs hung freeze no response multiple network interface card References: None
Problem Statement: This problem has been seen when connecting to Siemens or Honeywell Experion OPC Servers via Cim-IO for OPC. When loading an IQ application, the load for the application is successful but the following messages are seen in Manage: Warning validating tag: DBVL <Tag Name>: CIMIO_OPC_CONNECTSINK_FAILED, Failed to Connect Sink, the Dadvise Method Failed Other checks to validate this is the same issue: Cim-IO Test API is able to read the tags without issues; GET is successful The tags also have a good status on the DCS The default CIMIO Logical Device in the IQF is configured properly under View > Options
Solution: The root cause of this issue was found to be a communication lag between CIMIO and OPC. This can be resolved by enabling cache reads for ACO applications using this KB article: How do I configure cache reads for ACO-based APC software? https://esupport.aspentech.com/S_Article?id=000074039 For controllers and applications running once a minute, setting the Frequency setting to 15 seconds should work. If the controller or application cycle is 20 seconds, do not exceed a Frequency setting higher than 5 seconds. Keywords: IQ, load, CIMIO_OPC_CONNECTSINK_FAILED, Failed to Connect Sink, the Dadvise Method Failed, Warning validating tag References: None
Problem Statement: This KB addresses a problem reported often when moving to V12. The problem related is when calling history plots a window pop up asking for authentication credentials. This issue has been observed on Microsoft Edge and Internet Explorer
Solution: This problem is Related to a Bug in Internet Explorer and there is not plan to fix in the short term. However, some potential fixes can be suggested: 1.- Change Browser: it has been observed that Chrome does not deliver this issue, only Edge and IE. 2.- Consult KB 000098308 and KB 000091262 as they suggest a couple of steps that could help with this issue from IIS. Keywords: PCWS, IE, History Plots. References: None
Problem Statement: How to enable Aspen Simulation Workbook (ASW) in Excel?
Solution: To enable ASW in Excel, open Excel (as administrator), go to File | Options | Add-Ins. On the bottom of the list, select COM add-Ins and click Go. The COM Add-Ins window should open. Select the desired version of Aspen Simulation Workbook VX, click “OK” and restart your machine. Only one ASW version should be selected at a same time if multiple versions are installed on your PC. If ASW is activated successfully you will see the Aspen ASW ribbon. Key words ASW, enable ASW, COM Add-Ins Keywords: None References: None
Problem Statement: The current
Solution: described a strange behavior that can be encountered on a DMC3 controller. If you have an MV with an ET target and you change one of the limits to something outside the current value, the SS Target takes a big jump and calculates a value equal to twice the limit. Solution In the next example, the Low Limit for the MV is at 0. The SS Target and the Current Value is around the same value. The ET Service Switch is OFF and the Controller behaves as expected after the low limit changes from 0 to 10: But if the ET Service Switch is ON, and the same changes are made, the SS Target is calculated to be twice the value of the limit: This problem can be solved by changing the parameter for TargetAlgorithmType (EPSMVPMX) from 2 to 4 2 - Legacy interior point QP algorithm 4 - Active set method - very robust but can be slow for Composite-size problems. Shadow prices are calculated for both LP and QP subproblems. Keywords: DMCplus, External Targets, Steady-State References: None
Problem Statement: This KB article presents a simple
Solution: for the error “CV has invalid equal concern error factor ECECMM” that can cause the controller to stay turn OFF Solution This problem is trigger by a logic error in transform variables. The workaround consists of changing the ECEs Using EngUnit (XFRMECEC) from Yes to No to fix the problem. This problem can be reproduced on Simulate and requires that the controller is redeployed to see the changes on PCWS. Keywords: DMC3 Builder, ECECMM, XFRMECEC References: None
Problem Statement: There is a reported defect (578716) in Aspen DMC3 Builder versions V12.0 and prior where users may see a time shift in variable prediction plots for both Subspace and FIR parameter trials when running case identification. This is inconsistent with the Dataset view of the vectors and may also result in the wrong model identification results. This shifting only occurs when there are missing data points in the dataset. This is a defect in the way that DMC3 Builder regularizes a dataset within a case during the model identification process. As a result of this defect, users might see differences in model identification in V12.1 when compared to model identification results from prior versions. This defect was fixed in V12.1 release, therefore the V12.1 results are correct and those prior versions are not. To see missing data points in DMC3 Builder's Dataset view, plot the vector in Distributed view, then right click on the vector and change the style from Interpolate to Point. This makes it easier to identify the gaps in the full dataset. If you are experiencing this issue, the workaround below should resolve it.
Solution: The target fix for this defect is V12.1 release. Since the bug lies in the way DMC3 Builder resamples irregular datasets during case model identification, the workaround for this issue is to Resample the dataset in the Datasets view first and then use the resampled data in the case identification. Keywords: dataset, shift, model, results, different, dmc3, builder References: None
Problem Statement: Version 2.5.1 client applications are unable to see data on 3.0 or later InfoPlus.21 database. Version 3.0 and later client machines are able to access the data from the database.
Solution: The problem may be resolved by stopping and restarting the TSK_ORIG_SERVER using the InfoPlus.21 Manager. The reason for the difference in client behavior is that TSK_ORIG_SERVER provides remote access for pre 3.0 client applications. Newer clients access the database via other server tasks. Five Remote Procedure Call (RPC) Servers are built into InfoPlus.21, all of which enable remote access of the database by various client applications. These RPC servers are separated into five external tasks, which can be started or stopped from the IP.21 Manager. TSK_ADMIN_SERVER - provides remote access for the InfoPlus.21 Administrator and the InfoPlus.21 Definition Editor. TSK_APEX_SERVER - provides remote access for all clients in the Aspen Process Explorer family. TSK_EXCEL_SERVER - provides remote access for the Aspen Excel Add-in (legacy - atdata.xlam). TSK_ORIG_SERVER - provides remote access for all the pre-V3.0 clients. TSK_DEFAULT_SERVER - provides remote access for all clients not connecting to the other four servers. Keywords: access remote trend 107018 References: None
Problem Statement: This
Solution: provides examples about ad-hoc calculations that can be used in Aspen Process Explorer as well as samples on how to use time-based functions, math methods as well as logical and conditional statements. Solution atcl101, atcl102, atcl103 and atcai are tag names that come from the out-of-the-box simulation example with Aspen InfoPlus.21. The tags are used here only as examples and correspond to IP_AnalogDef records. Mathematical operators Operator Example + =atcl101 + atcl102 - =atcl101 – atcl102 * =atcl101 * atcl102 / =atcl101 / atcl102 ^ =atcl101^2 Combinations of mathematical operators can be done by using “( )” for example: =atcl101 + (atcl102 - atcl103) =(atcl101 + atcl102) / (atcl103 – atcl102) =(atcl101 + atcl102)^2 / (atcl103 - atcl101)^2 =((atcl101 + atcl102)^2 / (atcl103 - atcl101)^2) + atcl101 Constant values (integer or decimal) can be written also as part of calculation: =atcl101 * 10.25 =atcl101 * (10.25 + atcl102)^3 Time-Based Functions For a detailed explanation of time-based function please consult AspenCalc Help files. Function Description Example TimeShift Returns the value of a tag at a time specified by an offset =TimeShift(atcai, 1:00) RateOfChange Returns the rate of change of the tag in units per minute =RateOfChange(atcai) RateOfAvgChange Calculates the rate of change of the time-weighted average value of a tag in units per min = RateOfAvgChange(atcai) MovingAverage Calculates the time-weighted average value of a tag =MovingAverage(atcai, 1.1) LeadLag Implementation of the Lead-Lag filter used in control systems =LeadLag(atcai, 1:00, 1:00) Math Methods For a detailed explanation of math methods please consult AspenCalc Help files. Function Example Abs =Abs(atcai) Acos =Acos(atcai) Acosh =Acosh(atcai) Asin =Asin(atcai) Asinh =Asinh(atcai) Atan =Atan(atcai) Atanh =Atanh(atcai) Ceiling =Ceiling((atcai, 2) Cos =Cos(atcai) Cosh =Cosh(atcai) Degrees =Degrees(atcai) Even =Even(atcai) Exp =Exp(atcai) Fact =Fact(atcai) Float =Float(atcai) Floor =Floor(atcai, 2) Int =Int(atcai) Ln =Ln(atcai) Log =Log(atcai) Log10 =Log10(atcai) Max =Max(atcl101, atcl102) Min =Min(atcl101, atcl102) Mod =Mod(atck101, atcl102) Odd =Odd(atcai) Pi =atcai * PI() Radians =Radian(atcai) Rand =atcai * Rand() Round =Round(atcai, 2) Rounddown =Rounddown(atcai, 2) Roundup =Roundup(atcai, 2) Sign =Sign(atcai) Sin =Sin(atcai) Sinh =Sinh(atcai) Sqrt =Sqrt(atcai) Tan =Tan(atcai) Tanh =Tanh(atcai) Trunc =Trunc(atcai) Comparison and Conditional statements IF…Then…Else statements can be used in conjunction with comparison operators Function Example > = IF atcl101 > atcl102 THEN atcl101 < = IF atcl101 < atcl102 THEN atcl101 = = IF atcl101 = atcl102 THEN atcl101 <> = IF atcl101 <> atcl102 THEN atcl101 And =IF atcl102 < atcl101 AND atcl101>atcl103 THEN atcl101 Or =IF atcl102 < atcl101 OR atcl101>atcl103 THEN atcl101 Special characters and Sources Braces can be used as part of a formula to enclose tags names with characters that could interfere with the behavior of the calculation like all mathematical operators. Supposing a tag name like “atcl-101”: Correct formula Statement ={atcl-101} + atcl102 + atcl103 Incorrect formula Statement = atclc-101 + atcl102 + atcl103 It is possible also to use tags from different data sources. Supposing two data sources name like IP21DA1 and IP21DA2: =IPDA1:atcl101 + IP21DA2:atcl101 This can be combined with previous formula constructions rules like: =IPDA1:atcl101 + (IPDA2:atcl102 - atcl103) In this case, data source for tag atcl103 will be defined by value in Data Source column. Note - if a data source has special characters in the name (like the dash in this name: C-47SkyTrain ) then surround the data source name in braces / curly brackets: =PEGASUSBRIDGE:feed + {C-47SKYTRAIN}:feed Keywords: ad-hoc Process Explorer References: None
Problem Statement: How does a network license function within my organization?
Solution: A network license file will consist of token or non-token license keys. The Software License Manager (SLM) Server needs to be installed on a machine to host the network license file. This will allow the machine to be a license server. Users within the organization will have SLM Client Tools that are configured to connect to the license server. When the user opens an Aspen application, the machine will connect to the license server and check out the necessary license keys for the application to run. The user’s machine will need to be able connect to the license server through the organization’s internal network. NOTE: The user’s machine does not need access to the Internet to connect to the license server with the organization. Keywords: aspenONE SLM Sentinel RMS License Manager References: None
Problem Statement: Attached to this KB article is the Aspentech.BPC.S95WebServices_TestUI.exe test tool. The tool is provided for download and use “as is” without any documentation or support.
Solution: The attached BPC S95 Web Services test tool contains two tabs: The Commands tab that contains a list of available methods, and… The Results tab that contains the results of invoking various commands Click on each command to invoke a particular method. Web Service Method Arguments dialog opens. You must supply a Collection Identifier (such as Equipment.Equipment for instance), Timestamp and Item Values (such as Name, Description, etc...). Keywords: None References: None
Problem Statement: The following symptoms of the issue can be seen: Unable to start or restart the data collection for an application in Aspen Watch The RepositoryErrors.log file (found here on the Watch Server: C:\ProgramData\AspenTech\APC\Performance Monitor) shows this error: ---> System.ApplicationException: Failed Updating records: Exception Failure - Query timeout at AspenTech.ACP.DataPump.Watch.IP21DbHelper.ReloadHistoryRecords(AppWatchInfo appInfo)
Solution: This issue is caused by the SQL query for Aspen Watch Data Pump timing out. Please follow these steps to resolve it: On the Watch Server, navigate to this directory: C:\Program Files (x86)\AspenTech\APC\Performance Monitor Make a backup copy somewhere else (just for your reference) and then open this file to edit in a Notepad as administrator: WatchDataPump.exe.config Navigate down to the line that says: <!-- add key=operationTimeoutInSeconds value=90 / --> Then un-comment this line and increase the operationTimeoutinSeconds value to 270. This is what that should look like after your changes, a copy of the updated file is also attached here: <add key=operationTimeoutInSeconds value=270 /> You will need to restart the Aspen APC Performance Monitor Data Service for the changes to take effect. Wait 5 minutes and you should be able to stop and start collection for all controllers. Keywords: start, collection, aspen watch, query, timeout, operationtimeoutinseconds, watch, data, pump References: None
Problem Statement: What Calcscript formula can be used to perform a one hour rolling average on a tag in Aspen Calc?
Solution: The TagAverage command returns the average of a specified tag over a specified time period. Syntax: TagAverage( tagname, <start time>, <end time>, [Actual], [Stepped] ) The Now( ) command is used to return the current date and time. Syntax: Now ( ) equals the current time Now ( ) - X equals the current time minus X (X is defined as the number of days) Now ( ) - 0.041667 equals the current time minus one hour The TagAverage and Now( ) commands can be combined to calculate a rolling average. Syntax: a = TagAverage( b, Now( )-0.041667, Now( ), 0, 0 ) ** It's also possible to use a time display format, like hh:mm ** Example for 1 hour: Now ( )-01:00 a=TagAverage(b, Now( )-01:00, Now( ), 0, 0) Keywords: Now() Now ( ) Tagaverage Aspen Calc Rolling Average Calcscrip t 105249-2 Converted from 105249_Default.txt References: None
Problem Statement: Are' HistoryBackupDef' command entries activated simultaneously or in a sequence?
Solution: In records defined by HistoryBackupDef the following fields, if they are filled in with commands, are activated in this order: 1. Active ('Last Active Command' field) 2. Shifted/Changed ('Last Shift Command' field / 'Changed Command' field) 3. Post Backup ('Post Backup Command' field - located in fixed area of the record) Keywords: 130846-2 References: None
Problem Statement: What is one possible cause of the AFW Security Manager initialization failed message when launching AFW Security Manager?
Solution: You may see this error if both 'Aspen Framework' and 'Aspen Local Security' are installed on the same server. To correct this problem you need to remove one of these products. Since most Aspen InfoPlus.21 installations use Aspen Local Security, this article describes how to remove Aspen Framework. Use the Uninstall AspenTech Software link beneath the AspenTech program group to uninstall Aspen Framework. Reboot the server if prompted. Next perform an installation repair for Aspen Local Security to replace components that are in common between Aspen Framework and Aspen Local Security. Reboot the system if requested when the repair process completes. Keywords: 141965-2 References: None
Problem Statement: Aspen HYSYS V9.0 Stream Reporter (HSR 1.7)
Solution: HYSYS Stream Reporter (HSR) is an Excel spreadsheet utility that allows to import to a spreadsheet the material stream information such as conditions, properties and compositions, and also compare streams from different cases. HSR can report properties from the following phases: Overall, Vapour, Light and Heavy (Aqueous) Liquid, Combined Liquid and Solid. It also allows stream user variables and property correlations to be reported. It is also possible to create formulae in the output table. The user can save sets of properties or use one of the pre-built property sets. Streams from different HYSYS cases can be reported in the same stream table. Once a stream table has been generated it can be updated by pressing a single button. Stream tables can be moved to another Excel workbook whilst maintaining the ability to be updated. HSR takes the form of an Excel spreadsheet file with embedded Visual Basic for Applications (VBA) code that demonstrates how HYSYS can be accessed programmatically. The VBA source code is freely accessible and users are encouraged to learn from it and adapt it to their own needs. For V8.0 - V8.8 please see the KB Article 057412. For older versions see the KB Article 054553. Note This Automation application has been created by AspenTech as an example of what can be achieved through the object architecture of HYSYS. This application is provided for academic purposes only and as such is not subject to the quality and support procedures of officially released AspenTech products. Users are strongly encouraged to check performance and results carefully and, by downloading and using, agree to assume all risk related to the use of this example. We invite any feedback through the normal support channel at [email protected]. Keywords: HYSYS Stream Reporter, HSR References: None
Problem Statement: Why ASTM D86 distillation curve are empty for the stream?
Solution: The results of ASTM D86 can be empty if almost all components in the stream are individual light end components. For such a situation, D86 curve is not applicable to characterize this stream. The D86 curve is converted from TBP curve by API 3A11 method. Users can search this method via API manual. We need get the TBP curve in the valid range. In Aspen algorithm, it's required 50% TBP >=0 F. Otherwise, it is out of the range of the conversion. For a stream including only a finite set of individual small molecules, distillation curve is not a recommended item to evaluate the stream properties. Key words ASTM D86, distillation curve, D86 empty Keywords: None References: None
Problem Statement: How do I enable debug logs for Aspen Mtell applications?
Solution: This article will help the user to enable Debug logs on Aspen Mtell Agent Builder, Aspen Mtell System Manager, Mimosa Explorer, Aspen Mtell Agent Services, Aspen Mtell Training Services, Aspen Watch Dog Service, Aspen Mtell Asset Sync, Aspen Mtell Work Sync applications. Enable debug log on Aspen Mtell Agent Builder 1. Go to C:\Program Files\AspenTech\Aspen Mtell\Suite\Tools\Agent Builder\ folder 2. Open AssetHealthAnalyzer.exe.config file 3. Scroll down to <logger name=* minlevel=Info writeTo=logFile/> script 4. Replace the word Info with Debug 5. Save the file. 6. Close and open Aspen Mtell Agent Builder Enable debug log on Aspen Mtell System Manager 1. Go to C:\Program Files\AspenTech\Aspen Mtell\Suite\Tools\System Manager\ folder 2. Open Mtell.SystemManager.exe.config file 3. Scroll down to <logger name=* minlevel=Info writeTo=logFile/> script 4. Replace the word Info with Debug 5. Save the file. 6. Close and open Aspen Mtell System Manager Enable debug log on Mimosa Explorer 1. Go to C:\Program Files\AspenTech\Aspen Mtell\Suite\Tools\Mimosa Explorer\ folder 2. Open Mtelligence.MimosaExplorer.exe.config file 3. Scroll down to <logger name=* minlevel=Info writeTo=logFile/> script 4. Replace the word Info with Debug 5. Save the file. 6. Close and open Aspen Mtell Agent Builder Enable debug log on Aspen Mtell Agent Service 1. Go to C:\Program Files\AspenTech\Aspen Mtell\Mtell Agent Service\ folder 2. Open Mtell.Cbm.Service.exe.config file 3. Scroll down to <logger name=* minlevel=Info writeTo=logFile/> script 4. Replace the word Info with Debug 5. Save the file. 6. Restart Aspen Mtell Agent Service Enable debug log on Aspen Mtell Training Service 1. Go to C:\Program Files\AspenTech\Aspen Mtell\Training Service\ folder 2. Open Mtell.MachineLearningService.exe.config file 3. Scroll down to <logger name=* minlevel=Info writeTo=logFile/> script 4. Replace the word Info with Debug 5. Save the file. 6. Restart Aspen Mtell Training Service Enable debug log on Aspen Mtell Watch Dog Service 1. Go to C:\Program Files\AspenTech\Aspen Mtell\Suite\Watch Dog Service\ folder 2. Open Mtell.WatchDogService.exe.config file 3. Scroll down to <logger name=* minlevel=Info writeTo=logFile/> script 4. Replace the word Info with Debug 5. Save the file. 6. Restart Aspen Mtell Watch Dog Service Enable debug log on Aspen Mtell Asset Sync and Aspen Mtell Work Sync 1. Go to C:\Program Files\AspenTech\Aspen Mtell\Suite\MDM\ folder 2. Open Mtelligence.MdmPhysicalAssetSyncService.exe.config file 3. Scroll down to <logger name=* minlevel=Info writeTo=logFile/> script 4. Replace the word Info with Debug 5. Open Mtelligence.MdmRuntimeDataSyncService.exe.config file 6. Scroll down to <logger name=* minlevel=Info writeTo=logFile/> script 7. Replace the word Info with Debug 8. Save the file. 9. Restart Aspen Mtell Asset Sync Service 10. Restart Aspen Mtell Work Sync All the log files are locations in C:\ProgramData\AspenTech\Aspen Mtell\ folder Note: Enabling the debug log will increase the log file size very rapidly. Please make sure to revert the changes after debugging is completed. Keywords: debug log verbose log detail log References: None
Problem Statement: From V12.0 onwards, some properties inside Sulphur Recovery subflowsheets are marked as <empty>, for example mass heat of vaporization and the True VP and the Reid VP at 38.7°C.
Solution: These properties used to be calculated in the previous versions: Opposite to V12.0: This is because it was compared that the values predicted by the SULSIM Sub-Flowsheet property package were not reliable in comparison with other thermodynamic methods, so the properties were removed altogether. The workaround is determining these properties outside of the Sub-Flowsheet. Keywords: HYSYS, empty, sulphur, recovery, SULSIM, mass, heat, vaporization, vp, reid, true References: None
Problem Statement: How to delete rows or columns in an ASW table without breaking the link to the HYSYS file?
Solution: For deleting rows in the ASW without breaking the linking with the HYSYS file, follow the next steps: 1. Click on one cell of the table. The ASW variable Table appears. 2. Select Modify Table. 3. Simulation Workbook Table Wizard will appear. 4. Click either on the “Columns” or “Rows” tab, select the columns or rows you want to delete, and click on “Remove”, then click on “Finish. 5. Now you can erase the values of the columns that you Removed before without breaking the link with the HYSYS simulation. Keywords: Workbook, row, column, Table Wizard, excel, delete, ASW. References: None
Problem Statement: The Convergence Monitor does not update until the end of the run. How do I get it to update after every iteration?
Solution: The Convergence Monitor button in the Control Panel is used to to open a panel displaying the convergence history graphically. In the Control Panel toolbar click the Run Settings button and then, click the Options tab. If Dynamic update of calculation results is checked, then the graphical convergence output updates as the simulation runs. If it is not checked, it only updates at the end of the run. Keywords: None References: None
Problem Statement: Why doesn't my pressure profile in the RadFrac Profiles result form match the pressure drop specified on the Specifications | Pressure sheet? How does update pressure drop in RadFrac affect simulation results?
Solution: Tray and packing calculations in Column Internals do not affect column separation by default – they are supplementary calculations. However, options are available to use the computed pressure drop (under Column Internals) to update the column pressure profile. By selecting Rating in Column Internals, users have three options regarding incorporating hydraulics rating results into column simulation: Don't update pressure drop: rating results from Column Internals will NOT affect the column simulation, i.e. the results on the Column's Profiles | TPFQ sheet will be consistent with the input on the Specifications | Pressure sheet. Update pressure from top stage: rating results (pressure drop) from Column Internals WILL update column simulation, i.e. pressure drop on the Column's Profiles | TPFQ sheet will be consistent with the pressure drop calculation from Column Internals. And the pressure at Top Stage of the rated section will be the same as the user provided specification in Column's Specifications | Setup | Pressure sheet. Update pressure from bottom stage: rating results (pressure drop) from Column Internals WILL update column simulation, i.e. pressure drops on the Column's Profiles | TPFQ sheet will be consistent with pressure drop calculation from Column Internals. And the pressure at Bottom Stage of the rated section will be the same as the user provided specification in Column's Specifications | Setup | Pressure sheet. Keywords: Column Internals Column Hydraulics Pressure Profile References: None
Problem Statement: I am trying to model the power consumed by an agitated vessel, but there does not seem to be a way to do so. Is there any way to model electricity consumption for stirring tank?
Solution: Aspen Plus doesn’t calculate power consumption for stirring tank regardless it’s an agitated vessel or reactor. Nevertheless, users can take advantage of properties calculation used by Aspen Plus to create a correlation in a Calculator block for this purpose. An example file is provided to demonstrate how to achieve this goal. Users can revise the correlations in the example depending on the agitator type, operating conditions, and fluid characteristics. Keyword Agitator Impeller Propeller Turbine mixer Power / electricity consumption Keywords: None References: None
Problem Statement: This knowledgebase article documents the
Solution: to the following error message which may be returned from the Tag Browser: S95 can't be connected. BPC S95 Web service is not responding Solution If you are receiving this error, and your site has not implemented an S95 configuration, then disable the S95 Search option. Open the Tag Browser. In the Search menu make sure that 'S95 Search' is unchecked. Keywords: References: None
Problem Statement: This
Solution: talks about the possibility and suggestion for the installation of APC Desktop applications (DMC3 Builder, DMCplus Model, IQ config, PID Watch etc…) Solution The APC Desktop applications are normally installed under three scenarios: When Installing the APC Online Server Components When Installing the APC Performance Monitoring (AspenWatch ) Server When Installing only APC Desktop In the two first scenarios, the installation of the desktop tools as default and cannot be disabled because the communication and further configuration is done through the desktop applications (DMC3 Builder, and Watch Maker). One thing that you will notice on any of the scenarios mentioned before, is that is not possible to install desktop application separately. The media content by design compresses all desktop applications in a pack called APC desktop applications, this not only contains the media for those applications it also contains additional media required for some of these applications to work properly as MES package and AFW package. However, if some of these components are not required, they can be uninstalled. Furthermore, you can customize the windows start menu to isolate one specific APC Desktop application. Keywords: APC installation, APC Desktop applications References: None
Problem Statement: This knowledge base article describes how to prevent write failures to an OPC server when the error message The AsyncWrite Function Failed is reported in the cimio_msg.log.
Solution: The AsyncWrite() function is an OPC server function call which has been seen to fail on various OPC servers. In order to work around this problem, the option to perform synchronous writes was introduced into Cim-IO for OPC v2.3. Users running older versions of Cim-IO for OPC will need to upgrade to obtain this feature. To enable synchronous writes run OPCProperties.exe, which can be found in the following folder: Program Files\AspenTech\CIM-IO\io\cio_opc_api Next, check the 'Perform Synchronous Writes' option as shown below. Finally, restart the Cim-IO for OPC service (or the Cim-IO Manager service if your asyncdlgp starts from the cimio_autostart.bat file) for the change to take effect. From CIM-IO software version 7.3.0.3 onwards the setting for 'Perform Synchronous Writes' can be located and selected within the CIM-IO Interface Manager as shown below. In these versions the configuration will need to be saved as shown and the CIM-IO Interfaces will need to be restarted for this setting to take effect. Keywords: Sync Async Put Fail References: None
Problem Statement: This KB article provides the instructions to enable logging of Infoplus.21 OPC DA.
Solution: 1. Copy IP21DaServer.dll.config from ..\AspenTech\InfoPlus.21\db21\code\ to a temporary location such as C:\Tmp 2. Rename ..\AspenTech\InfoPlus.21\db21\code\IP21DaServer.dll.config to IP21DaServer.dll.config_bak to preserve a backup of original file. 3. Edit C:\Tmp\IP21DaServer.dll.config in notepad. 4. Please make the below changes in Red Color <threshold value=TRACE/> <root> <level value= TRACE/> <appender-ref ref=LogFileAppender /> </root> <appender name=LogFileAppender type=log4net.Appender.RollingFileAppender > . . . <conversionPattern value=C:\ProgramData\AspenTech\DiagnosticLogs\InfoPlus.21\IP21DAServer\IP21OPCServer.log /> . . . <filter type=log4net.Filter.LevelRangeFilter> <param name=LevelMin value= TRACE /> <param name=LevelMax value=FATAL /> </filter> 5. Copy C:\Tmp\IP21DaServer.dll.config and replace the one at ..\Program Files\AspenTech\InfoPlus.21\db21\code\IP21DaServer.dll.config 6. Use OPC Client application and connect to IP21DA Server to browse the tags. Note: If OPC DA client is already running, stop it first, then reopen it. This is to make sure that the new task started will read in the changes that were done in the config file. For detailed information on stopping and starting the IP21DA_Server.exe see article 000077908. 7. Send the IP21DAServer.log at C:\ProgramData\AspenTech\DiagnosticLogs\InfoPlus.21\IP21DAServer to AspenTech for analysis. Key Words: OPC DA Debug Logging IP21DA_SERVER Keywords: None References: None
Problem Statement: Is it possible to use Aspen InfoPlus.21 as an OPC server?
Solution: Yes. Aspen InfoPlus.21 includes an OPC-DA server that conforms to the OPC Foundation DA specifications. The OPC-DA specification describes standard COM interfaces that would allow an OPC-DA client application to request current process values and attributes (that is, tag data) from an OPC-DA server. An OPC-DA server knows how to fetch the requested process data, commonly known as tag data, from a process data source such as distributed control system (DCS) or process information system. OPC-DA client applications, whether developed by AspenTech or third parties, can then access InfoPlus.21 in the same way that they access other OPC-DA compliant process data sources. AspenTech applications that can access InfoPlus.21 via OPC-DA include Cim-IO for OPC, Aspen Calc, and Aspen Recipe. The format for specifying an InfoPlus.21 database field as an OPC item is: definition record.record name.field name For example: IP_AnalogDef.ATCL101.IP_INPUT_VALUE The OPC server name (as seen from OPC clients) is Aspen.InfoPlus21_DA. In the the following screen capture, the presence of the OPC server is shown in the Distributed COM Configuration Properties interface. It is possible to establish a connection to the Aspen.InfoPlus21_DA OPC server using any OPC client. The OPC client used in this example is Matrikon's OPC Explorer. The final screen capture shows the same Matrikon OPC client after it has been connected to the server. An OPC group and three items (tags) have been added. Keywords: OPC DA server client OPC DA References: None
Problem Statement: AspenTech's Knowledgebase article 000079263 comprehensively describes how to use Aspen InfoPlus.21 as an OPC server. It describes how it should be set up and gives examples of its use. This article discusses the actual executable that gets used on the IP.21 system, and how it gets Started and Stopped.
Solution: The executable is called IP21DA_Server.exe This is located in the following directory : C:\Program Files\AspenTech\InfoPlus.21\db21\code From the Windows Task Manager on the IP.21 box, under Processes, when started, it will be displayed as: IP21DA~1.EXE The DA server is a COM object and as such is launched whenever it is instantiated by a client. When a client instantiates it, a reference counter is incremented for each client. When the reference counter is greater than one the server process will hang around indefinitely. When the last client disconnects the counter will go to zero and after some period of time the DA Server process will go away on its own. It doesn't need to be started at reboot or by a service because it is registered as a COM object. Typically it is not a good idea to terminate a COM object unless you have a for sure way of knowing that there are no COM objects using it. If for some reason the client goes away and doesn't tell the server then it will stay in the queue forever and may have to be terminated with the Windows Task Manager. Keywords: OPC OPC-DA IP.21 References: None
Problem Statement: How do I establish connection between OPC HDA Server and Aspen Mtell server?
Solution: Aspen Mtell only support OPC HDA connections, it does not support OPC DA connections. You could either Install Aspen Mtell Adapter for Honeywell PHD on OPC HDA server or on Aspen Mtell Server. It's recommended to install the adapter on OPC HDA server to avoid opening DCOM ports. Installation of the Adapter on OPC HDA server or Aspen Mtell Server 1. Copy Aspen Mtell installation media on OPC HDA server or Aspen Mtell Server. 2. Run Setup.exe As Administrator 3. Select Honeywell PHD Sensor Adapter under the product selection page 4. Install the required pre-requisites (required .NET can be found in 3rd party Redistributables folder in the installation media). 5. Finish installation and reboot system as prompted. 6. To confirm if Aspen Mtell_PHD Application pool is installed, launch IIS Manager and select Application Pools, confirm if Aspen Mtel_PHD is listed 7. Select Sites and make sure port 80 http is configured 8. Expand Sites/Default Web Site/Aspentech/AspenMtell/Adapter and right click HoneywellPHD and select Manage Application and click Browse. 9. It should launch IE and the page should show ASP.NET Open required ports 1. If the adapter is installed on the OPC HDA Server then open HTTP port 80 between the Aspen Mtell Server and OPC HDA server. Port 80 also needs to be open between OPC HDA server and any client machine that wishes to view data. 2. If the adapter is installed on the Aspen Mtell Server then open all DCOM ports between the Aspen Mtell Server and OPC HDA server Configuring and checking connection between adapter and OPC server 1. Go to C:\inetpub\wwwroot\AspenTech\AspenMtell\Adapter\HoneywellPHD\bin\ folder and launch Mtell.Sensor.Honeywell.Phd.Configuration.exe 2. Enter the Server Name of the OPC HDA server 3. Enter the ProgID 4. Click Test button to check the connection 5. Click Save button Configuring and checking connection between Aspen Mtell System Manager and Adapter 1. Launch Aspen Mtell System Manager on the Aspen Mtell Server 2. Click on the Configuration tab. 3. Select Sensor Data Sources in Settings. 4. Click Add Data Source 5. Enter a Name for OPC HDA Server 6. Click on OK button. Select the following: Source - Plant Historian Historian - Honeywell PHD (OPC UA) 7. Insert the hostname or IP of the machine hosting the adapter into the highlighted region. Example: http://PHDServerName/AspenTech/AspenMtell/Adapter/HoneywellPhd/ 8. Click on the Test button to test the configuration between Aspen Mtell System Manager and the adapter. 9. Click on the Save 10. Click on the Map Sensors button 11. Click on the Refresh button. The Tags and their values will be displayed after clicking Refresh button, as below Keywords: OPC HDA connection OPC Server OPC historian connection References: None
Problem Statement: In Unscrambler, is it possible to have the constants for SNV or MSC computed over a different region from which a PLS model is computed?
Solution: Yes, it is just a matter of selecting a different range when doing the preprocessing. For autopretreatment, all pretreatment variables must be included as well. Keywords: None References: None
Problem Statement: This knowledge base article provides an example of how the Aspen InfoPlus.21 history data compression algorithm works.
Solution: Using an example of a 60 second time interval (which we will consider to be 1 time unit), the tag has IP_DC_SIGNIFICANCE set to 1 and the first four values are 25, 22, 20 and 18. The first value is always recorded. In fact, any data value following a bad status will be stored. When the second value is scanned, the slope of the line between the first value and second value is calculated using this formula: (Value2 - Value1) / (Time2 - Time1) or (22 - 25) / 1 which is -3 (negative three). When the third value (20) is scanned, a prediction of what the next value should be is calculated, using the slope above and the time interval. The basic calculation is: Last_recorded_value + (slope * time_interval) For the example, this is 25 + (-3 * 2) = 19. Then the absolute value of the difference between the new value and the predicted value is calculated. For the example this is 20 - 19 = 1. Since this result is not greater than the significance (1), the next to most recent value (22) is not sent to history but is replaced by the most recent value (e.g., in IP_Input_Value). When the fourth value (18) is scanned, a prediction of what the next value should be is calculated, using the slope above and the time interval. This is 25 + (-3 * 3) = 16. Then the absolute value of the difference between the new value and the predicted value is calculated. For the example this is 18 - 16 = 2. Since this result is greater than the significance (1), the next to most recent value (20) is sent to history and is replaced by the most recent value (18). Put another way, the most recent value breaks compression so the previous value is stored in History. The slope is calculated between 20 and 18 for use with the next data point scanned. Keywords: boxcar back slope References: None
Problem Statement: Although there are other options, the most common form of Data Compression used within IP.21 processing is the BoxCar Slope Algorithm. This is the one used by the standard database Definition records IP_AnalogDef and IP_DiscreteDef. Knowledge Base article How do value/slope calculations work with historical data compression? gives an excellent description of the way the Numeric or 'Slope' Algorithm works with relation to the Value portion of an incoming sample or occurrence. There are two other important factors in deciding exactly which samples are stored when compression is turned on. There is the fact that IP_AnalogDef and IP_DiscreteDef not only contain a Value, but also a Quality Status such as Good, Bad, Initial, etc. There is also the fact that each Value/Status pair has an associated Timestamp. These may be regularly spaced as with scheduled GETs from an Aspen Cim-IO system, or they may be irregularly spaced as with Unsolicited. This article discusses the checks that the software performs with relation to Quality Status and Time.
Solution: Here are the exact steps the software performs followed by some real examples. First: IP.21 obtains the maximum time interval from IP_DC_MAX_TIME_INT. IP.21 obtains the last recorded time from the first occurrence of IP_TREND_TIME. IP.21 obtains the 'previous' input value, 'previous' input quality status, and 'previous' input time from IP_INPUT_VALUE. IP_INPUT_QUALITY, and IP_INPUT_TIME. IP.21 updates IP_INPUT_VALUE, IP_INPUT_TIME, and IP_INPUT_QUALITY with new data. IP.21 obtains the 'new' input value, 'new' input quality status, and 'new' input time from IP_INPUT_VALUE, IP_INPUT_QUALITY, and IP_INPUT_TIME 1) Quality Status Test: If the new input quality status differs from the previous input quality status, two steps are performed : 1. IP.21 updates IP_VALUE, IP_VALUE_QUALITY, and IP_VALUE_TIME with the 'previous' input value, input quality status, and input time. This causes a new occurrence to shift into history. 2. IP.21 updates IP_VALUE, IP_VALUE_QUALITY, and IP_VALUE_TIME with the 'new' input value, input quality status, and input time. This causes a second new occurrence to shift into history. 2) Elapsed Time Test (1): Note-1 The Elapsed Time Tests are not performed if the Quality Status Test was positive (i.e. resulting in recordings). Note-2 The Elapsed Time Tests are not performed if the Slope Test was positive (i.e. resulting in a recording). IP.21 calculates the difference between the new input time and the last recorded time. If the maximum time interval exceeds zero but is less than the calculated time difference, AND IP_VALUE, IP_VALUE_QUALITY, and IP_VALUE_TIME are not already identical to the 'previous' input value, input quality status, and input time THEN IP.21 updates IP_VALUE, IP_VALUE_QUALITY, and IP_VALUE_TIME with the 'previous' input value, input quality status, and input time. This causes a new occurrence to insert into history. 3) Elapsed Time Test (2): Note that the Elapsed Time Test(2) will always be performed after the Elapsed Time Test(1) - whether it had positive or negative results. IP.21 calculates the difference between the new input time and the last recorded time (which may have changed after Elapsed Time Test (1). ) If the maximum time interval exceeds zero but is less than the calculated time difference: THEN IP21 updates IP_VALUE, IP_VALUE_QUALITY, and IP_VALUE_TIME with the 'new' input value, input quality status, and input time. This causes a new occurrence to shift into history. ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Real life examples A) Data incoming regularly every 5 seconds and the incoming value never changes. ? Up to and including 00:01:05 the Quality Status was 'Good' From 00:01:10 onwards the Quality Status was 'Bad' From 00:02:20 onwards the Quality Staus was 'Good'. Using the Quality Status Test above, The Values stored in History would have a Quality and Timestamp of :- Good 'the last time the Quality changed to Good' Good 00:01:05 Bad 00:01:10 Bad 00:02:15 Good 00:02:20 Good 'the last time before the Quality changed to Bad' B) Data incoming regularly every 5 seconds. The incoming Value never changes. The incoming Quality Status never changes. The IP_DC_MAX_TIME_INT = 1 minute Using the Elapsed Time Test(1) above, Values stored in History would have Timestamps exactly one minute apart, BUT, the value is not stored in History until the 'next' set of Value, Timestamp, and Quality appear in the 'input' value fields (i.e. 5 seconds later). C) The incoming Value never changes. The incoming Quality Status never changes. The IP_DC_MAX_TIME_INT = 1 minute. Data incoming at irregular time intervals as follows :- 00:00:00 00:00:30 00:01:15 00:02:30 00:05:00 ? At 00:00:00 being the very first value it will be stored in History At 00:00:30 The IP_Input Fields will be updated with the new values but both Elapsed Time Tests will fail At 00:01:15 Elapsed Time Test (1) Will cause the value at 00:00:30 to be stored At 00:02:30 Elapsed Time Test (1) will cause the value at 00:01:15 to be stored, and, Elapsed Time Test (2) will cause the value at 02:30 to be stored At 00:05:00 Elapsed Time Test (1) will fail. Elapsed Time Test (2) Will cause the value at 00:05:00 to be stored. Keywords: References: None
Problem Statement: This
Solution: frames the problem of a repeated prompt for logon credentials when HTP plots are enabled and history Plots are requested. The plot would appear showing no data and the windows prompt for logon would appear and will not access even if the right credentials are introduced. Solution To fix this is necessary to change Process Data to Anonymous Authentication. The way to do this is by following the next steps: 1.- Access to IIS manager for Windows Start 2.- Expand Default Web Sites and Select Web21 3.-from Web 21 select Process Data and from the right-side display look for the Authentication icon 4.- Double click on Authentication and change Anonymous Authentication from Disable to Enable and Disable Windows Authentication. 5.- After this open a command prompt using Admin rights and execute and IISRESET command. Keywords: PCWS, History Plot, IIS References: None
Problem Statement: How do you model batch drying?.
Solution: Starting in V11, the Dryer block available from the Solids tab of the Model Palette now has a Batch mode which allows convective dryers to be modeled in batch flowsheets. Batch mode can only be used if the main flowsheet is of Batch type or if the Dryer block is within a BatchProcess block. In batch mode, rather than the dimensions of the dryer determining the residence time, specify the batch feed time directly, or specify the batch charge amount which will determine the feed time based on the feed flow rate. The batch discharge time and the down time collectively determine the batch cycle time. Aspen Plus will calculate the total cycle time as the sum of: Feed time + Drying time + Discharge time + Down time An example of batch drying is attached. In V11, if the wet solid contains gas such as Nitrogen, the batch dryer calculation is not correct and can lead to errors. The work around is to use a Sep block to separate out the gas before the Dryer. This is illustrated in the file batchdrying_final-V11-Working.bkp. Keywords: None References: : VSTS 585697, 601190
Problem Statement: How can I enable or disable tag replication through Aspen SQLplus?
Solution: Use the following queries to update the tag replication flag for any Aspen InfoPlus.21 tag: To Enable FOR (SELECT RECID as ID from <Tag Name>) DO SETREPLIC(ID,1); END; Example 1 (this will enable replication for the record ATCAI): FOR (SELECT RECID as ID from ATCAI) DO SETREPLIC(ID,1); END; Example 2 (this will enable replication for ALL the IP_AnalogDef records - it is the same as right-clicking on IP_AnalogDef in the Administrator and choosing 'Enable replication for All'): FOR (SELECT RECID as ID from IP_AnalogDef) DO SETREPLIC(ID,1); END; To Disable FOR (SELECT RECID as ID from <Tag Name>) DO SETREPLIC(ID,0); END; Disabling is done in a similar fashion - what changes is that a zero ( 0 ) is passed as the second parameter to SETREPLIC (instead of a 1). It is in bold red above. Note: Ensure that 'Enable global replication for this IP 21 database' is checked in order to engage replication. This setting is visible by right-clicking the database and selecting 'Properties' followed by the 'Replication' tab: Keywords: SETREPLIC RECID 139213-2 References: None
Problem Statement: How do you determine/verify how many tokens are consumed by the Aspen InfoPlus.21 database?
Solution: Aspen InfoPlus.21 will consume tokens based on the point count specified in the InfoPlus.21 Administrator. You can determine the consumption of tokens for InfoPlus.21 by dividing the Total Licensed Points by the IP.21 Points Ratio. The value for the SLM_IP21_PointsRatio key can be found in the SLM License Profiler under the last column labeled Vendor Information: Total Licensed Points / IP.21 Points Ratio = Tokens Consumed 66,000 points * (1 token / 4000 points) = 16.5 Tokens are consumed as a whole (no fractions), so this IP.21 is using 17 tokens. Keywords: tokens IP.21 InfoPlus.21 total licensed points points ratio set point count 137894-2 References: None
Problem Statement: When implementing tag replication in Aspen InfoPlus.21 (IP.21), customers may place a firewall between the Aspen InfoPlus.21 systems. What communication ports need to be opened in the firewall for replication to work correctly?
Solution: InfoPlus.21 replication uses Microsoft Message Queueing (MSMQ) as the transport in Windows Communication Framework (WCF) and therefore relies on the MSMQ transport layer. The communication ports required specifically for MSMQ are documented in this Microsoft KB article: http://support.microsoft.com/kb/832017 (Original Article) https://docs.microsoft.com/en-US/troubleshoot/windows-server/networking/service-overview-and-network-port-requirements (Current Article) which lists the communication ports used by that application. Article excerpts: Message Queuing The Message Queuing system service is a messaging infrastructure and development tool for creating distributed messaging programs for Windows. These programs can communicate across heterogeneous networks and can send messages between computers that may be temporarily unable to connect to each other. Message Queuing helps provide security, efficient routing, support for sending messages within transactions, priority-based messaging, and guaranteed message delivery. System service name: MSMQ MSMQ Port Information Application Protocol Protocol Ports MSMQ TCP 1801 MSMQ UDP 1801 MSMQ-DCs TCP 2101 MSMQ-Mgmt TCP 2107 MSMQ-Ping UDP 3527 MSMQ-RPC TCP 2105 MSMQ-RPC TCP 2103 RPC TCP 135 These are the firewall ports that need to be opened for IP.21 replication to work. Keywords: InfoPlus.21 tag replication firewall MSMQ WCF 130208-2 References: None
Problem Statement: It is possible to enter a Lab Sample with a future timestamp using the DCS interface for an Aspen Inferential Qualities sensor, however, there are two situations that could happen: in the first one, the value of the sample is taken as zero and the Bias does not update; and in the second one, the application has the expected behavior of taking in account the Lab Sample when the future time comes.
Solution: The Production Control Web Server prevents you from entering a future time but the DCS input does not. IQ requires that the time a Lab Sample value is entered in to have already occurred, so it performs the Bias update using data from that time. There are two main situations that can be observed. Situation 1 - SAMPFLAG and AVLFLAG set to 1 before future timestamp occurs If the SAMPFLAG and AVLFLAG are both set to 1 for a future timestamp in the DCS the Combined Status goes to No Sample. If the SAMPFLAG is set to 1 first, with AVLFLAG still at 0, for a future timestamp the Combined Status goes to Waiting. If AVLFLAG set to 1 before the future timestamp occurs the Combined Status goes to No Sample. When the time occurs a value of 0 appears and the combined status stays at No Sample, the Combined Status remains at No Sample when a new sample is taken. The Combined Status goes to Update Skipped when the new lab value comes in, if the new sample and available flag come in for the new sample at the same time you see it go from No Sample to Update Skipped. A similar behavior to Situation 1 can be reproduced by running IQ extract and opening the LBU history file. This prevents IQ from writing to the LBU history file because it is open in another program and the update will go to No Sample. To see this open command prompt as an administrator and run the iq_extract command as below for your IQ and then enter in a lab value you will see the Combined Status go to No Sample when the AVLFLAG is set to 1. Situation 2 - SAMPFLAG set to 1 before future timestamp occurs and AVLFLAG set to 1 after future timestamp occurs If the SAMPFLAG is set to 1 while the AVLFLAG remains at 0 for a future timestamp in the DCS, the Combined Status goes to Waiting. If after the future timestamp occurs the AVLFLAG is set to 1. The Combined Status goes to Good, and the update occurs successfully. You can enter in a future timestamp from the DCS, but you have to wait until that time occurs to enter in the AVLFLAG. To reproduce this behavior, set the SAMPFLAG to 1 when you know a lab value will be coming in, but do not set the AVLFLAG to read that value until that timestamp has occurred. Or have operations only enter in the value after the time has occurred. Keywords: Aspen Inferential Qualities, Lab Sample, DCS References: None
Problem Statement: When trying to compile a C++ procedure in Aspen Custom Modeler v11 with Visual Studio 2017 Professional Edition, the following error may occur: Building DLL... No suitable version of Visual Studio was found - compilation aborted. See the online help for details of the supported compilers. 1 error , 0 warnings generating procedure code. See above for more details. In the MakeUserCode file, you may also see the line USING_VC15 is commented out (# is the comment character in this type of language): #!include <win32.mak> # Uncomment the following for DEBUG build, comment out for RELEASE build #DEBUG_BUILD = 1 # Uncomment this line for INTEL Fortran, comment out for any other, as yet unsupported, Fortran INTEL_FORTRAN = 1 #Uncomment this line for VC15 (VS 2017) compatibility (no other version is currently supported) #USING_VC15 = 1 !ifndef USING_VC15 !endif
Solution: ACM V11 is looking for the environment variable: VS2017INSTALLDIR. The typical value should be: C:\Program Files (x86)\Microsoft Visual Studio\2017\Professional To set the environment variable, you should browse with Windows Explorer to the System Properties: Then click Advanced system settings: Then click Environment Variables... Finally, you click the New button to add the new environment variable (for the user or the system, as you prefer. If you're the only user of the computer, this makes no difference). You need to make this change only once. You must restart Aspen Custom Modeler for the change to be effective. Note that you must install the MFC component when installing Visual Studio. Re-launch the Visual Studio Installer to select the Desktop development with C++ as shown below. Keywords: Visual Studio, C++, procedure, makeusercode References: None
Problem Statement: When trying to compile a C++ procedure in Aspen Custom Modeler v11 with Visual Studio 2017 Professional Edition, the following error may occur: Building DLL... No suitable version of Visual Studio was found - compilation aborted. See the online help for details of the supported compilers. 1 error , 0 warnings generating procedure code. See above for more details. In the MakeUserCode file, you may also see the line USING_VC15 is commented out (# is the comment character in this type of language): #!include <win32.mak> # Uncomment the following for DEBUG build, comment out for RELEASE build #DEBUG_BUILD = 1 # Uncomment this line for INTEL Fortran, comment out for any other, as yet unsupported, Fortran INTEL_FORTRAN = 1 #Uncomment this line for VC15 (VS 2017) compatibility (no other version is currently supported) #USING_VC15 = 1 !ifndef USING_VC15 !endif
Solution: ACM V11 is looking for the environment variable: VS2017INSTALLDIR. The typical value should be: C:\Program Files (x86)\Microsoft Visual Studio\2017\Professional To set the environment variable, you should browse with Windows Explorer to the System Properties: Then click Advanced system settings: Then click Environment Variables... Finally, you click the New button to add the new environment variable (for the user or the system, as you prefer. If you're the only user of the computer, this makes no difference). You need to make this change only once. You must restart Aspen Custom Modeler for the change to be effective. Note that you must install the MFC component when installing Visual Studio. Re-launch the Visual Studio Installer to select the Desktop development with C++ as shown below. Keywords: Visual Studio, C++, procedure, makeusercode References: None
Problem Statement: How to generate an x-y plot?
Solution: You can generate an x-y plot using any Binary Analysis in the Properties Environment. Let's take water-methanol as an example to illustrate the steps. 1) Define your components in the Components | Specifications form. 2) Select the appropriate property method. An activity coefficient model such as NRTL is a good choice for water-methanol mixture. 3) On the top Home ribbon, select Binary in the Analysis section NOTE: These Ribbon menu options are only available in the Properties Environment (not Simulation). 4) A new object called BINRY-1 will be created in the Analysis folder on the left navigation pane.The Binary analysis defaults to Txy. Note that the property method is the same as the global property method, but you can choose a different one here, from your list of previously selected methods. 5) Click on Run Analysis and view the T-x-y plot that is generated. Note that you can also change the pressure in the form shown above. 6) On the BINRY-1| Results section, on the navigation page, you can view the data points through which the plot has been produced. Here, on the Home ribbon at the top, click on the Plot section and you will see the option of y-x plot. Keywords: None References: None
Problem Statement: How to test Aspen Cim-IO Store and Forward
Solution: Access to the Aspen InfoPlus21 server and the Aspen Cim-IO server is required for this test. Aspen InfoPlus21 server Locate a tag that is currently collecting data using Aspen InfoPlus21 Administrator, Aspen SQLplus, or other method. Stop TSK_A_device name If using unsolicited transfers stop TSK_U_device name Note the time when the tag above stops receiving data. Aspen Cim-IO server Open the Cim-IO Interface Manager Select the Interface and open “Monitor Store&Forward Queue” Select the S&F Queue to monitor Note the S&F Mode and other statistics about the Queue files. Aspen InfoPlus21 server Re Start the task that was stopped Review the tag to see the data fill in up to current time and normal data collection begin. Aspen Cim-IO server Check “Monitor Store&Forward Queue” again. If data forwards and the tag is updating with current data then Store & Forward is working. If data does not forward then contact AspenTech Support for assistance. Keywords: None References: None
Problem Statement: How to verify that history repositories have sufficient cache buckets?
Solution: InfoPlus.21 is designed to not lose history. If InfoPlus.21 sees that there's a repository that is overflowing into an event.dat file, IP21 will start slowing down database access to allow the history system to catch up. It is very important that you have enough cache buckets in your File Set for all the tags that are being written into the File Set. To optimize IP21 performance, it is worth checking the Cache Size. Hence, for each repository, go to the Active File Set and right-click it, and select Summary. A File Set Summary window will pop up as below. Then look at the number highlighted in the upper right hand corner. Then, go to IP21 Admin, select a repository, right-click on the repository to check how many cache buckets there are? In this case, I am using TSK_DHIS as an example. It is important that you need to verify the number of cache size for the repository is larger than the number of tags being historized for the repository. You can update the number of cache size and restart IP21 database accordingly. This will help greatly in improving IP21 performance. Keywords: Cache Buckets, IP21 Performance, Poor Performance References: None
Problem Statement: How do I change the default folder for tag replication queue files?
Solution: The folder location for the replication queues can be changed for ReplicationPublisher but not for ReplicationSubscriber. The new location can be specified using a command line argument -path=FolderName in the InfoPlus21 Manager under the external task TSK_PUBR. For example, to change the publisher's queue location to D:\TMP\Replication, add this command line parameter for TSK_PUBR: -path=D:\TMP\Replication Note: Solution 000076293 has a list of the parameters that may be used when starting TSK_PUBR. Keywords: Replication Queue ReplicationPublisher ReplicationSubscriber Folder 139363-2 References: None
Problem Statement: What are the command line parameters available to control PUBLISHER behavior within Aspen InfoPlus.21 replication? These parameters could be applied to TSK_PUBR in the InfoPlus.21 Manager.
Solution: Switch Description -? or –help Displays the usage syntax. (the help text for some advanced options is not shown) -p[ath]:<directory> Directory path where the memory queue files reside. The default is the “Replication” folder in the group 200 directory for InfoPlus.21. -r[etentionPeriod]:<number> Retention period in hours to keep queued tag updates before they are considered expired and then discarded. The default is 72 hours old based on the current time. -x Causes the replication publisher task to use a fixed buffer allocation of size maxBufferSize. This applies when there is concern about possible memory fragmentation issues. Prior to v7.3 cumulative patch 4, this was the default behavior. This behavior is now disabled, by default. For example, without this setting, the replication publisher task will constantly re-allocate new message buffers to exactly fit the size of the message to be published. Note - the number of available parameters has been reduced from the amount available in prior releases (for example, 5 were removed prior to the release of InfoPlus.21 V11). The four listed above are the ones available in V11 and V12 of the Manufacturing products at the time of this writing (March 2021). Keywords: Replication Configuration 139975-2 References: None
Problem Statement: When accessing web client applications such as aspenONE Process Explorer (A1PE), an error message is displayed: HTTP/1.1 500 Server Error http://localhost/processexplorer/aspenONE.html http://localhost/ProcessExplorer/WebControls/PBItemDetails.asp?admin=true This issue affects all client users of the AspenTech web applications and only occurred after March 9, 2021. In the case of A1PE, the home workspace page displays correctly, the error message is only seen when clicking the Process Explorer action link. Rebooting the server does not resolve the issue.
Solution: This issue has been observed after the rollout of Microsoft security update KB5000800 - Cumulative security update for Internet Explorer: March 9, 2021: https://support.microsoft.com/en-us/topic/kb5000800-cumulative-security-update-for-internet-explorer-march-9-2021-b7b43be0-e9ef-48b6-b102-ed28fd89e0f2 The patch sets new permissions on a .dll file. A1PE and other web client applications are then hit with the HTTP/1.1 500 Server Error. Consequently, new permissions are required to resolve the headline issue. This can be done in a command prompt window opened with escalated permissions (Run as Administrator): cacls %windir%\system32\jscript.dll /E /R everyone cacls %windir%\syswow64\jscript.dll /E /R everyone Successful access to A1PE and other web client applications will now be possible. Keywords: Internal server error no longer able to show trends References: None
Problem Statement: After upgrade of Aspen Production Execution Manager from V8.5 to V11.1 ,MOC was unable to read / write data to Batch.
Solution: This issue happens because of configuration mistake on path.m2r_cfg file. Check BATCH_21_DSN is set correctly On the APEM server open the path.M2r_cfg file with Notepad. This is typically located in: C:\Program files(x86)\AspenTech\AeBRS\cfg_source BATCH_21_DSN =APRM Whenever a change is made to path.m2r_cfg files, it is necessary to run codify_all.cmd (located in the same directory) which codifies the content of the text file (i.e. writes compiled *.cfg files to the Program Files\AeBRS directory.) Keywords: MOC path.m2r_cfg BATCH_21_DSN =APRM Batch References: None
Problem Statement: Where do I find the results for Shell & Tube Heat Exchangers in Series?
Solution: EDR – Exchanger design & rating program is well capable of design of the Shell & Tube heat exchnager in series / parallel. User can easily observe the results for the inlet & outlet temperatures, heat exchanger duty, pressure drops for both the exchangers in series: The results available on TEMA Sheet is the overall (final) results of the heat exchaner including both the exchangers in series. Only user can find that under TEMA sheet, whether connected in parallel or in series. To check the individual exchanger results when its in Series like heat duty, temperatures, pressures, etc. user can check the same under: Results à Thermal/Hydraulic Summary à Performance à “Shell by Shell Conditions” Details available under help in EDR: When there are multiple shells in series, the table of shell by shell conditions lists the heat load in each shell, and the temperature, pressure, and quality (vapor mass fraction) of both the shell side and tube side streams at inlet to and outlet from each shell. The table is also produced when there is only a single exchanger, but in this case the information merely duplicates what is given elsewhere on the overall exchanger performance. The table also lists the mean shell metal and mean tube metal temperatures in each shell, together with the maximum and minimum tube metal temperatures. The mean temperatures are distance weighted averages, appropriate for use in mechanical design calculations. The corresponding values for the overall set of exchangers are given on the MTD & Fluxes tab under the Heat Transfer results section. Keywords: Shell in series, shell by shell conditions References: None
Problem Statement: It could appear as though the event.dat file is not unbuffering even after following all of AspenTech's instructions for bringing up the history task after it went down. How can one determine if the event.dat is actually unbuffering?
Solution: If the Aspen InfoPlus.21 (IP.21) history task goes down or can't keep up with data flow from the plant during a weekend or overnight, the event.dat file can grow quite large. Depending on the size of the event.dat, significant time could be required for this file to unbuffer. Moreover, the time for the event.dat to unbuffer will be regulated by the efficiency by which data can be written to the hard disk. When IP.21 is up and running after the history task has been restarted, not only is data written to disk from the event.dat, but data is also coming from the plant into the event.dat until the event.dat file completely unbuffers. This can really slow the system down. To assure that disk writes are done as efficiently as possible make sure the drive stays in an unfragmented state. It would also be advisable to close any unnecessary applications that perform writes to the disk. To determine whether or not the event.dat is unbuffering go to into the Aspen InfoPlus.21 Administrator and expand the node that represents the database. Expand the Historian container and look for the appropriate history repository. We will use TSK_DHIS in this example. Right-click on TSK_DHIS and choose 'Check Queue... Here is an example of what 'Check Queue...' produces: Although TSK_DHIS is used as an example here, look at the appropriate repository on your server. Examine the START and END Disk Overflow fields. The Start and End fields do not display the file size of the event.dat file, but rather, beginning and ending integers are displayed that represent starting and ending numbers of the event.dat file. In essence, the difference between the Start and End numbers should be monitored because the relative differences of these values can be correlated to an increasing or decreasing event.dat size. Furthermore, if the difference between Start and End is found to decrease, then the event.dat file is unbuffering. If the difference between the Start and End numbers is found to increase, then the event.dat file is not unbuffering. The table below illustrates one such example. Start End Difference 6,963,810 27,121,498 20,157,688 16,664,535 29,677,268 13,012,733 17,840,440 30,021,332 12,180,892 For the above case, the event.dat file is unbuffering because the differences between the Start and End fields are decreasing. Keywords: Event.dat Archive Repository Store & Forward Converted from 103995_Default.txt References: None
Problem Statement: This Knowledge Base article provides steps to invoke the Aspen InfoPlus.21 Health Monitor. The application has both a thick client (software that is specifically installed) and a thin client (an application that is accessed via a website).
Solution: The Aspen InfoPlus.21 Health Monitor is a Microsoft Windows thick client used to check the status of the components that make up Aspen InfoPlus.21, including the database itself, the tasks, the history subsystem, Aspen Cim-IO database-specific components, and a few important categories of supporting records, including SQLplus query records and those responsible for saving database snapshots and backing up history files. It even has some Windows server computer-specific items: The Health Monitor can be launched from the InfoPlus.21 Administrator by following the steps provided below: 1. Invoke the InfoPlus.21 Administrator. 2. In the left pane, right-click the top-level InfoPlus.21 tree node. 3. In the context menu, select 'Launch Health Monitor' to invoke the InfoPlus.21 Health Monitor client application. NOTE: Users of Aspen InfoPlus.21 Process Browser (formerly known as Aspen Web.21) can launch Aspen InfoPlus.21 Health Monitor web client by typing in the following URL into their web browser: <http://<Web21_Server_Name>/ip21healthmonitor> Keywords: 123619-2 References: None
Problem Statement: In regard to Aspen InfoPlus.21 replication, for how long will data buffer on the InfoPlus.21 PUBLISHING system? Answer By default, the data will buffer on the InfoPlus.21 PUBLISHING system for up to 72 hours before older data begins to be discarded. The speed of the unbuffering is a function of the network speed. Note:
Solution: 000076293 - What are the command line parameters available to control publisher behavior within Aspen InfoPlus.21 replication? has some additional details about switches which may be added to ReplicationPublisherNG.exe to observe and adjust those settings... Keywords: None References: None
Problem Statement: Is there an Aspen SQLplus function that will tell you if a value is an integer or text?
Solution: This example code below demonstrates how to determine if the input typed at the prompt is an integer or text by using the SET CONVERT_ERRORS option and then checking if CAST(x as INT) is NULL. The example code below may be copied into the top half of the SQLplus Query Writer and then run. The code will produce a prompt... Example code: local x char(200); set convert_errors 0; x = prompt('Type something'); if cast(x as int) is null then write 'Not an integer'; else write 'It is an integer'; end Keywords: convert_errors number text Converted from 116277_Default.txt References: None
Problem Statement: Is there a limit or maximum on the number of entries per role in Aspen Framework Security or Local Security?
Solution: There is not a limit. Keywords: None References: None
Problem Statement: Below error message shown in screenshot is returned when AFW Security Manager is being opened.
Solution: Furthermore, when browsing to security URL, http://<Server_Name>/AspenTech/AFW/Security/pfwauthz.aspx, in Internet Explorer, the below error page is returned.Solution Below is a checklist to troubleshoot on the issue. By no means is it exhaustive as different configuration or account used to log on application pool will result in different requirement. Ensure that Microsoft Internet Information Services (IIS) is started and running. This can be checked by issuing the command iisreset /status in a command prompt launch using Run as administrator. Starting with version 8.8 of the AFW Security, the application pool that is associated with the AFW/Security virtual directory is now ASP.NET v4.0 Aspen Security Pool. The Aspen Security Pool will be running using the built-in account ApplicationPoolIdentity. Ensure that IIS_IUSRS has following permissions to C:\inetpub\wwwroot\AspenTech\Afw folder. - Read & execute - List folder contents - Read Note: By default, the above permissions are inherited from C:\inetpub\wwwroot which should have IIS_IUSRS with the above-mentioned permissions at wwwroot folder level. After DCOM logging was enabled per knowledge based article How to enable extra DCOM logging for troubleshooting, the following message was found in the Windows Event Viewer showing that the identity running the Aspen Security Pool could not authenticate the logged in user: Based on the failure to authenticate the user, in IIS Manager the Identity of the Aspen Security Pool was changed to the same Administrator account as that which runs the IP.21 Task Service. Keywords: Could not connect to AFW Security Server via HTTP with the URL : http://<servername>/AspenTech/AFW/Security/pfwauthz.aspx Server Error in '/AspenTech/afw' Application Retrieving the COM class factory for component with CLSID {225E90A4-34E7-11D5-B9D5-00B0D053AAD5} failed due to the following error: 80070005 Access is denied. (Exception from HRESULT: 0x80070005 (E_ACCESSDENIED)) References: None
Problem Statement: How to access Chinese eLearning in V12 Aspen Plus/HYSYS ?
Solution: As V12 version released, there is a new feature in Aspen Plus/HYSYS that user can directly access to Web Example and Training with internent connection, it will directly connect to Aspen eLearning website in English version, then how to access Chinese eLearning website? 1, Change Aspen language to Chinese by Select Aspen Language Utility tool. 2, Open HYSYS/Plus, File | Open option | Training tab, click an example in training tab and click Yes to unlock eLearning as below, it will directly connect to Chinese version eLearning website. After that, please login with your eLearning account. Keywords: Chinese eLearning, Web Example References: None
Problem Statement: What is the license behavior for V8.x of Aspen CIM-IO?
Solution: Aspen Cim-IO V8.0 to V10.1 has the following licensing behavior: The Cim-IO Manager service attempts to check out a license on startup. · The Cim-IO Manager continues to run if either: a) The license request is granted. b) The license request was previously granted. - or - · The Cim-IO Manager will generate a service log of the denial message and will stop running after determining that: a) The license request is not granted. b) The license request has never been granted to an application on the computer hosting the Cim-IO Manager. License requirement was removed in V11.0. NOTE: The license behavior may differ with other MES products. Click on the product below to review its license behavior. InfoPlus.21 Process Explorer aspenONE Process Explorer Aspen Production Record Manager Keywords: CIM-IO, License behavior References: None
Problem Statement: This Knowledge Base article explains best practice to migrate Aspen Production Record Manager (APRM) database from Oracle to MS SQL Server.
Solution: You should use APRM's Backup/Restore features to backup the data from Oracle then import it into SQL Server. To access APRM's Backup/Restore feature open the APRM Administrator, expand out until you see Areas, right click on Areas, select Database. The restore feature has a known limitation and may not work without error if your database is large. You must incrementally backup and restore subsets of batches rather than the entire database in one go. The exact size of a set of batches that can be restored in one go is entirely dependent on your system configuration and you must determine this in your own testing. Make sure you thoroughly test all features of APRM when in SQL Server to ensure that you don't encounter any surprise when attempting to use the database in production. Finally, only consider upgrading the APRM version (using the Aspen Database Wizard) after testing the restored SQL Server database with the same version of APRM as used with the original Oracle database. Keywords: SQLServer Upgrade References: None
Problem Statement: TSK_DBCLOCK terminates when starting Aspen InfoPlus.21. The last message in TSK_DBCLOCK.OUT is Performing history synchronization. How can I determine the tag and repository responsible for the failure?
Solution: The article https://esupport.aspentech.com/S_Article?id=000096225 explains the meaning of the message Performing history synchronization found in TSK_DBCLOCK.OUT. To determined the tag and repository causing TSK_DBCLOCK to fail, open REGEDIT and find the key: HKEY_LOCAL_MACHINE\SOFTWARE\Aspentech\InfoPlus.21\NN.0\group200 where NN is the internal version number of Aspen InfoPlus.21. You can find determine the internal Aspen InfoPlus.21 version number by looking at the key HKEY_LOCAL_MACHINE\SOFTWARE\Aspentech\InfoPlus.21\CurrentVersion. Next, add the DWORD (32-bit) Value DEBUG_LOAD_SNAPSHOT to HKEY_LOCAL_MACHINE\SOFTWARE\Aspentech\InfoPlus.21\NN.0\group200 and set DEBUG_LOAD_SNAPSHOT to 1. Try restarting Aspen InfoPlus.21. If TSK_DBCLOCK fails, examine TSK_DBCLOCK.OUT which will contain messages for each point and repository being synchronized. The last message will be similar to: H/Sync: record id=2080, fld#=38 H/Sync: Sync record id=2080, repository name=TSK_DHIS In this example, the record ID causing the failure is 2080, and the repository is TSK_DHIS; unfortunately, the damaged file set is not listed. Using the instructions in theSolution https://esupport.aspentech.com/S_Article?id=000078513, repair the most recent file sets of the repository causing the error, and then try restarting Aspen InfoPlus.21. If InfoPlus.21 starts, then use REGEDIT to set DEBUG_LOAD_SNAPSHOT to 0 to disable history synchronization logging. Key Words TSK_DBCLOCK history synchronization Keywords: None References: None
Problem Statement: Can aspenONE Process Explorer (A1PE) be used if the license is a perpetual / non-token / older commercial model type? Or must the license be of the token / new commercial model type?
Solution: A1PE requires the use of a token license model. Keywords: None References: None
Problem Statement: In the Calculations window of AspenCalc client tool I can validate all calculations of a server by right clicking on server name and choosing the Validate InfoPlus.21 option in the environment menu. Can I do the same (checking all calculations of a server [for lost tags]) via SQLPlus?
Solution: You will need to load the CalcScheduler object in SQLPlus as follows: Go to the View menu Select Object Browser Load the Object CalcScheduler When you use the Object Browser and open CalcScheduler there is CalcCommands class. Expanding CalcCommands and scrolling down there are 2 Methods for Validating. Click on them to see the parameters. The help section within SQLPlus will help with the coding and here is an abbreviated example: local CalcCmd; local CalcObj; local CalcInt I4; local CalcName STRING; CalcName = 'test'; --replace this with the calculation that needs to be validated CalcCmd = createobject('CalcScheduler.CalcCommands'); CalcObj = CalcCmd.GetCalculationObject(CalcName); CalcCmd.Validate(CalcObj, CalcInt); Keywords: Validate SQLPlus References: None
Problem Statement: What could cause the Boiling Point Assay in the Performance Tab to not match the results in the Boiling Point Curve Utility?
Solution: This is caused by the characterization of the oil in the simulation. The results match in both places only if Aspen Petroleum Refining Refsys is being used to characterize the crude. If you are using the Oil Manager to characterize your crude, the results in the Performance tab> Profiles>Boiling Point Curves or the specifications in the Column (Monitor> Add specification) are not going to match with the Boiling Point Curve Utility or the Petroleum Assay Utility. If you are using Refsys to characterize your crude the results in the Performance tab> Profiles>Boiling Point Curves or the specifications in the Column (Monitor> Add specification) are not going to match with the Boiling Point Curve Utility but they will match with the Petroleum Assay Utility. If you want to use Refsys but you have used the Oil manager you can convert your assays to Refsys assays by clicking the button of Convert to Refsys Assays in the Oil Manager tab. After pressing that button Hysys will: 1. Use the Export Oil Manager Information view to map HYSYS Oil assays to RefSYS Assays. Each assay in the Oil Manager is listed in the left hand column. Use the RefSYS assay column to name the converted assay, and check the box to the right if you want to install that assay in HYSYS Petroleum Refining. 2. Use the Component List options to use the existing component list for the selected Oil Manager assay, or to Use the RefSYS default component list. 3. Use the Blend Options fields to map blend descriptions from the Oil Manager to feeders used by HYSYS Petroleum Refining. 4. If there are user properties present in the Oil Manager assays, click User Property To Petroleum Property Map, and use the Map User Property To Petroleum Property view to match the Oil Manager user property names to those used in HYSYS Petroleum Refining. 5. When set up is complete, click Convert. A confirmation prompt appears for each assay converted to HYSYS Petroleum Refining format. The assays will appear listed under Petroleum Assays Keywords: Boiling point curve, ASTM D86 References: None
Problem Statement: The Audit & Compliance Manager (AACM) web site does not display or function correctly unless you are using MS Internet Explorer as the web browser. Even then, the pages may render badly, for example an extended top margin could appear on the Event Query page: Clicking the Submit Query button on the Event Query page would most likely not respond in such a case. Even if it did, the response might be an error dialog titled Application Blocked by Java Security: This article aims to resolve these issues.
Solution: The Audit & Compliance Manager web site should be accessed using MS Internet Explorer 11 (IE11) when it has compatibility mode enabled for a URL with the AspenTech web server specified. To enable compatibility mode in IE11, first press the ALT key to reveal the menu bar, and then select Tools, Compatibility View settings from the menu. On the Compatibility View Settings dialog, either add the website to the Websites you've added to Compatibility View list box by clicking on the Add button or tick the Display intranet sites in Compatibility View check box. Having made this change, close the Settings dialog and return to and refresh the AACM web site. If you see the Java Application Blocked dialog when you click the Submit Query button, and assuming you have installed the Oracle Java Runtime Environment (JRE), you must make sure the AACM web site is added to the exception site list. The exception site list is managed on the Security tab of the Java Control Panel. To launch the Java Control Panel, select Java (32-bit) in the Windows Control Panel. You can then select Security tab on the Java Control Panel where you can then edit the exception site list: If you are unable to launch the Java Control Panel, you could edit a text file that lists all such excepted sites. The file can be found here: C:\Windows\Sun\Java\Deployment\exception.sites and simply consists of a list of site URLs containing the protocol and the web server name, eg. http://cjplt101 Keywords: AuditComplianceManager/EventQuery.asp ERROR: Cannot locate proxy which supports Remote Scripting. Was RSEnableRemoteScripting method invoked? Please wait while searching for records References: None
Problem Statement: Where can I find Sustainability sample cases in Aspen Plus V12.1?
Solution: AspenTech is focused on helping our customers improve the sustainability of the process industries. Aspen Plus V12.1 includes seven new sample files to facilitate this objective. The new samples can be found by clicking Examples on the Resources ribbon tab, or by navigating to the examples folder C:\Program Files\AspenTech\Aspen Plus V12.1\GUI\Examples. The examples include: Four new carbon capture examples in the Carbon Capture\Industrial Scale subfolder. Each of the new carbon capture models include the absorber and solvent regeneration columns as well as the heat recovery and trim exchangers operating in a closed loop. Rate-based distillation methods and rigorous column hydraulics are applied for accuracy. The new MEA models use the ENRTL-RK property method (electrolyte NRTL with RKS equation of state) together with equilibrium chemistry and rate-based reaction kinetics in the columns. Several key model parameters are drawn from AspenTech’s proprietary ACIDGAS unary and binary databases. Two new biofuels examples. Focus on the circular economy continues to spur research related to biofuel and biochemical application. These new examples provide resources demonstrating modeling techniques to characterize biomass and thermal conversion of biomass to bio-oil. A new alkaline electrolysis model for production of green hydrogen. Governments, research agencies, and major corporations are spending billions of dollars developing and scaling up new technologies to decarbonize the economy. One key technology is the production of “green hydrogen”, produced by electrolysis of water using electricity from renewable sources such as wind and solar power. This set of example files includes a rigorous custom model of an alkaline electrolysis cell and an Aspen Plus process model covering the ancillary equipment supporting the electrolysis unit. Keywords: Sustainability, Carbon capture, biofuel References: None
Problem Statement: This article described what are the recommended Browsers for PCWS for Different Versions.
Solution: For PCWS and up to Version 12 we recommend basically 2 browsers for all supported versions: Internet Explorer Version 11 Google Chrome (Version depend on APC Version) For APC V12 we also recommend Microsoft Edge (Chromium) which was released last year. Here is a summary of the Browsers we suggest using for V10, V11 and V12. APC Versión Internet Explorer V11 Google Chrome Microsoft Edge (Chromium) Microsoft Edge Safari Firefox V12 Supported Supported Supported N/A Not Supported Not Supported V11 Supported Supported Not Supported Supported Not Supported Not Supported V10 Supported Supported Not Supported Not Supported Not Supported Not Supported Keywords: DMC3, OPC, Delta V References: None
Problem Statement: When creating a direct OPC connection on Config Online Server and try to test using DMC3 builder you may get multiple errors showing that the connection fails between the DMC server and the OPC server. This article mentions a couple of things that we have observed may work to establish the connection. Notice that some of these suggestions come from experience in the case of the Delta V system.
Solution: Once the connection is created, we suggest verifying the next things: 1.- Verify that the account that is used for the Connection on the DMC server is the same account as the DeltaV Administrator. We have observed that this is also a requirement that other products as CIMIO often requires. 2.- In the Specific case of DMC3 Builder verify that the RTE service uses the DetalV Administrator account to run. After changing the user proceed to reboot the server. 3.- Make sure to uncheck the tag name cache from the configure online server Edit IO source window. For some DCS servers, we saw that Tag cache sometimes consumes a large amount of memory and eventually impacted performance. So, you might want to keep an eye on that and disable it if you start seeing memory-related issues. Keywords: DMC3, OPC, Delta V References: None
Problem Statement: How to get Relative Humidity in Aspen HYSYS material stream?
Solution: A new Relative Humidity correlation is available under the Standard correlation type on the Correlation Manager. The relative humidity of a vapor is the partial pressure of water vapor in a given vapor divided by vapor pressure of water at the given temperature. Keywords: Relative Humidity, partial pressure, etc. References: None
Problem Statement: Which are Heat Transfer Fluid Components available in Aspen HYSYS?
Solution: In HYSYS V12, the parameters for the following heat transfer fluid components were updated based on experimental data from the current literature: THEOL-55 THEOL-59 THEOL-66 THEOL-75 THEOL-LT THEOLD12 THEOLVP1 As a result, the property curves for these components now closely match the latest publications. Keywords: Heat Transfer Fluid Components, Therminol, Theol, etc References: “Heat Transfer Fluids.” Therminol, Eastman Chemical Company, www.therminol.com/heat-transfer-fluids.
Problem Statement: This
Solution: frames a way to increase the memory usage for AspenWatch . Solution The way to solve this problem is by increasing the number of Filesets. To know how many filesets you may need, you will need to check one of your current Filesets on the repository and check how often they get full and how often the change takes place. That could give you a rough estimation of how much you may need for a year. Note: the fileset limit per repository is 10000 Please refer to the following KB articles for more information: How to create a back-up of history files Link: https://esupport.aspentech.com/S_Article?id=000050329 How do I configure history repositories for optimal performance? Link: https://esupport.aspentech.com/S_Article?id=000061772 How to add empty filesets to an Aspen InfoPlus.21 database in order to add old history Link: https://esupport.aspentech.com/S_Article?id=000077596 Cannot exceed the maximum allowed number of file-sets set in history Link: https://esupport.aspentech.com/S_Article?id=000096502 Keywords: AspenWatch, Inforplus21. References: None
Problem Statement: Is it possible to migrate customer records from an AspenWatch to another AspenWatch or IP21 Server??
Solution: The easiest way to do this is by using RLD Files, to pass records from one AW to another. to use the RLD you can follow the next steps: Extract/Export RDL: 1.- In Aspen InfoPlus.21 Administrator do right-click on the database and select Save Records. 2.- Specify the Record Name that you want to extract or select find and the Aspen InfoPlus.21 record finder will show. Then specify a name and location and save the rld. file Note: Keep in mind that customized definition records and their customized fields need to be saved as well. On the destination machine, save your snapshot before attempting to load your RLD file, in case there are errors or problems. Then do a test load to see if the RLD file loads successfully. Refer below KB for more information: https://esupport.aspentech.com/S_Article?id=000079353 Import RLD: 1.- Open the new AW Server InfoPlus.21 Administrator. 2.- Select the AW database. 3.- Right click on the database name that will be modified. 4.- Select Load Record to display the Load Records dialog box. 5.- Use the browse button to locate and select the desired .RLD file. and click OK to load the records. 6.- A Results List dialog box will appear indicating the the records have been loaded successfully. Refer below KB for more information: https://esupport.aspentech.com/S_Article?id=000077937 Keywords: AspenWatch, RDL, Inforplus21 References: None
Problem Statement: One of the post-configuration steps for GDOT web viewers is the creation of the GDOTOnlineHistory and GDOTSimulationHistory databases using the SQL Management Studio, in each one it is needed to create a table called GDOThistory for the web functionalities such as trending. In the original SQL query, for the first parameters called AppKey and VarName, the maximum number of characters were set to 50 and 20 respectively with the VARCHAR instruction. This can lead to several problems with GDOT web viewers if the application counts with longer variable names like the ones coming from an Aspen Unified GDOT Builder model. For addressing this issue, the normal procedure is to recreate the table changing the maximum number of characters from AppKey and VarName to 255 and restart GDOT Offline and/or Online services respectively. However, this procedure leads to the loss of information that has been collected for shorter variables if the GDOT application has been running for some time.
Solution: In this procedure it is not needed to recreate the whole table for troubleshooting trending or fetching issues on GDOT web viewers. It consists of the following steps: Open Microsoft SQL Management Studio, connect to the SQL server and go to Tools, click on Options. Expand Designers and select Table and Database Designers, uncheck the box that says Prevent saving changes that require table re-creation. Click on OK. Locate which database you want to work with, GDOTOnlineHistory or GDOTSimulationHistory and expand it. In this case we are going to change the parameters of the table from GDOTOnlineHistory. Expand Tables and right-click on GDOThistory table, then click on Design. On the Data Type column, modify the number inside the brackets for the varchar instruction and change it to 255 for AppKey and VarName. It is possible to change the length in column properties down below in the Length parameter. Click on Save. New parameters are now being taken in account and long-named variables will start being historized without loosing the data from other variables. Keywords: Aspen GDOT, GDOT web viewer References: None
Problem Statement: When trying to open Aspen Unified V12 an error message shows up preventing the user to access the web page and even if user clicks on Retry button, the message does not disappear.
Solution: Go to Internet Information Services (IIS) Manager. In the column of the left side calle Connections, identify the respective user and server and open Sites, then Default Web Site, AspenTech and locate AspenUnified. Once you click on AspenUnified, some icons will show on the display, in the icons that corresponds to IIS find Handler Mappings and click on it. Then, in the right column a menu of Actions will be displayed, click on Edit Feature Permissions… Proceed to check the three boxes, allowing permissions for: Read, Script and Execute. After performing these steps, Aspen Unified web page will be provided with the privileges it did not count before and the page will be able to open without the error message showed at the beginning. Keywords: Aspen Unified, Error 1920 References: None
Problem Statement: Does aspenONE Process Explorer (a1PE) support day light savings time?
Solution: Yes, a1PE does support day light savings time and displays the data according to the local timezone set on the client system. No specific configuration is required. Keywords: A1PE Day Light Saving References: None
Problem Statement: Can the environment variable “PATH” that gets created during IP21 installation be eliminated? In this KB, we are referring to below PATH environment variable. C:\Program Files (x86)\Common Files\AspenTech Shared\ C:\Program Files\Common Files\AspenTech Shared\
Solution: If those were eliminated, IP21 will not function properly and may throw error messages such as below error screenshot The environment variable “PATH” can affect the way IP21 processes will behave on a computer. They are part of the environment in which a process runs. For example, a running IP21 process can query the value of the PATH environment variable to discover a suitable location to store temporary files, or to find the directory structure owned by the user running the process. This is with respect to all software, and is not limited to AspenTech’s. For example, Windows itself uses environment variables for its products (such as SQL) to run on Windows OS. This simply means that when you delete the following PATH environment variable C:\Program Files (x86)\Common Files\AspenTech Shared\ C:\Program Files\Common Files\AspenTech Shared\. Aspen IP21 will not run. If you have other software/processes/services that depend upon the above PATH environment variable, those will not run too. This has been tested with IP21, and we found out that IP21 was not running and throwing the above error message. Keywords: None References: None
Problem Statement: This
Solution: frames some additional information on the PID Watch Model ID parameters. Solution The Open, Closed and Hybrid options mainly works for the type of Data we collected from the plant: Open Loop: the loop was on Manuel Mode and changes were made on the OP to see the response of the PV Closed Loop: Loop was on Auto y changes were made on the SP Hybrid: combines both some information of the vectors was collected when the loop was on Manual mode (open Loop) and other part were collected when loop was on Auto (Closed Loop) Maxim Delay: is an optimal value that be provided for the Maximum Dead Time Search. If set to 0, the sub space ID algorithm will implicitly calculate the value (in case Dead Time does exist). If you select the value to be different than 0, then the value is used as the maximum limit in an explicit dead-time search algorithm applied before model identification Keywords: PIDWatch, Model Identification, PID Loop References: None
Problem Statement: The problem reported consist on not been able to open and use the plots in What-If Simulation, returning the following error: <html><body><div id=body_msglog><font face=Verdana size=4>Error 2</font></div></body></html>
Solution: This issue is caused by the number of future moves and the magnitude of some of the dependent variables. This is exceeding a 600 character limitation in the query. The workaround is to truncate the prediction array to some smaller number of points to allow the vector to fit within the 600 character limit. The attached field test will do this. To install the field test follow the next instructions: 1. Download the attached file (WhatIfSimulationPlotFix_v11_v12.rld) to the aspen watch server. 2. Run InfoPlus.21 Manager and choose the View -> Utilities menu option. 3. On the Records tab, choose Load and Use first free record id. 4. Click Browse and select the rld file. 5. Click Execute. Keywords: PCWS, What-If Simulation, AspenWatch References: None
Problem Statement: This
Solution: frames how multiple Parameter information can be extracted from CCF file. Solution By default, there is not a feature that helps us to take out the Description and typical Move from dpp file, CCF or Model File. However, we can use a third-party application as excel to extract this kind of information. In the case of the model or dpp files the information can be find repeated on the CCF file For example The CCF file contains a parameter for description and a parameter for the Typical Move that also exist on the dpp or model file. To implanted this, you need to follow the next steps: 1,- make a copy of my CCF file and change the extension of the copy to txt. 2.- By opening the CCF file you will be noticed it contains some special characters shows as: ~~~. These characters define the spaces on the CCF 3.- Change the ~~~ to tabs and save the file 3.- Open Microsoft Excel and import the file as TXT delimitated file, In this way excel will put every value on a different column. 4.- Finally you can apply some functions or Macros to filter and extract the desire information from the Model or CCF file. Keywords: CCF, DMCplus, Excel References: None
Problem Statement: This
Solution: frames a couple of ways to resolve the error “Could not load type “System.ServiceModel.Activation.HttpModule” when Aspen Unified is Launched Solution In chapter 7 of the Aspen Unified Installation Guide V12 called Troubleshooting, this problem is reported and suggest some steps to fix this problem. The steps are the following: General ReSolution To resolve the issue of having multiple versions of .NET installed and running IIS, ensure the correct version of .NET is running with the installed version of IIS. Solution for Windows 7 1. 1.- Open a command-line prompt with administrator privileges. 2. Navigate to the correct location by entering the following: cd\Windows\Microsoft.NET\Framework64\v4.0.30319 3. Enter the following command: aspnet_regiis.exe -iru 4. After enter: iisreset Solution for Windows Server 2012 R2 Proceed to: http://go.microsoft.com/fwlink/?LinkID=216771 Determine And additional workaround for this problem is do some changes on the Applicationhost.config file as follow: Find the section: <add name=ServiceModel type=System.ServiceModel.Activation.HttpModule, System.ServiceModel, Version=3.0.0.0, Culture=neutral, PublicKeyToken=b77a5c561934e089 preCondition=managedHandler /> And redefine the section by the following one: <add name=ServiceModel type=System.ServiceModel.Activation.HttpModule, System.ServiceModel, Version=3.0.0.0, Culture=neutral, PublicKeyToken=b77a5c561934e089 preCondition=managedHandler,runtimeVersionv2.0 /> Note You can find the Applicationhost.config file in the following location: %windir%\system32\inetsrv\config This issue occurs because the Applicationhost.config file for Windows Process Activation Service (WAS) has the previous mentioned section defined, and that section is incompatible with the .NET Framework 4.0. Keywords: Aspen Unified, NET framework References: None
Problem Statement: The following message appears when trying to add new repositories or file sets using the Aspen Infoplus.21 Administrator: There is not enough space in the IP21HISTADMIN shared memory This article explains how to resolve this error.
Solution: Aspen InfoPlus.21 allocates memory to administer repositories and file sets. The amount of memory needed depends on the number of repositories and file sets in use. Use the following procedure to increase the amount of memory reserved for this task. 1. Stop Aspen Infoplus.21. 2. Make a backup of the registry. 3. Open a Windows Command Prompt as an Administrator and enter REGEDIT. 4. Navigate to the key below depending on the installation type of your Aspen InfoPlus.21 server: For 64 Bit: HKLM\SOFTWARE\AspenTech\InfoPlus.21\nn.n\group200 For 32 Bit: HKLM\SOFTWARE\Wow6432Node\AspenTech\InfoPlus.21\nn.n\group200 (Where nn.n is the version number.) 5. Change the DWORD Value IP21HISTADMIN_SIZE_IN_KB based on the following formula: The default value is 5120 KB decimal (1400 hexadecimal) or 5 MB. Total Size = ( 680 + (numReps * 1136) + (numFilesets * 600) ) / 1024 Where numReps = Total number repositories needed numFilesets = Total number of file sets needed 7. Start Aspen Infoplus.21. Keywords: DiskHistoryDef Archive Repository number is undefined KB 123782-2 References: None
Problem Statement: PCWS switch between different baselines created from AW Maker or PCWS. This Procedure Describe the process to create multiple baselines and how to switch between them
Solution: On AspenWatch Maker, you go Tools > Baseline Management: You can select the controller that you want to work with. By selecting cases you can create new cases for the specified controller. Once a new case is selected some information needs to be specified. As Case Name, Description, Start Time, and End time. Then click on Add and then Create. You can repeat the same Steps for multiple Cases. On PCWS, go to History and select the controller list to select the one for which you create the baselines. Then you select one of the KPI created for that controller and click on the Value Click on Active Case and this will display a new menu that allows changing between baselines Keywords: AspenWatch, PCWS, KPI References: None