question
stringlengths 19
6.88k
| answer
stringlengths 38
33.3k
|
---|---|
Problem Statement: This knowledge base article provides steps to resolve the following error:
Product prefix not found
which may be encountered when applying a Microsoft Installer (MSI) patch to an aspenONE product. | Solution: This error is usually returned because the original product was not installed properly. Therefore, first verify that the product works as expected. If the product does not work properly theSolution is to un-install the product(s) that are causing your trouble.
It is recommended to uninstall the software through the Windows Control Panel Add/Remove Programs.
Once the product is uninstalled, reboot your computer, then delete the following registry key (if the un-installation does not delete it for you)
HKEY_LOCAL_MACHINE | Software | AspenTech
Note: If you have multiple Aspen products installed, delete the registry key that belongs to the product that is causing trouble for you, for example, if Hysys is giving you this error, then delete only the HKEY_LOCAL_MACHINE\SOFTWARE\AspenTech\Aspen HYSYS key. Deleting the whole AspenTech key will render all other products unusable.
Then, re-install the product that is causing you trouble, reboot your computer, then apply the patch.
If you are unable to use the Add/Remove Programs function, you will have to manually delete the registry key procedure above, as well as manually delete the products from your local hard drive under C:\Program Files\AspenTech.
For the same example, if hysys was causing issues, just delete the C:\Program Files\AspenTech\Aspen HYSYS V1x.x folder.
Note: If the original products were installed through a custom installation script you should un-install the products then re-install them from the aspenONE dvd set to rule out a problem with the custom installation script.
Keep the following in mind when applying aspenONE patches. This error has also been caused for the following reasons:
- You must be logged in with an account which has administrative privileges on the local computer to apply the patch
- Virus scanners should be disabled before the patch is applied
If uninstall of products is not possible and other ideas need to be attempted first, please see if this workaround approach is applicable in your case: WorkaroundSolution if Aspen Update Agent fails to install Aspen Cim-IO V12.2 after successfully installing Aspen Cim-IO V12
Keywords: reapply, uninstall, reinstall, product prefix not found
References: None |
Problem Statement: Does the debugger work with Aspen Plus and the Intel compiler? | Solution: It is possible to use the debugger with all versions of Aspen Plus. To use the debugger, you must have the same compiler as we used to compile the source code. Intel Fortran 9 was used to compile Aspen Plus 2006, 2006.5, V7.1, V7.2, and V7.3; Compaq/HP Visual Fortran was used for 2004.1 and earlier; Intel Fortran XE 2011 (12) was used for V7.3.2, V8.0, V8.2, V8.4, and V8.6; Aspen Plus V8.8, V9.0 and V10 are based on the Intel Fortran compiler 2013 SP1 and Microsoft Visual Studio 2013 Update 3 as a 32-bit program. Aspen Plus V11, V12, and V12.1 were based on the Intel Fortran compiler 2017 and Microsoft Visual Studio 2017; Aspen Plus V14 was compiled with the Intel Fortran Compiler 2021 and Microsoft Visual Studio 2019 as a 64-bit program.
If you do not have the compatible version of the compiler, add Fortran Write statements to your user routine at strategic locations.
The compiler's symbolic debugger can be invoked for the following categories of Fortran routines:
· Inline Fortran
· User models
· In-house models
The debugger works a bit differently with the Intel compiler from how it worked with the Compaq/HP compiler.
1. From the Aspen Plus program group, open the Aspen Plus Simulation Engine window.
2. Compile user routine(s) using the dbg option. Type:
ASPCOMP filename dbg
This creates an object file (filename.obj) with debug information built into the obj file.
3. Prepare the debug linking of the shared library containing your user routines by increasing the system diagnostics level reporting. You can do this in one of the following two ways.
Add the following to your simulation input file that uses your user code. The simulation input file is the inp file which can be generated by export menu in Aspen Plus GUI:
DIAGNOSTICS
HISTORY SYS-LEVEL=8
DEBUG DYNLINK=2
Alternatively, you can change the following line in %aprsys%\xeq\asplink.bat:
%asptmp% %1 %2 %3 %4 %5 %6 %7 %8 %9
to
%asptmp% debug %1 %2 %3 %4 %5 %6 %7 %8 %9
4. Also add the line
SIM IAP=1
near the top of the input file. This will cause IAP (Interactive Aspen Plus) to start up when you run the aspen command.
5. Run the input file that requires the user routine(s). For example, if the input file is gmutest.inp and the RunID is test, type:
aspen gmutest test
The .dll will be generated as well as the program debug database file, the .pdb file.
6. When you get to the IAP prompt, start the Developer Visual Studio. If Visual studio was started earlier, it might not recognize the newly generated .dll created in step 5, so it could fail to honor breakpoints in the source code files.
7. Click Tools, then Attach, and then attach the debugger to the apmain.exe process. During the attach process command, one may need to select code types explicitly. The default of automatic was working in 2006.5 and earlier. However, in V7.0, the default does not work anymore; users need to explicitly swich to native code type.
8. Open your Fortran source file in the debugger environment, and to test the work flow, put a break point on one line of the executable source code.
9. Allow Aspen Plus to continue using the RUN, GO, or NEXT commands at the A+> prompt. If you had some break points set up in your code in the studio debugger, or your code breaks down, the debugger would be automatically given focus and you can then check variable values and step through your code at will. Type HELP for a more complete list of interactive run mode commands in Aspen Plus. More debugging capabilities please check Microsoft web site or MSDN library for further information. Please note that Aspen officially does not offer support for 3rd party tools such as intel Fortran or Microsoft Visual Studio.
Keywords: None
References: None |
Problem Statement: It is possible to face issues when creating a new Aspen OnLine Project for new users. | Solution: This video provides step by step instructions on how to create an Aspen OnLine Project and how to avoid common mistakes.
Main topics covered in this video:
Starting Aspen OnLine Service.
Launching Aspen OnLine.
Creating a new Aspen OnLine Project.
Keywords: AOL, Project setup, Digital Twin, RTO
References: None |
Problem Statement: CVE-2021-44228 discovered on December 9, 2021, is a critical remote code execution vulnerability present in certain versions of Log4j, a common Apache logging component. Log4j includes JNDI (Java Naming and Directory Interface) component, which while being able to help fetch resources, could also be leveraged by attackers to remote code execution attacks.
As a new vulnerability in a popular, web-facing component, potential attackers are actively looking for locations to exploit. Urgent action is necessary to ensure users of this software remain secure. | Solution: Corporate security and response
While Log4j is a popular component, it is only used by Aspen Collaborative Forecasting and aspenONE Process Explorer (A1PE) products. AspenTech is working on a fix to address the issue. The following steps will help mitigate the issue.
AspenTech Patch Updates:
In response to log4j vulnerabilities (CVE-2021-44228, CVE-2021-45046, and CVE-2021-45105) reported by Apache, AspenTech has included the recommended log4j patches in these AspenTech Products.
aspenONE Process Explorer (A1PE):
V12.2: AspenCoreSearch_V12.0 for AspenOne Process Explorer version V12.0 (Includes Log4j version 2.17.1)
V12.0: AspenCoreSearch_V12.0 for AspenOne Process Explorer version V12.0 (Includes Log4j version 2.17.1)
V11.0: AspenCoreSearch_V11.0 for AspenOne Process Explorer version V11.0 (Includes Log4j version 2.17)
V10.1: AspenCoreSearch_V10.1.4 for AspenOne Process Explorer version V10.1.4 (Includes Log4j version 2.17)
Aspen Collaborative Forecasting:
V12:
Aspen Collaborative Forecasting V12.0.2 CP2 (Includes Log4j version 2.17.1)
Aspen Collaborative Forecasting V12.0.2.1 EP1 (Includes Log4j version 2.17)
V11:
Aspen Collaborative Forecasting V11.0.6.2 EP2 (Includes Log4j version 2.17.1)
Aspen Production Execution Manager (APEM):
V12.2: Aspen_Production_Execution_Manager_V12.2.0.4 For Aspen Production Execution Manager V12.2 (Includes Log4j version 2.17.1)
V11.0: Aspen_Production_Execution_Manager_V11.0.1_00751599 For Aspen Production Execution Manager V11.0.1 (Includes Log4j version 2.17.1)
V10.0: Aspen_Production_Execution_Manager_V10.0.3_00751598 For Aspen Production Execution Manager V10.0.3 (Includes Log4j version 2.17.1)
Determine if aspenONE Process Explorer (A1PE) is installed on the machine:
aspenONE Process Explorer (A1PE) is used in many AspenTech software, please follow the below steps to confirm if you have aspenONE Process Explorer installed.
1. Click Start Menu and type Appwiz.cpl to launch Programs and Features
2. Check to see if aspenONE Process Explorer or Aspen Collaborative Forecasting product is listed
Legacy Log4j version details:
In light of the recent disclosure of the Log4j (CVE-2021-44228) vulnerability, AspenTech is committed to addressing all Log4j security concerns and providing all customers the information needed to remain secure. While Log4j (CVE-2021-44228) is referring to a vulnerability in Log4j 2.x versions there are still considerations for earlier versions. Specifically, Log4j (CVE-2021-4104) impacts Log4j 1.x versions. Here is what AspenTech customers need to know regarding the security of Log4j 1.x as it is used by AspenTech products:
Log4j 1.x is used by Aspen Production Execution Manager (APEM), Aspen Role Based Visualization (RBV), and Aspen Weigh and Dispense. Exploiting Log4j (CVE-2021-4104) requires Java Message Service (JMS) to be used via the Log4j JMS Appender. None of the AspenTech products above use JMS or the Log4j JMS Appender, and are not vulnerable to Log4j (CVE-2021-4104) as distributed.
For clients that are interested in learning more or who are interested in validating the security of services currently running, we recommend the following informational article Red Hat CVE-2021-4104. Per the article users can search any systems running Log4j 1.x for Log4j configuration files and verify the configuration files do enable the JMS Appender.
Mitigation Overview – The following mitigations can be applied by a user to immediately mitigate risks. These will not be necessary after either the temporary or permanent patches have been applied. Outgoing connections from the server hosting Log4j are necessary for attackers to identify vulnerable servers. Restricting these outgoing communications will help mitigate an attacker’s ability to identify the vulnerability. For publicly accessible instances a firewall can be configured to block potentially malicious traffic of the nature required to exploit CVE-2021-44228. Common WAF providers are already working on deploying effective rules.
Mitigation using Firewalls and Security Tools from Vendors:
Microsoft blog: Guidance for Preventing, Detecting, and Hunting for CVE-2021-44228 Log4j 2 Exploitation
Cisco Talos Intelligence Group - Comprehensive Threat Intelligence: Threat Advisory: Critical Apache Log4j vulnerability being exploited in the wild
Palo Alto Networks blog: Apache log4j Vulnerability CVE-2021-4428: Analysis and Mitigations
CrowdStrike blog: Log4j2 Vulnerability Analysis and Mitigation Recommendations
IBM Security Intelligence blog: How Log4j Vulnerability Could Impact You
Tenable blog: CVE-2021-44228: Proof-of-Concept for Critical Apache Log4j Remote Code Execution Vulnerability Available (Log4Shell)
Broadcom's Symantec Enterprise blog: Apache Log4j Zero-Day Being Exploited in the Wild content
Splunk's blog: Log4Shell - Detecting Log4j Vulnerability (CVE-2021-44228) Continued
VMware Blog: Log4j Vulnerability Security Advisory: What You Need to Know
Investigating CVE-2021-44228 Log4Shell Vulnerability: VMWare Threat Research
Mitigations using log4J System Settings for A1PE:
These instructions provide a sample with an older version of Tomcat. In V12 you can change Tomcat but it will not change Solr directly. Note:
Solr does not run under Tomcat in V12.
In A1PE v12, the user would have to update both Solr and Tomcat.
In A1PE v11, the user could just update Tomcat which would update Solr
The design of Aspen CoreSearch (used in A1PE) changed in V11 to V12.
Instructions to update Settings
1. Open Services control panel
2. Scroll to Apache Tomcat – right click and select Properties
3. Note the path to executable
By default, its installed in C:\Program Files\Common Files\AspenTech Shared\TomcatX.X.XX\bin\ path
4. Go to the Tomcat installation folder and double click to run the tomcatXw.exe executable file. Where X is a version number
5. Click the Java tab:
6. Append the below text to within Java Options
-Dlog4j2.formatMsgNoLookups=true
7. Click Apply, then OK.
8. Restart Apache Tomcat Service
9. Browse to http://<servername>:8080/solr using login as admin/admin
10. Check that -Dlog4j2.formatMsgNoLookups=true appears in the JVM Args section
Detection Overview:
As with any web service, it’s important to have a plan for detecting if a breach has already occurred. Here is some guidance for preventing and detecting attacks.
Cybersecurity and Infrastructure Security Agency: Apache Log4j Vulnerability Guidance | CISA (https://www.cisa.gov/uscert/apache-log4j-vulnerability-guidance)
This has multiple links to detection and response including information to support multiple security partners. (e.g., Microsoft, Palo Alto, Cisco, IBM, Tenable, Crowdstrike, Splunk, VMware)
Guidance for preventing, detecting, and hunting for CVE-2021-44228 Log4j 2 exploitation - Microsoft Security Blog (https://www.microsoft.com/security/blog/2021/12/11/guidance-for-preventing-detecting-and-hunting-for-cve-2021-44228-log4j-2-exploitation/)
Palo Alto systems (https://unit42.paloaltonetworks.com/apache-log4j-vulnerability-cve-2021-44228/)
For further updates on this topic subscribe to Security Notifications. Customers can subscribe to security alerts using a valid AspenTech support account. To subscribe, login to the AspenTech Support Portal and modify your email preferences under the Update My Profile menu item. Select the Security Notifications checkbox to opt-in.
This knowledge bulletin will be updated once patches are released.
Keywords: Log4J
Log4shell
Apache log4j Vulnerability CVE-2021-4428
Product Security
Collaborative Forecasts
Logging
JNDI Attack
Supply Chain Attack
References: s
Cybersecurity and Infrastructure Security Agency: Apache Log4j Vulnerability Guidance | CISA (https://www.cisa.gov/uscert/apache-log4j-vulnerability-guidance)
CVE-2021-44228: Apache Log4j2 Zero-Day Exploited in the Wild (Log4Shell) | Qualys Security Blog (https://blog.qualys.com/vulnerabilities-threat-research/2021/12/10/apache-log4j2-zero-day-exploited-in-the-wild-log4shell)
SANS Internet Storm Center, first round (https://isc.sans.edu/forums/diary/RCE+in+log4j+Log4Shell+or+how+things+can+get+bad+quickly/28120)
SANS Internet Storm Center, follow up (https://isc.sans.edu/forums/diary/Log4j+Log4Shell+Followup+What+we+see+and+how+to+defend+and+how+to+access+our+data/28122)
Log4Shell: RCE 0-day exploit found in log4j 2, a popular Java logging package | LunaSec (https://www.lunasec.io/docs/blog/log4j-zero-day/)
Microsoft’s Response to CVE-2021-44228 Apache Log4j 2 – Microsoft Security Response Center (https://msrc-blog.microsoft.com/2021/12/11/microsofts-response-to-cve-2021-44228-apache-log4j2/) |
Problem Statement: Is it possible to model C25-C30 alkane paraffins as both liquids and solids depending on the temperature. Do I need to define a user component? | Solution: Solids can use the same component name and formula, but have a type of Solid and a different Component ID.
If you use the SFE (Solid Fluid Equilibrium) Wizard to set up the solid components for the alkanes.
Select all of the components that possibly exist as solids.
The solid components and a Chemistry object will be created which will handle the solid-liquid equilibrium for the components.
The Chemistry can be used with most property methods.
In the attached file, a Sensitivity Analysis was created to report the flows of the components at different temperatures. The file is attached. Property parameters can be adjusted if needed to fit experimental data.
Keywords: None
References: None |
Problem Statement: When I go to download the Summary | Solution: Report in Excel format, the downloaded zip file is empty. What should I do?
Solution
ThisSolution applies to models that have been migrated from PIMS-AO.
The settings that were used in PIMS-AO are also transferred to AUP when migrating the model. In this situation, the summarySolution report was renamed in PIMS-AO to something other than the default. When that happens, you can apply the workaround as follows:
Open the model in PIMS-AO
Open the Model Execution Dialog Box
In the “Tables” space, rename the SummarySolution to “Solution”.
After running any case, save the model and remigrate this model to AUP.
Keywords: None
References: None |
Problem Statement: What are the default TCP ports for Process Pulse connections? | Solution: Main Service = 8001
Calculation Engine = 8004
Data Collector = 8005
Rest Service = 9134
If the ports were changed from their defaults, you can check the settings by going to the directory C:\ProgramData\AspenTech\CAMO\EPP, then opening the following XML files and locating the Port fields
mssetting.xml (Main Service)
efsetting.xml (Calculation Engine)
dcfsetting.xml (Data Collector)
restsetting.xml (Rest Service)
If SQL server is on separate machine, the default port between Main Service and the SQL server is 1433
If Active Directory is enabled, the default LDAP port is 389
If a license server is used, the port between Process Pulse and the license server is 5093 UDP/TCP inbound and outbound
Keywords: None
References: None |
Problem Statement: When importing data from Excel (.xlsx) into ProMV, you may come across the error message microsecond must be in 0..999999. | Solution: This is caused by rounding errors which can occur when cells have several multiply/divide operations, such as a calculated date column which uses DATE+1/24/60/6 for 10 second intervals. Each division introduces a rounding error in Excel, which leads to inconsistent date values and causes issues during import.
Wordaround:
This can be resolved one of several ways:
Use =MROUND(formula, increment)
Save Excel file as CSV and re-import
This will be resolved in an upcoming patch
Keywords: None
References: None |
Problem Statement: How to solve issues where the link to the alert found in emails does not point to the correct server specified in system manager general
In some instances, after updating the Server Name in system manager the email link can still reference the old server causing the users to be directed to a web page that does not load. This can be caused by the General settings name going out of sync with the Reporting setting. | Solution: Open Mtell System Manager
Navigate to the Configuration tab and select the general option on the side menu
Confirm the name in the Server Name is correctly displayed
Navigate to Reporting on the side menu and make sure the server name for email links matches the Server Name In the general settings and Save
Restart Internet Information Services (IIS) by opening a command prompt as admin and typing in IISRESET
Emails will now be displayed correctly
Keywords: Web page not loading
Link not updating
Wrong server
References: None |
Problem Statement: How to solve issues where Process Pulse Web terminal does not load correctly showing Invalid Language code
Sometimes after installing Process Pulse web terminal the dashboard page is not reachable and navigating to its web page shows an error related to an invalid language code | Solution: Open Internet Information Services (IIS)
Navigate down the hierarchy and locate Machine name\Sites\Default Web Site\AspenTech\AspenProcessPulse
Where machine name is the name of your server
In the center panel select Authentication
Select and enable Anonymous Authentication
Keywords: Web page
Process pulse online
Yellow page
Server error
Can’t access
References: None |
Problem Statement: How to Decode and Encode the ProMV SQL connection string?
When configuring Aspen ProMV online, if the SQL server and the ProMV server are not two separate machines, when standard workflow does not work, as found in this article, it can be required to manually configure the connection string between the two servers. This is found in the C:\inetpub\wwwroot\AspenTech\AspenProMVUtils\ServiceUtility directory in the appsettings.json file. The connection string is encprypted and in this guide we shall highlight how to decode and encode this string | Solution: 1. Open Google Chrome
2. Open Dev Tools using F12 or Ctrl + Shift + I or Settings > More Tools > Development Tools
3. Navigate to the Console tab in Dev Tools
4. In the input line, you can enter javascript commands to run. Utilize the command “atob()” to convert the encoded string to plain text. “btoa()” converts the plain text to the encoded string. Quotes around both strings are required
5. You will see in the dialog box your command followed by a response that will hold the decrypted or encrypted string in plain texts
Keywords: Different servers
Reconfigure connection
Service utility
References: None |
Problem Statement: This article describes the recommended approach for backing up Aspen ProMV Online information. | Solution: Aspen ProMV Online Continuous and Batch (ProMV) uses many configuration files to store offline and online model information and uses Aspen InfoPlus.21 (IP.21) to store real-time information for deployed ProMV models. It is recommended to regularly back up the following files related to ProMV:
1. ProMV Offline Model Files - These files with .pmvx extension represent the original ProMV model and are created with the ProMV desktop application. Any models created should be saved in a user-specified drive for future access. It is recommended to rename any models that are exported for live deployment, so they are easier to identify for future access.
2. Microsoft SQL Server database backups - ProMV creates 13 Microsoft SQL databases upon installation that are used to store configuration information:
Best practices for backing up a Microsoft SQL database can be found on the Microsoft knowledge base: https://docs.microsoft.com/en-us/sql/relational-databases/backup-restore/back-up-and-restore-of-sql-server-databases?view=sql-server-ver15
3. ProMV Online Model Files - These files are used to store online configuration information for models that have been deployed live. The default file path is C:\ProgramData\AspenTech\Aspen ProMV Online\Continuous for continuous models and C:\ProgramData\AspenTech\Aspen ProMV Online\Batch for batch models. We recommended taking a backup by regularly copying the entire folders to a separate drive.
4. IP.21 snapshot - This contains the structure of the real-time IP.21 database used to store and process ProMV records. Please review AspenTech KB 50326 for best practices on backing up the IP.21 snapshot.
5. IP.21 history files - These contain the historical runtime records of all Aspen ProMV online models since the time of deployment. Please review AspenTech KB 96935 for best practices on backing up the IP.21 history files.
Keywords: ProMV Online backup
References: None |
Problem Statement: When you open Aspen Mtell System Manager, you get an error message that looks like this.
Connection to MIS Web Service [http://servername/AspenTech/AspenMtell/InteropServer/MIMOSA/OSAEAIManagement.asmx] failed.
[Error: The remote server returned an error: (500) Internal Server Error.]
To confirm the error you are experiencing is related to SQL:
Go to the following URL in a browser. Replace servername with the name of your server. http://servername/AspenTech/AspenMtell/InteropServer/MIMOSA
For a SQL Authentication error, you will see a page like the below picture.
Login failed for user ‘domain\servername$’.
This image shows WORKGROUP\APM$, but WORKGROUP will be replaced with your domain and APM with your server. | Solution: This error occurs when SQL is configured to only allow Windows Authentication and not SQL Server Authentication. Mtell requires SQL Server Authentication to be enabled.
Open SQL Server Management Studio and connect to the Mtell database server with an account that has sysadmin permissions
In the Object Explorer, right click on the server name and select Properties
Switch to the Security page
Make sure SQL Server and Windows Authentication mode is selected
Click OK
Click OK to acknowledge you need to restart the SQL Server
On the SQL Server, open SQL Server Configuration Manager
In the left pane, select SQL Server Services
Right-click the instance of SQL Server that hosts the Mtell database and select Restart
Open Aspen Mtell System Manager
Switch to the Configuration tab
Select Database Connection
Click the icon of a stack with a check mark to test the connection to the database
You should see a successful message. Click OK to acknowledge.
Click OK to acknowledge you need to restart Mtell
Close and reopen System Manager
You should no longer receive the error message when opening System Manager
Keywords: MIS Web Service Error
Internal Server Error
IIS Error
500
Mtell Database Connection
Integrated Security
Windows Authentication
SQL Server Authentication
References: None |
Problem Statement: When you open ProMV Online, you see an error about the ProMV Optimizer Service in System Health Alerts.
No connection could be made because the target machine actively refused it. (localhost:8098) | Solution: This error appears when the ProMV Optimizer Service isn’t running. To start the service, you can use the ProMV Service Launcher.
Open a Command Prompt as an Administrator
Run the following commands
cd “C:\Program Files\AspenTech\Aspen ProMV Launcher”
AspenTech.ProMV.ServiceLauncher.exe
If you receive a message that “Service AspenTech.ProMV.Cont.OptimizerService is started,” then return to ProMV Online and confirm the System Health Alert about the ProMV Optimizer Service has disappeared.
It’s also possible the ProMV Optimizer Service was not started because of an underlying issue with ASP.NET. If you receive an error message when running command 2b, follow the steps below.
The error should reference a missing framework version: The framework ‘missing framework’, version ‘missing version’ was not found. It will also specify a link where the missing framework can be downloaded. Follow the link to download and install the missing framework.
Return to the command prompt and run the AspenTech.ProMV.ServiceLauncher.exe command again
If you continue to get error messages about missing frameworks, repeat steps 3 and 4 until you receive the message that “Service AspenTech.ProMV.Cont.OptimizerService is started”
Return to ProMV Online and confirm the System Health Alert about the ProMV Optimizer Service has disappeared
Keywords: ProMV Optimizer Service
Online Optimizer error
Framework not found
References: None |
Problem Statement: This KB article describes the process for moving your ProMV Online installation to a new server. By following these steps, you will be able to keep the information related to all your agents. | Solution: Start by installing the same version of ProMV Online on the new server. If you want to upgrade to a newer version, you can do so normally once you have confirmed that all the data has been properly transferred to the new server.
Move your IP.21 database following the steps in this KB article. KB 000062839
Keep IP.21 turned off on both servers until a later step
As of V14.0.2, ProMV creates 13 Microsoft SQL databases upon installation that are used to store configuration information. The number of databases may differ in a different version. A backup of these databases should be taken and then restored to the new server.
Best practices for backing up a Microsoft SQL database can be found on the Microsoft knowledge base: https://docs.microsoft.com/en-us/sql/relational-databases/backup-restore/back-up-and-restore-of-sql-server-databases?view=sql-server-ver15
When restoring the databases, you may need to use the options Overwrite the existing database and Close existing connections to destination database and uncheck Take tail-log backup before restore.
You should also copy the folders that store online configuration information for agents that have been deployed live. The default file path is C:\ProgramData\AspenTech\Aspen ProMV Online\Continuous for continuous agents and C:\ProgramData\AspenTech\Aspen ProMV Online\Batch for batch agents. Copy these folders from the old server and paste them into the same location on the new server.
Migrate your security settings
If you are using ProMV in Unsecured mode, open Local Users and Groups by searching for lusrmgr.msc
Double click on the ProMV Online Administrators group
Record what accounts are members of that group and add them in the new server
If you are using ProMV in Secured mode, migrate your AFW Security settings by following this KB. KB 000094313
Edit the Web.config file located in C:\inetpub\wwwroot\AspenTech\AspenProMVContinuous\UserSvc and C:\inetpub\wwwroot\AspenTech\AspenProMVBatch\UserSvc. Change the “UseAFWSecurity” key from “false” to “true”.
Start IP.21 on the new server
Go to aspenONE Process Explorer Admin and scan your tags by clicking Start Scan
Open ProMV Online and confirm that the agents exist
If you find that the agents are not there or they aren’t updating after you start them, try running the ProMV Service Utility to connect ProMV to the restored information you copied to the new server. This KB article shows how to use the ProMV Service Utility. KB 000100594
Next, we need to point all of the agents to the new Data Source. For each agent, click the three dots and choose Edit & Deploy Agent.
Go to the Input Mapping page, and click the outwards arrow button to export the tag mappings
Give the file a name and click OK
Open the CSV file
Use CTRL+H to find and replace the old Data Source (the old server name) with the new Data Source (the new server name)
Save the file
Click the inwards arrow button to import the tag mappings
Select the CSV file you edited and click Open
Click Apply
Repeat for any mapped tags on the Sensor Validation Mapping and Output Lab Results Mapping pages
Deploy the agent
Confirm the agents are receiving and processing data
Keywords: ProMV Online Server Migration
Move ProMV Server
Restore ProMV Backup
References: None |
Problem Statement: When installing Aspen Process Pulse, you may come across the following error:
Windows could not start the Aspen Process Pulse - Main Service service on Local Computer. Error 1064: An exception occurred in the service when handling the control request
Additionally, in Event Viewer, you may encounter the error: Login failed for user '<user name>'. Reason: Failed to open the database '<dbname>' specified in the login properties [CLIENT: <ip address>]. | Solution: This error can occur for many reasons, so it is recommended to trigger this error and then immediately check the Event Viewer for information about the error. Below are several known causes andSolutions for this specific error.
During the installation, the SQL login user must have sysadmin access. If Windows Authentication is necessary for the SQL server, promote the Windows user NT AUTHORITY\SYSTEM as sysadmin on MS SQL server if hosted on the same computer as Main Service. However, it is generally recommended to use SQL Server Authentication to avoid permission-related issues while starting the Main Service.
Keywords: None
References: None |
Problem Statement: When trying to run Python scripts in Unscrambler, you may see the error message indicating the PyCamo library is not installed. | Solution: The PyCamo.tar.gz file is located by default under C:\Program Files (x86)\AspenTech\Aspen Process Pulse\Calculation Engine
Open Command Prompt as administrator or alternatively, open a terminal through a code editor such as VS Code which is connected to the Python environment
Change the active directory to the directory containing the PyCamo file: cd C:\Program Files (x86)\AspenTech\Aspen Process Pulse\Calculation Engine
Install PyCamo by executing pip install PyCamo.tar.gz or py -m pip install PyCamo.tar.gz if Python is not part of the PATH
Wait for the command to finish, and PyCamo is now installed
Keywords: None
References: None |
Problem Statement: How to create a Bad Tag Status based on Current Value for Aspen ProMV Online using Aspen Calc
Sometimes it can be useful to monitor the status of a tag in order to take appropriate action during alert management and model improvement | Solution: Open Aspen Calc
Go to Calculation Tab
Right-Click on Server and add New Folder
This is optional and I have a Bad Tag folder and a Bias folder in this example
Right-Click on Folder (or Server if no folder is added) and add New Calculation
Create the formula needed
In this example the status of the tag will be switched to bad if it exceeds a certain value
For options on status tags to use, use the Help and search under Quality Status
In this particular example, the Return Value has the Historian Tag and the value is the basis for considering it a bad tag (value of 1)
Use the following to define a historical tag
Change Bind Type to InfoPlus.21, Use the drop down on the Server Name to locate the Historian server and input the Historian tag name in the Record Name
Test the Calculation
In the example above, the Historian tag is above 0.5 and the status for the tag is now Bad (normally is good)
Schedule the Calculation
Create schedule that writes faster than current Historian – for this example historian write time is 20 sec, calculation write time is 1 sec (Pull check mark from calculation to the schedule you want to use)
Set up rules to eliminate needless writing of data
Change has to be more than 0.001, Max time between results will be 20 sec (to match historian scan time)
Adding a Tag Status Calculation can also be added to a current AspenCalc Calculation. The example below shows the tag status calc as part of a larger calc that incorporates a bias shift in a historian tag.
When the rule violates the criteria, it will turn to “NULL” and ProMV will ignore the tag value.
Keywords: Bad values
Tag removal
Model validation
References: None |
Problem Statement: After installing Aspen Process Pulse, the following error message appears in Aspen ProMV Online indicating an issue with the Solr service.
When reviewing Apache Tomcat service, the service appears to have stopped. After manually starting the Apache Tomcat service, it will return to a stopped state after refreshing the services console.
After checking catalina logs at C:\Program Files\Common Files\AspenTech Shared\Tomcat9.0.27\logs, we found a severe error entry as shown. | Solution: As the catalina error log above suggested, this indicates that port 8005 is being used by other applications. Tomcat listens on TCP port 8005 to accept shutdown requests by default. As you may have guessed, Aspen Process Pulse Data Collector also uses port 8005 by default.
To resolve this conflict, we will need to change either Tomcat shutdown port or Process Pulse Data Collector port.
(a) To change Tomcat shutdown port:
What could cause Error 1053 when restarting Apache Tomcat service
(b) To change Aspen Process Pulse Data Collector port:
1. Locate the following path and run UPPDataCollector.exe as administrator:
C:\Program Files (x86)\AspenTech\Aspen Process Pulse\Data Collector
2. Click Stop to stop the Data Collector service.
3. Enter a new port which is not used by other services. In this example, port 8006 is selected. You can check if this port is currently being used by running netstat -aon | findstr :8006 on command prompt. Select a different port if it is being used. After the new port number is entered, click Start to start the Data Collector service.
4. Go back to services.msc to start Apache Tomcat service.
5. In Aspen ProMV Online, the Solr error message should no longer appear.
Keywords: A1PE
Catalina
Unscrambler
listening port 8005
References: None |
Problem Statement: During normal ProMV operations, you may eventually come across the following error:
Could not allocate space for object 'dbo.tbl_AlertSsnObss'.'PK_tbl_AlertSsnObss' in database 'MVContAlertMgrDB' because the 'PRIMARY' filegroup is full. Create disk space by deleting unneeded files, dropping objects in the filegroup, adding additional files to the filegroup, or setting autogrowth on for existing files in the filegroup.
Even after increasing the storage space of the database server, the message remains and ProMV is unable to function correctly.
Database servers running MS SQL Express Edition have their database sizes limited to 10 GB. You can check your edition by:
Right-click the SQL Server (ex. APM) > Properties > General > Product, then verify if it is Microsoft SQL Server Express, Standard, or Developer
Then to check the current database size:
Right-click MVContAlertMgrDB (ex. MtellSuite131 is used below) > Reports > Standard Reports > Disk Usage, then check the Total Space Reserved by the database
If the results of these steps show that the database is running MS SQL Server Express Edition and the database size noted in the error message is stuck at 10 GB, this indicates that a standard or developer license will need to be purchased and deployed on the database server. This is a software limitation, not a hardwired limitation, so increasing the storage space will not resolve the issue. | Solution: For full SQL licenses, an increase in storage space is necessary. There are two possibleSolutions to resolving this error if Express is being used:
1. Purchase a full MS SQL license and deploy it on the SQL Express server
2. Customize storage time for alerts if old alerts are not needed
By default, ProMV Online keeps alert data for 2 years (730 days). You can change this value to be shorter or longer in the appsettings.json file, by editing the DataMaxStoreDays value. If your database has a small capacity (SQL Express has 10GB limit) and you experience a high volume of alerts, you may want to make the storage time significantly less (e.g. 60 days). If your database has a large capacity and you want to store alerts for a longer period of time, you can increase the storage time (e.g. 3 years or 1095 days). Note that after updating the appsettings.json file, you must restart IIS for the changes to take effect.
For Continuous processes, the appsettings.json file can be found under:
C:\inetpub\wwwroot\AspenTech\AspenProMVContinuous\AlertMgr
For Batch processes, the appsettings.json file can be found under:
C:\inetpub\wwwroot\AspenTech\AspenProMVBatch\AlertMgr
Keywords: ProMV
SQL Express
References: None |
Problem Statement: During the Process Pulse installation of services, you may come across: Error 1920 Service Aspen Process Pulse - Data Collector UPPDCFService failed to start. Verify that you have sufficient privileges to start system services. | Solution: Since admin privileges are required to initialize the installation in the first place, this error message may not always be accurate. Observed on an Azure environment, the default port 8005 for the Data Collector service was found to be occupied by a separate, already running service. In some environments, port 8005 may be used by Microsoft's Configuration Manager, which blocks the installation wizard from starting its service on the default port. More information can be found here - https://learn.microsoft.com/en-us/mem/configmgr/core/plan-design/hierarchy/ports
There are two potentialSolutions,
Change the default Data Collector port from 8005 to another unused port
Change the port of the running service or disable it entirely if the service is unnecessary
Then rerun the installation wizard and the services should be installed correctly.
Keywords: None
References: None |
Problem Statement: Mailbox error: The server response was 5.7.1 Unable to relay for <[email protected]> | Solution: On the machine which is hosting the SMTP server:
Go to Administrative Tools -> IIS 6.0 Manager
Right click SMTP Virtual Server then Click Properties
Select Access tab
Click Relay... in Relay Restrictions
Select Only the list below
Add 127.0.0.1 and the Mtell machine IP to the list
Check Allow all computers which successfully authenticate to relay, regardless of the list above
Click OK
Keywords: None
References: None |
Problem Statement: You may encounter the error ACR login failed when trying to download the Maestro Azure containers. This generally means that the key credentials are expired. | Solution: Download the updated keys through the eSupport Download Center
Go to C:\ProgramData\AspenTech\Aspen Mtell Maestro (by default) or your Maestro directory, rename the previous acrcreds.key file or move it elsewhere as backup, then copy the new acrcreds.key file to that directory.
Keywords: None
References: None |
Problem Statement: How to add If clause in Calculator block in Aspen Plus using Fortran? | Solution: Calculator block in Aspen Plus is used to add user defined calculation and formula in Aspen Plus. Calculator block uses excel and Fortran to perform its calculation.
You can add simple and complex formula such as addition, subtraction etc in calculator block using Fortran.
If we want to add “If Clause” statement in Aspen Plus, please follow following steps.
Let us consider the example, let’s say if we have two streams connected to the mixer and we want to write the condition that if the temperature of stream 1 is higher than 25Degree then the mole flow of stream 2 should be 0.
1. You can add IF statement in Fortran in calculator block in Aspen Plus. We just need to know how to write the Fortran code IF statement. You can find the details of it on internet and also in help option available in Aspen Plus.
2. Consider the simple flowsheet with stream S1 and S2 with different flow rate and connected to the mixer block. If temperature of S1 is more than 25 than mass flowrate of stream S2 will be 0.
3. Go to the calculator block and define the two different variable which you want to use in the calculator block: Mass Flowrate of stream S2 and temperature of stream S1 in the define block.
4. Go to the calculate tab to write the Fortran code for IF clause and condition as show in image below:
5. After writing this formula in calculator block if you change the temperature of S1 to more than 25 Degree then your flowrate of stream S2 will be automatically calculated and changed to 0.
Keywords: Calculator block, IF statement, Aspen Plus
References: s
https://esupport.aspentech.com/S_Article?id=000054389
https://esupport.aspentech.com/S_Article?id=000097523
https://esupport.aspentech.com/S_Article?id=000098117 |
Problem Statement: This video covers the relationship between different IP.21-related products and provides a high-level overview of their functionality. | Solution: Due to the myriad of products that support the IP.21 database, this video aims to clarify how common IP.21-related products are connected. General details on how data is provided to IP.21 and how IP.21 data is used by its layered products will be included.
Keywords: None
References: None |
Problem Statement: What kind of plots are available for heat integration in Aspen Energy Analyzer (AEA)? | Solution: Aspen Energy Analyzer has a wide variety of plots.
- Grand Composite Curve
The Grand Composite Curve is a plot of shifted temperatures versus the cascaded heat between each temperature interval (Linnhoff et. al. 1982). It shows the heat available in various temperature intervals and the net heat flow in the process (which is zero at the pinch).
- Composite Curve
A Composite Curve is a graphical combination (or composite) of all hot or cold process streams in a heat exchange network (Linnhoff et. al. 1982). The Composite Curve plot displays both the hot composite curve and cold composite curve on the same plot. The closest temperature difference between the hot and cold composite curves is know as the minimum approach temperature, ΔTmin. The composite curves are moved horizontally such that the minimum approach temperature on the plot equals the minimum approach temperature you specified.
- Shifted Composite Curve
The Shifted Composite Curve is constructed the same way as the Composite Curve (Linnhoff et. al. 1982). However, the Hot Composite Curve (HCC) is shifted down by ΔTmin/2 and the Cold Composite Curve (CCC) is shifted up by ΔTmin/2. The result is that the hot and cold composite curves meet at the pinch location.
- Balanced Composite Curve
The Balanced Composite Curve is constructed the same way as the Composite Curve (Linnhoff et. al. 1982). However, the Balanced Composite Curve is a combination of both the utility and process streams. The HCC is a composite of both the hot process and utility streams. The CCC is a composite of both the cold process and utility streams.
- Shifted Balanced Composite Curve
The Shifted Balanced Composite Curve is constructed the same way as the Shifted Composite Curve (Linnhoff et. al. 1982). However, the Shifted Balanced Composite Curve is a combination of both the utility and process streams. The Shifted HCC is a composite of both the hot process and utility streams. The Shifted CCC is a composite of both the cold process and utility streams The HCC and CCC touch each other at the pinch points on this curve.
- Utility Composite Curve
The Utility Composite Curve is constructed much like a GCC (Linnhoff et. al. 1982). However, in addition to the GCC, the Utility Composite Curve contains hot and cold utility streams. From the Utility Composite Curve plot, you can determine the minimum hot and cold utility requirements for the network as well as visually inspect how much of each utility contributes to the total utility target.
- Hot Driving Force Curve
The Hot Driving Force Curve plots the temperature difference (driving force) between the hot and cold composite curves (Linnhoff 1993). The temperature difference between the two composite curves are plotted against the hot stream temperature. It depicts how driving forces vary with hot temperatures.
- Cold Driving Force Curve
The Cold Driving Force Curve plots the temperature difference (driving force) between the hot and cold composite curves (Linnhoff 1993). The temperature difference between the two composite curves are plotted against the cold stream temperature. It depicts how driving forces vary with cold temperatures.
- User Supplied Utility Load Plot
The User Supplied Utility Load Curve is essentially a composite curve of the utility streams specified by the user.
Keywords: AEA, Pinch Analysis, Heat Integration, Targeting
References: Linnhoff, B., Townsend, D.W., Boland, D., Hewitt, G.F., Thomas, B.E.A., Guy, A.R., Marsland, R.H., A User Guide on Process Integration for the Efficient use of Energy, IChemE England, 1982.
Linnhoff, B., Pinch Analysis: A State-of-the-Art Overview, Trans IChemE, Part A, 71, 503-522, 1993. |
Problem Statement: Why am I getting Remote database links: Maximum open cursors exceeded error message? | Solution: The Maximum open cursors exceeded message relates to the query exceeding a fixed limit of 100 for the number of open statements on a single Desktop ODBC connection. Usually, it is hard to reach this limit using the ODBC OAM from Aspen InfoPlus.21 because it automatically closes statements when it is finished with them. Some causes of the problem are:
1) The customer has a large (>100) number of cached (PROTECTED field set to CACHED) query records that are executing on the same external task and use the same database link to access the IP.21 system. If this is the case, the customer can use multiple SQLplus external tasks and/or multiple link names to reduce the number of statements per connection. or
2) The customer has found some way to write a query that doesn't free the statement automatically.
Aspen SQLplus includes a DISCONNECT command that can be used to manually disconnect open network links. The syntax of the command is:
DISCONNECT link_name;
Keywords: no free cursors
disconnect
ODBC
References: None |
Problem Statement: What is meaning of the dashed lines on Grid diagram in Aspen Energy Analyzer (AEA) | Solution: All process and utility streams in the heat exchanger network (HEN) system are displayed in the Grid Diagram.
Initially, all process streams are shown dashed because their enthalpy is not satisfied. As the heat exchangers are placed in the streams, the streams become solid.
With this feature, you can visually determine the unbalanced heating or cooling requirement for the streams.
The process streams that have no heat transfer to or from any utility stream may not reach their Target outlet temperature.
These process streams are represented by dashed lines.
Keywords: AEA, Grid Diagram, dashed line, Heat Integration
References: None |
Problem Statement: What the meaning of color of the heat exchangers on Grid Diagram in Aspen Energy Analyzer (AEA)? | Solution: There are five possible colors for the heat exchanger on the Grid Diagram in AEA. Each color indicates a type of heat exchanger:
Color Definition
Blue
Defines that heat exchanger as a cooler. In other words, the heat exchanger is attached to a hot process stream and a cold utility stream.
Red
Defines that heat exchanger as a heater. The heat exchanger is attached to a cold process stream and a hot utility stream.
Grey
Defines that heat exchanger as a process-process exchanger. The heat exchanger is attached to two process streams.
Yellow
Defines that heat exchanger is currently infeasible. Does not indicate if the exchanger is a cooler, heater, or process-process exchanger.
Green
Defines that the heat exchanger has been added/modified by a retrofit action.
Keywords: AEA, Grid diagram, Heat Integration
References: None |
Problem Statement: This article provide example files of Aspen Energy Analyzer (AEA) V12.1. | Solution: Aspen Energy Analyzer Sample files are stored in local folder as follows:
C:\Program Files\AspenTech\Aspen Energy Analyzer V12.1\Samples
Keywords: AEA, energy cost, heat integration
References: None |
Problem Statement: This knowledge base article explains why the following error may be found in the CIMIO_MSG.LOG:
WNT Error=10054 An existing connection was forcibly closed by the remote host | Solution: This message is typically preceded by several socket-level errors
CIMIO_OPC_RECEIVE_FAILED Error Receiving from MMMM on Node YYYY
CIMIO_MSG_RECV_CHK, Error checking connections
CIMIO_SOCK_CHK_RECV_FAIL, Error peeking at socket
System Error, errno=2 No such file or directory
CIMIO_SOCK_RD_BROKEN_PIPE, Connection was terminated by peer
WNT Error=10054 An existing connection was forcibly closed by the remote host.
These messages indicate that the network connection is being closed by an external mechanism or program. When this occurs the data transfer stops briefly then the 10054 message is logged which indicates that the socket has been forcibly closed. The next socket is selected and communication subsequently resumes. These errors are logged because Aspen Cim-IO encounters periodic breaks in the network.
Microsoft has this error documented here: https://support.microsoft.com/en-au/topic/an-application-may-receive-the-10054-error-when-the-application-receives-data-from-a-connection-on-a-computer-that-is-running-windows-7-or-windows-server-2008-r2-if-a-tdi-filter-driver-is-installed-73fd74a9-51c2-3663-e965-e1187b1cfc92
Specifically error 10054 means:
Connection reset by peer.
An existing connection was forcibly closed by the remote host. This normally results if the peer application on the remote host is suddenly stopped, the host is rebooted, the host or remote network interface is disabled, or the remote host uses a hard close (see setsockopt for more information on the SO_LINGER option on the remote socket). This error may also result if a connection was broken due to keep-alive activity detecting a failure while one or more operations are in progress. Operations that were in progress fail with WSAENETRESET. Subsequent operations fail with WSAECONNRESET.
Keywords: collect
not getting data
stop
References: None |
Problem Statement: You may want to configure KPI Hierarchy in aspenONE Process Explorer (A1PE). This article provides an example to do so. | Solution: You may need to do the following configurations to show KPI Hierarchy in A1PE. Please refer to the attached document:
Open Aspen mMDM Administrator to create an empty database
Download attached KPIHierarchy_xml.zip and extract the xml file. Import the xml file into the mMDM database using Aspen mMDM Bulk Load
Create Hierarchy from Aspen mMDM Editor
Create Equipment
Create Data source
Create Hierarchy
Publish the Hierarchy
Set the <ShowKPIHierarchy> as true on : C:\inetpub\wwwroot\AspenTech\ProcessData\AtProcessDataREST.config
IISreset.exe
Open A1PE and go to menu tree. Expand Hierarchies to see the detailed information. The KPI hierarchy should be displayed same as the mMDM part. You could also create more KPI node /categories on mMDM.
Keywords: KPI
KPI Hierarchy
A1PE KPI Hierarchy
References: None |
Problem Statement: Given the following information about a Gas Recovery Unit (GRU), APC users can practice their understanding of scope definition by answering the questions.
GRU Control Objectives
Maximize Deethanizer Feed
Minimize Propane Loss To Fuel
Maximize Mixed C4 Production
Minimize Energy Consumption
Maintain Product Qualities
GRU Typical Constraints
Depropanizer DP
Cooling Capacity on Debutanizer and Deethanizer
Propane Production (Limited Containment)
Deethanizer Feed Availability
Project Questions
Given the information above, what is the controller scope and what economics-related questions should be asked?
What additional questions should be asked to be able to design a good controller?
What other instruments, not shown in the graphic, should be placed in the collect file?
What calculations should be considered?
How many controllers or subcontrollers should we build?
Any operational suggestions? | Solution: 1 - Economic Questions
Can feed be cut when propane containment is full?
How would you characterize the feed?
Does it all come from one conversion unit or from many?
Is the makeup consistent from day to day, season to season?
What is the internal value of:
Feed
Fuel Gas
Propane
Mixed Butane
Pentane and Higher
Steam
Specs on Propane and Mixed Butanes?
2 - Questions for Controller Design
How often does a feed switch occur?
Analyzers:
Are they in service?
How well do they work?
What is their accuracy, repeatability?
Cycle time?
How much inventory in the bottom of each column? {Why do we ask this question?}
How does operations know when the propane containment is full?
3 - Questions for Instruments
How is the Deethanizer accumulator level currently controlled (L2)?
How is the Debutanizer accumulator operated? No level transmitter is shown – is it a flooded accumulator with a “level” in the condenser? If so, is there a level transmitter on the condenser, or is level inferred from pressure?
Do any of the regulatory control valves saturate?
What is the maximum allowable operating pressure for the Depropanizer (dP), the Deethanizer and the Debutanizer (reduce cooling requirements at higher pressure)?
4 - Calculations to Consider
Condenser differential temperatures on the Debutanizer and the Deethanizer (will need additional temperature transmitters)
Pressure compensated tray temperatures
Some quality constraint on the Debutanizer bottoms stream (i.e. vapor pressure)
Analyzer NEWPV calculation
Analyzer predictors (or use Aspen IQ)
5 - Controllers and Subcontrollers
1 vs 2 Controllers:
Depends on the operating philosophy
2 Controller case: one controller on the Deethanizer with an objective to always maximize feed and one controller the Depropanizer and Debutanizer provided that you NEVER want to cut deethanizer feed to limit propane losses to fuel.
1 Controller case: If Operations is willing to cut deethanizer feed to limit propane losses to fuel gas, especially when against a liquid propane production limit, then build 1 controller. Building 1 controller also provides the flexibility to postpone the decision about cutting charge to control propane losses to fuel gas.
Key question is will operations ever want to cut charge to limit propane losses to fuel gas.
Subcontrollers:
How many operators?
Three subcontrollers allows operation of two of the three columns even when one column is in distress, e. g. flooded.
6 - Operational Suggestion
Potential Controller Behavior to Reduce Impact of Depropanizer Capacity Limitation/Bottleneck
The depropanizer capacity is a constraint on the system, measured by the dP.
If operations does not want to cut charge to limit propane losses to the fuel gas system, then it makes some sense to run the deethanizer fairly hot – get as much of the ethane out of propane as possible because this will increase the depropanizer dP. Also, rather than taking propane to the depropanizer and venting it into the fuel system, unload the depropanizer by venting propane off the top of the deethanizer and sending only as much propane to the depropanizer as can be sent to containment.
Keywords: gas, recovery, unit, apc, project, example, workshop, questions, exercise,
References: : APC105 V11 Appendix |
Problem Statement: After loading and attempting to start an Aspen APC Inferential Qualities (IQ) application using PCWS | Manage or APCManage, the IQ may fail to start with the message: Unable to checkout Aspen IQ license SLM_RN_APC_AIQINF.
This license key SLM_RN_APC_AIQINF is required specifically for the IQ Watch tool, which allows the user to enable Aspen Watch (AW) Performance Monitoring for IQ applications by starting data collection for IQs using Watch Maker or the PCWS | History tab | Collection Status. However, if the license being used is missing this key, the user will see this error message when trying to start the IQ application. Some users intentionally opt out of having this IQ Watch feature because it uses up additional tokens (more info on APC Common Token Usage can be found in KB 000094989).
The user may check if this license key is missing using the aspenONE SLM License Manager | License Profiler tool. More information on how to use the License Profiler can be found in KB 000094205. | Solution: If the license key for IQ Watch is missing, the user can do the following steps as a workaround to get the IQ running:
1. Open the IQF in IQconfig, right-click on the Inferential name and select Properties
2. In the Inferential Quality Properties window, uncheck the box for Enable Performance Monitoring in Aspen Watch (AWENB)
3. Then reload and start the IQ application using Manage
The aboveSolution will allow the IQ to start without looking for that missing license key. However, it will also prevent the user from starting data collection for this IQ application in Aspen Watch.
If the user does wish to use IQ Watch, a new license that includes this missing key needs to be requested from Customer Care by filling out this form (be sure to specify the license key that you would like to be added): License Request Form.
Keywords: IQ, fails, start, load, manage, license, key, missing, unable, checkout, SLM_RN_APC_AIQINF, AIQINF
References: None |
Problem Statement: The APC IO logging Service can be used to gather some connection information from the RTE controllers. However, it has been noticed that this service may fail to start after a sever upgrade from V12.1 to V14. This KB describes the steps to solve this problem. | Solution: The most common symptom of this problem will be an error message after trying to start the service that calls for a missing file on the server:
The reason for this is that Service failed to get registered on the correct path during the upgrade to V14.
To solve this problem, you will have to change the path to the correct one on the Server Registry Editor.
1.- Click on Windows Start and type Regedit
2.- Once the Regedit is open, locate the following path:
3.- Once on the service details, you will notice the service will be still pointing to the V12.1 file path:
4.- To solve the problem we will have to change the path to the correct V14 path. To do so, double click on ImagePath and change the path. It should look like this:
5.- Once this step is completed, try again to start the service. This time it should run.
6.- Optionally, this service may have stayed as the V12.1 name AspenTech APC IO Logging Service. To make this consistent with the intended V14 install, the user can change the same registry for the field DisplayName to be Aspen APC IO Logging Service. A reboot of the machine may be required to implement this change.
Keywords: DMC3, IO, Logging, service, V14, RTE
Internal
References: : Defect 913837
Target fix to be V12.1 CP2 EP3, V14 EP2, and next major release. |
Problem Statement: The aspenONE Process Explorer Admin page can be opened by using the following URL in a web browser opened on the aspenONE Process Explorer (A1PE) web server: http://localhost/ProcessExplorer/WebControls/PBItemDetails.asp?admin=true
On the aspenONE Home tab there is a table that can be used to configure action links that are then included on every users' aspenONE Process Explorer home workspace. When clicking the Save button, a warning icon may appear in the bar at the top of the page, clicking on the icon results in a panel with the message Access Denied.: | Solution: When clicking the Save button, you are actually making use of a Process Data REST command that does require you to have elevated access permissions. Nevertheless, this error message is known to occur when using MS Internet Explorer (IE) on a server which also has IE Enhanced Security Configuration switched ON (see ServerManager.exe):
You could try again by opening IE using elevated permissions (Run as Administrator), or use an up-to-date alternative supported web browser (MS Edge, Chrome, Safari).
Keywords: rights
public
All or part of the path is invalid.
Manage Search Data
References: None |
Problem Statement: When there is no reboiler in a RadFrac block the bottom pressure/liquid outlet stream pressure is not correct using the automatic pressure update functionality. The pressure is reported at the top of the stage. For example, the internal rating calculates a total pressure drop of 0.542 psi but the bottom stage pressure/liquid outlet stream pressure is only showing a pressure drop of 0.488 psi. This is the pressure drop when the last stage in the internals rating is not used. If the assumption that the pressure is reported at the top of the stage the pressure on the last stage might be correct but at least the bottom liquid outlet stream, coming out of that last stage, should show the pressure drop of the last stage, so reading a pressure equivalent to a pressure drop of 0.542 psi. This would cause problems if designing a heat exchanger using the bottoms stream pressure. | Solution: In the RadFrac rigorous distillation unit operation block, we use theoretical stages. Each stage has a single pressure. Therefore we can only report a single pressure for each stage. For a real stage, there is a pressure above the stage and a pressure below this stage. This can be modeled using Rate-based distillation. For equilibrium stages, the pressure can be updated from the top or from the bottom. For columns with no reboiler, it is probably best to update from the bottom.
The pressure drop calculated for stage N is the pressure difference between stage N and stage N+1. In columns without reboilers where the last stage is included in a pressure update section, the pressure drop for the last stage is not used because there is no stage below to receive the updated pressure.
It is possible to use the Update pressure drop from bottom stage option to set the pressure starting at the bottom and the top stage pressure drop will not be used.
Another workaround, if the column is RadFrac, it doesn't involve 3-phase calculations, and an Aspen Rate-Based Distillation license is available, is to add a dummy rate-based stage at the bottom of the column:
Increase the number of stages by 1.
Adjust feed locations for feeds to the bottom of the column. Adjust any other inputs as necessary. Note that the new bottom stage should not be part of any reaction section.
Set the Calculation type to Rate-Based on the Setup | Configuration sheet.
Create a rating section that contains only the last stage. Select a packing that doesn't require any additional input parameters. Use a small pack height. Mark this section as rate-based. Use a small number such as 1E-5 for Interfacial Area Factor and Heat Transfer Factor.
No appreciable reaction or separation occurs in the dummy stage, but its presence allows the pressure drop to be applied appropriately.
.
Keywords: None
References: : VSTS 835090 |
Problem Statement: When changing or upgrading a license file or license server, what are the steps required for APC applications running online? | Solution: The recommended practice for APC applications during license changes or upgrades depends on whether only the license file is getting replaced on the same license server or if the license server is changing as well, and the APC server must point to the new license server.
Scenario 1: Replacing License on the Same License Server
When an APC application starts, it creates Last Valid Timestamp files, which have file extensions *.LST and *.2ST, in these directories of the APC Online Server:
C:\ProgramData\AspenTech\APC\Online\sys\etc
C:\ProgramData\AspenTech\RTE\<version>\Clouds\Online\sys
These files contain the license information (cache) and if the connection to the license server is disrupted, then it will start the 10-day countdown grace period. You can see that the license server connection has been lost from running applications by checking the PCWS (web viewer) messages that will indicate the license is lost and this countdown has started.
Normally, when a user changes the license file on the license server, the APC applications may lose connection to the license server for that short time during the change. This will not disrupt the running applications though, for now, as the 10-day grace period will begin. Once the license replacement is complete, communications are reestablished and the license is reacquired, the countdown message will be removed from the controller.
Therefore, when the license is being replaced on the same license server, there is no need for the controller to be stopped and started again. It should automatically reacquire the license once it establishes a connection to the license server after the change is complete. You can confirm that the license was reacquired by viewing the PCWS messages as they should not be showing a countdown period anymore.
However, if for any reason the license was not re-acquired automatically (this can happen if, for example, the license configuration was changed), try the following troubleshooting steps:
(1) First try stopping and starting the APC applications from Manage before the grace period is over.
(2) If that still does not work, then:
Take a snapshot of the online applications if you do not already have a copy of the latest configuration.
Stop and delete the online loaded applications.
Delete the Last Valid Timestamp files mentioned above.
Deploy / load the applications as new and Start them. This will re-create those Last Valid Timestamp files again with the new license file cache.
Scenario 2: License Server is Changed
In this scenario, if you change to a new license server, this information is not updated in the controller so the license will not be reacquired automatically like Scenario 1. When a new license server is used, you need to stop and restart the applications so that the Last Valid Timestamp files are recreated with the new license server information.
Ask site operators to turn off any running APC applications (Master Request Switch should be OFF on the PCWS)
Stop all APC applications:
Optionally, you can use Watch Maker on the Aspen Watch Server (or from PCWS > History > Collection Status) to first stop data collection for all applications so data is not being collected while controllers are off
ACO Controllers and IQ Applications – use APCManage on the Online Server to stop controllers (OR from PCWS > Online > Manage)
RTE Controllers – use DMC3 Builder’s Online section to stop controller processes (OR from PCWS > Online > Manage)
Change License to New License Server
View AspenTech Knowledge Based Article How to Install a Network License File
Use the SLM License Manager > Configuration wizard on the APC servers to connect to the new license server
Start Online Applications
ACO Controllers and IQ Applications – use APCManage OR PCWS > Manage
RTE Controllers – use DMC3 Builder’s Online section OR from PCWS > Manage
If data collection was stopped, you can start it up again once the controllers are online
Validation Checks / Troubleshooting Tips:
Check that SLM Manager > License Profiler to verify that tokens are in use when applications are running online
Observe the PCWS messages for any licensing errors
Open a DMC3 Builder project and if there is a license issue, it will say the license was not acquired and the project won’t open
KB on where to find licensing logs for APC applications: https://esupport.aspentech.com/S_Article?id=000094884
KB on how to validate APC applications after license change: https://esupport.aspentech.com/S_Article?id=000072828
If license was upgraded for Aspen Watch and/or IP.21, you might run into a “License Denied” state in IP.21 and “License Lost: See Events” run status in Aspen Watch. Here is a KB on how to resolve this: https://esupport.aspentech.com/S_Article?id=000097685
Keywords: license, apc, upgrade, install, replace
References: None |
Problem Statement: Starting in V12.1, DMC3 Builder projects have a single-file project structure with the extension .dmc3proj, which allows for large project files to be compressed more than the folder structure was able to in previous versions. There has been one instance found where using the File | Save Project As command allows the newly created project to compress as expected when it is closed, but the original project remains to be a large file size.
Note: this issue has only been seen with some projects, not all DMC3 projects, so the | Solution: below may not be applicable in most cases.
Steps to reproduce, as tested in V12.1 and V14:
1. Before opening the project, note down the file size in File Explorer. In this example, the project is 39 MB.
2. Open the project in DMC3 Builder and note the file size increases to 64 MB as expected:
3. In DMC3 Builder, go to File and click Save Project As with a new project file name:
4. Close the DMC3 Builder project and notice how the new project compresses down to 39 MB but the original project remains at 64 MB:
Solution
The workaround for the issue is to simply open the original project, close the project, and choose Yes when prompted to Save. This should compress the original project file size down:
The target fix for this defect (813080) will be the next major release.
Keywords: save, as, dmc3, project, .dmc3proj, compress, file, size, increase
References: None |
Problem Statement: How to change the solving method and damping group in any column? | Solution: The Solving Method in a column allows you to select the columnSolution method. With the “inside-out” based algorithms, simple equilibrium and enthalpy models are used in the inner loop to solve the overall component and heat balances as well as any specifications. The outer loop updates the simple thermodynamic models with rigorous model calculations.
To change the solving method, open the parameters tab in the column property window and select the solver page on the left-hand side and change the solving method from the drop-down list.
Damping factor is used to control the step size for the solver. It helps in faster convergence of a column solver.
The Damping Factor is a multiplication factor that controls the step size used in the outer loop when updating the thermodynamic models in the inner loop. A Damping Factor of 0.4 -0.6 may provide a faster, more stable convergence for most of the column simulations.
To change the damping group, open the parameters tab of the column property window and select the solver page.
Keywords: Damping, solver setting, parameters, column
References: https://esupport.aspentech.com/S_Article?id=000100852
https://esupport.aspentech.com/S_Article?id=000087803 |
Problem Statement: How to change the transfer basis for an internal stream in any column? | Solution: For any column property window, the Flowsheet tab contains the following pages:
Setup
Variables
Internal Streams
Mapping
Lock
The Transfer Basis is significant only when the sub-flowsheet and parent flowsheet Property Packages are different. To change the transfer basis for an associated inlet or outlet stream in a column, you may have to open the flowsheet tab in the column property window and click on setup page as shown below:
Keywords: Column flowsheet, transfer basis, property packages
References: https://esupport.aspentech.com/S_Article?id=000087448 |
Problem Statement: What should be done to address the error FLASH CALCULATIONS FAILED TO CONVERGE IN MAXIMUM NUMBER OF ITERATIONS for an Acid gas column model using Acid Gas-Caustic package alongside PR in Aspen HYSYS? | Solution: The Acid Gas - Caustic Wash property package can be used to model the caustic wash process. During the caustic wash process, impurities such as H2S and light mercaptans are removed from light hydrocarbon feeds. This property package functions similarly to the other Acid Gas property packages. The primary difference is that caustic NaOH, rather than an amine or physical solvent, is used to remove acid gas from the feedstock.
The caustic wash system can be used for both liquid and vapor hydrocarbons. You can use the Absorber and Liquid-Liquid Extractor columns to simulate this process.
While modelling Acid gas column using both Acid gas caustic wash package as well as Peng Robinson, there are probabilities of encountering with error to solve the feed stream in the column environment.
Choosing the appropriate basis for the unit operation and material stream is an important step in such modelling cases. Along with that the following workaround helps to resolve the error on iterations:
Increase the number of iterations for flash convergence on the Caustic Wash package in the properties environment. This will get the feed to solve on the sub-flowsheet.
To change the number of iterations for a fluid package follow the steps as below:
Enter the properties environment
Select the fluid package and click on edit option
Move on to Phase Handling Tab and increase the maximum number iterations
Keywords: Caustic wash, acid gas, iterations, fluid package, liquid-liquid extractor
References: https://esupport.aspentech.com/S_Article?id=000088403 |
Problem Statement: This tutorial shows the steps of how the CreateVectors.sql query works on how it can be used to extract data from the Misc. tags on the Aspen Watch Server. | Solution: For this Tutorial example, we have included a PDF file with a details example on how to use the CreateVector.sql attached to thisSolution. The query is defaulted installed on any AspenWatch Server, but is missing you can ask for a copy to AspenTech APC support.
Keywords: AspenWatch, VEC Files, SQLplus
References: None |
Problem Statement: This article describes the steps to take to extract the logs from PCWS. | Solution: The short answer to this problem would be No. Unfortunately, there is not a direct way to extract the messages from PCWS.
However, there is one useful workaround that can be used to extract the same information from PCWS using a Windows tool called Event Viewer.
Here are the steps that you can follow to get the information of the controllers logged on PCWS:
1.- On the Online Server (DMC Server) go to windows Start and Select the Event Viewer application.
2.- Once is open select the RTE node, to gather the DMC3 Controllers information. Then, on the right Side select the option Save All Events As
3.- When saving the events, you can store the file as different options. You can save it as txt or csv to store in plain text.
4.- you can use the Filter log options, to decrease the number of messages to store. I suggest filtering by time, but you can search for specific messages.
4.- Here is an example of a filter log by the last hour as csv file on excel.
Keywords: PCWS, Logs, Event Viewer
References: None |
Problem Statement: This | Solution: frames to how to use a Token Counter Report Query for DMC3 and DMCplus applications.
Solution
1.- Copy the query on Aspen SQLplus on the AspenWatch server. The query script is the following, and can be found on the attached txt file.
---------------CONTROLLER INFORMATION
Keywords: None
References: None |
Problem Statement: Users have the option of making their PCWS web page to have read-only access if they want to restrict permissions on certain security levels, for example a web server on the DMZ. To do so, the user would have to follow the steps below:
For ACO (ccf/mdl) applications:
Open the C:\ProgramData\AspenTech\APC\Online\cfg\AspenTech.ACP.DMCplus.DataService.config file
Change allowManage=False and save the file
Restart the Aspen APC DMCplus Data Service
For RTE (DMC3 Builder) applications:
Open the C:\ProgramData\AspenTech\RTE\Vxx\Config\AspenTech.ACP.RTE.DataService.config file
Change allowManage=False and save the file
Restart the AspenTech APC Production Control RTE Service (in V14 and later, the service was renamed to Aspen APC Production Control RTE Service)
One side effect of making this change noted above, however, is that when trying to access the DMC3 Strategy page on this PCWS to view and change DMC3 Smart Tune configuration, they would see an error message like this:
In V12.1 and previous versions: Error: Exception caught. Proxy not connected.
In V14 and later versions: Data Service security prohibits access to this feature. | Solution: This behavior is by design. Having the DataService.config file (both for RTE and DMCplus) set to allowManage=False for the web server blocks ALL Manage Command interface actions from occurring when received from that web server. DMC3 Smart Tune also uses this interface, therefore the actions on the Strategy page used to view and change the Smart Tune configuration is also blocked. We cannot change this behavior for the Manage Command interface, due to its purpose of security restrictions.
Keywords: PCWS, apc, web, strategy, smart, tune, error, proxy, connected, data, service, security, restriction, permission, blocked, read, only, dmz
Internal reference: Defect 597139
References: None |
Problem Statement: This article described what HBK files and good practices to manage them. | Solution: Simple definition: HBK files are history backup files generated by IQ applications.
These files are created from multiple entries on the inferential (depending on lab or analyzer), and the main purpose is to keep a backup of the IQ application. They can also be used to extract data using IQ Extract to convert it into clc/txt files.
Normally they are located in the following URL:
APC\Online\sys\app\<appname> folder on the Online Server
However, one of the problems related to these files is that multiple files may get created and if we don’t keep an eye on them, it is highly possible that they end up filling the Hard Disk Space. The best way to manage these files is by keeping the files that will keep the most recent data of the IQ application.
The .hbk files can be deleted. We suggest deleting the old ones because those will be from a state of the application that may not be used anymore. Do not delete the most recent .hbk files along with the .his files in case they may need.
Note that the files can be deleted directly from the folder. The process should not interrupt the IQ performance.
Additional Information on HBK and HIS files:
Typically the size the history file is managed by setting the maximum number of records per entry the customer wants to store in history. This can be set in the CONFIG section of IQconfig with the corresponding parameter for each history file: AZUMAXREC, DCMAXREC, LBUMAXREC, LDCMAXREC. Each entry is discrete in that they do not have to be aligned in time with the other stored entries. When the number of records for an entry exceeds the maximum then the oldest record is replaced by a new record.
Something that we can try to reduce the size of these history files is to reduce the value of DCMAXREC to 1440. Another parameter that you can reduce in the CONFIG section of the IQF is the PLOTHOURS (default is 24 hours) which would also reduce the amount of storage the files take up. It It would be good to stop and delete the IQ files (please make a backup for your record). Sometimes you have to manually delete the folder from C:\ProgramData\AspenTech\APC\Online\sys\app to remove any residual file. Then on the IQF lower the value of DCMAXREC, PLOTHOURS, and any other parameters before loading the IQ again.
Keywords: IQ, hbk files, Online Server
References: None |
Problem Statement: When running Optimization in Aspen Utilities Planner user have an option to include StartUp and ShutDown cost for some equipment. However sometimes one can observe that, even if initial specification for the equipment is different: in one case Equipment status is “In Service” and in another “Shut Down”, the results for Optimization cost has not changed. Why StartUp and ShutDown cost are not taken into account in the optimization? | Solution: When running single period optimization Aspen Utilities Planner is not taking into account StartUp and ShutDown costs, as Initial specification of the equipment from Pre-Optimization simulation is not checked by the optimizer.
StartUp and ShutDown cost are included ONLY with multi period optimization, and only for periods: 2 and higher. For first specified period Aspen Utilities Planner assume that specification included in the Availabilities is the initial state of the equipment.
If you observe the issue with multiperiod optimization, then make sure to specify StartUpTime, ShutDownTime, StartUpCost and ShutDownCost in Optimizatiom_Settings Table. Select: True value for DoStartStop option, as shown in a screenshot below:
Keywords: Optimization, single period, StartUp, ShutDown
References: None |
Problem Statement: How to register for a Public Training classes through AspenTech eSupport web site? | Solution: AspenTech offers a comprehensive curriculum of training courses on how to apply our software products and technologies to achieve business and operational benefits. Our customers have choices of arranging for instructor-led, private training classes at their sites or registering for public classes at our regional training centers at various locations around the world. They can also select to attend classes from their own offices through the Internet (Virtual Classroom).
Public trainings will be guided by our experts. The training will be interactive, live-instructor led program; delivered in-person or virtually.
To register for public training, please check out the following link: AspenTech Training Center
Login to https://esupport.aspentech.com/
Select AspenTech University > Take a Class > Register for a class
Use the search option to easily find the upcoming training courses in the website.
Use the filter options such as Class Type, Start date, Location, Language etc to easily find the required product and courses.
Once you find the course of your interest, proceed clicking on the Register option and fill the necessary details requested and finally submit.
Important Note: If you are a corporate customer, and would like to pay later via company invoice, then please choose Other Payment Methods and proceed submitting your training registration.
When you select Other Payment Methods :
No payment will be required on the spot of registration.
AspenTech will share the payment invoice directly with your organization.
Key words
Training, Class, Register, Public, Virtual, Payment, Company Invoice
Keywords: None
References: training.aspentech.com
https://esupport.aspentech.com/ |
Problem Statement: When installing Tomcat, it is possible to change the default port number Tomcat usually installs on to some other port. This may cause Aspen Search not to work correctly.
This Knowledge Base article provides a list of files to check when the Tomcat port number was changed from default 8080 to some other port. This is to make sure the files have been updated accordingly. | Solution: Open the following files in a text editor, search for 8080 and replace with the new port number (8081 for example):
C:\inetpub\wwwroot\AspenTech\ProcessData\AtProcessDataREST.config
C:\inetpub\wwwroot\AspenTech\ProcessExplorer\WebControls\AtWebPlotsConfig.xml
C:\inetpub\wwwroot\AspenTech\aspenONE\App_Data\config.xml
C:\Program Files (x86)\Common Files\AspenTech Shared\TomcatX.X.XX\conf\server.xml
C:\Program Files (x86)\Common Files\AspenTech Shared\TomcatX.X.XX\appdata\scheduler\config\jobs\NPEScanData.xml
C:\Program Files (x86)\Common Files\AspenTech Shared\TomcatX.X.XX\appdata\scheduler\config\jobs\NPEScanFiles.xml
C:\Program Files (x86)\Common Files\AspenTech Shared\TomcatX.X.XX\appdata\scheduler\config\jobs\ScanAPRMDataSources.xml
C:\inetpub\wwwroot\AspenTech\Web21\WebControls\AtWebPlotsConfig.xml (this is for search within IP.21 Process Browser)
C:\inetpub\wwwroot\AspenTech\Web21\WebControls\AtDetailSearch.asp (this is for search within IP.21 Process Browser)
Keywords: Aspen Search
Tomcat
References: None |
Problem Statement: The Aspen Manufacturing Suite has the following two options for viewing data from a relational database within Aspen Process Explorer:
SQLA Tags
Aspen Process Data (RDBMS)
This knowledge base article discusses the differences in the way each application connects to the relational database. | Solution: When SQLA Tags are used, a single SQLConnectionDef record is created in the Aspen InfoPlus.21 database which contains the OLE DB connection information to your relational database. (See image below)
With the Aspen Process Data (RDBMS) option, you do not have to create a record in Aspen InfoPlus.21 to hold the database connection information. Rather you create the OLE DB relational database connection in the ADSA configuration for the new data source that is defined for the Aspen Process Data (RDBMS) connection.
(See image below)
Although Aspen Process Data (RDBMS) is initially a bit easier to set up than SQLA Tags, there are some advantages to using SQLA Tags. These advantages are:
If you use Oracle as the relational database, you must install the Oracle client software on each Aspen Process Explorer machine when using Aspen Process Data (RDBMS). This is because each Aspen Process Explorer client makes a separate connection to the relational database when Aspen Process Data (RDBMS) is used. When using SQLA Tags, the Oracle client only needs to be installed on the Aspen InfoPlus.21 server.
If you license your relational database based on the number of simultaneous connections to the database server, each Aspen Process Explorer client counts as a separate connection to the server when Aspen Process Data (RDBMS) is used. This is not the case with SQLA tags since there is only one client connection made to the relational database from the SQLConnectionDef record.
You must create a separate data source in ADSA with the Aspen Process Data (RDBMS) option. This means that an Aspen Process Explorer user will have to select the new data source in the trend plot to view the data from the relational database. When SQLA Tags are used, the user uses the same ADSA data source that points to the Aspen InfoPlus.21 data source.
Keywords: Advantage
Compare
real-time
References: None |
Problem Statement: How to use Power Stream in Aspen Plus? | Solution: Use an inlet power stream to supply an electrical power specification to a block that requires electrical power, such as an Electrolyzer block.
Use the following forms to enter data for work streams.
Use this form To do this
Input Specify power for power streams
Result View stream results for power streams
There is a new Electrolyzer unit operation to model the electrolysis of water to generate hydrogen. There are also electrical power streams to specify the power input into such operations. Power streams are not supported in EO.
Keywords: Power stream, Electrolyzer
References: None |
Problem Statement: Is the Power stream supported in Aspen Plus V14 Equation-Oriented mode? | Solution: Power streams are not supported in EO.
Keywords: Power stream, EO
References: None |
Problem Statement: How do I change my display name on Datasheets and Drawings during the Submit/Check/Issue Workflow? | Solution: The steps to change the display are described below.
Note: The modified Display Name will only affect the new Datasheets and Drawings, not the existing ones.
Solution
To modify the display name, go to the Administration tool (AZAdminTool)
Under the corresponding workspace, click on User
Right click the user that you want to modify the display name and select Properties
Under the Properties window, enter your desired name under Display Name
Reload your workspace and the next Submit/Check/Issue Datasheet and/or Drawing will use the new display name:
Keywords: Aspen Basic Engineer, Display Name
References: None |
Problem Statement: How do I create a Plant Breakdown Structure (PBS)? | Solution: Plant Breakdown Structure represents the geography of a plant. You can associate an object/Document/Datasheet to a plant structure like Unit or Area. First a template object of class PBSConfiguration needs to be created.
There are two ways to create a Plant Breakdown Structure:
1. Import a Plant Breakdown Structure available under Packages:
A template of standard Plant Breakdown Structure of Site, Plant, Area, Unit is provided in the package file PlantBreakdownStructure.zpkg. This file is installed in the default `package? folder. If you want to use standard structure, you can import it using the Explorer.
1. Click Folder | Import Package. The Import Package dialog box appears.
2. Select PlantBreakdownStructure from the list. (Click on Library to select the default directory)
3. Click Import.
2. Manually set up a Plant Breakdown Structure
You can use this method if you don't want to use standard structure. To create a new template:
1. Select the root level folder of the workspace, right-click, and select New | Pbs template config to create a new PBS template, e.g., Template1.
2. Select Template1, right-click, and select New | Pbs template container to create the top level container type name, e.g., Sites.
3. Click Sites and continue adding container types as necessary, e.g., Plants, Areas, Units, etc.
To create PBS folders using any of these templates:
1. After the templates have been created in a workspace, select the root level folder of the workspace, right-click, and select New | TemplateName::Toplevel folder type name.
2. You can create subfolders in the same way; you will be prompted with the available types as defined in the template.
3. Note: To be able to add objects into any level of the Plant Breakdown Structure: From the Template Structure, right click the level you would like to add objects to and select New | Pbs Template Item
Keywords: PBS, templates, structures
References: None |
Problem Statement: When I add an Engineering Fee to a Contractor, how does it appear in my indirect reports? | Solution: The Engineering Fee will appear in your indirect reports as part of your Contract Fees.
Keywords: Engineering Fee, Engineering, Fee, Contract Fees, Fees, Contract
References: None |
Problem Statement: Does equipment summary cost include the cost for all the identical items? | Solution: In Aspen Economic Evaluation, components such as pump of same design and installation specifications can be entered as identical items.
The equipment summary available in HTML, Reporter Report or Capital Cost Item Report includes the cost of all identical items.
Keywords: Equipment Summary, identical, items.
References: None |
Problem Statement: How do I model a cold box in Aspen Process Economic Analyzer? | Solution: If you are using Aspen Process Economic Analyzer (APEA) and would like to model a cold box, there are two approaches. APEA does not have a cold box, but Aspen Capital Cost Estimator (ACCE) does.
1. If you only have access to APEA, then you need to treat the cold box as a quoted Item. Enter it directly as quoted equipment, and enter the cost.
2. If you have access to ACCE, you can open your APEA file in ACCE, and use the ACCE area type for cold box. In the area give the details and it will then cost the cold box for you.
Keywords: cold box
References: None |
Problem Statement: Does the PFD in Aspen Process Economic Analyzer have a Find feature for the Components | Solution: Yes there is a Find feature in Aspen Process Economic Analyzer for the PFD Components.
1.) Open the PFD in Aspen Process Economic Analyzer
2.) In the Project Explorer window, make sure you are in the Project View tab.
3.) In the Project View tab, you will see your various Components. With the right mouse button, click on the one you want to find in the PFD. Hold the mouse button down and drag the mouse pointer from the Project View tab to the PFD. The PFD will have you Component appear on the right side.
Keywords: Find, Locate, PFD
References: None |
Problem Statement: For agitated tanks, other than the Machine Pulper components, I do not see a choice for seal type; what is the seal being used for Agitated Tanks? | Solution: We are using packed seals as the seals on the agitated tanks. The seal is part of the agitator package and the cost of that is based on a packed seal.
Keywords:
References: None |
Problem Statement: Can I do multiple item evaluations during one evaluation? | Solution: Yes - highlight the Components to be evaluated in the List View window, then use your right mouse button and click evaluate item
Keywords: None
References: None |
Problem Statement: If I create a Template in Aspen Capital Cost Estimator, can I use it in Aspen In-Plant Cost Estimator or Aspen Process Economic Analyzer? | Solution: Yes. That is one of the beauties of the new Template capability in V7.3 for the Economic Evaluation products. You can create a template in one of the products and use it in any of the other products!
Keywords: None
References: None |
Problem Statement: How can I adjust my piping and instrumentation for a component in Aspen Economic Evaluation products like Aspen Economic Analyzer, Aspen In-Plant Cost Estimator and Aspen Capital Cost Estimator? | Solution: You can adjust the piping and instrumentation for a component in two ways in Aspen Capital Cost Estimator. Either by using form based Pipe item detail from option menu of the component or by using graphical mode from the menu. In Aspen Process Economic Analyzer and in Aspen In-Plant Cost Estimator only user mode can be used.
Whenever a component is created in Aspen Process Economic Analyzer and in Aspen Capital Cost Estimator piping and instrumentation (P&ID) is automatically associated with the component. If you select load volumetric model from the option menu it will load the model in the form and you can see the detail from pipe item detail.
User mode:
You can select Pipe-Item Details from the Options menu of the component and it will take you to the form based adjustment.
After you select Pipe-Item Details from the Option menu you will see a form similar to the one below:
There are another method to adjust the P&ID in Aspen Capital Cost Estimator is by using piping and instrumentation graphical mode.
Graphical mode:
When you press the P&ID button it will take you to the page where you can see the piping and instrumentation that is associated with the component. The graphical page you see will be similar to what is shown below:
Using the symbol menu you can then modify the piping and instrumentation graphically.
Aspen Process Economic Analyzer will let you see the volumetric model from the graphical mode but you can not modify anything in there like you can do in Aspen Capital Cost Estimator.
Aspen In-Plant Cost Estimator will let you add piping and instrumentation by user mode pipe item detail but there is not any option of loading volumetric model with the component or the Graphical mode that can be used.
Keywords: P&ID, piping, instrumentation, graphical, user, mode
References: None |
Problem Statement: How do you specify the Engineering Start Date and the Construction Start Date in Aspen Process Economic Analyzer (APEA) and Aspen Capital Cost Estimator (ACCE)? | Solution: For Aspen Process Economic Analyzer (APEA):
The Engineering Start Date can be specified in the Project Basis View | Basis for Capital Costs | General Specs form. In APEA, the Construction Start Date is a calculated date which occurs after the Engineering Start Date and is a function of the number and type components in the estimate. A specific Construction Start Date can not be entered directly by a user.
For Aspen Capital Cost Estimator (ACCE):
Both the Engineering Start Date and the Construction Start Date can be specified under Project Basis View | Project Execution Schedule Settings | Adjust Schedule and Barcharts:
Keywords: Start Date, Start, Construction, Engineering, Construction Start Date, Engineering Start Date
References: None |
Problem Statement: Why do I get the following error when I try to model a combustor reactor as a furnace:
User Tag Number:
*Component Ref #: 32
ERROR> 'FU - 32' STRESS FOR MATERIAL A 214 IS ZERO AT A TEMPERATURE OF 1200.00
SYSTEM MAY NOT CONTAIN STRESS VALUES FOR THE MATERIAL. | Solution: The furnace models in our library are actually fire heaters, not combustion reactors. That means that we have inside the combustion chamber a set of tubes with the process stream to be heated. The material that we specify for the furnace in the input form is in fact the material for the tubes, and that is why we usually have an error with the material stress factor if we map the combustor as a furnace (the operating temperature for a combustor is usually higher than the maximum limit for the tube materials in a fire heater).
The furnace model already contains refractory lined materials by default. From our documentation: “The walls of the furnace are refractory (an insulating heat resistant material) lined”. The design temperature we input in the model refers to the temperature that the tubes are required to undertake. The value that will be used for its mechanical design (stress factor for ASME code). In the Icarus reference, the maximum temperature for a furnace box type is 850C (1500F), that corresponds to the use of an austenitic stainless steel, for example 316S.
We have three options:
1. Use the fire heater (furnace) model, select standard CS and lower the temperature within the material range. Because our real equipment model won’t contain the tubes, we can apply a material percent adjustment to account for that.
2. Map the combustor as a set of a vessel and a lining component. You will need to specify dimensions:
3. You can use instead a quoted item and add manually the vendor cost for that item.
Keywords: Error, Material Stress, Furnace
References: None |
Problem Statement: When I add PVC or CPVC piping, I don't see any bolts or flanges generated. How do I add them? | Solution: Economic Evaluation products do not generate bolts and flanges for PVC or CPVC piping. You can add the bolts and flange costs by using the Mat'l/Man-hour additions field on the Components Options menu.
Keywords: PVC, CPVC, bolts, flanges
References: None |
Problem Statement: How is Electrical circuit testing calculated in Aspen Capital Cost Estimator? | Solution: Electrical testing is calculated based the number of certain types of items that we perform testing on, multiplied by a factor to get man-hours.
We test on these Electrical items for SUBSTATIONS and for each AREA:
Low, Med and High Voltage Transformers
Electrical Circuits
Transmission lines
Grounding
Low, Med and High switchgears
The algorithm we use to calculate those manhours for Substations is the following:
(# of Low Voltage Transformers) 6.0 + (# of High Voltage and Medium Voltage Transformers) 8.0 + (# of Electrical Circuits) 4.0 + (# of Transmission lines) 15.0 + (# of Grounding items) 4.0 + (# of Low and Med Voltage Switchgears) 3.0 + (# of High Voltage Switchgears) * 4.0
The algorithm we use to calculate those manhours for each AREA is the following:
(# of Motors) 1.5 + (# of Electrical Circuits) 2.5 + (Total number of MCCs) 0.3 + (# of Low and Med Voltage Switchgears) 3.0 + (# of High Voltage Switchgears) 4.0 + (# of Low Voltage Transformers) 6.0 + (# of High Voltage and Medium Voltage Transformers) 8.0 + (# of Transmission lines) 15.0 + (# of Grounding items) * 0.15
Keywords: electrical
testing
circuit
References: None |
Problem Statement: How to set up a project so that motors developed in Aspen Process Economic Analyzer will be explosion-proof? | Solution: Explosion-proof can be specified under Project Basis View | Basis for Capital Costs | Design Basis | Electrical Specs, which is accessible in Aspen Capital Cost Estimator, but not in Aspen Process Economic Analyzer. For this reason, template generated in ACCE can be used to set this up for Aspen Process Economic Analyzer.
1. Open ACCE, click on File | New template, and give it a name, e.g. explosion-proof.
2. In the new template, go to Project Basis View | Basis for Capital Costs | Design Basis | Electrical Specs form. Specify 0 for Class and division or zone field. This will make all the motors in a project to be explosion proofed. Save the template and close ACCE.
3. Open APEA, create a new project, click on template button in the Create New Project window, and pick the template you want to use in the pop-up window, e.g. explosion-proof.
4. You can then enter new components and motors developed will be explosion proof, or import components from old projects by drag/drop
Keywords: Explosion proof
References: None |
Problem Statement: How is Instrumentation Testing calculated in Economic Evaluation products? | Solution: Instrumentation is a function of the following:
Number of pneumatic and electronic controllers
Number of thermocouples
Number of switch wire runs
Number of Local Control and Local Panel indicating and controlling loops.
We take the TOTAL COUNT of all the above items, divide by 2, then multiply that result by 3.5 hours to get the total number of Instrumentation testing hours.
Keywords: testing
instrumentation
calculated
calculate
References: None |
Problem Statement: How is Equipment Allowable Stress Calculated for a given Design Temperature and in accordance with the ASME Design Code? | Solution: Aspen Economic Evaluation materials of construction utilize the ASME materials database from Aspen Exchanger Design and Rating (EDR) to determine the allowable stress at the project equipment design temperature.
Aspen EDR ASME materials database contains allowable stress data at different temperatures as well as Minimum and Maximum temperature. For a given equipment design temperature that lies within the minimum and maximum range, Economic Evaluation will interpolate within the immediate temperatures to determine the temperature related allowable stress.
As an illustration, ASME carbon steel plates for pressure vessels grade material is available as A-516 (70) in Aspen Economic Evaluation and as SA-516 Gr 70 Steel Plt in Aspen EDR ASME materials database, and has a minimum and maximum temperature of -20F and 1000F respectively. The allowable stress tables are at 50F increments, starting from 500F, and measured at ksi.
For instance at 500 F, allowable stress is 20 ksi while ar 700 F allowable stress is 18.1 ksi. To determine the allowable stress at 520 F, Aspen Economic Evaluation will interpolate between 500 and 550 F to provide the corresponding allowable stress value.
Aspen Icarus
Keywords: None
References: guide, Chapter 28, provides the a list of all the materials of construction available in Aspen Economic Evaluation, showing the different country design codes including ASME. The material allowable stress tables are not published in this reference guide but can be accessed from the Aspen EDR program. |
Problem Statement: Is a structure generated when the OPEN area type is specified? | Solution: The following examples demonstrate the steel bulk items generated for a 50 ft. x 50 ft. OPEN area. There are several reports that can be viewed to analyze the steel cost and quantities that are generated for an OPEN area type, after evaluating the project. Two reports are used to illustrate.
1) Using the Interactive Reporter run the Standard Reports| Capital Cost Reports| Direct Costs| COA Summaries| Area-COA Summary to view the Steel Account total (500+) as below:
2) Another report to view is the Detailed Bulk Material And Field Manpower Listing report in the Evaluation Report:
Additional information can also be found beginning on page 36-4 of the Icarus
Keywords: None
References: manual. |
Problem Statement: I specified the distance to MCC on the area electrical specification as 150m. Aspen Capital Cost Estimator returns a warning Specified distance To the MCC for the area may generate oversized wire/cable for the area. However, equivalent distances have been specified for the instrumentation junction boxes but no warnings as to these lengths was generated. Why? | Solution: Electrical wire is a much higher voltage than instrument wire (480/4160 for electrical wire vs 24 volts for instrument wire) and as such the Icarus cost engine does not check voltage drops on instrument signal wire. We do however calculate voltage drops on electrical wire, and have to adjust the wire size to take this into account.
The warning is to have people check whether the length specified is correct and to make sure that they understand that the wire is sized for both required amperage and the voltage drop.
Keywords: voltage, resistance, electrical, voltage, control, cable, ohm, MCC, “specified distance to MCC…”
References: None |
Problem Statement: How do I import a fluid package from Aspen Properties or Aspen Plus files in to Aspen HYSYS? | Solution: If you are working with a custom property package with Fortran code, please see article 000101265.
Since Aspen HYSYS V7.0, the fluid package along with binary interaction parameters can be imported from either Aspen PRoperties or Aspen Plus into Aspen HYSYS (Not the PFD and objects).
If you want to transfer the components, methods and parameters from Aspen Properties or Aspen Plus, please follow these steps:
Create a Aspen HYSYS new file.
Go to Fluid Package Tab.
Click Import.
From File Type Select either Aspen Properties File or Aspen Plus (bkp) file.
Notes:
When you are using Aspen Properties within Aspen HYSYS, you do NOT need Aspen Properties license. While installing Aspen HYSYS, Aspen Properties will automatically install.
Only physical property-related information (components, property methods, route changes, user-entered parameters, Chemistry) will be imported into Aspen HYSYS. Neither stream, nor block, nor flowsheet, nor pfd information will be imported.
We allow both .bkp and .aprbkp to be imported for user convenience.
Advantages
In addition to refinery & Oil gas industry, Aspen HYSYS users now can model chemicals, pharmaceuticals, polymer and electrolytic process. For example Aspen Plus has NRTL-Electrolyte as fluid package. Originally Aspen HYSYS users have to link to OLI_Electrolyte a third party software to model such system. Aspen Properties also has the updated thermodynamics from NIST database.
Keywords: Import, Fluid package, Component List, Properties Environment, Aspen Properties Databank
References: None |
Problem Statement: We have created a fluid property package/model in Aspen Plus by writing our own customized Fortran code and compiling it using aspcomp/asplink to create a .dlopt file. With this new property package, we are able to successfully run a simulation in Aspen Plus. We now wish to run a HYSYS Dynamics simulation using the same property package. When we tried creating and importing a .aspbkp file (from Aspen Plus to HYSYS), the HYSYS Dynamic program is unable to load the package and it crashes.
Is there a way to have HYSYS Dynamic read or import Fortran files/.obj files/.dll file/dlopt file or directly import the 'user-defined' fluid property package? | Solution: If you are working with a standard property method without any customization, please see article 125517-2.
There are specific ways you could adjust the settings/files to include your user-defined property package that can be later read by HYSYS.
There are twoSolution approaches -
Solution A:
Put the dll file into: C:\Program Files (x86)\AspenTech\AprSystem V10.0\Engine\Inhouse and create a file in this folder: default.dlopt, add line:
%APRSYS%\Inhouse\*.dll
Note: If you move to V11 or later, the folder directory will be under Program files i.e. C:\Program Files\AspenTech\AprSystem12.1\Engine\Inhouse.
In C:\Program Files (x86)\AspenTech\AprSystem V10.0\Engine\Xeq, find aprsysfiles.def add one line at the end :
DLOPT: ${APRSYS}\inhouse\default.dlopt
Go to C:\Program Files (x86)\AspenTech\Aspen Properties V10.0\Engine\Xeq, find the fileaprsysfiles.def and also add one line at the end:
DLOPT: ${APRSYS}\inhouse\default.dlopt
Now, Aspen Properties can use this dll subroutine.
If want to use in Aspen Plus, then in C:\Program Files (x86)\AspenTech\Aspen Plus V10.0\Engine\Xeq, find aspfiles.def and one line at the end: DLOPT:
${APRSYS}\inhouse\default.dlopt
You can now import the Aspen Properties (*.aprbkp) or Aspen PLUS (*.bkp) file into HYSYS and the user-defined fluid package will work on HYSYS.
Solution B:
Create a file with a name such as aspen.dlopt and only need to add one line in this. In this case it could be below -
C:\Program Files (x86)\AspenTech\AprSystem V10.0\Engine\Xeq \test.dll
Copy this aspen.dlopt to both locations of C:\Program Files (x86)\AspenTech\Aspen Properties V10.0\Engine\Xeq and C:\Program Files (x86)\AspenTech\Aspen Plus V10.0\Engine\Xeq
Go to C:\Program Files(x86)\AspenTech\Aspen Properties V12.1\Engine\Xeq and change
file aprfiles.def by adding one line below.
DLOPT: ${ASPTOP}\xeq\aspen.dlopt
Go to C:\Program Files (x86)\AspenTech\Aspen Plus V10.0\Engine\Xeq and change file aspfiles.def by adding one line below -
DLOPT: ${ASPTOP}\xeq\aspen.dlopt
Copy the dll file i.e. test.dll and test.lib to C:\Program Files (x86)\AspenTech\AprSystem V10.0\Engine\Xeq.
Import the Aspen Properties or PLUS file into HYSYS as mentioned onSolution A.
Note: Aspen PLUS V10 is a 32-bit program and if your dlopt works on Aspen PLUS V10, it is not required to recompile the Fortran code. However, because Aspen Plus V11 or later versions were compiled with the Intel Fortran compiler 2017 and Microsoft Visual Studio 2017 as a 64-bit program, compiled user models MUST be rebuilt as 64-bit.
Keywords: Custom Property Package, Property Package Import.
References: None |
Problem Statement: What are the common methods to import Assay Files into HYSYS Petroleum Assay? | Solution: Assays of Aspen Assay Management in PIMS and Third Party Planning CSV are two common sources that users import into HYSYS Petroleum Assay. ThisSolution is to explain the workflows and tips to import those assays.
1. Assays of Aspen Assay Management in PIMS
a. The general workflow of bringing an Assay of AAM in PIMS into HYSYS/RefSYS looks like the following:
This is for people who has already have assays in the AAM in PIMS after V8.4. Please note we only have backward compatibility and not forward compatibility (you can only import .afam file of equal or earlier versions). This is the preferred workflow comparing with the one below.
b. Another similar workflow of this can look like as follows:
This workflow is essential the same as the .afam file. The only difference you are “exporting” or saving the assay to the AAM's Custom Library, which can be imported back to AAM in HYSYS/RefSYS. Also, you will need to move the .aal file to the corresponding file location. Move the file from (C:\users\USERNAME\AppData\Local\AspenTech\Aspen Assay Management\PIMSVERSION\UserAssayLibrary.aal) to (C:\users\USERNAME\AppData\Local\AspenTech\Aspen AssayManagment\HYSYSVERSION\UserAssayLibrary.aal). The same compatibility rule apply.
2. Third Party Planning CSV
The general workflow of bringing a Third Party Planning CSV into HYSYS/RefSYS looks like the following:
The Third Party Planning CSV should be the term of use here. It is commonly called PIMS CSV, Spiral Crude CSV, or may be Haverly CSV because they all have the common format. The CSV can be generated from the Third Party software such as Spiral or Haverly. The format should consist of the tags and the values in a Comma-Separated Value (CSV) file. The converter is already in the knowledge base, which can be downloaded from this link - Aspen Assay Converter. Also, the Assay Converter has been posted on Aspen Exchange.
Important Note: In general, users should never confuse the PIMS “Excel” format (with many tabs and columns) with the “PIMS CSV” (correct name being Third Party Planning CSV). Although it is possible that you can dissect PIMS Excel file into a format that looks like the Planning CSV, this is not usually done. The two workflows described above should serve any needs to take a PIMS-associated assay into HYSYS / RefSYS.
Key words
PIMS, CSV, Import Assay, Aspen Assay Converter, RefSYS, HYSYS Petroleum Assay, Spiral, Haverly
Keywords: None
References: None |
Problem Statement: When we create user-defined entries, they are set as Engineering values by default. This means that only Engineering roles can change them through PCWS, and thus will be ReadOnly for Operators. Here, we will review how to customize this setting and show examples of removing write permissions to Engineering roles for RTE (DMC3 Builder) and ACO (DMCplus) controllers. | Solution: Changing read-write permissions and display options for user-defined entries requires following a different set of instructions, depending on if the application is RTE or ACO based.
RTE (DMC3 Builder):
To set a user-defined entry as a read-only value for a Web Engineer role, simply locate the entry on the Calculations View from DMC3 Builder. Then, on the Properties window, set the SecurityAcces -> StandardEntry value to True:
Now, save the project and redeploy the application to see the change in PCWS.
ACO (DMCplus):
To remove write permissions from the ACOEngineer role to your DMCplus user-defined entries, you must create a DMCplus.user.display.config file using Notepad and change the permissions for each entry. Here's an example of what that file needs to look like:
<?xml version=1.0 encoding=utf-8?>
<acp.product.display productName=DMCplus>
<variableRoles>
<variableRole name=UserDefined multiplicity=One>
<variableRole name=Independent multiplicity=One>
<variableRole name=Independent multiplicity=OnePlus>
<entries>
<entry name=ENTRYTEST2 value=float>
<runtime isInputToApp=True isResultFromApp=False logMessageOnChange=True changePermission=None />
<display description=ENTRYTEST2 columnHeader=ENTRYTEST2 />
</entry>
</entries>
</variableRole>
</variableRole>
</variableRole>
</variableRoles>
</acp.product.display>
In this case, we declared a user-defined entry called ENTRYTEST2, which is an Independent variable from a DMCplus controller, and we also set the changePermission parameter to None, so it is ReadOnly for both Operator and Engineer Users in PCWS.
You can copy the above code to your .config file and make the appropriate changes to fit your controller. After checking your code is ready, please save the file to this directory: C:\ProgramData\AspenTech\APC\Web Server\Products\DMCplus, and finally restart the APC Web Data Provider Service to see the changes reflected in PCWS.
You can read more details about this (and find other examples) in the 'Application Entry Override' section of the Web Server Help file, as well as in these KBs:
How can I add write access to Aspen DMCplus user entries in the Aspen Production Control Web Server?
How to remove user defined entries from the operator detail view?
Keywords: PCWS, Web Server, ACOEngineer, Engineer, Operator, Role, Write Permission, Blue / Grey Cell, User-Defined Entry, APC
References: None |
Problem Statement: How do I customize items like Freight (Domestic and Ocean) with a different percent over material for various locations? | Solution: Aspen Capital Cost Estimator doesn't have a place for you to tell our software where you want to acquire your equipment from (domestic or international) or to specify the taxes, freight, etc for different locations (Mexico, US, UK, etc).
However, you may want to use our currency.dat file to specify various currency conversion rates for different locations.
For example, if you want to specify a higher cost for an equipment item, because you are acquiring it from Mexico, you can set up a currency conversion from 1.0 ($1=$1) to say 1.5 ($1USDollar=$1.5MEX).
Then, once you set up the currency.dat file for all the conversions for all the locations that you wish, you can then select the field ?Currency unit for material cost? in each of your equipment items, and that currency conversion rate will be used for that equipment item.
You can make it even easier, that once you set up your currency.dat file with all the conversion rates you wish for various locations, you can also select a Code of Account file from the project basis tab (choose the DEFAULT.COA file), and then, if you edit this Codes Of Account file and MODIFY the definitions section part, there is a column called ?Currency Unit for Material Cost?. Here, for each code of account that we have in the entire system, you can specify which currency conversion you want to use for every single code of account. The equipment codes of account are numbers 100-299.
Another option you may want to think about is setting up multiple Contractors for each location you wish (this could be time consuming) and then you can specify your different taxes, domestic freight and oceanic freight.
You can also use SCOPE (under Contracts in the project basis) to specify which contractor who will do the Engineering, Purchase Materials, and Installation of your equipment items. You have to have all your contractors set up first, in order to select the contractor number to do the work.
Keywords: domestic, ocean, freight, taxes, indirects
References: None |
Problem Statement: Under certain conditions, model prediction for a CV suddenly resets back to the Current Measurement value. No calculation, redeployment, or CV initialization has been performed during the prediction reset. | Solution: CV model prediction resets can happen when:
Predictions are initialized
Individual CV predictions are initialized (manually or triggered by calculations)
Accumulated Prediction Error reaches or exceeds the maximum limit of 9000
The CV has bad status
The CV has bad value
If none of the above caused the CV prediction to reset, the only other option is a hard-coded CV prediction reset that happens once per month. This periodic reset is a built-in feature that prevents the biased prediction from getting far out of sync with the current operating conditions, so it cannot be disabled. This reset will also be more noticeable when there is a large drift from the current measurement.
Keywords: Model prediction, CV reset, initialization, PCWS, Process Explorer, APC, DMCplus, DMC3
References: None |
Problem Statement: By default, The Web Data Provider service can host up to 20000 logging messages for the Production Control Web Server. However, we can modify the amount of PCWS log messages in different ways to store more logging information or save space on the server. | Solution: You can find the instructions on how to modify the number of PCWS messages in the maxMessagesTotal section of the Web Server Help:
In summary, to change the maxMessagesTotal parameter:
Use the Windows Services applet to stop the Aspen APC Web Provider Data Service on the Web server.
Open Notepad with admin rights and edit the WebDataProviderSvc.exe.config file, located in C:\Program Files (x86)\AspenTech\APC\Web Server\bin
Within the .exe.config file, locate maxMessagesTotal, and adjust the appropriate element line (<add key=maxMessagesTotal value=20000/>) so that it includes the new desired value.
Save changes to the .exe.config file, then restart all applicable services on the current server.
You can also change the maximum number of messages for a specific APC Online product. These steps are performed on the Online server:
For Aspen APC ACO (DMCplus), Inferential Qualities (IQ), and Nonlinear Controller applications:
In Windows Services, stop the Aspen APC Message Log Service.
Using Notepad, modify the maxMessagesInCache value for the appropriate product in the AspenTech.APC.MessageLog.Service.config file, located in C:\ProgramData\AspenTech\APC\Online\sys\etc
Save the file and then restart the Aspen APC Message Log Service.
This service is set to Manual Startup by default, but you can change it to Automatic if you want.
For Aspen APC RTE (DMC3 Builder) applications:
In Windows Services, stop the AspenTech Production Control RTE Service.
Use the Windows Registry Editor to create or locate the following key. Then edit it for the desired value:
HKLM/Software/WOW6432Node/AspenTech/RTE/<version_number>/MaxMessagesInCache
Key type: REG_SZ (string value)
Key value format: integer string (example: 20000)
Restart the AspenTech Production Control RTE Service on the current server.
To read more about the Message Log Service, you can go to the very end of the Customizing and Troubleshooting Data Services section in the same Help file:
Finally, starting on v12, there is a new Logging tab in the Configure Online Server application. Here, you can manage the AspenTech APC IO Logging Service. This logging feature can record (in a database) messages produced by or involving the DMC3 online applications hosted on the server, IO source communications, and the RTE Service.
Instructions for this new logging feature can be found by pressing the Help button.
See also KB article: “How to enable and use RTE Controller logging in V12 and later ”
Keywords: PCWS, logs, logging service, messages, APC, DMC3, DMCplus
References: None |
Problem Statement: How do I create a Scheduling Objective in APS? | Solution: If you're using equations to calculate values, then the necessary calculations have to be created (Model-> Calculations)
To create a new objective:
1. From the Event interface, click Model | Scheduling Objective to display the Schedulng Objectives dialog box.
Note: The list of existing objectives appears in the left side of the dialog box.
2. Click the New button.
3. In the Objective Details Tag field, modify the default tag OSn, to one that you want to assign to the objective.
4. In the Description field, enter a description.
5. In the Objective Type field select one of the following options:
Instantaneous – The instantaneous value is the value at the event boundary.
Cumulative – The cumulative value is the aggregated value of the instantaneous values from the start date of the model onward.
6. Click New Schedule Objective Value to add a new value row to the Schedule Objective Value table.
7. Add an initial Value in addition to Min and Max values to define the objective.
8. Add a calculation for values if desired.
a. Click in a Calc checkbox to create a calculation or to add an existing calculation.
b. Map the result variable to a flow or a throughput.
c. Once the calculation is added, click OK to close the Calculations dialog box.
9. Click OK on the Schedule Objectives dialog box to save your changes in memory and exit the dialog box. Your changes will not appear in the database until you save your model by clicking on the toolbar.
Keywords: None
References: None |
Problem Statement: Why am I not seeing any manhours for vessels with large volume? For example, I have a spherical tank of 2,100,100 gallon, that does not generate any manhours, but the manhours do get generated for smaller sizes (1000 gallons). Is this a bug or is there a reason for this? | Solution: No, this is not a bug and there is a good reason for this.
With very large volume tanks, the tank is being built on site because it is a subcontracted field-erected tank, thus no man-hours for a subcontracted field erected tank.
The break point is the Shop Fabrication Diameter in the Equipment Design basis screen. By default it is 14.5 FT.
So, when the diameter of the vessel goes over the Equipment Design basis default of 14.5 feet, then the vessel automatically switches to be a subcontracted field-erected tank, and this the tank is being built on site. All man-hours are included in the quote from the subcontractor.
All field erection costs (labor costs) are wrapped into material price; and for those users who use IPM, there is no subcontract, so only a labor hours field is listed..with no labor cost.
Using Aspen Capital Cost Estimator, you can control what your ?Design Basis Specs? at the Project Basis Level were set at for ?Shop Fabrication Equipment?....this can control whether the item is field or shop erected, depending on what your inputs are.
Keywords: manhours, man hour, man hours, vessels, tank, labor, cost
References: None |
Problem Statement: What percent full of water is used for foundation design for vessels? | Solution: In designing foundations, Aspen Capital Cost Estimator assumes that a vessel is completely filled (100%) with water for foundations without piling. The amount is reduced to half full (50%) for foundations with piling.
The foundation is designed for weight (vessel weight + weight of water) and horizontal load due to wind or seismic loads. In general, horizontal loading will be the governing factor for vessels.
Vessels receive an octagonal foundation by default. In many cases, the size of the foundation will not vary based on the water volume. However, the octagonal foundation sizing will change based upon the water volume in the following situations.
Very poor soil
Vessel is designed for seismic only
Wind speed is set to 0 (In such a case, the foundation is designed for weight alone.)
Keywords: foundation
water
fluid
piling
pile
References: None |
Problem Statement: How can I specify demolition as a specific indirect cost instead of calculating it as specific direct cost items. | Solution: On the indirects form in the Aspen In-Plant Cost Estimator Project Basis View or on the contractor definition form in Aspen Capital Cost Estimator enter the demolition cost in the Special Indirect Item 1 field.
This will then report the demolition cost under the project indirects.
If required for reporting purposes, one can also edit the COA definitions and rename COA 24 from Special Indirect Item 1 to Demolition Indirects .
Keywords: code of accounts, demo, demolition, indirects, indirect, COA
References: None |
Problem Statement: Using the Above Grade or Buried Pipe [Piping Plant Bulk] model, I am seeing more valves and fittings being generated and I cannot figure out why. | Solution: The reason why one would see additional valves and fittings is when the Pipe Routing Type field is used. When this field is used, the external fitting specs file will be used.
If using Aspen Capital Cost Estimator, the Pipe Routing Type is explained in the Icarus reference manual, as indicated by the Properties window when you place your cursor in the Pipe routing type field. (See page 18-9 of the Icarus reference manual).
You would specify the Pipe routing type when you are using the external fitting specs file (specified at the design basis level). Once a pipe routing type is specified, the system uses the data in the external fitting specs file for the valve rules, fitting rules, horizontal allowance, vertical allowance and the diameter size definition. Pipe routing type is meant to refer to the location of the piping. However, the above descriptions are for reference only; using the fitting specs file, the user controls the context of the rules and can assign meaning at his discretion.
Keywords: pipe fitting, pipe routing type, routing, FITSPC, external fitting, increased valves, increased fittings, more valves, more fittings
References: None |
Problem Statement: In the installation details reports, costs labeled Project / Area - GRADE appear. The following example comes from the Area - Installation Details standard report (Capital Cost Reports | Direct Costs | Installation Details | Area - Installation Details).
What is the source of these Project / Area - GRADE costs? | Solution: The area-level electrical costs are generated based upon the model used by Aspen Capital Cost Estimator. The system automatically generates the following bulk quantities for the default GRADE area type.
Perimeter lights every 50 feet.
Lights are supplied from 1 or more area panelboards, supplied from a disconnect switch in the MCC for this area.
A ground grid the length of the perimeter.
A concrete pad will be developed based on the dimensions of the area if the PAD option is selected.
To disable cost for the perimeter lighting and ground grid, go to the area electrical specs (in the Project View, double-click on the area or right-click the area and select modify). In the Area Electrical form, area lighting and grounding can be removed by selecting - in the Lights and grounding and Equipment grounding fields.
For additional information on area bulks, see the Introduction to Areas section of Chapter 36 in the Icarus
Keywords: project
area
electrical
bulks
References: |
Problem Statement: Why would Freight and Taxes not be calculated using relocation in Decision Analyzer? | Solution: Depending on the relocation country specified, there may be no taxes and freight calculated. For example, a relocation to Egypt, will show no taxes calculated for the first year.
If you create a sample project, and run the estimate on the project, look at the CASHFOW.ICS report, and look at the line item called 'Earnings Before Taxes'.
If there is a NEGATIVE number where it says 'Earnings Before Taxes', taxes will be set to 0. Taxes are not generated on negative earnings.
If you look in the other columns for that same line, which represent years 1, 2, 3, and so on?..
If the 'Earnings Before Taxes' is a positive number, then taxes will be calculated.
When starting a new project, whether it be METRIC or IP, the original total project cost is a positive number, and this is because the project location is set to the default METRIC or IP location.
An original METRIC project is based on Northwest UK, and a default IP project is based on US Gulf Coast. (Please refer to the Aspen Icarus
Keywords: relocation, location, freight, tax, taxes, analyzer
References: Manual, Chapter 36, pages 36-38 to 36-40. This section and Table T-1 shows you the 'System Default Values Affected by Selection of Country Base - Primary Effect of Selected Country Base.
When you change the project location to a new location, let's say for example, Cairo, Egypt, the new taxes, permits, etc. percentages are then used.
The calculation of Freight costs in Aspen Icarus Process Evaluator cannot be controlled by the user. Taxes and permits are not 'calculated'. Taxes and permits are set percentages, which come from the v2004 Richardson?s Cost Location data. If using Aspen Capital Cost Estimator, the Freight costs can be controlled using the Contractor form. One can also control taxes in this contractor form as well. |
Problem Statement: After using the Aspen Capital Cost Estimator Piping and Instrumentation Drawings (P&ID) editor for a component, I noticed that my component summary costs are different for that component. Why would this be happening? | Solution: There are two types of Piping and Instrumentation Drawings (P&IDs) associated with each piece of equipment.
A Fully Instrumented P&ID
Standard PID.
What happens when choosing to use a graphical P&ID
When you choose to open the graphical P&ID (or default P&ID) for a component it will be the Fully instrumented P&ID drawing. This is even true if you have Standard (STD) selected in the Project Basis tab under Instrumentation section.
To change the graphical P&ID, open the component and then from the drop down P&ID menu, choose Select and Open Alternate. Find the equipment type folder you are working with. Choose the appropriate P&ID (whether it be Standard or Full). All Full graphical P&IDs have the number '6' as a beginning. i.e. drawing #16 is the Standard drawing for the Centrifugal Pump, and drawing #616 is the fully instrumented drawing for the Centrifugal Pump.
If you wish to only open the Standard graphical P&ID, then you must select to Open Alternate and select the Standard Graphical P&ID.
What happens when not using a graphical P&ID
If there is no graphical P&ID drawing being used, then your setting in the Project Basis tab under Instrumentation will be used by the cost engine as the default drawing number....either the STD (Standard) or FULL (Fully Instrumented) drawing. You can tell what drawing is being used in the Item Details report you generate.
You can either change the instrumentation in the Project Basis View to FULL (full or heavy instrumented) or change the P&ID associated to the component to Standard (STD).
To change the instrumentation in the Project Basis View:
Project Basis View | Instrumentation Specs | Piping and Instrumentation Design (FULL)
Keywords: P&ID, FULL, instrumentation, component, costs, piping and instrumentation, piping, instrumented, drawing, pid, PID, instrument, graphical
References: None |
Problem Statement: How can I modify the Equipment Rental information generated for my Project? | Solution: ACCE has the capability of automatically adding Construction Equipment required to install the elements of the project.
However, this predicted equipment may not exactly represent the real values required for the project, therefore, the user has the ability of overwriting these system calculations.
In order to control the Construction Equipment, the user can access the Equipment Rental section on the Project Basis view form. This allows the user to either Add, Delete or Change the current Equipment Rental from the project.
In order to apply modifications fo this form follow these recommendations:
1. Always start by reading the system generated Equipment Rental, this will allow you to track the current results and verify if any equipment is missing and should be added, should be removed, or the price/days for one should be modified. Use either the CCP report or the Interactive reports for this.
2. Use the Add button to add multiple columns, each column allows you to apply a modification to a specific Equipment Rental.
3. The Item description allows you to identify the action each column applies (i.e. Include Ambulance).
4. The Contractor number allow you to control if the new changes will be applied to a single or multiple contractors (i.e. Costs assigned to Owner).
5. The Rental equipment number should be filled with the assistance of the Icarus
Keywords: Equipment, Rental, Database, Hprojid, Aprojid, Modify, Change
References: guide. This number allows the program to identify what Equipment will be Added/Modified. (i.e. 21, for ambulance)
- Go to Help | Documentation | Icarus Reference, switch to Chapter 32 (Construction Equipment).
- Use the table in the Construction Equipment chapter, the first column Equip No. shows the number that identifies each equipment.
6. Rental Action Code allows you to select if this is a new Equipment category will be Added (Add), if an Equipment will be removed from the project (Delete) or if the calculated days/ price will be adjusted (Change)
7. Use the Rental days required and Monthly rental rate fields to adjust ether the required days and the monthly price of the equipment in the column. |
Problem Statement: What is the new Scheduling Objectives Feature in APS V14? | Solution: The Scheduling Objectives has been added in APS V14 that allows you to define scheduling KPI's that can be trended as functions of time.
You can create equations that include entities from your model to develop cost, revenue, or margin calculations, to produce values that can be tracked during simulation and plotted as Schedule Objective trends. An additional benefit of Schedule Objectives is the ability to publish the results of these objectives in reports.
Keywords: None
References: None |
Problem Statement: How and when does the Aspen Capital Cost Estimator install Ladders and Platforms? Is there some sort of calculation where the length has to be a certain size? | Solution: Systems with automatic installation bulks generate platforms and towers. A tower begins with 1 360 degree platform for the first 25 feet [7.6M], or fraction thereof, of exposed height, plus 1 90 degree platform for every additional 25 feet [7.6M], or fraction thereof, of exposed height.
If the tower is in a structure, the system subtracts the height of the structure from the total tower height (Bottom T-T height + Skirt Height) to calculate the total exposed height.
Keywords: ladder
platform
ladders
platforms
References: None |
Problem Statement: Is it possible to add in drawings for one contractor who has two separate drawing phases? | Solution: The place to access drawings are under the Project Basis menu called PROJECT | ENGINEERING WORKFORCE. Once select the option called Drawing types.
Here is where you can select the Phase, Drawing number, Class, and Discipline.
Keywords: drawing
References: None |
Problem Statement: There are times when you may want to apply a fixed cost for all of Engineering. This | Solution: Document covers how to apply a fixed cost for all of Engineering.
Solution
Create or open a project in Aspen Capital Cost Estimator.
To make changes to Engineering, go to the Project Basis tab, and right-click on Engineering Workforce | By Phase.
In this Engineering adjustments, by phase form, is where you will need to set all phases to zero dollars and zero man-hours.
Enter an asterisk (*) in the field called Engineering Phase.
Enter in a zero (0) for both Man-hours and Cost.
Click APPLY to add in these changes. You should see those changes appear in the box above the form.
Next, enter a D into the field called Engineering Phase.
Next, enter the lump amount of Man-hours into the Man-hours field.
Finally, enter the lump sum of dollars for the Cost field.
Click APPLY to add in these changes. You should see these changes appear as well in the box above the form.
Where will I see these Engineering lump sum dollar and man-hour changes in my reports?
You will see these lump sum changes in the report called CONTRACT DATA SHEET (for each contract number) under the headings:
MISCELLANEOUS
DETAIL ENGINEERING MANHOURS
DETAIL ENGINEERING COST
Keywords: fixed
engineering
References: None |
Problem Statement: I was running the Indirect and Non-Field Summary report, and noticed that Indirect Report Acct 85 - Field Const. Supervision is a very large number! Where can I edit the Field Construction Supervision dollar figure in the project? | Solution: The Indirect Report Acct 85 - Field Const. Supervision can be modified in the Standard Basis under the Engineering Workforce either by Phase or by discipline.
Keywords: supervision
85
indirect
References: None |
Problem Statement: What percent of Direct Field Labor (DFL) does Aspen Capital Cost Estimator use for scaffolding? | Solution: Scaffolding can either be entered as a fixed number or as a percentage of DFL. If neither is entered, the default values are (according to country):
US 3%
UK 7%
EU 7%
JP 3%
Scaffolding costs or percentages can be entered under:
Project Basis View | Contracts | Contractors | (contractor name) | Scaffolding cost
or
Project Basis View | Contracts | Contractors | (contractor name) | Scaffolding percent DFL (Direct Field Labor).
Keywords: Scaffolding, indirects, contracts, contractors, contract, direct field labor, DFL
References: None |
Problem Statement: What is the difference between the X and the V option, on the volumetric model button on the Civil installation bulk? Both have the same message (X: Default civil model + any specified data) when I specify either the X or the V, I get the same results.
Is it supposed to work this way? | Solution: There is no difference between the X and V for the Volumetric model.
Keywords: volumetric, X, V, model, civil, requirement
References: None |
Problem Statement: What is the source of small block (SM BLOCK) PROJECT/AREA - AREA costs in Aspen Capital Cost Estimator detail reports?
For example, the following entries appear as PROJECT/AREA - AREA costs for the Absorption Area in the Gas Plant sample project. The costs come from the Area - Installation Details standard report generated on a Kbase 2004.0.6 system. | Solution: The PROJECT/AREA - AREA foundation items (SM BLOCK) in each area are based upon the miscellaneous concrete allowance field in the civil specs. The misccellaneous concrete allowance can be modified either at the area level (right-click on the area in the Project View, select Modify, and select Area Civil) or at the project level (edit Civil/Steel Specs in the Project Basis View).
The default miscellaneous concrete allowance is 15%. The SM BLOCK items will be removed if 0% is specified.
Keywords: civil
concrete
foundation
area
References: None |
Problem Statement: How do I split an event by Quantity? | Solution: Starting from V12, you can now split an event by time or quantity. The Split short-cut command has been renamed to Split by Time and a new split option, Split by Quant has been added to provide enhanced splitting capability.
You can use the Split by Quantity dialog box to enter the amount at which to split the event. The amount entered is included to the left of where the event is split. The remaining amount goes to the new event to the right.
To access this dialog box:
Access the appropriate event screen from the Events Screen node of the navigation tree Events tab.
Right-click the desired event and click Split by Quantity to display the Split by Quantity dialog box. The current event quantity is visible.
Enter the value you wish to use to determine where to split the event in the Left Event Quant field.
Keywords: None
References: None |
Problem Statement: How do I include the original source event information in the publish tables for the Pipeline fill events | Solution: Two new columns have been added to the pipeline publish tables to address the issue where pipeline line fill events do not capture/publish the original source event information. Note that fill information inside a pipeline does not contain the original event information. It is only the new batches that are pushed into the pipeline that will carry the information when published. Event Shipment and Event Batch IDs associated with the source can be viewed from both the crude and product Beginning Line Fill dialog boxes, if this information is available.
Be sure to run DB update to add EVENT_SHIPMENTID and EVENT_BATCH columns to the following tables as well as to make the additional table changes:
_PL_PIPELINE
_ZPL_PIPELINE
An additional table has been added to store source information, _EVENTS_BATCHES_MBO and an additional column, EVENT_SHIPMENTID, has been added to the _EVENTS_MBO table.
Keywords: None
References: None |
Problem Statement: Hitting Min/Max component tank limits is not desirable in refinery operations because it limits flexibility in online blend control and blend execution systems.
Blend recipes are often adjusted to build or draft inventories if too low or too high and during that process, de-optimizes the schedule, thus ending up in lost credits. This requires additional work because of inventory management and reporting work processes. Prior to V12, MBO drove tank inventories to either Min or Max limits and schedulers needed to micromanage the component tank trend limits every time they ran MBO, depending upon the component starting inventory and component rundowns. This was a manual, tedious, and iterative process. | Solution: New functionality was added in V12 to be applied only to component tanks. This feature allows you to define a deviation penalty when a component tank's inventory moves outside of a volume range. The range is defined by a minimum/maximum Target value that you provide as soft bounds. These limits are not hard bounds in tank or trend limits. You can define the penalty value from Blend Settings tab.
To implement this functionality, do the following:
On the Settings dialog box Blending tab, set the Target Inventory | Deviation penalty. This is the penalty that will be applied if component tank volumes exceed tank limits. This value syncs with the TARGET_INV_DEVIATION_PENALTY keyword. The higher the limit, the harder the target optimizer will try to keep the inventory level within the inventory target limits.
Select a component tank, and set desired Target Minimum and Tank Maximum soft bound values from the Tank dialog box.
Notes:
DBUpdate should be run prior to using this feature to add the appropriate columns to the TANKS table.
MBO enforces the defined penalty if values go above the target max or below the target min. These are not hard bounds.
The cost incurred by all tanks across the whole campaign will be included in the ECONOMIC SUMMARY section of the html report on the Inventory Target Deviation Cost row.
Target inventory limits can be trended.
Violations to the Target min or max will not be reported as infeasibilities.
Cautions:
In some cases:
The optimizer will not have enough degrees of freedom to keep the inventory within the target limits. In this case, the cost of violating these limits will keep increasing proportionally to any increases in the magnitude of the penalty.
Other penalty terms will be higher than the penalty for violating target limits; in these cases the inventory will violate the target limits even when it appears to have the possibility not to violate them.
Keywords: None
References: None |
Problem Statement: For an AUP model which has Russian descriptions for materials/capacities etc., when user runs the model in machine with regional Russian settings, AUP model HTML reports shows ??? for Russian characters. | Solution: The ? characters within the reports are a result of the Supervisor Service in AU running on a NETWORK user by default. Regional settings are in English by default in the NETWORK, whereas the local User settings have regional settings corresponding to the VM. To fix the issue, follow these steps.
Have a working build of AU and open Windows Services
Find the Service labeled Aspen Unified Agent Supervisor Service. If it is running, ensure that AU is closed, right-click the Service, and click Stop
3. Right-click the Service again and click Properties -> Log On
4. Enter the local user name and password, and click Apply
5. Back on the Supervisor Service, check that the Log On As column is the local user name
6. Open AUP and execute the model.
7. Open the FullSolution report following case execution. It should have the Cyrillic characters seen in the UI.
Keywords: None
References: None |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.