您的位置:首页 > 数据库 > Oracle

Optimizing Item Import Performance in Oracle Product Hub/Inventory

2014-10-08 21:37 183 查看


APPLIES TO:

Oracle Product Hub - Version 12.1.1 to 12.1.1 [Release 12.1]

Oracle Inventory Management - Version 12.1.1 to 12.1.1 [Release 12.1]

Oracle Item Master - Version 12.0.6 to 12.0.6 [Release 12]

Information in this document applies to any platform.


ABSTRACT

This document discusses the Import Catalog Items (ICI) program and how to achieve optimal performance for high volumes. It lists a number of functional areas that you can tune for optimal scalability. ICC processing and performance depends on which features
you use, as well as how you use and configure your system. This document identifies opportunities for setting up your system to scale for high volumes. It also lists and briefly discusses other item bulk import options available in Oracle Product Hub (also
known as Oracle Product Information Management Data Hub (PIMDH)).

Performance of ICI varies depending on what features you use. Performance also depends on system configuration. It is not possible to make absolute hardware recommendations because hardware requirements vary based on the features used and the processing volume.
Hardware requirements also vary depending on the time window allowed for processing. For instance, more powerful hardware is required to process 10,000 items in one hour than in three hours.

Although it is not possible to predict throughput for a given hardware configuration, it is possible to identify throughput observed with a specific hardware configuration. Refer to Performance Results section of this document for information about throughput
observed using a specific set of features and data. The data is not a benchmark.


HISTORY

Author: Oracle Product Hub Development Team

Create Date
30-Jul-2010

Update Date 19-Oct-2010

Expire Date Not Applicable


DETAILS

Bulk Import Capabilities in PIMDH

Product Hub uses the following methods for importing items in bulk:

Import Items (INCOIN)

Import Items is the core inventory program provided to upload items in bulk from interface tables. It is available to Inventory and PIMDH users. This program is for creating and/or updating basic items in bulk, without any of the associated PIMDH features such
as UDAs. You must load data into the items/revisions interface table before launching this program.
Import Catalog Items (ICI)

This program is a wrapper on top of the Import Items program and supports bulk load of additional PIMDH child entities, such as user defined attributes (UDAs) and associations. Details of the additional entities supported are discussed in the next section,
Functional Background and Process Flow for ICI. As with the Import Items program, you must load data into the interface table before launching this program.
Excel Upload (EGOIJAVA)

This feature enables business end users to enter data into Excel and upload it to PIMDH. The data from Excel is automatically loaded into an interface table, then the “EGO Spreadsheet Java Concurrent Program” launches and processes the interface records. This
program internally calls the same APIs called by Import Catalog Items, so the item import performance after uploading the data from Excel into interface tables is the same as the ICI program. Most of the discussion in this document also applies to this part
of the Excel upload process. Time required for loading data from Excel into the interface tables is additional.

In addition to loading items and their child entities, this program can also create change orders, new item requests and structures, all of which are out of scope of this document.
Batch Import using Import Workbench

Access the Import Workbench through the Product Information Management Data Librarian responsibility. In the Import Workbench, you can create batches, then add items to the batch, match the items against production to eliminate duplicates, and submit the batch
for loading the items into production tables. The Import Workbench internally calls the same EGOIJAVA program mentioned above to load items and their child entities.

Functional Background and Process Flow for ICI

This document specifically focuses on the functionality and performance of the Import Catalog Items program. As noted in the previous section, the Excel Upload and Import Workbench features are wrappers on top of this program, so the discussion also applies
to these programs


Functional Background

The Import Catalog Items program can import the following entities:

Items (both master and organization items)
Revisions
Item Intersections

Suppliers
Supplier Site
Supplier Site Store

User Defined Attributes at all data levels
Item People
Approved Manufacturer Lists
Source System Cross References

While validating and importing the above entities, the program accomplishes the following tasks based on your setup:

Defaulting of UDAs.
Applying templates.
Copying information from the source item.
Generating Item Number/Description using designated functions where applicable.
Applying Rules (assignment and validation rules).
Checking for change policies and creating change orders/NIRs where required.


Process Flow

The following list shows a sequence for data processing:

Process standard/style items

Preprocess interface records.
Run Rules.
Generate Item Number/Description if function generation is specified.
Process the item records to create items.
Call the Change Management pre-process API.
Call APIs to process child entities (such as UDAs, Intersections, AML, People).
Call the Change Management post-process API.

Process SKU items (same flow as step 1)

Preprocess interface records.
Run Rules.
Generate Item Number/Description if function generation is specified.
Process the items records to create items.
Call the Change Management pre-process API.
Call APIs to process child entities (such as UDAs, Intersections, AML, People).
Call the Change Management post-process API.

Architecturally, the Import Catalog Items program is divided into smaller programs to handle items and each item's entities. The processing is also split across the Java and PL/SQL layers. The validation and processing of entities such as items and UDAs
are mainly performed in PL/SQL while supporting activities like number/description generation and Rules are performed in the Java layer. The Java layer program processes each interface record individually due to the functional nature of Rules and Function
Generation.

The PL/SQL programs mainly handle the corresponding interface records in bulk rather than processing one record at a time. However, certain validations and processing, due to their functional nature, are performed on individual interface records or in a
smaller chunk of records.

Depending on the availability of resources, customers can submit multiple batches in parallel. Finding the optimal number of threads and the optimal batch size for processing in each thread on your system is an iterative process, which is described in the
next section.

Factors Affecting the Performance of Import Catalog Items

The performance of the Import Catalog Items program depends on numerous factors. This section discusses each of these factors in detail.

Number of items: Larger batches take a longer time to complete. However, having too large a batch size puts excessive memory requirements on the program, especially in cases involving rules and change management. The ideal batch size depends
on resource availability.
Transaction Type: You can submit batches with CREATE or UPDATE transaction types (or a combination of both). Performance varies with each transaction type. CREATE transactions have additional processing requirements, like creating revisions
and applying default values.
Number of UDAs Per Item: Each user defined attribute value you specify is represented as a separate row in the UDA interface table. Therefore, the volume of this table for a batch could potentially grow very large depending on the number
of items in the batch and the number of UDAs per item.

In addition to the UDAs provided by the user, rows are inserted into this interface table from templates and UDA defaults. Also, more UDAs may be added depending on the rules configuration for the Item Catalog Category/Attribute Groups participating in this
batch.
Associations and other child entities: Adding supplier associations, role assignments, AMLs and cross references in the same batch as the item creation batch requires additional processing and more time for the program to complete. Supplier
associations also add more UDAs to the interface table (and further processing time) depending on the setup.
Rules: You can define rules at the Attribute Group level or the Item Catalog Category level. Performance is affected by the number of rules, number of attributes participating in these rules, and the types of rules. As described earlier,
assignment rules add more records to the UDA interface table, causing additional processing time.

Rules processing is performed in a java program, unlike the rest of the import program which is a PL/SQL program. To avoid multiple database trips, all information required to process these rules is cached into the memory in chunks of 100 items at a time. This
makes rules a memory intensive process. If more memory is needed for this process, consider adding the following parameters to the concurrent program definition form: -Xms<memorysize>M -Xmx<memorysize>M. Also, review Document
362851.1, JVM: Guidelines to setup the Java Virtual Machine in Apps Ebusiness Suite 11i and R12, to ensure optimal performance.
Change Management requirements: If a batch is created using the Import Workbench function, you can select the change management option to add changes to a change order or add new items to a new item request. This is a java program that is
memory intensive and causes additional processing time.
Type of Items (Style/SKU): Due to the functional nature of SKU items, they are created after the Style items. In batches where SKU items coexist along with Style/Standard items, certain sections of the program are repetitively called, first
to process style/standard items, then again to process SKU items. Including SKU items in same batch as Style items also increases the number of records for processing in multiple interface tables, since UDAs, organizations, associations, and so on are defaulted
from the Style item to the SKU item.
Items for Child Organizations: Assigning items to child organizations requires the same amount of processing time as creating an item in the child organizations. If you create 1000 items and assign each item to two child organizations in
the same batch, the batch essentially creates 3000 items in the items table. In addition, extra processing, such as validating for master controlled attributes, copying master controlled attributes from master records, and so on, increases the processing time
for child organization item records. Consider these factors while measuring performance of batches.
Number of Instances of this Program Running: Under ideal circumstances, the time expected for a single ICI program to run should remain the same while running more than one ICI process in parallel. However, this is seldom the case since
each program shares the same resources with the other programs running in parallel. CPU and memory utilization determine the number of parallel threads. The CPU utilization ideally never exceeds 80% at any point when all threads are running.

In addition, despite best efforts to tune all SQLs, there still are a few queries that perform a Full Table Scan on the interface tables, mainly due to the data distribution. These queries degrade while running more threads of the program in parallel, since
each program scans through all records of the other batches processing at the same time.
Hardware configuration: The throughput of the ICI program greatly depends on hardware capabilities such as number of processors and the amount of memory in the configuration of the various tiers (dB Tier, Middle Tier, Concurrent Manager,
and so on). If your installation uses Real Application Clusters (RAC) with multiple nodes for each tier, Oracle recommends bringing down all nodes and retaining only one node per tier for best performance during the initial data load. Additionally, refer to
the "Concurrent Processing on RAC Information Center" section in Document 1304305.1, Concurrent Processing - Product Information Center (PIC), and ensure that
all dB parameters are optimized and the same for all nodes.

The above factors affect the performance of the item import programs. Next, we will discuss how to measure and, if required, tune the performance of the program.

Measuring and Tuning the Performance of the ICI Program

The process of tuning the performance of the ICI program involves multiple steps:

Ensuring the system configuration prerequisites are met.
Measuring the current performance of the ICI program.
Tuning and measuring the post-tuning performance.

The following sections discuss each of the above steps in detail.


System Configuration Pre-requirements

Before using PIMDH, ensure system configurations are performed as specified in the following My Oracle Support documents. Some of these documents are specific to the Linux x86-64 bit platform, but documents for other platforms as well are available in My
Oracle Support.


396009.1 - Database Initialization Parameters for Oracle Applications Release 12

761566.1 - Oracle Applications Installation and Upgrade Notes (Linux X86-64 bit)

361323.1 - HugePages on Linux

859397.1 - Integrating Oracle E-Business Suite Release 12 with Oracle Database Vault 11.1.0.7 (if you use 11gR2, review note 823587.1 instead)

466649.1 - Using Oracle 11g Release 1 (11.1.0.7) Real Application Clusters and Automatic Storage Management with Oracle E-Business Suite Release 12


The list of documents listed above is not a comprehensive list, and it is only indicative for a specific configuration. If you use a different platform or configuration, look for other similar documents that are relevant to your configuration

Verify the following init.ora parameters in your environment since they can impact the performance of batch processes.


"optimizer_dynamic_sampling" = 2

"optimizer_index_cost_adj" = 100

"_optim_peek_user_binds" = false

"_optimizer_extended_cursor_sharing" = NONE

"_optimizer_extended_cursor_sharing_rel" = NONE


Gather statistics regularly on your INV/EGO/APPLSYS schema using the Gather Schema Statistics program.


Measuring the Current Performance of the ICI Program

This section describes the steps needed to measure performance of the Import Catalog Items program.

Ensure the test environment is comparable in configuration to your final production instance configuration. Ensure the instance is isolated and not subjected to varying loads (for example: accessed by other users, other programs) during the time of the
test.
Ensure the performance measurement run is not the first run after a bounce of the database, middle tier, or concurrent manager. If you have bounced any of these components, perform a warm up run for a small batch of items. This eliminates overhead times
taken for activities such as class loading and query parsing during the actual performance measurement runs.
One of the factors impacting performance is the batch size. Ensure that you measure performance for a batch size most commonly used by your business process. However, Oracle recommends limiting batch sizes to functionally manageable limits to facilitate
error reviews and corrections due to warnings received. A functionally manageable size depends on the setup, which dictates the number of records in the interface table. For example, if you have 5 UDAs per item, you would have 25,000 records in the UDA interface
tables for a batch of 5,000 items. However, if you had 100 UDAs per item, a 5,000 item batch results in 500,000 records in the UDA interface table.
Ensure that statistics are gathered for the EGO, INV and APPLSYS schema regularly. If you have performed any major data load or data update activities recently, ensure that the statistics are collected on these schema again before measuring performance.
Ensure that the following interface tables are truncated and do not contain any data:

MTL_SYSTEM_ITEMS_INTERFACE
MTL_ITEM_REVISIONS_INTERFACE
MTL_INTERFACE_ERRORS
EGO_ITM_USR_ATTR_INTRFC
EGO_ITEM_ASSOCIATIONS_INTF
EGO_AML_INTF
EGO_ITEM_PEOPLE_INTF
EGO_ BULKLOAD_INTF

Ensure tracing is disabled.

Set the profile "Initialization SQL Statement - Custom" to Null.
Clear the “Enable Trace” check box in the Concurrent Programs form for Import Catalog Items.

Ensure statistics gathering is disabled .

Set the profile "EGO: Enable Stats Collection" to No.
Set the profile "EGO: Stats Collection Threshold" to 999999999.

Ensure logging is disabled.

Set the profile “INV:Debug Level” to Null.
Set the profile “FND: Debug Log Enabled” to No.

Load data into relevant interface tables. Ensure the quality of data you are using for the performance measurement is good. If too many records fail the program validation, the program performance degrades. Preferably, measure performance on a batch of
items that all process successfully. Also, ensure the data loaded is representative of your production data. For example, if you expect your production items to have an average of 100 UDAs each, load around the same number of UDAs per item during this measurement
run.

Keep the setup similar to your production setup. For example, if you plan to use 10 business rules involving 20 attributes in production, set up similar rules in the performance measurement run. In addition, follow these tips while setting up:

If rules functionality is enabled, then set up to run assignment rules before validation rules.
For bigger data loads, avoid using approval type rules since approvals significantly impact performance.
Consider creating separate batches for items with different properties (for example, items with a different set of attribute groups or rules written using different attribute groups).
For simple validations and assignments, perform these tasks before loading the data into interface tables instead of running the Rules framework during import.

Submit the Import Catalog Items program to process the data.
Once the program completes, ensure there are no errors or warnings. Note the start and end time of the program from the Concurrent Program log to get the elapsed time of the program.

The performance of multiple batches running in parallel is different from the performance measured for a single batch. To measure the performance of multiple batches running in parallel, follow the same steps outlined above, except load data for multiple
batches in step 8 and launch a program for each batch in step 9.

If the performance of the ICI program does not meet your business requirements, check the hardware configuration. If the performance is still unsatisfactory, try to tune it using the steps provided in the next section.


Tuning the Performance of ICI Program

Ensure the following patches are applied to your instance:

Patch 9811599:R12.EGO.C
Patch 9728386:R12.EGO.C
Patch 13434831:R12.EGO.C
Patch 13414358:R12.EGO.C
Patch 17207868:R12.EGO.C

In addition to tuning, performance predominately depends on the statistics of your instance. The steps below explain how to gather and maintain a good set of statistics. These steps consider import of items and UDAs. If your import includes other entities
(such as suppliers), include the relevant interface tables in these steps.

Decide on an optimal batch size that you plan to use for your data loads.
Truncate the following interface tables:

MTL_SYSTEM_ITEMS_INTERFACE

MTL_ITEM_REVISIONS_INTERFACE

EGO_ITM_USR_ATTR_INTRFC

MTL_INTERFACE_ERRORS
Load data into the interface tables for one batch (use the size of the batch chosen in step 1).
Run the ICI Program to process the batch.
Ensure there are no errors in this batch and that all records in all interface tables are processed successfully.
Create a backup of the three interface tables that have processed data.

CREATE TABLE msii_bak_batch1

AS (SELECT * FROM MTL_SYSTEM_ITEMS_INTERFACE);

CREATE TABLE miri_bak_batch1

AS (SELECT * FROM MTL_ITEM_REVISIONS_INTERFACE);

CREATE TABLE eiuai_bak_batch1

AS (SELECT * FROM EGO_ITM_USR_ATTR_INTRFC);

COMMIT;

Repeat steps 2 through 5 for another batch id.
Copy the data backed up in step 6 to the respective interface tables.

INSERT INTO MTL_SYSTEM_ITEMS_INTERFACE

(SELECT * FROM msii_bak_batch1);

INSERT INTO MTL_ITEM_REVISIONS_INTERFACE

(SELECT * FROM miri_bak_batch1);

INSERT INTO EGO_ITM_USR_ATTR_INTRFC

(SELECT * FROM eiuai_bak_batch1);

COMMIT;

Gather statistics on these two sets of processed data in the interface tables using the code below. This code includes some recommended tweaks to the statistics to achieve optimal performance.

DECLARE

l_schema VARCHAR2(30);

l_schema_status VARCHAR2(1);

l_industry VARCHAR2(1);

BEGIN

fnd_stats.load_histogram_cols('DELETE',

401,

'MTL_SYSTEM_ITEMS_INTERFACE',

'PROCESS_FLAG') ;

fnd_stats.load_histogram_cols('DELETE',

401,

'MTL_SYSTEM_ITEMS_INTERFACE',

'SET_PROCESS_ID') ;

IF (FND_INSTALLATION.GET_APP_INFO('EGO',

l_schema_status,

l_industry,l_schema))

THEN

FND_STATS.GATHER_TABLE_STATS(

OWNNAME => l_schema,

TABNAME => 'EGO_ITM_USR_ATTR_INTRFC',

CASCADE => True);

END IF;

IF (FND_INSTALLATION.GET_APP_INFO('INV',

l_schema_status, l_industry,

l_schema))

THEN

FND_STATS.GATHER_TABLE_STATS(

OWNNAME => l_schema,

TABNAME => 'MTL_SYSTEM_ITEMS_INTERFACE',

CASCADE => True);

FND_STATS.GATHER_TABLE_STATS(

OWNNAME => l_schema,

TABNAME => 'MTL_ITEM_REVISIONS_INTERFACE',

CASCADE => True);

END IF;

dbms_stats.set_column_stats('EGO',

'EGO_ITM_USR_ATTR_INTRFC',

'PROCESS_STATUS',

distcnt=>4,

density=>0.25);

dbms_stats.lock_table_stats('EGO', 'EGO_ITM_USR_ATTR_INTRFC');

dbms_stats.lock_table_stats('INV', 'MTL_SYSTEM_ITEMS_INTERFACE');

dbms_stats.lock_table_stats('INV', 'MTL_ITEM_REVISIONS_INTERFACE');

END;

Perform the steps listed in the “Measuring the Current Performance of the ICI Program” section to measure the performance of a single batch and multiple batches.
Back up the statistics into backup tables for future use in case of an instance refresh. This avoids running all of the above steps to collect a good set of statistics. You can upload the statistics from the backup files directly instead.

drop table ego.EIUAI_STATS;

exec dbms_stats.create_stat_table('EGO','EIUAI_STATS');

exec dbms_stats.export_table_stats('EGO',

'EGO_ITM_USR_ATTR_INTRFC',

null,

'EIUAI_STATS');

drop table inv.MSII_STATS;

exec dbms_stats.create_stat_table('INV','MSII_STATS');

exec dbms_stats.export_table_stats('INV',

'MTL_SYSTEM_ITEMS_INTERFACE',

null,

'MSII_STATS');

drop table inv.MIRI_STATS;

exec dbms_stats.create_stat_table('INV','MIRI_STATS');

exec dbms_stats.export_table_stats('INV',

'MTL_ITEM_REVISIONS_INTERFACE',

null,

'MIRI_STATS');

Dump the data from EIUAI_STATS, MSII_STATS, MIRI_STATS to files.

In case of a refresh, restore these tables from the file dumps and restore the statistics.

Functional Recommendations


Deferring Sync of Item Text Index

The ICI program synchronizes the Item Text Index (used for item simple searches) with the latest item information after creating/updating items. This process has contention issues when executed in parallel across multiple concurrent programs. For initial
data load scenarios, where you do not expect other users to search for items at that time, Oracle recommends that you customize the product code to comment the call to sync index, and do the same manually after each set of parallel threads are completed. To perform
this customization:

In INVPOPIF, comment call to INV_ITEM_PVT.SYNC_IM_INDEX.
Run your data load batches.
After a set of parallel batches, call INV_ITEM_PVT.SYNC_IM_INDEX manually.
After you complete the data load process, uncomment the change you made in step 1.

Note: Commenting the call to INV_ITEM_PVT.SYNC_IM_INDEX stops the Item Text Index sync as well as the iProcurement (ICX) Item Text Index sync. If you do not want to stop the Item Text Index sync, but do want to stop the iProcurement (ICX) Item
Text Index sync, then:

Set the profile option 'INV:Sync iProcurement Item Index' to No.
Execute the concurrent request 'Rebuild Catalog interMedia Index' manually to sync the Item Text Index with Oracle iProcurement.


Disabling ENI Sync

The items created through the ICI program are also synced into the ENI tables. If you do not use ENI, Oracle recommends that you disable this using the following steps.

Confirm that you do not use ENI by running the following query. The output of this query should not be "I".

SELECT FA.APPLICATION_SHORT_NAME, FA.APPLICATION_ID,

PI.STATUS, PI.PRODUCT_VERSION, PI.PATCH_LEVEL

FROM FND_PRODUCT_INSTALLATIONS PI,FND_APPLICATION FA

WHERE FA.APPLICATION_ID = PI.APPLICATION_ID

AND UPPER(APPLICATION_SHORT_NAME)='ENI

If the above query does not return an "I", then truncate table ENI_OLTP_ITEM_STAR. This disables ENI sync during item import.


Table Valueset

Queries used in a table valueset also affect the performance of the Import Catalog Items program. Ensure that the queries used in a valueset based on a table are tuned for optimal performance.

Troubleshooting



If you have implemented the tips explained in this document and you still have a performance issue, you may need development assistance. This section discusses the information you need to provide while logging a Service Request (SR) in order to involve the
Oracle Product Hub development team. When logging a SR, provide the following information in addition to the standard information provided in all SRs (such as DB version and RUP):

Your hardware configuration:

Are DB/Middle and Concurrent tiers co-located on the same box or on different boxes?
Number of CPUs in each box.
Number of cores per CPU.
Memory

Number of items that already exist in your instance.

select count(*) from mtl_system_items_b
select count(*) from mtl_system_items_tl
select count(*) from ego_mtl_sy_items_ext_b
select count(*) from ego_mtl_sy_items_ext_tl
select count(*) from mtl_item_catalog_groups_b

Number of items you are trying to load.
Number of items you are loading per batch.
Number of child entities (UDAs, Intersections, People, AML, etc.).
Number of batches you are trying to run in parallel.
Provide the following reports:

Concurrent Program Log(s) for a debug enabled run. To enable debugs:

Set profile FND: Debug Log Enabled to Yes.
Set profile FND: Debug Log Level to Statement.
Set profile INV:Debug Level to 102.

Tkprof of the trace generated for the request.

To enable the trace:

Set the profile "Initialization SQL Statement - Custom" to "BEGIN

FND_CTL.FND_SESS_CTL('','','','','','ALTER SESSION SET EVENTS ='||''''||'

10046 TRACE NAME CONTEXT FOREVER, LEVEL 8 '||''''); END;"
Deselect the “Enable Trace” check box in the Concurrent Programs form for Import Catalog Items.
Once the program is complete, run the query below to get a trace file name and trace location:

SELECT 'Request id: '||request_id ,

'Trace id: '||oracle_Process_id,

'Trace Flag: '||req.enable_trace,

'Trace Name:

'||dest.value||'/'||dbnm.value||'_ora_'||oracle_process_id||'.trc',

'Prog. Name: '||prog.user_concurrent_program_name,

'File Name: '||execname.execution_file_name|| execname.subroutine_name ,

'Status : '||decode(phase_code,'R','Running')

||'-'||decode(status_code,'R','Normal'),

'SID Serial: '||ses.sid||','|| ses.serial#,

'Module : '||ses.module

from fnd_concurrent_requests req, v$session ses, v$process proc,

v$parameter dest, v$parameter dbnm, fnd_concurrent_programs_vl prog,

fnd_executables execname

where req.request_id = &request_id

and req.oracle_process_id=proc.spid(+)

and proc.addr = ses.paddr(+)

and dest.name='user_dump_dest'

and dbnm.name='db_name'

and req.concurrent_program_id = prog.concurrent_program_id

and req.program_application_id = prog.application_id

and prog.application_id = execname.application_id

and prog.executable_id=execname.executable_id;

AWR report for exact duration of the program.
ADDM and ASH report for exact duration of the program.
Sar report for the duration of the program.

If you think your program runs for 20 minutes, then, on the DB tier of your instance, run the command “sar 2 600” in the unix prompt. Spool that output and upload it.
Top output for the duration of the program.

Run the top command in the DB nodes unix prompt every 30 seconds when the program is running and provide the output.

Performance Results

This section lists some performance data observed using various hardware configurations during a performance tuning exercise. These numbers are only indicative and are not meant to be a benchmark. You should not use them for sizing your hardware. As stated
earlier, the performance numbers can vary greatly depending on the hardware configuration, functional setup, and so on.


Test Case 1


Hardware Setup


DB Server (also hosts Concurrent Manager): 2 nodes RAC, including 2 CPUs, 4 cores each, with 32GB memory in each node.
Middle Tier: 2 nodes RAC, including 2 CPUs, 4 cores each, with 32GB memory in each node.


Functional Setup


Number of UDAs per item (including user entered and defaulted): 70
Number of rules: 27

Item Number is Function Generated.


Timing Data

Number of items per threadNumber of orgs to which each item is assignedNumber of threadsAverage time in minutes per thread
50001 (master org)19
10000

1 (master org)119
50001 (master org)424
50001 (master org)632
100001 (master org)453
100001 (master org)683
50002 (1 master,

1 child)
453
50002 (1 master,

1 child)
690


Test Case 2


Hardware Setup


DB Server (also hosts Concurrent Manager): Non RAC. 32 processors and 64 GB memory total.
Middle Tier: 1 processor with 32 cores each and 16 GB memory total.


Functional Setup


Number of UDAs per item (including user entered and defaulted): 30-40
No rules. No function generated item number/description.


Timing Data

Number of items per threadNumber of orgs to which each item is assignedNumber of threadsAverage time in minutes per thread
50001 (master org)16.5
50001 (master org)27.0
100001 (master org)111.0
100001 (master org)219.0
50001 (master org)411.0


Test Case 3


Hardware Setup


DB Server (also hosts Concurrent Manager): Non RAC. 34 processors with 4 cores each and 48 GB memory total.
Middle Tier: 3 FE servers (2 forms/web tiers, 1 Concurrent Manager tier), 2 processors with 4 cores each and 8 GB memory total.


Functional Setup


Number of UDAs per item (including user entered and defaulted): 50-60
No rules. No function generated item number/description.


Timing Data

Number of items per threadNumber of orgs to which each item is assignedNumber of threadsAverage time in minutes per thread
50001 (master org)110
50001 (master org)215
100001 (master org)120
100001 (master org)2
50001 (master org)4

Functional Setup


Number of UDAs per item (including user entered and defaulted): 50-60
Number of Rules: 600.
Item Number and Description generated using PLSQL function.

Number of items per threadNumber of orgs to which each item is assignedNumber of threadsAverage time in minutes per thread
10001 (master org)116
2501 (master org)45

文档详细信息


通过电子邮件发送此文档的链接

在新窗口中打开文档









类型:
状态:
上次主更新:
上次更新:
WHITE PAPER
PUBLISHED
2014-4-10
2014-4-10


相关产品


Oracle Product Hub

Oracle Inventory Management

Oracle Item Master


信息中心


Product Information Center: EBS
Product Information Management (PIM) [1371567.2]



Information Center: Receiving [1391367.2]



E-Business Suite Product Information
Center Index [444.2]



Information Center: Product Information
Management (PIM) [1310505.2]



文档引用


此文档无可用的引用。

最近查看


Optimizing
Item Import Performance in Oracle Product Hub/Inventory [1208945.1]



中文文档列表
- Oracle EBS (Enterprise Business Suite) [1553829.1]



"500
Internal Server Error - java.lang.NoClassDefFoundError: oracle.apps.fnd.profiles.Profiles" on 10G [567554.1]



APP-SQLAP-97733
The System Cannot Generate Tax Distributions [1285604.1]



Enter
Tax Only Invoice fails - APP-SQLAP-97733 [1551788.1]


显示更多

内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签: 
相关文章推荐