OATS - 12.1.0.1.0361 Issue with running a  scenario

Hi,
I am doing a comparison of Hits/sec and cpu usage for a scenario with OATS and LR.
I have remote load generator and app server.
With OATS, i am getting almost half the hits/sec and cpu as compared to lr.
Till now what i have noticed is that the time for the transaction in OATS is higher than that of LR. But the transaction contains one single URL only. So am not sure what is causing this increase.
I have checked the server response content size and it is same in OATS and LR.
Logging and think time are disabled.
I would like some help with this issue.
Thanks

this thread is closed

Similar Messages

  • HT3131 Are there any ventilation issues with running a macbook air in clamshell mode?

    Are there any ventilation issues with running a macbook air in clamshell mode?

    ...that makes sense to me. 
    but, why do so many MBA users (on this forum and others) claim that ventilation occurs through the keyboard?  Is this claim based on an old macbook design?
    for example, here's a related discussion (although fairly old): 
    11-08-2010, 12:16 PM
      #6 (permalink)
    SP Forsythe 
    Notebook Evangelist
    Join Date: Jul 2007 
    Location: California
    Posts: 660
    Rep Power: 14
    Re: Is it safe to use MBP with screen lid closed? Leave charge on always?
    Quote: 
    Originally Posted by tHE j0KER
    Actually I you shouldn't close the lid while running on an external screen. The keyboard of the Macbook is an air intake for the fan. Close the lid, and it could overheat.
    A common misconception of unknown origin. The intake and the exhaust for the cooling fan on the 13" MB and MBP are both located on the rear slot at the base of the hinge. In fact, you can actually see the divider that separates the intake flow from the output flow. If it were through the keyboard crevices, then an awful lot of overheats would result from people using impermeable keyboard covers, as well as Apple's warranty department would be flipping out over the Apple store carrying such covers. iSkin ProTouch FX Keyboard Cover for all MacBooks - Black Printed Keys on White - Apple Store (U.S.) 
    Does one think that Apple sells these only for use when the Notebook is off??? http://store.apple.com/us/product/TW...co=MTM3OTUwMDE Closing the lid, whilst operating the unit actually results in cooler operation due to reduction in power consumed by operating the display, which in supporting the on-board display generates heat far greater than simply powering the video port.. 
    Does one think that Apple sells these for use only when the MacBook is off? http://store.apple.com/us/product/TW...co=MTM3OTUwMDEApples direction for use is specifically for when using with an external display.
    Currently using:
    Apple MacBook Air 13" mid 2011 1.8GHz Core i7 4GB 256 GB SSD Lion & Ubuntu Linux via Fusion
    MacBook Air 11.6" late 2010 1.6 GHZ, 320M, 4GB 180 GB SSD Upgraded OWC), OS X Lion
    11-08-2010, 02:34 PM
      #7 (permalink)
    ajreynol 
    Notebook Virtuoso
    Join Date: Mar 2009 
    Location: Ann Arbor, MI
    Posts: 2,542
    Rep Power: 18
    Re: Is it safe to use MBP with screen lid closed? Leave charge on always?
    once again, I've tried a few of these keyboard covers. the Moshi keyboard cover is the ONLY one I can recommend. the others are too thick or change the keyboard experience too much.
    17" Apple MacBook Pro | i7 2720m | 160GB SSD + 750GB | 16GB | HD 6750M 1GB
    Dell 435MT | i7 920 | 10GB RAM | 7.64TB HDDs | HD 6970 | Win7+SL
    HP Elitebook 2710p Tablet PC | 1.8GHz C2D | 4GB RAM | 160GB HDD | X3100
    Apple iPhone 4 32GB | Apple iPad 64GB (Gen 3)
    Stop random laptop wakeup | 5K500.B bench data | How to Disable PowerMiser
    Disable Vaio beep when pressing volume or special keys
    11-08-2010, 03:34 PM
      #8 (permalink)
    doh123 
    Without ME its just AWESO
    Join Date: Feb 2009 
    Posts: 3,282
    Rep Power: 22
    Re: Is it safe to use MBP with screen lid closed? Leave charge on always?
    1. Closing the cover will cause more heat. This is not because of covering the keyboard very much (though it does help some heat be retained). It's mainly because of the shape of the hinge and the fact when closed it covers up the back vent a lot more. For the best cooling, it is best to have the screen open. Just run it as a dual monitor, but make the external the Primary monitor, and if you don't want to use the built in, just turn its backlight off and don't use it.
    2. The thing you plug into the wall is not a battery charger. The actual "charger" is built into the computer. It knows when to charge and when not to. If the little light on the power plug is amber, then its charging your battery. When its green, its just powering the laptop and NOT charging your battery at all.
    Mac OS X Gamer/Porter
    (We do exist!)
    Wineskin 2.5 is available!. Turn Windows apps into Mac apps for free!
    11-08-2010, 03:38 PM
      #9 (permalink)
    Wolfpup 
    Notebook Virtuoso
    Join Date: Jun 2007 
    Posts: 3,871
    Rep Power: 28
    Re: Is it safe to use MBP with screen lid closed? Leave charge on always?
    I'd leave the lid open, at least partially...yes it may be fine not doing that, but you are making heat dissipate worse, of course could even theoretically hurt the screen.
    As for the battery...well there's really only two choices, have it plugged in or not. As others mentioned, you can't overcharge the battery. It can be damaged a bit from heat, but of course the number one thing that's going to damage it is discharging it...so it's a no brainer-use it plugged in whenever possible, and try to charge it whenever possible when it's not plugged in.
    11-08-2010, 03:45 PM
      #10 (permalink)
    SP Forsythe 
    Notebook Evangelist
    Join Date: Jul 2007 
    Location: California
    Posts: 660
    Rep Power: 14
    Re: Is it safe to use MBP with screen lid closed? Leave charge on always?
    Quote: 
    Originally Posted by doh123
    1. Closing the cover will cause more heat. This is not because of covering the keyboard very much (though it does help some heat be retained). It's mainly because of the shape of the hinge and the fact when closed it covers up the back vent a lot more. For the best cooling, it is best to have the screen open. Just run it as a dual monitor, but make the external the Primary monitor, and if you don't want to use the built in, just turn its backlight off and don't use it.
    Nope. Apple would disagree with you on that one. Any "closure", which is, when compared to the amount of CFM, is insignificant. In fact, the opening size remains the same. it is only deflected at a slight angle when the lid is closed. Tilt your MBP and see. As well, shutting down the display lowers the heat being generated, even in the lower case. As I said, if it were a problem, Apple would not be selling stands designed to operate your unit in the closed position as the original poster of this thread proposes to do.
    Currently using:
    Apple MacBook Air 13" mid 2011 1.8GHz Core i7 4GB 256 GB SSD Lion & Ubuntu Linux via Fusion
    MacBook Air 11.6" late 2010 1.6 GHZ, 320M, 4GB 180 GB SSD Upgraded OWC), OS X Lion

  • Hgrid - Issue with an unusual scenario

    Hi All,
    I've a Hgrid component with two levels, one VO master and one VO details and a View Link that is created in runtime to join them. In my page, this Hgrid shows the results based on parameters from a search region (I'm not using query region since it doesn't fit with the requirements). Basically what I need now is to filter the details VO based on extra parameters passed in runtime. In order to meet this requirement I've tried to use createViewLinkBetweenViewObjects method as the example below:
    am.createViewLinkBetweenViewObjects("AnalyticShipLnToAllocsGridVL",
    "AnalyticAllocsGridVO",
    voShipLineGrid, shipLnLinkAttrs,
    voAllocsGrid, allocsLinkAttrs,
    "ITEM_CODE = :1 AND ORGANIZATION_ID = " + pOrganizationId); // VL attribute plus extra criteria
    I'm currently facing two issues with the above code:
    1. My query is grouped by ITEM_CODE and consequently I'm not selecting the ORGANIZATION_ID attribute, even though it is defined in the View I'm selecting. After executing the application with the above code, the following error comes out:
    Statement:
    SELECT * FROM (SELECT component_type,
    component_reference,
    item_code,
    SUM(estimated_amt) estimated_amt,
    SUM(actual_amt) actual_amt
    FROM inl_dashboard_overall_v
    GROUP BY item_code, component_type, component_reference) QRSLT
    WHERE ITEM_CODE = :1 AND ORGANIZATION_ID = 9246
    Error: java.sql.SQLException: ORA-00904: "ORGANIZATION_ID": invalid identifier
    2. To solve the first issue, I tried to use the VO method setNestedSelectForFullSql(false), but if I do so, the WHERE clause is built after the Group By and it results in a different error:
    Statement:
    SELECT component_type,
    component_reference,
    item_code,
    SUM(estimated_amt) estimated_amt,
    SUM(actual_amt) actual_amt
    FROM inl_dashboard_overall_v
    GROUP BY item_code, component_type, component_reference
    WHERE ITEM_CODE = :1 AND ORGANIZATION_ID = 9246
    Error: java.sql.SQLException: ORA-00933: SQL command not properly ended
    Can anybody please help me on how to solve and deal with such scenario?
    I'd appreciate any suggestion, idea or comments on this.
    Thanks and Regards,
    Eduardo

    Hi Kumar,
    Thanks for your answer but the "9246" in the organization_id is passed in runtime.. How can I handle this since this VO is the Detail in a View Link??
    And If I do this:
    SELECT component_type,
    component_reference,
    item_code,
    SUM(estimated_amt) estimated_amt,
    SUM(actual_amt) actual_amt
    FROM inl_dashboard_overall_v
    WHERE ITEM_CODE = :1 AND ORGANIZATION_ID = :2
    GROUP BY item_code, component_type, component_reference
    How can I use the View Link to set the binding parameter :2 ??
    Thanks,
    Eduardo

  • Are there issues with running CS 6 on Windows 8? (NOT the CC edition)

    I'm looking at getting a new computer and am wondering what issues I'll have using my CS6 suite if I get a Windows 8 machine. I know that the Creative Cloud versions are compatible, but what about the 'old' software?

    So, it seems my issue with the Limited Internet connection has been resolved. I didn't use my computer for two days and this afternoon when I turned it on I was still having issues. However, I went into the HP Support Assistant and it popped up that I had some updates to download and install, so I let it do its thing. It took a while because the Internet kept cutting out, but once the update for the Intel Wireless finally installed, my connection has STAYED CONNECTED. Finally!
    I don't know why when I went into the wireless properties it said "device is working properly" and when I tried to Update the driver it said "your driver is the latest version" or whatever.
    I suggest you go to the HP Support Assistant and check for updates in the "Updates and Tune-Ups" section. It was kind of random, but knock on wood my Wireless connection issue looks to be gone.
    Hope this helps
    P.S.
    Here's a screen shot of what the HP Support Assistant (it is the one with the big question mark logo) main screen looks like :

  • Issues with running datapump export

    Hi all!
    I am running in to an issue with data pump exports on 6 servers. They are running the following:
    Windows 2008 server R1 standard
    Oracle Enterprise Edition 11.1.0.7
    I am entering the following:
    expdp directory=DIRECTORY dumpfile=TEST1.dmp schema=ORCL
    I get the following error when doing the export:
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit
    Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Starting "PSSBACKUP"."SYS_EXPORT_SCHEMA_05": pssbackup/******** directory=PSS_B
    ACKUP dumpfile=testing2345.dmp schemas=cuba
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 481.6 MB
    Processing object type SCHEMA_EXPORT/USER
    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
    Processing object type SCHEMA_EXPORT/ROLE_GRANT
    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
    Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
    ORA-39014: One or more workers have prematurely exited.
    ORA-39029: worker 1 with process name "DW01" prematurely terminated
    ORA-31672: Worker process DW01 died unexpectedly.
    Job "PSSBACKUP"."SYS_EXPORT_SCHEMA_05" stopped due to fatal error at 14:24:16
    Oracle support was brought in on this but they were not able to figure it out. Can anyone help out?

    I got this from the trace file:
    *** 2010-10-01 14:23:10.760
    Exception [type: ACCESS_VIOLATION, UNABLE_TO_WRITE] [ADDR:0x0] [PC:0x2CFAF23, kuxgetHashCode()+121]
    Incident 60413 created, dump file: D:\APP\PSSADMIN\diag\rdbms\pss2437\pss2437\incident\incdir_60413\pss2437_dw01_3740_i60413.trc
    ORA-07445: exception encountered: core dump [kuxgetHashCode()+121] [ACCESS_VIOLATION] [ADDR:0x0] [PC:0x2CFAF23] [UNABLE_TO_WRITE] []
    I haven't been able to pull up any information indicating how to fix this error. I'm not sure where the [UNABLE_TO_WRITE] error is referring to.

  • Issues with running Server 2012R2 Essentials as an Offline Standalone Root CA?

    Hi everyone,
       I've searched this forum as well as Google and have not been able to find any concrete answers so I am hoping one of you experts will be able to assist me.  
       I have an all Windows 2008 server/domain enviroment.   I was looking at implementing a two-tier PKI with an offline, standalone root CA and an enterprise issuing CA (2008 member server).   Budgets are tight so I was hoping
    I could get away with using Server 2012R2 Essentials as the offline standalone root.  
       It is my understanding that 2012 Essentials is configured as a DC by default and needs to remain that way per licensing.   I know the recommended configuration for an offline root would be to have the server be in a workgroup and not
    on a domain.
       So question 1 is will 2012 Essentials work as an offline standalone root?
       Question 2 is will there be any issues with it running as described?   In other words will the fact that it is the sole DC in its own domain cause issues with its use as an offline root?  
       Thank you in advance for your help!

    The essentials experience role runs on server standard and is very different from the essentials product in both licensing and pricing. While you can indeed buy standard and deploy the essentials experience role as a "standalone" server, there
    would be ZERO benefit in a PKI offline root scenario the essentials role has no automation or configuration options in the dashboard for that use case, and pricing, you'd still be paying for server standard.
     The essentials product (or SKU) has the benefit of reduced cost, but cannot be deployed standalone and has enough other restrictions that it is not well suited for the given desired use.
    So either way, my answer stands. Essentials (prodct it role) is not the right tool for the job in this case.

  • Issue with running Lexmark Prestige Pro805 printer with iMac running 10.3.9

    Thought I'd share this with you in case anyone else runs into this problem.
    Recently bought a Lexmark Prestige Pro 805 printer to service a home network. Had no issues with the printer using 10.6 on MacBook Pro, 10.5 on Intel iMac and also running on Laptop with Windows XP all running off same network via USB connection into a Time Machine.
    However when tried running this off an iMac PPC using 10.3.9 although I could print after initial install subsequent prints after logging off and on hit the problem of "Unable to read CUPS raster data".
    Lexmark driver updates and firmware updates failed to fix issue and contacted Lexmark who advised me to do a complete uninstall and re-install but to same effect. They then determined this was a CUPS issue and I should contact Apple. Looked on this forum and found a message posted in Nov 2005 which suggested doing a disk repair off the original install disk in order to repair corrupted CUPS directories (simple to say after 6 years - where exactly did I put those disks?).
    Anyway did this and indeed both permissions and disks needed repair and it fixed the issue.
    As it took me a while to get to the answer hopefully by posting this if anyone else comes across this issue they can save some time!

    Thanks for posting your solution. May help another Mac user in the future.
     Cheers, Tom

  • Issue with running redistributeactivedatabases.ps1 in new-pssession

    Hi,
    I am stuck with the issue to run redistributeactivedatabases.ps1 in new-pssession.
    Is there any best pratice or some other script to redistribute the mailbox databases using new-pssession

    Hi Jrv,
    I am sorry for not being very clear.
    Also i know what i am trying is not very sane but i cannot help it as i got order that i have to achive this.
    Below are the exact suitation.
    We have JP1 server (Automation) which we use to reboot/failover/failback.
    I have to automate failover and failback scripts. I am able to achive failover part as it is simple command
    Move-ActiveMailboxDatabase -server servername
    and it is working.
    For failback and automatic rebalance mailbox database according to the activation preference i am using below script which i have to run using cmd.
    $mbxs =
    Get-MailboxDatabase
    | sort name
    ForEach($mbx
    in $mbxs){
        $MBdb=$mbx.Name
        $ServerHosting=$mbx.Server.Name
        if($mbx.activationPreference.value
    -eq 1)
            If
    ($ServerHosting -ne $ActivationPreference.Key.Name){
                Move-ActiveMailboxDatabase $MBdb
    -ActivateOnServer $ActivationPreference.Key.Name
    -confirm:$False
    But the problem i am getting is that script is not working when i am running this in cmd or in powershell.
    But if i run the same script on Exchange server using Exchange Management shell then there are no issues.
    So, i need your help so that i can run above script using cmd in new-pssession.
    Hopefully i am able to make some sense now.

  • Issue while running a scenario while using TTU in ODI

    Hi All,
    I have created an interface to load data from oracle to teradata. I have used LKM as SQL to Teradata (TTU). When i run this interface from UI (from my desktop) I receive an error telling
    "sh: fastload: command not found" which makes perfect sense.
    I then tried to create a scenario and run the scenario from a machine which has the utility installed. I get the below error.
    ./startscen.sh ORACLE2TD 001 TESTNEWCONTEXT
    OracleDI: Starting scenario ORACLE2TD 001 in context TESTNEWCONTEXT ...
    2010-11-15 22:53:42.017 ERROR ODI-1217 Session ORACLE2TD (399001) fails with return code 7000.
    2010-11-15 22:53:42.865 ERROR ODI-1134 Agent Internal encountered an error: ODI-1217: Session ORACLE2TD (399001) fails with return code 7000.
    2010-11-15 22:53:43.551 NOTIFICATION Session 399,001 ended with status E (ERROR)
    Status message is : Session Failed :399001 : ODI-1217: Session ORACLE2TD (399001) fails with return code 7000.
    ODI-1226: Step oracle2td fails after 1 attempt(s).
    ODI-1240: Flow oracle2td fails while performing a Loading operation. This flow loads target table ETL_CONTROL_GROUP_CUR_1.
    Caused By: org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 19, in ?
    ImportError: no module named threading
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
    at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:345)
    at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:169)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2374)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java:1615)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java:1580)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java:2755)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2515)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:534)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:449)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1954)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1818)
    at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor$2.doAction(StartScenRequestProcessor.java:559)
    at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:224)
    at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor.doProcessStartScenTask(StartScenRequestProcessor.java:481)
    at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor$StartScenTask.doExecute(StartScenRequestProcessor.java:1040)
    at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:114)
    at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$1.run(DefaultAgentTaskExecutor.java:49)
    at org.springframework.core.task.SyncTaskExecutor.execute(SyncTaskExecutor.java:49)
    at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor.executeAgentTask(DefaultAgentTaskExecutor.java:41)
    at oracle.odi.runtime.agent.processor.TaskExecutorAgentRequestProcessor.doExecuteAgentTask(TaskExecutorAgentRequestProcessor.java:93)
    at oracle.odi.runtime.agent.processor.TaskExecutorAgentRequestProcessor.process(TaskExecutorAgentRequestProcessor.java:84)
    at oracle.odi.runtime.agent.RuntimeAgent.startScenario(RuntimeAgent.java:318)
    at oracle.odi.StartScen.main(StartScen.java:192)
    If anyone has encountered a similar issue and able to solve it, please share your thoughts with me.
    Thanks in advance for your help.
    Thanks
    Ashish
    Edited by: 811686 on Nov 15, 2010 3:41 PM

    Thanks Cezar. I tried the same.
    sato:work/ashish/ODI/generic/agent_standalone/oracledi/agent/bin$./startscen.sh ORACLE2TD 001 TESTNEWCONTEXT -NAME=localagent
    OracleDI: Starting scenario ORACLE2TD 001 in context TESTNEWCONTEXT ...
    2010-11-16 23:54:54.388 ERROR ODI-1217 Session ORACLE2TD (401001) fails with return code 7000.
    2010-11-16 23:54:54.899 ERROR ODI-1134 Agent localagent encountered an error: ODI-1217: Session ORACLE2TD (401001) fails with return code 7000.
    2010-11-16 23:54:55.617 NOTIFICATION Session 401,001 ended with status E (ERROR)
    Status message is : Session Failed :401001 : ODI-1217: Session ORACLE2TD (401001) fails with return code 7000.
    ODI-1226: Step oracle2td fails after 1 attempt(s).
    ODI-1240: Flow oracle2td fails while performing a Loading operation. This flow loads target table ETL_CONTROL_GROUP_CUR_1.
    Caused By: org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 19, in ?
    ImportError: no module named threading
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
    at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:345)
    at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:169)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2374)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java:1615)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java:1580)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java:2755)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2515)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:534)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:449)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1954)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1818)
    at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor$2.doAction(StartScenRequestProcessor.java:559)
    at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:224)
    at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor.doProcessStartScenTask(StartScenRequestProcessor.java:481)
    at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor$StartScenTask.doExecute(StartScenRequestProcessor.java:1040)
    at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:114)
    at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$1.run(DefaultAgentTaskExecutor.java:49)
    at org.springframework.core.task.SyncTaskExecutor.execute(SyncTaskExecutor.java:49)
    at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor.executeAgentTask(DefaultAgentTaskExecutor.java:41)
    at oracle.odi.runtime.agent.processor.TaskExecutorAgentRequestProcessor.doExecuteAgentTask(TaskExecutorAgentRequestProcessor.java:93)
    at oracle.odi.runtime.agent.processor.TaskExecutorAgentRequestProcessor.process(TaskExecutorAgentRequestProcessor.java:84)
    at oracle.odi.runtime.agent.RuntimeAgent.startScenario(RuntimeAgent.java:318)
    at oracle.odi.StartScen.main(StartScen.java:192)
    sato:work/ashish/ODI/generic/agent_standalone/oracledi/agent/bin$ps -ef | grep -i localagent
    edwd 1311024 2274074 0 23:55:23 pts/15 0:00 grep -i localagent
    edwd 1573280 1545022 0 23:48:11 pts/11 0:00 /bin/sh ./agent.sh -NAME=localagent
    edwd 2396516 1573280 0 23:48:11 pts/11 1:50 /usr/java6_64/bin/java -Xms32m -Xmx256m -classpath ../lib/.:../drivers/.:../../../odi_misc/*.zip:../../../odi_misc/*:../lib/*.zip:../lib/*:../drivers/*.zip:../drivers/*:../lib/scripting/*.zip:../lib/scripting/*::/usr/java6_64/lib/tools.jar -Djava.security.policy=server.policy -Doracle.security.jps.config=./jps-config.xml -DODI_MASTER_DRIVER=oracle.jdbc.OracleDriver -DODI_MASTER_URL=jdbc:oracle:thin:@chasers.corp.apple.com:1722:edsd -DODI_MASTER_USER=edwsd -DODI_MASTER_ENCODED_PASS=d,yX5CMNzbGSWEfbYghXQabtk -DODI_SECU_WORK_REP=WORKREP1 -DODI_SUPERVISOR=SUPERVISOR -DODI_SUPERVISOR_ENCODED_PASS=f8yXwZyTPaaMainI5qdnIPFaf -DODI_USER=SUPERVISOR -DODI_ENCODED_PASS=f8yXwZyTPaaMainI5qdnIPFaf -DODI_JMX_PROTOCOL=rmi -Dorg.mortbay.log.class=oracle.odi.logging.AgentJettyLogger -Doracle.core.ojdl.logging.config.file=ODI-logging-config.xml oracle.odi.Agent -MASTER_DRIVER=oracle.jdbc.OracleDriver -MASTER_URL=jdbc:oracle:thin:@chasers.corp.apple.com:1722:edsd -MASTER_USER=edwsd -MASTER_ENCODED_PASS=d,yX5CMNzbGSWEfbYghXQabtk -WORK_REPOSITORY=WORKREP1 -ODI_SUPERVISOR=SUPERVISOR -ODI_SUPERVISOR_ENCODED_PASS=f8yXwZyTPaaMainI5qdnIPFaf -ODI_USER=SUPERVISOR -ODI_ENCODED_PASS=f8yXwZyTPaaMainI5qdnIPFaf -ODI_CONNECTION_RETRY_COUNT=0 -ODI_CONNECTION_RETRY_DELAY=7000 -NAME=localagent
    sato:work/ashish/ODI/generic/agent_standalone/oracledi/agent/bin$which fastload
    /usr/bin/fastload
    sato:work/ashish/ODI/generic/agent_standalone/oracledi/agent/bin$
    I have an agent localagent that is running on this machine, but when i run the scenario, it fails. Any further inputs will be very useful. Thanks

  • Issue with running QuickTime Windows.  Buffer Overrun Error - C++ Library .

    Initial problem was Buffer Overrun Error (C++ Library) when clicking on QuickTime after installation. IE. QT would not even open. http://support.microsoft.com/kb/831875#appliesto
    I took these steps:
    1. Tried to uninstall QuickTime by itself (it failed).
    2.
    3. Manually deleted apple, itunes and quicktime from the entire system (where ever it let me).
    4.
    5. Manually took out from the registration the apple stuff.
    6.
    7. Left the items in the recycle bin (in case there were any real issues and I needed something restored).
    8.
    9. Performed a registration cure (RegCure).
    10.
    11. Took off my entire Anti Virus.
    12.
    13. Dropped down separately the QuickTime and separately the iTunes on desktop.
    14.
    15. Tried to install QuickTime from one of the two saved files on my desktop, but encountered a serious fault
    16.
    17. It needed the QuickTime Installer to remove QuickTime itself, else it crapped out and nothing happened. This complained for a QuickTime.msi file which was a problem.
    18.
    19. Went to the recycle bin and restored only components which were marked QuickTime Installer.
    Removed the QuickTime instead of Repair.
    Went to the website and Installed QuickTime 7 directly.
    It opened on the desktop after installation.
    Installed iTunes separately from the desktop and it opened directly.
    Rebooted my pc.
    Enabled all my Security (McAFee).
    Opened one by one the QuickTime and Then the iTunes.
    Created a computer restore point with a narrative for the future.
    This was a very difficult task and required a lot of steps. I am glad you helped me with the removing part. Its great to have everything working again on my pc.
    I hope this was helpful - it took me ageges to fix.

    I'm experiencing exactly the same bug. Matter of fact, it's the first time in years that I've run across this kind of 'problem' when using non-beta software from a major player. Too bad. This really reflects poorly on Apple's credibility.

  • Issue with Product Allocation Scenario

    I am Trying out Product Allocation in APO (SCM 5.0).
    Relevant setup in SPRO - Product Allocation Object, Group, Procedure, Sequence etc., Planning Object Structure, Planning Area, CVC generation, Connection between Planning Area and Product Allocation object in GATP, assignment in Location Product Master (at Customer Location) Product Allocation Procedure, Check mode, ATP Group has been done in APO.
    After entering data in appropriate keyfigures (Allocated Qty) it shows up under GATP>> Reporting >> Product Allocations >> Time Series per CVC
    Have carried out ATP Simulation to check that for a given product at Customer the Product Allocation against a Required Quantity takes place correctly.
    Now trying to carry out the Product Allocation check by creating a Sales Order in ECC. In that case during the material availability check screen in APO, the product allocation does not show up.
    I believe it is related to master data settings for the product in ECC. My questions are as follows:
    1. In ECC the product exists at the manufacturing plant only (and not at customer as in APO). There in MRP3 tab Strategy Group is maintained. Is this along with Availability Check field responsible for determining the check instruction (combination of check mode + business event) in APO for the Product Allocation check.
    2. If so (which I guess it is) I can see the check instruction picked up in APO while creation of sales order in ECC is different from the check instruction I ran simulation with. How can I ensure the same check instruction is chosen (in which case I believe the availability check result will be same as the ATP Simulation carried out in APO).
    3. In APO I ran simulation for the product at customer location (and maintained the required ATP settings in the product master). In case of Sales Order creation in ECC how does system pick up the Check Mode and Business Event for the product (having Sold-to as customer and Supply Plant as the plant code) and use it in APO.
    4. Any other setting/configuration that is missed causing this behaviour.
    I guess some of the questions is a bit hazy but I would appreciate if someone can outline the Product Allocation Scenario. I am trying out based on Building Block S59 Demand Planning for Product Allocation but of course it does not detail out enough.
    Disclaimer: My knowledge of GATP is almost nil but am exploring.
    Thanks,
    Somnath

    Hi Somnath,
    It will be a pleasure to help you out.
    Firstly I will list out all the steps which I can think of:
    (1) R3-You have moved all the ATP customizing from R3 to APO via ATP integration model.
    (2) APO- goto SPRO>APO>GATP>General Settings>Maintain Check Instructions
    Here go to Check Mode- 041 and Business Event- A and maintain the following settings manually in development and migrate it all the way to Production
    I am assuming here that you want to first do an allocation check and then availability check**
    Product Allocation: Change from NO CHECK to First Check
    Product Check: Change from First Check to Second Check
    This check mode and business event is the most common one which is triggered while placing a sales order
    See the following blog for DETERMINATION OF CHECK MODE
    http://solution-timezone-issue.blogspot.com/
    There might be other check modes( 030, 050) which might also require this change. However you can judge that while placing a sales order and when you go to the availability screen, You have the tab - SCOPE OF CHECK CONTROL where you can see the Check Mode + Business Event
    Following guidelines can be used to modify Check Instrctions after Customizing Import from R/3 –
    1.     Business Event A*  is Sales Order Entry (The main scope of GATP) so define check instructions for all check modes since it is not possible to identify the requirement class that may come over on an order. For Business Event A* if availability check was being done in R3, then define check instructions with Allocation as first step, then Availability check as second step. Else if no availablity check in R3 then define check instructions with no availability and no allocation checks.
    2.     Business Event B* is  normally for deliveries but is also required for Rush Sales Orders. An R/3 configuration showed that 041 is the only requirement class that will come across for this business event. Hence defined only for 041 check mode. For Business Event B* if availability check was being done in R3, then define check instructions with Allocation as first step, then Availability check as second step. Else if no availablity check in R3 then define check instructions with no availability and no allocation checks.
    3.     Business Event 03 required for Inventory Transaction availability check such as MB1C. For such transactions, the APO Product Master Check Mode is used. Hence check instructions for all Check Modes are maintained. For Business Event 03 if availability check was being done in R3, then define check instructions with Availability check as first step. Else if no availablity check in R3 then define check instructions with no availability checks.
    (3) Maintain the Product Alllocation Procedure:
    <b>a-</b> Create a Product Alloc Object -- Example: PAO-001
    In this step, we define the product allocations objects - product allocations are saved per object for a characteristics combination in the product allocation group.
      Product Allocation Procedures can be used to maintain different Product Allocation Objects for different time validity periods.
      Generally Product Allocation Objects are used to model the different ‘reasons’ for the need for allocation – material shortage, capacity constraint, etc.
    <b>b-</b> Product Alloc Group--PAO-1234
    A product allocation group defines a group of characteristics against which allocation checking is done.
       Creating a Product Allocation Group creates a characteristics directory in the APO database and an area in Live Cache where the key figures are maintained.
       If ‘Check Planning Area’ is set to No then allocation and incoming order quantitiy data can be seen in table /SAPAPO/QTTAB. If the ‘Check Planning Area’ is set to Yes then no data is stored in this table and the allocation check is done directly against the data in the Planning area.
       Detials of allocationconsumption by sales orders are maintained in table /SAPAPO/QTVB in APO and in QTVB in R/3.
      If ‘Check Planning Area’ is set to ‘Yes’ and if there are users are in the change mode in the Planning Book, Sales Order Entry can get locked out.
      If ‘Check Planning Area’ is initially set to yes and then deactivated later you will no longer see data in /SAPAPO/AC42.
    Settings:
    Check Planning Area – Yes. The reason being that users are not expected to lock planning books as they would be defaulted into the display mode (through a development object). Also most allocation uploads will be through batch jobs. Setting it to ‘Yes’ reduces one step of moving data between the planning area and the allocation group.
    Check Date - Material Availability Date. Consumption of allocation quantity and checking of available allocation quantities is determined by the check date.
    Time Bucket Profile – Week
    Characteristics – Product Allocation Object (1), Product (2), Plant(3), Customer(4)
    Consumption Period – Bwd Consumption 1  
    Meaning that first the order will consume from current week allocation and then go 1 week backward to consume.
    <b>c-</b> Product Allocation Procedure --- PAO-4567
    Step Sequence:
    <i>Step</i> -
    <i>Step Name-</i> -
    Prod Alloc Group -
    Wild card
    1                      First Step             PAO-1234                     9
    <i>Alloc Procudure</i> - PAO-4567
    <i>Prod Alloc Step</i> -1
    <i>Control</i>
    <i>Valid from -
    Valid to---- -
    Object -
    Active -
    Factor</i>
                            12/31/2037      PAO-001    Check the Box                  1
    <b>d-</b> Maintain Connection to Planning Area (APO) 
    In this step, we maintain the assignment of the planning area to the product allocation group. We must make this assignment because characteristics of the product allocation group have different names from the info objects of the planning area.
      All characteristics of a product allocation group must be assigned to an info object. But multiple info objects can be assigned to one info object. It is not always necessary to assign every info object to a characteristic. It is possible for aggregation and disaggregation (depending on the direction of transfer) to take place with info objects that are not assigned.
    <i>Product Alloc Group -
    Planning Area -
    Version -
    Time Buckets</i>
    PAO-1234                         xxxxx                     000             W
    <i>Characteristic</i>
    KONOB - Prod Alloc Obj
    MATNR --- Material
    WERK -
    Plant
    XXXXX---- Customer
    <i>Key Figures</i>
    Order Qty - AMENGE
    Alloc Qty - KCQTY
    <b>e-</b>  Check the Product Alloc Settings
    This is only a check- Everything should be green and OK .
    (4) You have planned the Characteristic Combination on which you want to do ALLOCATION.
    <b>Create a new POS-</b> Which has your chracateristics as PRODUCT, PLANT,CUSTOMER, Product Alloc Object ( 9AKONOB)
    <b>Planning Area,-</b> Key figures required are Alloc Qty, Order Qty( Sales Order Qty)
    <b>PLannig Book, Data View</b> which has the following characteristics- Product, PLant
    Assuming you are allocating at the following CVC level- Product, PLant,Customer
    Step 1:
    /n/sapapo/MC62-Create Time Series- Basicallly your CVCs -
    Example: Product 1234, Plant INDIA, CUstomer WALKART
    Step 2:
    Move Characteristic Values Combination to Allocation Group (APO) 
    /SAPAPO/ATPQ_PAREA_K
    INput :
    Planning Area: xxxx
    Prod Alloc Group: PAO-1234
    (5) In the INtegration MOdel - Activate the ATP CHECK Integration model for the Plant and material
    THis is the KEY which decides whether availability should be done in R3 or APO
    Material Active in the integration model -- Means Availability( GLOBAL AVAILABILITY: Prod Alloc+Prod Availability) is done in APO
    Material not active in the integration model --- Means Availability Check is done in R3
    (6) APO MASTER DATA SETTINGS:
    Material Master:  On the ATP Tab, Assign Loc Dep Procedure ( Prod Alloc Procedure)
    Location Master: Calendar Tab> Create a shipping time stream calendar
    Additional Info: Please contact your R3-SD people as this involves a lot of understanding and I have not listed all the details here.
    This config shown here is a very basic one which not even 25% of the companies might be using.
    If you have any questions on this part- please see my business card and email me**
    (7) SCHEDULING: This is a major task which determines the route and transit times for calculating the
    Material Availability date
    Goods Issue Date
    Delivery Date
    on the sales order line item- schedule line
    I dont know whether you need info on this**There are many ways of doing this
    a- Configurable Scheduling
    b- COndition records
    c- User Exit
    You can find documentation in SPRO or else let me know*
    For simplicity- Let me take the easiest to create and difficult to maintain master data- Condition Records
    Step1 --Define Field Catalog for Condition Table (APO) 
    /SAPCND/AU01
    Require Plant and Route defined here in the field catalog
    Step 2- Define Condition Table for Scheduling (APO) 
    /SAPCND/AU03
    Create a new table- 901
    901 -  Pick Time based on Route with key as Route
    902- Route Time based on route
    Step 3- Maintain Access for Scheduling (APO) 
    /SAPCND/AU07
    Take the important ones- PICK and TRAN
    I am assuming again as I dont know what kind of Scheduling is maintained on R3 Side. Mostly 80% companies have only two lead times- Pick and Route Transit time.**
    PICK- Select pick and click on accesses > <i>Acno</i>-10-- <i>Table</i>-901
             >> Then save it > and go in again
    Select PICK and click on Fields> A popup will appear > Enter the access sequence as PICK--- Enter the access as 10  and click on the GREEn ARROW and save it
    TRAN- Select TRAN and click on accesses > <i>Acno</i>-10-- <i>Table</i>-902
             >> Then save it > and go in again
    Select TRAN and click on Fields> A popup will appear > Enter the access sequence as TRAN--- Enter the access as 10  and click on the GREEn ARROW and save it
    Now coming to maintaining the Condition records in APO EASY ACCESS
    /SAPCND/AU11
    Create  SCheduling Step for PICK : Enter condition type- PICK
    <i>Enter Route</i>- USCAN --- <i>Duration</i> -24:00 (Hrs)
    Create  SCheduling Step for TRAN: Enter condition type- TRAN
    <i>Enter Route</i>- USCAN --- <i>Duration</i> -48:00 (Hrs)
    Now Let me answer your questions:
    (1) Question 1- <i>In ECC the product exists at the manufacturing plant only (and not at customer as in APO).</i>     
    I did not undertsand what you mean here. As we can create a Material at the following locations in APO( Dist Center(DC), Manu Plant(PL), Cusomter, Vendor, etc)
    For GATP all we need is the material at a PLant location
    STrategy Group on MRP3 tab: This is purely used for determining the CHECK MODE value on APO Product Master.
    When this is blank on R3, the check mode will be blank on APO.
    However SAP recommends Putting in a strategy group value on R3 Material Master
    or
    a default Check MOde value for all those Products who have blank check mode-   041 is the best value * This piece SAP hasnt recommended but I found that there is No harm in anyway and it is MANDATORYfor inventory transactions like GOODS ISSUE otherwise you will get a QUEUE in SYSFAIL **
    Even if this is blank this has no effect on the GLOBAL ATP as the Check mode for a SD transaction is determined as follows
    98% of the SD transactions will generate a CHeck mode of 041 and Business Event- A ***
    Through the item category and the MRP type in the MRP view
    The item category is derived from the item category group in the Sales view of the material master record and from the order type of the sales document. Together with the MRP type in the MRP view of the material master record, the item category determines what requirements type is selected. The advantage of this method is that the selection of the requirements type depends on the business process (such as quotation, sales order, or repair).
    (2) Question 2 - I guess now you must have understood why when you do a simulation in APO, it is taking the Check mode from APO product Master
    and when you do a ATP Check from R3, the check mode(041) is different
    You dont need to ensure that they are the same- It doesnt matter in anyway.
    (3) Question 3- <i>In APO I ran simulation for the product at customer location (and maintained the required ATP settings in the product master).</i>I
    When you run ATP simulation- AC04, you generally run it for the plant as that is the supplying location and not the customer.
    <i>In case of Sales Order creation in ECC how does system pick up the Check Mode and Business Event for the product (having Sold-to as customer and Supply Plant as the plant code) and use it in APO.</i>
    PLease look above for this answer.
    (4) Question 4 - <i>Any other setting/configuration that is missed causing this behaviour.</i>
    To my best knowledge, whatever I could remember I have listed out. However there are some more steps which might be needed.
    Follow all the steps as I have mentioned and whenever you get struck somewhere, shoot me a message and I will help you.
    ONe more pointer here to what you have mentioned
    Relevant setup in SPRO - Product Allocation Object, Group, Procedure, Sequence etc., Planning Object Structure, Planning Area, CVC generation, Connection between Planning Area and Product Allocation object in GATP, assignment in Location Product Master (at Customer Location) Product Allocation Procedure, Check mode, ATP Group has been done in APO.
    After entering data in appropriate keyfigures (Allocated Qty) it shows up under GATP>> Reporting >> Product Allocations >> Time Series per CVC
    <u><i>Have carried out ATP Simulation to check that for a given product at Customer the Product Allocation against a Required Quantity takes place correctly.</i></u>
    Somnath, I believe you might have a wrong picture here. The concept of allocation is  Controlling Supply against huge Demand which is in this case coming from SALES ORDER
    When you place a sales order, the supplying location is your plant and that is where you would like to put in allocation quantities
    Say for example for a material 12345 - You have stock of 100 Qty in Plant KING fo this week and it shows as follows in the planning book
    Customers say are WALKART, GE-FUG, MACROGUN
    The CVCs which we have to create are as follows
    Similar to Demand PLanning**
    Material 12345 -PLant KING-Cust WALKART-Prod Alloc Obj
    Material 12345 -PLant KING-Cust GE-FUG-Prod Alloc Obj
    Material 12345 -PLant KING-Cust MACROGUN-Prod Alloc Obj
    And this is how it appears in the planing book when you go into DETAILS ALL
    You basically enter the Allocation Quantities just as how you enter forecast quantities in DP Book*
                            Wk1   |  WK2     WK3   WK4  WK5
    Alloc Qty                     |
    Material 12345                |
    Plant KING                    |
    Cust WALKART            50    |
    Cust GE-FUG             30    |
    Cust MACROGUN           20    |
    Incoming Order Qty               |
    Material 12345                   |
    Plant KING                       |
    Cust WALKART           <u>50</u>|
    Cust GE-FUG            <u>25</u> |
    Cust MACROGUN          <u>20</u> |
    <u>50</u> : Say the Cust WALKART has ordered 250 for this week, however only the allocated quantity -50 will be confirmed for him and the remaining 200 qty will be confirme at a later week when he is given the allocated amount.
    <u>25</u> :Similarily say this customer GE-FUG has ordered 20 qty. So his remaining allocation is basically 5 which cannot be used by any other customer.
    <u>20</u> : Say this customer MACROGUN has ordered exactly the allocated amount of 20. So his remaining allocation will be 0.
    This is the basic idea of how Allocation Concept is used in GATP.
    Ofcourse you can add the other functionalities like
    Rules Based ATP(LOcation Substitution, Product Substitution)
    Multi ATP( BOM-COmponents)
    Check Aginst Forecast
    Etc Etc
    Like any other module of SAP- this module has much to offer which unfortunately I cannot list here**
    GOOD LUCK and let me know If you need additional help.
    Regards
    KUMAR AYYAGARI
    Message was edited by:
            Kumar P Ayyagari

  • Can anybody Assist, please ? Issue with running Prem Pro 1.5 on 'new' pc with Win XP Pro

    After some years of faithful service my PC died and I am not able to get Premiere Pro 1.5 to work on the new PC. I had several HDD with my video files on and the C: drive died. I have reinstalled the original drives so all my files are recovered and I can see/open them in Premiere BUT I cannot setup to capture anything.
    System is Win XP Pro, Pentium Dual core, 2.7GHz, 2Mb RAM, Capture device is Canopus with DV Storm 2 drivers, Sony DR1000 deck connected by firewire.
    When I go to capture settings there is an IEEE1394 option and DV Capture. If I select Canopus the PD INSTANTLY reboots. I don't know where to start looking for the issue . . . is anybody able to help ?
    Would be much appreciated . . .

    Ian
    Did you replace the whole PC or just the boot drive?
    I may be totally wrong but I had a Storm card years ago and it worked with Premiere up to the old verions 6 or 6.5, Premiere 6.5 is when it  became realtime and hence the Storm card was no longer necessary for anything other than faster mpeg encoding
    I seem to remember that Canopus introduced Lets Edit which then became Edius at this time and they dropped bundling Abobe products so I am surprized that Premiere Pro 1.5 worked with the Storm
    If the Storm cannot be got to work and you do not have a firewire port on the Mobo go to your local PC component supplier and obtain a firewire PCI card with Texas chipset and install it. the card should cost about $20-$30.  The Storms worked a treat but the Mobo had to be right and it did not like certain chipsets on the boards, not only that the installation process was a total pain

  • Issue with running reports on the portal

    Hello Guyz
    1. I have a question regarding running reports on the portal.
    2. I have standard web templates that have been installed in BI.
    3. But in the portal, we are not able to run these reports.
    4. Do we need to create custom I-Views or can we install them from business content?
    5. Can someone explain the process to me AND also any other issues that we might have?
    Thanks.

    Hi Srinivas,
    Standard web templates doesnot get installed with attached queries.Create custom web templates in WAD and attach ur queries based on the type of display u like either tabular or graphical.In order to diplay in portal you have to attach webtemplates to iViews and then group iViews to worksets and attach to portal roles.Then add those roles to portal users.From BI Prospective u just create web-templates and give technical names to EP consultants if you have any then they can takecare of rest.If you want to call standard web template 0ANALYSIS_PATTERN from portal you need to have Business Explorer role added to ur user in EP.Using that template u can only open one query at a time and execute.Hope its clear
    Chandu

  • Issue with running pages through jDev, wls starts but no target url is provided.

    Hi all,
    I recently downloaded an older version of jDev for a new assignment - Studio Edition Version 11.1.1.6.0 - Build JDEVADF_11.1.1.6.0CLOUD_GENERIC_121118.1600.6229.
    I created a basic wls domain and I've tried running simple jspx pages, at which point a weblogic appears to have started without issue. However, my page never runs and no target domain is provided. Below is how my log appears and it just seems to hang. I'm my experience a window always popped up in my default browser with the domain and it is also shown in the log. Going to my expected default domain (http://127.0.0.1:7101/Home.jspx for example) results in a 404. I sure its something stupid, but I'm not sure how to address this issue. Any advice would be appreciated.
    *** Using port 7101 ***
    <Jul 7, 2013 5:01:48 PM EDT> <Notice> <Security> <BEA-090082> <Security initializing using security realm myrealm.>
    <Jul 7, 2013 5:01:52 PM EDT> <Notice> <WebLogicServer> <BEA-000365> <Server state changed to STANDBY>
    <Jul 7, 2013 5:01:52 PM EDT> <Notice> <WebLogicServer> <BEA-000365> <Server state changed to STARTING>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <LoggingService> <BEA-320400> <The log file C:\Users\dslack\AppData\Roaming\JDeveloper\system11.1.1.6.38.62.29\DefaultDomain\servers\DefaultServer\logs\DefaultDomain.log will be rotated. Reopen the log file if tailing has stopped. This can happen on some platforms like Windows.>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <LoggingService> <BEA-320401> <The log file has been rotated to C:\Users\dslack\AppData\Roaming\JDeveloper\system11.1.1.6.38.62.29\DefaultDomain\servers\DefaultServer\logs\DefaultDomain.log00017. Log messages will continue to be logged in C:\Users\dslack\AppData\Roaming\JDeveloper\system11.1.1.6.38.62.29\DefaultDomain\servers\DefaultServer\logs\DefaultDomain.log.>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <Log Management> <BEA-170027> <The Server has established connection with the Domain level Diagnostic Service successfully.>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <WebLogicServer> <BEA-000365> <Server state changed to ADMIN>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <WebLogicServer> <BEA-000365> <Server state changed to RESUMING>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <Server> <BEA-002613> <Channel "Default" is now listening on 127.0.0.1:7101 for protocols iiop, t3, ldap, snmp, http.>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <WebLogicServer> <BEA-000331> <Started WebLogic Admin Server "DefaultServer" for domain "DefaultDomain" running in Development Mode>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <WebLogicServer> <BEA-000365> <Server state changed to RUNNING>
    <Jul 7, 2013 5:02:19 PM EDT> <Notice> <WebLogicServer> <BEA-000360> <Server started in RUNNING mode>
    Regards - Dylan

    Hi dslack,
    Please delete the default domain from the following path. Close Jdev before deleting and restart , it will create a new domain.
    C:\Users\dslack\AppData\Roaming\JDeveloper\system11.1.1.6.38.62.29\DefaultDomain
    Thanks
    Sandeep

  • Issue with Run time parameter values

    Hi All,
    Am facing an error while loading EBS 11.5.10 Financial data. Problem is, my parameters in the SQL Qualifier query (run time parameters in the SQL query) were not being replaced with appropriate values during run time. This is happening for the task "SDE_ORA_Reverse_GLJournals_ImportReference_Extract" that have SQL query in 'mplt_BC_ORA_XactsimpGlrfExtract' mapplet.
    Ihave checked in parameter file and it have values for those Run time parameters. For some reasons, these values were not being picked up by that SQL query.
    Can someboday help me on this about possible reason(s) behind this?

    There are 5 tasks SDE_PSFT_PayrollFact_Total, SDE_PSFT_PayrollFact_Earnings, SDE_PSFT_PayrollFact_EarningsOther, SDE_PSFT_PayrollFact_Deductions, SDE_PSFT_PayrollFact_Taxes are failing all are with same error READER_1_1_1> PR_18012 SQL Error [
    ORA-00936: missing expression Database driver error...
    I verified the mapplet in that missing expression is $$LAST_EXTRACT_DATE value.
    If you know any workaround for this issue please suggest
    I am attaching one of the error logs.
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [PSFT_9_0_HCM] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [12] for mapping parameter:[$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_PSFT_PayrollFact_Total.$$INITIAL_EXTRACT_DATE].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_PSFT_PayrollFact_Total.$$LAST_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [DEFAULT] for mapping parameter:[$$TENANT_ID].
    DIRECTOR> TM_6014 Initializing session [SDE_PSFT_PayrollFact_Total_Full] at [Wed Dec 10 13:41:44 2008]
    DIRECTOR> TM_6683 Repository Name: [Oracle_BI_DW_Base]
    DIRECTOR> TM_6684 Server Name: [Oracle_BI_DW_Base_Integration_Service]
    DIRECTOR> TM_6686 Folder: [SDE_PSFT_90_Adaptor_Test]
    DIRECTOR> TM_6685 Workflow: [SDE_PSFT_PayrollFact_Total_Full]
    DIRECTOR> TM_6101 Mapping name: SDE_PSFT_PayrollFact_Total [version 1]
    DIRECTOR> TM_6827 [C:\Informatica\PowerCenter8.1.1\server\infa_shared\Storage] will be used as storage directory for session [SDE_PSFT_PayrollFact_Total_Full].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6703 Session [SDE_PSFT_PayrollFact_Total_Full] is run by 32-bit Integration Service [node01_HSCHBSCGN20031], version [8.1.1], build [0831].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 Session Sort Order: [Binary]
    MAPPING> TM_6156 Using LOW precision decimal arithmetic
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6307 DTM Error Log Disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> DBG_21075 Connecting to database [orcl], user [obaw]
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_PSFT_PayrollFact_Total_Full]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Wed Dec 10 13:41:45 2008)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Wed Dec 10 13:41:46 2008)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 32000000 bytes and Block size is 1280000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [HR90DEV], user [sysadm]
    READER_1_1_1> DBG_21438 Reader: Source is [HR90DEV], user [sysadm]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8147 Writer: Target is database [orcl], user [obaw], bulk mode [OFF]
    WRITER_1_*_1> WRT_8124 Target Table W_PSFT_PAYROLL_F_TMP :SQL INSERT statement:
    INSERT INTO W_PSFT_PAYROLL_F_TMP(EMPLID,EMPL_RCD,PAYGROUP,PAY_PERIOD_START_DT,PAY_PERIOD_END_DT,CHECK_DT,OFF_CYCLE,PAGE_NUM,LINE_NUM,SEPCHK,COMPANY,DEPTID,LOCATION_CD,PLAN_TYPE,BENEFIT_PLAN,BENEFIT_RCD_NBR,DEDCD,DED_CLASS,DED_SLSTX_CLASS,ADDL_NBR,PAY_EARN_START_DT,PAY_EARN_END_DT,ERNCD,COUNTRY,STATE,LOCALITY,TAX_CLASS,PAYROLL_TYPE,PAY_DETAIL_FLG,PAYROLL_AMOUNT,CHANGED_ON_DT,DATASOURCE_NUM_ID,INTEGRATION_ID,TENANT_ID,X_CUSTOM) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_PSFT_PAYROLL_F_TMP]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    READER_1_1_1> PR_18039 Application Source Qualifier [mplt_BC_PSFT_PayrollFact_Total.Sq_Payroll_Total] is reading from PeopleSoft.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Wed Dec 10 13:41:46 2008
    Target tables:
    W_PSFT_PAYROLL_F_TMP
    READER_1_1_1> PR_18008 Application SQ instance [mplt_BC_PSFT_PayrollFact_Total.Sq_Payroll_Total] User SQL Query [SELECT 
    PS_PAY_CHECK.COMPANY , 
    PS_PAY_CHECK.PAYGROUP , 
    PS_PAY_CHECK.PAY_END_DT , 
    PS_PAY_CHECK.OFF_CYCLE , 
    PS_PAY_CHECK.PAGE_NUM , 
    PS_PAY_CHECK.LINE_NUM , 
    PS_PAY_CHECK.SEPCHK , 
    PS_PAY_CHECK.EMPLID , 
    PS_PAY_CHECK.EMPL_RCD , 
    PS_PAY_CHECK.DEPTID , 
    'TOTAL_GROSS' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_GROSS SUMMARY_AMOUNT, 
    PS_PAY_CHECK.CHECK_DT , 
    PS_PAY_CHECK.LOCATION , 
    PS_PAY_CHECK.UPDATE_DT 
    FROM 
    PS_PAY_CHECK 
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND 
    PS_PAY_CHECK.PAY_END_DT >= 
    UNION ALL
    SELECT 
    PS_PAY_CHECK.COMPANY , 
    PS_PAY_CHECK.PAYGROUP , 
    PS_PAY_CHECK.PAY_END_DT , 
    PS_PAY_CHECK.OFF_CYCLE , 
    PS_PAY_CHECK.PAGE_NUM , 
    PS_PAY_CHECK.LINE_NUM , 
    PS_PAY_CHECK.SEPCHK , 
    PS_PAY_CHECK.EMPLID , 
    PS_PAY_CHECK.EMPL_RCD , 
    PS_PAY_CHECK.DEPTID , 
    'TOTAL_TAXES' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_TAXES SUMMARY_AMOUNT, 
    PS_PAY_CHECK.CHECK_DT , 
    PS_PAY_CHECK.LOCATION , 
    PS_PAY_CHECK.UPDATE_DT 
    FROM 
    PS_PAY_CHECK 
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND 
    PS_PAY_CHECK.PAY_END_DT >= 
    UNION ALL
    SELECT 
    PS_PAY_CHECK.COMPANY , 
    PS_PAY_CHECK.PAYGROUP , 
    PS_PAY_CHECK.PAY_END_DT , 
    PS_PAY_CHECK.OFF_CYCLE , 
    PS_PAY_CHECK.PAGE_NUM , 
    PS_PAY_CHECK.LINE_NUM , 
    PS_PAY_CHECK.SEPCHK , 
    PS_PAY_CHECK.EMPLID , 
    PS_PAY_CHECK.EMPL_RCD , 
    PS_PAY_CHECK.DEPTID , 
    'TOTAL_DEDUCTIONS' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_DEDUCTIONS SUMMARY_AMOUNT, 
    PS_PAY_CHECK.CHECK_DT , 
    PS_PAY_CHECK.LOCATION , 
    PS_PAY_CHECK.UPDATE_DT 
    FROM 
    PS_PAY_CHECK 
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND 
    PS_PAY_CHECK.PAY_END_DT >= 
    UNION ALL
    SELECT 
    PS_PAY_CHECK.COMPANY , 
    PS_PAY_CHECK.PAYGROUP , 
    PS_PAY_CHECK.PAY_END_DT , 
    PS_PAY_CHECK.OFF_CYCLE , 
    PS_PAY_CHECK.PAGE_NUM , 
    PS_PAY_CHECK.LINE_NUM , 
    PS_PAY_CHECK.SEPCHK , 
    PS_PAY_CHECK.EMPLID , 
    PS_PAY_CHECK.EMPL_RCD , 
    PS_PAY_CHECK.DEPTID , 
    'NET_PAY' SUMMARY_TYPE,
    PS_PAY_CHECK.NET_PAY SUMMARY_AMOUNT, 
    PS_PAY_CHECK.CHECK_DT , 
    PS_PAY_CHECK.LOCATION , 
    PS_PAY_CHECK.UPDATE_DT 
    FROM 
    PS_PAY_CHECK 
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND 
    PS_PAY_CHECK.PAY_END_DT >= ]
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Dec 10 13:41:46 2008]
    READER_1_1_1> PR_18012 SQL Error [
    ORA-00936: missing expression
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT
    PS_PAY_CHECK.COMPANY ,
    PS_PAY_CHECK.PAYGROUP ,
    PS_PAY_CHECK.PAY_END_DT ,
    PS_PAY_CHECK.OFF_CYCLE ,
    PS_PAY_CHECK.PAGE_NUM ,
    PS_PAY_CHECK.LINE_NUM ,
    PS_PAY_CHECK.SEPCHK ,
    PS_PAY_CHECK.EMPLID ,
    PS_PAY_CHECK.EMPL_RCD ,
    PS_PAY_CHECK.DEPTID ,
    'TOTAL_GROSS' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_GROSS SUMMARY_AMOUNT,
    PS_PAY_CHECK.CHECK_DT ,
    PS_PAY_CHECK.LOCATION ,
    PS_PAY_CHECK.UPDATE_DT
    FROM
    PS_PAY_CHECK
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND
    PS_PAY_CHECK.PAY_END_DT >=
    UNION ALL
    SELECT
    PS_PAY_CHECK.COMPANY ,
    PS_PAY_CHECK.PAYGROUP ,
    PS_PAY_CHECK.PAY_END_DT ,
    PS_PAY_CHECK.OFF_CYCLE ,
    PS_PAY_CHECK.PAGE_NUM ,
    PS_PAY_CHECK.LINE_NUM ,
    PS_PAY_CHECK.SEPCHK ,
    PS_PAY_CHECK.EMPLID ,
    PS_PAY_CHECK.EMPL_RCD ,
    PS_PAY_CHECK.DEPTID ,
    'TOTAL_TAXES' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_TAXES SUMMARY_AMOUNT,
    PS_PAY_CHECK.CHECK_DT ,
    PS_PAY_CHECK.LOCATION ,
    PS_PAY_CHECK.UPDATE_DT
    FROM
    PS_PAY_CHECK
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND
    PS_PAY_CHECK.PAY_END_DT >=
    UNION ALL
    SELECT
    PS_PAY_CHECK.COMPANY ,
    PS_PAY_CHECK.PAYGROUP ,
    PS_PAY_CHECK.PAY_END_DT ,
    PS_PAY_CHECK.OFF_CYCLE ,
    PS_PAY_CHECK.PAGE_NUM ,
    PS_PAY_CHECK.LINE_NUM ,
    PS_PAY_CHECK.SEPCHK ,
    PS_PAY_CHECK.EMPLID ,
    PS_PAY_CHECK.EMPL_RCD ,
    PS_PAY_CHECK.DEPTID ,
    'TOTAL_DEDUCTIONS' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_DEDUCTIONS SUMMARY_AMOUNT,
    PS_PAY_CHECK.CHECK_DT ,
    PS_PAY_CHECK.LOCATION ,
    PS_PAY_CHECK.UPDATE_DT
    FROM
    PS_PAY_CHECK
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND
    PS_PAY_CHECK.PAY_END_DT >=
    UNION ALL
    SELECT
    PS_PAY_CHECK.COMPANY ,
    PS_PAY_CHECK.PAYGROUP ,
    PS_PAY_CHECK.PAY_END_DT ,
    PS_PAY_CHECK.OFF_CYCLE ,
    PS_PAY_CHECK.PAGE_NUM ,
    PS_PAY_CHECK.LINE_NUM ,
    PS_PAY_CHECK.SEPCHK ,
    PS_PAY_CHECK.EMPLID ,
    PS_PAY_CHECK.EMPL_RCD ,
    PS_PAY_CHECK.DEPTID ,
    'NET_PAY' SUMMARY_TYPE,
    PS_PAY_CHECK.NET_PAY SUMMARY_AMOUNT,
    PS_PAY_CHECK.CHECK_DT ,
    PS_PAY_CHECK.LOCATION ,
    PS_PAY_CHECK.UPDATE_DT
    FROM
    PS_PAY_CHECK
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND
    PS_PAY_CHECK.PAY_END_DT >=
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT
    PS_PAY_CHECK.COMPANY ,
    PS_PAY_CHECK.PAYGROUP ,
    PS_PAY_CHECK.PAY_END_DT ,
    PS_PAY_CHECK.OFF_CYCLE ,
    PS_PAY_CHECK.PAGE_NUM ,
    PS_PAY_CHECK.LINE_NUM ,
    PS_PAY_CHECK.SEPCHK ,
    PS_PAY_CHECK.EMPLID ,
    PS_PAY_CHECK.EMPL_RCD ,
    PS_PAY_CHECK.DEPTID ,
    'TOTAL_GROSS' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_GROSS SUMMARY_AMOUNT,
    PS_PAY_CHECK.CHECK_DT ,
    PS_PAY_CHECK.LOCATION ,
    PS_PAY_CHECK.UPDATE_DT
    FROM
    PS_PAY_CHECK
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND
    PS_PAY_CHECK.PAY_END_DT >=
    UNION ALL
    SELECT
    PS_PAY_CHECK.COMPANY ,
    PS_PAY_CHECK.PAYGROUP ,
    PS_PAY_CHECK.PAY_END_DT ,
    PS_PAY_CHECK.OFF_CYCLE ,
    PS_PAY_CHECK.PAGE_NUM ,
    PS_PAY_CHECK.LINE_NUM ,
    PS_PAY_CHECK.SEPCHK ,
    PS_PAY_CHECK.EMPLID ,
    PS_PAY_CHECK.EMPL_RCD ,
    PS_PAY_CHECK.DEPTID ,
    'TOTAL_TAXES' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_TAXES SUMMARY_AMOUNT,
    PS_PAY_CHECK.CHECK_DT ,
    PS_PAY_CHECK.LOCATION ,
    PS_PAY_CHECK.UPDATE_DT
    FROM
    PS_PAY_CHECK
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND
    PS_PAY_CHECK.PAY_END_DT >=
    UNION ALL
    SELECT
    PS_PAY_CHECK.COMPANY ,
    PS_PAY_CHECK.PAYGROUP ,
    PS_PAY_CHECK.PAY_END_DT ,
    PS_PAY_CHECK.OFF_CYCLE ,
    PS_PAY_CHECK.PAGE_NUM ,
    PS_PAY_CHECK.LINE_NUM ,
    PS_PAY_CHECK.SEPCHK ,
    PS_PAY_CHECK.EMPLID ,
    PS_PAY_CHECK.EMPL_RCD ,
    PS_PAY_CHECK.DEPTID ,
    'TOTAL_DEDUCTIONS' SUMMARY_TYPE,
    PS_PAY_CHECK.TOTAL_DEDUCTIONS SUMMARY_AMOUNT,
    PS_PAY_CHECK.CHECK_DT ,
    PS_PAY_CHECK.LOCATION ,
    PS_PAY_CHECK.UPDATE_DT
    FROM
    PS_PAY_CHECK
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND
    PS_PAY_CHECK.PAY_END_DT >=
    UNION ALL
    SELECT
    PS_PAY_CHECK.COMPANY ,
    PS_PAY_CHECK.PAYGROUP ,
    PS_PAY_CHECK.PAY_END_DT ,
    PS_PAY_CHECK.OFF_CYCLE ,
    PS_PAY_CHECK.PAGE_NUM ,
    PS_PAY_CHECK.LINE_NUM ,
    PS_PAY_CHECK.SEPCHK ,
    PS_PAY_CHECK.EMPLID ,
    PS_PAY_CHECK.EMPL_RCD ,
    PS_PAY_CHECK.DEPTID ,
    'NET_PAY' SUMMARY_TYPE,
    PS_PAY_CHECK.NET_PAY SUMMARY_AMOUNT,
    PS_PAY_CHECK.CHECK_DT ,
    PS_PAY_CHECK.LOCATION ,
    PS_PAY_CHECK.UPDATE_DT
    FROM
    PS_PAY_CHECK
    WHERE
    PS_PAY_CHECK.PAYCHECK_STATUS IN ('A','F','R') AND
    PS_PAY_CHECK.PAY_END_DT >=
    Oracle Fatal Error]
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Dec 10 13:41:46 2008]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_PSFT_PAYROLL_F_TMP] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Wed Dec 10 13:41:46 2008
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_PSFT_PAYROLL_F_TMP (Instance Name: [W_PSFT_PAYROLL_F_TMP])
    WRT_8044 No data loaded for this target
    WRITER_1__1> WRT_8043 ****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [mplt_BC_PSFT_PayrollFact_Total.Sq_Payroll_Total] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [mplt_BC_PSFT_PayrollFact_Total.Sq_Payroll_Total] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_PSFT_PAYROLL_F_TMP] has completed. The total run time was insufficient for any meaningful statistics.
    MANAGER> PETL_24005 Starting post-session tasks. : (Wed Dec 10 13:41:47 2008)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Wed Dec 10 13:41:47 2008)
    MAPPING> TM_6018 Session [SDE_PSFT_PayrollFact_Total_Full] run completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [Sq_Payroll_Total] (Instance Name: [mplt_BC_PSFT_PayrollFact_Total.Sq_Payroll_Total])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_PSFT_PAYROLL_F_TMP] (Instance Name: [W_PSFT_PAYROLL_F_TMP])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_PSFT_PayrollFact_Total_Full] completed at [Wed Dec 10 13:41:48 2008]

Maybe you are looking for

  • Creating the Windows Service for the Database Instance

    Hello, I am using the PeopleTools 8.52 Installation for Oracle documentation and was in Task 6A-2: Creating the Windows Service for the Database Instance. After I type in oradim -new -sid ORCL -intpwd manager -startmode auto -pfile C:\Oracle\database

  • Inheritance in a Relational Database

    Hi Now I've some problem in mapping my class hierarchy in relational database table. How to express inheritance relationship in database? Any advice appreciated victor

  • Can i load a class in subdirectoy  inside a jar file using applet tag?

    hi every one.. thank you for reading ... i am really in dire need for the solution.. my problem is that i have a jar file contianing a package which inturn contains my applet class... i am trying to access this applet class using a applet tag in html

  • Labview run-time 8

    Tengo instalado en mi computador con Windows XP Labview 6.1, y queria abrir unos VI's que baje de internet pero no podia pues necesitaban ser leidos por lo menos en un Labview 7... Por lo que instale el LabVIEW Run-Time Engine 8.0 for Windows 2000/XP

  • Using form with mssql,access and oracle9i

    Is it a good idea to use forms 9i with mssql,access and oracle 9i as back end meaning using forms6i/9i as front end what happens with datatype defined in mssql tables like smallint etc SEJ