Chain 3.5 with 7.0

HI,
I had a chain in BW 3.5. I update the system to 7.0. It have today a TPD in the final area.
NOw, I want to create a InfoPaquet in the middle, but the systems give a message :" You have to create a infopaket in DWB'.
I created the infoPaket in the DWB but the error or the message appears to.
Any idea?
thanks a lot

Joerg:
What I can see is that your client stops at some point responding to AP/AAA server requests so the retries timr expires and auth process stops. I think it is better to check your macbook because the problem with it not with the wireless network.
Make sure EAP-TLS is correclty configured.
Make sure you get the latest driver for your Wireless adapter.
You can try other EAP methods to isolate further if the problem with all EAP types or with EAP-TLS only.
Was your device working fine before with same network?
I hope this will solve your issue.
Amjad

Similar Messages

  • How  to Implement a Chained Hash Table with Linked Lists

    I'm making a migration from C/C++ to Java, and my task is to implement a Chained Hash Table with a Linked List. My problem is to put the strings(in this case names) hashed by the table using de Division Metod (H(k)= k mod N) in to a Linked list that is handling the colisions. My table has an interface implemented(public boolean insert(), public boolean findItem(), public void remove()). Any Help is needed. Thanks for everyone in advance.

    OK. you have your hash table. What you want it to do is keep key/value pairs in linked lists, so that when there is a collision, you add the key/value pair to the linked list rather than searching for free space on the table.
    This means that whenever you add an item, you hash it out, check to see if there is already a linked list and if not, create one and put it in that slot. Then in either case you add the key/value pair to the linked list if the key is not already on the linked list. If it is there you have to decide whether your requirements are to give an error or allow duplicate keys or just keep or modify one or the other (old/new).
    When you are searching for a key, you hash it out once again and check to see if there is a linked list at that slot. If there is one, look for the key and if it's there, return it and if not, or if there was no linked list at that slot, return an error.
    You aren't clear on whether you can simply use the provided linked-list implementations in the Java Collections or whether you have to cobble the linked list yourself. In any case, it's up to you.
    Is this what you're asking?
    Doug

  • Step by Step documents for Process Chain Error Handling with Screen Shots

    Hi
    Is anybody having Step By Step Documents for Process Chain Error Handling with Screen Shots ?. Please forward it to me to my e-mail [email protected] .  I will reward points to u immediately.
    bye
    Aamira Khan

    Hi,
    You can find lots of topic related to process chain issues in SDN.Please make a search in SDN.
    https://www.sdn.sap.com/irj/sdn/advancedsearch?cat=sdn_library&query=process+chain&adv=true&sdn_author_name=&sdn_updated_on_comparator=ge&sdn_updated_on=&sortby=cm_rnd_rankvalue
    Regards.

  • Daisy chaining external HDD with FW 800? How do I do this?

    I have a new My Book 500G HDD hooked up to my iMac with FW 800. I want to daisy chain a new one to it. Do you have to then use a FW 400 between the two external drives? How do you do this? Does that slow it down noticeably. I use it for HDV editing. Thanks!

    Hi Dan,
    if I recall correctly than the FW800 MyBook has two FW800 ports, so that you can use one for connection to the Mac and one for daisy chain a second with FW800 to the first (assuming the second has also FW800).
    If the second has only FW400 than you have to use the FW400 port of the first.
    And since WD delivers all needed cables with the drives, you should have the needed cable to do so.
    When using FW400 you will have a slowdown, since the harddisk itself has a higher throughput than what FW400 can do.
    Regards
    Stefan

  • Untrusted server cert chain - while connecting with ldap

    Hi All,
    I am getting the following error while running a standalone java program in windows 2000+jdk1.3 environment to connect with LDAP.
    javax.naming.CommunicationException: hostname:636 [Root exception is ja
    vax.net.ssl.SSLException: untrusted server cert chain]
    javax.naming.CommunicationException: hostname:636. Root exception is j
    avax.net.ssl.SSLException: untrusted server cert chain
    at com.sun.net.ssl.internal.ssl.SSLSocketImpl.a(DashoA12275)
    at com.sun.net.ssl.internal.ssl.ClientHandshaker.a(DashoA12275)
    at com.sun.net.ssl.internal.ssl.ClientHandshaker.processMessage(DashoA12
    275)
    at com.sun.net.ssl.internal.ssl.Handshaker.process_record(DashoA12275)
    at com.sun.net.ssl.internal.ssl.SSLSocketImpl.a(DashoA12275)
    at com.sun.net.ssl.internal.ssl.SSLSocketImpl.a(DashoA12275)
    at com.sun.net.ssl.internal.ssl.AppOutputStream.write(DashoA12275)
    at java.io.OutputStream.write(Unknown Source)
    at com.sun.jndi.ldap.Connection.<init>(Unknown Source)
    at com.sun.jndi.ldap.LdapClient.<init>(Unknown Source)
    at com.sun.jndi.ldap.LdapCtx.connect(Unknown Source)
    at com.sun.jndi.ldap.LdapCtx.<init>(Unknown Source)
    at com.sun.jndi.ldap.LdapCtxFactory.getInitialContext(Unknown Source)
    at javax.naming.spi.NamingManager.getInitialContext(Unknown Source)
    at javax.naming.InitialContext.getDefaultInitCtx(Unknown Source)
    at javax.naming.InitialContext.init(Unknown Source)
    at javax.naming.InitialContext.<init>(Unknown Source)
    at javax.naming.directory.InitialDirContext.<init>(Unknown Source)
    at Test2.getProxyDirContext(Test2.java:66)
    at Test2.main(Test2.java:40)
    Any help would be appreciated
    Thanks in Advance
    Somu

    This got resolved when in the code the following
    System.setProperty("javax.net.ssl.tmrustStore", CertFileName);
    where cert file name is the filename with complete path.the file is a CA certificate of the LDAP server
    in X509 format

  • How to daisy chain thunderbolt devices with only one port

    I am using a macbook pro with one thunderbolt port. Quite a few thunderbolt devices have only a single thunderbolt port (don't support daisy chaining).
    Does anyone know of a thuderbolt hub that adds additional thunderbolt ports?
    If such a product does not yet exist, is it at least technically possible?
    Thanks!

    alexgrainger wrote:
    Quite a few thunderbolt devices have only a single thunderbolt port (don't support daisy chaining).
    Does anyone know of a thuderbolt hub that adds additional thunderbolt ports?
    If such a product does not yet exist, is it at least technically possible?
    (1) Thunderbolt does support daisy chaining up to 6 devices to the MBP's port (with some limitations). See
    <http://www.intel.com/content/www/us/en/io/thunderbolt/thunderbolt-technology-dev eloper.html>
    <http://docs.info.apple.com/article.html?path=Mac/10.7/en/mh35952.html>
    <http://www.macworld.com/article/158145/2011/02/thunderbolt_what_you_need_to_know .html>
    <http://en.wikipedia.org/wiki/Thunderbolt_(interface)>
    (2) Note that daisy chain and hub are not the same. However, my understanding is that, in principle, a Thunderbolt hub is feasible. I'm not aware if such a device is already on the market. I guess that whether or not Thunderbolt hubs will be eventually available depends on the demand.

  • Chaining a disk with a monitor

    Is it possible to use Thunderbolt in the following fashion? The LaCie disk of course has two Thunderbolt connectors so that it can be chained. The real question is if it will pass on the DP signal.
    MacBook Air -> Thunderbolt cable -> LaCie Thunderbolt disk -> DP mini to DVI adapter -> DVI-D Monitor?
    Bold = Device, Italic = cable.
    The reason for the adapter is simply that i have a DVI-D KVM switch to toggle between my work box and my gaming PC.

    I have a similar setup:
    MacBook Air --> Thunderbolt Cable --> LaCie Thunderbolt --> mini DP to HDMI adapter --> HDMI display with sound.
    The display works fine at 1920 by 1080 (1080P), but the sound does not. When connected directly
    MacBook Air --> Thunderbolt Cable -->  mini DP to HDMI adapter --> HDMI display with sound.
    The sound works fine and can be controlled through the computer or display. I tried adding a USB speaker, which works but the system does not alwayswakes and just quits when I reboot, have to unplug everything to get it back. I also connected speakers  to the audio out and the same thing happened. Now I am using Bluetooth speakers without any of the previous problems. I like the LaCie it is a 2 TB and very fast so I am living with the sound on Bluetooth.
    This may help provide some information that you can use for your situation.
    Steve C

  • Row chaining in table with more than 255 columns

    Hi,
    I have a table with 1000 columns.
    I saw the following citation: "Any table with more then 255 columns will have chained
    rows (we break really wide tables up)."
    If I insert a row populated with only the first 3 columns (the others are null), is a row chaining occurred?
    I tried to insert a row described above and no row chaining occurred.
    As I understand, a row chaining occurs in a table with 1000 columns only when the populated data increases
    the block size OR when more than 255 columns are populated. Am I right?
    Thanks
    dyahav

    user10952094 wrote:
    Hi,
    I have a table with 1000 columns.
    I saw the following citation: "Any table with more then 255 columns will have chained
    rows (we break really wide tables up)."
    If I insert a row populated with only the first 3 columns (the others are null), is a row chaining occurred?
    I tried to insert a row described above and no row chaining occurred.
    As I understand, a row chaining occurs in a table with 1000 columns only when the populated data increases
    the block size OR when more than 255 columns are populated. Am I right?
    Thanks
    dyahavYesterday, I stated this on the forum "Tables with more than 255 columns will always have chained rows." My statement needs clarification. It was based on the following:
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28318/schema.htm#i4383
    "Oracle Database can only store 255 columns in a row piece. Thus, if you insert a row into a table that has 1000 columns, then the database creates 4 row pieces, typically chained over multiple blocks."
    And this paraphrase from "Practical Oracle 8i":
    V$SYSSTAT will show increasing values for CONTINUED ROW FETCH as table rows are read for tables containing more than 255 columns.
    Related information may also be found here:
    http://download.oracle.com/docs/cd/B10501_01/server.920/a96524/c11schem.htm
    "When a table has more than 255 columns, rows that have data after the 255th column are likely to be chained within the same block. This is called intra-block chaining. A chained row's pieces are chained together using the rowids of the pieces. With intra-block chaining, users receive all the data in the same block. If the row fits in the block, users do not see an effect in I/O performance, because no extra I/O operation is required to retrieve the rest of the row."
    http://download.oracle.com/docs/html/B14340_01/data.htm
    "For a table with several columns, the key question to consider is the (average) row length, not the number of columns. Having more than 255 columns in a table built with a smaller block size typically results in intrablock chaining.
    Oracle stores multiple row pieces in the same block, but the overhead to maintain the column information is minimal as long as all row pieces fit in a single data block. If the rows don't fit in a single data block, you may consider using a larger database block size (or use multiple block sizes in the same database). "
    Why not a test case?
    Create a test table named T4 with 1000 columns.
    With the table created, insert 1,000 rows into the table, populating the first 257 columns each with a random 3 byte string which should result in an average row length of about 771 bytes.
    SPOOL C:\TESTME.TXT
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    INSERT INTO T4 (
    COL1,
    COL2,
    COL3,
    COL255,
    COL256,
    COL257)
    SELECT
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3)
    FROM
      DUAL
    CONNECT BY
      LEVEL<=1000;
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SET AUTOTRACE TRACEONLY STATISTICS
    SELECT
    FROM
      T4;
    SET AUTOTRACE OFF
    SELECT
      SN.NAME,
      SN.STATISTIC#,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SPOOL OFFWhat are the results of the above?
    Before the insert:
    NAME                      VALUE                                                
    table fetch continue        166
    After the insert:
    NAME                      VALUE                                                
    table fetch continue        166                                                
    After the select:
    NAME                 STATISTIC#      VALUE                                     
    table fetch continue        252        332  Another test, this time with an average row length of about 12 bytes:
    DELETE FROM T4;
    COMMIT;
    SPOOL C:\TESTME2.TXT
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    INSERT INTO T4 (
      COL1,
      COL256,
      COL257,
      COL999)
    SELECT
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3)
    FROM
      DUAL
    CONNECT BY
      LEVEL<=100000;
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SET AUTOTRACE TRACEONLY STATISTICS
    SELECT
    FROM
      T4;
    SET AUTOTRACE OFF
    SELECT
      SN.NAME,
      SN.STATISTIC#,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SPOOL OFFWith 100,000 rows each containing about 12 bytes, what should the 'table fetch continued row' statistic show?
    Before the insert:
    NAME                      VALUE                                                
    table fetch continue        332 
    After the insert:
    NAME                      VALUE                                                
    table fetch continue        332
    After the select:
    NAME                 STATISTIC#      VALUE                                     
    table fetch continue        252      33695The final test only inserts data into the first 4 columns:
    DELETE FROM T4;
    COMMIT;
    SPOOL C:\TESTME3.TXT
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    INSERT INTO T4 (
      COL1,
      COL2,
      COL3,
      COL4)
    SELECT
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3)
    FROM
      DUAL
    CONNECT BY
      LEVEL<=100000;
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SET AUTOTRACE TRACEONLY STATISTICS
    SELECT
    FROM
      T4;
    SET AUTOTRACE OFF
    SELECT
      SN.NAME,
      SN.STATISTIC#,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SPOOL OFFWhat should the 'table fetch continued row' show?
    Before the insert:
    NAME                      VALUE                                                
    table fetch continue      33695
    After the insert:
    NAME                      VALUE                                                
    table fetch continue      33695
    After the select:
    NAME                 STATISTIC#      VALUE                                     
    table fetch continue        252      33695 My statement "Tables with more than 255 columns will always have chained rows." needs to be clarified:
    "Tables with more than 255 columns will always have chained rows +(row pieces)+ if a column beyond column 255 is used, but the 'table fetch continued row' statistic +may+ only increase in value if the remaining row pieces are found in a different block."
    Charles Hooper
    IT Manager/Oracle DBA
    K&M Machine-Fabricating, Inc.
    Edited by: Charles Hooper on Aug 5, 2009 9:52 AM
    Paraphrase misspelled the view name "V$SYSSTAT", corrected a couple minor typos, and changed "will" to "may" in the closing paragraph as this appears to be the behavior based on the test case.

  • Best way to know correct forecast model - process chain set up with multiple forecast models

    Hi Experts,
    I need your help in selecting best forecast model for our company. We have some of the models already used for our company, and because of multiple models used it is taking very long time for process chain to finish. There is no existing documentation available on which model was used why initially. Please help me to make out forecasting process smooth.
    - What is the best way to know, which forecast model is correct and should be used for our forecasting process.
    - In case multiple forecasting models are really required to be used, please suggest ways to optimally schedule them in process chain.
    - At times we get messages like "not enough data available" for specific model - any way to avoid this.
    - How to optimally use parallel processing profiles forecasting process in process chain.
    - Things which should be avoided.
    Request your help, please share your experiences.
    Regards
    NB

    Hi Neelesh,
    There are many points you need to consider to redesign forecast process for your company/client.
    You need to select the best suited forecast model first depending on the business. This has to be well tested & agreed by business users. Complexity will be an outcome of this exercise with business users. Best id to give then a brief intro on all available models & then help them selection the best one as per their requirement.
    Auto selection models are generally more time taking & should be used only when you have no idea at all on the business/demand pattern.
    Run time will depend how you are clubbing the CVCs to get the forecast generated & also parallel processing. For parallel processing profile you will need to do trial & error testing along with help from Basis team on how many free dial up processes are available.
    Even you can run many forecast calculations in parallel if the product/cvcs are totally different. - As per my personal experience maximum run time reduction can be achieved here.
    Daily run is not advisable except only for businesses where you have too much dynamism in demand planning i.e. you expect the demands to be changed overnight. Most of the companies run forecast on monthly basis or at weekly basis at the max.
    "Not Enough data" will be a problem if you are having the irrelevant models used in forecast profiles. This means users are not bothered to maintain the needed data for he forecast calculations or they are not aware at all of the situation. Running such models on daily basis is not advised at all. Better users should use interactive forecasting & saving the results in such cases.
    Just to give a crude example we get forecast calculated on monthly basis for approximately 4 lac cvcs in less than 3 hrs using moving avg, seasonal linear regression, seasonal trend, croston models. We use parallel profiles also everywhere with 10 blocks & 500 cvc/block.
    Hope this helps. Let me know if you have nay more questions & also the results using any of this.
    Regards,
    Rahul

  • Export scheduler job but chain step and chain rule failed with ORA-24150 ORA-06512 during executed sql script.

    Hi Folks,
    I used expdp utility to export all Oracle scheduler jobs and chains with below method, after that generate sql script by impdp, later on executing sql script encountered some errors.
    Only chain step and chain rule for executing script. Does anyone bright me some light? Thanks!
    My env: Oracle 11g + Oracle Linux 5.5
    My steps as below:
    1. export(expdp) oracle scheduler job(chain)
    2. generate sql script by impdp.
    3. remove orginal scheduler job(chain)
    4. execute sql script
    5. job with no chain well but job with chain failed
    [oracle@linux1 ~]$ expdp scott/tiger directory=db_dump_dir dumpfile=scott_job.dmp include=procobj:\" in \(select \
    > name from sys.obj$ where type\# in \(46,59,66,67,68,69,72,74,79\)\)\"  schemas=scott
    Export: Release 11.2.0.1.0 - Production on Tue Dec 3 17:42:31 2013
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, Real Application Clusters, OLAP, Data Mining
    and Real Application Testing options
    Starting "SCOTT"."SYS_EXPORT_SCHEMA_01":  scott/******** directory=db_dump_dir dumpfile=scott_job.dmp include=procobj:" in (select name from sys.obj$ where type# in (46,59,66,67,68,69,72,74,79))" schemas=scott
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    Processing object type SCHEMA_EXPORT/POST_SCHEMA/PROCOBJ
    Master table "SCOTT"."SYS_EXPORT_SCHEMA_01" successfully loaded/unloaded
    Dump file set for SCOTT.SYS_EXPORT_SCHEMA_01 is:
      /u03/database/usbo/BNR/dump/scott_job.dmp
    Job "SCOTT"."SYS_EXPORT_SCHEMA_01" successfully completed at 17:42:54
    [oracle@linux1 ~]$ impdp scott/tiger sqlfile=scott_job.sql directory=db_dump_dir dumpfile=scott_job.dmp logfile=imp_scott_job.log
    Import: Release 11.2.0.1.0 - Production on Tue Dec 3 17:43:04 2013
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, Real Application Clusters, OLAP, Data Mining
    and Real Application Testing options
    Master table "SCOTT"."SYS_SQL_FILE_FULL_01" successfully loaded/unloaded
    Starting "SCOTT"."SYS_SQL_FILE_FULL_01":  scott/******** sqlfile=scott_job.sql directory=db_dump_dir dumpfile=scott_job.dmp logfile=imp_scott_job.log
    Processing object type SCHEMA_EXPORT/POST_SCHEMA/PROCOBJ
    Job "SCOTT"."SYS_SQL_FILE_FULL_01" successfully completed at 17:43:07
    [oracle@linux1 ~]$ more /u03/database/usbo/BNR/dump/scott_job.
    scott_job.dmp  scott_job.sql 
    [oracle@linux1 ~]$ more /u03/database/usbo/BNR/dump/scott_job.sql
    -- CONNECT SCOTT
    ALTER SESSION SET EVENTS '10150 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10904 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '25475 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10407 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10851 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '22830 TRACE NAME CONTEXT FOREVER, LEVEL 192 ';
    -- new object type path: SCHEMA_EXPORT/POST_SCHEMA/PROCOBJ
    BEGIN
    BEGIN
    dbms_rule_imp_obj.import_rule('"SCOTT"','"CHAIN_RULE_1"','1=1',NULL, 'First link in the chain.',0,NULL);
    END;
    COMMIT;
    END;
    BEGIN
    BEGIN
    dbms_rule_imp_obj.import_rule('"SCOTT"','"CHAIN_RULE_2"',':"CHAIN_STEP_1".COMPLETED = ''TRUE''',NULL, 'Second link in the chain.',0,NULL);
    END;
    COMMIT;
    END;
    BEGIN
    BEGIN
    dbms_rule_imp_obj.import_rule('"SCOTT"','"CHAIN_RULE_3"',':"CHAIN_STEP_2".COMPLETED = ''TRUE''',NULL, 'Third link in the chain.',0,NULL);
    END;
    COMMIT;
    END;
    BEGIN
    BEGIN
    dbms_rule_imp_obj.import_rule('"SCOTT"','"CHAIN_RULE_4"',':"CHAIN_STEP_3".COMPLETED = ''TRUE''',NULL, 'End of the chain.',0,NULL);
    END;
    COMMIT;
    END;
    BEGIN
    BEGIN
    dbms_rule_imp_obj.import_rule_set('"SCHED_RULESET$1"','"SCHED_EV_CTX$1"',NULL, 0);
    END;
    COMMIT;
    END;
    BEGIN
    dbms_scheduler.create_program('"TEST_PROC_1"','PLSQL_BLOCK',
    'BEGIN
                             INSERT INTO tb_schduler (id, descr, cr_date)
                             VALUES (tb_schduler_seq.NEXTVAL, ''test_proc_1'', SYSDATE);
                             COMMIT;
                           END;'
    ,0, TRUE,
    'Program for first link in the chain.'
    COMMIT;
    END;
    BEGIN
    dbms_scheduler.create_program('"TEST_PROC_3"','PLSQL_BLOCK',
    'BEGIN
                             INSERT INTO tb_schduler (id, descr, cr_date)
                             VALUES (tb_schduler_seq.NEXTVAL, ''test_proc_3'', SYSDATE);
                             COMMIT;
                           END;'
    ,0, TRUE,
    'Program for last link in the chain.'
    COMMIT;
    END;
    BEGIN
    dbms_scheduler.create_program('"TEST_PROC_2"','PLSQL_BLOCK',
    'BEGIN
                             INSERT INTO tb_schduler (id, descr, cr_date)
                             VALUES (tb_schduler_seq.NEXTVAL, ''test_proc_2'', SYSDATE);
                             COMMIT;
                           END;'
    ,0, TRUE,
    'Program for second link in the chain.'
    COMMIT;
    END;
    BEGIN
    dbms_scheduler.create_chain('"TEST_CHAIN_1"', evaluation_interval=>NULL, comments=>'A test chain.'
    , rule_set_name=>'"SCHED_RULESET$1"   '
    dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_1"', program_name=>'"TEST_PROC_1"');
    dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_2"', program_name=>'"TEST_PROC_2"');
    dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_3"', program_name=>'"TEST_PROC_3"');
    COMMIT;
    END;
    BEGIN
    dbms_scheduler.create_job('"TEST_CHAIN_1_JOB"',
    job_type=>'CHAIN', job_action=>
    'test_chain_1'
    , number_of_arguments=>0,
    start_date=>TO_TIMESTAMP_TZ('03-DEC-2013 05.38.56.718161000 PM +08:00','DD-MON-RRRR HH.MI.SSXFF AM TZR','NLS_DATE_LANGUAGE=english'), repeat_interval=>
    'freq=minutely; interval=2'
    , end_date=>TO_TIMESTAMP_TZ('03-DEC-2013 06.08.56.000000000 PM +08:00','DD-MON-RRRR HH.MI.SSXFF AM TZR','NLS_DATE_LANGUAGE=english'),
    job_class=>'"DEFAULT_JOB_CLASS"', enabled=>FALSE, auto_drop=>TRUE,comments=>
    NULL
    COMMIT;
    END;
    [oracle@linux1 ~]$ export ORACLE_SID=usbo
    [oracle@linux1 ~]$ sqlplus / as sysdba
    SQL*Plus: Release 11.2.0.1.0 Production on Tue Dec 3 17:44:43 2013
    Copyright (c) 1982, 2009, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, Real Application Clusters, OLAP, Data Mining
    and Real Application Testing options
    sys@USBO> show parameter db_name
    NAME                                 TYPE                              VALUE
    db_name                              string                            usbo
    sys@USBO> conn scott/tiger;
    Connected.
    --remove job and chain.
    scott@USBO> EXEC DBMS_SCHEDULER.drop_job(job_name => 'test_chain_1_job');
    EXEC DBMS_SCHEDULER.drop_chain (chain_name  => 'test_chain_1');
    EXEC DBMS_SCHEDULER.drop_program (program_name  => 'test_proc_1');
    EXEC DBMS_SCHEDULER.drop_program (program_name  => 'test_proc_2');
    EXEC DBMS_SCHEDULER.drop_program (program_name  => 'test_proc_3');
    PL/SQL procedure successfully completed.
    scott@USBO> @/u03/database/usbo/BNR/dump/scott_job.sql
    Session altered.
    Session altered.
    Session altered.
    Session altered.
    Session altered.
    Session altered.
    PL/SQL procedure successfully completed.
    PL/SQL procedure successfully completed.
    PL/SQL procedure successfully completed.
    PL/SQL procedure successfully completed.
    BEGIN
    ERROR at line 1:
    ORA-24150: evaluation context SCOTT.SCHED_EV_CTX$1 does not exist
    ORA-06512: at "SYS.DBMS_RULEADM_INTERNAL", line 28
    ORA-06512: at "SYS.DBMS_RULE_IMP_OBJ", line 40
    ORA-06512: at line 3
    PL/SQL procedure successfully completed.
    PL/SQL procedure successfully completed.
    PL/SQL procedure successfully completed.
    BEGIN
    ERROR at line 1:
    ORA-24141: rule set SCOTT.SCHED_RULESET$1 does not exist
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.DBMS_ISCHED", line 1694
    ORA-01403: no data found
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 1638
    ORA-06512: at line 5
    PL/SQL procedure successfully completed.

    Thanks all of you!
    Hi DK2010,
    I took some test that the data dict(dba_rule_sets/dba_evaluation_contexts) no any data returned after I had removed the job.
    So I tried to create evaluation context and re-executed script(only exception setion.) the first error has gone. For the second still have some issue.
    ---->no any returned
    scott@USBO> select * from dba_rule_sets where rule_set_owner='SCOTT';       
    no rows selected
    scott@USBO> select * from dba_evaluation_contexts WHERE evaluation_context_owner='SCOTT';
    no rows selected
    -->add new EVALUATION CONTEXT
    scott@USBO> exec DBMS_RULE_ADM.CREATE_EVALUATION_CONTEXT('SCOTT.SCHED_EV_CTX$1');
    PL/SQL procedure successfully completed.
    --->now it looks fine
    scott@USBO> BEGIN
      2  BEGIN
      3  dbms_rule_imp_obj.import_rule_set('"SCHED_RULESET$1"','"SCHED_EV_CTX$1"',NULL, 0);
      4  END;
      5 
      6  COMMIT;
      7  END;
      8  /
    PL/SQL procedure successfully completed.
    --->add new rule set, it prompt aleady exists
    scott@USBO> exec DBMS_RULE_ADM.CREATE_RULE_SET('SCOTT.SCHED_RULESET$1') 
    BEGIN DBMS_RULE_ADM.CREATE_RULE_SET('SCOTT.SCHED_RULESET$1'); END;
    ERROR at line 1:
    ORA-24153: rule set SCOTT.SCHED_RULESET$1 already exists
    ORA-06512: at "SYS.DBMS_RULEADM_INTERNAL", line 28
    ORA-06512: at "SYS.DBMS_RULE_ADM", line 138
    ORA-06512: at line 1
    -->chain rule still could not find.
    scott@USBO> @job_chain_rules.sql
    no rows selected
    -->rerun
    scott@USBO> BEGIN
      2  dbms_scheduler.create_chain('"TEST_CHAIN_1"', evaluation_interval=>NULL, comments=>'A test chain.'
      3  , rule_set_name=>'"SCHED_RULESET$1"   '
      4  );
      5  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_1"', program_name=>'"TEST_PROC_1"');
      6  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_2"', program_name=>'"TEST_PROC_2"');
      7  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_3"', program_name=>'"TEST_PROC_3"');
      8  COMMIT;
      9  END;
    10  /
    BEGIN
    ERROR at line 1:
    ORA-27477: "SCOTT.TEST_CHAIN_1" already exists 
    ORA-06512: at "SYS.DBMS_ISCHED", line 1148
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 1598
    ORA-06512: at line 2
    -->drop chain
    scott@USBO> exec dbms_scheduler.drop_chain('TEST_CHAIN_1');
    BEGIN dbms_scheduler.drop_chain('TEST_CHAIN_1'); END;
    ERROR at line 1:
    ORA-27479: Cannot drop "SCOTT.TEST_CHAIN_1" because other objects depend on it
    ORA-06512: at "SYS.DBMS_ISCHED", line 1319
    ORA-06512: at "SYS.DBMS_ISCHED", line 1222
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 1854
    ORA-06512: at line 1
    scott@USBO> exec dbms_scheduler.drop_chain('TEST_CHAIN_1',force=>TRUE);
    PL/SQL procedure successfully completed.
    scott@USBO> BEGIN
      2  dbms_scheduler.create_chain('"TEST_CHAIN_1"', evaluation_interval=>NULL, comments=>'A test chain.'
      3  , rule_set_name=>'"SCHED_RULESET$1"   '
      4  );
      5  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_1"', program_name=>'"TEST_PROC_1"');
      6  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_2"', program_name=>'"TEST_PROC_2"');
      7  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_3"', program_name=>'"TEST_PROC_3"');
      8  COMMIT;
      9  END;
    10  /
    BEGIN
    ERROR at line 1:
    ORA-24141: rule set SCOTT.SCHED_RULESET$1 does not exist   --->still returned no rule set
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.DBMS_ISCHED", line 1694
    ORA-01403: no data found
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 1638
    ORA-06512: at line 5
    scott@USBO> exec DBMS_RULE_ADM.CREATE_RULE_SET('SCOTT.SCHED_RULESET$1')
    PL/SQL procedure successfully completed.
    scott@USBO> BEGIN
      2  dbms_scheduler.create_chain('"TEST_CHAIN_1"', evaluation_interval=>NULL, comments=>'A test chain.'
      3  , rule_set_name=>'"SCHED_RULESET$1"   '
      4  );
      5  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_1"', program_name=>'"TEST_PROC_1"');
      6  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_2"', program_name=>'"TEST_PROC_2"');
      7  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_3"', program_name=>'"TEST_PROC_3"');
      8  COMMIT;
      9  END;
    10  /
    BEGIN
    ERROR at line 1:
    ORA-27477: "SCOTT.TEST_CHAIN_1" already exists
    ORA-06512: at "SYS.DBMS_ISCHED", line 1148
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 1598
    ORA-06512: at line 2
    scott@USBO> exec dbms_scheduler.drop_chain('TEST_CHAIN_1',force=>TRUE);
    PL/SQL procedure successfully completed.
    scott@USBO> BEGIN
      2  dbms_scheduler.create_chain('"TEST_CHAIN_1"', evaluation_interval=>NULL, comments=>'A test chain.'
      3  , rule_set_name=>'"SCHED_RULESET$1"   '
      4  );
      5  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_1"', program_name=>'"TEST_PROC_1"');
      6  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_2"', program_name=>'"TEST_PROC_2"');
      7  dbms_scheduler.define_chain_step('"TEST_CHAIN_1"', step_name=>'"CHAIN_STEP_3"', program_name=>'"TEST_PROC_3"');
      8  COMMIT;
      9  END;
    10  /
    BEGIN
    ERROR at line 1:
    ORA-24141: rule set SCOTT.SCHED_RULESET$1 does not exist
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.DBMS_ISCHED", line 1694
    ORA-01403: no data found
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 1638
    ORA-06512: at line 5
    Would you like to give me more clue?
    Thanks again.

  • Security Provider(JAAS chaine module)logging with SAP J2EE Agent Policy 2.2

    Hello,
    I have installed and configured a SAP J2EE Policy Agent 2.2 on a SAP J2EE Application Server 7.0 and installed an opensso 8.0.
    I have configured and deployed a jsp application with the descriptor containing the agent filter on the SAP J2EE.
    I have configured the security provider (based on JAAS) as follows:
    EvaluateTicketModule
    AmSAPWASLoginModule
    CreateTicketLoginModule
    I have then increased the severity check of logging on the SAP J2EE Application Server for:
    com.sap.security.server.jaas
    com.sap.engine.services.security
    But since the installation of the agent, I am not able to see the logs of the different logon modules and be sure that the CreateLoginModule has been done.
    I need to check that after the success of the agent login module, it goes through the CreateTicketLoginModule, how can I check that.
    Thanks,
    Tanguy Mezzano
    amFilter log:
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    AmFilter: now processing: Notification Task Handler
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    AmFilter: now processing: Port Check Task Handler
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    PortCheckTaskHandler: request is on valid port
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    AmFilter: now processing: FQDN Task Handler
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    FQDNHelper: Incoming Server Name: [myserver] Result: null
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    AmFilter: now processing: Not Enforced List Task Handler
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    NotenforcedURIHelper.isNotEnforced(/myApp.jsp) found in enforced cache
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    NotenforcedURIHelper.isNotEnforced(/myApp.jsp) => false
    11/13/2008 06:02:39:810 PM CET: SAPEngine_Application_Thread[impl:3]_35
    AmFilter: now processing: SSO Task Handler
    11/13/2008 06:02:40:122 PM CET: SAPEngine_Application_Thread[impl:3]_35
    SSOTaskHandler: SSO Validation successful for id=myUser,ou=user,dc=myCompany,dc=com
    11/13/2008 06:02:40:122 PM CET: SAPEngine_Application_Thread[impl:3]_35
    AmFilter: now processing: URL Policy Task Handler
    11/13/2008 06:02:40:310 PM CET: SAPEngine_Application_Thread[impl:3]_35
    URLPolicyTaskHandler: access allowed by AmWebPolicy
    11/13/2008 06:02:40:310 PM CET: SAPEngine_Application_Thread[impl:3]_35
    AmFilter: now processing: Redirect Check Result Handler
    11/13/2008 06:02:40:310 PM CET: SAPEngine_Application_Thread[impl:3]_35
    RedirectCheckResultHandler: removing redirect counter cookie
    11/13/2008 06:02:40:310 PM CET: SAPEngine_Application_Thread[impl:3]_35
    AmFilter: result =>
    FilterResult:
    Status : CONTINUE
    RedirectURL : null
    RequestHelper:
    null
    Data:
    null
    -----------------------------------------------------------

    I want to use my login page in BEA instead of Access Manager or Distributed Authentication Module. I think it is similar to the problem you have? Did you come up with a solution?
    My problem is described in more detail on this thread:
    http://forum.java.sun.com/thread.jspa?threadID=5197783

  • BW chains in Error status - processes within the chain not aligned with BW

    Hi,
    Have anyone encountered scenario where processes within BW process chain are not listed in the correct sequence when chain gone to Error status?
    If we need to skip to the next process, ignoring the current error process (w/o referring to the chain in BW), I would have picked the incorrect process to reply to Operator.
    Any thoughts?
    Thanks,
    Kevin Ng

    No answer
    I have impl. a workaround: I trigger event, which starts the precalculation of WebTemplates.
    p.

  • Supply Chain working with SAP

    Can Supply Chain co-exist with an SAP implementation? And, to what extent?
    Thanks

    Today, any integrated system must include 'supply chain' as a part of the system. It has to have a 'chain link' to the extent of input suppliers & output customers, and beyond.
    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by kruiz_us:
    Can Supply Chain co-exist with an SAP implementation? And, to what extent?
    Thanks<HR></BLOCKQUOTE>
    null

  • Process Chain that uses a DTP and locks up with CX_RSBK_REQUEST_LOCKED

    Hello Experts,
    I am making a program to start a PC inside a loop. This Loop has several dates and ODS ID's. The problem is that inside the loop I make a call to FM RSPC_API_CHAIN_START, after this I have a WAIT UNTIL 30 SECONDS command so that the process has enough time to finish loading data. Each time the loop starts I change the name of the .txt flat file that te PC is going to load and store it inside a Z table with a rutine inside the DTP calling it every loop cycle. And so every time the Process Chain starts it should load a different file into a specifyc ODS. The problem comes at the second loop in which through the RSA1 transaction I find out that the chain is locked with an exception CX_RSBK_REQUEST_LOCKED and the subsequent calls to the Process Chain simply don't start at all. One solution I found was to delete the last Petition Request ID in the table RSSTATMANPART.
    I use RSSM_DELETE_REQUEST to delete that request using that ID and the cube name.
    Now this doesn't work either.
    PROGRAM SUMMARY:
    LOOP WITH ODS THAT NEED LOADING INTO BW ODS
        DELETE PETITION REQUESTS FROM RSSTATMANPART WITH YELLOW OR RED STATUS.(So that PC can start)
        CHANGE FILE THAT DTP ROUTINE WILL LOAD
        START PROCESS CHAIN
        WAIT A CERTAIN AMOUNT OF TIME FOR PC TO FINISH
    ENDLOOP.
    Edited by: Antonio Alejandro Ortega Vergani on Apr 12, 2011 12:05 AM

    Hi saveen,
    I've checked the transaction sm50 to check the available processes and I have 8 BTC(batch), 5 DIA(dialog), 2 UPD(update), 1 ENQ(enqueue) and 1 SPO(spool). Are these the ones you are asking about? I assume you refer to BTC which I believe 8 is enough....correct me if Im wrong here as Im starting out with this please.
    Thing is that Im trying to automate this whole process (for loading various file types(headers.txt, details.txt, payments.txt, etc), for various centers and several months...this is why I want to automate the whole process, so I dont have to go manually through all this data loading. So what I did was try to start the chain inside a loop for every new combination of dates that DIDN'T EXIST in the ODS tables. The process chain is supposed to call a DTP that uses a routine to see which file is going to upload to BW but only does it the first time.
    The problem comes with the second cicle of the loop in which the petition request for the DTP is already green, nevertheless the next start of the chain doesn't come as there seems to be no creation of another request and the exception CX_RSBK_REQUEST_LOCKED in the transaction RSPC.
    Appart from this whenever I try to delete the last request so that this way it does create another request(rnr) in RSSTATMANPART through the use of RSPC_API_CHAIN_START the data in my ODS does not stay and is deleted.

  • Help with Process Chain - Stop remaining Processes if Package returns 0 rec

    Hi, i have a Process Chain that starts with a single InfoPackage.  Then it goes on to distribute the load to multiple DSO's and then onto DSO Activation.  I would like the Process Chain to stop right after the InfoPackage if no records are loaded. 
    Is this possible using the decision variant somehow?  Any details on how to achieve this are greatly appreciated!

    Hi ,
    The request is goin to red status when 0 records are pulled in. There is a option in info package so that even when 0 records come in the request status would be green.
    Open your info package , then in the menu goto "Scheduler > Treatment of warnings > Select Green" so that in future even if the info package pulls 0 records the status would be green . then the activation step in the process chain would also be carried out smoothly.
    In this way you need not manually change any request status to green . Saves you time.
    regards
    Venky

Maybe you are looking for