Issues while processing the BOMs using fm ''CSAP_MAT_BOM_MAINTAIN'

Hi Group,
we are facing issues while processing the BOMs using the basic type(Z-idoc type) for standard type BOMMAT04.
thing is that the segment 'E1STPOM' is defined to contain 1 to 99999 records in it.
when an Idoc is run(say with 150/200 segments of 'E1STPOM'), there was a standard error with error message id PIC01(number : 004) when the Idoc is processed with the Standard function module 'CSAP_MAT_BOM_MAINTAIN'.
the error says 'Error reading material FING_005 . Exception: 0' and Check whether the material number is blocked. but when checked, the Material seems not blocked.
kindly advise how to overcome this error and proceed further.
Regards,
Vishnu

Checj the material status, while its transfer from one server to another server the status should be different

Similar Messages

  • Issue while Processing the Huge File in BPEL

    Hi,
    We are facing an Issue while Processing a Hige file in BPEL Process (More than 1MB File). When i test the files with more than 1500 transactions (More than 1MB file) then the BPEL Process is automatically goes to OFF Mode or it goes to Perform Manually Recovery Queue.
    Even we are facing this issue in Production also so we are using UNIX Script to Split the file before place the file in BPEL Input directory.Any Pointers to resolve this issue will be helpful.
    Thanks,
    Saravana

    Hi,
    Please find the answers.
    1. Currently we are using SOA 10.1.2 Version and JDev10g
    2. we are using File Adapeter
    3. yes. We used debatching.
    4. Yes. I am able to recover from Manual Recovery Queue
    5. Please find the error message
    <2009-05-21 04:32:38,461> <DEBUG> <ESIBT.collaxa.cube.engine.dispatch> <Dispatcher::adjustThreadPool> Allocating 1 thread(s); pending threads: 1, active threads: 0, total: 83
    <2009-05-21 04:32:44,077> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> File : /harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B is ready to be processed.
    <2009-05-21 04:32:44,077> <INFO> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Processing file : /harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B
    <2009-05-21 04:32:44,077> <DEBUG> <ESIBT.collaxa.cube.activation> <AdapterFramework::Inbound> onBatchBegin: Batch 'bpel://localhost/ESIBT/BPELProcess_810~1.0//Input5162009.B2B_1242894594000' (/harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B) starting...
    <2009-05-21 04:32:44,077> <DEBUG> <ESIBT.collaxa.cube.translation> <TranslatorFactory::log> Inside TranslatorFactory
    <2009-05-21 04:32:44,078> <DEBUG> <ESIBT.collaxa.cube.translation> <TranslatorFactory::log> using version attribute = NXSD
    <2009-05-21 04:32:44,078> <DEBUG> <ESIBT.collaxa.cube.translation> <TranslatorFactory::log> loading xlator class...oracle.tip.pc.services.translation.xlators.nxsd.NXSDTranslatorImpl
    <2009-05-21 04:32:44,081> <DEBUG> <ESIBT.collaxa.cube.translation> <TranslatorFactory::log> class loaded
    <2009-05-21 04:32:44,081> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Created translator : oracle.tip.pc.services.translation.xlators.nxsd.NXSDTranslatorImpl@46908ae8
    <2009-05-21 04:32:44,098> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Setting up Control dir for debatching error recovery
    <2009-05-21 04:32:44,121> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Control dir for debatching error recovery : /opt01/app/ESIBT/oracle/esibt10gR2iAS/BPEL10gR2/iAS/j2ee/home/fileftp/controlFiles/localhost_ESIBT_BPELProcess_810~1.0_/inbound
    <2009-05-21 04:32:44,121> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Invoking inbound translation for : Input5162009.B2B
    <2009-05-21 04:32:44,121> <DEBUG> <ESIBT.collaxa.cube.translation> <NXSDTranslatorImpl::log> Starting translateFromNative
    <2009-05-21 04:32:44,139> <DEBUG> <ESIBT.collaxa.cube.translation> <NXSDTranslatorImpl::log> Done with translateFromNative
    <2009-05-21 04:32:44,139> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Completed inbound translation for : Input5162009.B2B
    <2009-05-21 04:32:44,139> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> isTextFile : true
    <2009-05-21 04:32:44,139> <INFO> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Translated inbound batch index 1 of file {Input5162009.B2B} with corrupted message count = 1
    <2009-05-21 04:32:44,139> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Error Reader created using charset :ASCII
    <2009-05-21 04:32:44,139> <INFO> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Sending message to Adapter Framework for rejection to user-configured rejection handlers : {
    fileName=/harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B, startLine=1, startColumn=1, endLine=-1, endCol=-1, Exception=ORABPEL-11167
    Error while reading native data.
    [Line=1, Col=70] Expected "\t" at the specified position in the native data, while trying to read the data for "element with name HDR_STORE_NUM", using "style" as "terminated" and "terminatedBy" as "\t", but not found.
    Ensure that "\t", exists at the specified position in the native data.
    <2009-05-21 04:32:44,139> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Setting batchId in NativeRecord to bpel://localhost/ESIBT/BPELProcess_810~1.0//Input5162009.B2B_1242894594000
    <2009-05-21 04:32:44,139> <WARN> <ESIBT.collaxa.cube.activation> <AdapterFramework::Inbound> [Read_ptt::Read(Object)] - onReject: The resource adapter 'File Adapter' requested handling of a malformed inbound message. However, the following bpel.xml activation property has not been defined: 'rejectedMessageHandlers'. Please define it and redeploy the business process. Will use the default Rejection Directory file:///opt01/app/ESIBT/oracle/esibt10gR2iAS/BPEL10gR2/iAS/integration/orabpel/domains/ESIBT/archive/jca/BPELProcess_810/rejectedMessages for now.
    <2009-05-21 04:32:44,140> <WARN> <ESIBT.collaxa.cube.activation> <AdapterFramework::Inbound> [Read_ptt::Read(Object)] - onReject: Sending invalid inbound message to Exception Handler:
    <2009-05-21 04:32:44,140> <INFO> <ESIBT.collaxa.cube.activation> <AdapterFramework::Inbound> Handing rejected message to DEFAULT rejection handler: file:///opt01/app/ESIBT/oracle/esibt10gR2iAS/BPEL10gR2/iAS/integration/orabpel/domains/ESIBT/archive/jca/BPELProcess_810/rejectedMessages since none of the configured rejection handlers [] succeeded.
    <2009-05-21 04:32:44,140> <DEBUG> <ESIBT.collaxa.cube.activation> <AdapterFramework::Inbound> Finished persisting rejected message to file system under the name: /opt01/app/ESIBT/oracle/esibt10gR2iAS/BPEL10gR2/iAS/integration/orabpel/domains/ESIBT/archive/jca/BPELProcess_810/rejectedMessages/INVALID_MSG_BPELProcess_810_Read_20090521_043244_0140.dat
    <2009-05-21 04:32:44,141> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Setting last error record to : -1
    <2009-05-21 04:32:44,141> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Translator has failed to translate any message from batch number: 1
    <2009-05-21 04:32:44,141> <INFO> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Message not published as translation failed: {
    File=/harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B, batchIndex=1, PublishSize=1
    <2009-05-21 04:32:44,141> <ERROR> <ESIBT.collaxa.cube.activation> <AdapterFramework::Inbound> onBatchFailure: Batch 'bpel://localhost/ESIBT/BPELProcess_810~1.0//Input5162009.B2B_1242894594000' (/harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B) has failed due to: ORABPEL-11167
    Error while reading native data.
    [Line=1, Col=70] Expected "\t" at the specified position in the native data, while trying to read the data for "element with name HDR_STORE_NUM", using "style" as "terminated" and "terminatedBy" as "\t", but not found.
    Ensure that "\t", exists at the specified position in the native data.
    <2009-05-21 04:32:44,141> <INFO> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Deleting file : /harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B after processing.
    <2009-05-21 04:32:44,141> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Deleting file : Input5162009.B2B
    <2009-05-21 04:32:44,141> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Deleted file : true
    <2009-05-21 04:32:44,141> <DEBUG> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Removing file /harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B from files to be processed Map.
    <2009-05-21 04:32:44,141> <INFO> <ESIBT.collaxa.cube.activation> <File Adapter::Inbound> Done processing File : /harbinger/prod/xlate/Outbound_Input_LDS_810/Input5162009.B2B
    <2009-05-21 04:33:09,698> <DEBUG> <ESIBT.collaxa.cube.engine.data> <ConnectionFactory::getConnection> GOT CONNECTION 4 Autocommit = false
    For this error message this shows due to some /t its not picking up the file. but even i am facing the same issue for all the files where load is huge.
    Thanks,
    Saravana

  • Issues while generating the output(using VF02- header- output)

    Hi Group,
    I have issues while generating output using the tcode VF02.
    the issue is that, I have got an output type which is configured for the medium "Special Function" and the o/p program generates PDF depending on certain criteria.
    now, when I try to "repeat output" option to generate output again, I was getting errors. it works sometimes by choosing the option "Successfully processed" then selecting the record and then saving it. if I do the same thing again, the output make be running into error or success, it is varying everytime.
    I am not able to understand why the process is not working correctly, when doing the "Repeat output" option on the "Successfully processed" record, the system is giving error instead of a Successful message. it is succeeding once and sometimes not.
    Kindly let me know your inputs if you have encountered same kind of issue.
    Regards,
    Vishnu.

    hi Group,
    when I checked the configuration part of the o/p type I am currently processing, I could see that, in customizing part of this o/p type, there was a checkbox for Multiple Issuing - which is not checked.
    when it is not checked, it should not be possible to run the o/ps for multiple times, but when I tried to generate o/ps again and again, in a few attempts I could to generate the PDFs and in some cases not. I don't know why there is no relevance for this check box Multiple Issuing.
    Could you please let me know why this issuing of o/p is working in some occassions and not working work sometimes, which is  having the configuration of checkbox Multiple Issuing - not being checked.
    thanks for your help in advance.
    Regards,
    Vishnu.

  • Issue while importing the dump using impdp

    **command:**
    While iam using to import the dump with below command iam facing an issue as below,
    impdp pdol2/pdol2@DOLPHDB directory=NEW_IMPORT_DIR table
    existsaction=TRUNCATE EXCLUDE=statistics dumpfile=exportTDOLPHIN2_30062011_pre
    batch_%u.dmp remap_schema=dolphin2:pdol2 logfile=exportTDOLPHIN2_30062011_pdol2.
    log parallel=1 transform=oid:n
    Import: Release 11.2.0.2.0 - Production on Wed Dec 14 18:29:37 2011
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit
    Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-31626: job does not exist
    ORA-31637: cannot create job SYS_IMPORT_FULL_01 for user PDOL2
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPV$FT_INT", line 798
    ORA-39080: failed to create queues "KUPC$C_1_20111214182937" and "" for Data Pum
    p job
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPC$QUE_INT", line 1530
    ORA-04063: package body "SYS.DBMS_AQADM_SYS" has errors
    ORA-06508: PL/SQL: could not find program unit being called: "SYS.DBMS_AQADM_SYS
    Thanks in Advance
    Sivakarthick

    Pl post details of OS version.
    Pl see if these MOS Docs can help
    563701.1 - Data Pump Job Fails With ORA-31626 ORA-31637 ORA-39080 ORA-04063 And ORA-06508
    345198.1 - Impdp or Expdp Fails With ORA-31626 and ORA-31637
    727804.1 - ORA-39080 Failed to Create Queues During Expdp or Impdp
    372961.1 - Ora-00832 No Streams Pool Created while running datapump
    HTH
    Srini

  • Issue while processing large message using schema exposed as WCFservice

    I am using biztalk 2010 and am facing one issue.
    I have a schema exposed as WCF service. Used the ecf pubulishing wizard for the same.
    When we post a large message using this WCF service it throws http 400 error.
    We have isolated the problem to webserver.
    Is there a way to update the MEXbindings to set the request message size.
    If anyone has faced similar problem before request your help

    Hi,
    Following settings in the web.config file of the WCF service may help.
    <bindings>
        <basicHttpBinding>
            <binding
    name="basicHttp" allowCookies="true"
    maxReceivedMessageSize="10000000"
    maxBufferSize="10000000"
    maxBufferPoolSize="10000000">
                <readerQuotas
    maxDepth="32"
    maxArrayLength="100000000"
    maxStringContentLength="100000000"/>
            </binding>
        </basicHttpBinding>
    </bindings>
    HTH
    Sumit
    Sumit Verma - MCTS BizTalk 2006/2010 - Please indicate "Mark as Answer" or "Mark as Helpful" if this post has answered the question

  • SSAS communication link failure issue while processing cube.

    Hi All,
    While processing the cube using full process option I am getting the issue communication link failure. If I process each dimension and measure group manually, I am able to succeed. If any one has faced this issue, please let me know.
    Thanks,
    Moorthy

    Please see:
    https://social.msdn.microsoft.com/Forums/sqlserver/en-US/e32efab7-f06d-405e-98ea-c6565c98e9bc/cube-is-getting-offline-multiple-times-in-a-day?forum=sqlanalysisservices

  • Issue while archiving the processed file in sender communication channel using SFTP adapter

    Hi All,
    In one of my scenario (File to IDOC), we are using SFTP sender communicationchannel.
    we are facing an issue while archiving the processed file. Some times PI processed the file successfully but unable to archive it and in the next poll PI process & archives the same file successfully which will creates duplicate orders in ECC.
    Please let us know how to resolve this issue.

    Hi Anil,
    Refer Archiving concepts in below links.
    http://help.sap.com/saphelp_nw73/helpdata/en/44/682bcd7f2a6d12e10000000a1553f6/content.htm?frameset=/en/44/6830e67f2a6d12e10000000a1553f6/frameset.htm
    http://scn.sap.com/docs/DOC-35572
    Warm Regards,
    DNK Siddhardha.

  • Issue while posting the invoice in background using the WF-BATCH user

    Hi Friends,
      I am facing an issue while posting the invoice in background using the WF-BATCH user. I am using a invoice approval workflow where in when the approver approvers the invoice the invoice document get posted using a background method, which uses BO FIPP and Method POST and i am returning the Message Text to my workflow container from this method. When i see the log an exception is rasied from this method with an error  message "V004: You are not authorized to change this document", but WF-BATCH is having SAP_ALL and SAP_NEW authorizations. If i try to post the invoice using the method from my user id it is getting posted. What could be the issue. Please advice.

    Hi Sapient,
    The Parameter, Roles would be different for the LOGIN USER and WF-BATCH.. So ask your administrator
    to set the Roles & Parameters similar to that of LOGIN USER to WF-BATCH.
    For further refrence check in SU01 giving the LOGIN USER and then check with WF-BATCH... you would
    find the difference...
    Hope this would help you..
    Good luck
    Narin

  • Issue While executing the Query for Pagination using ROWNUM with like

    Issue While executing the Query for Pagination using ROWNUM with like.
    Database is Oracle11G.
    Oracle Database Table contains 8-9 lakh records
    1) SQL equal (=)
    SELECT /*+ FIRST_ROWS(n) */ ROWNUM RNUM, A.* FROM LINE A
    WHERE A.REFERENCE = 'KMF22600920'
    Execution Time:- 0.00869245 seconds
    Returns 2 resultsets
    2) SQL like (one %)
    SELECT /*+ FIRST_ROWS(n) */ ROWNUM RNUM, A.* FROM LINE A
    WHERE A.REFERENCE = 'KMF22600920%'
    Execution Time:- 0.01094301 seconds
    Returns 2 resultsets
    3) SQL like (two%)
    SELECT /*+ FIRST_ROWS(n) */ ROWNUM RNUM, A.* FROM LINE A
    WHERE A.REFERENCE like '%KMF22600920%'
    Execution Time:- 6.43989658 seconds
    Returns 2 resultsets
    In Pagination, we are using Modified version of SQL Query 3) with ROWNUM as mentioned below :-
    4) SELECT * FROM (
    SELECT /*+ FIRST_ROWS(n) */ ROWNUM RNUM, A.* FROM LINE A
    WHERE REFERENCE like '%KMF22600920%' AND ROWNUM <= 20 ) WHERE RNUM > 0
    Execution Time:- Infinite
    ResultSets:- No as execution time is infinite
    a) Instead of like if we use = in the above query it is returning the 2 resultsets (execution time 0.02699282 seconds)
    b) Instead of two % in the above query, if use one example REFERENCE like 'KMF22600920%' it is returning the 2 resultsets (execution time 0.03313019 seconds)
    Issue:- When using two % in like in the above query i.e. REFERENCE like '%KMF22600920%' AND ROWNUM <= 20 ) , it is going to infinite.
    Could you please let us know what is the issue with two % used in like and rownum
    5) Modified version of Option1 query (move out the RNUM condition AND RNUM <= 20)
    SELECT * FROM (
    SELECT /*+ FIRST_ROWS(n) */ ROWNUM RNUM, A.* FROM LINE A
    WHERE REFERENCE like '%KMF22600920%' ) WHERE RNUM > 0 AND RNUM <= 20
    Execution Time:- 7.41368914 seconds
    Returns 2 resultsets
    Is the above query is best optimized query which should be used for the Pagination or still can improve on this ?

    This would be easier to diagnose if there was an explain plan posted for the 'good' and 'bad' queries. Generally speaking using '%' on both sides precludes the use of any indexes.

  • It did not insert the materials in the BOM using 'CSAP_MAT_BOM_MAINTAIN'

    Hi!
      We developed a ABAP technical maintenance BOM.
      In the program.
      Before, we got current BOM and eliminated all related items.
      Routine below.
         LOOP AT INTO it_stpo wa_stpo.
           MOVE-CORRESPONDING TO wa_stpo wa_stpo_i.
           wa_stpo_i-id_itm_ctg = wa_stpo-item_categ.
           wa_stpo_i-id_item_no = wa_stpo-item_no.
           wa_stpo_i-id_comp = wa_stpo-component.
           wa_stpo_i-fldelete = 'X'.
           APPEND TO wa_stpo_i it_stpo_i.
         ENDLOOP.
         CALL FUNCTION 'CSAP_MAT_BOM_MAINTAIN'
           EXPORTING
             Stuff = pv_matnr
             plant = p_werks
             bom_usage = p_tplis
             Alternative = p_altlis
             VALID_FROM = v_valid_from
             i_stko = pv_wa_stko
            TABLES
              t_stpo = it_stpo_i.
    After , we inserted new materials in the BOM using the same function .
    routine below
       LOOP AT INTO it_tab_i wa_tab_i ins_up_del WHERE = '1 ' OR
                                               ins_up_del = '2 ' .
            wa_stpo_i - item_categ = ' L' .
            wa_stpo_i -component = wa_tab_i - matnr_pri .
            v_menge_i = wa_tab_i - menge ( 9 ) .
            v_menge = v_menge_i / 1000 .
            WRITE TO v_menge wa_stpo_i - comp_qty .
            wa_stpo_i - comp_unit = " KG " .
            wa_stpo_i - issue_loc = " COS " .
            APPEND TO wa_stpo_i it_stpo_i .
          ENDLOOP .
    * Including new materials in the BOM .
          CALL FUNCTION ' CSAP_MAT_BOM_MAINTAIN '
            EXPORTING
              Stuff = pv_matnr
              plant = p_werks
              bom_usage = p_tplis
              Alternative = p_altlis
              VALID_FROM = v_valid_from
              i_stko = pv_wa_stko
            IMPORTING
              fl_warning = v_fl_warning
            TABLES
              t_stpo = it_stpo_i
            EXCEPTIONS
              error = 1
              OTHERS = 2 .
    But the new materials were not inserted.
    Could you help me what is missing in the routine ?
    Thanks.

    Hi!
    We found the problem.
    The problem was in the fields ID_ITM_CTG, ID_ITEM_NO and ID_COMP.
      They were not cleaned when we were inserting  the new materials in the BOM.

  • InfoPath form, rich text fields, "There was a form postback error" InvalidOperationException, There has been an error while processing the form

    Using InfoPath 2013 browser enabled form.
    I am getting the above error on ALL Infopath Designed Rich Text fields, where the "Cannot be blank" attribute is set.
    To reproduce it, I create a custom list and custom list form with InfoPath 2013. I add 2 Rich Text fields and enable "cannot be blank". To raise the error, I put some data in the RT field. Skip to another field (so focus is changed and a postback
    occurs), then back to original field to delete the contents (to raise the validation).
    I originally thought it was associated with the HTMLCHKR.DLL not being registered (and I have re-registered it just in case), but the exception I get from the ULS logs reads (it is from a list AFTER I have re-registered the DLL):
    There was a form postback error. (User: 0#.w|myDomain\jc, Form Name: Template, IP: , Request: h t t p ://MyWebApp/MySite/Lists/rtAfterHtmlCHkrReg/Item/newifs.aspx?List=2212ff41-77b4-445b-931b-d7e538c9da91&Source=h t t p://MyWebApp/MySite/Lists/rtAfterHtmlCHkrReg/AllItems.aspx&RootFolder=&Web=3db49106-bdca-47bb-b4cd-a549d2d86aa7,
    Form ID: urn:schemas-microsoft-com:office:infopath:list:-AutoGen-2015-01-16T21:51:48:853Z, Type: InvalidOperationException, Exception Message: No content generated as the result of the operation.) 8cc5e09c-3665-903b-575a-faaac506c40a
    I noticed that errors associated with the HTMLCHKR.DLL not being registered would have some sort of COM exception (example: TYPE_E_LIBNOTREGISTERED or REGDB_E_CLASSNOTREG)
    I also should mention that this problem started happening about 3 weeks ago. We have extended the web application to handle HTTPS on the intranet zone (we had a reverse proxy project that did not eventuate) - would that cause something? How can I do further
    checking?

    Hi,
    I have done a test in my SharePoint, and I met the same issue with you.
    I created a custom list and custom list form with InfoPath 2013. I added 2 Rich Text fields and enabled "cannot be blank".  I put some content in the RT field, then delete the contents, I got the error message:"there has been an
    error while processing the form."
    Here is a similar post said that executing the command: regsvr32 "C:\Program Files\Common Files\Microsoft Shared\
    OFFICE14\htmlchkr.dll" will solve the issue.
    https://social.msdn.microsoft.com/Forums/en-US/eb2e0f6e-c8e4-4e92-ac5e-a09d72759eda/rich-text-field-error-in-webform?forum=sharepointcustomizationprevious
    But I just disabled "cannot be blank", and it solved the issue.
    Best Regards,
    Lisa Chen
    Lisa Chen
    TechNet Community Support

  • Error While processing the COBRA qualified Life event

    Hi All,
    We have setup the Termination Life event as a COBRA qualifying life event.
    Also setup the a new monthly payroll as the default payroll and set the benefit assignment to Yes at BG level.
    I created an employee on 18 aug 2010 , processed new hire life event and enrolled that employee into the COBRA plan named "Be well Medical Plan".
    Then When I terminated the employee as on 18oct2010, While Processing The Termination life event, I am getting an Error saying “When determining pay periods, a payroll Id is required. This error has occurred in the following package : ben_distribute_rates.get_periods_between. "
    When I searched for this error in metalink ,it says it does not have a payroll id ,but when I check on 19oct2010 using the People->Total Comp Participation->Person Benefits Assignment window, he has got the default payroll attached to his benefits assignment.
    Could you please help me to reslove this issue?
    Your help is greatly appreciated.
    Thank you,
    Anuradha

    Hi Anuradha,
    This error comes in many scenarios and there are bug fixes also for them. Please check on all patches available for you apps patchset level having the file benelmen.pkb. You may find a fix for it. (This was a common issue sometime back with suspended/terminated assignments) Also, I suggest logging a SR with oracle asap if your terminated assignment and also the benefits assignment have the correct payrolls attached.
    Regards,
    Vinayaka

  • Issue while playing the content of Azure Media Services after deleting the channel.

    Hi,
    We are having the issue while playing the content of Azure Media Services after deleting the channel.
    Please find detail scenarios below:
    We have done live streaming of an event using Azure Media Services, event was around 5-6 hrs. After completion of event channel was deleted, when we checked content data
    it was only 1hr. Content has only last 1 hr. recording, we are unable to get full recoding.
    We divided event into two parts to avoid above issue. After completion of 1st part which was of 2 hrs., we stopped the streaming and deleted the channel. New channel was
    created and streaming was started for 2<sup>nd</sup> part, after completion of event channel was deleted. Now we are unable to play the content data for 1st part but able to play the 2nd part of event through content publish URL.
    We have done streaming of 3 hrs. event on 18th April with same process and full recoding was available. 
    Can you let me know what might be the issue?

    Hi,
    Please refer to Justin’s blog and see if it helps:
    http://www.justinkobel.com/post/2012/04/30/SharePoint-2010-Claims%E2%80%93Operation-is-not-valid-due-to-the-current-state-of-the-object.aspx
    As suggested by Alex, please provide more information about your process of changing password.
    Here is an article for updating passwords on SharePoint 2010:
    http://blogs.technet.com/b/seanearp/archive/2011/01/25/updating-passwords-on-sharepoint-2010.aspx
    Regards,
    Rebecca Tu
    TechNet Community Support

  • Error occurred while processing the "sales" partition of the "sales" measure group in the cube

    Hi 
    when i ran the job for processing the cube it showing the error like "error occurred while processing the sales partition of the sales measure group in the cube". but in log files no error massage was there. after get that error message we ran
    the cube manually
    at that time the cube executed  successfully. 
    my aim when the job was run it will be process automatically but it is not like that.
    can you suggest  the solution.
    thank you
    satyak248

    Hi Satyak248,
    According to your description, you get the error when using Windows Task to process a cube on a schedule, however can process the cube on SSMS successful manually, right?
    In your scenario, you can process the cube manually, the issue can be cause by Windows Task was not set correctly. So you can try to process the cube using SSIS package. The Analysis Services Processing Task in SQL Server Integration Services (SSIS) allows
    for the processing of one to many to all Analysis Services objects in an SSIS package. Once the SSIS package is created, then a job can be created within the SQL Server Management Studio which will allow for scheduling.
    http://www.mssqltips.com/sqlservertip/2994/configuring-the-analysis-services-processing-task-in-sql-server-2012-integration-services/
    Regards,
    Charlie Liao
    TechNet Community Support

  • Graphing engine is not responding.A fatal error occurred while processing the request. The server responded with: java.lang.NumberFormatException.

    Dear Team,
    I have implemented OBIA 7964 for HRA. Some reports throw the error Graphing engine is not responding.A fatal error occurred while processing the request. The server responded with: java.lang.NumberFormatException. It is not happened all reports. Only some of the reports showed this kind of error. Kindly help me to resolve this issue.
    We have used
    OBIA 7964
    OBIEE 11.1.1.7.140527
    Thanks & Regards,
    Gauthaman S

    Issue was, Font been defined as 7.5 (may be font size decimal not supported ).the problem solved

Maybe you are looking for

  • What is difference between Firefox OS and Android or iOS?

    What is difference between Firefox OS and Android or iOS?

  • CO product..How to allocate the cost

    Hi, Anybody can guide me and thro' some light on how can i allocate the cost to a Co-Product from the main header materials. What is the use of the tab " JOINT PRODUCT " in costing view in the material master,in the same tab  we have something called

  • Cannot connect to database using SQL*Plus

    Hi, I have Oracle 10g XE installed in my labtop and I cannot connect using SQL*Plus. I can connect using the broser User Interface though, which I was able to do after doing the following procedure to change the password of the sys account: -     ope

  • Get "screen junk" after newest update to 10.6.7

    I have a Mac Pro using dual monitor. Every 10 minutes or so the display on one (or sometimes both) monitors goes crazy. Here is a screenshot of what I mean by "screen junk." http://i55.tinypic.com/34rgo41.png It is getting really annoying and makes i

  • Why is my iMac slowing down?

    I've had my iMac for almost five years and it's been great, but for the last few months most of the apps have become very slow to open and operate. Seems like I get the SBOD almost every time I input a command. Are there steps I can take to clean up