Data Load process for 0FI_AR_4  failed

Hi!
I am aobut to implement SAP Best practices scenario "Accounts Receivable Analysis".
When I schedule data load process in Dialog immediately for Transaction Data 0FI_AR_4 and check them in Monitor the the status is yellow:
On the top I can see the following information:
12:33:35  (194 from 0 records)
Request still running
Diagnosis
No errors found. The current process has probably not finished yet.
System Response
The ALE inbox of BI is identical to the ALE outbox of the source system
or
the maximum wait time for this request has not yet been exceeded
or
the background job has not yet finished in the source system.
Current status
No Idocs arrived from the source system.
Question:
which acitons can  I do to run the loading process succesfully?

Hi,
The job is still in progress it seems.
You could monitor the job that was created in R/3 (by copying the technical name in the monitor, appending "BI" to is as prefix, and searching for this in SM37 in R/3).
Keep on eye on ST22 as well if this job is taking too long, as you may have gotten a short dump for it already, and this may not have been reported to the monitor yet.
Regards,
De Villiers

Similar Messages

  • Data load process for FI module

    Dear all,
    We are using BI7.00 and in one of our FI data source 0EC_PCA_1 we had data load failure, the cause for the failure was analysed and we did the following
    1) deleted the data from cube and the PSA
    2) reloaded (full load) data - without disturbing the init.
    This solved our problem. Now when the data reconciliation is done we find that there are doubled entries for some of the G/L codes.
    I have a doubt here.
    Since there is no setup table for FI transactions (correct me if i am wrong), the full load had taken the data which was also present in the delta queue and subsequently the delta load had also loaded the same data
    (some g/l which was available as delta).
    Kindly provide the funtioning of FI data loads. Should we go for a Down time and how FI data loads works without setup tables.
    Can experts provided valuable solution for addressing this problem. Can anyone provide step by step process that has to be adopted to solve this problem permenantly.
    Regards,
    M.M

    Hi Magesh,
    The FI datasources do not involve Setup tables while performing full loads and they do not involve outbound queue during delta loads.
    Full load happens directly from your datasource view to BI and delta is captured in the delta queue.
    Yes you are right in saying that when you did a full load some of the values were pulled that were also present in the delta queue. Hence you have double loads.
    You need to completely reinitialise as the full load process is disturbed. Taking a down time depends on how frequent the transactions are happening.
    You need to.
    1. Completely delete the data in BW including the initialisation.
    2. Take a down time if necessary.
    3. Reintialise the whole datasource from scratch.
    Regards,
    Pramod

  • How to design data load process chain?

    Hello,
    I am designing data load process chains for the first time and would like to get some general information on best practicies in that area.
    My situation is as follows:
    I have 3 source systems (R3 and two for which I use flat files).
    How do you suggest, should I define one big chain for all my loading process (I have about 20 InfoSources) or define a few shorter e.g.
    1. Master data R3
    2. Master data flat file system 1
    3. Master data flat file system 2
    4. Transaction data R3
    5. Transaction data file sys 1
    ... and execute one after another succesful end?
    Could you also suggest me any links or manuals on that topic?
    Thank you
    Andrzej

    Andrzej,
    My advise is to make separate chains for master & transaction data (always load in this order!) and afterwards make a 'master chain' where you insert these 2 chains one after the other (so: Start process -> Master data chain -> Transaction data chain).
    Regarding the separate chains; paralellize as much as possible (if functionally allowed). Normally, the number of parallel ('vertical') chains equals the nr of CPU's available (check with basis-person).
    Hope this provides you with enough info to start off with!
    Regards,
    Marco

  • How to automate the data load process using data load file & task Scheduler

    Hi,
    I am doing Automated Process to load the data in Hyperion Planning application with the help of data_Load.bat file & Task Scheduler.
    I have created Data_Load.bat file but rest of the process i am unable complete.
    So could you help me , how to automate the data load process using Data_load.bat file & task Scheduler or what are the rest of the file is require to achieve this.
    Thanks

    To follow up on your question are you using the maxl scripts for the dataload?
    If so I have seen and issue within the batch (ex: load_data.bat) that if you do not have the full maxl script path with a batch when running it through event task scheduler the task will work but the log and/ or error file will not be created. Meaning the batch claims it ran from the task scheduler although it didn't do what you needed it to.
    If you are using maxl use this as the batch
    "essmsh C:\data\DataLoad.mxl" Or you can also use the full path for the maxl either way works. The only reason I would think that the maxl may then not work is if you do not have the batch updated to call on all the maxl PATH changes or if you need to update your environment variables to correct the essmsh command to work in a command prompt.

  • Automate the data load process using task scheduler

    Hi,
    I am doing Automated Process to load the data in Hyperion Planning application with the help of Task Scheduler.
    I have created Data_Load.bat file but rest of the process i am unable complete.
    So could you help me , how to automate the data load process using Data_load.bat file & task Scheduler.
    Thanks

    Thanks for your help.
    I have done data loading using Data_load.batch file . the .bat file call to .msh file using this i have followed your steps & automatic data lload is possible.
    If you need this then please let me know.
    Thanks again.
    Edited by: 949936 on Oct 26, 2012 5:46 AM

  • Master data Reorganization Process for a single InfoObject

    Hallo Experts,
    mayby someone else had  this problem bevor us?
    SAP-BWProducion: Master data Reorganization Process for a single InfoObject (selfmade), other InfoObjects worked fine.
    SAP BW 7.0 SP20 or 21
    This element has daily attributes and the Q-Table has actual 180Mio entries. We startet reorganisation with 250Mio. entries.
    Because of the daily loading, we cancel the batchprocess in the evening and started it again in the morning. This was fine for about 5 days. We elliminate 80Mio entries.
    Our destination would be 80Mio entries left.
    Whenever we run this feature now  - in a process chain - no entries were deleted after 30 hours working, and there were no counting-stastics in SM50
         Sequentielles Lesen         
         Insert                       
         Update
         Delete
    The five days before, we saw Mio's of entries in this statistics.
    We have checked the Q-table and  there exist a number of records which have the same attribute values across several records and the date from - date to values of these records are in a row.
    RSRV is fine and we checked note 1234411 Master data reorganization - Performance improvement.
    There are no other notes helping us.
    Please give us a solution-idea.
    Thanks
    Santra

    Hi Santra,
    Could you please provide the solution for this?
    Thanks
    Vinod

  • Optimize the data load process into BPC Cubes on BW

    Hello Gurus,
    We like to know how to optimize the data load process and our scenario for this is that we have ECC Classic Ledger,  and we are looking for the best way to load data into the BW Infocubes from an ECC source.
    To complement the question above, from what tables the data must be extracted and then parsed to BW so the consolidation it´s done ?  also, is there any other module that has to be considered from other modules like FI or EC-CS for this?
    Best Regards,
    Rodrigo

    Hi Rodrigo,
    Have you looked at the BW Business Content extractors available for the classic GL? If not, I suggest you take a look. BW business content provides all the business logic you will normally need to get data out of ECC and into BW for pretty much every ECC application component in existence: [http://help.sap.com/saphelp_nw70/helpdata/en/17/cdfb637ca5436fa07f1fdc0123aaf8/frameset.htm]
    Ethan

  • Data loading sequence for 0FIGL_014

    Hi experts,
    Can you explain or brief the data loading sequence for 0FIGL_014?
    Regards.
    Prasad

    Hi,
    Following is my system configuration information
    Software Component      Release             Level                  Highest support                  Short Description
    SAP_BASIS                     620                    0058                   SAPKB62058                      SAP Basis Component
    SAP_ABA                        620                    0058                   SAPKA62058                      Cross Application Component
    SAP_APPL                       470                    0025                   SAPKH47025                       Logistics and Accounting
    SAP_HR                           470                    0030                   SAPKE47030                       Human Resources
    With the above configuration in R/3, I am not able to find the data source 0FI_GL_14.
    Can you please let me know what package to install and how to install?
    Regards.
    Prasad

  • Need data loader script for Buyer Creation for R12

    Hi ,
    Anybody has data loader script for Buyer creation in R12. Please copy paste one line in the reply on this message.
    Thanks

    Hi ,
    Anybody has data loader script for Buyer creation in R12. Please copy paste one line in the reply on this message.
    Thanks

  • Check data load performance for DSO

    Hi,
        Please can any one provide the detials, to check the data load performance for perticular DSO.
       Like how much time it took to load perticular (e.g 200000) records in DSO from R/3 system. The DSO data flow is in BW 3.x version.
    Thanks,
    Manjunatha.

    Hi Manju,
    You can take help of BW statistics and its standard content.
    Regards,
    Rambabu

  • Problems when creating data transfer process for  0EC_CPLAN (Compensation )

    Hi experts,
    I have the infoobject 0EC_CPLAN (Compensation Plan) Active in my BW sistem.
    I tri to create data transfer process for text and atributtes for it, but when i do this i get a "Target 0EC_CPLAN (Type IOBJA) is not active" error message.
    When i double-click on the texts or attributes i recive an  "Administration function is not possible for this InfoObject" information message and i can not active or edit it.
    What could this be? I am breaking my brain.
    Thank you all!

    Debugging i have seen this:
    15         CALL FUNCTION 'RSD_IOBJ_GET'
    16           EXPORTING
    17             i_iobjnm         = l_iobjnm
    18             i_objvers        = rs_c_objvers-active
    19             i_bypass_buffer  = rs_c_false
    20           IMPORTING
    21             e_s_vcha         = l_s_vcha
    22           EXCEPTIONS
    23             iobj_not_found   = 1
    24             illegal_input    = 2
    25             bct_comp_invalid = 3
    26             OTHERS           = 4.
    27         IF sy-subrc <> 0 OR l_s_vcha-infoarea IS INITIAL.
    28           MESSAGE i025(rsawbn).
    l_s_vcha-infoarea is Initial.
    Dues this mean that the object has no-infoarea??? In infoprovider i is under "enterprise compensation management" Does it give you a clue?
    Thanks!

  • Essbase data load process never terminating

    Hi
    We are using Essbase 11.1.2.1.
    We try to load a data file into a BSO Essbase application (essbase export format, so no dataload). We are using "execute in background" option.
    If the file contents no unknown member : it is correctly loaded.
    If the file contents an unknown member : the data load is never terminated : we have killed it after more than 15 hours, and the request is still in "terminating" so we have to kill the essbase process on the server. No error message or file.
    Thanks in advance for your help
    Fanny

    Hi
    I try to start this discussion again, because we still have the same problem, and no idea of why (no idea from the Oracle support too) and my client is waiting for an explanation :-/
    So, some more details :
    Context :
    I have an export (essbase format) from a database.
    If I import this file in a database with all the members used in the file => no problem. The dataload takes around 10 seconds.
    If I import this file in a database with all the members used in the file, excepted 1 => problem. The data load never ends and I have to kill the Essbase application on the Essbase server.
    I have done another test :
    I have imported the file in a database with all the members used in the file (so no problem).
    I have exported level-0 data in columns.
    I have imported this new file in the database with all the members used in the file, using a DLR => no problem. The data load takes around 10 seconds, with a dataload.err generated.
    So, John, you are right, the client shoud really use a column format export. But I need to explain to him why he cannot (not should not) use an export format.
    If using an export format is possible : I have to identify my problem and make the right modification to solve the problem.
    If it not possible : this is a bug, and it is not my responsability anymore and I can close the project!
    Thanks in advance for your great ideas!
    Fanny

  • EDT Cat. 15 Data loaded to SAP but failed when updating

    Hi,
    I have an structure to load data into BP using KCLJ with Cat 15.
    Here it is.
    AKTYP
    TYPE
    PARTNER
    ROLE1
    KUNNR
    KUNNR_EXT
    BU_GROUP
    FIBUKRS
    CHIND_ADDR
    NAME_CO
    NAME_ORG1
    NAME_ORG2
    STREET
    STR_SUPPL1
    STR_SUPPL2
    STR_SUPPL3
    LOCATION
    HOUSE_NUM1
    POST_CODE1
    CITY1
    COUNTRY
    REGION
    LANGU
    CHIND_TEL
    TEL_NUMBER
    TEL_EXTENS
    We had the data loaded to SAP with ROLE 000000 and TR0100 but somehow when we tried to update the address thru ROLE TR0100, we had a dump and failed to update. However, the update is good when we use ROLE 000000 to update.
    Anyone would have a clue of this?
    Thanks

    I am new to EDT also. I know Cat 15 which can load BP. SAP has more categories which can load more using KCLJ. Here's my experience.
    Under Tcode SIMGH, find IMG structure External Data Transfer for SAP Banking. Under that structure, you can find Display Required and Optional Entry Fields for SEM Banking. When you enter 15 in the Category box, you'll see a whole list of fields about BP. You can use these fields to create your own structure.
    After you created your structure, use Define Sender Structure under External Data Transfer for SAP Banking to define your structure. After that, it is done. You can try use KCLJ to load your BP.
    If you still have other issues, mostly will be the configuration.
    Enjoy.

  • EIS to Essbase data load process

    I have two user defined queries in a metaoutline. When i perform dataload operation from EIS, the Essbase application log says that dataload updated [] cells and Data Load Elapsed Time [] seconds but the EIS dataload process is still running. It seems that Essbase is loading the data only for the first query and completing dataload while EIS is still running second query.Has anyone else experienced this problem? Essbase and EIS both are 6.5.0.

    In EIS
    1) You have to define the Logical OLAP Model connecting to the relational source.It defines the joins between fact table and dimension tables.
    2) Based on the OLAP Model You have to create meta outline which defines the rules for loading members and data into essbase.

  • Data Loading Error for cube 0TCT_C22

    Dear Colleagues,
    I am working on BI 7.0 / SP09.
    I am loading technical content cube 0TCT_C22 from datasource 0TCT_DS22. Till PSA, there is no problem with data load. But from PSA to data target, data load fails. Went to monitor and it shows teh error "Error calling number range object 0TCTPRCSCHN for dimension D2 (  ). Message no: RSAU023".
    I tried to find the SAP notes fro this, but no success. Also checked the Dump and application logs, but nothing is there.
    Please advice ASAP.
    Regards
    PS

    Hi Pank,
    I just solved the very similar issue. Try what I did to see if it works for you.  For each dimension in each Infocube a number range is created. For some weird reason during the activation of the Infocube the number range for the dimension giving troubles was not created. Look for it in TCODE SNRO and you should find a number range per each dimension in the cube but the one giving you error.
    To solve it, (the easiest way I found) just add any characteristic to the problematic dimension. Activate the Infocube. After that, modify again you Infocube and remove the characteristic you just added to leave the dimension how you need it. Activate the Infocube again. By doing that you will force the regeneration of the dimension and with it the number range. You can chek in TCODE SNRO and the number range should be there. Try loading your data again and it should work.
    One thing I don't understand is why that number range sometimes is not created during activation
    Good luck, I hope you can solve it!!!
    Regards,
    Raimundo Alvarez

Maybe you are looking for

  • Text to anchored objects

    Hi, Somehow I think my question is quite simple to answer, but I've run out of ideas right now... Is it possible to place text in an anchord object and remove it from it's current story in one action? The words are all labeled (xml), but they need to

  • Itunes will not load all photos

    Suddenly, my Ipad 3 will not load all of the photos that are in a folder onto the iPad into an album.  I have a folder of 58 images and iTunes recognizes that.  When I sync to add the images, itunes loads only 53 of them.  I have repeatedly attempted

  • Very high resolution after update

    I just finished update after 2-weeks holiday, and... X start it very high resolution: 2048x1536. I have 1024x768 in the /etc/X11/xorg.conf, log shows that this file is used. How it is possible that higher mode is initialized? I use Linux for about 10

  • BAPI   for  Changing  Status of  Internal Order Tcode : KO02

    Hello Everybody , I want to change   System status  of  Internal Order   Tcode  :  KO02 . I.e  if  status  are  REL  (  Released  )     then  I  want to make it as  CLSD ( Closed )    and  Vice  Versa . In     AUFK   , AUFK-PHAS2    =    REL  (RELEAS

  • Save 2d array to disc

    I am trying to save 2d  collected data in array indicator to file on a button click , and resetting it  . I can see values filled in Array indicator , but when i save it to file . only three rows are beeing saved. what i am doing wrong ? Thanks in ad