Problem loading from DATA MART to ODS, SERVICE-API

Hi gurus,
I have a problem loading data from data mart to ODS, full load.
But, if i try extractor itself (test in RSA3) it works fine.
I already replicated, generated,check transfer rules....datamart but when i try to load data, I get this to messages:
Message no. R3005
"The InfoSource 8TEST_MM specified in the data request, is not defined in the
source system."
Message no. RSM340
Errors in source system.
BTW: This test system was copied from production system. And so far I had no problems with system, but i never tried loading from data marts.
Any ideas?
Regards, Uros

Thanks, for your answer.
I already did that and everything is fine, I see the infosource, and if I test the extractor it works fine, but the infopackage gives me above mentioned errors.
I already looked through notes and I couldn't find anything useful.
I forgot to mention that I generated export data source from transacional ODS.
Regards, Uros

Similar Messages

  • Table on delta loads from data mart

    Hi,
    I am loading data from two DSO's (let's call them A and B) to another DSO (B) in BI 7 with a BW3.5 delta infopackage.
    Now I want to know where I can find information on the timestamp or last request from the last delta load from A and B to C.
    So in fact I would like to know how the system knows which requests in A en B have not been loaded to C yet at the next delta load.
    In which table can I find the information for ODS A and B that is used by the system to define what data in the change log has been loaded or not to ODS C or other targets. (In fact I should have a comparison table in BW as ROOS* in R3)
    Thanks in advance!
    Kind regards,
    Bart

    Hi Guys,
    Thanks for the answers.
    I know how to check everything in the Workbench, but I want to know where the information of the delta is stored technically.
    Just for the sake of completeness:
    Due to some issues; several successive loads from A and B were correctly loaded into DSO C (the 'new' table), but could not be activated. It is not possible to do a repeat or anything else. I am not going into too much detail, but just take this for granted.
    The only way we can 'solve' the problem is to make the system believe that the 3 last loads (activated data) in A and the two last loads in B have not been loaded to C yet. Just deleting the last delta's in C and do a new delta from A and B to C will not work.
    Therefore I want to 'manipulate' the table that is being read by a delta load. If I can change the timestamp or request numbers in that table, I can make the system believe that some requests have not been loaded to C yet.
    Dirty, but it should work I think. But I am still figuring out what table contains information about the datamart data source (8A and 8B) and the last delta load to the targets.
    Hope this is more clear.
    Thanks in advance!
    Kind regards,
    Bart

  • Automatic loading from data marts

    Hi All,
    I have a cube which has a data flow wherein I have an ODS1 which gives data to ODS2 and then finally to the cube.
    I have put this entire process in a process chain.
    But many a times due to an erroneous record,. the ODS1 loading fails and then I have to manually correct the PSA and do a manual activation too in ODS1.
    I
    I now want to know whether the system will pick up the processes again from the process chain and do further loading automatically OR if there is a possibility of automated loading settings be defined in the infopackage so as to do subsequent data targets once I have done it successfully in ODS1???
    Also, if I do such a setting, then the job will be done in my username or BWALEREMOTE??
    Thanks,
    Sharmishtha Biswas

    Thnaks for ur reply,
    But it is possible that the data mart does it automatically??
    As I have observed that some of the subequent infoproviders get loaded automatically after I manually load and activate one data mart.
    I wanted to find an explanation for this..
    Thanks,
    sharmishtha

  • TSV_TNEW_PAGE_ALLOC_FAILED - BCS load from data stream task

    Hi experts,
    We had a short dump when executing BCS Load from Data Stream task. The message is: TSV_TNEW_PAGE_ALLOC_FAILED.
    No storage space available for extending an internal table.
    What happened? How we can solve this error?
    Thanks
    Marilia

    Hi,
    Most likely, the remedy for your problem is the same as in my answer to your another question:
    Raise Exception when execute UCMON

  • Can we use 0INFOPROV as a selection in Load from Data Stream

    Hi,
    We have implemented BW-SEM BPS and BCS (SEM-BW - 602 and BI 7 ) in our company.
    We have two BPS cubes for Cost Center and Revenue Planning and we have Actuals Data staging cube, we use 0SEM_BCS_10 to load actuals.
    We created a MultiProvider on BPS cubes and Staging cube as a Source Data Basis for BCS.
    Issue:
    When loading plan data or Actuals data into BCS (0BCS_C11) cube using Load from Data Stream method, we have performance issue, We automated load process in a Process Chain. Some times it take about 20 hrs for only Plan data load for 3 group currencies and then elimination tasks.
    What I noticed is, (for example/) when loading Plan data, system is also reading Actuals Cube which is not required, there is no selection available in Mapping or selection tab where I can restrict data load from a particular cube.
    I tried to add 0INFOPROV into databasis but then it doen't show up as selection option in the data collection tasks.
    Is there a way where I can restrict data load into BCS using this load option and able to restrict on cube I will be reading data from ?
    I know that there is a filter Badi available, but not sure how it works.
    Thanks !!
    Naveen Rao Kattela

    Thanks Eugene,
    We do have other characteristics like Value Type (10 = Actual and 20 = Plan) and Version (100 = USD Actual and 200 = USD Plan), but when I am loading data into BCS using Load from Data Stream method, the request goes to all the underlying cubes, which in my case are Planning cubes and Actual Cube, but I don't want request to goto Actual Cube when I am running only Plan load. I think its causing some performance issue.
    For this reason I am thinking if I can use 0INFOPROV as we use in Bex queries to filter the InfoProvider so that the data load performance will improve. 
    I was able to to bring in 0INFOPROV into DataBasis by adding 0INFOPROV  in the characteristics folder used by the Data Basis.
    I am able to see this InfoObject Data Stream Fileds tab. I checked marked it to use it in the selection and regenerated the databasis.
    I was expecting that now this field would be available for selection in data collection method, but its not.
    So If its confirmed that there is no way we can use 0INFOPROV as a selection then I would suggest my client for a redesign of the DataBasis it self.
    Thanks,
    Naveen Rao Kattela

  • Issue in loading data from data mart

    Actual load is from APO to BW is happening, but load gets failed when it is loaded from base cube to target cube (data mart), this flow is in 3.5.
    Attached document contains the system msg.

    Your load is still in process.
    Where load failure mesasge?
    Try to activate source 8DSO/Cube and later actiavte update rules thru prorgams.
    Can you check your job details from SM37.
    From info pack monitor, copy request id.
    Go SM37,Job name - <enter copied request id>.
    Use job status - Active/cancel/finished.
    Use proper data ranges.
    User - *

  • Long runtime for loading PCA data into PCA ODS

    hi all,
    hope you can help me regarding this problem:
    i have been trying to load the PCA ODS (standard) from R/3 but it is taking too long so i tried to limit the size by using key fields but no luck. i have tried to load to PSA even that is taking long for around 53440 records i am unable to figure out the problem, i have checked all the transfer rules and update rules every thing is fine.
    can anyone please help me
    thanx a lot in advance

    Hi there,
    don't know if you're loading line items, but following recommendation from SAP shortened our delta loads from 1.5 - 2 hours down to 2 minutes:
    To optimize the performance of data procurement when DataSource 0EC_PCA_3 accesses line item table GLPCA in the OLTP, you should set up an additional secondary index. This should contain the following fields:
    ·         RCLNT
    ·         CPUDT
    ·         CPUTM
    Kind regards
    /martin

  • Problem loading Image data with French accent in file name

    We have a very complex commercial application that has been working perfectly in English for over a year. We have now localized it to French and Spanish, and discovered a serious show-stopping problem. We need help fast, with the scheduled release being less than 2 weeks away...!
    We iterate over a folder to find all the images in it, then display those in a user window.  The code for loading the images is:
                for each (var name:String in filenames){
                    var image:SmoothImageTransformable = new SmoothImageTransformable();
                    var imagePath:String = dir + "/" + name;
                    image.source = imagePath;
                    trace("Adding Item to picture carousel with url " + image.source);
                    image.data = {directory:dir, filename:name};
                    image.isDragEnabled = true;
                    image.setStyle("horizontalCenter", 0);
                    image.setStyle("verticalCenter", 0);
                    _images.push(image);
    Our SmoothImageTransformable class is an internal class that inherits from Image.  This all works fine for image names like 'myFavoritePhoto', 'test', 'photo346', etc.  However, it fails to load the data if the file name is something like 'caméra'  Any name with a UTF-8 character causes the data load to fail.
    To be sure we weren't doing a load somewhere else in the code that I had overlooked, I added a "image.autoLoad = false' line in there.  That did exactly as expected -- all image squares then show a grey blank, with no image shown.  Removing that line (reverting to the original code, as shown above) causes all the images to work perfectly, except for those with French accent characters like the é.
    HELP!  Is this a known bug?  Is there some additional setting I need to make it understand the extended character set?  I can't find any mention about this at all in the documentation.

    @markerline:
    We have a fix!  Almost...   I don't understand one aspect of what I am seeing though.
    We have one test file named "camèra.jpg" that is an example of what I am unable to load into an Image.  I decided to examine the character codes at each letter, and this is what Flex has as the string:
    name = cam?ra.jpg
       [0] = c; 0x99
       [1] = a; 0x97
       [2] = m; 0x109
       [3] = ?; 0x232
       [4] = r; 0x114
       [5] = a; 0x97
       [6] = .; 0x46
       [7] = j; 0x106
       [8] = p; 0x112
       [9] = g; 0x103
    Note that the è prints as gibberish in the trace statement, and shows a numeric encoding of 0x232.  According to
         http://docstore.mik.ua/orelly/web2/action/appb_01.htm
    The unicode equivalent of Latin1 for è is 0xE8, which is equivalent to 0x232. That implies that the string is already Latin1, which I would think would mean that Windoze would recognize the file name and open it properly -- but it does not.
    I then tried taking the "camèra.jpg" string, pasting it into NotePad++ in a UTF8 document, then telling NotePad++ to display as ASCII.  The result is "camèra".  I then tried just forcing the file name within Flex to this new mangled string.  The result then is:
    MAGIC name = camèra
       [0] = c; 0x99
       [1] = a; 0x97
       [2] = m; 0x109
       [3] = ?; 0x195
       [4] = ?; 0x168
       [5] = r; 0x114
       [6] = a; 0x97
       [7] = .; 0x46
       [8] = j; 0x106
       [9] = p; 0x112
      [10] = g; 0x103
    Note this time the trace statement looks proper.  More important, the IMAGE DOES LOAD!  (Our display of the name to the user now looks like crap, but that just means I need to modify our class to have a separate 'name' property, rather than extracting it from the 'image.source')
    So... my remaining question -- what encoding did I just do through that NotePad++ incantation?  That site I quoted above said that it was already Latin1, which is what I thought I was trying to get to.  I clearly need to write a method to map from incoming (UTF8 I thought, but maybe Latin1?) to Windows file system (Latin1 I thought, but maybe ASCII).
    In other words, I am not sure what charcter set to come from or go to...  It turns out the code was already in the character set I had been trying to go to (which is why it wasn't working) and the set I seem to need to go to is not the one I thought...

  • I want to load from two datasources to ODS then ODS to infocube

    Hi Experts,
    I want to load the data from Two datasources to one ODS.
    Once i completed ODS loading i need to load to infocube.
    Please anybody tell me what're the steps to complete this task.
    Regards
    Suresh

    Hi,
    Once you are done creating the ODSs and InfoCube, create an update rule by right clicking the ODS -> create update rule and there you give your InfoSource1 name and create routines if necessary, likewise you would do the same for your InfoSource2 and load the data...
    After successful load from two sources do the samething in Cube, right click on the cube and create update rule but this time you have to mention ODS name instead of InfoSource name. create the routines if necessary then save & activate.
    then come out of the cube then go to ODS generate datasource and once it is done again right click on the ODS -> update ods data into targets -> Full / Initial -> InfoPackage will pop up and you are good to go with your load.
    Hope that helps!!
    ~ Vaishnav

  • Problem load master data

    Hi experts
    I am trying to load master data 0MATERIAL ATTR, there is no error when load into PSA but when I did load into IO it shows error as below:
    Too many error records - update terminated
    Example:
    Record 1 :0MATERIAL : Data record 1 ('000000020000000001 '): Version '000000020000000001 ' is not valiad RSDMD 194
    Please kindly advice if you have any idea how to fix the problem.
    Thanks in advance
    Angel

    Hi Angel,
    There might be a problem in data from source system. Check the tech spec of the info object like type length etc.in BW. then check with the data in PSA.
    Hope this will help.
    Assign Points if helpful
    Regards,
    Ramesh

  • Delta and Full load from a single Delta ODS

    I have delivered 0DS 0GT_DS01 that is being loaded from R/3 as a delta update. This ODS is then loaded to a delivered InfoCube 0GT_C01 as a delta.
    My question is this, can I create update rules and a Infopackage to do a full load from 0GT_DS01 ODS to my custom ODS ZGT_DS01?
    I kind of remember from my BW310 class that in release 3.5,  if I have an ODS to InfoCube Delta update all other updates from that ODS must be a delta also.
    Regards,
    Mike...

    Siva,
    I created my update rule without a problem.
    I created the Infopackage with my ZGT_DS01 ODS at the target, the update mode as FULL and I saved the InfoPackage. After SAVE I went to the drop-down menu, Scheduler ==> Initialization For This Datasource and there is a Delta present.
    The Infopackage works OK when I run is from RSA1, but when I put the Infopackage in Process ChainI get and error "Delta upload is only possible after successful initialization". What I see in the Manage screen for my ZGT_DS01 ODS is a request for a full load followed by a request for a Delta.
    Any Ideas why the infopackage works in RSA1 but not in a process chain?
    Regards,
    Mike...

  • Load from Infocube to an ODS.

    Hi All,
    I needed some information on the logic when you are loading from an Infocube to the ODS.
    I have 2 requests with 100,000 records each in an Infocube.All the keys(dimensions-chars) in the infocube match. The keyfigures in both the requests are different.When i execute  Full load to load the ODS i see only 100,000 records loaded into
    the ODS and the keyfigures have been aggregated.
    Verified the update rules for the keyfigure and the "Update Type" is Overwrite.
    Please help me understand if this is the way its supposed to be workign when loading from an infocube to an ODS.
    thanks in advance

    Hi sandesh,
    As your saying this is how it works"records are aggregated when data is extracted from cube even before DSO update rules comes into act".
    So in this case,i will never be able to overwrite the kefy figures even though i have the update type as "Overwrite".
    Ex- I have 2 reqs in the cube
                 Mat        Plant       KF1   
    Req1     1000        07           10
    Req2     1000        07             5
    Loaded this into the ODS .
    Then i have another request in the cube
    Req3      1000      07        1
    Layter when i load this Req3 into the ODS,how would it show up in.If its aggregated int he cube before its loaded intot he cube,then it will never use the overwrite functionality i ahve selected for the update rules.
    Please help me understand.

  • Deriving Time Characteristics from Date fields in ODS View

    Hi,
    I am creating ODS Views using HANA calculation views.
    is there any direct approach to derive time characteristics like cal week, cal quarter, cal month, calmonth 2, fiscal year, cal year etc from date fields?
    work around:
    Deriving them in HANA view in calculated columns and mapping them to IO created with respective time characteristics as reference.
    is there any standard and easy to use approach for this??
    Thanks in advance.

    Hi Ravi,
    Insteod deriving them in the ODS view you can derive them in tha Calculation View in SAP HANA itself using Time Dimention concept.
    Regards
    Jagan

  • Error message while loading data from data mart

    Please can some one help on this.
    While loading data from one cube to another (I am actually creating a backup of original cube) for 1 record I am getting the error message " internal error occured with time split" The long text of error message says Message no. RSAU101

    Take a look at this thread:
    Internal error occured with time split
    Hope it helps,
    Gilad

  • How to load data from data source to ODS

    Hi all,
    I am using BI 7.0.My requirement is to upload sales order item data in to the DSO(ods).Do I need to create info objects in ODS or by defalut r/3 fileds will come and fill the info objects when i assign my custom ODS.if i need to mailtain info objects in ODS plesae let me know what type of objects i need to maintian in keyfields,data fields and navigation attributs of ODS.
    Thanks
    Raj

    hi raj,
    u will have create infoobjects in BI corresponding to objects in source system.
    create a data source for the source system.
    go to proposal tab u will find fields listed in column.
    u can also watch a perview also.
    give ok
    & next tab is fields tab where u can map the source and target objects .
    u have a template option where u can assign a BI infoobject.
    once assigned the source fields name gets converted into BI name & all properties of BI objects overwrites the source fields.
    now for ODS.
    u have key field & data field.
    key fields usually has characteristics objects
    data field may have characteristics & keyfigure as its component
    navigation atributes u can deploy any attribute object here. providede this attribute has to be specified as navigation attributes in corresponding characteristics object.

Maybe you are looking for