BAdI UC_DATATRANSFER for BCS Mapping in "Load from Data Stream" method

Hello Everyone,
I need some help on finishing up the code for the UC_DATATRANSFER BAdI.
I have looked up in the SDN and other places, but could not get comprehensive breakdown of documentation except for the "F1" documentation available on the BAdI.
So, any help would be appreciated.
The Steps so far completed,
1. Have activated the BAdI and have created the filter value for the BAdI.
2. After the BAdI has been activated, I was able to go into the MAP method and have written the logic for profit center derivation from consolidation hierarchy.
The issue is there are four components for the Map method,
IT_DATA_SOURCE
IS_DATA_TARGET
ES_DATA_TARGET
ET_DATA_TARGET
The data is available from Source system in the table IT_DATA_SOURCE.
But this is not changeable as it is "Importing" type. Whereas the actual ET_DATA_TARGET which is passed over into FINALIZE method of the BAdI is not filled initially.
When I try to do a MOVE-CORRESPONDING from the IT_DATA_TARGET into ET_DATA_TARGET I continuously am getting the short dumps as both the tables length is not the same.
Did anyone else face the same issue as above when trying to do the BAdI implementation for Mapping.
I will really appreciate if any one can provide me a sample code if possible.
Let me know if you need additional information.
Thanks
Dharma.

Hello,
Thanks for looking into the question.
I already had tried doing that, I get the Short dump stating the object tables are not convertible.
When I looked into the table structures, I found out that the table structures "IS_DATA_TARGET", "ES_DATA_TARGET" & "ET_DATA_TARGET" belong to the same category in terms of these structures being flat structures or tables of length 484 as per the debugger.
Whereas the structure "IT_DATA_SOURCE" has the length 404.
Due to this reason when I say,
ET_DATA_TARGET = IT_DATA_SOURCE, I keep getting the short dumps.
Also, is your consolidation process legal or managerial.
Our Consolidation process is legal and we have the Company and Profit Center fields assigned to the Consolidation Unit role in the Data Basis definition.
Can you please let me know what is the structures length in your system.
Thanks
Dharma.

Similar Messages

  • Can we use 0INFOPROV as a selection in Load from Data Stream

    Hi,
    We have implemented BW-SEM BPS and BCS (SEM-BW - 602 and BI 7 ) in our company.
    We have two BPS cubes for Cost Center and Revenue Planning and we have Actuals Data staging cube, we use 0SEM_BCS_10 to load actuals.
    We created a MultiProvider on BPS cubes and Staging cube as a Source Data Basis for BCS.
    Issue:
    When loading plan data or Actuals data into BCS (0BCS_C11) cube using Load from Data Stream method, we have performance issue, We automated load process in a Process Chain. Some times it take about 20 hrs for only Plan data load for 3 group currencies and then elimination tasks.
    What I noticed is, (for example/) when loading Plan data, system is also reading Actuals Cube which is not required, there is no selection available in Mapping or selection tab where I can restrict data load from a particular cube.
    I tried to add 0INFOPROV into databasis but then it doen't show up as selection option in the data collection tasks.
    Is there a way where I can restrict data load into BCS using this load option and able to restrict on cube I will be reading data from ?
    I know that there is a filter Badi available, but not sure how it works.
    Thanks !!
    Naveen Rao Kattela

    Thanks Eugene,
    We do have other characteristics like Value Type (10 = Actual and 20 = Plan) and Version (100 = USD Actual and 200 = USD Plan), but when I am loading data into BCS using Load from Data Stream method, the request goes to all the underlying cubes, which in my case are Planning cubes and Actual Cube, but I don't want request to goto Actual Cube when I am running only Plan load. I think its causing some performance issue.
    For this reason I am thinking if I can use 0INFOPROV as we use in Bex queries to filter the InfoProvider so that the data load performance will improve. 
    I was able to to bring in 0INFOPROV into DataBasis by adding 0INFOPROV  in the characteristics folder used by the Data Basis.
    I am able to see this InfoObject Data Stream Fileds tab. I checked marked it to use it in the selection and regenerated the databasis.
    I was expecting that now this field would be available for selection in data collection method, but its not.
    So If its confirmed that there is no way we can use 0INFOPROV as a selection then I would suggest my client for a redesign of the DataBasis it self.
    Thanks,
    Naveen Rao Kattela

  • TSV_TNEW_PAGE_ALLOC_FAILED - BCS load from data stream task

    Hi experts,
    We had a short dump when executing BCS Load from Data Stream task. The message is: TSV_TNEW_PAGE_ALLOC_FAILED.
    No storage space available for extending an internal table.
    What happened? How we can solve this error?
    Thanks
    Marilia

    Hi,
    Most likely, the remedy for your problem is the same as in my answer to your another question:
    Raise Exception when execute UCMON

  • Problem loading from DATA MART to ODS, SERVICE-API

    Hi gurus,
    I have a problem loading data from data mart to ODS, full load.
    But, if i try extractor itself (test in RSA3) it works fine.
    I already replicated, generated,check transfer rules....datamart but when i try to load data, I get this to messages:
    Message no. R3005
    "The InfoSource 8TEST_MM specified in the data request, is not defined in the
    source system."
    Message no. RSM340
    Errors in source system.
    BTW: This test system was copied from production system. And so far I had no problems with system, but i never tried loading from data marts.
    Any ideas?
    Regards, Uros

    Thanks, for your answer.
    I already did that and everything is fine, I see the infosource, and if I test the extractor it works fine, but the infopackage gives me above mentioned errors.
    I already looked through notes and I couldn't find anything useful.
    I forgot to mention that I generated export data source from transacional ODS.
    Regards, Uros

  • Error while loading Reported Financial Data from Data Stream

    Hi Guys,
    I'm facing the following error while loading Reported Financial Data from Data Stream:
    Message no. UCD1003: Item "Blank" is not defined in Cons Chart of Accts 01
    The message appears in Target Data.  Item is not filled in almost 50% of the target data records and the error message appears.
    Upon deeper analysis I found that Some Items are defined with Dr./Cr. sign of + and with no breakdown.  When these items appear as negative (Cr.) in the Source Data, they are not properly loaded to the target data.  Item is not filled up, hence causing the error.
    For Example: Item "114190 - Prepayments" is defined with + Debit/Credit Sign.  When it is posted as negative / Credit in the source data, it is not properly written to the target.
    Should I need to define any breakdown category for these items?  I think there's something wrong with the Item definitions OR I'm missing something....
    I would highly appreciate your quick assistance in this.
    Kind regards,
    Amir

    Found the answer with OSS Note: 642591..... 
    Thanks

  • Error in code page mapping for Source system whil loading the data from ECC

    HI Gurus,
    I am working in a implementation project, Recently our BI sand box is up, when i am doing my load from 0comp_code_attr it is throwing an error "Error in code page mapping for source system"(This is my first load from ECC).
    In details tab it is showing as the data is sent from the source system but the data is not reaching to the PSA.
    Please let me know if there are any settings needs to be made.
    Many thanks in Advance
    Jagadeesh

    HI V,
    Thanks for your quick response. I did it but it didn't resolve the issue. since the system id which i am having is 3 digits(LEC) but there it is taking only 2 digits, so i clicked on the button called Propose system ids, it has praposed LE, but the issue is stil there.
    Do we need to do any settings in LBWE??
    Thanks and Regards
    Jagadeesh

  • Process for Flat file delta load from Application server to transactional??

    I want to do a flat file extraction in delta loads from application server and that should be loaded into a transactional cubes scheduling a process chain.
    Can any one help with ABAP code and step by step process for this?
    Thank you
    Devi

    Hi Devi,
    As per your explaination, you want to load a list of files from application server using process chain.
    You can do this with below steps.
    1) Create a Event and call event to run Process chain multiple times.
    2) Put all filenames in a file say, source-File
    3) Write a rooting to read filename from Source-File and trigger InfoPackage
    4) At the end of the process chain create new customized program which will delete first line from Source-File so that it will take next file in next run as well as delete currently loaded file. Once this is done trigger event again to start process chain again.
    Regards,
    Ganesh

  • Table on delta loads from data mart

    Hi,
    I am loading data from two DSO's (let's call them A and B) to another DSO (B) in BI 7 with a BW3.5 delta infopackage.
    Now I want to know where I can find information on the timestamp or last request from the last delta load from A and B to C.
    So in fact I would like to know how the system knows which requests in A en B have not been loaded to C yet at the next delta load.
    In which table can I find the information for ODS A and B that is used by the system to define what data in the change log has been loaded or not to ODS C or other targets. (In fact I should have a comparison table in BW as ROOS* in R3)
    Thanks in advance!
    Kind regards,
    Bart

    Hi Guys,
    Thanks for the answers.
    I know how to check everything in the Workbench, but I want to know where the information of the delta is stored technically.
    Just for the sake of completeness:
    Due to some issues; several successive loads from A and B were correctly loaded into DSO C (the 'new' table), but could not be activated. It is not possible to do a repeat or anything else. I am not going into too much detail, but just take this for granted.
    The only way we can 'solve' the problem is to make the system believe that the 3 last loads (activated data) in A and the two last loads in B have not been loaded to C yet. Just deleting the last delta's in C and do a new delta from A and B to C will not work.
    Therefore I want to 'manipulate' the table that is being read by a delta load. If I can change the timestamp or request numbers in that table, I can make the system believe that some requests have not been loaded to C yet.
    Dirty, but it should work I think. But I am still figuring out what table contains information about the datamart data source (8A and 8B) and the last delta load to the targets.
    Hope this is more clear.
    Thanks in advance!
    Kind regards,
    Bart

  • Status in the request for a infocubo to load from a ods.

    I have a ods that after loading data has to upgrade a infocubo, but after conducting the load on the infocubo, although this has been done correctly, is not the traffic lights green, yellow staying until I manually put it in green. Please, which must be realized to automatically after loading the request of infocubo gets green?.
    Thanks.

    Hi,
    Go to infocube - Manage - In menubar - Envoirnment -  Set quality status to O.k ..
    Check whether flag is enabled for quality status to o.k ?? if not enable the flag..

  • Code for  File upload / down load from WD java application

    Hi experts ,
    I am creating a WD java application , I want to add Picture upload and piucture download ( from / To ) KM . I also want to specify the path ( where it has to be uploaded in KM ) and picture name . can i get any code so that i can use it in WD application .
    Regards
    Abhay

    First you need to make sure your web dynpro application has permissions to access the file in the first place, but you could use the standard java.io.File class:
    File file = new File("/path/file.sh");
    file.setExecutable(false);
    file.setReadable(false);
    file.setWritable(false);

  • Automatic loading from data marts

    Hi All,
    I have a cube which has a data flow wherein I have an ODS1 which gives data to ODS2 and then finally to the cube.
    I have put this entire process in a process chain.
    But many a times due to an erroneous record,. the ODS1 loading fails and then I have to manually correct the PSA and do a manual activation too in ODS1.
    I
    I now want to know whether the system will pick up the processes again from the process chain and do further loading automatically OR if there is a possibility of automated loading settings be defined in the infopackage so as to do subsequent data targets once I have done it successfully in ODS1???
    Also, if I do such a setting, then the job will be done in my username or BWALEREMOTE??
    Thanks,
    Sharmishtha Biswas

    Thnaks for ur reply,
    But it is possible that the data mart does it automatically??
    As I have observed that some of the subequent infoproviders get loaded automatically after I manually load and activate one data mart.
    I wanted to find an explanation for this..
    Thanks,
    sharmishtha

  • Oracle BI Publisher no columns loaded from data model

    Hello Guys,
    I use Oracle BI Publisher 11g. Have created a data model by using
    the JDBC thin connection to a 11g database. The creation of the data model
    worked fine using a SQL query. Now in the report creation wizard, I select
    the "guide me" option to see how nice the wizard is.
    The problem comes here, on the second window I should select the columns
    but I only read "no columns available". Why is that?

    Hi Metalray
    For BIP report sample XML data is importent. with sample data you cannot generate the BIP reports.
    Follow the below steps, you can achieve your expected BIP output
    After crating data set in the data medel save the data model then click on
    --get XML output option on right side of the top-corner.
    -- then selet no. for the nuber of rows to return dropdown option,
    --then click on Run. Then save the sample data by clicking on Save as sample data option.
    --then it will take you to the data model page, now again save the data model
    --and use it for creating the reports.
    Now you were able to see the available columns in the the second window of guide me option
    please mark if it helps you..

  • Removing zeros from data stream

    Hi
    I have incoming data (plz see attached diagram which shows the 2 states of the for loop '0' and default) which is a 1-D array of 64 bit real data .....it goes through the loop which removes zeros from the array.
    As I have 3 elements in the 1-D array (call them x,y,z ), the loop works very well for all values of 'y' and 'z' .
    However when x gets towards 1 and below say 10e-3 (it never is a negative number), the loop is rounding everything to 1 and then when the value gets between 1 and 0 (10e-3 for example) the loop leaves it out altogether and I just get y,z  values saved.
    Plz help me sort this out .
    Cheers
    Baz
    Attachments:
    zeros.PNG ‏25 KB
    zeros.PNG ‏25 KB

    If you want a tolerance comparison, try the attached VI.  I rewrote it from one in vi.lib after an earlier post.
    This account is no longer active. Contact ShadesOfGray for current posts and information.
    Attachments:
    CheckForEquality(DBL).vi ‏23 KB

  • SEM BCS - Step for Load from Datastream

    Hi,
    What are steps to configuration "Load from DataStream".
    What kind of infocube that i should use?
    Thanks in advance.

    Hi,
    There are rather lot threads reg the issue.
    For example:
    Re: BCS configuration
    Re: How is data loaded into the BCS cubes?
    Re: Which info object should be used for Financial statement line (BPS and BCS)
    Re: SEM-BCS: Load from data stream
    Feel free to ask if something is not clear.

  • SEM-BCS:Data Stream Upload

    Hi! All.
    I am facing an issue in Data Stream upload.....The Target field 0Company is 6 char. and the source field 0Company code is 4 char. in length....the system gives an error ...the value Target field exceeds source field..use info object with greater length.....however, upon mapping Company to another infoobject...with 6 char..used instead of 0comp-Code...the system returns yet
    another error that its not coming from source system or source system can't be determined i.e. the new info object.......
    I was thinking changing target field length to four...any work around this issue...If the target field is changed i.e. 0company what implications will it have or just changing the field in data model do the trick??
    Thanks for your input....
    Victor

    Hello Viktor,
    If i understood well your problem i faced the same thing in the past.
    When costumizing the load from data stream include the lenght to 4 characters or try to include an offset of 2.
    Hope is helps.
    If yes award points.
    Best regards,
    João Arvanas

Maybe you are looking for

  • How to find out tcode from a program that can run the same report?

    Hello all, I have a program name and I can see it is successfully proccessed. Now if I want to get the report via a tcode, how can I get it? How can I find the tcode from the program itself? Regards,

  • I am unable to see my ipod, ipad or iphone as a device in itunes.

    I am unable to see my ipod, ipad or iphone as a device in itunes.  It has worked in the past but no longer.  Has anyone experienced this?

  • Time capsule netbios

    ciao a tutti.. ieri sera ho acquistato la tanto famigerata Time Capsule da 2Tb, ottima con mac, perfetta in tutto.. solo che con il mio pc sto avendo qualche problema ad interfacciare l'hard disk per condividere file,musica e quant'altro.. L'errore c

  • Question about synchronized singleton classes

    Hi, I have a singleton class, with only 1 method (which is static), I want access to that method to be synchronized, but someone suggested that I make my getInstance() method synchronized as well. Is this neccessary, or is it overkill? thanks!

  • LAN turn off after few time while Satellite L300D-10B is working

    The main problem is that when I trying to download something long, after few time (30 min to 1 hour or something near value) the network card is turning off and does not work again when I disable/enable it. It only began work when I reestart my noteb