Update Picking date, Loading date, Planned GI date etc on outbound delivery

Hi All,
I'm a bit stuck.  I'm trying to perform some updates against outbound delivery documents, at the header level, from within a user exit of a shipment document.
Essentially, based on some logic and data on the shipment I need to apply some changes to the header dates on LIKP for the deliveries attached to the shipment.  I've looked at BAPI_OUTB_DELIVERY_CHANGE but this does not allow me to change the fields I need to.
So, does anyone know how I can update the following fields on LIKP:-
Picking date - LIKP-KODAT
Picking time - LIKP-KOUHR
Loading date - LIKP-LDDAT
Loading time - LIKP-LDUHR
Planned GI date - LIKP-WADAT
Planned GI time - LIKP-WAUHR
Delivery date - LIKP-LFDAT
Delivery time - LIKP-LFUHR
I'm not sure I can acutally do this or whether it is advisable to do so but have been asked to make it work!
Thanks in advance,
Gareth.

>
Kshamatha Eda wrote:
> Why not pass the values in EXTENSION1 and implement BADI: SMOD_V50B0001 and pass the values in to the header in method EXIT_SAPLV50I_010 and see if it works. If i am not mistaken this should probably hit only when EXTENSION1 is populated in the BAPI.
Hi,
Thank you very much for the suggestion - I'll give this a try and see how I get on.  I was hoping there may be something like this but couldn't find a way of doing it and haven't really done much in this area before.
Thanks
Gareth.

Similar Messages

  • How can i add the dimensions and data loading into planning apllications?

    Now please let me know how can i add the dimensions and data loading into planning apllication without manuallly?

    you can use tools like ODI or DIM or HAL to load metadata & data into planning applications.
    The data load can be done at the Essbase end using rules file. But metadata changes should flow from planning to essbase through any of above mentioned tools and also there are many other way to achieve the same.
    - Krish

  • Best practices for loading apo planning book data to cube for reporting

    Hi,
    I would like to know whether there are any Best practices for loading apo planning book data to cube for reporting.
    I have seen 2 types of Design:
    1) The Planning Book Extractor data is Loaded first to the APO BW system within a Cube, and then get transferred to the Actual BW system. Reports are run from the Actual BW system cube.
    2) The Planning Book Extractor data is loaded directly to a cube within the Actual BW system.
    We do these data loads during evening hours once in a day.
    Rgds
    Gk

    Hi GK,
    What I have normally seen is:
    1) Data would be extracted from APO Planning Area to APO Cube (FOR BACKUP purpose). Weekly or monthly, depending on how much data change you expect, or how critical it is for business. Backups are mostly monthly for DP.
    2) Data extracted from APO planning area directly to DSO of staging layer in BW, and then to BW cubes, for reporting.
    For DP monthly, SNP daily
    You can also use the option 1 that you mentioned below. In this case, the APO cube is the backup cube, while the BW cube is the one that you could use for reporting, and this BW cube gets data from APO cube.
    Benefit in this case is that we have to extract data from Planning Area only once. So, planning area is available for jobs/users for more time. However, backup and reporting extraction are getting mixed in this case, so issues in the flow could impact both the backup and the reporting. We have used this scenario recently, and yet to see the full impact.
    Thanks - Pawan

  • Error Regarding Data loading in Planning and budgeting cloud service.

    Hi ,
    I was able to load data in planning and budgeting cloud service before 1 month.
    I loaded via Administration -> Import and Export -> Import data from File.
    Today i loaded the same file to the cloud instance for the same entity that i loaded after clearing the existing data.
    I am getting error while validating itself as
    \\Unrecognized column header value(s) were specified for the "Account" dimension: "Apr", "May", "Jun", "Jul", "Aug", "Sep", "Oct", "Nov", "Dec", "Jan", "Feb", "Mar", "Point-of-View", "Data Load Cube Name". (Check delimiter settings, also, column header values are case sensitive.)
    I checked the source file. Everything is correct. I actually loaded the same file before and i was able to load.
    Does anyone know the problem behind. Can anyone give me suggestion please
    Thanks in Advance
    Pragadeesh.J

    Thanks for your response John.
    I had Period and Year dimension in Columns
    I changed the layout of year dimension in period to POV and loaded the data file. I was able to load without any errors.
    I changed the layout again to the original form as before and loaded. I didn't get any errors again.
    It worked somehow.
    Thank you
    Cheers John

  • Data loading from planning forms into two different essbase application

    Hi All,
    I have 2 different planning application into 2 different planning and essbase server but the application repository for these applications is same which is lying into 3rd server. Due to this when I add the member into any application it will automatically reflect into another application because the application repository is same. When I refresh the applications the meta data is loaded into there corresponding Essbase server.
    My problem is:
    When I enter the data figure into the planning form it is going into there corresponding Essbase cube. But I want when user enter the data into 1 planning application form it should be reflected into both Essbase application , LIKE METADATA which is not happening. We are doing this for fail over support.

    Hi,
    I am sorry but that is not how I look at, the point of failure for a planning application would the web server, repository, essbase database...
    If it is the web server element then surely you would have two planning web servers pointing to one planning application and one repository.... then if one web server went down to switch to the other web server?
    You won't get a planning application to write to two seperate essbase applications, if you want failover on essbase then you will need to look at alternative methods.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Transactional Data Load without Respective Master Data in Place

    Hi Folks,
    Could anyone let me know if i could load transactional data without loading the respective master data. If yes, well could you give me a scenario when you do it. Also could you let me know the option of how to load transactional data without master data in place.
    Best Regards,
    Prithvi

    Hi,
    You can load transaction data without loading master data.
    You have the option in the info package to load data even if no master data exists..
    You may use it in cases where your transfer rules/update rules or routines donot refer to master data tables to populate transaction data..
    Ashish

  • Full load - Every month data loading without deleting previous data

    Dear SDN,
    I am loading (Full load) Data from 001.2005 to 010.2007 using 0FI_GL_1 with the selection option Fiscal Year/Period (From : 001.2005 and To : 010.2007) ...
    From next month (011.2007) onwards, when loading data everytime it deletes entire data and again loading (Since i have kept an option in Infopackage -- Datatargets Tab -- Delete entire content of datatarget)...
    My doubt is wheather without deleting the whole data, can we load from next month data onwards...so, if that is the case shall i use Fiscal year/Period range ?? (or) should not use Fiscal year/Period range...
    But the client needs data only from 001.2005 onwards not beyond that
    Please help me to resolve this issue..
    Help will be greatly appreciated with points...
    Thanks....

    Hello Venkat,
    Unfortunately you can not do that way, because in period 11.2007 (assume this is the period that you want to newly load to cube.) there might be posting in previous period. And these postings made to previous period should also reflect to cube. Only way is delete and reload by using full.
    Other option you might use a ODS before sending to cube. you might every time load full data, ODS key fields might be checked by system and neccesary inserts and updates will be done by system, so there wont be duplications. And finally only new data will be sent to cube as delta.
    Sarhan.

  • Flat file data load - ODS - Look up data in Startroutine

    Hi All,
    There is a requirement that I have two ODS say 1) ABC and XYZ.
    For Both ODS , We load flat file data.
    First we load data to ABC for current fiscal period
    During the data load to XYZ, We lookup part nos data in ABC for current fiscal period , if data there , for those part nos we load XYZ Flat file data.
    My requirement is that , I need to fetch part nos data from ABC for Fiscal period < current fisacl period.  Then I need to extract already loaded data for those part nos from XYZ and Mark the flag field as Non reportable .
    How can we achieve this ?
    Please advice
    Thanks
    Ajay

    Hi ,
    Thanks for your reply.
    I have done so. When I add lines of My Internal table data to DATA_PACKAGE, It gives the syntax error that both structures are not unique. We have several fields in XYZ ods but they are not in Communication structure.
    Thats the problem. Also , if I write the code , will it be executed for each datapackage. I mean , for each datapacket process , my code will fetch whole data from ABC and then from XYZ. Repetion will be there?
    Any other logic can i use ?
    Thanks
    Ajay

  • Problem in getting data in SNP Planning Area Data Source

    Hi All,
    Problem Statement: To fetch data from SNP Planning Area for reporting in SAP BW.
    Approach: Followed the standard SAP Help Document. Link:http://help.sap.com/saphelp_scm40/helpdata/en/e0/9088392b385f6be10000000a11402f/content.htm
    First Step: To create a DataSource on the Planning Area (Z9ASNP02) and Aggregate(9AMALO).
    For that
    Go to SAP Easy Access>Click on /SAPAPO/MSDP_ADMIN/ you are on S&DP Administration>Select the planning Area Z9ASNP02 double click on it
    You are on Change Planning Area Screen>Go to Extras left click>DP/SNP Data Extraction Tool-->Click on Generate Data Source, Provide the
    Data Source Name & Select the Aggregate 9AMALO and click on continue. Data Source is generated.
    But now when i go for Test Data Source-->Extractor Checker Screen and then providing the version 000 and then when i click on extraction Button
    It gives a Message "0 data records selected".
    Thanks in Advance,
    rahul

    Hi Somnath,
    Please correct me if my understanding is wrong; you are saying that, we cannot extract data from order series keyfigures. In that case how many key figures you maintain for time series, performance miight come down for planning area if you maintain huge no of keyfigures.
    Is there any way to extract data from Order series Planning area?
    Thanks & Regards,
    Jagadeesh.

  • Log on data load through a BW data flow

    Dears,
    I am requesting to all of you who have already implemented this type of functionality. I am trying to find the easiet way, with less complexity, to implement a log through an existing BW data flow.
    I mean data load by an infopackage give some log on right and wrong records within the monitor, how can I utilize this information? is there a specific table which stored each record and their message? Or a program has to be implemented which will publish laoding status in a specific table?
    Thanks for your quick feedback,
    LL

    Hi Ludovic
    The monitor messages are only written if there is some problem in the record processing. You can only find information for those records which have problem or if the processing during the routines encountered some problem.
    What you can do to capture messages is write one transfer routine and amend the monitor messages table rsmonmess for the same.
    Also,please check the tables starting with RSMO*
    regards
    Vishal

  • Data load error for master data

    hi
    both errors i need to fix manually or any automatic is there
    psa i need to edit that record and run again infopackage.
    i need keep that infoobject(BIC/ZAKFZKZ ) alpha conversion at infoobject level.
    Errors in values of the BOM explosion number field:
    Diagnosis
        Data record 5584 & with the key '001 3000000048420000 &' is invalid in
        value 'BOX 55576\77 &' of the attribute/characteristic 0RT_SERIAL &.
    InfoObject /BIC/ZAKFZKZ contains non-alpha compliant value 1

    Hi Suneel,
    I feel that the symbol '&' is cauing the issue in ur data loads..
    So try to edit the same at the PSA and then load the data from PSa to the target..
    Also u can use RSKC to handle these special Chars..
    For more information u can search SDN..
    Thanks
    Hope this helps

  • BW data loading to cube (delete data before load)

    I have a cube which load data from a SQL based source system.
    I would like to have a schedule job which first delete current year's data in the cube, then extract and load current year's data from SQL server to cube.
    How can it be done? Thanks.

    Hi,
    If you are using Process chains then use...
    Use DELETE_FACTS tcode and genarate the report name i.e. select Data target and Genarate Program then execute it will genarate the program, so create varient for that program and then use it in process chain. i.e.
    Start
      |
    Delete using that Program (with varient on year/month)
      |
    Load data
    Thanks
    Reddy

  • Missing Standard Dimension Column for data load (MSSQL to Essbase Data)

    This is similar error to one posted by Sravan -- however I'm sure I have all dimensions covered -- going from MS SQL to SunOpsys Staging to Essbase. It is telling me missing standard dimension, however I have all accounted for:
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last): File "<string>", line 23, in ? com.hyperion.odi.essbase.ODIEssbaseException: Missing standard dimension column for data load
    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
    at java.lang.reflect.Method.invoke(Unknown Source)
    I'm using multiple time period inputs -- BegBalance,Jul,Aug,Sep,Oct,Nov,Dec,Jan,Feb,Mar,Apr,May,Jun (target has all of those in place of Time Periods)
    I'm using hard coded input mapping for Metric, Scenario, Version, HSP_Rates and Currencies. -> 'Amount', 'Actual', 'Final', 'HSP_InputValue','Local' respectively.
    The only thing I can think of is that since I'm loading to each of the months in the Time Periods dimension (the reversal was set up to accomodate that)... and now its somehow still looking for that? Time Periods as a dimension does not show up in the reversal -- only the individual months named above.
    Any ideas on this one??

    John -- I extracted the data to a file and created a data load rule in Essbase to load the data. All dimensions present and accounted for (five header items as similar here) and everything loads fine.
    So not sure what else is wrong -- still getting the missing dimension error.
    Any other thoughts?? Here's the entire error message. Thanks for all your help on this.
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 23, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Missing standard dimension column for data load
         at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
         at org.python.core.PyMethod.__call__(PyMethod.java)
         at org.python.core.PyObject.__call__(PyObject.java)
         at org.python.core.PyInstance.invoke(PyInstance.java)
         at org.python.pycode._pyx8.f$0(<string>:23)
         at org.python.pycode._pyx8.call_function(<string>)
         at org.python.core.PyTableCode.call(PyTableCode.java)
         at org.python.core.PyCode.call(PyCode.java)
         at org.python.core.Py.runCode(Py.java)
         at org.python.core.Py.exec(Py.java)
         at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.k(e.java)
         at com.sunopsis.dwg.cmd.g.A(g.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: Missing standard dimension column for data load
         at com.hyperion.odi.essbase.ODIEssbaseDataWriter.validateColumns(Unknown Source)
         ... 32 more
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Missing standard dimension column for data load
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.k(e.java)
         at com.sunopsis.dwg.cmd.g.A(g.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)

  • Data load issue with export data source - BW 3.5

    Hi,
    We are facing issues in loading data with the help of export data source.
    We have created export data source of 0PCA_C01 cube. With the help of this export datasource,  we are loading data to other custom cube. Scenario is working fine in development server.
    But when we transported objects to quality server data is not getting loaded to custom target cube.
    It is extracting zero records.  All transports are ok and we have generated export datasource in quality before transports .Also regenerated export datasource after transport and activated infosource, update rule via RS* programs.  Every object is active but data is not getting extracted.
    RSA3 for 80PCA_C01 datasource isn't extracting any record in Quality. Records getting extracted in development.   We are in BW 3.5 with patch level 19.
    Please guide us to resolve the issue.
    Thanks,
    Aditya

    Hi
    Make sure that you have relevant Role & Authorization at Quality/PRS.
    You have to Transport the Source Cube first and then Create a Generate Export Data Source in QAS. Then, replicate data sources for BW QAS Soruce System. Make sure this replicated Data Source in QAS. Only then can transport new update rules for second cube.
    Hope it helps and clear

  • Problem in data loading in 0PM_DS05 using data source 0PM_MEASUREMENT

    Hi Gurus,
    I have to load data in OPM_DS05 ODS using 0PM_MEASUREMENT data source.
    Now the problem what I am facing is, there is a field equipment number which was there in data source but not mapped to infoobject then I mapped that field
    EQUNR(R3 field) to 0EQUIPMENT(BI infobject) after this  when I try to load data error is comming and when I check the dump I understand that there is some routine written in update rules which refers master data of 0EQUIPMENT thus what I got is I have to load master data first then the ODS but when I load master data then request is not comming in the ATTRIBUTES 0EQUIPMENT.
    I am using 0EQUIPMENT_ATTR data source to load data to 0EQUIPMENT.
    Please help me out from the situation what should I do and also suggest whether it is necessary to load master data for 0EQUIPMENT.
    Your efforts will be appreciated.

    Hi,
    in generally it is always a good idea to load the master data first. In your case there are still some questions.
    a) did you activate the master data after loading?
    b) what kind of short dump are you getting while loading the transactional data?
    c) what is the routine doing?
    I guess your loading issues are more related to the routine than anything else.
    regards
    Siggi

Maybe you are looking for

  • ITunes stalls when trying to update my photos - can it be fixed?

    From what I've seen of 7.0, there are a lot of problems, and I've searched these posts but haven't found anything on my specific problem. Basically, when I try to update my iPod and I have the "sync photos from folder" checked, iTunes never finished

  • Fios Network Terminal power consumption

    Hello, Does anyone know how much power my Alcatel lucent ( I 211m h ) (3fe51852aaac) Fios network terminal uses.  We live in a multi-family house situation, and the other party pays for basement power, where the box is plugged in. Thank you very much

  • Exception message during planned order creation

    I am coming across a situation wherein am creating a planned order for Semi-fininsh good, but the problem is that am getting an exception msg "15" against the planned order created whihc states that "Reschedule out". Can somebody tell me the signific

  • Vertical Text in Tabels

    I wonder how I can flip text in a single cell of a table to vertical orientation? Thanks for any support.

  • Integrating SAP UI5 application in Portal through LPD_Cust gives error

    Hi Team, In our scenario we are calling sap UI5 application in portal through LPD_Cust . We receive error SAP UI5 mobile is not supported in your browser . When we call same application through portal directly it opens properly( In both cases we are