Moving KM data (incl. for ex. modified)

Hi,
I would like to move KM data of a document store to another folder including all metadata. It's important, that the original author and modified date are not changed. The data should be moved in the same portal for example from /room_extensions/room1 to /room_extensions/room2. 
Is this possible?
For background-information: We've got compatibility-problems with our Collaboration Rooms prior EP 6.0 SP 2 Patch 5 and we would like to migrate them. This change should be as less visible as possible to the endusers.
Best regards
Sandra

Hi Sandra,
You can use the KM Repository Java API and the
<a href="https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/javadocs/Pre%20NW04%20SP2/KM/com/sapportals/wcm/repository/IResource.html">IResource Interface</a> to instanciate your resources and use the copy method (public IResource copy(RID destinationRID,...) to create your new resources.
Regards
Message was edited by: Laurent Bouziani

Similar Messages

  • SSRS 2008 Column Chart with Calculated Series (moving average) "formula error - there are not enough data points for the period" error

    I have a simple column chart grouping on 1 value on the category axis.  For simplicity's sake, we are plotting $ amounts grouping by Month on the category axis.  I right click on the data series and choose "Add calculated series...".  I choose moving average.  I want to move the average over at least 2 periods.
    When I run the report, I get the error "Formula error - there are not enough data points for the period".  The way the report is, I never have a guaranteed number of categories (there could be one or there could be 5).  When there is 2 or more, the chart renders fine, however, when there is only 1 value, instead of suppressing the moving average line, I get that error and the chart shows nothing.
    I don't think this is entirely acceptable for our end users.  At a minimum, I would think the moving average line would be suppressed instead of hiding the entire chart.  Does anyone know of any workarounds or do I have to enter another ms. connect bug/design consideration.
    Thank you,
    Dan

    I was having the same error while trying to plot a moving average across 7 days. The work around I found was rather simple.
    If you right click your report in the solution explorer and select "View Code" it will give you the underlying XML of the report. Find the entry for the value of your calculated series and enter a formula to dynamically create your periods.
    <ChartFormulaParameter Name="Period">
                      <Value>=IIf(Count(Fields!Calls.Value) >= 7 ,7, (Count(Fields!Calls.Value)))</Value>
    </ChartFormulaParameter>
    What I'm doing here is getting the row count of records returned in the chart. If the returned rows are greater than or equal to 7 (The amount of days I want the average) it will set the points to 7. If not, it will set the number to the amount of returned rows. So far this has worked great. I'm probably going to add more code to handle no records returned although in my case that shouldn't happen but, you never know.
    A side note:
    If you open the calculated series properties in the designer, you will notice the number of periods is set to "0". If you change this it will overwrite your custom formula in the XML.

  • Order Date should not be modifiable by the user

    Hi,
    I want  to do ordered date is don't modifiable by the user when they selected  order type in Order Management Super User Responsibility ,Please help me  If anybody done earlier I am not aware of form
    personalization .
    Thanks&Regards,
    Mannala Krishna.

    Hi Nagesh
    Thanks for your replay,
    I had done this through form personalization and moved into Prod it is working fine
    Thanks,
    Krishna Mannala

  • Excel Upload and moving the data to dynamic itab

    Hi Folks,
    I am building a fieldcatalog dynamically based on the table name and then creating a dynamic internal table based on the fieldcatalog.Now I have a excel sheet having the fields in the same structure as the fieldcatalog that I built.I want to upload the data in the excel to an internal table having the structure of the fieldcatalog I built.I am having a problem in moving the data from the excel to the itab (dynamic itab built based on the fieldcatalog)
    1.Building the fieldcatalog.
    data:gt_fieldcat    TYPE lvc_t_fcat.
      ls_fieldcat-tabname    = p_tabname.
      ls_fieldcat-fieldname  = p_fieldname.  
      ls_fieldcat-inttype    = 'C'.
      ls_fieldcat-outputlen  = wa_dfies-outputlen.   
      ls_fieldcat-coltext    = wa_dfies-fieldtext.
      APPEND: ls_fieldcat TO gt_fieldcat.
      CLEAR : ls_fieldcat.
    2.Building the dynamic itab
    DATA:gp_table       TYPE REF TO data.
    FIELD-SYMBOLS: <gt_table> TYPE table,
    CALL METHOD cl_alv_table_create=>create_dynamic_table
        EXPORTING
          it_fieldcatalog = gt_fieldcat
        IMPORTING
          ep_table        = gp_table.
      ASSIGN gp_table->* TO <gt_table>.
    3.Now I have an excel with a similar structure with data.Uploading the excel data into itab.
    DATA: g_t_xltab  TYPE STANDARD TABLE OF ty_xltab,
          it_rawdata TYPE truxs_t_text_data,
      CALL FUNCTION 'TEXT_CONVERT_XLS_TO_SAP'
        EXPORTING
    *     I_FIELD_SEPERATOR    =
          i_line_header        = 'X'
          i_tab_raw_data       = it_rawdata
          i_filename           = xls_filename
        TABLES
          i_tab_converted_data = g_t_xltab[]
        EXCEPTIONS
          conversion_failed    = 1
          OTHERS               = 2.
      IF sy-subrc <> 0.
        MESSAGE ID sy-msgid TYPE sy-msgty NUMBER sy-msgno
                WITH sy-msgv1 sy-msgv2 sy-msgv3 sy-msgv4.
      Endif.
    *Move the data that got uploaded from excel into
    *respective internal tables.
        LOOP AT g_t_xltab
        here I want move the data into an internal table having the structure I built using
        the fieldcatalog/dynamic itab ie    <gt_table>.
        Endloop.
    I tried using ASSIGN COMPONENT sy-index of <gt_table> and move-corresponding but in vain as I am not that much familiar with dynamic itab using field symbols.Kindly let me know how the excel data can be moved to dynamic itab.
    Thanks,
    K.Kiran.

    Hi,
    you can try some logic like this. Please note this is just an example
    TYPE-POOLS:truxs.
    TYPES:BEGIN OF ty_xltab,
          field TYPE string,
          END OF ty_xltab.
    DATA:wa TYPE ty_xltab.
    DATA:lv_len TYPE i.
    DATA:startpos TYPE i.
    FIELD-SYMBOLS:<fs_line> TYPE mara.
    FIELD-SYMBOLS:<fs> TYPE ANY.
    DATA: g_t_xltab  TYPE STANDARD TABLE OF ty_xltab,
          it_rawdata TYPE truxs_t_text_data.
    CALL FUNCTION 'TEXT_CONVERT_XLS_TO_SAP'
        EXPORTING
    *     I_FIELD_SEPERATOR    =
          i_line_header        = 'X'
          i_tab_raw_data       = it_rawdata
          i_filename           = ' '
        TABLES
          i_tab_converted_data = g_t_xltab[]
        EXCEPTIONS
          conversion_failed    = 1
          OTHERS               = 2.
    IF sy-subrc  <> 0.
      MESSAGE ID sy-msgid TYPE sy-msgty NUMBER sy-msgno
              WITH sy-msgv1 sy-msgv2 sy-msgv3 sy-msgv4.
    ENDIF.
    startpos = 0.
    LOOP AT g_t_xltab INTO wa.
      DO.
        ASSIGN COMPONENT sy-index OF STRUCTURE <fs_line> TO <fs>.
        IF sy-subrc = 0.
          DESCRIBE FIELD <fs> OUTPUT-LENGTH lv_len.
          IF lv_len > 0.
            <fs> = wa-field+startpos(lv_len).
            startpos = startpos + lv_len.
          ENDIF.
        ELSE.
          append <fs_line> to <fs_tab>.
          EXIT.
        ENDIF.
      ENDDO.
    ENDLOOP.
    Sorry my answer is not for you requirement..Please Ignore.
    Your solution can be easily solved by looking at the concept Assign component . Please search
    Edited by: Keshav.T on Nov 7, 2011 1:42 PM

  • Data Migration for Open Purchase Order

    Hi, All,
    Is there anyone know how to Count the volume for Open Purchase Order. What's the normal strategy for the Data Migration and Cut-over stage?
    My client want to know how many Open Purchase Order in the legacy system and then determine manual or automatic data migration. If manual, how to do? If automatic, how to do? Because all materials and vendors, plants are different number. How to track? How to find out to match between new and old?
    Thank you very much

    JC,
    Sounds a bit early to be making decisions about the realization phase.  It doesn't sound like you have finished the Blueprinting phase yet, much less the testing phase.
    Anyhow, in my experience I typically use LSMW (Legacy system migration workbench) to load MM master data (material masters), Inventory (WIP, RM, FG, etc) Purchasing Master data (Vendors, Purchase Info Records, Source Lists, Quota Arrangements), and Purchasing transactional documents (POs, PurReqs, Scheduling Agreements, etc).  Depending on the complexity and volume of data, it  may be necessary to write custom programs to load the data.  You will find this out during your requirements gathering.
    It is uncommon but possible to load all of these data manually.  I have never run across a client that wants to pay a consultant's hourly rate to sit at a terminal to peck away loading master data, so if the client intends to have his own users enter the data manually, the project manager should make provision that there will be qualified TRAINED client employees available for this data entry.  I did help with a portion of a conversion once manually; of Sales Credits, but there were only about 30 SD docs to load.   I did this the evening before go-live day, while I was waiting for some of my LSMW projects to complete in the background.
    A good opportunity to 'practice' your data loads is right after you have completed your development and customization, and you have gotten the approval from the client to proceed from the pilot build to the full test environment.  Once you have moved your workbench and customization into the client's test environment, but before integration testing, you can mass load all, or a substantial portion of your conversion data into the qual system.  You can treat it like a dry run for go-live, and fine tune your processes, as well as your LSMW projects.
    Yes, it is good practice to generate comparisons between legacy and SAP even if the client doesn't ask for it. For Purchase orders on the SAP side, you could use any of the standard SAP Purchasing reports, such as ME2W, ME2M, ME2C, ME2L, ME2N.  If these reports do not meet the requirements of the client, you could write a query to display the loaded data, or have an ABAPer write a custom report.
    You didn't ask, but you should also do comparisons of ALL loaded data - including master data.
    It sounds like you are implying that the client wants YOU to extract the legacy data.  For an SAP consultant, this is not very realistic (unless the legacy system is another SAP system).  Most of us do not understand the workings of the myriad legacy systems.  The client is usually expected to produce one or more legacy system technical experts for you to liase with.  You normally negotiate with the technical expert about every facet of of the data migration.  In addition, you will liase with business users, who will help you and the implementation team to logically validate that the final solution (turnkey SAP production system, fully loaded with data) will meet the client's business needs.
    Finally, you mentioned how do you track the mapping of master data between legacy and SAP.  There are many ways to do this.  I normally try to get the legacy person do the conversion on his end, eg, when he gives you the load file, you would like to have already translated the master data and inserted the SAP relevant values into the file.  If this is not possible, I usually use MS Access databases to maintain a master map, and I perform the mapping on a PC.  If your data package is small, you can probably get by using MS Excel or similar.
    Good Luck,
    DB49

  • Error while executing Initialize with Data Transfer for 0FI_GL_10

    Hello All,
    Post Pre prod refresh, our timestamp for the datasource 0FI_GL_10 got reset. Due to which our deltas did not bring any records to the BW system.
    First we did an Initialize without data transfer for all the BW related datasources. The deltas were then set properly for all datasources except 0FI_GL_10.
    We then raised a message to SAP and they suggested to run the 'Initialize with data transfer' for 0FI_GL_10 so that the timestamp is set and accordingly the deltas are fixed.
    The issue now is we are getting the following error message while running INIT with data transfer.
    Job terminated in source system --> Request set to red
    Message no. RSM078
    We have copied the data till July 1st week of 2011.
    Please advice. The issue is very critical.
    Thanks & Regards
    Sneha

    Hi Arvind
    Thanks for your inputs.
    Please find below the details of the short dump.
    Runtime Errors         DBIF_RSQL_SQL_ERROR
    Exception              CX_SY_OPEN_SQL_DB
    Date and Time          09/07/2011 11:25:32
    Short text
         SQL error in the database when accessing a table.
    What can you do?
         Note which actions and input led to the error.
         For further help in handling the problem, contact your SAP administrator
         You can use the ABAP dump analysis transaction ST22 to view and manage
         termination messages, in particular for long term reference.
    How to correct the error
         Database error text........: "ORA-01652: unable to extend temp segment by 128
          in tablespace PSAPTEMP"
         Internal call code.........: "[RSQL/FTCH/FAGLFLEXT ]"
         Please check the entries in the system log (Transaction SM21).
         If the error occures in a non-modified SAP program, you may be able to
         find an interim solution in an SAP Note.
         If you have access to SAP Notes, carry out a search with the following
         keywords:
         "DBIF_RSQL_SQL_ERROR" "CX_SY_OPEN_SQL_DB"
    Information on where terminated
        Termination occurred in the ABAP program "GP_GLX_FAGLFLEXT" - in
         "FETCH_TO_ISTRUCTURE".
        The main program was "SBIE0001 ".
        In the source code you have the termination point in line 903
        of the (Include) program "GP_GLX_FAGLFLEXT".
        The program "GP_GLX_FAGLFLEXT" was started as a background job.
        Job Name....... "BIREQU_4N3PZQ12IA0X0PYGEA85IG39S"
        Job Initiator.. "BIWREMOTE"
        Job Number..... 11203300
        The termination is caused because exception "CX_SY_OPEN_SQL_DB" occurred in
        procedure "FETCH_TO_ISTRUCTURE" "(FORM)", but it was neither handled locally
         nor declared
        in the RAISING clause of its signature.
        The procedure is in program "GP_GLX_FAGLFLEXT "; its source code begins in line
        840 of the (Include program "GP_GLX_FAGLFLEXT ".

  • No international data plan for one year-???

    Here is the background.  I moved over to Verizon about 4 months ago from Alltel.  I did this because I wanted to be able to have a blackberry with international BBm'ing.  I did not go with ATT because there coverage was spotty in my area.  Since the move, I would say my Alltel coverage was 3x better then the cuurent verizon in my house even with the network extender ( BB don't work well with network extender compare to droids, according to the VZW reps).  I did not stay with Alltel because they did not have the storm2.  I needed 6 lines and a mifi and the local verizon rep told me no problem that they could do one primary line and 5 additonal line plus the mifi but after I signed up they realized that I maxed out at 5 lines on one account and they made me make my 6th line a primary line and then added the mifi to the 2nd account.  More costs per month.  Within a month of having my VZW account, one of my kid went to China and we were able to add the international data plan for her trip so she could BBM us.  No problems.
    Here is the problem.  I am going abroad in 2 days and would like to put the international data plan on 2 phones for the trip.  The rep says he cannot do becaue the polciies have changed with the "xerox" buyout and they won't let me do it for the following reasons.
    1.  I have not had my phones for more than one year even though I paid a deposit on them
    2. I have 7 lines on one account- not true I have one account with 1 primary line and 4 additional lines and I have another account with one primary line and the mifi account.  they are all billed on the same statement.  If they are all "one account" then VZW owes me a refund on the 2nd account's primary line as they cannot have it both ways. 
     We came to verizon for the purpose of BBM’ing when going international and now we cannot use it.  I am paying over $500.00 a month and have not been late on any payments, paid a deposit on the phones, am paying for 2 primary lines and did this a few weeks ago for my daughter and now I cannot add the international option again for a week of vacation abroad?  This is really upsetting and pretty much the last straw if it does not get resolved before this trip.  I should have stayed at Alltel or went to ATT and got an I phone because this is no better.    The rep I talked too was bedeviled by this policy and went two levels up in supervisors and was denied.  They are supposed to call me to resolve this issue tonight or in the morning. 
    {word filter avoidance}

    Update:
    I got a call from my local rep at the store where I bought all the phones.  She said they called corporate and all blocks on my international data plan have been removed?  I will stop by this afternoon and confirm that it has all been done???? I still cannot believe how many hoops I will have had to jump through to get this done as it was done 2 months ago, if it actually gets done.
    If it does not get done then i will be requesting to be let out of my contract for breach of contract on Verizon's part as it was never told to me that I could not have international service for one year after signing up and we were able to do it within a month of signing up initially.  Will post updates

  • Trying to change the data source for a Crystal Report.

    <p>The method below represents my best attempt to programatically change the data source of a Crystal Report. The goal is to have a routine that will update the data source for reports after they have been distributed to production servers. So far I have not been successful in saving the report back to the CMS. No exceptions are thrown, but when I view the Database Configuration of the report in the CMC nothing has changed.
    </p>
    <p>
    Am I missing a step, or is there another way to accomplish this?
    </p>
    <p>
    Thank you.
    </p>
    <hr />
    <pre>
    private void test(String reportName)
       throws SDKException, ReportSDKException, java.io.IOException
       IInfoObjects newInfoObjects;
       IInfoObject reportObj;
       ReportClientDocument clientDoc = new ReportClientDocument();
       DatabaseController dc;
       PropertyBag pBag;
       PropertyBag logonProps;
       ConnectionInfo newConInfo;
       ConnectionInfo oldConInfo;
       ConnectionInfos conInfos;
       int connOptions = DBOptions._ignoreCurrentTableQualifiers + DBOptions._doNotVerifyDB; //0;
       Fields connFields = null;
       String queryStr = "Select * From CI_INFOOBJECTS " +
          "Where SI_NAME='wfr.rpt' AND SI_KIND='CrystalReport' AND SI_INSTANCE=0";
       newInfoObjects = getCms().executeQuery(queryStr);
       if(newInfoObjects.size() > 0)
          reportObj = (IInfoObject)newInfoObjects.get(0);
          try
             clientDoc = getCms().getReportAppFactory().openDocument(
                reportObj
                , OpenReportOptions._refreshRepositoryObjects
                , java.util.Locale.US);
             dc = clientDoc.getDatabaseController();
             conInfos = dc.getConnectionInfos(null);
             for(int i = 0; i < conInfos.size(); ++i)
                oldConInfo = (ConnectionInfo)conInfos.getConnectionInfo(i);
                newConInfo = (ConnectionInfo)oldConInfo.clone(true);
                pBag = newConInfo.getAttributes();
                pBag.putStringValue("QE_ServerDescription", "alio");
                logonProps = new PropertyBag();
                logonProps.putStringValue("Trusted_Connection", "false");
                logonProps.putStringValue("Server", "alio");
                pBag.put("QE_LogonProperties", logonProps);
                newConInfo.setUserName("admin");
                newConInfo.setPassword("password");
                dc.replaceConnection(
                   oldConInfo
                   , newConInfo
                   , connFields
                   , connOptions);
          catch(ReportSDKServerException Ex)
             String msg = "A server error occured while processing the " + reportObj.getKind()
                + " object, " + reportObj.getTitle() + " (" + reportObj.getCUID() + "), from the CMS.";
             Utility.errorOut(msg, Ex);
          catch(Exception Ex)
             String msg = "An error occured while processing the " + reportObj.getKind()
                + " object, " + reportObj.getTitle() + " (" + reportObj.getCUID() + "), from the CMS.";
             Utility.errorOut(msg, Ex);
          finally
             clientDoc.save();
             getCms().commitToInfoStore(newInfoObjects);
             clientDoc.close();
    </pre>
    Edited by: Mark Young on Sep 10, 2009 2:16 PM

    <style type="text/css">
    /<![CDATA[/
        body
            font-size: 1.125em;
              font-family: helvetica,arial,"sans-serif";
          .code{font-family: "courier new",courier,mono,monospace}
          .bi{font-style: italic; font-weight: bold;}
    /]]>/
    </style>
    <p>Justin,</p>
    <p>
    Thank you for the reply. Time constraints have not allowed me to post back to this tread
    till now. I will try your suggestion. My assumption is that <i>Save the report back to the
    info store</i> refers to <span class="code">IInfoStore.commit(IInfoObjects)</span>.
    </p>
    <p>
    I'm afraid that I do not understand why I don't want to change the report client document,
    or why <i>successfully exporting the report with the new login/password</i> is not what I
    want to do. Any explanation on that statement would be appreciated.
    </p>
    <p>
    I did find a way to accomplish my goal. It involved adding the SSOKEY property to the
    logon property bag. Below you'll see my revised code which modifies the report logon and
    server. I have no idea what
    this does, and SAP support has not been able to tell me why it works. However, what I
    discovered is that if I changed the report option, <b>Database Configuration -> When
    viewing report:</b>, in the CMS to <span class="bi">Use same database logon as when report
    is run</span> from <span class="bi">Prompt the user for database logon</span>, then the
    SSOKEY property had been added to the logon property bag having an empty string as its
    value. This allowed me to successfullyupdate and save the modified logon back to the CMS.
    </p>
    <p>
    So I took a chance and added code to always add the SSOKEY property with an empty string
    as its value, and I could then successfully modify and save the report's logon info
    and server. Again, I don't know what this means, but it has worked so far. If anyone has
    some insight or comments, either are welcome. Thank you in advance.
    </p>
    <br />
    <hr />
    <pre>
    private void changeDataSourceOfAWFCrystalReports()
       throws SDKException, ReportSDKException, java.io.IOException
       IInfoObjects newInfoObjects = null;
       IInfoObject reportObj = null;
       IReport curReport = null;
       ReportClientDocument clientDoc = new ReportClientDocument();
       DatabaseController dbController;
       PropertyBag pBag;
       PropertyBag logonProps;
       ConnectionInfo newConInfo;
       ConnectionInfo oldConInfo;
       ConnectionInfos conInfos;
       int connOptions = DBOptions._ignoreCurrentTableQualifiers + DBOptions._doNotVerifyDB;
       Fields connFields = null;
       String outputStr;
       int numOfReports;
       int numOfQueryPages;
       double progressIncrementPerPage = 30;
       int progressIncrementPerReport = 0;
       // Path query to reports is in a .properties file.
       String queryStr = getAppSettingsFile().getWscAwfCrystalReportPathQuery();
       try
          // Executes IInfoStore.getPageingQuery() and generates a list of queries.
          getCms().setPathQueryQueries(queryStr, 100);
          numOfQueryPages = 0;
          // Gets a List&lt;String&gt; of the IPageResult returned from IInfoStore.getPageingQuery().
          if(getCms().getPathQueryQueries() != null)
             numOfQueryPages = getCms().getPathQueryQueries().size();
          if(numOfQueryPages &gt; 0)
             // Use 30% of progress bar for the following loop.
             progressIncrementPerPage = Math.floor(30.0/(double)numOfQueryPages);
          for(int queryPageIndex = 0; queryPageIndex &lt; numOfQueryPages; ++queryPageIndex)
             // Gets the IInfoObjects returned from the current page query
             newInfoObjects = getCms().getPathQueryResultSetPage(queryPageIndex);
             numOfReports = newInfoObjects.size();
             if(newInfoObjects != null && numOfReports &gt; 0)
                progressIncrementPerReport =
                   Math.round((float)Math.floor(progressIncrementPerPage/(double)numOfReports));
                for(int reportIndex = 0; reportIndex &lt; numOfReports; ++reportIndex)
                   reportObj = (IInfoObject)newInfoObjects.get(reportIndex);
                   curReport = (IReport)reportObj;
                   clientDoc = getCms().getReportAppFactory().openDocument(
                      reportObj
                      , OpenReportOptions._refreshRepositoryObjects
                      , java.util.Locale.US);
                   dbController = clientDoc.getDatabaseController();
                   conInfos = dbController.getConnectionInfos(null);
                   for(int conInfosIndex = 0; conInfosIndex &lt; conInfos.size(); ++conInfosIndex)
                      oldConInfo = (ConnectionInfo)conInfos.getConnectionInfo(conInfosIndex);
                      newConInfo = (ConnectionInfo)oldConInfo.clone(true);
                      pBag = newConInfo.getAttributes();
                      pBag.putStringValue(
                         "QE_ServerDescription"
                         ,getConfigFile().getDBDataSourceConnections());
                      logonProps = new PropertyBag();
                      logonProps.putStringValue("Trusted_Connection", "false");
                      <b>logonProps.putStringValue("SSOKEY", "");</b>
                      logonProps.putStringValue(
                         "Server"
                         ,getConfigFile().getDBDataSourceConnections());
                      pBag.put("QE_LogonProperties", logonProps);
                      newConInfo.setUserName(getConfigFile().getUNVConnectionUserName());
                      newConInfo.setPassword(getConfigFile().getUNVConnectionPasswordDecrypted());
                      dbController.replaceConnection(
                         oldConInfo
                         , newConInfo
                         , connFields
                         , connOptions);
                      newConInfo = (ConnectionInfo)conInfos.getConnectionInfo(conInfosIndex);
                   } // end for on conInfosIndex
                   clientDoc.save();
                } // end for on reportIndex
             } // end if on newInfoObjects
          } // end for on queryPageIndex
       } // end try
       catch(ReportSDKServerException Ex)
          // handle...
       catch(Exception Ex)
          // handle...
       finally
          getCms().commitToInfoStore(newInfoObjects);
          if(clientDoc != null)
             clientDoc.close();
    </pre>

  • Can´t create a date attribute for basic asset

    Hi
    I created a basic asset where I want to have a date attribute to pick up a date for calendar.
    My question is: How can I create a date attribute for a basic asset? For what I understand I have to create an element attribute in my descriptor file like this
    <PROPERTY NAME="imagedate" DESCRIPTION="Image date">
      <STORAGE TYPE="TIMESTAMP" LENGTH="8"/>
      <INPUTFORM TYPE="ELEMENT" WIDTH="24" MAXLENGTH="48" REQUIRED="NO" DEFAULT="" INSTRUCTION="Format: yyyy-mm-dd hh:mm"/>
      <SEARCHFORM DESCRIPTION="Image date" TYPE="ELEMENT" WIDTH="48" MAXLENGTH="128"/>
    </PROPERTY>
    Then, when I create a new instance of this asset, this attribute appears with this error:
    Date: Unable to find element OpenMarket/Xcelerate/AssetType/Oportunidade/ContentForm/imagedate
    But the element (imagedate.xml) is there!!
    I request help.

    It seems that the table does not have the entries, maybe b/c we are on R3 4.7.  Could you tell me what the entries are suppossed to be to allow the Asset and Cost Center lookup?  The search does work for WBS element though and I don't see anything specific to WBS element in that table.  In any case if you know the entries that would be great, also this table is not modifiable via SM30, here are the entries that I have in that table.
    AUF_NETNR     PLM_HELPVALUES_AUTHCHECK
    BU_PARTNER     BUPA_BAPI_F4_AUTHORITY
    DAENR     PLM_HELPVALUES_AUTHCHECK
    DOKNR     PLM_HELPVALUES_AUTHCHECK
    EQUNR     PLM_HELPVALUES_AUTHCHECK
    KUNNR     PARTNER_BAPI_F4_AUTHORITY
    LIFNR     PARTNER_BAPI_F4_AUTHORITY
    MATNR     PLM_HELPVALUES_AUTHCHECK
    NW_AUFNR     PLM_HELPVALUES_AUTHCHECK
    PARNR     PARTNER_BAPI_F4_AUTHORITY
    PS_POSID     PLM_HELPVALUES_AUTHCHECK
    PS_PSPID     PLM_HELPVALUES_AUTHCHECK
    QMNUM     PLM_HELPVALUES_AUTHCHECK
    S_AGNCYNUM     INT_FLBOOK_F4_AUTHORITY
    S_AIRPORT     INT_FLBOOK_F4_AUTHORITY
    S_BOOK_ID     INT_FLBOOK_F4_AUTHORITY
    S_CARR_ID     INT_FLBOOK_F4_AUTHORITY
    S_CITY     INT_FLBOOK_F4_AUTHORITY
    S_CONN_ID     INT_FLBOOK_F4_AUTHORITY
    S_COUNTNUM     INT_FLBOOK_F4_AUTHORITY
    S_COUNTR     INT_FLBOOK_F4_AUTHORITY
    S_CUSTOMER     INT_FLBOOK_F4_AUTHORITY
    S_FLCONN     INT_FLBOOK_F4_AUTHORITY
    S_FLCONN1     INT_FLBOOK_F4_AUTHORITY
    S_FLCONN2     INT_FLBOOK_F4_AUTHORITY
    S_PLANETYE     INT_FLBOOK_F4_AUTHORITY
    S_TRNUM     INT_FLBOOK_F4_AUTHORITY
    TPLNR     PLM_HELPVALUES_AUTHCHECK
    USCOMP     SUSR_BAPI_F4_AUTHORITY
    XUBNAME     SUSR_BAPI_F4_AUTHORITY

  • Default Data Server for Hypersonic SQL in ODI 11g

    Hi All,
    I installed ODI 11g 11.1.1.6.0 for Win 7 with 10g DB R2 and WLS 10.3.5 successfully without any issues.
    Created DEV_ODI_REPO Repository also without any issues.
    Created Master Repository snpm1 and Work Repository as snpw1.
    Created the Agents (Physical and Logical) and tested successfully.
    Created the data servers for respective technologies
    1) Technology Oracle -
        Data Server - ORACLE_ORCL_LOCAL 
        Physical Schemas - SALES_DEV & SALES_PROD and
        Logical Schema as ORACLE_ORCL_LOCAL_SALES
    2) Technology FILE -
        Data Server - FILE_GENERIC
        Physical Schema - FILE_GENERIC.E:/oracle/Middleware/Oracle_ODI1/oracledi/demo/file
        Logical Schema - FILE_DEMO_SRC
    3) Technology - XML
         Data Server - XML FILE
         Physical Schema - GEO_D
        Logical Schema - GEO_DIM
    4) Technology - Hypersonic SQL
         Data Server - HSQL   (Is this the correct name or some other name is recommended ? )
         Physical Schema - HSQL_LOCALHOST_20001_Default and HSQL_LOCALHOST_20001_Default
        Logical Schema - HSQL_DEMO_SRC and HSQL_DEMO_TARG
    All the Mappings are done properly using Development & Production context for Physical and Logical Schemas.
    I am able to create the Models for Oracle, FIle and XML with successful Reverse Engineering (standard with development context )
    with all the data displayed correctly.
    The issue starts here
    The Main issue is for Hypersonic SQL where I am not able to get the Reverse Engineering.
    Technology - Hypersonic SQL  I have defined the data server as HSQL
    Is this correct data server or some other name for data server is suggested ?
    I am able to start the startdemo.bat file successfully opening the 3 dos prompt windows as
    a) Oracle DI Demo - Repository Server
    b) Oracle DI Demo - Source Server
    c) Oracle DI Demo - Target Server
    Later I make  entries on the JDBC tab od HSQL - Data Server as
    JDBC Driver : org.hsqldb.jdbc.JDBCDriver
    JDBC Url : jdbc:hsqldb:hsql://localhost
    Test Connection - ok with successful connection.
    Prior to this I already run the script created schemas and inserted data to the SRC_ tables in my 10g Database as
    ORDERS - containing the Tables as SRC_CITY, SRC_CUSTOMER, SRC_PRODUCT, SRC_REGION.....
    In the Designer I create a new Model as HSQL_SRC
    Technology - Hypersonic SQL
    Logical Schema - HSQL_DEMO_SRC as defined earlier in logical schema.
    Reverse Engineer Tab - Selected as Standard   Context selected as Development from drop down list.
    When I right click on HSQL_SRC model for reverse engineer, it executes, however it does not display any tables inserted
    as shown for ORDERS Schema.
    Please correct me where I am wrong in giving the name for data server ? or some other place.
    Thanks
    Ajaz

    Hi,
    according to docs
    http://docs.oracle.com/cd/E23943_01/integrate.1111/e12644/hypersonic_sql.htm#ODIKM590
    i see
    JDBC Driver: org.hsqldb.jdbcDriver
    JDBC URL: jdbc:hsqldb:hsql://<host>:<port>
    could you modify
    JDBC Driver : org.hsqldb.jdbc.JDBCDriver
    JDBC Url : jdbc:hsqldb:hsql://localhost
    according to docs?
    Thanks

  • Date problem for a report (urgent)

    hi all ,
    i am developing a report for asset aquisition of fi module.
    in this my client has requirement like this-
    1. if i enter date in selection screen ie 31.03 2006 or less than that
         then i should get all the records from 01.04.2005 to the date which i have
        entered in selction screen ie might be 31.03.2006 or less than that.
      2. but if i enter date in selection screen any date more  than 31.03.2006
         then i should get all the records from 01.04.2006 to the date which i have
        entered in selction screen .
    ex- if i have entered 22.12.2006 then i should get record from
          01.04.2006 to 22.12.2006.
    bcoz indian fiscal year starts from 1st april and ends at 31st march.
    so plz give the solution.
    Thanks & Regards
    Sanjeev

    hi mr reddy,
    thanks for the reply.
    here i am sending my code just make the modification in it wherever
    it is required, bcoz i m also using some date logic for it.
    so i request u to modify my code and check it that wethere its working
    correctly after applying ur logic.
    bcoz right now if i am entering any date  suppose 22.12.2006
    then i am getting all the record from 01.04.2005 to 22.12.2006.
    but it should display only from 01.04.2006 to 22.12.2006.
    but ya one more thing i want to make it in ur notice that
    if i have entered 31.03.2006 then it should show
    records from 01.04.2005 to 31.03.2006.
    so both of my above requirement should fulfill out of which one i am alraedy getting.
    check my code and do the modification as per ur logic.
    REPORT Z_FA_ACQ  NO STANDARD PAGE HEADING
                     LINE-SIZE 400
                     LINE-COUNT 65(3).
    TYPE-POOLS : SLIS.
    TABLES : ANLA,                         "ASSET MASTER RECORD-SEGMENT
             ANLZ,                         "time dependent asset allocations
             ANLC,
             ANEK.
    DATA : BEGIN OF T_ANLA OCCURS 0,
           BUKRS LIKE ANLA-BUKRS,
           ANLN1 LIKE ANLA-ANLN1,
           ANLN2 LIKE ANLA-ANLN2,
           AKTIV LIKE ANLA-AKTIV,
           TXT50 LIKE ANLA-TXT50,
           ZUGDT LIKE ANLA-ZUGDT,
           MENGE LIKE ANLA-MENGE,
           MEINS LIKE ANLA-MEINS,
           ANLKL LIKE ANLA-ANLKL,
           END OF T_ANLA.
    DATA : BEGIN OF T_ANLZ OCCURS 0,
           BUKRS LIKE ANLZ-BUKRS,
           ANLN1 LIKE ANLZ-ANLN1,
           ANLN2 LIKE ANLZ-ANLN2,
           GSBER LIKE ANLZ-GSBER,
           KOSTL LIKE ANLZ-KOSTL,
           STORT LIKE ANLZ-STORT,
           RAUMN LIKE ANLZ-RAUMN,
           KFZKZ LIKE ANLZ-KFZKZ,
           WERKS LIKE ANLZ-WERKS,
           END OF T_ANLZ.
    DATA : BEGIN OF T_ANEK OCCURS 0,
           BUKRS LIKE ANEK-BUKRS,
           ANLN1 LIKE ANEK-ANLN1,
           ANLN2 LIKE ANEK-ANLN2,
           BELNR LIKE ANEK-BELNR,
           BUDAT LIKE ANEK-BUDAT,
           XBLNR LIKE ANEK-XBLNR,
           SGTXT LIKE ANEK-SGTXT,
           GJAHR LIKE ANEK-GJAHR,
           END OF T_ANEK.
    DATA : BEGIN OF T_ANLC OCCURS 0,
           BUKRS LIKE ANLC-BUKRS,
           ANLN1 LIKE ANLC-ANLN1,
           ANLN2 LIKE ANLC-ANLN2,
           ANSWL LIKE ANLC-ANSWL,
           NAFAP LIKE ANLC-NAFAP,
           AAFAP LIKE ANLC-AAFAP,
           AFABE LIKE ANLC-AFABE,
           GJAHR LIKE ANLC-GJAHR,
           END OF T_ANLC.
    DATA : BEGIN OF IT_FINAL OCCURS 0,
            ANLN1 LIKE ANLA-ANLN1,
            ANLN2 LIKE ANLA-ANLN2,
            AKTIV LIKE ANLA-AKTIV,
            TXT50 LIKE ANLA-TXT50,
            ZUGDT LIKE ANLA-ZUGDT,
            MENGE LIKE ANLA-MENGE,
            MEINS LIKE ANLA-MEINS,
            GSBER LIKE ANLZ-GSBER,
            KOSTL LIKE ANLZ-KOSTL,
            STORT LIKE ANLZ-STORT,
            RAUMN LIKE ANLZ-RAUMN,
            KFZKZ LIKE ANLZ-KFZKZ,
            BELNR LIKE ANEK-BELNR,
            BUDAT LIKE ANEK-BUDAT,
            XBLNR LIKE ANEK-XBLNR,
            SGTXT LIKE ANEK-SGTXT,
            ANSWL LIKE ANLC-ANSWL,
            NAFAP LIKE ANLC-NAFAP,
            AAFAP LIKE ANLC-AAFAP,
            ANLKL LIKE ANLA-ANLKL,
            GJAHR LIKE ANEK-GJAHR,
            END OF IT_FINAL.
    *added for date logic by sanjeev
    DATA:V_LOW LIKE SY-DATUM,
        V_HIGH LIKE SY-DATUM,
        V_YEAR(04) TYPE N.
    *end of date logic
    DATA : LINE_COLOR(4) TYPE C.
    DATA : FIELDCATALOG TYPE SLIS_T_FIELDCAT_ALV WITH HEADER LINE,
           GD_LAYOUT TYPE SLIS_LAYOUT_ALV.
    DATA: EVENT  TYPE SLIS_ALV_EVENT OCCURS 0 WITH HEADER LINE.
    SELECTION-SCREEN : BEGIN OF BLOCK B_FA WITH FRAME TITLE TEXT-001.
    SELECT-OPTIONS : S_BUKRS FOR ANLA-BUKRS DEFAULT 'ML',
                     S_ANLN1 FOR ANLA-ANLN1,
                     S_ANLN2 FOR ANLA-ANLN2,
                     S_ANLKL FOR ANLA-ANLKL,
                     S_AKTIV FOR ANLA-AKTIV,
                     S_GSBER FOR ANLZ-GSBER,
                     S_KOSTL FOR ANLZ-KOSTL,
                     S_WERKS FOR ANLZ-WERKS,
                     S_STORT FOR ANLZ-STORT,
                    S_BUDAT FOR ANEK-BUDAT.
    SELECTION-SCREEN : END OF BLOCK B_FA.
    SET PF-STATUS '100'.
    added for date logic by Sanjeev
    IF S_BUDAT-HIGH IS INITIAL.
    V_YEAR = S_BUDAT-LOW+0(4).
    V_YEAR = V_YEAR - 1.
    CONCATENATE V_YEAR '0401' INTO V_LOW.
    V_HIGH = S_BUDAT-LOW.
    ELSE.
    V_LOW = S_BUDAT-LOW.
    V_HIGH = S_BUDAT-HIGH.
    ENDIF.
    end of logic
    SELECT BUKRS ANLN1 ANLN2 AKTIV TXT50 ZUGDT MENGE MEINS ANLKL
           FROM ANLA
           INTO CORRESPONDING FIELDS OF TABLE T_ANLA
           WHERE ANLN1 IN S_ANLN1
           AND   ANLN2 IN S_ANLN2
           AND   BUKRS IN S_BUKRS
           AND   ANLKL IN S_ANLKL
           AND   AKTIV IN S_AKTIV
           AND   ZUGDT >= V_LOW  " added to get the date logic work
           AND   ZUGDT <= V_HIGH. " added to get the date logic work
    CHECK : NOT T_ANLA[] IS INITIAL.
    SELECT BUKRS ANLN1 ANLN2 GSBER KOSTL STORT RAUMN KFZKZ
           FROM ANLZ
           INTO CORRESPONDING FIELDS OF TABLE T_ANLZ
           FOR ALL ENTRIES IN T_ANLA
           WHERE ANLN1 = T_ANLA-ANLN1
           AND   ANLN2 = T_ANLA-ANLN2
           AND   BUKRS = T_ANLA-BUKRS
           AND   GSBER NOT IN S_GSBER
           AND   KOSTL IN S_KOSTL
           AND   WERKS IN S_WERKS.
    SELECT BUKRS ANLN1 ANLN2 BELNR BUDAT XBLNR SGTXT GJAHR
           FROM ANEK
           INTO CORRESPONDING FIELDS OF TABLE T_ANEK
           FOR ALL ENTRIES IN T_ANLA
           WHERE ANLN1 = T_ANLA-ANLN1
           AND   ANLN2 = T_ANLA-ANLN2
           AND   BUKRS = T_ANLA-BUKRS
           AND   BUDAT >= V_LOW    " added to get the date logic work
           AND   BUDAT <= V_HIGH.  " added to get the date logic work
    SELECT BUKRS ANLN1 ANLN2 ANSWL NAFAP AAFAP AFABE GJAHR
           FROM ANLC
           INTO CORRESPONDING FIELDS OF TABLE T_ANLC
           FOR ALL ENTRIES IN T_ANLA
           WHERE ANLN1 = T_ANLA-ANLN1
           AND   ANLN2 = T_ANLA-ANLN2
           AND   BUKRS = T_ANLA-BUKRS
           AND   AFABE = '01'.
    SORT T_ANLC BY BUKRS ANLN1 ANLN2.
    SORT T_ANLZ BY BUKRS ANLN1 ANLN2.
    LOOP AT T_ANLA.
      MOVE : T_ANLA-ANLN1 TO IT_FINAL-ANLN1,
             T_ANLA-ANLN2 TO IT_FINAL-ANLN2,
             T_ANLA-AKTIV TO IT_FINAL-AKTIV,
             T_ANLA-TXT50 TO IT_FINAL-TXT50,
             T_ANLA-ZUGDT TO IT_FINAL-ZUGDT,
             T_ANLA-MENGE TO IT_FINAL-MENGE,
             T_ANLA-MEINS TO IT_FINAL-MEINS.
      READ TABLE T_ANLZ WITH KEY BUKRS = T_ANLA-BUKRS
                                 ANLN1 = T_ANLA-ANLN1
                                 ANLN2 = T_ANLA-ANLN2 BINARY SEARCH.
      IF SY-SUBRC EQ 0.
        MOVE : T_ANLZ-GSBER TO IT_FINAL-GSBER,
               T_ANLZ-KOSTL TO IT_FINAL-KOSTL,
               T_ANLZ-STORT TO IT_FINAL-STORT,
               T_ANLZ-RAUMN TO IT_FINAL-RAUMN,
               T_ANLZ-KFZKZ TO IT_FINAL-KFZKZ.
      ENDIF.
      READ TABLE T_ANLC WITH KEY BUKRS = T_ANLA-BUKRS
                                 ANLN1 = T_ANLA-ANLN1
                                 ANLN2 = T_ANLA-ANLN2 BINARY SEARCH.
      IF SY-SUBRC EQ 0.
        MOVE : T_ANLC-ANSWL TO IT_FINAL-ANSWL,
               T_ANLC-NAFAP TO IT_FINAL-NAFAP,
               T_ANLC-AAFAP TO IT_FINAL-AAFAP.
      ENDIF.
    added to get the date logic work
    LOOP AT T_ANEK WHERE BUKRS = T_ANLA-BUKRS AND ANLN1 = T_ANLA-ANLN1 AND
        ANLN2 = T_ANLA-ANLN2 AND BUDAT >= V_LOW AND BUDAT <= V_HIGH.
        MOVE : T_ANEK-BELNR TO IT_FINAL-BELNR,
               T_ANEK-BUDAT TO IT_FINAL-BUDAT,
               T_ANEK-GJAHR TO IT_FINAL-GJAHR,
               T_ANEK-XBLNR TO IT_FINAL-XBLNR,
               T_ANEK-SGTXT TO IT_FINAL-SGTXT.
        APPEND IT_FINAL.
      ENDLOOP.
    ENDLOOP.
    PERFORM BUILD_FIELDCATALOG.
    PERFORM DISPLAY_ALV_REPORT.
    PERFORM BUILD_LAYOUT.
          FORM BUILD_FIELDCATALOG                                       *
    FORM BUILD_FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'ANLN1'.
      FIELDCATALOG-SELTEXT_M = 'ASSET MAIN NO'.
      FIELDCATALOG-COL_POS   = 0.
      FIELDCATALOG-OUTPUTLEN = 14.
      FIELDCATALOG-EMPHASIZE = 'X'.
      FIELDCATALOG-KEY       = 'X'.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'ANLN2'.
      FIELDCATALOG-SELTEXT_M = 'SUB NO'.
      FIELDCATALOG-COL_POS   = 1.
      FIELDCATALOG-OUTPUTLEN = 7.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'AKTIV'.
      FIELDCATALOG-SELTEXT_M = 'CAP.DATE'.
      FIELDCATALOG-COL_POS   = 2.
      FIELDCATALOG-OUTPUTLEN = 10.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'TXT50'.
      FIELDCATALOG-SELTEXT_M = 'NAME'.
      FIELDCATALOG-COL_POS   = 3.
      FIELDCATALOG-OUTPUTLEN = 50.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'ZUGDT'.
      FIELDCATALOG-SELTEXT_M = 'POSTING DATE'.
      FIELDCATALOG-COL_POS   = 4.
      FIELDCATALOG-OUTPUTLEN = 10.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'MENGE'.
      FIELDCATALOG-SELTEXT_M = 'QTY'.
      FIELDCATALOG-COL_POS   = 5.
      FIELDCATALOG-OUTPUTLEN = 13.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'MEINS'.
      FIELDCATALOG-SELTEXT_M = 'UNITS'.
      FIELDCATALOG-COL_POS   = 6.
      FIELDCATALOG-OUTPUTLEN = 3.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'GSBER'.
      FIELDCATALOG-SELTEXT_M = 'BA'.
      FIELDCATALOG-COL_POS   = 7.
      FIELDCATALOG-OUTPUTLEN = 4.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'KOSTL'.
      FIELDCATALOG-SELTEXT_M = 'COST CENTER'.
      FIELDCATALOG-COL_POS   = 8.
      FIELDCATALOG-OUTPUTLEN = 13.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'STORT'.
      FIELDCATALOG-SELTEXT_M = 'LOCATION'.
      FIELDCATALOG-COL_POS   = 9.
      FIELDCATALOG-OUTPUTLEN = 10.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'RAUMN'.
      FIELDCATALOG-SELTEXT_M = 'ROOM'.
      FIELDCATALOG-COL_POS   = 10.
      FIELDCATALOG-OUTPUTLEN = 10.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'KFZKZ'.
      FIELDCATALOG-SELTEXT_M = 'LICENSE PLATE NO'.
      FIELDCATALOG-COL_POS   = 11.
      FIELDCATALOG-OUTPUTLEN = 20.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'BELNR'.
      FIELDCATALOG-SELTEXT_M = 'DOC NO'.
      FIELDCATALOG-COL_POS   = 12.
      FIELDCATALOG-OUTPUTLEN = 13.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'BUDAT'.
      FIELDCATALOG-SELTEXT_M = 'POSTING DATE'.
      FIELDCATALOG-COL_POS   = 13.
      FIELDCATALOG-OUTPUTLEN = 15.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'XBLNR'.
      FIELDCATALOG-SELTEXT_M = 'REF NO'.
      FIELDCATALOG-COL_POS   = 14.
      FIELDCATALOG-OUTPUTLEN = 25.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'SGTXT'.
      FIELDCATALOG-SELTEXT_M = 'TEXT'.
      FIELDCATALOG-COL_POS   = 15.
      FIELDCATALOG-OUTPUTLEN = 50.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'ANSWL'.
      FIELDCATALOG-SELTEXT_M = 'ACQ VAL'.
      FIELDCATALOG-COL_POS   = 16.
      FIELDCATALOG-OUTPUTLEN = 13.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'NAFAP'.
      FIELDCATALOG-SELTEXT_M = 'ORD.DEP'.
      FIELDCATALOG-COL_POS   = 17.
      FIELDCATALOG-OUTPUTLEN = 13.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
      FIELDCATALOG-FIELDNAME = 'AAFAP'.
      FIELDCATALOG-SELTEXT_M = 'UNPLAN DEP'.
      FIELDCATALOG-COL_POS   = 18.
      FIELDCATALOG-OUTPUTLEN = 13.
      APPEND FIELDCATALOG TO FIELDCATALOG.
      CLEAR FIELDCATALOG.
    ENDFORM.
    PERFORM ALVDISPLAY.
    FORM ALVDISPLAY.
    EVENT-FORM = 'USERCOMMAND'(002).
    EVENT-NAME = 'USER_COMMAND'(001).
    APPEND EVENT.
    ENDFORM.
    FORM DOWNLOADEXCEL.
    DATA : B_FILE1 LIKE RLGRAP-FILENAME.
    ENDFORM.
          FORM BUILD_LAYOUT                                             *
    FORM BUILD_LAYOUT.
      GD_LAYOUT-ZEBRA = 'X'.
      GD_LAYOUT-INFO_FIELDNAME = 'LINE_COLOR'.
    ENDFORM.
          FORM DISPLAY_ALV_REPORT                                       *
    FORM DISPLAY_ALV_REPORT.
      CALL FUNCTION 'REUSE_ALV_LIST_DISPLAY'
          EXPORTING
            I_INTERFACE_CHECK        = ' '
             I_CALLBACK_PROGRAM       = SY-REPID
            i_callback_pf_status_set = 100
            i_callback_user_command  = sy-ucomm
            I_STRUCTURE_NAME         =
             IS_LAYOUT                = GD_LAYOUT
               IT_FIELDCAT              = FIELDCATALOG[]
            IT_EXCLUDING             =
            IT_SPECIAL_GROUPS        =
            IT_SORT                  =
            IT_FILTER                =
            IS_SEL_HIDE              =
            I_DEFAULT                = 'X'
            I_SAVE                   = ' '
            IS_VARIANT               = ' '
            IT_EVENTS                =
            IT_EVENT_EXIT            = EVENT[]
            IS_PRINT                 =
            I_SCREEN_START_COLUMN    = 0
            I_SCREEN_START_LINE      = 0
            I_SCREEN_END_COLUMN      = 0
            I_SCREEN_END_LINE        = 0
       IMPORTING
            E_EXIT_CAUSED_BY_CALLER  =
            ES_EXIT_CAUSED_BY_USER   =
           TABLES
                T_OUTTAB                 = IT_FINAL
           EXCEPTIONS
                PROGRAM_ERROR            = 1
                OTHERS                   = 2.
    ENDFORM.

  • Data source for Deposit Advice (XML) and Live CheckWrite (XML) reports

    Hello All,
    We need to customize the Deposit Advice (XML) and Live Check Write(XML) reports to accomodate a few business requirements. We already have customizations for these reports in non-XML versions. For Ex: USPAYPST.rdf for Deposit Advice report. As these non-XML RDF reports are going to be unsupported by Oracle soon, we need to customize the XML versions of these report available in R12.1.3.
    As these reports are kicked off by the PYUGEN spawned program, therefore I need information on the data source for these reports as soon as possible.
    Looking forward to your prompt help.
    Thanks,
    Rahul

    Now, I have anew problem here. I modified the USPAYPST.rdf report for Deposit Advice and then changed concurrent program output to XML from Text.
    I attached a RTF template to the concurrent program to have the output in XML.
    When I run Deposit Invoice Report from the SRS window, it finishes off normally and log file says the number of records processed but no output file is generated.
    I also tried looking into the XML generated through View XMl and there too no output file is generated.
    I am confused if the Deposit Advice program can only generate Text output. As PYUGEN is the executable for this report, I have no idea if XML file can even be generated.
    All pointers to this issue are welcome.
    Thanks,
    Rahul

  • How to determine the creation date/time for a file?

    The important operating systems maintain both a creation date/time and last modified date/time for files. But in the File class there is only a lastModified() method. How does one determine the creation date/time for a file?

    As far as i know, there is no way to know creation time, since it is a OS dependant information.

  • Using Oracle Stored Proc as Data Source for PowerBuilder Data Window Object

    Hi,
    I'm trying to use an Oracle stored procedure as the data source for a PowerBuilder data window object. I'm using the following (which is a variation of an example given on Sybase's web site, with the appropriate database/column name changes and use of %TYPE):
    ======================
    CREATE OR REPLACE PACKAGE pkg_dw_sp_example
    AS
    TYPE typ_example_rec1 IS RECORD(
    v_comcodsa st_sa.comcodsa%TYPE,
    v_comnamsa st_sa.comnamsa%TYPE,
    v_cocitysa st_sa.cocitysa%TYPE);
    TYPE cur_example_cur1 IS REF CURSOR;
    RETURN typ_example_rec1;
    END;
    ======================
    CREATE OR REPLACE PROCEDURE sp_dw_example1
    (arg_teamid IN st_sa.comcodsa%TYPE,
    cur_team IN OUT pkg_dw_sp_example.cur_example_cur1)
    AS
    BEGIN
    OPEN cur_team FOR
              SELECT comcodsa,
              comnamsa,
              cocitysa
              FROM st_sa
              WHERE st_sa.comcodha = arg_teamid;
    END sp_dw_example1;
    ======================
    The package is created okay, but the procedure is saved with compilation errors. Unfortunately, Oracle does not indicate what the errors are (great product), so I have no idea what it is complaining about.
    Questions:
    1) Can anyone point me to/or provide a better example?
    2) The examples in books I have seen all would have declared the procedure in the package specification - is the above valid?
    3) This is a PowerBuilder specific question, but if the package does not show up in the stored procedure list when creating a data window object, how does one use it as a data source?
    Oracle 8.1.7
    PowerBuilder 8.0.1
    Thanks,
    Ed Holloman

    1) Everything looks fine with your code. You did not mention the environment you are compiling your code in, but if it is SQL*Plus, you can type 'show errors' after the compile and it will list the errors and line numbers for you.
    2) You can include the procedure in the package - in which case you would have a procedure declaration in the spec and the actual procedure in the body. You would no longer include the 'create or replace' - the declaration and procedure would just start with 'procedure'. The spec would look include:
    PROCEDURE sp_dw_example1
      (arg_teamid IN st_sa.comcodsa%TYPE,
       cur_team OUT cur_example_cur1);Note the removal of the package reference (since we are now IN the package) and the second parameter is just OUT instead of IN OUT.
    3) Not much, but I know that in PB8, our developers have said they have visibility now to the package when creating a DW - something they didn't have in PB7. Previously, we had to create a dummy standalone procedure that they would select and they would later modify the code to include the package prefix.
    Hope this helps a bit.

  • Should cancelled mailbox moves result in any moved mail data removed from target server?

    I was moving mailboxes from 1 DB to another. 3 out of 120 were left running for 12 hours; they were the biggest ones and it was very slow to process the items. There were also reports of slowness on the server so I had no choice but to cancel the moves.
    My question is should cancelled mailbox moves result in any moved mail data being removed from the target database? It doesn't appear that my disk as increased in space again so I'm wondering if it's there as whitespace maybe? There is an event logged to say
    the mailbox has been deleted from the target.
    Would appreciate if someone can clear this up and possibly explain exactly what occurs during this type of scenario.
    Thanks.

    Hi,
    Thank you for your kindly remind.
    In Exchange 2000, 2003 and 2007 mailboxes become unavailable during movement, when a mailbox move takes place, the mailbox isn't actually moved, it's just copied to the destination location. Once that copy has finished, AD is updated to point to the new location
    and the old mailbox is deleted from the source EDB file.
    The most different between Exchange 2010 and legacy version in move mailbox is that allows end-users to be online in their email accounts. On completing the move users just needs to reopen their Outlook clients.
    Mailbox Move Requests are a new feature set of Exchange 2010 and requires running a series of Cmdlets to perform asynchronous online mailbox moves with the help of a service agent called Mailbox Replication Service (MRS). MRS is available on all Exchange 2010
    Client Access Servers.
    Best Regards,
    Allen Wang

Maybe you are looking for

  • ITunes 10 Quit Unexpectedly

    I just installed iTunes 10 via Software Update and now I cannot open iTunes at all. I've tried uninstalling and reinstalling, but no luck. Any ideas? Process: iTunes [296] Path: /Applications/iTunes.app/Contents/MacOS/iTunes Identifier: com.apple.iTu

  • Pass value from previous interactive report to create record form

    Hi All, I am an APEX newbie and have an APEX 3.2.1 application that is still being developed that is working reasonably well. The problem I am having is the following: I have 3 interactive reports. These are defined on their respective tables, namely

  • Can Any One Please Help on this Requirement

    Hi All ,             I am trying to creating a vendor in XK01 in Address screen im Enterying country as US , But System is showing this error (Incompete Address ; enter country ) im giving all the details , can any one explainn me where i am doing ro

  • Cannot produce log in servlet

    Hi everyone, I am using OAS 4.0.8 on Solaris 2.6. I cannot produce log file with content. The file is there but with zero byte. The result is the same for servlet log API(I do include init (super.init..) and wrb logger API. Thank you for your time. n

  • Business Transaction Event(BTE)

    hi what is  Business Transaction Event  and how to use this.any links available for this.