SSM 7.0 - Copy SAP BI data into PAS or Create Dim. Model viewing BI Data

I am using SSM 7.0 with SAP BI.  Is there some best practice / direction / comments on making a decision between:
#1 - Duplicate SAP BI data into PAS Dimensional Model
- Create an InfoCube in SAP BI
- Create the model connection required by BI Connection
- Copy the required SAP BW/BI InfoCube information into a PAS Model using the BI Connector
- Create KPIs based on this model connection
or
#2 - Make SAP BI data visible in PAS Dimensional Model (no data stored in PAS)
- Create an InfoCube in SAP BI
- Create a Dimensional Model based on this InfoCube using a Link ID and the SCHEMA commands
- Create a model connection to make data in InfoCube available in SSM
- Create KPIs based on this model connection
My thinking is that if the volume of data is low then not actually having to go to the trouble of duplicating the data in SSM would work well - with the benefit that the information is updated as soon as the SAP BI InfoCube is updated.
(At the moment I have tried #2 but so far not failed to create an KPIs - I am wondering if the Key Figures in the InfoCube need to use a particular naming standard)

Hi Mark,
See the link below for a Best practice document that is available here in SDN. Although you can use queries that already exist in BW, it is usually better to create new queries adjusted to the kind of information you want to provide in SM. There are a few rules you need to follow when creating such queries, but there is actually no naming convention for its Key-Figures.
[Best Practiceu2026 The BI Connector in SSM 7.0|https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/701b8da0-3135-2b10-929d-936e43164b06]
Hope this helps!
Ricardo

Similar Messages

  • Populating data into internal table created dynamically

    Hi all,
    I have created an internal table dynamically using the method
    CALL METHOD CL_ALV_TABLE_CREATE=>CREATE_DYNAMIC_TABLE
        EXPORTING
          IT_FIELDCATALOG = IT_FIELDCAT2[]
        IMPORTING
          EP_TABLE        = L_TABLE.
      ASSIGN L_TABLE->* TO <IT_TABLE>.
      CREATE DATA wa_table LIKE LINE OF <it_table>.
      ASSIGN wa_table->* TO <wa_table1>.
    when I tried to append data into the internal table <it_table>
    as below
    ex :  <wa_table1>-prueflos = wa_item-prueflos.
    it is giving the error as 'The data object <wa_table1> has no structure therefore no component called prueflos'.
    can anybody tell me how to append data into a dynamically created internal table.
    Thanks,
    Sudheer

    Hi Sudheer,
    You need to use the ASSIGN COMPONENT command for this purpose. Check this code:
    FIELD-SYMBOLS: <t_dyntable> TYPE STANDARD TABLE,  u201C Dynamic internal table name
                   <fs_dyntable>,                     u201C Field symbol to create work area
                   <fs_fldval> type any.              u201C Field symbol to assign values 
    * Assume your internal table has fields COL1, COL2, COL3 etc
      DO p_cols TIMES. 
        index = sy-index.
        MOVE sy-index TO wa_colno.
        CONCATENATE 'COL'
                    wa_colno
               INTO wa_flname. 
    * Set up fieldvalue
        CONCATENATE 'VALUE' index INTO
                    fieldvalue.
        CONDENSE    fieldvalue NO-GAPS. 
        ASSIGN COMPONENT  wa_flname
            OF STRUCTURE <fs_dyntable> TO <fs_fldval>.
        <fs_fldval> =  fieldvalue. 
      ENDDO. 
    Hope this solves your problem!
    Cheers,
    Shailesh.

  • Merging Excel data into a form created in LiveCycle

    Hello,
    I've made a form in LiveCycle.  The first of the 5 pages I want to pre-populate with data before I distribute the final form.  My data is in Excel and I have ensured that the header titles match the field names on the form.  Could someone help me to get the data into the form please.  I'm pretty new to xml and have found following the Help difficult.
    Thanks
    Scott

    Please re-ask in the LiveCycle Designer forum  http://forums.adobe.com/community/livecycle/livecycle_es/livecycle_designer_es  That is where people who use the Designer product hang out.

  • How to writing the data into the file at the same frequency of data acquisition using myRIO

    Hi everyone,
    I have a question regarding data acquisition fequency and data recording frequency of myRIO. Hope you guys can help me out.
    Basically, I want to acquire voltage input at analog input 0 at the frequency of 1kHz and then write the data into a file (tdms format). 
    However, I always found there were only 55 or 56 data points recorded every second in the data file (see the excel sheet). 
    To confirm my data acquisition was performed at the correct frequency, I added a small function in the main loop to indicate the time spent between two acquisition events. 
    To my surprise, the period of data acquistion is correct (1ms or 1kHz) but there are only 55 or 56 data point per second recorded in the data file.
    How can I record every data point acquired by the analog input?
    Thank you!
    P.S. I am very new to myRIO. How can I manually set the system time for myRIO? The default time of my myRIO is wrong. 
    Best,
    Tengyang
    Attachments:
    test result.xlsx ‏16 KB
    Main with timed loop.vi ‏122 KB
    test result.xlsx ‏16 KB

    Have a look at the Jakarta POI project, they have a Java API for creating Word documents.
    http://jakarta.apache.org/poi/hwpf/index.html

  • Loading data into csv file based on the count of data in table in ODI

    Hi,
    I have a requirement like i have data in table and need to be loaded in csv file.
    If the count of the data in the table exceeds the csv data limit then i need to call the same procedure again to
    generate one more csv file and load the remaining data into it.my data in the table may vary all the time.
    I am new to this ODI.Can anyone tell any ways to do it in ODI or tell me any logic to do this?
    Thanks in advance..
    Edited by: 883410 on May 22, 2012 12:01 AM

    What limit on the csv file are you talking about. There isn't a limit (unless you are on Unix where the filesize is limit is 2GB). Are you talking about opening it in earlier versions of Excel where the display limit is 60,000 rows as this is irrelevant for your operation?

  • How2 load xml into datagrid, and create realtime search of datagrid data

    I wanted to be able to load xml data into a datagrid and then create a searchbox that would search in real time (as person typed) from any part of the datagrid.  Does anyone know of any as2.0 component that can do that easily?  If not, i would like to try to tackle this the old fashioned and use this to teach myself as2.0 advanced coding.  can someone point to a step by step or explain how i might get this accomplished?  from loading the xml to putting the node info into the datagrid, to being able to search it in real time.  This is part of a project that i was given that has some serious consequences if i botch it.  Please help!  Thanks so much!

    import fl.controls.DataGrid;
    import fl.controls.ScrollPolicy;
    import fl.data.DataProvider;
    import flash.net.URLLoader;
    import flash.events.Event;
    var dp:DataProvider = new DataProvider();
    var dg:DataGrid = new DataGrid();
    dg.columns=["User","Xp","Rank"];
    addChild(dg);
    var urlloader:URLLoader = new URLLoader();
    urlloader.addEventListener(Event.COMPLETE,loadcompleteF);
    urlloader.load(new URLRequest("whatever.xml"));
    function loadcompleteF(e:Event):void {
        var xml:XML=XML(e.target.data);
        for (var i:int = 0; i < xml..Name.length(); i++) {
            dp.addItem({User:xml..Name[i].User.text(), Xp:xml..Name[i].Xp.text(), Rank:xml..Name[i].Rank.text()});
        dg.dataProvider=dp;

  • Error when create classification model by Oracle Data Miner 10.1.0.2

    Hello every body!
    Please help me!
    I have created a classification model by Oracle Data Mining 10.1.0.2, Oracle 10g Release 1 (10.1.0.3) with following options:
    Single record per case
    Adaptive Bayes Network algorithms
    SingleFeatureBuild
    When finished, i receive an error with following detail content:
    ORA-40101: Data Mining System Error ODM_ABN_MODEL-ODM_ABN_BUILD--20002
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
    ORA-06512: at "DMSYS.ODM_ABN_MODEL", line 458
    ORA-06512: at "DMSYS.ODM_ABN_MODEL", line 5664
    ORA-40101: Data Mining System Error ODM_ABN_MODEL-ODM_ABN_BUILD--20012
    ORA-06512: at line 1
    Hope you help me!
    Thanks!

    Hi,
    Can you provide instructions that recreate this problem using sample data we provide in the SH schema?
    Thanks, Mark

  • Edit Date updated for whole catalog (how to view edit date?)

    I suddenly found out that my publish service that exports newly edited photos suddenly contained all my photos, and I have no idea how that happened. It seems the Edit Date has been changed, but I can not find any way to view the date, so it's difficult to pinpoint exactly when it happened.
    So - two questions:
    How could that happen? I have not made any mass updates and not changed the development version either
    How to view the Edit Date?
    Thanks

    Well, a good guess would be: something is getting (or did get) changed.
    If its not a plugin, then its something else (at the risk of stating the obvious...)
    ChangeManager can provide a detailed list of all develop settings and metadata changes between any two states of a photo, so if its develop settings or metadata that's changing, then you can see exactly what.
    Some kinds of changes ChangeManager can not detail - if the changes are falling in the "mystery" class, at least you'd have it narrowed down... - may be worth trying...
    PS - Have you checked Metadata Status Date in the library panel - thats the closest thing I know of to the edit-date that is exposed by the Lightroom UI - you cant filter or collect on it, but you can at least view it.
    R

  • How can i get pr header note data into po while creating a po using pr no

    Hi all,
    Can any one tell me how can i get the PR header note's data in PO when i am creating a PO using that PR.
    and what is the name of table in which the header note's data is copied.
    Thanks
    Edited by: Saurabh.Shrivastava on Jan 17, 2012 3:54 PM

    Hi,
    Unfortunately I have to tell you that the described behaviour is not
    an Error but SAP Standard behaviour.
    Please see note 448814 question 7.
    In SAP Standard it is not foreseen to take over the header text of a
    PREQ to a PO !
    You can try to use BADI ME_REQ_HEADER_TEXT .
    I have found a document which gives coding for Purchase Requisition Header Long Text using Badi - ME_PROCESS_REQ_CUST:
    http://wiki.sdn.sap.com/wiki/display/ABAP/PurchaseReq.HeaderLongTextusingBadi-ME_PROCESS_REQ_CUST
    I hope it can help you!
    Best Regards,
    Arminda Jack

  • How do you save data into an excel file while myRIO is acquiring data? I tried saving it using "Write to file" but it doesn't work for some reason.

    I am acquiring cosine wave and a pulse wave as input and I want to store their peak to peak values into an excel file. "Write to File" is not working for it. Is there any other vi which can be used for data logging?
    Thank you for your help.

    Hi Ssheoran,
    Can you provide more detail when you say that the Write to File VI doesn't work? Is there an error given? Or can you just not find the file on your computer? Keep in mind using this file in a Real-Time VI on the myRIO will save files on the myRIO. You will then have to transfer to your PC. Please view the following video as a guide for saving files and transferring them to your computer: (http://www.youtube.com/watch?v=BuREWnD6Eno). Hope this helps.
    Best Regards,
    Roel F.
    Applications Engineer
    National Instruments

  • Error loading essbase data into flat files through DIM

    Hi All,
    While running the session in DIM 8.1.1 which will load data from Hyperion essbase (11.1.1.3) to flat file it fails.
    From the log file it shows failes to load the HASReader plugins.
    All the environmental variables are properly set.
    Still I am facing this issue.
    Could anymore suggest me what I am missing.
    Thanks,

    Hi Avs sai,
    By default, the columns in the destination flat file will be non-Unicode columns, e.g. the data type of the columns will be DT_STR. If you want to keep the original DT_WSTR data type of the input column when outputting to the destination file, you can check
    the Unicode option on the “Choose a Destination” page of the SQL Server Import and Export Wizard. Then, on the “Configure Flat File Destination” page, you can click the Edit Mappings…“ button to check the data types. Please see the screenshot:
    Regards,
    Mike Yin
    TechNet Community Support

  • Role for create service and view Master Data

    Hey people !
    Im using CRM 4.0 and want to assign some role.
    The role should be able to do all the transaction in service and masterdata.
    Someone got any tip?

    You can use SAP delivered Interaction Center Manager or Service Manager role.
    Thanks,
    Thirumala.

  • Load one text file with 12 periods' data into 12 different periods at once?

    Hi guys,
    In one swoop, can we load one .txt file with 12 periods of data into 12 different periods?
    The scenario:
    Budget data is required for monthly comparative reporting with actuals, so we have 12 periods in our Budget version.
    From a non-SAP system we get one .txt file containing 12 periods worth of budget data,
    - it is in the correct format (and we don't want to create risk by opening and editing it) so it is loaded into each period and the other 11 periods of irrelevant data are obviously ignored.
    Some extra tasks (such as validation, cashflow calculations) are then performed per period.
    Because this has to be repeated 12 times, this can be a time consuming process
    We now have multiperiod monitor functionality (from EHP2) and I see how it works for the automatic tasks (very well).
    I'm aware that the guide says that manual tasks will be ignored during the automatic run, this is true.
    However, if I remember correctly EC-CS used to allow it with upoad files, so i was expecting it in BCS 6.02
    - is there anyway to load a file containing 12 periods worth of data into 12 individual periods in all at once?
    (NB we still have an improvement to the previous situation, the user can scroll between periods more quickly and load the file 12 times, then go back to the start and run all auto tasks at once)..
    One thought was to use a file server location with a hardcoded filename but his would require work beyond my expertise.
    All suggestions welcome

    Hi
    I wil suggest mapping the path based on a common location is the best way and then map the same logically in the Upload method
    If you are using Citrix then it becomes all the more easy to have a commmon location
    Rgds
    Dheeraj

  • Problems with special characters uploading data into oracle database

    Greetings
    I have a problem uploading data into oracle tables from sybase datasource. The data that I want to upload is in Spanish, for example when I have a varchar field with the data 'Consultoría', in oracle table the data upload with interrogation symbols.
    I have my source and target datastores configured as follows.
    Any suggestion? Thank you for your time

    Chack section 10 Locales and Multi-byte Functionality in the SAP Data Services Reference Guide for an exhaustive description of NLS settings.
    As a summary, following settings are required:
    locales in the datastore definitions match the actual locale used in the associated databases
    the locale of the DS engine allows for source to target translation
    the character set in the target database supports storage of accented characters

  • Facing  problem while loading data into XI

    Hi All,
    I am facing a peculiar problem while loading data into XI when we load small amount of data say 5-10 records these records gets processed, but when we try to load a bit higher volume say 100-500 which comes out to be around 7 record s per second then the record processing gets stalled. That is it gives a message that "recorded for outbound processing".
    So can any body throw some light on it and can tell us why we rare facing such problem and how can we resolve that problem.
    Thanks in Advance.
    Regards
    rahul

    Hi rahul,
    Deregistering and Registering the Queue is a temporary solution.
    The better way to this would be to Tune your Xi system as already mentioned by Vijaya as mentioned in the guide ,
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/70ada5ef-0201-0010-1f8b-c935e444b0ad
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/7fdca26e-0601-0010-369d-b3fc87d3a2d9
    Regards,
    Bhavesh

Maybe you are looking for

  • How can I create a new color in the calendar?

    Good day, I wanted the to create a new color in the default calendar from Apple, but I do not know how that goes. Perhaps someone can explain to me how this works. I would be very grateful for any advice.

  • IOS6 problem

    Even you can download iOS 6 to iPhone 3GS but you can not download to an iPad 1 ! is it fair ? i dont think so. Is there any answer for this situation ?

  • Store links to uploaded files in mysql database

    Hello. What I'm trying to do is, let a user upload a file to my server, and lets other users download those files. I don't want to store the files in a database, because that will slooow things down. What I was wondering is how I would be able to sto

  • New video format on older Mac

    I'm looking at getting a new camcorder that has ACVHD format, but I have a G4 powerpc (tower) that runs at 400 Mhz and has 800 RAM. iMovie 3. Can I download, or am I nuts? Will it just take a long time, or is it impossible? Thanks.

  • How do i add a smaller video into the corner of a full-screened video?

    i had a video all ready and running and what-not and i wanted to add another video into the corner of it to show a different angle. how can i do this with premiere cs6?