Master data load from 2 source system

Hi all,
        I am working on a project which has 2 source systems.
Now I have to load master data from 2 source systems without making 0LOGSYS as compunding
attribute. Because all the objects which i want to use compunding attribute are reference
objects for lot of other info objects. So, i dont want to change the business content
objects structure. Please guide me
Thanks

Hi,
I cube there is nothing to do with compounding attribute, it will be handled in MD IO level.  As your requirement to separate the transaction happen in different source system then make add another info object as Zsource and make it as nav. attribute then design your report based on nav. attribute.
Note that this separation is only possible in Material Related transaction not for others.
Thanks
BVR

Similar Messages

  • Regarding master data loading for different source systems

    Hi Friends,
    I have an issue regarding master data loading.
    we have two source systems one is 4.6c and another is ecc 6.0.
    First i am loading the master data from 4.6c to bi7.0.
    Now this 4.6c is upgraded to ecc6.0.
    *In 4.6c and ecc6.0c master data is changing.
    After some time there is no 4.6c, only ecc 6.0 is there.
    Now if i load master data from ecc6.0 to bi7.0 what will happen.
    Is it possible ?
    Could you please tell me?
    Regards,
    ramnaresh.

    Hi ramnaresh porana,
    Yes, its possible. You can load data from ECC.
    Data will not change, may be you may get more fields in datasource at r/3 side, but BW/BI side no change in mappings structures are same. So data also same.
    You need to take care of Delta's before and after upgrade.
    Hope it Helps
    Srini

  • How to Load Master Data Text  from Multiple Source Systems

    Situation:  Loading vendor master (text) from two systems.  Some of the vendor keys/number are the same and so is the description.  I understand that if I was loading Attributes, I could create another attribute such a source system id and compound it to 0Vendor.  However, I don't believe you can do the same when loading text. 
    Question:  So, how can I load duplicate vendor keys and description into BW using the *_TEXT datasource so they do not overwrite?
    Thanks!

    Hi,
    Don't doubt, it'll work!
    Sys ID as compound to 0Vendor.
    If you doubt about predefined structure of the *_TEXT datasource and direct infosource, then use a flexible infosource. Maybe you'll have to derive the sys ID in a routine or a formula.
    Best regards,
    Eugene

  • Master Data Loading from APO--- BW System

    Hi Experts,
    I am loading master  data from the APO to bW System,after loading data into the BW System ,I found the number of data records are mismatching
    ex:APO (8340) --->BW (9445)
    and that to I tried to delet the data in InfoObject,
    only some records are deleting
    and one more doubt
    in APO i have info object ZA9MATNR CHAR 40
    With Conversion Routine PRODU
    in BW i have the same info object ZA9MATNR CHAR 40 with Conversion Routine ALPHA.
    and I have checked the Check box in transfer rules also,may be becoz of this I am getting more records.
    Could any body help me
    Thanks in Advance...
    Regards,
    Nagamani.

    Hi,
    Ideally it is impossible to get the increased count of data records when you load it into BW unless you are using the RETURN TABLE concept.
    Normally the record count decreases with the records with same keys gets overwritten/added.
    Coming to your second question, ALPHA conversion routine has nothing to do with the record count. It appends the ZEROS to the value coming in that InfoObject. For example, if the lenght of the object is 5 and data is coming only in 3 digits, it will append the ZEROS to make its length 5.
    Hope this info helps you.
    Regards,
    Yogesh.

  • BPC:: Master data load from BI Process chain

    Hi,
    we are trying to automatize the master data load from BI.
    Now we are using a package with:
    PROMPT(INFILES,,"Import file:",)
    PROMPT(TRANSFORMATION,%TRANSFORMATION%,"Transformation file:",,,Import.xls)
    PROMPT(DIMENSIONNAME,%DIMNAME%,"Dimension name:",,,%DIMS%)
    PROMPT(RADIOBUTTON,%WRITEMODE%,"Write Mode",2,{"Overwirte","Update"},{"1","2"})
    INFO(%TEMPNO1%,%INCREASENO%)
    INFO(%TEMPNO2%,%INCREASENO%)
    TASK(/CPMB/MASTER_CONVERT,OUTPUTNO,%TEMPNO1%)
    TASK(/CPMB/MASTER_CONVERT,FORMULA_FILE_NO,%TEMPNO2%)
    TASK(/CPMB/MASTER_CONVERT,TRANSFORMATIONFILEPATH,%TRANSFORMATION%)
    TASK(/CPMB/MASTER_CONVERT,SUSER,%USER%)
    TASK(/CPMB/MASTER_CONVERT,SAPPSET,%APPSET%)
    TASK(/CPMB/MASTER_CONVERT,SAPP,%APP%)
    TASK(/CPMB/MASTER_CONVERT,FILE,%FILE%)
    TASK(/CPMB/MASTER_CONVERT,DIMNAME,%DIMNAME%)
    TASK(/CPMB/MASTER_LOAD,INPUTNO,%TEMPNO1%)
    TASK(/CPMB/MASTER_LOAD,FORMULA_FILE_NO,%TEMPNO2%)
    TASK(/CPMB/MASTER_LOAD,DIMNAME,%DIMNAME%)
    TASK(/CPMB/MASTER_LOAD,WRITEMODE,%WRITEMODE%)
    But we need to include these tasks into a BI process chain.
    How can we add the INFO statement into a process chain?
    And how can we declare the variables?
    Regards,
    EZ.

    Hi,
    i have followed your recomendation, but when i try to use the process /CPMB/MASTER_CONVERT, with the parameter TRANSFORMATIONFILEPATH and the root of the transformation file as value, i have a new problem. The value only have 60 char, and my root is longer:
    \ROOT\WEBFOLDERS\APPXX\PLANNING\DATAMANAGER\TRANSFORMATIONFILES\trans.xls
    How can we put this root???
    Regards,
    EZ.

  • Master data Load from Oracle table

    Hi,
    My master data comes from an Oracle table. The table has both attributes and texts combined. How do I load the Master data into the text and attribute info objects thru Direct Upload of Master Data?
    Is it necessary to go in for Flexible Upload in this case?
    Regards

    Hi,
    you can create two views on the table. One for the text and the other one for the attributes. Create your datasources on these views and assign them to your infoobject.
    regards
    Siggi

  • Error : Data load from relation source (SQL Server 2005) to Essbase Cube

    Hi All,
    I am looking help from you. I am trying to load data from SQLServer 2005 table to Essbase Cube using IKM SQL to Hyperion Essbase (Metadata) Modules.
    I am getting below error. Let me know if i am missing some things.
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 61, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Invalid value specified [RULES_FILE] for Load option [null]
         at com.hyperion.odi.essbase.ODIEssbaseMetaWriter.validateLoadOptions(Unknown Source)
         at com.hyperion.odi.essbase.AbstractEssbaseWriter.beginLoad(Unknown Source)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
         at org.python.core.PyMethod.__call__(PyMethod.java)
         at org.python.core.PyObject.__call__(PyObject.java)
         at org.python.core.PyInstance.invoke(PyInstance.java)
         at org.python.pycode._pyx3.f$0(<string>:61)
         at org.python.pycode._pyx3.call_function(<string>)
         at org.python.core.PyTableCode.call(PyTableCode.java)
         at org.python.core.PyCode.call(PyCode.java)
         at org.python.core.Py.runCode(Py.java)
         at org.python.core.Py.exec(Py.java)
         at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.g.y(g.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Invalid value specified [RULES_FILE] for Load option [null]
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.g.y(g.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)

    ODI Step: Prepare for Loading Step:
    from java.util import HashMap
    from java.lang import Boolean
    from java.lang import Integer
    from com.hyperion.odi.common import ODIConstants
    from com.hyperion.odi.connection import HypAppConnectionFactory
    # Target planning connection properties
    serverName = "HCDCD-HYPDB01"
    userName = "admin"
    password = "<@=snpRef.getInfo("DEST_PASS") @>"
    application = "BUDGET01"
    database = "PLAN1"
    portStr = "1423"
    srvportParts = serverName.split(':',2)
    srvStr = srvportParts[0]
    if(len(srvportParts) > 1):
    portStr = srvportParts[1]
    # Put the connection properites and initialize the essbase loader
    targetProps = HashMap()
    targetProps.put(ODIConstants.SERVER,srvStr)
    targetProps.put(ODIConstants.PORT,portStr)
    targetProps.put(ODIConstants.USER,userName)
    targetProps.put(ODIConstants.PASSWORD,password)
    targetProps.put(ODIConstants.APPLICATION_NAME,application)
    targetProps.put(ODIConstants.DATABASE_NAME,database)
    targetProps.put(ODIConstants.WRITER_TYPE,ODIConstants.DATA_WRITER)
    print "Initalizing the essbase wrapper and connecting"
    pWriter = HypAppConnectionFactory.getAppWriter(HypAppConnectionFactory.APP_ESSBASE, targetProps);
    tableName = "BUDGET01_PLAN1"
    rulesFile = r"ActLd"
    ruleSeparator = "Tab"
    clearDatabase = "None"
    calcScript = r""
    maxErrors = 1
    logErrors = 1
    errFileName = r"E:\OraHome_ODI\Error\Budget01Plan1.err"
    logEnabled = 1
    logFileName = r"E:\OraHome_ODI\Error\Budget01Plan1.log"
    errColDelimiter = r","
    errRowDelimiter = r"\r\n"
    errTextDelimiter = r"'"
    logHeader = 1
    commitInterval = 1000
    calcOnly = 0
    preMaxlScript = r""
    postMaxlScript = r""
    abortOnPreMaxlError = 1
    # set the load options
    loadOptions = HashMap()
    loadOptions.put(ODIConstants.CLEAR_DATABASE, clearDatabase)
    loadOptions.put(ODIConstants.CALCULATION_SCRIPT, calcScript)
    loadOptions.put(ODIConstants.RULES_FILE, rulesFile)
    loadOptions.put(ODIConstants.LOG_ENABLED, Boolean(logEnabled))
    loadOptions.put(ODIConstants.LOG_FILE_NAME, logFileName)
    loadOptions.put(ODIConstants.MAXIMUM_ERRORS_ALLOWED, Integer(maxErrors))
    loadOptions.put(ODIConstants.LOG_ERRORS, Boolean(logErrors))
    loadOptions.put(ODIConstants.ERROR_LOG_FILENAME, errFileName)
    loadOptions.put(ODIConstants.RULE_SEPARATOR, ruleSeparator)
    loadOptions.put(ODIConstants.ERR_COL_DELIMITER, errColDelimiter)
    loadOptions.put(ODIConstants.ERR_ROW_DELIMITER, errRowDelimiter)
    loadOptions.put(ODIConstants.ERR_TEXT_DELIMITER, errTextDelimiter)
    loadOptions.put(ODIConstants.ERR_LOG_HEADER_ROW, Boolean(logHeader))
    loadOptions.put(ODIConstants.COMMIT_INTERVAL, Integer(commitInterval))
    loadOptions.put(ODIConstants.RUN_CALC_SCRIPT_ONLY,Boolean(calcOnly))
    loadOptions.put(ODIConstants.PRE_LOAD_MAXL_SCRIPT,preMaxlScript)
    loadOptions.put(ODIConstants.POST_LOAD_MAXL_SCRIPT,postMaxlScript)
    loadOptions.put(ODIConstants.ABORT_ON_PRE_MAXL_ERROR,Boolean(abortOnPreMaxlError))
    #call begin load
    pWriter.beginLoad(loadOptions)
    Execution step from opartor:
    Read rows: 0
    Insert/Delete/updat rows: 0
    ODI Step: Load Data Into Essbase
    from com.hyperion.odi.common import ODIConstants
    from com.hyperion.odi.connection import HypAppConnectionFactory
    from java.lang import Class
    from java.lang import Boolean
    from java.sql import *
    from java.util import HashMap
    # Get the select statement on the staging area:
    sql= """select C1_ACCOUNT "Account",C3_TIMEPERIOD "TimePeriod",C4_LOBS "LOBs",C5_TREATY "Treaty",C6_SCENARIO "Scenario",C7_VERSION "Version",C8_CURRENCY "Currency",C9_YEAR "Year",C10_DEPARTMENT "Department",C11_ENTITY "Entity",C2_DIVLOC "DivLoc",C12_DATA "Data" from OdiMapping.dbo.C$_0BUDGET01_PLAN1Data where      (1=1) """
    srcCx = odiRef.getJDBCConnection("SRC")
    stmt = srcCx.createStatement()
    srcFetchSize=30
    stmt.setFetchSize(srcFetchSize)
    rs = stmt.executeQuery(sql)
    #load the data
    stats = pWriter.loadData(rs)
    #close the database result set, connection
    rs.close()
    stmt.close()
    ODI Step: Report Statistics
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 2, in ?
    Essbase Writer Load Summary:
         Number of rows successfully processed: 1
         Number of rows rejected: 0
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.h.y(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    Where am i doing wronge for data load in Essbase? How I will find correct information to load data into Essbase?

  • ODS data Load from multiple sources.... Help

    Hello every one,
    I have a situation:
    I have a ODS where out of 150 fields 100 are coming from flat files, 30 from other ODSs and 20 are from Master Data Tables (some are attributes of Master Data InfoObjects).
    I am thinking i can do the following. Please advise if this is fine or if there is any other way of doing it for better performance e.t.c
    1 Load the Flat File Fields based on Flat File DataSource/InfoSource/UpdateRules.
    2. Load the other ODS fields/data in the Start Routine of the above Update Rules ???
    3. Load the Master Data objects through Master Data InfoSources
    Please confirm if i can do this. Also is there any order that i have to maintain for the same..... and finally
    When i declare the Masterdata Object (say 0MATERIAL) as one of my data object in the ODS and some the fields as the attributes of this 0MATERIAL, if some one is already loading data into the 0MATERIAL, should i load again or not, because i remember Master Data stores data in separate tables.. ... Please remove my confusion )
    All useful answers will be given full points.
    Thanks a lot ...

    Hi:
    <i>1 Load the Flat File Fields based on Flat File DataSource/InfoSource/UpdateRules.
    2. Load the other ODS fields/data in the Start Routine of the above Update Rules ???
    3. Load the Master Data objects through Master Data InfoSources</i>
    THe third one is only possible if those InfoObjects are InfoProviders. This is becasue if the I-Objects are direct update, you cannot use those infosources for any other laods but InfoObjects.
    <i>Please confirm if i can do this. Also is there any order that i have to maintain for the same.....</i>
    Order is needed only if th ekey figures are Overwrite and Business Requests the order.
    <i>and finally
    When i declare the Masterdata Object (say 0MATERIAL) as one of my data object in the ODS and some the fields as the attributes of this 0MATERIAL, if some one is already loading data into the 0MATERIAL, should i load again or not, because i remember Master Data stores data in separate tables.. ... Please remove my confusion )</i>
    Didn't understand this part clearly.
    DO you have the Attributes of 0material in ODS and populating thme in ODS?
    Ram C.

  • Hierarchy for Master Data Load from BW

    Hi All
    I'm loading Master data from BW infoobject 0Customer. There is no Hierarchy maintained for it in BW.
    Rather than creating hierarchy manually in BPC, I want to include mapping in transformation/conversion file which should roll master data records coming from BW to ALL_CUSTOMERS (member manually maintained in BPC). ALL_CUSTOMERS will be parent node for the hierarchy in BPC. No other levels of hierarchy is required. All the data should have one parent.
    Something like PARENT=*STR(ALL_CUSTOMERS)
    Please share your ideas? Help is very much appreciated.
    Thanks.

    Hi Asif
    If you are laoding master data from BW infoobject, you will not be able to load /specify any value for ParentH1/hierarchy node as BW /BPC does not allow that.
    An option would be to
    1) load hierarchy from a flat file or
    2) to maintain  hierarchy in BW amd load into BPC the normal way
    Regards
    Surabhi

  • Master data loading from an XML file

    Gurus,
    Is it possible to load Master data using XML files as source??

    Hi Niteen,
    how r u ?? hope u r doing fine, yes you can load it !!
    follow this link ..
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/21d8aa90-0201-0010-5e83-a3798b9a5ee0
    Or
    XML data
    regards ..
    Message was edited by:
            SAUGATA SEN

  • Article Master Data Extraction from IS-Retail System

    Hi,
    I have one IS-Retail System with me. I want to extract Article Master data from IS-Retail System.
    Is there any specific procedure to do this? If yes, can anybody tell me the procedure?
    Waiting for your response...
    Cheers,
    Yash K

    Hi Yash,
    you can use transaction BD10 to extract articles. Note that you should change to message type "ARTMAS" instead of "MATMAS".
    Best regards
    Michael

  • Reagarding Master Data loading from ECC

    Hi i want to activate Master Data  for Customer and Material from ECC ,Can any one  give me the clear steps ,and also  i want  to know what is the source for these Material And Customer in ECC.

    Hi
    The datasources are 0CUSTOMER_ATTR and 0MATERIAL_ATTR you can find these in RSA6 if activated under SAP-R/3->LO->LO-IO.
    Assign these to 0CUSTOMER and 0MATERIAL infoobjects in BW for a direct update or you can use the following infosources for flexible update
    0CUSTOMER_ATTR
    0MATERIAL_ATTR
    Also use texts datasources 0CUSTOMER_TEXT & 0MATERIAL_TEXT
    If the above mentioned datasources are not activated (NOT VISIBLE IN RSA6), use RSA5 transaction code to activate them.
    Hope you are familiar with BW and good luck
    Regards

  • Vendor master data load from flat file

    Hi Experts,
    I am trying to load data from flat file. I am confused which info object i should use. My requirement is for info cube 0BBP_C01.
    As i found for this info cube only two info objects are available: 0BBP_VENDOR and 0VENDOR. But in my flat file i have following fields: Vendor code, Vendor Name, Ultimate Parent, City, Country, Minority Status, Vendor Payment Terms. which all are together not present in either of info objects.
    Please suggest which info object i must use. or should i create Z info object for my requirement?
    Thanks in advance.
    Regards,
    Niranjan Chechani

    Hi Kiran,
    I am loading data through flat file. I am not getting your point like how to get attributes present in the R/3 system which I am not using in this case.
    Hi rvc,
    I have created two characteristics z info objects i.e. ZU_PARENT for Ultimate parent and ZVEN_PAY for Vendor Payment Terms.
    When i am trying to add these two as attributes and add it as Navigational attribute, I am getting the error "Characteristic 0VENDOR: The attibutes SID table(s) could not be filled" (Message no. R7586)
    Please suggest am i wrong somewhere?
    Thanks and Regards,
    Niranjan Chechani
    Edited by: Niranjan Chechani on Nov 28, 2011 12:15 PM

  • Master Data load from Info Object

    Hello,
    Iam tryting to load masterdata from info object using the standard package in 7.5 NW. My requirement is similar to the scenario 3b in the How to load masterdata doc(to prefix and strip the Compounding obj). This is the ssue i am currently facing.
    1. When i validate the transformation file , I get an error saying "Duplicate member id'12345'
    2. I have spaces in my node name. How am i supposed to give the technical name in the Member id field of the hiererchy tab in the data manager set selection options.
    3. text nodes not getting converted to member values.
    note:  I  am able to  get past this with out giving the seletion for the hierarchy in the DM package and running the package succesfully ,but i dont see the text nodes getting converted to member values.
    Please provide me some useful insights. had anyone had a similar issue??
    Thanks,
    Simi

    Simi,
    Did you checked YES to import the Text Nodes ... and even the Short Text in Language Tab under the SET SELCTION of the info object.

  • Automate Master data loads in BPC7(MS)

    I need to update dimension members in SAP BPC7 (MS) from an external source.
    The source system for my current project is BW, but if you have done any automated master data loads from any other system or even from a "temp" table please post your solution or tips.
    Thanks!
    Mez

    Hi Mezert,
    Would like to chime in with some of the things I've learned about this process over 3-4 years I've spent with BPC/Outlooksoft:
    -- Automating maintenance of the dimension/master data might be the biggest single weakness of the BPC solution on the MS platform.  It's one of the big reasons that the Netweaver platform basically replaces DTS/SSIS and has more direct linkage to BI/BW cubes and master data (InfoCubes).
    -- It is also one a very common issue and in fact, you will have to deal with what the old-timers in BI call "slowly-changing dimensions" at almost every client eventually.  For instance, you have a list of proucts/materials that every so often someone adds an item to.  ...Well, before you can enter anything for it in BPC, you have to add the dimension member, of course.
    -- So far, this challenge has been met by bringing in specialist DTS/SSIS experts in most cases who can employ the special "Custom Tasks" and build one of these packages to run in an automated way to populate the tables inside SQL Server and then execute the Admin MakeDim task to process the cube/application.
    -- Usually the development/testing/deployment of these solutions is a non-trivial undertaking.  It can be done, but there are peculiar challenges.
    -- One example of a challenge is what happens if you need a dimension with > 65,000 members?  Well, it can be created and populated by a DTS package, however, when you check the package and the custom task from the old Outlooksoft stuff out, you see that it CLEARS the entire dimension table before re-filling it.  ...Yep, it wouldn't allow an "Append" function to run. 
    -- A second example of a peculiar challenge is that unless you update the Excel Dimension Sheet, you will have to do all the updates in SQL Server Mgmt. Studio and really won't be able to touch the BPC Admin again for that dimension.
    Apologies for such a long post, but this is a fairly significant issue and I have to do a lot of research to piece together some of these things.  ...If anyone else has something to add as a best-practice, please post.

Maybe you are looking for