Loading transaction data using an attribute field in the BW cube

I am trying to load transaction data into BPC.  All but one of my dimensions are referencing technical fields.  One of my fields is referencing an attribute of a technical field.  The technical field is called 0MAT_PLANT and the attribute of that field is 0PROFIT_CTR.  In my mapping section of my transaction file I set the PROFIT_CTR dimension to 0MAT_PLANT__0PROFIT_CTR.  The transaction file validates and processes fine but when I try to import I get the error message: 0MAT_PLANT__0PROFIT_CTR is not a valid command or column 0MAT_PLANT__0PROFIT_CTR does not exist in source
Does anyone know how to complete this successfully?
THANK YOU!
Karen B. Thibodeau

Hi,
If is mapped fine and still getting same errors, maybe you can try to check if the IO attribute it's checked in the options in the data package.
Regards

Similar Messages

  • Deleting master data after loading transactional data using flat file

    Dear All,
    I have loaded transaction data  into an infocube using a flat file . While loading DTP i have checked the  option "load  transactional data with out master data exists" . So transactional data is loaded even if no master data is there in BW.
    While loading the flat file, I made a mistake for DIVISION Characteristic  where original master data value is '04000' , but i loaded the transactional data with value '4000' .Then i later realized after seeing the data from the infocube and deleted the request. then i reloaded data with value '04000'. Till now every thing is fine.
    But when I see the master data for DIVISION , i can see a new entry  with value '4000'.
    My question is how to delete the entry value '4000' from DIVISION. I tried deleting manually this entry from 'maintaining masterdata' , but it is not allowing me to do so .
    I have also checked if any transactional data exists for that value '4000' , as i said earlier I have deleted the transactional data with that values. even tried to delete the entries from the master data table, but i donot see a option to delete entries there.
    Please suggest me on this.
    Regards,
    Veera

    Hi,
    Goto RSA1 right click on the Info object and select Delete Master data. This will delete the master data unused existing in the table.
    If this master data is not used any where else just delete the master data completely with SID option.
    If even this doesnt work you can delete the complete table entire in SE14. But this will wipe out the entire table. Be sure if you wanna do this.
    Hope this helps
    Akhan.

  • Load transaction data from ECC to BPC 10.1 using US GAAP starter kit SP3

    I need to understand a bit more the process to load transactional data into BPC 10.1. We have US GAAP Starter Kit SP 3. Below is an screenshot from the config guide:
    It explains how transactional data can be extracted from ECC system to the SAP Netweaver BW data source 0FI_GL_10 and then transform it
    to get it loaded into BPC. The Objects /PKG/FC_C01 and PKG/FC_DS01 are just mentioned here as reference because they come in the RDS for Financial Close and Disclosure Management, which not all companies have.
    I believe the upwards data flow should be from Data Source 0FI_GL_10 to Data Store Object 0FIGL_O10 and then to InfoCube 0FIGL_R10. There is also a data flow that goes from 0FI_FL_10 to InfoCube 0FIGL_R10. Can anyone with experience with US GAAP starter kit answer this?
    Thank you.

    Hello, we were able to load actuals to our environment with the US GAAP Starter Kit, SP03. I followed the Operating Guide document up to section 5.2 and ran the Consolidation Data Manager Package with no issue. We are using the A20 Input  and A20 Consolidation process flows based on flows F00, F10, F20, F30, F99,  etc... According to the documentation, the Statemnet of Cash Flow and Changes  in Equity should be automatically calculated and available from the changes
    in Balance Sheet accounts once Consolidation is completed . However, when I  ran the corresponding reports, they bring no data.
    We loaded actual  data for the whole 2013 and Jan 2014. Our intention is
    to run the first  consolidation for Jan 2014. The closing balance for period 12.2013 in flow 99  was copied to flow 00 2014 (opening balance). Flows 20 and 30 were updated with the corresponding balance sheet movements (increase/decrease), according to the delivered controls in the starter kit. However, cash flow is still showing no results.
    I found the following text in the operating guide,  but I am not clear if I am missing a step. Can you please clarify? This is  telling me that I need to copy my 01.2014 opening balance (F00) to 12.2013  closing balance (F99, which is done by the copy opening data manager package, but in the opposite direction, from Y-1 F00 to Y F99) and in addition to also copy that balance to 12.2013 F00 (previous year  opening balance)??
    "5.2.2 First Consolidation
    When operating the  consolidation for a given Scope for the first time in the application, it is  necessary to populate and process the prior year-end time period for this Scope, in addition to the first required consolidation period. This is because the flows dedicated to the scope change analysis are properly populated by the consolidation engine provided that at least one automatic journal entry has been detected in the consolidation used as opening consolidation, which by default is the prior year end (December, Y-1). To accomplish this, it is recommended that you copy all input-level opening data of the first required consolidation (flow “F00”) into the closing data  (“F99”) and opening data (“F00”) of the prior year-end time period, including intercompany breakdowns.
    This breakdown by intercompany in the balance  sheet triggers automatic eliminations, which are carried over to the closing position. After the consolidation process, the closing data of the prior year
    end time period will contain the appropriate source data for the opening data of the following consolidation, notably the opening automatic journal entries.

  • Can we use LSMW to load Transaction data

    Hi.
    Can we use LSMW to load transactional data.Or it is used only for Master Data.
    With Best Regards
    Mamatha

    Hi,
    you can do upload of transactional data using,
    Standard Programs,
    Batch Input Recording,
    BAPI method,
    IDoc,
    best regards,
    Thangesh

  • Need more Info about "Load transactional data when master data not loaded"

    Hi,
    Can you please explain me this option in the infopackage " Load transactional data when master data is not loaded"....
    Say i load a transactional data record which has a material no. AAAXX.
    In the fact table, the material no. is replaced with the corresp. DIM ID.
    Now, assume that there is no entry for this Material no.
    AAAXX in the master data table...so no DIM ID for it..
    How is it then stored in the fact table ?
    Hope i have managed to explain the scenario..
    Thanks in advance,
    Punkuj

    Hello Punkuj K,
    How r u ?
    No, if the entry for that Material Number "AAAXX" is not there in the Master Data then it will create a SIDs & DIMs ID for that & Transaction Data will be loaded.
    Use
    Choose this indicator if you want to always update the data, even if no master data for the navigation attributes of the loaded records exists. The master data is generated from the loaded transaction data. The system draws SIDs. You can subsequently load the master data.
    Dependencies
    The texts, attributes, and hierarchies are not recognized by the system. Load the texts / attributes / hierarchies for the corresponding InfoObjects separately.
    This function corresponds to the update. Possible errors when reading the master data into the update rules are not associated with this indicator.
    Recommendation
    We recommended loading the master data into the productive systems in advance. Select this indicator especially in test systems.
    Best Regards....
    Sankar Kumar
    +91 98403 47141

  • Can i load transaction data thro flat files without loading masters?

    Hi all,
    is it possible to load history data using flat files in sales cube without having loaded any masters?
    thanks

    Hi
    You can load the transaction data without loading the master data. You have to check the option in update rules (update also if no master data exists).
    In reporting you will not be able to see the attribute and text data. Its always adivceable to load master data first and run apply hierarchy/attribute change and load the transaction data.
    AHP: Nice to see you after a long time
    REGards
    Rak

  • Error while loading Transactional data from NW BW Infoprovider

    Hi,
      I am trying to load the transactional data using the delivered "Transactional data from NW BW Infoprovider" package and getting error message Error occurs when loading transaction data from other cube .
    Below is the transformation file content
    *OPTIONS
    FORMAT = DELIMITED
    HEADER = YES
    DELIMITER = TAB
    AMOUNTDECIMALPOINT = .
    SKIP = 0
    SKIPIF =
    VALIDATERECORDS=NO
    CREDITPOSITIVE=YES
    MAXREJECTCOUNT= 9999999999999
    ROUNDAMOUNT=
    *MAPPING
    ACCOUNT = 0ACCOUNT
    BUSINESSTYPE = *NEWCOL(NOBTYPE)
    BWSRC = *NEWCOL(R3)
    COMPANYCODE = 0COMP_CODE
    DATASRC = *NEWCOL(R3)
    FUNCTIONALAREA = *IF(0FUNC_AREA = *STR() then *STR(NOFA);0FUNC_AREA)
    GAME = *NEWCOL(NOGAME)
    INPUTCURRENCY = *IF(0CURRENCY = *STR() then *STR(NOCURR) ; 0CURRENCY)
    PROBABILITY = *NEWCOL(NOPROB)
    PRODUCT = *NEWCOL(NOPROD)
    PROFITCENTER = 0PROFIT_CTR
    PROJECT = *NEWCOL(NOPROJECT)
    TIME = 0FISCPER
    VERSION = *NEWCOL(REV0)
    WEEKS = *NEWCOL(NOWEEK)
    SIGNEDDATA= 0AMOUNT
    *CONVERSION
    Below is the Error Log
    /CPMB/MODIFY completed in 0 seconds
    /CPMB/INFOPROVIDER_CONVERT completed in 0 seconds
    /CPMB/CLEAR completed in 0 seconds
    [Selection]
    InforProvide=ZPCA_C01
    TRANSFORMATION= DATAMANAGER\TRANSFORMATIONFILES\R3_TD_P&L.xls
    CLEARDATA= Yes
    RUNLOGIC= No
    CHECKLCK= Yes
    [Messages]
    Task name CONVERT:
    No 1 Round:
    Error occurs when loading transaction data from other cube
    Application: ACTUALREPORTING Package status: ERROR
    Does anyone had this issue before or is there something that needs to be looked in the BW delivered task ?
    ANy Inpout will be appreciated.
    Sanjay

    Hello Guru,
    Currently i am getting below error while loading costcenter master data from BW to BPC.
    Task name MASTER DATA SOURCE:
    Record count: 189
    Task name TEXT SOURCE:
    Record count: 189
    Task name CONVERT:
    No 1 Round:
    Info provider  is not available
    Application: ZRB_SALES_CMB Package status: ERROR
    Anybody can tell me, if i have missed anything ???
    Regards,
    BI NEW
    Edited by: BI  NEW on Feb 23, 2011 12:25 PM

  • Loading transaction data from flat file to SNP order series objects

    Hi,
    I am an BW developer and i need to provide data to my SNP team.
    Can you please let me know more about <b>loading transaction data (sales order, purchase order, etc.,) from external systems into order based SNP objects/structure</b>. there is a 3rd party tool called webconnect that gets data from external systems and can give data in flat file or in database tables what ever required format we want.
    I know we can use BAPI's, but dont know how. can you please send any <b>sample ABAP program code that calls BAPI to read flat file and write to SNP order series objects</b>.
    Please let me know ASAP, how to get data from flat file into SNP order based objects, with options and I will be very grateful.
    thanks in advance
    Rahul

    Hi,
    Please go through the following links:
    https://forums.sdn.sap.com/click.jspa?searchID=6094493&messageID=4180456
    https://forums.sdn.sap.com/click.jspa?searchID=6094493&messageID=4040057
    https://forums.sdn.sap.com/click.jspa?searchID=6094493&messageID=3832922
    https://forums.sdn.sap.com/click.jspa?searchID=6094493&messageID=4067999
    Hope this helps...
    Regards,
    Habeeb
    Assign points if helpful..:)

  • Problem in loading transactional data from to 0MKT_DSO1(ods) to 0MKTG_C01

    Hi,
    I am trying to load lead transaction data to the standard Crm lead management cube from ODS.There is a problem while loading transaction data from 0MKT_DSO1(ods) to the infocube 0MKTG_C01 as the field 0STATECSYS2(CRM STATUS)  is set to 10 in ods -meaning incorrect transaction. This feild is not there in the infocube.
    There is a routine in the cube that deletes data records with (0statecsys2) set to 10.
    THIS field is not coming in the transaction.
    so, where can i see the master data in crm source system? and why is that feild getting set to 10 ?
    thanks in advance!

    Thanks for the reply..
    I have checked the Fact table which shows
    1. packet Dimension
    2. Time dimension
    3. Unit dimension.
    I have kept the 0CALDAY as the time characteristics.
    Sample data i have loaded from ODS to Cube.
    Sample data in ODS.
    Sales order No_____0CALDAY_____AMOUNT
    800001___________12/02/2009____15
    I have loaded this data in Cube with Full Upload.
    Data in Cube.
    Sales order No_____0CALDAY_____AMOUNT
    800001___________12/02/2009____15
    Again i am loading the same data to cube
    Data in cube after loading.
    Sales order No_____0CALDAY_____AMOUNT
    800001___________12/02/2009____15
    800001___________12/02/2009____15
    The data is duplicated and it is not cumulating.
    Am i missing anything on this.
    Pls help..
    Thanks,
    Siva.

  • Error Loading Transactional Data into Cube(0PCA_C01)

    Hi Guys,
        I am trying to Install the following cubes from Business Content. 0PCA_C01 / 0PCA_C02 (Profit Center Analysis). Everything got replicated. I am trying to load transaction data now. I created Infopackage and loaded the data. Its running for a long time. It still says " not yet completed/warning ". If i try to see the content of the cube, I am getting the following errors.
    "Your user master record is not sufficiently maintained for object Authorization Object3.
    System error: RSDRC / FORM AUTHORITY_CHECK USER NOT AUTHORIZED 0PCA_C01 0PCA_C01
    System error: RSDRC / FUNC RSDRC_BASIC_CUBE_DATA_GET ERROR IN RSDRC_BASIC_QUERY_DATA_GET 0PCA_C01 64
    System error: RSDRC / FORM DATA_GET ERROR IN RSDRC_BASIC_CUBE_DATA_GET 0PCA_C01 64"
    Also if i try to change something in the infopackage, it says "Init. select. for field name  currently running in". I guess its because the job is still running.
    Please let me know if i missed something.
    Raj

    Hi Raj
    This seems to be an authorization case.
    i guess you are in BW Dev system.
    Go to SU01. Enter your user id and display.Go to Tabs ROle and PRofile.
    Ideally in development, a developers id shud have access to all devevlopment activities. So check with basis folks and get access to the relevant profiles. If you can get SAP_ALL, then perfect!!
    Prakash
    Assignin points is a way of saying thanks on SDN!!

  • Can i load transactional Data from ge DS without having load M Data before

    Hello,
    I want to load transactional data from a generic data source.
    I just want to know whether it would be successful, after the master data of 0material i load in the PSA couldn't be loaded in the infoObject (000000XXXXXXX -> material number error) by the DTP (see Thread BI - BI general - User arnaud).
    If i can load the transactional data without having loaded the master data in the info Object, is it possible that it loads also the master data simultaneous.
    Thank you very much for helping
    AR.

    Loading transactions before master data can result in data elements not being populated in the InfoProvider in some cases, hence the "Best Practice" recommendation to load master data first.
    For example, if you have update/transformation rules that go read master data to get additional data to enhance the transactional data, or you map master data attributes  during the transaction load, the master data attributes could be missing values if you don't load master data first.

  • Load Transactional Data Current Year: 0PM_OM_OPA_1

    Hi Experts,
    We are doing load monitoring on dailiy basis, As we are doing some testing in the BW Development system, we are loading data into the cube 0PM_C01. we want to load the current year data (2007-2008). but whenever we are trying to load through the InfoPackage "Load Transactional Data Current Year: 0PM_OM_OPA_1" it doesn't delete previous load requests. It was initially configured with this option not selected. we recently changed it, but it still doesn't delete. The only difference we see between this load and others that do delete is that this one is executed from a process chain rather than an InfoPackage Group. What do we need to do to make it delete previous load requests?
    And moreover the data is updated in the data target with two infosources namely 0PM_OM_PA_1 and 0PM_OM_PA_2.
    Because you eliminated use of the four conditions, it deleted the delta initialization and delta update requests loaded from InfoSource 0PM_OM_OPA_2.  We don't want it to do that!!!  I'm not sure whether we had any test data in the cube that had been loaded from the flat file, but I imagine it would delete those as well.  I believe we need all four conditions checked,
    So please help me in this regard. points will be assigned.
    Regards
    sekhar

    Hi experts,
    Please help in this regard.
    we are loading the data from two insources to one data target. two insources are :
    0PM_OM_OPA_1 and 0PM_OM_OPA_2 , as they are loading the  current year data.
    we have scenario like this please give your valauble inputs.
    I am still perplexed by the "Overlapping" option.  Why doesn't the "Same or more comprehensive" option work?  The requests we want it to replace have identical Data Selection because they were loaded by the same InfoPackage.  The InfoPackage "Load Transactional Data Current Year: 0PU_IS_PS_2" (ZPAK_3RUTSXU2BWPBKELMBE0T6NS3U) uses the "Same or more comprehensive" option and it does delete previous requests.  What is different about 0PM_OM_OPA_1?
    points will be assigned for the speedy reply.
    thanks
    sekhar

  • BPC 7.5 NW: Loading transaction data from infocube

    Hello, I am trying to load transaction data from an infocube into a BPC application using a package based on /CPMB/LOAD_INFOPROVIDER.
    The master data (cost center and cost element) are already loaded. As they are compounded, we have added in the key the CO area (eg ID=0CO_AREA+ID)
    I am facing an issue that the system ads leading 0 in front of the values of those master data comming from the cube.
    So the data looks like CO_AREA00000000000cost_center.
    As the data where correctly loaded in the dimensions (CO_AREAcost_center), the loading fails.
    How can we remove those 'O'?
    Thanks!

    Hi Anton,
    We use this in most of our projects; since auditing is one of the important requirements.
    You should enable the auditing for selective categories only. With each transaction data entered, it will keep a log, which can increase the size a lot, and hence impacting the performance.
    If you are enabling the log, you should frequently purge the audit data as well. However, you should take a backup of this and then purge it.
    Hope this helps.

  • "Error occurs when loading transaction data from other model" - BW loading into BPC

    Hi Experts,
    I'm having a problem with my data loading from BW, using the standard Load InfoProvider Selections data manager package.
    If I run for a period without data it succeeds (with warning) but if there is data to be extracted I get the following error:
    Task name CONVERT:
    No 1 Round:
    Error occurs when loading transaction data from other model
    model: AIACONS. Package status: ERROR
    As it runs ok when there isn't data it appears there is something preventing the movements of data out of the cube itself, rather then a validation issue.
    Has anyone encountered similar or have any ideas as to the problem?
    Best,
    Chris

    Hi Vadim,
    It's not specific to the transformation file as I have tried with others for the same BW cube and get the same result.
    We get a warning when we try and validate the transformation file:
    "Error occurs when loading transaction data from other model".
    This only appears in the validation pop up and doesn't throw up any warnings about the transformation file itself.  The validation log says:
    Validate  and Process Transformation File Log
    Log creation time
    3/7/2014 16:09
    The result of validation of the
      conversion file
    SUCCESS
    The result of validation of the
      conversion file with the data file
    FAIL
    Validation Result
    Validation Option
    ValidateRecords = NO
    Message
    Error occurs when loading transaction data from other model
    Reject List
    I can't find any errors anywhere else.
    Best,
    Chris

  • How to load transaction data

    Hi,
    Can anyone share how to load transaction data from BI Cube to BPC Cube. How do i map key figures in BI to BPC in NW version. Pls share any document that explains.
    Thanks
    Veeru

    For example, I've a BI Actual Cube with two key figures quantity and amount. I would like to copy this data to a BPC cube. Since BPC cube has only signed date (one key figure), how do we transform the data from BI cube to BPC. Can we able to achive this using transformation and  conversion file in NW. Is SAP providing standard packages that convert KF's in BI cube to BPC dimension members.
    Whats the approach you are doing for the scenarios where you are converting multiple key figures into BPC dimensions. I did not find a good document that explains this process.
    Thanks for sharing any information.
    Regards
    Veeru

Maybe you are looking for

  • Downlading Material stocks and its Value, storage location wise

    Dear Experts, As we are reimplementing SAP, we need to download the a)     stocks Plant and storage location wise for all the materials. Bcoz stocks are maintained in different storage locations. b)     Value of stocks for each material in the Plant

  • Object reference NULL exception is occuring at the time of iterating the projectCustomFieldRows

    Hi , I'm gettign object reference NULL exception in the following for each statement.  Please help me to resolve this issue. foreach (SvcProject.ProjectDataSet.ProjectCustomFieldsRow row in prjDataSetValues.ProjectCustomFields.Rows)                  

  • Time to time its giving error in the FTP connection

    Hi Gurus, Im doing File to File Interface. Im using FTP for File transfer and the message protocol is 'File'.  Im using the connection security as 'FTPS(FTP Using SSL/TLS) for control and data connection' and set the default command order.     I have

  • Regarding connection reset exception in socket

    hi, I am using java socket on one side and c socket on other side iam trying to send data and recieve from it but i got error socketexception connection reset by the peer, Environment which i am using to open a socket is websphere studio test environ

  • ADR3 on MacBook cannot see clients' user name

    I have just installed ARD 3.2 admin on a MacBook running 10.4.10, and it cannot access two servers on our network, one running 10.3.9 and the other with 10.4.9. Both have the 3.2.1 client installed. The admin console sees the clients but shows that t