Loading data with attribute dimension.

I have a currency attribute dimension asociated to Entities dimension and I like to upload data using the associated attribute dimension. I have a file with a column with the entity name and also a column with the attribute name. Could somebody tell me how to define the attribute column in the loading rule?
Edited by: user5170363 on Mar 22, 2012 12:55 PM

I guess, Attribute dimensions are not count while you load the data ..Data file always have the standard dimension members...
So you can filter your entities according to your attribute and load only on those entities...
Vivek Jaiswal

Similar Messages

  • How to load data to attribute dimension

    hi gurus,
    I have two questions:
    1. could I enter data into attribute dimension using SmartView?
    2. how could i load data into attribute dimension using load rules?
    Regards,
    Daniel

    Hi,
    1.
    You can't send data back into Essbase if your POV has an attribute specified. The attribute will calculate (by default it does a sum) your POV dynamically and only bring back data which has that attribute assigned to it.
    If you want to send data back into Essbase then the easiest way would be to keep the same POV and just remove the reference to the attribute. Smartview will then let you send data back into the Essbase cube.
    2.
    Same answer as point 1. Attributes are ways of adding more detail to members that already exist. You dont' load data into the attribute, you load data into the member that the attribute is assigned to.
    Hope this helps,
    Gee

  • Sql@loader-704  and ORA-12154: error messages when trying to load data with SQL Loader

    I have a data base with two tables that is used by Apex 4.2. One table has 800,000 records . The other has 7 million records
    The client recently upgraded from Apex 3.2 to Apex 4.2 . We exported/imported the data to the new location with no problems
    The source of the data is an old mainframe system; I needed to make changes to the source data and then load the tables.
    The first time I loaded the data i did it from a command line with SQL loader
    Now when I try to load the data I get this message:
    sql@loader-704 Internal error: ulconnect OCISERVERATTACH
    ORA-12154: tns:could not resolve the connect identifier specified
    I've searched for postings on these error message and they all seem to say that SQL Ldr can't find my TNSNAMES file.
    I am able to  connect and load data with SQL Developer; so SQL developer is able to find the TNSNAMES file
    However SQL Developer will not let me load a file this big
    I have also tried to load the file within Apex  (SQL Workshop/ Utilities) but again, the file is too big.
    So it seems like SQL Loader is the only option
    I did find one post online that said to set an environment variable with the path to the TNSNAMES file, but that didn't work..
    Not sure what else to try or where to look
    thanks

    Hi,
    You must have more than one tnsnames file or multiple installations of oracle. What i suggest you do (as I'm sure will be mentioned in ed's link that you were already pointed at) is the following (* i assume you are on windows?)
    open a command prompt
    set TNS_ADMIN=PATH_TO_DIRECTOT_THAT_CONTAINS_CORRECT_TNSNAMES_FILE (i.e. something like set TNS_ADMIN=c:\oracle\network\admin)
    This will tell oracle use the config files you find here and no others
    then try sqlldr user/pass@db (in the same dos window)
    see if that connects and let us know.
    Cheers,
    Harry
    http://dbaharrison.blogspot.com

  • SQL * Loader : Load data with format MM/DD/YYYY HH:MI:SS PM

    Please advice how to load data with format MM/DD/YYYY HH:MI:SS PM into an Oracle Table using SQL * Loader.
    - What format should I give in the control file?
    - What would be the column type to create the table to load data.
    Sample data below;
    MM/DD/YYYY HH:MI:SS PM
    12/9/2012 2:40:20 PM
    11/29/2011 11:23:12 AM
    Thanks in advance
    Avinash

    Hello Srini,
    I had tried with the creation date as DATE datatype but i had got an error as
    ORA-01830: date format picture ends before converting entire input stringI am running the SQL*LOADER from Oracle R12 EBS front-end.
    the contents of my control file is
    LOAD DATA
    INFILE "$_FileName"
    REPLACE
    INTO TABLE po_recp_int_lines_stg
    WHEN (01) = 'L'
    FIELDS TERMINATED BY "|"
    OPTIONALLY ENCLOSED BY '"'
    TRAILING NULLCOLS
    INDICATOR                POSITION(1) CHAR,
    TRANSACTION_MODE          "TRIM(:TRANSACTION_MODE)",
    RECEIPT_NUMBER               "TRIM(:RECEIPT_NUMBER)",
    INTERFACE_SOURCE          "TRIM(:INTERFACE_SOURCE)",
    RECEIPT_DATE               "TO_CHAR(TO_DATE(:RECEIPT_DATE,'MM/DD/YYYY'),'DD-MON-YYYY')",
    QUANTITY               "TRIM(:QUANTITY)",
    PO_NUMBER               "TRIM(:PO_NUMBER)",
    PO_LINE_NUMBER               "TRIM(:PO_LINE_NUMBER)",
    CREATION_DATE               "TO_CHAR(TO_DATE(:CREATION_DATE,'MM/DD/YYYY HH:MI:SS AM'),'DD-MON-YYYY HH:MI:SS AM')",
    ERROR_MESSAGE                   "TRIM(:ERROR_MESSAGE)",
    PROCESS_FLAG                    CONSTANT 'N',
    CREATED_BY                      "fnd_global.user_id",
    LAST_UPDATE_DATE                SYSDATE,
    LAST_UPDATED_BY                 "fnd_global.user_id"
    {code}
    My data file goes like
    {code}
    H|CREATE|123|ABC|12/10/2012||||
    L|CREATE|123|ABC|12/10/2012|100|PO12345|1|12/9/2012  2:40:20 PM
    L|CORRECT|123|ABC|12/10/2012|150|PO12346|2|11/29/2011 11:23:12 AM{code}
    Below is the desc of the table
    {code}
    INDICATOR             VARCHAR2 (1 Byte)                         
    TRANSACTION_MODE        VARCHAR2 (10 Byte)                         
    RECEIPT_NUMBER             NUMBER                         
    INTERFACE_SOURCE        VARCHAR2 (20 Byte)                         
    RECEIPT_DATE             DATE                    
    QUANTITY             NUMBER                    
    PO_NUMBER             VARCHAR2 (15 Byte)                         
    PO_LINE_NUMBER             NUMBER                         
    CREATION_DATE             TIMESTAMP(0)                         
    ERROR_MESSAGE             VARCHAR2 (4000 Byte)                         
    PROCESS_FLAG             VARCHAR2 (5 Byte)                         
    CREATED_BY             NUMBER               
    LAST_UPDATE_DATE        DATE               
    LAST_UPDATED_BY             NUMBER     {code}
    Thanks,
    Avinash                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Is it possible to load data from attribute of I.O. to text of the same I.O.

    Hi guys,
    is it possible to load data from attribute of InfoObject to text of the same InfoObject.
    I have this requierement: the text (description long) of InfoObject should be filled defends on the content of one attribute of this InfoObject.
    Is it possible via Update Rule? (we are using BW3.5)
    And how?
    Thanks in adance!
    Regards

    HI,
    1st) activate your object as data target and do the assignment in the transfer or update rules.
    --- How can I active my InfoObject (Text if the InofoObject)as data target?
    Right Click your info area and choose " Insert characteristics as a Datatarget" and give your info object Id.
    2) create a generic datasource on the p-table of your object and assign that datasource as text datasource to your object.
    --- How can I create a generic datasource on the p-table of your object ?
    For Eg: your info object name is like this
    ZIO_CID
    For this P table will be like this
    /BIC/PZIO_CID
    In RSO2 tcode Generic datasources create datasource based on this table .
    Hope it helps
    Regards,
    Arun.M.D

  • C# load xml data with attributes to datagridview; update data and save back to the xml

    Hi guys,
    i currently have this XML as data
    <DEALS>
    <DEAL ID="22" ISBN="0-7888-1623-3" Screenplay="Mary Kaplan" Title ="Mr Gentleman" Director = "Jonathan Jones">
    <ALLOCATION DEMANDE="5000" CODE="72" PRICE="25.00">2500</ALLOCATION>
    <ALLOCATION DEMANDE="7000" CODE="75" PRICE="35.00">4000</ALLOCATION>
    </DEAL>
    <DEAL ID="23" ISBN="1-7988-1623-3" Screenplay="Joe Doe" Title ="Nothing Much" Director = "Listentome">
    <ALLOCATION DEMANDE="3300" CODE="72" PRICE="15.00">2500</ALLOCATION>
    </DEAL>
    </DEALS>
    I load the data with the code below:
    i use xDocument to load the DealData.xml and then put it in ToString.
    //outside of the form
    public static class XElementExtensions
    public static DataTable ToDataTable(this XElement element)
    DataSet ds = new DataSet();
    string rawXml = element.ToString();
    ds.ReadXml(new StringReader(rawXml));
    return ds.Tables[0];
    public static DataTable ToDataTable(this IEnumerable<XElement> elements)
    return ToDataTable(new XElement("Root", elements));
    //in the form
    private void button2_Click(object sender, EventArgs e)
    var xDocument = XDocument.Load("DealData.xml");
    string txtxml = xDocument.ToString();
    StringReader reader = new StringReader(txtxml);
    XDocument doc1 = XDocument.Load(reader);
    var res = doc1.Descendants("DEAL").ToList().Where(x => x.Attribute("ID").Value == "22").Descendants("ALLOCATION");
    dataGridView1.DataSource = res.ToDataTable();
    Now my question is:
    I would like to update the data from the DataGridview with the Attribute("ID").Value == "22" and save the data back to the XML.
    For example, after I load my data on the datagridview, we only pickup anything that is in the node
    <DEAL ID="22" ISBN="0-7888-1623-3" Screenplay="Mary Kaplan" Title ="Mr Gentleman" Director = "Jonathan Jones">
    <ALLOCATION DEMANDE="5000" CODE="72" PRICE="25.00">2500</ALLOCATION>
    <ALLOCATION DEMANDE="7000" CODE="75" PRICE="35.00">4000</ALLOCATION>
    </DEAL>
    I updated the datagridview below with DEMANDE = 9000 CODE = 66 PRICE = 24.77 AND ALLOCATION = 1200
    I want be able to extract the data back in the xml as a child of the DEAL ID = "22"
    So that the XML file looks like this 
    <DEALS>
    <DEAL ID="22" ISBN="0-7888-1623-3" Screenplay="Mary Kaplan" Title ="Mr Gentleman" Director = "Jonathan Jones">
    <ALLOCATION DEMANDE="5000" CODE="72" PRICE="25.00">2500</ALLOCATION>
    <ALLOCATION DEMANDE="7000" CODE="75" PRICE="35.00">4000</ALLOCATION>
    <ALLOCATION DEMANDE="9000" CODE="66" PRICE="24.77">1200</ALLOCATION> <-! this is the new line !->
    </DEAL>
    <DEAL ID="23" ISBN="1-7988-1623-3" Screenplay="Joe Doe" Title ="Nothing Much" Director = "Listentome">
    <ALLOCATION DEMANDE="3300" CODE="72" PRICE="15.00">2500</ALLOCATION>
    </DEAL>
    </DEALS>
    Is there a way to achieve that?
    I have been searching and reading in the books but i cannot find a solution for this.
    Thank you
    Please do not forget to click “Vote as Helpful” if the reply helps/directs you toward your solution and or "Mark as Answer" if it solves your question. This will help to contribute to the forum.

    I would think of something like the below, the id is passed as static you need to change this
    protected virtual void button1_Click(object sender, EventArgs e)
    //Get DataTable from DGV datasource
    DataTable dt = new DataTable();
    dt = (DataTable)dataGridView1.DataSource;
    string file1 = "<?xml version=\"1.0\" encoding=\"utf-8\" ?><DEALS><DEAL ID=\"22\" ISBN=\"0-7888-1623-3\" Screenplay=\"Mary Kaplan\" Title =\"Mr Gentleman\" Director = \"Jonathan Jones\"><ALLOCATION DEMANDE=\"5000\" CODE=\"72\" PRICE=\"25.00\">2500</ALLOCATION><ALLOCATION DEMANDE=\"7000\" CODE=\"75\" PRICE=\"35.00\">4000</ALLOCATION></DEAL><DEAL ID=\"23\" ISBN=\"1-7988-1623-3\" Screenplay=\"Joe Doe\" Title =\"Nothing Much\" Director = \"Listentome\"><ALLOCATION DEMANDE=\"3300\" CODE=\"72\" PRICE=\"15.00\">2500</ALLOCATION></DEAL></DEALS>";
    StringReader reader = new StringReader(file1);
    XDocument doc1 = XDocument.Load(reader);
    //Remove all elements related to the id being populated into the Grid
    doc1.Descendants("DEAL").ToList().Where(x => x.Attribute("ID").Value == "22").Descendants("ALLOCATION").Remove();
    //loop through datatable and create new xelement to be added to the xdocument
    foreach (DataRow dr in dt.Rows)
    XElement xe = new XElement("ALLOCATION",
    new XAttribute("DEMANDE", dr[0].ToString()),
    new XAttribute("CODE", dr[1].ToString()),
    new XAttribute("PRICE", dr[2].ToString()));
    xe.Value = dr[3].ToString();
    doc1.Descendants("DEAL").ToList().Where(x => x.Attribute("ID").Value == "22").FirstOrDefault().Add(xe);
    Fouad Roumieh

  • SmartView 11.1.2.2-drill down with attribute dimensions

    In Smartview 11.1.2.2.300, we have noticed that if we exclude attribute dimensions from our report, then we are able to drill down to Bottom level on all our reports; but if there are attribute dimensions listed, then it throws errors. Is it an option to drill down to bottom level, while keeping the attribute dimensions in the report? Thanks.
    Edited by: 937685 on Jan 17, 2013 9:35 AM

    According to Oracle that still relates to a timeout error: SmartView Error "XML Load Error: DTD is prohibited [ID 1059964.1]
    Update the registry on the client machine with the following settings:
    [HKEY_CURRENT_USER\Software\Microsoft\Windows\CurrentVersion\Internet Settings]
    "ReceiveTimeout"=dword:00075300
    "KeepAliveTimeout"=dword:000493e0
    "ServerInfoTimeout"=dword:000493e0
    You may need to reboot, I've seen it make a difference without the need.
    Is this an FDM connection?
    Anyway try that.
    Steve

  • 'Allow Duplicate Members' not compatible with Attribute  Dimensions

    When setting the 'Allow Duplicate Members' option on there is a problem if the duplicate member is used as an Attribute. When attempting to apply the member filter to use the attribute Essbase throws an error: Cannot uniquely identify a member, the name [L] is duplicated.
    In my case L is duplicated across two dimensions [L=LOCAL as a BOOK_CODE and L=LIABILITY as an ACCOUNT_TYPE].
    The ACCOUNT_TYPE is an attribute dimension related to the ACCOUNTS dimension. Both the excel addin and Financial Reports have a problem as their member selection interfaces don't fully qualify the selection.
    This seems like a bug.

    When you sync over the cable, it is one set of calendars in one section labeled "From My Mac"  when you sync with iCloud, that is store in a completely separate area labeled "iCloud."  The "From My Mac" calendars are stored on the phone as just "regular" calendars.  The iCloud account are "CalDav" calendars, which syncs with iCloud.  The old mobileme calendars did do what you wanted, sync over the cable and then to mobileme, and that lead to huge data corruption and why Apple made the new mobileme calendars which sync over CalDav only (and cannot talk to the cable) to protect the calendars.  The phone sees them as two separate calendars.
    Your best bet is if you want to use calendars on your G4 is to just use the iCloud.com web interface.  I'm also not familiar with the whole drop box thing, so I can't comment on that.

  • Want to load data with selection criteria

    Hi Everyone,
    I want to load data from X ODS to Y ODS , X ods is a datasource to Y ods.
    On Y ods i don't have any data loaded.
    But on X ods i am having 10 requests with 200000 records.
    In X ods i am having a request with 2 records.
    I want to load the 2 records request to Y ODS.to check data on Y ods
    Can anyone help me in solving these, b'cos i am new to BW, It's urgent Please.
    Can you tell me step by  step navigation.

    Hi,
    Just select Full upload; it will bring the InfoPackage and then in the Selection tab give the range value. If this is required only one time then this method is fine or full load is fine, otherwise you will have to write a code to pick the records. If you frequently want to load data from one ODS to another ODS then better go for init and then from next time onwards do the delta load. If you don't want to provide selection in the InfoPackage then the other way is to load all the data from X to Y and do selective deletion on Y ODS.
    Hope this helps.
    PB

  • How can I load data with Scripts on *FDM* to a HFM target System????

    Hi all!
    I need help because I can´t find a good guide about scripting on FDM. The problem I have is the next one.
    I have on mind to load my data with data load file in FDM to a HFM target system, but I would like to load an additional data using an event script, ie after validate. I would need any way to access to HFM system though FDM Scripts, is it possible??
    If so, It would be wonderful to get a data from HFM with any Point of View, reachable from FDM Scripts in order to load or getting any data.
    I´ve looking for a good guide about scripting in FDM but I couldn´t find any information about accessing data on HFM target system, does it really exist?
    Thanks for help

    Hi,
    Take a look at the LOAD Action scripts of your adapter. This might give you an idea.
    Theoretically it should be possible to load data in an additional load, but you need to be very careful. You don't want to corrupt any of the log and status information that is being stored during the load process. The audit trail is an important feature in many implementations. In this context it might not be a good idea to improve automation and risk compliance of your system.
    Regards,
    Matt

  • How to load data with carriage return through DRM action script ?

    Hello,
    We are using DRM to manage Essbase metadata. These metadata contain a field for member formula.
    Currently it's a string data type property in DRM so we can't use carriage return and our formula are really hard to read.
    But DRM support other data type property : memo or formatted memo where we can use carriage return.
    Then in the export file, we have change the record delimiter to an other character than CRLF
    Our issue : we are regularly using action script to load new metadata => How to load data properties with carriage return using action script ? There is no option to change the record delimiter.
    Thanks!

    Hello Sandeep
    here what I want to do through action script : loading a formula that use more on than one line:
    Here, I write my formula using 4 lines but action script cannot load since one line = 1 record.
    ChangeProp|Version_name|Hier_name|Node_name|Formula|@round(
    qty*price
    *m05*fy13

  • Create a script to update, deploy, load data with version 9.3.1

    Hi,
    I'd like to know if it's possible run a script to execute:
    - Import dimensions by profile (by EPMA)
    - deploy an application (by EPMA)
    - load rule to load data (by essbase)
    - Calculation script (by essbase)
    we are using the version 9.3.1
    Thanks in advanced.
    Roby

    Hi,
    I don't believe there any batch command functionality for EPMA until version 11.
    It is possible you can try and create a scheduled taskflow in EPMA to import a dimension and then reploy, finally you could get it run a batch script which contained a maxl script to load and calculate essbase data.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Performance problem with attribute dimensions - please help!

    Hi everybody,
    I run OBIEE reports on top of EssBase 9.3 ASO cube.
    My cube has 8 regular dimensions and 10 attribute dimensions.
    I am running a query based on:
    - 3 dimensions (TIME, BATCH, MODEL),
    - 2 attributes of BATCH ( SHIFT, SHIFT_STATUS )
    - and one measure.
    The query runs 3 minutes (!) and returns 190 rows.
    If I remove the attributes, the query runs 4 seconds for the same number of result rows.
    My MDX:
    With
    set [TIME6] as 'Filter([TIME].Generations(6).members,
    ([TIME].currentmember.MEMBER_ALIAS = "2009/01" OR [TIME].currentmember.MEMBER_Name = "2009/01"))'
    set [TIME7] as 'Generate([TIME6], Descendants([TIME].currentmember, [TIME].Generations(7), leaves))'
    set [BATCHES] as '[DIM_BATCH_HEADERS].Generations(5).members'
    set [SHIFT2] as '[SHIFT].Generations(2).members'
    set [SHIFT_STATUS2] as '[SHIFT_STATUS].Generations(2).members'
    set [MODEL2] as 'Filter([MODEL].Generations(2).members, ([MODEL].currentmember.MEMBER_ALIAS = "SNF_PCK" OR [MODEL].currentmember.MEMBER_Name = "SNF_PCK"))'
    select
    {[Accounts].[PROD_ACTUAL_QTY_KG]
    } on columns,
    NON EMPTY
    {crossjoin ({[TIME7]},
    crossjoin ({[BATCHES]},
    crossjoin ({[SHIFT2]},
    crossjoin ({[SHIFT_STATUS2]},{[MODEL2]}))))}
    properties ANCESTOR_NAMES, GEN_NUMBER on rows
    from [COP_MAAD.Cop_Maad]
    Thank you in advance,
    Alex
    Edited by: AM_1 on Feb 26, 2009 6:35 AM
    Edited by: AM_1 on Feb 26, 2009 7:03 AM

    Hi Glenn,
    I though that the reason might be the crossjoin with a very large set. The main dimension in this query has ~40000 members.
    Is there in EssBase MDX some analog for NONEMPTYCROSSJOIN() function that exists in MSAS?
    As these two attributes belong to the dimension that is used in the query, for every member of the dimension, there is one and only one member of each attribute. Therefore I though that Essbase should "know" how to behave.
    Thanks for an idea. I will try to increase the space for aggregation - hope this will help...
    Best Regards,
    Alex

  • Load data with SQL Loader link field between CSV file and Control File

    Hi all,
    in a SQL Loader control file, how do you specify link with field in CSV file and Control file?
    E.g. if I wat to import the record in table TEST (col1, col2, col3) with data in csv file BUT in different position. How to do this?
    FILE CSV (with variable position):
    test1;prova;pippo;Ferrari;
    xx;yy;hello;by;
    In the table TEST i want that col1 = 'prova' (xx),
    col2 = 'Ferrari' (yy)
    col3 = default N
    the others data in CSV file are ignored.
    so:
    load data
    infile 'TEST.CSV'
    into table TEST
    fields terminated by ';'
    col1 ?????,
    col2 ?????,
    col3 CONSTANT "N"
    Thanks,
    Attilio

    With '?' mark i mean " How i can link this COL1 with column in csv file ? "
    Attilio

  • Load data with Delta

    Hello Expert,
    I loaded data from SAP R/3 into PSA using Delta mode and found that a new record which was  just created was not loaded into SAP BI. What could be possible?
    Step of loading
    1. Initial without data into PSA
    2. Full load with criteria Fiscal Period (e.g. 001.2009 - 016.2009) into PSA
    3. Load data from PSA into DSO with update mode = Delta
    4. Create a new transaction from SAP R/3
    5. Load data from SAP R/3 into SAP BI-PSA with update mode = Delta
    Expected Result: A new record should be loaded into PSA.
    Actual Result: There was no record loaded.
    After initial loading without data, is it necessary to full load with all fiscal period. And then load with Delta mode?
    I have never seen this kind of problem before.
    Is anyone familiar with this kind of problem? How did you resolve it?
    Any suggestion would be appreciated.
    Thank you very much
    W.J.

    Hi,
    Is your Datasource is Logistics? if so how the job has been scheduled in LO cockpit (hourly / daily)
    Did you check the record in Delta Queue(RSA7) ?? 
    After initial loading without data, is it necessary to full load with all fiscal period. And then load with Delta mode?
    I have never seen this kind of problem before.
    If you have specific selections you have to do it by repair full load .. and it doesn't impact delta loads

Maybe you are looking for