Dtw gives special character while importing the data

Hi all
I m trying to update business partner and my template file is tab delimited but I don't know why my all business partner updated with special character " like this "" even I tried to update from CSV also ??
Can anybody help me to remove or update this special character from business partner name as well as bill to address , block, building, city  by query or any other way ??
Thanks
Ricky

Hi ricky 1,
Before you upload the CSV/TXT file, open it up in a Notepad.
You will see the "hidden" quotation marks that don't appear in Excel.
Use the replace function to remove them, save it, then upload again.
Kind Regards,
Nick Lakasas

Similar Messages

  • Error while importing the data

    Hi All,
    In fdm i am trying to Import the data but i am getting the below error while importing text file.
    Error: An error occurred importing the file.
    Detail: Data access error.
    also the below is the log found ..
    2012-02-14-01:04:42
    User ID...........     admin
    Location..........     mine
    Source File.......     C:\Hyperion\products\FinancialDataQuality\Mineapp\Test2\Inbox\mine\load_demo2.txt
    Processing Codes:
    BLANK............. Line is blank or empty.
    ESD............... Excluded String Detected, SKIP Field value was found.
    NN................ Non-Numeric, Amount field contains non numeric characters.
    RFM............... Required Field Missing.
    TC................ Type Conversion, Amount field could be converted to a number.
    ZP................ Zero Suppress, Amount field contains a 0 value and zero suppress is ON.
    Create Output File Start: [2012-02-14-01:04:42]
    [TC] - [Amount=NN]     Accounts,Market,Year,Product,Scenario,Data
    [Blank] -      
    Excluded Record Count..............1
    Blank Record Count.................1
    Total Records Bypassed.............2
    Valid Records......................2
    Total Records Processed............4
    Begin SQL Server (BCP) Process (2): [2012-02-14-01:04:42]
    [RDMS Bulk Load Error Begin]
         Message:      (-2147217900) - Data access error.
         See Bulk Load File:      C:\Hyperion\products\FinancialDataQuality\Mineapp\Test2\Inbox\tWadmin56465844032.tmp
    [RDMS Bulk Load Error End]
    Please help me out i am really helpless and dont know how to solve this error.

    When you perform a bulk load, you basically tell SQL Server to execute a command local to the SQL server. What that means is that if you pass it a parameter that says load the file :
    C:\Hyperion\products\FinancialDataQuality\Mineapp\Test2\Inbox\mine\load_demo2.txt
    SQL Server is going to look on it's C drive for the file.
    Unless you have SQL Server running on the FDM server, this is going to fail obviously.
    The 'right' thing to do is setup your FDM app using a share name. (i.e. \\FDMSERVER\FDMAPPS\Mineapp instead of c:\Hyperion\products\FinancialDataQuality\Mineapp).
    If you are relatively handy with FDM, there is a config file (in XML) that has the path information for the apps and you just need to change it from the hardcoded C drive reference to the share you create.)
    Your other option is to NOT load the fiels as bulk. If you have a lot of large data files being loaded, this would be a bad idea. If you have relatively small load files every month and not a lot of users on the system, this may not impact performance. I would update the path information if it were me.
    Charles

  • NULL error while importing the data

    Hai everybody. In one of our branches there was some problem. To simulate the situation. i advised them to send the backupdata. While importing data from a dump file, i am getting the erro message
    "cannot insert NULL into (“APGB5168”. “D010014”. “LBRCODE”)’.
    Can i change the NULL constraint and import OR can i import first only the table strucuture and modify the constraint and can import the data afterwards.
    Pl can any one suggest me how to import the data from this dump file so that i can simulate the situation.
    Thanks - JayaDev(JD)

    Hi,
    It means you have NOT NULL column and you are trying to insert NULL data.
    Can i change the NULL constraint and import OR can i import first only the table strucuture and modify the constraint and can import the data afterwards.It all depends on you and your business needs. If you are sure that removing the constraint will have no harm on your application and data then you can, but think why it has NOT NULL constraint if data has NULL values. Either your table data is incorrect or your table design. Rectify the problem and take action accordingly.
    Cheers,
    Avinash

  • Getting error while importing the data using loadercli

    Hello,
    I want to copy the data using export/ import  via (loadercli).
    scenario:
    1) I have 2 server and i want to export data from old server and import in to new MAXDB server using loadercli.
    I have done this once and it went fine , but now I want to do the export import again so i will have te latest data on the new server.
    when i tried to do that it gives error that the table already exist . Can i use loadercli to import hte data again ?
    can any one help me in this ?
    Regards,
    Bhavesh

    > Do you want to add and/or update the data in the already existing tables or do you want to replace the content completely?
    >
    > so in that way :
    > bot the options are fine what ever take less time.
    Sorry mate, but YOU have to know what you want here.
    I gave you an easy to follow set of steps.
    As you don't seam to mind the outcome, just might just use them...
    > I wanted to know weathe i can use the  loadercli for thie export import or not? if yes then is there any new steps to do before i do the export import?
    We had this discussion before...
    >
    > For that the easiest option would be just to drop the tables of SAPR3 and run the import again.
    >
    > For ease of use you could also just do:
    > - logon as superdba
    > - drop user SAPR3
    > - create user SAPR3 password SOMEPW not exclusive dba
    >
    > After these steps you can easily pump the data into the database again.
    >
    > So here in th above given steps , i am creating a new SAPR3 user and why it is not exclusive dba ?
    >  i already have that user SAPR3 can i use the same.
    Yes, you do have the SAPR3 user.
    But you don't seem to like to read documentation or learn about how the tools work or anything like that.
    Therefore I gave you s simple way to reach your goal.
    Of course it's possible to reuse the user.
    But then you would have to deal with already existing tables, already existing data etc.
    You don't seem to be able to do that. So, the easy steps might be better suited for your needs.
    regards,
    Lars

  • Getting error while importing the data to scsm

    Hi,
    We are planing to iport data from another Ticketing tools .For this i have created a .CSV and .XML .While iporting i am getting below error
    And the Xml File is 
    ================================================================
    <CSVImportFormat>
      <Projection Type='System.workItem.Incident.ProjectionType">
        <Seed>
          <Class Type="System.WorkItem.incident">
            <Property ID="Id"/>
            <Property ID="Source"/>
            <Property ID="TierQueue"/>
            <Property ID="Department"/>
            <Property ID="Classification"/>
            <Property ID="Title"/>
            <Property ID="Urgency"/>
            <Property ID="Impact"/>
            <Property ID="Assignedto"/>
            <Property ID="Description"/>
            <Property ID="Site"/>
            <Property ID="Priority"/>
            <Property ID="Asset"/>
            <Property ID="ChangeIncidentStatus"/>
            <Property ID="CreatedDate"/>
            <Property ID="ActualEndDate"/>
            <Property ID="ClosedDate"/>
            <Property ID="ResolutionDescription"/>
            <Property ID="SharepointBug"/>
            <Property ID="ClientContact:"/>
            <Property ID="RequestorCategory"/>
            <Property ID="CaseReason"/>
            <Property ID="Prod"/>
            <Property ID="ParentAccount"/>
            <Property ID="CustomerSupportAccount"/>
            <Property ID="Hardware"/>
            <Property ID="Browser"/>
            <Property ID="InternetConnection"/>
            <Property ID="ProblemAffectsOthers"/>
            <Property ID="IssueonDifferentWorkstation"/>
            <Property ID="PatientSpecific"/>
            <Property ID="SalesforceCase"/>
            <Property ID="TFS"/>
            <Property ID="StepstoReproduce"/>
            <Property ID="RelatedPatients"/>
            <Property ID="AffectedDevice"/>
            <Property ID="ReasonforOnHold"/>
            <Property ID="CustomerSupportAgent"/>
            <Property ID="FirstResponseDate"/>
            <Property ID="LastModified (Internal)"/>
            <Property ID="RequiredBy"/>
         </Class>
       </Seed>
       <Component Alias="AffectedUser">
         <Seed>
           <Class Type="System.Domain.User">
             <Property ID="FirstName"/>
    <Property ID="Domain"/>
             <Property ID="UserName"/>
           </Class>
         </Seed>
       </Component>
      </Projection>
    </CSVImportFormat>
    =====================================================================
    Please let me know that error in the xml file.
    Regards, H@ri

    Thanks ,I have been gone through <Projection Type='System.WorkItem.Incident.ProjectionType">.
    Now i am getting many errors while importing data.I have customized fields in servicedesk and i have created same fields in scsm incident form.When i am imported
    data i am getting below error.
    Please let me know how can i import customized feilds data to SCSM without issues.
    Regards, H@ri

  • How to import the data to Data manager

    Hi All,
    I am new to MDM and need your help in handling the following issue
    In our setup we are sending the Customer master data from R/3->XI->MDM (R/3-> XI IDOCs,  XI->MDM XML files).
    with the help of extraction job in R/3 IDOCs r generated and XI has onverted these IDOCs to XML Files and created at MDM Server. Now my Question is how to import the Data from these XML files to Data Manager !!!!
    Note : Import Map settings are done already in the system.
    Can any one help me the steps requeired to get this data to Data manager .
    and am I need to do some activity in MDM to import the data to Data manager each and every time or is there any automatic way of doing this .
    Please guide me.
    Thanks in advance!

    Hi MDM User,
    Generally there are 2 cases while importing the data,
    1) Initial Load
    2) Delta Load
    Initial Load means importing the data into MDM at first time.This is usually done by manual importing
    without using import server.
    Delta Load means importing the data into MDM when we have changes to existing data or added new data
    to existing data.This is usually done by automatically using import server.
    To manually import the data,just create a remote system and port in console and login into import manager
    using remote system.Create ,save the map and execute the map.There is no other configurations needed.
    To automatically import the data without manual intervention,we go for import server.
    To let this happen we need to set some configuarations in mdis.ini file of import server.
    Give the repository name,username and password which are set in console for that repository and also
    give schedule time based on which import server picks the file from ready folder.
    Go to the import manager , change the map and import action depends on your requriment and save that
    map.
    and also you need to set port type as inbound and automatic and select the map which you saved earlier in
    MDM console.
    As soon as we start the import server, it picks the file from ready folder of MDM server based on
    configurations which we set in mdis.ini file and executes the map.
    Here i am giving complete path of ready folder.
    SAP MDM 5.5> Server>Distributions>bp1bocap080.bp.co_MSQL> BP3_PoC_Customer> Inbound>Siebel> SIEBEL_OB_CUS_SIEBELCUS01> Ready)
    This path has to be given in target directory of receiver communication channel of XI so that it puts XML
    file in the above path.
    For better understanding,go through below blogs.
    /people/balas.gorla/blog/2007/02/05/r3-xi-mdm-outbound-scenario
    /people/balas.gorla/blog/2006/09/27/mdm-xi-r3-integration
    Hope it helps
    Reward points,if found useful
    Thanks
    Narendra

  • Error While importing the transaction data

    All SAP BPC Gurus,
      I need your help in resolving this error,  I have encountered this below error while uploading (importing) the transaction data of (Non-Reporting) application,  Would you please help me resolving this error.  I don't know if i'm doing anything wrong or is anything wrong with the setup.
    I used DataManager in EXCEL BPC and Ran the IMPORT Transaction Package.
    /CPMB/MODIFY completed in 0 seconds
    /CPMB/CONVERT completed in 0 seconds
    /CPMB/CLEAR completed in 0 seconds
    [Selection]
    FILE= DATAMANAGER\DATAFILES\IFP_BPCUSER121\rate data file.csv
    TRANSFORMATION= DATAMANAGER\TRANSFORMATIONFILES\SYSTEM FILES\IMPORT.XLS
    CLEARDATA= No
    RUNLOGIC= No
    CHECKLCK= No
    [Messages]
    Task name CONVERT:
    XML file (...ANAGER\TRANSFORMATIONFILES\SYSTEM FILES\IMPORT.TDM) is empty or is not found
    Cannot find document/directory
    Application: PLANNING Package status: ERROR

    are you using the standard "Import" data package?
    Check the code in the Advanced tab of the Import data package
    Check your Transformation file is in correct format.
    code in Advaced tab should be as  below:
    PROMPT(INFILES,,"Import file:",)
    PROMPT(TRANSFORMATION,%TRANSFORMATION%,"Transformation file:",,,Import.xls)
    PROMPT(RADIOBUTTON,%CLEARDATA%,"Select the method for importing the data from the source file to the destination database",0,{"Merge data values (Imports all records, leaving all remaining records in the destination intact)","Replace && clear datavalues (Clears the data values for any existing records that mirror each entity/category/time combination defined in the source, then imports the source records)"},{"0","1"})
    PROMPT(RADIOBUTTON,%RUNLOGIC%,"Select whether to run default logic for stored values after importing",1,{"Yes","No"},{"1","0"})
    PROMPT(RADIOBUTTON,%CHECKLCK%,"Select whether to check work status settings when importing data.",1,{"Yes, check for work status settings before importing","No, do not check work status settings"},{"1","0"})
    INFO(%TEMPNO1%,%INCREASENO%)
    INFO(%ACTNO%,%INCREASENO%)
    TASK(/CPMB/CONVERT,OUTPUTNO,%TEMPNO1%)
    TASK(/CPMB/CONVERT,ACT_FILE_NO,%ACTNO%)
    TASK(/CPMB/CONVERT,TRANSFORMATIONFILEPATH,%TRANSFORMATION%)
    TASK(/CPMB/CONVERT,SUSER,%USER%)
    TASK(/CPMB/CONVERT,SAPPSET,%APPSET%)
    TASK(/CPMB/CONVERT,SAPP,%APP%)
    TASK(/CPMB/CONVERT,FILE,%FILE%)
    TASK(/CPMB/CONVERT,CLEARDATA,%CLEARDATA%)
    TASK(/CPMB/LOAD,INPUTNO,%TEMPNO1%)
    TASK(/CPMB/LOAD,ACT_FILE_NO,%ACTNO%)
    TASK(/CPMB/LOAD,RUNLOGIC,%RUNLOGIC%)
    TASK(/CPMB/LOAD,CHECKLCK,%CHECKLCK%)
    TASK(/CPMB/LOAD,CLEARDATA,%CLEARDATA%)

  • Error while loading the data from PSA to Data Target

    Hi to all,
         I'm spacing some error while loading the data to data target.
    Error :  Record 1 :Value 'Kuldeep Puri Milan Joshi ' (hex. '004B0075006C0064006500650070002000500075007200690
    Details:
    Requests (messages): Everything OK
    Extraction (messages): Everything OK
    Transfer (IDocs and TRFC): Errors occurred
          Request IDoc : Application document posted
          Info IDoc 2 : Application document posted
          Info IDoc 1 : Application document posted
          Info IDoc 4 : Application document posted
          Info IDoc 3 : Application document posted
          Data Package 1 : arrived in BW ; Processing : Data records for package 1 selected in PSA - 1 er
    Processing (data packet): Errors occurred
          Update PSA ( 2462  Records posted ) : No errors
          Transfer Rules ( 2462  -> 2462  Records ) : No errors
          Update rules ( 2462  -> 2462  Records ) : No errors
          Update ( 0 new / 0 changed ) : Errors occurred
          Processing end : Errors occurred
    I'm totally new to this issue. please help to solve this error.
    Regards,
    Saran

    Hi,
    I think you are facing an invalid character issue.
    This issue can be resolved by correcting the error records in PSA and updating it into the target. For that the first step should be to identify if all the records are there in PSA. You can find out this from checking the Details tab in RSMO, Job log , PSA > sorting records based on status,etc. Once its confirmed force the request to red and delete the particular request from the target cube. Then go to PSA and edit the incorrect records (correcting or blanking out the invalid entries for particular field InfoObject for the incorrect record) and save it. Once all the incorrect records are edited go to RSA1>PSA find the particular request and update to target manually (right click on PSA request > Start update immediately).
    I will add the step by step procedure to edit PSA data and update into target (request based).
    In your case the error message says Error : Record 1 :Value 'Kuldeep Puri Milan Joshi '. You just need to conver this to Capital letter in PSA and reload.
    Edit the field to KULDEEP PURI MILAN JOSHI in PSA and push it to target.
    Identifying incorrect records.
    System wont show all the incorrect records at the first time itself. You need to search the PSA table manually to find all the incorrect records.
    1. First see RSMO > Details > Expand upate rules / processing tabs and you will find some of the error records.
    2. Then you can go to PSA and filter using the status of records. Filter all the red requests. This may also wont show the entire incorrect records.
    3. Then you can go to PSA and filter using the incorrect records based on the particular field.
    4. If this also doesnt work out go to PSA and sort (not filter) the records based on the particular field with incorrect values and it will show all the records. Note down the record numbers and then edit them one by one.
    If you want to confirm find the PSA table and search manually."
    Also Run the report RS_ERRORLOG_EXAMPLE,By this report you can display all incorrected records of the data & you can also find whether the error occured in PSA or in TRANSFER RULES.
    Steps to resolve this
    1. Force the request to red in RSMO > Status tab.
    2. Delete the request from target.
    3. Come to RSMO > top right you can see PSA maintenace button > click and go to PSA .
    4.Edit the record
    5. Save PSA data.
    6. Got to RSA15 > Search by request name > Right click > update the request from PSA to target.
    Refer how to Modify PSA Data
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/40890eda-1b99-2a10-2d8b-a18b9108fc38
    This should solve your problem for now.
    As a long term you can apply some user exit in source system side or change your update rules to ensure that this field is getting blanked out before getting loaded in cube or add that particular char to permitted character list in BW.
    RSKC --> type ALL_CAPITAL --> F8 (Execute)
    OR
    Go to SE38 and execute the program RSKC_ALLOWED_CHAR_MAINTAIN and give ALL_CAPITAL or the char you want to add.
    Check the table RSALLOWEDCHAR. It should contain ALL_CAPITAL or the char you have entered.
    Refer
    /people/sap.user72/blog/2006/07/23/invalid-characters-in-sap-bw-3x-myths-and-reality-part-2
    /people/sap.user72/blog/2006/07/08/invalid-characters-in-sap-bw-3x-myths-and-reality-part-1
    /people/aaron.wang3/blog/2007/09/03/steps-of-including-one-special-characters-into-permitted-ones-in-bi
    http://help.sap.com/saphelp_nw04/helpdata/en/64/e90da7a60f11d2a97100a0c9449261/frameset.htm
    For adding Other characters
    OSS note #173241 – “Allowed characters in the BW System”
    Thanks,
    JituK
    Edited by: Jitu Krishna on Mar 22, 2008 1:52 PM

  • Error coming while activating the data request at DSO level

    Hi to all,
    please can anybody tell me the solution for that.
    While activation the data request at DSO level  , i am getting error .
    Value 'R2S-11.60UML  ' (hex. '5200320053002D00310031002E003600300055004D004C00A0') of char
    Error when assigning SID: Action VAL_SID_CONVERT table 0MATERIAL
    Process 000001 returned with errors
    this error is coming under SID Generation .
    I shall be thnakfull to you for this.
    Regards
    Pavneet Rana

    Hi Rana,
    Activation failed because of the special characters in the Omaterial
    Delete the request in target
    edit it in the specific record in PSA and load it back
    In your case i think its a space problem u can see a space after L
    delete the space for the record and load it again
    Regards,
    MADhu

  • Getting an error while fetching the data and bind it in the Tree table

    Hi All,
    I am getting an error "A navigation paths parameter object has to be defined - " while fetching the data and bind it in the Tree table.
    Please find the code and screenshot below
    var oModel = new sap.ui.model.odata.ODataModel("../../../XXXX.xsodata/", true);
    var oTable = sap.ui.getCore().byId("table");
    oTable.setModel(oModel);
    oTable.bindRows({
        path: "/Parent",
        parameters: {expand: "Children"}
    Can anyone please give me a suggestion to rectify this?
    Thanks in Advance,
    Aravindh

    Hi All,
    Please see the below code. It works fine for me.
    var oController = sap.ui.controller("member_assignment");
    var oModel = new sap.ui.model.odata.ODataModel("../../../services/XXXX.xsodata/", true);
    var Context = "/PARENT?$expand=ASSIGNEDCHILD&$select=NAME,ID,ASSIGNEDCHILD/NAME,ASSIGNEDCHILD/ID,ASSIGNEDCHILD/PARENT_ID";
    var oTable = sap.ui.getCore().byId("tblProviders");
    oModel.read(Context, null, null, true, onSuccess, onError);
    function onSuccess(oEventdata){
        var outputJson = {};
        var p = 0;
        var r = {};
        try {
            if (oEventdata.results){
                r = oEventdata.results;
        } catch(e){
            //alert('oEventdata.results failed');
        $.each(r, function(i, j) {
            outputJson[p] = {};
            outputJson[p]["NAME"] = j.NAME;
            outputJson[p]["ID"] = j.ID;
            outputJson[p]["PARENT_ID"] = j.ID;
            outputJson[p]["DELETE"] = 0;
            var m = 0;
            if (j.ASSIGNEDCHILD.results.length > 0) {
                $.each(j.ASSIGNEDCHILD.results, function(a,b) {
                outputJson[p][m] = { NAME: b.NAME,
                                     ID : b.ID,
                                     PARENT_ID: b.PARENT_ID,
                                     DELETE: 1};
                m++;
            p++;
        var oPM = new sap.ui.model.json.JSONModel();
        oPM.setData(outputJson);
        oTable.setModel(oPM);
    function onError(oEvent){
        console.log("Error on Provider Members");
    oTable.bindRows({
        path:"/"
    Regards
    Aravindh

  • No data Exists error while retrieving the data from a table

    Hi Everyone!
    I am getting "No Data Exists" error while retrieving the data from a table....where i need to check...if possible please give me example link.....please help me regarding this

    Hi !
    thanks for ur response...
    I have written vo.executeQuery for the table...My page is running but i m not getting the data....I have to select two LOV's and when I click on Go button the data has to display in the table...After the selection of LOV's when I click on Go in the table "No data exists" message is appearing...the table has the data and the query is also executing...please give info where i did the mistake....

  • Error while importing the export in new environment

    All,
    I have getting AIP-11052: Writing following objects: Transport Protocol Parameter failed due to following constraint violation: TRANSPORTPROTOCOLPARAMETER_UK. UNIQUE CONSTRAINT VIOLATION
    while importing the export. Has anyone encountered this issue? or Can anyone provide pointers as to why this might be occuring?
    Any help would be greatly appreciated!
    Dipal

    Hello Dipal,
    The erros seems to be because of some UNIQUE key constraint issue while importing the export. This may be due to some conflicting import data which is causing this voilation. If it is a possibility, please purge the entire repository and have a clean import or else please analyse the UI log to find some more details.
    Rgds,Ramesh

  • Getting Error while accessing the data using odata service

    Hi All,
    Iam new to SAP FIORI. 
    Iam getting the below error while accessing the data using odata service.
    "Failed to load resource: the server responded with a status of 404 (Not found)"
    "No 'Access-Control-Allow-Origin' header is present on the requested resource. Origin "
    i have tried all the solutions like changing the url pattern "proxy/htttp".
    and disabled - security in chrome (Chrome is Updated version).
    i tried with IE still got the same problem.
    And installed all the required software in eclipse
    While installing GWPA plugin i got the following error.
    let me know if any one have idea.
    Thanks in advance.

    > Do you want to add and/or update the data in the already existing tables or do you want to replace the content completely?
    >
    > so in that way :
    > bot the options are fine what ever take less time.
    Sorry mate, but YOU have to know what you want here.
    I gave you an easy to follow set of steps.
    As you don't seam to mind the outcome, just might just use them...
    > I wanted to know weathe i can use the  loadercli for thie export import or not? if yes then is there any new steps to do before i do the export import?
    We had this discussion before...
    >
    > For that the easiest option would be just to drop the tables of SAPR3 and run the import again.
    >
    > For ease of use you could also just do:
    > - logon as superdba
    > - drop user SAPR3
    > - create user SAPR3 password SOMEPW not exclusive dba
    >
    > After these steps you can easily pump the data into the database again.
    >
    > So here in th above given steps , i am creating a new SAPR3 user and why it is not exclusive dba ?
    >  i already have that user SAPR3 can i use the same.
    Yes, you do have the SAPR3 user.
    But you don't seem to like to read documentation or learn about how the tools work or anything like that.
    Therefore I gave you s simple way to reach your goal.
    Of course it's possible to reuse the user.
    But then you would have to deal with already existing tables, already existing data etc.
    You don't seem to be able to do that. So, the easy steps might be better suited for your needs.
    regards,
    Lars

  • While creating the Data source for Sql server am getting this error?

    Hi i new to Power BI sites and office 365,
    For my first step 
    1. i have successful creaeted Gateway in Office 365
    2.Now i have creating  the Data source for that gateway using Sql server R2
    3.While Creating the Data source using the Corresponding gateway i have given and next phase set Credentials  phase i have noticed one pop up window and it will shows me like
    Data source Settings and  below in that window it will show's u the things like 
    Datasource Name.
    Loading......
    Credentials type:
    Privacy Level:
    Failed to verify parameter
    I will get this results finally  ,so hw should i achive this probelm ,Please let me know if any one knows
    So kindly give me the proper answer for this.
    Regards

    Any suggestions for Chinnarianu?
    Thanks!
    Ed Price, Azure & Power BI Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • While deleting the data from cube..getting the following erros...

    while deleting the data from cube the below error occured in develoment..any help
    i have 4 requets.....2 are deltedd but no records for those...
    2 are not deleted and they have some data..
    please give me u r inputts pelase
    Performing check and potential update for status control table
    Message no. RSM1490
    Diagnosis
    If data is loaded into an InfoCube, or existing data is edited (aggregated/compressed/deleted/got from a DataMart), then there is a change in  the potential reportability of the data , or the possibility of deleting data by request, or of aggregating or compressing.
    This status of each one of these Cubes is stored in a status table, that is updated when there is any change to the status of a request in the Cube.
    The system now analyzes the requests in the Cube and compares the calculated status with the status table.
    If deviations from the status table arise then you are given the option of adjusting the status table.
    System Response Procedure Procedure for System Administration

    Hi,
    Did you tried by right click on cube and delete data? are those requests are compressed or roll up?

Maybe you are looking for

  • Report obsolete returns diff results when backups are UNAVAIL vs AVAIL

    Hello from Nova Scotia, Canada: I'm wondering if someone out there who lives and breaths rman (10gR2) for a living can shed some light... We are finding that report obsolete (as well as delete obsolete) are returning different results depending on if

  • Feature Suggestion: Auto Copy Libraries into Project from GUI

    I'm surely not the only one to think of this but I know I would find it very useful. When going to the Libraries section in the project settings under the desired configuration, we choose from all the libraries Jdev knows about. Of course when we shu

  • How do i put photoshop elements 10 on both pc and mac

    if i purchase the downladable version of photoshop elements 10 for my mac, do i have the option of also downloading it to my pc without purchasing it again?

  • How to install hard drive Macbook Pro 10.6.3

    I have a Macbook Pro Os X 10.6.3 Snow Leopard. I was told that I have a failed hard drive. It costs about $195 for repair at my local shop. I read that I can purchase my own hard drive from BestBuy and have geek squad to install it. The guy says it c

  • Planned and actual consumption for each confirmation

    Dear All, There is a production order. it has two operations. i have made three confirmations of each operation for that order. now i want to see the planned and actual consumption for each confirmation. how is it possible. However i know that on ord