Data from a deleted datasource

Hi guys,
I've deleted a datasource and since all data loaded with this datasource seems to be deleted in my cube when I do a report on these. Nevertheless, the requests of these are still in the cube.
Who can say me if the data have been really deleted ? If no, the way to retrieve them ?
MANY THANKS...

Thanks for your answer...
Unfortunately the data in the PSA have been deleted in the same time I deleted the Datasource.
Is it normal that the system deletes data already loaded when we delete the datasource?

Similar Messages

  • Simple example of a combobox displaying data from a CF datasource

    Can anyone point me to a simple example of a Flex 3 combobox that displays data from a ColdFusion datasource?  I cannot seem to find a simple example.  As always, thanks!

    Thanks for the response, but the example you provided is more complex than I need.  For example, I have a simple table accessed through ColdFusion.  The table has one filed with 7 records.  I need a flex page to display that data in DataGrid, ComboBox and List controls.  The Datagrid displays the 7 records but the ComboBox and List are blank.  How do I get the values from the database to display in the BomboBox and List controls?
    Below is my code:
    <?xml version="1.0" encoding="utf-8"?>
    <mx:Canvas xmlns:mx="http://www.adobe.com/2006/mxml"
                    creationComplete="initComponent()">
        <mx:Script>
        <![CDATA[
            import mx.collections.ArrayCollection;
            import mx.rpc.events.ResultEvent;
            import mx.rpc.remoting.RemoteObject;
             private var currentIndex:int = 0;
            private var _key:Object;
            [Bindable]
            public function get key():Object
                    return this._key;
            public function set key(key:Object):void
                    this._key = key;
             private function initComponent():void
                    refreshList(null);
             public function refreshList(event:Event):void
                    this.dataManager.getRoles(this.key);
            private function getRoles_result(event:ResultEvent):void
                    this.roleList.dataProvider = event.result as ArrayCollection;
                     this.roleList.selectedIndex = this.currentIndex;
        ]]>
        </mx:Script>
            <mx:RemoteObject
                id="dataManager"
                showBusyCursor="true"
                destination="ColdFusion"
                source="ifqgtfuses.com.vwRoles">
                <mx:method name="getRoles" result="getRoles_result(event)" />
            </mx:RemoteObject>
       <mx:Canvas width="100%" height="100%" x="0" y="0">
            <mx:DataGrid id="roleList"
                x="10" y="10">
                <mx:columns>
                    <mx:DataGridColumn dataField="vchar_role" headerText="Role" />
                </mx:columns>
            </mx:DataGrid>
            <mx:ComboBox x="10" y="160" id="rolesComboBox" dataProvider="{roleList.selectedIndex}"/>
            <mx:List x="120" y="10" id="rolesList" dataProvider="{roleList.selectedIndex}"/>
       </mx:Canvas>
    </mx:Canvas>

  • Recovering Data from a deleted Mac partition

    Through no fault of my own, (an IT department at a well know digital data storage company that will remain nameless) I had all the partitions deleted from my hard drive. No data was damaged, but the Mac partition and the FAT 32 I had windows 7 running on with bootcamp were both deleted. My hard drive is now 250 gigs of unallocated space.
    Is there any easy and inexpensive way to recover that data? I'm pretty sure I can't just "undelete" the deleted partitions but it would be nice to get my data off so I don't have to go from the DVD back ups from christmas time.
    Anyway I found Disk Warrior and Drive Genius that claim to recover data but they are both about $100.
    Any ideas?
    Here's the irony of the situation, I was using the utility given me by the IT department while trying to setup a NAS device so I can create regular time machine backups in case anything happened to my lappy. So by trying to get a system working that would help me protect my data, it deleted my data. I would have been fine without it. GRRRR...

    Hi, Take the HDD out and connect it to another system then run R-Studio against it. It can recover FAT partitions and I have used it many times with great sucess.

  • How can I access data from accidentally deleted user?

    I was given a MacBook Pro by a family member. I set myself as the Administrator, but could not get rid of the previous owner's name on the Home folder.
    In my attempts I accidentally deleted something that made the entire computer unaccessable. I can boot up, however, all the work and applications I installed are unaccessable. I can't even open Preferences.
    I used TechTool Pro to start up and explore the drive, and found that the previous owner's name was on a folder with a red-circled minus sign. The data (considerable) was inside, but I can't access it.
    Is there any way I can get access to that data? I slaved the laptop to my iMac and transferred the folder (with 32 gb of data).
    When I tried to open it I got the message: "The folder "sandy....." can't be opend because you don't have permission to see its contents." I used Get Info on the folder and changed Sharing and Permissions to Read and Write, but still no access.
    What can I do, if anything to gain access to that folder?

    http://docs.info.apple.com/article.html?path=Mac/10.7/en/mchlp2968.html

  • Extracting data from multiple datasources through DAC

    We have to extract data from two Oracle datasources (GL datasource 1 and GL datasource 2) by DAC client to one Datawarehouse. We have configured DAC for one datasource and it is working fine. How can we add another datasource into the same DAC client.
    Neeraj

    You will need to update the Source DB connections in both the DAC and Informatica Workflow before you want to load the DW with the other GL Database. You will contnuously have to switch these configuration settings or stand up another DAC Server, Repository and Client that point to the same Datawarehouse.

  • Can I load data from 2 datasource to single infosource.

    Hi
    Please let me know , can we load the data from 2 -3 datasource to single infosource, in 3.5 data flow.
    for eg : 2LIS_11_VITM, 2LIS_12_VCITM and 2LIS_13_VDITM feeding data to sigle infosurce ?? How to handal such requirment. ??
    Appreciate your help.
    Regards
    Ashutosh D

    Hi,
    It is  recommend that you use an InfoSource if you want to connect a number of different DataSources to a target and the different DataSources have the same business rules.
    In the transformation, you can align the format of the data in DataSource with the format of the data in the InfoSource.
    The required business rules are applied in the subsequent transformation between the InfoSource and the target. You can make any changes to these rules centrally in this one transformation, as required.
    Revert back for further queries.
    Regards,
    Rajkandula

  • Unable to extract data from 0HR_PA_PD_2 datasource

    Hi,
    I am extracting the data from 0HR_PA_PD_2 (Appraisals) datasource to the cube 0PAPD_C02 (Appraisals).Though the data available in R/3 (83 records,checked in RSA3) it is giving the message "No data available in the source system".I checked the RFC connection,IDOCs,SM58 for any message.But everything is fine.In the status tab ,it is showing "Info IDoc received with status 8".But the data is available in R/3.
    Please help me if any one faced this type of issue or is there any other steps has to be followed.
    Thanks & Regards,
    Dinakar.

    Hi Dinakar,
    Check Roberto's reply in this thread... It might help you...
    Re: Problem in InfoPackage for Billing Document Item Data(2LIS_13_VDITM)
    Also check these related threads...
    Info IDoc received with status 8. - extracting data from R/3 -co-pa
    Info IDoc received with status 8. - extracting data from R/3
    error while doing init load ..Idoc received with status 8
    Info Idoc received with status 8
    Regards,
    KK.

  • What is the best way to append data from one field to another?

    I have the following table, table1:
    Name Null? Type
    MAIL_ID NOT NULL NUMBER(10)
    LAST_NAME VARCHAR2(45)
    FIRST_NAME VARCHAR2(45)
    MIDDLE_INITIAL VARCHAR2(1)
    ADDRESS_1 VARCHAR2(45)
    CITY VARCHAR2(35)
    STATE VARCHAR2(2)
    ZIP VARCHAR2(10)
    REMARKS VARCHAR2(200)
    The table has duplicate entries that need to be removed. The records that will be removed need the
    data in the Remarks column appended to the Remarks data of the record that is not deleted.
    For example, the following listing shows a sample of the duplicate records.
    Mail ID Last Name First Name M Address City St ZIP Remarks
    189 BROWN STEPHEN 6706 MOESER LN EL CERRITO CA 94530-2909 Sf7#s124,f16#d7996(NML)[Cl#117][Ml#1649][NMf1#d288][NCf9#d319][SNl#e62]
    211023 BROWN STEPHEN B 6706 MOESER LN EL CERRITO CA 94530 RLl#a12047[IDl#i398]
    287796 BROWN STEPHEN B 6706 MOESER LN EL CERRITO CA 94530 SNl#e1163
    The following listing shows how the kept record should appear after the duplicate records are deleted.
    Mail ID Last Name First Name M Address City St ZIP Remarks
    189 BROWN STEPHEN 6706 MOESER LN EL CERRITO CA 94530-2909 Sf7#s124,f16#d7996(NML)[Cl#117][Ml#1649][NMf1#d288][NCf9#d319][SNl#e62]RLl#a12047[IDl#i398]SNl#e1163
    I have the process of deleting duplicates working but have yet to determine the best way to move
    the Remarks data from the deleted records to the preserved record.
    I know there are probably various ways to approach this.
    Any suggestions will be greatly appreciated!
    Here is the sql for deleting duplicates.
    DELETE FROM table1
    WHERE mail_id in (SELECT mail_id FROM table1
              where not first_name = 'Null' and
    not last_name = 'Null' and
              not city = 'Null' and
              not state = 'Null'and
    not last_name = 'Anon'
              minus
              select min(mail_id) from table1
              group by first_name, last_name, city, state, address_1, organization, title);
    THANKS in advance!!!!

    Here's quick and dirty example probably a better way to do it, but this is what I came up with quickly.
    My table looks like this:
    MAIL_ID LAST FIRST PHONE REMARKS
    123 Ruff Shawn 555-555-5555 Called 10-10-04
    135 Ruff Shawn 555-555-5555 Called 10-12-04
    201 Ruff Shawn 555-555-5555 Called 10-19-04
    The code below will concatenate the remarks column from the rows, and delete the 135 and 201 rows, then update the 123 row with the concatenated remarks.
    declare
    l_remarks varchar2(500);
    l_min_mail_id number;
    begin
    select min(mail_id) into l_min_mail_id
    from test
    group by last, first, phone;
    select remarks into l_remarks from test where mail_id = l_min_mail_id;
    for i in (select mail_id, remarks from test
         where last = 'Ruff'
              and first = 'Shawn'
              and phone = '555-555-5555'
              and mail_id <> l_min_mail_id)
    loop
    l_remarks := l_remarks||','||i.remarks;
    delete from test where mail_id = i.mail_id;
    end loop;
    update test set remarks = l_remarks where mail_id = l_min_mail_id;
    commit;
    end;
    Hope this helps.

  • Problem While loading Master data from R/3

    Hi All,
    When i am trying to load masters data from R/3 datasources to BW Info objects in Info provider menu data is coming to PSA but going to dta targets. the data is matching with the data in R/ 3 but
    Missing message: Request received
    Missing message: Number of sent records
    Missing message: Selection completed
    Data Package 1 : arrived in BW ; Processing : Selected number does not agree with transferred n
    Info IDoc 1 : sent, not arrived ; IDoc ready for dispatch (ALE service)
    and error message says ERROR 4 in the update.
    Can you please help me in this regard . Its Urgent.

    Hi,
    Check for pending IDOCS in BD87 or SM58 and execute them.
    In RSMO, select the particular load you want to monitor.
    In the menu bar, Environment >>> Transact. RFC >>> Select whichever is required, BW or Source System.
    In the next screen you can select the Execute button and the IDOCS will be displayed.
    Refer this Tcodes for IDOC status.
    WE02 Display IDoc
    WE05 IDoc Lists
    WE06 Active IDoc monitoring
    WE07 IDoc statistics
    WE08 Status File Interface
    WE09 Search for IDoc in Database
    WE10 Search for IDoc in Archive
    WE23 Verification of IDoc processing
    IDOC
    www.erpgenie.com/sapgenie/docs/ale_whitepaper.doc
    Check Note 561880 - Requests hang because IDocs are not processed.
    OR
    Transact RFC error 
    tRFC Error - status running Yellow for long time (Transact RFC will be enabled in Status tab in RSMO).
    Step 1:  Goto Details, Status get the IDoc number,and go to BD87 in R/3,place the cursor in the RED IDoc entroes in tRFC
    queue thats under outbound processing and click on display the IDOC which is on the menu bar.
    Step 2: In the next screen click on Display tRFC calls (will take you to SM58 particular TRFC call)
    place the cursor on the particular Transaction ID and go to EDIT in the menu bar --> press 'Execute LUW'
    (Display tRFC calls (will take you to SM58 particular TRFC call) ---> select the TrasnID ---> EDIT ---> Execute LUW)
    Rather than going to SM58 and executing LUW directly it is safer to go through BD87 giving the IDOC name as it will take you
    to the particular TRFC request for that Idoc.
    Thanks,
    JituK

  • MySQL move date from one table to another

    I was wondering if there is a MySQL command that will let me move a selected row of data from one table to another. both tables have the same columns and declaration type (one table is actually an archived table on old data)
    example
    I wasnt to move all data in Table1 where the date is greater than 30 days old to Table 2
    -- so the result should be...import all rows to Table 2 where the date is greater than 30 days old..and delete all date from Table 1 that is greater than 30 days.
    currently..I'm doing three process
    1) get all row that is greater than 30 days  
        "SELECT * FROM Table1 WHERE TO_DAYS(NOW()) - TO_DAYS(dateField) > 30"
    2) insert data into Table2
        while (res.hasNext())
              TableData data = ..... // .get row
              dataList.add(data);
              for (int i = 0; i < dataList.size(); i++){
                    pstm.setString.....
                    pstm.addBatch()
              pstm.executeBatch();
    3) delete data from Table1
        "DELETE FROM Table 1 WHERE  TO_DAYS(NOW()) - TO_DAYS(dateField) > 30"

    for this app..losing a few rows does not
    impact on how we analyze the data.That's what everyone always tells me too. But 99% of the time they come back and want to know why the cannot balance and/or validate the data between two runs taken only minutes from each other.
    I've seen people puzzle over data for days that they swear they ran the exact same utility for their tests, but they were in fact using live data, and additional data had accrued but since all they had to do was execute the a script without parameters (they didn't put in a stop time), they got two different answers and it always, and I mean always confuses people. Be safe, and put the option in for and end date/time, then when they waste days trying to figure out why the two different observations gave them different numbers, they cannot blame you (because you gave them the option)!
    My 2 cents for the day...

  • Extract data from Traderu0092s and Scheduleru0092s Workbench

    Hi Guys,
       Anybody had experience extract data from Trader’s and Scheduler’s Workbench? Is there standard extractor or everything has to be csotomized?

    Can you explain your scenario a little more as FIM is not really a source of data.  FIM does however have data storage for its mapping rules, job logs, and even transactional data which is access by EPM apps to  "drill-to-origin".  
    Rather FIM is a middle ware application to transfer data between applications.  Working underneath FIM is Data Services.  You can use Data Services to move data from any source into BW.  Or you can use BW's own ETL capabilities with UD Connect or DB Connect to bring data from an outside datasource.
    Best regards,
    [Jeffrey Holdeman|http://wiki.sdn.sap.com/wiki/display/profile/Jeffrey+Holdeman]
    SAP Labs, LLC
    BusinessObjects Division
    Americas Customer Solutions Adoption (CSA) team

  • Protecting certain backups from being deleted?

    say, i have a time machine backup. and its running full. the oldest backup i have is on 20th december 2010. i wanna know if there's a way to prevent that or other certain dates from being deleted?

    auturmis wrote:
    say, i have a time machine backup. and its running full. the oldest backup i have is on 20th december 2010. i wanna know if there's a way to prevent that or other certain dates from being deleted?
    No.
    As Matt says, it sounds like you may be doing something rather dangerous. Please clarify just what's going on, and we can provide some help.

  • Delete Data in Infocube which come from the same datasource before loading

    Hi,experts
    We would like to delete the data in infocube which come from the same datasource--0CO_OM_CCA_1  before we load it again.
    and would like to using a process chain to do it.but it seems difficult in BI7.0
    Could you give some suggestion,thanks
    Best Regards
    Jero

    Hi Ian,
    have you tried to use the "Get Delta only Once" flag in the DTP.
    Please see for example thread
    <a href="https://forums.sdn.sap.com/thread.jspa?threadID=370427">https://forums.sdn.sap.com/thread.jspa?threadID=370427</a>
    for some details.
    Nevertheless, the process type 'Delete overlapping requests' is most of the times needed in conjunction with this flag and should work. If in doubt please send me a private EMail containing the customer problem number. Maybe you have a specfic scenario.
      Cheers
         SAP NetWeaver BI Organisation

  • Deleting data from a cube

    Hi,
    I added a info object to a dimension in a cube. now i want to delete the info object from the cube. But since the cube has data in it, i cannot do it unless i delete all the data from the cube. When i select 'delete data', i'm given 2 options, delete fact table only and 2nd option delete fact table and dimension table. what is the difference between the two and what option should i select.
    i then need to load the cube back from ODS. the cude right now is setup in the proce4ss chain to do a delta load every day. would this be affected after i delete the data and load it again from the ODS?
    Thanks
    Sameer

    The way I understand your question is...
    1) You have data in the cube from which you have to remove an infoobject
    Ans: To acheive this, you have to delete the data from the cube
    2) You have deleted the data and then deleted the IO, activated the cube and the update rules from the ODS which feeds this cube
    3) Now you have to reload the deleted data from the ODS. But you already did an init and you are loading delta's from this ODS to the cube through process chain.
    Ans: As you've already deleted the whole data from the cube, do a full load from the ODS (you can do this even when delta initialization is done) to your cube by creating a new infopackage under the export datasource of the ODS. This will capture all the records from the ODS. Then to be on the safe side so as not to load any duplicate records...go the infopackage that you've used to do the full load (you can even create a new Info package)...SCHEDULER -> Initilization options for source system -> Delete the existing INIT -> Then do Init again...
    Leave the delta infopackage as is and let the process chain take care of the deltas from now on...
    Hope this helps. Let me know I am missing something here...

  • Load Data with 7.0 DataSource from Falt file to Write Optimized DSO

    Hi all,
    we have a problem loading data from flat file using the 7.0 datasource.
    We have to load a flat file (monthly) into a WO DSO. The infopackage load file in full mode into the Datasource (PSA) and the DTP Load in delta mode data from datasource into the WO DSO.
    When i load the second file in the Datasource, the DTP load all data present in the Datasource and not only the new one as aspected using Delta mode.
    Has anyone any tips to help me?
    Thank you for help.
    Regards
    Emiliano

    Hi,
    Iam facing the similar problem.
    Iam using Write Optimized DSO and i have got only 1 req in PSA (have deleted all previous req from PSA and DSO).
    When iam doing a delta load from PSA to DSO, i expect to see only that 1 req to get loaded into DSO.
    But, its picking up the data from 3 other reqests and doubling the records...
    Can you please help me, how did you managed to get out of that isue?
    Cheers,
    Nisha

Maybe you are looking for