Finding some tool for loading data

I will finding some tools for loading, migrate
data from another dbms like Access, DBF Files, Excel
to Oracle Database
Please suggess me
Thank you
Kachornsak

I suppose the various sources will have various optimal solutions.
Like in case you want to import from excel I would suggest to save your data to a csv and use import. (imp)
For access files the best solution is the Oracle Migration Workbench.
Best Regards,
Kris

Similar Messages

  • FDMEE Import error "No periods were identified for loading data into table 'AIF_EBS_GL_BALANCES_STG'

    Hi,
    We are having trouble while importing one ledger 'GERMANY EUR GGAAP'. It works for Dec 2014 but while trying to import data for 2015 it gives an error.
    Import error shows " RuntimeError: No periods were identified for loading data into table 'AIF_EBS_GL_BALANCES_STG'."
    I tried all Knowledge docs from Oracle support but no luck. Please help us resolving this issue as its occurring in our Production system.
    I also checked all period settings under Data Management> Setup> Integration Setup > Global Mapping and Source Mapping and they all look correct.
    Also its only happening to one ledger rest all ledgers are working fine without any issues.
    Thanks

    Hi,
    there are some Support documents related to this issue.
    I would suggest you have a look to them.
    Regards

  • Steps for loading data into the infocube in BI7, with dso in between

    Dear All,
    I am loading data into the infocube in BI7, with dso in between. By data flow looks like....
    Top to bottom:
    InfoCube (Customized)
    Transformation
    DSO (Customized)
    Transformation
    DataSource (Customized).
    The mapping and everything else looks fine and data is also seen in the cube on the FULL Load.
    But due to some minor error (i guess), i am unable to see the DELTA data in DSO, although it is loaded in DataSource through process chains.
    Kindly advise me, where did i miss.
    OR .. Step by step instructions for loading data into the infocube in BI7, with dso in between would be really helpful.
    Regards,

    Hi,
    my first impulse would be to check if the DSO is set to "direct update". In this case there is no Delta possible, because the Change log is not maintained.
    My second thought would be to check the DTP moving data between the DSO and the target cube. If this is set to full, you will not get a delta. It is only possible to create one DTP. So if you created one in FULL mode you can't switch to Delta. Just create the DTP in Delta mode.
    Hope this helps.
    Kind regards,
    Jürgen

  • Process chains for loading data to target is not functioning

    Hi SAPians,
    Recently, we have upgraded the firmware on IBM P590 with the help of IBM on Saturday i.e. 06/12/2008 (The firmware of P5-590 was SF235_209. It was upgraded to SF240_XXX) and since then the process chains for loading data to targets are not functioning properly.  We have stopped all the SAP services, database services & OS services from our end and services have been rebooted after firmware upgrade.
    However, the problem with the process chains that load transaction data and hierarchies is solved but the chains that load master data are still not working as scheduled.
    We tried to load the master data manually, by running DTP to load data to an object analysis code (attributes) the request remained YELLOW. After refreshing it several times, the request turned into RED. The error message was - PROCESSING TERMINATED (screenshot 1 attached).
    To mitigate this we tried deleting the bad request and it prompted with the below message:
    "Cannot lock request DTPR_4C37TCZGDUDX1PXL38LVD9CLJ of table ZANLYSIS for deletion (Message no. RSMPC178)" (screenshot 2 attached)
    Please advise us how the bad request should be deleted?
    Regards,
    Soujanya

    Hi Sowjanya,
    Follow the below procedure to make yellow request to RED:
    Use SE37, to execute the function module RSBM_GUI_CHANGE_USTATE
    From the next screen, for I_REQUID enter that request ID and execute.
    From the next screen, select 'Status Erroneous' radiobutton and continue.
    This Function Module, change the status of request from Green / Yellow to RED.
    Once it is RED, you can manually delete that request...
    Releasing LocK
    Gott Transaction Code SM12 and release the lock.
    Hope It helps you.
    Regardss,
    Nagaraju.V

  • When I use 32 bit version I easily find plugins but for Nighty 64 bit I can not find some plugin for this browser what can i do I don't want to return back to 32 bit please advice

    When I use 32 bit version of Mozilla I easily find plugins but for Nighty 64 bit I can not find some plugin for this browser what can i do
    I don't want to return back to 32 bit I love Nighty 64 bit and I'm satisfied with performance
    what can I do please advice ?

    Many plugin vendors haven't updated their plugins to work with x64, or you have to download a special x64 version (Flash and Java for example). Which plugins is it that you are having problems with?

  • HT1918 i want to change the country from USA to EGYPT in the itunes but i get a message saying, u have remaining balance which is 0.44 USD, how can i find some thing for 0.44 USD to spend this sum so i can switch my account to the right store??

    i want to change the country from USA to EGYPT in the itunes but i get a message saying, u have remaining balance which is 0.44 USD, how can i find some thing for 0.44 USD to spend this sum so i can switch my account to the right store??

    Buy something for more than the balance and charge the remainder to a credit card, or write Apple and ask them to keep the balance.
    iTunes Customer Service Contact - http://www.apple.com/support/itunes/contact.html > Get iTunes support via Express Lane > iTunes > iTunes Store

  • Procedure for loading data through 0CO_OM_CCA_9 without taking downtime .

    Dear Friends ,
    We have already maintained the flow from 0CO_OM_CCA_9 to Infocube 0COOM_C02 . Now i enhanced two fields in the datasource which i want to load into the cube . The infocube already contains data from 2008 onwards . Now i know that for controlling datasources SAP ECC/R3 downtime is not needed . Requirement is to load the historical data again for the two enhanced fields also from 2008 onwards .
    Kindly let me know the steps for doing as the delta mechanism is already activated for this flow ?
    Thanks & Regards ,
    Ashutosh Singh

    Hi All,
    I am able to create this in client 100 but not in 120. I created this program by copying the source code from sap note and run in test mode in client 100 of development server. I am getting the message that "there is no inconsistency in table BWMO2_TIMEST".
    1. I am using client for loading data 120 not 100.
    Please let me know the further steps.
    Thanks,
    Harry

  • ........ Expert  suggestion needed for Loading Data in DB

    Hi all
    The current requirement we have is to Load data from Excel File to oracle Databases.
    The Excel file has around 30 Sheets , each corresponds to a table, ( means 30 tables in the database) .
    Currently we are using sqlldr commands to load data from CSV files to the Oracle Database. The only problem is that sometime the DBA has to go out or is busy workign on somthing else, so May i kindy get some expert suggestions on how to automate this DataLoading Task.
    Somebody has suggested to My Manager ( who is not aware of oracle and IT at all) that data can be loaded via ODBC. It is suggested to him that all we need to do is to place the CSV files on the Server at a particular folder and Oracle will do the rest.
    I am not that proficient in Data loading methods.. so may i know any technique which will simplify/automate the Task.
    I mean how can we automate so that every time , the the sql loader scripts run at command promt. I think data base ( Oracle ) as nothing to do with command promt i feel. isnt it ..
    Kindly have ur expert./experienced suggestions.
    I would be highly grateful to all
    Regards

    To automate sqlldr scripts, you would usually write a OS script file that will periodically looks for files in a given directory and run the sqlldr commands:
    - on Windows: .bat or .wsh files to be scheduled with "at" command or windows scheduled tasks tool
    - on Unix: a script shell in the crontab.
    Much more complicate but without any OS script:
    - write a PL/SQL code to read, parse the file to be loaded using UTL_FILE
    package
    - this PL/SQL code must also generate INSERT statements and process errors ...
    - this PL/SQL code can be scheduled with DBMS_JOB package to run a periodic intervals.
    Message was edited by:
    Pierre Forstmann

  • Where to find best practices for tuning data warehouse ETL queries?

    Hi Everybody,
    Where can I find some good educational material on tuning ETL procedures for a data warehouse environment?  Everything I've found on the web regarding query tuning seems to be geared only toward OLTP systems.  (For example, most of our ETL
    queries don't use a WHERE statement, so the vast majority of searches are table scans and index scans, whereas most index tuning sites are striving for index seeks.)
    I have read Microsoft's "Best Practices for Data Warehousing with SQL Server 2008R2," but I was only able to glean a few helpful hints that don't also apply to OLTP systems:
    often better to recompile stored procedure query plans in order to eliminate variances introduced by parameter sniffing (i.e., better to use the right plan than to save a few seconds and use a cached plan SOMETIMES);
    partition tables that are larger than 50 GB;
    use minimal logging to load data precisely where you want it as fast as possible;
    often better to disable non-clustered indexes before inserting a large number of rows and then rebuild them immdiately afterward (sometimes even for clustered indexes, but test first);
    rebuild statistics after every load of a table.
    But I still feel like I'm missing some very crucial concepts for performant ETL development.
    BTW, our office uses SSIS, but only as a glorified stored procedure execution manager, so I'm not looking for SSIS ETL best practices.  Except for a few packages that pull from source systems, the majority of our SSIS packages consist of numerous "Execute
    SQL" tasks.
    Thanks, and any best practices you could include here would be greatly appreciated.
    -Eric

    Online ETL Solutions are really one of the biggest challenging solutions and to do that efficiently , you can read my blogs for online DWH solutions to know at the end how you can configure online DWH Solution for ETL  using Merge command of SQL Server
    2008 and also to know some important concepts related to any DWH solutions such as indexing , de-normalization..etc
    http://www.sqlserver-performance-tuning.com/apps/blog/show/12927061-data-warehousing-workshop-1-4-
    http://www.sqlserver-performance-tuning.com/apps/blog/show/12927103-data-warehousing-workshop-2-4-
    http://www.sqlserver-performance-tuning.com/apps/blog/show/12927173-data-warehousing-workshop-3-4-
    http://www.sqlserver-performance-tuning.com/apps/blog/show/12927061-data-warehousing-workshop-1-4-
    Kindly let me know if any further help is needed
    Shehap (DB Consultant/DB Architect) Think More deeply of DB Stress Stabilities

  • Tools for load testing SharePoint 2007

    Please advice a tools, which i can use for load testing a SharePoint 2007. We had one issue with performance and want now to detect a reason. So i need to simulate a load testing with 500-600 users for one SharePoint page.
    I have found some information, that i can use a Visual Studio Test Edition, but my level of MSDN subscriber has just a Visual Studio. May be there is another free tools for this...

    Hi Evgeny,
    SharePoint test data load tool (WSSDW.exe) can be used for your scenario, this article should help you to start with-
    Tools for performance and capacity planning (Office SharePoint Server)
    http://technet.microsoft.com/en-us/library/cc262198(v=office.12).aspx
    To download the tool - http://sptdatapop.codeplex.com/
    Bill Baer has good blog on this -
    http://blogs.technet.com/b/wbaer/archive/2007/08/02/stress-testing-microsoft-office-sharepoint-server-2007-windows-sharepoint-services-3-0.aspx
    http://blogs.msdn.com/b/sanjaynarang/archive/2009/04/18/sharepoint-performance-stress-load-testing.aspx
    http://blogs.msdn.com/b/edglas/archive/2007/02/17/sharepoint-load-tests.aspx
    another tool to consider - Web Capacity Analysis Tool
    http://technet.microsoft.com/en-us/magazine/2008.04.utilityspotlight.aspx?pr=blog
    http://support.microsoft.com/kb/231282/en-us
    http://blogs.iis.net/thomad/archive/2010/05/11/using-the-wcat-fiddler-extension-for-web-server-performance-tests.aspx
    Hope this helps!
    Ram - SharePoint Architect
    Blog - SharePointDeveloper.in
    Please vote or mark your question answered, if my reply helps you

  • Problem in finding the datasources for master data for 0TCT_MC11

    Hi,
    I want to load the data in 0TCT_MC11(BI7) , which is one of the technical content multiprovider. I have a datasource to load data in it , but I am not able to find the data source for the master data objects. hence I am stopped in doing loading to this MP. Can some body let me know where can I get the Information about the Datasources to load master data.
    Thanks,
    Jeetu

    Hi Srilatha,
    While reading data from 0TCT_MC11 through query I am getting the following message...(query name 0TCT_MC11_Q0130)
    "Error while reading data; navigation is possible
    "Value XY2CLNT400AHPL_ACCT"
    "You are trying to load the invalid characteristic value 1. (hexidecimal representation 0044004200320043004C004E00540036003000300049004F00)."
    Please let me know if you have any suggestions for this.
    Thanks, Jeetu

  • Help for loading data from flat file it is an imp for me!!

    Hi All,
    I am working on loading data. I am getting data from Informix as '|' (pipe) delimited.
    Each file consists of 20-25 columns. But I need few columns information only for this load. Thanks in advance for your reply.
    Thanks,
    RM

    go for fixed length loading. refer to some of the sample scripts
    (ref: search for *.ctl in find files )

  • Unable to find file error while loading data from text file to Oracle

    Hi,
    I am having a interface where i am loading a data of Text file to Oracle.
    But when i am trying to do this i am getting following error.
    ODI-1227: Task SrcSet0 (Loading) fails on the source FILE connection SAPMM.
    Caused By: java.sql.SQLException: File not found: d:/mdb/#General.get_filename
         at com.sunopsis.jdbc.driver.file.FileResultSet.<init>(FileResultSet.java:160)
         at com.sunopsis.jdbc.driver.file.impl.commands.CommandSelect.execute(CommandSelect.java:57)
         at com.sunopsis.jdbc.driver.file.CommandExecutor.executeCommand(CommandExecutor.java:33)
    SAPMM is connection name.
    I am using get_filename to get the filename and it is fetching correct value as the this variable refreshes in previous step of this interface.
    KM used for loading is File to SQL
    What would be cause of this error?
    Thanks,
    Mahesh

    Hi,
    Did a testing and following are the result
    I have
    A> created package having steps like
    1. Declaration of variable v_filename.
    2. Refreshing variable v_filename.
    3. Execution of Interface which gets the file name from v_filename and load into target table
    Package executes successfully.
    B> created package having steps like
    1. Declaration of variable v_filename.
    2. Refreshing variable v_filename.
    3. Scenario of Interface which gets the file name from v_filename and load into target table
    Package executes with erre as it is not able to find the file
    C> created package having steps like
    1. Declaration of variable v_filename.
    2. Refreshing variable v_filename.
    3. Execution of Interface which gets the file name from v_filename and load into target table
    4. Now create a scenario of the package , use the generated scenario in another package say main_package
    Execution of main_package is successful.
    Thanks,
    Sutirtha

  • How to find Info Source for Export Data source in Data Marts node

    Hi
                  I need to load data from ODS to Info Cube. I created the Export Data source for the ODS. I can see the Export Data Source but in the Data Marts node of Info Source i cannot find the Info Source for the Export Data source i created. I replicated Data sources in the BW source system. I also tried to use Insert Lost Nodes from the Context Menu of the InfoSource node but nothing worked. Please let me know what i need to do to see the Info source in the Data Marts.
    Thanks
    Padma

    In the infosource tab in RSA1 - use settings --> display generated objects
    you will be able to see the datamart infosources..

  • Options for loading data from legacy mainframe to SAP CRM and ECC

    Hello All,
    I am new to SAP.
    As part of data conversion planning from legacy mainframe to SAP CRM and ECC systems, I need to know the different tools available for loading the data and comparative analysis of the tools by showing the features. Please also describe the process to evaluate the tools (prototyping, technical discussions etc) and make the recommendations.
    I would also appreciate any information on testing tools like (eCATT) and any third party tools (like Informatica)
    I know I am asking for lot of things here. But, I really appreciate any information reagrding this.
    Thanks,
    Balu

    Hi
    Data Migration(Conversions) mainly involves the following Major Steps
    Discovery->Extract->Cleanse->Transform->Load
    Discovery-> Involves identifying various source systems available where the actual data is residing.
    Extract->Extract the data from the source systems(legacy systems)
    Cleanse->Cleanse the data by enriching it,De-duplications etc
    Transform->Transform the data into SAP Loadable format
    Load->Upload the data into SAP Using the Load Programs(Eg,BDC,LSMW,Data Transfer Workbench,BAPIs etc)
    Also i would request you to visit http://www.businessobjects.com/solutions/im/data_migration.asp
    Cheers,
    Hakim

Maybe you are looking for

  • HT204053 I have stored music on my iPad & I am unable to get it on my iPhone or Mac, help?

    I have purchased music which is on my iPad, the problem I am having is I cannot sync it to my Mac or iPhone?  I have followed all directions & still cannot sync it through the cloud.  Help?

  • Dynamic JDBC URls at Runtime

    Hi Experts, I did my application in jdev 11.1.1.6 and deployed using Jdbc Url connections and it worked successfully (Actually I m working with 2 diff Database connectiosn in my application) But the thing is my company need me to do something like I

  • Interfaces in abap programming

    Hi,     Please tell me what is an interface in abap program and also let me know how many types are there with ex's. Thanks Ali

  • Adobe air installer problem. app behaving strange

    Hi all i developed a desktop app in flex..In this i set the alpha of a text input /text area to alpha=0 .the app is behaving different..on different pc's..i executed to debug the app on my pc on setting the alpha=0 for text input the background becom

  • Cannot install due to registry errors.

    *NOTE TO SUPPORT STAFF: I DID NOT VIOLATE YOUR "SIGNATURE" RULES IN MY LAST THREAD, I DIDN'T EVEN INCLUDE A SIGNATURE. PLEASE ACTUALLY READ THE THREAD BEFORE YOU GO DELETING THINGS FOR NO REASON. ALL LINKS IN THIS THREAD ARE RELEVANT TO THE SUBJECT A