DSE Data Import Question

Hi,
When we receive data back from DSE after cleansing & normalization, are all the fields copied over the un-classified records (transaction) based on the key ? Or is it driven by the mapping between Classified DSO and Detail DSO ?
According to that mapping not all the fields are mapped from Inbound DSO to Detail DSO, hence the question.
Regards

Yes, this data gets loaded to the classified DSO object in the inbound layer and depending on what data needs to be loaded you have to update the transformation between this DSO and the detail layer DSO object.  Out of the box this transformation may or may not fit your needs, we expect customer to change this mapping based on their DSE process.
Regards,
Rohit

Similar Messages

  • Organize by Date - Import question

    What;s the best practice for importing by Date
    1. 2005/2005-12-17
    or the one that makes separate folders for year, month and date
    2. 2005/12/17
    I'm making a new library for my new D200, so I want to make a right choise

    I organize this way:
    Each folder has the complete date and a "brief" description of the content.
    (The description can be quite long but that's not a problem).
    Each day can have several different folders.
    This makes it also easy to sort out a specific group of images on the disk.
    And you can quite easily transfer you library to another app if you, forbid it, ever would get tired of LR.
    And as said above, you can easily browse by date with the metadata browser
    070521 - Flower
    070521 - Birds

  • Data import for users of forms created with Livecycle Designer

    Hello,
    I have seen several posts regarding data import for forms created by Livecycle Designer but nothing that helps with something I am trying to accomplish.  I can create a data connection and import information in a form but what I would like to do is import data, then send the pdf for completion to a user.  There are a few data elements that I have available and the rest of the information comes from from the user.  The problem I run into is once I create a data connection, the pdf is ALWAYS looking for the source file for that data.  I simply want to prepopulate some fields and send to the various users for completion.  Any help would be greatly appreciated.
    Thanks!

    Which type of Data Connection are you trying to create?
    XML Schema, Sample Data File or WSDL?
    Creating any one of first two types(mentioned above) will only create schema and will never import any data into PDF.
    If you create the WSDL connection, you can surely import data (i.e. prepopulate data) into your PDF and forward it for users review/fill.
    If I misunderstood your question, please get me clarified.
    Nith

  • Data Import Wizard from Excel on DATE column

    I have exported data to a csv in Oracle SQL Developer and one of the columns are of an Oracle Date datatype. In opening up Excel I see the Date column represented as 15-MAY-01 12.00.00.000000000 AM.
    I am now going through the Data Import Wizard to create a new table and import the data. When I get to Step 3 of 4 it reads the data and assigns it a VARCHAR2. I want the column to be Date or Timestamp. In this case I choose timestamp and for format entered DD-MON-YY hh.mi.ss.fffffffff AM.
    When I go to Step 4 of 4 to and click the Verify button I get:
    "Verifying if the Date columns have date formats" Date columns DOT_LOGN_EFDT, have invalid or null date formats.
    I am using Oracle SQL Developer 2.1.0.63 Build 63.73.
    Any idea what I put in the format window to read the excel value of 15-MAY-01 12.00.00.000000000 AM correctly?

    I am not able to reproduce your problem. When I export a date column, I do not get the time portion. This is the expected behavior. When re-importing the date, I therefore have no problem. I have tried manipulating the tools->preferences->database->nls->date format mask to include the time (and fractional seconds), but export still only exports the date portion. The fractional seconds is not valid for a date field, which explains the import error you are getting. The question remains as to why the exported data has the time. Can you describe your exact steps for doing the export?

  • Using Data Import Handler

    This question was posted in response to the following article: http://help.adobe.com/en_US/ColdFusion/10.0/Developing/WSe61e35da8d318518-1acb57941353e8b4 f85-7ffe.html

    This page should mention that the Data Import Handler is not available in ColdFusion 10 Standard (so available only in Enterprise, Trial, or Developer editions.)

  • Data Import from Excel - Users Logged on

    Hello all,
    I am new the forums and am learning SAP from a user and administrator perspective so please excuse if some of my questions seem obvious.  I appreciate everyone's help very help
    I have the task of setting up two new company profiles in addition to our current one.
    I am at the stage of importing Item Master and Business Partner data.
    1) What is the preferred method?  DTW vs. Import from Excel?  Is one more suitable then the other for this purpose?
    2) The help file for importing via excel states "If SAP Business One is running on a network, you have ensured that other users will be logged off the system during the import."  Could someone please explain:
    - What are the risks of running DTW or imports while users are logged in?
    - If users are logged into the live version of our company but I am testing DTW/excel imports from a TEST copy of that company are there still risks or is this okay?
    - If users are logged into one database but I am performing data imports on a different company database (for example I would be the only one logged into the companies currently being set up - is this okay even though users are logged into SAP server but accessing a different company?)
    Thanks very much!
    Richard

    Hi Kiwi,
    Referring to the Rules of enagement you have to post one thread per question. Please adhere to the rules from your next post.
    DTW is the very much Preferred method you have simulation to see if you record are find to be imported before importing to the DB, using DTW Import from Excel will only help importing  BP and Item master data. If you have to make any updates then it will be a task again. In DTW you can save the template and keep adding data and changes can be updated from the template and re-imported
    Please refer to the link below to "Differences between the Import From Excel and Data Transfer Workbench templates" the screen shots for DTW are from version before 2007A SP01 PL09.
    [http://wiki.sdn.sap.com/wiki/display/B1/DifferencesbetweentheImportFromExcelandDataTransferWorkbenchtemplatesandwhathappenswhentheyare+interchanged.]
    1. When users are logged on It can cause some inconsistency, ensure no user is logged on the live DB before performing any major DTW import
    2. It will be ok if you are Test importing data to test DB on the same SQL server.
    Regards,
    Rakesh N

  • Data warehousing question/best practices

    I have been given the task of copying a few tables from our production database to a data warehousing database on a once-a-day (overnight) basis. The number of tables will grow over time; currently it is 10. I am interested in not only task success but also best practices. Here's what I've come up with:
    1) drop the table in the destination database.
    2) re-create the destination table from the script provided by SQL Developer when you click on the 'SQL' tab while you're viewing the table.
    3) INSERT INTO the destination table from the source table using a database link. Note: I am not aware of any columns in the tables themselves which could be used to filter added/deleted/modified rows only.
    4) After data import, create primary key and indexes.
    Questions:
    1) SQL Developer included the following lines when generating the table creation script:
    <table creation DDL commands>
    then
    PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING
    STORAGE (INITIAL 251658240 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    TABLESPACE "TBLSPC_PGROW"
    it generated this code snippet for the table, the primary key and every index.
    Is this necessary to include in my code if they are all default values? For example, one of the indexes gets scripted as follows:
    CREATE INDEX "XYZ"."PATIENT_INDEX" ON "XYZ"."PATIENT" ("Patient")
    -- do I need the following four lines?
    PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS
    STORAGE(INITIAL 60817408 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    TABLESPACE "TBLSPC_IGROW"
    2) Anyone with advice on best practices for warehousing data like this, I am very willing to learn from your experience.
    Thanks in advance,
    Carl

    I would strongly suggest not dropping and recreating tables every day.
    The simplest option would be to create a materialized view on the destination database that queries the source database and to do a nightly refresh of that materialized view. You could then create a materialized view log on the source table and then do an incremental refresh of the materialized view.
    You can schedule the refresh of the materialized view either in the materialized view definition, as a separate job, or by creating a refresh group and adding one or more materialized views.
    Justin

  • Important Questions to ask when developing in Workflow

    Hello All,
      I have two questions.
      I am not a Workflow developer, but I've been asked to find out information about creating a Workflow process.
    Here is what <b>I</b> need to know: What are important questions to ask the end-user when creating a workflow. What needs to be taken into consideration?
    Here is an example of the kind of thing we need to do:
    Send an email on a weekly basis to different Project Managers a Summary of Project Charges = (maybe CN41?)
    Each Project Manager will have different requirements, so  I guess one question will be: "What info do you want from this report?". What else do I need to know?
    My second question is this: I should be able to output this as an Excel, correct? How much can I customize this? I haven't done much SAP -> Excel work. Only when I go to System -> List -> Save -> Local File ->Spreadsheet.
    Thanks for the help.
    Don Fields

    Hello Don!
    WF is quite a complex topic and someone could write a book to answer your questions - actually there is already a good book available -> check: Practical Workflow for SAP
    http://www.sap-press.com/product.cfm?account=&product=H950
    Looking at your example their might not even be a need for a workflow but rather a job that runs periodically to send the required information via email to project managers.
    Workflow is mostly used if user actions are required e.g. confirm, approve, reject a request or notifications need to be sent to someone after an event or user action occured.
    Check this link for getting more information on how to download data to excel:
    http://www.sapdevelopment.co.uk/reporting/email/attach_xls.htm
    Harald

  • Data-importing functionality

    Hi,
    want some help to know if there are any functionality existing to use the data importing feature which is now available as a part of html DB spreadsheet import ,in the html DB developed appliation.
    Thanks,
    Vidhya.s

    hi vidyha--
    the short answer to your question is, "not yet", but you can find the more complete explanation at the bottom of...
    Bulk Insert, SQL*Loader style
    ...regards,
    raj

  • UNDOTBS during Data Import (datapump)

    Hi,
    I want to know how much space does the data import occupies on the undotbs tablespace? will it consume the same exact space (size) of the table being imported? I am using impdp for importing data. (data_only).
    Also, i want to know when will ORACLE use UNDOTBS and TEMP tbs.
    Your assistance would be greatly appreciated!
    Thanks!!
    Christian George Medija

    IMPDP just does the plain inserts that is what makes datapump platform independant. So if it just doing plain insert , it will definitely use Undo Tablespace. Now the question is how much?
    Impdp consume more undo tablespace because it has to run some additional metadata and some master table queries during import.BUt if you are just importing using parameter data_only, you might not go above your table size or might go little above that. For example, if you have 5GB of table, you might go 5.2 GB.
    Temporary tablespace: Impdp definitely use Temporay tablespace at lot of places while importing. One is when creating indexes, it has to do a sorting which uses you temporary tablespace.

  • 3 important questions about SAP HANA

    Hello Community!
    I have got 3 important questions about SAP HANA.
    1.
    - Our Company has all up about 150 SAP - Systems.
    - 12 Systems are ERP-Systems.
    => We consider about SAP HANA at the moment.
    => Is it possible to put all the 12 ERP-Systems in ONE SAP HANA or do I need one seperate HANA for each ERP-System?
    => If it is possible, can you tell me the disadvantages when I use 1 HANA for the 12 ERP-Systems.
    => Is there a Storage-Limit in the HANA (I think no, because of 64 bit).
    => Can I still make realtime-analytics with 1 HANA for 12 SAP Systems?
    2.
    => Can I run ONE HANA with different non-SAP-Systems?
    3.
    => Is there a Demo-Center from SAP or IBM, where I can test and see the HANA before I buy it?
         (I do not mean the Fujitsu TDS Demo Center)
    It would be nice, when you would answer all my questions.
    Thank you.

    I have got 3 important questions about SAP HANA.
    1.
    - Our Company has all up about 150 SAP - Systems.
    - 12 Systems are ERP-Systems.
    => We consider about SAP HANA at the moment.
    => Is it possible to put all the 12 ERP-Systems in ONE SAP HANA or do I need one seperate HANA for each ERP-System? Yes it is quite possible to put all the ERP/CRM/SCM or other components on one HANA. You need to check for pricing/users/RAM (storage)
    => If it is possible, can you tell me the disadvantages when I use 1 HANA for the 12 ERP-Systems.
    There are no disadvantages as of now.. improve existing planning, forecasting, pricing
    optimization and other processes
    => Is there a Storage-Limit in the HANA (I think no, because of 64 bit).
    There is no limit on Stotrage: You need check how much Data you need on RAM (Real time processing) Storage.Minimum you can buy is 64 GB and multiple of it for (RAM)
    => Can I still make realtime-analytics with 1 HANA for 12 SAP Systems?
    Yes certainly..
    2.
    => Can I run ONE HANA with different non-SAP-Systems?
    YEs you can run HANA for 3rd party S/W or non-sap systems for your real time anlytics.
    It has ODBC and JDBC both and you can combine it with other systems.
    3.
    => Is there a Demo-Center from SAP or IBM, where I can test and see the HANA before I buy it?
    (I do not mean the Fujitsu TDS Demo Center)
    This you need to check with your local SAP Service channel..
    It would be nice, when you would answer all my questions.
    Thank you.
    PY

  • Data archieving  questioner required

    Dear All,
    We have been approched by one of our cleint for DATA ARCHIVING from R/3 system.
    Management has pushed my name in this.
    Requirement is to prepare DATA ARCHIVING QUESTIONER template.
    Can please anybody help me out in this regard from MM point of view.
    Thanking you in advance.
    Regards
    Nasir Chapparband.

    Hi,
    Refer following link;
    [http://itmanagement.earthweb.com/datbus/article.php/3109221]
    SAP Data Archiving
    1.0 Introduction to Enterprise Data Archiving
    Currently, a large number of enterprises use SAP R/3 as a platform for integration of business processes. The continuous usage of SAP results in huge amounts of enterprise data, which is stored in SAP R/3. With passage of time, the new and updated data is entered into the system while the old data still resides in the SAP enterprise system.
    Since some of the old data is critical, it cannot be deleted. The difficulty is keeping the data you want, and deleting the data you do not want. Hence, a SAP database keeps on expanding rapidly and enterprise systems, which have limited data retention abilities for a few years, suffer from problems such as data overflow, longer transaction processing times, and performance degradation.
    The solution of this problem has led to the concept of Data Archiving in SAP. Data Archiving removes out-of-date data from the SAP database that the R/3 system does not need online, but can be retrieved on a later date, if required. This data is known as archived data and is stored at an offline location. Data Archiving not only consistently removes data from the database but also ensures data availability for future business requirements.
    One rule of thumb is that in a typical SAP enterprise system, the ratio of data required to be online and instantly accessible to old data, which could be archived, and stored offline is 1:6. For example, if an enterprise has 2100 GB of SAP database, the online data, which is frequently used by SAP users will be 300 MB and the rest (1800 MB) will be scarcely used and hence can be archived.
    1.1 Data Archiving u2013 Features
    It provides a protection layer to the SAP database and resolves underperformance problems caused by huge volumes of data. It is important that SAP users should keep only minimal data to efficiently work with database and servers. Data archiving ensures that the SAP database contains only relevant and up-to-date data that meet your requirements.
    Data archiving uses hardware components such as hard disks and memory. For efficient data archiving, minimum number of disks and disk space should be used.
    It also reduces the system maintenance costs associated with the SAP database. In the SAP database there are various procedures such as, data backup, data recovery, and data upgrade.
    SAP data archiving complies with statutory data retention rules that are common and well-proven techniques.
    SAP data archiving can be implemented in two ways. In the next section both options will be discussed in detail.
    Also refer following link;
    [SAP Data Archiving Tutorial|http://www.thespot4sap.com/articles/SAP_Data_Archiving_Overview.asp]

  • C100 Data Import Utility - download where?

    Hi there!
    I lost the DVD that came with my C100. Can I download the Data Import Utility somewhere? I can't seem to find it anywhere on Canon's pages...

    Hi o-shows!
    Thanks for posting.
    The Data Import Utility is licensed software provided by the Pixela Corporation.  They have it available on their website for download.  Clicking HERE will take you to the download page.  There is a download link at the bottom of the page you can click after reading over the instructions.
    This didn't answer your question or issue? Find more help at Contact Us.
    Did this answer your question? Please click the Accept as Solution button so that others may find the answer as well.

  • Data import from EBS failed via FDMEE in fdm . Getting error message as "Error connecting to AIF URL.

    FDM Data import from EBS failed via FDMEE after roll back the 11.1.2.3.500 patch . Getting below error message in ERPI Adapter log.
    *** clsGetFinData.fExecuteDataRule @ 2/18/2015 5:36:17 AM ***
    PeriodKey = 5/31/2013 12:00:00 AM
    PriorPeriodKey = 4/30/2013 12:00:00 AM
    Rule Name = 6001
    Execution Mode = FULLREFRESH
    System.Runtime.InteropServices.COMException (0x80040209): Error connecting to AIF URL.
    at Oracle.Erpi.ErpiFdmCommon.ExecuteRule(String userName, String ssoToken, String ruleName, String executionMode, String priorPeriodKey, String periodKey, String& loadId)
    at fdmERPIfinE1.clsGetFinData.fExecuteDataRule(String strERPIUserID, String strDataRuleName, String strExecutionMode, String strPeriodKey, String strPriorPeriodKey)
    Any help Please?
    Thanks

    Hi
    Getting this error in ErpiIntergrator0.log . ODI session ID were not generated in ODI / FDMEE. If I import from FDMEE its importing data from EBS.
    <[ServletContext@809342788[app:AIF module:aif path:/aif spec-version:2.5 version:11.1.2.0]] Servlet failed with Exception
    java.lang.RuntimeException
    at com.hyperion.aif.servlet.FDMRuleServlet.doPost(FDMRuleServlet.java:76)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
    at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
    at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:301)
    at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:27)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:57)
    at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:119)
    at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:324)
    at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:460)
    at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)
    at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)
    at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:57)
    at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:163)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:57)
    at weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:27)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:57)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3730)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3696)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
    at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2273)
    at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2179)
    at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1490)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:256)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:221)

  • Data Import ERROR

    Hi Guys
    I am facing an error when uploading the items through Data Import.
    Just i am uplaoding only Item Code and Item Description .
    but error is coming like this " Row Number 1: Invalid Code [Items - Chapter ID] "
    Also when uploading the BP master data
    i am uploading BP name and Addresses like street, block, state and country. in state i am giving delhi
    but error is coming like this " Business Partner Row Number 6: Character number is greater than allowed [Business Partner - Bill-to State] , 'Delhi' "
    Pls help me on this.
    Regards
    Giridharan V

    hi
    i tried to upload the data's through DTW. The thing i did
    RecordKey     ItemCode  ItemName
    RecordKey     ItemCode          ItemName
    1                       i011                 a
    2                       i002                  b
    3                       i003                    c
    in the ".csv " format..
    Bur error is coming like this
    " Invalid code (OITM_ Chapter ID ) Application defined or object degfined error 5171 "
    Edited by: Giri dharan on Jul 4, 2008 10:33 AM

Maybe you are looking for

  • Voice memos gone after ios 7 upgrade

    The voice memos on my phone are gone following the update to IOS 7.  Does anyone know why they disappeared and how I can get them back on the phone.  I have an iphone 5.  I backed up before the update but I do not see the voice memos in my itunes lib

  • How can I send an audio file via mms that isn't a MEMO?

    Can someone please help? Am i missing something here? i have an Iphone 4 and I cannot find how to do this.... Any help greatly appreciated, thanks.

  • Short dump with Where-used list

    Hi All, I'm making some research on specific InfoObjects, and in some cases I want to know where are they being used, but when I try to use the icon for "Where-used list using Data Warehouse" I end up with an error, a short dump. Message class is: RS

  • Sum in where clause

    can i use sum in where clause eg select from table_name id_klient,name,lastname,sum(number_of_products) from Klients k,Sales s where k.id_klient=s.id_klient and sum(number_of_products)>100

  • Report (sales order not delibered-not reserved)

    Hello to everybody! Hoe it's going after hollidays? I need some help to know if there is any standard option to get a report in which we obtain,per material, the sales orders which haven't been still delivered and which have not any reservation assoc