Data Type for Process Flow... PB with Date?

I've got a problem by passing parameters in process flow.
I have a mapping with a parameter DATE_EXEC (data type : DATE) and a default value that is TO_DATE('20/01/2007' , 'dd/mm/yyyy') . My mapping is working good when i launch it.
I have a process flow which contains the mapping. This process has a parameter DATE_EXEC (data type : DATE). I bind the 2 DATE_EXEC. But when i launch my mapping the value is not recognized, I try with :
- TO_DATE('20/01/2007' , 'dd/mm/yyyy')
- 20/01/2007
- 2007.01.20
- 2007-01-20
My question is what are the data type in process flow? They are not ORACLE TYPE.
For example , a parameter in a mapping which is a VARCHAR2 must be input between quotes but if you bind it to a parameter of a process flow which is a STRING (not ORACLE Data type) , you must input it without quotes?
Anybody has some rules about that?
I apologize for my english, i'm a french people.

Here is some information on the literal quote or not quote query and what I think you need to do at the end, hope it helps. Not exactly intuitive...since the flow designer (you) have to know what is a PLSQL object and what is not.
1. Literal = FALSE
When Literal = FALSE is set then the value entered must be a valid PL/SQL expression which is evaluated at the Control Center e.g.
'Hello World!'
22 / 7
2. Literal = TRUE
When Literal = TRUE then the value is dependent on the the type of Activity. If the activity is a PL/SQL object i.e. Mapping or Transformation, then the value is PL/SQL snippet. The critical difference here is that the value is macro substituted into the call for the object. The format of the value is identical to that entered as default value in the Mapping editor. e.g.
'Hello World!'
sysdate()
If the activity type is not a PL/SQL object then the value is language independent. e.g.
Hello World
3.1427571
What you should try......
Check the map activity parameter in your process flow to see if literal is false (an expression), set it to false and then try using your TO_DATE('20/01/2007' , 'dd/mm/yyyy') expression, deploy your flow and execute. Alternatively the user guide defines the DATE type for flow with the format YYYY-MM-DD so you can have the parameter value as '2007-01-20' use literal equal to true and remember and quote your value.
Cheers
David

Similar Messages

  • J1INUT Error - No data exist for processing with the given selection option

    Hi Guru's,
    I am using transaction J1INUT utilization of provision of TDS on Services for which in have made the provision with the help of J1INPR, But when I am executing J1INUT transaction .The following error message is displaying:
    No data exist for processing with the given selection options
    I have followed the below steps.
    1) ME21N - OP Creation
    2) ML81N - Service Entry
    3) J1INPR - Provision of TDS
    4) MIRO -  Invoice Posting
    I have checked the Table J_1IEWTPROV In that system is updating the table also. Even I have activated table TRWCA for field IND
    But still I am getting the same error. Any suggestions to resolve this.
    Appreciate your inputs. Thanks in Advance
    Regards,
    DeepaK

    Hi Deepak,
    Refer the below link and follow the steps - Provision for Taxes on Service Recieved.
    Re: Provision for Taxes on Service Recieved
    Hope it may useful to you.
    Regards,
    Govind Bhaskaran

  • Proper Data type for transferring XML from ABAP to XI

    Dear Colleagues,
    I have a question on how to wrap XML data, that itself contains an encoding
    declaration, so that it will be properly transferred to XI.
    XML Generation: "always use X"
    I have searched a lot on SDN and other resources but could not find
    information, describing how a web service should wrap unicode-data when
    making it available to XI. The XML processing group emphasizes that all generation of XML should be done using X-bases (HEX) datatypes in ABAP and preferably using UTF-8 as ancoding. This by itself is not a problem.
    But transferring UTF-8-XML - through which interface?
    When a webservice provides an XML Stream and wants to transfer it to XI then there is a bit of a paradox situation, because it has to choose between using an X-based or Character-based type for its interface
    -> as Binary data?
    - when using X-based data types in the interface, I have control of the character encoding, but I expect that the XML-information will probably be wrapped in base64 encoding and then the XI will receive a base64 encoded
    XML-block instead of "plain" XML.
    I did not find any reference information on how the soap interface of XI and the WAS behave when transferring X-based data, however.
    -> as Character data?
    - on the other hand when using Character-based data types (e.g. String), I have no control of the codepage being used. I have then only the choice to generate the default encoding of the application server. Maybe this encoding
    will be transformed later to a different encoding, but usually during this process the header, which declares the encoding in use will not be adapted as well.
    So if I have a unicode machine, it will generate UTF-16 - based XML. I have no other choice, if I use a character-based type for output. And this will also be stated in the header of the XML stream. If the XML stream
    leaves the system it will probably be transformed to UTF-8, but the XML-header will still state the content to be UTF-16, because the header will probably not be auto-converted together while adapting the Character
    encoding.
    Because of these issues I am quite insecure, on how to generate and wrap UTF-8 - based XML, so that I can properly hand it over to XI, ensuring that
    it stays UTF-8 and that XI can really unwrap it automatically.
    Does anyone have a tip in which type to wrap XML-based data on the ABAP stack so it will be usable in XI?
    Thank you very much for your attention and time.
    Yours
    Christian Steinert

    Something like this should work (with a little tweek I imagine)
    This would be in the 'Command on Target':
    OdiFtpGet "-HOST=<%=odiRef.getInfo("SRC_DSERV_NAME")%>" "-USER= '<%=odiRef.getInfo("SRC_USER_NAME")%>'" "-PASSWORD= <%=odiRef.getInfo("SRC_PASS")%>" "-REMOTE_DIR=<%=odiRef.getInfo("SRC_SCHEMA")%>" "-REMOTE_FILE=<%=odiRef.getTable("L", "COLL_NAME", "W")%>" "-PASSIVE_MODE=YES"
    Set the Technology of the 'Command on Source' as File, and the technology of the 'Command on Target' as ODI Tools
    Edited by: Bos on Jun 30, 2011 11:53 AM

  • IU Elim. "No data found for processing using current selection conditions""

    Dear Experts,
    While Executing task of Interunit elimination  in Consolidation Montior  I am getting Message "No data found for processing using current selection conditions"
    Ex. is
    A)
    In Unit X
    GL (399999) Account     Dr. 65000 (Customer Recon. Ac.)   (with Trading Parter X)
    GL (499999) Rev.A/c           65000 (with Trading Parter X)
    In Unit Y
    GL (199999) Exp. A/c     Dr. 65000   (with Trading Parter Y)
    GL (299999).Account           65000 (Vendor Recon. Ac.) (with Trading Parter Y)
    B) GLs in info cube in 0FIGL_C01 are :-
    GL Account---CCode--Trading PartnerDebit--
    Credit
    199999--YY65000-----00000
    299999--YY00000-----65000
    399999--XX65000-----00000
    499999--XX00000-----65000
    In COnsolidation WorkBench
    1) I have created Document Type
    2) Method-
      In  General Tab
       a) Two SIded Selection
        b) Per Transaction Currency Selected
    In Selection Tab
    1St Selection
    GL Account = 299999 (Customer Recon. Account)
    Company    = X
    Trading Partner = X
    1St Selection
    GL Account = 399999 (Vendor Recon. Account)
    Company    = Y
    Trading Partner = Y
    Difference Tab
    a) Post Diff to "Unit from Selection 1"
    b) Key Figure "Period Value GC
    c) Check Limit Per Difference Row
    Other Differnce
    GL Account = 100099 (Other GL)
    Currencyce Diff
    GL Account = 100510 (Other GL)
    Question :-
    1) Is the posting is appropriate and does it attracts IU Elimination?
    2) The Infocube Details are correct?
    3) Any Config issue
    May any one suggest, Why Am I not able to get the data?
    Thanks
    Rakesh Shrivastav

    Dear Sir,
    Following are the View at my end in context to your suggestion
    1. check the BCS totals data to ensure trading partner is included.
    THis is the view of Source Info Cube 0FIGL_C01
    GL Account---CCode--Trading PartnerDebit--
    Credit
    199999--YY65000-----00000
    299999--YY00000-----65000
    399999--XX65000-----00000
    499999--XX00000-----65000
    2. Execute the task for the cons group that includes both cons units X and Y
    The COns Group Is XYZ
    X- Cons Unit
    Y- Cons Unit
    In Cons Monitor I am executing Test run at XYZ level
    3. Although the trading partner for cons unit X should be Y and vice versa, the elimination should still occur with the cons unit X and trading partner X records.
    Same as the query description
    4. make sure that the items 199999, 299999, 399999 and 499999 are included in the elimination method for either selection 1 or selection 2.
    The Method Selection Tab View is
    1St Selection
    GL Account = 299999,199999
    Company = Y
    TP = Y
    2Nd Selection
    GL Account = 399999,499999
    Company = X
    TP = X
    What is your View On That?

  • Query regarding the data type for fetcing records from multiple ODS tables

    hey guys;
    i have a query regarding the data type for fetcing records from multiple ODS tables.
    if i have 2 table with a same column name then in the datatype under parent row node i cant add 2 nodes with the same name.
    can any one help with some suggestion.

    Hi Mudit,
    One option would be to go as mentioned by Padamja , prefxing the table name to the column name or another would be to use the AS keyoword in your SQL statement.
    AS is used to rename the column name when data is being selected from your DB.
    So, the query  Select ename as empname from emptable will return the data with column name as empname.
    Regards,
    Bhavesh

  • Analysing Task Audit, Data Audit and Process Flow History

    Hi,
    Internal Audit dept has requested a bunch of information, that we need to compile from Task Audit, Data Audit and Process Flow History logs. We do have all the info available, however not in a format that allows proper "reporting" of log information. What is the best way to handle HFM logs so that we can quickly filter and export required audit information?
    We do have housekeeping in place, so the logs are partial "live" db tables, and partial purged tables that were exported to Excel to archive the historical log info.
    Many Thanks.

    I thought I posted this Friday, but I just noticed I never hit the 'Post Message Button', ha ha.
    This info below will help you translate some of the information in the tables, etc. You could report on it from the Audit tables directly or move them to another appropriate data table for analysis later. The concensus, though I disagree, is that you will suffer performance issues if your audit tables get too big, so you want to move them periodically. You can do this using a scheduled Task, manual process, etc.
    I personally just dump it to another table and report on it from there. As mentioned above, you'll need to translate some of the information as it is not 'human readable' in the database.
    For instance, if I wanted to pull Metadata Load, Rules Load, Member List load, you could run a query like this. (NOTE: strAppName should be equal to the name of your application .... )
    The main tricks to know at least for task audit table are figuring out how to convert times and determing which activity code corresponds to the user friendly name.
    -- Declare working variables --
    declare @dtStartDate as nvarchar(20)
    declare @dtEndDate as nvarchar(20)
    declare @strAppName as nvarchar(20)
    declare @strSQL as nvarchar(4000)
    -- Initialize working variables --
    set @dtStartDate = '1/1/2012'
    set @dtEndDate = '8/31/2012'
    set @strAppName = 'YourAppNameHere'
    --Get Rules Load, Metadata, Member List
    set @strSQL = '
    select sUserName as "User", ''Rules Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
          cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
       from ' + @strAppName + '_task_audit ta, hsv_activity_users au
       where au.lUserID = ta.ActivityUserID and activitycode in (1)
            and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + '''
    union all
    select sUserName as "User", ''Metadata Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
          cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
       from ' + @strAppName + '_task_audit ta, hsv_activity_users au
       where au.lUserID = ta.ActivityUserID and activitycode in (21)
            and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + '''
    union all
    select sUserName as "User", ''Memberlist Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
          cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
       from ' + @strAppName + '_task_audit ta, hsv_activity_users au
       where au.lUserID = ta.ActivityUserID and activitycode in (23)
            and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + ''''
    exec sp_executesql @strSQLIn regards to activity codes, here's a quick breakdown on those ....
    ActivityID     ActivityName
    0     Idle
    1     Rules Load
    2     Rules Scan
    3     Rules Extract
    4     Consolidation
    5     Chart Logic
    6     Translation
    7     Custom Logic
    8     Allocate
    9     Data Load
    10     Data Extract
    11     Data Extract via HAL
    12     Data Entry
    13     Data Retrieval
    14     Data Clear
    15     Data Copy
    16     Journal Entry
    17     Journal Retrieval
    18     Journal Posting
    19     Journal Unposting
    20     Journal Template Entry
    21     Metadata Load
    22     Metadata Extract
    23     Member List Load
    24     Member List Scan
    25     Member List Extract
    26     Security Load
    27     Security Scan
    28     Security Extract
    29     Logon
    30     Logon Failure
    31     Logoff
    32     External
    33     Metadata Scan
    34     Data Scan
    35     Extended Analytics Export
    36     Extended Analytics Schema Delete
    37     Transactions Load
    38     Transactions Extract
    39     Document Attachments
    40     Document Detachments
    41     Create Transactions
    42     Edit Transactions
    43     Delete Transactions
    44     Post Transactions
    45     Unpost Transactions
    46     Delete Invalid Records
    47     Data Audit Purged
    48     Task Audit Purged
    49     Post All Transactions
    50     Unpost All Transactions
    51     Delete All Transactions
    52     Unmatch All Transactions
    53     Auto Match by ID
    54     Auto Match by Account
    55     Intercompany Matching Report by ID
    56     Intercompany Matching Report by Acct
    57     Intercompany Transaction Report
    58     Manual Match
    59     Unmatch Selected
    60     Manage IC Periods
    61     Lock/Unlock IC Entities
    62     Manage IC Reason Codes
    63     Null

  • Data Type for a Longitude/Latitude Column in an Oracle dB Table?

    Greetings All,
    What is the recommended data type for a Longitude(Lon) or Latitude(Lat) Column of an Oracle dB Table? I am creating a signs inventory table in our Oracle database to be used as a data warehouse for our signs inventory. The Lon and Lat fields along with some other data will be populated from a MS access file exported from our Geodatabase. Currently the data type for the Lon and Lat fields of the MS access file is defined as "Double". A couple of Examples of the Lon and Lat data are as follows:
    Lat                       Lon
    59.4564555909     -135.3133542922
    64.841125            -148.01808
    ...What data type should these two columns be defined in the new Oracle database table??? Should the data type be defined as FLOAT(126), NUMBER(10, 6), or else? Any suggestions/help on this would be greatly appreciated.
    If the data type is defined as NUMBER(10,6), the first example of Lon "-135.3133542922" will be truncated to "-135.313354" or would I receive an error while trying to import the data from an access file to the Oracle dB?
    Thanks in advance for any/all the help.

    Orchid,
    Is SDO_GEOMETRY available in Oracle 10g?Yup. I'm still on 10g as well; there's no shame in that.
    How do I format my lat/long's into WKT and use the SDO_GEOMETRY constructor?The March 16th example shows the steps, but I take it you're trying to get your data from MS Access into Oracle. Is that right? If so, consider these options:
    OPTION 1, When you have < 50k records
    Simply write a query that returns INSERT statements that you can run against Oracle.
    For example, let's say you have a table in MS Access defined as msaCoords (ID, LongX, LatY).
    Let's say, your destination table in Oracle is named oraCoords(ID NUMBER, Geometry MDSYS.SDO_GEOMETRY). Then,
    SELECT "INSERT INTO oraCoords (ID, Geometry) VALUES (" & ID & ", SDO_GEOMETRY('POINT (" & LongX & "," & LatY & "'), 8307);" FROM msaCoords;
    -- "MS-ACCESS QUERY RESULTS"
    INSERT INTO oraCoord (ID, Geometry) VALUES (1, SDO_GEOMETRY('POINT (-135.3133542922,59.4564555909'), 8307);
    INSERT INTO oraCoord (ID, Geometry) VALUES (2, SDO_GEOMETRY('POINT (-148.01808,64.841125'), 8307);Save those MS-ACCESS QUERY RESULTS to a text-file (D:\MyPts.sql). Launch SqlPlus and type @D:\MyPts.sql
    Once it's done, type Commit into SqlPlus and you're done.
    OPTION 2, when you have > 50k records
    Use SqlLoader. There's an example posted { here | http://forums.oracle.com/forums/thread.jspa?messageID=9412123&#9412123 }
    Cheers,
    Noel

  • How to create the Data Type for the receiver end?

    Hello Expertise
    I wanted to Create a Data type for  Receiver end in PI.
    Sender end is ECC.
    Receiver end is SRM
    Here's the Data type for Sender end  as
    Can anyone help to create the Data type for Receiver end i.e SRM.
    Thanks in advance.
    Regards
    karan

    Hi Karan,
    Please don't call Sir, we are just trying to help each other.
    Please go through below scenario's and check your configuration.
    ABAP Client Proxy to ABAP Server Proxy Scenario - Process Integration - SCN Wiki
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/e0ac1a33-debf-2c10-45bf-fb19f6e15649?overridelayout=true
    http://wiki.scn.sap.com/wiki/display/XI/Step-by-step+FTP+to+ABAP+Proxy
    regards,
    Harish

  • With 10.7.2 update iCal does not allow you to set "Reminders" by draging events from "All Day Events" into "Reminders" is there another way to do this instead of having to type out the entire reminder with date, time, and type?

    With 10.7.2 update iCal does not allow you to set "Reminders" by draging events from "All Day Events" into "Reminders" is there another way to do this instead of having to type out the entire reminder with date, time, and type?
    With Lion 10.7.1. you where able to drag events from "All Day Events" into the "Reminders" bar to create upcoming reminders.

    Exactly the same question I was about to post!
    Great being able to sync reminders (well overdue) however if I have to re-type a calener envent into reminders it's a waste of time!! 
    Come on Apple!!
    Just need an option for the calender event to add to reminder or the old drag to add to reminder functionality back, Please???

  • Naive Bayes Training - CLOB as output data type for JSON string?

    Hello everyone,
    My training model outputs a large JSON string that doesn't fit into one row, so the string is split across multiple rows. Default - or only possible output data type for that matter - is varchar according to the official documentation on PAL. Is there any chance I could use CLOB for the output?
    Regards
    Henry

    The property you provided is for applying an NB model, not for building one.
    For build, you should have used Sample_NaiveBayesBuild property file.
    I ran the NB apply sample program with your property, and was able to get past the point where you have problem.
    Please double check if you really used the property you provided and if it matches to the program you wanted to use.

  • Data Type for File adapter

    hi all,
    we have a scenario which is idoc-xi-file
    the output data type should be something like this:
    [Data.1]
    Promo,1000,,,,,,,
    PromoD,2000,,yup
    Item,1,IGRP,,,,,,FTY,3
    Tier,1,,,,,,,ITEM,,
    Itemy,1,IGRP,,,,,,FTY,1,A_,100,,ELST
    we have to use file content conversion parameters for this as well as this needs to be converted to a CSV file format(third party system).
    any help in declaring this kind of data type(structure) and handling file content conversion parameters would be highly appreciated.
    for data type if any anybody can help me build the structure or give an example!!!
    would be gr8!!!
    thanks in advance
    ahmed

    Hi Parvez,
    For file content conversion, you can refer the below link:-
    http://help.sap.com/saphelp_nw04/helpdata/en/d2/bab440c97f3716e10000000a155106/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/d2/bab440c97f3716e10000000a155106/frameset.htm
    I hope this helps.
    Regards.
    Praveen

  • Labview How to specify 1d array of clusters as data types for variant to data

    Hi, I'm new to labview. Can anyone tell me how to specify 1d array of clusters as data types for variant to data?

    First of all, you should be sure that there is such a data type within the variant; otherwise, you will run into errors.
    I recommend you to create the cluster and create a type definition from it. Then drop an array shell from the array palette and drop the cluster type into that array.
    Connect that constant to the data type input of the Variant To Data function.
    Norbert
    CEO: What exactly is stopping us from doing this?
    Expert: Geometry
    Marketing Manager: Just ignore it.

  • How to change data type for a field?

    Hi Experts,
    I have one std screen having one field having numeric type. So it will accept only numeric value. Remember I am using standard transaction.
    Now my user wants to enter the special character like + or - in that field(having data type numeric).
    How i can change the data type for the particular field  so that it will also accept + or -?
    Plz reply me as earliest as possible.
    Thanks in advance.

    Hi salil ,
    u have to change / create Domain of  that Data Element.
    but u have to check the Implications. i have done the same thing , but that one is Customer defined one , i got nearly 20 short dumps while doing the TP.
    regards
    Prabhu

  • Default date type for IT0041 in action

    Hi,
    I configure feature DATAR to set 01 as default date type for IT0041, but when I do hiring action, default date type will be 01,02,12.
    Don't know where to change the default.

    Hi,
    Please verify if there is any dynamic action in table V_T588Z set-up for same which may cause override over default.
    Also, verify if your DATAR feature is set up correctly for relevant country grouping and activated.
    Thanks,
    Ameet

  • How  create data type for coulmn name having special characters in IR

    hi ,
    any idea how to create how to do create data type for coulmn name having special characters

    Hi Amit,
    Can you ellaborate ?
    Actually if you are reading a file(or anything else) which contains special characters, the data type can be created as done for other interfaces. But it is in the adapters that we actually take care of handling these special characters.
    ShabZ

Maybe you are looking for

  • Helping w/printing issues

    Hi, I have some files that were recently sent to a professional printer.. I used Illustrator CS & InDesign CS to design my work... This is what the printer said (I have no idea what this means I've never had this issue before). "We need you to set al

  • How can I record a cassette on CD?  New 27 inch iMac does not have mini plug/mic input

    I want to transfer music from a cassette to a CD using my new iMac but the iMac does not have a mini plug input for a mic.  I understand that I need to get an adaptor to use a USB port.  I bought a Syba SD-CM-UAUD USB Stereo Audio Adaptor and tried t

  • Looking for a FM to download entirely PO/CTR data from R/3

    Hi all gurus, as told in subject, I'm looking for a way to dowload all the data of a contract or a purchase order into a local file on my client. For those who are familiar with SRM, there's a transaction, BBP_PD, which - given a contract or po numbe

  • Business system from SLD not visible in Integration Directory

    Hi, As on the project I do not have access to the SLD, another team created new business systems there. However, when I try in the Integration Builder(Configuration) to assign the newly created systems, I cannot see them on the "Select business syste

  • Embed a jsf page in a jsp page -

    Is there some way to included a jsf page inside a jsp page and code form action correctly ? In my JSP page I have the following: <jsp:include page="../user/unMapTrans.faces" /> The page loads correctly, however, when I submit the form I get "Cannot f