Changes to way Adobe reports data

Hello,
I'm using Adobe Acrobat 7.0 to collect data  from a Dynamic .pdf that I created in Adobe Designer 7.0.  When I  initiate the data file collection workflow, the resulting excel  spreadsheet reports the output in the following format, which I realize is the "hierarchy" structure of the document:
F[0].P#[0].x#[0]
What values could I change in Designer for a check mark box, a radio button, or a text box that the user fills in to make it easier to comprehend which question they are looking at?
Thank you for any help you can give me.

Please ask in the Designer forum: http://forums.adobe.com/community/livecycle/livecycle_previous_versions

Similar Messages

  • [svn] 4607: Change the way we handle setItemAt in DataGroup.. we now do it synchronously again.

    Revision: 4607
    Author: [email protected]
    Date: 2009-01-21 10:45:20 -0800 (Wed, 21 Jan 2009)
    Log Message:
    Change the way we handle setItemAt in DataGroup..we now do it synchronously again. However, this has the consequence that, even temporarily, you are not allowed to have the save visual element listed twice in the data collection.
    Get parity with Halo behavior and Flash behavior by allowing addElement()/addElementAt() to be called on an element already in the list. When this happens just shift the indices around.
    Small ASDoc fix in GroupBase. Also, change the way we handle data being set int he default complex item renderer...this way we're not using binding, and the code is also easier to understand.
    QE Notes: Write tests for addElement/addElementAt on items already in the group.
    Doc Notes: None
    Bugs: SDK-17909, SDK-18189
    Reviewer: Hans
    tests: checkintests, mustella DataGroup, List, Group, FxContainer, FxDataContainer
    Ticket Links:
    http://bugs.adobe.com/jira/browse/SDK-17909
    http://bugs.adobe.com/jira/browse/SDK-18189
    Modified Paths:
    flex/sdk/trunk/frameworks/projects/flex4/src/mx/components/DataGroup.as
    flex/sdk/trunk/frameworks/projects/flex4/src/mx/components/Group.as
    flex/sdk/trunk/frameworks/projects/flex4/src/mx/components/baseClasses/GroupBase.as
    flex/sdk/trunk/frameworks/projects/flex4/src/mx/skins/spark/FxDefaultComplexItemRenderer. mxml
    flex/sdk/trunk/frameworks/projects/wireframe/src/wireframe/FxDefaultComplexItemRenderer.m xml

    Hi blabla12345,
    (untested and without warranty)
    replace this line:
    const sSaveCUBE = "CUBE";
    with this:
    const sSaveCUBE = "cube";
    Have fun

  • Change default date format of reports data source

    Hi all,
    I need to convert an ugly EBS standard charactermode report (Oracle Reports) into the Xml-Publisher format. I'd like to use the existing report as the xml data source as all data is available in the old report.
    The problem is that the date format used by reports is determined by NLS_DATE_FORMAT (et al). This date format unfortunately is not xml compliant (dd.mm.yyyy in our case) and so Xml-Publisher date related features (e.g. sort by date) will not work .
    The report/template is submitted via the submit concurrent request form.
    Is there a way to make the underlying reports produce xml conformant dates without resorting to wild substring operations in the template or changing the EBS standard report?
    Thanks
    Christoph

    RajaramanV wrote:
    Can i change the nls parameter of a particular column(date field) in a table?if i set that it is apllied to all the tables in the database..No. Please understand that all dates, in all tables are stored in the same, oracle-defined, format. NLS_date_format affects how that internal date is converted to a character string for presentation to humans, or how to interpret a character string (received ,ultimately, from a human) for conversion to that internal format. It (NLS_DATE_FORMAT) can be set at several levels, but table/column is not one of them.
    Please read the link I posted in my earlier response. It will clear up your confusion.
    Edited by: EdStevens on Feb 1, 2012 6:55 AM

  • Need to change report data source on reports in a copied universe

    We are setting up a new testing environment.  So I made a copy of our BO universe so we could have two different testing universes.  I then created a copy of the data source database that will be used as the data source for the reports in the copied universe.  The two different databases have the exact same tables, columns and indexes as the development database.   The only difference will be the data inside the tables. 
    I now need to change the reports to use the new universe since the reports were written using the universe instead of using the database.  I am not changing anything else in the reports.  I just need them to run against a different Oracle database. 
    So I go into Crystal Reports version 11.5.0.313 and open the report up.  I then go into Database fields in Field Explorer to change the data source.  I then go to the Set Datasource Location screen.  I open up the universe label by clicking on the plus sign and I select the copied universe.  When I do that, the Business Objects Query panel screen opens up.  Because I have switched universes, Crystal Reports wants me to reselect the columns and rebuild the query.
    We have over three thousand reports so we are trying to avoid rebuilding them.  I am looking for a better way to change what database a report runs against. 
    Note: All of the reports will be run ion demand.  Nothing will be scheduled.
    Any suggestions and information will be greatly appreciated. 
    Has anyone run into this?   What was your solution to this problem?

    Hi Joe,
    Moved your post to the Universe Forum.
    First you are using the original release of XI R2. You need to upgrade to Service Pack 6 by:
    Run License Manager first to get the keycode if you don't have it on paper somewhere.
    Then download these and uninstall then run full build first:
    https://smpdl.sap-ag.de/~sapidp/012002523100011802732008E/crxir2_sp4_full_build.exe
    https://smpdl.sap-ag.de/~sapidp/012002523100013876392008E/crxir2win_sp5.exe
    https://smpdl.sap-ag.de/~sapidp/012002523100015859952009E/crxir2win_sp6.exe
    Try again
    Don

  • Asset history sheet Report date wants to change as selectoption

    Hi All,
    I have Requirement in Asset history sheet transaction S_ALR_87011990. In this i need to change Report date as select option now this is parameter.The selection screen declaration has done in logical data base ADA.
    Can u please suggest me how can we do this.
    Regards
    Madhu

    Hi,
    Can you be more specific about your requirement ? Because as per your Q there are two scenarios.
    1) Standard Logical data base
    2) Custom i.e. the one you created.
    In case of standard logical data base, goto SLDB --> name of the LDB ---> Goto the source code tab and in the source code search for enhancements to modify the screen elements. This way you can get the select-options for the parameters (provided any enhancements exist)
    If its a custom LDB, GOTO SLDB--> name of the LDB --> Source code, check for the screen elements code, here
    declare a tables statment
    ex.
    TABLES : SPFLI
    SELECT-OPTIONS S_CARRID FOR SPFI-CARRID
    instead of
    PARAMETERS:
    P_CARRID TYPE SPFLI-CARRID.
    This way the parameters option can be changed to select-options
    Thanks
    Sharath

  • Is there a way to report on Prospects being changed to a Sold-to

    I need a way to report on when a Prospect is changed to a Sold-to . . . is there something available on SAP?
    Our Finance group uses transaction OV50 to see when our Sales Group has created a Sold-to, so they can create the Finance views.  They check by Account Group and Date.  The problem is when a Prospect is changed it retains it's original date, so it's created on date is say 01/01/2000, and the change to a Sold-to is 9/9/09 . . . The report would have to be run wide open to see this is new.  They seem to have alot of customer masters they do not maintain.  So, is there a table where the changed date is stored, that I can report on.  I tried using OV51 the run time is very long.  Any ideas?
    Thanks,

    We create Sales Prospects, Acct Group 0005.  They are created with minimal data.  When the Prospects becomes a customer, we will convert to a Sold-to, Acct Group 001, by using transaction XD07.  Sales will do this so there is only the Sales "owned" views . . . like if they were to create thru VD01.  This means there are no Company Code views, XD01.  So I am looking for a report that will let Finance know when this conversion takes place so they can maintain their views, as well as the Credit Managment Master.  Today when Acct Group 0001 is created they run transaction OV50 to see what was recently created . . . daily or weekly they may run the report.  The problem with the Sales Prospect conversation, the created on date is in the past, so since they are either putting in a day or weekly date range, they never know when it was converted.  OV50 does have a changed on.  OV51 does, but has a significant run time, just for a day.  Any ideas?  Or any ideas where this data is stored, so a possible query could be done?

  • [svn:fx-4.x] 14837: Changing the way data tips are rendered in when mirrored.

    Revision: 14837
    Revision: 14837
    Author:   [email protected]
    Date:     2010-03-18 00:26:27 -0700 (Thu, 18 Mar 2010)
    Log Message:
    Changing the way data tips are rendered in when mirrored.
    Single data tip shows to left in RTL now which shows to right in LTR
    Callouts connecting datatips to chart items are also modified.
    QE notes:
    Doc notes:
    Bugs: FLEXDMV-2347 ( Datatips cluterred in bubble chart when layoutDirection=rtl)
    FLEXDMV-2348 ( Position of datatips not changing WRT layoutDirection)
    Reviewer:
    Tests run: checkintests
    Is noteworthy for integration:
    Ticket Links:
        http://bugs.adobe.com/jira/browse/FLEXDMV-2347
        http://bugs.adobe.com/jira/browse/FLEXDMV-2348
    Modified Paths:
        flex/sdk/branches/4.x/frameworks/projects/datavisualization/src/mx/charts/chartClasses/Ch artBase.as

    Setting Lightroom's preview size to larger than screen resolution is guaranteed to slow things down. I have a dual monitor system whith Lightroom showing the loupe on one screen and the grid on the other. The screen sizes are 1920x1080 and 2560x1440. Following your practice for previews would bring my system to a crawl and I have an entry level workstation.All my previews are of standard 1024x768 in size. The only delays in viewing come when Lightroom is asked to display an image where there is no preview. Once the preview has been created, browsing is near instantaneous.
    I beg to differ! The reason is that a preview can be used for any size, which is smaller than that of the preview. Scaling down a preview can be done with little or not loss in quality compared to getting the full resulution and scale that down. Scaling up is not possible.
    In your case, you will not use the previews of 1024x768 at all for full screen display. Instead you will force L3 to create bigger previews the first time you open the photo. You could try this by generating the standard previews for new imported photos before you start browsing. Then try to browse. You will probably get the "Loading" message, which indicates that the preview was not used and another one with higher resolution will be made. Then try the same experiment, setting the size of the standard preview to 2048. If, and only if, the size of your display area is not more than 2048 pixels wide and you display a photo in lanscape orientation, you will be able to browse without delay after generating the standard previews.
    If I understand your setup correctly, you display the photos in a window of 2560x1440 pixels. I suspect that only 1:1 previews could be used for that. I think Adobe should increase the maximum size of the previews to a value larger than 2048, so that you are not forced to use only the 1:1 previews.
    I agree, that when the previews have been generated, browsing is near instantaneous.

  • Trying to change a pre-defined report. Date format is wrong, 20140217 should be 17.02.2014.

    The pre-defined report in sccm "Computers with specific software registered in Add Remove Programs" has a query like this:
    Select DISTINCT sys.Netbios_Name0, fcm.SiteCode,  sys.User_Domain0, sys.User_Name0, sys.Operating_System_Name_and0, arp.DisplayName0,
    arp.InstallDate0  
    FROM fn_rbac_R_System(@UserSIDs)  sys 
    JOIN v_Add_Remove_Programs arp ON sys.ResourceID = arp.ResourceID  
    JOIN v_FullCollectionMembership fcm on sys.ResourceID=fcm.ResourceID 
    WHERE DisplayName0 like @filterwildcard and fcm.CollectionID=@CollID
    I have highlighted the part that I added to the report.
    I have inserted an extra column in Report Builder, and added the value inside the column and the install date can be viewed in the report. However the date format is shown like this: 20140217 - I want it to be 17.02.2014.
    So I try to change this via right-clicking on the cell in Report Builder, Text Box Properties, Number. Here I choose Date and the correct format. However it does not change the date format when saving and running the report again. It just stays the same. 
    Kthxbai

    The format can be change if it was a date. But if you look that field it is NOT a date, it is a string and therefore will not  change to the data format that you want.
    On top of this, there is a bigger problem. Not ever setup records the date or in the same format so.....
    http://www.enhansoft.com/
    Sorry - I am not sure what you mean by the last statement. "Not ever setup records the date or in the same format so.."
    - Would it be possible to elaborate? Thank you very much.
    Kthxbai

  • Best way to mark data that has changed when it is edited in  a spark datagrid?

    I am not sure if this is the best way to mark data that has changed (isModifiedClientSide ) when it is edited in  a spark datagrid:
    <GridColumn  width="140" headerText="Margin "dataField="margin" editable="false"  editable.editMode="true"  >
                        <itemRenderer>
                            <fx:Component>
                                <DefaultGridItemRenderer textAlign="right"   background="true" backgroundColor="#ffffff"  alpha="1.0" color="#000000" >
                                    <focusOut>
                                        <![CDATA[
                                        this.data.isModifiedClientSide = 1;
                                        ]]>
                                    </focusOut>
                                </DefaultGridItemRenderer>
                            </fx:Component>
                        </itemRenderer>
                    </GridColumn>

    I tried this but the value commit doesnt't get fired bro:
        <GridColumn  width="140" headerText="Margin (disabled)" headerText.editMode="Margin (editing)" dataField="margin" editable="false"  editable.editMode="true"  >
                        <itemEditor>
                            <fx:Component>
                                <DefaultGridItemEditor>
                                    <valueCommit>
                                        <![CDATA[
                                        this.data.isModifiedClientSide = 1;
                                        ]]>
                                    </valueCommit>
                                </DefaultGridItemEditor>
                            </fx:Component>
                        </itemEditor>
                    </GridColumn>
    what do you think?

  • Multiple data sets: a common global dataset and per/report data sets

    Is there a way to have a common dataset included in an actual report data set?
    Case:
    For one project I have about 70 different letters, each letter being a report in Bi Publisher, each one of them having its own dataset(s).
    However all of these letters share a common standardized reference block (e.g. the user, his email address, his phone number, etc), this common reference block comes from a common dataset.
    The layout of the reference block is done by including a sub-llayout (rtf-file).
    The SQL query for getting the dataset of the reference block is always the same, and, for now, is included in each of the 70 reports.
    Ths makes maintenance of this reference block very hard, because each of the 70 reports must be adapted when changes to the reference block/dataset are made.
    Is there a better way to handle this? Can I include a shared dataset that I would define and maintain only once, in each single report definition?

    Hi,
    The use of the subtemplate for the centrally managed layout, is ok.
    However I would like to be able to do the same thing for the datasets in the reports:
    one centrally managed data set (definition) for the common dataset, which is dynamic!, and in our case, a rather complex query
    and
    datasets defined on a per report basis
    It would be nice if we could do a kind of 'include dataset from another report' when defining the datasets for a report.
    Of course, this included dataset is executed within each individual report.
    This possibility would make the maintenance of this one central query easier than when we have to maintain this query in each of the 70 reports over and over again.

  • Cube with different reports data

    Hi
    We have 3 different Reports based on a Multi which has only one Infocube.
    We have new design policy that any New different report that needs to be created should be based only on this Multi and all data should be loaded only into this only infocube on which this Multi is based.
    So for the existing 3 reports,we have three different database views in source database.Fullload will be done from source every weekend(previous week's request will be deleted every week) of snapshot data into three different DSOs from these three different datasources.(Its new policy that every report should have its own DSO ).All these 3 DSOs should be connected to only one cube on which our reporting Multiprovider is based.So from these 3 DSOs,data will be loaded into cube in fullload every weekend(previous week's request will be deleted)
    Now my question is:
    1.when all the data is going to cube from three different views and DSOs(with different keyfigures but some same characteristics in 3 DSOs),how can our 3 different reports pick data that is relevant for it when all data is pooled together in one big cube?Doesnot aggregation happens?
    (When a report is run,does OLAP processor goes into cube and picks only those keyfigures and characteristics in cube thats relevant for that report and by that way,avoids other data??)
    2.I need to create one new report.For this I need to create one new view and one new DSO.So,data will be loaded from view to DSO and this DSO will be connected to existing cube(new dataflow until cube from source view).But our cube doesnot contain one keyfigure.It contains all other characteristics and keyfigures from DSO.
           a)How can I add this new keyfigure from new DSO to cube without disturbing existing data in cube.I donot need to load historical data for this new keyfigure and new report.
          b)I am going to transport this from development.So what care do I need to take while transporting changes to Production cube which contains lots of data and many reports running on it?
    Thanks.

    *1.when all the data is going to cube from three different views and DSOs(with different keyfigures but some same characteristics in 3 DSOs),how can our 3 different reports pick data that is relevant for it when all data is pooled together in one big cube?Doesnot aggregation happens?
    (When a report is run,does OLAP processor goes into cube and picks only those keyfigures and characteristics in cube thats relevant for that report and by that way,avoids other data??)*
    It will depend on what Key Figures you are displaying in your report. If a KF is coming from only 2 DSOs, data from those 2 DSOs will be displayed.  And you are right in saying that OLAP processor will pick only releavnt KFs and Characteristics but it will not filter based on characteristic value e.g. if a characteristis is coming only from DSO 1 and not from DSO2, but KF is coming from both, so for the data in DSO 2 that characteristic will be displayed as #. If you want to filter data, you can put filters in the report.
    Another good option can be to store the DSO name as a characeteristic in your cube, it might be helpful if you want to make a reprot based on data from a particular DSO, then you can put the filter based on DSO name.
    *2.I need to create one new report.For this I need to create one new view and one new DSO.So,data will be loaded from view to DSO and this DSO will be connected to existing cube(new dataflow until cube from source view).But our cube doesnot contain one keyfigure.It contains all other characteristics and keyfigures from DSO.
    a)How can I add this new keyfigure from new DSO to cube without disturbing existing data in cube.I donot need to load historical data for this new keyfigure and new report.
    b)I am going to transport this from development.So what care do I need to take while transporting changes to Production cube which contains lots of data and many reports running on it?*
    You are adding a new KF which is coming only from new DSO, so you do not need to worry about a lot of things, you can move your changes to procudtion and load data from this new DSO. Data which is already there in the cube will not be impacted.
    Regards,
    Gaurav

  • Finding changes in HR infotypes by date.

    Hi All,
    I am required to develop a report giving * changes in HR infotypes by date.
    *I have made use of two Function Modules, HR_INFOTYPE_LOG_GET_LIST and
    HR_INFOTYPE_LOG_GET_DETAIL. I am able to correctly get the table giving
    output of first function module HR_INFOTYPE_LOG_GET_LIST. However, when I am
    passing each record of table retrieved as output of first function module to
    second function module, it is taking a lot of time to process and finally
    gives a dump at the SELECT SQL statement that is responsible for retrieving the data from PCL4 cluster.
    Regards,
    Ameet

    Hi,
    Thanks for the reply.
    Yes, thats true. My requirement is based on report RPUAUD00 infotype audit report.
    Based on this report and Functional Specification given to me, I am required to make use of two Function Modules mentioned earlier.
    I have written a code given below:-
    DATA: t_change_list TYPE pldoc_key_tab,
    w_change_item TYPE pldoc_key.
    CALL FUNCTION 'HR_INFOTYPE_LOG_GET_LIST'
    EXPORTING
    TCLAS = 'A'
    begda = p_aedtm
    endda = p_aedtm
    TABLES
    infty_logg_key_tab = t_change_list
    DATA: t_before TYPE prelp_tab,
    t_after TYPE prelp_tab,
    w_before TYPE prelp,
    w_after TYPE prelp.
    LOOP AT t_change_list INTO w_change_item.
    CALL FUNCTION 'HR_INFOTYPE_LOG_GET_DETAIL'
    EXPORTING
    logged_infotype = w_change_item
    TABLES
    infty_tab_before = t_before
    infty_tab_after = t_after.
    ENDLOOP.
    Now, when I loop at t_after into its corresponding workarea, just to know the records populated by means of following code.
    LOOP AT t_after INTO w_after.
    CASE w_after-infty.
    WHEN '0001'.
    WHEN '0007'.
    etc..........
    ENDCASE.
    ENDLOOP.
    It gives me dump when report is run, and the dump error screen locates this error at point select * from PCL4.....
    I hope that my problem is more clear to you now. Please help me out to understand what would be causing to give the dump.
    Also I want to ask in case I dont want to use standard PNP selection screen. Inspite of that I want to use the only three select options viz Changed By, Changed On and PERNR from the RPUAUD00 infotype audit report. How can I achieve this. Is there a way to do this.
    Regards,
    Ameet

  • Report data binding error with date values

    I have CF7.02 with a Microsoft Visual FoxPro 9.0 SP1 Database
    that I connect to using ODBC (FoxPro Driver 6.01.8630.01). I send
    my sql results to a CF Report Builder 7.02 PDF report and it works
    fine. If I dump the date values before I change them, they look
    like the following: {ts '2004-12-20 00:00:00'} . However I have
    tried a number of ways of manipulating the date before sending it
    to the report, but I continue to get errors. I don't care what
    format they go to the report in, since the report reformats them
    anyway. I checked to make sure that none of the dates were null.
    For example <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDateTime(#ldCFPrcInputDate#) /> yields values in the
    {ts '2004-12-20 00:00:00'} format when I dump the query results to
    screen. Where:
    ldCFPrcInputDate = 12/20/2004
    CreateODBCDateTime(ldCFPrcInputDate) = {ts '2004-12-20
    00:00:00'}
    ReportQuery.PrcInputDate[lcCurRow] = {ts '2004-12-20
    00:00:00'}
    I get the error:
    Report data binding error Unable to get value for field
    'prcinputdate' of class 'java.util.Date'.
    coldfusion.runtime.OleDateTime -> Date
    Not using the CreateODBCDate function for example <cfset
    ReportQuery.PrcInputDate[lcCurRow] = #ldCFPrcInputDate# /> I
    get:
    Report data binding error Unable to get value for field
    'prcinputdate' of class 'java.util.Date'.
    java.lang.String -> Date
    Here are some of my failed attempts:
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDateTime(#ldCFPrcInputDate#) />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    #ldCFPrcInputDate# />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    DateFormat(CreateODBCDate(#ldCFPrcInputDate#),'mm/dd/yyyy') />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDate(DateFormat(#ldCFPrcInputDate#,'mm/dd/yyyy') />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDate(DateFormat(#ldCFPrcInputDate#,'mm-dd-yyyy') />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    Trim(createODBCDateTime(ldtmpdate)) />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDate({05-07-2006}) />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDate('{05-07-2006}') />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDate(05/07/2006) />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDate('05/07/2006')/>
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    CreateODBCDate(parseDateTime('05/07/2006')) />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    parseDateTime(CreateODBCDateTime(05/07/2006)) />
    <cfset ReportQuery.PrcInputDate[lcCurRow] =
    parseDateTime(CreateODBCDateTime('#ldCFPrcInputDate#')) />
    <cfset ReportQuery.PrcInputDate[lcCurRow] = '{05-07-2006}'
    />
    <cfset ReportQuery.PrcInputDate[lcCurRow] = '{05/07/2006}'
    />
    <cfset ReportQuery.PrcInputDate[lcCurRow] = '05/07/2006'
    />
    I also did some googling without success. Any help is
    appreciated.

    You may want to make sure that the column is not included in
    the query variable list in your report cfr file.
    The newest version of the Report Builder, which may be
    installed on the server, no longer tolerates vars appearing in that
    list of they are not in the query itselft.

  • Region sql query(updateable report) data entry in non db columns

    In a tabular report type updateable report
    What is the best way to provide data entry into a non db column a) (to accept parameters for some on demand pl/sql processes)?
    Only when setting a) as a standard report column the db insert/update transaction can be processed
    When changing a) to text field the db insert/update transaction fails with en error (screen/db are not in sync which is not actually the case)
    a) should not be part of the db insert/update transactions but still allow data entry
    How can this be achieved?
    Thanks
    Peter

    Hi Peter,
    I think the easiest solution is to use an updateable view instead of the direct table access.
    Here is an example:
    CREATE OR REPLACE VIEW V_EMPLOYEES
    AS
    SELECT EMPLOYEE_ID
         , FIRST_NAME
         , LAST_NAME
         , FIRST_NAME||' '||LAST_NAME AS FULL_NAME
      FROM EMPLOYEES
    CREATE OR REPLACE TRIGGER TRG_V_EMPLOYEES
        INSTEAD OF INSERT OR UPDATE OR DELETE
        ON V_EMPLOYEES
        REFERENCING NEW AS new OLD AS old
        FOR EACH ROW
    BEGIN
        IF UPDATING
        THEN
            UPDATE EMPLOYEES
               SET FIRST_NAME = :new.FIRST_NAME
                 , LAST_NAME  = :new.LAST_NAME
             WHERE EMPLOYEE_ID = :old.EMPLOYEE_ID
        -- insert and delete is not allowed
        ELSIF INSERTING
        THEN
            NULL; -- there would be a insert code
        ELSIF DELETING
        THEN
            NULL; -- there would be the delete code
        END IF;
    END;
    /FIRST_NAME||' '||LAST_NAME AS FULL_NAME in the example view simulates your function call.
    There might be a possibility to do it in APEX as well, but that was the first solution which came into my mind.
    Patrick
    Check out my APEX-blog: http://inside-apex.blogspot.com
    Check out the ApexLib Framework: http://apexlib.sourceforge.net

  • Adobe Form Data Storage

    When a user fills up a PCR Adobe Form from the portal where does the data or information is stored?
    There should be a fundamental way that SAP stores Adobe form data. I want to understand how and where this data is placed in SAP prior to it updating the employee’s infotypes so I can retrieve it and report on SAP and non-SAP fields.
    Same concept as in interface: data is sent from a file to SAP to update the back end. The file is stored separately from the back end up date data so it can be referenced later for support or other requirements. Same for the Adobe form data – it needs to be maintained the same way for future reference and current reporting.

    Well take the case of adobe forms based on web dynpro technology. Here the data is stored in context which are bean like structures and can store and transfer data to any system.
    Thanks,
    Harish

Maybe you are looking for

  • Photo sharing issues between iPad mini 2 and Macbook Pro

    Hi I'm trying to share two albums from iPhoto on my macbook to my iPad mini 2. However, nothing seems to be working, iCloud and Photostream aren't communicating between the two devices, bluetooth keeps telling me that my iPad is not equipped to recei

  • Mouse.hide() errors

    I am using AS 3 Flash CS4, WIndows XP SP3. I am unable to use the Mouse.hide(); method in many of my scripts. If I add Mouse.hide() to a preexisting script, I get the following error; line1 1061: Call to a possibly undefined method hide through a ref

  • How can I combine post to two different tables with one form?

    I am trying to create a "phantom" record in a table to be used later in a report. How can I create a form and properly $_POST to two different tables? Thanks for any suggestions!

  • Finding table name on which a trigger fired.

    I need my trigger to report the name of the table and the DML event that caused the "for each row" trigger to fire. How do I have the trigger identify this information without hardcoding the table name and event into the trigger ?

  • Import old film photos as tif, capture date set to 12/31/03 (1903)

    Lightroom 5.5 on Mac Mountain Lion, 8gb memory, files on external usb 2.0 hard drive. Imported old film images as tifs.  Looking in metadata the capture date on all images shows as 12/31/1903 (12/31/03?).  Where did this come from as there is no capt