Ideas on best approach - Product Catalogue, Orders to appear in a UWL

I am looking for some guidance/ideas on how best to approach the following:
I have a request to create a product catalogue (only to be accessed internally), that will allow staff to order products.
Once they have selected the products and quantities that they want, they will submit the order, and the the order should appear as an item in the UWL for a user. This user will then fill the order, and ship the goods to the user that ordered them.
I'm not sure which technology I would be best to use here, so any suggestions on this would be gratefully received.
Thanks in advance
Regards
Richard

Has no-one done anything like this?
Has anyone done a product catalog?

Similar Messages

  • Any ideas on best approach?

    I need ideas on how to best compare a users input against my randomly generated characters. ie does the users input of say, "abc", match what the computer has generated? how can i return the information to the user?

    if (computerGenerated.equals(userEntered)) {
        System.out.println("They match!");
    else {
        System.out.println("Computer had " + computerGenerated);
    }Mark

  • Best approach to creating a TOC for product catalog using data merge

    What is the best approach for creating a TOC for a product catalog (over 1,000 items) using Data Merge?
    The TOC would contain the product Categories. 
    So for example, Category A items could go from pages 1 - 3, and Category B items would start at pg 4, but if new items were added to Category A, then Category B may start from pg 6. 
    From the Data Source, there are 5 Data Fields I've chosen to be displayed.  If this were a regular digital print document, I could use the Paragraph Style method for creating a TOC, but if I make any one of the Data Fields a certain Paragraph Style and use that for the TOC, it'll populate the TOC with that Data Field for all the items. 
    Any suggestions?

    Peter Spier wrote:
    TOC is not interactive in the ID file, though it can be in a PDF that you export (there's a checkbox to create PDF bookmarks). You might want to think about using Cross-references (rather than hyperlinks, I think) to build the TOC. You have to do it manually, but once done it should maintain itself, whereas a TOC is built automatically, but must be regenerated after you edit the doc.
    One caveat witih TOCs created from cross-references: Although changing the text of an x-ref source paragraph (for example from "Patatas and tamatas" to "Tomatoes and Potatoes,"and/or when the source paragraph flows to the next or previous page) update automatically or when invoking "Update cross-references," MOVING a cross-reference source paragraph to a location before or after another source paragraph, does not change their sequence in the pseudo-TOC. You'll need to manually move the reference in the pseudo-TOC to the correct position in the sequence of cross-refs. So, put the task of checking the order of x-refs in the pseudo-TOC on your before hand-off check list.
    HTH
    Regards,
    Peter
    Peter Gold
    KnowHow ProServices

  • Whats the best software for photo editing and creating product catalogues that can be saved as PDF's

    Whats the best software for photo editing and creating product catalogues that can be saved as PDF's

    You are asking two different questiions:
    1. What's a good photo editor? Answer to that here is probably obvious.
    2. What's a good desktop publishing/page layout program for creating PDF files for production? Answer, not PSE. It's a photo editor, not a page layout program.

  • Best approach for handling a Repair business

    Hi All,
    A company provides Repair Services for various branded Products. The manufacturers send their Products for repair along with replacement Parts. The repaired Products are despatched back to the Original Manufacturer with details of Repair (nature of repair, Parts consumed etc.,). The Products are serialized and hence repair details has to be provided for each Unit of the Product.
    After repair, Labour and / or Material Charges (billing) are claimed with the Manufacturer (the company may also use some of their own Parts for repair). Actual cost incurred for Repair is also to be determined.
    Following metrics are to be taken care of:
    Technician’s Performance in terms of no of units repaired per day, % Yield (Out of 100 Units, recovery is 80) Technician wise, Cost of Repair – Actual cost incurred for Repair / No of Units repaired
    What would be the best approach to handle the above scenario in SAP R/3 v 4.7 ?
    Request feedback.
    Thanks.
    Raj

    Hi Raj,
               the steps for reparis process is as follows
    1.Repair order process
    2.Technical check and reparis processing
    3.outbound delivery and billing
    Repair order process
    1.create the sales order type RAS-repari/service and include the service product and serviceable material over there and save the order
    2. open the sales order in change mode go to menu sales document and deliver change the delivery type to LR i.e return delivery
    3. assign serial number for the serviceable materials
    the path is go to extras-->serial numbers assign and save
    4. go to VL02N and post the goods reciept that means we are taking the goods in to our stores. Then check IQ03 thr which we can display serial number master record and verify.
    Technical check and reparis processing
    1. open sales order in change mode and here we will recognize how the returns were posted
    go to item view-->repair tab in recieved field you can observe the recieved items
    2. below that Business decision: start repair
    you will get the repari procedure these things we can set it in customization.
    the path is sales and distributionsalessales documentscustomer services-returns and repari procedure--define repair procedure
    we have to assign repair procedure to item category
    so that whenever you will put the material through item category we will get our own repair procedure.
    then in the business decision
    we can select how many u want to repair and how many u want to do outbound and how many u want to scrap for scrap items uhave to issue credit memo after this save the sales order
    open sales order in change mode go to editdisplay criteria-all items
    double click on the item sothat you can see the smorder i.e. service order (which gets generated automatically) which we will assign in requirement class.
    Then go to service order confirm the operations and techo the order
    then last step is
    outbound delivery and billing
    we have to send the repaired device back
    go to va02
    and see as soon as we techo the order an outbound delivery is created in repari tab. Perfrom outbound delivery i.e. salesdocument--deliver(need ot remove default delivery type) and save.
    create transfer order and goods issue is posted.
    then go to DP90
    billing request process individually
    give your sales document number
    press save billing request
    then go to sales document billing
    invoice will get generated save it
    settle the service order to sales order thr kO88 and settle the repair order thr VA88
    still if you have any doubts mail me and reward me with points if it is useful to you
    Regards
    Satish

  • What is the best approach to process data on row by row basis ?

    Hi Gurus,
    I need to code stored proc to process sales_orders into Invoices. I
    think that I must do row by row operation, but if possible I don't want
    to use cursor. The algorithm is below :
    for all sales_orders with status = "open"
    check for credit limit
    if over credit limit -> insert row log_table; process next order
    check for overdue
    if there is overdue invoice -> insert row to log_table; process
    next order
    check all order_items for stock availability
    if there is item that has not enough stock -> insert row to
    log_table; process next order
    if all check above are passed:
    create Invoice (header + details)
    end_for
    What is the best approach to process data on row by row basis like
    above ?
    Thank you for your help,
    xtanto

    Processing data row by row is not the fastest method out there. You'll be sending much more SQL statements towards the database than needed. The advice is to use SQL, and if not possible or too complex, use PL/SQL with bulk processing.
    In this case a SQL only solution is possible.
    The example below is oversimplified, but it shows the idea:
    SQL> create table sales_orders
      2  as
      3  select 1 no, 'O' status, 'Y' ind_over_credit_limit, 'N' ind_overdue, 'N' ind_stock_not_available from dual union all
      4  select 2, 'O', 'N', 'N', 'N' from dual union all
      5  select 3, 'O', 'N', 'Y', 'Y' from dual union all
      6  select 4, 'O', 'N', 'Y', 'N' from dual union all
      7  select 5, 'O', 'N', 'N', 'Y' from dual
      8  /
    Tabel is aangemaakt.
    SQL> create table log_table
      2  ( sales_order_no number
      3  , message        varchar2(100)
      4  )
      5  /
    Tabel is aangemaakt.
    SQL> create table invoices
      2  ( sales_order_no number
      3  )
      4  /
    Tabel is aangemaakt.
    SQL> select * from sales_orders
      2  /
            NO STATUS IND_OVER_CREDIT_LIMIT IND_OVERDUE IND_STOCK_NOT_AVAILABLE
             1 O      Y                     N           N
             2 O      N                     N           N
             3 O      N                     Y           Y
             4 O      N                     Y           N
             5 O      N                     N           Y
    5 rijen zijn geselecteerd.
    SQL> insert
      2    when ind_over_credit_limit = 'Y' then
      3         into log_table (sales_order_no,message) values (no,'Over credit limit')
      4    when ind_overdue = 'Y' and ind_over_credit_limit = 'N' then
      5         into log_table (sales_order_no,message) values (no,'Overdue')
      6    when ind_stock_not_available = 'Y' and ind_overdue = 'N' and ind_over_credit_limit = 'N' then
      7         into log_table (sales_order_no,message) values (no,'Stock not available')
      8    else
      9         into invoices (sales_order_no) values (no)
    10  select * from sales_orders where status = 'O'
    11  /
    5 rijen zijn aangemaakt.
    SQL> select * from invoices
      2  /
    SALES_ORDER_NO
                 2
    1 rij is geselecteerd.
    SQL> select * from log_table
      2  /
    SALES_ORDER_NO MESSAGE
                 1 Over credit limit
                 3 Overdue
                 4 Overdue
                 5 Stock not available
    4 rijen zijn geselecteerd.Hope this helps.
    Regards,
    Rob.

  • What is the best adobe product to build an HD website moving background?

    I want to design a High Definition moving background for my website.  What would be the best adobe product to use?  I tried After Effects but the quality lowers when i integrate it.

    ... as with anything, you have to ask what the purpose of the 'web application' is: are you developing a portal system, a data collection form with or without reporting capabilities, logins, accounts, etc. - or an internal web application meant to handle simple business tasks with small amounts of data?
    The Oracle product offerings can run from simple (APEX) to more complex cases (ADF, WebCenter Framework, etc.) - it all depends on what you're trying to accomplish.
    A good place to start would be just to download the JDeveloper IDE - JDeveloper supports a really wide array of Oracle technologies so you can get a feel for what might be the best 'tool' for your use case.... and JDeveloper even supports a number of standard web app approaches (e.g. Apache Struts) if you're familiar with those kind of tools.

  • What's the best way to catalogue my media before import?

    I've got a new computer on the way and I wondered if some of you guys could help me set up a good file structure. I've been on a mac for the last 10 years and one thing that has got way out of hand is the way I import and save media from card. Although I've never lost connections to any media in the last decade it just looks a mess and it take me a while to locate thing because it has grown out of control.
    Can anyone point me in the direction of some good links to read about this? A new computer means a fresh start and before I start editing on Premire Pro I want to make sure I've got the best possible file structure in place for my media.
    All the best
    Kevin

    JIm, I assume that ... if one uses a camera with multi-part recording of data such as say, AVCHD, that renaming would be a right doggie. In that circumstance, then yes, I can see a "best practice" of never renaming those files.
    Again, coming from stills, "best practice" for cataloguing includes extensive discussions of proper renaming protocol, as the name is considered part of the searchable metadata of each individual file. Often, commercial photogs include their name (or at least initials) and the date of the job in every named file. Of course, those files get used individually, sent off to clients for use, and having the name on the file itself is considered ... wise.
    Within a portrait/wedding clientele, it's a good way to ensure at a glance that all the files in a folder being archived are where they ought to be, which of course, dsc_19843.nef doesn't help with at all. So it's a handy way to ensure that say, all the images going into a project or order for X client are actually images of X client. Of course, under the main name of the client/session, there'll be the original files (typically raw) in one folder, processed-for-use jpegs in another, and perhaps some tiff's or psd's with art-work still in layers in a third.  But this folder use is quite different from video use.
    So ... other than "it's considered best practice" ... I wondery why it's considered "best practice". Not that I'm doubting it at all, especially as you are one of the people around here with oft the best advice. Just curious as to the rationale.
    Neil

  • R/3 4.7 to ECC 6.0 Upgrade - Best Approach?

    Hi,
    We have to upgrade R/3 4.7 to ECC 6.0
    We have to do th DB, Unicode and R/3 upgrade. I want to know what are the best approaches available and what risks are associated iwth each approach.
    We have been considering the following approaches (but need to understand the risk for each approach).
    1) DB and Unicode in 1st and then R/3 upgrade after 2-3 months
    I want to understand that if we have about 700 Include Program changing as part of unicode conversion, how much of the functional testing is required fore this.
    2) DB in 1st step and then Unicode and R/3 together after 2-3 months
    Does it makes sense to combine Unicode and R/3 as both require similar testing? Is it possible to do it in 1 weekend with minimum downtime. We have about 2 tera bytes of data and will be using 2 systems fdor import and export during unicode conversion
    3) DB and R/3 in 1st step and then Unicode much later
    We had discussion with SAP and they say there is a disclaimer on not doing Unicode. But I also understand that this disclaimer does not apply if we are on single code page. Can someone please let us know if this is correct and also if doing Unicode later will have any key challenges apart from certain language character not being available.
    We are on single code pages 1100 and data base size is about 2 tera bytes
    Thanks in advance
    Regards
    Rahul

    Hi Rahul
    regarding your 'Unicode doubt"' some ideas:
    1) The Upgrade Master Guide SAP ERP 6.0 and the Master Guide SAP ERP 6.0 include introductory information. Among other, these guides reference the SAP Service Marketplace-location http://service.sap.com/unicode@sap.
    2) In Unicode@SAP can you find several (content-mighty) FAQs
    Conclusion from the FAQ: First of all your strategy needs to follow your busienss model (which we can not see from here):
    Example. The "Upgrade to mySAP ERP 2005"-FAQ includes interesting remarks in section "DO CUSTOMERS NEED TO CONVERT TO A UNICODE-COMPLIANT ENVIRONMENT?"
    "...The Unicode conversion depends on the customer situation....
    ... - If your organization runs a single code page system prior to the upgrade to mySAP ERP 2005, then the use of Unicode is not mandatory. ..... However, using Unicode is recommended if the system is deployed globally to facilitate interfaces and connections.
    - If your organization uses Multiple Display Multiple Processing (MDMP) .... the use of Unicode is mandatory for the mySAP ERP 2005 upgrade....."
    In the Technical Unicode FAQ you read under "What are the advantages of Unicode ...", that "Proper usage of JAVA is only possible with Unicode systems (for example, ESS/MSS or interfaces to Enterprise Portal). ....
    => Depending on the fact if your systems support global processes, or depending on your use of Java Applications, your strategy might need to look different
    3) In particular in view of your 3rd option, I recommend you to take a look into these FAQs, if not already done.
    Remark: mySAP ERP 2005 is the former name of the application, which is named SAP ERP 6.0, now
    regards, and HTH, Andreas R

  • What is the best approach to handle multiple FK with single table.

    If two tables are joined with each other with more than one ways, for example
    MAIN table is (col1, col2,....coln, person_creator_id, person_modifier_id)
    PERSON table is (person_id, name, address,........ phone) etc
    At database level PERSON_CREATOR_FK and PERSON_MODIFIER_FK are defined.
    Objective is to create a report that shows
    col1, col2...coln, person creator name, person modifier name
    If above two objects are imported with FKs in a EUL and discoverer plus is used to create above report. On first inclusion of person name discoverer plus will ask you to pick the join (provided the checkbox to disable this feature is not checked). Once you pick 'person creator' join it will never allow you to pick person modifier name.
    One solution is two create a custom folder with query like
    select col1, col2,...coln,
    pc.name, pc.address,.... pc.phone
    pm.name, pm.address,.... pm.phone
    from main m,
    person pc,
    person pm
    where m.person_id_creator = pc.person_id
    and m.person_id_modifier = pm.person_id
    Second solution is to import the PERSON folder twice in EUL (optionally named one as perosn_creator and other as person_modifier) and manually define one join per table. i.e. join MAIN with PERSON_CREATOR on person_creator_fk and join MAIN with PERSON_MODIFIER table using person_modifier_fk.
    Now discoverer plus will let you drag Name from each person folder without needing to resolve multiple joins.
    Question is, what approach is better OR is there a better way?
    With solution 1 you will not be able to use functions on folder items.
    With solution 2 there is a EUL design overhead of including same object multiple times and then manually defining all join (or deleting unwanted joins), and this could be a problem when you have person_modifier and person_creator in nearly all tables. It could be more complicated if person table is further linked other tables and users want to see that information too. (for instance, if person address is stored in LOCATION table joined with location_id and user want to see both creator address and modifier address....now you will have to create multiple LOCATION folders).
    A third solution could be to register a function in discoverer that return person name when person_id is passed. This will work perfectly for above requirement but a down side is the report will run slower if they need filters on person names (then function will be used in where clause). Also, this solution is very specific to above scenario, it will not work if you want the report developer the freedom to pick any attribute from person table (lets say, person table contain 50 attributes then its not a good idea to register 50 functions).
    Any comments/suggestion will be appreciated.
    thanks

    Hi
    In a roundabout way you have really answered your own question :-)
    In my opinion, the best approach, although by all means not the only approach - see below) would be to have the object loaded as two folders with one join going to the first folder and the second join to the other folder. You would of course name the folders appropriately.
    Here's a workflow that I use all of the time and one that I teach when I'm giving Discoverer Administrator training. It might help you:
    1. Bring in the PERSON folder to begin with
    2. Make all necessary adjustments to bring it up to deployment standard. These adjustments would be: folder name (E.g PERSON_CREATOR), item names, item placement, default positions, default aggregation and so on.
    3. Create or assign the required lists of values
    4. Create any required calculations
    5. Create any required conditions
    6. Create the first join from this folder to MAIN.
    7. Click on the heading for the folder and press CTRL-C.
    8. Click on the heading for the business area and press CTRL-V. A second copy of the folder, complete with all of the adjustments you made earlier will be inserted into the business area.
    Note: joins are not copied, everything else is.
    9. Rename this folder to say PERSON_MODIFIED
    10. Rename the items as appropriate
    11. Add a join from this folder to MAIN - you're done
    Other ideas that I have used and work well would be to use a database view or create a complex folder. Either will work, In both cases you would need to join on some other column other than the ones you referred earlier.
    I hope this helps
    Best wishes
    Michael

  • Best approach for IDOC - JDBC scenario

    Hi,
    In my scenarion I am creating sales order(ORDERS04) in R/3 system and which need to be replicated in a SQL Server system. I am sending the order to XI as an IDoc and want to use JDBC for sending data to SQL Server. I need to insert data in two tables(header & details). Is it possible without BPM?  Or what is the best approach for this?
    Thanks,
    Sri.

    Yes, this is possible without the BPM.
    Just create the Corresponding Datatype for the insertion.
    if the records to be inserted are different, then there wil be 2 different datatypes ( one for header and one for detail).
    Do a mutlimapping, where your Source is mapped into the header and details datatype and then send using the JDBC sender adapter.
    For the strucutre of your Datatype for insertion , just check this link,
    http://help.sap.com/saphelp_nw04/helpdata/en/7e/5df96381ec72468a00815dd80f8b63/content.htm
    To access any Database from XI, you will have to install the corresponding Driver on your XI server.
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/3867a582-0401-0010-6cbf-9644e49f1a10
    Regards,
    Bhavesh

  • Best approach to return Large data ( 4K) from Stored Proc

    We have a stored proc (Oracle 8i) that:
    1) receives some parameters.
    2)performs computations which create a large block of data
    3) return this data to caller.
    compatible with both ASP (using MSDAORA.Oracle), and ColdFusion (using Oracle ODBC driver). This procedure is critical in terms of performance.
    I have written this procedure as having an OUT param which is a REF CURSOR to a record containing a LONG. In order to make this work, at the end of the procedure I have to store the working buffer (an internal LONG variable) into a temp table, and then open the cursor as a SELECT from the temp table.
    I have tried to open the cursor as a SELECT of the working buffer (from dual) but I get the error "ORA-01460: unimplemented or unreasonable conversion requested"
    I suspect this is taking too much time; any tips about the best approach here? is there a resource with REAL examples on returning large data?
    If I switch to CLOB, will it speed the process, be compatible with callers, etc ? all references to CLOB I saw use trivial examples.
    Thanks for any help,
    Yoram Ayalon

    We have a stored proc (Oracle 8i) that:
    1) receives some parameters.
    2)performs computations which create a large block of data
    3) return this data to caller.
    compatible with both ASP (using MSDAORA.Oracle), and ColdFusion (using Oracle ODBC driver). This procedure is critical in terms of performance.
    I have written this procedure as having an OUT param which is a REF CURSOR to a record containing a LONG. In order to make this work, at the end of the procedure I have to store the working buffer (an internal LONG variable) into a temp table, and then open the cursor as a SELECT from the temp table.
    I have tried to open the cursor as a SELECT of the working buffer (from dual) but I get the error "ORA-01460: unimplemented or unreasonable conversion requested"
    I suspect this is taking too much time; any tips about the best approach here? is there a resource with REAL examples on returning large data?
    If I switch to CLOB, will it speed the process, be compatible with callers, etc ? all references to CLOB I saw use trivial examples.
    Thanks for any help,
    Yoram Ayalon

  • Best approach to "migrate" from BEX reports to Webi reports ?

    Hello,
    i have read lots of documents regarding best practices on how to built webi reports and universes on top of BW.
    But i can't find any document about best approach, not in performance way of thinking but in best way of using reports.
    i mean: when end users are coming from bex reports (where they can drill down through hierarhies and use free filters ) to webi reports (where layout is quite beatiful and user can change it easely), this is not the same way of consuming reports.
    I come from BO world and are new on reporting on top of BW.
    for me webi is good for quite static layout reporting where data is clear and available.of course you can have prompts for interactivy and more accurate reporting.  Drill down is just a functionality but is not the real purpose of the report tool.
    So ,according to me there is a gap between both tools (BEX and WEBI) but end users are the same.
    So i 'm wondering if you have any feedback for the best approach to build webi reports where end users are coming from bex reporting.
    And how to choose between prompts, drill-down (with available filters on top of the window), fold/unfold and input controls or just having diffrent levels of hierarchies in the table/ section/ breaks but without drill down (because if you drill down, report begins weird with diffrent levels) ...?
    So , if you have any feedback , advise....
    thanks in advance,
    Rgds,

    Hi,
      WEBI don't replace BEX reports, is for different audience, in fact BEX is for OLAP reports and analysis.
      You can find some answer in this page
    [FAQ: The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI & Business Objects Roadmap|FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap [original link is broken]|FAQ]
    spercific for What is the future of the BEx Query Designer? you can read here
    [FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap [original link is broken]#section11] and here [FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap [original link is broken]#section3]
      The idea is to use the rigth tool for the rigth job.
      You can find more information here [http://www.sdn.sap.com/irj/sdn/edw], [http://www.sap.com/solutions/sapbusinessobjects/index.epx], [http://www.sap.com/solutions/sapbusinessobjects/newsevents/index.epx], [http://www.sap.com/community/flash/BusinessIntelligenceAGuideforMidsizeCompanies.pdf]
    I hope this help you.
    Best regards.

  • What are the best approaches for mapping re-start in OWB?

    What are the best approaches for mapping re-start in OWB?
    We are using OWB repository 10.2.0.1.0 and OWB client 10.2.0.1.31. The Oracle version is 10 G (10.2.0.3.0). OWB is installed on Linux.
    We have number of mappings. We built process flows for mappings as well.
    I like to know, what are the best approches to incorportate re-start options in our process. ie a failure of mapping in process flow.
    How do we re-cycle failed rows?
    Are there any builtin features/best approaches in OWB to implement the above?
    Does runtime audit tables help us to build re-start process?
    If not, do we need to maintain our own tables (custom) to maintain such data?
    How did our forum members handled above situations?
    Any idea ?
    Thanks in advance.
    RI

    Hi RI,
    How many mappings (range) do you have in a process flows?Several hundreds (100-300 mappings).
    If we have three mappings (eg m1, m2, m3) in process flow. What will happen if m2 fails?Suppose mappings connected sequentially (m1 -> m2 -> m3). When m2 fails then processflow is suspended (transition to m3 will not be performed). You should obviate cause of error (modify mapping and redeploy, correct data, etc) and then repeat m2 mapping execution from Workflow monitor - open diagram with processflow, select mapping m2 and click button Expedite, choose option Repeat.
    In re-start, will it run m1 again and m2 son on, or will it re-start at row1 of m2?You can specify restart point. "at row1 of m2" - I don't understand what you mean (all mappings run in Set based mode, so in case of error all table updates will rollback,
    but there are several exception - for example multiple target tables in mapping without corelated commit, or error in post-mapping - you must carefully analyze results of error).
    What will happen if m3 fails?Process is suspended and you can restart execution from m3.
    By having without failover and with max.number of errors=0, you achieve re-cycle failed rows to zero (0).This settings guarantee existence only two return result of mapping - SUCCSES or ERROR.
    What is the impact, if we have large volume of data?In my opinion for large volume Set based mode is the prefered processing mode of data processing.
    With this mode you have full range enterprise features of Oracle database - parallel query, parallel DML, nologging, etc.
    Oleg

  • Best approach to add Z custom field to IC Agent Inbox search and results view

    Hi Experts,
    We are having a requirement to add a Z custom field to IC Agent Inbox search and results view. I got multiple forums and ideas, but looking for the best approach for handling this. I am sure, you experts, would have already done this.
    Thanks in advance.
    Regards
    Siva

    Hi Sivakumar,
    AET is the best way by far to create a custom field in this area. It is easy and simple.
    Also, field once added in one business object it can be used at different objects as well.
    There is also a demo available for AET on sdn.
    Please let me know if any more help is required.
    Thanks,
    Bhushan

Maybe you are looking for