Printing help -- Best approach

Hi, I've been working on a database application and have been having trouble finding the best way to print the table (with some header info or something). Since I already wrote an export to html function I've been using that as a basis for printing. As of right now basically adding the html code to a styled text area and printing that. It prints, but gets cutoff and doesn't handle extra pages.
I'm been searching around the web and following the sun tutorial quite a bit, but I'm not sure I'm even taking the write approach to printing the data from the DB. I really just want it to take the info from a query and print it to a printer. Can someone point me in the right direction or give me some suggestions?

Thanks, I had seen something about the netbeans iReport plugin before. I'll take a look at a bit more, but this is probably the way to go. The report doesn't have to have specific page breaks, but it does need a heading on each page.

Similar Messages

  • HOW TO/BEST APPROACH: Printing CMS-1500 medical claim forms from C# winforms app..

    Hello everyone. I am currently developing a winforms app in c# and need to print a medical claim form: CMS-1500 (ANSI837/X12) look for best approach.. But first I need to know best forum to post this?
    Thanks!

    Steven and Jim,
    I know this is an old post, so it's probably a long shot, but I am also looking for some sample code to output a cms1500.  My wife runs a 1-person counseling company and is a provide for the state of Ohio, which is switching to cms1500 for claims.  
    I made an application for her ages ago to manage claim submission, but now need to switch it over to output cms1500.  Any advice you could offer would be greatly appreciated!
    Regards,
    Mike

  • What is the best approach to handle multiple FK with single table.

    If two tables are joined with each other with more than one ways, for example
    MAIN table is (col1, col2,....coln, person_creator_id, person_modifier_id)
    PERSON table is (person_id, name, address,........ phone) etc
    At database level PERSON_CREATOR_FK and PERSON_MODIFIER_FK are defined.
    Objective is to create a report that shows
    col1, col2...coln, person creator name, person modifier name
    If above two objects are imported with FKs in a EUL and discoverer plus is used to create above report. On first inclusion of person name discoverer plus will ask you to pick the join (provided the checkbox to disable this feature is not checked). Once you pick 'person creator' join it will never allow you to pick person modifier name.
    One solution is two create a custom folder with query like
    select col1, col2,...coln,
    pc.name, pc.address,.... pc.phone
    pm.name, pm.address,.... pm.phone
    from main m,
    person pc,
    person pm
    where m.person_id_creator = pc.person_id
    and m.person_id_modifier = pm.person_id
    Second solution is to import the PERSON folder twice in EUL (optionally named one as perosn_creator and other as person_modifier) and manually define one join per table. i.e. join MAIN with PERSON_CREATOR on person_creator_fk and join MAIN with PERSON_MODIFIER table using person_modifier_fk.
    Now discoverer plus will let you drag Name from each person folder without needing to resolve multiple joins.
    Question is, what approach is better OR is there a better way?
    With solution 1 you will not be able to use functions on folder items.
    With solution 2 there is a EUL design overhead of including same object multiple times and then manually defining all join (or deleting unwanted joins), and this could be a problem when you have person_modifier and person_creator in nearly all tables. It could be more complicated if person table is further linked other tables and users want to see that information too. (for instance, if person address is stored in LOCATION table joined with location_id and user want to see both creator address and modifier address....now you will have to create multiple LOCATION folders).
    A third solution could be to register a function in discoverer that return person name when person_id is passed. This will work perfectly for above requirement but a down side is the report will run slower if they need filters on person names (then function will be used in where clause). Also, this solution is very specific to above scenario, it will not work if you want the report developer the freedom to pick any attribute from person table (lets say, person table contain 50 attributes then its not a good idea to register 50 functions).
    Any comments/suggestion will be appreciated.
    thanks

    Hi
    In a roundabout way you have really answered your own question :-)
    In my opinion, the best approach, although by all means not the only approach - see below) would be to have the object loaded as two folders with one join going to the first folder and the second join to the other folder. You would of course name the folders appropriately.
    Here's a workflow that I use all of the time and one that I teach when I'm giving Discoverer Administrator training. It might help you:
    1. Bring in the PERSON folder to begin with
    2. Make all necessary adjustments to bring it up to deployment standard. These adjustments would be: folder name (E.g PERSON_CREATOR), item names, item placement, default positions, default aggregation and so on.
    3. Create or assign the required lists of values
    4. Create any required calculations
    5. Create any required conditions
    6. Create the first join from this folder to MAIN.
    7. Click on the heading for the folder and press CTRL-C.
    8. Click on the heading for the business area and press CTRL-V. A second copy of the folder, complete with all of the adjustments you made earlier will be inserted into the business area.
    Note: joins are not copied, everything else is.
    9. Rename this folder to say PERSON_MODIFIED
    10. Rename the items as appropriate
    11. Add a join from this folder to MAIN - you're done
    Other ideas that I have used and work well would be to use a database view or create a complex folder. Either will work, In both cases you would need to join on some other column other than the ones you referred earlier.
    I hope this helps
    Best wishes
    Michael

  • Best approach for IDOC - JDBC scenario

    Hi,
    In my scenarion I am creating sales order(ORDERS04) in R/3 system and which need to be replicated in a SQL Server system. I am sending the order to XI as an IDoc and want to use JDBC for sending data to SQL Server. I need to insert data in two tables(header & details). Is it possible without BPM?  Or what is the best approach for this?
    Thanks,
    Sri.

    Yes, this is possible without the BPM.
    Just create the Corresponding Datatype for the insertion.
    if the records to be inserted are different, then there wil be 2 different datatypes ( one for header and one for detail).
    Do a mutlimapping, where your Source is mapped into the header and details datatype and then send using the JDBC sender adapter.
    For the strucutre of your Datatype for insertion , just check this link,
    http://help.sap.com/saphelp_nw04/helpdata/en/7e/5df96381ec72468a00815dd80f8b63/content.htm
    To access any Database from XI, you will have to install the corresponding Driver on your XI server.
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/3867a582-0401-0010-6cbf-9644e49f1a10
    Regards,
    Bhavesh

  • Best approach to return Large data ( 4K) from Stored Proc

    We have a stored proc (Oracle 8i) that:
    1) receives some parameters.
    2)performs computations which create a large block of data
    3) return this data to caller.
    compatible with both ASP (using MSDAORA.Oracle), and ColdFusion (using Oracle ODBC driver). This procedure is critical in terms of performance.
    I have written this procedure as having an OUT param which is a REF CURSOR to a record containing a LONG. In order to make this work, at the end of the procedure I have to store the working buffer (an internal LONG variable) into a temp table, and then open the cursor as a SELECT from the temp table.
    I have tried to open the cursor as a SELECT of the working buffer (from dual) but I get the error "ORA-01460: unimplemented or unreasonable conversion requested"
    I suspect this is taking too much time; any tips about the best approach here? is there a resource with REAL examples on returning large data?
    If I switch to CLOB, will it speed the process, be compatible with callers, etc ? all references to CLOB I saw use trivial examples.
    Thanks for any help,
    Yoram Ayalon

    We have a stored proc (Oracle 8i) that:
    1) receives some parameters.
    2)performs computations which create a large block of data
    3) return this data to caller.
    compatible with both ASP (using MSDAORA.Oracle), and ColdFusion (using Oracle ODBC driver). This procedure is critical in terms of performance.
    I have written this procedure as having an OUT param which is a REF CURSOR to a record containing a LONG. In order to make this work, at the end of the procedure I have to store the working buffer (an internal LONG variable) into a temp table, and then open the cursor as a SELECT from the temp table.
    I have tried to open the cursor as a SELECT of the working buffer (from dual) but I get the error "ORA-01460: unimplemented or unreasonable conversion requested"
    I suspect this is taking too much time; any tips about the best approach here? is there a resource with REAL examples on returning large data?
    If I switch to CLOB, will it speed the process, be compatible with callers, etc ? all references to CLOB I saw use trivial examples.
    Thanks for any help,
    Yoram Ayalon

  • Best approach to "migrate" from BEX reports to Webi reports ?

    Hello,
    i have read lots of documents regarding best practices on how to built webi reports and universes on top of BW.
    But i can't find any document about best approach, not in performance way of thinking but in best way of using reports.
    i mean: when end users are coming from bex reports (where they can drill down through hierarhies and use free filters ) to webi reports (where layout is quite beatiful and user can change it easely), this is not the same way of consuming reports.
    I come from BO world and are new on reporting on top of BW.
    for me webi is good for quite static layout reporting where data is clear and available.of course you can have prompts for interactivy and more accurate reporting.  Drill down is just a functionality but is not the real purpose of the report tool.
    So ,according to me there is a gap between both tools (BEX and WEBI) but end users are the same.
    So i 'm wondering if you have any feedback for the best approach to build webi reports where end users are coming from bex reporting.
    And how to choose between prompts, drill-down (with available filters on top of the window), fold/unfold and input controls or just having diffrent levels of hierarchies in the table/ section/ breaks but without drill down (because if you drill down, report begins weird with diffrent levels) ...?
    So , if you have any feedback , advise....
    thanks in advance,
    Rgds,

    Hi,
      WEBI don't replace BEX reports, is for different audience, in fact BEX is for OLAP reports and analysis.
      You can find some answer in this page
    [FAQ: The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI & Business Objects Roadmap|FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap [original link is broken]|FAQ]
    spercific for What is the future of the BEx Query Designer? you can read here
    [FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap [original link is broken]#section11] and here [FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap [original link is broken]#section3]
      The idea is to use the rigth tool for the rigth job.
      You can find more information here [http://www.sdn.sap.com/irj/sdn/edw], [http://www.sap.com/solutions/sapbusinessobjects/index.epx], [http://www.sap.com/solutions/sapbusinessobjects/newsevents/index.epx], [http://www.sap.com/community/flash/BusinessIntelligenceAGuideforMidsizeCompanies.pdf]
    I hope this help you.
    Best regards.

  • Best approach -To create RTF template having more than 50 tables.

    Hi All,
    Need your help.I am new to BI publisher. Currently we are using BIP 11g.
    I want to develop.rtf template having lots of layout and images.
    Data is coming from different tables (example : pulling from around 40 tables). When i tried to pull data from 5 tables by joining tables. It takes more time using data model in BI publisher 11g saved in xml and used in word doc.
    Could you please suggest best approach  weather i need to develop .rtf template via data model or query to generate a report.
    Also please suggest / guide me .
    Regards & Thanks in advance.

    it's very specific requirements
    first of all it's relate to logic behind
    as example 50 tables are related ? or 50 independent tables ? or may be 5 related and another independent ?
    based on relation of tables you create sql statement(s)
    how many sql statement(s) you'll have lead to identify ways to get data, as example, by package or trigger etc
    kim size of resulting select statement(s)
    if size say 1mb it's must be fast to get report but for 1000mb it can consume many time
    also kim what time it's not only to select data but to merge data and template
    looks like experimenting and knowing full logic of report is only ways to get needed output in projection of data and time

  • What are the best approaches for mapping re-start in OWB?

    What are the best approaches for mapping re-start in OWB?
    We are using OWB repository 10.2.0.1.0 and OWB client 10.2.0.1.31. The Oracle version is 10 G (10.2.0.3.0). OWB is installed on Linux.
    We have number of mappings. We built process flows for mappings as well.
    I like to know, what are the best approches to incorportate re-start options in our process. ie a failure of mapping in process flow.
    How do we re-cycle failed rows?
    Are there any builtin features/best approaches in OWB to implement the above?
    Does runtime audit tables help us to build re-start process?
    If not, do we need to maintain our own tables (custom) to maintain such data?
    How did our forum members handled above situations?
    Any idea ?
    Thanks in advance.
    RI

    Hi RI,
    How many mappings (range) do you have in a process flows?Several hundreds (100-300 mappings).
    If we have three mappings (eg m1, m2, m3) in process flow. What will happen if m2 fails?Suppose mappings connected sequentially (m1 -> m2 -> m3). When m2 fails then processflow is suspended (transition to m3 will not be performed). You should obviate cause of error (modify mapping and redeploy, correct data, etc) and then repeat m2 mapping execution from Workflow monitor - open diagram with processflow, select mapping m2 and click button Expedite, choose option Repeat.
    In re-start, will it run m1 again and m2 son on, or will it re-start at row1 of m2?You can specify restart point. "at row1 of m2" - I don't understand what you mean (all mappings run in Set based mode, so in case of error all table updates will rollback,
    but there are several exception - for example multiple target tables in mapping without corelated commit, or error in post-mapping - you must carefully analyze results of error).
    What will happen if m3 fails?Process is suspended and you can restart execution from m3.
    By having without failover and with max.number of errors=0, you achieve re-cycle failed rows to zero (0).This settings guarantee existence only two return result of mapping - SUCCSES or ERROR.
    What is the impact, if we have large volume of data?In my opinion for large volume Set based mode is the prefered processing mode of data processing.
    With this mode you have full range enterprise features of Oracle database - parallel query, parallel DML, nologging, etc.
    Oleg

  • Best approach to add Z custom field to IC Agent Inbox search and results view

    Hi Experts,
    We are having a requirement to add a Z custom field to IC Agent Inbox search and results view. I got multiple forums and ideas, but looking for the best approach for handling this. I am sure, you experts, would have already done this.
    Thanks in advance.
    Regards
    Siva

    Hi Sivakumar,
    AET is the best way by far to create a custom field in this area. It is easy and simple.
    Also, field once added in one business object it can be used at different objects as well.
    There is also a demo available for AET on sdn.
    Please let me know if any more help is required.
    Thanks,
    Bhushan

  • Best approach -Tabs based ADF Tree left side navigation with Dynamic Regions with out UI Shell

    Hi,
    Somebody can help for the best approach to implement the following requirement.
    Req: When the user select the ADF Tree left side navigation menu, each menu will open as multiple tabs(Dynamic Tabs) in right side content area with out UI Shell Template.
    I completed the
    Step-1: From the Model project, I can able to render ADF Tree in the using view and view links. I can get the adf tree which is having 3 menu items. Each menu item having 2 sub menu's.
    I took each menu item as one(1) taskflow, each taskflow will have two(2) fragments.
    Total I have 3 task flows as Menu Items and 6 fragments for sub menu's.
    Step-2:  My question is How do I implement Tab based the ADF tree navigation (left side area to dynamic regions in content area) through dynamic regions? Please provide the steps in view layers.

    Than ks for your response.
    This is working fine for ADF Tree navigation with dynamic regions if the taskflow having only one fragment. if the taskflow having more than one fragments, this will not work. The following conditions are always satisfies one page fragment of either "employees" or "departments" task flow.  If the "employees" task flow have 2 page fragments, it's not work even you pass parameters through routers.
    public TaskFlowId getDynamicTaskFlowId() {
    if (currentTaskFlowID == null ||
    currentTaskFlowID.equalsIgnoreCase(“employees”)) {
    return TaskFlowId.parse(employeetaskFlowId);
    if (currentTaskFlowID != null &&
    currentTaskFlowID.equalsIgnoreCase(“departments”)) {
    return TaskFlowId.parse(departmetaskFlowId);
    return TaskFlowId.parse(employeetaskFlowId);
    My question is "Same use case with Dynamic Tabs" when the user click on any adf tree node.

  • Best approach to develop DataTypes

    Hey guys
    suppose i m doin a complex File scenario in which i have a flat file either on the sender or receiver side.what is the best approach to develop the data type for the file structure,is it possible to generate an XML and with the help of that create a data type?
    I actually want to know the professional way of generating datatypes (i m pretty sure in real world scenarios,we get very complex file structures).
    thanx
    Ahmad

    Hi,
    In many cases, if your structure is very complex, you can not get direct nested xml after content conversion. In that case, in the mapping we need to handle the generation of nested strutcure. So you use Java or xslt mapping etc.. if it is not possible via graphical mapping. Also you can do this in the adapter module.
    Here you go with good example- Generic Structure-
    /people/sravya.talanki2/blog/2005/08/16/configuring-generic-sender-file-cc-adapter
    Also file content conversion - limitations-
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/50061bd9-e56e-2910-3495-c5faa652b710
    Rgds,
    Moorthy

  • Best approach with jsp

    hi,
    i have a scenario that's sending data from jsp.
    i have two options, soap and client java proxy.
    what is best approach for this?,  becasue jsp is calling java which contains coding to call webservice,  so this is similar with client java proxy.
    if you are me,  what is the best?  soap or cleint java proxy?  eventually, soap also proxy.
    please guide me what is big point?  i do not want to get you with basically different point that i know that client java proxy is ejb. soap is wsdl...
    thanks & regards;
    ogawa
    null

    Hi,
    As you already have a J2EE application ( a JSP ), i would suggest that you use Java Proxies for the same to integrate with XI. Proxies allow you to have a better perfromance and so, I would suggest to go with Java Proxies.
    For Java proxies this how to guide would help,
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/7d4db211-0d01-0010-1e8e-9b07fc2113ab
    Regards
    Bhavesh

  • Best approach for sending SMS through workflow

    Hi Experts,
    I have a scenario where in I have to send SMS as a part of an escalation workflow process.I wanted to ask to the forum as to what is the best approach to send sms from workflow.The information regarding the users and the numbers is to come from a table.So how do I integrating the workflow with the SMS.
    I am stuck with this problem and need urgent help with this..Any help would be sincerely appreciated.
    Thanks and Regards,
    Narsingh

    Hi Narsingh Panigrahi,
    Have you checked this function module as i suggested previously.
    Function Module--TELEPHONE_FROM_TO
    Since configuration is not done in our system i cant test this function module
    In this function module they have used this table TSP01
    The fields are
    RQTELELAN--country
    RQTELENUMTelecommunications partnerfrom
    RQTELENUMETelecommunications partnerto
    Function Module--TELEPHONE_FROM_TO
    In this function module they have used this table TSP01
    The fields are
    RQTELELAN--country
    RQTELENUMTelecommunications partnerfrom
    RQTELENUMETelecommunications partnerto
    When you are entering the telephone number you have give prefix country code ( for US it is 011) and at the end of the number you have to give the to country code that is ( for IN it is 91).
    Try this and let me know.
    Regards,
    Balaji E.

  • Best approach for multi-team/multi-projects.

    Hi,
    I'm looking for the best approach to handle multi-teams/multi-projects scenario. We have 20 development groups and over 300 products. Each products on it's own Schedule.
    Product X can be assign to Group A, but at some point, it can be assign to Group B.
    We are currently using TFS 2012, but will be upgrading to 2013 soon.
    Based on many reading, we are thinking to create only one Team Projects to ease management.
    In it, we will create a team for each development group, but we will not create an associated area path with the name of the team.
    - Group A
    - Group B
    - Group C
    Than, we will create an Area for each product.
    - Product X
    - Product Y
    - Product Z
    and, we will create multi-level of iterations to match each Schedule.
    - Product X
       - Release 1
          - Sprint 1
          - Sprint 2
    - Product Y
        - Release 1
           -Sprint 1
        - Release 2
           -Sprint 1
    The main issue, we have with this approach is that we can't use the Backlog or the Task Board effectivelly, as there is no way to filter per areas and iterations.
    Reading "How do I change the underlying query for the task board (and backlog board) on TFS Preview", this doen't seam to be possible in TFS 2012.
    In TFS 2013, "The Agile Portfolio Management: Using TFS to support backlogs across multiple teams" was introduced. Will this help to solve the problem?
    We would create a management team for each development group.
    We would create an agile team with an associated area for each product.
    The only thing that I couldn't find in the documentation is how to re-assign an agile team to another management team. Is this possible?
    Also can each agile team have their specific itérations, if so will it roll up properly to the management team?
    Regards
    SYSOTI
    PS: Sorry couldn't post the links of the quoted text as I get the message: Body text cannot contain images or links until we are able to verify your account. ;-(

    Hi SYSOTI,
    Based on your description, seems the area path is not configured properly hence you can't use the Backlog or the Task Board effectivelly.
    From the
    Agile Portfolio Management: Using TFS to support backlogs across multiple teams, the area path is set as agile team which is a consist of team members but not a product name. For you scenario, you can set the area path name as your product name to
    identify the associate products for work items.  And the groups you mentioned for products in the team projects are sub-group of contributors.
    Seems there is no need to create a management team for each development group since management team might be in a higher level to view the progress for all of the work across the agile teams. Certianly, you can create multiple management teams, but the management
    teams will be able to view works for all agile teams. 
    If you have multiple teams and products, you can create a team project for each product if the products don't have much relationship. However, it's OK to manage the projects for multiple products in the same team project. And working within a single team
    project also have benifits, you ccan check this
    blog for more information.
    Best regards,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • What´s the best approach to work with Excel, csv files

    Hi gurus. I got a question for you. According to your experience what's the best approach to work with Excel or csv files that have to be uploaded through DataServices to you datawarehouse.
    Let's say your end-user, who is not a programmer, creates a group of 4 excel files with different calculations in a monthly basis, so they can generate a set of reports from their datawarehouse once the files have been uploaded to tables in your DWH. The calculations vary from month to month. The user doesn't have a front-end to upload the excel files directly to Data Services. The end user needs to keep a track of which person uploaded the files for a determined month.
    1. The end user should place their 4 excel files in a shared directory that will be seen by DataServices.
    2. DataServices will execute certain scheduled job that will read the four files and upload them to the Datawarehouse at a determined time, lets say at 9:00pm.
    It makes me wonder... what happens if the user needs to present their reports immediately so they can´t wait until 9:00pm.  Is it possible for the end user to execute some kind of action (out of the DataServices Environment) so DataServices "could know" that it has to process those files right now, instead of waiting for the night schedule?
    Is there a way that DS will track who was the person who uploaded those files?
    Would it be better to build a front-end for the end user so they can upload their four files directlyto the datawarehouse?
    Waiting for your comments to resolve this dilemma
    Best Regards
    Erika

    Hi,
    There are functions in DS that captures the input files automatically. You could use file_exists() or wait_for_file() option to do that. Schedule the job to run every certain minute and if the file exists then run. This could be done by using a certain file name with date and timestamp etc or after running move the old files to archive and DS wait for new files to show up.
    Check this - Selective Reading and Postprocessing - Enterprise Information Management - SCN Wiki
    Hope this helps.
    Arun

Maybe you are looking for

  • HT202879 Where can I find a tutorial for Numbers on Maverick?

    I need help learning Numbers. I have a very basic understanding of MS Excel. Is there a tutorial targeted to financial spreadsheets?

  • Making money using FCE/FCP (I'm getting ahead of myself!)

    I received a quick and friendly response to my last question so I think I will try this again. Is it hard making a living in the film/video industry as an editor? I am at a four year university and I have to decide my major next month. Originally I w

  • N85-how to use SIP setting and use Voice over IP t...

    Dear Nokia user In Nokia N85 setting there is an option foe SIP setting.I want to know how can I use this option to register in a SIP proxy server and have an VOIP converstion over internet by my N85 small_nokia_supporter

  • Sending mail for failed idocs

    hi is there any standard function module to send mail to the receipients if any idocs has been failed during the process. Please revert back. Points will be awarded for the appropriate answers. Thanks & Regards Varalakshmi.K

  • Contacts Smart Groups Still Broken in Mountain Lion

    I have been waiting for Mountain Lion, in hope that Address Book / Contacts bugs would be fixed.  Regretfully, they have not.  Smart groups using the query "not a member of" still do not work.  It would have been wonderful it it did. Jack