If the system prameter has value bigger than 1 trap the current copie value

Hi.I have a report in which there is a requirement to be printed in two copies.So I put in the initial value of system parameter copies value 2.But I must print in the margin at the bottom of the page a label with TITLE-1 if we are in the first copie!!!, and a label with TITLE-2 if we are in the second copie.The report builder now always has the value copies=2.Is there some way to control the current copie value which means something like current page number in analogy current copie?
Thanks in advance thrylos97.

Hello,
The parameter COPIES is treated by the "Printer Driver" and not by the "Data Model" / "Data Layout"
If you want to be able to print Copy #1 / Copy #2 on the output, you have to create a repeating frame enclosing everything,
base this repeating frame on a sql query :
select "Copy 1" from dual union all select 'Copy 2' from dual
and put the Copy # on the Layout
Regards
Edited by: dsegard on Mar 30, 2009 11:26 AM

Similar Messages

  • I used scripting brigde to add a movie that has size bigger than 5GB, exactly after two minutes iTunes return a failed, but the processing of the file is actually added to iTunes Library successfully. The copying take more than 5 minutes to complete. Why?

    I used scripting brigde to add a movie that has size bigger than 5GB, exactly after two minutes iTunes return a failed, but the processing of the file is actually added to iTunes Library successfully. The copying take more than 5 minutes to complete. Why the iTunes Scripting Brigde returned failed when it is actually success? It occurred exactly 2 minutes after submit the request to Scripting Brigde. Is this 2 minutes related to the Apple Event time out? if it does, how do I get around this problem? thx

    I can tell you that this is some of the absolutely worst customer service I have ever dealt with. I found out from a store employee that when they are really busy with calls, they have third party companies taking overflow calls. One of those companies is Xerox. What can a Xerox call center rep possibly be able to authorize on a Verizon account?  I'm Sure there is a ton of misinformation out there due to this. They don't note the accounts properly or so everyone can see them. I have been transferred before and have asked if they work for Verizon or a third party also and was refused an answer so, apparently they aren't required to disclose that information. I spent a long time in the store on my last visit and it's not just customers that get the runaround. It happens to the store employees as well and it's beyond frustrating.

  • HT1657 I rented a movie on Itunes last night and I watched half of it before falling asleep. Today I tried to open the movie and watch the last half but I cannot find it. It has been less than 24 hours since I rented it. Where can I find the movie?

    I rented a movie on Itunes last night and I watched half of it before falling asleep. Today I tried to open the movie and watch the last half but I cannot find it. It has been less than 24 hours since I rented it. Where can I find the movie?

    I found my answer in another Thread and it worked... so for anyone who needs it, here it is:
    https://discussions.apple.com/message/19134562#19134562
    Had the same problem and spoke with Apple Support. They had me reset the System Management Controller:
    Shut down the computer. With the computer off, on the left hand side of the keyboard press shift+control+option along with the power button simultaneously and release. Wait 5-10 seconds and restart the computer.
    This solved my problem and I was immediately able to purchase, download, and play an HD movie from iTunes w/o any error message.

  • I am having a problem with my cs6 running very slow and when i save i am getting an error message that says, "This document is bigger than 2 gb (the file may not be read correctly by some TIFF viewers.) please help

    I am having a problem with my cs6 running very slow and when i save i am getting an error message that says, "This document is bigger than 2 gb (the file may not be read correctly by some TIFF viewers.) please help

    wen6275,
    It looks like you're actually using a camera or phone to take a photo of your monitor.
    That's not what is commonly known as a "screen shot". 
    It looks like you're on a Mac.  Hitting Command+Shift+3 places a capture of the contents of your entire monitor on your Desktop; hitting Command+Shift+4 will give you a cross-hairs cursor for you to select just the portion you want to capture.
    I'm mentioning this because I fear I'm not construing your original post correctly when you type "I am working with a large files [sic], currently it has 149 layers, all of which are high resolution photographs", so I'm wondering if there's some similar use of your own idiosyncratic nomenclature in your description of your troublesome file.
    Perhaps I'm just having a major senior moment myself, but I'm utterly unable to imagine such a file.  Would you mind elaborating?

  • PS CS6 on the mac pro: the bruch in hard and soft is the circle of the bruch bigger than then the effect. How can i bring it to normale?       /Users/jorisneyt/Desktop/Schermafbeelding 2014-11-07 om 10.04.59.png

    PS CS6 on the mac pro: the bruch in hard and soft is the circle of the bruch bigger than then the effect. How can i bring it to normale?

    Go to System Preferences>Accessibility>Display and the set the Cursor Size to Normal

  • Why are the PDF files in v27 bigger than in v25?

    I have just created a v27 version of my single iOS app, which has been v25 before. Now the app has a size of about 550 MB instead of about 430 MB. I found out that the PDF files of the slideshows are bigger  than before. What is the reason for this change and is there a way to get smaller file sizes again?
    Kind regards
    Michael
    No idea? At least someone of the staff should know why the DPS creates larger files now.
    There is another issue which indicates that Adobe doesn't care about file sizes. In slideshows the DPS always uses PNG files (if you don't choose PDF files, which are larger now, as I have told you). Of course it is because of transparency, but it would be easy to let the user choose between PNG and JPG like Twixl does.
    Message was edited by: Michael Kokoscha

    You should report this via Gold Support who will investigate this. Minor file size changes will always occure on releases, however since you report the size change is significant, there may be some issue and needs further testing.

  • How can I make invoice when down payment is bigger than item value ?

    Dear all,
    please, I need make invoice with bigger down payment than the item value.
    How can I make invoice when down payment clearing value is bigger than the item value ?
    When I do Release To Accounting, I have got a massage:
    "Total of down payment to be billed too high. See billing document item XXXXX"
    Please, how can I solve this problem?
    Thank you
    Peter

    Hello,
      I would like to underline that this is not a problem, but a standard, and in my opinion rightful behaviour of SAP: in a dowpayment process, when you create the invoice document, you should deduct only an amount less or equal to the invoice amount. And in the following invoices you will deduct the remaining downpayment amounts still open.
    Also under al legal and fiscal point of view, I guess how would you print the invoice? With a minus value?
    Best regards,
    Andrea

  • Other system has higher release than XI syst. in landscape

    Hi,
    We are running an XI 3.0/NW04 system and we're currently planning to upgrade our BW 3.5/NW04 system to BI 7.0/NW04s.
    In the Master for NW04s you can read the following :
    "For PI, it is a prerequisite that no other system in your system landscape has a higher release than the PI system. If you want to upgrade or install an application in your system landscape, you first have to make sure that the current release of the PI system is on the same release level ‒ if required, you have to upgrade the PI system first to the new or a higher release."
    After having read this, my understanding is that we have to upgrade our XI system before upgrading the BW system.
    What are the consequences/implications running a system landscape where a system has a higher release than the XI-system ?
    Thanks in advance for help and hints.
    Rgds,
    Christian

    Hi Christian,
    The implications are, PI as the highest release is the only supported combination.  It is possible that other combinations may work, but they are not tested by SAP and, in the case of any issues, SAP Support will tell you to upgrade your PI system to the highest release.
    Best Regards,
    Matt

  • Index size (row_nums) is bigger than the tables row

    Hi everyone,
    I'm encountering some strange problems with the CBO in Oracle 10.2.0.3 - it's telling me that I have more rows in the indexes than there are rows in the tables.
    I've tried all combinations of dbms_stats and analyse and cannot understand how the CBO comes up with such numbers. I've even done a "delete statistics" and
    Re-analysed the table and indexes but it doesn't help.
    The command I used is variations of the following:
    exec
    DBMS_STATS.GATHER_TABLE_STATS(ownname=>'MBS',tabname=>'READINGTOU', -
    estimate_percent=>dbms_stats.auto_sample_size,method_opt=>'FOR COLUMNS PROCESSSTATUS',degree=>2);
    EVEN TRIED
    exec sys.dbms_utility.analyze_schema('MBS','ESTIMATE', estimate_percent => 15);
    I've even used estimate_percent of 50 and still getting lower numbers for the table.
    Initially I was afraid that since the index is larger than the table, the index would never be used. So the question is, does it really matter that the indexes' num_rows is bigger than the tables' num_rows? What is the consequence of this? And how do I get the optimizer to correct the differences in the stats. The table is 30G in size and growing, so a COMPUTE is out of the question.
    but have the same problem in dev..and i did the COMPUTE in dev...get the same thing... I have more rows in the indexes than there are rows in the tables
    Edited by: user630084 on Mar 11, 2009 10:45 AM

    Is your issue that you are having problems with the execution plans of queries referencing these objects? Or is your problem that you are observing more num_rows in the index than in the table when you query the data dictionary?
    If it's the latter then there's really no concern (unless the estimates are insanely inconsistent). The statistics are estimates and as such, will not be 100% accurate, though they should do a reasonable job of representing the data in your system (when they don't, then you have an issue, but we've seen nothing to indicate that as of yet).

  • LOB segment size is 2 times bigger than the real data

    here's an interesting test:
    1. I created a tablespace called "smallblock" with 2K blocksize
    2. I created a table with a CLOB type field and specified the smallblock tablespace as a storage for the LOB segment:
    SCOTT@andrkydb> create table t1 (i int, b clob) lob (b) store as
    t1_lob (chunk 2K disable storage in row tablespace smallblock);
    3. I insert data into the table, using a bit less than 2K of data for the clob type column:
    SCOTT@andrkydb> begin
    2 for i in 1..1000 loop
    3 insert into t1 values (mod(i,5), rpad('*',2000,'*'));
    4 end loop;
    5 end;
    6 /
    4. Now I can see that I have an average of 2000 bytes for each lob item:
    SCOTT@andrkydb> select avg(dbms_lob.getlength(b)) from t1;
    AVG(DBMS_LOB.GETLENGTH(B))
    2000
    and that all together they take up:
    SCOTT@andrkydb> select sum(dbms_lob.getlength(b)) from t1;
    SUM(DBMS_LOB.GETLENGTH(B))
    2000000
    But when I take a look at how much is the LOB segment actually taking, I get a result, which is being a total mystery to me:
    SCOTT@andrkydb> select bytes from dba_segments where segment_name = 'T1_LOB';
    BYTES
    5242880
    What am I missing? Why is LOB segment is being ~2 times bigger than it is required by the data?
    I am on 10.2.0.3 EE, Solaris 5.10 sparc 64bit.
    Message was edited by:
    Andrei Kübar

    thanks for the link, it is good to know such thing is possible. Although I don't really see how can it help me..
    But you know, you were right regarding the smaller data amounts. I have tested with 1800 bytes of data and in this case it does fit just right.
    But this means that there is 248 bytes wasted (from my, as developer, point of view) per block! But if there is such an overhead, then I must be able to estimate it when designing the data structures. And I don't see anywhere in the docs a single word about such thing.
    Moreover, if you use NCLOB type, then only 990 bytes fits into a single 2K chunk. So the overhead might become really huge when you go over to gigabyte amounts...
    I have a LOB segment for a nclob type field in a production database, which is 5GB large and it contains only 2,2GB of real data. There is no "deleted" rows in it, I know because I have rebuilt it. So this looks like a total waste of disk space... I must say, I'm quite disappointed with this.
    - Andrei

  • How do i make the whole veiw of the computer smaller? it seems bigger than the whole screen

    how do i make the whole veiw of the computer smaller? it seems bigger than the whole screen

    often, this happens when you blunder into Accessibility/ Universal Access features, such as screen Zooming.
    System Preferences > Accessibility/ Universal Access > Seeing

  • Bug in MarlinCore script when value of rangesize attribut bigger than 20

    Hi,
    I have web application based in BC4j/Struts/uix,
    and I use <jbo:datasource> tag.
    I get javascript error in IE version6 when I change the value of rangesize attribut to number bigger than 20.
    the error happen at line 823 in MarlinCore.js :
    a0.submit();
    anyone have idea about the problem !!!
    thanks in advance
    Med

    It isn't a bug in MarlinCore. You just need to set the <form>s method to POST.

  • Problem happening again. The maximum report processing jobs limit configured by your system administrator has been reached.

    We have started receiving load report failed error again in one of our production servers. This problem occurred some months ago and on that time we found that in the code reportdocument.close method was not called so we added into the code and set the print job limit to 400 and problem went away for couple of months but now it has started again.
    Load report failed.
       at CrystalDecisions.ReportAppServer.ReportClientDocumentWrapper.EnsureDocumentIsOpened()
       at CrystalDecisions.CrystalReports.Engine.ReportDocument.Load(String filename, OpenReportMethod openMethod, Int16 parentJob)
       at CrystalDecisions.CrystalReports.Engine.ReportDocument.Load(String filename)
       at AppSuite.WebApp.CrystalViewer.LoadReport(Boolean bRefresh)
       at AppSuite.WebApp.CrystalViewer.Page_Load(Object sender, EventArgs e)
       at System.Web.UI.Control.OnLoad(EventArgs e)
       at System.Web.UI.Control.LoadRecursive()
       at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint)
    The maximum report processing jobs limit configured by your system administrator has been reached.
       at CrystalDecisions.ReportAppServer.ClientDoc.ReportClientDocumentClass.Open(Object& DocumentPath, Int32 Options)
       at CrystalDecisions.ReportAppServer.ReportClientDocumentWrapper.Open(Object& DocumentPath, Int32 Options)
       at CrystalDecisions.ReportAppServer.ReportClientDocumentWrapper.EnsureDocumentIsOpened()
    This is our code
    Private m_oReportDocument As New ReportDocument
        Private ReadOnly Property ReportId() As Integer
            Get
                Return Integer.Parse(Me.Request.QueryString("r"))
            End Get
        End Property
        Private Property CacheKey() As String
            Get
                Dim sCacheKey As String = Me.Request.QueryString("ck")
                If sCacheKey Is Nothing OrElse sCacheKey.Length = 0 Then
                    sCacheKey = CStr(Me.ViewState("CacheKey"))
                End If
                Return sCacheKey
            End Get
            Set(ByVal Value As String)
                Me.ViewState("CacheKey") = Value
            End Set
        End Property
        Private Sub Page_Load(ByVal sender As System.Object, ByVal e As System.EventArgs) Handles MyBase.Load
            If Not Me.IsPostBack Then
                LoadReport()
            End If
        End Sub
        Private Sub btnPDF_Click(ByVal sender As System.Object, ByVal e As System.EventArgs) Handles btnPDF.Click
            'Redirect to the PDF Viewer passing it the ReportId and CacheKey
            Me.Response.Redirect(String.Format(ReportHelper.PDFViewerURL, Me.ReportId, Me.CacheKey))
        End Sub
        Private Sub btnRTF_Click(ByVal sender As System.Object, ByVal e As System.EventArgs) Handles btnRTF.Click
            'Redirect to the RTF Viewer passing it the ReportId and CacheKey
            Me.Response.Redirect(String.Format(ReportHelper.RTFViewerURL, Me.ReportId, Me.CacheKey))
        End Sub
        Private Sub btnExcel_Click(ByVal sender As System.Object, ByVal e As System.EventArgs) Handles btnExcel.Click
            'Redirect to the Excel Viewer passing it the ReportId and CacheKey
            Me.Response.Redirect(String.Format(ReportHelper.ExcelViewerURL, Me.ReportId, Me.CacheKey))
        End Sub
        Private Sub crvMain_Navigate(ByVal source As Object, ByVal e As CrystalDecisions.Web.NavigateEventArgs) Handles crvMain.Navigate
            LoadReport()
        End Sub
        Private Sub crvMain_Search(ByVal source As Object, ByVal e As CrystalDecisions.Web.SearchEventArgs) Handles crvMain.Search
            LoadReport()
        End Sub
        Private Sub crvMain_ViewZoom(ByVal source As Object, ByVal e As CrystalDecisions.Web.ZoomEventArgs) Handles crvMain.ViewZoom
            LoadReport()
        End Sub
        Private Sub crvMain_Drill(ByVal source As Object, ByVal e As CrystalDecisions.Web.DrillEventArgs) Handles crvMain.Drill
            LoadReport()
        End Sub
        Private Sub crvMain_DrillDownSubreport(ByVal source As Object, ByVal e As CrystalDecisions.Web.DrillSubreportEventArgs) Handles crvMain.DrillDownSubreport
            LoadReport()
        End Sub
        Private Sub crvMain_ReportRefresh(ByVal source As Object, ByVal e As CrystalDecisions.Web.ViewerEventArgs) Handles crvMain.ReportRefresh
            LoadReport(True)
        End Sub
        Private Sub LoadReport()
            LoadReport(False)
        End Sub
        Private Sub LoadReport(ByVal bRefresh As Boolean)
            If Common.CouldBeMultiDB(User.Identity.Name) AndAlso TypedSession.OverrideCompany.Length > 0 Then
                Common.Settings.OverrideCompany = TypedSession.OverrideCompany
            End If
            'Get the report data
            Dim dtReport As DataTable = ReportHelper.GetReportData(Me.CacheKey, bRefresh)
            'If there is data to display bind it to the Crystal Viewer
            If dtReport.Rows.Count > 0 Then
                With m_oReportDocument
                    .Load(ReportHelper.GetReportPath(Me.ReportId))
                    .SetDataSource(dtReport)
                    .PrintOptions.PaperSize = Common.Settings.CrystalPaperSize
                End With
                crvMain.ReportSource = m_oReportDocument
            Else
                'Hide the controls and display a message if there is no data
                crvMain.Visible = False
                btnPDF.Visible = False
                btnExcel.Visible = False
                lblNoResults.Visible = True
            End If
        End Sub
        Protected Sub Page_Unload(ByVal sender As Object, ByVal e As System.EventArgs) Handles Me.Unload
            m_oReportDocument.Close()
            m_oReportDocument.Dispose()
        End Sub
    Can any one tell if we are doing anything wrong in our code. We don't use sub reports any more however we do use paging and some of our reports have 200+ pages. Also is it possible to find out print job limit and concurrent users by writing some code.
    Thanks

    I suspect it's not the code - after all it worked(?). It is much more likely that, it is the load. Remember that the report engine can only handle three concurrent requests at any one time. Also, remember that you can run out of print jobs real fast as subreports also count as print job. E.g.; say someone added a report that has 4 subreports in the detail section, the report runs a 100 records, you will be at 401 print jobs and thus over the limit. You can keep increasing the print job limit, but that means you are loading the server and eventually you will bring it down too...
    It may be time to consider, web farms or Report Application Server to scale up(?).
    Getting to the latest SP is never a bad idea either.
    And have a look at the scalability suggestions in this doc:
    Crystal Reports 2008 Component Engine Scalability
    - Ludek
    Senior Support Engineer AGS Product Support, Global Support Center Canada
    Follow us on Twitter

  • Limit the no of segments per IDOC to less than the current default value

    Hi friends,
    please help me in the following issue.
    The possibility to limit the number
    of segments per IDOC to less than the current default value.
    Purpose is to limit the traffic of IDOCs between systems.
    How to solve this issue to avoid traffic
    Thanks in advance.
    Regards
    Sandeep

    Hi Kiran,
    Actually we are using standard ones at that time we are getting prolems. When using custom developed ALE's
    we are putting segment size is less.To decrease segment size is there any filtering at the sender side? To change segment size may be nedd to take help form SAP.
    But if have filtering at any level from sender side need to confirm before consult SAP.
    Regards
    Sandeep

  • When my system upgrade from os x 10.9.4 to os x 10.10.1, the time of shot down in my system is very long rather than pervious os. my system is macbook pro 11.1

    when my system upgrade from os x 10.9.4 to os x 10.10.1, the time of shot down in my system is very long rather than pervious os. my system is macbook pro 11.1.
    please inform me about this problem

    Jan  8 18:36:31 localhost kernel[0]: Previous shutdown cause: 5
    Jan  8 18:37:06 Jashan-Mr-2.local shutdown[927]: reboot by apll:
    Jan  8 18:37:06 Jashan-Mr-2.local shutdown[927]: SHUTDOWN_TIME: 1420729626 658562
    Jan  8 18:39:44 localhost com.apple.xpc.launchd[1] (com.parallels.desktop.launchdaemon): Service has increased maximum shutdown time to 150 seconds. This will create a poor experience for the user.
    Jan  8 18:39:45 localhost kernel[0]: Previous shutdown cause: 5
    Jan  8 19:02:58 Jashan-Mr-2.local shutdown[1286]: halt by apll:
    Jan  8 19:02:58 Jashan-Mr-2.local shutdown[1286]: SHUTDOWN_TIME: 1420731178 850680
    Jan  8 19:02:58 Jashan-Mr-2 com.apple.xpc.launchd[1] (com.apple.xpc.launchd.domain.system): System reboot initiated by: shutdown.1286<-sessionlogoutd.1285<-launchd.1
    Jan  8 19:13:09 Jashan-Mr-2 kernel[0]: Previous shutdown cause: 5
    Jan  8 20:11:44 Jashan-Mr-2.local shutdown[1927]: halt by apll:
    Jan  8 20:11:44 Jashan-Mr-2.local shutdown[1927]: SHUTDOWN_TIME: 1420735304 825738
    Jan  8 20:11:44 Jashan-Mr-2 com.apple.xpc.launchd[1] (com.apple.xpc.launchd.domain.system): System reboot initiated by: shutdown.1927<-sessionlogoutd.1926<-launchd.1
    Jan  9 12:06:41 localhost com.apple.xpc.launchd[1] (com.parallels.desktop.launchdaemon): Service has increased maximum shutdown time to 150 seconds. This will create a poor experience for the user.
    Jan  9 12:06:41 Jashan-Mr-2 kernel[0]: Previous shutdown cause: 5
    Jan  9 12:25:15 Jashan-Mr-2.local shutdown[1868]: halt by apll:
    Jan  9 12:25:15 Jashan-Mr-2.local shutdown[1868]: SHUTDOWN_TIME: 1420793715 291375
    Jan  9 12:25:15 Jashan-Mr-2 com.apple.xpc.launchd[1] (com.apple.xpc.launchd.domain.system): System reboot initiated by: shutdown.1868<-sessionlogoutd.1867<-launchd.1
    Jan  9 12:28:47 localhost com.apple.xpc.launchd[1] (com.parallels.desktop.launchdaemon): Service has increased maximum shutdown time to 150 seconds. This will create a poor experience for the user.
    Jan  9 12:28:48 Jashan-Mr-2 kernel[0]: Previous shutdown cause: 5

Maybe you are looking for