Large number of movieclip slowdown the application

Hi,
I have a scrollbar function.In the content area i attached movieclips.Those moviclips have dynamic textfield.The scroll bar working for small list of contents.up to 500 or 600 moviclips
When i attached 2000 or larger number of movieclip the Entire flash get slows. How to i fasten them with same number of movieclip.
why flash player slow when attaching large number of movieclip? it takes near by 58% cpu usage at this time.What is the reason for this?
Thanks,
Siva

imm also if you put much media inside the flash library (Audio,images, videos) those will be uploaded on Ram,,,
so instead you can export your library as external library
or
use AS3 loader to load your media, immutablility will be fine here!! ^_^
Goodluck

Similar Messages

  • How do i go about deleting a large number of files at the same time?  where's the easiest place to do it?

    how do i go about deleting a large number of files at the same time?  where's the easiest place to do it?

    A bit vague as to what you intend, but the simple answer is to select all the files you want to delete then either drag to the Trash or CTRL- or RIGHT-click on the selection and choose Move to Trash from the contextual menu.

  • How to add a large number of keywords to the e-mail filter?

    Hello.
    I would like to know how to add a large number of keywords to a filter.
    The thing I want to accomplish, is it to add around 4000 e-mail addresses to a filter list, which checks the body of incoming e-mails, which get forwarded to me.
    I don't want to outright delete them, but I would love it if it detects that the forwarded message contains one of the e-mail addresses, it would add a tag to the message.
    Is it in any way possible to make a filter like this, which doesn't slow Thunderbird down to ass-crawl speed?
    I tried to copy the whole list into the small filter tab, but It had no discernible effect on my messages, since some of the previously received ones, which I was sure contained the keywords, were not tagged. All it did was make the program super slow and I was forced to delete the filter.

    You can look at creating a exclusion crawl rule:
    http://technet.microsoft.com/en-us/library/jj219686(v=office.15).aspx
    You can also modify your content source starting addresses and remove onedrive:
    http://technet.microsoft.com/en-us/library/jj219808(v=office.15).aspx
    Blog | SharePoint Field Notes Dev Tools |
    SPFastDeploy | SPRemoteAPIExplorer

  • LENGTH function for a large number returns 40 not the number of digits

    In SQL*Plus:
    SQL> select length(12345678901234567890123456789012345678901234567890)lngth
    2 from dual;
    LNGTH
    40
    SQL> select length('12345678901234567890123456789012345678901234567890')lngth
    2 from dual;
    LNGTH
    50
    It seems that the implicit conversion from number to char in the first query causes an unexpected result. From the documentation of the LENGTH function at http://download.oracle.com/docs/cd/B19306_01/server.102/b14200/functions181.htm#i79330:
    "If you omit fmt, then n is converted to a VARCHAR2 value exactly long enough to hold its significant digits."
    Perhaps Oracle considers that only 40 digits are significant (?)

    Additional reason.
    There is implicit data conversion as like to_char(number).
    SQL> set numwidth 50
    SQL> select 12345678901234567890123456789012345678901234567890 num from dual;
                                                   NUM
    12345678901234567890123456789012345678900000000000
    SQL> select length(12345678901234567890123456789012345678901234567890) numlen from dual;
                                                NUMLEN
                                                    40
    SQL> select to_char(12345678901234567890123456789012345678901234567890) numlen from dual;
    NUMLEN
    1.2345678901234567890123456789012346E+49
    SQL> select length(to_char(12345678901234567890123456789012345678901234567890)) numlen from dual;
                                                NUMLEN
                                                    40
    SQL> select to_char(12345678901234567890123456789012345678901234567890,
      2                '09000000000000000000000000000000000000000000000000')
      3  num from dual;
    NUM
    12345678901234567890123456789012345678900000000000
    SQL> select length(to_char(12345678901234567890123456789012345678901234567890,
      2                '09000000000000000000000000000000000000000000000000'))
      3  num from dual;
                                                   NUM
                                                    51
    SQL> select to_char(12345678901234567890123456789012345678901234567890,
      2                'fm09000000000000000000000000000000000000000000000000')
      3  num from dual;
    NUM
    12345678901234567890123456789012345678900000000000
    SQL> select length(to_char(12345678901234567890123456789012345678901234567890,
      2                'fm09000000000000000000000000000000000000000000000000'))
      3  num from dual;
                                                   NUM
                                                    50

  • Large attachment in mail stalls the application. Please help!

    I accidentally tried to send a 10mb email. I realized the issue immediately and tried to delete the email from Outbox, which didn't show up till later. And when it did I deleted the large email in the Outbox. Or so I thought.
    Next, my Internet got knocked out at some point and I had to reconnect.
    Now, I'm back on line. Mail is not working at all on its own. When I try to send a test mail to myself, the Sent button gets busy and it reproduces the same copy of the 10mb email under On My Mac category and in Sent Messages. And my test email would sit in the Outbox.
    Meanwhile nothing else is working. I have no mail coming in and obviously cannot send anything out. Please help.
    BTW, I have already tried to go into Finder/home/library/Mail and tried deleting Outbox.mbox and Sent Messages.mbox. When I relaunch Mail, the same problem persists.
    If you know how to fix this, I would love to know. Thanks!

    Hello leaflet:
    Try a couple of things:
    Select the mailbox giving you problems. Select mailbox>rebuild from the drop down menu.
    If that does not help, move the preference file (com.apple.mail.plist) to the trash. Do not empty the trash just yet. Restart mail and see if the problem is corrected. You will need to reenter your personal preferences.
    Barry

  • PIXMA MP610: Can Printing a large number of copies damage the printer head

    Hi!I want some suggestion, if some one give me nice suggestion it will be best for meI bought Canon Mp610 , for my thesis (which contain 400 pages inculding color and black pages), and require 10 copies meanwhile, 4000 pagesI decide print to thesis with 1, or in 5 actions , I need suggestion which  will be best for me  to print collective 5 copies or print one or two copy then give it rest for 15-30 minutes and then give action for other copy, but is time consuming,  but my colleague said me if you print 5 copies in signle or multiple actions, it  damage the printer head, and you will loss your printer. He said me to use Laer jet for black portion and for color use inkjet printer. I am worried about his suggestion because i will waste my lot pages which will be print in black then I will replace them with color pages. Please give me your kind suggestion Please  RegardsMuqeet Soomro 

    I know what you mean, this has to happen every month, and for some reason (orders from the bosses), i have to be at work until the whole job has finished printing incase a batch goes wrong so we can imediatly start reprinting, it used to be only from the Friday night until early Monday morning, so I'm also trying to get the job done a bit quicker, but i can't see that happening.
    The problem is that its all together 150 000 documents that we need to print that will be split up into these batches (within 5 days, monthly), the printer that we are printing to is an OCE, so it can handle big batches. the thing is that we use to print it from an AS400 through DOCOUT, so while the printer was printing one of these batches, it was busy spooling the next batch, which made it a bit quicker, but now this process has to move over to JAVA, and if i can the program to spool the jobs, i can start spooling while the previous one is printing. i even thought about writing a batch through the socket, in theory its almost like spooling, but i don't know how much RAM the OCE has.
    I also thought of merging the POSTSCRIPTS into one document, but that will take alot of time and space.

  • Which is the best way for posting a large number of records?

    I have around 12000 register to commit to dababase.
    Which is the best way for doing it?
    What depends on ?
    Nowadays I can't commit such a large number of register..The dabatase seems hanged!!!
    Thanks in advance

    Xavi wrote:
    Nowadays I can't commit such a large number of registerIt should be possible to insert tens of thousands of rows in a few seconds using an insert statement even with a complex query such as the all_objects view, and commit at the end.
    SQL> create table t as select * from all_objects where 0 = 1;
    Table created.
    Elapsed: 00:00:00.03
    SQL> insert into t select * from all_objects;
    32151 rows created.
    Elapsed: 00:00:09.01
    SQL> commit;
    Commit complete.
    Elapsed: 00:00:00.00
    I meant RECORDS instead of REGISTERS.Maybe that is where you are going wrong, records are for putting on turntables.

  • Number of user sessions in the Application Server for a config entry

    Hi All
    I am using 10g AS, with 9i DB.
    My requirement is that -
    i need to find the total number of users of the application in any moment, and to
    display that information to the users. say, config=abc, super users need to find all
    those who are using the link from the application itself.
    I can see that information from the 10 g Enterprise Manager -> Forms -> User Sessions.
    But is there any means of transfering that information to forms.
    Can this be done by using PJC's ?
    Please post your solutions/suggestions
    Thanks in Advance
    Deepa

    Any help/ideas Please
    Deepa

  • Passing large number of parameters

    In chapter 6 of the reports 6i documentation it states that one of the advantages of using a ref cursor is to "avoid the use of lexical parameters in your reports". I am not clear how it does this. Can anyone explain how this would work?
    In my application, I would like to print an exception report (ie use run_report) when an item in the base report exceeds a threshold amount. I would rather not pass all the subtotal values, as that would require passing a large number of parameters to the exception report. I can do it, but it seems like there should be an easier way. Passing a parameter of type RECORD would do it, but isn't allowed, right?
    null

    1. In some cases, you can avoid the use of lexical parameters in your reports
    with "static" ref cursor.
    Example for "static" ref cursor:
    1.1 Stored package
    CREATE OR REPLACE PACKAGE report_static IS
    TYPE type_ref_cur_sta IS REF CURSOR RETURN dept%ROWTYPE;
    FUNCTION func_sta (p_order_by VARCHAR2) RETURN type_ref_cur_sta;
    END;
    CREATE OR REPLACE PACKAGE BODY report_static IS
    FUNCTION func_sta (p_order_by VARCHAR2) RETURN type_ref_cur_sta IS
    ref_cur_sta type_ref_cur_sta;
    BEGIN
    IF p_order_by = 'dname' THEN
    OPEN ref_cur_sta FOR
    SELECT * FROM dept ORDER BY dname;
    ELSE
    OPEN ref_cur_sta FOR
    SELECT * FROM dept ORDER BY deptno;
    END IF;
    RETURN ref_cur_sta;
    END;
    END;
    1.2 Query PL/SQL in Reports
    function QR_1RefCurQuery return report_static.type_ref_cur_sta is
    begin
    return report_static.func_sta (:p_order_by);
    end;
    2. But if you need (for example) dynamic where, you (practically)
    can't use "static" ref cursor. In this case, you can use "dynamic" ref cursor
    (available from Oracle 8.1.5).
    Example for "dynamic" ref cursor (note that Reports need
    "static" ref cursor type for building Report Layout):
    2.1 Stored package
    CREATE OR REPLACE PACKAGE report_dynamic IS
    TYPE type_ref_cur_sta IS REF CURSOR RETURN dept%ROWTYPE; -- for Report Layout only
    TYPE type_ref_cur_dyn IS REF CURSOR;
    FUNCTION func_dyn (p_where VARCHAR2) RETURN type_ref_cur_dyn;
    END;
    CREATE OR REPLACE PACKAGE BODY report_dynamic IS
    FUNCTION func_dyn (p_where VARCHAR2) RETURN type_ref_cur_dyn IS
    ref_cur_dyn type_ref_cur_dyn;
    BEGIN
    OPEN ref_cur_dyn FOR
    'SELECT * FROM dept WHERE ' | | NVL (p_where, '1 = 1');
    RETURN ref_cur_dyn;
    END;
    END;
    2.2 Query PL/SQL in Reports
    function QR_1RefCurQuery return report_dynamic.type_ref_cur_sta is
    begin
    return report_dynamic.func_dyn (:p_where);
    end;
    Regards

  • How to let the user know the application is still running during SQLite reads

    I'm using an SQLite database to read and write to my application. In some places it has to do very large reads and writes and the application comes to a standstill for like 10 seconds while this is happening. I've tried calling showBusyCursor(), but it only works if I call that function and then set a timer to execute the sql query. But the busy cursor also freezes once the connection starts.
    Why doesn't the showBusyCursor work during this time? Are there any other things I can do to the application to ensure the user that it is not frozen and just processing?

    Keep in mind that I haven't actually implemented this yet, but my path is headed right in that direction in the next few days. From wh
    at I have read you can access the database in two modes: synchronous and asynchronous. From your description it sounds like you are accessing the database in synchronous mode, which will stop the application thread until the database thread is completed.
    To get around that limitation you will want to run your database accesses in asynchronous mode, so that the application will continue running while the database is being accessed. What you do then is implement a listener on the database that will call a function that will hide the waiting cursor or image. Then you start the waiting indicator in the application. Once the database is done, it will trigger your listener and remove the waiting indicator, allowing the user to once again interact with your app.
    Hope this helps!
    ~Mike

  • Some of the applications on my computer were downloaded by a different Apple ID so I am unable to update them. Is there a way to look up who originally purchased them?

    I'm sorry if this doesn't make very much sense but it's an incredibly aggravating issue.
    My computer was purchased by a separate account with iPhoto, Pages, Keynote, Numbers, and other additional applications already installed. I have been trying to figure out which account it is by checking old e-mails, trying different account names, and more. I was wondering if there was a way I could see what account purchased my computer/applications in the first place - like some kind of purchase history based on the serial number of Mac or the applications having some way to retreive that information. But I can't seem to do it on my own. Any recommendations? I really don't want to delete and repurchase all of the applications. Not only that, but my iPhoto is unable to open due to my recent update to the OSX Mavericks, and I don't want to lose those photos! 
    Thanks so much!

    You can't merge accounts. But you can check your purchase history:
    iTunes Store & Mac App Store: Seeing your purchase history and order numbers
              http://support.apple.com/kb/HT2727
    Also, what may seem stupid to you... may be a protection of privacy to others.

  • Large number of event Log entries: connection open...

    Hi,
    I am seeing a large number of entries in the event log of the type:
    21:49:17, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [81.154.101.160:51163] CLOSED/TIME_WAIT ppp0 NAPT)
    21:49:15, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [81.154.101.160:41820] ppp0 NAPT)
    Are these anything I should be concerned about? I have tried a couple of forum and Google searches, but I don't quite know where to start beyond pasting the first bit of the message. I haven't found anything obvious from those searches.
    DHCP table lists 192.168.1.78 as the desktop PC on which I'm writing this.
    Please could you point me in the direction of any resources that will help me to work out if I should be worried about this?
    A slightly longer extract is shown below:
    21:49:17, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [81.154.101.160:51163] CLOSED/TIME_WAIT ppp0 NAPT)
    21:49:15, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [81.154.101.160:41820] ppp0 NAPT)
    21:49:15, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [81.154.101.160:51163] CLOSED/SYN_SENT ppp0 NAPT)
    21:49:11, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [213.205.231.156:51027] TIME_WAIT/CLOSED ppp0 NAPT)
    21:49:03, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [178.190.63.75:55535] CLOSED/SYN_SENT ppp0 NAPT)
    21:49:00, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [2.96.4.85:23939] TIME_WAIT/CLOSED ppp0 NAPT)
    21:48:59, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [78.144.143.222:21617] CLOSED/TIME_WAIT ppp0 NAPT)
    21:48:58, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [41.218.222.34:28188] ppp0 NAPT)
    21:48:57, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [41.218.222.34:28288] CLOSED/SYN_SENT ppp0 NAPT)
    21:48:57, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [86.132.123.255:18048] ppp0 NAPT)
    21:48:57, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [86.132.123.255:54199] CLOSED/SYN_SENT ppp0 NAPT)
    21:48:55, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [86.144.91.49:60704] ppp0 NAPT)
    21:48:55, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [80.3.100.12:50875] TIME_WAIT/CLOSED ppp0 NAPT)
    21:48:45, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [78.150.251.216:57656] ppp0 NAPT)
    21:48:39, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [78.150.251.216:56975] CLOSED/SYN_SENT ppp0 NAPT)
    21:48:29, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [79.99.145.46:8368] CLOSED/SYN_SENT ppp0 NAPT)
    21:48:27, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [90.192.249.173:45250] ppp0 NAPT)
    21:48:16, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [212.17.96.246:62447] ppp0 NAPT)
    21:48:10, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [82.16.198.117:49942] TIME_WAIT/CLOSED ppp0 NAPT)
    21:48:08, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [213.205.231.156:51027] CLOSED/SYN_SENT ppp0 NAPT)
    21:48:04, 11 Mar.
    IN: ACCEPT [57] Connection closed (Port Forwarding: TCP 192.168.1.78:14312 <-->86.128.58.172:14312 [89.153.251.9:53729] TIME_WAIT/CLOSED ppp0 NAPT)
    21:47:54, 11 Mar.
    IN: ACCEPT [54] Connection opened (Port Forwarding: UDP 192.168.1.78:14312 <-->86.128.58.172:14312 [80.3.100.12:37150] ppp0 NAPT)

    Hi,
    Thank you for the response. I think, but can't remember for sure, that UPnP was already switched off when I captured that log. Anyway, even if it wasn't, it is now. So I will see what gets captured in my logs.
    I've just had to restart my Home Hub because of other connection issues and I notice that the first few entries are also odd:
    19:35:16, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49250->173.194.78.125:5222 on ppp0)
    19:34:45, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49266->173.194.34.101:443 on ppp0)
    19:34:31, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49250->173.194.78.125:5222 on ppp0)
    19:34:31, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49266->173.194.34.101:443 on ppp0)
    19:34:04, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49462->199.59.149.232:443 on ppp0)
    19:33:46, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49250->173.194.78.125:5222 on ppp0)
    19:33:46, 12 Mar.
    IN: BLOCK [12] Spoofing protection (IGMP 86.164.178.188->224.0.0.22 on ppp0)
    19:33:45, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49266->173.194.34.101:443 on ppp0)
    19:33:39, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49462->199.59.149.232:443 on ppp0)
    19:33:33, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49463->199.59.149.232:443 on ppp0)
    19:33:29, 12 Mar.
    IN: BLOCK [15] Default policy (UDP 111.252.36.217:26328->86.164.178.188:12708 on ppp0)
    19:33:16, 12 Mar.
    IN: BLOCK [15] Default policy (TCP 193.113.4.153:80->86.164.178.188:49572 on ppp0)
    19:33:14, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49266->173.194.34.101:443 on ppp0)
    19:33:14, 12 Mar.
    IN: BLOCK [15] Default policy (TCP 66.193.112.93:443->86.164.178.188:44266 on ppp0)
    19:33:14, 12 Mar.
    ( 164.240000) CWMP: session completed successfully
    19:33:13, 12 Mar.
    ( 163.700000) CWMP: HTTP authentication success from https://pbthdm.bt.mo
    19:33:05, 12 Mar.
    BLOCKED 106 more packets (because of Default policy)
    19:33:05, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49462->199.59.149.232:443 on ppp0)
    19:33:05, 12 Mar.
    IN: BLOCK [15] Default policy (TCP 213.1.72.209:80->86.164.178.188:49547 on ppp0)
    19:33:05, 12 Mar.
    BLOCKED 94 more packets (because of Default policy)
    19:33:05, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:33:05, 12 Mar.
    IN: BLOCK [15] Default policy (TCP 199.59.148.87:443->86.164.178.188:49531 on ppp0)
    19:33:05, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49250->173.194.78.125:5222 on ppp0)
    19:33:04, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49266->173.194.34.101:443 on ppp0)
    19:33:04, 12 Mar.
    ( 155.110000) CWMP: Server URL: https://pbthdm.bt.mo; Connecting as user: ACS username
    19:33:04, 12 Mar.
    ( 155.090000) CWMP: Session start now. Event code(s): '1 BOOT,4 VALUE CHANGE'
    19:32:59, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49266->173.194.34.101:443 on ppp0)
    19:32:54, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49462->199.59.149.232:443 on ppp0)
    19:32:53, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:32:52, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49463->199.59.149.232:443 on ppp0)
    19:32:51, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49266->173.194.34.101:443 on ppp0)
    19:32:48, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:32:47, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49266->173.194.34.101:443 on ppp0)
    19:32:46, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:32:46, 12 Mar.
    BLOCKED 4 more packets (because of First packet is Invalid)
    19:32:45, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49461->199.59.149.232:443 on ppp0)
    19:32:44, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:32:44, 12 Mar.
    BLOCKED 1 more packets (because of First packet is Invalid)
    19:32:43, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49398->193.113.4.153:80 on ppp0)
    19:32:42, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:32:42, 12 Mar.
    BLOCKED 3 more packets (because of First packet is Invalid)
    19:32:42, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49277->119.254.30.32:443 on ppp0)
    19:32:41, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:32:41, 12 Mar.
    BLOCKED 1 more packets (because of First packet is Invalid)
    19:32:41, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:32:38, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49280->119.254.30.32:443 on ppp0)
    19:32:36, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49330->173.194.67.94:443 on ppp0)
    19:32:34, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49463->199.59.149.232:443 on ppp0)
    19:32:30, 12 Mar.
    IN: BLOCK [15] Default policy (TCP 66.193.112.93:443->86.164.178.188:47022 on ppp0)
    19:32:30, 12 Mar.
    ( 120.790000) CWMP: session closed due to error: WGET TLS error
    19:32:30, 12 Mar.
    ( 120.140000) NTP synchronization success!
    19:32:30, 12 Mar.
    BLOCKED 1 more packets (because of Default policy)
    19:32:29, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49458->217.41.223.234:80 on ppp0)
    19:32:28, 12 Mar.
    OUT: BLOCK [65] First packet is Invalid (TCP 192.168.1.78:49280->119.254.30.32:443 on ppp0)
    19:32:26, 12 Mar.
    ( 116.030000) NTP synchronization start
    19:32:25, 12 Mar.
    OUT: BLOCK [15] Default policy (First packet in connection is not a SYN packet: TCP 192.168.1.78:49442->74.125.141.91:443 on ppp0)
    19:32:25, 12 Mar.
    OUT: BLOCK [15] Default policy (TCP 192.168.1.78:49310->204.154.94.81:443 on ppp0)
    19:32:25, 12 Mar.
    IN: BLOCK [15] Default policy (TCP 88.221.94.116:80->86.164.178.188:49863 on ppp0)

  • Large number of deadlocks seen after upgrade to 2.4.14

    We upgraded the BDB version to 2.4.14 and are using the latest 4.7 release. Without code change on our part we are seeing large number of deadlocks with the new version.
    Do let me know if more information is needed.
    BDB: 106 lockers
    BDB: 106 lockers
    BDB: 106 lockers
    BDB: 106 lockers
    BDB: Aborting locker 8000a651
    BDB: MASTER: /m-db/m-db432816-3302-4c30-9dd0-e42a295c970c/master rep_send_message: msgv = 5 logv 14 gen = 149 eid -1, type log, LSN [825][1499447]
    BDB: 107 lockers
    BDB: Aborting locker 8000a652
    BDB: 107 lockers
    BDB: MASTER: /m-db/m-db432816-3302-4c30-9dd0-e42a295c970c/master rep_send_message: msgv = 5 logv 14 gen = 149 eid -1, type log, LSN [825][1500259] perm
    BDB: MASTER: will await acknowledgement: need 1
    BDB: 106 lockers
    BDB: 106 lockers
    BDB: Aborting locker 8000a65a
    BDB: Aborting locker 8000a658
    BDB: MASTER: got ack [825][1500259](149) from site rit-004:10502
    BDB: 105 lockers
    BDB: 103 lockers
    BDB: Container - 5e69b5cf184b41ef8f0719e1b0f944a1.bdbxml - Updating document: 5ca1ab1e0a0571bf048c6e298618c7048c6e2ec315a3
    BDB: 104 lockers
    BDB: Container - 5e69b5cf184b41ef8f0719e1b0f944a1.bdbxml - Updating document: 5ca1ab1e0a0571bf048c6e298618c7048c6e2ec35d5d

    Also an interesting observation the replica process which not doing any thing except keeping up with the master is eating 3/4 times more cpu than the master when i am creating and updating records in the xml db.
    On a 4 cpu setup, the master process takes about half a cpu where as the replica is chewing upto 2 CPUs
    What is the replica doing which is this cpu intensive?!?!?!

  • Need to blacklist (block) a large number of domains and email addresses

    I tried to import a large number of domains into a transport rule (new/set-transportrule -name xxx -domainis $csv -quarantine $true). 
    I added a few domains manually and tested to make sure the rule was working.  It was.  Then I ran the cmdlet to import a small number of domains (4).  That also worked.  Then I ran the cmdlet on a large number of domains.  The cmdlet
    worked with no error message (once I got the list of domains well under 4096 bytes).  I waited a day to see if the domains would show up in the rule.  But the imported domains did not show up.
    Is there a better solution to blocking large numbers of domains and email addresses?  Or is the transport rule the only option? 

    Since you do not want to crop your images to a square 1:1 aspect ratio changing the canvas to be square will not make your images square they will retain their Aspect Ratio and  image size will be changer to fit within your 1020 px square. There will be a border or borders on a side or two borders on opposite sides.   You do not need a script because Photoshop ships with a Plug-in script to be used in Actions.   What is good about Plugins is the support Actions.  When you record the action the plug-in during action recording records the setting you use in its dialog into  the actions step.  When the Action is played the Plug-in use the recorded setting an bypasses displaying its dialog. So the Action can be Batch.  The Action you would record would have two  Steps.   Step 1  menu File>Automate>Fit Image... in the Fit Image dialog enter 1020 in the width and height  fields.  Step 2 Canvas size enter 1020 pixels in width and height  not relative leave the anchor point centered it you want even borders on two sides set color to white in the canvas size dialog. You can batch the action.
    The above script will also work. Its squares the document then re-sizes to 1020x1020  the action re-sizes the image to fit with in an area 1020 x 1020 then add any missing canvas. The script like the action only process one image so it would also need to be batched. Record the script into and action and batch the action. As the author wrote. The script re size canvas did not specify an anchor point so the default center anchor point is uses  like the action canvas will be added to two sides.

  • Rman backup failure, and is generating a large number of files.

    I would appreciate some pointers on this if possible, as I'm a bit of an rman novice.
    Our rman backup logs indicated a failure and in the directory where it puts its files there appeared a large number of files for the 18th, which was the date of the failure. Previous days backups generated 5 files of moderate size. When it failed it generated between 30 - 40 G of files ( it looks like one for each database file ).
    The full backup is early monday morning, and the rest incremental :
    I have placed the rman log, the script and a the full directory file listing here : http://www.tinshed.plus.com/rman/
    Thanks in advance - George
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734055071_s244_s1
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734055096_s245_s1
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734573008_s281_s1
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734055045_s243_s1
    -rw-r----- 1 oracle dba 524296192 Jan 18 00:03 database_f734055121_s246_s1
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734055020_s242_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054454_s233_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054519_s234_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054595_s235_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054660_s236_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054725_s237_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054790_s238_s1
    -rw-r----- 1 oracle dba 209723392 Jan 18 00:02 database_f734055136_s247_s1
    -rw-r----- 1 oracle dba 73408512 Jan 18 00:02 database_f734055143_s248_s1
    -rw-r----- 1 oracle dba 67117056 Jan 18 00:02 database_f734055146_s249_s1
    -rw-r----- 1 oracle dba 4194312192 Jan 18 00:02 database_f734054855_s239_s1
    -rw-r----- 1 oracle dba 2147491840 Jan 18 00:02 database_f734054975_s241_s1
    -rw-r----- 1 oracle dba 3221233664 Jan 18 00:02 database_f734054920_s240_s1
    drwxr-xr-x 2 oracle dba 4096 Jan 18 00:00 logs
    -rw-r----- 1 oracle dba 18710528 Jan 17 00:15 controlfile_c-1911789030-20110117-00
    -rw-r----- 1 oracle dba 1343488 Jan 17 00:15 database_f740621746_s624_s1
    -rw-r----- 1 oracle dba 2958848 Jan 17 00:15 database_f740621745_s623_s1
    -rw-r----- 1 oracle dba 6415990784 Jan 17 00:15 database_f740620829_s622_s1
    -rw-r----- 1 oracle dba 172391424 Jan 17 00:00 database_f740620814_s621_s1

    george3 wrote:
    Ok, perhaps its my understanding of RMAN that is at fault. From the logs :
    Starting recover at 18-JAN-11
    channel m1: starting incremental datafile backup set restore
    channel m1: specifying datafile copies to recover
    recovering datafile copy file number=00001
    name=/exlibris1/rmanbackup/database_f734055020_s242_s1
    recovering datafile copy file number=00002
    name=/exlibris1/rmanbackup/database_f734055045_s243_s1
    it seems to make backup copies of the datafiles every night, so the creation of these large files is normal ?Above results indicate that you have full (incremental level 0) backup(all datafiles copies ) and there happen update/recover (applying incremental level 1) backup.So there was happen applying */exlibris1/rmanbackup/database_f734055045_s243_s1* inremental backup to full(level 1) backup.And size should be normal
    Why is it making copies of the datafiles even on days of incrementals ?
    Because after getting level 1 backup and need applying and every day will apply one incremental backup.

Maybe you are looking for

  • EJB @OneToOne

    Hi guys, I'm new with EJB 3.0 and probably I'm going to ask a stupid question. I want to map a simple unidirectional one-to-one relationship between two entities: Customer and Address. This is the Customer entity @Entity @Table(name="cliente") public

  • EP- CRM- BAB-BSP-Applications problem

    Hi Guys, When we are trying to copy the existing CRM BSP application to new BSP applicaiton using BAB [Bluprint Application Builder], we are encountering the following error " Object R3tr Tabu CRMC_Blueprntc, referred to by the object key, does not e

  • XFCE4 systemload plugin doesn't show graphs.

    Hello, I just installed the xfce4-systemload-plugin, which is supposed to show me a nice little bar/graph showing me how much RAM / CPU I'm using at the time. The plugin does show, but I don't see anything in the graphs/bar. When I hover my mouse ove

  • Elements 12 - editor won't load

    Brand new, just got Elements 12 for Christmas.  Have uninstalled and restarted.  Organizer opens, but editor will not.  The welcome page says my trial has expired.  I assume that is for the free 30 trial for premiere, which I did not opt to install. 

  • How do you erase an iphone that is dead?

    My iphone died, won't take a charge anymore and only displays the empty battery symbol. I have a new phone on the way and will most likely hand the dead phone over to the tech guys (it's a work phone) but I want to clear the phone first. Who knows wh