Check point query information

Hello Team,
one of my production server i am continually getting this error:
[CHECK POINT FREE SPACE QUERY TIME OUT]
this information we are getting through third party tools and IT team raise ticket to DBA team,
Note:we checked sql server error log  due to this issue but there is no information for
this issue each and every think works fine.
how we can resolve this issue ,this alert getting daily
4 or 5 times.

Hello Team,
one of my production server i am continually getting this error:
[CHECK POINT FREE SPACE QUERY TIME OUT]
this information we are getting through third party tools and IT team raise ticket to DBA team,
Note:we checked sql server error log  due to this issue but there is no information for
this issue each and every think works fine.
how we can resolve this issue ,this alert getting daily
4 or 5 times.
From where did you get this error, as you said you have went through the SQL errorlog & no issues reported, then it is not related to the SQL database server, Also as you said it is 3rd party tool, then suggesting to check what that 3rd party tool it is
for & where it configured & for what(like monitoring,application level query etc) & what inside described it.
Thanks, Rama Udaya.K (http://rama38udaya.wordpress.com) ---------------------------------------- Please remember to mark the replies as answers if they help and UN-mark them if they provide no help,Vote if they gives you information.

Similar Messages

  • MARS and Check Point Firewall Logging

    Hi,
    I have added my Check Point CMA object to MARS, but am not getting seeing any log information.  My CLM is a separate server (child enforcement module), which is discovered OK when the intial CMA discovery takes place in MARS.  I have configured the Log Info settings for the CLM entry in MARS with the SIC details for the Check Point MARS and CLM objects.
    I've created a simple query to gather outbound ftp data (for which there is lots) and I am seeing nothing when running this query in MARS.  The associated CLM log shows plenty of entries.  I am keen to be able to get some historical logging data via MARS, so any help to resolve this issue would be appreciated.
    Many thanks
    Liam

    Liam;
      CS-MARS<>Check Point integration can be very tricky and is very dependent on the versions of software involved.  You may be able to find out some additional insight into the process by raising the CS-MARS logging level for Check Point and monitoring the output.  This is accomplished from the CS-MARS CLI:
    [pnadmin]$ pnlog setlevel cpdebug
    You can then view the messages via the CLI as well:
    [pnadmin]$ pnlog show cpdebug
      If this does not shed any light on the communication between CS-MARS and the Check Point devices, it would be best to open a service request with TAC to further troubleshooting can be performed.
    Scott

  • App server check Points

    Hello Guys
    Can any one pass on a document that will explain what are the check points that to be monitored on a daily basis for a Application Server?
    We have several issues where users are complained about load balancing / Citrix / connection issues ?
    If any one can pass me a document about the health check of any server, area where we need to monitored / mindful on a daily basis.
    Thanks
    BI

    Hi,
    When you are speaking about user logs, what logs are you reffering?
    BPC logs, Citrix logs, Windows logs?
    for BPC logs you can use the follow note:1328702
    User profile path is define into Citrix when the installation was perfomed.
    Person which perform the installation of CITRIX server can provide you this information.
    Kind Regards
    Sorin Radulescu

  • How to check which query is consuming most resources....

    Hi guys how can i check which query is consuming most resources , run by which user and how to kill that....
    Khurana

    1)     At the first stage, login to the server and at the OS level I run commands like TOP and see the overall performance of the Server. From this one can know the Total CPU Memory, CPU Usage, Memory Available, and Swap Memory Available and how busy the CPU is.
    2)     Once you identify the top processes in the CPU, relate them to ORACLE using V$PROCESS.SPID = OS Process ID to get the V$PROCESS.ADDR and join this with the V$SESSION.PADDR to get the V$SESSION.SID and SERIAL# columns.
    3)     Then enable the Session Trace using DBMS_SYSTEM.SET_SQL_TRACE_IN_SESSION.
    4)     look on the Trace File using the TKPROF. During this sort the Trace File using various options like EXEELA/DSK, PRSELA/DSK, FCHELA /DSK.
    5)     Once you identify the expensive SQL statements, inform the user about it and KILL the respective job.

  • NET8의 LOCAL NAMING METHODS 이용관련 CHECK POINTS & TROUBLESHOOTING

    제품 : SQL*NET
    작성날짜 : 2000-12-15
    ===================================================================
    Net8의 Local Naming Methods 이용관련 Check Points & Troubleshooting
    ====================================================================
    Net8 관련 내용중 Local Naming Methods(sqlnet.ora와 tnsnames.ora
    를 이용해서 DB에 Connect하는 방법)에 관한 일반적인 Check Points 및
    Troubleshooing 사항을 정리해보도록 하겠다.
    [Client Connection : Local Naming Methods]
    1. Local Naming method resolves service names by using the local configuration files tnsnames.ora and sqlnet.ora.
    2. One of the benefits of the Local Naming method is that it provides a simple method for resolving service name addresses. It is easy to configure local naming by using a GUI called the Net8 Assistant. When you configure a client machine for local naming by using the Net8 Assistant, a tnsnames.ora file is generated. To configure all the other client machines on the network that need to use the same database services, you can simply copy this file to the client machines.
    3. Another benefit of using the local-naming method is that it resolves service name across networks running different protocols.
    [Local Naming : Configuration]
    1. The local-naming method resolves service names by using the information configured and stored on each individual client. Local naming is most appropriate for simple distributed networks with a small number of services that change infrequently.
    2. Local naming requires the configuration of two local files, tnsnames.ora and sqlnet.ora. The tnsnames.ora file contains the address needed to direct a connection request to the specified listener on the specified node by using the specified database. The sqlnet.ora file stores information about the selected naming method.
    3. Local naming is easy to configure by using the Net8 Assistant. The Net8 Assistant is implemented in Java and is packed with the Java Runtime Environment. You can run the Net8 Assistant on any platform on which Net8 is installed.
    4. When you configure an Oracle8 server to use the local-naming method for service names resolution, the client-side configuration files tnsnames.ora and sqlnet.ora are automatically generated. After the configuration is complete, you copy these two files from the server to the client machines.
    [Client Files : Parameters]
    1. The tnsnames.ora and sqlnet.ora files are generated at the default location ORACLE_HOME\net80\admin.
    2. The first parameter in the tnsnames.ora file specifies the service name and the domain name for the client.
    3. The tnsnames.ora file contains a parameter called DESCRIPTION. This parameter contains information about the connect descriptor for the client.
    4. The ADDRESS parameter in the tnsnames.ora file specifies the network address of the host to which the client is connected. If multiple addresses are specified, use the keyword ADDRESS_LIST before the ADDRESS parameter.
    5. Another parameter in the tnsnames.ora file is the CONNECT_DATA parameter. This parameter specifies the SID of the database to which the client is connected.
    6. The DEFAULT_DOMAIN parameter of the sqlnet.ora file specifies the domain from which the client most often requests names. When this parameter is set, the domain name is automatically appended to the service name in a connect string.
    7. The DIRECTORY_PATH parameter is another key parameter of the sqlnet.ora file. This parameter specifies the names resolution method to be used for the client-server connection.
    8. The sqlnet.ora file also contains a parameter called DEFAULT_ZONE.
    This parameter specifies the region to which the client belongs.
    [Troubleshooting the Client Side]
    These are common client-side problems and the error messages associated with
    each error code.
    ORA-12154: "TNS: could not resolve service name"
    ORA-12198: "TNS: could not find path to destination"
    ORA-12203: "TNS: unable to connect to destination"
    ORA-12533: "TNS: illegal ADDRESS parameters"
    1. The ORA-12154 error occurs when Net8 cannot locate the connect descriptor specified in the tnsnames.ora configuration file. To solve this problem, first verify that a tnsnames.ora file exists and is accessible. Next, verify that multiple copies of the tnsnames.ora file are not present. After checking the existence of the tnsnames.ora file, verify that the service name specified in the connect string is mapped to a connect descriptor in the tnsnames.ora file. Verify also that the file does not contain any syntaz errors. The ORA-12154 error also occur if there is more than one copy of the sqlnet.ora file. Verify that duplicate copies of the sqlnet.ora file are not present.
    2. The ORA-12198 and ORA-12203 errors occur when the client is unable to find the required database. To solve this problem, first verify that you have correctly entered the service name that you want to use. Next verify that the service name parameters in the ADDRESS section of the connect descriptor in the tnsnames.ora file are correctly defined. Then, verify that the tnsnames.ora file is stored in the correct directory. The ORA-12198 and ORA-12203 errors can also occur if the listener on the remote node is not running. Verify that the listener on the remote node has started and is running. If the listener is not running, start the listener by using the Listener Control(LSNRCTL) utility.
    3. The ORA-12533 error occurs if the protocol-specific parameters in the ADDRESS section of the designated connect descriptor in the tnsnames.ora file are incorrect. To solve this problem, use the correct protocol-specific parameters in the ADDRESS section of the connect descriptor.
    4. The ORA-12545 error occurs when the listener on the remote node cannot be contacted. This may happen if the values of the ADDRESS parameter in the tnsnames.ora file and the listener.ora file are incorrect. In this case replace the incorrect values with the correct values in both the files. The ORA-12545 error can also occur if the listener on the remote node is not
    started. To verify whether or not the listener is started, determine its status with the STATUS command of the LSNRCTL utility. If necessary, start the listener on the remote node with the START command.

    제품 : SQL*NET
    작성날짜 : 2000-12-15
    ===================================================================
    Net8의 Local Naming Methods 이용관련 Check Points & Troubleshooting
    ====================================================================
    Net8 관련 내용중 Local Naming Methods(sqlnet.ora와 tnsnames.ora
    를 이용해서 DB에 Connect하는 방법)에 관한 일반적인 Check Points 및
    Troubleshooing 사항을 정리해보도록 하겠다.
    [Client Connection : Local Naming Methods]
    1. Local Naming method resolves service names by using the local configuration files tnsnames.ora and sqlnet.ora.
    2. One of the benefits of the Local Naming method is that it provides a simple method for resolving service name addresses. It is easy to configure local naming by using a GUI called the Net8 Assistant. When you configure a client machine for local naming by using the Net8 Assistant, a tnsnames.ora file is generated. To configure all the other client machines on the network that need to use the same database services, you can simply copy this file to the client machines.
    3. Another benefit of using the local-naming method is that it resolves service name across networks running different protocols.
    [Local Naming : Configuration]
    1. The local-naming method resolves service names by using the information configured and stored on each individual client. Local naming is most appropriate for simple distributed networks with a small number of services that change infrequently.
    2. Local naming requires the configuration of two local files, tnsnames.ora and sqlnet.ora. The tnsnames.ora file contains the address needed to direct a connection request to the specified listener on the specified node by using the specified database. The sqlnet.ora file stores information about the selected naming method.
    3. Local naming is easy to configure by using the Net8 Assistant. The Net8 Assistant is implemented in Java and is packed with the Java Runtime Environment. You can run the Net8 Assistant on any platform on which Net8 is installed.
    4. When you configure an Oracle8 server to use the local-naming method for service names resolution, the client-side configuration files tnsnames.ora and sqlnet.ora are automatically generated. After the configuration is complete, you copy these two files from the server to the client machines.
    [Client Files : Parameters]
    1. The tnsnames.ora and sqlnet.ora files are generated at the default location ORACLE_HOME\net80\admin.
    2. The first parameter in the tnsnames.ora file specifies the service name and the domain name for the client.
    3. The tnsnames.ora file contains a parameter called DESCRIPTION. This parameter contains information about the connect descriptor for the client.
    4. The ADDRESS parameter in the tnsnames.ora file specifies the network address of the host to which the client is connected. If multiple addresses are specified, use the keyword ADDRESS_LIST before the ADDRESS parameter.
    5. Another parameter in the tnsnames.ora file is the CONNECT_DATA parameter. This parameter specifies the SID of the database to which the client is connected.
    6. The DEFAULT_DOMAIN parameter of the sqlnet.ora file specifies the domain from which the client most often requests names. When this parameter is set, the domain name is automatically appended to the service name in a connect string.
    7. The DIRECTORY_PATH parameter is another key parameter of the sqlnet.ora file. This parameter specifies the names resolution method to be used for the client-server connection.
    8. The sqlnet.ora file also contains a parameter called DEFAULT_ZONE.
    This parameter specifies the region to which the client belongs.
    [Troubleshooting the Client Side]
    These are common client-side problems and the error messages associated with
    each error code.
    ORA-12154: "TNS: could not resolve service name"
    ORA-12198: "TNS: could not find path to destination"
    ORA-12203: "TNS: unable to connect to destination"
    ORA-12533: "TNS: illegal ADDRESS parameters"
    1. The ORA-12154 error occurs when Net8 cannot locate the connect descriptor specified in the tnsnames.ora configuration file. To solve this problem, first verify that a tnsnames.ora file exists and is accessible. Next, verify that multiple copies of the tnsnames.ora file are not present. After checking the existence of the tnsnames.ora file, verify that the service name specified in the connect string is mapped to a connect descriptor in the tnsnames.ora file. Verify also that the file does not contain any syntaz errors. The ORA-12154 error also occur if there is more than one copy of the sqlnet.ora file. Verify that duplicate copies of the sqlnet.ora file are not present.
    2. The ORA-12198 and ORA-12203 errors occur when the client is unable to find the required database. To solve this problem, first verify that you have correctly entered the service name that you want to use. Next verify that the service name parameters in the ADDRESS section of the connect descriptor in the tnsnames.ora file are correctly defined. Then, verify that the tnsnames.ora file is stored in the correct directory. The ORA-12198 and ORA-12203 errors can also occur if the listener on the remote node is not running. Verify that the listener on the remote node has started and is running. If the listener is not running, start the listener by using the Listener Control(LSNRCTL) utility.
    3. The ORA-12533 error occurs if the protocol-specific parameters in the ADDRESS section of the designated connect descriptor in the tnsnames.ora file are incorrect. To solve this problem, use the correct protocol-specific parameters in the ADDRESS section of the connect descriptor.
    4. The ORA-12545 error occurs when the listener on the remote node cannot be contacted. This may happen if the values of the ADDRESS parameter in the tnsnames.ora file and the listener.ora file are incorrect. In this case replace the incorrect values with the correct values in both the files. The ORA-12545 error can also occur if the listener on the remote node is not
    started. To verify whether or not the listener is started, determine its status with the STATUS command of the LSNRCTL utility. If necessary, start the listener on the remote node with the START command.

  • Modbus Ethernet Slave seems to fail when the Master query information simultaneously for more than 16 Coils or Discrete Inputs.

    I'm developing an control application by using the Modbus TCP/IP for communicating with the host computer.
    I'm using the Modbus Library for LabVIEW
     (http://sine.ni.com/apps/we/niepd_web_display.display_epd4?p_guid=F1582737BACF5CA8E0340003BA7CCD71&p_node=DZ52363&p_source=External).
    The LabVIEW MB Ethernet Slave seems to fail when the MB Ethernet Master
    query information simultaneously for more than 16 Coils or Discrete
    Inputs.
    The Modbus Application Protocol Specification (http://www.Modbus-IDA.org) state that:
    1- The function code 0x01 'Read Coilsis' is designed to read from 1 to 2000 contiguous status of coils in a remote device.
    2- The function code 0x02 'Read Discrete Inputs' is designed to read
    from 1 to 2000 contiguous status of discrete inputs in a remote device.
    To test the behaviour of the LabVIEW MB Ethernet Slave I have modified
    the 'MB Ethernet Example Slave.vi' by adding more Coils and Discrete
    Inputs. I have added up ONLY to 20 Coils and 20 Discrete Inputs.
    By using the 'MB Ethernet Example Master.vi' the answer to the MB query
    is correct for the first 16 Coils or Discrete Inputs. Starting from the
    17th Coil or Discrete Input the MB Ethernet Slave doesn't give any
    answer.
    By shifting the start point address the answer remain correct for the first 16 Coils or Discrete Inputs.
    I have tested the slave by using also a third part MB Ethernet Master and the behaviour is the same.
    1) There is a reason for this behaviour?
    2) There is a Modbus Library for LabVIEW that satisfy the  'Modbus
    Application Protocol Specification' managing from 1 to 2000 contiguous
    status of Coils or Discrete Inputs?
    Thanks,
    Asper

    I have the same problem with the MB Ethernet.
    I need 40 Outputs.
    The first 16 work perfectly. The next 8 coils do not work, The next 8 work, The next 16 do not work,The next 8 work,The next 24 do not work And work for the last 8. Conclusion, to activate the 40 outputs, I needed 88 coils.
     Because 48 coil, do nothing Can someone help me and i cant implemet the jim solution

  • Error Message: The attempt to connect to the report server failed. Check your connection information and that the report server is a compatible version. The request failed with HTTP status 404: Not Found.

    I have a web page that contains a ReportViewer control.  I am trying to display a report, which is an .rdl file located on the SSRS server, in this ReportViewer control.  I have set the ReportPath and ReportServerUrl correctly.  I am
    getting an error message.
    Am I suppose to use an .rdlc file rather than a .rdl file?  Does the web server configuration need to use a certain account?
     I am getting the following error message:
    The attempt to connect to the report server failed. Check your connection information and that the report server is a compatible version.
    The request failed with HTTP status 404: Not Found.

    Hi bucaroov,
    The error "The request failed with HTTP status 404: Not Found." means the ReportServerURL configured in the ReportViewer control is invalid.
    Please follow these steps to solve the issue:
    Logon the Report Server machine.
    Open the Reporting Services Configuration Manager.
    Copy the Report Server URL from 'Web Services URL'.
    Logon the application server(in this case, it is the server that host the web page), check if we can use the URL we got from step 3 to access the Report Server. If so, please replace the ReportServerURL in the ReportViewer control with this URL. If it is
    not available, could you please post the error message.
    Additionaly, we don't need to provide the extension for a server report. The ReportPath should be like: /<reports folder>/<report name>
    For more information, please see:
    Walkthrough: Using the ReportViewer Control in Remote Mode:
    http://msdn.microsoft.com/en-us/library/ms251669(VS.80).aspx
    If you have any more questions, please feel free to ask.
    Thanks,
    Jin Chen
    Jin Chen - MSFT

  • I am getting the following error when I open my Safari Browser.the proceedure entry point sqlite3_wal_ check point could not be located in the dynamic link library SQLITE3.DLL

    I am getting the following error when I open my Safari Browser.the proceedure entry point sqlite3_wal_ check point could not be located in the dynamic link library SQLITE3.DLL

    Try going back to a restore point >  System Restore - Microsoft Windows

  • Check point in fund management for budget and creation of po

    I have a Budget for a perticular expenses say Rs 100000/-, so the po should not be created for more than Rs 100000/- as per the budget. but the po is been created for more than the budget amount.
    pls. let me know where the check point has to be maintain.
    Thanks
    Santosh Visave
    Moderator: Please, search SDN

    Hi ,
    First of all as a primary step you can do the following :
    1. is the commitment item getting derived in the PO.
    2. Is the commitment item non statistical ( means you have not put the tick on statistical commitment item ) in FMCIA .
    3. check the budget amount via the report FMRP_RW_BUDCON for the budget amount and released amount .
    Hope this would help .
    Cheers ,
    Dewang

  • (ID 30101 Details: Data error (cyclic redundancy check) (0x80070017)) _More information

    Hi,
    Can anyone please help with this?  Our Tape Drive died and we got a new one from our Maintenance Company. Since then large backups to tape are failing.  When the jobs are small they seem to go through OK but when the job is greater than around
    200GB then it fails with the error below.
    I tried updating the tape drive driver but Windows said that the drive wasn't functioning properly any more.
    The only difference between the old faulty drive and the new drive is the firmware.  We used to have A422 in the old drive.  The new drive has  D8D5.  Can someone please help as it's driving me crazy.
    "Affected area:
    IBM TotalStorage 3573 Tape Library
    Occurred since: 03/03/2015 10:19:37
    Description: Library drive IBM ULTRIUM 5 HH 3580 TAPE DRIVE (1068006186) in IBM TotalStorage 3573 Tape Library is not functioning and library jobs may fail until the drive is repaired.
     The drive is not functioning for the following reason:
     (ID 3303)
    DPM encountered a critical error while performing an I/O operation on the tape Physical Servers-LT-MonthlyBackup-00000235 (000086L4) (Barcode - 000086L4) in Drive IBM ULTRIUM 5 HH 3580 TAPE DRIVE (1068006186).
    (ID 30101 Details: Data error (cyclic redundancy check) (0x80070017))
    More information
    Recommended action:
    Retry the operation. If the problem persists, contact your hardware vendor.
    No action required
    Resolution: To dismiss the alert, click below
    Inactivate"
    Regards
    Willie

    Hi,
    It sounds like another faulty drive - You can try to reproduce the problem outside of DPM using some external utilities.   If you get an error before the tape fills you can use net helpmsg errorcode to see what the error was.
    Download the DPMerasetape.zip file from the following link and extract to c:\temp folder.
    https://onedrive.live.com/?cid=b03306b628ab886f&id=B03306B628AB886F%21524&sc=documents
    The utilities are not that user friendly, but here are the basics.
    Always Stop DPMLA Service prior to running MCT.EXE Commands.
      NET STOP DPMLA
    C:\> mct-x64.exe -p
    Opening changer \\.\Changer0
         ********** Changer Parameters **********
             Number of Transport Elements : 1
             Number of Storage Elements : 50
             Number of Cleaner Slots : 0
             Number of of IE Elements : 0
             Number of NumberDataTransferElements : 6
             Number of Doors : 0
             First Slot Number : 0
             First Drive Number : 0
             First Transport Number : 0
             First IEPort number : 0
             First Cleaner Slot Address : 0
             Magazine Size : 0
             Drive Clean Timeout : 600
      Flags set for the changer :
             CHANGER_BAR_CODE_SCANNER_INSTALLED
             CHANGER_POSITION_TO_ELEMENT
             CHANGER_STORAGE_DRIVE
             CHANGER_STORAGE_SLOT
             CHANGER_DRIVE_CLEANING_REQUIRED
             CHANGER_VOLUME_IDENTIFICATION
             CHANGER_VOLUME_SEARCH
             CHANGER_SERIAL_NUMBER_VALID
     Changer can move from Slot to :
                     Slot
                     Drive
     Changer can move from Drive to :
                     Slot
                     Drive
     Changer is Capable of positioning transport to Slot.
     Changer is Capable of positioning transport to Drive.
    C:\> mct-x64.exe -d
    Opening changer \\.\Changer0
    Product Data for Medium Changer device :
      Vendor Id    : STK
      Product Id   : L180
      Revision     : 030
      SerialNumber : 3077520000
    For MCT utility we have the  -m [MOVE] command to move media around inside the library.
    -m [ElemType-T] Transport# [ElemType-Source] S_lot#/D_rive# [ElemType-Destination] S_lot#/D_rive#
    Get / view command syntax for –m (move) command for changer 0
    C:\>mct-x64 0 -m
    Opening changer \\.\Changer0
    MoveMedium : mct -m t N s\d N s\d N   [Where s/d means Slot or Drive and N is ZERO based].
    Some Examples:
    mct-x64 -m t 0 s 0 d 0    (Using transport-0, move media from slot-0  to drive-0)
    mct-x64 -m t 0 d 0 s 0    (Using transport-0, move media from drive-0 to slot-0)
    mct-x64 -m t 0 s 0 s 100  (Using transport-0, move media from slot-0  to slot-100)
    mct-x64 -m t 0 d 0 d 1    (Using transport-0, move media from drive-0 to drive-1)
    mct-x64 -m t 0 s 0 ie 0   (Using transport-0, move media from slot-0  to IEPort 0)
    Once you move a tape into a drive, use mytape commands Loadtape, taperewind, locktape, Disable hardware compression, Set block size to 65536 (64K), writeforspanning.
    You need the symbolic name for the tape drive you loaded media into - look in the DPM console by clicking the tape drive and look at the details for
    \\.\tape########.  use that in the following command.
    Mytape.exe \\.\Tape2147483638
    Status: Getting the handle for \\.\Tape2147483638...Success
    TapeConsole_1.0>taperewind">\\.\Tape2147483638>TapeConsole_1.0>taperewind
    Status: Rewinding Tape ...Success
    TapeConsole_1.0>setdriveinfo">\\.\Tape2147483638>TapeConsole_1.0>setdriveinfo
    Hardware error correction  [y]-Enable / [n] Disable : y
    Hardware data compression  [y]-Enable / [n] Disable : N   (BE SURE TO DISABLE)
    Data padding  [y]-Enable / [n] Disable : n
    Setmark reporting   [y]-Enable / [n] Disable : n
    Number of bytes between the end-of-tape warning and the physical end of the tape: 0
    Status: Setting Drive Information...Success
    TapeConsole_1.0>writeforspanning">\\.\Tape2147483638>TapeConsole_1.0>writeforspanning
    Status: Writing onto tape...Failed !!!
    Error_ID reported: 1100                 (net helpmsg 1100
    = The physical end of the tape has been reached.
    Number of bytes written: 983040     (Ignore bytes written, we'll get physical tape position later)
    Giving up
    Time taken: 15788ms
    TapeConsole_1.0>taperewind">\\.\Tape2147483638>TapeConsole_1.0>taperewind
    Status: Rewinding Tape ...Success
    REPEAT
    TapeConsole_1.0>erasetape">\\.\Tape2147483638>TapeConsole_1.0>erasetape s
    Short erase / Long Erase [s/l]:Status: Erasing the tape...Success
    \\.\Tape2147483646...Success
    c:\>mct-x64.exe -m t 0 d 0 s 0
    Opening changer \\.\Changer0
    Source is a Drive
    Destination is a Slot
    Move : Transport - 0, Src - 0, Dest - 0
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread. Regards, Mike J. [MSFT]
    This posting is provided "AS IS" with no warranties, and confers no rights.

  • The procedure entry point GetLogicalProcessor Information could not be located in the dynamic link library KERNEL32.dll

    When I installed CS6 trial version and started from start menu and clicked Adobe photoshop CS6, I get this message.
    The procedure entry point GetLogicalProcessor Information could not be located in the dynamic link library KERNEL32.dll
    How to resolve it?

    Either your OS is damaged, or doesn't meet the minimum requirements for Photoshop CS6.

  • I can sign on to the iTunes store using my ID but when I want to check my billing information (to change the existing info) it wont accept my password? What do I do?

    I can sign on to the iTunes store using my ID but when I want to check my billing information (to change the existing info) it wont accept my password? What do I do?

    Adding Open DNS codes to your Network Preferences, should give good results in terms of speed-up as well as added security, (including anti-phishing and redirects) (Full information about Open DNS is here: http://www.opendns.com/home/nobloat ) and further independent information can be read here:
    http://reviews.cnet.com/8301-13727_7-57338784-263/free-dnscrypt-tool-enhances-ma c-web-security/?tag=mncol;txt
    and here:
    http://www.macworld.com/article/1146064/troubleshootdns.html?t=234
    Open System Preferences/Network. Double click on your connection type, or select it in the drop-down menu, and in the box marked 'DNS Servers' add the following two numbers:
    208.67.222.222
    208.67.220.220
    (You can also enter them if you click on Advanced and then DNS)
    Sometimes reversing the order of the DNS numbers can be beneficial in cases where there is a long delay before web pages start to load, and then suddenly load at normal speed:
    http://support.apple.com/kb/TS2296

  • Performance tuneup for a special DB (disable locking, check-pointing,...)

    Hi,
    I have simple database contains key/value records. The program is a multi-thread application that iterate over records. Each worker thread read a record and after some calculations, replace it. The records are completely independent from each other. The following is my DBController class that share between all threads. Is there any considerations to achieve best performance?, For example I don't want any locking, check-pointing, caching,...overheads. I just want to achieve THE BEST PERFORMANCE to store and retrieve each record independently.
    import gnu.trove.*;
    import java.io.*;
    import com.sleepycat.bind.tuple.*;
    import com.sleepycat.je.*;
    public class DBController {
         private class WikiSimTupleBinding extends TupleBinding<TIntObjectHashMap<TIntDoubleHashMap>> {
              // Write an appropriate object to a TupleOutput (a DatabaseEntry)
              public void objectToEntry(TIntObjectHashMap<TIntDoubleHashMap> object, TupleOutput to) {
                   try {
                        ByteArrayOutputStream bout = new ByteArrayOutputStream();
                        ObjectOutputStream oout = new ObjectOutputStream(bout);
                        oout.writeObject(object);
                        oout.flush();
                        oout.close();
                        bout.close();
                        byte[] data = bout.toByteArray();
                        to.write(data);
                   } catch (IOException e) {
                        e.printStackTrace();
              // Convert a TupleInput(a DatabaseEntry) to an appropriate object
              public TIntObjectHashMap<TIntDoubleHashMap> entryToObject(TupleInput ti) {
                   TIntObjectHashMap<TIntDoubleHashMap> object = null;
                   try {
                        byte[] data = ti.getBufferBytes();
                        object = (TIntObjectHashMap<TIntDoubleHashMap>) new java.io.ObjectInputStream(
                                  new java.io.ByteArrayInputStream(data)).readObject();
                   } catch (Exception e) {
                        e.printStackTrace();
                   return object;
         private Environment myDbEnvironment = null;
         private Database db_R = null;
         private WikiSimTupleBinding myBinding;
         public DBController(File dbEnv) {
              try {
                   // Open the environment. Create it if it does not already exist.
                   EnvironmentConfig envConfig = new EnvironmentConfig();
                   envConfig.setAllowCreate(true);
                   myDbEnvironment = new Environment(dbEnv, envConfig);
                   // Open the databases. Create them if they don't already exist.
                   DatabaseConfig dbConfig = new DatabaseConfig();
                   dbConfig.setAllowCreate(true);
                   db_R = myDbEnvironment.openDatabase(null, "R", dbConfig);
                   // initialize Binding API
                   myBinding = new WikiSimTupleBinding();
              } catch (DatabaseException dbe) {
                   // Exception handling goes here
                   dbe.printStackTrace();
         private final byte[] intToByteArray(int value) {
              return new byte[] { (byte) (value >>> 24), (byte) (value >>> 16), (byte) (value >>> 8), (byte) value };
         public void put(int id, TIntObjectHashMap<TIntDoubleHashMap> repository) {
              try {
                   DatabaseEntry theKey = new DatabaseEntry(intToByteArray(id));
                   DatabaseEntry theData = new DatabaseEntry();
                   myBinding.objectToEntry(repository, theData);
                   db_R.put(null, theKey, theData);
              } catch (Exception dbe) {
                   // Exception handling goes here
                   dbe.printStackTrace();
         public TIntObjectHashMap<TIntDoubleHashMap> get(int id) {
              TIntObjectHashMap<TIntDoubleHashMap> repository = null;
              try {
                   // Create a pair of DatabaseEntry objects. theKey is used to perform the search. theData is used to store the
                   // data returned by the get() operation.
                   DatabaseEntry theKey = new DatabaseEntry(intToByteArray(id));
                   DatabaseEntry theData = new DatabaseEntry();
                   // Perform the get.
                   if (db_R.get(null, theKey, theData, LockMode.DEFAULT) == OperationStatus.SUCCESS) {
                        // Recreate the data repository
                        repository = myBinding.entryToObject(theData);
                   } else {
                        System.out.println("No record found for key '" + id + "'.");
              } catch (Exception e) {
                   // Exception handling goes here
                   e.printStackTrace();
              return repository;
         public void close() {
              // closing the DB
              try {
                   if (db_R != null)
                        db_R.close();
                   if (myDbEnvironment != null)
                        myDbEnvironment.close();
              } catch (DatabaseException dbe) {
                   // Exception handling goes here
                   dbe.printStackTrace();
    }

    If you are writing and you need to recover in a reasonable amount of time after a crash, you need checkpointing.
    If multiple threads may access a record concurrently, you need locking.
    If you need good read performance, you need as large a JE cache as possible.
    If you want to tune performance, the first step is to print the EnvironmentStats (Environment.getStats) periodically, and read the FAQ performance section. Try to find out if your app's performance is limited by CPU or I/O.
    If you are reading records in key order, then you'll get better performance if you also write them in key order.
    I'm not sure why you're using a TupleBinding to do Java object serialization If you want Java serialization, try using a SerialBinding.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Server does not return metadata, check the query.

    Hi,
    I need your help in using report designer of Bex, The issue is when we try to open a query in the report designer we get a Error Server does not return metadata, check the query.
    Thanks in Advance
    Kind Regards
    Phutane

    Hi,
    All the yellow and red lights will have an effect on query performance or execution.  Read up on them as there are too many to explain via this forum.
    There is a document on SDN on query performance.  Some useful links:
    [https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/9f4a452b-0301-0010-8ca6-ef25a095834a]
    [http://help.sap.com/saphelp_nw70/helpdata/en/41/b987eb1443534ba78a793f4beed9d5/frameset.htm]
    [http://help.sap.com/saphelp_nw70/helpdata/en/d9/31363dc992752de10000000a114084/frameset.htm]
    [http://help.sap.com/saphelp_nw04/helpdata/en/2e/caceae8dd08e48a63c2da60c8ffa5e/frameset.htm]
    [https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/e0501cb9-d51c-2a10-239c-c141a22c45a6]
    Cheers...

  • Server does not return query metadata check the query

    hi,
    i want know how to  use bex query designer,
    when i  insert a data provider,the message is appear "server does not return query metadata check the query",
    who can help me ?
    thanks a lot !
    addition: my current entironment is: gui710 with sp4,bi710 with sp2.

    Hi,
    All the yellow and red lights will have an effect on query performance or execution.  Read up on them as there are too many to explain via this forum.
    There is a document on SDN on query performance.  Some useful links:
    [https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/9f4a452b-0301-0010-8ca6-ef25a095834a]
    [http://help.sap.com/saphelp_nw70/helpdata/en/41/b987eb1443534ba78a793f4beed9d5/frameset.htm]
    [http://help.sap.com/saphelp_nw70/helpdata/en/d9/31363dc992752de10000000a114084/frameset.htm]
    [http://help.sap.com/saphelp_nw04/helpdata/en/2e/caceae8dd08e48a63c2da60c8ffa5e/frameset.htm]
    [https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/e0501cb9-d51c-2a10-239c-c141a22c45a6]
    Cheers...

Maybe you are looking for

  • Trouble importing photos from Panasonic camera

    I'm getting error messages when attempting to transfer photos from my Panasonic camera to my Mac. Details: * Running Lion 10.7.4 on MacBook Pro * iPhoto '09 version 8.1.2 * Brand new Panasonic DMC-ZS20 * SanDisk 4 GB SDHC card, which I formatted usin

  • How do I find out or know if everything is backed up before I upgrade to ios 6?

    before I update to ios6 how can I make triple sure that all my music, photos, contacts, notes, and messages are all backed up? in settings I have my iPhone 4 programmed to back up to the iCloud - but how do I know for sure that everything is there? I

  • Error in A/P invoice -non deduct tax a/c missing

    Dear All While adding A/P invoice from base GRPO where it contains frieght an error is displaying:non deduct tax account is missing for stock item. In frieght the stock option and distribution method is none only,but A/P invoice getting added with ou

  • Why can't I install Creative Cloud on my second computer?

    I'm having trouble install Creative Cloud on my 2nd computer -- MacBook Pro. I download the app and clicked on the Installer, and when I initializing the installer, I get this message: Adobe Creative Cloud is needed to resolve this problem. However,

  • What is a PRAM reset and why did I have to do it ?

    My very new (4 days) 20" Aluminium iMac was getting quite noisy, so much so that I called for support today. The Apple tech explained how to do a 'P.RAM' reset, which in fairness does seem to have resolved the issue. (something to do with a FAN, I th