Read the c2 log file of the sql server using java

Hi All,
i want to read the c2 log file using the core java. how is it possible ? if anybody knows about this please give me the sample code to help me.
i am also searching on net but i am not getting any result about this? so please help me to doing this task?
awaited person

Hi All,
i want to read the c2 log file using the core java. how is it possible ? if anybody knows about this please give me the sample code to help me.
i am also searching on net but i am not getting any result about this? so please help me to doing this task?
awaited person

Similar Messages

  • How do i connect to sql server using java?

    hi ever body
    i need to connect to sql server using java
    what is the driver i will use?
    thanks

    Microsoft make a jdbc driver and it is readily available through MSDN. Here is a link http://msdn2.microsoft.com/en-us/data/aa937724.aspx. Whatever jdbc driver you are looking for, Google it, and if it exists you'll find it in 10 seconds.

  • Is it possible to lock the keyboard of a client from a server using java

    please explain wheterit is possible to lock the keyboard of a client from a server using java

    You want to process code on one machine, and thereby lock the keyboard on another machine? No, that's not possible. It is extremely far from possible.
    Of course, if the client is running software with security holes in it you might hack into it and crash the thing. This will lock up the keyboard pretty good. I hope that's not what you want ...
    Or are you talking about a setup where you already have code running on the client, and some sort of communication between client and server? In that case what you need to know is whether it is possible to lock the keyboard at all. Once you have figured that out, it is trivial to add the communication code to have the server software tell the client software to lock the keyboard.
    So what do you mean with "lock the keyboard"? It's pretty easy to remove/disable all keyboard related listeners in your own application. It's a lot harder (and AFAIK impossible with pure java) to disable alt-tabbing out of the application. And impossible, except from exploiting security holes, to lock the ctrl-alt-delete-combination on windows machine.

  • How to access sql server using java script

    Hi, Is it possible to connect to sql server in java script. I'm using microsoft jdbc driver for connectivity to sql server.
    Thanks

    Hi, Is it possible to connect to sql server in java
    script. I'm using microsoft jdbc driver for
    connectivity to sql server.If you want to ask about Javascript, you are in the wrong place. Go and find some JavaScript forum. Actually, you could connect to database from javascript using Ajax or the like mainly because you could have Java code behind the scenes, but then that too is not something I would want to discuss here. So, go find the correct place for the topic.

  • Steps to move Data and Log file for clustered SQL Server

    Hi guys 
    we have Active'passive SQL 2008R2 cluster environment.
    looking for steps to move Data and log files from user Database  and System Database for  SQL Server Clustered Instance. 
    Currently Data and log  files resides on same drive for user and system Databases..
    Thanks
    Please Mark As Answer if it is helpful. \\Aim To Inspire Rather to Teach A.Shah

    Try the below link
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/468de435-3432-45c2-a50b-23519cd2686e/moving-the-system-databases-in-a-sql-cluster?forum=sqldisasterrecovery
    -Prashanth

  • Log File Issue In SQL server 2005 standard Edition

    We have database of size 375GB .The data file has 80 GB free space within .When trying to rebuild the index we had 450 GB free space on the disk where Log file is residing.The rebuild index activity failed due to space issue.added more space and got the
    job done successfully
    The Log file has grow up to 611GB to complete the rebuild index.
    version :SQL server 2005 Standard Edition .Is ther a way to estimate the space required for rebuild index in this version.
    I am aware we notmaly allocate 1.5 times of data file.But in this case It was totaly wrong.
    Any suggestion with examples would be appreciated.
    Raghu

    OK, there's a few things here.
    Can you outline for everybody the recovery model you are using, the frequency with which you take full, differential and transaction log backups.
    Are you selectively rebuilding your indexes or are you rebuilding everything?
    How often are you doing this? Do you need to?
    There are some great resources on automated index maintenance, check out
    this post by Kendra Little.
    Depending on your recovery point objectives I would expect a production database to be in the full recovery mode and as part of this you need to be taking regular log backups otherwise your log file will just continue to grow. By taking a log backup it will
    clear out information from inactive VLF's and therefore allow SQL Server to write back to those VLF's rather than having to grow the log file. This is a simplified version of events, there are caveats.
    A VLF will be marked as active if it still has an open transaction in it or there is a HA option that still requires that data to be available as that data has not been copied to another node yet.
    Most customers that I see take transaction log backups every 15 - 30 minutes, but this really does depend upon how much data your company can afford to lose. That's another discussion for another day.
    Make sure that you take a transaction log backup prior to your job that does your index rebuilds (hopefully a smart job not a sledge hammer job).
    As mentioned previously swapping to bulk logged can help to reduce the size of the amount of information logged during index rebuilds. If you do this make sure to swap back into the full recovery model straight after and perform a full backup. There are
    problems with the ability to do point in time restores whilst in the bulk logged recovery model, so you need to reduce the amount of time you use it.
    Really you also need to look at how your indexes are created does the design of them lead to them being fragmented on a regular basis? Are they being used? Are there better indexes out there that can help performance?
    Hopefully that should put you on the right track.
    If you find this helpful, please mark the post as helpful,
    If you think this solves the problem, please propose or mark it an an answer.
    Please provide details on your SQL Server environment such as version and edition, also DDL statements for tables when posting T-SQL issues
    Richard Douglas
    My Blog: Http://SQL.RichardDouglas.co.uk
    Twitter: @SQLRich

  • Log file shrinking in SQL server

    Hi,
    I have a log file with initial size of 80 GB in C drive.
    Now we are having space issue in C drive, so i tried to shrink the log file, but its not reducing below initial size.
    Is it the default behavior or i missed anything while shrinking ?
    I used DBCC SHRINKFILE option to shrink it.
    How can i change the initial size of the log file ?
    If it has been set as 80 GB , because of that am i not able to free space in C drive ?
    Thanks,
    Vinodh Selvaraj

    Hello,
    Please check first the log reuse wait state of the databases, may you have to run an additional log backup before you can shrink the log file
    select name, log_reuse_wait_desc
    from sys.databases
    order by name
    Olaf Helper
    [ Blog] [ Xing] [ MVP]

  • How i connect to SQL server using java?

    hi every body
    i need to connect to a sql server database using java
    i do not know sql driver in java
    plz help me
    thanks

    This is certainly the wrong forum for that question. Search in the JDBC forum. This has been asked and answered many times.

  • How to get pdf file from sap presentation server using java connector

    Hi Friends,
    with the below code i am able to get po details in pdf in presentation server.
    DATA : w_url TYPE string
           VALUE 'C:\Documents and Settings\1011\Solutions\web\files\podet.pdf'.
    CALL FUNCTION 'ECP_PDF_DISPLAY'
            EXPORTING
              purchase_order       = i_ponum
           IMPORTING
      PDF_BYTECOUNT        =
             pdf                  = file  " data in Xsting format
    *Converting Xstring to binary_tab
          CALL FUNCTION 'SCMS_XSTRING_TO_BINARY'
            EXPORTING
              buffer                = file
      APPEND_TO_TABLE       = ' '
    IMPORTING
      OUTPUT_LENGTH         =
            TABLES
              binary_tab            = it_bin " data in binary format
    **Downloading into PDF file
          CALL FUNCTION 'GUI_DOWNLOAD'
            EXPORTING
      BIN_FILESIZE                    =
              filename                        = w_url
              filetype                        = 'BIN'
             TABLES
              data_tab                        = it_bin
    when i am using java connector , to retirve the file from presentation server , the follwoing error i am getting...
    init:
    deps-jar:
    compile-single:
    run-single:
    com.sap.mw.jco.JCO$Exception: (104) RFC_ERROR_SYSTEM_FAILURE: Error in Control Framework
            at com.sap.mw.jco.rfc.MiddlewareRFC$Client.nativeExecute(Native Method)
            at com.sap.mw.jco.rfc.MiddlewareRFC$Client.execute(MiddlewareRFC.java:1244)
            at com.sap.mw.jco.JCO$Client.execute(JCO.java:3842)
            at com.sap.mw.jco.JCO$Client.execute(JCO.java:3287)
            at PdfGen.<init>(PdfGen.java:35)
            at PdfGen.main(PdfGen.java:78)
    Java Result: 1
    BUILD SUCCESSFUL (total time: 1 second)
    i debugged too, problem with <b>gui_download......</b>
    I am very glad to all with your suggestions!!
    Regards,
    Madhu..!!

    Hi
    You can try to create an external command (transaction SM69).......sorry I've forgotten,,,,they works on application
    How do you call CL_GUI_FRONTEND_SERVICES=>EXECUTE?
    Max
    Edited by: max bianchi on Oct 13, 2011 10:27 AM

  • Warnings in the Upgrade Log file for OBIEE 11G from 10G

    Hi,
    I am new to OBI 11g.
    I have installed simple type OBI 11g (version - 11.1.1.5.0) on a single server(Windows 2003, RAM-4GB,CPU-2.8GHz Dual Core ). I used upgrade assistant to upgrade the RPD & Webcat from 10.1.3.4.1 version to 11g, but I got about 50K warnings in the upgarde log file for the web catalog. They were basically 2 types as below -
    1. [2012-02-24T13:17:48.255-05:00] [BIEE] [WARNING] [] [upgrade.BIEE] [tid: 14] [ecid: 0000JMk1k3lDkZKcyToYpk1FHuer000005,0] Invalid columnID reference 'c10' in views!
    Does someone know this warning and if this requires a fix or it would not impact any reports or dashboards? This error is about 25K times in the upgrade log file.
    2. [2012-02-24T12:20:04.162-05:00] [BIEE] [WARNING] [] [upgrade.BIEE] [tid: 14] [ecid: 0000JMk1k3lDkZKcyToYpk1FHuer000005,0] Removed the following node from the element 'view': [[
    <saw:selector xmlns:saw="com.siebel.analytics.web/report/v1.1" columnID="c20" prompt="true"><saw:choice type="column" formula="&quot;- Year&quot;.&quot;MVSR Doses R12M&quot;"><saw:tableHeading><saw:caption><saw:text>Month</saw:text></saw:caption></saw:tableHeading><saw:columnHeading><saw:caption><saw:text>Doses</saw:text></saw:caption></saw:columnHeading></saw:choice><saw:choice type="column" formula="&quot;- Year&quot;.&quot;MVSR Doses YTD&quot;-&quot;- Year&quot;.&quot;MVSR Doses Prior YTD&quot;"><saw:tableHeading><saw:caption><saw:text>Year</saw:text></saw:caption></saw:tableHeading><saw:columnHeading><saw:caption><saw:text>YTD Dose Volume Change</saw:text></saw:caption></saw:columnHeading></saw:choice></saw:selector>
    I have checked this error on Oracle support web and it says that this warning is fine as there have been some features in 10G that would not be in the OBI 11G and are not required. I am still not convinced why there should be any warning for this type of issue.
    Please let me know if someone has faced these warnings in when doing the upgrade of the 10g RPD and Webcat to 11g together.
    Just as note, the services start fine and I could login to the application with users and see the dashboards and reports although I did not check all the dashboards and reports as this is a POC before doing the actual upgrade.
    Edited by: user11255710 on Feb 24, 2012 1:44 PM

    Hi
    Can you please refer me to the links on oracle web support which explained that these warnings are ok and can be ignored. We are also facing the same issue.
    Thanks in advance.

  • Getting IO exception when trying to view the joblog.log file

    Hi,
    We are unable to view the joblog.log file. The joblog file is created in server path, but when we trying to view/read/download the joblog.log file we are getting JCS-108004: An IO exception occurred accessing a local file error.
    server path: xxx/log/processerver_.log.
    But we are able to view the reports through redwood for these extensions (*. html, *.csv. *. xml.).
    server path: xxx/lis/sys_.html
    Can anybody help on this issue.
    Thanks,
    Suresh.

    Dear Suresh,
    Your Question is not clear to me. what i understood is you can abel to check log in redwood , and unable to open in server at os level? if it is correct, login to server with cps installation user, ten i hope you can able to open files, check the file permissions in init file in $JCS_HOME/admin directory, check the permissions given, and also check listener port value, is it correct?
    thanks and regards
    Muhammad Asif

  • Message in alert log file during the startup of the DB

    Hi,
    i have two instances, i have the same warning in the alert log file during the startup of the DB:
    Oracle instance running on a system with low open file descriptor limit.
    Tune your system to increase this limit to avoid severe performance degradation.
    The db_files parameter was at 1024 for the two instances, i have set it to 300 for the first instance and 550 for second, the message for the first instance was droped(it's ok), but the message was there in the alert log for the second instance.
    The two instance are on the same machine.
    The number of the open file limit on the linux machine is equal to 1024.
    I have shutdown the first instance (with db_files=300) the problem persits for the second instance (with db_files=550).
    My question is how to determinate the right value of db_files parameter to avoid this message.
    Regards.
    Message was edited by:
    learn

    Increase the max number of open file on Linux or decrease db_files.
    f the OS limit is 1024 db_files would have to be >= 472 for the message to be printed. t is printed if an instance is configured to access more datafiles than the OS-specific limit on the number of files that can be open in a single process. In this case the server will re-cycle the file descriptors.
    Vadim Bobrov
    Oracle Database Tools
    http://www.fourthelephant.com

  • Can't read NI-SPY log file

    Hi there,
    Our customer sent us an NI-SPY log file that captured GPIB operation with his PC and our instrument for trouble shooting. I tried to open the .SPY file but could not. NI-SPY apparently opens the .SPY file but actually does not show anything. (See attached .SPY file that could not be read on my PC, and a screen shot of NI-SPY.)
    The SPY log file is the one captured on the customer PC that installs PCI-GPIB with NI-488.2M v1.6 running Win98(probably).
    I tried to open it with my two PCs but both were not possible. Both PC do not have NI-GPIB card but installs NI-VISA 2.6.
    PC1:
    Win2k(admin log-on), NI-VISA 2.6 with Tulip passport enabled, Agilent PCI-GPIB.
    NI-SPY version is 1.5.0.
    PC2:
    WinMe, NI-VISA 2.6, no G
    PIB card.
    NI-SPY version is 1.5.0.
    I found there was a similar Q&A with the URL below but it could not help.
    http://exchange.ni.com/servlet/ProcessRequest?RHIV​EID=101&RNAME=ViewQuestion&HOID=506500000008000000​34230000&ECategory=Instrument+Connectivity.GPIB
    What is causing the problem?
    Any help will be appreciated.
    Thanks
    Makoto Kondo, Kikusui Electronics Corp.
    Attachments:
    ibic.spy ‏5 KB
    spy_screen.jpg ‏25 KB

    Hi,
    I was able to open the spy file. I don't know why it's not showing in yours. Maybe it has something to do with the localized version of the driver and OS.
    What's the language of the OS and Driver installed in your computer?. And on the computer that made the capture?
    In any case, I'm sending you a screenshot and an exported tes file of the capture.
    DiegoF.
    National Instruments.
    Attachments:
    ibic.bmp ‏210 KB
    ibic.txt ‏2 KB

  • Will RMAN delete archive log files on a Standby server?

    Environment:
    Oracle 11.2.0.3 EE on Solaris 10.5
    I am currently NOT using an RMAN repository (coming soon).
    I have a Primary database sending log files to a Standby.
    My Retention Policy is set to 'RECOVERY WINDOW OF 8 DAYS'.
    Question: Will RMAN delete the archive log files on the Standby server after they become obsolete based on the Retention Policy or do I need to remove them manually via O/S command?
    Does the fact that I'm NOT using an RMAN Repository at the moment make a difference?
    Couldn't find the answer in the docs.
    Thanks very much!!
    -gary

    Hello again Gary;
    Sorry for the delay.
    Why is what you suggested better?
    No, its not better, but I prefer to manage the archive. This method works, period.
    Does that fact (running a backup every 4 hours) make my archivelog deletion policy irrelevant?
    No. The policy is important.
    Having the Primary set to :
    CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON STANDBYBut set to "NONE" on the Standby.
    Means the worst thing that can happen is RMAN will bark when you try to delete something. ( this is a good thing )
    How do I prevent the archive backup process from backing up an archive log file before it gets shipped to the standby?
    Should be a non-issue, the archive does not move, the REDO is transported and applied. There's SQL to monitor both ( Transport and Apply )
    For Data Guard I would consider getting a copy of
    "Oracle Data Guard 11g Handbook" - Larry Carpenter (AKA Dr. Paranoid ) ISBN 978-0-07-162111-2
    Best Oracle book I've read in 10 years. Covers a ton of ground clearly.
    Also Data Guard forum here :
    Data Guard
    Best Regards
    mseberg
    Edited by: mseberg on Apr 10, 2012 4:39 PM

  • How can I get an accurate count of computers (workstations) logged in to my SQL Server database?

    We license our software by number of workstations.
    I have a query that I have used for years to get an accurate count of the workstations logged in to my SQL Server database. For simplicity, all users use the same login name/password. This is built in to the script that attaches to the DB. They
    have access only to that DB with the exception of
    USE [Master] GRANT VIEW SERVER STATE to MyUser
    The query that has been working is below:
    Basically this relies on the mac address (Master.dbo.sysprocesses.net_addre
    SELECT COUNT(Users) AS UserCount FROM (SELECT COUNT(Master.dbo.sysprocesses.hostname) AS Users FROM Master.dbo.sysprocesses LEFT OUTER JOIN Master.dbo.sysdatabases ON Master.dbo.sysdatabases.dbid = Master.dbo.sysprocesses.dbid WHERE (Master.dbo.sysdatabases.name = 'MyDatabase') GROUP BY Master.dbo.sysprocesses.net_address) AS UserCount_1
    ss), since both Workstation names and ip addresses can be duplicated.
    Recently this has stopped working at a number of customers. Suddenly individual workstations are showing multiple net addresses for the same workstation causing a substantial overcount of users. This may be related to SQL Server 2012 - not sure.
    What I need is a very reliable way to get a count of workstations logged in to my database.
    If anyone can tell me why I am suddenly getting multiple net_addresses for each workstation and how to prevent that that would be one possible solution.
    Otherwise if anyone can give me a rock solid way to get a workstation count other than the above that would be great. Our largest customer is 50 users by the way.
    Thanks in advance for any help.
    Kim

    I think your count is getting skewed because of duplicate rows in sysprocesses because of parallelism. Add DISTINCT to your query:
    SELECT COUNT(Users) AS UserCount
    FROM
    ( SELECT COUNT(DISTINCT Master.dbo.sysprocesses.hostname) AS Users
    FROM Master.dbo.sysprocesses LEFT OUTER JOIN Master.dbo.sysdatabases
    ON Master.dbo.sysdatabases.dbid = Master.dbo.sysprocesses.dbid
    WHERE (Master.dbo.sysdatabases.name = 'MyDatabase'
    GROUP BY Master.dbo.sysprocesses.net_address) AS UserCount_1
    Satish Kartan www.sqlfood.com

Maybe you are looking for