GRC FF 5.1 BACKGROUND JOB WAS NOT SCHEDULED/LOG FILE NOT YET GENERATED

I've had FF installed in prod for 4 days. I just now got around to running the /VIRSA/ZVFATBAK job. However it only generates the logs for the last 24 hours. I know when you schedule the batch run to run hourly this will be resolved but I wanted to run the /VIRSA/ZVFATBAK job manually the first time because I knew it would take longer than an hour. Is there a variant I can use in SE38 to get these earlier logs generated?
thanks for your help
Dave wood

You can create your variants in SE38. It does not come up with the programs. For Spools, you can get upto 3 days of logs while for the work logs, it will be a week max. But, i think you can change the system parameters in such a way you can retrieve the logs of the jobs for more than a week.
Thanks
Sudhan Shan

Similar Messages

  • BACKGROUND JOB WAS NOT SCHEDULED/LOG FILE NOT YET GENERATED

    Hello,
    To generate the log report, /VIRSA/ZVFATBAK program is scheduled on hourly basis but some time report doesn't get generated and if we see the background job then it shows sucessfully finished.
    If we see the maually the log report for FFID then below error message is displayed.
    " BACKGROUND JOB WAS NOT SCHEDULED/LOG FILE NOT YET GENERATED"
    Can anyone guide me to solve the issue.
    Thanks in advance.
    Best Regards,
    Prashant Dubey

    Hi,
    once chk the status of the job by selecting that and check job status(cltr+shift_f12)
    since it was periodically scheduled job there will be a RELEASED job after every active job..
    so try to copy that into another job using copy option and give some new name which u have to remember...
    the moment u copy u can find the same copied job in SCHEDULED status...
    from here, try to run it again on hourly basis....
    After copying the job u can unschedule the old released job to scheduled otherwise 2 will run at a time...
    rgds,

  • Error in batch processing: background job cannot be scheduled

    Error in batch processing: background job cannot be scheduled in transaction code jpmg0. How to resolve the same?

    Hello,
    Check system log and also check work process trace, to know the error.
    I feel that it could be Authorization issue, you don't have sufficient authorization to do.
    Hope this helps.
    Regards
    venkata
    Edited by: venkata emandi on Sep 12, 2011 8:13 AM

  • SQL Server 2012 Reorg Index Job Blew up the Log File

    We have a maintenance plan that nightly (1) runs dbcc checkdb on all databases, (2) reorgs indexes on all databases, compacting large objects, (3) updates statistics, etc. There are three user databases, one large, one medium, one small. Usually it uses
    a little more than 80% of the medium database's log, set to 6,700 MB. Last night the reorg index step caused the log to increase to almost 14,000 MB and then blew up because the maximum file size was set to 14,000 MB, one of the alter index commands failed
    because it ran out of log space. (Dbcc checkdb step ran successfully.) Anyone have any idea what might cause this? There is one update process on this database, it runs at 3 AM. The maintenance plan runs at 9 PM and completes by 1 AM. The medium database has
    a 21,000 MB data file, reserved space is at about 10 GB. This is a SQL 2012 Standard SP 2 running on Windows 2012 Server Standard.

    I personally like to shrink the log files once the indexes have been rebuilt and before switching back to full recovery, because as I'm going to take a full backup afterwards, having a small log file reduces the size of the backup.
    Do you grow them afterwards, or do you let the application waste time on that during peak hours?
    I have not checked, but I see no reason why the backup size would depend on the size of the log file - it's the data in the data file you back up, not the log file.
    I would say this is highly dubious.
    Erland Sommarskog, SQL Server MVP, [email protected]
    Yeah I let the application allegedly "waste" a few milisseconds a day autogrowing the log file. Common, how long do you think it takes for a log file to grow a few GB on most storage systems nowadays? As long as you set an appropriate autogrow
    interval so your log file doesn't get too fragmented (full of VLFs), you'll be perfectly fine in most situations.
    Lets say you have a logical disk dedicated to log file storage, but it is shared across multiple databases within the instance. Having allocated space for the log files means there will be not much free space left in the disk in case ANY database needs more
    space than the others due to a peak in transactional workload, even though other databases have unused space that could have been used.
    What if this same disk, for some reason, is also used to store the tempdb log file? Then all applications will become unstable.
    These are the main reasons I don't recommend people blindly crucify keeping log files small when possible. I know there are many people who disagree and I'm aware of their reasons. Maybe we just had different experiences about this subject. Maybe people
    just haven't been through the nightmare of having a corrupted system database or a crashed instance because of insuficient log space in the middle of the day.
    And you are right about the size of the backup, I didn't put it correctly. It isn't the size of the backup that gets smaller (although the backup operation will run faster, having tested this myself), but the benefit from backing up a database with a small
    log file is that you won't need the extra space to restore it in a different environment such as a BI or DEV server, where recuperability doesn't matter and the database will be on simple recovery mode.
    Restoring the database will also be faster.
    Just because there are clouds in the sky it doesn't mean it isn't blue. But someone will come and argue that in addition to clouds, birds, airplanes, pollution, sunsets, daltonism and nuclear bombs, all adding different colours to the sky, this
    is an undocumented behavior and should not be relied upon.

  • Cannot publish get error message - log file not being created

    When trying to publish a FlashHelp project, I get an error
    message window that says "Publishing has been cancelled. Failed to
    create file: (project name).log "
    When I click okay in the message window, the publishing
    process stops. However, if I look in the SSL folder, I see the log
    file. It is a text file.
    I had this problem in January 09 but it seemed to be an issue
    with the password and path in the FTP command window. I fixed it
    and it worked fine. However, I haven't published since the end of
    January. Now, when I try to publish, it is giving me the same error
    message. I checked and reviewed the FTP window fields and they are
    fine. But I'm still getting the error message and can't publish.
    Why?
    I need to get this problem fixed ASAP and ensure that it
    doesn't occur again. What's strange is that I've got 3 other
    projects and this is the only 1 that gets this error message.

    Yes, the generation worked. I checked the log file that
    worked from the time it worked before and it seems to be the same
    as the log file that is generated when I get the error message.
    I created a new FlashHelp layout and got the same error
    message. What's really weird is there is a log file in the SSL
    folder but when you click OK in the error message, it stops the
    publish function.
    Last time I had to blow away the cpd file as if this was a
    corrupt project. But that gets to be painful. As I use templates to
    put change dates in the footers of topics and templates get lost
    when you blow away the cpd.
    Any other thoughts?

  • Could not open log file in Win2000 sp3 Message in console

    I have configured log file for ias. Then restarted my Application server. It is giving Error message "Could not open log file ''logs/logVal.10817" in kas window. But logs files are there (created by the server itself) in the logs floder.
    I have configured logs for two systems.
    One system is noted all messages in the file
    Second system did not not any messages.
    But the files are existed in the logs folder in both the systems.
    I need to configure logs for ias,Kjs and KXS also.
    Please suggest me regarding this.
    thanks
    sudheer

    Hi,
    I'm not sure what operation are you trying to, can you please confirm that ? Please check what kind of messages did you try to log, is it only errors ? or all errors & warning or all messages ?. If it were only errors and warnings, then there is a possibility that the server did not encounter any of this, due to which the log file can be empty.
    Regards
    Raj

  • Job number from alert log file to information

    Hello!
    I have a question about job numbers in alert log file. Today one of our Oracle 10g R2 [10.2.0.4] RAC nodes crashed. After examining alert log file for one of the nodes I saw a lot of messages like:
    Tue Jul 26 11:52:43 2011
    Errors in file /u01/app/oracle/admin/zeme/bdump/zeme2_j002_28952.trc:
    ORA-12012: error on auto execute of job *20627358*
    ORA-12705: Cannot access NLS data files or invalid environment specified
    Tue Jul 26 11:52:43 2011
    Errors in file /u01/app/oracle/admin/zeme/bdump/zeme2_j001_11018.trc:
    ORA-12012: error on auto execute of job *20627357*
    ORA-12705: Cannot access NLS data files or invalid environment specified
    Tue Jul 26 11:52:43 2011
    Errors in file /u01/app/oracle/admin/zeme/bdump/zeme2_j000_9684.trc:
    ORA-12012: error on auto execute of job *20627342*
    ORA-12705: Cannot access NLS data files or invalid environment specified
    After examining trc files I have found no further information about error except session ids.
    My question is: how to find what job caused these messages to appear in alert log file.
    How do I map number in alert log file to some "real" information (owner, statement executed, schedule)?
    Marx.

    Sorry for the delay
    Try this to find the job :
    select job, what from dba_jobs ;
    How do I find NLS_LANG version?SQL> show parameter NLS_LANG
    Do you mean ALTER SESSION inside a job?I meant anywhere, but your question is better.
    ORA-12705 - Common Reasons and How to Resolve Them [ID 158654.1]
    If OS is Windows lookout for NLS_LANG=NA in the registry
    Is it possible you are doing this somewhere ?
    ALTER SESSION SET NLS_DATE_FORMAT = 'RRRR-MM-DD\"T\"HH24:MI:SS';NLS database settings are superseded by NLS instance settings
    SELECT * from NLS_SESSION_PARAMETERS;
    These are the settings used for the current SQL session.
    NLS_LANG could be set in a profile for example.
    NLS_LANG=_AMERICA.WE8ISO8859P1     ( correct )
    NLS_LANG=AMERICA.WE8ISO8859P1 ( Incorrect )
    you need to set the "_" as separator.
    Windows
    set NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P1
    Unix
    export NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P1
    mseberg
    Edited by: mseberg on Jul 28, 2011 3:51 PM
    Edited by: mseberg on Jul 29, 2011 4:05 AM

  • Could not load plugins: File not found. Any solution to that?

    I am using a macbook-pro 2009 which, recently upgraded to Mavericks. Since the upgrading I have been unable to open "video clips" from several sites, with the message : "Could not load plugins: File not found". I have made sure of getting the latest Adobe flashplayer and Java, but the problem is still there. Does anyone experience the same? Any solution, please .....

    If you use extensions (Firefox/Tools > Add-ons > Extensions) that can block content (e.g. Adblock Plus, NoScript, Flash Block, Ghostery) then make sure that such extensions aren't blocking content.
    Start Firefox in <u>[[Safe Mode|Safe Mode]]</u> to check if one of the extensions (Firefox/Tools > Add-ons > Extensions) or if hardware acceleration is causing the problem (switch to the DEFAULT theme: Firefox/Tools > Add-ons > Appearance).
    *Do NOT click the Reset button on the Safe Mode start window.
    *https://support.mozilla.org/kb/Safe+Mode
    *https://support.mozilla.org/kb/Troubleshooting+extensions+and+themes

  • Background Job cancelling with error Data does not match the job definition

    Dear Team,
    Background Job is getting cancelled when I run a Job on periodically but the same Job is executing perfectly when I run it manually(repeat scheduling) .
    Let me describe the problem clearly.
    We have a program which picks up files from an FTP server and posts the documents into SAP. We are scheduling this program as a background Job daily. This Job is running perfectly if the files contain no data. But if the file contains data the JOb is getting cancelled with the following messages.
    And also the same Job is getting executed perfectly when repeat scheduling is done ( even for files with data).
    Time     Message text                                                                       Message class Message no. Message type
    03:46:08 Job PREPAID_OCT_APPS2_11: Data does not match the job definition; job terminated        BD           078          E
    03:46:08 Job cancelled after system exception ERROR_MESSAGE                                      00           564          A
    Please help me in resolving this issue.
    Thanks in advance,
    Sai.

    hi,
    If you have any GUI function modules used in the program of job
    you cannot run it in background mode.

  • Background job in Released status and does not run. What is the problem?

    Hi Experts,
        I ran a ABAP report as a background job. But when I see the job status in SM37 I found that it is "Released" but the job does not run further, It remains in "Released" state. I have given the correct variant and start condition "immediate" . Moreover there is no spool and log as well.
    The ABAP report runs fine in forground. It used to run in background few days back.
    What could be the problem? Kindly help!
    Thanks
    Gopal

    Hi Bob,
        But I checked after 2 days still the job status is in Released state.  There were some other jobs that ran successfully. So atleast after two days I hope the job should have run. I can't even delete it. What shall I do?
    Thanks
    Gopal

  • Can we run in Background job which can pick a file from front end ?

    Hi All,
    I would like to run a report in Background job , for this i need to pick a file from front end services(Local drive or share folder)
    Is there any alternate FM=>GUI_UPLOAD  to upload data from front end services or any other solution for this?
    I am getting error=> "No batch" and " Cannot perform frontend function in batch input mode"
    Could you please help on this ?
    Thanks for your help in advance.
    Regards,
    Prasad.

    Hello Prasad,
    IF you keep your file on a shared folder on your front end, you can access it via OPEN DATASET if your application server has access to that path.
    You can test it out by asking Basis to try and access your front end file from the app server itself.

  • Log shipping is not restoring log files ata particular time

    Hi,
    I have configured log shipping and it restores all the log files upto a particular time after which it throws error and not in consistent sate. I tried deleting and again configuring log shipping couple of times but no success. Can any one tell me 
    how to prevent it as I have already configured log shipping from another server to same destination server and it is working fine for more than 1 year? The new configuration is only throwing errors in restoration job.
    Thanks,
    Preetha

    Message
    2014-07-21 14:00:21.62    *** Error: The log backup file 'E:\Program Files\MSSQL10_50.MSSQLSERVER\MSSQL\Backup\Qcforecasting_log\Qcforecasting_20140721034526.trn' was verified but could not be applied to secondary database 'Qcforecasting'.(Microsoft.SqlServer.Management.LogShipping)
    2014-07-21 14:00:21.62    Deleting old log backup files. Primary Database: 'Qcforecasting'
    2014-07-21 14:00:21.62    The restore operation completed with errors. Secondary ID: '46b20de0-0ccf-4411-b810-2bd82200ead8'
    2014-07-21 14:00:21.63    ----- END OF TRANSACTION LOG RESTORE     -----
    The same file was tried thrice and it threw error all the 3 times.
    But when I manually restored the Qcforecasting_20140721034526 transaction log it worked. Not sure why this is happening. After the manual restoration it worked fine for one run.ow waiting for the other to complete.
    This seems strange to me error points to fact that backup was consistent but could not be applied because may be restore process found out that log backup was not it correct sequence or another log backup which was accidentally taken can be applied.
    But then you said you can apply this manually then this must be related to permission. Because if it can restore manually it can do it with job unless agent account has some permission issue.
    Ofcourse more logs would help
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
    My TechNet Wiki Articles

  • Cannot conect to Job server after redirecting log files

    Platform: Windows
    Software: Data Services 3.0 (Data Integrator 12.0.0)
    Issue: When redirecting the Log files to go to another drive on the windows Server, we are unable to reconnect to the Job Server with Designer.
    We did the following to re-direct the Log files:
    In the DSConfig.txt file on the machine where the jobserver runs, they will find a section similar to the following:
    [AL_JobServer]
    AL_JobServerPath=C:\PROGRA1\BUSINE1\DATAIN~1.7\bin\al_jobserver.exe
    AL_JobServerLoadBalanceDebug=FALSE
    AL_JobServerLoadOSPolling=60
    AL_JobServerSendNotificationTimeout=60
    AL_JobServerLoadRBThreshold=10
    AL_JobServerLoadAlwaysRoundRobin=FALSE
    AL_JobServerAdditionalJobCommandLine=
    ServiceDisplayName=Data Integrator Service
    AL_JobServerName1=WestJS
    AL_JobServerPort1=3500
    AL_JobServerRepoPrefix1=
    AL_JobServerLogDirectory1=
    AL_JobServerBrokerPort1=
    AL_JobServerAdapterManager1=
    AL_JobServerEnableSNMP1=
    AL_JobServerName2=EastJS
    AL_JobServerPort2=3501
    AL_JobServerRepoPrefix2=
    AL_JobServerLogDirectory2=
    AL_JobServerBrokerPort2=
    AL_JobServerAdapterManager2=
    AL_JobServerEnableSNMP2=
    (As with any DSConfig.txt edits, I always tell customers to be very careful, to only follow instructions they receive from customer assurance, and to make backups of the prior file in case they need to revert.)
    Note that there is a u201CWestu201D jobserver defined with the 1 suffix, and an u201CEastu201D jobserver defined with the 2 suffix.  Both jobservers will by default log into %LINK_DIR%\log\ into a directory named after their jobserver name.  If the customer wants to redirect the logs for the u201CWestu201D jobserver to a different path, add the following to this section:
    AL_JobServerLogReDir1=D:\dierrorlogs\Log
    Note the use of the suffix 1 u2013 it means that it will be a part of the settings for the u201CWestu201D jobserver.  The u201CEastu201D jobserver will still log to %LINK_DIR%\log\EastJS while the u201CWestu201D jobserver will log to D:\dierrorlogs\Log.
    All we did was add the one single line AL_JobServerLogReDir1="E:\Program Files\BusinessObjects\dierrorlogs\Log"
    Can anyone give us some sort of suggestion or idea as to why we lost our connection to our Job Server with this change?  We DID restart the service.
    Thanks!  Ken

    I am not sure if this Parameter actually works ? I have to check that
    can you see the JobServer process in the task manager ? check the windows event log for any errors for the JobServe
    since you modified the path, I don't think it will be generating any log file for Job server
    what's the status for the job server in Desginer and Management console ?

  • Data Services 4.0 Designer. Job Execution but empty log file no matter what

    Hi all,
    am running DS 4.0. When i execute my batch_job via designer, log window pops up but is blank. i.e. cannot see any trace messages.
    doesn't matter if i select "Print all trace messages" in execution properties.
    Jobserver is running on a seperate server. The only thing i have locally is just my designer.
    if i log into the Data Services management console and select the job server, i can see trace and error logs from the job. So i guess what i need is for this stuff to show up in my designer?
    Did i miss a step somewhere?
    can't find anything in docs about this.
    thanks
    Edited by: Andrew Wangsanata on May 10, 2011 11:35 AM
    Added additional detail

    awesome. Thanks Manoj
    I found the log file. in it relevant lines for last job i ran are
    (14.0) 05-11-11 16:52:27 (2272:2472) JobServer:  Starting job with command line -PLocaleUTF8 -Utip_coo_ds_admin
                                                    -P+04000000001A030100100000328DE1B2EE700DEF1C33B1277BEAF1FCECF6A9E9B1DA41488E99DA88A384001AA3A9A82E94D2D9BCD2E48FE2068E59414B12E
                                                    48A70A91BCB  -ek********  -G"70dd304a_4918_4d50_bf06_f372fdbd9bb3" -r1000 -T1073745950  -ncollect_cache_stats
                                                    -nCollectCacheSize  -ClusterLevelJOB  -Cmxxx -CaDesigner -Cjxxx -Cp3500 -CtBatch  -LocaleGV
                                                    -BOESxxx.xxx.xxx.xxx -BOEAsecLDAP -BOEUi804716
                                                    -BOEP+04000000001A0301001000003F488EB2F5A1CAB2F098F72D7ED1B05E6B7C81A482A469790953383DD1CDA2C151790E451EF8DBC5241633C1CE01864D93
                                                    72DDA4D16B46E4C6AD -Sxxx.xxx.xxx -NMicrosoft_SQL_Server -Qlocal_repo  coo ds local
                                                    repo_azdzgq4dnuxbm4xeriey1_e" -l"C:\Program Files (x86)\SAP BusinessObjects\Data Services/log/js01/tip coo ds local
                                                    repo_azdzgq4dnuxbm4xeriey1_e/trace_05_11_2011_16_52_27_9__70dd304a_4918_4d50_bf06_f372fdbd9bb3.txt" -z"C:\Program Files
                                                    (x86)\SAP BusinessObjects\Data Services/log/js01/tip coo ds local
                                                    repo_azdzgq4dnuxbm4xeriey1_e/error_05_11_2011_16_52_27_9__70dd304a_4918_4d50_bf06_f372fdbd9bb3.txt" -w"C:\Program Files
                                                    (x86)\SAP BusinessObjects\Data Services/log/js01/tip coo ds local
                                                    repo_azdzgq4dnuxbm4xeriey1_e/monitor_05_11_2011_16_52_27_9__70dd304a_4918_4d50_bf06_f372fdbd9bb3.txt" -Dt05_11_2011_16_52_27_9
                                                    (BODI-850052)
    (14.0) 05-11-11 16:52:27 (2272:2472) JobServer:  StartJob : Job '05_11_2011_16_52_27_9__70dd304a_4918_4d50_bf06_f372fdbd9bb3' with pid '148' is kicked off
                                                    (BODI-850048)
    (14.0) 05-11-11 16:52:28 (2272:2072) JobServer:  Sending notification to <inet:10.165.218.xxx:56511> with message type <4> (BODI-850170)
    (14.0) 05-11-11 16:52:28 (2272:2472) JobServer:  AddChangeInterest: log change interests for <05_11_2011_16_52_27_9__70dd304a_4918_4d50_bf06_f372fdbd9bb3> from client
                                                    <inet:10.165.218.xxx:56511>. (BODI-850003)
    (14.0) 05-11-11 17:02:32 (2272:2472) JobServer:  RemoveChangeInterest: log change interests for <05_11_2011_16_52_27_9__70dd304a_4918_4d50_bf06_f372fdbd9bb3> from client
                                                    <inet:10.165.218.xxx:56511>. (BODI-850003)
    (14.0) 05-11-11 19:57:45 (2272:2468) JobServer:  GetRunningJobs() success. (BODI-850058)
    (14.0) 05-11-11 19:57:45 (2272:2468) JobServer:  PutLastJobs Success.  (BODI-850001)
    (14.0) 05-11-11 19:57:45 (2272:2072) JobServer:  Sending notification to <inet:10.165.218.xxx:56511> with message type <5> (BODI-850170)
    (14.0) 05-11-11 19:57:45 (2272:2472) JobServer:  GetHistoricalLogStatus()  Success. 05_11_2011_16_52_27_9__70dd304a_4918_4d50_bf06_f372fdbd9bb3 (BODI-850001)
    (14.0) 05-11-11 19:57:45 (2272:2472) JobServer:  GetHistoricalLogStatus()  Success. 05_11_2011_16_52_27_9__70dd304a_4918_4d50_bf06_f372fdbd9bb3 (BODI-850001)
    it does not look like i have any errors with respect to connectivity? ( or any errors at all....)
    Please advise on what, if anything you notice from log file and/or next steps i can take.
    thanks.

  • Not all iTunes Files not showing up on External HDD

    My old TC (the flat model from back to 2008 I think it was) broke down the other night so I removed the WD 2.0TB HDD (containing over 30'000 music files) to run this as an external HDD. I had swapped the original 500Gb with a 2TB drive which worked super for a couple of years.
    Anyway, before all of this happened all songs showed up on iTunes and each single one could be played correctly from iTunes (without the ! should the file not be found) so all files were located through iTunes to the correct directory. Only one directory for my Music files...
    Now, when I connect the HDD through an USB cradle (nothing has been touched on the drive) only a fraction of the files actually show up in the directory which is very odd indeed? I believe it was the same when the thing was actually working but didn't give it a think as it was doing what it was expected to do.
    Next step was that I copied the whole HDD over onto my MAC and sizewise all (?) the files came over but still not all are showing...
    So, how can I make sure all the files are being included as I want to move them all over to my new 2TB TC?

    Thanks Limnos for taking your time with this.
    Here's what happened
    1) All my Music was stored on a TC 2.0TB WD hard disk which I shared with my partner.
    2) All music in iTunes were mapped to this specific location and only this location.
    3) When opening the folder on the TC only a fraction of the actual music folders were available to see (i.e. Artist The Wonder Stuff folder was present and in this folder only 2 album folders were found when in fact there should have been 12)
    4) All songs in iTunes did play correctly so if i wanted to listen to one of the The Wonder Stuff songs they would all play (even albums that did not show up in the artist folder from above).
    5) When clicking show file in finder, the file would show up and then again when going back manually into the folder as above.
    6) TC dies on me. No more power.
    7) Took out the HDD and put it in a cradle to extract the songs.
    8) Right click on the folder shows there are about 35'000 files which sounds right BUT only a fraction of these can be seen.
    9) Now I cannot do the show in finder through iTunes as the directory no longer exists (the TC died and I threw it away, except for the HDD).
    10) The files are on the HDD but I cannot see but a fraction of them when opening the folder...
    I hope this makes it a little more clear.
    I honesty do not know what to do as I know the files are there but cannot see them and thus map them in iTunes... :-(

Maybe you are looking for

  • How to return more than one string?

    Hi,below is my code: @WebMethod(operationName = "CheckBooking") public String CheckBooking(@WebParam(name = "ID") String ID) { try { Class.forName("sun.jdbc.odbc.JdbcOdbcDriver"); Connection con = DriverManager.getConnection("jdbc:odbc:Flight"); // S

  • Problem injecting ejb in war

    This problem have dogged me for one week but cant seem to know where the problem is. i am creating an enterprise with war, ejb and jar. i want both web clients and application client to share the code in ejb to ease maintainablility. i have a jsf bac

  • Two iTunes Libraries two types of audio files?

    I like to rip my CDs to AIFF for Hi-Fi listening. However, I like to have more storage on my iPod. Can I keep one iTunes library with AIFF files and the other with Apple Lossless for iPod syncing? Both on the same computer.

  • Upgrade to mySAP ERP 2005 with Unicode

    Dear Experts, We are currentl looking at upgrading our 4.5B system to mySAP ERP2005 and have the following issues. 1.Our DB size is currently 2.4 TB and with Unicode we expect more....what should be the best strategy to control the DB size apart from

  • Bug? - Copying Controls in Group Removes Binding

    [LabVIEW 2010] When I copy a control binding remains intact however when I copy the controls in a Group binding is removed. Cheers -JG Unable to display content. Adobe Flash is required. Certified LabVIEW Architect * LabVIEW Champion