Organize JDev for large applications

Hello JDev team,
In order to properly manage very large applications and different versions of the same application (like maintenance and development) in the same environment, there are various settings that would find their more appropriate place at the level of the workspace.
I already saw some other threads discussing similar topics, I would only like to put them together and try to get some schedules :-)
1. One problem is that there are no nested workspaces, and you don't have good "visibility" outside your own workspace (one example, you cannot create a BC Client Data Model unless the BC project is in the same workspace).
Nested workspaces would be great for properly organizing a large application. If you implement this, please don't forget the visibility "details".
2. A quite different problem is related to the definition of libraries. Once you have to manage in parallel, in the same environment, different versions of the same application, things become hard to manage.
Most of the projects should have their associated library defined, as there are relations between projects. But a library entry is absolute, so it stands for one version. When you have to switch to another version (which is also under SCM), there are different physical paths on the disk, so different values for the libraries. And you have to change (one by one or directly in libraries.xml) their definitions.
This would be solved naturally if the libraries were defined for a workspace. And would also provide the possibility to have relative paths rather than absolute ones, which is easier to manage.
3. Base classes. It is nice to have this in the global Preferences, but, as somebody else noticed before me, it might be necessary sometimes to have different base classes for different applications, all in the same development environment. This comes similar to having base classes defined (also) at the level of the workspace.
4. Debugger/Breakpoints/Default Action. This is another kind of preference I would like to have at the level of the workspace, or even for each project. Especially because of the "Group of Breakpoints" section.
5. Deployment. Here, a new option "Relative to Workspace Directory" would do. Or, allow the definition of the default deployment directory for each workspace.
6. Web Browser/Proxy. Well, here my problem might be more specific. We access internet resources via a proxy, so when it comes to accessing Oracle resources (like Check for Updates...) from JDeveloper, the proxy setting is useful, even required.
But, in our intranet development, we do not use a proxy (not the same, anyway). So, for normal development purposes, there should be no proxy.
In this case, what I would like would be to have the global setting in the IDE preferences used for global resources, while the setting in the workspace (if any) will be used for the application-specific resources (e.g. servlet URL).
I am quite sure I did not cover all the possible enhancement topics, but I hope things will be checked carefully.
Apart from adding valuable features for developers, the approach of moving certain preferences from the IDE to a "work component" (workspace, project) would also help team-work.
Once one team member will set some preference(s) for a workspace (and put it under SCM), this setting(s) will be automatically available to all other team members.
Could you please comment my issues and tell me which of them (with the solutions that you foresee) could integrate 9.0.4?
Thank you,
Adrian

Hi Adrian,
Thanks for the detailed feedback.
I agree with you that there are significant issues when using project files at the moment in a team environment, particularly when an SCM system is involved. And since a recent survey indicated that the majority of JDeveloper users use some kind of SCM system, it's particularly important that we address this in future releases.
At the moment, I'm putting together some suggestions for future changes in this area from the point of view of the feature team who develop SCM support in JDeveloper. Your ideas are useful input into this process. Sadly, it's unlikely that any major changes to the way project files are stored will make it into 9.0.4, since 9.0.4 has already gone into "bugfix only" mode, and the focus of the 9.0.4 release has been on performance improvements and stability. However, we are looking at this issue for a future release (possibly 9.0.5, although these discussions are at a very early stage).
I've included comments on some of your specific points below.
1. Nested workspaces
There was some discussion about whether we should use nested projects a long time ago (i.e. get rid of the concept of a workspace altogether, and just allow projects to contain, and presumably inherit properties from, other projects). This is currently being tracked as ER 2048949. It requires a substantial change to the architecture of JDeveloper and could have a massive effect on components across the IDE. I believe that a lot of the issues with project files can be solved in other ways without going to the extreme of using nested projects / workspaces.
2. Library Definitions
Project-scope libraries should hopefully make this a whole lot easier in 9.0.3. I also like the idea of defining libraries at the workspace level so that they can be shared between projects; alternatively, using a "library definition" file as a separate, version-controllable resource that can be referred to by a project. I've filed ER 2689043 to track this.
3. Base Classes
This is a BC4J thing, and a bit outside my area of knowledge :) I'll pass your suggestions along.
4. Debugger Preferences
I'm not sure what you mean by "group of breakpoints section"
5. Deployment
Filed ER 2689052
6. Web Browser/Proxy
You might be able to make use of the "Exceptions" field in the proxy settings to help with this. Within Oracle, we have exactly the same issue, but normally use an exception list that looks something like this: "localhost|127.0.0.1|*.oracle.com|*.oraclecorp.com|*.oracleads.com"
One of the areas I'm interested in seeing improved is the concept of "configurations" that we currently have in project files. For example, it would be nice to define a workspace-scope configuration that's inherited by projects. Also nice would be the ability to allow read-only configurations (that are source controlled as part of the checked in project definition) and user/view-specific configurations (which would not be version controlled). The user/view specific configurations could inherit preferences from the read-only configurations, and the user could override things in a way that wouldn't affect other users (e.g. in the "Paths" preferences, change the output directory to one on their own local machine) without having to check out the project file from source control.
Thanks,
Brian
JDev Team

Similar Messages

  • How to use Source Code Control for Large Application?

    Hi, All!
    I would like to collect knowledge about "best practice" examples for using Source Code Control and project organization for relative large application (let's say approx 1000 SubVIs).
    Tools used:
    LabVIEW 8.0
    CVS Server
    PushOK CVS Proxy Client
    WinCVS
    With LabVIEW 8 we can organize large project pretty well. This described in article Managing Large Applications with the LabVIEW Project.
    I have read this article too: Using Source Control Software with LabVIEW In this Article Source Safe used, but with PushOK all looks nearby the same and works (some tricks for compare function are required).
    Example. Two developers working together on same project. Internally project is modular, so one developer will work with module "Analysis", and another one with "Configuration" without interferences. These modules placed into Subfolders as shown in example above.
    Scenario 1:
    Developer A started with modification of module "Analysis". Some files checked out. He would like to add some SubVIs here. So, he must also perform check out for the project file (*.lvproj), otherwise he cannot add anything into project structure.
    Developer B at the same time would like to add some new functions into module "Configuration". He also needed to check out project file, but this file already checked out by Developer A (and locked). So, he must wait until lvproj file will be checked in. Another way is mark *.lvproj files as text files in PushOK, but then one of developers will get conflict message by checking in and then merging will be necessary. This situation will coming very often, because in most cases *.lvproj file will be checked out all the time.
    Question: Which practice is better for such situation? Is Libraries better than folder for large project?
    Scenario 2:
    Developer C joined to the team. First, he must get complete project code for starting (or may be at least code of one Library, which assigned to him).
    Question: How it can be done within LabVIEW IDE? Or WinCVS (or other SCC UI) should be used for initial checkout?
    Scenario 3:
    Developer D is responcible for Build. Developers A,B,C have added lot of files into modules "Analysis", Configuration" and "FileIO". For building he need to get complete code. If our project splitted into folders, he should get latest *.lvproj first, then newly added SubVIs will appear in Project Explorer, then he should expand tree, select all SubVIs and get latest versions for all. If Project organized in Libraries, he must do the same for each library, isn't?.
    Question: Is this "normal way", or WinCVS should be used for this way? In WinCVS its possible with two mouseclicks, but I prefer to get all code from CVS within LabVIEW IDE recursively...
    That was a long post... So, if you already working with LabVIEW 8 with SCC used for large project, please post your knowledge here about project structure (Folders or Libraries) and best practices, its may be helpful and useful for all of us. Any examples/use cases/links etc are appreciated.
    Thank you,
    Andrey

    Regarding your scenarios:
    1. Using your example, let's say both developers checked out version 3
    of the project file. Assuming that there are only files under the
    directories in the example project, when Developer A checks in his
    version of the project, there will be new files in one section of the
    project separate from where Developer B is working. Developer B,
    notices that there is now a version 4 of the project. He needs to
    resolve the changes so will need to merge his changes to the latest
    version of project file. Since the project file is a text file, that is
    easy to do. Where an issue arrises is that after Developer B checks in
    his merged changes, there is a revision 5. When Developer A and B go to
    make another change, they get the latest version which will have the
    merged changes to the project file but not the referenced files from
    both Developer A and B. So when A opens version 5, he sees that he is
    missing the files that B checked in and visa versa. Here is where the
    developers will needs to manually use the source control client and,
    external to LabVIEW, get those new files.
    Where libraries help with the above scenario is that the library is a
    separate file from the project so changes made to it outside of the
    project do not require the project to be modified. So this time, the
    developers are using a single project again which time time references
    two libraries. The developers check out the libraries, make changes to
    the libraries, and then check those changes in. So when each developer
    opens the project file, since it references the project file, the
    changes to the library will be reflected. There is still the issue of
    the new files not automatically coming down when the latest version of
    the library is obtained. Again, the developers will needs to manually
    use the source control client and, external to LabVIEW, get those new
    files. In general, you should take advantage of the the modularity that
    libraries provide.
    2. As noted in the above scenario, there is no intrinsic mechanism to
    get all files referenced by a LabVIEW project. Files that are missing
    will be noted. The developer will then have to use the source control
    provider's IDE to get the initial contents of the project  (or library).
    3. See above scenarios.
    George M
    National Instruments

  • Organization Calendar for all applications

    Hi, it is possible to declare one time an Organization Calendar and refer it from each BPM process in the domain ?
    I ask this because it does not make sense to have a calendar for each process if all processes belongs to the same organization ...

    Hi Frank,
    We're still using your handy DBLoginModule. I don't see what would be changed in the java2.policy file to give everyone the RMIPermission and the jazn.jar app is remarkably unhelpful in assigning the perm to a role. Can you give a hint of what you suggest could be added to the java2.policy? Or some documentation with more info?
    Thanks!

  • Newest MBA or MBP for large applications at the same time?

    Love the newest MBA and really want it - the reviews are so fantastic - but I am simply afraid of making the same mistake I've made in the past and ordering the less powerful machine. I currently have a Dell and an HP, and neither will allow me to use photoshop CS5 BY ITSELF with any efficiency, let alone while open with other applications such as internet explorer. The quality of webcam conversations with Skype is also very poor, despite a very fast internet connection. I am fed up and converting to Mac, just not sure which one to get - but I need it to be the best for PS and Skype. Advice from those who have used the newest MBA? Thanks!

    Ferrell, I have done what you suggested. I clicked onto the Spaces icon in the upper control panel and assigned spaces to six apps: Mail, Safari, iPhoto, iMovie, Pages, and iWeb.
    The problem is that if I am one app and I want to put something into that app, then the minute my mouse cursor goes to get that other thing, the original application into which I want the thing put GETS WHOOSHED OFF THE DESKTOP. What I want is for the original app to remain open whilst I go to retrieve a file or folder, and stay open while I put that thing into the original app.
    How does one turn off Spaces? I don't mean to delete all the names from the settings.I mean, just turn it off for a while so I can do something which requires for two apps to be open at the same time????
    ~ Lorna in Southern California

  • Javafx 1.2 ready for large application?

    I try to use javafx to build a dashboard which may contain 5000 cells, I used Text, ImageView, Stack, HBox, VBox to build this component, which just like Table in HTML.
    I didnt notice that would be a problem in the begining, when more and more cells to be renderd, the program become slower, when cells up to about 500, I got out of memory error.
    So I try to make simple nodes using Text, HBox, then put 5000 cells into VBox, and configure jvm parameter to use 512M memory, then run this simple test. It takes about 1 and a half minutes to startup.
    I dont think javafx is acceptable for my dashboard, then I try to build simple program with java swing, using JLabel, JPanel, BoxLayout, JScrollPane, and the same has 5000 rows. This program takes about 3 seconds to startup and dont need to configure jvm parameter!
    And, I notice that javafx1.2 do not have scrollpane component, how to build this component with javafx? will javafx provide this component in the next version?

    Use [http://code.google.com/p/crudfx/|http://code.google.com/p/crudfx/]
    It includes:
    1. Layouts
    2. Grid/Table
    3. Tree
    4. Text/Password components
    5. Localization tools
    6. User preferences util
    7. Tools for asynchronous task
    8. Bridge to Web Services (XML, JSON)
    9. Bridge to SQL Servers (JDBC)
    10. Google Map component
    11. Sound synthesizer
    12. Other tools

  • Flow chart for FPGA application

    This is a general question but important: how should I setup the strategy to develop a LabView FPGA/RT/Windows application? In text-based programming languages flowcharts become useful to sort out where to start and where to end, but when I am developing Labview VIs in three spaces, FPGA VI, RT host VI, and Windows host VI which one would be the head and tail? Should one VI be completed then  the next one started or develop small portions of every VI at a time? I am using the later method which is not always efficient especially for large applications.
    I am wondering if any one has come across with those questions and can share his/er experiences.
    Thanks.

    I think the perfect place for you to start is the NI CompactRIO Developers Guide.  This can also be helpful if its a non-CompactRIO application but uses Windows, RT and FPGA.  
    Personally I like the idea of writing your FPGA VI first, and get it acquiring the data you need.  Then you can write the RT VI that accesses the data from the FPGA VI and does the processing.  Once you have this processing done, you can finish off in the Windows VI for User interface.  Of course as you go along you will have to go backwards to add modifications to the lower code to make communication possible, but it usually works out pretty good for me..
    David_L | Certified LabVIEW Architect
    LabVIEW Tools Network | LabVIEW Tools Network Developer Center

  • Download FREE NI TestStand Toolkit for Large Applicatio​n Developmen​t at NI Labs!

    In hopes of giving you a sneak peek of what we're
    thinking about adding to NI TestStand in the future, we have added the NI TestStand Toolkit for Large Application Development to NI Labs! The toolkit currently includes two features with more planned for the future:
    NI TestStand Type Differ - compares and merges differences between types and helps you
    manage type conflicts easily
    NI TestStand Batch Type Differ - automates the process of using the NI TestStand Type Differ on multiple master and updated files
    Again, we are considering adding these features
    to NI TestStand and want to get your feedback. Download the toolkit at NI Labs and tell us what you think!
    Regards,
    James M.

    Hi It looks interesting. But what about putting a little effort in the "View > Sequence File > Display File Hierarchy" that I think is a cool small tool. But When you print the hierarchy, the legend i missing. Alternative is to include the names of the sequence files in the hierarcy view. /Christian  

  • Warning:The EXPORT data cluster is too large for the application buffer.

    Hi Friends,
    I am getting following warning messages whenever I click on costing in "Accounting" tab.
    1. Costing data may not be up to date  Display Help
    2.  No costing variant exists in controlling scenario CPR0001
    3.  An error occurred in Accounting (system ID3DEV310)
    4. The EXPORT data cluster is too large for the application buffer.
      I can create project automatically from cprojects. PLan costs, budget and actuals maintain in WBS elements can be visible in cproject Object Links.
    Your reply is highly appreciated.
    Regards,
    Aryan

    Hi;
    Please, check the Note --> 1166365
    We are facing the same problem in R3, but apply this no fix it.
    Best regards.
    Mariano

  • What design aids for large-scale application (using LV6i) exist?

    I have a large-scale application that includes analog and digital I/O, motion control, multiple temperature readings, Ethernet communication, RS-232, DDE and ActiveX controls for communicating with other commercial software. We have to improve the system performance and ease the pain of maintaining and upgrading. What aids are there for large-scale application design and development?

    This doesn't exactly count as development "tools" but I can send you copies of three papers that I found when I was just getting started with LV. They can show you how to think about your problem--and that is really the hardest (and most important) part.
    Contact me directly and I'll email them to you. The archive is to large to post.
    Mike...
    Certified Professional Instructor
    Certified LabVIEW Architect
    LabVIEW Champion
    "... after all, He's not a tame lion..."
    Be thinking ahead and mark your dance card for NI Week 2015 now: TS 6139 - Object Oriented First Steps

  • Bc4j; large application approach

    To follow up on an offline conversation.
    If one were to build a Very Large Application or Application Suite using JDev, does the following make sense as an approach?
    1. Set up a project with all your Domains, entity objects and View objects to be used across all other projects. This is maintained by someone (like a dba) who puts validations, defaults, rules, and the like here. Periodically this could be generated into a library for use by other projects. ( Hmmm... I don't know how to do that yet... )
    2. Set up projects for each of the applications. The entity/view objects and such DO NOT appear directly in this project ( and you might not want all the programmers to have updateable access to 'em anyway... it needs to go through some sort of command/control structure ). They use the library to add their own special view objects and such... and build the application.
    Will that work? Is that the way it should be put together?

    Business Components can be packaged up for deployment and reuse like any java classes.
    You can easily create a JAR file for all of the common entity objects, associations, domains, and base view objects of use to multiple applications. You can use the standard JDeveloper deployment wizard for this, and indicate that you want to create a "Simple Archive File". What will get JAR'ed up will be all of the XML and *.class files that comprise your components.
    Then, any other BC4J project can include this "application core" JAR file as a library and then import all of the components (READ ONLY). All of the core components are available for reuse, but the imported components cannot be modified in the importing project.
    See the help topic:
    Developing Business Components[b]
    --[b]Working with Business Components Projects, ...
    ----Ways to Import Existing Business Components
    for more details.
    Each new application you build will likely need some application-specific view objects, so I wouldn't work under the assumption that all view objects need to live under the "core" project.

  • Oracle 10g Discoverer Reports & export to xls fails for large reports

    Hi ,
    We have following configurations:
    1: RHEL 5.4
    2: Discoverer :Version 10.1.2.48.18
    3: Oracle10g Apps Version : Version 10.1.2.0.2
    Issue:
    Most of small reports works fine ....but when large discoverer reports are executed the page
    keeps on refreshing for 15-20 hours but no output ....same for export to xls ......
    But same reports works fine in oracle9i for same data voulme....
    Observations:
    When on linux with top command the processes are monitored its observed that discoverer process
    dis51ws dies for large reports after 1-2 minutes ...& the page keeps on refreshing but no output....
    for 1-2 minutes it consumes 50-80% cpu utilisation then process disappears & cpu 80% idle ...
    It seems that as 10g Apps is installed on RHEL 5.4 ...non certfified OS causing an issue...
    Can any one adds more inputs in this regards......
    we have checked logs : below are log details :
    Below logs gives "Logkeys: exceptions discoiv.servlet_exceptions" for this report ...
    1:
    OC4J~OC4J_BI_Forms~default_island~1:
    10/04/11 12:11:29 Oracle Application Server Containers for J2EE 10g (10.1.2.0.2) initialized
    10/04/11 12:11:59 Using oracle.reports.util.EnvironmentGlobal class
    10/04/11 14:23:37 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:23:38 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114240115278
    10/04/11 14:25:42 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:25:42 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114254315439
    10/04/11 14:28:53 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:28:54 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114285615691
    10/04/11 14:29:13 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:29:13 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114291315728
    10/04/11 14:32:35 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:32:35 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114323615949
    10/04/11 14:32:48 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:32:48 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114324815982
    10/04/11 14:34:44 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:34:44 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114344616128
    10/04/11 14:34:55 Logkeys: exceptions discoiv.servlet_exceptions
    10/04/11 14:34:55 Discoverer Model - 10.1.2.48.18
    Session ID:2010041114345516155
    10/04/11 14:36:25 Tutalii: /oracle10gas/app/oracle10g/discoverer/lib/discoverer5.jar archive
    2:
    Discoverer~SessionServer~12
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Active Eul: EULADMIN
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    Calling GetData on Preference Repository
    DCSCORBAInterface::Delete called
    DCSCORBAInterface destructor called
    DCSCORBAInterface::Delete called
    DCSCORBAInterface destructor called
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    ASSERT [email protected]:436
    3:
    application.log
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Finding async request action forward
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Calling externalize
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_TOOL: dvtb xk-1ml versionzw1.0w kyxdvtbyxbisltyxbicho vzwwjyxjbisltyxjdvtby
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_VIEW: dv xk-1ml versionzw1.0w kyxdv bazw0w cszw25wyxpc vzw1wjyxjdvy
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_VIEW: lc
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_HS: dvhs
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.WorksheetModel.getStateString EXT_DS: dv_ds &qls_z!2=New GL Report.Clndr Id&arq=false&qls_x!14=Sheet 1.Closing Balance (Credit)&qls_x!3=New GL Report.Voucher No&qls_z!3=New GL Report.Frm Prd&fm=xml&qls_z!4=New GL Report.To Prd&qls_x!2=New GL Report.Prd Desc&qls_x!11=Sheet 1.Debit Amount&qls_x!4=New GL Report.Gl Voucher No&qls_x!9=Sheet 1.Opening Balance Debit&tss_s!0=New GL Report.Acct Id,lh,group,false&qls_x!1=New GL Report.Acct Sdesc&qls_z!1=New GL Report.Loc Desc&qls_x!10=Sheet 1.Opening Balance (Credit)&qls_x!12=Sheet 1.Credit Amount&aow=false&qls_x!7=New GL Report.Shrt Code&qls_x!13=Sheet 1.Closing Balance (Debit)&qls_x!6=New GL Report.Pmt Rct Dt&qls_x!8=New GL Report.Dtl Nrtn&sss=true&qls_x!5=New GL Report.Voucher Dt&qls_x!0=New GL Report.Acct Id&qls_z!0=New GL Report.Loc Id&ddsver=1
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.model.ViewerModelImpl.externalize [EXTERN_STATE]: $wksht$%24dv_ds%24%26qls_z%212%3DNew+GL+Report.Clndr+Id%26arq%3Dfalse%26qls_x%2114%3DSheet+1.Closing+Balance+%28Credit%29%26qls_x%213%3DNew+GL+Report.Voucher+No%26qls_z%213%3DNew+GL+Report.Frm+Prd%26fm%3Dxml%26qls_z%214%3DNew+GL+Report.To+Prd%26qls_x%212%3DNew+GL+Report.Prd+Desc%26qls_x%2111%3DSheet+1.Debit+Amount%26qls_x%214%3DNew+GL+Report.Gl+Voucher+No%26qls_x%219%3DSheet+1.Opening+Balance+Debit%26tss_s%210%3DNew+GL+Report.Acct+Id%2Clh%2Cgroup%2Cfalse%26qls_x%211%3DNew+GL+Report.Acct+Sdesc%26qls_z%211%3DNew+GL+Report.Loc+Desc%26qls_x%2110%3DSheet+1.Opening+Balance+%28Credit%29%26qls_x%2112%3DSheet+1.Credit+Amount%26aow%3Dfalse%26qls_x%217%3DNew+GL+Report.Shrt+Code%26qls_x%2113%3DSheet+1.Closing+Balance+%28Debit%29%26qls_x%216%3DNew+GL+Report.Pmt+Rct+Dt%26qls_x%218%3DNew+GL+Report.Dtl+Nrtn%26sss%3Dtrue%26qls_x%215%3DNew+GL+Report.Voucher+Dt%26qls_x%210%3DNew+GL+Report.Acct+Id%26qls_z%210%3DNew+GL+Report.Loc+Id%26ddsver%3D1%24dv%24xk-1ml+versionzw1.0w+kyxdv+bazw0w+cszw25wyxpc+vzw1wjyxjdvy%24wd%24false%24lc%24%24dvtb%24xk-1ml+versionzw1.0w+kyxdvtbyxbisltyxbicho+vzwwjyxjbisltyxjdvtby%24wvs%241101%24dvhs%24$cn$&vct=svd&cnk=cf_a101$ap$%26df%3D%26l%3D%26s%3D%26nc%3D%26dl%3D$expl$&sp=&node=&event=focus&state=(117)&root=63&wbt=2$prid$NEW_GL_REPORT%2F31$ctyp$viewer
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Storing state
    10/04/11 14:37:25 discoverer: [INFO] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Externalize Perf: 8ms
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Saving Attributes
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Saving ApplicationRequest
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Checking for SSO mode
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Saving errors, messages
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Returning final forward: "/ExportProgress.uix" redirect: "false"
    10/04/11 14:37:25 discoverer: [TRACE] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute ----------------------- End Request --------------------------
    10/04/11 14:37:25 discoverer: [INFO] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.framework.ApplicationController.execute Total Request Time in AppCtrl: 18
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] org.apache.struts.action.RequestProcessor.processForwardConfig processForwardConfig(ForwardConfig[name=long running operation,path=/ExportProgress.uix,redirect=false,contextRelative=false])
    10/04/11 14:37:25 discoverer: [DEBUG] [AJPRequestHandler-ApplicationServerThread-17] oracle.discoverer.applications.viewer.view.DiscovererPageBroker.isCacheable Setting cacheable to: true
    4: XML log:
    log2010041114285615691.xml
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcstim.cpp</FILE_NAME> <LINE_NUMBER>184</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:29:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:29:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="NOTIFICATION"></MSG_TYPE><MSG_LEVEL>4</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[Timer started.]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcstim.cpp</FILE_NAME> <LINE_NUMBER>162</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:29:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:29:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="TRACE"></MSG_TYPE><MSG_LEVEL>5</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[Return DCSModelInterface::SendReceiveData(kScheduleInterface, inTable, outTable)
    outTable = DCITable
         Length=0
    ]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcsmdli.cpp</FILE_NAME> <LINE_NUMBER>259</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:29:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:29:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="NOTIFICATION"></MSG_TYPE><MSG_LEVEL>4</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[DCSModelInterface::SendReceiveData(kScheduleInterface)]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS </MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcsmdli.cpp</FILE_NAME> <LINE_NUMBER>226</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:29:13 2010
    </LOG_TIME> <LOG_SIZE>0</LOG_SIZE> <EXTRA_INFO><MethodEnd duration="0.1" sizeChange="0" >
    real 0m0.1s
    user 0m0.900s
    sys 0m0.109s
    </MethodEnd></EXTRA_INFO> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:37:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="NOTIFICATION"></MSG_TYPE><MSG_LEVEL>4</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[Timer stopped.]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcstim.cpp</FILE_NAME> <LINE_NUMBER>184</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:37:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    <MESSAGE><HEADER><TSTZ_ORIGINATING>2010-04-11T14:37:13+00:00</TSTZ_ORIGINATING> <COMPONENT_ID>DISCOVER</COMPONENT_ID> <MSG_TYPE TYPE="NOTIFICATION"></MSG_TYPE><MSG_LEVEL>4</MSG_LEVEL> <HOST_ID>2010041114285615691</HOST_ID> <MODULE_ID>DCS</MODULE_ID> </HEADER> <PAYLOAD><MSG_TEXT><![CDATA[Timer started.]]></MSG_TEXT> <SUPPL_DETAIL><![CDATA[
    <MSG_GROUP>DCS</MSG_GROUP> <PROCESS_ID>15691</PROCESS_ID> <FILE_NAME>dcstim.cpp</FILE_NAME> <LINE_NUMBER>162</LINE_NUMBER> <THREAD_ID>-1283799360</THREAD_ID> <LOG_TIME>Sun Apr 11 14:37:13 2010
    </LOG_TIME> ]]>
    </SUPPL_DETAIL> </PAYLOAD> </MESSAGE>
    Regards,

    Hi ,
    as per Note: 466697.1 ....its an memory error....& need to increase MaxVirtualDiskMem and MaxVirtualHeapMem
    But we already have slowly increased MaxVirtualDiskMem and MaxVirtualHeapMem to below very high values ...but the issue remains same.......
    as per note we are getting Logkeys: exceptions discoiv.servlet_exceptions error but
    after that we are not getting below error ............
    Unexpected error in state machine: java.lang.OutOfMemoryError
    Hence I presume that its different issue rather than memmory....
    Below are pref.txt values..........
    CacheFlushPercentage          = 25          # Percent of cache flushed if the cache is full. valid values 0 - 100%.
    MaxVirtualDiskMem          = 9294967296     # Maximum amount of disk memory allowed for the data cache. Should be greater than or equal to MaxVirtualHeapMem
    MaxVirtualHeapMem          = 4294967296     # Maximum amount of heap memory allowed for the data cache.
    QueryBehavior = 0          # Action to take after opening a workbook (0 = Run Query Automatically, 1 = Don't Run Query, 2 = Ask for Confirmation)
    Also we have raised an SR & as per SR .............all below settings are tried as per SR except for applying a recent patch....
    ====================================================================
    Discoverer performance is largely determined by how well the database
    has been designed and tuned for queries.
    A well-designed database will perform significantly better than a poorly
    designed database.
    Workbook design can also affect query performance.
    1. Apply latest Discoverer patch as documented in <<Note:237607.1>>
    'ALERT: Required and Recommended Patch Levels For All Discoverer Version'.
    2. Increase the maximum JVM heap memory:
    In general, the default values for the minimum heap (-Xms) memory and
    maximum heap (-Xmx) memory are sufficient.
    However, if your organization consistently runs large Discoverer queries
    then you may benefit from increasing the maximum heap memory from the
    default values.
    This is recomended if your users are typically running large queries via
    Discoverer Viewer.
    Increasing the JVM memory can help to avoid "java.lang.OutOfMemoryError"
    in Discoverer Viewer:
    Please see <<Note 563960.1>>, Best Practice: Configuring The
    OC4J_BI_Forms JVM For
    Discoverer Viewer/Portlet 10g Performance And Stability for specific
    details.
    3. Disable Query Prediction:
    Query Prediction provides an estimate of the time required to retrieve
    the information in a query.
    The Query Prediction appears before the query and consumes time.
    Edit the <oracle_home>/discoverer/util/pref.txt on the middle-tier
    server and set:
    QPPEnable=0
    Also set:
    QPPObtainCostMethod = 0
    4. Uncheck the 'Enable fantrap detection' checkbox.
    When the box is checked, every query generated by Discoverer is
    interrogated. Discoverer will detect a fan trap and rewrite the query to
    ensure that the aggregation is done at the correct level.
    Please refer to <<Note:210553.1>>, Oracle BI Discoverer: Fan Trap
    Resolution - Correct Results
    Everytime for more information on fantraps.
    To disable, in Plus go to Tools -> Options -> Advanced -> Fan Trap
    settings.
    In Viewer go to Preferences and uncheck the box.
    5. Disable Materialized Views/Summaries
    In pref.txt add parameter:
    MaterializedViewRedirectionBehavior = 0
    Value equal to 0 ensures that Materialized View (MV) Redirection is
    always performed when MVs are available.
    6. Improve query performance by optimizing the SQL.
    In pref.txt modify/add following parameters:
    SQLFlatten = 0
    SQLItemTrim = 0
    SQLJoinTrim = 0
    UseOptimizerHints = 0
    If value of SQLFlatten is 1 then Discoverer will merge inline views in
    the query SQL where ever possible.
    In case of SQLItemTrim, Discoverer will remove unused folder items from
    the query SQL where possible and for SQLJoinTrim Discoverer will remove
    unnecessary joins from the query where possible.
    UseOptimizerHints will add Optimizer hints to SQL if set >
    0.Unnecessarily making Discoverer perform these checks consumes
    resources and rather than increasing the performance may reduce the same. So unless you feel these checks have to be performed depending on
    requirements, assign zero to these parameters.
    7. When Discoverer builds a query, Discoverer makes a database security
    check to confirm that the user has access to the tables referenced in
    the folders. Avoiding this check can save time.
    So in pref.txt underDatabase section add:
    ObjectsAlwaysAccessible = 1
    8. Whenever you are creating conditions, ensure that you match the Case.
    This in turn can reduce the time Discoverer spends on changing the Case
    and matching.
    For example:
    option Upper(Department) in (Upper('VIDEO SALES')
    9. Ensure that summaries are refreshed periodically in Discoverer
    Administrator.
    10. Increase the amount of memory available for the Discoverer data cache. Please refer to <<Note 245752.1>>, Explaining Oracle BI Discoverer
    Session Memory Management
    And Server Cache Settings.
    11. Performance may be enhanced by enabling OracleAS Web Cache.
    ==================================================================
    Thanks for your reply....................
    Regards,

  • Does on-premises OneDrive for Business ***REQUIRE*** an Organization ID for synchronization to local systems?

    We have an on-premises SharePoint 2013 SP1 deployment in our lab.  We access the server through three methods: direct LAN access inside the firewall, over the Internet via Forefront TMG, and over the Internet via DirectAccess.  In all three access
    cases, almost everything functions exactly as expected.
    One exception is OneDrive for Business (the on-premises version).  After creating my MySite, I can browse my OneDrive for Business, upload/download files, preview documents, share with other SharePoint users, etc.  That stuff works properly. 
    But if I click the Sync link to establish a synchronization connection between my OneDrive for Business and my desktop system, the "We're getting things ready to sync..." dialog gets overridden by a OneDrive for Business error box - "We can't
    connect to the specified SharePoint site...".
    We see this behavior whether inside or outside the firewall.  I am logged into Windows, Office 2013 applications, and SharePoint using the same credentials: my Windows domain account.
    Running Wireshark and Fiddler on my workstation, I see conversations being sent to live.microsoft.com when I click on the Sync link.  This leads me to believe that *ALL* such sync connections are established by looking up the user's account in the Microsoft
    cloud (even if the credentials are for an on-premises domain) and, further, that because of this, businesses must create an official in-cloud Organization ID for their domain even if they are not actually using Office365 services (e.g. even if they are only
    using on-premises SharePoint) if they want to sync with on-premises OneDrive for Business.
    I have seen almost zero documentation about this, and certainly nothing that states that you MUST register for an Organization ID even if you are only using on-premises SharePoint and on-premises OneDrive for Business, yet this seems to be the behavior I
    see and that theory seems to be supported by my network traces.
    So...
    1) Can anyone confirm or deny this theory, or at least give a lot more detail about how SharePoint attempts to "find" the service when you click on the Sync link?
    2) If you HAVE successfully set up on-premises OneDrive for Business and have NOT had to register an Organization ID, can you please share the details of how you configured things?
    Many thanks for any help or pointers...
    tkarp

    I Just checked pre-requisites for Onedrive on-prem:
    To support OneDrive for Business document libraries in SharePoint Server 2013, you must have:
    A SharePoint Server 2013 My Site Host.
    Configured the User Profile Service application.
    The user must have an existing My Site with the personal site collection created.
    To support synchronizing document libraries to a user's client workstation with the OneDrive for Business Windows Sync client, you must have:
    Either
    Office 2013 Standard or Office 2013 Professional Plus, or the stand-alone OneDrive for Business Windows Sync client.
    An Office 365 subscription that includes the Office 2013 applications.
    Client workstations running Windows 7 or Windows 8.
    If you have to purchase Office 365 then you will need Organization ID which will be required for Document SYNC, or yous should have Office 2013 installed on system.
    If this helped you resolve your issue, please mark it Answered

  • How to configure ENV and DB for multithreaded application?

    Hi,
    From document, I know DB_THREAD must be checked for both ENV and DB, but , I don't know which one is best choice for multithreaded application while facing DB_INIT_LOCK and DB_INIT_CDB. In my application, there maybe multi readers and writers at the same time, should I use DB_INIT_LOCK instead of DB_INIT_CDB? what other flags should I use?
    DB_INIT_CDB provides multiple reader/single writer access while DB_INIT_LOCK should be used when multiple processes or threads are going to be reading and writing a Berkeley DB database.
    Thanks for your seggestions and answers.

    Thanks for the explanation,
    The Berkeley DB Concurrent Data Store product
    allows for multiple reader/single writer access
    to a database. This means that at any point in time,
    there may be either multiple readers accessing a
    database or a single writer updating the database.
    Berkeley DB Concurrent Data Store is intended for
    applications that need support for concurrent updates
    to a database that is largely used for reading.
    If you are looking to support multiple readers and
    multiple writers then take a look at the Transactional
    Data Store product
    (http://download.oracle.com/docs/cd/E17076_02/html/programmer_reference/transapp.html)
    In this case the Environment is typically opened with:
    DB_INIT_MPOOL, DB_INIT_LOCK, DB_INIT_LOG, and DB_INIT_TXN.
    Let me know if I missed any of your question.
    Thanks,
    Sandra

  • Slow performance for large queries

    Hi -
    I'm experiencing slow performance when I use a filter with a very large OR clause.
    I have a list of users, whose uid's are known, and I want to retrieve attributes for all users. If I do this one at a time, I pay the network overhead, and this becomes a bottleneck. However, if I try to get information about all users at once, the query runs ridiculously slow - ~ 10minutes for 5000 users.
    The syntax of my filter is: (|(uid=user1)(uid=user2)(uid=user3)(uid=user4).....(uid=user5000))
    I'm trying this technique because it's similar to good design for oracle - minimizing round trips to the database.
    I'm running LDAP 4.1.1 on a Tru64 OS - v5.1.

    This is a performance/tuning forum for iPlanet Application Server. You'd have better luck with this question on the Directory forum.
    The directory folks don't have a separate forum dedicated to tuning, but they answer performance questions in the main forum all of the time.
    David

  • Separate CSS files for Differenct Applications in the Same Workspace

    I am working in APEX 4.0, using Theme 4 (Topaz), primarily with the region template "Top Bar." The application I am working on has gotten so large that I decided to break it into two applications. So I copied the original application and then deleted the pages that I did not want in the second application. I then modified the authentication for a single login. One of the nice things about the Top Bar template is that you can color the background. I had a blue background in the first application and wanted to use a differenct color in the other. So I created a second CSS file and uploaded it to the Cascading Style Sheets area of the Shared Components file for the second application. Turns out that both it and the style sheet for the first application now show up in the CSS file area. The CSS file for the second application has no effect. Here is the CSS code:
    .rc-content-main
          background-color: #E5E5AA;
         }I also tried giving the region on each page in the second application a static id and then identifying them in the CSS file by the id. That didn't work either, at least not yet. Sometimes it takes the server a while to refresh.
    Question: For two applications in the same workspace, can there be separate CSS files? If so, how do I set them up so that each responds only to one of the applications?
    Edited by: Doug on Nov 14, 2012 4:54 AM
    Edited by: Doug on Nov 14, 2012 7:16 AM

    Hello,
    If you upload your file to Shared Components >> Cascading Style Sheetst then:
    edit your page template in the header section and just before +*</head>*+ tag  add :
    <link rel="stylesheet" src="#WORKSPACE_IMAGES#MY_FILE.css" type="text/css" />And that will be applied for all pages that have the same Page Template.
    You can edit a page and put this in the header section to be applied only for a that page:
    <style>
    <link rel="stylesheet" src="#WORKSPACE_IMAGES#MY_FILE.css" type="text/css" />
    <style>If the file is on your web server then replace #WORKSPACE_IMAGES# with #IMAGE_PREFIX#
    This should be helpful to you:
    http://docs.oracle.com/cd/E23903_01/doc/doc.41/e21674/concept_sub.htm
    Best Regards,
    Fateh
    If you believe that my answer is correct or helpful to you, then please mark it as helpful or correct...

Maybe you are looking for