Problem in production environment

Hi,
Using Designer and Headstart to generating forms!
I have an application that runs on Client/server, I am using a template and .pll, .plx and .olb. and the application runs good!
For installing the same application on another environment, I have copied the *.fmx , *.pll, *.plx, *.olb,
I have added the necessary paths to the registry!
When running this application on this environement,
the toolbar does not appear under the menu on the main form!!!!
Have I need to copy the template(the horizontal toolbar is a property of the template and its set correctly) ???? or what is the problem?????
Thanks for help!!!

I think your case requires a SR with Oracle.
Idan.

Similar Messages

  • Problem on developing SSRS report in a production environment

    Any problem could  be caused if have to develop SSRS report in a production environment ?
    If so, how it can be prevented?
    Thanks!

    Hi ,
    As per my understanding, Normally our environment setup contains Development,testing,preproduction,production and its varies from company to company.
    We develop SSRS report in development environment and after testing we deploy it to production report server.
    Development environment use development database for creating report like your database IP and credential will use development environment
    setup. and production environment use production database for creating report like your database IP and credential will use production environment setup.
    Thanks
    Neha Prajapati Please Mark This As Answer or vote for Helpful Post if this helps you to solve your question/problem.

  • How to port database changes from development to a production environment

    How do I port database changes from the development to the production environment?
    I am using v8 and have always had to redo everything using the schema manager all over in the production environment. Is there an easy way to generate a script, for example to dump the database changes on the development machine to be executed later on the production machine?

    This should already be a clearly defined change control process. Once a procedure, function, package, trigger, or whatever completes the testing rounds, it should be promoted to production.
    Forgive me if it seems I'm trivializing, but I don't see the problem, just copy the object(s) from your software library (or development) into production using whatever tool works best or has been chosen. If you are doing data copies then you have various options again including good old export/import.

  • Problem with Path Managed property in alternate zones - also we have identified problem with Product catalog in alt zones

    Hi All,
    I'll start "one step back". Note we have resolved this issue but it is worth mentioning:
    Yesterday we set up the product catalog. We were having a lot of inexplicable problems with it initially. After much troubleshooting, we found that it was because we were browsing the SharePoint site via the "Internet Zone" rather than the "Default
    Zone" when we "Connected" our site to the product catalog. Issues experienced:
    1) "Catalog Item URL Format" displayed error "Properties <Managed Property> specified by the shared catalog could not be found in the search schema" when attempting to connect to the catalog, despite that managed property being
    configured correctly. The catalog would connect despite the error, and the navigation worked fine.
    2) The search Result Source for the catalog would not return any results. Configuring the query and selecting "Advanced Mode" would show SPSiteURL:http://externalURL.domain.com. Changing SPSiteURL manually to
    http://InternalURL resulted in results being returned as expected.
    After much troubleshooting, we found that if were browsing SharePoint using the default zone (http://InternalURL) when we connected the Product Catalog, the result source then worked properly (internally and externally).
    Ok, so now onto our current problem:
    The Product Catalog appeared to be working properly (internally and externally) after figuring out that we needed to connect it from the Default Zone. However, we have found a remaining glitch, which doesn't appear to be Product-Catalog-specific. Consider
    the following scenario:
    1) We click a category in our Product Catalog, such as Electronics
    2) A list of our electronic devices are presented via the "Category" display page
    3) If we click on a specific item, a different URL is returned internally vs externally. We only get the friendly managed navigation URL when browsing via the Default zone. And of course totally different pages load based on which URL you get.
    Internal (when browsing via Default zone):
    http://InternalURL/catalog/PRODUCT-CATEGORY/PRODUCT-NAME
    External (when browsing via Internet zone):
    http://ExternalURL.domain.com/catalog/Lists/Products/DispForm.aspx?ID=1
    We have reviewed the display template, and the "Path" managed property is simply being used to render the link. Furthermore, we have configured the Content Search web part's "Property Mappings" to show the "Path" MP, and it
    displays the same (wrong) result externally.
    This tells us that the Path MP is not correctly rendering "Friendly URLs" when using managed navigation.
    Any help/ideas?
    Our environment:
    SharePoint Version: 15.0.4667.1000
    Default Zone: http://InternalURL
    Internet Zone: http://ExternalURL.domain.com
    We are NOT using host named site collections
    We ARE crawling the Default zone URL in our content source
    Thanks,
    Tommy

    Hi,
    We have now also observed the same issue with a Managed Property of type “Hyperlink/Picture”. The picture returns the wrong URL externally.
    This post describes the same issue, however the poster found that setting UseAAMMapping=True doesn’t work in 2013:
    http://sharepoint.stackexchange.com/questions/104806/set-the-useaammapping-property-of-a-managedproperty-object-map-an-url-of-a-h
    We checked, and our Picture MP had UseAAMMapping=False. We set it to True, did a full crawl, and see no change in behavior. Path already had UseAAMMapping=True.
    More info:
    http://macslui.blogspot.com/2013/02/sharepoint-2010-problem-in-image-result.html
    https://camerondwyer.wordpress.com/2014/08/04/beware-sharepoint-2013-search-results-and-the-listurl-property/#comment-2606
    Thanks,
    Tommy

  • Problem in Production as Login Credentials changed

    Hi,
    This is a scenario from IDOC to JDBC.
    This is a case arised due to blocking of the id, which we were previously  using in XI for getting connected from R/3 systems, ie the RFC destinations from R/3 to XI were using that login credentials.
    So the scenarios failed.
    Again we reset the ID with the same password. But the scenario is failing.
    The idocs are reaching to XI , but the the pipeline steps are not executed. It is saying " INBOUND INTERFACE NOT FOUND"
    This is a problem in Production, and we ve not touched any interface or mapping.Previously the scenario was working fine.
    Pls give me the solution and we 'll give 500 points:-)
    Regards,
    Navneet

    Quick Checks:
    1. Check the status of User in R/3 which you have used in the RFC destination i.e SM59 and IDOC Adapter ? if it is locked , then unlock it. If password is the problem, then try to change the password.
    Don't reset it ...
    2. Also check out the status in XI, SM59... what is the user used ? is it correct etc... by testing the connection
    3. I think the IDOC Metadata is having problem.. Go to IDX2 and reimport the metadata in XI.
    Also note that, it is prefer to use Service User in SM59, and IDOC adapter instead of dialog use in the Production environment. so that no problem of locking etc
    Hope this helps`
    Rgds Moorthy

  • Weblogic 8.1 Server log size increase in Production environment

    Hi,
    Issue:: One of the log file is increasing in size and exceeding beyond the size mentioned in the configuration file resulting in application outage.
    Issue description:
    We are having problems with the log size in the Weblogic 8.1 server. The fileminsize has been mentioned in the config.xml.
    New log files like MYsvr.log00001,MYsvr.log00002, MYsvr.log00003, MYsvr.log00004 etc are also being generated appropriately when the max file size has been reached. But simultaneously, one of the files is growing in size, exceeding the limit mentioned in the configuration file. Eg.. the MYsvr.log00001 file is 800MB in size while the other files(MYsvr.log00002, MYsvr.log00003 etc are 10MB in size)
    This increase in size of the log has been resulting in an application outage.
    More Details:
    1. Server: BEA Weblogic 8.1 server
    2. Log size is fine in other environements. This is a problem only in the production environment.
    3. The entry in the config.xml is as follows:
    <Server ListenPort="6313" Name="MYsvr" NativeIOEnabled="true" TransactionLogFilePrefix="./">
    <ServerStart Name="MYsvr"/>
    <Log FileMinSize="10000" FileName="MYsvr.log" Name="MYsvr"
    NumberOfFilesLimited="true" RotationType="bySize"/>
    <SSL Name="MYsvr"/>
    <ServerDebug Name="MYsvr"/>
    <WebServer Name="MYsvr"/>
    <ExecuteQueue Name="default" ThreadCount="15"/>
    <KernelDebug Name="MYsvr"/>
    </Server>
    Could you please help with this issue ?
    Thank you.

    Can someone please provide a solution for the issue

  • Firewall problem in domain environment

    I have built two domains for testing purposes. Having deployed domain controllers, exchange servers, sccm/scom servers, sql servers along with some client computers I noticed that I had problems accessing some of servers/clients - I could not manage
    some of them directly with manage command from domain controller or access them via unc path. Some of them I could not ping neither. I was able to solve these problems by changing inbound firewall rules on these machines thus
    allowing some connections such as smb-in, dcom-in . . . In my production environment (I have been working there as system engineer for almost seven years) I have never had these problems - any domain member, whether it has been server or client, was easily
    accessible (managed from dc, unc, ping, . . .). I could deploy GPO with all necessary settings for inbound rules but it should be done automatically - as soon as machine is joined to the domain it must be accessible by using at least common protocols such as
    dcom, smb for managing or simple file copy operation. I checked my production environment again and there were no GPOs for altering default firewall settings on domain member computers so I have no idea why this is happening in my testing domains.

    No one but me has access to these machines. Also Symantec Endpoint Protection software is installed on these machines as it is the case with my production machines where everything is functioning flawlessly - as I mentioned I can access all my machines
    in production domain via computer management, smb, ping . . . One of my test domains has FFL/DFL Windows Server 2008 R2,  the same as my production domain, the other one has FFL/DFL Windows Server 2012 R2 and it is created for learning purposes. As I
    said, in both test domains, all domain computers have SEP installed - the same version and configuration as on my production machines. I have not done anything related to firewall in my test domains on problematic machines - I installed OS on them, joined
    to the domain, installed SEP and afterwards I have worked with specific product machine was created for - SCCM/SCOM, Exchange, SQL . . . servers and their clients.

  • RE: Production Environment Definition

    Brad,
    We use connected environments so that we do not have a single point of
    failure.
    We use multiple environments and connect them together in a star topology
    for reliability of service. Our servers (23 in total) sit out at branches
    in the back of beyond and the WAN connections between the servers are
    unreliable. One needs a reliable connection to the Name Service which sits
    on each Environment Manager. We have thus created 23 connected
    environments with an Environment Manager on each LAN. Connected
    environments are still a bit buggy but Tech Support is currently working on
    fixing the last of the problems. We are still on ver 2H15 for this reason.
    Disadvantages of this topology are that making distributions take a long
    time because referenced partitioning cannot be scripted in fscript and
    econsole only connects to one environment at a time.
    There is a Fort&eacute; consultant in Denver called Pieter Pretorius who has had a
    lot of experience with our connected environments. It may be worth
    chatting to him.
    Regards,
    Richard Stobart
    Technical Consultant for Fort&eacute;
    E-mail [email protected]
    Quick-mail: [email protected]
    Voice: (+ 27 83) 269 1942
    (+27 11) 456 2238
    Fax: (+ 27 83) 8269 1942
    -----Original Message-----
    From: Brad Wells [SMTP:[email protected]]
    Sent: Tuesday, February 10, 1998 11:52 PM
    To: 'Forte Users - Sage'
    Subject: Production Environment Definition
    Hello again,
    We are just starting to look at what it will take to setup a production
    Forte environment. I have some general questions regarding
    considerations that may affect the environment definition and thought
    maybe some of the more experienced users could share some thoughts on the
    following:
    1) What factors lead to the creation of multiple production environments?
    a. How many environments should you use in a production situation?
    b. Do people create separate environments for separate business units?
    c. Are there performance improvements to be had by restricting the
    number of server and client nodes included in a single environment?
    d. How do the performance benefits of multiple environments compare to
    the additional complexity of managing and maintaining multiple connected
    environments?
    The initial need is for an environment that will service approximately 50
    clients and contain a couple of server nodes (database and service
    related). However, as the environment grows, it could easily grow to a
    size of 600 clients encompassing approximately 15-20 server nodes.
    At this point in time, there is no need for the failover support of
    connected environments, but this is something we will need to add as the
    environment absorbs applications with high reliability needs. Should the
    environments be setup and connected right away or can this be easily
    added on an "as needed" basis? What other recommendations would you
    make?
    Has anyone taken advantage of Forte consulting services in defining the
    production environment? Where you satisfied with the results of the
    service?
    Thanks.
    Bradley Wells
    [email protected]
    Strong Capital Management, Inc
    http://www.strong-funds.com

    On Tue, 10 Feb 98 13:52:00 PST Brad Wells <[email protected]>
    writes:
    At this point in time, there is no need for the failover support of
    connected environments, but this is something we will need to add as
    the
    environment absorbs applications with high reliability needs. Should
    the
    environments be setup and connected right away or can this be easily
    added on an "as needed" basis? What other recommendations would you
    make?
    From the Forte Systems Management point of view, you can add them "asneeded"
    fairly easily.
    Now from the application source code point of view, implementing
    Fail/Over support
    is a different story... You will need to check your SO's dialog
    durations, handle
    DistributedAccessExceptions, "warm-up" your distributed references for
    F/O,
    design a solution for restoring global transient data, do lots of
    testings etc...
    So implementing Fail/Over is not only related to systems-management
    issues, it can
    have some influence on your application(s) source code.
    Hope this helps,
    Vincent Figari
    You don't need to buy Internet access to use free Internet e-mail.
    Get completely free e-mail from Juno at http://www.juno.com
    Or call Juno at (800) 654-JUNO [654-5866]

  • How to download run time ABAP dump in production environment

    Hello,
    I am getting run time ABAP dump in production environment. Please guide me onHow to download run time ABAP dump in production environment?
    Also, please tell me how best can analyze that dump. Like there are tools like gdb, purify etc. to analyze problems in other languages.
    Regards,
    Sameer.

    Hi,
    Go to ST22, and Double click on the error line, It will show you the detailed analysis.
    See the section How to correct the error for details on error correction and the instructions to download the log( you can download in the format you want - HTML would be a nice option).
    And there are no tools to analyze errors but ST22 has a large data regarding the location of the error, the reason and the source code which caused. So you need to use the debugger to reproduce the error and check the source code,
    Regards
    Karthik D

  • Proces chain problem in production

    Hello friends,
    i have faced the process chain problem in the production as "Status change of process ZD_BDIT_ACTIVATE" save status and trigge event if appropiate ? . and displayed message is " ABAP/4 Processor :DBIF_DSQL2_DEFAULT_CR_ERROR" . i want help to solved this problem and what exactly the problem is?
    actually this problem came to activation and its indicated red but when i went to data target and then i show that activation is not started or not trigger. please give me what is exactly problem and what is the proper solution.
    please reply me as soon as possible.

    Quick Checks:
    1. Check the status of User in R/3 which you have used in the RFC destination i.e SM59 and IDOC Adapter ? if it is locked , then unlock it. If password is the problem, then try to change the password.
    Don't reset it ...
    2. Also check out the status in XI, SM59... what is the user used ? is it correct etc... by testing the connection
    3. I think the IDOC Metadata is having problem.. Go to IDX2 and reimport the metadata in XI.
    Also note that, it is prefer to use Service User in SM59, and IDOC adapter instead of dialog use in the Production environment. so that no problem of locking etc
    Hope this helps`
    Rgds Moorthy

  • Production environment, please help...

    Hi,
    we have a large BC4J web based application.
    we plan to put it production to allow to 40 CONCURENT USERS to connect to...
    after testing the application in stand-alone OC4J. it works well, nut at some moments, the server print the message
    like out of memory and it becomes realy slow. i restarted the OC4J
    like java -Doc4j.userThreads=true -Xms100m -Xmx800m oc4j.jar to increase the heap and i hope that this will
    solve the problem.
    no for the production environment :
    hardware
    Sun server
    4 x CPU x 450
    4 G on RAM.
    software
    Sun Solaris 8
    Oracle 9iAS R 2 (9.0.2.0.0)
    so what is the good configuration for the OC4J in 9iAS to get our system support all users. should we
    use - Doc4j.userThreads=true -Xms100m -Xmx800m even whith 9iAS.
    thank you for your help.

    El Bachir,
    The heap size is dependent upon your application and no. of concurrent users you are using. Yes, you can specify the the heap size for OC4J also in an Oracle9iAS instance from the EMWebsite by navigating OC4J Instance ->Server Properties -> specify the parameters for the Java command.
    For details how to determine the heapsize please look at the Performancing Guide. Here is the link
    http://otn.oracle.com/docs/products/ias/doc_library/90200sol_otn/perf.902/a95102/optj2ee.htm#1012258
    regards
    Debu Panda
    Oracle
    Hi,
    we have a large BC4J web based application.
    we plan to put it production to allow to 40 CONCURENT USERS to connect to...
    after testing the application in stand-alone OC4J. it works well, nut at some moments, the server print the message
    like out of memory and it becomes realy slow. i restarted the OC4J
    like java -Doc4j.userThreads=true -Xms100m -Xmx800m oc4j.jar to increase the heap and i hope that this will
    solve the problem.
    no for the production environment :
    hardware
    Sun server
    4 x CPU x 450
    4 G on RAM.
    software
    Sun Solaris 8
    Oracle 9iAS R 2 (9.0.2.0.0)
    so what is the good configuration for the OC4J in 9iAS to get our system support all users. should we
    use - Doc4j.userThreads=true -Xms100m -Xmx800m even whith 9iAS.
    thank you for your help.

  • Struggling with oracle clustered production environment

    Hi Experts,
    I have a requirement to upgrade Oracle Data Integrator(ODI) from 10.1.3.5 to 11.1.1.6.3 We have a Clustered production environment where N1 will be up when N2 be down and viceversa.
    Here N1 and N2 are the ODI servers as well as DB(11g Release 2) servers. They both access the SHARED CLUSTERED database.
    From ODI we will generally point the Oracle clustered IP(Virtual IP) which will internally point either N1 or N2 whichever is active.
    ODI application wise we are clear about the procedure.
    Having some issues on DB related activities.
    1. Should I break the cluster definitely? Cant I do the activity without breaking the cluster?
    2. Do I need to point N1, N2, Clustered IP (Virtual IP) while doing the activities?
    3. Since its a clustered database, do I need to db related activities once or twice? (Twice means, manually on both the servers)
    4. As they are using same file structures (RAC), If the Virtual IP points N1 by default, assume that I create two new users and log in is success. Now i manually point N2 and what will happen if I try to log in to the two new users from N2. will it work?
    5. If it will not work, then what would be the solution for this?
    Please someone suggest about the clustered environment production problem.
    As this is high priority, early response would be highly appreciated.
    Need Expert's suggestions for the sequence of steps that to be carried out for a successful migration in clustered environment.
    Many thanks in advance..!

    This is an ODI question.
    Please discontinue your question here and continue it in your duplicate thread here Struggling with oracle clustered production environment
    Apart from that, ODI is a front-end tool. If you have any clue about RAC: for ODI it doesn't matter whether your database is RAC or not. the issue is you installed a front-end tool in a RAC environment.
    Finally:
    As this is high priority, early response would be highly appreciated.For high priority questions there is paid support. This is a forum of volunteers.
    Asking for high priority is IMO insulting and rude.
    Sybrand Bakker
    Senior Oracle DBA

  • ORACLE NOSQL INSTALLATION ON PRODUCTION ENVIRONMENT

    Hi ,
    I have already installed nosql in my test server using kvlite. My doubt is how to setup ORACLE NoSQL on production environment ?
    Please suggest .
    Thanks
    Edited by: 861012 on Sep 3, 2012 12:54 PM

    Hi Kamran,
    I am using doc - http://docs.oracle.com/cd/NOSQL/html/AdminGuide/Oracle-NoSQLDB-Admin.pdf
    Here on page-26 while "Create the Remainder of your Storage Nodes" i am getting following error -
    kv-> plan -execute -name "Deploy n02" deploy-sn 1 node02 5000
    Exception while executing command plan: Oracle NoSQL DB 11gR2.1.2.123 oracle.kv.impl.fault.OperationFaultException: Plan 9[Deploy n02] finished in state ERROR. Problem during plan execution: Exception creating connection to: node02; nested exception is:
    java.net.NoRouteToHostException: No route to host
    kv->
    I have already configure iptables to allow network traffic to port 5000. Still i am getting this error java.net.NoRouteToHostException: No route to host
    Please suggest how to resolve this.
    Thanks
    Edited by: 861012 on Sep 4, 2012 3:42 PM

  • NullPointerException in JHSLov in Production Environment

    Hi,
    I am using Jdeveloper 11.1.1.2.0 and JHeadstart 11.1.1.2.29 and I am facing the following problem:
    A java.lang.NullPointerException is occured in a jhs lov only in the deployment in a production environment, despite the fact that locally the same lov works flawlessly.
    Specifically, I have an updatable VO (through an entity) that has reference in another entity. The VO has access in two attributes of the referential entity and the lov is based on one of them, while the other has its returned value.
    The lov is a single jhs lov based on a vo from a readonly query and is appeared in a table (being placed in a tab region). The specific lov is used and tested in the local wls without any malfunctions.
    The real problem starts when I deployed the application in another wls. Following the same case as locally, when I try to access this specific lov, I get the following exception on pressing the ok button in lov pop-up:
    oracle.jbo.JboException: java.lang.NullPointerException, message=null
    at oracle.adf.model.binding.DCBindingContainer.reportException(DCBindingContainer.java:405)
    at oracle.adf.model.binding.DCBindingContainer.reportException(DCBindingContainer.java:467)
    at oracle.adf.model.binding.DCControlBinding.reportException(DCControlBinding.java:202)
    at oracle.jbo.uicli.binding.JUCtrlValueBinding.getAttributeFromRow(JUCtrlValueBinding.java:739)
    at oracle.jbo.uicli.binding.JUCtrlValueBinding.getAttribute(JUCtrlValueBinding.java:1314)
    at oracle.jheadstart.controller.jsf.bean.LovItemBean.getLovFieldValueInTable(LovItemBean.java:698)
    at sun.reflect.GeneratedMethodAccessor961.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    You should also take into consideration that almost the same jhs lov is used in another region in the jsf page without this problem, neither locally, nor on the deployment.
    Moreover, the version of the wls on which the application is deployed is *10.3.3.0* (different from the local wls of v.10.3.2.0)
    This matter is of high importance to me so I am really looking forward for a reply.
    Thank you in advance

    Zelda,
    To see what is going wrong, it is probably easiest to create a subclass of LovItemBean, override method getLovFieldValueInTable, and add some debugging statements.
    The method looks like this:
    public Object getLovFieldValueInTable()
    if (getLovFieldValueExpr() != null)
    return JsfUtils.getExpressionValue(getLovFieldValueExpr());
    JUCtrlValueBinding rowData =
    (JUCtrlValueBinding) getTableBean().getBindingCollectionModel().getRowData();
    String attribute = getLovFieldBinding().getAttributeDef().getName();
    Object value = rowData.getAttribute(attribute);
    return value;
    So, maybe the attribute variable is null, or has an invalid value.
    Steven Davelaar,
    Jheadstart team,

  • Datasource not replicated in Production Environment

    Hi all,
    I activate a datasource 2lis_04_p_comp in SAP R/3 development environment and replicate it in BW development environment. All works are OK.
    Then I transport this datasource in R/3 production environment a first time, then a second time after several modifications on my datasource. All transports and activation (first and second wave of transport on this datasource) are OK between development and production environment (on R/3).
    2LIS_04_P_COMP is same (on R/3) between development and production environment.
    But when I want replicate 2LIS_04_P_COMP in BW production environment, several fields are not replicated. Only fields wich are transported in first wave are replicated in BW production environment. Fields which are added (after datasource modifications) and transported in second wave of datasource transports, are not replicated (Datasource contains these fields on R/3 development and production environment). Whereas in development environment all fields are correctly replicated.
    I think, my problem stems from replication in BW production environment.
    How can I replicated all fields in my BW production environment ?
    Thanks,
    Rgds,
    Sébastien

    hi,
    Can u check the transport log for 2nd time transport from R/3 Dev to R/3 Prod...
    Did u see all the changes in R/3 prod for DS 2LIS_04_P_COMP ..
    try this again..
    R/3 prod
    RSA6--> Select 2LIS_04_P_COMP Click on Truck icon -->
    BW Prod
    RSA1--> Source Sys tab > Select ur SS> RightClk and Replicate Datasources
    If u still missing some fields .. come back to us
    Thanks

Maybe you are looking for

  • BADI/User exit for Purchase Requisition from MRP

    I am looking for a BADI or any enhancement when a purchase requisition is created or modified from MRP .ie. MD50 transaction. Can anyone help ? THanks in advance.

  • Time Machine back up to MyBook (2TB).

    When Time Machine backs up to WD MyBook (2TB), it transfers 100GB of my iMac HD (500GB capacity) each time. Consequently, MyBook is full. How do I get Time Machine to only add changes/alterations to the next backup?

  • Premiere Pro CC 2014 crashing

    Please help. I work on a Mac Pro just recently updated to OS X Yosemite. When I go to open a project that a client sent me I start to relink the files. When I click on the first one it immediately pulls up the "Premiere has quit unexpectedly". And ev

  • Is three decimal possible in SAP

    From the first day, we were told by our consultant that we could only use 2 decimal . We have some unit cost with three decimals. Is there any way to customize decimal? Please kindly advise. Thanks, Linda

  • The "File" class does not accept my paths

    Hi , I need yo creat an instance of a directory/file using the File class , but it seems to not recognize my paths: using : File f = new File("c:/Dir1/dir2/file.ext"); or: File f = new File("c:/Dir1/dir2"); allthow all do exist , when I use the f.exi