Issue in Mapping _PO Create

Hi XI experts ,
Thsi is  XML (SOAP) mapped to PO BAPI(ZMM_IBX_RFC_CREATE_CABIN_ORDER) .Standard BAPI  with commit statement.
The External partner through xml sends the data  of :Purchase Order
The XMl contains the COST CENTRE , GL ACCOUNT  , WBS ELEMENT information.
-Take a case where  the Account Category is : ZNWBS (Project in ME23N PO) ,  WBS_ELEM_N is sent .
Eg: When :<ReferenceTypeCoded>AccountCategory</ReferenceTypeCoded> =ZNWBS , there is WBS DATA SENT:
<ReferenceTypeCoded>Other</ReferenceTypeCoded>
                <ReferenceTypeCodedOther>WBS_ELEM_E</ReferenceTypeCodedOther>
                <PrimaryReference>
                  <Reference>
                    <RefNum>U.888206.0</RefNum>
-Similary when <ReferenceTypeCoded>AccountCategory</ReferenceTypeCoded> =ZNCC (in ME23N), Cost centre data is sent:
<ReferenceTypeCodedOther>COST_CTR</ReferenceTypeCodedOther>
                <PrimaryReference>
                  <Reference>
                    <RefNum>1222</RefNum>
- In Case of 50% , 50% split cost centre data :
1 . 50% -- > have Account Category :ZNWBS , and  only WBS_ELEM_E, NOT COST_CTR
2. 50%  have Account Category :ZNCC , have  COST_CTR
I applied this logic for WBS_ELEM_E in XI:
ReferenceTypeCodeOtHER| EQUALS (TEXT )| WBS_ELEM_E | If Then | remove context | Split vy value(Each Value)| WBS_ELEMNT( Fiield of BAPI)
Now the issue is WBS ELEMNT  gets pouplated under :PO_ACCOUNT of first 50% item  and COST CENTRE data lso populated in fisrt item of :PO ACCOUNT .Here COST centre data is of second 50%  Line Item , so it is supposed to be populating in second line item not in first .
Since i have used :Remove Context  (same as logic of WBS_ELEMENT), it is supressing all the values, and single value of COST_CTR instead of passing to 2nd line item , it is passing to first lineitem .
Hence there is mismatch of data .
Ideally : 1 line item should be populated with : WBS _LEMENT
             2 Line item should be populated with :COST_CTR element.
Kindly provide your inputs

I have used this logic in WBS_ELEMNT:
ReferenceType Coded Other| EQUALS (text) | If   without else|   RefNum ||  Remove Context | SplitbyValue | WBS_ELEMENT
This is the Result:
1     Percentage     Cost Centre     WBS ELEMENT     Remarks
     50%     X     A     Working
     50%     Y     A     
2     Percentage     Cost Centre     WBS Element     Remarks
     50%          B     Working
     50%          C     
3     Percentage      Cost Centre     WBS ELEMENT     Remarks
     20%     Blank--No Data     A     "Not Working -Value X, is getting populating 
instead of Null Value"
     50%     X     B     
     30%     Y     B     
Your Input is appreciated.
Best Regards,
Vara

Similar Messages

  • Issue in mapping.. Source message expect with prefix ns0 in input message!

    Issue in mapping.. Source message expect with prefix ns0 in input message
    Hi All,
    I used the XSD structure which I got from partner which starts like below.
    <xsd:schema xmlns:xsd="http://www.w3.org/2001/XMLSchema" targetNamespace="http://xxxx.com/DirectSales/CustomerData" xmlns:prodata="urn:schemas-progress-com:xml-prodata:0001" xmlns="">
    <xsd:element name="CustomerRequest" prodata:proDataSet="true">
    I'm receiving the messages through SOAP adapter and the message looks like below.
    <?xml version="1.0" encoding="UTF-8" standalone="yes" ?><!-- Technical Routing --> <CustomerRequest xmlns='http://XXXX.com/DirectSales/CustomerData' xmlns:xsd='http://www.w3.org/2001/XMLSchema' xmlns:SOAP-ENV='http://schemas.xmlsoap.org/soap/envelope/' xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance' xmlns:SE='http://schemas.xmlsoap.org/soap/encoding/'>
    The message is failing during mapping. When i checked in message mapping, I found that the input source message is expecting with prefix "ns0" like <ns0:CustomerRequest...
    how I can solve this issue? Is there anyway I can add the prefix "ns0" before it comes to mapping step.?
    Thanks
    Deepthi

    Hi Guys,
    I tried to use the below paramter.
    anonymizer.acceptNamespaces         'http://XXXX.com/DirectSales/CustomerData  ns0
    The ns0 is coming in all the tags including segments and fields. I just want ns0 to come only at the root level ie <ns0:customerReq..>
    it is coming like below...
    - <ns0:CustomerRequest xmlns:ns0="http://XXXXXX.com/DirectSales/CustomerData">
    - <ns0:ttCustomer>
      <ns0:branch-num>1424</ns0:branch-num>
      <ns0:cust-num>121</ns0:cust-num>
      <ns0:contact-code>3</ns0:contact-code> 
       </ns0:ttCustomer>
      </ns0:CustomerRequest>
    I want the message to come as
    - <ns0:CustomerRequest xmlns:ns0="http://XXXXXX.com/DirectSales/CustomerData">
    - <ttCustomer>
      <branch-num>1424</branch-num>
      <cust-num>121</cust-num>
      <contact-code>3</contact-code> 
       </ttCustomer>
      </CustomerRequest>
    Kindly help me with the paramater value  to be passed for anonymizer.acceptNamespaces.
    Thanks
    Deepthi

  • Issue in mapping Operation subcontracting scenario

    Hi all,
    I will explain my scenario of subcontracting, i am getting issue when mapping the scenario in SAP.
    Scenario: For manufasturing 1 SF my client require 4 operations. All the raw materials required will be issued in the first operation. After first operation he will send the product manufactured to outside vendor and again he receives it after some value additiion. He continues the next two steps and finally manufactures SF which inturn will be used to manufacture his FG.
    Configuration: I have assigned a purchase info record in the operation detail. I have not given any WC to that operation because the cost should be captured only from the purchase order during the GR from vendor. After the confirmation of the first operation i am getting a purchase requisition for operation without a material.
    Issue: I am not able to show the material which he is sending to the vendor as my purchase requisition doesnot contain any materials.
    Please suggest me any possible solution to outcome this scenario.
    Thanx in advance
    Regards,
    Krishna Chandra

    Hi all,
    i got my the answer for my scenario by this link
    http://forums.sdn.sap.com/thread.jspa?threadID=1805266
    Regards,
    Krishna Chandra

  • CO-PA for good issue documents are not created

    Dear All
    We are using both account & cost based CO-PA , actually the problem is this that when we are doing the sales cycle, after billing my CO-PA document created  but the same when I do delivery COGS document not created  ,actually for sales I am taking the sales data from condition type(assignment of condition type to CO-PA value filed) and for COGS , I have assigned the COGS  GL accounts in TRANSACTION Code OKB9 (automatic determine the profitability segment) also the same GL accounts I have assigned  with value field in
    Assignment PA transfer structure
    Now when I am doing sales cycle , CO-PA document has been created when I am doing billing but when I am doing Good issue , CO-PA not created  but if I go and check the GL account of  COGS account I found the all detail of my profitability segment which hit this GL account
    Please help in this regard.
    Regards
    Kashif

    I assume that problem relates to billing in make-to-stock scenario. Since COPA document is automatically created with billing. Please note the following -
    1. COGS GL account is not created as cost element.
    2. There is no CO / COPA document at the time of delivery.
    3. Condition type ‘VPRS’ should exist in SD pricing. It is updated form material master.
    4. Condition type ‘VPRS’ should be assigned to COPA value field (just like PR00 / price).
    5. Revenue and COGS value fields are updated at the time of billing.
    Hope it clarifies.
    Regards,

  • PB2 issue Bing Maps says 'Unable to determine your current location'

    PB2 issue Bing Maps says 'Unable to determine your current location'.
    please help..

    That happens if you are indoors and the App canot locate your GPS location.
    You may have to take your PB outside for a minute or so so it can get a fix on your location.
    Have you tried Google Maps or the MapApp from App World that also uses Google Maps?
    They are much better than Bing Map.
    DC-IT

  • Multi Mapping issue - *Split mapping created no messages*

    Hi,
    The scenario I am trying to test is a multi mapping scenario where I am trying to split one source message and create two target messages by using two different inbound service interfaces, one for each message type.
    I am on PI 7.1 and when I test message mapping and operation mapping using the payload from SXMB_MONI, it is successful. Whereas when I test the scenario end to end I am getting the following error messages:
    Runtime error
    *Split mapping created no messages*
      Start tag ns0:MessagesStart tag ns0:Message1Close tag ns0:Message1Start tag ns0:Message2Close tag ns0:Message2Close tag ns0:Messages
    Could someone please help
    Cheers,
    S

    Hi,
    (1)First change the occurrence of target message to 0..unbounded.
    (2) Do the mapping and test it under test tab of mapping whether your mapping is working fine creating target records.
    (3) Copy the source code from mapping and remove the unwanted tags "as Mark said " those were created due to multimapping.
    Then test that file and that will work.
    You can follow this link .
    Link :Message Splitting
    Thanks
    Rabi
    Edited by: RabiPanda on Nov 17, 2011 8:42 AM

  • Issue while trying to create Sales Order from Inbound Idoc ORDER05

    Hi All,
    I am trying to create a sales order through an Inbound idoc ORDERS05 created out of a Purchase order. The header pricing conditions from the Purchase orders gets populated in E1EDK05 segment and this needs to go to the Sales order header conditions while trying to create the Sales order.
    But when the fields are getting incorrectly mapped to the Sales order. Issue  being, the segment E1EDK05-BETRG contains the condition value. But this fields gets populated to the KOMV-KWERT (condition value) screen field on the Sales order header conditions. But this should have got mapped to field KOMV-KBETR (Rate). If I look at the Purchase order, E1EDK05-BETRG is the field getting fetched from field KOMV-KBETR on the Purchase order. So this should have correspondingly got mapped to KOMV-KBETR on the Sales order as well.
    This looks like a standard behavior. But if any of you have encountered this issue before please let me know if any Correction instructions are available for the same. If this some issue in Pricing conditions configuration, do let me know.

    Hi Paaavan
    You can set a break-point include "FV45PFAP_VBAP_BEARBEITEN_VORBE" that is called by both the SAP Standard and BAPI where the data is prepared and checked in the internal table XVBAP. Compare both SAP standard VA01 and your 'Z' program for what may be different in XVBAP.
    I think the problem my be with regards to the business_object number that is hardcoded in the standard function module. It has given me issues before and hence I tend to use the function module  SD_SALESDOCUMENT_CREATE to create sales documents with.
    Regards,
    Marius

  • Convert Date format from YYYYMMDD to MM/DD/YYYY --issue with Mapping

    Hi All ,
    I have IDOC(DESADV.DELVRY03-> file ) scenario which i am working on . While debugging noticed that the field :
    E1EDT13\NATNZ -Data Transformation(from yyyymmdd to mm/dd/yyyy)-Target .This works
    But say if IDOC doesn't pass any value (For Good's issue) I get Error during debugging : RuntimeException in Message-Mapping transformation: Unparseable date: ""
    I did check thsi blog : Convertinf Date format from YYYY-MM-DD to MMDDYYYY  but this is  about the Data Transformtaion.
    I tried using constants , create if , -these function doesn't work /not applicle in this case
    Could you please provide  inputs as to how to handle two condition :
    1. With date -- then transform
    2.Else , pass nothing

    Hi Varalakshmi,
    Plese try to use UDF for date conversion and include logic which would check the Space value from the source structure.
    so if there is space return as such
    else.
    perform the conversion logic.
      Convertinf Date format from YYYY-MM-DD to MMDDYYYY
    (The conversion logic has been provided by micheal in the thread you have provided in the query)
    Thanks,
    Ram.

  • Issue in deploying webservice created from a JWS file.

    Hi,
    I tried deploying the helloworld sample webservice. Here are the files that i used.
    h1. HelloWorldImpl .java
    package com.agile.ws.helloworld.v1;
    import java.security.Principal;
    import javax.annotation.Resource;
    import javax.jws.WebMethod;
    import javax.jws.WebParam;
    import javax.jws.WebResult;
    import javax.jws.WebService;
    import javax.jws.soap.SOAPBinding;
    import javax.xml.ws.WebServiceContext;
    import weblogic.jws.Policy;
    @WebService(name="HelloWorld", serviceName="HelloWorldService", targetNamespace="http://xmlns.oracle.com/AgileServices/Core/HelloWorld/V1")
    @Policy(uri="policy:Wssp1.2-2007-Saml2.0-Bearer-Https.xml")
    @SOAPBinding(style=SOAPBinding.Style.DOCUMENT, use=SOAPBinding.Use.LITERAL, parameterStyle=SOAPBinding.ParameterStyle.WRAPPED)
    public class HelloWorldImpl {
    @Resource private WebServiceContext ctx;
    @WebResult(name="response", targetNamespace="http://xmlns.oracle.com/AgileServices/Core/HelloWorld/V1")
    @WebMethod()
    public HelloWorldVO sayHello(@WebParam(name="IntegerInput", targetNamespace="http://xmlns.oracle.com/AgileServices/Core/HelloWorld/V1") HelloWorldVO request) {
    Principal prp = ctx.getUserPrincipal();
    String username = prp.getName();
    String message = "User Name:" + username + "\tMessage:" + request.getMessage();
    HelloWorldVO response = new HelloWorldVO();
    response.setMessage(message);
    System.out.println(message);
    return response;
    h1. HelloWorldVO .java
    package com.agile.ws.helloworld.v1;
    public class HelloWorldVO {
    private String message;
    public void setMessage(String message) {
    this.message = message;
    public String getMessage() {
    return message;
    h1. build.xml
    <project default="generateFromWsdl">
         <taskdef name="wsdlc" classname="weblogic.wsee.tools.anttasks.WsdlcTask"/>
         <taskdef name="jwsc" classname="weblogic.wsee.tools.anttasks.JwscTask" />
         <target name="generateFromWsdl">
              <wsdlc srcWsdl="wsdl_files/HelloWorldService.wsdl" destJwsDir="output/compiledWsdl"
                        destImplDir="output/impl_output" packageName="com.agile.ws.hello.v1" type="JAXWS" />
         </target>
         <target name="buildService">
              <jwsc srcdir="src" destdir="output/wsdlcEar">
                   <jws file="com/agile/ws/helloworld/v1/HelloWorldImpl.java" type="JAXWS"/>
              </jwsc>
         </target>
    </project>
    h1. Issue
    I am able to run both the targets and the war gets generated. I copied the contents of application.xml (given below) to the existing application's application.xml file and also copied the war to the appropriate location.
    h3. application.xml contents
    <module id="WebModule_helloworld">
    <web>
    <web-uri>HelloWorldPort.war</web-uri>
    <context-root>core/services/saml</context-root>
    </web>
    </module>
    When i start the weblogic server i am getting the following exception
    <Dec 1, 2011 11:56:13 PM IST> <Error> <HTTP> <BEA-101216> <Servlet: "HelloWorldServiceServlethttp" failed to preload on startup in Web application: "core/services/saml".
    weblogic.wsee.server.ServerURLNotFoundException: Cannot resolve URL for protocol http/https
    at weblogic.wsee.server.ServerUtil.getHTTPServerURL(ServerUtil.java:204)
    at weblogic.wsee.jaxws.HttpServletAdapter.checkEndpointPolicy(HttpServletAdapter.java:129)
    at weblogic.wsee.jaxws.HttpServletAdapter.<init>(HttpServletAdapter.java:105)
    at weblogic.wsee.jaxws.HttpServletAdapter.<init>(HttpServletAdapter.java:80)
    at weblogic.wsee.jaxws.JAXWSServlet.registerEndpoint(JAXWSServlet.java:139)
    Truncated. see log file for complete stacktrace
    <Dec 1, 2011 11:56:13 PM IST> <Error> <Deployer> <BEA-149231> <Unable to set the activation state to true for the application 'AgilePLM'.
    weblogic.application.ModuleException: [HTTP:101216]Servlet: "HelloWorldServiceServlethttp" failed to preload on startup in Web application: "core/services/saml".
    weblogic.wsee.server.ServerURLNotFoundException: Cannot resolve URL for protocol http/https
    at weblogic.wsee.server.ServerUtil.getHTTPServerURL(ServerUtil.java:204)
    at weblogic.wsee.jaxws.HttpServletAdapter.checkEndpointPolicy(HttpServletAdapter.java:129)
    at weblogic.wsee.jaxws.HttpServletAdapter.<init>(HttpServletAdapter.java:105)
    at weblogic.wsee.jaxws.HttpServletAdapter.<init>(HttpServletAdapter.java:80)
    at weblogic.wsee.jaxws.JAXWSServlet.registerEndpoint(JAXWSServlet.java:139)
    at weblogic.wsee.jaxws.JAXWSServlet.init(JAXWSServlet.java:64)
    at weblogic.wsee.jaxws.JAXWSDeployedServlet.init(JAXWSDeployedServlet.java:54)
    at javax.servlet.GenericServlet.init(GenericServlet.java:241)
    at weblogic.servlet.internal.StubSecurityHelper$ServletInitAction.run(StubSecurityHelper.java:283)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
    at weblogic.servlet.internal.StubSecurityHelper.createServlet(StubSecurityHelper.java:64)
    at weblogic.servlet.internal.StubLifecycleHelper.createOneInstance(StubLifecycleHelper.java:58)
    at weblogic.servlet.internal.StubLifecycleHelper.<init>(StubLifecycleHelper.java:48)
    at weblogic.servlet.internal.ServletStubImpl.prepareServlet(ServletStubImpl.java:539)
    at weblogic.servlet.internal.WebAppServletContext.preloadServlet(WebAppServletContext.java:1985)
    at weblogic.servlet.internal.WebAppServletContext.loadServletsOnStartup(WebAppServletContext.java:1959)
    at weblogic.servlet.internal.WebAppServletContext.preloadResources(WebAppServletContext.java:1878)
    at weblogic.servlet.internal.WebAppServletContext.start(WebAppServletContext.java:3153)
    at weblogic.servlet.internal.WebAppModule.startContexts(WebAppModule.java:1508)
    at weblogic.servlet.internal.WebAppModule.start(WebAppModule.java:482)
    at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:425)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
    at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:119)
    at weblogic.application.internal.flow.ScopedModuleDriver.start(ScopedModuleDriver.java:200)
    at weblogic.application.internal.flow.ModuleListenerInvoker.start(ModuleListenerInvoker.java:247)
    at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:425)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
    at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:119)
    at weblogic.application.internal.flow.StartModulesFlow.activate(StartModulesFlow.java:27)
    at weblogic.application.internal.BaseDeployment$2.next(BaseDeployment.java:636)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
    at weblogic.application.internal.BaseDeployment.activate(BaseDeployment.java:205)
    at weblogic.application.internal.EarDeployment.activate(EarDeployment.java:58)
    at weblogic.application.internal.DeploymentStateChecker.activate(DeploymentStateChecker.java:161)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.activate(AppContainerInvoker.java:79)
    at weblogic.deploy.internal.targetserver.BasicDeployment.activate(BasicDeployment.java:184)
    at weblogic.deploy.internal.targetserver.BasicDeployment.activateFromServerLifecycle(BasicDeployment.java:361)
    at weblogic.management.deploy.internal.DeploymentAdapter$1.doActivate(DeploymentAdapter.java:51)
    at weblogic.management.deploy.internal.DeploymentAdapter.activate(DeploymentAdapter.java:200)
    at weblogic.management.deploy.internal.AppTransition$2.transitionApp(AppTransition.java:30)
    at weblogic.management.deploy.internal.ConfiguredDeployments.transitionApps(ConfiguredDeployments.java:240)
    at weblogic.management.deploy.internal.ConfiguredDeployments.activate(ConfiguredDeployments.java:169)
    at weblogic.management.deploy.internal.ConfiguredDeployments.deploy(ConfiguredDeployments.java:123)
    at weblogic.management.deploy.internal.DeploymentServerService.resume(DeploymentServerService.java:180)
    at weblogic.management.deploy.internal.DeploymentServerService.start(DeploymentServerService.java:96)
    at weblogic.t3.srvr.SubsystemRequest.run(SubsystemRequest.java:64)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:178)
    at weblogic.servlet.internal.WebAppModule.startContexts(WebAppModule.java:1510)
    at weblogic.servlet.internal.WebAppModule.start(WebAppModule.java:482)
    at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:425)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
    at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:119)
    Truncated. see log file for complete stacktrace
    Caused By: weblogic.wsee.server.ServerURLNotFoundException: Cannot resolve URL for protocol http/https
    at weblogic.wsee.server.ServerUtil.getHTTPServerURL(ServerUtil.java:204)
    at weblogic.wsee.jaxws.HttpServletAdapter.checkEndpointPolicy(HttpServletAdapter.java:129)
    at weblogic.wsee.jaxws.HttpServletAdapter.<init>(HttpServletAdapter.java:105)
    at weblogic.wsee.jaxws.HttpServletAdapter.<init>(HttpServletAdapter.java:80)
    at weblogic.wsee.jaxws.JAXWSServlet.registerEndpoint(JAXWSServlet.java:139)
    I am trying with Weblogic 10.3.5 version. Any pointers to solve the issue would be of great help.
    Thanks,
    Arul

    Hi,
    I believe that the problem you are having is that you have a delimited file, of which you want to pick columns from position 2,3,5. In this case, ODI will pick the first 3 columns of a delimited file regardless of position.
    For example, if you a tab delimited file with c1,c2,c3,c4,c5 columns, and you want only columns c2,c3,c5 - when mapping these in an ODI interface, and executing, you will actually pick up the data from c1,c2,c3 as these are the first three columns in the file (reading from left to right). You can ignore "columns" on the right hand side of a file, but not the left. E.g delimited file with c1,c2,c3,c4,c5. Only pick columns c1,c2 will give you data for the first 2 columns
    Create a temporary table to load all the data from the file, and use you temp table to extract the data you require. Or you could get the file created with the first three columns as the columns you require.
    Cheers
    Bos
    Edited by: Bos on Jan 18, 2011 1:06 PM

  • 3 issues with mapped-superclass

    I created a superclass for my entities which is a simple pojo with 4 properties. I mapped it in JPA with mapped-superclass.
    I want the 'ownerId' and 'created' fields to be read-only on the API, in other words, the pojo shouldn't have setters for those 2 fields.
    TLE objects violently to the absence of those setters. I throws a NPE (see below).
    Secondly, I decided I could live without the read-only requirement, so I put the setters on the pojo, set the mapping's access= to access="FIELD" so that the setters are not called. Out of curiosity, I threw exceptions in the setters just to make sure.
    TLE insists on calling the setters, regardless of the 'access' mode I choose. It's always PROPERTY access which is used. FIELD is ignored.
    So I can live with that too. However once I managed to deploy the whole JPA and started testing it, I realised the prePersist event listener isn't called as it is should be according to the mapping I defined. The prePersist callback is in the mapped-superclass entity mapping and calls a method on the superclass. However it isn't called.
    Does anybody here use mapped-superclass without trouble, or is there an example app out on the net demonstrating its use without issues?
    This is TLE build 58g-fcs, btw.
    Regards
    Adam
    java.lang.NullPointerException
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.objects.MetadataMethod.getSetMethodName(MetadataMethod.java:69)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.MetadataAccessor.getSetMethodName(MetadataAccessor.java:330)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.MetadataAccessor.setAccessorMethods(MetadataAccessor.java:575)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.BasicAccessor.processDirectToFieldMapping(BasicAccessor.java:182)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.BasicAccessor.process(BasicAccessor.java:160)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.ClassAccessor.processAccessor(ClassAccessor.java:528)
         at oracle.toplink.essentials.internal.ejb.cmp3.xml.accessors.XMLClassAccessor.processAccessors(XMLClassAccessor.java:371)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.ClassAccessor.processMappedSuperclass(ClassAccessor.java:1128)
         at oracle.toplink.essentials.internal.ejb.cmp3.xml.accessors.XMLMappedSuperclassAccessor.process(XMLMappedSuperclassAccessor.java:67)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.ClassAccessor.processMappedSuperclasses(ClassAccessor.java:1138)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.accessors.ClassAccessor.process(ClassAccessor.java:495)
         at oracle.toplink.essentials.internal.ejb.cmp3.xml.accessors.XMLClassAccessor.process(XMLClassAccessor.java:322)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.MetadataProcessor.processMappingFile(MetadataProcessor.java:568)
         at oracle.toplink.essentials.internal.ejb.cmp3.metadata.MetadataProcessor.processMappingFiles(MetadataProcessor.java:533)
         at oracle.toplink.essentials.ejb.cmp3.persistence.PersistenceUnitProcessor.processORMetadata(PersistenceUnitProcessor.java:368)
         at oracle.toplink.essentials.internal.ejb.cmp3.EntityManagerSetupImpl.predeploy(EntityManagerSetupImpl.java:607)
         at oracle.toplink.essentials.internal.ejb.cmp3.JavaSECMPInitializer.callPredeploy(JavaSECMPInitializer.java:171)
         at oracle.toplink.essentials.internal.ejb.cmp3.JavaSECMPInitializer.initPersistenceUnits(JavaSECMPInitializer.java:239)
         at oracle.toplink.essentials.internal.ejb.cmp3.JavaSECMPInitializer.initialize(JavaSECMPInitializer.java:255)
         at oracle.toplink.essentials.ejb.cmp3.EntityManagerFactoryProvider.createEntityManagerFactory(EntityManagerFactoryProvider.java:155)
         at javax.persistence.Persistence.createFactory(Persistence.java:172)

    https://glassfish.dev.java.net/issues/show_bug.cgi?id=4414

  • ***FONT ISSUES: How do I create a font repository/server??? ***

    Before I explain my font problem, let me first say that I hope that SOMEONE out there/here has good knowledge of fonts, (storage/management/etc.) and can take the time to help me out with my problem. I'd be ETERNALLY grateful for a solution to this problem!! Let me also explain the workflow setup I am coming from. I work at a production company that has 3 edit suites, accessing media from a SAN, managed by FibreJet software. We are running Apple Mac Pro's with 2 x 3 GHz Dual-Core Intel Xeon chips, and 8 GB 667 MHz DDR2 FB-DIMM memory. The 3 stations can have READ access to any one drive on the SAN at the same time with no problem, but only one station at a time can have WRITE access to a drive. We each have our own fonts (~2500) on our local systems, and we are all using Extensis Suitcase Fusion 12.1.6 to manage our fonts.
    Over the past few months, a couple of projects have reminded us that we have to address a problem in the way we manage our fonts. A project was started in Edit 1 (E1). Footage was loaded, music added and graphics created all in E1. The graphics were made in the Title 3D generator. Near the end of the project, a change had to be made and the original editor and edit suite (E1) was unavailable, so another editor needed to make the change in E2. This is a workflow that we use all the time, and usually don't have a problem. However, this time, when the graphic change was made, either the font that was originally used was missing from E2, or it was not activated. In either case, FCP simply substituted another font in it's place. Unlike Photoshop, FCP does not give any warning when this happens, it just changes the font, so unless you are familiar with the project, you may not know that the change occurred.
    As I said, these problems have reminded us that we have to address a problem in the way we manage our fonts. One solution I have thought of would be to consolidate all our fonts onto one of our SAN media drives, and always read that drive. We would then add fonts from this drive, instead of adding from our individual local drives, and hopefully avoiding the "font-switch" problem in the future.
    Questions arise, such as: Will Extensis read a font from a SAN drive, or does it have to be local? Will there be a problem with the fact that the drive will be READ only, as opposed to having WRITE access, since all three of us cannot write to the same drive at the same time? Should we just use Apple's own FontBook, instead of Extensis Suitcase, or do we even look into something like Linotype's FontExplorer X, (which I stumbled upon during searches for solutions to our problem)?
    If anyone has any answers to my questions, or any solutions we haven't thought of, I'd be very grateful for your help!

    Will Extensis read a font from a SAN drive, or does it have to be local?
    Here's a bit of information on a SAN setup.
    Storage Area Network (SAN) is a high-speed subnetwork of shared storage devices. A storage device is a machine that contains nothing but a disk or disks for storing data. A SAN's architecture works in a way that makes all storage devices available to all servers on a LAN or WAN. As more storage devices are added to a SAN, they too will be accessible from any server in the larger network. In this case, the server merely acts as a pathway between the end user and the stored data.
    Because stored data does not reside directly on any of a network's servers, server power is utilized for business applications, and network capacity is released to the end user.
    So as I understand it, your Mac's will all see data on a SAN drive array the same as it would a drive that is part of any LAN or WAN. In other words, if activating fonts from a drive on a typical server works, then doing so from a SAN should work the same.
    That said, I've never had the opportunity to try it. The easiest way to find out would be to simply drag some fonts into Suitcase that you know are on the SAN storage and see if the activate as you would expect. An issue that may come up is that multiple users may not be able to activate the same fonts at the same time. That wouldn't specifically be a write issue since Suitcase doesn't need to perform a write when opening fonts (other than to update the font cache files on the local drive), but rather if the server considers the files "locked" (in use) to the first user to access them. All depends on how the server is set up, or how Suitcase handles fonts it currently has active. That is, will the server lock the files, or Suitcase, if either?
    One solution I have thought of would be to consolidate all our fonts onto one of our SAN media drives, and always read that drive. We would then add fonts from this drive, instead of adding from our individual local drives, and hopefully avoiding the "font-switch" problem in the future.
    That would be my first suggestion whether you can read simultaneously read the fonts from the SAN drive or not. I imagine each project is kept in its own folder. Within that, all fonts should be kept together in a single subfolder. An obvious name for that folder is good, such as "Fonts_jobnumber". If Suitcase and/or the server will allow all of you to access the same fonts at the same time, just activate them from the SAN. If not, copy the fonts to your local drive first and then activate.
    Should we just use Apple's own FontBook, instead of Extensis Suitcase, or do we even look into something like Linotype's FontExplorer X, (which I stumbled upon during searches for solutions to our problem)?
    Font Book is a very weak font manager. It's only advantage in a case like this is that every font you open gets copied to the local drive into one of the two Fonts folder you can designate in its preferences. Then you're always opening the fonts locally. The bad part is that when you use hundreds or thousands of fonts, they all get copied to your Mac as you add fonts. This increases startup times, sometimes dramatically as Font Book has to check all of those fonts during startup against its database to see which should be active and which not. The larger that database gets, the more likely it is to get corrupted.
    Suitcase is a far more elegant solution for heavy font usage since you can activate fonts in place. They don't need to be copied anywhere to use them. FontExplorer X can be set up to do the same thing. Depending on what features of Suitcase you use, you may still prefer it over FontExplorer X, even though the latter is free.

  • Strange Database issues while trying to create new user through EM...

    I'm attempting to configure my local Oracle 11g database in order to create a Content Server user in order to setup a Fatwire installation. I'm getting some strange issues with my database that I was hoping maybe someone could help me out with.
    The guide I'm following starts with this:
    +1. Determine the Console Server port:+
    a. Open the emoms.properties file in a text editor. The file is located in:
    +<ora_home>/<servername>_<SID>/sysman/config/+
    b. Find the line,
    oracle.sysman.emSDK.svlt.ConsoleServerPort
    and make a record of the port number value at the end of the line.
    When I open the emoms.properties file in Notepad++, I'm unable to read the contents - it's all symbols/jibberish.
    The next step says this:
    +2. Log in to the Oracle Enterprise Manager console:+
    a. Execute the following command: emctl status dbconsole
    When I execute that command I get the following error:
    Environment variable ORACLE_UNQNAME not defined. Please set ORACLE_UNQNAME to database unique name.
    The third step says this:
    Open a browser and go to the URL highlighted in bold in step a above. If you see
    a “Security Mismatch” error, ignore it (the error appears if you are using a selfsigned
    certificate).
    I skipped to this step because I'm assuming the link it wants me to go to is my EM console (https://localhost:1158/em/console/aboutApplication), however when I browse to that URL I get the following error:
    503 Service Unavailable
    Service is not initialized correctly. Verify that the repository connection information provided is correct.
    Clearly, something isn't right here. What's going on? Any help would be appreciated.

    Alfonso Vicente wrote:
    Did you try what I suggested?Yes, thanks for the advice. Now I'm having some additional issues. When I run emctl status dbconsole I get the following error:
    OC4J Configuration issue. C:\app\john.smith\product\11.2.0\dbhome_1/oc4j/j2ee/OC4
    J_DBConsole_ORCL-SMITH.company.com_localhost_orcl not found.
    This is appearing after I made the appropriate changes to the emoms.properties file and set ORCL_UNQNAME to "localhost_orcl" <-- is that correct? the path to my emoms.properties is:
    C:\app\john.smith\product\11.2.0\dbhome_1\localhost_orcl\sysman\config
    Here is my emoms.properties file, does this look right?
    #Mon Mar 19 17:56:05 EDT 2012
    oracle.sysman.emSDK.svlt.ConsoleServerName=localhost\:1158_Management_Service
    oracle.sysman.eml.mntr.emdRepPwd=2ad171ddf6ce91aebf1c2f2c371ee5a315591a1be21f37965a88d54126d7528aa1091f543569e58b
    emdrep.ping.pingCommand=/bin/ping -c 3 -w 30 <hostname>
    em_oob_shutdown=false
    LargeRepository=false
    oracle.sysman.eml.mntr.emdRepPort=1521
    em_email_address=
    oracle.sysman.eml.mntr.emdRepPwdSeed=8622625238442107993
    em_oob_crash=false
    em.oms.dumpModules=omsThread,repos
    oracle.sysman.emRep.dbConn.statementCacheSize=50
    oracle.sysman.db.isqlplusUrl=%ISQLPLUS_PROTOCOL%\://%ISQLPLUS_HOST%\:%ISQLPLUS_PORT%/isqlplus/dynamic
    em_oob_startup=false
    oracle.sysman.emSDK.svlt.ConsoleServerPort=4889
    em_from_email_address=_NOT_AVAILABLE_
    oracle.sysman.eml.mntr.emdRepPwdEncrypted=TRUE
    oracle.sysman.db.isqlplusWebDBAUrl=%ISQLPLUS_PROTOCOL%\://%ISQLPLUS_HOST%\:%ISQLPLUS_PORT%/isqlplus/dba/dynamic
    oracle.sysman.emSDK.svlt.ConsoleServerHost=localhost
    em_from_email_name=_NOT_AVAILABLE_
    oracle.sysman.emSDK.svlt.ConsoleServerHTTPSPort=1159
    em_email_gateway=
    oracle.sysman.eml.mntr.emdRepServer=localhost
    oracle.sysman.eml.mntr.emdRepSID=NO_VALUE
    oracle.sysman.eml.mntr.emdRepConnectDescriptor=(DESCRIPTION\=(ADDRESS_LIST\=(ADDRESS\=(PROTOCOL\=TCP)(HOST\=localhost)(PORT\=1521)))(CONNECT_DATA\=(SERVICE_NAME\=orcl)))
    oracle.sysman.emkeyfile=C:\app\john.smith\product\11.2.0\dbhome_1\localhost_orcl\sysman\config\emkey.ora
    em.ip.ui.enable=true
    oracle.sysman.eml.mntr.emdRepUser=sysman
    oracle.sysman.emSDK.svlt.PublicServletEnabled=true

  • Mapping tool creates new columns one at a time

    Hi All,
    I'm using Kodo 3.2.2 with MySql 4.1. I have a fairly large table (over
    2million rows) representing one persistent class. I need to add several
    new fields to the object, which will require a refresh mapping on the
    database to add new columns for the fields.
    It seems when I run refresh mapping on the database, Kodo is adding each
    new field individually in seperate ALTER TABLE statements, rather than
    adding all the columns in one ALTER TABLE statement.
    With MySQL and large tables, this can really increase the time required to
    refresh the mapping - I've tried with MSSQL and that database must handle
    ALTER TABLE commands differently because it doesn't take a fraction of the
    time that MySQL takes.
    Is this a bug (I doubt it), but rather a feature improvement?
    Thoughts?
    Thanks,
    Brendan

    Thanks Abe for the reply - I've been meaning to look at the process of
    creating SQL scripts to perform the refresh mappings on our databases.
    I'll have a look at this process before we talk contracting!
    Thanks,
    Brendan
    Abe White wrote:
    Unfortunately, we don't have any immediate plans to optimize ALTER TABLEsince
    it's not a runtime operation, and as you say most DBs don't have a problemwith
    it. Also, note that the documentation shows how to get Kodo to create SQL
    scripts rather than directly modify the DB; you could easily edit thescripts
    Kodo generates to make them more efficient before piping the SQL to MySQLusing
    its client app. (Note that Kodo 3.3 introduces the ability to generate SQL
    scripts directly from the mapping tool, rather than having to take an extrastep
    through the schema tool).
    If getting Kodo to combine ALTER statements is important to you, you couldalso
    contact [email protected] about contracting us to expedite this work.

  • Issue in mapping compounding attribute

    Hello All,
    I have info objects 0CO_AREA and 0PROFIT_CTR in the target whereas in the source i have only 0PROFIT_CTR and please note that CO_AREA is compounded with 0PROFIT_CTR.
    In the transformation, I have kept constant value for 0C0_AREA and info object 0PROFIT_CTR is mapped with the source field.
    I facing an issue with the below error message while activating the transformations.
    "Characteristic 0CO_AREA compounded to 0PROFIT_CTR is missing in transformation source"
    Is it mandatory to have 0CO_AREA in the source when we are using compounding info object relation or is there any work around ?
    Thanks in advance.

    Hi peppy ,
    The Compund characteristic is complementry to the parent characteristic so whenever you use the characteristic you have to use the compound characteristic. it is mandatory to have 0CO_AREA in the source when we are using compounding info object relation
    To upload the master data you also need to have compound characteristic in the source you're uploading because it is part of the compound.It is necessary to get correct data in target.This concept is basically used for maitaining master data .
    Regards,
    Jaya

  • Issue with G_FXX  while creating a Process

    Hello Friends,
    I have a Manual Tabular form made with APEX_ITEM, I am trying to create a process to perform DML operations based on the values from the G_FXX arrays of APEX_APPLICATION.
    Below is the error I am getting
    ORA-06550: line 9, column 30: PLS-00302: component 'G_F51' must be declared ORA-06550: line 9, column 1: PL/SQL: Statement ignoredCorresponding Line of the Insert Process:
    for i in 1..apex_application.g_f51.count
    loop
        if ( apex_application.g_f51(i) is not null) then
            update CP_INTERFACE_SCH_INFO
         set INTF_NAME = apex_application.g_f52(i),
           Corresponding Region code :
    select apex_item.hidden(50,CAP_PLAN_ID) ||
    apex_item.hidden(51,LINE_NO) ||
    apex_item.text(52,INTF_NAME) INTF_NAME,
    apex_item.select_list_from_query(53,TYPE,'SELECT TYPE,TYPE_ID FROM
    and so onThe form is generated properly and I could check the labels showing up correctly.
    I have multiple other regions and processes on the same page with similar syntax working with out issues. I have spent quite some time, but could not think of any problem...
    Please help me ...

    Hi Dir,
    Thanks for responding.
    In my case :- I have multiple regions, all are tabular forms made with APEX_ITEM each having 7 ~ 8 cols. So, I have run out of arrays. I was updating the collections on specific region SAVE buttons.[I created collections On Load-Before Header with APEX_COLLECTION.CREATE_TRUNCATE_COLLECTION]. I wanted to load actual tables from collections upon hitting the final SUBMIT button.
    Please let me know if there is a better practice to handle this situation.
    Thanks..

Maybe you are looking for

  • Windows disc does not update after resizing boot camp partition

    When I bought my iMac, it came with a 500 GB HDD. I used bootcamp to make a 200GB windows partition and a 300GB Mac partition. The other day I bought a 250GB SSD, and moved everything from my 300GB Mac partition onto the SSD. My plan was to use the w

  • Some images not displaying in some browsers

    Hi, I'm creating a newsletter in Dreamweaver using the design view. The finished thing works perfectly in Internet Explorer but when I view it in Firefox or Safari certain images are not displaying. There's no red cross or sign at all of why they're

  • Problem in converting projects

    Hi, Earlier, whenver I converted .hpj to .xpj - the BMP files used in the .hpj projects used to get converted to .gif/.jpeg automatically based on the selected option. Recently, I got my hard disk changed. So, all the installations are new. Today, I

  • Custom transaction codes?

    Hi All, Is there any possibility in CRM to create our own transaction code? I want to create my own transaction code to create BP - Person - new role -> contracted employee as default. Thanks and Best Regards, Abdul

  • Hierarchy  to  Flat  Structure

    Hi all.. I  need to map a Hirarchial Struture to a Flat Structure ...in  my souce node HDR can appear multiple times and below that one sub node STS also can appear multiple times ... i need a flat targer stucture of source of same kind .. what is th