Data mapping between sybase and xml

I want to make a data mapping between sybase relational data result set and xml.
I am using the function ForXmlTree for this purpose, I have the entire documentation about the synatx and usage of it, but i need to know what support does JAVA have for it.
What files need to be installed .
This is an example of using the function
java jcs.xmlutil.ForXmlTree �i forxmltree-spec [-o output-script-file] \
[-x output-document-file  -S server-name]
does anyone know where can i find the jcs.xmlutil package. If so let me know.
Thanks in advance
Sandeep

PRPS-PSPNR = AFVU-PROJN.
also you can use this fm.
BAPI_PROJECT_GETINFO

Similar Messages

  • Reference Library for Converting Between LabVIEW and XML Data (GXML)

    Please provide feedback, comments and questions on the Reference Library for Converting Between LabVIEW and XML Data (GXML) in this thread.
    The latest version of the NI GXML Library is availble in VIPM on the NI LabVIEW  Tools Network repository.

    Francesco, Thank you for the feedback.  With this component it was my intention to make a more "terse" version of the LabVIEW Flatten to XML VI that was also supported on RT and that gave the user more flexbility regarding the structure of the parsing type definition. I think you are right that the XML parser is not compliant to section 2.11 of the XML spec.  The parser does specifically looking for a #D#A and this appears to be an oversight on my part.  Please confirm for me, the specifcation is saying that the XML parser should be able to recognize three possibilities as an "end of line" character: #D#A, #D, or #A.  Am I reading this right?There are more efficient (and in some cases much more efficient) ways of sharing data between LabVIEW and LabVIEW: some examples are flattened binary strings and the datalog binary format.  XML is slower than these optons but the upside is that it is human readable.  Furthermore XML is inherently hierarchical which is convenient for complex data structures like clusters of arrays of clusters, etc.  If you don't care about human readability then you are correct XML doesn't make as much sense.I will return to the GXML source code and try to fix this in the near future but I would hope that instead of creating yet another custom VI from scratch that you could reuse what I have provided for you.  I included enough documentation in the source code so that users could make some modificiations themselves. The target application for this reference library was LabVIEW to LabVIEW communication.  As such I documented the schema on the dev zone document from a LabVIEW perspective.  It includes all the supported datatypes and all the supported data structures (cluters, arrays, multidimensional arrays, clusters of multidimensional arrays, etc.)  I do see some value in making a more conventional XML spec but the time investment required didn't really line up with my intended use case. Were there any other downsides to GXML that I have missed?  Best Regards, Jeff TippsSystems Engineer - Sound and VibrationMessage Edited by Jeff T. on 04-21-2010 10:09 AM

  • Users mapping between EP and ABAP system

    Hello
    I'd like to ask for some guidance in my quest
    Current situation looks like this:
    I've configured UME in AS Java to work with LDAP as read only data source. Then I've configured SPNego to run SSO - It works, users from MS AD can log into portal.
    Now I have application in WD which authorizes via EP/AD - works fine.
    And next step is users mapping between AD and ABAP backend (serving some BAPI's for WD app)
    I've found a bunch of help pages starting from
    http://help.sap.com/saphelp_nwce711/helpdata/en/0b/d82c4142aef623e10000000a155106/frameset.htm
    But somehow it's quite complicated to achieve this mapping. I've tried to set RFC destinations logon type to user mapping but without succes.
    Can anyone point me to some more clear example or give path to configure this scenario? Is there a way of configuring this with NWA or some XML file editing is required?
    Any help will be appreciated.
    BTW: whole environment is in version 7.11
    Best regards
    Maciej

    There is no equivalent to SPNEGO on the ABAP side.
    If your goal is to propagate the user, then possible options are:
    -> Wait for SAML 2.0 or invest now in a SAML 1.0 provider.
    -> Use the same kerberos ticket for the EP as what your ABAP system will accept: route = SNC and 3rd party libraries.
    -> Issue SAP logon tickets for the ABAP system from the EP, and use these in your WDA.
    Another option is to expose the service with saved logon data in the ICF. If the service is just a wrapper for the BAPI, then you can also consider using trusted RFC between the service and the backend, but this might not be acceptable for your service.
    I have only done experimental stuff with this and some of the above is not released yet. Also consider the consequences, even if it "does work"...
    Cheers,
    Julius

  • Role mapping between Portal and Back end systems

    I am new to SAP EP.
    I just want to know how the mapping between portal and back end system happens.
    Scenario : There is a role in ECC system...say FI India. Now there is a request by the FI team that they want to access this role from Portal. In this case, please tell me how the security team will do it. Because I guess, it has to be done by the security team.

    Hi,
    Usually the role from backend is uploaded to portal then it will be seen as Group and we need to assign our portal roles to this group. Please refer [this|http://help.sap.com/saphelp_nw73/helpdata/en/d6/7859ec80df46738e23ccb4f4c8c502/content.htm].
    Regards,
    Samir

  • Explicity mapping between ActionScript and Java objects for the BlazeDS Messaging Service

    The BlazeDS documentation shows how to explicitly map between ActionScript and Java objects. For example, this works fine for RPC services, e.g.
    import flash.utils.IExternalizable;
    import flash.utils.IDataInput;
    import flash.utils.IDataOutput;
    [Bindable]
    [RemoteClass(alias="javaclass.User")]
    public class User implements IExternalizable {
            public var id : String;
            public var secret : String;
            public function User() {
            public function readExternal(input : IDataInput) : void {
                    id = input.readObject() as String;
            public function writeExternal(output : IDataOutput) : void {
                    output.writeObject(id);
    and
    import java.io.Externalizable;
    import java.io.IOException;
    import java.io.ObjectInput;
    import java.io.ObjectOutput;
    import java.util.HashSet;
    import java.util.Iterator;
    import java.util.Set;
    public class User implements Externalizable {
        protected String id;
        protected String secret;
        public String getId() {
            return id;
        public void setId(String id) {
            this.id = id;
        public String getSecret() {
            return secret;
        public void setSecret(String secret) {
            this.secret = secret;
        public void readExternal(ObjectInput in) throws IOException,
                    ClassNotFoundException {
            id = (String) in.readObject();
        public void writeExternal(ObjectOutput out) throws IOException {
            out.writeObject(id);
    If I called an RPC service that returns a User, the secret is not sent over the wire.  Is it also possible to do this for the messaging service? That is, if I create a custom messaging adapter and use the function below, can I also prevent secret from being sent?
    MessageBroker messageBroker = MessageBroker.getMessageBroker(null);
    AsyncMessage message = new AsyncMessage();
    message.setDestination("MyMessagingService");
    message.setClientId(UUIDUtils.createUUID());
    message.setMessageId(UUIDUtils.createUUID());
    User user = new User();
    user.setId("id");
    user.setSecret("secret");
    message.setBody(user);
    messageBroker.routeMessageToService(message, null);

    Hi Martin. The way that AMF serialization/deserialization works for BlazeDS is the same regardless of which service is being used, so yes that code will work for messaging as well. On the server, the serialization/deserialization of messages happens at the endpoint. For an incoming message for example, the endpoint deserializes the message and then hands it off to the MessageBroker which decides which service/destination to deliver the message to.
    That was a good question. Thanks for asking it. Lots of people are used to doing custom serialization/deserialization with the RPC services (RemoteObject/RemotingService) but I'm not sure everyone realizes they can do this for messaging as well.
    -Alex

  • Difference in idoc adapter header mapping between SP14 and SP19

    Hi All,
    i have a very strange problem.......i have 2 XI systems : XI-sandbox and XI-dev.....
    XI-sandbox is on XI3.0 SP19 whereas XI-dev is on XI3.0 SP14...........
    I have a file->XI->idoc scenario which is same on both XI-sandbox and XI-dev....
    in receiver idoc adapter, i am using header mapping for sender service and in the sender service i am using XPATH of a field of control rec of idoc.........
    this scenario works fine on XI-sandbox which is on SP19 and the idocs are sent to SAP R/3 sandbox...........but this scenario on XI-dev gives error in SXMB_MONI in XI-dev of Call adapter - "No party conversion found".........
    So my question is that is there any difference in idoc adapter header mapping between SP14 and SP19.........
    Thanks,
    Rajeev Gupta
    Message was edited by:
            RAJEEV GUPTA

    ><i>so i cant use apply control rec values from payload.........</i>
    rajeev,
    I know this has nothing to do with the original problem but please do bear.
    If I undertood you posts correct, the Sender Partner Name is set in the Idoc Control record in the field SNDPRN in the mapping using some mapping logic.
    Now, this is the partner name with which the idoc is to be posted to thge R3 system.
    if yes, then like I have pointed above, using "Take Sender From Payload" in the Receivcer Idoc Adapter is a better option as pointed in the SAP help as well.
    As you are already populating the idoc control ecord with SNDPRN just make the SNDPRT as LS and then select "Take Sender from Payload" in the receiver Idoc adapter and the Sender Partner Name will not be taken from SDNPRN just as you want.
    This is different from "Apply Control Records From Payload" .
    Regards
    Bhavesh
    PS : you can ignore this reply if the situatios is too late to make this change and thanks for the patience.

  • Data Replication Between Sqlserver and Oracle11g using materialized view.

    I have Sqlserver 2005 as my source and oracle11g as my target.I need to populate the target daily with change data from source.
    for that we have created a dblink between sqlserver and oracle and replicated that table as a Materialized view in Oracle.
    problem we are getting here is Fast refresh option is not available.each day it will pick full data from the source.
    is there any way to use Fast refresh in this scenario??
    Thanks in advance.
    Regards,
    Balaram.

    Pl do not post duplicates - Data Replication Between Sqlserver and Oracle11g using materialized view.

  • Data streaming between server and client does not complete

    Using an ad-hoc app, data streaming between server
    and client does not complete as it supposed to be.
    The process runs well in solaris 5.8, however under 5.9
    we have found the characters stream buffer length limitation
    is around 900 to 950 characters (by default we are using 3072
    characters).
    Example:
    - We are transfering HTML file, which will be displayed
    in the App client, with buffer=3072, the HTML only displayed / transfered
    as xxxxxxxx characters, but with buffer=900 the HTML is displayed properly,
    in this case, the only problem that we have is the file transfer will
    eventually longer than usual.
    - There is another case, where we have to transfer information (data) as stream
    to the client. A long data stream will not appear at all in the client.
    Any ideas why the change between 5.8 and 5.9 woudl cause problems?
    The current app-driver that we are using is compiled using Solaris 5.6,
    if possible we would like to have use of a later version, which is compiled using Solaris 5.9, do you think this will probably solve our problem?
    Thanks
    Paul

    Does this have anything to do with Java RMI? or with Java come to think of it?

  • Secure the file/data transfer between XI and any third-party system

    Hi All,,
    I would like to use to "secure" SSH on OS Level the file/data transfer between XI and any third-party system Run OS Command before processing and OS command After processing. right now my XI server installed on iSeries OS.
    with ISeries we can't call the Unix commands hope we need to go for AS400 (CL) Programming. If we created the AS400 programm how i can call that in XI.
    If any one have idea pls let me know weather it will work or not.
    Thanks in adavance.
    Venkat

    Hi,
    Thanks for your reply.
    I have red some blogs like /people/krishna.moorthyp/blog/2007/07/31/sftp-vs-ftps-in-sap-pi to call the Unix Shell script in XI.
    But as i know in iSeries OS we can write the shell script we need to go for AS400 programe. If we go with AS400 how we need to call that programe and it will work or not i am not sure there i need some help please.
    Thanks,
    Venkat

  • Data mismatch between 10g and 11g.

    Hi
    We recently upgraded OBIEE to 11.1.1.6.0 from 10.1.3.4.0. While testing we found data mismatch between 10g and 11g in-case of few reports which are including a front end calculated column with division included in it, say for example ("- Paycheck"."Earnings" / COUNT(DISTINCT "- Pay"."Check Date")) / 25.
    The data is matching for the below scenarios.
    1) When the column is removed from both 10g and 11g.
    2) When the aggregation rule is set to either "Sum or Count" in both 10g and 11g.
    It would be very helpful and greatly appreciated if any workaround/pointers to solve this issue is provided.
    Thanks

    jfedynic wrote:
    The 10g and 11.1.0.7 Databases are currently set to AL32UTF8.
    In each database there is a VARCHAR2 field used to store data, but not specifically AL32UTF8 data but encrypted data.
    Using the 10g Client to connect to either the 10g database or 11g database it works fine.
    Using the 11.1.0.7 Client to go against either the 10g or 11g database and it produces the error: ORA-29275: partial multibyte character
    What has changed?
    Was it considered a Bug in 10g because it allowed this behavior and now 11g is operating correctly?
    29275, 00000, "partial multibyte character"
    // *Cause:  The requested read operation could not complete because a partial
    //          multibyte character was found at the end of the input.
    // *Action: Ensure that the complete multibyte character is sent from the
    //          remote server and retry the operation. Or read the partial
    //          multibyte character as RAW.It appears to me a bug got fixed.

  • Data sync between oracle and sql server

    Greetings Everyone,
    Your expert views are highly appreciable regarding the following.
    We at work are evaluation different solutions to achieve data synchronization between oracle and sql server data bases. Data sync i mentioned here is for live applications. We are runnign oracle EBS 11i with custom applications and intending to implement a custom software based on .NET and SQL Server. Now the whole research is to see updates and data changes whenever happens between these systems.
    I googled and found Oracle Golden Gate, Microsoft SSIS, Wisdom Force from Informatica....
    If you can pour in more knowledge then it's great.
    Thank You.

    Most of the work involved has to be done through scripts and there is no effective GUI to implement OGG.However using commands is not vey togh and they are very intutive.
    These are the steps, from a high level:
    1.Get the appropriate GG Software for your source and target OS.
    2.Install GG on source and target systems.
    3.Create Manager and extract processes on source system
    4.Create Manager and replicat processes on target system
    5.Start these processes.
    First try to achieve uni-directional replication. Then Bi-directional is easy.I have implemented bi-directional active active replication using Oracle DBs as source and target. Refer to Oracle installation and admin guides for more details.
    Here is a good article that may be handy in your case.
    http://www.oracle.com/technetwork/articles/datawarehouse/oracle-sqlserver-goldengate-460262.html
    Edited by: satrap on Nov 30, 2012 8:33 AM

  • Standard XML schema for Vendor data exchange between SAP and other system

    Is there a SAP standard way of XML schema that we exchange between SAP and other system? Please let me know.
    Thanks.

    See SAP Interface Repository (http://ifr.sap.com).
    My proposal is to leave old SAP connectors staff and use SAP Exchange Infrastructure. There is a support of industry XML standards in XI 3.0 like xCBL.

  • Transporting data between sybase and oracle

    I want to know how do you transport data from sybase to oracle. The table are created in sybase and we want to move
    everything over to Oracle8i

    I want to know how do you transport data from sybase to oracle. The table are created in sybase and we want to move
    everything over to Oracle8i .if sysbase has a facility to export the tables as comma separated text files, export them and then user sql *loader to load the data into oracle.
    you can check out oracle utilities documentation for sql *loader case studies.
    if there is a huge amount of data, you can try oracle trasnparent gateway for sybase. thank you..:>)

  • Password Field not mapped between Request and Provisioning Form

    Hi to all. I'm working with OIM 11g. I've faced a strange issue. I'm not sure I'm working properly, so let me explain you my case. In my installation I've got the SSH connector, which is correctly connected with the physical resources. I've loaded the resource dataset ProvisionResourceSSH User bundled with the connector. Consider now that the user "goofy", with "ALL USERS" role, tries to make a request Provision Resource SSH User (Request Based provisioning). He fills in all the field in the appropriate manner, but when OIM triggers "Create User" provisioning task, after the required approval process, the password field is always blank (although goofy filled it in!!!).
    I've thought: "ok, it seems a role trouble". And effectively, if goofy has got also the role "REQUEST ADMINISTRATORS", the provisioning form shows the password field correctly valued (as goofy stated in his request).
    Note that all the fields are correctly mapped between request dataset and the provisioning form (I'm using the original dataset and the original provisioning form installed by the connector). So all other fields filled by goofy on the request form (request based provisioning) are correctly passed to the provisioning form. All the fields, except for the password.
    Am I wrong in something? How Could I make possible to pass the data filled on the request for the password field to the provisioning form even if the requester has not the role "REQUEST ADMINISTRATORS"?
    Thank you in advance for the help.

    This sure seems goofy! ;-) ... can you try making the ALL Users have all the permissions on the Resource Object and the Process Form and test it out? Also try from the backend at the database and see if the table has NULL for the password field? ... What's the type of password field in the dataset and the process form? Encrypted/Secret at both ends or mis match? Try making them plain text both the places as well.
    -Bikash

  • Help, about data synchronization between C++ and PHP API!

    I use Berkeley DB XML as my Server's database, the client access the database via http by C++ API. I don't close the XmlManager and XmlContainer after read and write the database for better performance. However, I provide another way to manipulate the database via web by PHP. After I updated the data by PHP, I found that I couldn't catch the update.
    If I close the XmlManager and XmlContainer after read and write the database every time, the problem disappeared. But I can't do that for performance.
    How can I solve this problem? Thanks!

    First of all, thank you for your attention.
    I don't share the same environment for the two processes, but I seted the same configure flags on the environment. The flags as follows:
    DB_CREATE|DB_INIT_LOCK|DB_INIT_LOG|DB_INIT_MPOOL|DB_INIT_TXN|DB_RECOVER;
    The C++ code as follows:
    //C++ code
    UINT32 envFlags = DB_CREATE|DB_INIT_LOCK|DB_INIT_LOG|DB_INIT_MPOOL|DB_INIT_TXN|DB_RECOVER;
    INT32 lRet = 0;
    string ctnName = "rls_services.bdbxml";
    string docName = "index.xml";
    CHAR acXQuery[256] = {0};
    CHAR acXmlDoc[] = "<test>C++ test</test>";
    //from      
    DbEnv *pDbEnv  = new DbEnv(0);
    XmlManager *pxmlMgr = NULL;
    pDbEnv->open("/usr/local/xdms", envFlags, 0);
    pxmlMgr = new XmlManager(pDbEnv, DBXML_ADOPT_DBENV);
    lRet = pxmlMgr->existsContainer(ctnName);
    if(0 == lRet)
         pxmlMgr->createContainer(ctnName);     
    XmlContainer xmlCtn = pxmlMgr->openContainer(ctnName);
    //to
    // the code between 'from' and 'to' is a seperate function
    XmlQueryContext xmlQC = pxmlMgr->createQueryContext();
    XmlUpdateContext xmlUC = pxmlMgr->createUpdateContext();
    sprintf(acXQuery, "doc(\"%s/%s\")", ctnName, docName);
    XmlQueryExpression xmlQE = pxmlMgr->prepare(acXQuery, xmlQC);
    XmlResults xmlResult = xmlQE.execute(xmlQC);
    XmlDocument xmlDoc;
    if (xmlResult.hasNext())
    xmlDoc = xmlCtn.getDocument(docName);
    xmlDoc.setContent(acXmlDoc);
    xmlCtn.updateDocument(xmlDoc, xmlUC);
    else
    xmlDoc = pxmlMgr->createDocument();
    xmlDoc.setName(docName);
    xmlDoc.setContent(acXmlDoc);
    xmlCtn.putDocument(xmlDoc, xmlUC);
    // I don't close the Container and Manager for performance
    The PHP code as follow:
    <php?
    $DB_DIR = "/usr/local/xdms";
    $env = new Db4Env();
    $enFlags = DB_CREATE|DB_INIT_LOCK|DB_INIT_LOG|DB_INIT_MPOOL|DB_INIT_TXN|DB_RECOVER;
    $env->open("/usr/local/xdms", $enFlags, 0);
    $xmlManager = new XmlManager($env, 0);
    $ctnName = 'rls_services.bdbxml';
    $docName = 'index.xml';
    $docContent = '<test>PHP test</test>';
    if (!$xmlManager->existsContainer($ctnName))
         return ;
    $xmlCtn = $xmlManager->openContainer($ctnName);
    $xmlQC = $xmlManager->createQueryContext();
    $xmlUC = $xmlManager->createUpdateContext();
    $acXQuery = "doc('".$ctnName.'/'.$docName."')";
    $xmlQE = $xmlManager->prepare($acXQuery, $xmlQC);
    $xmlResult = $xmlQE->execute($xmlQC);
    if ($xmlResult->hasNext())
         $xmlDoc = $xmlCtn->getDocument($docName);
         $xmlDoc->setContent($docContent);
         $xmlCtn->updateDocument($xmlDoc, $xmlUC);
    else
         $xmlDoc = $xmlManager->createDocument();
         $xmlDoc->setName($docName);
         $xmlDoc->setContent($docContent);
         $xmlCtn->putDocument($xmlDoc, $xmlUC);
    unset($xmlDoc);
    unset($xmlCtn);
    ?>
    The code between 'from' and 'to' excute only one time when server start.
    After the server started, I write data by PHP. I can read out the data that
    writed by PHP just now, but I can't read out the data by C++ because I don't
    close the XmlContainer and XmlManager.
    After I restarted the server, I can readout the data by C++. But I can't
    open and close database for each request for performance as a server.
    what should I do ? Thank you!
    Whether I express my question clearly or not?

Maybe you are looking for

  • SRM PO- R/3 PO..Closing issues

    Hi, We are on SRM 5.0 and ECS Scenario We have an issue here. SRM PO has three line items. Replicated fine to the backened item 1 item 2 item 3 Item 1 is fully received and invoiced.(In backend R/3). Item 2 No goods to receive . (Tried no further con

  • Reformatting an external hard drive from mac to pc

    i have a 320GB simple tech external hard drive that i used for my mac book pro. My mac book pro broke down completely so i bought an Asus laptop which runs Windows 7. I go to plug in the external hard drive, however i cant access it. It does show up

  • Stuck on waiting

    My purchases page keeps trying to reload and keeps skipping.  Purchases are stuck on waiting and I can't resume or stop the downloads.  Any ideas?

  • What Kind of Plastic is Used on the GTX 970? Is it Painted Black?

    Hello, I'm going to be painting the black parts of the MSI GTX 970 GAMING 4G card I have, and so I was wondering about the cooling system they have on it. There's probably a specific name for the thing that I'm referring to, but I'm basically paintin

  • À QUELLE ADRESSE E-MAIL PEUT-ON SUGGÉRER DE NOUVEAU PRODUIT À APPLE???

    BONJOUR, JE VOUDRAIS QUE VOUS VÉRIFIEZ AVEC LES DIRIGEANT DE BSIDE U SI C'EST POSSIBLE D'ÉLARGIR LES TERRITOIRES QU'ILS POURRAIENT DESSERVIR. ILS ONT INVENTER UNE GPS LOCALISATEUR SPÉCIALEMENT POUR LES PERSONNES EN PERTES D'AUTONOMIE. MON MARI A ÉTÉ