SMODILOG and Trasnport Objects Issues

Hi Experts,
I want to clarify on the  following areas.
1) In SMODILOG Table, is there any possibility to have custom objects like Z/Y ,
If so , why it is existed in that table ,Any reason for this ?.
2) If we Transport Objects will that transport objects are saved as Modification Assistant ?
Please provide clarity on this.
Thanks & Regards,

1. This indeed is possible, like for example BaDI implementations. Reason: really wouldn't know, but I guess because all of these changes can in  a way be seen as extending (modifyng) SAP standard??
2. This question doesn't make any sense. You have to rephrase your question, and clearify what it is you are actually looking for.
The modification assistent is used to help you get a good overview of all modifications made. So during an upgrade this will come in handy.

Similar Messages

  • Issues in persisting dynamic entity and view objects using MDS

    Hi All,
    I'm trying to create dynamic entity and view objects per user session and to persist these objects throughout the session, I'm trying to use MDS configurations(either file or Database) in adf-config.xml.
    I'm facing following two errors while trying to run the app module:
    1. MDS error (MetadataNotFoundException): MDS-00013: no metadata found for metadata object "/model/DynamicEntityGenModuleOperations.xml"
    2. oracle.mds.exception.ReadOnlyStoreException: MDS-01273: The operation on the resource /sessiondef/dynamic/DynamicDeptEntityDef.xml failed because source metadata store mapped to the namespace / DEFAULT is read only.
    I've gone through the following links which talks about the cause of the issue, but still can't figure out the issue in the code or the config file. Please help if, someone has faced a similar issue.
    [http://docs.oracle.com/cd/E28271_01/doc.1111/e25450/mds_trouble.htm#BABIAGBG |http://docs.oracle.com/cd/E28271_01/doc.1111/e25450/mds_trouble.htm#BABIAGBG ]
    [http://docs.oracle.com/cd/E16162_01/core.1112/e22506/chapter_mds_messages.htm|http://docs.oracle.com/cd/E16162_01/core.1112/e22506/chapter_mds_messages.htm]
    Attached is the code for dynamic entity/view object generation and corresponding adf-config.xml used.
    ///////////App Module Implementation Class/////////////////////////
    public class DynamicEntityGenModuleImpl extends ApplicationModuleImpl implements DynamicEntityGenModule {
    private static final String DYNAMIC_DETP_VO_INSTANCE = "DynamicDeptVO";
    * This is the default constructor (do not remove).
    public DynamicEntityGenModuleImpl() {
    public ViewObjectImpl getDepartmentsView1() {
    return (ViewObjectImpl) findViewObject("DynamicDeptVO");
    public void buildDynamicDeptComp() {
    ViewObject internalDynamicVO = findViewObject(DYNAMIC_DETP_VO_INSTANCE);
    if (internalDynamicVO != null) {
    System.out.println("OK VO exists, return Defn- " + internalDynamicVO.getDefFullName());
    return;
    EntityDefImpl deptEntDef = buildDeptEntitySessionDef();
    ViewDefImpl viewDef = buildDeptViewSessionDef(deptEntDef);
    addViewToPdefApplicationModule(viewDef);
    private EntityDefImpl buildDeptEntitySessionDef() {
    try {
    EntityDefImpl entDef = new EntityDefImpl(oracle.jbo.server.EntityDefImpl.DEF_SCOPE_SESSION, "DynamicDeptEntityDef");
    entDef.setFullName(entDef.getBasePackage() + ".dynamic." + entDef.getName());
    entDef.setName(entDef.getName());
    System.out.println("Application Module Path name: " + getDefFullName());
    System.out.println("EntDef :" + entDef.getFileName() + " : " + entDef.getBasePackage() + ".dynamic." + entDef.getName());
    entDef.setAliasName(entDef.getName());
    entDef.setSource("DEPT");
    entDef.setSourceType("table");
    entDef.addAttribute("ID", "ID", Integer.class, true, false, true);
    entDef.addAttribute("Name", "NAME", String.class, false, false, true);
    entDef.addAttribute("Location", "LOCATION", Integer.class, false, false, true);
    entDef.resolveDefObject();
    entDef.registerSessionDefObject();
    entDef.writeXMLContents();
    entDef.saveXMLContents();
    return entDef;
    } catch (Exception ex) {
    System.out.println(ex.getLocalizedMessage());
    return null;
    private ViewDefImpl buildDeptViewSessionDef(EntityDefImpl entityDef) {
    try {
    ViewDefImpl viewDef = new oracle.jbo.server.ViewDefImpl(oracle.jbo.server.ViewDefImpl.DEF_SCOPE_SESSION, "DynamicDeptViewDef");
    viewDef.setFullName(viewDef.getBasePackage() + ".dynamic." + viewDef.getName());
    System.out.println("ViewDef :" + viewDef.getFileName());
    viewDef.setUseGlueCode(false);
    viewDef.setIterMode(RowIterator.ITER_MODE_LAST_PAGE_FULL);
    viewDef.setBindingStyle(SQLBuilder.BINDING_STYLE_ORACLE_NAME);
    viewDef.setSelectClauseFlags(ViewDefImpl.CLAUSE_GENERATE_RT);
    viewDef.setFromClauseFlags(ViewDefImpl.CLAUSE_GENERATE_RT);
    viewDef.addEntityUsage("DynamicDeptUsage", entityDef.getFullName(), false, false);
    viewDef.addAllEntityAttributes("DynamicDeptUsage");
    viewDef.resolveDefObject();
    viewDef.registerSessionDefObject();
    viewDef.writeXMLContents();
    viewDef.saveXMLContents();
    return viewDef;
    } catch (Exception ex) {
    System.out.println(ex.getLocalizedMessage());
    return null;
    private void addViewToPdefApplicationModule(ViewDefImpl viewDef) {
    oracle.jbo.server.PDefApplicationModule pDefAM = oracle.jbo.server.PDefApplicationModule.findDefObject(getDefFullName());
    if (pDefAM == null) {
    pDefAM = new oracle.jbo.server.PDefApplicationModule();
    pDefAM.setFullName(getDefFullName());
    pDefAM.setEditable(true);
    pDefAM.createViewObject(DYNAMIC_DETP_VO_INSTANCE, viewDef.getFullName());
    pDefAM.applyPersonalization(this);
    pDefAM.writeXMLContents();
    pDefAM.saveXMLContents();
    ////////adf-config.xml//////////////////////
    <?xml version="1.0" encoding="windows-1252" ?>
    <adf-config xmlns="http://xmlns.oracle.com/adf/config" xmlns:config="http://xmlns.oracle.com/bc4j/configuration" xmlns:adf="http://xmlns.oracle.com/adf/config/properties"
    xmlns:sec="http://xmlns.oracle.com/adf/security/config">
    <adf-adfm-config xmlns="http://xmlns.oracle.com/adfm/config">
    <defaults useBindVarsForViewCriteriaLiterals="true"/>
    <startup>
    <amconfig-overrides>
    <config:Database jbo.locking.mode="optimistic"/>
    </amconfig-overrides>
    </startup>
    </adf-adfm-config>
    <adf:adf-properties-child xmlns="http://xmlns.oracle.com/adf/config/properties">
    <adf-property name="adfAppUID" value="TestDynamicEC-8827"/>
    </adf:adf-properties-child>
    <sec:adf-security-child xmlns="http://xmlns.oracle.com/adf/security/config">
    <CredentialStoreContext credentialStoreClass="oracle.adf.share.security.providers.jps.CSFCredentialStore" credentialStoreLocation="../../src/META-INF/jps-config.xml"/>
    </sec:adf-security-child>
    <persistence-config>
    <metadata-namespaces>
    <namespace metadata-store-usage="mdsRepos" path="/sessiondef/"/>
    <namespace path="/model/" metadata-store-usage="mdsRepos"/>
    </metadata-namespaces>
    <metadata-store-usages>
    <metadata-store-usage default-cust-store="true" deploy-target="true" id="mdsRepos">
    <metadata-store class-name="oracle.mds.persistence.stores.file.FileMetadataStore">
    <property name="metadata-path" value="/tmp"/>
    <!-- <metadata-store class-name="oracle.mds.persistence.stores.db.DBMetadataStore">
    <property name="jndi-datasource" value="jdbc/TestDynamicEC"/>
    <property name="repository-name" value="TestDynamicEC"/>
    <property name="jdbc-userid" value="adfmay28"/>
    <property name="jdbc-password" value="adfmay28"/>
    <property name="jdbc-url" value="jdbc:oracle:thin:@localhost:1521:XE"/>-->
    </metadata-store>
    </metadata-store-usage>
    </metadata-store-usages>
    </persistence-config>
    </adf-config>
    //////////////////////////////////////////////////////////////////////////////////////////////////////////

    Hi Frank,
    I m trying to save entity and view object xml by calling writeXMLContents() and saveXMLContents() so that these objects can be retrieved using the xmls later on.
    These methods internally use MDS configuration in adf-config.xml, which is creating the issue.
    Please share your thoughts on resolving this or if, there is any other way of creating dynamic entity/view objects for db tables created at runtime.
    Nik

  • Is there any BAPI/FM to create a trasnport request and add objects into it?

    Hi Experts,
    Could anyone please assist me in finding a BAPI/FM to create a transport request and add objects to it.
    I could find a BAPI BAPI_REQUEST_CREATE, it creates trasnport request but doesn't have option to add objects.
    Thanks
    Yogesh Gupta

    Create a custom Function Module , in that write a BDC to create a transport request and add objects to it.
    Thats it.

  • Newbie question about entity and view objects

    Hi everyone,
    My first ADF application in JDeveloper is off to a difficult start. Having come from a forms background, I know that it is necessary avoid using post-query type lookups in order to have full filtering using F11/Ctrl+F11. This means creating an CRUDable view and getting as much of the lookup data as possible into the view without losing the ability to modify the data. In JDeveloper, I do not know how to build my data model to support that. My thought was to start with a robust updateable view as my main CRUD EO and then create a VO on top of that with additional EOs or VOs. But, I have found that I cannot add VOs to another VO. However, if I link the VOs then I have a master-detail which will not work for me.
    For example, I have two joins to CSI_INST_EXTEND_ATTRIB_V shown in the queries below and need to have that show in the table displaying the CRUD VO’s data. It seemed that the best way to do this is to create a CSI_INST_EXTEND_ATTRIB_V entity object and view object. The view object would have two parameters, P_INSTANCE_ID and P_ATTRIBUTE name. Both the building and the unit are needed on the same record, and this is not a master-detail even though it might look that way. (A master-detail data control will not work for me because I need all this data to appear on the same line in the table.) So, I need help figuring out the best way to link these to the main (CRUD) view; I want as much of this data as possible to be filterable.
    select
    cieav.attribute_value
    from
    csi_inst_extend_attrib_v cieav
    where cieav.instance_id = p_instance_id
    and cieav.attribute_code = 'BUILDING NAME'
    select
    cieav.attribute_value
    from
    csi_inst_extend_attrib_v cieav
    where cieav.instance_id = p_instance_id
    and cieav.attribute_code = 'UNIT NAME'
    Ultimately, I need to display a ton of data in each record displayed in the UI table, so a ton of joins will be needed. And, I need to be able to add records using the UI table also.
    James

    Hi Alejandro,
    Sorry if I caused confusion with my first post. What I had in mind assumed that I could have a single CSI_INST_EXTEND_ATTRIB_V EO with a BuildingVO and UnitVO on top of it. So, I wrote the queries individually to show how I would invoke each view. When I realized that confused the issue, I rewrote the query to explain things better.
    Now having seen your 2 queries. You need to create 2 EO. One for each table. Then create an association between the 2 aeO (this will be the join you are talking about). Then, you need to create a VO based on one of the EO and after you can modify and add the second EO (in here you select the join type).
    After your done with this, youll have 1 VO that supports CRUD in both tables.
    There were three tables in the query: CIEAV_BUILDING, CIEAV_UNIT, and T -- the main CRUD table. When you say that I should create two EOs, do you mean that they are to be for CIEAV_BUILDING and CIEAV_UNIT? Or, CIEAV and T? Assuming CIEAV and T, that sounds like it would allow me to show my building or unit on the record but not both.
    By the way, everything is a reference except for the main CRUD table.
    Look forward to hearing from you. Thanks for your help (and patience).

  • Beginners Guide sound installation -- OSS and ALSA objective features

    Beginners Guide sound installation -- OSS and ALSA objective features needed
    Hi,
    In most wiki pages i just edit the things that i think need editing. Just look at the oss history
    In all those cases i didn't really saw a need to ask for permission.. that would kinda destroy the wiki idea.
    However i want to change the sound instructions in the beginners guide so i made a copy of the entire guide and the part that i changed there is: http://wiki.archlinux.org/index.php/Use … ling_Sound now i have a slight issue there. As you can see i'm in favor of OSS and because you can spot that it's not objective. Now i would ask 2 things.
    1. Could you all post your features of OSS and ALSA
    2. Once that's done can i have permission to place that section in the beginners guide? that will also include removing the sound installation from the beginners guide which i already did in my version
    My personal reason to do this. i've read this  and am since then pro OSS and against the ALSA/PulseAudio combo. i think that combo needs to get out of linux (most notably Fedora and Ubuntu) ASAP. and oss needs to go back into the linux kernel
    Thank you for your time,
    Mark

    ngoonee wrote:
    I'm an alsa/pulse user, so I'll give a bit of the 'other side'.
    Alsa/pulse
    Pros:-
    network sound
    advanced connection of sink/source (including merging sinks)
    bluetooth support!
    highly supported by existing apps, either through directly supporting Pulse or through its alsa plugin (you should not need to recompile properly-written apps which do not assume they should write audio data directly to hardware, I think I only needed to recompile mpd on my system)
    Cons:-
    Setup isn't the easiest. Can't comment vs OSS4 because I haven't tried it
    OSS4
    Pros:-
    Everyone seems to say sound quality is better. I guess that's because they're comparing it with dmix alsa. Use pulse with alsa and you should not notice any difference in sound quality though.
    Cons:-
    Most apps nowadays default output to alsa. Meaning OSS plays them using an alsa plugin.
    USB support is admittedly skimpy.
    EDIT: Having read the sound article you referred to, my only comment is that the writer really has it in for Pulse... 3 seconds latency, where'd he get that from? I use pulse for audio recording (when I'm lazy to fire up JACK) and while there IS latency, its definitely in the ms range.
    Thanx for the input
    Gen2ly wrote:
    Gen2ly wrote:...As a side note, do you need libflashsupport here???...
    markg85, libflashsupport isn't needed. [1]
    pacman -Ql oss | grep flash
    If you don't know, please don't put in wiki, this could cause unnecessary problems.  As for the mms section:
    If your stream sounds ugly in totem like it did with me then you could try to play it with another codec like ffmpeg (mplayer). That "fixed" the issue for me. This will not fix the issue that somehow pops up in gstreamer when playing MMS streams but it will give you the option to play it with good sound quality. Playing it in mplayer is simple:
    # mplayer mmsh://yourstreamurl
    Could you fix this?  ffmpeg is not a codec .  Also define ugly, and what is somehow?
    markg85 wrote:Thanx a lot for your feedback. i will certainly use it when i make more edits.
    As for the things you didn't know. As soon as i fully understand how i can get a microphone working in OSS i will add that to the wiki as well. Unless you already know it.. in that case, feel free to add it.
    For the mic, I did get mine going.  Can't remember just how I did mine (sorry, think I had to disable one of the inputs),  but do remember to prevent it from passing through the speakers had to disable "Misc Microphone".
    # ossmix
    Selected mixer 0/High Definition Audio ALC888
    Known controls are:
    jack.green.mode <front|rear|center/LFE|side|pcm4|input> (currently front)
    jack.green [<leftvol>:<rightvol>] (currently 29.9:29.9 dB)
    jack.green.mute ON|OFF (currently OFF)
    jack.black.mode <front|rear|center/LFE|side|pcm4|input> (currently center/LFE)
    jack.black [<leftvol>:<rightvol>] (currently 29.9:29.9 dB)
    jack.black.mute ON|OFF (currently OFF)
    jack.orange.mode <front|rear|center/LFE|side|pcm4|input> (currently rear)
    jack.orange [<leftvol>:<rightvol>] (currently 29.9:29.9 dB)
    jack.orange.mute ON|OFF (currently OFF)
    jack.gray.mode <front|rear|center/LFE|side|pcm4|input> (currently pcm4)
    jack.gray [<leftvol>:<rightvol>] (currently 29.9:29.9 dB)
    jack.gray.mute ON|OFF (currently OFF)
    jack.pink.mode <front|rear|center/LFE|side|pcm4|input> (currently input)
    jack.pink [<leftvol>:<rightvol>] (currently 19.9:19.9 dB)
    jack.pink.mute ON|OFF (currently OFF)
    jack.fp-pink.mode <front|rear|center/LFE|side|pcm4|input> (currently front)
    jack.fp-pink [<leftvol>:<rightvol>] (currently 29.9:29.9 dB)
    jack.fp-pink.mute ON|OFF (currently OFF)
    jack.blue.mode <front|rear|center/LFE|side|pcm4|input> (currently input)
    jack.blue [<leftvol>:<rightvol>] (currently 29.9:29.9 dB)
    jack.blue.mute ON|OFF (currently OFF)
    jack.fp-green.mode <front|rear|center/LFE|side|pcm4|input> (currently front)
    jack.fp-green [<leftvol>:<rightvol>] (currently 29.9:29.9 dB)
    jack.fp-green.mute ON|OFF (currently OFF)
    record.mix.mute.mic1 ON|OFF (currently OFF)
    record.mix.mute.fp-mic1 ON|OFF (currently OFF)
    record.mix.mute.linein1 ON|OFF (currently OFF)
    record.mix.mute.fp-headphone1 ON|OFF (currently OFF)
    record.mix.mute.green1 ON|OFF (currently OFF)
    record.mix.mute.black1 ON|OFF (currently OFF)
    record.mix.mute.orange1 ON|OFF (currently OFF)
    record.mix.mute.gray1 ON|OFF (currently OFF)
    record.mix.mute.input-mix1 ON|OFF (currently OFF)
    record.mix1 [<leftvol>:<rightvol>] (currently 38.9:38.9 dB)
    record.mix.mute.mic2 ON|OFF (currently OFF)
    record.mix.mute.fp-mic2 ON|OFF (currently OFF)
    record.mix.mute.linein2 ON|OFF (currently OFF)
    record.mix.mute.fp-headphone2 ON|OFF (currently OFF)
    record.mix.mute.green2 ON|OFF (currently OFF)
    record.mix.mute.black2 ON|OFF (currently OFF)
    record.mix.mute.orange2 ON|OFF (currently OFF)
    record.mix.mute.gray2 ON|OFF (currently OFF)
    record.mix.mute.input-mix2 ON|OFF (currently OFF)
    record.mix2 [<leftvol>:<rightvol>] (currently 4.4:2.9 dB)
    misc.mic [<leftvol>:<rightvol>] (currently 0.0:0.0 dB)
    misc.fp-mic [<leftvol>:<rightvol>] (currently 46.4:37.4 dB)
    misc.linein [<leftvol>:<rightvol>] (currently 38.9:38.9 dB)
    misc.fp-headphone [<leftvol>:<rightvol>] (currently 34.4:34.4 dB)
    misc.green [<leftvol>:<rightvol>] (currently 34.4:38.9 dB)
    misc.black [<leftvol>:<rightvol>] (currently 38.9:38.9 dB)
    misc.orange [<leftvol>:<rightvol>] (currently 38.9:38.9 dB)
    misc.gray [<leftvol>:<rightvol>] (currently 40.4:41.9 dB)
    misc.input-mix <mic|fp-mic|linein> (currently mic)
    misc.front-mute ON|OFF (currently OFF)
    misc.input-mix-mute1 ON|OFF (currently OFF)
    misc.front1 [<leftvol>:<rightvol>] (currently 43.4:43.4 dB)
    misc.front2 <front|input-mix> (currently front)
    misc.rear-mute ON|OFF (currently OFF)
    misc.input-mix-mute2 ON|OFF (currently OFF)
    misc.rear1 [<leftvol>:<rightvol>] (currently 4.4:4.4 dB)
    misc.rear2 <rear|input-mix> (currently rear)
    misc.center/lfe-mute ON|OFF (currently OFF)
    misc.input-mix-mute3 ON|OFF (currently OFF)
    misc.center/lfe1 [<leftvol>:<rightvol>] (currently 41.9:41.9 dB)
    misc.center/lfe2 <center/LFE|input-mix> (currently center/LFE)
    misc.side-mute ON|OFF (currently OFF)
    misc.input-mix-mute4 ON|OFF (currently OFF)
    misc.side1 [<leftvol>:<rightvol>] (currently 35.9:35.9 dB)
    misc.side2 <side|input-mix> (currently side)
    misc.pcm4-mute ON|OFF (currently OFF)
    misc.input-mix-mute5 ON|OFF (currently OFF)
    misc.pcm41 [<leftvol>:<rightvol>] (currently 25.4:25.4 dB)
    misc.pcm42 <pcm4|input-mix> (currently pcm4)
    vmix0-enable ON|OFF (currently ON)
    vmix0-rate <decimal value> (currently 48000) (Read-only)
    vmix0-channels <Stereo|Multich> (currently Stereo)
    vmix0-src <Fast|Low|Medium|High|High+|Production|OFF> (currently Medium)
    vmix0-outvol <monovol> (currently 25.0 dB)
    vmix0-invol <monovol> (currently 25.0 dB)
    vmix0.pcm8 [<leftvol>:<rightvol>] (currently 19.9:19.9 dB) ("knotify4")
    vmix0.pcm9 [<leftvol>:<rightvol>] (currently 25.0:25.0 dB)
    vmix0.pcm10 [<leftvol>:<rightvol>] (currently 25.0:25.0 dB)
    vmix0.pcm11 [<leftvol>:<rightvol>] (currently 25.0:25.0 dB)
    For libflashsupport on the same page you linked it clearly states:
    #  Flash V9 and V10 require libflashsupport to output sound via OSS. Typically a 32-bit version of the library is required.
    # Flash V10 has a 64-bit version which requires a 64 bit libflashsupport.
    Also i tested it with and without libflashsupport. On archlinux (x64 running here) there most certainly is a need for libflashsupport when you want to have sound in your flash. And yes i tested the archlinux OSS version and the mercurial version (running now) bith need it  installed manually! On my pc sound in flash didn't work without it but did with it. So, no not removing from the wiki as it's needed. But i see you removed it for me! please do NOT do that if you didn't even verified it. I use flash 10 x64 and i need it!
    As for the ffmpeg "codec" changed it to backend.
    And i did get the microphone working near perfect: http://www.4front-tech.com/forum/viewtopic.php?p=13192
    Now for some news you all might like.
    On my school i need to do an investigation to whatever i want and i'm heavily thinking about investigating the pros/cons of alsa compared to oss (or oss compared to alsa). That investigation will take from monday next week till next mondey till friday 23 of oktober. In that investigation i'm going to do some in depth look of alsa and oss and that will include the usability as well.
    Following up on that investigation i will spend another 8 weeks on my school making a volume control application that can be used with alsa and oss and easily expandable with other sound systems. The goal of this is to make one sound application that can manage (in the first place) alsa and oss. oss is going to be implemented and alsa is probably going to be dummy implemented because it's likely way to much for me to implement both.
    Before you get to exited, both projects (investigating and making the application) are just made up today and i just don't know if both will get accepted by my school. I asked one teacher and he liked the idea a lot and could potentially have a value for the sound management under linux. Once i do get this started i will involve the community (YOU!) with this since this project can't be done without the community specially the investigation.
    And once i start and have something to tell/ask i will blog about it on http://blog.mageprojects.com
    edit::
    And this idea already got dumped. read more a few posts down or click: http://bbs.archlinux.org/viewtopic.php? … 34#p612634
    Last edited by markg85 (2009-09-03 17:51:45)

  • What is the difference between delivery date and actual goods issue date?

    Hi ,
    Can you please tell me the difference between Delivery date and Actual goods issue date ?
    Delivery data can be be greater than goods issue date?
    Thanks,
    Ajay

    Hi Ajay,
                  Delivery date means its the original date on which the delivery should be done.Actual goods Issue date is the date on which goods to be delivered according to the purchase order.There can be delay in the delivery so we maintain 2 objects.
    Yes delivery date can be greater than Goods Issue date.
    Hope this clears
    Regards
    Karthik

  • JAVA Calling Oracle Function and Returning OBJECT

    HI,
    I am working as a developer in java/j2ee project.
    I am facing one issue:
    I need to call Oracle function from java code. Oracle User define function is residing in oracle package and returning Object which contains data.
    Can you please help me
    With Best Regards

    golduniya wrote:
    I need to call Oracle function from java code. Oracle User define function is residing in oracle package and returning Object which contains data.
    Can you please help meIt requires a great deal of Oracle jdbc driver specific code.
    [http://download-west.oracle.com/docs/cd/B10501_01/java.920/a96654/oraint.htm#1012664]

  • ALE and IDocs Relevant Issue

    Dear Gurus,
    I created a Master Data (MD)u2018testmrtu2019 via MM01 in our DEV environment.
    Now I want to transfer it into QAS environment via ALE mode. (The ALE configuration work was done.)
    The MD was successfully sent from DEV via BD10. (The light was green via WE05 in DEV.)
    But Iu2019ve failed when I receiving the MD in QAS and can see error logs via BD73.
    Error Log:
    Message Type: MATMAS
    IDoc Status: 51
    Status Grouping: Red Lights.
    Status text: Application document not posted.
    Double Click 'Application document not posted.' and expand the 'Status Records' of 51, I can see:
    'Function module not allowed: AFS_RETAIL_ARTMAS_IDOC_INPUT'.
    Have you come across this problem? How to solve the problem?
    Thank you very much!
    Jason

    Hi ,
    For the Error:  Function module not allowed: AFS_RETAIL_ARTMAS_IDOC_INPUT'
    Error Log:
    Idocs would be failing in 51 status in Error state.
    for
    Message Type: MATMAS
    IDoc Status: 51
    Status Grouping: Red Lights.
    Status text: Application document not posted.
    Diagnosis
    The function module AFS_RETAIL_ARTMAS_IDOC_INPUT'
    and the application object type  which were determined are not valid for this IDoc.
    Procedure
    1. Please check that the process code in the inbound partner profile is correct.
    2. If this is the case, you should check the ALE inbound methods ALE inbound methods for the process code
    and see whether the specified function module and application object type are correct.
    3. If this is also the case, then the function module and the application object type are not permitted
    for the logical message type, message variant, message function and basis type that are contained
    in the IDoc control record.
    You should check whether the correct values have been assigned to these fields in the control record.
    If they do have the correct values, then the assignment to the function module and the
    application object type needs to be maintained.
    Please check and revert in case u still face the issue .
    Regards,
    Shravan A

  • Problem while loading texts and authorization objects file in RAR

    Hi all,
           i am getting internet explorer error while loading the texts and authorization objects text files in RAR .actually we uploaded rule file before this,does this step causes any error ?if so how to resolve this error.do i need to remove all rules/risks and then load text and authorization files? is there any shortcut to renove all risks generated in one shot? please reply me soon to resolve this.
    Thanks,
    Joseph.

    Hi Joseph,
    Please make sure to convert both the files in UTF-8 encoding format and then try to upload the files again. This should resolve the issue and if not then please paste the logs here.
    Regards
    Harleen

  • How to combine Session Facade and Transfer object?

    Hello All!
    I'm working on an enterprise application. Presentation layer is a stand alone client, business logic is build on the Glassfish v2.1 and MySQL is used as a database. The client is connection to the GlassFishj server remotely using EJBs.
    I have problems with business logic architecture.
    Here is the brief description of backend application architecture design:
    1. Session Facade pattern is used to simplify the client and application server interface and to provide application layers between backend (http://java.sun.com/blueprints/corej2eepatterns/Patterns/SessionFacade.html).
    2.Transfer Object pattern to define update transfer objects strategy in order to decrease network overhead during client and application server interactions and to provide version control for objects. Transfer objects are designed as simple java business serializable objects. (http://java.sun.com/blueprints/corej2eepatterns/Patterns/TransferObject.html)
    3. Originally the backend application consisted of three modules: users, storage and orders, but at the end I have decided to divide my application into the following parts - assortments, map, menu, orders, transactions, users.
    4. All MySQL database transactions are via JDBC using procedures. No use of entity beans.
    Questions:
    1. I have some doubts about using Session Facade and Transfer object patterns at the same time. At first I'd mike to cite the definitions of the patters from the SUN official web site.
    * Use a session bean as a facade to encapsulate the complexity of interactions between the business objects participating in a workflow. The Session Facade manages the business objects, and provides a uniform coarse-grained service access layer to clients.
    * Use a Transfer Object to encapsulate the business data. A single method call is used to send and retrieve the Transfer Object. When the client requests the enterprise bean for the business data, the enterprise bean can construct the Transfer Object, populate it with its attribute values, and pass it by value to the client.
    * So, if I use Transfer Object along with Session Facade, it makes some difficulties with object version control, because I 2 or
    3 transfer objects controls with the 1 bean class. The best option for Transfer object Pattern is that each transfer object should have its own bean class to provide ability of object's version control. In the case it can bring the network overhead because of frequent remote calls caused by the large number of the bean classes.
    * So, should I use the both patterns? If yes, how to manage the interaction the patterns. If no, which one to use.
    2. E.g. I have a huge list of the Order objects and each Order object consists of other complicated objects. So, would I have trouble to transfer that list over network? If yes, how to manage it.
    Thank you!
    Astghik

    Astghik wrote:
    Hello All!
    I'm working on an enterprise application. Presentation layer is a stand alone client, business logic is build on the Glassfish v2.1 and MySQL is used as a database. The client is connection to the GlassFishj server remotely using EJBs.
    I have problems with business logic architecture.
    Here is the brief description of backend application architecture design:
    1. Session Facade pattern is used to simplify the client and application server interface and to provide application layers between backend (http://java.sun.com/blueprints/corej2eepatterns/Patterns/SessionFacade.html).
    I would simply recommend establishing a service tier. Your services should be stateless. You can go the extra mile and have a session facade, but in the majority of cases, coding to an interface for your service accomplishes the same goals.
    2.Transfer Object pattern to define update transfer objects strategy in order to decrease network overhead during client and application server interactions and to provide version control for objects. Transfer objects are designed as simple java business serializable objects. (http://java.sun.com/blueprints/corej2eepatterns/Patterns/TransferObject.html)
    The idea of the transfer object is very similar to the Command pattern. I think if you investigate that pattern, it will be more obvious. The transfer object reduces network latency by consolidating all the parameters into an object, ideally, this also consolidates multiple method calls. If you combine a transfer object (or command object) with a service tier, you get the best of both worlds. The service can delegate calls to helper objects (or other services or components) using the data in the transfer / command object.
    3. Originally the backend application consisted of three modules: users, storage and orders, but at the end I have decided to divide my application into the following parts - assortments, map, menu, orders, transactions, users.
    The is your domain. It will vary from application to application. The principles above are more general (e.g., patterns and architectural tiers) and should apply to most domains. However, your actual use case may require something different.
    4. All MySQL database transactions are via JDBC using procedures. No use of entity beans.
    Consider using something like iBatis or Spring's JDBC templating to make your life easier with JDBC.
    Questions:
    1. I have some doubts about using Session Facade and Transfer object patterns at the same time. At first I'd mike to cite the definitions of the patters from the SUN official web site.
    * Use a session bean as a facade to encapsulate the complexity of interactions between the business objects participating in a workflow. The Session Facade manages the business objects, and provides a uniform coarse-grained service access layer to clients.
    * Use a Transfer Object to encapsulate the business data. A single method call is used to send and retrieve the Transfer Object. When the client requests the enterprise bean for the business data, the enterprise bean can construct the Transfer Object, populate it with its attribute values, and pass it by value to the client.
    * So, if I use Transfer Object along with Session Facade, it makes some difficulties with object version control, because I 2 or
    3 transfer objects controls with the 1 bean class. The best option for Transfer object Pattern is that each transfer object should have its own bean class to provide ability of object's version control. In the case it can bring the network overhead because of frequent remote calls caused by the large number of the bean classes.
    * So, should I use the both patterns? If yes, how to manage the interaction the patterns. If no, which one to use.
    Versioning is a separate issue. Generally, the more coarsely grained your transfer / command object is, the more changes are likely to impact dependent objects.
    Your command or transfer object does not have to be a vanilla JavaBean, where you are basically creating a bean that has data from other objects. You can simply use your command / transfer object to encapsulate already existing domain objects. I see no need to map to a JavaBean with what you have described.
    Generally, a method signature should be understandable. This means that many times it is better to pass the method, say, two coarsely grained objects than a signature with a dozen primitives. There are no hard and fast rules here. If you find a method signature getting large, consider a transfer / command object. If you want one service to delegate calls to a number of other services, you can also create a transfer / command object to furnish the controlling service with the data it needs to invoke the dependent services.
    2. E.g. I have a huge list of the Order objects and each Order object consists of other complicated objects. So, would I have trouble to transfer that list over network? If yes, how to manage it.
    This is a large, open-ended question. If you are going to display it to a user on a screen, I do not see how you avoid a network transfer with the data. The general answer is to not pass the data itself but rather a token (such as a primary key, or a primary key and a start and stop range such as you see on a Google search result). You do want to limit the data over the network, but this comes at a cost. Usually, the database will receive additional load. Once that becomes unacceptable, you might start putting things into session. Then you worry about memory concerns, etc. There is no silver bullet to the problem. It depends on what issues you are trying to address and what trade-offs are acceptable in your environment.
    Thank you!
    AstghikBest of luck.
    - Saish

  • List of Modified and New objects between two consecutive TRs/ TOCs

    Hi gurus,
    My requirement is to create a Z program that will give me a report of the objects that have been modified or newly created between two Transport Requests or Two TOCs containing the objects.
    I had created the program, but facing these difficulties -
    1. While comparing the objects in the two TRs./ TOCs , I tried comparing the Program, Id and Object definition. but learned that, when we are creating an object say Report, then the object entry after putting in a TR (say TR1)is :
    Object Name | Pgm_Id | Object Def
    ZXXYY         | R3TR    | PROG
    Next, I modified this report and added a comment. Then I put this object in another TR say (TR 2) and released it. But, now when I check the entry of this object in Table E071 for the TR , I find this  -
    Object Name | Pgm_Id | Object Def
    ZXXYY         | LIMU    | REPS
    So, you see,  the object entry has changed when I modified it.
    So, when I go to compare the two TRs for new and modified objects, I have only the Name of the Object (ZXXYY) , to compare and state that this object has been modified.
    2. This difficulty arise out of the first -
        When I tried to compare objects with their Names only, then I found out that for Tables, (for eg.) , the modification of a Table field has the name of the table as object name, when put in to a TR. Same is the case while modifying a Table's attributes.
    Hence, If I modified a table's attribute, my program shows the table as modified instead of showing the tables attribute as modified.
    Kindly propose a solution to fix this issue.

    Hi Thomas, thanks a lot for your help.
    Solution: I checked first if the pgmid and object column entries matched . If not then I called the FM 'TR_CHECK_TYPE' and passed the details of the object in table E071, whose original pgmid and object entry is to be fund out. This returned me the original PGMID and OBJECT, which I compared with the entries of the object in the target TOC. If matched, I declared it as modified entry.

  • What is different between conflict object and lingering object in AD

    Hi
    I found lots of conflict object in my Forest on AD 2k8R2 with FFL & DFL 2k3 with dsquery filter "(name=*\0ACNF:*).
    But Only I can see 4 to 5 AD object in ADUC.
    How can I found others.
    Also these conflict object are replicated with AD replication? and due to these conflict object, can be any issue in AD replication like slow replication or delay replication?
    If I delete them then how can I delete and which command or tool like LDP?
    Also Is conflict AD Object and lingering objects means are same?
    Please suggest and guide..
    See below link of my conflict object ..
    http://sdrv.ms/19DUTbi

    IT means... If I don't find duplicate or conflict object in ADUC then I no need to worry about it as after AD tombstone lifetime period, these conflict object would be delete automatically ...?
    Conflict objects can be duplicate or lingering objects both & you need to worry. Consider an example, if you have a system named abc & you try to join another system with the same name from other site DC where connection b/w other DC's are down &
    ABC information is not present in that local DC, it will allow you to join system in the domain with the same name, then the object joined later vl have CNF added.
    Another case, two new systems are joined together from two different AD site with the same name, it will allow you to join & once replication occurs, you will find duplicate name with CNF.
    Same problem happens, when you connect old DC which has crossed TSL. WHat happens here is you have most recent DC with the deleted objects, when you connect old DC which has not received replication & crossed TSL, again during replication, you will see
    objects with CNF which will be lingering objects.
    So, CNF objects can be duplicate & lingering objects both & in both the case, i'll worry. Lingering objects are not visible directly, you can find out using repadmin tool. It is really complex to clean them if they exists in the ready only partition
    of the GC.
    http://technet.microsoft.com/en-us/library/cc738018%28v=ws.10%29.aspx
    Awinish Vishwakarma - MVP
    My Blog: awinish.wordpress.com
    Disclaimer This posting is provided AS-IS with no warranties/guarantees and confers no rights.

  • Renaming default Datacenter, Cluster, and Datastore object names in EVO:RAIL

    Hello,
    The Datacenter, Cluster, and vSAN Datastore objects in my newly-installed EVO:RAIL appliance all default to a label of "MARVIN-...". Is it safe to rename these labels appropriate to my environment, such as a geographic location for my Datacenter object? I suspect that it is okay because the underlying UID will not change, but am looking for confirmation from someone who has successfully done this in a lab or production environment and who has also been successful in applying EVO:RAIL patches afterward.
    Thanks in advance.

    Just to close the loop, I successfully renamed the datacenter, cluster, and vSAN objects in vCenter Server.
    I kept a close watch on Log Insight at each step, and did log one error that I found suspicious following the object rename of the vSAN datastore:
    2015-01-29T19:03:27.158Z <My 4th Host's FQDN> Hostd: [783C2B70 info 'DiskLib'] DISKLIB-LINK :DiskLinkGetSpaceUsedInfo: Failed to get the file system unique id for file vsan://118eb054-f123-469a-d542-XXXXXXXXXXXX, using id as "remote"
    Where in my case the 4th host happens to currently be the HA master. I'm guessing this was just a transient error as it hasn't reappeared in the last 30 minutes, but I'll keep an eye out for it. There was no corresponding error viewed in vCenter Server, and I seem to be able to refresh the vSAN storage counter without issue.
    Thanks again.

  • Table for Customizing and BASIS objects

    Hi all,
    Is there any table/view which stores a list of all customizing and BASIS objects created in the SAP system (something like table TADIR for programs)?
    I should be able to query against that table and find if a particular customizing objects exists or not.
    Regards,
    Saurabh

    Hi Saurabh,
    here in table SMODILOG...if u fatch data with condition...
    where ( OBJ_TYPE = 'CUSO' or
               OBJ_TYPE = 'ACGR' ) and
              OBJ_NAME <> Y*string or
              OBJ_NAME <>  z*string.
    This will get you the objects that you want.
    <b>Reward helpful answers with points,</b>
    Regards,
    Tejas

  • Daily activity help for script... and investgating some issue

    hi
    i am new to apps.i fil i need some script to solve some daily activity issues and solving some issue.
    can anyone help me or list of issue(s).I know it may different from site to site but i want very common list.so that issues could be resolved quickly.
    regards

    hsawwan wrote:
    Hi,
    invalid object lists and number.Query dba_objects for invalid objects and compile it either manually or via adadmin.yes. i know but i need to schedule this with scripts so that it will compile jar files,invld obj etc.
    >
    failing patches.Check the patch log file.
    diagonostics failed patches.Check the patch log file as well as the workers log file.where is the worker logfile?
    >
    backup of archivelog and database.Test the backup periodically.how?
    >
    gathering statistics for apps and dadatabase users.Schedule "Gather Schema Statistics" concurrent program on regular basis.option is dbms_job right?
    >
    specific concurrent manager/log lists for specific requests.Check the CM log file along with the failed request log file.i notice FND<sum number>.log usually created for this but for specific log file is needed to query from db.for this reason i want to use some scripts.
    >
    shrinking apps tablespaces to utilise the spaces and enhancing the performance of apps.Shrinking tablespaces does not enhance performance.
    may not enhance performance but may enhance the disk space.
    See (Note: 122669.1 - How to Perform a Health Check on the Database).
    monitoring alerts.
    monitoring forms and report and generating trace files.You can do from OAM.
    monitoring triggers Write an alert to notify you in case the status of the triggers changed.
    startup total apps and shutdowns total apps (this actually sysadmins of clients want)Usually, we review the startup/shutdown scripts log files.
    total cloning process within a night.You will have to review the log files.
    Thanks,
    Hussein

Maybe you are looking for