OPTIARC DVD RW AD-7170A issues with reading DL DVDs

Hi,
I just wanted to let you know, that after unsuccsesful lens cleanings I did the fw flash.
I have a MacPro3.1 on 10.5.8 OSX. I used the http://liggydee.cdfreaks.com/page/en/Optiarc-AD-7170A/ 102cbt_rpc1.bin bin file for flashing the drive. And flashed it with the http://binflash.cdfreaks.com/ Binflash utility within the terminal.
Here are the results:
BEFORE FLASH
OPTIARC DVD RW AD-7170A:
  Firmware Revision:    1.NC
  Interconnect:    ATAPI
  Burn Support:    Yes (Apple Shipping Drive)
  Cache:    2048 KB
  Reads DVD:    Yes
  CD-Write:    -R, -RW
  DVD-Write:    -R, -R DL, -RW, +R, +R DL, +RW
  Write Strategies:    CD-TAO, CD-SAO, CD-Raw, DVD-DAO
  Media:
  Type:    DVD+R DL
  ID:    MKM 003
  Blank:    No
  Erasable:    No
  Overwritable:    No
  Appendable:    Yes
  Write Speeds:    2.4x, 4x, 6x, 8x
AFTER FLASH
Optiarc DVD RW AD-7170A:
  Firmware Revision:    1.02
  Interconnect:    ATAPI
  Burn Support:    Yes (Generic Drive Support)
  Profile Path:    None
  Cache:    2048 KB
  Reads DVD:    Yes
  CD-Write:    -R, -RW
  DVD-Write:    -R, -R DL, -RAM, -RW, +R, +R DL, +RW
  Write Strategies:    CD-TAO, CD-SAO, CD-Raw, DVD-DAO
  Media:
  Type:    DVD+R DL
  ID:    MKM 003
  Blank:    No
  Erasable:    No
  Overwritable:    No
  Appendable:    No
Now everything works! Shame on the Apple, that users have to solve Apple's problems on their own...

I've been trying with the above mentioned tools to no avail. I keep getting a 'command not found' reply in terminal. I've double clicked on the Binflash utility and it listed the commands in terminal
Last login: Wed Nov  7 11:17:26 on ttys001
macpro-5:~ #########$ /Users/########/Downloads/necflash ; exit;
Binflash - NEC version - (C) by Liggy and Herrie
Visit http://binflash.cdfreaks.com
Usage:
necflash {-scan} [-v]
necflash {-dump} [-v] [-s] {filename} {device}
necflash {-flash} [-v] [-s] {filename} {device}
necflash {-identify} [-s] {device} {Filename}
necflash {-querybt} {device}
necflash {-setbt} {Mediatype} {Booktype} {Temp|Perm} {device}
necflash {-setbtrw} {Booktype} {Device}
necflash {-clearbt} {device}
necflash {-queryrpc} {device}
necflash {-resetrpc} {device} {Vendor|Global}
necflash {-version}
-scan     - Scan for supported devices
-dump     - Dump firmware and write to file
-flash    - Flash firmware and write to file
-identify - Identify unknown drives
-querybt  - Query booktype information
-setbt    - Change booktype settings
-setbtrw  - Change booktype on current DVD+RW disc
-clearbt  - Clear booktype settings and counter
-queryrpc - Query region status
-resetrpc - Reset region counter on RPC2 drives
-version  - Display version information
-v     - Verbose mode
-s     - Skip disclaimer
logout
[Process completed]
and said logout. I logged out and in again and tried the commands with the below results.
macpro-5:~ ####### necflash {-scan} [-v]
-bash: necflash: command not found
What am I doing wrong?
Cheers
Lars
MacPro 3,1
OPTIARC DVD RW AD-7170A:
  Firmware Revision:          1.NC
  Interconnect:          ATAPI
  Burn Support:          Yes (Apple Shipping Drive)
  Cache:          2048 KB
  Reads DVD:          Yes
  CD-Write:          -R, -RW
  DVD-Write:          -R, -R DL, -RW, +R, +R DL, +RW
  Write Strategies:          CD-TAO, CD-SAO, CD-Raw, DVD-DAO
  Media:          To show the available burn speeds, insert a disc and choose File > Refresh Information

Similar Messages

  • DVD/CD-RW drive won't read a DVD+R/DL disc

    Hi,
    Has anybody had problems with reading a DVD+R/DL disc on a Macbook with a MATSHITACD-RW CW-8221 revision GA0J drive. A client sent me a Fuji 4X (manufactured by Ritek) DVD+R/DL disc with about 5.43G bytes of data.
    I know the drive on my Macbook can only burn CDs, and that seems OK. It is supposed to be able to read any DVD format, though write any DVD format. Commercial DVDs and single layer DVDs seem to be OK, but the DL dvd just spits out after a few seconds. I get the same reaction in Windows vVsta on the Macbook, installed with Bootcamp.
    The same DVD+R/DL disc reads just fine on an Acer 3680 notebook with a cheap CD-RW/DVD pop-out style drive. Nero's utility prrogram, on the Acer, says the DVD+R/DL is OK, but notes it is not finalized - which it isn't supposed to be - as the client is planning to add more data in another session. Based on what I've read on OSTA.org's web site, the +R format (unlike the -R format) should be transportable between disc readers without needing to be finalized.
    I am glad I have an external hard disk drive to read the files on the Acer and move them to the Macbook.
    Has anybody else run into this and/or found a flash update for the Macbook's optical drive?
    Bill

    Have the same problem. I've spent last day to find the cause of this problem, but I still have no ideas how to make my MacBook C2D read DVD-R DL DVDs. I was trying to read blank Verbatim DVD-R DL 4x and Verbatim DVD-R 8x filled with data but both are still just spits out after 15-20 seconds. And during spinning the disc MacBook's optical drive emits a few beeps (very quiet but easy to hear).
    Sory for my english it's not my native language

  • Dynamic Link (Adobe Premiere Pro to Encore) issues with burning a DVD from the project file

    Hello,
    I am running Adobe Suite CS 5.1 on my Mac:
    2.8 GHz Quad Core
    6GB Memory
    I've been having issues with a few projects; using Dynamic Link, directly from my project in Premier, to send them to Encore:
    When attempting to burn a DVD, Encore will 'hang up' and not burn; I rarely burn to DVD, usually I'm always encoding for web, and just always use the 'default settings' in Encore.
    A few things to note:
    *In my project I'm using MOV files, from a Canon 7D, and MP4 files, from a Sony EX-1. 
    *I am working off of 3 separate hard drives; very large projects with over 5TBs of media files.
    *I noticed that after inserting a DVD, it caused one of my hard drives to eject; this happened once, as far as I know.
    *This only happens when I use Dynamic Link to send to Encore, via my Premiere Pro project file.
    *My Project settings are always: XDCam HD-1080p/24p
    Please help.
    Thanks

    So you must downrez at some point in the workflow. Some users have problems and/or think they get better quality out of direct PR/AME exports than using DL for this purpose.
    I would test the workflow by creating three short test sequences - one with a clip from one camera, a second from the other, and a third with both in a sequence using the same settings you are using in your edit. I would use the "drag a clip" to the new item method for making each of the first two sequences. By short I mean 30 seconds to 2 minutes with a mixture of still and action.
    Export using whatever method you pick of those recommended, but use CBR 8 instead of VBR etc. Bring the resulting files into Encore. Yes, look at the files exported on the computer, and look at Encore preview, etc, but also burn to a disk and watch on multiple TVs. There are lots of issues in best quality.

  • Terrible Quality issue with MPEG 2 (DVD- encode plz HELP!!!

    Hello,
    I edited something short in FCP 5 and wanted to make a DVD but the resulted mpeg-2 file has grabage all over. Not just square artefacts but a lot of problems like an old VHS maybe (but it isn't video noise either). It's more like a DV tape that would be ******.
    I tried to recompress in DV and the resulting quicktime .mov file is fine. But every time I encode it in any of the DVD presets it is terrible. Also tried MPEG-1. Same.
    As for the 2 pass encodes they crash my computer...
    Please help Obi-Wan Compressor you are my only hope.

    If you're running some flavor of Compressor 2, then try using the One Pass setting (but not the One-Pass VBR setting) at 6.5 Mbps. If you're still having issues, other things to try:
    up the One Pass bit rate to 7.5 Mbps
    Alter GOP to Open (from the default Closed)
    Alter GOP structure to IP (from the default IBBP)
    If I'm having issues with Compressor 2's Two Pass encodes, I can usually save my hide with a One Pass 7.5 Mbps Open IBBP GOP encode. If that doesn't work, Plan B is to do a One Pass 6.5 Mbps Closed IP GOP encode.
    Or are you using Compressor 1?

  • Issue with "read by other session" and a parallel MERGE query

    Hi everyone,
    we have run into an issue with a batch process updating a large table (12 million rows / a few GB, so it's not that large). The process is quite simple - load the 'increment' from a file into a working table (INCREMENT_TABLE) and apply it to the main table using a MERGE. The increment is rather small (usually less than 10k rows), but the MERGE runs for hours (literally) although the execution plan seems quite reasonable (can post it tomorrow, if needed).
    The first thing we've checked is AWR report, and we've noticed this:
    Top 5 Timed Foreground Events
    Event     Waits     Time(s)     Avg wait (ms)     % DB time     Wait Class
    DB CPU           10,086           43.82     
    read by other session     3,968,673     9,179     2     39.88     User I/O
    db file scattered read     1,058,889     2,307     2     10.02     User I/O
    db file sequential read     408,499     600     1     2.61     User I/O
    direct path read     132,430     459     3     1.99     User I/OSo obviously most of the time was consumed by "read by other session" wait event. There were no other queries running at the server, so in this case "other session" actually means "parallel processes" used to execute the same query. The main table (the one that's updated by the batch process) has "PARALLEL DEGREE 4" so Oracle spawns 4 processes.
    I'm not sure how to fix this. I've read a lot of details about "read by other session" but I'm not sure it's the root cause - in the end, when two processes read the same block, it's quite natural that only one does the physical I/O while the other waits. What really seems suspicious is the number of waits - 4 million waits means 4 million blocks, 8kB each. That's about 32GB - the table has about 4GB, and there are less than 10k rows updated. So 32 GB is a bit overkill (OK, there are indexes etc. but still, that's 8x the size of the table).
    So I'm thinking that the buffer cache is too small - one process reads the data into cache, then it's removed and read again. And again ...
    One of the recommendations I've read was to increase the PCTFREE, to eliminate 'hot blocks' - but wouldn't that make the problem even worse (more blocks to read and keep in the cache)? Or am I completely wrong?
    The database is 11gR2, the buffer cache is about 4GB. The storage is a SAN (but I don't think this is the bottleneck - according to the iostat results it performs much better in case of other batch jobs).

    OK, so a bit more details - we've managed to significantly decrease the estimated cost and runtime. All we had to do was a small change in the SQL - instead of
    MERGE /*+ parallel(D DEFAULT)*/ INTO T_NOTUNIFIED_CLIENT D /*+ append */
      USING (SELECT
          FROM TMP_SODW_BB) S
      ON (D.NCLIENT_KEY = S.NCLIENT_KEY AND D.CURRENT_RECORD = 'Y' AND S.DIFF_FLAG IN ('U', 'D'))
      ...(which is the query listed above) we have done this
    MERGE /*+ parallel(D DEFAULT)*/ INTO T_NOTUNIFIED_CLIENT D /*+ append */
      USING (SELECT
          FROM TMP_SODW_BB AND DIFF_FLAG IN ('U', 'D')) S
      ON (D.NCLIENT_KEY = S.NCLIENT_KEY AND D.CURRENT_RECORD = 'Y')
      ...i.e. we have moved the condition from the MERGE ON clause to the SELECT. And suddenly, the execution plan is this
    OPERATION                           OBJECT_NAME             OPTIONS             COST
    MERGE STATEMENT                                                                 239
      MERGE                             T_NOTUNIFIED_CLIENT
        PX COORDINATOR
          PX SEND                       :TQ10000                QC (RANDOM)         239
            VIEW
              NESTED LOOPS                                      OUTER               239
                PX BLOCK                                        ITERATOR
                  TABLE ACCESS          TMP_SODW_BB             FULL                2
                    Filter Predicates
                      OR
                        DIFF_FLAG='D'
                        DIFF_FLAG='U'
                  TABLE ACCESS          T_NOTUNIFIED_CLIENT       BY INDEX ROWID    3
                    INDEX               AK_UQ_NOTUNIF_T_NOTUNI    RANGE SCAN        2
                      Access Predicates
                        AND
                          D.NCLIENT_KEY(+)=NCLIENT_KEY
                          D.CURRENT_RECORD(+)='Y'
                      Filter Predicates
                        D.CURRENT_RECORD(+)='Y' Yes, I know the queries are not exactly the same - but we can fix that. The point is that the TMP_SODW_BB table contains 1639 rows in total, and 284 of them match the moved 'IN' condition. Even if we remove the condition altogether (i.e. 1639 rows have to be merged), the execution plan does not change (the cost increases to about 1300, which is proportional to the number of rows).
    But with the original IN condition (that turns into an OR combination of predicates) in the MERGE ON clausule, the cost suddenly skyrockets to 990.000 and it's damn slow. It seems like a problem with cost estimation, because once we remove one of the values (so there's only one value in the IN clausule), it works fine again. So I guess it's a planner/estimator issue ...

  • Issue with reading emails state.

    Dear community,
    in my company, we have BB to bring a very good 24hrs service to our customer, but the thing is that we started to have an issue with our BB and the Microsoft Office 365 and Exchange service.
    We have BlackBerry Enterprise service activated on each blackberry, so we receive every email in our BBs.
    But, when we read an email, after you go back to the mailbox, it appears as un-read again. Then, if you read it again, and go back to the mailbox, it appears as read.
    DO you know why it's happening?.
    Thanks in advance,

    Hi,
    Use TRY & CATCH statements before OPEN DATASET...
    For exact syntax check F1 help on try & catch...
    TRY.
      CATCH.
        OPEN DATASET..
      endcatch.
    endtry.
    Hope it helps!!
    Rgds,
    Pavan

  • How can I download my issues with Reader's Digest when each time a pop appears and reads " the item you tried to buy is no longer available"?

    I can no longer download my purchased Reader's Digest magazine since August as it requires updating the RD app. So when i try to download the RD app a pop up appears and reads " the item you tried to buy is no longer available". What does this mean and what do I do? I have 2 pending reader's digest issues that needs to be downloaded.

    I don't have an answer for you but it may help to have a look at the "More Like This" section at the far right of this page. Reading those additional threads may get you some idea of the actual issue and how to resolve it. It could be that problems were found with the app and it was temporarily pulled for some reason which may be the reason for the "the item you tried to buy is no longer available" message.

  • Mac / Adobe Compatibility Issues with Reader X1 & Acrobat X1

    I'm using Acrobat X1  11.0.10 and Reader 11.0.10 with Mac OS X ( 10.8.5).
    In Reader, when I click the "show next page" down arrow, it takes around 5 seconds to jump to the next page. (scrolling to the next page via the scroll bar works fine).
    In Acrobat, when I use the text edit tools, it can take up to 10 seconds before I can edit a word (remove or add a letter).Oftentimes, Acrobat will highlight the whole paragraph rather than just the word I select. (I think this is due to the lag time.) When I backspace to remove a word, nothing happens ... for several seconds. Then letters begin disappearing.
    I reverted from Mavericks and Yosemite back to Lion (or Mountain Lion - I forget what they called 10.8.5) in the hopes of fixing this issue. When I say reverted back. I did a full install. Wiped the hard drive, loaded 10.8.5 and started this Mac as if it were new.
    Is there a version of Reader and Acrobat that work with Mac OS X ( 10.8.5)?

    Hi,
    Acrobat 11 is compatible with MAC 10.8.5. So, there are no compatibility issues.
    Please let me know if this happens with every PDF or some specific one?
    If specific, then please tell me what is the size of your PDF file.
    You might also try repairing Acrobat from the Help menu.
    Regards,
    Anubha

  • Issues with exporting to DVD

    I have a collection that I want to burn to a DVD. I have also chosen a custom naming sequence, that begins with the sequence number 1. The files are output to a folder, although the sequence doesn't start with the number 1. It goes sequence.jpg, then sequence-2.jpg, etc. More importantly, when the files are burned to the DVD, they retain their original file name instead of the new sequence name. I've wasted several DVD's trying to sort this out. Any suggestions?
    Ken D

    I have tried dragging. I am pretty sure the problem lies in the Ken Burns effect being on when I imported the pics. I'm still looking into it, but it seemsthat it has something to do with the fact that the original file is put in the trash. I have imported still clips as an experiment and the trasitions work fine. The reason I imported the stills with the "ken burns" on is because I read that it increases the quality when the movie is imported into IDVD since it is now treated as a video clip anot a still. The Ken burns effect 's movement isn't really used in my movie. I have the start and end set with the same specs.

  • Modbus - Issues with Reading Holding Registers

    I am trying to read the holding registers on a RS485 2 wire modbus power monitor (Veris Industries E50C2).I am receiving bytes from the device so there appears to be communication. However, the response string is the exact same string as my command string. For instance, I am sending the hexadecimal string 0103 0000 0005 in order to read the first 5 holding registers, but I am receiving this exact same string back via the read VISA read function. I have attached an image of my VI. I would expect to see the response as "0103 (the device ID and command) followed by the number of bytes to follow and the contents of the holding registers.
    The VI is running on a cRIO 9012 with 9111 chassis. I am using the 9871 c-series RS485/RS422 interface module.
    Any ideas what the issue may be?
    Kind Regards
    Adam
    Attachments:
    Modbus VI.png ‏15 KB

    I indeed manage to fix the issue for myself.
    I used the same method you are starting right now, i conencted the NI9871 to the rs485 adapter given by the supplier and tried sending and reading my own signals.
    After much hassle and some changes to your (and my) initial vi setup as shown in your first post, everything worked out fine:
    (see the attached picture)
    1. I used the NI modbus VIs, but it worked as well with the normal visa commands and building the modbus command on your own, but  way easier with the NI VIs.
    2. i added the function to resize the buffer to its maximum size (not sure if thats actually helping with anything, but better be safe then sorry).
    3. Change the wire mode for the 9871 to rs485/wire2-auto with a property node.
    4. remove the large delay between write and read
    5. Last but not least, the one thing that changed the fact that i only received my own echo:
    I changed the connection of the Rx+/Tx+ and Rx-/Tx- around, basically connecting the 2wire sensor the other way around the manufacturer said it should be connected to the master device.
    (6. Recheck the amount of stop bits, all modbus device i use in my lab are using 2 stop bits, but that could be a coincidence)
    Maybe  something of that list helps you
    Attachments:
    Modbustest01.jpg ‏95 KB

  • Compusa brand dvd-r, anyone having issues with them?

    Last September I bought a 50 pack of 2x DVD-r disks, Compusa brand. I've been trying to burn some disk images with my new macbook pro, which I created with idvd. I'm not having much luck with them. I get around one good burn after going through three or for disks.
    Anyone else have poor results with their disks? Could someone recommend good quality disks? It's a 4x drive.

    I use Memorex 8x single layer discs, Part # 3202 5601 - they burn fast and all have verified 100%

  • Incompatibility issues with Dual Layer DVD+R  help?

    I think I've made a serious mistake in believing I could author a Dual layer dvd for replication and sale on my own. I've spent over 14 months on this film, which is working perfectly in Encore, complete with extensive menus and transitional movies.
    I've done my best to transcode the media in the most compatible way (keeping bitrate below 7mbps and using dolby digital audio format) BUT I'm still having bad luck with MOST dvd players.
    The DVD plays properly on my cheap sony player, but both my toshiba and humax/tivo players failed to read the disc. (multiple tries/burns....or coasters, now) I've read that there are common problems with dual layer +R.
    MY QUESTION IS: what is the BEST POSSIBLE WAY to author/burn this project for mass replication? Can I really achieve the result I'm looking for with Encore or will I have to give up and go to a professional authoring house with DLT capability?
    Will I get the best result by burning from encore straight to disc OR burning an image or folder, then using a third party burning program?
    Thanks SO MUCH for any responses. This process has been soul-crushing.

    >The DVD plays properly on my cheap sony player, but both my toshiba and humax/tivo players failed to read the disc. (multiple tries/burns....or coasters, now) I've read that there are common problems with dual layer +R.
    Marc, the only mistake I think you are making is in assuming that set top players can handle a written DVD+R DL disc.
    Support in older players is at best sketchy.
    If the disc is working properly in at least one player, then you should be good to go.
    Replication will produce a different type of disc. One that
    i will
    play properly.
    Not all burners support book typing, and unless you know what you're doing here do not use these "fudges". It's best using Plextor, and even Benq burners. The freeware utility IMGBurn has this ability, and it also tells you what burners do this. I've never needed it personally, as my Pioneer burners & Denon players have no trouble with +R DL media.
    What you really need is a player that you
    i know
    plays DVD+R DL discs properly. Once you know your player is good with the format, then all other inconsistencies are player related, not authoring, and not your problem. There is simply no way at all you can guarantee any form of compliance with +R DL media, as there is nothing in the specs that says any set top player has to read these discs. Or -R DL where support is generally even worse.
    Whilst most players will read -R single layer & +R single layer, again there is no mandatory requirement to do so.
    You will need to get these
    i replicated
    and not duplicated. Duplication (burning) will result in a lot of returns, and a lot of wasted money.

  • Issue with reading a xml file from xsl

    Hi,
    When I am trying to read a xml file from xsl, I am getting unwanted output.
    Following is the XSL:
    <?xml version="1.0" encoding="UTF-8" ?>
    <?oracle-xsl-mapper
      <!-- SPECIFICATION OF MAP SOURCES AND TARGETS, DO NOT MODIFY. -->
      <mapSources>
        <source type="XSD">
          <schema location="../xsd/B2BMarketProperties.xsd"/>
          <rootElement name="ReceipentIDType" namespace="http://www.example.org"/>
        </source>
      </mapSources>
      <mapTargets>
        <target type="XSD">
          <schema location="../xsd/B2BMarketProperties.xsd"/>
          <rootElement name="ReceipentIDType" namespace="http://www.example.org"/>
        </target>
      </mapTargets>
      <!-- GENERATED BY ORACLE XSL MAPPER 11.1.1.4.0(build 110106.1932.5682) AT [TUE DEC 03 16:06:03 EST 2013]. -->
    ?>
    <xsl:stylesheet version="1.0"
                    xmlns:bpws="http://schemas.xmlsoap.org/ws/2003/03/business-process/"
                    xmlns:xp20="http://www.oracle.com/XSL/Transform/java/oracle.tip.pc.services.functions.Xpath20"
                    xmlns:mhdr="http://www.oracle.com/XSL/Transform/java/oracle.tip.mediator.service.common.functions.MediatorExtnFunction"
                    xmlns:bpel="http://docs.oasis-open.org/wsbpel/2.0/process/executable"
                    xmlns:oraext="http://www.oracle.com/XSL/Transform/java/oracle.tip.pc.services.functions.ExtFunc"
                    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
                    xmlns:ns0="http://www.example.org"
                    xmlns:dvm="http://www.oracle.com/XSL/Transform/java/oracle.tip.dvm.LookupValue"
                    xmlns:hwf="http://xmlns.oracle.com/bpel/workflow/xpath"
                    xmlns:xsl="http://www.w3.org/1999/XSL/Transform"
                    xmlns:med="http://schemas.oracle.com/mediator/xpath"
                    xmlns:ids="http://xmlns.oracle.com/bpel/services/IdentityService/xpath"
                    xmlns:bpm="http://xmlns.oracle.com/bpmn20/extensions"
                    xmlns:xdk="http://schemas.oracle.com/bpel/extension/xpath/function/xdk"
                    xmlns:xref="http://www.oracle.com/XSL/Transform/java/oracle.tip.xref.xpath.XRefXPathFunctions"
                    xmlns:xsd="http://www.w3.org/2001/XMLSchema"
                    xmlns:ora="http://schemas.oracle.com/xpath/extension"
                    xmlns:socket="http://www.oracle.com/XSL/Transform/java/oracle.tip.adapter.socket.ProtocolTranslator"
                    xmlns:ldap="http://schemas.oracle.com/xpath/extension/ldap"
                    exclude-result-prefixes="xsi xsl ns0 xsd bpws xp20 mhdr bpel oraext dvm hwf med ids bpm xdk xref ora socket ldap">
      <xsl:output method="xml" version="1.0" encoding="UTF-8" indent="yes"/>
      <xsl:variable name="ReceipentID" select="document('../xsd/B2BMarketProperties.xml')"/>
      <xsl:template match="/">
        <ns0:ReceipentIDType>
        <xsl:for-each select="$ReceipentID">
          <ns0:ReceipentID>
            <xsl:value-of select="$ReceipentID"/>
          </ns0:ReceipentID>
          </xsl:for-each>
        </ns0:ReceipentIDType>
      </xsl:template>
    </xsl:stylesheet>
    Following is the XML ( B2BMarketProperties.xml)
    <?xml version="1.0" encoding="UTF-8" ?>
    <ReceipentIDType xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
                     xsi:schemaLocation="http://www.example.org B2BMarketProperties.xsd"
                     xmlns="http://www.example.org">
      <ReceipentID>123</ReceipentID>
      <ReceipentID>345</ReceipentID>
    </ReceipentIDType>
    The output i am getting with this code is
    <?xml version="1.0" encoding="UTF-8"?>
    <ns0:ReceipentIDType xmlns:ns0="http://www.example.org">
        <ns0:ReceipentID>123345</ns0:ReceipentID>
    </ns0:ReceipentIDType>
    But, I need output in the following format
    <ns0:ReceipentIDType xmlns:ns0="http://www.example.org">
        <ns0:ReceipentID>123</ns0:ReceipentID>
         <ns0:ReceipentID>345</ns0:ReceipentID>
    </ns0:ReceipentIDType>
    Could you guys let me know what i am doing wrong. Any help would be appreciated.
    Thanks,

    This worked for me :
      <xsl:template match="/">
        <ns0:ReceipentIDType>
          <xsl:for-each select="document('B2BMarketProperties.xml')/*:ReceipentIDType/*:ReceipentID">
            <xsl:variable name="count" select="position()"/>
            <ns0:ReceipentID>
              <xsl:value-of select="document('B2BMarketProperties.xml')/*:ReceipentIDType/*:ReceipentID[$count]"/>
            </ns0:ReceipentID>
          </xsl:for-each>
        </ns0:ReceipentIDType>
      </xsl:template>

  • Issue with reading attachment in sender mail adapter

    Hi, I am working on Email to file scenario, where I have to connect to exchange server and take the attachment of the email (xml file) and save it on to PI server. I have configured sender mail communication channel with following details:
    Protocol: POP3
    Message protocol: XIPAYLOAD
    Mail attributes: keep attachments
    Module:
    1     localejbs/AF_Modules/PayloadSwapBean     Local Enterprise Bean     1
    2     localejbs/sap.com/com.sap.aii.adapter.mail.app/XIMailAdapterBean     Local Enterprise Bean     mail
    1     swap.keyName     Content-Description
    1     swap.keyName     Content-Description
    1     swap.keyValue     attachment,filename="MailAttachment-1.xml"
    1     swap.keyValue     MailAttachment-1
    Also, configured receiver file adapter.
    Since I donu2019t have to read mail content , I havenu2019t created any design components in ESR. I have just refrred to dummy namespace, interface (both sender and receiver). in receiver determination, interface determination, sender agreement and receiver agreements.
    It was working fine, not sure what changes happened. Now I could see the attachment payload in message monitor.
    Under Inbound message->payloads->
    MailMessage ( text/xml;charset=utf-8 )
    MailAttachment-1 ( text/xml;name="****************")
    But the fille that is getting created doesnu2019t have the contents of the attachment (xml file) .  It has some other information, shown below:
    Directory:  ******
    Name:       ********.xml
          =NextPart_002_01CB2C5B.B033B325#
    content-type:text/plain;charset="us-ascii"#
    content-length:30#
    content-transfer-encoding:quoted-printable#
    <<**********.xml>>=20#
          =NextPart_002_01CB2C5B.B033B325#
    content-type:text/html;charset="us-ascii"#
    content-length:#
    content-transfer-encoding:quoted-printable#
          =NextPart_002_01CB2C5B.B033B325  #
    Can you please let me know what is going wrong here.

    using following parameters in module resolved my problem:
    MOdule name                                                                           Type                                                            Module Key
    AF_Modules/PayloadSwapBean                                            Local Enterprise Bean                                      transform sap.com/com.sap.aii.adapter.mail.app/XIMailAdapterBean     Local Enterprise Bean                                      mail
    Module Key                                    Parameter Name                              Parameter value
    transform                                        swap.keyName                                Payload-Name
    transform                                        swap.keyName                                 MailAttachment-1

  • Issue with read statement with one more key missing in mapping

    Hi All ,
    I have such data in two internals table :
    IT_bdc
    vbeln            posnr
    90000593     10
    90000576     10
    90000672     10
    90000672     20
    90000672     30
    it_konv
    kbetr          vbeln
    6250          90000576
    12160000          90000593
    500000          90000672
    600000          90000672
    700000          90000672
    My current program statement is :
    LOOP AT it_bdc.
    READ TABLE it_konv WITH KEY
          vbeln = it_bdocs-vbeln.
      currency =   it_konv-waers.
    endloop.
    as you can see the posnr is missing in it_konv how can i modify this read statement so
    that vbeln posnr from it_bdc should get correct kbetr from it_konv.
    Kindly help in this mapping.

    Hi
    sort it_konv by vbeln
    then
    loop at it_bdc.
    read table it_konv with key vbeln = it_bdc-vbeln binary search.
    if sy-subrc = 0.
    perform your logic/task.
    endif.
    endloop.
    also it depends what you want to do after reading it_konv.
    in my logic if there is a vbeln in it_konv which s present in it_bdc then sy-subrc will be 0
    and you can perform your logic.
    and if there will be no matching vbeln in it_konv then sy-subrc will not be 0.
    check the values in debugging.
    Thanks
    Lalit

Maybe you are looking for