How to use md5 Checksums to verify thousands of copied files

Hello,
So I need to be able to verify the succesful copy of thousands of files using md5 checksums.
is there a program for this or a string of code for terminal?
Please help.
Cheers.

I cover how to do that in the link above, but to quickly do this, run the following command on both folder paths:
find -s ~/Downloads -type f -exec md5 -q {} \; | md5
This will generate a list of md5 checksums for each file in the directory (in this case, the Downloads directory in your home folder, but you can change this to whatever you would like), and then create a summary checksum of the collective checksums.
In my case, the above command yields the following checksum:
42cf9632cf763cad84c633f0cc7c93bb
Doing this on my Documents folder yields the following checksum:
782845544f0ecf8a7ae5fa66197d6abd
You can then compare these to see if they are different. In this case they are, indicating the folders do not have the same contents.
Keep in mind this may take a while to run, as checksums will have to be computed for each file in the directory. While this is a raw command, you can save it in a script and then simply run the script with an input argument to make it easier. For example, the following script should work:
#!/bin/bash
find -s $1 -type f -exec md5 -q {} \; | md5
Saving this in a text file, naming it something like "dirsum" and then making it executable in a path location such sa /bin (or a custom added path) will make it easily accessible, so you can run the same command by simply typing the following in the Terminal (in this case to calculate a checksum of the Downloads folder):
dirsum ~/Downloads

Similar Messages

  • How to use TCP Checksum Off-load in my network driver

    I am seeking infomation on how to use TCP Checksum Off-load to tell the Solaris 8 stack not to handle it for both receives and transmits (and let our NIC handle this compute-intensive task instead).
    There is mention that Sun GigabitEthernet Adapter cards, Release 1.1 supports Sol-2.6 use of NIC features: TCP checksum offload & byte swapping
    There are limited bits of info in include files. For example
    sys/strick.h indicates use of:
    inetcksum_t's ick_xmit
    stream.h has:
    struct datab's unsigned u16:16; /* used to store hw-calculated cksum
    #define STRUIO_IP 0x04 /* IP checksum stored in db_struioun */
    #define STRUIO_ZC 0x08 /* mblk eligible for zero-copy */
    strsub.h has:
    struct zero_copy_kstat
    Also /include/inet/tcp.h has tcp_sum for outbound.
    Thank you. -Weimin Tchen

    VenK7337,
    Could you show your python code. so we know what your are "writing" to the ethernet port?
    That way we can see what you are receiving.
    Parsing the incoming data (from the TCP-read) depends heavily on the device that sends it, and can not generically be described. LabVIEW has many byte (and even bit) manipulation functions to convert many different data formats to its own build in formats.
    So after the TCP listener is connected, you are constantly reading from the established connection (until it gets broken of course). More advanced example would be the internet toolkit if oyu have it.
    From the read characters (and I hope you designed a protocol with a clear starting character, ending character and maybe even a build in checksum) you parse the data and perform you action, and of course generate a reply. Again the internet toolkit is a good example. It parsed the input as it comes it, based on the HTTP format. Then generates the reply based on the request received.
    These days I would suggest not to use binairy encoded numerics. Try and use XML formatted data. Yes, it causes a lot of overhead. But typically this is not an issue and makes the code a lot more portable and maintainable. Also makes it easier to interface with other languages/platforms.
    Umless of course you are looking at Khz data rates, then XML is not th preferred choice.
    Hope this helps...

  • How to make CUCM as a TFTP server , then copy files to Voice Gateway ?

    how to make CUCM as a TFTP server , then copy files to Voice Gateway ? anyone knows?

    Hi,
    Please check the following link
    http://www.cisco.com/c/en/us/td/docs/voice_ip_comm/cucm/admin/6_1_1/ccmsys/cucm-accm-611/a02tftp.html#wp1023004
    Understanding How Devices Access the TFTP Server
    You can enable the IP phones and gateways to discover the TFTP server IP address in one or more of the following ways, depending on the device type:
    •Gateways and phones can use DHCP custom option 150.
    Cisco recommends this method. With this method, you configure the TFTP server IP address as the option value.
    •Gateways and phones can use DHCP option 066.
    You may configure either the host name or IP address of the TFTP server as the option value.
    •Gateways and phones can query CiscoCM1.
    Ensure the Domain Name System (DNS) can resolve this name to the IP address of the TFTP server. Cisco does not recommend this option because it does not scale.
    •You can configure phones with the IP address of the TFTP server. If DHCP is enabled on the phone, you can still configure an alternate TFTP server IP address locally on the phone that will override the TFTP address that was obtained through DHCP.
    •Gateways and phones also accept the DHCP Optional Server Name (sname) parameter.
    •The phone or gateway can use the value of Next-Server in the boot processes (siaddr).
    Devices save the TFTP server address in nonvolatile memory. If one of the preceding methods was available at least once, but is not currently available, the device uses the address that is saved in memory.
    You can configure the TFTP service on the first node or a subsequent node, but usually you should configure it on the first node. For small systems, the TFTP server can coexist with a Cisco Unified Communications Manager on the same server.
    HTH
    Manish

  • How to do md5 checksum in PL/SQL in 8i

    Is there any function in PL/SQL for md5 checksum?

    Oracle8i Enterprise Edition Release 8.1.7.3.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.3.0 - Production
    SQL> SET SERVEROUTPUT ON;
    SQL> BEGIN
      2    DBMS_OUTPUT.PUT_LINE (
      3      'md5 hash => ' ||
      4      DBMS_OBFUSCATION_TOOLKIT.MD5 (
      5     input => UTL_RAW.CAST_TO_RAW ('plain text')));
      6  END;
      7  /
    md5 hash => 31BC5C2B8FD4F20CD747347B7504A385
    PL/SQL procedure successfully completed.
    SQL>

  • How to use old archieve logs with a new control file

    Environment:
    ORACLE_BASE="/export/home/oracle"
    ORACLE_HOME="/export/home/oracle/product/8.1.6"
    NLS_LANG=".UTF8"
    2 partitions:
    i) /data1 -- contains important datafiles (OS striping on 3 hard
    disks)
    ii) /export/home -- contains the oracle program, and SYS/SYSTEM
    datafiles
    Problem:
    -/data1 cannot be read/mount (damaged)
    -oracle failed
    Action Performed:
    -reinstalled OS
    -mount /export/home successfully (all oracle system files,
    instance init files exist)
    -/data1 is an empty partition
    -created the oracle user, and its groups
    -chown recusively for the $ORACLE_BASE directory
    -set all the oracle environment variables
    -attempted to start the instance, but failed due to a control
    file was
    missing
    -since control files were set to be mirrored, i copied a control
    file somewhere
    from /export/home/oracle/oradata/<SID>/control1.ctl to
    /data1/oracle/oradata/<SID>/control2.ctl (i.e. have them back to
    their original locations)
    -the instance failed to start as well, since the datafile set in
    the control files couldn't be found
    -this forced me to re-create the control file... b4 i re-create
    a new control file, i backed up the old one
    -once the control file was created, the database can be started,
    but, to oracle, all achieve log information are lost (although
    the achieve log's physical files r still there)
    -i imported the important data from a dump file that was
    exported a week ago b4 the system failure
    -since we r using the new control file, the redo logs switch
    contains no achieve log information that the old control file
    has, so running "recover database" doesn't do anything
    Purpose:
    since the dump file is a week old, i'd like to get the data
    after my last export and b4 the system failure. the database was
    run in arhieve log mode, how can i recover those data with the
    new control file
    Question:
    -how can we create a new control file that can drive the old
    achieve logs?
    -can we convert the achieve log data (.dbf format) into text
    format?
    -can we still use the old control files to start the database?
    -what's a suggested solution if we'd like to re-construct the
    database up to the moment b4 the system failure in another
    server?
    thx

    user3930585 wrote:
    I am in an unenviable position, with an unsupported database.
    We are running Oracle 9i on Windows XP. We are upgrading soon to Oracle 11g on a newer platform, but need to get our development environment working first.
    We lost a system that was running our development database without having a database export. The C drive was placed into a new system as the D drive.
    I have loaded Oracle 9i on the C drive, but I have been unable to determine how to point it to the existing data files on the D drive. My search skills may be the limiting factor here...
    We cannot simply load the drive as C, since the hardware is different.
    What are the steps to point the new database software at the data files on the D drive? Or, how do I copy the old data files into the new Oracle Home and have them recognized properly?
    Are you stating that you don't know how to use COPY command?
    Can you recreate same directory structure on new C drive as exist on old C drive?
    Can you then drag & drop copies of the files?

  • Since upgrading to OSX 10.6.8, I am unable to use scp on a remote machine to copy files into my iMac.

    Any suggestions?
    However, I can use scp to copy files out from my iMac (10.6.8) .
    I can ssh into the iMac OK.
    I created a .profile in my iMac home directory with the single line "export TERM=xterm-color".
    This makes a difference to how scp behaves when attempting to copy into the iMac, but it still fails.
    Thanks in advance for any help.  Peter R
    Here is the scp debug output (on the remote machine that I'm copying a file from, to the iMac):
    ===
    Executing: program /usr/bin/ssh host molika.ucsd.edu, user prowat, command scp -v -t Desktop/FromPelican/Outs201201.pdf
    Sun_SSH_1.1.3, SSH protocols 1.5/2.0, OpenSSL 0x0090704f
    debug1: Reading configuration data /etc/ssh/ssh_config
    debug1: Rhosts Authentication disabled, originating port will not be trusted.
    debug1: ssh_connect: needpriv 0
    debug1: Connecting to molika.ucsd.edu [137.110.243.133] port 22.
    debug1: Connection established.
    debug1: identity file /home/nona/.ssh/identity type -1
    debug1: identity file /home/nona/.ssh/id_rsa type -1
    debug1: identity file /home/nona/.ssh/id_dsa type -1
    debug1: Remote protocol version 1.99, remote software version OpenSSH_5.2
    debug1: match: OpenSSH_5.2 pat OpenSSH*
    debug1: Enabling compatibility mode for protocol 2.0
    debug1: Local version string SSH-2.0-Sun_SSH_1.1.3
    debug1: use_engine is 'yes'
    debug1: pkcs11 engine initialized, now setting it as default for RSA, DSA, and symmetric ciphers
    debug1: pkcs11 engine initialization complete
    debug1: Failed to acquire GSS-API credentials for any mechanisms (No credentials were supplied, or the credentials were unavailable or inaccessible
    Unknown code 0
    debug1: SSH2_MSG_KEXINIT sent
    debug1: SSH2_MSG_KEXINIT received
    debug1: kex: server->client aes128-ctr hmac-md5 none
    debug1: kex: client->server aes128-ctr hmac-md5 none
    debug1: Peer sent proposed langtags, ctos:
    debug1: Peer sent proposed langtags, stoc:
    debug1: We proposed langtags, ctos: i-default
    debug1: We proposed langtags, stoc: i-default
    debug1: SSH2_MSG_KEX_DH_GEX_REQUEST sent
    debug1: expecting SSH2_MSG_KEX_DH_GEX_GROUP
    debug1: dh_gen_key: priv key bits set: 124/256
    debug1: bits set: 1007/2048
    debug1: SSH2_MSG_KEX_DH_GEX_INIT sent
    debug1: expecting SSH2_MSG_KEX_DH_GEX_REPLY
    debug1: Host 'molika.ucsd.edu' is known and matches the RSA host key.
    debug1: Found key in /home/nona/.ssh/known_hosts:1
    debug1: bits set: 1043/2048
    debug1: ssh_rsa_verify: signature correct
    debug1: newkeys: mode 1
    debug1: set_newkeys: setting new keys for 'out' mode
    debug1: SSH2_MSG_NEWKEYS sent
    debug1: expecting SSH2_MSG_NEWKEYS
    debug1: newkeys: mode 0
    debug1: set_newkeys: setting new keys for 'in' mode
    debug1: SSH2_MSG_NEWKEYS received
    debug1: done: ssh_kex2.
    debug1: send SSH2_MSG_SERVICE_REQUEST
    debug1: got SSH2_MSG_SERVICE_ACCEPT
    debug1: Authentications that can continue: publickey,keyboard-interactive
    debug1: Next authentication method: publickey
    debug1: Trying private key: /home/nona/.ssh/identity
    debug1: Trying private key: /home/nona/.ssh/id_rsa
    debug1: Trying private key: /home/nona/.ssh/id_dsa
    debug1: Next authentication method: keyboard-interactive
    *** Password requested & supplied
    debug1: Authentication succeeded (keyboard-interactive)
    debug1: fd 5 setting O_NONBLOCK
    debug1: fd 6 setting O_NONBLOCK
    debug1: fd 7 setting O_NONBLOCK
    debug1: channel 0: new [client-session]
    debug1: send channel open 0
    debug1: Entering interactive session.
    debug1: ssh_session2_setup: id 0
    debug1: channel request 0: env
    debug1: Sending command: scp -v -t Desktop/FromPelican/Outs201201.pdf
    debug1: channel request 0: exec
    debug1: channel 0: open confirm rwindow 0 rmax 32768
    ~
    debug1: channel 0: read<=0 rfd 5 len 0
    debug1: channel 0: read failed
    debug1: channel 0: close_read
    debug1: channel 0: input open -> drain
    debug1: channel 0: ibuf empty
    debug1: channel 0: send eof
    debug1: channel 0: input drain -> closed
    debug1: channel 0: write failed
    debug1: channel 0: close_write
    debug1: channel 0: output open -> closed
    debug1: channel 0: rcvd eof
    debug1: client_input_channel_req: channel 0 rtype exit-status reply 0
    debug1: channel 0: rcvd close
    debug1: channel 0: almost dead
    debug1: channel 0: gc: notify user
    debug1: channel 0: gc: user detached
    debug1: channel 0: send close
    debug1: channel 0: is dead
    debug1: channel 0: garbage collecting
    debug1: channel_free: channel 0: client-session, nchannels 1
    debug1: fd 0 clearing O_NONBLOCK
    debug1: fd 1 clearing O_NONBLOCK
    debug1: fd 2 clearing O_NONBLOCK
    debug1: Transferred: stdin 0, stdout 0, stderr 0 bytes in 0.9 seconds
    debug1: Bytes per second: stdin 0.0, stdout 0.0, stderr 0.0
    debug1: Exit status 0

    On looking at the secure.log on the iMac with OSX10.6.8, I see these entries at each attempt to scp into the iMac:
    Dec 29 11:06:20 molika sshd[7248]: in pam_sm_authenticate(): Failed to determine Kerberos principal name.
    Dec 29 11:06:26 molika sshd[7246]: Accepted keyboard-interactive/pam for prowat from 137.110.243.132 port 35571 ssh2
    Dec 29 11:06:26 molika com.apple.SecurityServer[26]: Session 0x236a14 created
    Dec 29 11:06:26 molika com.apple.SecurityServer[26]: Session 0x236a14 attributes 0x20
    Dec 29 11:06:27 molika com.apple.SecurityServer[26]: Session 0x236a14 dead
    Dec 29 11:06:27 molika com.apple.SecurityServer[26]: Killing auth hosts
    Dec 29 11:06:27 molika com.apple.SecurityServer[26]: Session 0x236a14 destroyed
    I don't know what to make of this.
    Peter R

  • How to use MD5 format of string

    i want to change a password string to md5 format
    how can i do it

    http://www.twmacinta.com/myjava/fast_md5.php

  • How to use IMAQ extract color panel in a LabView file?

    I want to change a color image into a greyscale image. Therefore I want to use the function IMAQ extract color panels but I don't know how to insert it in my LabView application.
    Please, can anyone give me some advice ?!!!
    Thx, Birgit

    Birgit,
    If you want all of the color planes, you can use IMAQ ExtractColorPlanes. Allocate an 8-bit image for each of the color planes. Wire the original color image and each of the 8-bit images for the color planes on the left and bottom of the vi. The outputs on the right side will be the same images, and they will each contain the desired color plane. The input at the top of the vi controls which color planes are extracted (RGB, HSI, etc.) Grayscale will probably be the I plane in HSI.
    If you just want a grayscale image, use IMAQ ExtractSingleColorPlane instead. Wire the color image to Image Src, and wire an 8-bit image to Image Dest. Select the color plane using the top input, and it will come out on the right side.
    Bruce
    Bruce Ammons
    Ammons Engineering

  • How to use preprocess​or directives (#define) in C++ header file with LabVIEW 8.2

    I have a C++ header file that contains around 2000 preprocessor directives:
    #define MEM_1   0xC
    #define MEM_2   0xD
    #define MEM_3   0x18
    I want to be able to "access" these memory offsets by identifier name (MEM_1) in my LabVIEW program like I would in a C++ program.  I do not want the overhead of parsing through the header file and storing all the offsets into an array or similar structure. 
    I've written a simple Win32 console program to return the memory offset given the identifier (see code below), and created a DLL to use with my LabVIEW program.  In the console program, you notice that I can call a function and pass in the identifer name, and get the offset back correctly:
    getOffset(MEM_1);
    In LabVIEW, I was hoping to be able to pass in the identifier (MEM_1) but was unsure what datatype to use.  In my C++ code, I defined the parameter as an int.  But in LabVIEW, I can't enter in MEM_1 as an int.   Can someone advise on how to do this?  Or if there is an alternate way to use #define's from external code inside LabVIEW?
    #include "stdafx.h"
    #include "scrmem.h"
    #include "stdio.h"
    void getOffset (int var);
    int _tmain(int argc, _TCHAR* argv[])
     getOffset(MEM_1);
    canf("%d");
     return 0;
    void getOffset (int var)
     printf("The address of MEM_1 is %x", var); 

    kaycea114 wrote:
    Hi,
    Where do you think I should use the string? 
    The way that getOffset is currently defined in the DLL, I have to connect an integer input into the LabVIEW function.  This prevents me from entering in: MEM_1 as the input to the LabVIEW function.
    Are you suggesting that I change getOffset to receive a String parameter ("MEM_1")?  Does that mean I need to do a string compare (line by line) through the header file to get the offset?  It seems like doing this search through the header file would degrade performance, but if that's the only work around, then I'll do it.
    Please advise.
    Well, what you want to do is indeed entering a string and getting back the assigned integer. That is what the C preprocessor is doing too although there it is done only once at the preprocessor stage of course and not at runtime anymore. But LabVIEW is not a C preprocessor.
    What you did so far seems to be to define getOffset() that accepts an enum that needs to be created from the C source code to then return the assigned constant. That's of course not very helpful.
    And writing a VI that could parse the C header file and create a name/constant array is really a lot easier than doing the same in C. You don't even need to parse the file each time again, but can instead cache them in an uninitialized shift register (LV2 style global).
    Even more easy would be to create from that data a ring control using property nodes and save it as custom control and voila you have the most direct lookup you can get in LabVIEW and it works just as comfortable as using the define in C code. It would mean that you need to seperate your header file possibly into several different files to only get related constants into the same ring control, but that is easily done.
    Rolf Kalbermatter
    Rolf Kalbermatter
    CIT Engineering Netherlands
    a division of Test & Measurement Solutions

  • How to use XQuery Exist() Function on a SSIS XML file

    I have a Package Inventory table that has an XML Column named CurrentPackageXML. This XML column contains the SSIS Package XML. I need to use the XQuery Exist() function to determine if the
    errorRowDisposition="IgnoreFailure" exists.
    Below I have a basic SQL statement, but I have no idea how to write the Correct XQuery Expression. In my example below, I just tried to see if the DTS:Name Property existed, but that XQuery Expression is invalid also. It is not what I need in the end, but
    I was just trying to get anything to work.
    Below my SQL, I have included a snip-it of the SSIS package XML where I can find the errorRowDisposition. It is to the far right, so you may have to scroll to the right to see it. I bolded the text so it is easier to find.
    Finally, once I have got a SQL statement that checks if errorRowDisposition exists = False, I need to check for every occurrence of  errorRowDisposition exists = False.
    Would there be an easier way to do this by converting the XML to a varchar(max) then searching for all instances of  errorRowDisposition exists = False?
    SELECT CurrentPackageXML.exist('/DTS:Executable/DTS:Property[DTS:Name]')
    FROM [dbo].[PackageInventory]
    <inputs>
    <input id="43" name="OLE DB Destination Input" description="" hasSideEffects="true" dangling="false" errorOrTruncationOperation="Insert" errorRowDisposition="FailComponent" truncationRowDisposition="NotUsed">
    <inputColumns>
    <inputColumn id="113" name="" description="" lineageId="110" usageType="readOnly" errorOrTruncationOperation="" errorRowDisposition="NotUsed" truncationRowDisposition="NotUsed" externalMetadataColumnId="95" mappedColumnId="0" />
    <inputColumn id="172" name="" description="" lineageId="167" usageType="readOnly" errorOrTruncationOperation="" errorRowDisposition="NotUsed" truncationRowDisposition="NotUsed" externalMetadataColumnId="94" mappedColumnId="0" />
    </inputColumns>
    <externalMetadataColumns isUsed="True">
    <externalMetadataColumn id="50" name="DateCreated" description="" precision="0" scale="0" length="0" dataType="dbTimeStamp" codePage="0" mappedColumnId="0" />
    <externalMetadataColumn id="51" name="ProcessedCount" description="" precision="0" scale="0" length="0" dataType="i4" codePage="0" mappedColumnId="0" />
    <externalMetadataColumn id="52" name="ErrorCount" description="" precision="0" scale="0" length="0" dataType="i4" codePage="0" mappedColumnId="0" />
    <externalMetadataColumn id="90" name="AcknowledgementID" description="" precision="0" scale="0" length="0" dataType="i4" codePage="0" mappedColumnId="0" />
    <externalMetadataColumn id="91" name="Date" description="" precision="0" scale="0" length="0" dataType="dbTimeStamp" codePage="0" mappedColumnId="0" />
    <externalMetadataColumn id="92" name="FileSeqNumber" description="" precision="0" scale="0" length="0" dataType="i4" codePage="0" mappedColumnId="0" />
    <externalMetadataColumn id="93" name="FileType" description="" precision="0" scale="0" length="50" dataType="str" codePage="65001" mappedColumnId="0" />
    <externalMetadataColumn id="94" name="FileName" description="" precision="0" scale="0" length="50" dataType="str" codePage="65001" mappedColumnId="0" />
    <externalMetadataColumn id="95" name="DateReceived" description="" precision="0" scale="0" length="0" dataType="dbTimeStamp" codePage="0" mappedColumnId="0" />
    </externalMetadataColumns>
    </input>
    </inputs>
    Ryan P. Casey • <a href="http://www.R-P-C-Group.com">www.R-P-C-Group.com</a>

    Try this and let know what errors you get.  Note, this is a rough draft of one of a series of queries that I have started writing to document SSIS using SQL XQuery.  I read a file in my version (the one I tested) and get the results.
    -- Get the flat file destinations and their mappings
    with xmlnamespaces ('www.microsoft.com/SqlServer/Dts' as DTS
    ), compflow as (
    SELECT
    task.xml.value('@DTS:ObjectName', 'varchar(200)') as TaskName
    , task.xml.value('@DTS:ExecutableType', 'varchar(200)') as TaskType
    , task.xml.value('@DTS:refId', 'varchar(1000)') as TaskRefId
    --, task.xml.query ('.') as Node
    , component.xml.value ('@name', 'varchar(200)') as ComponentName
    , component.xml.value ('@refId', 'varchar(1000)') as ComponentRefId
    , component.xml.value ('@componentClassID', 'varchar(1000)') AS componentClassID
    --, component.xml.value ('count(inputs/input)', 'int') AS NumberOfInputs
    --, component.xml.value ('count(outputs/output)', 'int') AS NumberOfOutputs
    --, component.xml.value ('count(outputs/output[@isErrorOut="true"])', 'int') AS NumberOfErrorOutputs
    , component.xml.query('.') as ComponentNode
    FROM [dbo].[PackageInventory] pk
    cross apply pk.CurrentPackageXML.nodes('//DTS:Executable[@DTS:ExecutableType="Microsoft.Pipeline"]') as task (xml)
    OUTER APPLY task.xml.nodes('DTS:ObjectData/pipeline/components/component') as component (xml)
    WHERE [PackageInventory_ID] = 13
    , inputNode as (
    SELECT c.*
    , input.xml.value ('@name', 'varchar(200)') as InputName
    , input.xml.value ('@refId', 'varchar(1000)') as InputRefId
    , input.xml.value ('@hasSideEffects', 'varchar(1000)') as InputHasSideEffects
    , input.xml.value ('count(inputColumns/inputColumn)', 'int') AS NumberOfInputColumns
    , input.xml.value ('count(externalMetadataColumns/externalMetadataColumn)', 'int') AS NumberOfExternalMetadataColumns
    , input.xml.value ('@errorOrTruncationOperation', 'varchar(1000)') as InputerrorOrTruncationOperation
    , input.xml.value ('@errorRowDisposition', 'varchar(1000)') as InputerrorRowDisposition
    , col.xml.value ('@errorOrTruncationOperation', 'varchar(1000)') as InputColumnErrorOrTruncationOperation
    , col.xml.value ('@errorRowDisposition', 'varchar(1000)') as InputColumnErrorRowDisposition
    from compFlow c
    OUTER APPLY c.ComponentNode.nodes ('component/inputs/input') as input (xml)
    OUTER APPLY input.xml.nodes ('inputColumns/inputColumn') as col (xml)
    , inputAnal as (
    select
    TaskName
    ,TaskType
    ,TaskRefId
    ,ComponentName
    ,ComponentRefId
    ,componentClassID
    ,InputName
    ,InputRefId
    ,InputHasSideEffects
    , Null as IsError
    ,NumberOfInputColumns
    ,NumberOfExternalMetadataColumns
    ,InputerrorOrTruncationOperation
    ,InputerrorRowDisposition
    ,InputColumnErrorOrTruncationOperation
    ,InputColumnErrorRowDisposition
    , count(*) as NumRec
    from inputNode
    group by TaskName
    ,TaskType
    ,TaskRefId
    ,ComponentName
    ,ComponentRefId
    ,componentClassID
    ,InputName
    ,InputRefId
    ,InputHasSideEffects
    ,NumberOfInputColumns
    ,NumberOfExternalMetadataColumns
    ,InputerrorOrTruncationOperation
    ,InputerrorRowDisposition
    ,InputColumnErrorOrTruncationOperation
    ,InputColumnErrorRowDisposition
    , xoutputNode as (
    SELECT
    TaskName
    ,TaskType
    ,TaskRefId
    ,ComponentName
    ,ComponentRefId
    ,componentClassID
    , xout.xml.value ('@name', 'varchar(200)') as OutputName
    , xout.xml.value ('@refId', 'varchar(1000)') as OutputRefId
    , xout.xml.value ('@hasSideEffects', 'varchar(1000)') as OutputHasSideEffects
    , xout.xml.value ('@isErrorOut', 'varchar(1000)') as OutputIsError
    , xout.xml.value ('count(outputColumns/outputColumn)', 'int') AS NumberOfOutputColumns
    , xout.xml.value ('count(externalMetadataColumns/externalMetadataColumn)', 'int') AS NumberOfExternalMetadataColumns
    , xout.xml.value ('@errorOrTruncationOperation', 'varchar(1000)') as OutputerrorOrTruncationOperation
    , xout.xml.value ('@errorRowDisposition', 'varchar(1000)') as OutputerrorRowDisposition
    , col.xml.value ('@errorOrTruncationOperation', 'varchar(1000)') as OutputColumnErrorOrTruncationOperation
    , col.xml.value ('@errorRowDisposition', 'varchar(1000)') as OutputColumnErrorRowDisposition
    from compFlow c
    OUTER APPLY c.ComponentNode.nodes ('component/outputs/output') as xout (xml)
    OUTER APPLY xout.xml.nodes ('outputColumns/outputColumn') as col (xml)
    , outputAnal as(
    select
    TaskName
    ,TaskType
    ,TaskRefId
    ,ComponentName
    ,ComponentRefId
    ,componentClassID
    ,OutputName
    ,OutputRefId
    ,OutputHasSideEffects
    ,OutputIsError
    ,NumberOfOutputColumns
    ,NumberOfExternalMetadataColumns
    ,OutputerrorOrTruncationOperation
    ,OutputerrorRowDisposition
    ,OutputColumnErrorOrTruncationOperation
    ,OutputColumnErrorRowDisposition
    , count (*) as NumberOfRows
    from xoutputNode
    group by TaskName
    ,TaskType
    ,TaskRefId
    ,ComponentName
    ,ComponentRefId
    ,componentClassID
    ,OutputName
    ,OutputRefId
    ,OutputHasSideEffects
    ,OutputIsError
    ,NumberOfOutputColumns
    ,NumberOfExternalMetadataColumns
    ,OutputerrorOrTruncationOperation
    ,OutputerrorRowDisposition
    ,OutputColumnErrorOrTruncationOperation
    ,OutputColumnErrorRowDisposition
    select 'Input' as RowType, i.* from inputAnal i
    union all
    select 'Output' as RowType, o.* from outputAnal o
    Russel Loski, MCT, MCSE Data Platform/Business Intelligence. Twitter: @sqlmovers; blog: www.sqlmovers.com

  • How to use Oracle Reports in an enterprise app, wrt file i/o?

    How are people architecting their enterprise applications to make use of Oracle Reports, with respect to transmission of the output?
    Do you have session beans which call the Oracle Reports JSP pages, and have them write the output to file names, the names of which you then transmit to the web layer? Alternatively, is there any practical way to "stream" the output directly from the EJB layer to the web layer and to the client?

    Hello David,
    When you call the report server via an URL, with DESTYPE=CACHE, the output is directly sent to the web browser.
    I am not sure to well understand you question, but if you want to control the report from a session bean, you can write a bean that call report server with an URL and get the output. Today, we do not have any EJB to control Oracle Reports.
    I am interested by the type of beans/architecture you would like to see in Reports, could you please send me a detailled example of a type of bean you would like to have ? ([email protected])
    Regards
    Tugdual Grall

  • How to use RAC database connection details in oc4j.xcfg files

    Hi all,
    we are using oc4j.xcfg files for the application module configuration for an ADF application. and one of the database has been configured on RAC. Any input on how to specify the RAC database details in the oc4j.xcfg file specially for the host name parameter.
    The example entry like in oc4j.xcfg is like
    ?xml version = '1.0' encoding = 'UTF-8'?>
    <BC4JConfig>
    <AppModuleConfigBag>
    <AppModuleConfig name="AppModuleLocal">
    <AppModuleJndiName>dbta.model.AppModule</AppModuleJndiName>
    <DeployPlatform>LOCAL</DeployPlatform>
    <JDBCName>connectDB</JDBCName>
    <ApplicationName>dbta.model.AppModule</ApplicationName>
    <jbo.project>Model</jbo.project>
    </AppModuleConfig>
    </AppModuleConfigBag>
    <ConnectionDefinition name="connectDB">
    <ENTRY name="ConnectionName" value="connectDB"/>
    <ENTRY name="URL" value="jdbc:oracle:thin:<user>/<pass>@<server>:<Port>:<SID>"/>
    <ENTRY name="HOSTNAME" value="<server>"/>
    <ENTRY name="JDBC_PORT" value="<port>"/>
    <ENTRY name="SID" value="<SID>"/>
    <ENTRY name="ConnectionType" value="JDBC"/>
    <ENTRY name="JdbcDriver" value="oracle.jdbc.driver.OracleDriver"/>
    <ENTRY name="ORACLE_JDBC_TYPE" value="thin"/>
    <ENTRY name="DeployPassword" value="true"/>
    </ConnectionDefinition>
    how to specify the failover nodes details in the above entry for host name etc.
    Thanks
    Navin

    Here you are; all you want at
    http://www.oracle.com/technology/sample_code/tech/java/sqlj_jdbc/files/9i_jdbc/OCIdriverTAFSample/OCIdriverTAFSample.java.html

  • How to Use iSetup to create new environment in copy of clean master.

    Hi,
    I want to use iSetup to create a complete new environment. No manual setup has been done yet in the copy of the master (no Business Group, No OU, no Chart of Acc: nothing). Source environment is being setup rightnow by functional consultants
    Can one please provide me with a manual or guideline how to start this challenging job. Maybe a best practise can be used ?
    In the end we want to setup an environment based on a clean master.
    ( practical data: eBS 12.1.2 on linux )
    Thanks in advance
    Guido

    Hi Guido,
    You can get in touch with [email protected] who is the product manager for iSetup. He would able to assist you to go forward in the right direction.

  • Report question: how to use external table to bring data from excel file?

    i need to import excel data into oracle report, how to do that? thanks.

    Hi... What is the error message you get when using external tables?
    You must create a control center connection in the Connection Explorer.
    I recommend you read this:
    http://download-east.oracle.com/docs/cd/B31080_01/doc/owb.102/b28223/toc.htm
    Look for the following sections:
    - Using External Tables (and the "External Tables versus Flat File Operators" sub section)
    - About Flat File Modules
    - Chapter 16, Deploying Target Systems
    Regards,
    Marcos

  • How to use third party library (like .dll and .lib files) in flascc?

    I have a number of .dll and .lib files of the  third party library, which are generated from C++ code.
    I read about flascc on other web site says:
    "If you want to use third party library in your Flascc project,you must have either the source code or the linkable file which is compiled of the third party library   "
    Dose the linkable file mean .dll file or .lib file?
    AND
    How can I use the fuctions in them in my flascc project.?

    Native code is not interchangeable between platform.
    It means that you can't use library that wasn't build specifically for flascc.
    And when they say linkable files they mean libraries compiled for flascc, not any libraries.
    So if your libs are for windows, no luck, you have to recompile everything

Maybe you are looking for

  • SharePoint Active Directory version support

    Dear All, My active directory is on windows server 2003 and now planning to install SharePoint 2013. I have windows server 2012 with seperate hardware and want to install Sharepoint on this server. I have some queries relates on the Sharepoint deploy

  • Is there a quick way to delete all songs on my iPhone 4, running 5.0.1?

    The last time i synced my iPhone on iTunes i must have chosen the wrong option to put 2 new songs onto my Playlist (on iTunes) and then i chose Autofill. What happened was all/most of my songs duplicated. i did accidentally chose to the duplication o

  • DVD Player bookmarks and clips.

    I did an erase and install of Leopard on my machine, which previously had Tiger on it. I want to restore my DVD Player bookmarks and clips from my backup. Can anyone Tell me which files I need to restore? I found the following thread-which is dated-o

  • ORA-12535: TNS:operation timed out

    First thing first I need to vpn to Data Center before able to connect to any of the servers in Data Center. The other thing is do not ask me why my company set up is so strange. Let's call this DB server Machine B DB Server Oracle9i Release 9.0.1.1.1

  • No Viewsets in WDP for ABAP so....?

    How do I go about displaying 2 or more views in the same window? Is this possible thru a workaround of some sort? I tried embedding 2 views in one window but the default inbound plug only fires one of the windows, and I can't set both the views to be