Bundled Tomcat File Structure and Sun Studio Enterprise 8

I am having trouble using form authorization with Enterprise 8,
the bundled Tomcat and MySql.
Request clarification on what tomcat-users.xml is used by Ent 8.
I see the following in my configuration:
1) apache-tomcat-5.5.15-embed/conf/tomcat-users.xml
2) apache-tomcat-5.5.15/conf/tomcat-users.xml
3) C:\Program Files\Sun\jstudio_ent8\ide\enterprise1\jakarta-tomcat-5.5.7\conf\tomcat-users.xml
4) C:\Document and Settings\motel\.jstudio\Ent8\jakarta-tomcat-5.5.7_base\conf\tomcat-users.xml
I assume the first two are some type of windows cached.
Further, #3 I assume is initial or default tomcat-user file.
So is #4 the file that I want to edit so that my IDE can work with Tomcat?
Thank you for any help and guidance,
Craig

I assume the first two are some type of windows cached.
Further, #3 I assume is initial or default tomcat-user file.I would agree with you; regarding the first two files, it is not clear how they came about since they don't seem to be in either the ide directory or ide-user directory; and also because the directories seem to point to a different version of tomcat, namely 5.15 instead of the ide-bundled 5.7... It is even possible that the first two are not related to JSE ide at all...
Regarding 3 and 4, your guess is right; 3 is the initial installation directory. This directory is not used by the ide at all for actually working with the tomcat server; this is because the ide-userdir is customizable via the command-line. You can have several user directories at the same time and switch between them by running './runide.exe -userdir <this-sessions-user-dir>'. Naturally the different sessions cannot use the same tomcat server since that would lead to conflicts. So the ide first copies the server installation to the current user directory and then runs the tomcat from there.

Similar Messages

  • Confused between Netbeans and Sun Studio Enterprise

    Hi all,
    Acually I am very confused between Netbeans and Sun Studio Enterprise/Creator.
    Why sun did not concentrate on one IDE as Microsoft is doing with Visual Studio?
    Now, I do not know what to learn....Netbeans or Sun Studio !!!
    What make me more confused is that both are supported oficially by Sun. And both are free for SDN members!!!!!
    So, what is the difference???!!!
    regards

    I tried it back when it was called Forte, and it seamed pretty
    identical to netbeans. the only difference was that Forte could
    only be used for personal use without a license.

  • INFORMIX 9 and SUN Studio Enterprise Configuration

    Attention all informix to java programmers
    Architecture regulations obliges the use of the J2EE platform. Sun ONE
    Application server 7 is the appserver to use and Informix Dynamic Server
    9.22 the database. This is not the most mainstream combination.
    Especially since Sun ONE doesn't support the Informix database.
    It turned out that you can use this combination, but you need to apply
    some workarounds.
    First you need to know which properties to set for the datasource. This
    is the list:
    IfxIFXHOST=foo.bar.com
    PortNumber=1527
    User=Krusty
    Password=L3k4N1%j
    ServerName=foobarifx_ol
    DatabaseName=testdb
    The values are examples. Be careful of the syntax. The first letter
    should be a capital!
    The second adjustment is a little harder. Because Informix syntax isn't
    fully ANSI92 compliant (it doesn't support right joins),
    the persistence manager of Sun ONE generates illegal SQL when
    fulfilling it's CMP responsabilities.
    To solve this problem you need to add a property file to the
    appserv-cmp.jar.
    This lib is part of the Sun ONE server libraries and you can find it at
    {Sun ONE INSTALLDIR}/lib.
    The file name should be "INFORMIX DYNAMIC SERVER.properties" and it
    should go under the folder
    com/sun/jdo/spi/persistence/support/sqlstore/database.
    This is the content of the file:
    FOR_UPDATE=for update
    LEFT_JOIN=LEFT OUTER JOIN
    RIGHT_JOIN=,
    RIGHT_JOIN_PRE=outer
    IS_NULL=IS NULL
    IS_NOT_NULL=IS NOT NULL
    RTRIM=trim (trailing \" \" from
    RTRIM_POST=)
    QUOTE_CHAR_START=
    QUOTE_CHAR_END=
    Having this information up front would have saved me many hours of frustration. I hope anyone else using Informix 9 and J2EE finds this helpful.

    i follow this steps , but i still have problems now i got this exception
    [19/Jul/2005:20:17:53] INFO ( 3812): CORE3282: stdout: java.sql.SQLException: com.informix.asf.IfxASFException: Attempt to connect to database server (srvr_cannes_tcp) failed.
    i think is the ServerName variable did you know which is the problem; i'm still frustrated

  • Need help understanding Time Capsule file structure and how to get it back

    I have the original Time Capsule on which I backup both my Mac Pro and my wife’s Macbook. When Snow Leopard came out, I successfully used the ‘Restore from Time Machine’ feature on my Mac Pro so I know it has worked. However, my wife’s MacBook harddrive died the other day and I was trying to do the ‘Restore from Time Machine’ and all it would find was a backup from April (when I put her in a new larger drive). Time Machine would not find any backup files newer that April. She stated that she had seen the the Time Machine backup notices regularly and as recent as the day the haddrive died (Nov. 23) so I figured that I should have no problem. Here is what I have found in my trouble shooting and what leads to my questions below.
    This is the file structure I found: (note that our ID’s are ‘Denise’ and ‘John’)
    *Time Capsule* (the drive as listed in my Finder window sidebar under ‘shared’)
    >Folder called ‘Time Capsule’ (when logged in as either ‘Denise’ or ‘John’)
    >>Denise Sparsebundle
    >>>Backup of Denise’s iBook (mounted image)
    >>>>Folder called ‘Backups.backupdb’
    >>>>>Folder called ‘Denise’s iBook
    >>>>>>Single folder with old April backup (not the right files)
    >>John Sparsebundle
    >>>Backup of John’s Mac Pro (mounted image)
    >>>>Folder called ‘Backups.backupdb’
    >>>>>Folder called ‘John’s Mac Pro’
    >>>>>>Folders containing all my backup files
    >Folder Called ‘Denise’ (if logged as ‘Denise’)
    >>Denise’s Sparsebundle (a disk image)
    >>>Backup of Denise iBook (the mounted image. Name from old machine)
    >>>>Backups.Backupdb
    >>>>>Denise’s iBook (Contains the backup I need)
    >Folder Called ‘John’ (if logged in as ‘John’)
    >> (empty)
    For some reason, my wife’s backup files are stored within a folder located at the top level of the Time Capsule drive called ‘Denise’, however, mine are within a folder called ‘Time Capsule’ which is at the same level as the ‘Denise’ folder.
    For some reason, when trying to use Time Machine to recover, it bypasses the top level ‘Denise’ folder which contains the correct files and goes into the ‘Time Capsule’ folder and finds the outdated backup files.
    I would assume that both my backup files and my wife’s should be at the same level of the Time Capsule.
    I was eventually able to use Migration Assistant to recover the files after installing a fresh OS and mounting the correct Sparsebundle image.
    So, my question, how do I get this fixed so that if I have to recover a drive with Time Capsule, it will find the correct files.
    Sorry for the long post and thanks in advance for any help.

    John Ormsby wrote:
    What I was trying to determine is why different backups for one particular machine are located at different file structure levels on my Time Capsule and why the most resent one for that machine is at a different level that for my other machine. Also, what is the correct level that both machines backups should be at.
    well John, first can you clarify if you are on 10.4.8 as your profile suggests ? if so, i'm wondering how you can use TM at all because TM was introduced with Leo. if you're not on 10.4.8, please update your profile. if you could, please also indicate details of your machine such as available RAM, processor, etc.
    second, what OS is your wife's machine running ?
    third, i frankly don't know too much about TM's file structure or if there indeed is such a thing as the correct one. however, i do know that it is best to leave TM to do its thing (as long as it's working). FWIW, though off-topic, you may want to have a look at this read for further information.
    last but not least, i see TM backups to my TC only as a part of my backup strategy. the backbone of my strategy is to create bootable clone of my startup disk(s) using e.g. Carbon Copy Cloner on a regular basis. while TM is capable of doing a full system restore, i don't trust it as much as a working clone. i use TM to retrieve a file i accidentally trashed a week ago but you cannot boot from a TM backup if your startup disk goes belly up.
    If I have missed this information in either the FAQs or the Troubleshooting article, I would greatly appreciate being pointed to it .
    i expect you didn't miss anything and i'm sorry you didn't find any help there. perhaps if Pondini (author of the FAQ and troubleshooting user tips) reads this thread, you will get the necessary advice on the file structure and more besides.
    good luck to you !

  • File structure and hierarchy in Aperture.

    Hi, everyone.
    I am trying to import folders with images into Aperture and, much to my surprise, have found that this process is everyting but intuitive and straightforward.
    My library (outside of Aperture) is made up of folders each corresponding to a day and location. These folders are labeled using the data and location as part of their name.
    As I try to import these folders into Aperture I can't seem to import them into a single folder or project as sub-folders or sub-domains of that project. I have tried to create a project and them import the folders into it but Aperture won't import them into that project and doesn't allow me to drag and drop it either once it has been imported. I also tried to create albums and folders but haven't been successful.
    How does this file structure in Aperture work in terms of hierarchy ? It certainly isn't structured the way the Finder is or any other file system I have seen to date. Creating folders, organizing them and bringing any type of data into these folders should be a simple process but in Aperture it doesn't seem to be.
    Am I doing something wrong or is Aperture trying to re-invent the wheel ?
    Is there any tutorial I can watch or read on how to work with Aperture's file structure and import folders and folders into it ?
    Thank you in advance.

    Regarding your specific problem I suggest treating each existing dated folder of image files as a single Project in Aperture. In Aperture a Project is a specific time-based concept that may or may not jive with what you previously considered a project. For instance I may have in my mind that shooting all the highest peaks in every state is a "project," but that would be inappropriate as an Aperture Project. Instead each peak might be a Project, but the pix of all the peaks would be pointed to as an Album.
    The way I look at it conceptually:
    Aperture is a database (DB), and each image file lives in one Project.
    Albums are just collections of Pointers that point to individual image files living in one or more Projects. Since they just contain pointers, albums can be created or deleted at will without affecting image files. Very powerful. And Albums of pointers take up almost zero space, so they are fast and do not make the Library size grow.
    Keywords can be applied to every image separately or in batches. Keywords are hugely powerful and largely obviate the need for folders. Not that we should never use folders, just that we should use folders only when useful organizationally - - after first determining that using keywords and albums is not a better approach.
    As one example imagine the keyword "flowers."  Every image of a 100,000 images Library that has some flowers in it has the keyword flowers. Then say we want to put flowers in an ad, or as background for a show of some kind, or to print pix for a party, or even just to look for an image for some other reason. We can find every flower image in a 100k-image database in 2 seconds, and in another few seconds create an Album called "Flowers" that points to all of those individual images.
    Similarly all family pix can have a keyword "family" and all work pix can have a key word "work." Each individual pic may have any number of keywords. Such pic characteristics (work, family, flowers, etc.) should not be organized via folders.
    So by using keywords and albums we can have instant access to every image everywhere, very cool. And keywords and albums essentially take up no space in the database.
    Another approach is to use a folder "Family" for family pix, a folder "Flowers" for flowers pix and another folder "Work" for work pix. IMO such folders usage is a very poor approach to using an images database (probably stemming from old paper or film work practices). Note that one cannot put an image with family in a field of flowers at a work picnic in all three folders; but it is instant with keywords.
    HTH
    -Allen

  • IWeb file structure and naming system

    I am trying to publish my iWeb directly to my domain (www.jessewarren.net) using the FTP direct publish function. However, iWeb is applying an overly complex naming system.
    For example, its publishing not only the page names, but also the name of the Site in iWeb, after the original URL. Here's an example: http://jessewarren.net/testing/Site/Blank.html
    First, I need everything to be on the domain. Then, I only need the name of the individual pages to be published, not the Site Name itself. I already have a site name, and that's my domain.
    Any other advice on how to control file structure and naming in iWeb?

    This is not incorrect and it is the way iWeb publishes and has done ever since iWeb 06. When you publish from iWeb to a local folder, what you get is your Site folder with your site name and a separate index file.
    If you don't want your site name included, then what you need to do is rather than using the direct ftp from iWeb, select ftp to a local folder instead. Your site will then be published to your desktop. Then download either Cyberduck or Transmit for free - these are dedicated ftp programs. Then all you do is load just the contents of your site folder to your public_html folder on your server and your url will then be http://www.domain.com/pagename.html.
    Be aware though, that if you are loading like this, you are breaking iWeb site structure and RSS feed will not be available to you if you need to use it.
    If you want to continue to use the built in iWeb ftp, then you could try naming your site public_html and then leaving the directory blank and then uploading like this and because you have named your site public_html and are uploading to the same place, your public_html folder on your server will be replaced with your website, thus eliminating the site name.
    I have tried this method once myself with my own site and it does work, however, I still prefer to publish to a folder and then upload to my sever using Cyberduck.

  • How to change Makefile file name in sun studio for linux 11?

    Greetings,
    For historical reasons, our makefiles are called Make.bld
    Sun Studio refuses to accept these as makefiles.
    It'd be a big pain to modify our entire build system to take makefile as the name.
    Is there a way to instruct sun studio to accept a different filename for makefiles ?
    Thanks,
    Bill

    There is no supported way of getting the Sun Studio IDE (or NetBeans) to recognize
    Make.bld as a Makefile. However, the solution Nik pointed out with a .nbattrs file will
    work.
    Here is a complete file which, if copied to each directory containing a Make.bld file
    will work. However, if the directory already contains a .nbattrs file you should edit
    the existing one and add the <fileobject>...</fileobject> lines rather than overwriting
    the entire file.
    <?xml version="1.0" encoding="UTF-8"?>
    <!DOCTYPE attributes PUBLIC "-//NetBeans//DTD DefaultAttributes 1.0//EN"
    " http://www.netbeans.org/dtds/attributes-1_0.dtd">
    <attributes version="1.0">
    <fileobject name="Make.bld">
    <attr name="MAKEFILE_TYPE" stringvalue="foo"/>
    </fileobject>
    </attributes>
    Note: I split the DOCTYPE line myself. It should be a single line in the file (although
    its probably OK if split, I didn't verify that so I'll suggest a single line).
    I should warn you that the next release of the Sun Studio IDE will no longer use
    .nbattrs files and you'll need a different solution. A preliminary version of the next
    release is currently available by downloading Sun Studio Express 3. The IDE is
    significantly better than the current one! You also get a lot of the same features
    by downloading NetBeans 5.5 with the C/C++ Development Pack.
    Gordon

  • BOOST and Sun Studio 10

    C++ 5.7 in Sun Studio 10 now compiles most of BOOST 1.32.0, when the compiler is updated with the current patches from
    http://developers.sun.com/prodtech/cc/downloads/patches/ss10_patches.html
    Two known bugs remain, and should be fixed in a future patch.
    Some of the BOOST compilation failures are due to errors in the BOOST configuration for Sun C++. We have submitted corrections to the BOOST site, and we hope all of them will be included in version 1.33.0.
    In the mean time, you can apply the following corrections to version 1.32.0:
    1. Get this correction from the BOOST bug tracker:
    http://sourceforge.net/tracker/index.php?func=detail&aid=1155439&group_id=7586&atid=107586
    2. Also get this correction from the BOOST bug tracker:
    http://sourceforge.net/tracker/index.php?func=detail&aid=1165980&group_id=7586&atid=107586
    3. Edit build.sh according to these diffs: *** build.sh.orig     Fri Mar 18 06:09:44 2005
    --- build.sh     Fri Mar 18 06:41:07 2005
    *** 150,160 ****
          sunpro)
    !     if test -r /opt/SUNWspro/bin/cc ; then
              BOOST_JAM_TOOLSET_ROOT=/opt/SUNWspro/
          fi
    !     if test -r $BOOST_JAM_TOOLSET_ROOTbin/cc ; then
    !         export PATH=$BOOST_JAM_TOOLSET_ROOTbin:$PATH
          fi
          BOOST_JAM_CC=cc
    --- 150,161 ----
          sunpro)
    !     if test -z "${BOOST_JAM_TOOLSET_ROOT}" -r /opt/SUNWspro/bin/cc ; then
              BOOST_JAM_TOOLSET_ROOT=/opt/SUNWspro/
          fi
    !     if test -r ${BOOST_JAM_TOOLSET_ROOT}bin/cc ; then
    !         PATH=${BOOST_JAM_TOOLSET_ROOT}bin:${PATH}
    !         export PATH
          fi
          BOOST_JAM_CC=cc
          ;;

    Here is a small but complete test program, together with the command line of the compiler. This is the patch level of the compiler we use here:
    CC: Sun C++ 5.7 Patch 117830-02 2005/03/30Test program (graph_test.cpp):
    #include <iostream>                  // for std::cout
    #include <utility>                   // for std::pair
    #include <algorithm>                 // for std::for_each
    #include <boost/graph/graph_traits.hpp>
    #include <boost/graph/adjacency_list.hpp>
    #include <boost/graph/dijkstra_shortest_paths.hpp>
    using namespace boost;
    int main(int,char*[])
      enum { A, B, C, D, E, N };
      const int num_vertices = N;
      const char* name = "ABCDE";
      typedef adjacency_list<vecS, vecS, bidirectionalS> Graph;
      Graph g(num_vertices);
      typedef graph_traits<Graph>::vertex_iterator vertex_iter;
      std::pair<vertex_iter, vertex_iter> vp;
      vp = vertices(g); /** Error **/
      return 0;
    }Command line:
    CC -I./include/boost -v graph_test.cpp
    ###     command line files and options (expanded):
    ### -I/export/rpms/401/tools/inst/include/boost -v graph_test.cpp
    ### CC: Note: NLSPATH = /export/home2/SUNs10/SUNWspro/prod/bin/../lib/locale/%L/LC_MESSAGES/%N.cat:/export/home2/SUNs10/SUNWspro/prod/bin/../../lib/locale/%L/LC_MESSAGES/%N.cat
    /export/home2/SUNs10/SUNWspro/prod/bin/ccfe -y-o -ygraph_test.o -I/export/rpms/401/tools/inst/include/boost -y-fbe -y/export/home2/SUNs10/SUNWspro/prod/bin/fbe -y-xarch=generic -y-verbose -O0 -ptf /tmp/00299%1.%2 -ptx /export/home2/SUNs10/SUNWspro/prod/bin/CC -ptk "-I/export/rpms/401/tools/inst/include/boost -v  -xs " -D__SunOS_5_8 -D__SUNPRO_CC=0x570 -Dunix -Dsun -Dsparc -D__sparc -D__sparcv8plus -D__unix -D__sun -D__BUILTIN_VA_ARG_INCR -D__SVR4 -D__SUNPRO_CC_COMPAT=5 -D__SUN_PREFETCH -xdbggen=no%dwarf2+stabs -y-s -xdbggen=incl -y-xmemalign=8i -I-xbuiltin -xldscope=global -instlib=/export/home2/SUNs10/SUNWspro/prod/lib/libCstd.a -I/export/home2/SUNs10/SUNWspro/prod/include/CC/Cstd -I/export/home2/SUNs10/SUNWspro/prod/include/CC -I/export/home2/SUNs10/SUNWspro/prod/include/CC/rw7 -I/export/home2/SUNs10/SUNWspro/prod/include/cc -y-comdat graph_test.cpp -s /tmp/ccfe.00299.0.s >&/tmp/ccfe.00299.1.err
    rm /tmp/ccfe.00299.0.s
    /export/home2/SUNs10/SUNWspro/prod/bin/stdlibfilt -stderr </tmp/ccfe.00299.1.err
    "graph_test.cpp", line 22: Error: Could not find a match for boost::vertices<boost::T>(boost::adjacency_list<boost::vecS, boost::vecS, boost::bidirectionalS, boost::no_property, boost::no_property, boost::no_property, boost::listS>) needed in main(int, char**).
    1 Error(s) detected.
    rm /tmp/ccfe.00299.1.err
    rm graph_test.oAny help would be greatly appreciated! If you need any more information, please let me know.
    Thanks in advance,
    Andreas

  • Error building using Coverity Prevent and Sun Studio 11

    We recently started using a static code-checking tool called Coverity Prevent. When we first started using it, we were on a very old verison of the Forte Compiler (Forte 6 update 2). Builds completed fine through Prevent. In the last month, we've upgraded to Sun Studio 11 (can't go to 12 because we're using Solaris 8). Now I can't get a build to run completely through Prevent. Prevent has a preprocessor that's a bit more strict than most compilers. I'm getting the following errors in a lot of places.
    "/vobs1/cots/ForteDeveloper/SUNWspro/prod/include/CC/rw7/rw/defs.h", line 316: error #101:
    "istream" has already been declared in the current scope
    class _RWCLASSTYPE istream;
    ^
    "/vobs1/cots/ForteDeveloper/SUNWspro/prod/include/CC/rw7/rw/defs.h", line 317: error #101:
    "ostream" has already been declared in the current scope
    class _RWCLASSTYPE ostream;
    ^
    "/vobs1/cots/ForteDeveloper/SUNWspro/prod/include/CC/rw7/rw/defs.h", line 318: error #101:
    "ios" has already been declared in the current scope
    class _RWCLASSTYPE ios;
    ^
    "/vobs1/cots/ForteDeveloper/SUNWspro/prod/include/CC/Cstd/iostream.h", line 4: error #101:
    "istream" has already been declared in the current scope
    using std::istream;
    ^
    "/vobs1/cots/ForteDeveloper/SUNWspro/prod/include/CC/Cstd/iostream.h", line 9: error #101:
    "ostream" has already been declared in the current scope
    using std::ostream;
    ^
    I've googled a bit, and most of the solutions involve someone mistakenly putting "using namespace std" in a header file. I can't find any case where we do this in our code. One specific file that is failing includes one local header, and that header only includes <fstream.h> and <rw/cstring.h>. It seems like RW and std are conflicting with each other somewhere, but I can't figure out where. The compiler command looks like:
    /vobs1/cots/ForteDeveloper/SUNWspro/bin/CC -xildoff -instances=static -mt -g -fsimple
    -xlibmil -xO3 -xtarget=ultra -DUSE_VFU_LL_C -DXRT -KPIC -library=rwtools7_std -x
    libmopt -DRW_MULTI_THREAD -I/vobs1/cots/rogue/workspaces/SOLARIS7/SUNPRO50/12d -c c4i_file_stream.cpp
    We're still pointing at an older version of RW for the DBTools libraries (rebuilt using the compiler's internal Tools++), but we're using the compiler's internal Tools++ headers and libraries for everything else.

    It sounds like Prevent is not processing the #include directives, or possibly some other preprocessing directives, the same way as the Sun C++ compiler.
    The C++ Users Guide documents how #include directives are processed, and in particular the order in which directories are searched.
    [http://docs.sun.com/app/docs/doc/819-5267/bkaso?a=view]
    If Prevent provides similar documentation, you can compare the docs. If you see a difference, maybe Prevent has some options to control the behavior to make it work like the Sun compiler.
    Note: The CC driver passes some -I directives to the C++ compiler proper (ccfe) that do not appear on the CC command line that you write. The exact -I options depend on other options on the command line. Possibly Prevent is not using the same set of implicit -I directives.
    If you can't resolve the #include directive issue, you might need to talk to Coventry tech support.
    We at Sun are happy to work with 3rd-party vendors to help them get their products working correctly on Sun.

  • Difference between Sun Studio 12u1 and Sun Studio Express...

    Hi,
    on my open solaris:
    SunOS opensolaris 5.11 snv_111b i86pc i386 i86pc
    I've installed Sun Studio Express to get the C/C++ compilers. What is the difference with the Sun Studio 12 suite (which is about 200 MB larger than the Express one)? Also I saw there are patches for Sun Studio 12 (last one Jul 20 2009) that might have a fix for a problem I just reported. How can I install such patches?
    Patch: http://sunsolve.sun.co.jp/search/document.do?assetkey=1-21-126498-14-1
    Cheers, Fons.

    Studio Express is a pre-release or beta release. Some features may be missing or not functional. It cannot be patched. It can be replaced by a later release.
    Sun Studio 12 update 1 is an actual, supported release, later than the most recent (as of this date) Studio Express. If you install SS12u1 from packages (using the installer GUI for example), you can download and install patches for it as they become available. (Installations from a tar file cannot be patched; they can be replaced.)

  • Import Manager Usage : Approaches for developing Import file structure and text validations

    Hi Experts,
    We are having 50+ import maps. We have provided option to users to drop the files for data import. Currently Import Manager(7.1 Sp08) does not have capability of pre-import validation of file data such as
    a. file structure - number of columns specific to import map
    b. file text validations - special characters, empty lines, empty cells
    c. Uniqueness of the records in the file
    For this, we are planning to build temporary folder(port specific) in which user drops in the file. We use the custom development to do above mentioned validations and then move the files to actual import ports.
    Please let us know if you have worked on similar requirements and how you have fulfilled such requirements.
    Regards,
    Ganga

    Hi Ganga,
    Assuming you have a well defined xsd and are getting valid xmls from source in the Inbound Port of MDM. Also,you have a Primary key in form of External ID (say).
    So just by making and defining a XSD you get most of what you want in your questions a and b.
    Now if you wish to use PI to drop files in the inbound port then you can build all the validations in PI itself and you would not need Staging table.
    Otherwise,you can have another table (preferably Main table) in the same repository or other dummy repository where records are created on import based on External ID.
    Here you can launch an MDM workflow on import of these records and run assignments to replace unwanted characters and Validations to give error for rejecting some records based on the data quality level desired.Once unwanted characters are removed and data is validated it can be syndicated using a syndication step in the Workflow.So records which fail are not sent and which pass are sent to a outbound port.
    From the outbound port PI or some other job can pick the file from this outbound folder and drop to Inbound folder of the same repository which imports to the required Primary Main table.Here again you have the option to leverage validations in PI and further check if data is fine.
    Once this activity is done you can delete the records from the staging table.
    Thanks,
    Ravi

  • File structure and pictures

    I had the data from my PC transferred over to the Mac. Photos were in a "My pictures" folder and then I had a number of sub folders by event or year, etc. Now everything seems to be in a folder labeled "iPhoto Library" and when I click on it it opens iPhoto, so I can't really see the files themselves, change file names, set up sub-folders, etc. How am I supposed to manage these files? Do I have to do it through iPhoto? My understanding was that changing names and descriptions in iPhoto doesn't actually change the file name. Additionally, I'm not thrilled with the iPhoto organizes things.

    First, change your iPhoto settings so that this doesn't happen again: Go to iPhoto-> Preferences-> Advanced and uncheck Copy items to the iPhoto library. This way, iPhoto leaves your photos alone and you can do whatever you want with them.
    That's not quite true... If you rename a file, move it, change the path to it then iPhoto will lose track of it. If you edit it directly instead of via the iPhoto Interface then iPhoto will not know of the changes, and the view in the iPhoto Window will not reflect what you've done. Precisely the only thing that running in the mode (Referenced mode) will do is allow you to see the actual folders of images.
    Other things to consider:
    1. Importing and deleting pics are more complex procedures. You have to to put the files where they will be stored before importing them. When you delete them you'll need to remove the files from the HD yourself.
    2. You cannot move or rename the files on your system or iPhoto will lose track of them on systems prior to 10.5 and iPhoto 08. Even with the later versions issues can still arise if you move the referenced files to new volumes or between volumes.
    3. Most importantly, migrating to a new disk or computer can be much more complex.
    4. Because iPhoto has no tools for managing Referenced Files, if, for some reason, the path to the photos changes then you could find yourself resolving aliases for +each photo in the Library+ one by one.
    My own opinion:
    I've yet to see a good reason to run iPhoto in referenced mode unless you're using two photo organiser
    here's how to get inside the iPhoto Library bundle. Right-click it and choose Show Package Contents. A new window opens with a bunch of folders. That's the iPhoto library. Links to your photos are under Originals.
    And here's how to actually +use iPhoto+ to access the images:
    *For Users of 10.5 and later*
    You can use any Open / Attach / Browse dialogue. On the left there's a Media heading, your pics can be accessed there. Command-Click for selecting multiple pics.
    Uploaded with plasq's Skitch!
    (Note the above illustration is not a Finder Window. It's the dialogue you get when you go File -> Open)
    You can access the Library from the New Message Window in Mail:
    Uploaded with plasq's Skitch!
    *For users of 10.4 and later* ...
    Many internet sites such as Flickr and SmugMug have plug-ins for accessing the iPhoto Library. If the site you want to use doesn’t then some, one or any of these will also work:
    To upload to a site that does not have an iPhoto Export Plug-in the recommended way is to Select the Pic in the iPhoto Window and go File -> Export and export the pic to the desktop, then upload from there. After the upload you can trash the pic on the desktop. It's only a copy and your original is safe in iPhoto.
    This is also true for emailing with Web-based services. However, if you're using Gmail you can use iPhoto2GMail
    If you use Apple's Mail, Entourage, AOL or Eudora you can email from within iPhoto.
    If you use a Cocoa-based Browser such as Safari, you can drag the pics from the iPhoto Window to the Attach window in the browser.
    *If you want to access the files with iPhoto not running*:
    For users of 10.6 and later:
    You can download a free Services component from MacOSXAutomation which will give you access to the iPhoto Library from your Services Menu. Using the Services Preference Pane you can even create a keyboard shortcut for it.
    For Users of 10.4 and later:
    Create a Media Browser using Automator (takes about 10 seconds) or use this free utility Karelia iMedia Browser
    Other options include:
    1. *Drag and Drop*: Drag a photo from the iPhoto Window to the desktop, there iPhoto will make a full-sized copy of the pic.
    2. *File -> Export*: Select the files in the iPhoto Window and go File -> Export. The dialogue will give you various options, including altering the format, naming the files and changing the size. Again, producing a copy.
    3. *Show File*: Right- (or Control-) Click on a pic and in the resulting dialogue choose 'Show File'. A Finder window will pop open with the file already selected.
    *To use a 3rd Party Editor*
    You can set Photoshop (or any image editor) as an external editor in iPhoto. (Preferences -> General -> Edit Photo: Choose from the Drop Down Menu.) This way, when you double click a pic to edit in iPhoto it will open automatically in Photoshop or your Image Editor, and when you save it it's sent back to iPhoto automatically. This is the only way that edits made in another application will be displayed in iPhoto.
    Note that iPhoto sends a copy+ of the file to Photoshop, so when you save be sure to use the Save command, not Save As... If you use Save As then you're creating a new file and iPhoto has no way of knowing about this new file. iPhoto is preserving your original anyway.
    Regards
    TD

  • Redhat 9 and Java studio enterprise 7

    Did anybody install Java Studio Enterprise 7 with sun application server 7 on Red Hat 9.0 Linux.
    If so do I need to set some special parameters in the environment ?
    Since I tried, but became unsuccessful.
    Thank you in advance.

    Truly speaking Java Studio Enterprise 7 is supported on the following platforms:
    - Microsoft Windows XP Professional SP1a
    - Microsoft Windows 2000 Professional SP4
    - Solaris� 9 OS Update 7 (x86)
    - Solaris� 8 OS Update 7 (UltraSPARC� III)
    - Solaris� 9 OS Update 7 (UltraSPARC� III)

  • GCC and Sun Studio 12 compatibility ?

    What is the level of compatibility between g++ for SPARC systems and the Sun Studio C++ compilers for SPARC ?
    If I build a shared library using g++ for SPARC is it safe for objects built with the Sun C++ compilers to use these shared libraries ?

    They are completely incompatible, there is no way you can mix them.

  • File Structure and DW FTP

    Hello,
    I find I like Dreamweaver's built in FTP client.  So long as I FTP from the same local directory, the files transfer to the root at the server.  However, when I FTP from another directory, DW will construct a tree on the remote server to mirror the source of that file on the local machine.  It would suit me much better if DW simply transfered that / those files to whatever directory I have opened on the server.  If that's possible, would you mind directing me towards that toggle or selector(s)?  Thank you,
    Peace,
    Dave

    I do not choose to organize the files on my machine in a way that suits the software I use.  I have a directory for the files that are involved in my webwork.  Sub-directories for text, image, and related files come next, and so forth.  I don't want a single directory with a jumbled mass of files, related only by project name.  Additionally, I frequently use the same files over and over, such as my logo, banner, and related objects.  Finally, some files are served from my server, while others at servers of a more proprietary nature, like eBay, and Amazon, and those do require absolute links.
    So, the question isn't about how I compensate for anything.  Rather, it's to know why a free piece of software like FileZilla is able to accommodate my needs, while a well apointed package like Dreamweaver is not?  A second question; why does it matter to you at all how I do things? You have sufficiently demonstrated your contempt in another thread.  My question was answered in the second post of this one, and all this supplementary drivel this morning is mere passive aggression, and it is pointless.
    Peace,
    Dave

Maybe you are looking for