Distribute to Multiple Locations on File System

Hi,
Is there any way to schedule a publication to multiple file system directories at the same time?  When I use the "File System" destination option, it only allows for one directory. I'd like to be able to publish to multiple directories without having to create multiple publications or editing one's destination over and over again.
Thanks,
Cameron

not inside BOE, you'll have to do this outside  - maybe using a cron/batch job that will copy files from one location to others
p.s.
And Why? do you need to do this is also interesting

Similar Messages

  • Best Practice for Distributed TREX NFS vs cluster file systems

    Hi,
    We are planning to implement a distributed TREX, using RedHat on X64, but we are wondering which could be the best practice or approach to configure the "file server" used on the TREX distributed environment. The guides mention file server, that seems to be another server connected to a SAN exporting or sharing the file systems required to be mounted in all the TREX systems (Master, Backup and Slaves), but we know that the BI accelerator uses OCFS2 (cluster file systems) to access the storage, in the case of RedHat we have GFS or even OCFS.
    Basically we would like to know which is the best practice and how other companies are doing it, for a TREX distributed environment using either network file systems or cluster file systems.
    Thanks in advance,
    Zareh

    I would like to add one more thing, in my previous comment I assumed that it is possible to use cluster file system on TREX because BI accelerator, but maybe that is not supported, it does not seem to be clear on the TREX guides.
    That should be the initial question:
    Aare cluster file system solutions supported on plain TREX implementation?
    Thanks again,
    Zareh

  • Getting XML data of SSIS Package (which is located in File System) from a query

    Hi,
    We have around 200 SSIS packages deployed to File System. These pacakges are being scheduled using SQL Agent job.
    So, currently I have a list of all the packages and their respective File Path.
    Note : **They are NOT deployed to Server**.
    Now, my task is to search all the packages which has a specific connection in it.
    From the below blog, I got to know we can achieve this by querying the XML data of the SSIS Package.
    https://www.simple-talk.com/sql/ssis/-exploring-ssis-architecture-and-execution-history-through-scripting/.
    But in the above article they are getting the XML definition of the Package from msdb(as it was deployed to server).
    In my requirement, SSIS Packages are in File System. Is there any way to to get the XML definition of these packages using a query.
    Please let me know if the above requirement is not clear or if any further information is required.
    Thanks
    Raksha

    Hi Visakh & Praveen,
    Thanks for the response. 
    @Visakh: The query provided by you gives the information about all the connection strings in the package.
    @Praveen: The query in the link provided by you lists the values of all the properties of the package like MaxerrorCount, CreatorName, CreatorDate and other properties.
    I think, we may have to slightly modify this query to get the connection string values
    Praveen, nice to hear from you after a long time and that too in this forum !! :) Hope you are doing good.
    Thanks
    Raksha

  • Cache of a File System

    Would appreciate it if someone could validate my thought process:
    I'm looking to put together a large distributed cache and use the file system (shared by all cluster nodes) as long-term storage. My approach has been to: use a distributed scheme with POF serialization, and to create a custom cache store in order to write-behind to Berkeley db
    In order to speed up the process of seeding/bulk-loading the cache at start-up I thought that it would be a good idea to partition the data and to write to different files/directories based on my own fairly primitive partitioning strategy, and then to load the data similar to this example: [Performing Distributed Bulk Loading|http://coherence.oracle.com/display/COH35UG/Pre-Loading+the+Cache#Pre-LoadingtheCache-PerformingDistributedBulkLoading]
    I later realized that coherence also provides a bdb-store-manager.
    My question is:
    1. Does the overall approach make sense? I'd like to optimize this as much as possible so if you can think of a better/faster way then please let me know
    2. Can I achieve the same by just using the bdb-store-manager and not implementing my own Cache Store?
    Thanks,
    Kevin

    user10766240 wrote:
    Would appreciate it if someone could validate my thought process:
    I'm looking to put together a large distributed cache and use the file system (shared by all cluster nodes) as long-term storage. My approach has been to: use a distributed scheme with POF serialization, and to create a custom cache store in order to write-behind to Berkeley db
    You have to have a separate Berkeley DB for each partition they need to be "mounted" and "unmounted" when partitions move between nodes.
    You also have to write a partition listener handling the PARTITION_TRANSFER_* events with "mounting" and "unmounting" the per-partition Berkeley DB instance appropriately as the partition is transferred.
    In order to speed up the process of seeding/bulk-loading the cache at start-up I thought that it would be a good idea to partition the data and to write to different files/directories based on my own fairly primitive partitioning strategy, and then to load the data similar to this example: [Performing Distributed Bulk Loading|http://coherence.oracle.com/display/COH35UG/Pre-Loading+the+Cache#Pre-LoadingtheCache-PerformingDistributedBulkLoading]
    It should be possible to adapt that example to your needs, e.g. get the keys for all entries in the same partition should be fairly easy by going to the appropriate per-partition BDB instance.
    I later realized that coherence also provides a bdb-store-manager.
    My question is:
    1. Does the overall approach make sense? I'd like to optimize this as much as possible so if you can think of a better/faster way then please let me knowYes.
    2. Can I achieve the same by just using the bdb-store-manager and not implementing my own Cache Store?
    Thanks,
    KevinNo. The BDB store manager is not able to initialize the cache from a non-empty BDB database. It recreates an empty BDB database on restart, thereby you would lose all your data before you could have preloaded it, even if you could configure the BDB store manager to access your data files.
    Best regards,
    Robert

  • Distributing the GOA to multiple locations in the same backend

    Hello Folks,
    We are implementing SRM 5.0 (SAPKIBKT10 Support Package level) with Extended Classic Scenario, with only one backend system (R3 V4.6b) integrated with the SRM system.
    Queries:
    At GOA header - Distribution,  Example is shown below....
    Backend Purchase Organization 1 - Location A
    Backend Purchase Organization 1 - Location B
    When we distribute one GOA to the backend with multiple locations for the same Purchase organiation.
    In the backend system two idocs are getting generated with the same contract number. While processing the first Idoc the system is creating the contract but while processing the second idoc the system is changing the previously created contract, instead of creating a new contract.
    So, in standard does the system creates actually creates two contract numbers for two different locations?
    Can somebody throw some light on the logic of the distribution
    Regds
    Amith
    Edited by: Amith Varma on Feb 26, 2008 4:51 AM

    Hi Amith,
    this is normal behaviour of the system in the case POrgs are set up accross company codes. In the Contract IDoc the Comany Code is derived from the location. When the system finds the same company code, it does not create an additional contract, but just updates the already distributed contract.
    Cheers,
    Claudia

  • How to package a .war file for multiple locations...

    Sorry if this is in the wrong section, but I wasn't sure where to place it.
    We have a webapp (.war) that we would like to package to distribute to multiple remote locations. We have to assume that the people installing this webapp have no Java/J2EE experience and will not have the technical know-how to be able to handle packaging the .war file themselves.
    They do, however, need to be able to make configuration changes to the JDBC settings in the web.xml, the logging settings in log4j.xml, and configuration changes in the context.xml.
    Since these files reside inside the .war, what is the best way to package this application so they can make changes to those files and package it to deploy on their application servers?
    Thanks!

    Well, the best way is to move that configuration out of those XML files and manage them directly from you application.
    Then you can use the Java Preferences API to actually maintain the configuration data, and that data is stored outside of your application completely in some JRE dependent way (Registry on Windows, and the users home directory on Unix for Sun's JRE).
    You can also modify your webapp to manage the configuration directly and have it rewrite the XML files itself. The Java Preferences API is better just because the config changes will survive if you happen to update the WAR file with a new version. Of course, you can still leverage the Preference API, get your "defaults" from the supplied XML files in the WAR, and then if your application detects a version change (the one in the Preferences is different from the one in the WAR), the app can update the XML files, and pop up a page telling the user to restart the application to see the new changes.
    Finally, you can also write a Swing app that presents the few options that the user might need, and then it has the smarts to update the WAR file for them (it's simply a ZIP file and Java can read/write those). That way you have a simple wizard they can play with.
    Also, for things like log4j, you can bundle that seperate from the WAR, and let them place it either on the CLASSPATH of their container, or in one of their shared directories (like shared/classes on Tomcat for example).
    I'm fond of the Preferences technique with a Webapp configuration, as you can make your WAR almost plug and play. They plop in the war, app fires up, they hit the URL and there's a config page staring at them.
    But all of these options require work outside of your application.
    You can also perhaps get one of those Installer programs, and right an installer for them. You can explode the WAR in to any directory, and then have the container deploy it from there.

  • Aim to process all files in folders on desktop to run through photoshop and save in multiple locations

    Aim to process all files in folders on desktop to run through photoshop and save in multiple locations
    Part one:-
    Gather information from desktop to get brand names and week numbers from the folders
    Excluding folders on desktop beginning with "2" or "Hot"
    Not sure about the list of folders
    but I have got this bit to work with
    set folderPath to "Hal 9000:Users:matthew:Desktop:DIVA_WK30_PSD" --<<this would be gained from the items on the desktop
    set {oldTID, my text item delimiters} to {my text item delimiters, ":"}
    set folderName to last text item of folderPath
    set my text item delimiters to "_WK"
    set FolderEndName to last text item of folderName
    set brandName to first text item of folderName
    set my text item delimiters to "_PSD"
    set weekNumber to first text item of FolderEndName
    set my text item delimiters to oldTID
    After running this I have enough information to create folders in multiple locations, (i need to know where they are so that photoshop can later save them in those multiple locations
    So I need the following folders created
    Locally
    Hal 9000:Users:matthew:Pictures:2011-2012:"WK" + weekNumber
    Hal 9000:Users:matthew:Pictures:2011-2012:"WK" + weekNumber: brandName
    Hal 9000:Users:matthew:Pictures:2011-2012:"WK" + weekNumber: brandName: brandName + "_WK" + weekNumber + "_LR" --(Set path for Later)PathA
    Hal 9000:Users:matthew:Pictures:2011-2012:"WK" + weekNumber: brandName: brandName + "_WK" + weekNumber + "_HR"--(Set path for Later)PathB
    Network
    Volumes:GEN:Brands:Zoom:Brands - Zoom:Upload Photos:2012:"Week" + weekNumber
    Volumes:GEN:Brands:Zoom:Brands - Zoom:Upload Photos:2012:"Week" + weekNumber:brandName + "_WK" + weekNumber + "_LR"  --(Set path for Later)PathC
    Volumes:GEN:Website_Images --(no need to create folder just set path)PathD
    FTP (Still as a normal Volume) So like another Network
    Volumes:impulse:"Week" + weekNumber
    Volumes:impulse:"Week" + weekNumber:Brand
    Volumes:impulse:"Week" + weekNumber:Brand:brandName + "_WK" + weekNumber + "_LR"  --(Set path for Later)PathE
    Volumes:impulse:"Week" + weekNumber:Brand:brandName + "_WK" + weekNumber + "_HR"  --(Set path for Later)PathF
    I like to think that is end of Part 1
    Part 2
    Take the images  (PSD's) from those folders relevant to the Brand then possibly run more applescript that opens flattens and then saves it in the locations above.
    For example….
    An image in folder DIVA_WK30_PSD will then run an applescript in Photoshop, lets call it DivaProcessImages within this we then save to PathA, PathB, PathC, PathD, PathE, PathF the folder path of C should therefore look like this
    Volumes:GEN:Brands:Zoom:Brands - Zoom:Upload Photos:2012:Week30:DIVA_WK30_LR and of course save the image as original filename.
    Then from the next folder
    An image in folder Free_WK30_PSD will then run an applescript in Photoshop, lets call it FreeProcessImages within this we then save to PathA, PathB, PathC, PathD, PathE, PathF the folder path of C should therefore look like this
    Volumes:GEN:Brands:Zoom:Brands - Zoom:Upload Photos:2012:Week30:Free_WK30_LR and of course save the image as original filename.
    The photoshop applescript i'm hoping will be easier as it should be a clearer step by step process without any if's and but's
    Now for the coffee!!

    Hi,
    MattJayC wrote:
    Now to the other part, where each folder was created (and those that already existed) how do I set them as varibles?
    For example,
    set localBrandFolder_High_Res to my getFolderPath(brandName & "_WK" & weekNumber & "_HR", localBrandFolder)
    This line was used to create more than one folder as it ran though the folders on the desktop. The next part is I will need to reference them to save files to them.
    You can use a records
    Examples
    if you want the path of localBrandFolder_High_Res  of "Diva", if "Diva" is the second folder of the Desktop
    You get the path with this : localBrandFolder_High_Res of record 2 of myRecords
    if you want the path of localWeekFolder  in the first folder of the Desktop
    You get the path with this : localWeekFolder of record 1 of myRecords
    Here is the script
    set myRecords to {}
    set dtF to paragraphs of (do shell script "ls -F ~/Desktop | grep '/' | cut -d'/' -f1")
    repeat with i from 1 to number of items in dtF
        set this_item to item i of dtF
        if this_item does not start with "2_" and this_item does not start with "Hot" then
            try
                set folderPath to this_item
                set {oldTID, my text item delimiters} to {my text item delimiters, ":"}
                set folderName to last text item of folderPath
                set my text item delimiters to "_WK"
                set FolderEndName to last text item of folderName
                set brandName to first text item of folderName
                set my text item delimiters to "_PSD"
                set weekNumber to first text item of FolderEndName
                set my text item delimiters to oldTID
            end try
            try
                set this_local_folder to "Hal 9000:Users:matthew:Pictures:2011-2012"
                set var1 to my getFolderPath("WK" & weekNumber, this_local_folder)
                set var2 to my getFolderPath(brandName, var1)
                set var3 to my getFolderPath(brandName & "_WK" & weekNumber & "_LR", var2)
                set var4 to my getFolderPath(brandName & "_WK" & weekNumber & "_HR", var2)
                --set up names to destination folders and create over Netwrok including an already exisiting folder
                set this_Network_folder to "DCKGEN:Brands:Zoom:Brand - Zoom:Upload Photos:2012:"
                set var5 to my getFolderPath("WK" & weekNumber, this_Network_folder)
                set var6 to my getFolderPath(brandName, var5)
                set var7 to my getFolderPath(brandName & "_WK" & weekNumber & "_LR", var6)
                set website_images to "DCKGEN:Website_Images:"
                --set up names to destination folders and create over Netwrok for FTP collection (based on a mounted drive)
                set this_ftp_folder to "Impulse:"
                set var8 to my getFolderPath("Week" & weekNumber, this_ftp_folder)
                set var9 to my getFolderPath(brandName, var8)
                set var10 to my getFolderPath(brandName & "_WK" & weekNumber & "_LR", var9)
                set var11 to my getFolderPath(brandName & "_WK" & weekNumber & "_HR", var9)
                set end of myRecords to ¬
      {localWeekFolder:var1, localBrandFolder:var2, localBrandFolder_Low_Res:var3, localBrandFolder_High_Res:var4, networkWeekFolder:var5, networkBrandFolder:var6, networkBrandFolder_Low_Res:var7, ftpWeekFolder:var8, ftpBrandFolder:var9, ftpBrandFolder_Low_Res:var10, ftpBrandFolder_High_Res:var11}
            end try
        end if
    end repeat
    localBrandFolder_High_Res of record 2 of myRecords -- get full path of localBrandFolder_High_Res in the second folder of Desktop
    on getFolderPath(tName, folderPath)
        tell application "Finder" to tell folder folderPath
            if not (exists folder tName) then
                return (make new folder at it with properties {name:tName}) as string
            else
                return (folder tName) as string
            end if
        end tell
    end getFolderPath

  • Multiple Locations in Receiver File channel

    Hi Experts,
                      Can i give multiple locations to pick the file in receiver filec hanel depending on some condition.
    Thanks
    Ajay Garg

    Hi,
    Definately you can....
    in sender Communication channel check the ADVANCED SELECTION FOR SOURCE FILES.......
    Thanks,
    Madhu

  • Multiple files picked and to zipped at target file system

    Hi
    In my scenario multiple files to be picked from FTP server by using Sender File Adapter then it has to be Zipped at the receiver file system.
    Please help me in this scenario.
    Sri

    Hi,
    >> In my scenario multiple files to be picked from FTP server by using Sender File Adapter then it has to be Zipped at the receiver file system.
    First you need to pick multiple files by using advance selection source at file sender adapter.
    Once it' is  picked use module name is: AF_Modules/PayloadZipBean in receiver file system for zipping files
    Pls ref the blog :
    /people/stefan.grube/blog/2007/02/20/working-with-the-payloadzipbean-module-of-the-xi-adapter-framework
    Abid Mazhar.

  • Delete multiple files from multiple locations using config file

    Hello All,
    I am fairly new to .net and I have to make a windows application to delete files older than 90 days from multi ple locations. these multiple locations cant be hard coded and have to be configurable, also the number of days to delete the files have to be
    configurable.
    I think the answer to this would be an app.config file. but dont know how to use the same.
    can anyone tell me about this and if possible help me out with a code snippet.?

    Add settings like picture below.  Add the string by press the double dot icon.
    The use the code below
    static void Main(string[] args)
    int deleteDays = Properties.Settings.Default.DeleteDays;
    StringCollection folders = Properties.Settings.Default.Folders;
    foreach (string folder in folders)
    DirectoryInfo info = new DirectoryInfo(folder);
    FileSystemInfo[] deleleteFiles = info.EnumerateFiles()
    .Where(x => x.LastWriteTime < DateTime.Today.AddDays(deleteDays))
    .ToArray() ;
    foreach(FileSystemInfo file in deleleteFiles)
    File.Delete(file.FullName);
    jdweng

  • I want my file adapter to poll multiple locations in soa 10g

    Hi All
    I had a requirement that "I want my file adapter to poll multiple locations in soa 10g".I came to know that in soa 11g it is possible.
    I want this to be done in soa 10g.
    plz help me.
    thanks,
    krishna

    You can have pick activity and have separate ftp adapter for each folder, but you will be using the same jndi in all pick activity.
    In that way you can have multiple ftp adapter but uses the same jndi by increasing the performance.
    User will be configured to a home location, example as below.
    say a ftp user "adminuser" is configured to a default folder c:/ftp_files
    You can create multiple folders under the same folder like below.
    c:/ftp_files/location_a
    location_b
    location_c
    location_d
    For each location have a separate ftp adatper, but all of them uses the same jndi while defining the path you have to change according to the above folder.
    Let me know how this goes.
    - It is considered good etiquette to reward answerers with points (as "helpful" - 5 pts - or "correct" - 10pts).
    Thanks,
    Vijay

  • BEA-141230 Could Not locate descriptor file for System Resource: DS name

    Guys,
    When I start the Admin and Managed server ..I get below warning message.
    *<Warning> <Management> <BEA-141230> <Could Not locate descriptor file for System Resource : datasource1.>*
    I know that weblogic server is unable to locatie datasource1.xml file at <SOA_HOME>/user_projects/domains/base_domain/config/jdbc
    Is there any way to to get rid off from this error? I tried to delete datasource from weblogic Admin console, but get below error:
    java.lang.NullPointerException
    All changes have been activated. No restarts are necessary.
    Errors were encountered while performing this delete operation.
    Thanks,
    Ab

    Are you starting servers with the root user? make sure that using whichever user you are starting server, it has full access on MW_Home and it's sub-directories. You may consider raising a support SR.
    Regards,
    Anuj

  • CAF service: import web services using Remote Location / File System

    Hallo Experts
    I use the NetWeaver DevStudio 7.1 SP5.
    I have a CAF project and in this project I want to import one WebService as external service. I have chosen the option "Remote location/ File System" in the import web service wizard, then I enter the URL in the following step. But at the end of wizard I got the error "WSDL could not be downloaded because Server returned HTTP response code 403 for URL". The web service I've deployed and tested on the server. It runs well. Any Hints?
    Thanks in advance
    Kind Regards
    Ping

    Hallo
    I found out: if I enter one URL "http://xxxx.xx.ch:51000/ExampleService/ExampleBean?wsdl&mode=ws_policy". Afterwards I got the error "http://xxxx.xx.ch:51000/ExampleService/ExampleBean?wsdlmode=ws_policy". & ist not correct displayed. Is this the problem that I got 403 Error... It is a bug in SP5
    Thanks if you can tell me some work-arround.
    Ping

  • File System Location properties: view path (OWB 11..2.0.1 Windows)

    Newbee question:
    From the tutorial "Handling Flat File and COBOL Copybook Sources in Mappings", I imported the metadata for ETL_OWNER. It contains file locations such as SRC_FLAT_FILES_LOCATION but I cannot find any way to view its path in the file system. I created my own file locations where I can specify a path in the Create File System Location wizard, but afterwards, I can neither inspect nor change the path.
    Thanks for help
    Sebastian

    Thank's for help, I am simply not used to the user interface. The Connections Explorer seems to have disappeared in 11gR2 but the Locations node still exists, listing all file locations. However, the property inspector does not display the file's path, somewhat counterintuitively. Double-klicking the file location opens the Edit window containing the path.

  • I want the location of the file system in SharePoint.

    Hi All,
    In SharePoint some files are stored in file systems and some files are stored in content data base.
    Now i want the location of the file system in SharePoint.
    Thanks in advance.

    Hi Anil,
    In SharePoint 2010 files are stored at
    C:\Program Files\Common Files\Microsoft Shared\Web Server Extensions\14. This directory represents core SharePoint files
    Below link will help understand the14 hive structure in SharePoint 2010
    http://www.learningsharepoint.com/2010/06/07/14-hive-and-other-sharepoint-2010-directories/
    http://mysharepointwork.blogspot.in/2010/02/sharepoint-14-hive-directory-structure.html
    Thanks
     Somnath Matere

Maybe you are looking for