Best approach for storing XML files long term

We have a legal requirement to store the XML files that relate to SAP billing documents for 10 years. (This is a "digital seal" requirement in Mexico.) The files will be sent to SAP server by the third party software. We'll attach them as GOS objects to the billing documents, but we need another option for safekeeping the files for 10 years. Files need to be accessible somewhat easily if there is an audit down the road. We expect to have about 500 files per month, 5k each.
The simplest solution that comes to mind is just to copy the files to a designated folder (or rather several folders by year or month) and backing it up regularly, but I'm wondering if there could be another solution.
Has anyone had a similar requirement and what approach would you recommend? I'm also checking with our Basis support, but thought I'd ask here as well.
Thank you.

You may also check here: http://help.sap.com/saphelp_nw70/helpdata/en/26/72a1376f6cd903e10000009b38f8cf/content.htm

Similar Messages

  • Best method for keeping iMovie projects long term?

    I read through 30 pages of posts as I think this is a pretty basic question, but didn't see it directly addressed, so please forgive me...
    Some of my movies are mammoth in size (as much as 30GB for 1 movie), which I've learned are due to iMovie handing onto full clip size even when only 10 seconds of it makes it into the final film...
    I am starting to move the movies off the iMac onto an external HD, but don't know what is the most economical way to keep a copy digitally without sacrificing a lot of quality? Full Quality QT? as iDVD project instead of iMovie project?
    Any suggestions are most appreciated. Thanks to all.
    Message was edited by: toddleu

    Hello Toddieu,
    I am sure that there are better ways to store your projects, but I have struggled with the exact same question that you have and here are my solutions. I like iMovie 6, but I hate the "non-destructive" editing which makes your projects huge. So, this may seem strange, but I start my movies in iMovie 4. I import all of my raw footage and then do the first round of crude editing which, as you mention, sometimes reduces a 5 minute clip, down to a couple of 10 or 20 second clips. By the time I am done with this process, I usually reduce an hour's worth of "raw" footage down to around 30 to 40 minutes worth of "good stuff". Then I close iMovie 4, copy the movie file (just to be safe), and then open the copy using iMovie 6. When you do this, iMovie 6 will tell you that this movie will be permanently changed to the new format, which is what you want, so click OK. If the movie opens correctly, and everything looks good, then I continue editing the rest of my movie using iMovie 6 (fine cuts, transitions, titles, music, effects, chapters, etc). In the end, using this technique, an hour's worth of edited video uses about 12 gbs of HardDrive. I have several external HDs and I store all of my finished movies on these externals as iMovies (full quality). You don't want to save as DVD projects, because these are compressed files so you will lose some quality.
    Like I said, I am sure there are easier ways to do this, and I am interested to hear what other's are doing, but this works for me.
    Lincoln

  • XML SCHEMA registration for XML TYPE (storing XML files in Oracle 10g)

    I have created the XML Schema for the XML file stored in Oracle 10g and also added this Schema into the database. I have related that schema with the column in the table which contains the XML file. When i execute the query to fetch the data from the stored file i am getting a blank resultset. Is registering the XML Schema is necessary, if yes then please let me know the process of doing it. I have tried following steps to register Schema, but it is not working
    Step1:
    DECLARE
    v_return BOOLEAN;
    BEGIN
    v_return := dbms_xdb.createFolder('/home/');
    v_return := dbms_xdb.createFolder('/home/DEV/');
    v_return := dbms_xdb.createFolder('/home/DEV/xsd/');
    v_return := dbms_xdb.createFolder('/home/DEV/messages/');
    v_return := dbms_xdb.createFolder('/home/DEV/employees/');
    COMMIT;
    END;
    STEP 2:
    Connecting To XML DB
    Step3:
    Register XML schema
    I am failing to execute step number 2 and hence not able to register the schema also.

    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by sudeepk:
    If a java exception is thrown probably during ur install u might have forgotten
    grant javauserpriv to scott;
    grant javasyspriv to scott;
    Thanks
    [email protected]
    <HR></BLOCKQUOTE>
    Thank you!!!

  • Best Approach for Creating an Extract File?

    Now that we are switching to PeopleSoft, I need to create a file each week based upon data in PeopelSoft that is modified before being written to the file. What is considered the best way to do this? Write some PS SQR, App Engine program? Batch program in Java (I'd be abel to do this the fastest but maybe it is not the best choice). I don't really know what to do, as I've never extracted data from PS before. Thanks.
    Ken

    I recommend creating an Application Engine and using a File Layout.
    Here is the [PeopleBooks information on File Layouts with the File API|http://download.oracle.com/docs/cd/E13292_01/pt849pbr0/eng/psbooks/tpcr/htm/tpcr17.htm#g037ee99c9453fb39_ef90c_10c791ddc07__1454] . PeopleBooks has some pretty decent [code examples|http://download.oracle.com/docs/cd/E13292_01/pt849pbr0/eng/psbooks/tpcr/htm/tpcr17.htm#g037ee99c9453fb39_ef90c_10c791ddc07__1429] . In particular, look at the WriteRecord example.
    One of the advantages to a file layout is that you just define what fields you want in the file, and the PeopleCode takes care of formating the file. To create the file layout definition, you will be adding a new definition in Application Designer --> File > New ... File Layout. Here is [some information in PeopleBooks|http://download.oracle.com/docs/cd/E13292_01/pt849pbr0/eng/psbooks/tapd/htm/tapd21.htm#g037ee99c9453fb39_ef90c_10c791ddc07__7705] about that part. You set the file type (CSV, Flat, or XML) on the File Layout properties (click the properties button on the toolbar). If you have a Record definition in the system that resembles the format (# of fields) in the file, you can just insert that record into the File Layout. Otherwise, you have to insert the fields one by one. For each field, double click on the field name to set the properties for the field such as the position in a flat file or the tag name for an XML file.
    In the Application Engine, you will do most of the work in a PeopleCode action. If you don't already have all of the data in one table, I recommend creating a Staging table (staging record). Then, add a step to the App Engine program and a SQL action or two that will load all your data into that one staging table.
    In your PeopleCode action, you will use PeopleCode's SQL object to loop through the rows of the staging table. This is in the WriteRecord example, but here is the [SQL object documentation|http://download.oracle.com/docs/cd/E13292_01/pt849pbr0/eng/psbooks/tpcr/htm/tpcr39.htm#g037ee99c9453fb39_ef90c_10c791ddc07__1848] for more information.
    I hope I have pointed you to enough resources to help you figure it out. Post back more questions if anything is unclear or you get stuck.

  • What is the Best Approach for System Defined Fields and Default values

    Hi ,
    Please let me know that what can be the best approach for providing a Default value for the System defined fields when creating a User and How can we hide the System defined fields at the time of User creation

    You cannot provide default values for any attributes defined in the FormMetaData.xml file. You can only provide default values for fields defined in the User Defined Fields and supply a default value.
    You can using entity adapters to populate some of the values, but you must supply an Organization because there is an entity adapter that you cannot modify that verifies the organization name.
    -Kevin

  • Best approach for RFC call from Adapter module

    What is the best approach for making a RFC call from a <b>reciever</b> file adapter module?
    1. JCo
    2. Is it possible to make use of MappingLookupAPI classes to achieve this or those run in the mapping runtime environment only?
    3. Any other way?
    Has anybody ever tried this? Any pointers????
    Regards,
    Amol

    Hi ,
    The JCo lookup is internally the same as the Jco call. the only difference being you are not hardcoding the system related data in the code. So its easier to maintain during transportation.
    Also the JCO lookup code is more readable.
    Regards
    Vijaya

  • Best Practice for storing PDF docs

    My client has a number of PDF documents for handouts that go
    with his consulting business. He wants logged in users to be able
    to download the PDF docs for handouts at training. The question is,
    what is the 'Best Practice' for storing/accessing these PDF files?
    I'm using CF/MySQL to put everything else together and my
    thought was to store the PDF files in the db. Except! there seems
    to be a great deal of talk about BLOBs and storing files this way
    being inefficient.
    How do I make it so my client can use the admin tool to
    upload the information about the files and the files themselves,
    not store them in the db but still be able to find them when the
    user want's to download them?

    Storing documents outside the web root and using
    <cfcontent> to push their contents to the users is the most
    secure method.
    Putting the documents in a subdirectory of the web root and
    securing that directory with an Application.cfm will only protect
    .cfm and .cfc files (as that's the only time that CF is involved in
    the request). That is, unless you configure CF to handle every
    request.
    The virtual directory is no safer than putting the documents
    in a subdirectory. The links to your documents are still going to
    look like:
    http://www.mysite.com/virtualdirectory/myfile.pdf
    Users won't need to log in to access these documents.
    <cfcontent> or configuring CF to handle every request
    is the only way to ensure users have to log in before accessing
    non-CF files. Unless you want to use web-server
    authentication.

  • Best approach for uploading document using custom web part-Client OM or REST API

    Hi,
     Am using my custom upload Visual web part for uploading documents in my document library with a lot of metadata.
    This columns contain single line of text, dropdownlist, lookup columns and managed metadata columns[taxonomy] also.
    so, would like to know which is the best approach for uploading.
    curretnly I am  trying to use the traditional SSOM, server oject model.Would like to know which is the best approach for uploading files into doclibs.
    I am having hundreds of sub sites with 30+ doc libs within those sub sites. Currently  its taking few minutes to upload the  files in my dev env. am just wondering, what would happen if  the no of subsites reaches hundred!
    am looking from the performance perspective.
    my thought process is :
    1) Implement Client OM
    2) REST API
    Has anyone tried these approaches before, and which approach provides better  performance.
    if anyone has sample source code or links, pls provide the same 
    and if there any restrictions on the size of the file  uploaded?
    any suggestions are appreciated!

    Try below:
    http://blogs.msdn.com/b/sridhara/archive/2010/03/12/uploading-files-using-client-object-model-in-sharepoint-2010.aspx
    http://stackoverflow.com/questions/9847935/upload-a-document-to-a-sharepoint-list-from-client-side-object-model
    http://www.codeproject.com/Articles/103503/How-to-upload-download-a-document-in-SharePoint
    public void UploadDocument(string siteURL, string documentListName,
    string documentListURL, string documentName,
    byte[] documentStream)
    using (ClientContext clientContext = new ClientContext(siteURL))
    //Get Document List
    List documentsList = clientContext.Web.Lists.GetByTitle(documentListName);
    var fileCreationInformation = new FileCreationInformation();
    //Assign to content byte[] i.e. documentStream
    fileCreationInformation.Content = documentStream;
    //Allow owerwrite of document
    fileCreationInformation.Overwrite = true;
    //Upload URL
    fileCreationInformation.Url = siteURL + documentListURL + documentName;
    Microsoft.SharePoint.Client.File uploadFile = documentsList.RootFolder.Files.Add(
    fileCreationInformation);
    //Update the metadata for a field having name "DocType"
    uploadFile.ListItemAllFields["DocType"] = "Favourites";
    uploadFile.ListItemAllFields.Update();
    clientContext.ExecuteQuery();
    If this helped you resolve your issue, please mark it Answered

  • Design Patterns, best approach for this app

    Hi all,
    i am starting with design patterns, and i would like to hear your opinion on what would be the best approach for this app. 
    this is basically an app for data monitoring, analysis and logging (voltage, temperature & vibration)
    i am using 3 devices for N channels (NI 9211A, NI 9215A, NI PXI 4472) all running at different rates. asynchronous.
    and signals are being processed and monitored for logging at a rate specified by the user and in realtime also. 
    individual devices can be initialized or stopped at any time
    basically i'm using 5 loops.
    *1.- GUI: Stop App, Reload Plot Names  (Event handling)
    *2.- Chart & Log:  Monitors Data and Start/Stop log data at a specified time in the GUI (State Machine)
    *3.- Temperature DAQ monitoring @ 3 S/s  (State Machine)   NI 9211A
    *4.- Voltage DAQ monitoring and scaling @ 1K kS/s (State Machine) NI 9215A
    *5.- Vibration DAQ monitoring and Analysis @ 25.6 kS/s (State Machine) NI PXI 4472
    i have attached the files for review, thanks in advance for taking the time.
    Attachments:
    V-T-G Monitor_Logger.llb ‏355 KB

    mundo wrote:
    thanks Will for your response,
    so, basically i could apply a producer/consummer architecture for just the Vibration analysis loop? or all data being collected by the Monitor/Logger loop?
    is it ok having individual loops for every DAQ device as is shown?
    thanks.
    You could use the producer/consumer architecture to split the areas where you are doing both the data collection and teh analysis in the same state machine. If one of these processes is not time critical or the data rate is slow enough you could leave it in a single state machine. I admit that I didn't look through your code but based purely on the descriptions above I would imagine that you could change the three collection state machines to use a producer/consumer architecture. I would leave your UI processing in its own loop as well as the logging process. If this logging is time critical you may want to split that as well.
    Mark Yedinak
    "Does anyone know where the love of God goes when the waves turn the minutes to hours?"
    Wreck of the Edmund Fitzgerald - Gordon Lightfoot

  • What is the best approach for combining events?

    When I work on a wedding my current workflow involves creating a compound clip for each section of the video (e.g. reception, ceremony, dancing etc). Then I add the compound clip 'sequences' into a single project to add the chapter markers and export to a single master file.
    I like the idea of managing each section in a project rather than a compound clip now that projects are part of the library in 10.1, but is there a good way to combine multiple projects (for each section) into a single master project, or would I still need to copy the contents of each project and paste in the master project?
    Maybe I am best to continue with my current workflow.

    Just saw the discussion title - should have said "What is the best approach for combining projects"?

  • XML publisher report not generating output for huge XML files

    Changed Depreciation Projections Report output type to XML.
    Defined a Data Definition and a new Data Template (RTF) for this report.
    Ran the Depreciation Projection Report to generate the XML output.
    Ran the XML Report Publisher report to generate teh PDF/Excel output of the above report.
    Output generated for smaller XML files. When XML size is big, the program is running for hours without generating the output.
    Teh RTF template is basically a matrix report in which the number of columns in the report is based on the number of periods the report is run for.
    The same is not working in the Desktop version also. The system is hanging when i try to view the preview pdf.
    The XML file size is approximately 33 MB.
    Please let me know if there is any way we can increase the memory size to see the output.
    Thanks,
    Ram.

    for publisher use Category: E-Business Suite

  • What's the best approach for handeling about 1300 connections in Oracle.

    What's the best approach for handling about 1300 connections in Oracle 9i/10g through a Java application?
    1.Using separate schema s for various type users(We can store only relevant data with a particular schema.     Then No. of records per table can be reduced by replicating tables but we have to maintain all data with a another schema     Then we need update two schema s for a given session.Because we maintain separate scheama for a one user and another schema for all data and then there may be Updating problems)
    OR
    2. Using single schema for all users.
    Note: All users may access the same tables and there may be lot of records than previous case.
    What is the Best case.
    Please give Your valuable ideas

    It is a true but i want a solution from you all.I want you to tell me how to fix my friends car.

  • Best approach for IDOC - JDBC scenario

    Hi,
    In my scenarion I am creating sales order(ORDERS04) in R/3 system and which need to be replicated in a SQL Server system. I am sending the order to XI as an IDoc and want to use JDBC for sending data to SQL Server. I need to insert data in two tables(header & details). Is it possible without BPM?  Or what is the best approach for this?
    Thanks,
    Sri.

    Yes, this is possible without the BPM.
    Just create the Corresponding Datatype for the insertion.
    if the records to be inserted are different, then there wil be 2 different datatypes ( one for header and one for detail).
    Do a mutlimapping, where your Source is mapped into the header and details datatype and then send using the JDBC sender adapter.
    For the strucutre of your Datatype for insertion , just check this link,
    http://help.sap.com/saphelp_nw04/helpdata/en/7e/5df96381ec72468a00815dd80f8b63/content.htm
    To access any Database from XI, you will have to install the corresponding Driver on your XI server.
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/3867a582-0401-0010-6cbf-9644e49f1a10
    Regards,
    Bhavesh

  • What are the best approaches for mapping re-start in OWB?

    What are the best approaches for mapping re-start in OWB?
    We are using OWB repository 10.2.0.1.0 and OWB client 10.2.0.1.31. The Oracle version is 10 G (10.2.0.3.0). OWB is installed on Linux.
    We have number of mappings. We built process flows for mappings as well.
    I like to know, what are the best approches to incorportate re-start options in our process. ie a failure of mapping in process flow.
    How do we re-cycle failed rows?
    Are there any builtin features/best approaches in OWB to implement the above?
    Does runtime audit tables help us to build re-start process?
    If not, do we need to maintain our own tables (custom) to maintain such data?
    How did our forum members handled above situations?
    Any idea ?
    Thanks in advance.
    RI

    Hi RI,
    How many mappings (range) do you have in a process flows?Several hundreds (100-300 mappings).
    If we have three mappings (eg m1, m2, m3) in process flow. What will happen if m2 fails?Suppose mappings connected sequentially (m1 -> m2 -> m3). When m2 fails then processflow is suspended (transition to m3 will not be performed). You should obviate cause of error (modify mapping and redeploy, correct data, etc) and then repeat m2 mapping execution from Workflow monitor - open diagram with processflow, select mapping m2 and click button Expedite, choose option Repeat.
    In re-start, will it run m1 again and m2 son on, or will it re-start at row1 of m2?You can specify restart point. "at row1 of m2" - I don't understand what you mean (all mappings run in Set based mode, so in case of error all table updates will rollback,
    but there are several exception - for example multiple target tables in mapping without corelated commit, or error in post-mapping - you must carefully analyze results of error).
    What will happen if m3 fails?Process is suspended and you can restart execution from m3.
    By having without failover and with max.number of errors=0, you achieve re-cycle failed rows to zero (0).This settings guarantee existence only two return result of mapping - SUCCSES or ERROR.
    What is the impact, if we have large volume of data?In my opinion for large volume Set based mode is the prefered processing mode of data processing.
    With this mode you have full range enterprise features of Oracle database - parallel query, parallel DML, nologging, etc.
    Oleg

  • Unable to establish a for loop in custom rtf template for seeded xml file

    Hi Experts ,
    I am Unable to establish a for loop in custom rtf template for seeded xml file. i am using <?:for-each:G_BSALINE?> FORM FIELD .....<?:FOR-EACH?END> .AM i missing something?
    Please help.
    Thanks

    Hi,
    Need more information about your XML data structure; however based on the code you provided, the synax does not look right; you have some colon sign where it is not needed etc... You could use the following (although not quite sure if you want to use Form Filed to group by or just display), but look at the example provided below.
    <?for-each-group:G_BSALINE?> FORM FIELD .....<?end for-each-group?>
    Example code:
    <?for-each-group:G_INVOICE_NUM;./INV_TYPE?>
    xxxxxxxxxx
    <?end for-each-group?>
    Thanks!

Maybe you are looking for

  • Document 4500016031   does not contain any selectable items

    Hi Friends, I am doing MIGO(godds receipt) and adopted following procedure for MIGO. ME5J----->(Entered my Proj. defn---> executed Select material T-20610. Click the "Assign automatically" pushbutton. Assign materials T-20620 and T-20630 the same way

  • Problem with running rhcl.exe in batch file

    Hi, I'm running Robohelp 8 and trying to use rhcl.exe to generate multiple help projects. I'm having a problem with the folder where the rhcl.exe is generating the web help. It seems to ignore the "-o" switch and gets the generation folder from the p

  • Depreciation run with special period

    We require 2 set of FS reports:- 1.) FS as at 14/12/2007 2.) FS as at 31/12/2007 We have use the following period to derive the the reports 1.) Period 12 (Transactions 1st - 14th Dec 2007) 2.) Period 14 ( Transactions 15th - 31st Dec 2007) Q1.)Should

  • Slow login after sleep or slow authentication when click on lock in prefs

    takes about 30 seconds to login after sleep or clicking on the lock during a system preference change 9/16/09 6:33:32 PM com.apple.launchd.peruser.501[120] (com.apple.pbs) Throttling respawn: Will start in 10 seconds 9/16/09 6:34:12 PM System Prefere

  • Actions won't save

    I am using Photoshop Elements 13 and have uploaded paid actions and every time I close or the darn program freezes I have to start all over uploading the actions again! Time consuming and frustrating.  I know someone must have the answer?  Thanks!