Script Logic VS Data Transformation File

Hi all,
   I'm new to SAP BPC. I have knowledge of SAP BW.
   I can see conversion file, which we are referring in data transformation file. which we can use for mapping and conversion of external data into internal data.
  How data transformation file different form script logic? Are we going to refer script logis in Data transformation file for each required dimension?
  Can any of you give me clarity on how to place script logic and data transformation file in BPC data management flow.
  I will really applicate all your help!!!
Thanks
Ben.

Nilanjan,
   I have a another quick question...
  suppose my bpc application has 5 dimensions. Out of the 5 dimensions, 4 dimensions data i'm getting directly from SAP BW. assume 1 dim, i need to extract by doing look up at different table which also reside in BW.
   how to populate data for DIM 5.
   I got your point that data transformation file purely for field mapping. suppose DIM5 if i want to populate from script logic, wht do i need map in Transformation file. I hope you got my point.
   My question if how to populate a DIM in BPC using lookup approach.
Thanks,
Ben.

Similar Messages

  • Data Transformation File

    Hi,
       Can any of you explain me the purpose of Data Transformation File ???
       How to use Script logic based on BADI in SAP BPC? Do we need to call in Default.lgf ???
       If i want to apply some custom logic while populating data using BADI, do i need to call that BADI in Data Tranformation file?
    Thanks,
    Ben.

    Hi Ben,
    There is no possibility of calling userdefined scripts or BADI from transformation files.The main use of transformation file is to map data from source to destination.
    If you require data load based on some filter criteria there are certain functionalities available for transformation file creation which u can find in the below link.
    http://help.sap.com/saphelp_bpc70sp02/helpdata/en/66/ac5f7e0e174c848b0ecffe5a1d7730/frameset.htm
    Hope this helps,
    Regards,
    G.Vijaya Kumar

  • Execute script logic using data manager

    Hi all,
    Could anyone explain me how to run a script logic using the data manager?
    The only way I get it working is adding the script logic into the defaul.lgl file and running the default formulas package via data manager.
    How can I run an specific script logic without using the default.lgf file?
    Thanks and regards,
    Ana

    HI Ana,
    You will have to create a SSIS package that will run the logic for you. A good example of this is the ICELIM package that comes standard with AppShell. The easiest is way for me is to copy the package and renaming it to what you want to call it in the following folder: C:\BPC\Data\Webfolders\<AppSet>\<Application>\DataManager\PackageFiles\System Files
    You will then have to organize the package list and add the package and modify the advanced properties of the package and point it to the correct logic file. I always use the lgf file rather than the lgx file, but it is up to you.
    Lemme know if you need more info.
    Regards,
    Andries

  • "Exp" utility script with current date dmp file format

    Dear All,
    I want to creat a window batch file script "oralce database 11gr2" in such a way that dmp file automatically created in c:\test folder with oralce _sid and current date.
    C:> exp file=c:\test\ora-sid_date_format.dmp owner=test
    Any help would be appreciated.
    Regards,

    Microsoft Windows XP [Version 5.1.2600]
    (C) Copyright 1985-2001 Microsoft Corp.
    C:\Documents and Settings\comp>exp system/sm file=c:\%ORACLE_SID%_%date%_forma
    t.dmp owner=test
    ^C
    C:\Documents and Settings\comp>exp system/sm@test file=c:\%ORACLE_SID%_%date%_
    format.dmp owner=test
    Export: Release 10.2.0.2.0 - Production on Wed May 19 12:41:11 2010
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit
    Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    About to export specified users ...
    . exporting pre-schema procedural objects and actions
    . exporting foreign function library names for user TEST
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions for user TEST
    About to export TEST's objects ...
    . exporting database links
    . exporting sequence numbers
    . exporting cluster definitions
    . about to export TEST's tables via Conventional Path ...
    . . exporting table AGENTSINFO 7 rows exported
    . . exporting table ARCHIVE_SETT 3 rows exported
    . . exporting table ARC_CPAYMENT 0 rows exported
    . . exporting table ARC_CREDIT 2596 rows exported
    . . exporting table ARC_DEBET 4 rows exported
    . . exporting table ARC_DPAYMENT 4 rows exported
    . . exporting table BRANCHESCASH 2 rows exported
    . . exporting table CANCELCOUPON_OPRCODES 17 rows exported
    . . exporting table CHECK_PAYMENTREQUEST 36 rows exported
    . . exporting table CORPCODES 0 rows exported
    . . exporting table CREDIT_MESS 0 rows exported
    . . exporting table CREDIT_PAYMENT 0 rows exported
    . . exporting table CURRENCIES 1 rows exported
    . . exporting table DEBET_MESS 0 rows exported
    . . exporting table DEBET_PAYMENT 0 rows exported
    . . exporting table DIRECTORY 1 rows exported
    . . exporting table MAILSETTINGS 2 rows exported
    . . exporting table PAYCOUPON_OPRCODES 22 rows exported
    . . exporting table PAYMENT_OPRCODES 21 rows exported
    . . exporting table PAYMENT_REQUEST 53 rows exported
    . . exporting table PAYMENT_RESPONSE 53 rows exported
    . . exporting table REP_CANCELCOUPON 0 rows exported
    . . exporting table REP_PAYCOUPON 70 rows exported
    . . exporting table REP_VERIFYCOUPON 121 rows exported
    . . exporting table REQ_CANCELCOUPON 0 rows exported
    . . exporting table REQ_PAYCOUPON 108 rows exported
    . . exporting table REQ_VERIFYCOUPON 124 rows exported
    . . exporting table VERIFYCOUPON_OPRCODES 18 rows exported
    . exporting synonyms
    . exporting views
    . exporting stored procedures
    . exporting operators
    . exporting referential integrity constraints
    . exporting triggers
    . exporting indextypes
    . exporting bitmap, functional and extensible indexes
    . exporting posttables actions
    . exporting materialized views
    . exporting snapshot logs
    . exporting job queues
    . exporting refresh groups and children
    . exporting dimensions
    . exporting post-schema procedural objects and actions
    . exporting statistics
    Export terminated successfully without warnings.
    C:\Documents and Settings\chinar>date
    The current date is: 19.05.2010
    Enter the new date: (dd-mm-yy)
    C:\Documents and Settings\comp>SET ORACLE_SID
    ORACLE_SID=SB
    C:\Documents and Settings\comp>
    and result is SB_19.05.2010_format.dmp

  • SQL insert scripts generation from repository data xml file

    Hi All,
    I want to generate sql insert scripts from repository data xml file which contain <add-item>elements.
    Example: data xml file is contains
    <add-item item-descriptor="user" id="741">
    <set-property name="gender"><![CDATA[female]]></set-property>
    <set-property name="emailStatus"><![CDATA[__NULL__]]></set-property>
    <set-property name="receiveEmail"><![CDATA[__NULL__]]></set-property>
    <set-property name="email"><![CDATA[[email protected]]]></set-property>
    <set-property name="password"><![CDATA[a694cb1d6d011eddc444acb77338c1df6a22bac5de26330e2fe1d20f15f97ed6]]></set-property>
    <!-- rdonly derived <set-property name="abandonedOrderCount"><![CDATA[0]]></set-property> -->
    <set-property name="locale"><![CDATA[de_DE]]></set-property>
    <set-property name="userType"><![CDATA[1]]></set-property>
    <set-property name="autoLogin"><![CDATA[__NULL__]]></set-property>
    <set-property name="lastName"><![CDATA[Springford]]></set-property>
    <set-property name="member"><![CDATA[true]]></set-property>
    <set-property name="dateOfBirth"><![CDATA[7/18/1951 00:00:00]]></set-property>
    <!-- rdonly derived <set-property name="age"><![CDATA[61]]></set-property> -->
    <set-property name="registrationDate"><![CDATA[5/16/1996 00:00:00]]></set-property>
    <set-property name="login"><![CDATA[sally]]></set-property>
    <set-property name="lastActivity"><![CDATA[6/9/1999 00:00:00]]></set-property>
    <set-property name="homeAddress"><![CDATA[741]]></set-property>
    <set-property name="passwordSalt"><![CDATA[sally]]></set-property>
    <set-property name="generatedPassword"><![CDATA[__NULL__]]></set-property>
    <set-property name="firstName"><![CDATA[Sally]]></set-property>
    <set-property name="lastPasswordUpdate"><![CDATA[__NULL__]]></set-property>
    </add-item>
    Thanks
    Siva

    HI Rohan,
    Thanks for reply..
    I have checked the startSQLRepository script by using this we can export or import repository data from xml to repository.But in startSQLRepository there is no attribute to generate sql
    insert scripts from data xmls but there is a option to get the sql scripts(-ouputSQL).

  • Script logic error - Invalid object name 'tblStatus'

    Hi all,
    Running BPC 5.1 SP5 on SQL 2005.
    Has anyone ever seen this error message before? I only get the error message when choosing the run-time option in the package, to honor work status settings. This problem just appeared in the last few days, and I believe that a month or two ago, users were running the same logic (with the same yes-honor-work-status option selected).
    We did recently upgrade from 5.1 SP3 to 5.1 SP5, and I'm wondering if this may be a cause? That's just a hypothesis though.
    Invalid object name 'tblStatus'. in:
    select distinct d.VERSION,d.ENTITY,d.TIMEID
    from #RESULT_333729 d  where signeddata<>0
    and exists (select * from tblStatus s where [APP]='CONSOL' and d.VERSION=s.CATEGORY and d.ENTITY=s.ENTITY and d.TIMEID=s.TIME and [STATUS]>0)
    I understand what it means, and of course there is no table in the appset database tblStatus. Work status locks are kept in tblConsolLock (where Consol is my app name). And in that table, the field name is StatusCode, not Status. So it appears that this is some vestige of code from an earlier version of BPC / O'soft.
    This error occurs in a piece of script logic, that includes both basic SQL WHEN/IS/*REC's and also a *RUNALLOCATION. The error is thrown within the RUNALLOCATION code block. The complete run-up to the error message is shown below.
    -- Calculate difference
    insert into #DIFF_333729 ([ACCOUNT],[AFFILIATE],[CURRENCY],[DATASRC],[ENTITY],[PRODHIER],[TIMEID],[VERSION],signeddata)
    select [ACCOUNT],[AFFILIATE],[CURRENCY],[DATASRC],[ENTITY],[PRODHIER],[TIMEID],[VERSION],amtWHERE as signeddata
    from #WHERE2_333729
    select [ACCOUNT],[AFFILIATE],[CURRENCY],[DATASRC],[ENTITY],[PRODHIER],[TIMEID],[VERSION],sum(signeddata) as signeddata
    into #RESULT_333729 from #DIFF_333729
    group by [ACCOUNT],[AFFILIATE],[CURRENCY],[DATASRC],[ENTITY],[PRODHIER],[TIMEID],[VERSION]
    --Time to calculate and count differences (45459 found):2.5 sec.
    select distinct d.VERSION,d.ENTITY,d.TIMEID
    from #RESULT_333729 d  where signeddata<>0
    and exists (select * from tblStatus s where [APP]='CONSOL' and d.VERSION=s.CATEGORY and d.ENTITY=s.ENTITY and d.TIMEID=s.TIME and [STATUS]>0)
    call 1 completed and data posted in 23.6 sec.
    Run completed in 23.7 sec.
    End time --->11:01:05 AM  -  Date:1/16/2009
    Invalid object name 'tblStatus'. in:
    select distinct d.VERSION,d.ENTITY,d.TIMEID
    from #RESULT_333729 d  where signeddata<>0
    and exists (select * from tblStatus s where [APP]='CONSOL' and d.VERSION=s.CATEGORY and d.ENTITY=s.ENTITY and d.TIMEID=s.TIME and [STATUS]>0)
    Here is the original script logic (from the LGX file -- the LGF file has a whole bunch of variables in it that would only confuse the matter more). It works just fine when the
    *RUN_ALLOCATION @A_1||
    USING/TOTAL|
    ACCOUNT WHAT=659210,659310,659320,659410,659510,659610,659620,659630,659710,659810,659820,659910; WHERE=<<<; USING=PCAYTD510000; TOTAL=<<<|
    PRODHIER WHAT=PRODHIERL1A;WHERE=[PRODHIERLEVEL]='5'; USING=<<<; TOTAL=<<<|
    DATASRC WHAT=PCAASSESSSTAGE1,PCAASSESSSTAGE2; WHERE=PCAASSESSALLOC1; USING=INPUTCONSOL; TOTAL=<<<|
    ENTITY WHAT=[COMPANYCODE]='7470' AND [PCAASSESSDRIVERBP]='PCAYTD510000' AND
      ([PCAASSESSMETHODBP] = 'ALLOCAUTO' OR  [PCAASSESSMETHODBP] = 'ALLOCMANUAL');WHERE=<<<;USING=7470.BASE;TOTAL=<<<|
    Anyone experience this same error message? I searched for SAP support notes and found nothing to do with tblStatus or similar.
    Thanks for your help,
    Tim

    Hi Sorin,
    Thanks, you've confirmed what I suspected.
    Yes, we always process & save apps & dims on any upgrade.
    We've reproduced the problem in all 3 of the environments with the customer's appset. I haven't yet reproduced in ApShell, but will do so & report it as a bug.
    And just out of curiousity -- was the other message you saw here in this EPM forum? Before I posted my question, I searched both here, and in the support notes, for "tblStatus" and a few other terms about work status lock errors, etc., and came up with 0 results.
    Thanks again,
    Tim

  • Automate excel circular references - script logic - BPC

    Hi,
    I need to automate in BPC the calculation of financial debt that contains excel circular references, so you have to enable the calculation Automatic iterations properly to work.
    I would like to automate the process through a script logic.
    I appreciate any help or suggestions.
    Regards,
    Fernando Chávez.

    Dear Fernando Chávez,
    I just want to give you simple suggestions because I don't know your case exactly. The script logic is similiar with SQL script. You could put your logic script in default.lgf or make new file .lgf through BPC Administration Console. If you put it in default.lgf, the SAP BPC will execute it when user sent values through BPC Excel.  You also could run the script logic through Data Manager Packages, its runs depend on user execution.
    If we are talking about circular and complex formulas in Ms.Excel, I suggest you put your calculation in another worksheet and send values through BPC Excel. Before you define a script logic, the number one is you should to consider performance because its really critical.
    Kind Regards,
    Wandi Sutandi

  • How to load several transformation files with a single action

    Hi everybody,
    We are loading data from BI cube into BPC cube. We are working on SAP BPC 7.0 version and we have designed several transformation files in order to load each key figure we need.
    Now, we want to load all the transformation files executing only one action. Which one is the best way to do it?
    We thought that it would be possible to build a single process chain, where we would call the target cube and all the transformation files. In this way, the administrator only has to execute once a package that would execute the process chain. We don't want the administrator to execute several times a package looking for the different transformation files.
    How can we do it? Is there any example or document related to it?
    Any idea out there?
    Kind regards
    Albert Mas

    HI SCOTT,
    I AM FACING A PROBLEM WHEN I RUN 2 ROUNDS IN ONE TRANSFORMATION FILE...
    I need to distribute a source field in to BPC through making 2 conversion files... following is the data
    Transformation file
    *OPTIONS
    FORMAT = DELIMITED
    HEADER = YES
    DELIMITER = ,
    AMOUNTDECIMALPOINT = .
    SKIP = 0
    SKIPIF =
    VALIDATERECORDS=YES
    CREDITPOSITIVE=YES
    MAXREJECTCOUNT=
    ROUNDAMOUNT=
    CONVERTAMOUNTWDIM=ZOUTPUT
    *MAPPING
    CATEGORY=*NEWCOL(ACT)
    PAO=0COSTCENTER
    TIME=0FISCYEAR
    ZOUTPUT=0FUNDS_CTR
    SIGNEDDATA=0DEB_CRE_LC
    *CONVERSION
    PAO=PAO_CONVER.XLS
    ZOUTPUT=ZOUTPUT_CONVER.xls
    *OPTIONS
    FORMAT = DELIMITED
    HEADER = YES
    DELIMITER = ,
    AMOUNTDECIMALPOINT = .
    SKIP = 0
    SKIPIF =
    VALIDATERECORDS=YES
    CREDITPOSITIVE=YES
    MAXREJECTCOUNT=
    ROUNDAMOUNT=
    CONVERTAMOUNTWDIM=ZOUTPUT
    *MAPPING
    CATEGORY=*NEWCOL(ACT)
    PAO=0COSTCENTER
    TIME=0FISCYEAR
    ZOUTPUT=0FUNDS_CTR
    SIGNEDDATA=0DEB_CRE_LC
    *CONVERSION
    PAO=PAO_CONVER.XLS
    ZOUTPUT=AMOUNT_CONVER.XLS
    Conversion file 1 (PAO=PAO_CONVER.XLS)
    EXTERNAL
    INTERNAL
    ID0001
    F08001
    ID0002
    F08001
    ID0003
    F08001
    DG0001
    F08001
    DG0002
    F08001
    Conversion file 2 (ZOUTPUT=ZOUTPUT_CONVER.xls)
    ID0001
    FX01
    VALUE*1
    ID0002
    FX01
    VALUE*1
    ID0003
    FX01
    VALUE*.40
    DG0001
    FX02
    VALUE*1
    DG0002
    FX02
    VALUE*1
    Conversion file 3 (ZOUTPUT=AMOUNT_CONVER.XLS)
    EXTERNAL
    INTERNAL
    FORMULA
    ID0003
    FX02
    VALUE*.60
    I am getting the following error
    [Start validating transformation file]
    Validating transformation file format
    Start validation transformation 1/2
    Validating options...
    Validation of options was successful.
    Validating mappings...
    Validation of mappings was successful.
    Validating conversions...
    Validation of the conversion was successful
    Start validation transformation 2/2
    Validating options...
    Validation of options was successful.
    Validating mappings...
    Validation of mappings was successful.
    Validating conversions...
    Validation of the conversion was successful
    Creating the transformation xml file. Please wait...
    Transformation xml file has been saved successfully.
    Begin validate transformation file with data file...
    [Start test transformation file]
    Validate has successfully completed
    ValidateRecords = YES
    Reject count: 0
    Record count: 6
    Skip count: 0
    Accept count: 6
    0COSTCENTER is not a valid command or column 0COSTCENTER does not exist in source
    Validation with data file failed

  • How to run stored procedure IC Data / to define according script logic file

    We want to execute the stored procedure IC Data before the IC Booking. Could anybody tell me, how we have to define the script logic file in detail?
    We try it with this code (that runs without any error, but 0 rows are calculated, 0 rows are updated)
    *Run_stored_procedure=spicdata('%App%','%C_Category_Set%','%time_set%','','%entity_set%','','','Input','I','%logtable%','%scopetable%')
    *commit
    We also had tried with this code:
    *Run_stored_procedure=spicdata('%App%','%C_Category_Set%','%time_set%','','%entity_set%','','','','','%logtable%','%scopetable%')
    *commit

    step 1.
    Create a store procedure in back end.
    step2.
    Create a ssis package to execute the store procdure.
    step3.
    call this ssis pacakge in the data manager package and run the package.

  • Greater Than Function in Script Logic File

    BPC Experts,
    I've got a pretty simple script logic file to calculate salary amounts, overtime, etc.  The purpose of this logic is for forecasting.  I have two referenced dimensions, TIME and SCENARIO, where TIME.MONTHNUM is equal to its relative month number (eg: Jan monthnum = 1), and SCENARIO.CURRMONTH is equal to the relative amount of actual months data, (eg Jan currmonth = 0, Feb = 1) because if you are completing a February forecast, you have one month of actual.
    The ACTUAL scenario is never touched, but after a month closes, we copy the ACTUAL data to, for example, FEB_FCST.  So, after January closes, its actuals are copied to the FEB_FCST scenario so we can complete an actual/forecast (one month actual, 11 months forecast).
    Currently, the default logic skips anything in the ACTUAL scenario, by stating "*WHEN SCENARIO, *IS <> "ACTUAL" yada yada yada.
    However, when the default logic runs on the forecast scenarios, it takes the same inputs from the months that are copied over from the ACTUAL scenario and adds to the original amount, essentially doubling the value--ultimately causing an incorrect actuals number in the forecast scenario.
    My script right now looks like this:
    *XDIM_MEMBERSET DATASRC=INPUT
    *XDIM_MEMBERSET PRODUCT=NO_PRODUCT
    *XDIM_MEMBERSET SHIFT=NO_SHIFTS
    *XDIM_MEMBERSET MEASURES=PERIODIC
    *WHEN SCENARIO
    *IS <> "ACTUAL"
         *WHEN TIME.MONTHNUM
         *IS > SCENARIO.CURRENTMNTH
              *WHEN ACCOUNT
              *IS "SALARIED_MANPOWER"
                   *REC(EXPRESSION=((([ACCOUNT].[SALARIED_AVG_WAGE] * (1 + [ACCOUNT].[SALARY_TIMEAHALF]))* [ACCOUNT].[SALARIED_MANPOWER])),ACCOUNT="01100")
              *ENDWHEN
    *ENDWHEN
    *ENDWHEN
    *COMMIT
    When it hits line 9 (*IS > SCENARIO.CURRENTMNTH) during validation, it errors.  How can I use a "greater than" function to dictate whether or not a given scenario should run default logic on a specific month?
    If not, is there a different/better way to do it?
    Thank you!
    ABF

    Hi Alex
    Take this sample logic, check the properties in your Time dimension to trouble shoot your issue.
    *SELECT(%CAT_VAR%, "ID", CATEGORY, ID= PLAN_APRIL)
    *XDIM_MEMBERSET CATEGORY = ACTUAL, PLAN, %CAT_VAR%
    *SELECT(%CATMTH%, "STARTMTH", CATEGORY, ID= %CAT_VAR%)
    *SELECT(%ACT_PERIOD%, "ID", TIME, MONTHNUM < %CATMTH% AND  LEVEL = MONTH AND YEAR = 2010)
    *SELECT(%PLAN_PERIOD%, "ID", TIME, MONTHNUM >= %CATMTH% AND LEVEL = MONTH AND YEAR = 2010)
    *XDIM_MEMBERSET TIME= %ACT_PERIOD%
    *XDIM_MEMBERSET CATEGORY=ACTUAL
    *WHEN CATEGORY
    *IS ACTUAL
    *REC(EXPRESSION=%VALUE%, CATEGORY = %CAT_VAR%)
    *ENDWHEN
    *XDIM_MEMBERSET TIME= %PLAN_PERIOD%
    *XDIM_MEMBERSET CATEGORY=PLAN
    *WHEN CATEGORY
    *IS PLAN
    *REC(EXPRESSION=%VALUE%, CATEGORY = %CAT_VAR%)
    *ENDWHEN
    Thanks

  • From where i can run script logic file.

    Dear All,
    I am new to BPC.
    I have a code for currency translation in my Script logic (Default.lgf).
    From where i should run this file. Can any help me.
    thanks a lot in advance,
    Regards,
    Satish.

    Typically the Currency rates are stored in the "RATE" application. This application is what the script logic or Business rule will call during the conversion process.  The rates are only stored by MONTH and by TYPE in a typical planning application.  If you need more granular rate details, then you may need to make you time dimension by weeks or days, or add more "TYPES" that are assigned to ACCOUNTs in th ecorressponding account dimension.
    But in a simple planning world, most companies will use a monthly rate, one for an AVG and one for END account types, enter this into the RATE application, and run the Defualt logic on submission of data to compile the FX value.  Whne you wish to view the results, make certain that you select the correct REPORTING CURRENCY in the report. Many time users forget that data is "normally" entered to an ENTITY,, the ENTITY defines the currency it assumes data is entered, so the plan is loaded to the "LC" member, and converted to USD; or entered in LC, and converted to RS.
    But the RATE application will store all the rates required and you may need to open the INPUT template in that application to load the data at the correct intersections.
    Hope this helps.

  • Copying the data and ignoring the Script logic

    Hi,
    I have some data in Actual category for 2010.APR and I would like to COPY the same
    (through DM) to Budget category with my default script logic file.  When I am trying to
    Do the same it is copying the same data into Budget category, ignoring my script logic
    Actually it should increase the value.
    Can anybody help me in this regard.
    Thanks in adv.
    Raghu B.S.

    The problem it is into script logic.
    When you insert records you are inserting records with category Budget.
    But your default script logic is inserting new recotrds just in case if the input has category Actual.
    So you just to do a correction into your script logic to change the statement
    *WHEN CATEGORY
    *IS "ACTUAL"
    with
    *WHEN CATEGORY
    *IS "BUDGET"
    In this way the default script logic will be executed and after the copy you will have also the records generated by script logic.
    Regards
    Sorin Radulescu
    Edited by: Sorin Radulescu on Aug 5, 2010 2:15 PM

  • The export file from a calc script - naming and date/time stamp

    Here is a simple calc script to export data.
    2 questions:
    1. Is there an easy way to add a date/time stamp on the name of the data file so that it does not overwrite the older ones every time?
    2. I tried to specify the path so that it write to another server. "\\mfldappp011\E:\Exp_AW.txt". It's not working. How should I specify the path ?
    Fix (@Relative("Yeartotal",0),"Actual","Working",&ActualYear);
    Dataexport "file" "," "C:\Exp_AW.txt" "#MI";
    EndFix;
    Edited by: user9959627 on Sep 7, 2012 11:25 AM

    Probably easiest to call the maxl script from a command line script, then rename the exported file to include the tme stamp and copy/move it to a location.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • How to insert data from file into matlab script node

    I have interfaced input data from file to be processed using matlab script node. But the problem is that I could not capture the multiple data into matlab script node and to convert it into matrix. Further to this I need to process the data by plotting graphs from it. Thank you in advance for the advice

    Zarina,
    To clarify your problem, you have a script node contaning your Matlab code. Are you then using the standard LV functions to load in your data from a file and pass it into the script node?
    Regards
    Tristan

  • Shell Script- To FTP the latest file with the date stamp

    I have a Solaris 10 based system, where it generate some application files (multiple) in XML format and the file name comprises of the current date.
    File Format is as follows;
    CX-FIL-20070624000000-2-8452536d-000133.xml
    Where 20070624 repesents the curent date of the file
    I want to FTP these files to another server (Solaris 10 based Sun Machine) by comparing the file name with the current date.
    Please let me know how I can do this by using a shell script.

    Assuming you want to ftp the files with today's datestamp, you could match the files you want like so:
    CX-FIL-`/bin/date +"%Y%m%d"`*
    Use that in your script to generate the file list to be transferred...
    -Rob

Maybe you are looking for