Distributing the data files

Hi All,
We are working on the redistribution process of data files in SQL level. We knew that by doing the empty files options through shrink we could achieve it.  But I would like to know, Is there any things which I need to consider in terms of number of data files(Any Formula to arrive a total number of data files for the better performance in SQL database level).
In my environment, the current database size is 420GB and its running with 24 data files, SAP advised us to minimize the data files because the data files are not in order (Like 1 data file having GB and some having 40GB and few having 90GB) - So, SAP recommended us to maintain the data file size as common and distribute the data equally among the files.
So, wondering the data file size and also number according to our environment (It would be kind if you suggest on it)
Database size :420GB
Database version : SQL 2005
CPU core : 4 cores running and its a Xeon processor
Regards
Vijay

Hi,
But I would like to know, Is there any things which I need to consider
in terms of number of data files (Any Formula to arrive a total number of data files
for the better performance in SQL database level).
You can refer Topic "Number and Size of SQL Server Data Files" of this [Best Practice document|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/4ab89e84-0d01-0010-cda2-82ddc3548c65&overridelayout=true] (Page no.37 to 44 ). I hope you will get all the required information ,which can be considered for File Management in SQL server.
Also you can refer the following SAP Notes to get more information.
Note 987961 - FAQ: SQL Server I/O performance
Note 363018 - File management for SQL Server
Note 1488135 - Database compression for SQL Server
Regards,
Bhavik G. Shroff

Similar Messages

  • SQL Loader and foreign characters in the data file problem

    Hello,
    I have run into an issue which I can't find an answer for. When I run SQL Loader, one of my control files is used to get file content (LOBFILE) and one of the fields in the data file has a path to that file. The control file looks like:
    LOAD DATA
    INFILE 'PLACE_HOLDER.dat'
    INTO TABLE iceberg.rpt_document_core APPEND
    FIELDS TERMINATED BY ','
    doc_core_id "iceberg.seq_rpt_document_core.nextval",
    -- created_date POSITION(1) date "yyyy-mm-dd:hh24:mi:ss",
    created_date date "yyyy-mm-dd:hh24:mi:ss",
    document_size,
    hash,
    body_format,
    is_generic_doc,
    is_legacy_doc,
    external_filename FILLER char(275) ENCLOSED by '"',
    body LOBFILE(external_filename) terminated by EOF
    A sample data file looks like:
    0,2012-10-22:10:09:35,21,BB51344DD2127002118E286A197ECD4A,text,N,N,"E:\tmp\misc_files\index_testers\foreign\شیمیایی.txt"
    0,2012-10-22:10:09:35,17,CF85BE76B1E20704180534E19D363CF8,text,N,N,"E:\tmp\misc_files\index_testers\foreign\ลอบวางระเบิด.txt"
    0,2012-10-22:10:09:35,23552,47DB382558D69F170227AA18179FD0F0,binary,N,N,"E:\tmp\misc_files\index_testers\foreign\leesburgis_á_ñ_é_í_ó_ú_¿_¡_ü_99.doc"
    0,2012-10-22:10:09:35,17,83FCA0377445B60CE422DE8994900A79,binary,N,N,"E:\tmp\misc_files\index_testers\foreign\làm thế nào bạn làm ngày hôm nay"
    The problem is that whan I run this, SQL Loader throws an error that it can't find the file. It appears that it can't interpret the foreign characters in a way that allows it to find that path. I have tried adding a CHARACTERSET (using AL32UTF8 or UTF8) value in the control file but that only has some success with Western languages, not the ones listed above. Also, there is no set of defined languages that could be found in the data file. It essentaially could be any language.
    Does anyone know if there is a way to somehow get SQL Loader to "understand" the file system paths when a folder and/or file name could be in some other langauge?
    Thanks for any thoughts - Peter

    Thanks for the reply Harry. If I try to open the file in various text editors like Wordpad, Notepad, GVIM, andTextpad, they all display the foreign characters differently. Only Notepad comes close to displaying the characters properly. I have a C# app that will read the file and display the contents and it renders it fine. If you look at the directory of files in Windows Explorer, they all are displayed properly. So it seems things like .Net and Windows have some mechanism to understand the characters in order to render them properly. Other applications, again like Wordpad, do not know how to render them properly. It would seem that whatever SQL Loader is using to "read" the data files also is not rendering the characters properly which prevents it from finding the directory path to the file. If I add "CHARACTERSET AL32UTF8" in the control file, all is fine when dealing with Western langauges (ex, German, Spanish) but not for the Eastern languages (ex. Thai, Chinese). So .... telling SQL Loader to use a characterset seems to work, but not in all cases. The AL32UTF8 is the characterset that the Oracle database was created with. I have not had any luck if I try to set the CHARACTERSET to whatever the Thai character set is, for example. There problem there though is that even if that did work, I can't target specific lagauages because the data could come from anywhere. It's like I need some sort of global "super set" characterset to use. It seems like the CHARACTERSET is the right track to follow but I am not sure, and even if it is, is there a way to handle all languages.
    Thanks - Peter

  • Clear Data package - "The data file is empty." Error

    Hi,
    When I run the Clear package in Data Manager, I get the error "The data file is empty". We are running BPC 7.5 SP3. There are no calculated members selected. Any suggestions?
    Thanks.
    Tom

    Hi Tom,
    If i'm not mistaken when we run the clear data package, BPC inserts -ve values for the combination of dimensions selected.
    For ex if you have selected time as 2010.DEC it would reverse out all these values by inserting negative values which would then total up to zero.
    So i still think it could be a problem with the Transformation file.
    P.S. Maybe you can try using the Clear from Fact table package.
    Santosh

  • I have written a binary file with a specific header format in LABVIEW 8.6 and tried to Read the same Data File, Using LABVIEW 7.1.Here i Found some difficulty.Is there any way to Read the Data File(of LABVIEW 8.6), Using LABVIEW 7.1?

    I have written a binary file with a specific header format in LABVIEW 8.6 and tried  to Read the same Data File, Using LABVIEW 7.1.Here i Found some difficulty.Is there any way to Read the Data File(of LABVIEW 8.6), Using LABVIEW 7.1?

    I can think of two possible stumbling blocks:
    What are your 8.6 options for "byte order" and "prepend array or string size"?
    Overall, many file IO functions have changed with LabVIEW 8.0, so there might not be an exact 1:1 code conversion. You might need to make some modifications. For example, in 7.1, you should use "write file", the "binary file VIs" are special purpose (I16 or SGL). What is your data type?
    LabVIEW Champion . Do more with less code and in less time .

  • Data Integrator has no read permissions to get the data file using FTP

    Hi,
    I wonder if you could help.
    We have installed the data integrator and are using FTP to get the data file from the SAP server to the DI server. The files are created by the SAP admin user with permissions 660. The FTP user is not in the sapsys group so cannot read the files.
    Has anyone come accross this problem and how did you solve it?
    Many thanks.

    Hi,
    you might want to put you entry into the EIM forum where also the Data Integrator topics are being handled:
    Data Services and Data Quality
    Ingo

  • How can I put the signature into the data file.

    Hi all,
    I want to put the signature into the data file, mean the data file will be on the top and the signature will be at the buttom.how can i do that?
    Thanks.
    example:-
    dataFile.txt
    Name=...
    Address=...
    Signature=6A07C70E....123FEAB(Hex)

    Thanks for your reply.
    First, i have a keypair (public key and private key).
    I use the private key to initialize the Signature and
    use the Signature to sign a data file ( dataFile.txt) .
    So when i need to comfirm the data file not being modify
    I have to verify the data file with the Signature and the public key.
    I hardcoded the public key in my program so my client (receiver of data)
    need only the Signature file and the data file (2 files).
    So how can i combine the Signature file and the data file into 1 file?

  • Pls help me to count the no of occurence of each word in the data file

    I want to count the no of occurence of each word the data file..
    anybody help me out to solve this ,,
    I have attached my code hereby...
    import java.io.*;
    import java.util.*;
    import java.util.Collections;
    import javax.print.attribute.standard.Media;
    class WordCounter1 {
    String s;
    String word;
    public static void main (String[] args){
         WordCounter1 simpleCounter= new WordCounter1();
    try
    File firstFile=new File("C:/Documents and Settings/janmala/file4.txt");
    simpleCounter.countWords(firstFile);
    catch(Exception e)
    System.err.println(e);
    void countWords(File inputFile)
    int count=0;
    int key=0;
    LinkedHashMap lhm=new LinkedHashMap();
    try{
    FileInputStream fin=new FileInputStream(inputFile);
    DataInputStream din=new DataInputStream(fin);
    while(din.available()!=0){
    s=din.readLine();
    StringTokenizer st=new StringTokenizer(s);
    while(st.hasMoreTokens())
    word=st.nextToken();
    lhm.put(word,count);
    if(lhm.containsKey(word))
    lhm.get(word);
    count++;
    System.out.println(word);
    System.out.println(count);
    int a=lhm.size();
    System.out.println("The size of Hash map"+a);
    Set set = lhm.entrySet();
    Iterator it = set.iterator();
    while(it.hasNext())
    Map.Entry me = (Map.Entry)it.next();
    System.out.println(me.getValue()+":"+me.getKey());
    catch(Exception e)
    System.err.println(e);
    }}

    jan07 wrote:
    anybody help me out to solve this ,,Solve what?
    Don't just dump a pile of unformatted code and expect someone to magically fix it for you.
    Post code, hightlight it and click the CODE button to retain formatting
    Include error messages if you get them
    Ask an actual question

  • Where is the data file for the address book located, I need to restore mine from backup but don't know where it is.

    Where is the data file for the address book located, I need to restore mine from backup but don't know where it is to do so.  Putting my Mac Pro onto icloud synching contacts resulted in duplicate and triplicates of the same cards.  I have a few THOUSAND contacts and the last time this happened was on mobile me.  that had a function where you could replace info on the mobile me server with info on the computer.  I see no way to do that on the new cloud.

    In the stupidly hidden user's Library->/Application Support/AddressBook/ labeled as AddressBook-v22.abcddb
    For TM restores, see Pondini's TM FAQs for starters, since I don't use TM.

  • Where are the data files stored?

    I regularly back up using Time Machine, but when I reformatted my hard drive and then restored it, I could not find the data files for iWeb websites I have created.  Looking for help, I saw a reference to a subdirectory named "Domain", I found that but could not see any files for the four different websites I have created.  Are these data files kept hidden?  I am not very familiar with MAC file systems, so please respond in a MAC for Dummies language.  Thanks.

    In Lion and Mountain Lion the Home/Library folder is now invisible. To make it permanently visible enter the following in the Terminal application window: chflags nohidden ~/Library and hit the Enter button - 10.7: Un-hide the User Library folder.  Then it will be easy to fine the domain file in your Users/Home/Library/Application Support/iWeb folder.
    To open your domain file in Lion or Mountain Lion or to switch between multiple domain files Cyclosaurus has provided us with the following script that you can make into an Applescript application with Script Editor. Open Script Editor, copy and paste the script below into Script Editor's window and save as an application.
    do shell script "/usr/bin/defaults write com.apple.iWeb iWebDefaultsDocumentPath -boolean no"delay 1
    tell application "iWeb" to activate
    You can download an already compiled version with this link: iWeb Switch Domain.
    Just launch the application, find and select the domain file in your Home/Library/Application Support/iWeb folder that you want to open and it will open with iWeb. It modifies the iWeb preference file each time it's launched so one can switch between domain files.
    WARNING: iWeb Switch Domain will overwrite an existing Domain.sites2 file if you select to create a new domain in the same folder.  So rename your domain files once they've been created to something other than the default name.
    OT

  • Concatenating the data file(Mbrs) before loading.. due to change in Outline

    We have 8 Dimensions(11.1.1.3)
    The data file is coming from the source system in this format from 3 yrs as below ..
    D1 D2 D3 D4 D5 D6 Product, Curriency
    . , . ,. , . , . , ., a , USD
    . , . , ., . , . , . , a , EUR
    . , . , . , . , . , ., b , GBP
    . , . , . , . , . , . , b , INR
    Now the product name as been changed in outline as
    a_USD
    a_EUR
    b_GBP
    b_INR
    So, Is there any way in the hyperion suite(like in rules file or any other), where I can concatenate Product and currency and get the file as
    D1 D2 D3 D4 D5 D6 , Product , Curriency
    . . . . . . , a_USD , USD
    . . . . . . , a_EUR , EUR
    . . .. . . . , b_GBP , GBP
    . . . . . . , b_INR , INR
    Please do let me know
    Thanks in Advance.
    Edited by: 838300 on Sep 27, 2011 9:00 AM

    While what Mehmet wrote is correct, if this is anything more than a quick and dirty fix, may I suggest you go down another ETL path? (Yes, this is Cameron's load rule rant coming right at you, but in abbreviated form.)
    ETL in a load rule is great to get the job done, but is a pain to maintain. You would be (unpleasantly) surprised how these have a way of growing. Have you given a thought to fixing it at the source or doing some true ETL in a tool like ODI, or staging it in SQL and doing the mods there? I know, for a simple(ish) change, that seems overkill, but load rules for the purposes of ETL are Evil.
    Regards,
    Cameron Lackpour
    P.S. For anyone who must see me go bonkers over this, see: http://www.network54.com/Forum/58296/thread/1311019508/Data+Load+Optimization+-Headervs+Column

  • Where are the data files?

    Hi,
    I'm completely new to Berkeley db and i have some basic questions regarding architecture which i couldnt find in docs. Where does BDB store its data? Is there a default file location? Also, is there a way i can specify maximum file size? I'm developing an application which will store many giga bytes of data and I dont want the data files to grow above the OS limit. I'm using windows as the OS.
    regards,

    Hello,
    This question is probably deserving of its own thread, but I'll answer briefly here. As you've noted, BDB XML itself it designed and tuned to storage and query of XML. Because BDB XML is built on Berkeley DB, you can use DB proper transactionally along with BDB XML. In an application such as yours, you could store XML in BDB XML, and create (application-provided) references between those documents and other data stored in Berkeley DB. In addition, you might want to store some data in regular files (streaming video, large binary data, etc).
    You can store binary data in BDB XML as either document metadata or encoded into ascii, but for the scale you want I would not recommend that approach.
    Regards,
    George

  • Clear Data Manager Package Error "The data file is empty."

    Hi,
    When I run the Clear data package in Data Manager, I receive the error "The data file is empty." I selected a very specific set of dimension values (none are calculated) and am on BPC 7.5 SP3. I subsequently turned on debugging to troubleshoot, but do not see any obvious issues leading the the error message. The log file with debugging turned on is below. Any help would be greatly appreciated!
    Thanks.
    Tom
    TOTAL STEPS  3
    1. Export_Zero:        completed  in 1 sec.
    2. Load Cube:          Failed  in 0 sec.
    3. Clear:              completed  in 0 sec.
    [Selection]
    ENABLETASK= Yes
    CHECKLCK= Yes
    (Member Selection)
    Category: ACTUAL
    Time: 2010.C_SEP
    Affiliate: az_swhd
    Account: Donor_DART_ID_1
    Functional: Benchmark_F
    Report: Cons
    Restriction: AnyRestricted
    [Messages]
    The data file is empty. Please check the data file and try again.
    [EvModifyScript Detail]
    12-28-2010  17:30:05 - Debug turned ON
    INFO(%TEMPFILE%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)
    TASK(EXPORT_ZERO, APPSET, ESMetrics)
    TASK(EXPORT_ZERO, APP, CONSOLIDATED)
    TASK(EXPORT_ZERO, USER, NESSGROUP\tbardwil)
    TASK(EXPORT_ZERO, FILE, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)
    TASK(EXPORT_ZERO, SQL,
    select [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM ( SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED') UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTWBCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')  UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFAC2CONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')) as ZeroTable  group by [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID] OPTION(MAXDOP 1)
    TASK(EXPORT_ZERO, DATATRANSFERMODE, 2)
    TASK(LOAD CUBE, APPSET, ESMetrics)
    TASK(LOAD CUBE, APP, CONSOLIDATED)
    TASK(LOAD CUBE, USER, NESSGROUP\tbardwil)
    TASK(LOAD CUBE, FILE, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)
    TASK(LOAD CUBE, DATATRANSFERMODE, 4)
    TASK(LOAD CUBE, DMMCOPY, 0)
    TASK(LOAD CUBE, PKGTYPE, 0)
    TASK(LOAD CUBE, CHECKLCK, 1)
    TASK(CLEAR COMMENTS, APPSET, ESMetrics)
    TASK(CLEAR COMMENTS, APP, CONSOLIDATED)
    TASK(CLEAR COMMENTS, USER, NESSGROUP\tbardwil)
    TASK(CLEAR COMMENTS, DATATRANSFERMODE, 0)
    TASK(CLEAR COMMENTS, SELECTIONORFILE, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)
    TASK(CLEAR COMMENTS, ENABLETASK, 1)
    TASK(CLEAR COMMENTS, CHECKLCK, 1)
    INFO(%ENABLETASK%, 1)
    INFO(%CHECKLCK%, 1)
    INFO(%SELECTION%, [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED'))
    INFO(%TOSELECTION%, [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED'))
    INFO(%APPSET%, ESMetrics)
    INFO(%APP%, CONSOLIDATED)
    INFO(%CONVERSION_INSTRUCTIONS%, )
    INFO(%FACTCONVERSION_INSTRUCTIONS%, )
    INFO(%SELECTIONFILE%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\FROM_51_.TMP)
    INFO(%TOSELECTIONFILE%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\TO_51_.TMP)
    INFO(%DEFAULT_MEASURE%, PERIODIC)
    INFO(%MEASURES%, Periodic,QTD,YTD)
    INFO(%OLAPSERVER%, ETSCSAP047940.EASTER-SEALS.ORG)
    INFO(%SQLSERVER%, ETSCSAP047940.EASTER-SEALS.ORG)
    INFO(%APPPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\)
    INFO(%DATAPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\DataFiles\)
    INFO(%DATAROOTPATH%, C:\BPC\Data\WebFolders\)
    INFO(%SELECTIONPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\SelectionFiles\)
    INFO(%CONVERSIONPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\ConversionFiles\)
    INFO(%TEMPPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\)
    INFO(%LOGICPATH%, C:\BPC\Data\WebFolders\ESMetrics\Adminapp\CONSOLIDATED\)
    INFO(%TRANSFORMATIONPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\TransformationFiles\)
    INFO(%DIMS%, [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[Time])
    INFO(%FACTDIMS%, [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID])
    INFO(%CATEGORY_DIM%, [Category])
    INFO(%TIME_DIM%, [Time])
    INFO(%ENTITY_DIM%, [Affiliate])
    INFO(%ACCOUNT_DIM%, [Account])
    INFO(%CURRENCY_DIM%, )
    INFO(%APP_LIST%, Consolidated,ES_INC,GrantMgmt,LegalApp,LRate,Ownership,Rate)
    INFO(%ACCOUNT_SET%, DONOR_DART_ID_1)
    INFO(%AFFILIATE_SET%, AZ_SWHD)
    INFO(%CATEGORY_SET%, ACTUAL)
    INFO(%FUNCTIONAL_SET%, BENCHMARK_F)
    INFO(%REPORT_SET%, CONS)
    INFO(%RESTRICTION_SET%, ANYRESTRICTED)
    INFO(%TIME_SET%, 2010.C_SEP)
    INFO(%ACCOUNT_TO_SET%, DONOR_DART_ID_1)
    INFO(%AFFILIATE_TO_SET%, AZ_SWHD)
    INFO(%CATEGORY_TO_SET%, ACTUAL)
    INFO(%FUNCTIONAL_TO_SET%, BENCHMARK_F)
    INFO(%REPORT_TO_SET%, CONS)
    INFO(%RESTRICTION_TO_SET%, ANYRESTRICTED)
    INFO(%TIME_TO_SET%, 2010.C_SEP)
    INFO(DATAMGRGLOBALBPU, )
    INFO(DATAMGRGLOBALCLIENTMACHINEID, ETSCWLT048794)
    INFO(DATAMGRGLOBALERROR, )
    INFO(DATAMGRGLOBALPACKAGEINFOR, )
    INFO(DATAMGRGLOBALPACKAGENAME, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\PackageFiles\System Files/Clear.dtsx)
    INFO(DATAMGRGLOBALSEQ, 51)
    INFO(DATAMGRGLOBALSITEID, )
    INFO(MODIFYSCRIPT, DEBUG(ON)<BR>PROMPT(SELECTINPUT,[CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED'),,"SELECT THE MEMBERS TO CLEAR",[Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[Time])<BR>PROMPT(RADIOBUTTON,1,"DO YOU WANT TO CLEAR COMMENTS ASSOCIATED WITH DATA REGIONS IN BPC?",1,{"YES","NO"},{"1","0"})<BR>PROMPT(RADIOBUTTON,1,"SELECT WHETHER TO CHECK WORK STATUS SETTINGS WHEN DELETING COMMENTS.",1,{"YES, DELETE COMMENTS WITH WORK STATUS SETTINGS","NO, DO NO DELETE COMMENTS WITH WORK STATUS SETTINGS"},{"1","0"})<BR>INFO(C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp,C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempfdla_51_.tmp)<BR>TASK(EXPORT_ZERO,APPSET,ESMetrics)<BR>TASK(EXPORT_ZERO,APP,CONSOLIDATED)<BR>TASK(EXPORT_ZERO,USER,NESSGROUP\tbardwil)<BR>TASK(EXPORT_ZERO,FILE,C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)<BR>TASK(EXPORT_ZERO,SQL,
    select [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM ( SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED') UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTWBCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')  UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFAC2CONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')) as ZeroTable  group by [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID] OPTION(MAXDOP 1)
    )<BR>TASK(EXPORT_ZERO,DATATRANSFERMODE,2)<BR>TASK(LOAD CUBE,APPSET,ESMetrics)<BR>TASK(LOAD CUBE,APP,CONSOLIDATED)<BR>TASK(LOAD CUBE,USER,NESSGROUP\tbardwil)<BR>TASK(LOAD CUBE,FILE,C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)<BR>TASK(LOAD CUBE,DATATRANSFERMODE,4)<BR>TASK(LOAD CUBE,DMMCOPY,0)<BR>TASK(LOAD CUBE,PKGTYPE,0)<BR>TASK(LOAD CUBE,CHECKLCK,1)<BR>TASK(CLEAR COMMENTS,APPSET,ESMetrics)<BR>TASK(CLEAR COMMENTS,APP,CONSOLIDATED)<BR>TASK(CLEAR COMMENTS,USER,NESSGROUP\tbardwil)<BR>TASK(CLEAR COMMENTS,DATATRANSFERMODE,0)<BR>TASK(CLEAR COMMENTS,SELECTIONORFILE,C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)<BR>TASK(CLEAR COMMENTS,ENABLETASK,1)<BR>TASK(CLEAR COMMENTS,CHECKLCK,1)<BR>BEGININFO(
    select [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM ( SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED') UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTWBCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')  UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFAC2CONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')) as ZeroTable  group by [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID] OPTION(MAXDOP 1)
    )<BR><BR><BR><BR><BR><BR><BR>SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 AS SIGNEDDATA FROM ( SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 AS SIGNEDDATA FROM TBLFACTCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED') UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 AS SIGNEDDATA FROM TBLFACTWBCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')  UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 AS SIGNEDDATA FROM TBLFAC2CONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')) AS ZEROTABLE  GROUP BY [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID] OPTION(MAXDOP 1)<BR><BR><BR><BR>ENDINFO<BR><BR><BR>)
    Edited by: Tom Bardwil on Dec 28, 2010 5:20 PM

    You can greatly improve your chance of receiving a helpful answer to your question if you state the version (MS or NW) and the release (5.1, 7.0, 7.5) of BPC which you are using.
    Also notice the sticky [note|Please do not post BPC, SSM or FI/CO questions here!; at the top of this forum whereby we announced new dedicated forums for BPC which are the proper place to post your questions regarding BPC in the future to be able to reach the right audience for your question.
    Thanks and best regards,
    [Jeffrey Holdeman|http://wiki.sdn.sap.com/wiki/display/profile/Jeffrey+Holdeman]
    SAP Labs, LLC
    BusinessObjects Division
    Americas Applications Regional Implementation Group (RIG)

  • How to store the sampling rate into the header of the data file?

    I want to store the data parameters such as sampling rate to the data file. Now I am using the Write Labview Measurement File VI, the data is saved, but like the sampling rate this kind of parameter is not in the file. How can I store this parameter?

    If you want the sampling rate to appear in a different location of the header, then it and it's subVIs can be modified to write anyway you want. Right click on the Express VI and slect Open Front Panel. Then you have a VI that can be modified. If you want to use an unmodified Write LabVIEW Measurement File, you could use that, read the whole file back in with one of the file read functions, insert a string with the sampling rate and then write the whole thing back out again. It might be simpler though, to use Write Characters to File to create your own header and then use Write to Spreadsheet File to write the data. There is also the function Export Waveforms to Spreadsheet File on the Waveform>Waveform File I/O. This uses a slightly different format than a .lvm file. It too can be modified if you don't like the default header.

  • Reading the data file into XI

    Hi,
    I am facing a problem in XI. I have a file Adapter which fetches the data file from the appln server. For the testing purpose we kept the adapter disabled and placed th files in the directory. When activating the adapter, it fetches all files and brings to XI.
    Now the adapter is enabled. If I am trying place the data file in the application server while the adapter is enabled, the file simply disappeared/archived as per the setting in adapter. But the data is not coming into XI.
    This problem occurs only when the adapter is on and placing the files in the server.
    Can any one help me on this?
    Thanks.
    Jahir.

    Hi,
    http://xiserver:port/MessagingSystem/monitor/monitor.jsp
    have a look at this link and check if you have any errors
    over there
    Regards,
    michal
    <a href="/people/michal.krawczyk2/blog/2005/06/28/xipi-faq-frequently-asked-questions"><b>XI / PI FAQ - Frequently Asked Questions</b></a>

  • Address Book in Tiger 10.4.3 keeps losing track of the data file

    Address Book in Tiger 10.4.3 keeps losing track of the data file associated with it and creates a new file without the data I used to have.
    Any methodology to follow to correct this issue?

    I think Import would be better, or Migration Assistant, I'm thinking AB changed formats from 10.3 to 10.4, but the stuff for 10.4 For Address Book...
    Files...
    /Users/YourUserName/Library/Preferences/com.apple.AddressBook.plist
    /Users/YourUserName/Library/Preferences/com.apple.AddressBook.abd.plist
    Folders...
    /Users/YourUserName/Library/Application Support/AddressBook/
    /Users/YourUserName/Library/Caches/com.apple.AddressBook/

Maybe you are looking for

  • ScreenSaver in 10.8.2

    I recently upgraded from 10.6 and can't get the screensaver to do what I really liked with the old release. In the old release, I could select the "Pictures" and "Shuffle" and set it up so the screensaver cycled randomly through the various photos in

  • How can 'Consolidated' Report takes OutPut of 'Line Item' Report

    Hi I have 'Line Item Detail Report'(REPORT1) which gives the latest records . I want use these detail records as a input records to another consolidated report(REPORT2). I mean , REPORT1 - contains 'lastest detailed line items' . REPORT2 - contains '

  • Changing Pagesize in Reports 10 On Solaris platform

    All, I use the reports builder on a Windows machine, then copy the .rdf file over to a Solaris 10 environment. I use the rwconvert to convert from an rdf to rep file on Unix. One of the reports I'm working on needs legal paper (8.5 X 15). I have a si

  • Element input values

    Hi, I have a recurring element that has input parameters as Eligible Amount and Pay Value. The Eligible Amount will vary greatly every month and some times the user may not enter any amount in the input values(as required. But If I'm entering a value

  • ITunes Match: How can I tell if a matched song is from the correct album?

    Since iTunes Match uses my album art and metadata, is there a way to see the album art of the matched item from iTunes? I have ripped hundreds of classical music CDs through the years into iTunes, and they all have album art (not from iTunes Match).