Checking no of records in a cube.

Hi All,
I am trying to check the number of records in a cube. I tried two ways .
1) Using listcube -> rowcount -> total
2) RSA1 -> Manage -> Count no of records for each request.
I am getting different result for both . Can anyone help me on this?
Regards,
Sagar

Why Added Records ??
Ans :- Because there will be default aggreagation ...
These two will always be different
Think like
REQ1 has CHARA CHARB KF1 KF2
REQ2 also has CHARA CHARB KF1 KF2
When i see the records from LISTCUBE
I will just see i record for CHARA and CHARB with Key Figures will summed up ..
I hope u understood the issue
Cheers!!
Amit

Similar Messages

  • How to check No of Records in INfo Cube

    Hi Experts
    My Infocube consists of records in Millions...and i performed selective Deletion.
    Now can i check the No of records in the Info Cube
    Please update

    To fully estimate the volume of data in a cube, I'd also suggest that you check the number of records in each dimension table of the cube.
    go t-code listschema
    enter cube name
    all relevant tables will be displayed. right click on a dim table -> call se16
    on this page hit button "number of entries"
    hope this helps...

  • How to find number of records in a cube and ODS....

    Hi,
    How do we find total number of records in a cube and ODS?
    Is there any Tcode for this ?
    From the content it is difficult to get the number of records, if it is more in number.
    Thanks,
    Jeetu

    Hello ,
              Please check the following thread,
    Number of records in a infocube
    hope it helps,
    assign points if helpful.

  • Delete blank records from the cube

    Hello;
    We just upgraded from 3.5 to 7.0.  When we started running our queries for the Inventory cube, we were getting an error message of "Error:  The validity interval has the initial value as lower limit".
    When I checked the cube, I found records that were blank, such as no dates, plants, etc..  I then tried to see if I can delete those records, but I cannot find a way to do this.  Can someone tell me how to delete the blank records or any records from the cube directly?  When I run the queries from the ODS, I do not get any errors at this time.
    Any information, suggestions or HELP is greatly appreciated.
    Thanks,
    Maximina Barry

    Hello Maximina,
    You could try writing in some code that would delete the records which have some fields as blank in the start routine of the transformation that connects the ODS to the cube . So that the next time you load data from the ODS to the cube .. they would stay deleted .
    Hope it helps.
    Just to add ...
    refer -> Need Sample Code to delete records from Cube using  SE38 Editor
    Thanks,
    ~Vj
    Message was edited by:
            Vijay Gopinath

  • Show record number in cube report

    Hi,
    There is record number for ODS and infoset, where I can get record number for Cube? I remember I can use formulary, however I can not find it.
    Thanks

    hi,
    not sure this is what you are looking
    count
    try in selection screen, mark 'output number of hits', and entry Max. no. of hits with a big number, estimate will bigger then the total rows you will get.
    after execute, there is one column 1ROWCOUNT which's filled with '1', click this column and click 'Total' (icon like E), the last row of table will show the total rows.
    or check how to count doc
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/7e58e690-0201-0010-fd85-a2f29a41c7af
    hope this helps.

  • End Routine - Modify a record in the cube

    Hello Guys,
    In the end routine I have to update a field . The transformation is the same cube to the cube. I want the record to be modified like add a value to a field in the end routine. Since it is a cube it creates a new record instead of overwriting the existing record. Is there anyway I can modify the record. I know I can make the existing record 0.00 and then create a new record with the new value.Is there any other solution.
    For example:This is the existing record in the cube
    sales order           Item No                      Backlog Amount                      Indicator
    1000                     10                              1000.00
    After applying the end routine it has 2 records ( I modify the record with the indicator value)
    sales order           Item No                      Backlog Amount                      Indicator
    1000                     10                              1000.00                                  
    1000                     10                              1000.00                                    REV
    After applying the end routine I need the record to be like overwrite(similar to DSO)
    sales order           Item No                      Backlog Amount                      Indicator
    1000                     10                              1000.00                                   REV
    How to achieve the above result in end routine.
    Thanks
    Senthil

    Hi there,
    Since you create new records in end routine in the InfoCube, why not delete the old ones?
    You can use the
    delete RESULT_PACKAGE where ...
    Therefore deleting the old records after inserting the new ones.
    Diogo.

  • Check for duplicate record in SQL database before doing INSERT

    Hey guys,
           This is part powershell app doing a SQL insert. BUt my question really relates to the SQL insert. I need to do a check of the database PRIOR to doing the insert to check for duplicate records and if it exists then that record needs
    to be overwritten. I'm not sure how to accomplish this task. My back end is a SQL 2000 Server. I'm piping the data into my insert statement from a powershell FileSystemWatcher app. In my scenario here if the file dumped into a directory starts with I it gets
    written to a SQL database otherwise it gets written to an Access Table. I know silly, but thats the environment im in. haha.
    Any help is appreciated.
    Thanks in Advance
    Rich T.
    #### DEFINE WATCH FOLDERS AND DEFAULT FILE EXTENSION TO WATCH FOR ####
                $cofa_folder = '\\cpsfs001\Data_pvs\TestCofA'
                $bulk_folder = '\\cpsfs001\PVS\Subsidiary\Nolwood\McWood\POD'
                $filter = '*.tif'
                $cofa = New-Object IO.FileSystemWatcher $cofa_folder, $filter -Property @{ IncludeSubdirectories = $false; EnableRaisingEvents= $true; NotifyFilter = [IO.NotifyFilters]'FileName, LastWrite' }
                $bulk = New-Object IO.FileSystemWatcher $bulk_folder, $filter -Property @{ IncludeSubdirectories = $false; EnableRaisingEvents= $true; NotifyFilter = [IO.NotifyFilters]'FileName, LastWrite' }
    #### CERTIFICATE OF ANALYSIS AND PACKAGE SHIPPER PROCESSING ####
                Register-ObjectEvent $cofa Created -SourceIdentifier COFA/PACKAGE -Action {
           $name = $Event.SourceEventArgs.Name
           $changeType = $Event.SourceEventArgs.ChangeType
           $timeStamp = $Event.TimeGenerated
    #### CERTIFICATE OF ANALYSIS PROCESS BEGINS ####
                $test=$name.StartsWith("I")
         if ($test -eq $true) {
                $pos = $name.IndexOf(".")
           $left=$name.substring(0,$pos)
           $pos = $left.IndexOf("L")
           $tempItem=$left.substring(0,$pos)
           $lot = $left.Substring($pos + 1)
           $item=$tempItem.Substring(1)
                Write-Host "in_item_key $item in_lot_key $lot imgfilename $name in_cofa_crtdt $timestamp"  -fore green
                Out-File -FilePath c:\OutputLogs\CofA.csv -Append -InputObject "in_item_key $item in_lot_key $lot imgfilename $name in_cofa_crtdt $timestamp"
                start-sleep -s 5
                $conn = New-Object System.Data.SqlClient.SqlConnection("Data Source=PVSNTDB33; Initial Catalog=adagecopy_daily; Integrated Security=TRUE")
                $conn.Open()
                $insert_stmt = "INSERT INTO in_cofa_pvs (in_item_key, in_lot_key, imgfileName, in_cofa_crtdt) VALUES ('$item','$lot','$name','$timestamp')"
                $cmd = $conn.CreateCommand()
                $cmd.CommandText = $insert_stmt
                $cmd.ExecuteNonQuery()
                $conn.Close()
    #### PACKAGE SHIPPER PROCESS BEGINS ####
              elseif ($test -eq $false) {
                $pos = $name.IndexOf(".")
           $left=$name.substring(0,$pos)
           $pos = $left.IndexOf("O")
           $tempItem=$left.substring(0,$pos)
           $order = $left.Substring($pos + 1)
           $shipid=$tempItem.Substring(1)
                Write-Host "so_hdr_key $order so_ship_key $shipid imgfilename $name in_cofa_crtdt $timestamp"  -fore green
                Out-File -FilePath c:\OutputLogs\PackageShipper.csv -Append -InputObject "so_hdr_key $order so_ship_key $shipid imgfilename $name in_cofa_crtdt $timestamp"
    Rich Thompson

    Hi
    Since SQL Server 2000 has been out of support, I recommend you to upgrade the SQL Server 2000 to a higher version, such as SQL Server 2005 or SQL Server 2008.
    According to your description, you can try the following methods to check duplicate record in SQL Server.
    1. You can use
    RAISERROR to check the duplicate record, if exists then RAISERROR unless insert accordingly, code block is given below:
    IF EXISTS (SELECT 1 FROM TableName AS t
    WHERE t.Column1 = @ Column1
    AND t.Column2 = @ Column2)
    BEGIN
    RAISERROR(‘Duplicate records’,18,1)
    END
    ELSE
    BEGIN
    INSERT INTO TableName (Column1, Column2, Column3)
    SELECT @ Column1, @ Column2, @ Column3
    END
    2. Also you can create UNIQUE INDEX or UNIQUE CONSTRAINT on the column of a table, when you try to INSERT a value that conflicts with the INDEX/CONSTRAINT, an exception will be thrown. 
    Add the unique index:
    CREATE UNIQUE INDEX Unique_Index_name ON TableName(ColumnName)
    Add the unique constraint:
    ALTER TABLE TableName
    ADD CONSTRAINT Unique_Contraint_Name
    UNIQUE (ColumnName)
    Thanks
    Lydia Zhang

  • Number of records in BPC Cube

    Dear Friends,
    I am working in EPM10 NW/BW 7.31.
    I just want to know  how can I find number of records in BPC cube from the back end (i.e using T-Code RSA1)
    As I am not a BW expert, any simple suggestion will be appreciated.
    Thanks,
    Rahul

    Hi Rahul,
    The way I do it is right click on the cube, then display data.
    goto the bottom section of the menu where you see the following:
    tick output number of hits
    also change the max no, of hits to a high number
    then run
    in the output screen you sum up the first column, it will give you the total number of records in the cube
    Andy

  • Max number of records in a cube & architectural issues

    Hi,
    Sorry if my question was already done but i can't find the same question with the research button (may be i have not the good words for search, I'm not english).
    I am on a BIG BIG IP project. The forecast volume of planned records is about 1.000.000.000 records a year. So we choose to split records toward many cubes. 
    1) Is there a maximum number of record supported by a cube to be planned? We planned to put 100.000.000 records maximum on a cube to be planned. Is it too much?
    2) If I make 100 cubes (one for each organizational entity) with 10.000.000 records per cube and if I "plug" a planning layout on these 100 cubes with a multiprovider, will IP:
    spread time to search the good cube to write in (due to the selected entity) in the 100 cubes (too much time !), or,
    search directly in the good cube (thanks to user exit that match cube with selected entity), so the response time will be about the same as the one for 1 layout plugged on 1 cube of 10.000.000 records?
    Thanks a lot, and sorry for my english language level
    Georges

    Hi Georges,
    Having too many records in the cube should not be very detrimental for performance of planning application as long as you can ensure that the data volume that you fetch at one go is restricted to reasonable limits, using restrictions in filters (the more restrictive the better). Take care of this while modelling both your planning functions/sequences and input-ready queries.
    I understand that you'll need to create a multiprovider for reporting purposes, but if you dont need the data from more than one cube for planning purpose, it will be better to create the aggregation level (and rest of the planning model) on top of individual cube. In case, you want to use the same planning function/input queries for multiple cubes (which will probably be the case), you can create the aggregation levels on the multiprovider but make sure you restrict the characteristic 'infoprovider' properly in the filter restrictions to avoid the function reading unnecessary data from many cubes.
    Hope this helps.

  • When we drilldown at document level it displays one record for each cube

    Dear Experts .
    The users want to see the report by status for the order detail report at document level it displays one record for each cube .
    They want to see the single line for the order quantity , Amount, Billing quantity delivery and amount along with the Sales order number , billing document number billing date , delvert date ..et in signgle line. ( Sale order DeliveryBilling)
    When we are doing on Multiprovider recording displaying from each cube.
    We have tried with Constant selections and it is not giving the required results m since these want to see by status. Due to huge data performance reasons we re unable to use it .
    Please provide your valuable inputs.
    Thanks,
    Mahesh.

    Mahesh,
    If you want to get detail data with respect to Sales order - delivery - Billing, on top of Multiprovider you can come up with Constant selection.
    But I am not clear that why constant selection is not working properly, can I have some light on that.
    Probably, you are struggling with multiple records .
    I suggest you to go with end routine and map all fields to one IC and do reporting on that single IC, then your requirement will fulfill.
    But here we need to take care of performance issues and process chain scheduling wile loading deltas.
    If Multiprovider is not working properly , then try with Infoset(If you have more than 2 ICs it not possible).
    Regards,
    rvc

  • Getting error:  *cannot check out a record*  in URM

    We have recently implemented URM and it seems to be working fine as far as checking in records and the dispositions working. The problem occurs if someone tries to check out a record. An error comes up saying that a record cannot be checked out. Am I missing something here? Shouldn't I be able to check out a record, update it and then check in a new version?

    No, that's the point of records, they are not alterable.

  • Cannot check In workflow record

    Hi Guru's ,
    We are having some issues in Workflow
    When i process the workflow  Named "PurchOrg Data"as WorkflowNext step-Stop,record is still in cheked out state ,although the stop step has Record checkout property as ChekIn.
    And now there is no workflow in Workflow tab.So to checkin this record ,I right clicked on record Workflow Add to job-- Purch Org Data.
    Now it gives me error "Record is currently usedin Workflow".
    How do I check In this record (as there is no workflow in workflow tab)???
    Also I cannot delete it as it is in checkout state..
    Did anyone had this issue befor..
    Regards
    Vikrant M Kelkar

    HI Vikrant
    You need to join the checkout process first. when the record was checked out we have an option of using checkout exclusive or non exclusive.
    In case the record has been checked out exclusive the owner of the checkout needs to allow another user to join the checkout process.
    In case record is checked out non exclusive anyone can join the checkout process and the owner need not invite to join.
    hope this helps-
    Best regards
    Ravi

  • Check which data records have been updated or created by the business rule.

    Dear all,
       Is there any way to check which data records have been updated or created by the business rule?
       For ex:
       When I run a data package which execute some standard store procedure and the log shows :
    Executing SPRUNCONVERSION  'LEGALAPP', [ACTUAL], [GGGG],'GLOBAL','SPSCOPE_42412','SPLOG_125272'
    SPRunConversion Version 2.06
    20071200 - 210 Rows Calulated
    20071200 - 84 Rows Updated
    Anyway to see the data content of "84 Rows Updated "?
    Best Regards,
    Jeff

    Yes each aplication has 3 fact tables:
    FACTWB:Real time storage, when you send data it is stored here and this is linked to the FACTWB partition in OLAP which is a ROLAP partition to enable real time updates of the data.
    FAC2: Short term storage. If too many records are in the FACTWB the aplication becomes slow because these records need to be queried by the OLAP engine everytime new records are send to the database. To clean up te real time storage you can do a lite optimize, which will move the FACTWB records to the FAC2 partition. this lite optimize can be done even when users are in the system.
    FACT: this is the long term storage. When your reporting is finished you can run a full optimize to clean up both FACTWB and FAC2 tables and store it in the FACT table. This contains all your historical records.
    Evertime you run a reort, SAP BPC is looking at all 3 partitions and combines the data of the 3 partitions to present the correct value in your report. Remember that BPC is posting differences in the WB and not the final values. So if you have for example a record with 10 in the FAC2 partition and you send 12 to the database for that intersection, BPC will post only 2. When running a report it combines the 10 and the 2 which makes 12.
    Hope this makes it more clear.
    Joost
    Edited by: Joost Hoppenbrouwers on May 6, 2008 12:19 PM : Seems that James and I where writing pretty much the same post at the same time

  • Sortdump when checking number of records of an SAP table

    When trying to check number of records in an SAP table from SE16, that ends with shortdump as below.
    SQL error in the database when accessing a table.
    Database error text........: "601 Could not continue scan with NOLOCK due to
    data movement."
    Internal call code.........: "RSQL/COUN//BIC/B0000116000 "
    Thank you in advance for quick response.

    Hi,
    please follow note 142731 and do a complete database check. This error is often caused by database corruptions.
    Best regards
      Clas

  • How to restrict Overwritten records of Info-Cube  in Query

    Hi ,
    In query i am able to see the overwritten/replaced records with Zero values as well the same records with new values i.e. same record with in 2 lines one with zero value and other with replaced value.How i can restrict that only new record should be shown in Query.
    With regards,
    Abhishek Kumar

    As in InfoCube if a new record is inserted with same key fields it add the previous record with -ve values and new record with new values having same key fields.Now i can see these records in Info-Cube but when it comes to query it shows all the records ,one with the zero value and the same with the new values.Question is how i can restrict the record with zero value.
    Abhishek Kumar

Maybe you are looking for

  • Midlet Signing

    Hello, Questions - Is trusted third party signing enough to reach a significant amount of JSR-75 compatible mobiles, with a good user experience ? -What is the best third party for signing midlets ? - Is Manufacturer & Operator Sining feasible for a

  • How do you clone a hard drive with boot camp?

    My daughter got a new 13" Macbook pro in July.  She is still running Lion.  Se started to exerience system freezes.  She called Applye support which had her reset the PRAM(?).  The the system would no longer boot.  When she went away to college, I se

  • The ADS call has failed - Error

    Dear all I have created one web dynpr application with interactive forms followed by saptechical.com, If i run the web dynpr application, it thrown an error "WebDynpro Exception: The ADS call has failed. You can find information about the cause in th

  • Tree with Icons

    hi, i need to hava a tree with icons in apex is it possible?

  • Global Address List extended fields from mobile phone

    When on an exchange2010 server, search the GAL works fine, however, you don't get to see the entire GAL Object properties.  Only telephone & email fields are visible. what about address? Example: find out a colleague had a baby and want to send a gif