Query slow in testing but quick in production

my query is slow in testing but very quick in production.help me..

Hi,
Check whether your Prod is Same as you Testing Environment.
Whether Parameters in Both DB are same or not Mainly Stats are in Update and same or not
One more thing is Plan of Query, how it is working in Prod and Test Server.
that gives the idea how the query Executed. Check out these things.
- Pavan Kumar N

Similar Messages

  • BAPI works in test but not in production

    3rd party software uses BAPI to input data into SAP. This works in test but not in production. We have checked that user 3rd party software uses have same rights in both systems. BAPI is standard SAP BAPI. We can connect to production with SM59.
    Is there anything in SAP that could cause this?

    No. Better you check with the PATCH Levels.. etc to know is there any changes in the BAPI (BASIS people can help you) > Otherwise go to SYSTEM in the MENUBAR> STATUS--> Compare the Status there
    After that check the data once...
    Reward if helpful...

  • Query running in Quality but not in Production

    Hi Folks,
    we are running a query AQICSDUSERS=====MOVEMENTLIST4= which used to get executed previously,but all of a sudden now after running for an hour or so it is coming back to  sap easy access screen without displaying the report.It is running and giving the report in Quality but not in Production.
    What could be the reason?
    K.Kiran.

    Hi Folks,
    From DUMP I got this:-
    <b>You should usually execute long-running programs as batch jobs.
    If this is not possible, increase the system profile parameter
    "rdisp/max_wprun_time".</b>
    I want to know how to increase the system profile parameter
    "rdisp/max_wprun_time".
    Thanks,
    K.Kiran.

  • Sql query slow from application but executes in 1 min in Toad

    Hello,
    When I execute a big select query from DB itself , it gives results in a min.
    But when called from .Net application it takes 10 mins to execute.
    But after trying to excute no of times, it suddenly works faster sometimes.

    Hello,
    When I execute a big select query from DB itself , it gives results in a min.
    But when called from .Net application it takes 10 mins to execute.
    But after trying to excute no of times, it suddenly works faster sometimes.

  • Hi in my sql query i applied like condition (like '%TEST') but it is taking long time

    Hi in my sql query i applied like condition (like '%TEST') but it is taking long time. I applied indexes also,but still i'm facing the same problem. In my databse nearly 2,00,000 records their.

    Hi Manikandan
    Is there a difference in performance between running the query in BEx and WebI?
    have you aggregates in place on the BEx side of things?
    When you say its taking too long to open the report, have you a variable screen coming up for the report and is that what is taking the time or is it the report execution.
    With regards
    Gill

  • Help!! I'm using a Mid 2012 MBP retina. for the past 2 years my write speeds has been consistently on 400MB/s via black magic test. but now it only have 100MB/s going down to 80MB/s. what should I do???? getting slower...

    Help!! I'm using a Mid 2012 MBP retina 2.3Ghz i7 with 8GB of ram 256SSD. for the past 2 years my write speeds has been consistently on 400MB/s and up via black magic test. but now it only have 100MB/s going down to 80MB/s. what should I do???? I have already cleaned and even restored my MBP still it's on 100MB/s. please help. I have to finish tasks on a timely manner.

    No. This requires completely repartitioning the SSD. You can do that only by using the Network Recovery method or the method I outlined. My method preserves all your data and doesn't require doing a slow Network Recovery. But if you wish to do that instead. The purpose of the whole exercise is to "clean" the SSD of garbage collected during use that tends to cause the SSD to slow down as you've described.
    Install OS X Using Internet Recovery
    Be sure you backup your files to an external drive or second internal drive because the following procedure will remove everything from the hard drive.
    Boot to the Internet Recovery HD:
    Restart the computer and after the chime press and hold down the COMMAND-OPTION- R keys until a globe appears on the screen. Wait patiently - 15-20 minutes - until the Recovery main menu appears.
    Partition and Format the hard drive:
    Select Disk Utility from the main menu and click on the Continue button.
    After DU loads select your newly installed hard drive (this is the entry with the mfgr.'s ID and size) from the left side list. Click on the Partition tab in the DU main window.
    Under the Volume Scheme heading set the number of partitions from the drop down menu to one. Click on the Options button, set the partition scheme to GUID then click on the OK button. Set the format type to Mac OS Extended (Journaled.) Click on the Security button. Set Zero Data - one-pass, click on the OK button. Click on the Partition button and wait until the process has completed. Quit DU and return to the main menu.
    Reinstall OS X: Select Reinstall OS X and click on the Install button. Be sure to select the correct drive to use if you have more than one.
    Note: You will need an active Internet connection. I suggest using Ethernet if possible because it is three times faster than wireless.
    This should restore the version of OS X originally pre-installed on the computer.

  • A SQL tuning issue-sql runs much slower in test than in production?

    Hi Buddies,
    I am working on a sql tuning issue. A sql runs much slower in test than in production.
    I compared the two explain plans in test and production
    seems in test, CBO refuses to use index SUBLEDGER_ENTRY_I2.
    we rebuile it and re-gether that index statistcs. run, still slow..
    I compared the init.ora parameters like hash_area_size, sort_area_size in test, they are same as production.
    I wonder if any expert friend can show some light.
    in production,
    SQL> set autotrace traceonly
    SQL> SELECT rpt_horizon_subledger_entry_vw.onst_offst_cd,
    2 rpt_horizon_subledger_entry_vw.bkng_prd,
    3 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    4 rpt_horizon_subledger_entry_vw.jrnl_id,
    5 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    6 rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    7 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    8 rpt_horizon_subledger_entry_vw.crprt_melob_cd AS corp_mlb_cd,
    rpt_horizon_subledger_entry_vw.onst_offst_cd, SUM (amt) AS amount
    9 10 FROM rpt_horizon_subledger_entry_vw
    11 WHERE rpt_horizon_subledger_entry_vw.bkng_prd = '092008'
    12 AND rpt_horizon_subledger_entry_vw.jrnl_id = 'RCS0002100'
    13 AND rpt_horizon_subledger_entry_vw.systm_afflt_cd = 'SAFF01'
    14 GROUP BY rpt_horizon_subledger_entry_vw.onst_offst_cd,
    15 rpt_horizon_subledger_entry_vw.bkng_prd,
    16 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    17 rpt_horizon_subledger_entry_vw.jrnl_id,
    18 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    19 rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    20 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    21 rpt_horizon_subledger_entry_vw.crprt_melob_cd,
    22 rpt_horizon_subledger_entry_vw.onst_offst_cd;
    491 rows selected.
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=130605 Card=218764 B
    ytes=16407300)
    1 0 SORT (GROUP BY) (Cost=130605 Card=218764 Bytes=16407300)
    2 1 VIEW OF 'RPT_HORIZON_SUBLEDGER_ENTRY_VW' (Cost=129217 Ca
    rd=218764 Bytes=16407300)
    3 2 SORT (UNIQUE) (Cost=129217 Card=218764 Bytes=35877296)
    4 3 UNION-ALL
    5 4 HASH JOIN (Cost=61901 Card=109382 Bytes=17719884)
    6 5 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_ACCNT'
    (Cost=2 Card=111 Bytes=3774)
    7 5 HASH JOIN (Cost=61897 Card=109382 Bytes=14000896
    8 7 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACC
    OUNT' (Cost=2 Card=57 Bytes=1881)
    9 7 HASH JOIN (Cost=61893 Card=109382 Bytes=103912
    90)
    10 9 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cost=
    34 Card=4282 Bytes=132742)
    11 9 HASH JOIN (Cost=61833 Card=109390 Bytes=7000
    960)
    12 11 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGE
    R_ENTRY' (Cost=42958 Card=82076 Bytes=3611344)
    13 12 INDEX (RANGE SCAN) OF 'SUBLEDGER_ENTRY_I
    2' (NON-UNIQUE) (Cost=1069 Card=328303)
    14 11 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_
    LINK' (Cost=14314 Card=9235474 Bytes=184709480)
    15 4 HASH JOIN (Cost=61907 Card=109382 Bytes=18157412)
    16 15 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_ACCNT'
    (Cost=2 Card=111 Bytes=3774)
    17 15 HASH JOIN (Cost=61903 Card=109382 Bytes=14438424
    18 17 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACC
    OUNT' (Cost=2 Card=57 Bytes=1881)
    19 17 HASH JOIN (Cost=61899 Card=109382 Bytes=108288
    18)
    20 19 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cost=
    34 Card=4282 Bytes=132742)
    21 19 HASH JOIN (Cost=61838 Card=109390 Bytes=7438
    520)
    22 21 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGE
    R_ENTRY' (Cost=42958 Card=82076 Bytes=3939648)
    23 22 INDEX (RANGE SCAN) OF 'SUBLEDGER_ENTRY_I
    2' (NON-UNIQUE) (Cost=1069 Card=328303)
    24 21 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_
    LINK' (Cost=14314 Card=9235474 Bytes=184709480)
    Statistics
    25 recursive calls
    18 db block gets
    343266 consistent gets
    370353 physical reads
    0 redo size
    15051 bytes sent via SQL*Net to client
    1007 bytes received via SQL*Net from client
    34 SQL*Net roundtrips to/from client
    1 sorts (memory)
    1 sorts (disk)
    491 rows processed
    in test
    SQL> set autotrace traceonly
    SQL> SELECT rpt_horizon_subledger_entry_vw.onst_offst_cd,
    2 rpt_horizon_subledger_entry_vw.bkng_prd,
    3 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    4 rpt_horizon_subledger_entry_vw.jrnl_id,
    5 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    6 7 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    8 rpt_horizon_subledger_entry_vw.crprt_melob_cd AS corp_mlb_cd,
    9 rpt_horizon_subledger_entry_vw.onst_offst_cd, SUM (amt) AS amount
    10 FROM rpt_horizon_subledger_entry_vw
    11 WHERE rpt_horizon_subledger_entry_vw.bkng_prd = '092008'
    12 AND rpt_horizon_subledger_entry_vw.jrnl_id = 'RCS0002100'
    AND rpt_horizon_subledger_entry_vw.systm_afflt_cd = 'SAFF01'
    13 14 GROUP BY rpt_horizon_subledger_entry_vw.onst_offst_cd,
    15 rpt_horizon_subledger_entry_vw.bkng_prd,
    16 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    17 rpt_horizon_subledger_entry_vw.jrnl_id,
    18 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    rpt_horizon_subledger_entry_vw.crprt_melob_cd,
    rpt_horizon_subledger_entry_vw.onst_offst_cd; 19 20 21 22
    no rows selected
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=92944 Card=708 Bytes
    =53100)
    1 0 SORT (GROUP BY) (Cost=92944 Card=708 Bytes=53100)
    2 1 VIEW OF 'RPT_HORIZON_SUBLEDGER_ENTRY_VW' (Cost=92937 Car
    d=708 Bytes=53100)
    3 2 SORT (UNIQUE) (Cost=92937 Card=708 Bytes=124962)
    4 3 UNION-ALL
    5 4 HASH JOIN (Cost=46456 Card=354 Bytes=60180)
    6 5 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACCOU
    NT' (Cost=2 Card=57 Bytes=1881)
    7 5 NESTED LOOPS (Cost=46453 Card=354 Bytes=48498)
    8 7 HASH JOIN (Cost=11065 Card=17694 Bytes=1362438
    9 8 HASH JOIN (Cost=27 Card=87 Bytes=5133)
    10 9 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cos
    t=24 Card=87 Bytes=2175)
    11 9 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_
    ACCNT' (Cost=2 Card=111 Bytes=3774)
    12 8 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_LI
    NK' (Cost=11037 Card=142561 Bytes=2566098)
    13 7 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGER_EN
    TRY' (Cost=2 Card=1 Bytes=60)
    14 13 INDEX (UNIQUE SCAN) OF 'SUBLEDGER_ENTRY_PK'
    (UNIQUE) (Cost=1 Card=1)
    15 4 HASH JOIN (Cost=46456 Card=354 Bytes=64782)
    16 15 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACCOU
    NT' (Cost=2 Card=57 Bytes=1881)
    17 15 NESTED LOOPS (Cost=46453 Card=354 Bytes=53100)
    18 17 HASH JOIN (Cost=11065 Card=17694 Bytes=1362438
    19 18 HASH JOIN (Cost=27 Card=87 Bytes=5133)
    20 19 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cos
    t=24 Card=87 Bytes=2175)
    21 19 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_
    ACCNT' (Cost=2 Card=111 Bytes=3774)
    22 18 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_LI
    NK' (Cost=11037 Card=142561 Bytes=2566098)
    23 17 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGER_EN
    TRY' (Cost=2 Card=1 Bytes=73)
    24 23 INDEX (UNIQUE SCAN) OF 'SUBLEDGER_ENTRY_PK'
    (UNIQUE) (Cost=1 Card=1)
    Statistics
    1134 recursive calls
    0 db block gets
    38903505 consistent gets
    598254 physical reads
    60 redo size
    901 bytes sent via SQL*Net to client
    461 bytes received via SQL*Net from client
    1 SQL*Net roundtrips to/from client
    34 sorts (memory)
    0 sorts (disk)
    0 rows processed
    Thanks a lot in advance
    Jerry

    Hi
    Basically there are two kinds of tables
    - fact
    - lookup
    The number of records in a lookup table is usually small.
    The number of records in a fact table is usually huge.
    However, in test systems the number of records in a fact table is often also small.
    This results in different execution plans.
    I notice again you don't post version and platform info, and you didn't make sure your explain is properly idented
    Please read the FAQ to make sure it is properly idented.
    Also using the word 'buddies' is as far as I am concerned nearing disrespect and rudeness.
    Sybrand Bakker
    Senior Oracle DBA

  • Lightroom 2.4 Is Too Slow To Use As A Professional Product

    Hello -
    I would like to know how to get Lightroom to respond in less than 4-8 seconds for almost any task.
    From returning to Grid mode (4.5 seconds) to adjusting a routine crop and angle (5-7 seconds per move, 60-90 seconds total) to simply shift-selecting three to fifty photos (4-9 seconds) no matter what I do in Lightroom, it is worse that Photoshop 1.0 on a Mac Plus. Really.
    Returning to Lightroom from another app - 7 seconds.
    Getting the menus to drop down - 4 seconds.
    My lightroom settings:
    standard preview size: 1680
    Quality: High
    Discard 1:1: Never
    write changes to XMP: off
    Catalog: 127 mb on F drive (24 gig free)
    Catalog: 13.5k photos, 95 gig on F drive
    Cache: 6 gig on F drive
    NVIDIA settings: performance over quality
    My computer:
    HP 8730 elitebook
    4 gig ram
    XP Pro, SP2
    Proc: Core 2 Duo T9400, 2666 MHz (10 x 267)
    DirectX    4.09.00.0904 (DirectX 9.0c)
    Chipset: Cantiga PM45
    Video: NVIDIA Quadro FX 3700M  (1024 MB)
    (2x) 250 gig 7200 rmp HD
    Before anyone says to export to a new catalog and re-import, etc, use a smaller catalog, etc, be advised that I have done all of that. I have imported only 250 images to a new catalog, on a newly installed OS, with Lightroom as the ONLY application. Still it acts this sluggish.
    I have scoured the boards for solutions, and having tried all the varied fixes to no avail, I really quite strongly feel that a person should not EVER be in the position or running some of the most powerful available hardware and STILL have to wait interminable seconds for Lightroom to respond.
    This software, as it stands now on the PC platform (unless I am missing something quite obvous) is absolutely unusable in its present state.
    Can someone from Adobe or a board guru please respond in kind to my request for help?

    Photo_op8 wrote:
    BradKaye wrote:
    I'd have to agree with the subject of this 100%, even though I'm not experiencing anywhere near the levels of lag ellsworth is on any of my primary workhorses.
    1st-Gen 17" MacBook Pro 2.4Ghz Intel Core 2 Duo with Hi-Rez Screen (about 2 years old)
    1st-Gen MacPro (Dual Twin-Core 3Ghz, 3x, 10k-RPM 150GB Raptor RAID-0, partitioned to System/Scratch/Files (in that order) 16GB RAM...blah blah (about 2.2 years old)
    3 week old MacPro  2x2.66 Quad-Core MacPro, 16GB RAM, Multiple Partitioned (System/Scratch/Files)  Internal 4x-1.5TB RAID-0, 4x 1.5TB eSATA RAID-0, 2x NVIDIA GT 120's, 30" NEC 3090WQXi, 26" LaCie 526, 24" Apple LED Display
    OS 10.5.8 on all machines.
    Let me repeat-
    Partitioning ONE drive to contain system files, application, cache, catalog and photos=SLOW.
    It really doesn't matter that the drive is 7200 or 10k rpm if ALL items reside on the same drive. The MacPro has four bays. One drive for system and catalog, second for cache, third for photos=FAST.
    Wow, I'm glad you repeated that, and used capitalized text so that I would be sure to understand.
    Clearly, the benchmarks I have off of the various evolutions of my last 3, $10k computers must be in error. (single system disk, RAID-0 system disk, 10k RPM system disk, RAID-0 3x -10K rpm system disks, multiple scratch volumes on different drives, singular scratch volumes on RAIDed drives, tested via benchmark programs and application specific batch tests, etc.etc.)
    More importantly as an issue of forum protocol, you didn't actually read my post.
    Brad Kaye wrote:
    I partition about 10% of the outside of a drive as separate and put the information I need to get to most quickly which tends to be the system folder then scratch disk in a another separate partition and then I leave the rest on that drive as gobs of 'B' space for media and other crap files, reserving the first 10%-30% of the outside of a different drive (or RAID-0 sets) for my files and Lightroom catalogs.
    I'm throwing 370 mb/sec between two individual sets of 4 drive RAID-0 sets.  Single 7200rpm drives sling about 80 mb/sec.  My drive arrangement isn't the reason Lightroom IS TOO SLOW TO USE AS A PROFESSIONAL PRODUCT.
    Also, since I referenced and linked the former software engineer turned photo FREAK in my post, whose blog I started reading last year corroborating the decisions I've made in my previous system setups with immaculate documentation and testing procedure your response even more seems like the entire reason I rarely bother posting in community forums.  You seem to be here to spout, not to contribute and learn.
    Take a look at all of the info contained here: (I've already read it, all of it, and more)
    Diglloyd Mac Performance Guide
    and if you still think I'm wrong, lets have an informed debate about it.  Really.
    Otherwise, lets please keep this topic moving forward with the issues Adobe needs to address next to make Lightroom a better product for professional photographers, and specifically, help out ellsworth999 who started this topic, who seems to have a big helping of problems on his plate. I can't speak to him directly of his problems since I'm not using the Windows version of the software.

  • Sql query slow in new redhat enviornment

    We just migrated to a new dev environment in Linux REDHAT5, and now the query is very slow, and I used the TOAD to run the query, it took like 700 msecond to finish, however from any server connection, the sql query takes hours to finish.
    I checked toad monitor, it said need to increase db_buffer_cache and shared pool too small.
    Also three red alert from toad is:
    1. Library Cache get hit ratio: Dynamic or unsharable sql
    2. Chained fetch ratio: PCT free too low for a table
    3. parse to execute ratio: HIgh parse to execute ratio.
    App team said it ran real quick in the old AIX system, however I ran it in old system, and monitored in the toad, it gave me all same 5 red alerts in old system, and it did provide query results a lot quicker though.
    Here is the parameters in the old system (11gr1 on AIX):
    SQL> show parameter target
    NAME TYPE VALUE
    -------------------------------- archive_lag_target integer 0
    db_flashback_retention_target integer 1440
    fast_start_io_target integer 0
    fast_start_mttr_target integer 0
    memory_max_target big integer 0
    memory_target big integer 0
    pga_aggregate_target big integer 278928K
    sga_target big integer 0
    SQL> show parameter shared
    NAME TYPE VALUE
    -------------------------------- hi_shared_memory_address integer 0
    max_shared_servers integer
    shared_memory_address integer 0
    shared_pool_reserved_size big integer 31876710
    shared_pool_size big integer 608M
    shared_server_sessions integer
    shared_servers integer 0
    SQL> show parameter db_buffer
    SQL> show parameter buffer
    NAME TYPE VALUE
    -------------------------------- buffer_pool_keep string
    buffer_pool_recycle string
    db_block_buffers integer 0
    log_buffer integer 2048000
    use_indirect_data_buffers boolean FALSE
    SQL>
    In new 11gr2 Linux REDHAT parameter:
    NAME TYPE VALUE
    ----------- archive_lag_target integer 0
    db_flashback_retention_target integer 1440
    fast_start_io_target integer 0
    fast_start_mttr_target integer 0
    memory_max_target big integer 2512M
    memory_target big integer 2512M
    parallel_servers_target integer 192
    pga_aggregate_target big integer 0
    sga_target big integer 1648M
    SQL> show parameter shared
    NAME TYPE VALUE
    ----------- hi_shared_memory_address integer 0
    max_shared_servers integer
    shared_memory_address integer 0
    shared_pool_reserved_size big integer 28M
    shared_pool_size big integer 0
    shared_server_sessions integer
    shared_servers integer 1
    SQL> show parameter buffer
    NAME TYPE VALUE
    ----------- buffer_pool_keep string
    buffer_pool_recycle string
    db_block_buffers integer 0
    log_buffer integer 18857984
    use_indirect_data_buffers boolean FALSE
    SQL>
    Please help. Thanks in advance.

    846422 wrote:
    why need ddl? we have a sql query slow.The DDL shows the physical structure of the table and physical storage characteristics. All relevant in performance tuning.
    As for the SQL query being slow. It is not.
    You have not provided any evidence that it is slow. And no, comparing performance with a totally different system is not a valid baseline for comparison. (most cars have 4 wheels, a gearbox and a steering wheel - but that does not mean you can compare different cars like a VW Beetle with a VW Porsche)
    What is slow? What are the biggest wait states for the SQL? What does the execution plan say?
    You have not defined a problem - you identified a symptom called "+query is slow+". You need to diagnose the condition by determining exactly what the SQL qeury is doing in the database. (and please, do not use TOAD and similar tools in an attempt to do this - do it properly instead)

  • MDX query runs on SQL but does not populate 'Measures' in Powerpivot

    I ran MDX query on SQL 2012 for looking the products and their shipping status
    The code is given below:
    WITH 
    set Date1 AS 
         strtomember( "[Time].[2014].[Q4].[2014-Jun].[" +format(now()-2, "MM") + "/"+format(now()-2, "dd") + "/" +format(now()-2, "yyyy") + "]")
    set Item1 as
    { [Item].[ItemID].ALLMEMBERS}
    SELECT NON EMPTY { [Measures].[Shipped Qty], [Measures].[Shipped Amt], [Measures].[Ordered Qty], [Measures].[Ordered Amt] }
    ON COLUMNS, 
    Date1*[Item1] ON ROWS FROM [ABCD]
     WHERE ( 
                 [Channel].&[OL], 
                 [Geography].&[A.9])
    The query gives desired output when run in SQL. But when I run the same query in while pulling data in PowerPivot, Dimensions i.e. Date and Item are populated but measures were not populated. Can anyone please help me understand how to populated Measures
    in PowerPivot?
     

    Hi apoorv,
    It was hard to reproduce this issue on my test environment. I would suggest you try to use a simple MDX query to check if it will retrun the correct measures data. Please see the screenshot below:
    In addition, you can also create a SQL Server Analysis Services trace to capture some events for furhter investigation.
    Use SQL Server Profiler to Monitor Analysis Services:http://technet.microsoft.com/en-us/library/ms174946.aspx
    Hope this helps.
    Elvis Long
    TechNet Community Support

  • Query slow down when added a where clause

    I have a procedure that has performance issue, so I copy some of the query and run in the sql plus and try to spot which join cause the problem, but I get a result which I can figuer out why. I have a query which like below:
    Select Count(a.ID) From TableA a
    -- INNER JOIN other tables
    WHERE a.TypeID = 2;
    TableA has 140000 records, when the where clause is not added, the count return quite quick, but if I add the where clause, then the query slow down and seems never return so I have to kill my SQL Plus session. TableA has index on TypeID and TypeID is a number type. When TablA has 3000 records, the procedure return very quick, but it slow down and hang there when the TableA contains 140000 records. Any idea why this will slow down the query?
    Also, the TypeID is a foreign key to another table (TableAType), so the query above can written as :
    Select Count(a.ID) From TableA a
    -- INNER JOIN other tables
    INNER JOIN TableAType atype ON a.TypeID = atype.ID
    WHERE atype.Name = 'typename';
    TableAType table is a small table only contains less than 100 records, in this case, would the second query be more efficient to the first query?
    Any suggestions are welcome, thanks in advance...
    Message was edited by:
    user500168

    TableA now has 230000 records and 28000 of them has the TypeID 2.
    I haven't use the hint yet but thank you for your reply which let me to to run a query to check how many records in TableA has TypeID 2. When I doing this, it seems pretty fast. So I begin with the select count for TableA only and gradually add table to join and seems the query is pretty fast as long as TableA is the fist table to select from.
    Before in my query TableA is the second table to join from, there is another table (which is large as well but not as large as TableA) before TableA. So I think this is why it runs slow before. I am not at work yesterday so the query given in my post is based on my roughly memory and I forget to mention another table is joined before TableA, really sorry about that.
    I think I learn a lesson here, the largest table need to be in the begining of the select statement...
    Thank you very much everyone.

  • M9 Very Slow Charging Even With Quick Charge 2.0 Chargers

    I have a continuing problem with very slow charging even with Quick Charge 2.0 chargers. For example last night I had the M9 plugged into a fast charger for over 12 hrs and it went from 5% battery to 87%.   I checked  the Settings - Battery usage history - and it displays "NO APPLICATIONS FOUND" - very strange...
    Something is very wrong. I have multiple 2 amp and quick chargers all with poor charging times. I have tried each of these:
    https://www.amazon.com/gp/product/B00FW1GE9S/ref=oh_aui_detailpage_o03_s00?ie=UTF8&psc=1[/
    https://www.amazon.com/gp/product/B00L40YJ8O/ref=oh_aui_detailpage_o03_s00?ie=UTF8&psc=1[/
    https://www.amazon.com/gp/product/B00TAOEKPS/ref=oh_aui_detailpage_o03_s00?ie=UTF8&psc=1[/
    https://www.amazon.com/gp/product/B00HP0FOCM/ref=oh_aui_detailpage_o03_s00?ie=UTF8&psc=1[/
    The Ampere app displays varying charger power from 15 mA to 880 mA so I don't think it is accurate.
    Testing with multiple chargers point to the problem in the M9. This is maddening. Seriously considering returning the M9.
    Any feedback?
    ++md++

    Perhaps this helps:
    http://www.xperiablog.net/2015/07/02/these-sony-xperia-devices-support-qualcomm-quick-charge-2-0/

  • Sql query slowness due to rank and columns with null values:

        
    Sql query slowness due to rank and columns with null values:
    I have the following table in database with around 10 millions records:
    Declaration:
    create table PropertyOwners (
    [Key] int not null primary key,
    PropertyKey int not null,    
    BoughtDate DateTime,    
    OwnerKey int null,    
    GroupKey int null   
    go
    [Key] is primary key and combination of PropertyKey, BoughtDate, OwnerKey and GroupKey is unique.
    With the following index:
    CREATE NONCLUSTERED INDEX [IX_PropertyOwners] ON [dbo].[PropertyOwners]    
    [PropertyKey] ASC,   
    [BoughtDate] DESC,   
    [OwnerKey] DESC,   
    [GroupKey] DESC   
    go
    Description of the case:
    For single BoughtDate one property can belong to multiple owners or single group, for single record there can either be OwnerKey or GroupKey but not both so one of them will be null for each record. I am trying to retrieve the data from the table using
    following query for the OwnerKey. If there are same property rows for owners and group at the same time than the rows having OwnerKey with be preferred, that is why I am using "OwnerKey desc" in Rank function.
    declare @ownerKey int = 40000   
    select PropertyKey, BoughtDate, OwnerKey, GroupKey   
    from (    
    select PropertyKey, BoughtDate, OwnerKey, GroupKey,       
    RANK() over (partition by PropertyKey order by BoughtDate desc, OwnerKey desc, GroupKey desc) as [Rank]   
    from PropertyOwners   
    ) as result   
    where result.[Rank]=1 and result.[OwnerKey]=@ownerKey
    It is taking 2-3 seconds to get the records which is too slow, similar time it is taking as I try to get the records using the GroupKey. But when I tried to get the records for the PropertyKey with the same query, it is executing in 10 milliseconds.
    May be the slowness is due to as OwnerKey/GroupKey in the table  can be null and sql server in unable to index it. I have also tried to use the Indexed view to pre ranked them but I can't use it in my query as Rank function is not supported in indexed
    view.
    Please note this table is updated once a day and using Sql Server 2008 R2. Any help will be greatly appreciated.

    create table #result (PropertyKey int not null, BoughtDate datetime, OwnerKey int null, GroupKey int null, [Rank] int not null)Create index idx ON #result(OwnerKey ,rnk)
    insert into #result(PropertyKey, BoughtDate, OwnerKey, GroupKey, [Rank])
    select PropertyKey, BoughtDate, OwnerKey, GroupKey,
    RANK() over (partition by PropertyKey order by BoughtDate desc, OwnerKey desc, GroupKey desc) as [Rank]
    from PropertyOwners
    go
    declare @ownerKey int = 1
    select PropertyKey, BoughtDate, OwnerKey, GroupKey
    from #result as result
    where result.[Rank]=1
    and result.[OwnerKey]=@ownerKey
    go
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • The movies from my Iphone 5 are not coming across on Apple TV. The little spinning beach ball appears like it is thinking.  It use to work but quick working a month ago.

    The movies from my Iphone 5 are not coming across on Apple TV. The little spinning beach ball appears like it is thinking.  It use to work fine but quick working a month ago. I mirror from my phone to the Apple TV.  Still images and everything I do on the phone is mirrored on to the TV except when I want to show a YouTube video or any video and then the screen on the phone and the TV screen go black and the spinning beach ball appears.  The videos open right up on the phone if I do not have mirroring on and going through the Apple TV.  I just downloaded the latest update to the phone it did not fix the problem.  Please help.  Thanks, Elvis G

    Hello Elvis G,
    After reviewing your post, I have located an article that can help in this situation. It contains a number of troubleshooting steps concerning device issues. You may want to consider a reset, then testing to see if the issue is resolved:
    Turn your iOS device off and on (restart) and reset
    http://support.apple.com/kb/ht1430
    How to reset
    You should reset your device as a last resort and only if you can't restart it.
    To reset, press and hold both the Sleep/Wake and Home buttons for at least 10 seconds, until you see the Apple logo.
    Thank you for contributing to Apple Support Communities.
    Cheers,
    BobbyD

  • Cloning a test CMS from a Production environment - process?

    Hi there - I am attempting to clone a test environment from a production CMS system database.
    ENVIRONMENT
    BO Version: XI3.1
    O/S: Linux
    Web Server: JBOSS 4.3
    Database platform: Oracle 10g
    BOXI SETUP
    1. BOXI environment type on Production (clustered with two nodes.)
    *. Linux servernames for each BOXI node=Prod100 and Prod 101.
    *. JBOSS ports on both nodes=15000. CMS port on both nodes=6400.
    *. Database Sid=BOCMS@PROD
    2. BOXI environment type on Perf (clustered with two nodes)
    *. Linux servernames for each BOXI node=Perf100 and Perf 101.
    *. JBOSS ports on both nodes=15100.
    *. CMS port on both nodes=6400.
    *. Database Sid=BOCMS@PERF
    Process and requirement
    For each project, as part of performance testing, we copy the production database (CMS / system database) in order to clone it. The idea is that we have a copy of production which we can then use for performance testing. We do not need to rebuild the BOXI 3.1 server (that remains there), but we need to rebuild system database. In version five we would copy the BO_REPADMIN schema into the PERF database and then change the connections in Supervisor. Can this process also be done in XIR3.1 ?
    Options considered so far
    1. The DBAs recommended just copying the BOXI CMS / System Database tables from PROD to the PERF database. My question is - don't we then need to change the settings in the CMS database (PERF) to somehow so point to the correct physical servers / correct SIA etc.? Could I just alter th FRS references and use the cmsdbsetup.sh to alter all the details or are there other steps required (manally altering ccm.config) ?
    2. Copying the CMS from the PROD to the PERF CMS schema. However it seems like this is not an option as the Admin guides stipulates:
    "However, you cannot change the BusinessObjects Enterprise installation path, operating system, and machine name"
    As the installation path for the BOXI performance environment is different to production and they reside on different physical machines it sounds to me that this isn't an option, right?
    3. Use the cmsdbsetup.sh script - create a blank new CMS on the PERF database then migrate all production objects from PROD to PERF using the Import Wizard.
    Could anyone recommend further options? Thanks in advance.

    Hi;
    Can any one explain the detaild process about Operating level snapshot to prepare a test setup from a production databaseMainly; We are using IBM disk and our disk team take snapshot of disk and we can work on those snapshot than storage team again revert back system orginal state.
    As you can see its performed by storage team contact wiht your vendor.
    If you dont have storage team than I suggest see my below post:
    http://heliosguneserol.wordpress.com/2011/12/08/11g-snapshot-standby-for-real-time-testing/
    Regard
    Helios

Maybe you are looking for

  • Macbook pro won't connect with iphoto library

    One of the reasons I upgraded to my new Macbook Pro in the first place is that iPhoto was hanging up or crashing on my old G5 tower, and now I can't seem to access my old iphoto library that I had backed up on my exteranl drive. More accurately, it c

  • Error in a page with a master-detail:Unable to resolve a Validator instance

    Hi all, I am using Jdeveloper 11.1.1.2 I have a page with a master table and a detail table. When I try to select a row in the some tables (master or detail), I have this error: <AbstractValidatorTag><createValidator> Unable to resolve a Validator in

  • Date dimension fragmentation from 2 different database connections

    Hi Everyone, In obiee 10g, is it possible to implement fragmentation on a date dimension from 2 different physical database connections (not schemas)? For example: Physical connection & dimension 1: physical1.obiee_rpd.w_mcal_day_d.date physical conn

  • PDF attachments not showing

    Getting really annoyed now. A client of mine has 6 curve 9300s. The issue is when they are sent emails from macs containing PDFs. There appears to be no attachment. The issue is not the size of the PDF (tried very small PDF). Its not the mac itself.

  • Issues with Whatsapp in lumia 720

    I jus found a bug in Lumia 720. When ever i pause a song in the music player n i resume it later by clicking on the paused song icon title .. It directs me to home page of whatsapp ! .. How do i fix this ??? Also found Compatability issues with whats