How to force a parallel dml?

I have a DB 11g r2, and I use parallel queries (parallel delete) for some huge tables.
But, sometimes queries just don't want to go parallel..
to force parallelism, I use hint /*+ parallel(table_name, no_of_parallel_procs, no_of_instances) */ .-.. some_sql_condition..
note that same query works fine in some enviroments, so there's no error in sql hint, and there's no background processes (AFAIK)
db init parameters are:
parallel_degree_limit=cpu
parallel_degree_policy=manual
parallel_max_servers=20
parallel_server=false
parallel_threads_per_cpu=2
I could provide more database parameters if necessary.

Slave processes, which are required for DML, could be in use at the time of the DML running;
while the DELETE is executing, open another session and run the following:
select slave_name
      ,status, cpu_secs_total
from v$pq_slave;If this figure appears too low or no rows are returned then PARALLEL mode will not occur.

Similar Messages

  • How to make Auto Parallelism work?

    Hello,
    We are running Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production. We have upgraded from 11gR1 recently. We are building a data warehouse in the instance of Oracle I am writing about. Heretofore we had been using hints in queries to take advantage of parallelism. We were hoping to achieve a performance boost by using the new "automatic degree of parallelism" support in 11gR2, but are having some problems understanding how to use it properly. From reading the Oracle white paper on parallelism http://www.oracle.com/us/products/database/twp-parallel-execution-fundamentals-133639.pdf , we had thought that by setting the database initialization parameters properly, and removing the query-level hints, and removing the parallel specifications in the table declarations, that Oracle would automatically determine the correct level of parallelism, and apply it. Instead, we have found that the removal of the hints and declarations results in a single-thread execution for our data-load queries, which load 100s of millions of rows, involving joins among tables that each contain more than 90 million rows. Reapplying the hints significantly improves our throughput. During the time we are running the data loads, the data load is the only significant user on our system, and our goal is to complete execution as quickly as possible.
    The PLSQL packages that load the data issue a "ALTER SESSION ENABLE PARALLEL DML " before running the loads.
    Here are our database initialization parameter settings that relate to parallelism:
    NAME TYPE VALUE
    fast_start_parallel_rollback string LOW
    parallel_adaptive_multi_user boolean FALSE
    parallel_automatic_tuning boolean FALSE
    parallel_degree_limit string CPU
    parallel_degree_policy string AUTO
    parallel_execution_message_size integer 16384
    parallel_force_local boolean FALSE
    parallel_instance_group string
    parallel_io_cap_enabled boolean FALSE
    parallel_max_servers integer 128
    parallel_min_percent integer 0
    parallel_min_servers integer 8
    parallel_min_time_threshold string AUTO
    parallel_server boolean FALSE
    parallel_server_instances integer 1
    parallel_servers_target integer 128
    parallel_threads_per_cpu integer 2
    recovery_parallelism integer 4
    Any help you can provide to help us use parallelism more effectively in 11gR2 would be gratefully appreciated.

    One common mistake is to not run the IO CALIBRATE before enabling AUTODOP=AUTO
    you may need to tweak this value or set it manually to something like (0,200,...) -- Check MOS for more on this
    this requires a bounce of all instancesAlso: there are thresholds that need to be surpassed:
    .e.g the default 10 seconds for parallel_min_time_threshold would suppress parallelism attempts for shorter queries vs. longer
    If default load has thousand of short queries this may not be the parameter for you: However you can decrease the threshold to a lower value
    such as 5 seconds or less.
    EM is a good place to monitor parallelism in real time ( or historically)
    - You might consider testing and reviewing if the above is relevant
    Last- Trace files including 10046 or 10053 should show AUTODOP being used
    - If this is LIMITED but the parameter is AUTO, you probably have no io calibrate stats
    Take care
    CL

  • Oracle 11.2 - Perform parallel DML on a non partitioned table with LOB column

    Hi,
    Since I wanted to demonstrate new Oracle 12c enhancements on SecureFiles, I tried to use PDML statements on a non partitioned table with LOB column, in both Oracle 11g and Oracle 12c releases. The Oracle 11.2 SecureFiles and Large Objects Developer's Guide of January 2013 clearly says:
    Parallel execution of the following DML operations on tables with LOB columns is supported. These operations run in parallel execution mode only when performed on a partitioned table. DML statements on non-partitioned tables with LOB columns continue to execute in serial execution mode.
    INSERT AS SELECT
    CREATE TABLE AS SELECT
    DELETE
    UPDATE
    MERGE (conditional UPDATE and INSERT)
    Multi-table INSERT
    So I created and populated a simple table with a BLOB column:
    SQL> CREATE TABLE T1 (A BLOB);
    Table created.
    Then, I tried to see the execution plan of a parallel DELETE:
    SQL> EXPLAIN PLAN FOR
      2  delete /*+parallel (t1,8) */ from t1;
    Explained.
    SQL> select * from table(dbms_xplan.display);
    PLAN_TABLE_OUTPUT
    Plan hash value: 3718066193
    | Id  | Operation             | Name     | Rows  | Cost (%CPU)| Time     |    TQ  |IN-OUT| PQ Distrib |
    |   0 | DELETE STATEMENT      |          |  2048 |     2   (0)| 00:00:01 |        |      |            |
    |   1 |  DELETE               | T1       |       |            |          |        |      |            |
    |   2 |   PX COORDINATOR      |          |       |            |          |        |      |            |
    |   3 |    PX SEND QC (RANDOM)| :TQ10000 |  2048 |     2   (0)| 00:00:01 |  Q1,00 | P->S | QC (RAND)  |
    |   4 |     PX BLOCK ITERATOR |          |  2048 |     2   (0)| 00:00:01 |  Q1,00 | PCWC |            |
    |   5 |      TABLE ACCESS FULL| T1       |  2048 |     2   (0)| 00:00:01 |  Q1,00 | PCWP |            |
    PLAN_TABLE_OUTPUT
    Note
       - dynamic sampling used for this statement (level=2)
    And I finished by executing the statement.
    SQL> commit;
    Commit complete.
    SQL> alter session enable parallel dml;
    Session altered.
    SQL> delete /*+parallel (t1,8) */ from t1;
    2048 rows deleted.
    As we can see, the statement has been run as parallel:
    SQL> select * from v$pq_sesstat;
    STATISTIC                      LAST_QUERY SESSION_TOTAL
    Queries Parallelized                    1             1
    DML Parallelized                        0             0
    DDL Parallelized                        0             0
    DFO Trees                               1             1
    Server Threads                          5             0
    Allocation Height                       5             0
    Allocation Width                        1             0
    Local Msgs Sent                        55            55
    Distr Msgs Sent                         0             0
    Local Msgs Recv'd                      55            55
    Distr Msgs Recv'd                       0             0
    11 rows selected.
    Is it normal ? It is not supposed to be supported on Oracle 11g with non-partitioned table containing LOB column....
    Thank you for your help.
    Michael

    Yes I did it. I tried with force parallel dml, and that is the results on my 12c DB, with the non partitionned and SecureFiles LOB column.
    SQL> explain plan for delete from t1;
    Explained.
    | Id  | Operation             | Name     | Rows  | Cost (%CPU)| Time     |    TQ  |IN-OUT| PQ Distrib |
    |   0 | DELETE STATEMENT      |          |     4 |     2   (0)| 00:00:01 |        |      |            |
    |   1 |  DELETE               | T1       |       |            |          |        |      |            |
    |   2 |   PX COORDINATOR      |          |       |            |          |        |      |            |
    |   3 |    PX SEND QC (RANDOM)| :TQ10000 |     4 |     2   (0)| 00:00:01 |  Q1,00 | P->S | QC (RAND)  |
    |   4 |     PX BLOCK ITERATOR |          |     4 |     2   (0)| 00:00:01 |  Q1,00 | PCWC |            |
    |   5 |      TABLE ACCESS FULL| T1       |     4 |     2   (0)| 00:00:01 |  Q1,00 | PCWP |            |
    The DELETE is not performed in Parallel.
    I tried with another statement :
    SQL> explain plan for
    2        insert into t1 select * from t1;
    Here are the results:
    11g
    | Id  | Operation                | Name     | Rows  | Bytes | Cost (%CPU)| Time     |    TQ  |IN-OUT| PQ Distrib |
    |   0 | INSERT STATEMENT         |          |     4 |  8008 |     2   (0)| 00:00:01 |        |      |            |
    |   1 |  LOAD TABLE CONVENTIONAL | T1       |       |       |            |          |        |      |            |
    |   2 |   PX COORDINATOR         |          |       |       |            |          |        |      |            |
    |   3 |    PX SEND QC (RANDOM)   | :TQ10000 |     4 |  8008 |     2   (0)| 00:00:01 |  Q1,00 | P->S | QC (RAND)  |
    |   4 |     PX BLOCK ITERATOR    |          |     4 |  8008 |     2   (0)| 00:00:01 |  Q1,00 | PCWC |            |
    |   5 |      TABLE ACCESS FULL   | T1       |     4 |  8008 |     2   (0)| 00:00:01 |  Q1,00 | PCWP |            |
    12c
    | Id  | Operation                          | Name     | Rows  | Bytes | Cost (%CPU)| Time     |    TQ  |IN-OUT| PQ Distrib |
    |   0 | INSERT STATEMENT                   |          |     4 |  8008 |     2   (0)| 00:00:01 |        |      |            |
    |   1 |  PX COORDINATOR                    |          |       |       |            |          |        |      |            |
    |   2 |   PX SEND QC (RANDOM)              | :TQ10000 |     4 |  8008 |     2   (0)| 00:00:01 |  Q1,00 | P->S | QC (RAND)  |
    |   3 |    LOAD AS SELECT                  | T1       |       |       |            |          |  Q1,00 | PCWP |            |
    |   4 |     OPTIMIZER STATISTICS GATHERING |          |     4 |  8008 |     2   (0)| 00:00:01 |  Q1,00 | PCWP |            |
    |   5 |      PX BLOCK ITERATOR             |          |     4 |  8008 |     2   (0)| 00:00:01 |  Q1,00 | PCWC |            |
    It seems that the DELETE statement has problems but not the INSERT AS SELECT !

  • How can I implement parallel processing feature in Oracle 9i?

    I need to load (and transform) millions of data from Table1 to Table2. How can I speed the process using parallel processing. The server has 4 processors (running on Win2k Advanced server).
    Can anyone guide me what are the exact steps required to enable parallel loading. I've already partitioned Table1 into 4 parts.

    insert /*+ PARALLEL(8) */ into table A select * from table B ;
    or alter table A parallel 8 ;
    or alter session enable parallel dml ;
    Look for PARALLEL on http://tahiti.oracle.com for more infos.
    Fred

  • How to force refresh of data through browser or PDF?

    We have the dashboard set to refresh every minute.  We are pulling the data using XML from DB.  When we are in browser and clear the browser cache and then reload the .swf... the data is updated.  We haven't been able to figure out how to force the cache-clear and data refresh with either swf or pdf.
    Your help is greatly appreciated.

    Hi Jeff,
    Is the XML coming from a web page or a web server?
    If yes then you can give this a go. To stop the caching mark your web page with extra tags to say it has expired.
    HTML page example:
    <HEAD>
        < META HTTP-EQUIV="PRAGMA" CONTENT="NO-CACHE" />
        < META HTTP-EQUIV="EXPIRES" CONTENT="0" />
    </HEAD>
    JSP example:
    <%
      // Stop Internet Explorer from caching the results of this page.
      // We do this so that every time Xcelsius calls this page it see the latest results.
      response.setHeader("Cache-Control","no-cache"); //HTTP 1.1
      response.setHeader("Pragma","no-cache"); //HTTP 1.0
      response.setDateHeader ("Expires", 0); //prevents caching at the proxy server 
    %>
    Regards,
    Matt

  • How do i use parallel compression for 4 drum tracks in Logic pro 7.2.3 Please

    how do i use parallel compression for 4 drum tracks in Logic pro 7.2.3 Please. my version is 7.2.3 and my son did his drums on 4 tracks. 2 OH's 1 Kick and 1 Snare. Were confused because we cant figure out how to do parallel compression on the older logic version 7.2.3?

    I can't remember when the Aux tracks became the new buss tracks, so I'll spell this out using busses.
    The way I do it, is I assign all my drum tracks to a buss, say buss 1.
    Then, on buss 2, I assign it's INPUT to buss 1.
    So buss 1 (all the drums) are feeding into buss 2 simultaneously.
    Put a compressor on buss 2, turn the fader down all the way, then as the track is playing, bring buss 2's fader up slowly, until you hear the effect you're after.
    You could also do this using sends on the 4 drum tracks, set to feed the buss with the compressor on it.
    I hope I haven't forgotten anything particular to that version of Logic. See if this helps...

  • How do I clone Parallels?

    I need to clone Parallels (I have Parallels Desktop 7) separately from the Mac HD so backing up the computer won't work for me.
    How do I clone Parallels, meaning everything including files etc?
    I have internet banking set up on Parallels (Windows 7) and that was really difficult to set up so it's important that I clone Parallels exactly so that I can migrate it to the new computer and do internet banking on it.
    Also, after cloning, how do I restore the clone to my computer?
    I am moving from Lion to Snow Leopard by the way.
    The Parallels is on the Lion machine but I want to migrate to the Snow Leopard machine.
    I cannot use Carbon Copy Cloner to clone the HD including the Parallels file because there is a problem with that. I have to clone and migrate and restore Parallels separately.
    I will use an external disk to store the clone on.
    Thanks for help with this.

    Thanks for the instructions.
    I saw "Clone" in the menu of Parallels. And when I clicked on it, it asks where you want to save the clone. So that's why I thought you could clone Parallels.
    The support person outlined almost the same steps you did but she didn't really go into what to do after copying the Parallels folder into the Documents folder.
    When you'll start 'Parallels ' app., open this folder (from pull-down menu).
    In the Parallels menu, there is a "migrate" option. Should I use this?
    Also you can do online backup using Acronis. This appears in the menu.
    I will try your steps, Star1. Thanks.

  • How to force a new password in portal with LDAP user? external users

    With an external portal (used by agents that do not work for you or reside in your office), company policy is for password to be changed every qtr.
    If the users are creating as LDAP users how to force them to change their password when required?
    Is this a custom application that needs to be written so when they log into the portal if the qtr has expired the portal ask them to enter a new password that becomes valid for the next qtr.
    Versus internally deleting and emailing all the users a new password?

    Hi Glenn,
    We are getting one problem when we are creating user in LDAP and login with that user in  Portal that time we are getting Password change screen , but when we create a user in LDAP and change the password of that user in LDAP then when the user tries to  Login to portal that time we are not able to see the password change screen.
    But again if we change the password of that user through Portal we are able to see change password screen.
    can you help on this how we can force the user to change password when we are changing password in LDAP or in SAP System.
    Regards
    Trilochan

  • How to force the "Bluetooth Communicat​ions Port" to be one of COM1 to COM8 ports?

    Dear Lenovo Community, Happy Holidays to you all and wish you a great happy new year. Recently purchased a Bluetooth OBDII device and have difficulty making it to work with its provided software on my T61 (running original XP Home). My short story and question/problem is that I can open "My Bluetooth Places" and pair with the OBDII device as an "OBDII SPP Dev", but my T61 assigns serial port COM19 to it. The OBDII software which came with the device only can let user set to one of the COM1 to COM8 ports and in the properties of Bluetooth pairing, there is no way that I can select which COM port to use. I looked at the Device Manager and I do see these COM port assignments: COM4,5,6,7: Sierra Wireless (the HSDA modem in the laptop which I have never used BTW) COM 9,10,11,12,13,14,15,16,17: Bluetooth Serial Port COM 18, 19: Bluetooth Communications Port and I don't see anything for COM1,2,3, and 8 My question is how to force the computer/OS to assign one of the COM1 to COM8 ports to my device upon pairing? Can I disable the Sierra Wireless model from the COM ports list and hope this will happen? Thanks for your help and inputs beforehand. Regards, AL

    Hi, AL_K
    Have you attempted to change the port number in device manager itself? If you navigate to Device Manager and open the list of Ports, you can right-click on the device you wish to assign a different port number. After right-clicking, click Properties. There should be a tab called Port Settings. In here, you should find a setting to manually assign a port number.
    Good luck, and let me know how it goes,
    Adam
    Did someone help you today? Press the star on the left to thank them with a Kudo!
    If you find a post helpful and it answers your question, please mark it as an "Accepted Solution!" This will help the rest of the community with similar issues identify the verified solution and benefit from it.

  • How to find the parallelism degree of a query that has been finished?(10.2)

    We use Resource Manager, developers think that they can not get the parallelism they need for their queries because of the Resource Manager settings. I need to show them how many degree of parallelism each of their query have used, so I need a dictionary view to query this information. But couldn't find any relevant historical query information for this one, is there any dba_hist version of v$px_session on 10.2?
    Thank you.

    Thanks for this idea but I couldn't come to a point with it, any addtional comments are highly appreciated;
    SELECT qc_session_id,
           sql_id,
           to_char(sample_time,
                   'dd.mm.yyyy hh24:mi:ss'),
           COUNT(*) potantial_dop
      FROM dba_hist_active_sess_history
    GROUP BY qc_session_id,
              sql_id,
              to_char(sample_time,
                      'dd.mm.yyyy hh24:mi:ss')
    ORDER BY 4 DESC;

  • Adobe Acrobat 9 Standard is still seeing my old scanner install.  How to force it to see new scanner install.

    Adobe Acrobat 9 Standard is still seeing my old scanner install.  How to force it to see new scanner install?  Scanner is recognized by and works fine in Windows.  Acrobat however still sees Fujitsu fi-6130dj #3.  I need it to see Fujitsu fi-6130dj which is what is listed in Device Manager.  (Windows XP SP3).
    Thank you for any suggestions.

    To provide more detail and list what was done to resolve the issue but there may be a better way.
    Background.  When you install a USB device to a PC it is listed in Windows Device Manager as installed hardware.  If you take the USB plug for any device and plug it into a different USB port on the PC the PC will install drivers again for that device.  In Device Manager, if you set a system environment variable for device manager to show non-present devices you can see all of the times this may have occurred.  In this specific instance there were five installations in Device Manager for one piece of hardware, a Fujitsu fi-6130 scanner.  The first listing was Fujitsu fi-6130dj, then Fujitsu fi-6130dj #2, Fujitsu fi-6130dj #3 and so on.  Adobe Acrobat only saw the third instance of these drivers being installed.  Once I cleaned Device Manager of all Fujitsu scanner installs I was able to have a single set of drivers installed for this Fujitsu but Adobe still recognized the third instance of the driver install and I could not figure out how to force it to see what Device Manager was seeing, a single instance of the scanner being installed (with no #3 behind it).
    Steps performed to try to resolve.
    Repair install of Adobe - no luck - still sees the #3 instance of the scanner
    Deactivation, uninstall, reboot, reinstall, reactivation of Adobe - no luck, same behavior as above
    Complete removal of all scanner related software.  This included the ISIS drivers in Add/Remove Programs and all other related software.  This also included deleting the contents of the TWAIN_32 directory found within the root of the Windows directory and also included deleting the scanner from Device Manager.
    Rebooted.
    Reinstalled complete packages of both TWAIN and ISIS scanners - with the scanner disconnected.  Powered the scanner back on and let Windows install the drivers for the hardware.  Only at this point did Adobe see the new install of the Fujitsu scanner and no longer looked for the #3 instance of that scanner install.

  • How to force using the adobe reader to view the PDF by javascript / PHP / HTML?

    I use the standard way to embed a pdf to my website, however, when I open it the browser will choose the default pdf viewer instead of the adobe reader. I can change it manually by looking at the browser option but it would not be nice if the website visitor need to do this manually. So I would like to know how to force the browser to open the PDF in adobe by using some HTML header / javascript / php or any parameter in object tag. Thanks.
    <object id = 'zoomPage' type='application/pdf' data= '" + sourceFolder + "SingTao/2013/08/17/0/0/A/Content/pdf/pdf_" + currZoomPageNo + ".pdf#toolbar=1&navpanes=0&statusbar=0&messages=0&pagemode=read'><p>The PDF can not display</p></object>

    I know this problem related to the browser behaviour so I would target only :Chrome firefox and IE 9+ (it is ideal if IE8 can also adopt the solution) thanks

  • How to Force enable, silent updates for Adobe flash in windows 8

    How to Force enable updates, silently for Adobe flash in windows 8 using group policy?
    I have followed this Article:
    http://helpx.adobe.com/flash-player/kb/administration-configure-auto-update-notification.h tml
    http://gpyall.com/archives/disable-adobe-flash-updates-on-64-bit-windows-with-group-policy /
    This looks promising but not working.
    Also, I cannot even manually create a file, in Win 8  (C:\Windows\SysWOW64\Macromed\Flash)
    So how would group policy can put mms.cfg in this location with following vaule:
    AutoUpdateDisable=0
    SilentAutoUpdateEnable=1
    Thanks in Advance.

    You will find more information in http://wwwimages.adobe.com/www.adobe.com/content/dam/Adobe/en/devnet/flashplayer/pdfs/flas h_player_11_7_admin_guide.pdf

  • How to force my Web part to run regardless of users permissions

    I have created the following custom permission , which will allow users to Create items without being able to view,edit them:-
    $spweb=Get-SPWeb -Identity "http://vstg01";
    $spRoleDefinition = New-Object Microsoft.SharePoint.SPRoleDefinition;
    $spRoleDefinition.Name = "Submit only";
    $spRoleDefinition.Description = "Can submit/add forms/files/items into library or list but cannot view/edit them.";
    $spRoleDefinition.BasePermissions = "AddListItems, ViewPages, ViewFormPages, Open";
    $spweb.RoleDefinitions.Add($spRoleDefinition);
    $spweb.Dispose();
    then inside my "Issue Tracking List" i stop inheriting permission from team site , and i define the following permission for all users:-
    now users can add items and they can not view them ,, which is perfect :).
    But now i wanted to add a custom web part to my Create form which will hide certain fields if the user is not within specific group ,the web part looks as follow:-
    protected override void OnInit(EventArgs e)
    base.OnInit(e);
    InitializeControl();
    using (SPSite site = new SPSite(SPContext.Current.Site.Url))
    using (SPWeb web = site.OpenWeb())
    web.AllowUnsafeUpdates = true;
    SPGroup group = web.Groups["Intranet Visitors"];
    bool isUser = web.IsCurrentUserMemberOfGroup(group.ID);
    if (!isUser)
    SPList myList = web.Lists.TryGetList("Issue List");
    SPField titleField = myList.Fields.GetField("Category");
    titleField.Hidden = true;
    titleField.ShowInEditForm = false;
    titleField.ShowInNewForm = false;
    titleField.ShowInDisplayForm = false;
    titleField.Update();
    myList.Update();
    // web.AllowUnsafeUpdates = false;
    else
    SPList myList = web.Lists.TryGetList("Issue List");
    SPField titleField = myList.Fields.GetField("Title");
    titleField.Hidden = false;
    titleField.Update();
    myList.Update();
    // //web.AllowUnsafeUpdates = false;
    web.AllowUnsafeUpdates = false;
    then i deploy the web part and i add it to the Create form. but after doing so user are not able to create items and they will get the following error:-
    Sorry this site has not been shared with you
    so can anyone advice how to force my web part to run , without checking the users permissions or with minimal permssions ?

    in this case, use the elevated privileges to read/add/edit items with elevated privileges with below code.
    but make sure the page which you add this web part have at least read access to all user.
    SPSecurity.RunWithElevatedPrivileges(delegate()
    using (SPSite site = new SPSite(web.Site.ID))
    // implementation details omitted
    More: http://msdn.microsoft.com/en-us/library/microsoft.sharepoint.spsecurity.runwithelevatedprivileges.aspx
    Bistesh
    Ok after adding :-
    SPSecurity.RunWithElevatedPrivileges(delegate()
    users with the following permissions can create items:-
    "AddListItems, ViewPages, ViewFormPages, Open";
    and they can not edit/read them, which is great. but i am facing a caching problem , because if user is inside the "Intranet visitor" he will be able to see Category field as mentioned in my code, but if i remove him from the "Intranet Visitor"
    he still can see the field,, although in the web part i specify not to display the Category column if the user is not inside the "Intranet visitor " group... here is my current code:-
    protected override void OnInit(EventArgs e)
    base.OnInit(e);
    InitializeControl();
    SPSecurity.RunWithElevatedPrivileges(delegate()
    using (SPSite site = new SPSite(SPContext.Current.Site.Url))
    using (SPWeb web = site.OpenWeb())
    web.AllowUnsafeUpdates = true;
    SPGroup group = web.Groups["Intranet Visitor"];
    bool isUser = web.IsCurrentUserMemberOfGroup(group.ID);
    if (!isUser)
    SPList myList = web.Lists.TryGetList("Risk & Issue Management");
    SPField titleField = myList.Fields.GetField("Category");
    titleField.Hidden = true;
    titleField.ShowInEditForm = false;
    titleField.ShowInNewForm = false;
    titleField.ShowInDisplayForm = false;
    titleField.Update();
    myList.Update();
    // web.AllowUnsafeUpdates = false;
    else
    SPList myList = web.Lists.TryGetList("Risk & Issue Management");
    SPField titleField = myList.Fields.GetField("Category");
    titleField.Hidden = false;
    titleField.ShowInEditForm = true;
    titleField.ShowInNewForm = true;
    titleField.ShowInDisplayForm = true;
    titleField.Update();
    myList.Update();
    web.AllowUnsafeUpdates = false;
    so can you advice please ? is this a caching problem, or once the user add at-least single item he will be able to see all columns ?

  • How to map expdp parallele process to output file

    How to map expdp parallel process to its output file while running...
    say i use expdp dumpfile=test_%U.dmp parallel=5 ..
    Each parallele process writing to its related output file.. i want to know the mapping in run time...

    I'm not sure if this information is reported in the status command but it's worth a shot. You can get to the status command 2 ways:
    if you are running a datapump job from a terminal window, then while it is running, type ctl-c and you will get the datapump prompt. Either IMPORT> or EXPORT>
    IMPORT> status
    If you type status, it will tell you a bunch of information about the job and then each process. It may have dumpfile information in there.
    If you run it interactively, then you need to attach to the job. To do this, you need to know the job name. If you don't know, you can look at sys.dba_datapump_jobs if prived, or sys.user_datapump_jobs if not prived. You will see a job name and a schema name. Once you have that, you can:
    expdp user/password attach=schema.job_name
    This will bring you to the EXPORT>/iMPORT> prompt. Type status there.
    Like I said, I'm not sure if file name information is specified, but it might be. If it is not there, then I don't know of any other way to get it.
    Dean

Maybe you are looking for

  • Sequential file naming for repeating action on same file

    I need to export all the hands of a clock to animate the individual hour and minute hands in the software I'm using. So, I'm trying to set up an automated process to save 60 clock hands so that I don't have to do this manually for all future clock de

  • TypeError: beforeNode has no properties-MX query builder

    Hi everybody Hope i can find some support in here.sad that this bundle is discontinued. Anyways,am working on Classifieds Ad tutorial found here: http://www.interaktonline.com/Documentation/MXKollection/077000_tutorialclassifiedadssite. htm Am stuck

  • Result to Text : Spaces

    hi~ I am writing a query to result to text: select db_name(dbid),dbid from sys.sysprocesses but there is alot of spaces after the column of db_name... how to trim them ??

  • Cannot Collect BI Content DSO object for Installation to Version A

    Hi Gurus, I tried to simulate the installation of BI Content DSO object 0BBP_SCA. When I grouped by "In Data Flow Before" or "In Data Flow Before and Afterwards", I cannot transfer the selected object to the collected column. An error message says th

  • SSIS Migration between Servers

    Good Day, I am in the process of moving my ssis package development to a new machine.  I have zipped up all of my development/solution files and put them on the new machine...put them in place...but when I try to test them within Visual Studio I am g