Archiving-Need clarification

Hi, I would like to know what exactly is the difference between archiving a material ledger document and a material ledger index. there are two diff objs available for these, what is the diff and if i am archiving material ledger index, will the ledger doc be impacted?

hi,
check these links...
http://www.sap-img.com/basis/delete-remove-of-idocs-from-r3.htm
http://help.sap.com/saphelp_nw04/helpdata/en/0b/5d193ad0337142e10000000a11402f/content.htm
http://help.sap.com/saphelp_nw04/helpdata/en/b4/fb3340284e7a56e10000000a1550b0/content.htm

Similar Messages

  • Need Clarification On Internal tables in Start Routine

    Hi,
    I have intenal table some IT_A which deletes the requests which are populated in to it.
    Now I need to populate requests into IT_A from  another internal table like IT_B which are of different structure.
    I have three fields in common in both internal tables like RNR,Timestamp n SID with different field names
    I Need Clarification that if i move the contents of the three fields to IT_A from IT_B by spcifying these three fields.Will the internal table IT_A deletes the Requests?
    Thanks,
    Sriram.

    Hi Sriram,
    As mentioned IT_A deleted the request loaded into it.
    if you move the contents of the three fields to IT_A from IT_B by spcifying the three fields after the deletion step for internal table IT_A takes place then it wont get deleted.
    But it would be deleted if the the contents are moved before the deletion step takes place
    regards,
    mahesh

  • Need clarification on issue with tablespaces

    Hi All,
    I needed clarification on this issue with tablespaces.I am following a guideline which says "Presently,The general rule of thumb is to concentrate on tablespaces that are 90% or greater used and have less than 2 GB of free space"
    My Scenario:-Now i have tablesace which is 97.23 % and the free space is 10 gb.Please advise me that whats the best next step I can take in this respect?
    I am very new to this and just following the guideline,so can you all please explain me some details about this.
    Thanks

    Well, the guideline says "*and* have less than 2 GB of free space"
    Since that tablespace as 10GB of free space, the guideline says that you do not need to concentrate on it. Eh ?
    I would be careful with guidelines that are hard-coded in this manner. You need to know the context in which the guideline was set.
    Assuming that your database/application doesn't grow suddenly by 1GB or larger (or create temporary segments / objects of 1GB or larger), the "2 GB of free space" might make sense. Then, again, it depends on the extent allocation type. What size are extents getting allocated.
    You should check with your organisation's senior DBAs on what happens if you don't concentrate on a tablespace because it is outside the guidelines !

  • Need clarification on Bigfile Tablespaces

    In the following Oracle Documentation Lirbary PDF,
        Oracle® Database
        Concepts
        10g Release 2 (10.2)
        B14220-02
        October 2005
    section
        Overview of Tablespaces
        Bigfile Tablespaces (page: 3-5)
    it says,
        Benefits of Bigfile Tablespaces
        * Bigfile tablespaces can significantly increase the storage capacity of an Oracle database. Smallfile tablespaces can contain up to 1024 files, but bigfile tablespaces contain only one file that can be 1024 times larger than a smallfile tablespace. The total tablespace capacity is the same for smallfile tablespaces and bigfile tablespaces. However, because there is a limit of 64K database for each database, a database can contain 1024 times more bigfile tablespaces than smallfile tablespaces, so bigfile tablespaces increase the total database capacity by 3 orders of magnitude. In other words, 8 exabytes is the maximum size of the Oracle database when bigfile tablespaces are used with the maximum block size (32k).
    I need clarification on how to arrive at 8 exabytes ?
    1024 x 32k x 64,000 ??
    According to the exerpt above, there's no mention of maximum number of Operating System blocks per extent. Unless this was assumed knowledge ... how do I get 8 exabytes ?
    And if "a database can contain 1024 times more bigfile tablespaces than smallfile tablespaces", then what's the upper limit on smallfile tablespaces ? -- was this sentence referring to the number of datafiles per smallfile tablespace ? ...
    O_o
    Thanks !
    Message was edited by:
    mvanle

    Hi,
    According to [url http://download-west.oracle.com/docs/cd/B19306_01/server.102/b14237/limits002.htm#i287915]Physical Database Limits page, a bigfile tablespace contains only one datafile or tempfile, which can contain up to approximately 4 billion ( 232 ) blocks. The maximum size of the single datafile or tempfile is 128 terabytes (TB) for a tablespace with 32K blocks and 32TB for a tablespace with 8K blocks. In resume, A bigfile tablespace is a tablespace containing a single datafile that can be as large as 128 terabytes (TB), depending on the block size of the tablespace. In conjunction with setting the initialization parameter DB_FILES to the maximum value of 65,635 the total size of the database can be more than 8 exabytes (EB)
    >>how do I get 8 exabytes ?
    You can calculate the maximum amount of space (M) in a single Oracle database as the maximum number of datafiles (D) multiplied by the maximum number of blocks per datafile (F) multiplied by the tablespace block size (B):
    M = D * F * B. Therefore, the maximum database size, given the maximum block size and the maximum number of datafiles, is:
    65,535 datafiles * 4,294,967,296 blocks per datafile * 32,768 block size = 9,223,231,299,366,420,480 = 8EB.Cheers
    Legatti

  • BPM, Workflow and Netweaver - Need Clarification

    Hi Guru,
    I am new to workflow, BPM and Netweaver.  I have several questions about those concepts.
    1. What is/are the different between workflow and webflow.  Which scenario should take workflow into consideration? Which scenario webflow can be applied to?
    2. What is/are the different between workflow and BPM.  If I am going to implement workflow in a company, do I need to implement business workflow as well as BPM?
    3. I need clarification on Netweaver platform and concept.
    4. What is/are the different between workflow in R/3 and workflow under Netweaver?
    5. If I am going to implement workflow integrated with R/3 and Outlook email, do I need to buy new wrokflow for Netweaver and Netweaver platform or alternatively, I can use business workflow module under R/3 system?
    Sorry for many questions asked. I am new to those products.  I am now working on software selection for workflow technology.  My company is going to implement new workflow to client.  Thank you very much.
    Cheers,

    Please ask only one (or closely related) questions per thread. This makes it a lot easier to get a good structure in the database of previously answered questions. While we are on the subject of previously answered questions, I think you should have a look at them....
    My suggestion is therefore:
    1. Close this thread
    2. Read the Frequently Answered Questions. Before you ask (here are many workflow answers)
    3. Search the forum.
    4. Create new threads if you have questions afterwards.

  • Log file switch (archiving needed)

    Hi,
    My database is on windows 2003, 10.2.0.4, recently I have been getting the following wait events on a regular basis, LOG FILE SWITCH (ARCHIVING NEEDED)...My redo files are 50MB with 5 groups...I have changed the log_archive_max_process from 2 to 3, but still the problem persists...can anyone help me regarding the same?As to what other changes can be done?and this happens only during the time of BOD and EOD.
    Thanks,
    AJ

    Hi Jonathan,
    Thanks for the reply, I am not worried abt the archiving thing for the time being,as my database was working absolutely fine before the new query that was added in the Begin of Day process. So I am trying to tweak the query.
    Folllowing is the procedure which gets called during the Begin of Day:
    CREATE OR REPLACE procedure Lms_Pr_Bod_Update_Tmp (p_import_type in number)
    as
    v_update_script varchar2(2000);
    v_upper_start_tag varchar2(10); -- san_22-apr-2009
    v_upper_end_tag varchar2(10); -- san_22-apr-2009
    begin
    declare
    cursor cur_update_tmp is
    select -- this query is for all of lov type of data
    a.destination_column_name, a.column_to_update , b.mapped_lov_syscode lov_syscode,
    c.destination_table_name,'LOV_DATA_SYSCODE' select_attribute,'LMS_LOV_DATA_MAP' select_table,
    'SOURCE_UNIQUE_ID' where_condition,'LOV_SYSCODE' lov_condition,'LOV' att_type
    from lms_import_column_info a
    inner join lms_attribute_master b
    on a.attribute_syscode = b.attribute_syscode
    and (a.column_to_update is not null or trim(a.column_to_update) = '')
    inner join lms_import_type_master c
    on a.import_type_syscode = c.import_type_syscode
    where a.import_type_syscode = p_import_type
    and b.mapped_lov_syscode is not null
    union all
    select -- this query is for all of dummy attribute
    a.destination_column_name, a.column_to_update , 1 lov_syscode,
    d.destination_table_name,c.attribute_name_internal select_attribute,
    case
    when b.applicable_for = 'INS' then 'LMS_ENTITY_INSTRUMENT'
    when b.applicable_for = 'ACC' then 'LMS_ENTITY_ACCOUNT'
    end case,
    b.attribute_name_internal where_condition,'1' lov_condition,'DUMMY' att_type
    from lms_import_column_info a
    inner join lms_attribute_master b
    on a.attribute_syscode = b.attribute_syscode
    and (a.column_to_update is not null and trim(a.column_to_update) <> ' ')
    and b.attribute_type = 'DUM' and dummy_column_type = 'FILT'
    inner join lms_attribute_master c
    on b.dummy_mapped_attribute_pk = c.attribute_syscode
    inner join lms_import_type_master d
    on a.import_type_syscode = d.import_type_syscode
    where a.import_type_syscode = p_import_type;
    begin
    v_upper_start_tag := 'UPPER('; -- san_22-apr-2009
    v_upper_end_tag := ')';          -- san_22-apr-2009
    for cr_update_tmp in cur_update_tmp loop
    if p_import_type = 4 and cr_update_tmp.column_to_update IN ('INSTRUMENT_SYSCODE','ACCOUNT_SYSCODE')
    and Upper(cr_update_tmp.destination_table_name) = 'LMS_ENTITY_TRANSACTION_TMP' then
    v_upper_start_tag := ''; -- san_22-apr-2009
    v_upper_end_tag := ''; -- san_22-apr-2009
    end if;
    v_update_script := 'UPDATE ' || cr_update_tmp.destination_table_name || ' A SET ' || ' A.' || cr_update_tmp.column_to_update || ' =
    (select ' || CR_UPDATE_TMP.SELECT_ATTRIBUTE || ' from ' || CR_UPDATE_TMP.SELECT_TABLE || '
    where ' || V_UPPER_START_TAG || CR_UPDATE_TMP.WHERE_CONDITION || V_UPPER_END_TAG || '=' || V_UPPER_START_TAG || CR_UPDATE_TMP.DESTINATION_COLUMN_NAME || V_UPPER_END_TAG ||
    ' AND ' || cr_update_tmp.lov_condition || ' = ' || cr_update_tmp.lov_syscode || ')';
    v_upper_start_tag := 'UPPER('; v_upper_end_tag := ')';
    execute immediate (v_update_script);
    end loop;
    end;
    end;
    Following is the flow of the query:
    1. A temporary table is created where in the updates can be made.
    2. Rows are inserted into ths table from the source table
    3. Updates are performed on this table
    4. Updates are then copied to the source table
    5. This procedure is alled twice, so before getting called for the second time, table is truncated.
    Thanks,
    AJ

  • Need clarification for these names, R/3, WAS, NetWeaver

    Hi All
    I posted the question on WAS preview installation, then I realized I should have put it here. Here is the link to that post:
    Need clarification for these names, R/3, WAS, NetWeaver
    Thanks again for any input
    Xueqing
    Added the link to the first post
    Message was edited by: xueqing Han

    Hi Xueqing,
    First of all, sorry for any confusion that we have caused you. I hope I can give you an answer that will clear up the confusion. Sorry, but it is a long explanation of the development of two application solution but history tends to be very long winded!
    It is <b>not</b> true SAP R/3 Enterprise equals SAP Web AS, and I'll hopefully explain why:
    In the beginning (at least in the client server world) SAP only ship SAP R/3. The technology layer under SAP R/3 was called SAP Basis. There was only SAP Basis under SAP R/3.
    SAP then started to deliver other software solutions that were not included in or built on the SAP R/3 software. These included SAP Business Information Warehouse, SAP APO, SAP SEM, SAP CRM, and the list goes on.
    These solutions needed to run on technology layer (like SAP R/3 did). SAP Basis was the obvious choice for this because of the common technology layer providing DB/OS abstraction and a coding environment.
    <b>~2001:</b>
    Later there came the need to have SAP Basis support the web and its web standards and other programming languages. This radically changed what SAP Basis was and we decided to rename the new technology layer to SAP Web Application Server (SAP Web AS). So the SAP Basis name was retired and SAP Basis 4.6D was the last release called SAP Basis.
    The new release and the "technology change" means that all the applications mentioned above now ran on SAP Web AS. The first release of SAP Web AS was 6.10.
    <b>~2002:</b>
    SAP Web AS was first used in SAP R/3 Enterprise release .
    The important fact is that SAP R/3 Enterprise runs on SAP Web AS. SAP R/3 Enterprise = SAP R/3 business applications + new business functions + SAP Web AS.
    <b>~2003:</b>
    I hope the above explanation is clear, because technology takes another major change. It was realized that SAP now had a collection of business solutions/applications (SAP R/3, mySAP CRM, mySAP SCM, etc) and a collection of technology solutions (SAP Web AS, SAP BW, SAP EP, etc). The technology requirements for the business solutions did not end at SAP Web AS, they needed portal, data warehouse, knowledge management capabilities, etc to develop business solutions on.
    To address this SAP made the decision to combine all the technology solutions and tools into one single platform. This made complete sense for developers (SAP, Partners and customers). This single platform is called SAP NetWeaver. It includes all the old individual components of SAP Web AS, SAP BW, SAP KM, SAP EP, etc).
    <b>~2004:</b>
    I think you can guess what the next step is. Yes, the new release of SAP R/3. Since the release of SAP R/3 Enterprise and the release of SAP NetWeaver, SAP R/3's name changes. It is now called mySAP ERP as it includes a lot of applications that were previously sold separately (like SAP SEM, MSS, ESS, etc).
    So now mySAP ERP runs on SAP NetWeaver (yes everything that was in SAP Basis and then in SAP Web AS is still there but SAP NetWeaver has so much more now).
    Also with the release of SAP NetWeaver, SAP starts to stop using the old technology component names - you will not hear us talk about SAP Web AS, SAP BW, etc anymore, just SAP NetWeaver releases.
    In addition all the other SAP business applications also run on SAP NetWeaver, so the latest version of mySAP CRM, mySAP SRM, and mySAP SCM all run on SAP NetWeaver.
    So to simplify the explanation, it would be :
    <b>Evolution of SAP R/3:</b>
    SAP R/3 (up to release 4.6c) -> SAP R/3 Enterprise (releases 1.00 through 2.00 -> mySAP ERP (2004 onward)
    <b>Evolution of Basis:</b>
    SAP Basis (up to release 4.6d) -> SAP Web AS (up to release 6.40 which was included in SAP NetWeaver '04) -> SAP NetWeaver (2004 onward)
    SAP NetWeaver and mySAP ERP have <u>their own release cycles</u>. mySAP ERP always has an underlying technology release that it is built on (this is a SAP NetWeaver release)
    I hope this helps,
    Mike.
    <b></b>

  • Need clarification regarding the test cable-diagnostics tdr command

    Hello,
    I've read about the test cable-diagnostics tdr command but I need clarification on the examples listed below to make sure that I am providing the right answer to my co-workers.
    Example 1:
    CXXX7SW17#show cable-diagnostic tdr int g0/20
    TDR test last run on: July 16 10:23:00
    Interface Speed Local pair Pair length        Remote pair  Pair status
    Gi0/20    auto  Pair A     N/A                        N/A                Normal
                           Pair B     72   +/- 10 meters  N/A                Open
                           Pair C     75   +/- 10 meters  N/A                Short/Crosstalk 
                           Pair D     74   +/- 10 meters  N/A                Short/Crosstalk
    Does this example mean that there's a cable length issue in the line which is causing the device that it's connected to not to work properly?
    Example 2:
    CXXX2SW140#show cable-diagnostics tdr int g0/21
    TDR test last run on: July 16 09:16:22
    Interface Speed Local pair Pair length        Remote pair Pair status
    Gi0/21    100M  Pair A     N/A                      Pair A           Normal
                              Pair B     N/A                      Pair B           Normal
                              Pair C     N/A                     Pair C           Normal
                              Pair D     N/A                     Pair D           Normal
    Does this example state that the cable line is okay for use?
    Example 3:
    CXXX1SW19#show cable-diagnostics tdr int g0/22
    TDR test last run on: July 16 06:36:53
    Interface Speed Local pair Pair length            Remote pair Pair status
    Gi0/22    auto    Pair A       1    +/- 10 meters  N/A               Open
                              Pair B     39   +/- 10 meters   N/A               Open
                              Pair C     72   +/- 10 meters   N/A               Open
                              Pair D     1    +/- 10 meters    N/A               Open
    Does this example mean that there isn't a device connected on the other end? No pin-contact?
    Thank you very much for any help you could provide.
    S

    I found this article here at supportforums that seemed like the best explanation I've read so far for TDR info.
    Hope that helps.

  • BW Archiving- Need help as soon as possible

    Hi,
    Can Archiving on ODS be implemented using an ABAP program? If the data is loaded on to flat files and if it is required to reload it back to ODS is ABAP a better choice. If any one has worked on this issue earlier please provide me the information as soon as possible. Any input on this issue will be of great help.
    Thanks,
    SP

    Hi sudheer,
    I have gone through few of your requests to SDN network . and found out thta you have worked on  "sap bw archiving ". So, I need a small help from you . need to clear out some of my doubts with sap bw archiving issues . So could you please send me your e-mail id . so that i can be in contact with you . As i am new to SDN .
    Thanking you in advance . I would appreciate your help .
    or Anyone who have experience with "SAP  BW archiving " could you please send me your mail id . Thanks for your help .
    Cheers,
    shailu.

  • ZCM 11 SP3 and Windows 8.1, need clarification...

    I dont know if I can't read, but I need som clarification...
    When SP3 arrived, there were problems with Win 8.1.
    I have customers that have machines, that were upgraded to 8.1 and it have worked with the already installed 11.2.3a -agent, even if its not supported. They deployed ZCM 11.3-agent to this machines that were upgraded to 8.1, and they all crached, never started again. I made tests in my test-environment, same result.
    Last week I started to check and found the update "ZENworks 11.3.0a Windows 8.1 Update" (https://www.novell.com/support/kb/doc.php?id=7014805 / http://download.novell.com/Download?...d=0yMdXrTonF8~). Downloaded that yesterday, and was to test it today. Looked for the instructions again today, and now its not availble anymore, Obsolete. I have not deplyed it yet, so that is no problem.
    There are since yesterday a new update, "ZCM 11.3.0 WIN8.1 Patch 866736" (http://download.novell.com/Download?...d=OvBLs9qZhrU~).
    But in the instructions on this, it talks about it talks about how to update machines already updated. Here the text mentions "11.3.0_WIN8.1". What is that? Is it the now Obsolete "ZENworks 11.3.0a Windows 8.1 Update"? Or machined patched with the standard "Update for ZENworks (11 SP3)", that was created during SP3 install? If reffering to the last, it cant be done because in both my customers production environment and my testenvironment the Win 8.1 machined crached and never came up.
    Next two methods in the description describes how to update "Windows 8.1 Update for ZENworks (11 SP3)", if already imported (zman supf "Windows 8.1 Update for ZENworks (11 SP3)" ZCM_11.3.0_WIN8.1_20140404_866736.zip).
    But for those who have not imported that, what is the way to go? You cant download "Windows 8.1 Update for ZENworks (11 SP3)" anymore, and the patch seems to be for that one?
    What is the way to get Win 8.1 to works with ZCM 11.3?
    Can anyone clarify this?
    I dont know if this belongs in agent forum or server-forum, but I start here.
    /Stefan

    CRAIGDWILSON wrote:
    > New Versions of those patches are being rolled out.
    >
    > Normally it happens at the same time, but looks like timing was
    > slightly off.
    >
    >
    >
    > On 4/30/2014 11:31 AM, Niels Poulsen wrote:
    > > stesjo wrote:
    > >
    > > >
    > > > So, at this point, it means that there is no way to get ZCM 11.3
    > > > to work with Windows 8.1?
    > > >
    > > > Old patches removed, and just patches for the removed patches
    > > > available?
    > > >
    > > > One can assume, that there is a reasom why the 8.1 Update is
    > > > removed, and made obsolete?
    > > > /Stefan
    > >
    > > ... One would think so, yes. Not sure what's the reason...
    > >
    Cool :-)
    Niels
    A true red devil...
    If you find this post helpful, please show your appreciation by
    clicking on the star below
    A member must be logged in before s/he can assign reputation points.

  • B2B Inbound Error : Need Clarification

    Hi All,
    I would like to illustrate a particular use case for B2B Inbound Errors and if there is any workaround available to get over the problem. We are using B2B version 10.1.2
    Use Case Details</strong1. Consider the scenario wherein we have a Inbound EDI File which has 100 transactions in it. B2B reads the file and automatically debatches into individual transactions and processes them separately.
    2. Out of the 100 transactions in the input file, if say 5 transactions are error transactions, we have observed that all the 100 transactions error out in B2B.
    3. The 95 correct transactions fail with the error "General Validation Error" and the 5 error transactions have the exact error details in the B2B Message Reports.
    4. Ideally, B2B should error out only the 5 error transactions and process rest of the 95 transactions without allowing them to fail with "General Validation Error".
    I am sure that most of us must have faced this error. Could someone please let me know the folllowing:
    1. Is there any setting in B2B by which we can enable it to error out only the transactions which have valid errors in a single file
    2. Is there any other workaround that we can take to avoid this issue.
    The reason we need to have a solution for this is because in our Production B2B environment we receive inbound files from Trading Partners which have hundreds of transactions in each file.
    If, even a single transaction has error, then the whole file errors out and it is quite cumbersome to browse through the B2B Message Reports to get the exact error transaction because all the correct transactions would failed with error "General Validation Error".
    Please let me know the inputs.
    Thanks,
    Dibya

    Hi Ramesh,
    Thanks for the clarification and information. We have already set the parameter OneErrorAllError = true and my understanding was that if we set this parameter to TRUE, it means that if one transaction in the OUTBOUND batch errors out then the whole batch will error out.
    I was not aware that it also holds true for INBOUND. Is there any other way that we can set this parameter based on direction (INBOUND or OUTBOUND).
    We would like to have this parameter set for OUTBOUND and disabled for INBOUND Transactions.
    Please let me know. Thanks Again.
    Regards,
    Dibya

  • Need clarification regarding select query

    Hi,
    I need a little clarification regrding a Select senario
    I want to select data from table which have been minupulated between a certian date like between 1-DEC-10 to 31-DEC-10 and note that table does not have any time/date column. I've applied the following query to do this.
    select * from TABLE_NAME where sysdate between to_date('01-DEC-10') AND to_date('31-DEC-10');
    Would it work fine because I've tried it against a table and it returned me nothing however DML occur between time period.
    Regards,
    Abbasi

    Abbasi wrote:
    Hi,
    I need a little clarification regrding a Select senario
    I want to select data from table which have been minupulated between a certian date like between 1-DEC-10 to 31-DEC-10 and note that table does not have any time/date column. I've applied the following query to do this.
    select * from TABLE_NAME where sysdate between to_date('01-DEC-10') AND to_date('31-DEC-10');
    Would it work fine because I've tried it against a table and it returned me nothing however DML occur between time period.
    Regards,
    AbbasiAFAIK without log mining and auditing this is not possible.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/logminer.htm

  • Need clarification for cold backup and recovery

    Hello Everyone ..
    I have much confusion on cold backup and recovery topic.
    Already i posted a scanario regarding this and i want to know some clarity points to understand.
    1. I had cold backup two days before
    2. I am creating an object *(EMP)* and inserting some records then issue log switch continiously.
    3. Manually i removed all physical files (datafiles , control files , log files , redolog files ...)
    4. then i issued STARTUP FORCE MOUNT;
    5. here i am getting error identifying control file.
    6. cold backup two days before all files restored
    6. then i issue recover database using backup controlfile until cancel;
    7. i am trying to issue select * from emp; i am getting error
    select * from emp
    *ERROR at line 1:
    ORA-00942: table or view does not existMy QUESTION IS
    After cold backup finished , i created emp table ..
    when restoring old back " There is no emp table"
    *- then , can i recover emp table ? - If so , please explain about this ..*
    Note : REF - LINK https://forums.oracle.com/forums/message.jspa?messageID=11056341#11056341
    - Above link says what i did ? . In this thread i am asking concept & logic ..
    please do NOT consider as "*DUPLICATE*"

    When you recover the database using backup controlfile until cancel, you have to apply archived logs. Did you apply all of them?
    If the create command was in the online redo, not yet archived, and you lost the online redo, then you won't have it. This is why you want to have redo multiplexed, it is a critical piece.
    In some recovery scenarios, if you haven't lost the online redo, you need to specify those files as if they were archives.
    It is normally easier to just be sure you have the latest controlfiles and let Oracle figure it out automatically (which is why you want to use online RMAN backups rather than cold backups). But yes, it is important to understand what is going on for different scenarios. I used to have manual standbys, and one of the disaster instruction sets explained to try to get online redo over to the standby to lose as few transactions as possible. That could still be useful for some snapshot hardware type scenarios, assuming the business is too cheap to do proper failovers.

  • Unchecking Print Resolution - need clarification, need Jeff Schewe

    Seeking clarification on using the print module in LR. Have read several threads in which Jeff Schewe (Jeff, are you there?)suggests that you use your file as a master file in LR and that if you uncheck the print resolution box, you can print multiple sizes of an image by merely changing the print size in LR, provided you uncheck the print resolution box.
    Want to make sure that I've got this workflow right. You do your edits in LR and process your raw file. Then bring the file into CS3 for whatever other tweaks are needed. Add a layer and do your output sharpening on this layer, but don't go up to the Image, Image Size command. Do your sharpening without resizing or resampling.
    My RAW files come into CS3 with a resolution of 240 pixels/inch. This is the native resolution?? After sharpening in CS3 (without resizing or resampling), bring the file back into LR. In the print module, uncheck print resolution (LR shows the resolution as being 240 ppi)and enter the height and width of the print size (not the paper size, but the print size) in the cell size. Go into page setup (Windows) which will take you into your print driver. On my Epson r2400 this is where I choose sheet vs roll, paper type, print quality, click ICM, then click off. Ok my way out of the Epson driver back to LR. On the white page layout in the center of the print module, LR will give me a new ppi reading. In my workflow, I typically print at 8x12 inches, so now LR is telling me that it is at 275 ppi. Pick profile and rendering intent in color management. As long as the read out for my print size in the white center section of lightroom reads between 180-480ppi, then at this point I am good to go and click print.
    Have I got it? Am I missing something? Please, enlighten me.
    Will LR provide a superior print to CS3 or is it six of one, half dozen of the other?
    Thanks,
    Cynthia

    First off, it's poor form to address posts in a public forum to single individuals...this is a forum and ALL members should be encouraged to post...this is akin to asking for an answer to a question be sent directly by email.
    The resolution of raw files is xxxxpixels by xxxxpixels. That's the only way to think of resolution in Lightroom. As a result, the image may be said to be 240, 300, 360 or 480 PPI depending on the size the image will print. A small print, held closer will need more PPI than a large print hung on a wall. So the resolution of a raw file will vary depending on the size of the final cropped & printed image. As long as the PPI of the image size is between 180-480 PPI you really have no reason for up or down sampling...
    When I work on an image in Photoshop from Lightroom, I'll have an image print size in mind. I use softproofing for final tweaks to the image-often applied locally and set the PPI (Image Size without resample) to the size I want then use PhotoKit Sharpener for the resulting pixel density. Then hit save and print from Lightroom.
    The -Edit file in Lightroom has the pixel density I speced in Photoshop and the softproofing tweaks as well. I then print form Lightroom-not so much because Lightroom is capable of producing BETTER prints...but because I find it quicker and easier to print with LR's print environment.
    If I want to make a different sized print (or a print on different papaer requiring a different profile) I'll open the tiff image into Photoshop from Lightroom using the Edit Original...my original tiff file with all layers pops open, I'll make whatever changes I need and hit save again...so that one master tiff file becomes my print mater file with all the layers I may need to either use, or turn off if I don't need them. The same image, if printed 4x6 will have different pixel density than a 13x19" print, thus I'll have tow separat sets of output shaprening. But I don't need to resample the image, I resize without resampling checked to change the ratio and density.
    And yes, all of this with Lightroom's print module set to no resampling and no sharpening...minor changes in the image cell size in LR don't require going back into Photoshop for resizing...

  • Need Clarification on sga_target and sga_max_size

    HI,
    I need some clarification in SGA_TARGET and SGA_MAX_SIZE.
    I have the parameter like below.
    SGA_MAX_SIZE=10G
    SGA_TARGET=9G
    And I spread the 9G to all components like (DB_CACHE,SHARE_POOL etc.,).
    My doubt, Incase db need the memory more than 9GB, Whether it automatically take the extra 1G from sga_max_size
    or we have to change the sga_target to 10G.

    Unless and untill, we set the sga_taget=10G, The extra 1G(from sga_max_size) is not used.
    Am i correct?No - its wrong. Any change in the value of SGA_TARGET affects only the sizes of the auto-tuned components. If you increases its value then increased memory will be distributed in only among the components controlled. So, yes 1GB will be used, because have sga_max_size=10G. If you decrease the value then reduced memory is taken back by the auto-tuning policy from one or more of the auto-tuned components.
    If SGA_MAX_SIZE is greater than SGA_TARGET, you can increase SGA_TARGET without restarting the instance. Otherwise, you'd need to shutdown and restart the instance if you wanted to increase SGA_TARGET.
    Regards
    Girish Sharma

Maybe you are looking for