Configuration parameters in OWB mapping

Hi ,
I have a requrement where in I need to find out some details of an OWB mapping through a SQL script.
For a mapping when we "configure" the Table operators like entering the Schema name. Which table in OWBRT_SYS holds these details basically I need to query the OWBRT_SYS.TABLE_XXXXX to find out for a given mapping weather the Schema name has been entered or not.
Please suggest.
Thanks,
Jay

Re your first question, you are probably hitting a bug (3035080) that makes OWB generate false warnings. Please ignore the warnings and try to deploy and run the process flow.
As for your second question - I cannot reproduce this on 9.2.0.2.8. The configuration properties are updated corectly in the SQLLOADER map and I get the same values in the process flow. By the way, why are you not using external tables instead of the SQLLoader maps?
Regards:
Igor

Similar Messages

  • Array of Parameters in OWB mapping

    How do i pass an array of parameters to an OWB mapping.

    Try this and similarly you can get all operator.
    OMBRETRIEVE MAPPING 'MAP_NAME' GET TABLE OPERATORS
    OMBRETRIEVE MAPPING 'MAP_NAME' GET VIEW OPERATORS
    OMBRETRIEVE MAPPING 'MAP_NAME' GET TRANSFORMATION OPERATORS
    OMBRETRIEVE MAPPING 'MAP_NAME' GET TABLE_FUNCTION OPERATORS
    Cheers
    Nawneet
    Edited by: Nawneet_Aswal on Aug 30, 2010 8:43 AM

  • Reset non-default parameters in configure dialog of the mapping

    OWB 9.0.4
    Why after Reconcile inbound on table or cube in any mapping is reset non-default parameters (ex. hints, database link) in configure dialog of the mapping?

    Dmitry,
    The behavior is intended... And the reason for that is: we could not change it, but then sometimes you do want to change it, sometimes you don't. We could provide flexibility, but then you may have to answer or confirm the selection... which you don't want to do too often. In this case we change, and I am hoping it does not create too many issues for you. Let us know if it does, and we can evaluate the enhancement request.
    Thanks,
    Mark.

  • VLD-1124: Schema and/or Database link physical configuration parameters

    Hi!
    The warning message:
    VLD-1124: Schema and/or Database link physical configuration parameters should not be set for Table1.
    Is it possible to eliminate this warning message ? Anyway , Thanks for your response!
    Bye!

    Hi guys,
    I am facing a strange problem. The code generated for my mapping has NOT used the dblinks. My OWB version is 10.1.0.4
    This is my problem in brief.
    I have installed OWB newly and started to do a sample task. I created a very simple one to one table population mapping from source to the target schema. Both are in the same database.
    When i generated the code for the mapping, it gave a warning 'VLD-2771: System privileges may not allow extraction from source EMP'.
    When i checked the code which was generated, i couldnot see any dblinks associated with my source table( which seems strange)
    CURSOR "INGRP_c" IS
    SELECT
    "EMP_SRC_TRG_CONN"."EMPNO" "EMPNO",
    "EMP_SRC_TRG_CONN"."ENAME" "ENAME",
    "EMP_SRC_TRG_CONN"."JOB" "JOB",
    "EMP_SRC_TRG_CONN"."MGR" "MGR",
    "EMP_SRC_TRG_CONN"."HIREDATE" "HIREDATE",
    "EMP_SRC_TRG_CONN"."SAL" "SAL",
    "EMP_SRC_TRG_CONN"."COMM" "COMM",
    "DEPT_SRC_TRG_CONN"."DEPTNO" "DEPTNO_1",
    "DEPT_SRC_TRG_CONN"."DNAME" "DNAME",
    "EMP_SRC_TRG_CONN"."DEPTNO" "DEPTNO",
    "DEPT_SRC_TRG_CONN"."LOC" "LOC"
    FROM "SCOTT"."EMP" "EMP_SRC_TRG_CONN" ;
    In brief this is the process i have done.
    Source schema : SCOTT & Target schema : TRG_SCHEMA
    1) I have created source (for SCOTT) and target(for TRG_SCHEMA) modules.
    2) I have also created DBLinks, Locations and Connector from source to the target locations.
    3) I registered both the source location and target locations.
    4) Validated, Generated and Deployed the Connector from source to the target.
    I was unable to trace the error. Did i miss anything in the configuration? or during the installation of OWB.
    Though it is a very old post, i hope someone can help me out here.
    Thanks in Advance,
    Sri

  • Allowing parallel processing of cube partitions using OWB mapping

    Hi All,
    Iam using an OWB mapping to load a MOLAP cube partitioned on TIME dimension. I configured the OWB mapping by checking the 'Allow parallel processing' option with the no.of parallel jobs to be 2. I then deployed the mapping.The data loaded using the mapping is spread across multiple partitions.
    The server has 4 CPU's and 6 GB RAM.
    But, when i kick off the mapping, i can see only one partition being processed at a time in the XML_LOAD_LOG.
    If i process the same cube in AWM, using parallel processing, i can see that multiple partitions are processed.
    Could you pls suggest if i missed any setting on OWB side.
    Thanks
    Chakri

    Hi,
    I have assigned the OLAP_DBA to the user under which the OWB map is running and the job started off.
    But, it failed soon with the below error:
    ***Error Occured in __XML_MAIN_LOADER: Failed to Build(Refresh) XPRO_OLAP_NON_AGG.OLAP_NON_AGG Analytic Workspace. In __XML_VAL_MEASMAPS: In __XML_VAL_MEASMAPS_VAR: Error Validating Measure Mappings. In __XML_FND_PRT_TO_LOAD: In __XML_SET_LOAD_STATUS: In ___XML_LOAD_TEMPPRG:
    Here is the log :
    Load ID     Record ID     AW     Date     Actual Time     Message Time     Message
    3973     13     SYS.AWXML     12/1/2008 8:26     8:12:51     8:26:51     ***Error Occured in __XML_MAIN_LOADER: Failed to Build(Refresh) XPRO_OLAP_NON_AGG.OLAP_NON_AGG Analytic Workspace. In __XML_VAL_MEASMAPS: In __XML_VAL_MEASMAPS_VAR: Error Validating Measure Mappings. In __XML_FND_PRT_TO_LOAD: In __XML_SET_LOAD_STATUS: In ___XML_LOAD_TEMPPRG:
    3973     12     XPRO_OLAP_NON_AGG.OLAP_NON_AGG     12/1/2008 8:19     8:12:57     8:19:57     Attached AW XPRO_OLAP_NON_AGG.OLAP_NON_AGG in RW Mode.
    3973     11     SYS.AWXML     12/1/2008 8:19     8:12:56     8:19:56     Started Build(Refresh) of XPRO_OLAP_NON_AGG.OLAP_NON_AGG Analytic Workspace.
    3973     1     XPRO_OLAP_NON_AGG.OLAP_NON_AGG     12/1/2008 8:19     8:12:55     8:19:55     Job# AWXML$_3973 to Build(Refresh) Analytic Workspace XPRO_OLAP_NON_AGG.OLAP_NON_AGG Submitted to the Queue.
    Iam using AWM (10.2.0.3 A with OLAP Patch A) and OWB (10.2.0.3).
    Can anyone suggest why the job failed this time ?
    Regards
    Chakri

  • Logging OWB mapping execution in Shell script

    Hi,
    I am executing a OWB mapping from a shell script like this
    $OWB_SQLPLUS MY_WAREHOUSE plsql MY_MAPPING "," ","
    I want to log this mapping execution process into a file.
    Please let me know if this will work:
    $OWB_SQLPLUS MY_WAREHOUSE plsql MY_MAPPING "," "," >> LOGFIL.log
    I will just be using this log file to track all the execution and use it for logging purpose.
    If this wont work, please tell me the proper way to do this...
    Thanks.

    Avatar,
    ">>" is the Unix operator that will redirect output and append to a particular file, so what you have should work if you're executing it from the shell prompt. Although I don't know specifically what OWB_SQLPLUS and MY_WAREHOUSE are.
    In my company, we have the call to the owb script inside another script. For example, file x contains the following line:
    sqlplus repository_user/pwd@database @sqlplus_exec_template.sql repository_owner location task_type task_name custom_params system_params
    Then at the prompt, we enter:
    nohup x > x.log &
    And the mapping or workflow executes.
    Jakdwh,
    Are you redirecting your output to a file so you can see why it's returning a '3'? The log file will usually tell you where the error occurred. I don't know what your input parameters for your mapping is, but the script is pretty picky about the date format. Also, even if you don't have any input parameters, the "," still has to be sent into the script.
    Hope this helps,
    Heather

  • How to configure some of VMX file configurations / VM Configuration Parameters via vCloudDirector REST API

    Hi
    We need to configure some settings for VM using vCloudDirector REST API other wise which can be down in 2 following ways.
    1. On VCenter -> select VM  ( in Power Off Mode ) -> Edit Settings -> Options -> Advanced -> General -> Configuration Parameters
         Add row with values like below
               ethernet0.ctxPerDev = 1
    2. -OR- directly login to Esxi Shell .. Open VMX file ( of the VM  - which can be located in /etc/vmware/hostd/vmInventory.xml ) and  add entries manually.
          Follow this KB http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1714
          for editing VMX file.
    We would like to know how this can be achieved using vCloud Director REST. We are using 5.5 versions for both VCD and VCenter.
    Thanks for help.
    Regards,
    Murali G D

    This level of modification is not available via vCloud Director's API afaik.  A similar question came up recently about modifying the amount of Video Memory on a VM, which you have to create some sort of custom workflow directly to vCenter/ESXi about.

  • Problem in OWB map for source change capture

    The problem is as follows:-
    Our OWB map between two different databases, captures change in data at source and populates them in target. It is a one to one map. The map is failing to process some rows from source database though it gives no:of errors as 0 in the audit view thus we infer that the map is running successfully.
    In our case, we have a composite key ( Instance_id, Version_number) at source as well as target. One instance_id can has more than one version number. We have observed that the view is picking up all the version_number for a particular instance_id but the map is unable to populate the last version_number to the target table.
    But when this map was executed in another environment and it is successfully populating all the data in the target.
    The brief description of the map is as follows:-
    The map has a view which captures all the changes in data (insert and update) .From this view we filter the rows which were already processed in previous map run (with help of watermark). Then we sort the rows based on DML type and watermark . These rows are then mapped to a target table. The loading type of the map is INSERT/UPDATE.

    Hi Nagesh,
    When we are doing I18N for that Application. Go to Navigator Tab
    Here Select Particular Projects--à src-àPackages-àsap-àvijay
    To internationalize the Web Dynpro application, copy the automatically generated *.xlf files and save them under a new name in the same directory.
    The new name must meet the following convention:
    u2022 .xlf
    For example, if you are creating *.xlf files for German,
    Use the language key de.
    Here Click on OK
    After that we can edit and translate these new *.xlf files in the S2X Editor.
    ApplyTemFirstView.wdview_de.xlf-àClick on Resource Text Tab
    Select Particular Text and Change the language to German-àClick on EDIT Button
    Here Enter German Lang-àClick on OK
    Now Go to Web Dynpro Explorer Tab
    Select Project-àRC Click on Rebuild Project
    Select Project-àRC Click on Reload
    Then Deploye the Application
    How to Check in Explorer. It is converting to German Language or not
    Open Internet Explorer
    Tool-àInternet Options -àClick on Languages Button-àClick on ADD Button
    Select German Language--àClick on OK
    Now We will Check in Portal
    Created Web Dynpro iView in Portal. That iView assign to the particular WorkSet-àAssign to Role
    Select that user Can Change Language
    Regards
    Vijay Kalluri

  • Case statement in OWB mapping

    I am trying to use a case statement in an OWB mapping. I am attempting to use it in expression builder within a filter object. Each time I try to validate the expression in the expression builder I get an error:
    Line 4, Col 6:
    PL/SQL: ORA-00920: invalid relational operator
    Below is the code used in the expression builder window:
    CASE A.ATTRIBUTE_NAME
    WHEN 'CA COMPLETION DATE' THEN RTRIM(A.C_CODE_VALUE)
    ELSE NULL
    END
    Any help with what is happening would be appreciated.
    Thank you.
    Matt

    Thank you both for you help. I initially tried to use the expression operator but was confused since the expression builder in the expression operator has the drop down list with the available SQL clauses greyed out. In the OWB documentation it states that the drop down will contain SQL clauses that are appropriate for that expression type so when I saw it was greyed out I assumed that those clauses where not valid for this operator. Sure enough though when I pasted in my Case statement it verified correctly.
    Thanks again,
    Matt

  • Call a shell script from a OWB Mapping or Process Flow

    Hi,
    is it possible to call a shell script from a owb mapping or Process Flow and how can i do this?
    Greetings

    Check this :-
    http://blogs.oracle.com/warehousebuilder/2010/05/how_to_use_user_defined_activity_in_owb_process_flow.html
    Cheers
    Nawneet

  • How to Execute a OWB Mapping or a Process from SQLPLUS editor?

    I need to know how can we execute a OWB mapping from SQLPLUS editor or from a Stored Procedure.

    Hi,
    You can use the MDL export/import utility for this, but as you are importing into the same repository you will end up with importing into the same project again.
    So for this you can follow the workaround as:
    ->Take export of the required mapping from the project OWB_B
    -> Rename the Project OWB_B to some thing like OWB_B_1(or any other name), note that you have to rename both logical and physical name of the project
    -> Rename the project OWB_A to OWB_B and if you import the MDL all the objects will be imported to OWB_B then you can rename back the project to OWB_A
    -> Finally rename the project OWB_B_1 to OWB_B (to its original name).
    Regards,
    Pnreddy

  • JSF 1.2 web.xml configuration parameters

    Actually, I know these JSF 1.2 web.xml configuration parameters:
    javax.faces.CONFIG_FILES
    javax.faces.DEFAULT_SUFFIX
    javax.faces.LIFECYCLE_ID
    javax.faces.STATE_SAVING_METHOD
    Are there any more available?
    (I have to study its performance implications in conjunction with RichFaces ones).

    The JSF 1.2 specification document would contain the information you're looking for.

  • Transformations not occuring properly using PIVOT operator in OWB mapping.

    Hi ,
    In our project we have a OWB Map UII_D_MAP_SPC_CUST_FAULT_NATTR which is used to populate UII_CUSTOMER_FAULT table in our target side.
    The source table which populate UII_CUSTOMER_FAULT is SR_S_TABLE_N_ATTRIBUTEVALUE (a materialized view in another External Db)
    This Mview has 2 columns: S_N_STRINGVALUE & N_NAME.
    When N_NAME is 'FAULT_REPORT_CODE' the corresponding S_N_STRINGVALUE value should populate the REPORT_CODE field in UII_CUSTOMER_FAULTin the target sde.
    This source to target transformations is being done by using UNPIVOT operator in this mapping UII_MAP_SPC_CUST_FAULT_NATTR.
    In ideal case when S_N_STRINGVALUE is NOT NULL, REPORT_CODE value should be populated with the value of source field value (ie the value of S_N_STRINGVALUE) and when S_N_STRINGVALUE is NULL , REPORT_CODE should be populated with 'XX'.
    But in some cases REPORT_CODE value in UII_CUSTOMER_FAULT table is populated as NULL when S_N_STRINGVALUE is NOT NULL
    which should not happen.
    We suspect that there is some prpblem in the UNPIVOT operator, but we are not able to track down the exact location where it is failing. Can you please help in resolving this problem?
    Shall we remove Unpivot operator and use CASE statement in some package that will be called through Expression operator?
    Please advise.
    Regards
    Arinjit Dhar

    Arinjit,
    Have got any solution forthis. Today I ran into exactly the same problem. If you have got the solution, can you please post it in the forum.
    Ott Karesz,
    I have posted my SQL which got generated. The problem is when I am running the SQL just for the particular data set, it is giving the output properly. But what confuses me more is, when I run the mapping just for few records, it is populating the values correctly, but when running the mapping for all the records, then it populates null values for those records.
    Generated SQL:
    INSERT
    INTO
    "TB_PIPE_1"
    ("ASSETNUM",
    "SITEID",
    "TYPE",
    "STATUS",
    "YEARLAID",
    "MATERIALCODE",
    "PRESSURECODE",
    "DIAMETER",
    "METHODLAID",
    "JOINTTYPE",
    "SDRCOL",
    "LENGTHLAID",
    "LENGTHDIGITISED",
    "MEASUREUNITID")
    SELECT
    "UNPIVOT"."ASSETNUM" "ASSETNUM$1",
    "UNPIVOT"."SITEID" "SITEID$1",
    "UNPIVOT"."CLASSIFICATIONID" "CLASSIFICATIONID$1",
    "UNPIVOT"."STATUS" "STATUS$1",
    "UNPIVOT"."YEAR_LAID" "YEAR_LAID$1",
    "UNPIVOT"."MATERIAL" "MATERIAL$1",
    "UNPIVOT"."PRESSURE" "PRESSURE$1",
    "UNPIVOT"."DIAMETER" "DIAMETER$1",
    "UNPIVOT"."METHODLAID" "METHODLAID$1",
    "UNPIVOT"."JOINTTYPE" "JOINTTYPE$1",
    "UNPIVOT"."SDR" "SDR$1",
    "UNPIVOT"."LENGTHLAID" "LENGTHLAID$1",
    "UNPIVOT"."LENGTHDIGITESED" "LENGTHDIGITESED$1",
    "UNPIVOT"."MESUREUNIT" "MESUREUNIT$1"
    FROM (SELECT
    "ASSETNUM" "ASSETNUM",
    "SITEID" "SITEID",
    "STATUS" "STATUS",
    "YEAR_LAID" "YEAR_LAID",
    "CLASSIFICATIONID" "CLASSIFICATIONID",
    MIN(CASE WHEN "ASSETATTRID" = 'MATERIAL' THEN "ALNVALUE" ELSE NULL END) "MATERIAL",
    MIN(CASE WHEN "ASSETATTRID" = 'PRESSUREREGIME' THEN "ALNVALUE" ELSE NULL END) "PRESSURE",
    MIN(CASE WHEN "ASSETATTRID" = 'DIAMETER' THEN "NUMVALUE" ELSE NULL END) "DIAMETER",
    MIN(CASE WHEN "ASSETATTRID" = 'METHODLAID' THEN "ALNVALUE" ELSE NULL END) "METHODLAID",
    MIN(CASE WHEN "ASSETATTRID" = 'JOINTTYPE' THEN "ALNVALUE" ELSE NULL END) "JOINTTYPE",
    MIN(CASE WHEN "ASSETATTRID" = 'SDR' THEN "ALNVALUE" ELSE NULL END) "SDR",
    MIN(CASE WHEN "ASSETATTRID" = 'LENGTHLAID' THEN "NUMVALUE" ELSE NULL END) "LENGTHLAID",
    MIN(CASE WHEN "ASSETATTRID" = 'LENGTHDIGITISED' THEN "NUMVALUE" ELSE NULL END) "LENGTHDIGITESED",
    MIN(CASE WHEN "ASSETATTRID" = 'DIAMETER' THEN "MEASUREUNITID" ELSE NULL END) "MESUREUNIT"
    FROM (SELECT
    "MV_PIPE"."ASSETNUM" "ASSETNUM",
    "MV_PIPE"."SITEID" "SITEID",
    "MV_PIPE"."STATUS" "STATUS",
    "MV_PIPE"."YEAR_LAID" "YEAR_LAID",
    "MV_PIPE_SPEC"."ASSETATTRID" "ASSETATTRID",
    "MV_PIPE_SPEC"."NUMVALUE" "NUMVALUE",
    "MV_PIPE_SPEC"."ALNVALUE" "ALNVALUE",
    "MV_PIPE_SPEC"."MEASUREUNITID" "MEASUREUNITID",
    "MV_PIPE"."CLASSIFICATIONID" "CLASSIFICATIONID"
    FROM "MV_PIPE_SYN" "MV_PIPE",
    "MV_PIPE_SPEC_SYN" "MV_PIPE_SPEC" WHERE "MV_PIPE"."ASSETNUM" in ('466651706','606703143') and ( "MV_PIPE"."ASSETNUM" = "MV_PIPE_SPEC"."ASSETNUM" )) "OUTGRP1"
    GROUP BY
    "ASSETNUM", "SITEID", "STATUS", "YEAR_LAID", "CLASSIFICATIONID") "UNPIVOT"

  • Syntax error in profile relation while retreiving configuration parameters

    Trying to get our Directory Services properly functioning after an IP change. Running
    # kadmin.local -q listprincs
    yesterday revealed that all service principals weren't set up properly. Dig dig dig... found that I might (since this server was originally 10.3--now it's 10.4.4) try
    # slapconfig -kerberize diradmin
    That pointed me in the direction of using Workgroup Manager to fix a couple things with the KerberosClient record for the server. Did that. Reran the command. It complained about an improperly formatted configuration file because it wiped the contents of edu.mit.Kerberos (took out the DNS bits). Then ran
    # kerberosautoconfig -u
    which resulted in
    MergeLoggingInfo: creating default logging info
    At this point, edu.mit.Kerberos looked good. Then I tried
    # /usr/sbin/kdcsetup -f /LDAPv3/127.0.0.1 -w -a diradmin -p ** -v 1 REALM
    which resulted in
    Contacting the Directory Server
    Authenticating to the Directory Server
    Creating Kerberos directory
    Creating KDC Config File
    Creating Admin ACL File
    Creating Kerberos Master Key
    Creating Kerberos Database
    Syntax error in profile relation while retreiving configuration parameters
    SendInteractiveCommand: failed to get pattern
    Creating Kerberos Admin user
    kadmin.local: Syntax error in profile relation while initializing kadmin.local interface
    Adding kerberos auth authority to admin user
    Creating keytab for the admin tools
    kadmin.local: Syntax error in profile relation while initializing kadmin.local interface
    Adding KDC & kadmind to launchd
    Adding the new KDC into the KerberosClient config record
    edu.mit.kadmind: Already loaded
    com.apple.kdcmond: Already loaded
    Finished
    Now when I try
    # kadmin.local -q listprincs
    I get
    Authenticating as principal root/admin@REALM with password.
    kadmin.local: Syntax error in profile relation while initializing kadmin.local interface
    And now I'm getting this in kdc.log
    Feb 07 09:41:26 od.master.fqdn krb5kdc[21234](info): AS_REQ (7 etypes {18 17 16 23 1 3 2}) xx.xx.xx.xx: LOOKINGUPCLIENT: diradmin@REALM for krbtgt/REALM@REALM, Database is locked or in use--try again later
    Feb 07 09:41:26 od.master.fqdn krb5kdc[21234](info): DISPATCH: repeated (retransmitted?) request from xx.xx.xx.xx, resending previous response
    Help!
    Thanks,
    Rich
    ps. The typo in the error message is MIT's, not mine. I guess their engineering departments are better than their English department.
    Xserve G5   Mac OS X (10.4.4)  

    Look at the two config files:
    /Library/Preferences/edu.mit.Kerberos
    /var/db/krb5kdc/kdc.log
    the Database is locked or in use message is due to a missing lock file.
    Look in /var/db/krb5kdc/ for principal.kadm5.lock
    Hope this gets you started
    - Leland

  • SQL Query of an OWB map (row based)

    If I trace a session, execute an OWB map (row based), will the trace file contain the actual SQL query ?
    The problem with me is that when I am executing this row -based OWB map, it is throwing me an error CursorFetchMapTerminationRTV20007 BUT ( plus taking a long time) when I am taking out the intermediate SQL insert query,it is working fine ( and also within a very short period of time)
    Execution status = COMPLETE
    message text = ORA-06502: PL/SQL: numeric or value error: character string buffer too small
    message text = CursorFetchMapTerminationRTV20007
    No. task errors = 0
    No. task warnings = 2
    No. errors = 1
    Since this OWB map (Truncate Insert)  is row based hence I cannot cannot get the back end query from the OWB generated pl/sql package so wondering if I trace the session, check the trace file, may be I will able to see the exact SQL query generated. But wanted to confirm the same.

    Yes, the actual SQL run in the session will be in the trace file.

Maybe you are looking for

  • Issue with Line Graph in SSRS

    I am pulling my hair our trying to format a line graph. Below is my current report -  The Line Graph I have created has a Category Group of "YearLogged" and "MonthLogged". The series I have added is an expression - =Count(IIF(Fields!InvoicePaid_.Valu

  • Update query which taking more time

    Hi I am running an update query which takeing more time any help to run this fast. update arm538e_tmp t set t.qtr5 =(select (sum(nvl(m.net_sales_value,0))/1000) from mnthly_sales_actvty m where m.vndr#=t.vndr# and m.cust_type_cd=t.cust_type and m.cus

  • Help! This is driving me crazy. CASE statement report syntax error (missing operator)

    Hi Can anyone help with this SQL statement please. It uses a CASE statement but I keep getting back a syntax error (missing operator) error message. The SQL statement works without the CASE statement, so everything else is fine - its just when I put

  • Has the ability to add markers to a clip during capture been added to Premiere CS6?*

    I do live captures for video assist with PP5.5 and asked for the ability for markers to be added on the fly while capturing a while back. Just wondering if this has been added to PP6 or not. Use it all the time with Avid MC, and find it difficult to

  • 0FI_AP_6 assigning to DSO --Issue

    Hi, I have this std Datasource 0FI_AP_6. This when i try to assign to a custom DSO (copy of Std DSO 0FIAP_O06) i am getting infopackage assignment as well which was used in earlier datamodel.. i do not wan to have old infopackage in the new DSO assig