Problem with the creation of the date server with the technology Mysql

Hi,
I have installed date integrator 11.1.1.5.
I have installed oracle database 11.2
I have installed mysql 5.1.41 with xamp
I have created the repository master and work.
I have created data server for oracle .
ok all correct
I have created the date server for mysql but I cannot insert the data for jdbc (why ? where insert data ?) when except it creates me the date server.
but when I try to add a physical schema me from this error and I don't see the list of the schema.
java.lang.NoClassDefFoundError: bsh/EvalError
     at com.sunopsis.jdbc.driver.JavaDriver.connect(JavaDriver.java:80)
     at oracle.odi.jdbc.datasource.DriverManagerDataSource.getConnectionFromDriver(DriverManagerDataSource.java:409)
     at oracle.odi.jdbc.datasource.DriverManagerDataSource.getConnectionFromDriver(DriverManagerDataSource.java:385)
     at oracle.odi.jdbc.datasource.DriverManagerDataSource.getConnectionFromDriver(DriverManagerDataSource.java:352)
     at oracle.odi.jdbc.datasource.DriverManagerDataSource.getConnection(DriverManagerDataSource.java:331)
     at oracle.odi.jdbc.datasource.LoginTimeoutDatasourceAdapter$ConnectionProcessor.run(LoginTimeoutDatasourceAdapter.java:217)
     at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
     at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
     at java.util.concurrent.FutureTask.run(FutureTask.java:138)
     at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
     at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
     at java.lang.Thread.run(Thread.java:662)
how can I resolve my problem?
thank you for help and cooperation.
gabriele

Hi,
A sub asset master can be created with AS11 only.
You cannot create a asset sub number under a main asset number, in the case, where different set of dep areas needs to be maintained.
When you create a sub asset, all the dep areas will be copied to sub asset's master from main asset master. Since this set of dep areas cannot be different from main asset to sub asset.
Hopw you understood this.
Thanks,
Srinu.

Similar Messages

  • Get the month from a date column with the calculated column

    I am trying to get the month from a date field with the help of calculated column. However I get this syntax error whenever I want to submit the formula:
    Error 
    The formula contains a syntax error or is not supported. 
    the default language of our site is German and [datum, von] is a date field.

    Hi,
    I have created two columns
    Current MM-YY
    Calculated (calculation based on other columns)
    Today
    Date and Time
    Current MM-YY is calculated value with formula as
    =TEXT(Today,"mmmm")
    But the output shows as December instead of May.
    I have tried =TEXT([Datum, von];"mmmm") but no help.
    I am trying to populated the column automatically with current month..ex: if its May the field should show May, next month it should show June an so on.
    Any kind help is grateful.
    Regards,
    Pradeep

  • I just purchased the Mac Mini Server with the intention of replacing Lion Server with Lion for the first hard drive while installing Windows 7 thru Bootcamp on the second hard drive. Is it possible for me to do this? Thanks.

    I just purchased the Mac Mini Server with the intention of replacing Lion Server with Lion for the first hard drive while installing Windows 7 thru Bootcamp on the second hard drive. Is it possible for me to do this? Thanks.

    I would use Parallels or VMWare and use the second drive to hold the virtual machine. That way the VM only uses the space that it needs and you can still use the remainder of the drive for other things, not to mention that you would not need to reboot to run windows.
    Also Time Machine does not back up a boot camp partition.

  • HT5439 Has anyone had the Network time and date error with Apple TV when traveling?

    Has anyone had the Network time and date error with Apple TV when traveling?

    Jasion,
    Maybe a bit beside the question, but it is always recommended here to open/save from/to own hard disk, and to copy from/to networks and removable media. You may be lucky, at least for a while, but it is better to be safe than sorry.
    In addition to the (far greater) risk of file corruption, some issues are mentioned here:
    http://helpx.adobe.com/illustrator/kb/illustrator-support-networks-removable-media.html

  • Connect to a data server with topology manager

    Hi! I am new to ODI...
    I've created a master and work repository. Now I'm in topology manager to create a data server from Physical Architecture. I succeed in connecting to this data server with "Oracle Toad" using a service name...so I select JDBC driver and then I insert jdbc:oracle:oci8:@<tns_alias> (with the right service name).
    Then it give me the message that it can connect to the data server, but when i click ok, it says me:
    {color:#ff0000}java.lang.UnsatisfiedLinkError: no ocijdbc10 in java.library.path
    at java.lang.ClassLoader.loadLibrary(Unknown Source)
    at java.lang.Runtime.loadLibrary0(Unknown Source)
    at java.lang.System.loadLibrary(Unknown Source)
    at oracle.jdbc.driver.T2CConnection.loadNativeLibrary(T2CConnection.java:3013)
    at oracle.jdbc.driver.T2CConnection.logon(T2CConnection.java:228)
    at oracle.jdbc.driver.PhysicalConnection.<init>(PhysicalConnection.java:361)
    at oracle.jdbc.driver.T2CConnection.<init>(T2CConnection.java:142)
    at oracle.jdbc.driver.T2CDriverExtension.getConnection(T2CDriverExtension.java:79)
    at oracle.jdbc.driver.OracleDriver.connect(OracleDriver.java:595)
    at com.sunopsis.sql.SnpsConnection.u(SnpsConnection.java)
    at com.sunopsis.sql.SnpsConnection.a(SnpsConnection.java)
    at com.sunopsis.sql.SnpsConnection.testConnection(SnpsConnection.java)
    at com.sunopsis.sql.SnpsConnection.testConnection(SnpsConnection.java)
    at com.sunopsis.graphical.frame.b.jz.dP(jz.java)
    at com.sunopsis.graphical.frame.b.jz.bD(jz.java)
    at com.sunopsis.graphical.frame.bo.bz(bo.java)
    at com.sunopsis.graphical.frame.b.jz.em(jz.java)
    at com.sunopsis.graphical.frame.b.jz.<init>(jz.java)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
    at java.lang.reflect.Constructor.newInstance(Unknown Source)
    at com.sunopsis.graphical.frame.bb.b(bb.java)
    at com.sunopsis.graphical.tools.utils.swingworker.v.call(v.java)
    at edu.emory.mathcs.backport.java.util.concurrent.FutureTask.run(FutureTask.java:176)
    at com.sunopsis.graphical.tools.utils.swingworker.l.run(l.java)
    at edu.emory.mathcs.backport.java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:665)
    at edu.emory.mathcs.backport.java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:690)
    at java.lang.Thread.run(Unknown Source)
    {color}
    My environment variable for Oracle are:
    C:\oracle\ora92\bin;C:\Programmi\Oracle\jre\1.3.1\bin;
    I succeed in connecting with OCI using Toad...why can't I connect with ODI? Need I to install another component? (and which?)
    Sorry for my english...please help me! thanks....
    Alessia
    Edited by: user1762560 on 26-feb-2009 9.06

    jdbc:oracle:oci8:@tns_alias
    it doesn't contain the sign '< ' or '>'

  • The installed version of SQL Server could not be verified or is not supported. Verify that the computer and the installed version of SQL Server meet the minimum requirements for installation

    Hi Guys,
    Currently i am trying to install SCOM2012R2 with SQL 2012 SP1. When i proceed with SCOM installation, getting an error
    "The installed version of SQL Server could not be verified or is not supported. Verify that the computer and the installed version of SQL Server meet the minimum requirements for installation, and that the firewall settings are correct. See the Supported
    Configurations document for further information.
    SQL Collation i selected as per System Center2012 guide SQL_Latin1_General_CP1_CI_AS
    Can some one please help me on this?
    Regards
    Kris
    Kris

    Hi,
    Please check logs for further information. Let us know what the logs says:
    http://www.systemcentercentral.com/opsmgr-2012-my-installation-failed-what-log-files-do-i-read-and-where-can-i-find-them/
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • Two or more productid will be accquired by same customerid, by same shipvia, on the same day of the week of shipped date. i want the simple query for this.

    consider this situation,
     Two or more productid will be accquired by same customerid, by same shipvia, on the same  day of the week of shipped date. i want the simple query for this.
    my tables are  from northwind:
    [orders] = OrderID, CustomerID, EmployeeID, OrderDate, RequiredDate, ShippedDate, ShipVia, Freight, ShipName, ShipAddress, ShipCity, ShipRegion, ShipPostalCode, ShipCountry.
    [orders details] = OrderID, ProductID, UnitPrice, Quantity, Discount.
    i tried some but it is not exact, it gives wrong result.
    select pd.CustomerID,pd.ProductID, pd.no_of_time_purchased, sd.ShipVia, sd.same_ship_count, shipped_day from
    select ProductID,o.CustomerID,COUNT(productid) as no_of_time_purchased
    from orders o join [Order Details] od on o.OrderID=od.OrderID group by ProductID,o.CustomerID
    having count(od.ProductID) >1) pd
    join
    (select OrderID,customerid, shipvia, count(shipvia)as same_ship_count, DATENAME(DW,ShippedDate)as shipped_day from orders
    group by customerid, ShipVia, ShippedDate having COUNT(ShipVia) > 1 ) sd
    on sd.CustomerID=pd.CustomerID

    Hi,
    I think I have a solution that will at least give you a clue how to go about it. I have simplified the tables you mentioned and created them as temporary tables on my side, with some fake data to test with. I have incldued the generation of these temporary
    tables for your review.
    In my example I have included:
    1. A customer which has purchased the same product on the same day, using the same ship 3 times,
    2. Another example the same as the first but the third purchase was on a different day
    3. Another example the same as the first but the third purchase was a different product
    4. Another example the same as the first but the third purchase was using a different "ShipVia".
    You should be able to see that by grouping on all of the columns that you wich to return, you should not need to perform any subselects.
    Please let me know if I have missed any requirements.
    Hope this helps:
    CREATE TABLE #ORDERS
     OrderID INT,
     CustomerID INT,
     OrderDate DATETIME,
     ShipVia VARCHAR(5)
    CREATE TABLE #ORDERS_DETAILS
     OrderID INT,
     ProductID INT,
    INSERT INTO #ORDERS
    VALUES
    (1, 1, GETDATE(), 'ABC'),
    (2, 1, GETDATE(), 'ABC'),
    (3, 1, GETDATE(), 'ABC'),
    (4, 2, GETDATE() - 4, 'DEF'),
    (5, 2, GETDATE() - 4, 'DEF'),
    (6, 2, GETDATE() - 5, 'DEF'),
    (7, 3, GETDATE() - 10, 'GHI'),
    (8, 3, GETDATE() - 10, 'GHI'),
    (9, 3, GETDATE() - 10, 'GHI'),
    (10, 4, GETDATE() - 10, 'JKL'),
    (11, 4, GETDATE() - 10, 'JKL'),
    (12, 4, GETDATE() - 10, 'MNO')
    INSERT INTO #ORDERS_DETAILS
    VALUES
    (1, 1),
    (2, 1),
    (3, 1),
    (4, 2),
    (5, 2),
    (6, 2),
    (7, 3),
    (8, 3),
    (9, 4),
    (10, 5),
    (11, 5),
    (12, 5)
    SELECT * FROM #ORDERS
    SELECT * FROM #ORDERS_DETAILS
    SELECT
     O.CustomerID,
     OD.ProductID,
     O.ShipVia,
     COUNT(O.ShipVia),
     DATENAME(DW, O.OrderDate) AS [Shipped Day]
    FROM #ORDERS O
    JOIN #ORDERS_DETAILS OD ON O.orderID = OD.OrderID
    GROUP BY OD.ProductID, O.CustomerID, O.ShipVia, DATENAME(DW, O.OrderDate) HAVING COUNT(OD.ProductID) > 1
    DROP TABLE #ORDERS
    DROP TABLE #ORDERS_DETAILS

  • Approach when the used Live cache data area crosses the threshold

    Hi,
    Could any of you please let me know the detailed approach when the used Live cache data area crosses the threshold in APO system?
    The approach I have as of now is :
    1) When it is identified that data cache usage is nearly 100%, check for hit rate for OMS data in data cache in LC10 .Because generally hit rate for OMS data in data cache should be atleaset 99.8% and Data Cache usage should be well below 100%.
    2) To monitor unsuccessful accesses to data cache choose refresh and compare value now and before unsuccessful accesses result in physical disk I/O and should generally be avoided.
    3) The number of OMS data pages (OMS Data) should be much higher than the number of OMS history pages (History/Undo).A ratio of 4:1 is desirable. If OMS history has nearly the same size as OMS data, use Problem AnalysisPerformanceOMS versions to find out if named consistent views (Versions) are open for a long time. Maximum age should be 8hrs.
    4)If consumption of OMS heap and data cache is large, one reason may be a long running transaction simulation that accumulates heap memory and prevents the garbage collector from releasing old object images.
    5) To display existing transactional simulations in LC10,use Problem AnalysisPerformanceOMS versions and SM04 to find out user of corresponding transaction and may be required to cancel the session after contacting user if the version open for long time..
    Please help me by providing additional information on the issue.
    Thanks,
    Varada Reddy.

    Hi Mayank, sorry, one basic question - are you using some selection criteria during extraction? If yes, then try extraction without the selection criteria.
    If you maintain selection based on, let's say, material, you need to use the right number of zeros as prefix (based on how you have defined the characteristic for material) otherwise no records would be selected.
    Is this relevant in your case?
    One more option is to try to repair teh datasource. In the planning area, go to extraction tools, select the datasource, and then choose the option of repair datasource.
    If you need more info, pls let me know.
    - Pawan

  • Detailed approach when the used Live cache data area crosses the threshold

    Hi,
    Could any of you please let me know the detailed approach when the used Live cache data area crosses the threshold in APO system?
    The approach I have as of now is :
    1) When it is identified that data cache usage is nearly 100%, check for hit rate for OMS data in data cache in LC10 .Because generally hit rate for OMS data in data cache should be atleaset 99.8% and Data Cache usage should be well below 100%.
    2) To monitor unsuccessful accesses to data cache choose refresh and compare value now and before unsuccessful accesses result in physical disk I/O and should generally be avoided.
    3) The number of OMS data pages (OMS Data) should be much higher than the number of OMS history pages (History/Undo).A ratio of 4:1 is desirable. If OMS history has nearly the same size as OMS data, use Problem AnalysisPerformanceOMS versions to find out if named consistent views (Versions) are open for a long time. Maximum age should be 8hrs.
    4)If consumption of OMS heap and data cache is large, one reason may be a long running transaction simulation that accumulates heap memory and prevents the garbage collector from releasing old object images.
    5) To display existing transactional simulations in LC10,use Problem AnalysisPerformanceOMS versions and SM04 to find out user of corresponding transaction and may be required to cancel the session after contacting user if the version open for long time..
    Please help me by providing additional information on the issue.
    Thanks,
    Varada Reddy.

    Hi Mayank, sorry, one basic question - are you using some selection criteria during extraction? If yes, then try extraction without the selection criteria.
    If you maintain selection based on, let's say, material, you need to use the right number of zeros as prefix (based on how you have defined the characteristic for material) otherwise no records would be selected.
    Is this relevant in your case?
    One more option is to try to repair teh datasource. In the planning area, go to extraction tools, select the datasource, and then choose the option of repair datasource.
    If you need more info, pls let me know.
    - Pawan

  • How to get from Apple (officially) details on my iphone 5 (knowing the serial number): model, date of sale, the date of activation, etc.

    How to get from Apple (officially) details on my iphone 5 (knowing the serial number): model, date of sale, the date of activation, etc. I bought a new iphone 5 on ebay.com - he was used. For a decision on my case ebay asks for this information ...

    Apple will never reveal that info to you, since you were not the originally purchaser. However, you can get the date of sale by entering the SN here:
    https://selfsolve.apple.com/agreementWarrantyDynamic.do
    Warranty starts on the date of sale, which is most likely the date of activation, but not necessarily so.

  • Query in Data server of XML Technology

    Hello,
    We are doing Database to XML Transformation.we are achieving this successfully.
    But we are facing problem in XSD placeholder in target data server i.e XML Data server
    First time we mentioned absolute path of XSD and got target datastore by reverse engineer XSD structure.
    if i export my project and model into another machine.Then at another machine if i run my package i am getting error like
    Caused By: java.sql.SQLException: As no XML and no DTD were provided, the XML schema cannot be created
    1)
    My confusion is here, why every time during execution there is a need of XSD,because XSD represent the structure of XML file
    And that structure ,i already imported in another machine i.e target model
    2)
    if there is need , then Can we embedd XSD with importing of project and model so that whenever we export that project model in another machine ,my XSD also imported .
    3) OR
    Is there any default location of data server of XML technology where agent will go for finding XSD ?
    if more information required,let us know.
    please suggest.

    I haven't tried this myself but give this a go:
    1. In Word, highlight the Title Row
    2. Then go to Table > Repeat Header Rows. In Word 2007+ the Repeat Header Rows button is under the Layout section.
    3. Save, upload and re-run.
    Do the table headers now repeat?
    I believe this post will also help you: Header table to repeat on every page

  • Problems in the creation of a MOLAP Cube with DBMS_AWM.

    I want to create a MOLAP Cube with the package DBMS_AWM. So I create the ROLAP Cube and the Dimensions with the Enterprise Manager Website, and all works perfectly. Then I executed the code to create the multidimensional dimensions and the multidimensional cube (awm dimensions/cubes), but I had some problems in the first dimension.
    This dimension has the name D_DESTIN, and has the following hierarchy:
    +DESTIN_DESC_H3
    |
    +---DESTIN_DESC_H2
    |
    +------DESTIN_DESC_H1
    |
    +---------DESTIN_DESC_H0
    The name of the hierarchy is H_D_DESTIN.
    The following code is the code that I used to create the first multidimensional dimension:
    set serveroutput on
    execute cwm2_olap_manager.set_echo_on;
    execute dbms_aw.execute ('aw create ''WTESTE''');
    execute dbms_awm.create_awdimension('EXEMPLO_OLAP', 'D_DESTIN', 'EXEMPLO_OLAP','WTESTE', 'WD_DESTIN');
    execute dbms_awm.create_awdimload_spec('D_DESTIN_LOAD', 'EXEMPLO_OLAP', 'D_DESTIN', 'FULL_LOAD');
    execute DBMS_AWM.SET_AWDIMLOAD_SPEC_PARAMETER ('D_DESTIN_LOAD','EXEMPLO_OLAP','D_DESTIN','UNIQUE_RDBMS_KEY','NO');
    execute dbms_awm.refresh_awdimension('EXEMPLO_OLAP', 'WTESTE', 'WD_DESTIN', 'D_DESTIN_LOAD');
    commit;
    execute cwm2_olap_manager.set_echo_off;
    execute cwm2_olap_manager.end_log
    When I executed the code above, I got the following error:
    PL/SQL procedure successfully completed.
    SP2-0103: Nothing in SQL buffer to run.
    PL/SQL procedure successfully completed.
    AMD-00001 created AWDimension "EXEMPLO_OLAP.WTESTE.WD_DESTIN"
    PL/SQL procedure successfully completed.
    AMD-00001 created AWDimLoad_Spec "D_DESTIN_LOAD.EXEMPLO_OLAP.D_DESTIN"
    PL/SQL procedure successfully completed.
    AMD-00002 set AWDimLoad_Spec_Parameter "D_DESTIN_LOAD.EXEMPLO_OLAP.D_DESTIN"
    UNIQUE_RDBMS_KEY to "NO"
    PL/SQL procedure successfully completed.
    ERROR Create_AWDimension. Problem refreshing dimension:
    WD_DESTIN
    Error
    Validating Dimension Mappings WD_DESTIN.DIMENSION. Key Expression
    DWH.D_DESTIN.DESTIN_KEY for Mapping Group
    WD_DESTIN.H_D_DESTIN.DESTIN_DESC_H0.DWH_D_DESTIN_WD_DESTIN_H_D_DESTIN
    DESTINDESC_H0.DIMENSIONMAPGROUP, Level WD_DESTIN.DESTIN_DESC_H0.LEVEL,
    Hierarchy WD_DESTIN.H_D_DESTIN.HIERARCHY is Incorrectly Mapped to
    RDBMS.
    (AW$XML) AW$XML
    In SYS.AWXML!__XML_HANDLE_ERROR PROGRAM:
    BEGIN dbms_awm.refresh_awdimension('EXEMPLO_OLAP', 'WTESTE',
    'WD_DESTIN', 'D_DESTIN_LOAD'); END;
    ERROR at line 1:
    ORA-06510: PL/SQL: unhandled user-defined exception
    ORA-06512: at "OLAPSYS.DBMS_AWM", line 1012
    ORA-06512: at line 1
    Commit complete.
    PL/SQL procedure successfully completed.
    PL/SQL procedure successfully completed.
    I don’t know what is wrong. The ROLAP Cube is valid according to Oracle Enterprise Manager Website, and it is possible to consult its data with "OracleBI Spreadsheet Add-In"
    What is wrong?
    Regards,
    Rui Torres

    I executed the same code in a different user and the MOLAP Cube was created successfully.
    But, I don’t know what is the privilege/role, that permits this second user to create a MOLAP cube with the package DBMS_AWM.
    The privileges/roles of the first user are:
    ROLES
    ======
    CONNECT
    OLAP_DBA
    OLAP_USER
    OWBR_EXEMPLO_OLAP
    OWB_EXEMPLO_OLAP
    SYSTEM PRIVILEGES
    =================
    ALTER SESSION     
    CREATE ANY PROCEDURE     
    CREATE DATABASE LINK     
    CREATE DIMENSION     
    CREATE INDEXTYPE     
    CREATE MATERIALIZED VIEW     
    CREATE PROCEDURE     
    CREATE PUBLIC DATABASE LINK     
    CREATE PUBLIC SYNONYM     
    CREATE ROLE     
    CREATE SEQUENCE     
    CREATE SESSION     
    CREATE SYNONYM     
    CREATE TABLE     
    CREATE TRIGGER     
    CREATE TYPE     
    CREATE VIEW     
    DROP ANY PROCEDURE     
    DROP PUBLIC SYNONYM     
    EXECUTE ANY PROCEDURE     
    GLOBAL QUERY REWRITE     
    SELECT ANY TABLE     
    SYSDBA     
    UNLIMITED TABLESPACE
    OBJECTS PRIVILEGES
    ==================
    Object      Privilege     |Schema |Object     
    =======================================================
    SELECT               |SYS     |DBA_ROLE_PRIVS     
    EXECUTE               |SYS     |DBMS_LOCK     
    SELECT               |SYS     |DBMS_LOCK_ALLOCATED     
    EXECUTE               |SYS     |DBMS_OBFUSCATION_TOOLKIT     
    EXECUTE               |SYS     |DBMS_SNAPSHOT     
    SELECT               |SYS     |V_$LOCK     
    SELECT               |SYS     |V_$MYSTAT     
    SELECT               |SYS     |V_$SESSION     
    SELECT               |SYS     |V_$SYSTEM_PARAMETER
    The privileges/roles of the second user are:
    ROLES
    ======
    AQ_ADMINISTRATOR_ROLE          
    DBA          
    MGMT_USER
    SYSTEM PRIVILEGES
    =================
    CREATE MATERIALIZED VIEW     
    CREATE TABLE     
    GLOBAL QUERY REWRITE     
    SELECT ANY TABLE     
    UNLIMITED TABLESPACE
    OBJECTS PRIVILEGES
    ==================
    Object Privilege     |Schema     |Object     
    =============================================
    EXECUTE               |SYS     |DBMS_ALERT     
    EXECUTE               |SYS     |DBMS_AQ     
    EXECUTE               |SYS     |DBMS_AQADM     
    EXECUTE               |SYS     |DBMS_AQELM     
    EXECUTE               |SYS     |DBMS_AQ_IMPORT_INTERNAL     
    EXECUTE               |SYS     |DBMS_DEFER_IMPORT_INTERNAL     
    EXECUTE               |SYS     |DBMS_REPCAT     
    EXECUTE               |SYS     |DBMS_RULE_EXIMP     
    EXECUTE               |SYS     |DBMS_SYS_ERROR     
    EXECUTE               |SYS     |DBMS_TRANSFORM_EXIMP     
    ALTER               |SYS     |INCEXP     
    DEBUG               |SYS     |INCEXP     
    DELETE               |SYS     |INCEXP     
    FLASHBACK          |SYS     |INCEXP     
    INDEX               |SYS     |INCEXP     
    INSERT               |SYS     |INCEXP     
    ON COMMIT REFRESH     |SYS     |INCEXP     
    QUERY REWRITE          |SYS     |INCEXP     
    REFERENCES          |SYS     |INCEXP     
    SELECT               |SYS     |INCEXP     
    UPDATE               |SYS     |INCEXP     
    ALTER               |SYS     |INCFIL     
    DEBUG               |SYS     |INCFIL     
    DELETE               |SYS     |INCFIL     
    FLASHBACK          |SYS     |INCFIL
    Which privilege/role permits the second user to create a MOLAP cube?
    Regards,
    Rui Torres

  • Problem to calculate the coherence (using NetworkFunction-VI) with only 1 row of data for each, the stimulus and response input

    Hello,
    I am trying to calculate the coherence of a stimulus and response
    signal using the Network Functions (avg) VI's. The problem is that I
    always get a coherence of "1" at all frequencies. This problem is
    already known (see KnowledgeBase document: Why is the Network Functions (avg) VI's Coherence Function Output "1"?).
    My trouble is that the described solution (-> the stimulus and response input matrices need to have at least two rows to get non-unity coherence values) doesn't help me much, because I only have one array of stimulus data and one array of response values.
    Thus, how can I fullfil the 'coherence-criteria' to input at least two rows of data each when I just have one row of data each?
    Any hint or idea is very much appreciated. Thanks!
    Horst

    With this weird board layout, I'm not sure whether you were asking me, but, on the assumption that you were, here goes:
    I found no need to use the cross-power spectrum and power spectrum blocks
    1... I was looking for speed.
    2... I already had the component spectral data there, for other purposes. From that, it's nothing but addition and multiplication.
    3... The "easy" VIs, assume a time wave input, as I recall. Which means they would take the same spectrum of the same timewave several times, where I only do it once.
    I have attached PNGs of my code.
    The PROCESS CHANNEL vi accepts the time wave and:
    1... Removes DC value.
    2... Integrates (optional, used for certain sensors).
    3... Windows (Hanning, etc. - optional)
    4... Finds spectrum.
    5... Removes spectral mirrors.
    6... Scales into Eng. units.
    7... From there, you COULD use COMPLEX-TO-POLAR, but I don't care about the phase data, and I need the MAG^2 data anyway, so I rolled my own COMPLEX-TO-MAG code.
    The above is done on each channel. The PROCESS DATA vi calls the above with data for each channel. The 1st channel in the list is required to be the reference (stimulus) channel.
    After looping over each channel, we have the Sxx, Syy, and Sxy terms. This code contains some averaging and peak-picking stuff that's not relevant.
    From there, it's straightforward to ger XFER = Sxy/Sxx and COHERENCE = |Sxy|^2 / (Sxx * Syy)
    Note that it uses the MAGNITUDE SQUARED of Sxy. Again, if you use the "easy" stuff, it will do a square-root operation that you just have to reverse - it is obtained faster by the sum of the squares of the real and imag parts.
    Hope this helps.
    Steve Bird
    Culverson Software - Elegant software that is a pleasure to use.
    Culverson.com
    Blog for (mostly LabVIEW) programmers: Tips And Tricks
    Attachments:
    ProcessChannel.png ‏25 KB

  • Problem with SQL*Loader and different date formats in the same file

    DB: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    System: AIX 5.3.0.0
    Hello,
    I'm using SQL*Loader to import semi-colon separated values into a table. The files are delivered to us by a data provider who concatenates data from different sources and this results in us having different date formats within the same file. For example:
    ...;2010-12-31;22/11/1932;...
    I load this data using the following lines in the control file:
    EXECUTIONDATE1     TIMESTAMP     NULLIF EXECUTIONDATE1=BLANKS     "TO_DATE(:EXECUTIONDATE1, 'YYYY-MM-DD')",
    DELDOB          TIMESTAMP     NULLIF DELDOB=BLANKS          "TO_DATE(:DELDOB, 'DD/MM/YYYY')",
    The relevant NLS parameters:
    NLS_LANGUAGE=FRENCH
    NLS_DATE_FORMAT=DD/MM/RR
    NLS_DATE_LANGUAGE=FRENCH
    If I load this file as is the values loaded into the table are 31 dec 2010 and 22 nov *2032*, aven though the years are on 4 digits. If I change the NLS_DATE_FORMAT to DD/MM/YYYY then the second date value will be loaded correctly, but the first value will be loaded as 31 dec *2020* !!
    How can I get both date values to load correctly?
    Thanks!
    Sylvain

    This is very strange, after running a few tests I realized that if the year is 19XX then it will get loaded as 2019, and if it is 20XX then it will be 2020. I'm guessing it may have something to do with certain env variables that aren't set up properly because I'm fairly sure my SQL*Loader control file is correct... I'll run more tests :-(

  • Issue to send email (using sp_send_dbmail) with an attachement located in other server within the same network

    I have two servers. One is for databases (DB) with the mail server configured and the other is for SSIS without mail server. Both servers are in the same network. On the SSIS server the SSIS package exports data from the DB server table
    to xls file. The file is saved in SSIS server. When I created a task to send the email (using sp_send_dbmail) to attach that file, I got the error message:
    Error: Executing the query "EXECUTE [msdb].[dbo].[sp_send_dbmail]
     @profile_na..." failed with the following error: "Attachment file H:\test.xls is invalid.".
    Any help will be appreciated. Thanks.
    A Fan of SSIS, SSRS and SSAS

    Where are you running this?
    If its in the other server (DB) you need to use UNC path to the SSIS server folder where file resides
    Also the executing account should have access to the folder as well as the file. So If you're running this from a job make sure you either provider folder and file access to sql agent job service account or create a separate proxy account with the required
    permissions and configure the job to use the account
    see
    http://www.databasejournal.com/features/mssql/article.php/3789881/Proxy-Accounts-in-SQL-Server.htm
    http://www.codeproject.com/Articles/28918/SQL-Server-Agent-Proxy
    Please Mark This As Answer if it solved your issue
    Please Vote This As Helpful if it helps to solve your issue
    Visakh
    My Wiki User Page
    My MSDN Page
    My Personal Blog
    My Facebook Page

Maybe you are looking for

  • We couldn't release any transport request!

    Hi experts, We could't release any transport request on DEV from today. We could do it until yesterday. We have never changed any settings. Except release, we can use SAP system, and we can import released transport request to QAS . Also I could tp a

  • How to get passed Apple Activation Lock with an Apple ID and email that no longer exists

    I recently got my mother's IPhone second hand. She had her Apple ID, and Find my Iphone feature registered under her old work email address, neither of these no longer exist, and so I am stuck and can't unlock the phone. Any Advice?

  • Printer drivers...software downloaded, now how do i activate it?

    had to have os reinstalled; backed up with time machine; printer still shows up, but error message tells me to install software.  cannot use cd (apparently) so downloaded all canon drivers from apple. now what??? (coming from windows, i find macbook

  • Dented Macbook Pro 13 Inch

    Hello, I was at school with my Macbook Pro (We have to take them to school), and someone was walking past my desk and knocked my Macbook of the table. Of course I was very annoyed, and to make things worse it was open as it fell and hit the ground. T

  • HT4623 iOS 8 version 8.1.2 issues with the calendar displaying wrong TIME ZONE!

    I just updated to iOS 8 using version 8.1.2 and am having issues with the calendar.  I use my GMAIL calendar and sync to my iPhone and now any calendar entries that I add from my phone display the current time zone AND GMT time.  Entries that were al