Load failed due the spl char

Hello Gurus,
I am working on BW 3.5 version,  While loading the data to different ODS's. The loads are getting fail becoz of the below error.
*Value 'PPS2AEGN16ZZZSFBA  ' (hex. '505053324145474E31365A5A5A53464241A0 ') of characteristic 0MATERIAL contains invalid*
the problem with the above material  value is it  included space(blank) at the end.
For this i have included the FM 'SCP_REPLACE_STRANGE_CHARS'  in the Transfer Rules to ignore all the special Char for the 0material Info object. but still i am facing the same problem.
Kindly let me know if there is another solution for this. or i am missing any thing.
Thanks and Regards
Edited by: KK on Aug 13, 2009 12:10 AM

Hello thanks for ur response,
Space is on the right side of the material. hope i can maintain at both left and right sides bcoz to overcome the problem in future.
what i hav to include for the same kindly let me know
Regards

Similar Messages

  • BW Data Load error due to special char's

    Hi Experts,
    Our BW System is a Uni Code, Master data load din BW failed due to special chars, I corrected the field manually in PSA and loaded the
    data.
    But I would like to know the reason for failure.
    Issue for failure is term SOCIÉTÉGÉN…..My understanding its French and BW should accept it…why is it failing?
    Thanks in advance

    Hi User,
    The alphabet ' É ' which you have mentioned is a special Character.
    To avoid this failure in future , Please add  " É" this alphabet in RSKC t-code which would help to avoid
    special characters.
    Please revert in case you need any further details.
    Thanks & Regards,
    RDS

  • Incremental load fails with the error LM_44127 Failed to prepare the task

    Guys,
    I have created a custom mapping and cretaed a execution plan for this mapping in the DAC. The full load completes successfully. But when ever the incremental lod is run , i am getting the below error and the task fails(The SDE load completes sucessfully , but the SIL load fails with the below error).
    LM_44127     Failed to prepare the task
    Please help!!!

    i googled it..
    http://datawarehouse.ittoolbox.com/groups/technical-functional/informatica-l/lm_44127-failed-to-prepare-task-when-running-workflow-in-informatica-86-on-aix-3199309
    you can try for better links now.. !!

  • ODI-1228: Task SrcSet0 (Loading) fails on the target MICROSOFT_EXCEL connec

    Hi
    i am trying to exporting the database table data into excel sheet i am getting the below error.
    ODI-1228: Task SrcSet0 (Loading) fails on the target MICROSOFT_EXCEL connection Ms_EXCE_TGT.
    Caused By: java.sql.SQLException: [Microsoft][ODBC Excel Driver] 'C$_0Location' is not a valid name. Make sure that it does not include invalid characters or punctuation and that it is not too long.
    Code:
    create table "C$_0Location"
         C1_POSITION_DEFINITION_ID     VARCHAR(255) ,
         C2_SUMMARY_FLAG__________     VARCHAR(255) ,
         C3_ENABLED_FLAG__________     VARCHAR(255)
    Edited by: 985498 on May 15, 2013 4:06 AM

    The odbc driver doesn't seem to like the special character "$" int eh table name.
    Why is the C$ table being created on through the odbc driver. You cannot have non-sql technology for staging. Making the staging area different than target should help in your case.

  • ODI-1228: Task SrcSet0 (Loading) fails on the target SUNOPSIS_ENGINE connec

    Hi,
    When we are trying to load data from MSSQL Server to Hyperion Essbase i am getting below error
    ODI-1228: Task SrcSet0 (Loading) fails on the target SUNOPSIS_ENGINE connection SUNOPSIS_MEMORY_ENGINE.
    Caused By: java.sql.SQLException: statement is not in batch mode
    *     at org.hsqldb.jdbc.Util.sqlException(Unknown Source)*
    *     at org.hsqldb.jdbc.Util.sqlException(Unknown Source)*
    *     at org.hsqldb.jdbc.Util.sqlExceptionSQL(Unknown Source)*
    *     at org.hsqldb.jdbc.JDBCPreparedStatement.executeBatch(Unknown Source)*
    *     at oracle.odi.runtime.agent.execution.sql.SQLCommand.end(SQLCommand.java:267)*
    *     at oracle.odi.runtime.agent.execution.sql.SQLExecutor.endExecution(SQLExecutor.java:156)*
    *     at oracle.odi.runtime.agent.execution.sql.SQLExecutor.endExecution(SQLExecutor.java:1)*
    *     at oracle.odi.runtime.agent.execution.DataMovementTaskExecutionHandler.handleTask(DataMovementTaskExecutionHandler.java:113)*
    Thanks
    Rama

    as written here
    https://kr.forums.oracle.com/forums/thread.jspa?threadID=2421461
    Please us the no_batch_update or nobu as false in the JDBC URL of the Topology Navigator. &nobu=false

  • (Loading) fails on the target SUNOPSIS_ENGINE connection MEMORY_ENGINE.

    Hi
    I have run my first interface in ODI 11g and come up with the below error during the Loading KM phase. Can anyone suggest a way to fix?
    The background is that we are migrating from a 10g in a SQL Server environment to a 11g in an Oracle env. I export/imported the work rep and logical schemas. I imported the physical connections one by one since I couldn't import the master rep.
    ODI-1228: Task SrcSet0 (Loading) fails on the target SUNOPSIS_ENGINE connection MEMORY_ENGINE.
    Caused By: java.sql.SQLException: unknown token
    +     at org.hsqldb.jdbc.Util.sqlException(Unknown Source)+
    +     at org.hsqldb.jdbc.JDBCPreparedStatement.<init>(Unknown Source)+
    +     at org.hsqldb.jdbc.JDBCConnection.prepareStatement(Unknown Source)+
    +     at sun.reflect.GeneratedMethodAccessor67.invoke(Unknown Source)+
    +     at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)+
    +     at java.lang.reflect.Method.invoke(Method.java:597)+
    +     at oracle.odi.core.datasource.dwgobject.support.OnConnectOnDisconnectDataSourceAdapter$OnDisconnectCommandExecutionHandler.invoke(OnConnectOnDisconnectDataSourceAdapter.java:200)+
    +     at $Proxy2.prepareStatement(Unknown Source)+
    +     at oracle.odi.runtime.agent.execution.sql.SQLCommand.doInitializeStatement(SQLCommand.java:83)+
    +     at oracle.odi.runtime.agent.execution.sql.SQLCommand.getStatement(SQLCommand.java:117)+
    +     at oracle.odi.runtime.agent.execution.sql.SQLCommand.getStatement(SQLCommand.java:111)+
    +     at oracle.odi.runtime.agent.execution.sql.SQLCommand.execute(SQLCommand.java:158)+
    +     at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:102)+
    +     at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:1)+
    +     at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)+
    +     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2906)+
    +     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2609)+
    +     at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:537)+
    +     at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:453)+
    +     at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1740)+
    +     at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:338)+
    +     at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:214)+
    +     at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:272)+
    +     at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:263)+
    +     at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:822)+
    +     at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:123)+
    +     at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)+
    +     at java.lang.Thread.run(Thread.java:662)+
    Caused by: org.hsqldb.HsqlException: unknown token
    +     at org.hsqldb.error.Error.error(Unknown Source)+
    +     at org.hsqldb.error.Error.error(Unknown Source)+
    +     at org.hsqldb.ParserBase.read(Unknown Source)+
    +     at org.hsqldb.ParserDQL.readNewSchemaObjectName(Unknown Source)+
    +     at org.hsqldb.ParserDDL.compileCreateTable(Unknown Source)+
    +     at org.hsqldb.ParserDDL.compileCreate(Unknown Source)+
    +     at org.hsqldb.ParserCommand.compilePart(Unknown Source)+
    +     at org.hsqldb.ParserCommand.compileStatement(Unknown Source)+
    +     at org.hsqldb.Session.compileStatement(Unknown Source)+
    +     at org.hsqldb.StatementManager.compile(Unknown Source)+
    +     at org.hsqldb.Session.execute(Unknown Source)+
    +     ... 27 more+

    Still getting the error 'Unknown Token' when using the SUNOPSIS_MEMORY_ENGINE (LKM SQL to SQL).
    I have updated the KM to 11g
    I can run successfully using an Oracle db for staging, so it is just using the SUNOPSIS_MEMORY_ENGINE that causes a problem.
    Cheers

  • ODI-1228: Task SrcSet0 (Loading) fails on the target ORACLE connection

    Hi all. I am using ODI for the first time. Trying to import a flat file data into Oracle table.
    Flat file has 4 columns Empno,Ename,Deptno,Salary [2 rows]
    so is the Target Table in Oracle [All are VARCHAR2 data type] --Empty table
    I am using these KMs: CKM Oracle , IKM SQL CONTROL APPEND, LKM FILE TO SQL.
    When I execute this Interface, it fails with the following error message:
    ODI-1228: Task SrcSet0 (Loading) fails on the target ORACLE connection [Connection_Name]
    Caused By: java.sql.SQLSyntaxErrorException: ORA-00904: : invalid identifier
    I checked the Code and Source Code is null , Target Code is: Create Table schema.Load_{table_name} ( )
    With no columns in it.
    Can anyone help? I am using Oracle ODI 11G (11.1.1)
    Thanks in advance

    Hi,
    if i remember well ( i haven't ODI opened) LKM FILE TO SQL. needs to have all mapping done into staging area. Click on a column, check Staging Area symbol (hammer), save and re-execute.
    Remember to check that your mapping is enabled (just click on a column, first checkbox in property inspector)
    If it doesn't work please make some screenshot or at least post your execution code.
    Let us know

  • While loading master data to infoobject Load failed due to Duplicate record

    Hi Experts,
    While loading master data to the infoobject load failed .
    The error it is showing is 24 Duplicate record found. 23 recordings used in table.
    Pls help me to solve this issue
    Thanks in Advance.
    Regards,
    Gopal.

    In infopackage settings u will find a checkbox for 'delete duplicate records'.
    I think it appears beside the radio button for 'To PSA',and also tick checkbox for 'subsequent update to data targets'.
    This will remove the duplicate records(if any) from the PSA before they are processed further by transfer and update rules.
    Use this and reload master data.
    cheers,
    Vishvesh

  • X120e update BIOS failed "Driver Loading Failed. The Utility cannot proceed"

    I was tried to update Bios to 1.12 but not successed..
    It always came out this message " Driver Loading Failed. The Utility cannot proceed."..
    I tried to update under safe mode, still not work..
    I tried to use Winflash, it came out "Cannot open new BIOS file"..
    There's only one method that I didn't try is that burn a ISO CD and connect to External CDROM.. but I don't have External CDROM..
    The reason that I update Bios is that I want to install Toolbox; otherwise, I won't take this risk..
    For futhermore information click following link..
    http://forums.lenovo.com/t5/ThinkVantage-Technolog​ies/This-version-of-Lenovo-ThinkVantage-Toolboox-i​...

    Sounds like something wonky with your Windows install. I'd avoid doing a BIOS update through such an install.
    W520: i7-2720QM, Q2000M at 1080/688/1376, 21GB RAM, 500GB + 750GB HDD, FHD screen
    X61T: L7500, 3GB RAM, 500GB HDD, XGA screen, Ultrabase
    Y3P: 5Y70, 8GB RAM, 256GB SSD, QHD+ screen

  • Data Load Fails due to duplicate records from the PSA

    Hi,
    I have loaded the Master Data Twice in to the PSA.  Then, I created the DTP to load the data from the PSA to the InfoProvider.  The data load is failing with an error "duplicate key/records found".
    Is there any setting that I can configure by which even though I have duplicate records in PSA, I can be successfully able to load only one set of data (without duplicates) in to the InfoProvider?
    How can I set up the process chains to do so?
    Your answer to the above two questions is appreciated.
    Thanks,

    Hi Sesh,
    There are 2 places where the DTP checks for duplicates.
    In the first, it checks previous error stacks. If the records you are loading are still contained in the error stack of a previous DTP run, it will throw the error at this stage. In this case you will first have to clean up the previous error stack.
    The second stage will clean up duplicates  across datapackeages, provided the option is set in your datasource. But you should note that this will not solve the problem if you have duplicates in the same datapackage. In that case you can do the filtering yourself in the start routine of your transformation.
    Hope this helps,
    Pieter

  • Load failed due to Overflow converting from 853156+16

    Hi All,
    Load is getting failed from one DSO to another DSO with following error message. and its AP AR load
    Runtime error while executing rule -> see long text
    Message no. RSTRAN301
    Diagnosis
    An error occurred while executing a transformation rule:
    The exact error message is:
    Overflow converting from '8.85315e+16'
    The error was triggered at the following point in the program:
    Z_BFOD_A_AR_LOOKUP_ITEM_E 278
    System Response
    Processing the data record has been terminated.
    Procedure
    The following additional information is included in the higher-level node of
    the monitor:
    Transformation ID
    Data record number of the source record
    Number and name of the rule which produced the error
    Thanks,
    Asim

    Hey,
    Ask the developer that wrote the end routine to check the code where the error occurs. First check if too much data is transferred and if it can be reduced. Then also check if the object needs to be extended 460652 - Extending key figures in BW.
    Regards,
    Michael

  • Master data load failed due to duplicate records .

    hello friends ,
    need some help .
    I am loading the master data from soruce sys , and it is throwing error of duplicate 56 records.
    I repeated the step , but found the same error once again .
    i could not find out the duplicate record , as thr are more than 24000 records , and in this 56 are duplicate . and this duplicate also looks like same.
    when i click on error records , it is showing me the below procedure .
    maintain the attribute in Psa SCREEN .
    I could not find the duplicate records , could you please let me know how can i maintain this .
    Regards

    Hi ,
    Reload the masterdata by cheking ignoreduplicate records check box.since the master data has overwriting capability the duplicate records will be overwritten
    cheers,
    Swapna.G

  • Flat file data load failed with the below error

    hi guys,
    when the scheduling load(info package) from the flat file ,got below error..
    Error 'The argument 'INR' cannot be interpreted as a number ' on assignment field /BI
    thanks

    Hi,
    It seems that the flat file and data source are not in sync.... check your flat file format and prepare it in sync with your data source.
    I guess you have an amount field followed by currency in your flat file, but in your data source currency field is not next to the amount KF or vice versa. It could also be that the KF you have defined is not of Amount type, and hence it doesn't expect the currency field.
    Hope this helps
    Godhuli

  • When I hit 84% downloaded on the free trial download it says down load failed and the learn more tab says the following;"Installation completed though some optional components failed to install correctly. (6)" Not really sure what that means but any help

    Not sure what it means but if I can still use it and someone knows how,that would be great.

    Code 6 & Code 7 http://helpx.adobe.com/creative-suite/kb/errors-exit-code-6-exit.html

  • Need help finding the row where BI load failed

    Hi all,
    We use an ETL tool to load data from our staging database into our production data warehouse environment. The environment is Oracle 10g Release 1 on Red Hat Enterprise Linux Server. The ETL tool is Business Objects Data Integrator. At hand the issue is that the data load failed due to an Oracle error
    ORA-12899: value too large for column
    When I examined the table definition in question, the load fails on one of the table columns in staging since the column is a VARCHAR2 with length of 5 and the ETL load wants to load a value for the table of length 13.
    One possible solution is to change the column definition to accomodate the data load. However, since this is a production system, I cannot do this right now.
    Is there a method to find exactly what row it failed on during the ETL load for the table in question?
    Thanks!
    B. Scott

    1) Yes, DML error logging is nice feature but it is New Feature in Oracle Database 10g Release 2.
    2) Some tools (e.g. Oracle Warehouse Builder) enable you to run your "mappping" (ETL job) in 2 modes - set based (single transaction INS/UPD/DEL/MERGE) or in row based mode (PL/SQL loop, row by row processing). OWB even enables failover mode (I failed in set-based mode, let's failover to row-based mode to find out where is the problem; the errors will be reported in log; you can set maximum number of errors before the mapping terminates with failure status).
    3) If you do not have this functionality in your tool than I suggest: Try to understand the transformation and investigate the source data. You can usually find the culprit in the source data. Sometimes you can even extract the SQL behind the transformation and run CTAS (create table as select) and find the exceptions.
    Regards,
    Pavel

Maybe you are looking for

  • Why don't these videos display in Final Cut Pro X?

    I am very new to using my iMac and to final cut pro. I can't figure out why these videos don't display video in the application. They play fine with audio in FCP and they are perfectly watchable through quicktime on the Mac. I have a screenshot, just

  • [SOLVED]Starup sript not getting executed in systemd

    Hello All, I have a script(rpi_user.sh) which has to be executed every time when I start my computer. But the startup script is not executing. Following are the steps which I have done. I am using systemd. I upgraded my system successfully today. The

  • Strange issues with iOS5 and non-working Buttons

    I upgraded my iPhone 4 to the new iOS yesterday. Since then: In the app store, under updates when I click the button to update the apps that need updating the button does not work. ALSO In Newsstand I click on the store to look at what's available. I

  • When i plug my Iphone into my Asus Laptop it gives me Blue screen of death

    When i plug my Iphone 4s into my Asus Laptop (k53s) it gives me Blue screen of death instantly :/

  • My iPod Touch 4G Lock/Sleep Button Will Not Work

    I recently had an incident with my iPod Touch 4G where the screen was shattered to the point of no use. I sent my iPod Touch off to a Non Apple repair place (good example of why you should NEVER do this) due to me not having enough money to get apple