Backup utility call failed

Hello,
I upgraded recently the Oracle database from 10.2.0.4 to 11.2.0.2. (on AIX )
Since then, the backups do not more function correctly.
As in the BRBACKUP Action Log for Database, the following errors occured:
- BR0232E 0 of 90 files saved by backup utility
- BR0231E Backup utility call failed
- Cannot reopen backint log file as stderr
So, I checked the permissions for /usr/sap/DIS/SYS/exe/run/backint, the path and permission are set properly.
I checked the also log '/nsr/applogs/backintSID.log' and I noted that the system does not write any more in this file since last successful backup (befor the Upgrade)
Many Thanks in advance for your help
NB : according to the SAP Note 948294, I deleted the options cio, dio in /etc/filesystems
Best regards
Latif

>> @Orkun
My current Br*tools version is : BR0651I BRTOOLS 7.10 (30). I upgraded an other DEV System before, it has the same BRTOOLS version as my test system, and the backup works well.
Firstly, rename backintSID.log to backintSID.log.1 and start a new backup. By doing so, the system will try to create a new trace file. If it is not able to create trace file, check the permissions of its directory(/nsr/applogs). At the end of this operation, could you provide full trace?
Secondy, even you are able to execute the brtools 7.10 on your dev system, I still strongly recommend update it to 7.20. Check the note 12741 - Current versions of BRTools and SAPDBA
Thirdly, is brbackup forked by a 3rd party backup application or manually at the OS level by orasid user?
Best regards,
Orkun Gedik
Edited by: Orkun Gedik on Aug 9, 2011 1:32 PM

Similar Messages

  • BR0231E Problem with Legato bACKUP utility

    Hi,
    I'm try to restore my database with brrestore and I'm get the following error:
    BR0280I BRRESTORE time stamp: 2008-03-02 10.12.16
    BR0229I Calling backup utility with function 'restore'...
    BR0278I Command output of 'E:\usr\sap\DOP\SYS\exe\run\backint.exe -u DOP -f restore -i E:\oracle
    \DOP\sapbackup\.rdxjtccy.lst -t file -p E:\oracle\DOP\920\database\init.off.DOP.utl':
    BR0375E Backup utility has reported an error while restoring file M:\ORACLE\DOP\SAPDATA10\BTABI_
    7\BTABI.DATA7
    BR0280I BRRESTORE time stamp: 2008-03-02 10.16.24
    BR0279E Return code from 'E:\usr\sap\DOP\SYS\exe\run\backint.exe -u DOP -f restore -i E:\oracle\
    DOP\sapbackup\.rdxjtccy.lst -t file -p E:\oracle\DOP\920\database\init.off.DOP.utl': 2
    BR0374E 0 of 1 file restored by backup utility
    BR0280I BRRESTORE time stamp: 2008-03-02 10.16.24
    BR0231E Backup utility call failed
    I' have windows 2003 + SAP 4.7 Oracle 9.0.2
    Please help
    Edited by: Luis Minaya on Mar 2, 2008 3:55 PM

    Hello,
    Legato log files might tell you more about the root cause.
    And here's just a guess: Files permissions have to be considered as well. Did you do backup and restore as the same operating system user? If users are different, then different permissions may be a problem.

  • Online Backup from Tape failed

    Hi All
    We are trying to restore the online backup from tape the backup device is Netbackup Utility
    Please find below the log
    BR0280I BRRESTORE time stamp: 2011-01-07 16.13.44
    BR0229I Calling backup utility with function 'restore'...
    BR0278I Command output of '/usr/sap/QC4/SYS/exe/run/backint -u PC4 -f restore -i /oracle/QC4/sapbackup/.reeypwkq.lst -t file -p /oracle/QC4/102_64/dbs/initQC
    4.utl':
    Program:                /usr/sap/QC4/SYS/exe/run/backint 5.0GA
    Input File:             /oracle/QC4/sapbackup/.reeypwkq.lst
    Profile:                /oracle/QC4/102_64/dbs/initQC4.utl
    Function:               RESTORE
    BR0386E File '/oracle/PC4/sapdata1/pc4_1/pc4.data1' reported as not found by backup utility
    BR0386E File '/oracle/PC4/sapdata2/pc4_103/pc4.data103' reported as not found by backup utility
    BR0386E File '/oracle/PC4/sapdata1/pc4_64/pc4.data64' reported as not found by backup utility
    BR0386E File '/oracle/PC4/sapdata2/pc4_116/pc4.data116' reported as not found by backup utility
    BR0386E File '/oracle/PC4/sapbackup/cntrlPC4.dbf' reported as not found by backup utility
    BR0280I BRRESTORE time stamp: 2011-01-07 16.13.47
    BR0279E Return code from '/usr/sap/QC4/SYS/exe/run/backint -u PC4 -f restore -i /oracle/QC4/sapbackup/.reeypwkq.lst -t file -p /oracle/QC4/102_64/dbs/initQC4
    .utl': 2
    BR0374E 0 of 161 files restored by backup utility
    BR0280I BRRESTORE time stamp: 2011-01-07 16.13.47
    BR0231E Backup utility call failed
    BR0406I End of file restore: reeypwkq.rsb 2011-01-07 16.13.47
    BR0280I BRRESTORE time stamp: 2011-01-07 16.13.47
    BR0404I BRRESTORE terminated with errors
    Please help us out in rsloving this error
    Regards
    John N

    >
    John Namala wrote:
    > BR0278I Command output of '/usr/sap/QC4/SYS/exe/run/backint -u PC4 -f restore -i /oracle/QC4/sapbackup/.reeypwkq.lst -t file -p /oracle/QC4/102_64/dbs/initQC
    Hi,
    you are restoreing PC4 database on system QC4.
    This requires a netbackup permission configuration to allow client QC4 to access the backup from system PC4.
    Volker

  • BR0229I Calling backup utility with function 'backup'...

    Hi All,
    I am having a problem with my backups. It is like the backup is hanging at step: BR0229I Calling backup utility with function 'backup'... as per the backup log.
    Please help as this is preventing me from taking any backup including redo log backup.
    OS: Windows
    DB: Oracle
    SAP: ECC5
    Below is a copy of the backup log.
    E:\oracle\PS1\origlogA\cntrl\cntrlPS1.dbf
    BR0061I 105 files found for backup, total size 372323.979 MB
    BR0143I Backup type: offline_force
    BR0130I Backup device type: util_file
    BR0109I Files will be saved by backup utility
    BR0126I Unattended mode active - no operator confirmation required
    BR0280I BRBACKUP time stamp: 2010-01-24 15.08.49
    BR0307I Shutting down database instance PS1 ...
    BR0280I BRBACKUP time stamp: 2010-01-24 15.09.25
    BR0308I Shutdown of database instance PS1 successful
    BR0280I BRBACKUP time stamp: 2010-01-24 15.09.25
    BR0229I Calling backup utility with function 'backup'...
    Regards
    Jackson

    Hi,
    If You are performing backup using third party software, then Please check init<SID>.utl file to get more information about the parameters which BACKINT uses to communicate with your Third party Backup software utility.
    Please execute the following commands and post the O/P.
    brbackup -q
    or
    brbackup -i force -v <tape name>
    Regards,
    Bhavik G. Shroff

  • Backup using backint fails for maxdb

    Hi All, I have configured backint for backup of maxdb for content server 640. I configured it as per the documents available, created the configuration fiale and the parameter file. Created the backup medium in dbmgui. Now when i try to run the backup using the pipe am getting the above mentioned error. Please find below the dbm.ebp log for the same...
    more dbm.ebp 2009-10-22 02:06:08 Setting environment variable 'TEMP' for the directory for temporary files and pi pes to default ''. Setting environment variable 'TMP' for the directory for temporary files and pip es to default ''. Using connection to Backint for MaxDB Interface. 2009-10-22 02:06:08 Checking existence and configuration of Backint for MaxDB. Using configuration variable 'BSI_ENV' = '/sapdb/CFC/lcbackup/apoatlas.env' as path of the configuration file of Backint for MaxDB. Setting environment variable 'BSI_ENV' for the path of the configuration fil e of Backint for MaxDB to configuration value '/sapdb/CFC/lcbackup/apoatlas.env' . Reading the Backint for MaxDB configuration file '/sapdb/CFC/lcbackup/apoatl as.env'. Found keyword 'BACKINT' with value '/sapdb/CFC/db/bin/backint'. Found keyword 'INPUT' with value '/tmp/backint4sapdbCFC.in'. Found keyword 'OUTPUT' with value '/tmp/backint4sapdbCFC.out'. Found keyword 'ERROROUTPUT' with value '/tmp/backint4sapdbCFC.err'. Found keyword 'PARAMETERFILE' with value '/sapdb/CFC/lcbackup/param.cfg' . Found keyword 'TIMEOUT_SUCCESS' with value '1800'. Found keyword 'TIMEOUT_FAILURE' with value '1800'. Finished reading of the Backint for MaxDB configuration file. Using '/sapdb/CFC/db/bin/backint' as Backint for MaxDB program. Using '/tmp/backint4sapdbCFC.in' as input file for Backint for MaxDB. Using '/tmp/backint4sapdbCFC.out' as output file for Backint for MaxDB. Using '/tmp/backint4sapdbCFC.err' as error output file for Backint for MaxDB . Using '/sapdb/CFC/lcbackup/param.cfg' as parameter file for Backint for MaxD B. Using '1800' seconds as timeout for Backint for MaxDB in the case of success . Using '1800' seconds as timeout for Backint for MaxDB in the case of failure . Using '/sapdb/data/wrk/CFC/dbm.knl' as backup history of a database to migra te. Using '/sapdb/data/wrk/CFC/dbm.ebf' as external backup history of a database to migrate. Checking availability of backups using backint's inquire function. Check passed successful. 2009-10-22 02:06:08 Checking medium. Check passed successfully. 2009-10-22 02:06:08 Preparing backup. Setting environment variable 'BI_CALLER' to value 'DBMSRV'. Setting environment variable 'BI_REQUEST' to value 'NEW'. Setting environment variable 'BI_BACKUP' to value 'FULL'. Constructed Backint for MaxDB call '/sapdb/CFC/db/bin/backint -u CFC -f back up -t file -p /sapdb/CFC/lcbackup/param.cfg -i /tmp/backint4sapdbCFC.in -c'. Created temporary file '/tmp/backint4sapdbCFC.out' as output for Backint for MaxDB. Created temporary file '/tmp/backint4sapdbCFC.err' as error output for Backi nt for MaxDB. Writing '/sapdb/CFC/lcbackup/pipe1 #PIPE' to the input file. Writing '/sapdb/CFC/lcbackup/pipe2 #PIPE' to the input file. Prepare passed successfully. 2009-10-22 02:06:08 Creating pipes for data transfer. Creating pipe '/sapdb/CFC/lcbackup/pipe1' ... Done. Creating pipe '/sapdb/CFC/lcbackup/pipe2' ... Done. All data transfer pipes have been created. 2009-10-22 02:06:08 Starting database action for the backup. Requesting 'SAVE DATA QUICK TO '/sapdb/CFC/lcbackup/pipe1' PIPE,'/sapdb/CFC/ lcbackup/pipe2' PIPE BLOCKSIZE 8 NO CHECKPOINT MEDIANAME 'BACKINT_ONLINE1'' from db-kernel. The database is working on the request. 2009-10-22 02:06:09 Waiting until database has prepared the backup. Asking for state of database. 2009-10-22 02:06:09 Database is still preparing the backup. Waiting 1 second ... Done. Asking for state of database. 2009-10-22 02:06:10 Database is still preparing the backup. Waiting 2 seconds ... Done. Asking for state of database. 2009-10-22 02:06:12 Database has finished preparation of the backup. The database has prepared the backup successfully. 2009-10-22 02:06:12 Starting Backint for MaxDB. Starting Backint for MaxDB process '/sapdb/CFC/db/bin/backint -u CFC -f back up -t file -p /sapdb/CFC/lcbackup/param.cfg -i /tmp/backint4sapdbCFC.in -c >>/tm p/backint4sapdbCFC.out 2>>/tmp/backint4sapdbCFC.err'. Process was started successfully. Backint for MaxDB has been started successfully. 2009-10-22 02:06:12 Waiting for end of the backup operation. 2009-10-22 02:06:12 The backup tool is running. 2009-10-22 02:06:12 The database is working on the request. 2009-10-22 02:06:14 The backup tool process has finished work with return co de 2. 2009-10-22 02:06:17 The database is working on the request. 2009-10-22 02:06:27 The database is working on the request. 2009-10-22 02:06:42 The database is working on the request. 2009-10-22 02:07:02 The database is working on the request. 2009-10-22 02:07:15 Canceling Utility-task after a timeout of 60 seconds ela psed ... OK. 2009-10-22 02:07:17 The database has finished work on the request. Receiving a reply from the database kernel. Got the following reply from db-kernel: SQL-Code :-903 The backup operation has ended. 2009-10-22 02:07:17 Filling reply buffer. Have encountered error -24920: The backup tool failed with 2 as sum of exit codes. The database request was canceled and ended with error -903. Constructed the following reply: ERR -24920,ERR_BACKUPOP: backup operation was unsuccessful The backup tool failed with 2 as sum of exit codes. The database request was canceled and ended with error -903. Reply buffer filled. 2009-10-22 02:07:17 Cleaning up. Removing data transfer pipes. Removing data transfer pipe /sapdb/CFC/lcbackup/pipe2 ... Done. Removing data transfer pipe /sapdb/CFC/lcbackup/pipe1 ... Done. Removed data transfer pipes successfully. Copying output of Backint for MaxDB to this file. -
    Begin of output of Backint for MaxDB (/tmp/backint4sapdbCFC.out)- -
    Data Protection for mySAP(R) Interface between BR*Tools and Tivoli Storage Manager - Version 5, Release 4, Modification 0.0 for Linux x86_64 - Build: 303 compiled on Nov 16 2006 (c) Copyright IBM Corporation, 1996, 2006, All Rights Reserved. BKI0008E: The environment variable BI_CALLER is not set correctely. The current value is "DBMSRV" usage: backint -p  [-u ] [-f ] [-t ] [-i ] [-o ] [-c] where:  backint utility user  backup | restore | inquire | password | delete  file | file_online  parameter file for backup utility  name of a text file that defines the objects default: STDIN  Pool for processing messages and the results of the executed function. default: STOUT BKI0020I: End of program at: Thu 22 Oct 2009 02:06:14 AM EDT . BKI0021I: Elapsed time: 01 sec . BKI0024I: Return code is: 2. -
    End of output of Backint for MaxDB (/tmp/backint4sapdbCFC.out)- - Removed Backint for MaxDB's temporary output file '/tmp/backint4sapdbCFC.out '. Copying error output of Backint for MaxDB to this file. - Begin of error output of Backint for MaxDB (/tmp/backint4sapdbCFC .err) - End of error output of Backint for MaxDB (/tmp/backint4sapdbCFC.e rr)--
    Removed Backint for MaxDB's temporary error output file '/tmp/backint4sapdbC FC.err'. Removed the Backint for MaxDB input file '/tmp/backint4sapdbCFC.in'. Have finished clean up successfully.
    Also, is there any specification about the user permissions and about how the backup should be run?

    Hi Lars,
    I understand that its a clumpsy over here, but i already have raised an OSS message and SAP said, that they cannot support this issue with backint. If you can provide me with an email id, i can send you the log files which would be easy to read.
    My issue is that am not able run backup for maxdb of content server 640 using the backint tool.
    I have created the configuration file and the parameter file as per the specifications from http://maxdb.sap.com/doc/7_7/a9/8a1ef21e4b402bb76ff75bb559a98a/content.htm and http://maxdb.sap.com/doc/7_7/50/075205962843f69b9ec41f34427be7/content.htm.
    THe server is registered to the TSM server. Now when i run the wizard to take the backup using the backint tool, it gives the error "Begin of output of Backint for MaxDB (/tmp/backint4sapdbCFC.out)- -
    Data Protection for mySAP(R) Interface between BR*Tools and Tivoli Storage Manager - Version 5, Release 4, Modification 0.0 for Linux x86_64 - Build: 303 compiled on Nov 16 2006 (c) Copyright IBM Corporation, 1996, 2006, All Rights Reserved. BKI0008E: The environment variable BI_CALLER is not set correctely. The current value is "DBMSRV" usage: backint -p  [-u ] [-f ] [-t ] [-i ] [-o ] [-c] where:  backint utility user  backup | restore | inquire | password | delete  file | file_online  parameter file for backup utility  name of a text file that defines the objects default: STDIN  Pool for processing messages and the results of the executed function. default: STOUT BKI0020I: End of program at: Thu 22 Oct 2009 02:06:14 AM EDT . BKI0021I: Elapsed time: 01 sec . BKI0024I: Return code is: 2. -
    End of output of Backint for MaxDB (/tmp/backint4sapdbCFC.out)- - Removed Backint for MaxDB's temporary output file '/tmp/backint4sapdbCFC.out '. Copying error output of Backint for MaxDB to this file. - Begin of error output of Backint for MaxDB (/tmp/backint4sapdbCFC .err) - End of error output of Backint for MaxDB (/tmp/backint4sapdbCFC.e rr)--
    Removed Backint for MaxDB's temporary error output file '/tmp/backint4sapdbC FC.err'. Removed the Backint for MaxDB input file '/tmp/backint4sapdbCFC.in'. Have finished clean up successfully."
    I think this should be fine to read...
    Krishna KK

  • Trex installation error FSL-00001  System call failed. Error 3

    I am installing trex on multiple hosts for a Production setup. Trex is version 7.0 patch level 40. All hosts are Windows servers running on "Windows Server 2003 R2 Enterprise x64 Edition, Service Pack 2" Distribution of trex instances are as follows,
    Index Server =1
    Slave Server =1
    Backup Server =1
    File Server =1
    I started of with installing Trex global filesystem which resulted in following error,
    ERROR 2008-10-23 15:19:27
    FSL-00001  System call failed. Error 3 (The system cannot find the path specified.
    ) in execution of system call 'FindFirstVolumeMountPoint' with parameter (
    ?\Volume{51f489f8-8ef6-11dd-886f-001e0beae60e}\), line (91) in file (synxcfsmit.cpp).
    ERROR 2008-10-23 15:19:27
    MUT-03025  Caught ESyException in Modulecall: The system cannot find the path specified.
    ERROR 2008-10-23 15:19:27
    FCO-00011  The step getMounts with step key |NW_TREX_Prepare|ind|ind|ind|ind|0|0|NW_GetSidNoProfiles|ind|ind|ind|ind|1|0|NW_getMounts|ind|ind|ind|ind|0|0|getMounts was executed with status ERROR .
    Installation is performed by administrator user and so has full access rights. Has anyone encountered such an error before.
    Regards,
    Sandeep

    I uninstalled the Python application via the Windows "Add/Remove" programs utility and that seems to have resolved the issue.
    TREX has been installed successfully.

  • My Fantom backup drive is failing. I want to copy the time machine contents to a new backup drive and continue on as if nothing has happened.  Can I do this and if so how?  I have intel iMac running Mountain Lion.  Fantom Drive via Firewire 800.

    My Fantom backup drive is failing and I have to send it in for warranty repair.  Before I do that I want to copy the time machine contents to a new drive and then use it for the backup, continuing on as if nothing had gone wrong with the first drive.  Can I do this?  Is it as simple as duplicating the first drive and replacing it with the second? New drive arrives Tuesday PM and both drives have dual firewire 800 ports.  

    I have used the Disk Utility to verify the contents of the backup drive so I don't think I have any corrupt files.  The drive just takes a minute or so to spin up to speed.  If I turn it off and then back on, it beeps at me for a few seconds and then starts it's slow process of trying to spin.  I think it's a power problem not a data problem.  I have my fingers crossed that it holds on until Tuesday PM when the new drive arrives.  I'm thinking that my two options are to duplicate the bad drive or partition the new drive and put the old backups on one partition and start afresh on the second partition.  Any input?
    And thanks to both of you for the valuable information so far!

  • Event ID: 5014, 5004 The DFS Replication Service is stopping communication with partner / Error 1726 (The remote procedure call failed.)

    I'm replicating between two servers in two sites (Server A - Server 2012 R2 STD, Server B - Server 2008 R2) over a VPN (Sonicwall Firewall).  Though the initial replication seems to be
    happening it is very slow (the folder in question is less than 3GB).  I'm seeing these in the event viewer every few minutes:
    The DFS Replication service is stopping communication with partner PPIFTC for replication group FTC due to an error. The service will retry the connection periodically.
    Additional Information:
    Error: 1726 (The remote procedure call failed.)
    and then....
    The DFS Replication service successfully established an inbound connection with partner PPIFTC for replication group FTC.
    Here are all my troubleshooting steps (keep in mind that our VPN is going through a SonicWall <--I increased the TCP timeout to 24 hours):
    -Increased TCP Timeout to 24 hours 
    -Added the following values on both sending and receiving members and rebooted server
    HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Services\Tcpip\Parameters
    Value =DisableTaskOffload
    Type = DWORD
    Data = 1
    Value =EnableTCPChimney
    Type = DWORD
    Data = 0
    Value =EnableTCPA
    Type = DWORD
    Data = 0
    Value =EnableRSS
    Type = DWORD
    Data = 0
    ---------------------------------more troubleshooting--------------------------
    -Disabled AntiVirus on both members
    -Made sure DFSR TCP ports 135 & 5722 are open
    -Installed all hotfixes for 2008 R2 (http://support.microsoft.com/kb/968429) and rebooted
    -Ran NETSTAT –ANOBP TCP and the DFS executable results are listed below:
    Sending Member:
    [DFSRs.exe]
      TCP    10.x.x.x:53            0.0.0.0:0             
    LISTENING       1692
    [DFSRs.exe]
      TCP    10.x.x.x:54669        
    10.x.x.x:5722          TIME_WAIT       0
      TCP    10.x.x.x:54673        
    10.x.x.x:5722          ESTABLISHED     1656
     [DFSRs.exe]
      TCP    10.x.x.x:64773        
    10.x.x.x:389           ESTABLISHED     1692
    [DFSRs.exe]
      TCP    10.x.x.x:64787        
    10.x.x.x:389           ESTABLISHED     1656
     [DFSRs.exe]
      TCP    10.x.x.x:64795        
    10.x.x.x:389           ESTABLISHED     2104
    Receiving Member:
    [DFSRs.exe]
      TCP    10.x.x.x:56683        
    10.x.x.x:389           ESTABLISHED     7472
     [DFSRs.exe]
      TCP    10.x.x.x:57625        
    10.x.x.x:54886         ESTABLISHED     2808
    [DFSRs.exe]
      TCP    10.x.x.x:61759        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61760        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61763        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61764        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61770        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61771        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61774        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61775        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61776        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61777        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61778        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61779        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61784        
    10.x.x.x:52757         ESTABLISHED     7472
    [DFSRs.exe]
      TCP    10.x.x.x:63661        
    10.x.x.x:63781         ESTABLISHED     4880
    ------------------------------more troubleshooting--------------------------
    -Increased Staging to 32GB
    -Opened the ADSIedit.msc console to verify the "Authenticated Users" is set with the default READ permission on the following object:
    a. The computer object of the DFS server
    b. The DFSR-LocalSettings object under the DFS server computer object
    -Ran
    ping <var>10.x.x.x</var> -f -l 1472 and got replies back from both servers
    -AD replication is successful on all partners
    -Nslookup is working so DNS is working
    -Updated NIC drivers on both servers
    - I ran the following to set the Primary Member:
    dfsradmin Membership Set /RGName:<replication group name> /RFName:<replicated folder name> /MemName:<primary member> /IsPrimary:True
    Then Dfsrdiag Pollad /Member:<member name>
    I'm seeing these errors in the dfsr logs:
    20141014 19:28:17.746 9116 SRTR   957 [WARN] SERVER_EstablishSession Failed to establish a replicated folder session. connId:{45C8C309-4EDD-459A-A0BB-4C5FACD97D44} csId:{7AC7917F-F96F-411B-A4D8-6BB303B3C813}
    Error:
    + [Error:9051(0x235b) UpstreamTransport::EstablishSession upstreamtransport.cpp:808 9116 C The content set is not ready]
    + [Error:9051(0x235b) OutConnection::EstablishSession outconnection.cpp:532 9116 C The content set is not ready]
    + [Error:9051(0x235b) OutConnection::EstablishSession outconnection.cpp:471 9116 C The content set is not ready]
    ---------------------------------------more troubleshooting-----------------------------
    I've done a lot of research on the Internet and most of it is pointing to the same stuff I've tried.  Does anyone have any other suggestions?  Maybe I need to look somewhere
    else on the server side or firewall side? 
    I tried replicating from a 2012 R2 server to another 2012 server and am getting the same events in the event log so maybe it's not a server issue. 
    Some other things I'm wondering:
    -Could it be the speed of the NICs?  Server A is a 2012 Server that has Hyper-V installed.  NIC teaming was initially setup and since Hyper-V is installed the NIC is a "vEthernet
    (Microsoft Network Adapter Multiplexor Driver Virtual Switch) running at a speed of 10.0Gbps whereas Server B is running a single NIC at 1.0Gbps
    -Could occasional ping timeout's cause the issue?  From time to time I get a timeout but it's not as often as the events I'm seeing.  I'm getting 53ms pings.  The folder
    is only 3 GB so it shouldn't take that long to replicate but it's been days.  The schedule I have set for replication is mostly all day except for our backup times which start at 11pm-5am.  Throughout the rest of the time I have it set anywhere from
    4Mbps to 64 Kbps.  Server A is on a 5mb circuit and Server B is on a 10mb circuit. 

    I'm seeing the same errors, all servers are running 2008 R2 x64. Across multiple sites, VPN is steady and reliably.
    185 events from 12:28:21 to 12:49:25
    Events are for all five servers (one per office, five total offices, no two in the same city, across three states).
    Events are not limited to one replication group. I have quite a few replication groups, so I don't know for sure but I'm running under the reasonable assumption that none are spared.
    Reminder from original post (and also, yes, same for me), the error is: Error: 1726 (The remote procedure call failed.)
    Some way to figure out what code triggers an Event ID 5014, and what code therein specifies an Error 1726, would extremely helpful. Trying random command line/registry changes on live servers is exceptionally unappealing.
    Side note, 1726 is referenced here:
    https://support.microsoft.com/kb/976442?wa=wsignin1.0
    But it says, "This RPC connection problem may be caused by an unstable WAN connection." I don't believe this is the case for my system.
    It also says...
    For most RPC connection problems, the DFS Replication service will try to obtain the files again without logging a warning or an error in the DFS Replication log. You can capture the network trace to determine whether the cause of the problem is at the network
    layer. To examine the TCP ports that the DFS Replication service is using on replication partners, run the following command in a
    Command Prompt window:
    NETSTAT –ANOBP TCP
    This returns all open TCP connections. The connections in question are "DFSRs.exe", which the command won't let you filter for.
    Instead, I used the NETSTAT command as advertised, dumping output to info.txt:
    NETSTAT -ANOBP TCP >> X:\info.txt
    Then I opened Excel and manually opened the .TXT for the open wizard. I chose fixed-width fields based on the first row for each result, and then added a column:
    =IF(A3="Can not", "Can not obtain ownership information", IF(LEFT(A3,1) = "[", A3&B3&C3, ""))
    Dragging this down through the entire file let me see that row (Row F) as the file name. Some anomalies were present but none impacted DFSrs.exe results.
    Finally, you can sort/filter (I sorted because I like being able to see everything, should I choose to) to get just the results you need, with the partial rows removed from the result set, or bumped to the end.
    My server had 125 connections open.
    That is a staggering number of connections to review, and I feel like I'm looking for a needle in a haystack.
    I'll see if I can find anything useful out, but a better solution would be most wonderful.

  • Call fails after last update

    Can anyone help me, after the last update from Apple to my iphone 3g I get call fail after call fail. It is getting embarassing at work where I rolled these phone out.

    Standard troubleshooting is
    1)Reset: reset your iPhone by pressing the home and sleep buttons until you see the Apple logo, ignoring the slider. Takes about 5-15 secs of button holding and you won't lose any data or settings.
    2)Restore as new or from backup: see page 141 of the manual for details...http://manuals.info.apple.com/enUS/iPhone_UserGuide.pdf
    Note the last sentence of the Restore paragraph... "In iTunes, you can also restore from a backup without deleting data stored on iPhone."
    Message was edited by: paulcb

  • XI AF API call failed. Module exception: Java Mapping

    Hi Experts,
    I have faced a typical error for one of my interface. The Scenario is SOAP to SOAP and we are performing SOAP Lookup  before sending the data to Target web service. We are using one Java Mapping  'FileIns_lookupLoad.java --- com/fi/' and performing the SOAP Lookup from within the code. In the Lookup Communication channel we are using 'Axis' as message protocol and passing couple of values through Module Key parameters. We have same ESR and ID objects in Development, Quality and Production (as per Version ID and History).
    Now the problem is Lookup is failing only in Development but Quality and Production is working fine. I have replaced Quality CC URL with the development WSDL URL, but that worked fine. That means Development Web Service is good. We are getting the below error in Dev:
    <SAP:Stack>StreamTransformationException triggered by application mapping program com/fi/FileIns_lookupLoad; Look Up Failed</SAP:Stack>
    Trace : *
    <Trace level="1"
    type="T"> Some Thing Wrong in LookUpError when calling an adapter by using the communication channel CC_IN_SOAP_GEInsuranceLoadLookup1 (Party: , Service: BusService_GE, Object ID: 4214805c52893ef9b0b3f0ef0902fe9e) XI AF API call failed. Module exception: 'while trying to invoke the method org.apache.axis.types.URI.toString() of an object returned from com.sap.xi.XI.Message._30.QualifiedName.getNamespace()'. Cause Exception: 'while trying to invoke the method org.apache.axis.types.URI.toString() of an object returned from com.sap.xi.XI.Message._30.QualifiedName.getNamespace()'. </Trace>
      <Trace level="1" type="T">*** END APPLICATION TRACE ***</Trace>
      <Trace level="1" type="T">Java mapping com/fi/FileIns_lookupLoad has thrown a StreamTransformationException. Thrown: com.sap.aii.mapping.api.StreamTransformationException: Look Up Failed at com.fi.FileIns_lookupLoad.execute(FileIns_lookupLoad.java:282) at com.fi.FileIns_lookupLoad.transform(FileIns_lookupLoad.java:74) at com.sap.aii.ib.server.mapping.execution.JavaMapping.executeStep(JavaMapping.java:92) at com.sap.aii.ib.server.mapping.execution.Mapping.execute(Mapping.java:60) at com.sap.aii.ib.server.mapping.execution.SequenceMapping.executeStep.................................................
    ................................................................................<Trace level="1" type="T">Application mapping program com/fi/FileIns_lookupLoad throws a stream transformation exception: Look Up Failed Thrown: com.sap.aii.ib.core.mapping.execution.ApplicationException: Application mapping program com/fi/FileIns_lookupLoad throws a stream transformation exception: Look Up Failed at com.sap.aii.ib.server.mapping.execution.JavaMapping.executeStep(JavaMapping.java:95) at com.sap.aii.ib.server.mapping.execution.Mapping.execute(Mapping.java:60) at com.sap.aii.ib.server.mapping.execution.SequenceMapping.executeStep(SequenceMapping.java:40) at com.sap.aii.ib.server.mapping.execution.Mapping.execute
    I have already checked with Basis and as per them JDK and Java versions are identical in Dev, QAS and Prod. It seems something is wrong with 'XI AF API' which we are calling from the channel. If we totally ignore Lookup process and send direct data, it is working fine. Target CC is also using Axis.
    Sequence in CC: afreq ->xireq ->wssec2 ->xires ->afres
    For any more information please let me know.
    Thanks,
    Nabendu.

    Hi Anupam,
    The Java Mapping code is same in Dev , QAS and Prod. Also the versions of the JAR is same.
    Please find the code below.
    package com.fi;
    import java.util.HashMap;
    import com.sap.aii.mapping.api.AbstractTrace;
    import com.sap.aii.mapping.api.AbstractTransformation;
    import com.sap.aii.mapping.api.StreamTransformation;
    import com.sap.aii.mapping.api.StreamTransformationConstants;
    import com.sap.aii.mapping.api.MappingTrace;
    import com.sap.aii.mapping.api.StreamTransformationException;
    import com.sap.aii.mapping.api.DynamicConfiguration;
    import com.sap.aii.mapping.api.DynamicConfigurationKey;
    import com.sap.aii.mapping.api.TransformationInput;
    import com.sap.aii.mapping.api.TransformationOutput;
    import java.io.File;
    import java.io.FileInputStream;
    import java.io.FileOutputStream;
    import java.io.IOException;
    import java.io.InputStream;
    import java.io.OutputStream;
    import java.util.*;
    import javax.xml.parsers.DocumentBuilder;
    import javax.xml.parsers.DocumentBuilderFactory;
    import javax.xml.parsers.ParserConfigurationException;
    import javax.xml.transform.Result;
    import javax.xml.transform.Source;
    import javax.xml.transform.Transformer;
    import javax.xml.transform.TransformerConfigurationException;
    import javax.xml.transform.TransformerException;
    import javax.xml.transform.TransformerFactory;
    import javax.xml.transform.dom.DOMSource;
    import javax.xml.transform.OutputKeys;
    import javax.xml.transform.stream.StreamResult;
    import org.w3c.dom.*;
    import org.w3c.dom.views.AbstractView;
    import org.xml.sax.SAXException;
    import java.io.ByteArrayInputStream;
    import com.sap.aii.mapping.lookup.*;
    /*import com.sap.aii.utilxi.hmis.server.HmisEnvironment.Accessor;
    //import com.sap.aii.utilxi.lock.api.LockServiceException;
    //import com.pmintl.pppimes.RFCLookup.javamapping.*;
    //import java.io.FileInputStream;
    //import java.io.FileOutputStream;*/
    import java.text.DateFormat;
    import java.text.SimpleDateFormat;
    //import java.text.ParseException;
    import java.util.Date;
    import java.io.*;
    public class FileIns_lookupLoad extends AbstractTransformation{
            private Map param = null;
            private AbstractTrace trace = getTrace();
            String senderService = null;
           String inParamChannel = null;
           String inParamBusService = null;
              //3. Each JAVA Mapping using program 7.1 API must implement the method
              // transform(TransformationInput in, TransformationOutput out)
              // as oppose to execute Method in earlier version.
               public void transform(TransformationInput arg0, TransformationOutput arg1) throws StreamTransformationException {
              // TODO Auto-generated method stub
              //4. An info message is added to trace. An instance of trace of object is obtained by calling
              // the getTrace method of class AbstractTransformation
                    inParamChannel = arg0.getInputParameters().getString("COMM_CHANNEL");
                    inParamBusService = arg0.getInputParameters().getString("BUS_SERVICE");
                    getTrace().addInfo("Input Parameter: " + inParamChannel);
                    getTrace().addInfo("Input Parameter: " + inParamBusService);
                    this.execute(arg0.getInputPayload().getInputStream(),
                          arg1.getOutputPayload().getOutputStream());
              //5. Input payload is obtained by using arg0.getInputPayload().getInputStream()
          * @param args
         public static void main(String[] args) {
                // TODO Auto-generated method stub
         public void execute(InputStream in, OutputStream out)
                   throws StreamTransformationException {
              // TODO Auto-generated method stub
              try
                   //Get The Trace
    //               trace = (AbstractTrace)param.get(StreamTransformationConstants.MAPPING_TRACE );
                   trace = getTrace();
                  getTrace().addInfo("Java Mapping Started");
                   DocumentBuilderFactory ifactory = DocumentBuilderFactory.newInstance();
                   DocumentBuilder ibuilder = ifactory.newDocumentBuilder();
                   Document IDoc  = ibuilder.parse(in);
                   Document TDoc = ibuilder.newDocument();
                   String mrnVal = "";
                   trace.addInfo("Preparing Target Doc");
                   Element message = TDoc.createElementNS("urn:Medtronic.com:CATS_Patient_Transactions","n1:Mt_Insurance");
                   Node tRoot = TDoc.appendChild(message);
                   //trace.addInfo(tRoot.getNodeName());
                   Element sRoot = (Element)IDoc.getFirstChild();
                   //trace.addInfo(sRoot.getNodeName());
                   NodeList nl = sRoot.getChildNodes();
               // trace.addInfo("Len"+nl.getLength());
    ////               for(int i=0;i<nl.getLength();i++)
    //                    Node n1 = TDoc.importNode(nl.item(i), true);
    //                    trace.addInfo("Adding Child Nodes");
    //                    trace.addInfo(nl.item(i).getNodeName());
    //                    tRoot.appendChild(n1);
                   NodeList lmrn = sRoot.getElementsByTagName("Mrn");
                   if(lmrn != null)
                        Node n1 = TDoc.importNode(lmrn.item(0), true);
                        Element e1 = (Element)lmrn.item(0);
                        //trace.addInfo(e1.getNodeName());
                        mrnVal = e1.getTextContent();
                        tRoot.appendChild(n1);
                   NodeList lfsc = sRoot.getElementsByTagName("Fsc");
                   if(lfsc != null)
                        for(int i=0;i<lfsc.getLength();i++){
                             Element n1=TDoc.createElement(lfsc.item(i).getNodeName());
                             tRoot.appendChild(n1);
                             Node fieldch = TDoc.createElement("Fields");
                             n1.appendChild(fieldch);
                             NodeList chl = lfsc.item(i).getChildNodes();
                             trace.addInfo("len11"+chl.getLength());
                             for(int j=0;j<chl.getLength();j++){
                                  if(!chl.item(j).getNodeName().equals("FscNumber")&&!chl.item(j).getNodeName().equals("Priority")){
                                       if(chl.item(j).hasChildNodes()){
                                            trace.addInfo(chl.item(j).getNodeName());
                                       Element fscfield = TDoc.createElement("FSCField");
                                       fieldch.appendChild(fscfield);
                                       NodeList FUQl = chl.item(j).getChildNodes();
                                       for(int p =0;p<FUQl.getLength();p++){
                                            //trace.addInfo(" i: "+i+" J: "+j+" P: "+p);
                                            if(FUQl.item(p).getNodeName().equals("FUQNumber")){
                                                 //Node tFUQ = TDoc.createElement("FUQNumber");
                                                 //tFUQ.setTextContent(FUQl.item(p).getTextContent());
                                                 fscfield.setAttribute("FUQNumber", FUQl.item(p).getTextContent());
                                            if(FUQl.item(p).getNodeName().equals("Value")){
                                                 Node tVal = TDoc.createElement("Value");
                                                 tVal.setTextContent(FUQl.item(p).getTextContent());
                                                 fscfield.appendChild(tVal);
                                                 //Node tName = TDoc.createElement("Field");
                                                 //tName.setTextContent(chl.item(j).getNodeName());
                                                 fscfield.setAttribute("Field", chl.item(j).getNodeName());
    //                                   Node fuqch = chl.item(j).getFirstChild();
    //                                   Node valch = fuqch.getNextSibling();
    //                                   if(fuqch!=null){
    //                                        Node tFUQ = TDoc.createElement("FUQNumber");
    //                                        tFUQ.setTextContent(fuqch.getTextContent());
    //                                        fscfield.appendChild(tFUQ);
    //                                   if(valch!=null){
    //                                        Node tVal = TDoc.createElement("Value");
    //                                        tVal.setTextContent(valch.getTextContent());
    //                                        fscfield.appendChild(tVal);
    //                                        Node tName = TDoc.createElement("Filed");
    //                                        tName.setTextContent(chl.item(j).getNodeName());
    //                                        fscfield.appendChild(tName);
                                  else{
                                       //Node numch =TDoc.importNode(chl.item(j), true);
                                       //Element e1 =(Element)n1;
                                       if(chl.item(j).getNodeName().equals("FscNumber")){
                                            n1.setAttribute("Number", chl.item(j).getTextContent());
                                       if(chl.item(j).getNodeName().equals("Priority")){
                                            n1.setAttribute("Priority", chl.item(j).getTextContent());
                             //tRoot.appendChild(lfsc.item(0));
    //                         Element e1 = (Element)lfsc.item(0);
    //                         trace.addInfo(e1.getNodeName());
    //                         mrnVal = e1.getNodeValue();
    /////Start of Look Up Code////
                   //Preparing Input String
                   String lookUpRequest = "<LoadPatientFullInsuranceDetail><Mrn>"+mrnVal+"</Mrn></LoadPatientFullInsuranceDetail>";
                   trace.addInfo("Request");
                   trace.addInfo(lookUpRequest);
                        SystemAccessor acc;
                        Channel channel;
                        Payload lookupResult;     
                        try
                             channel = LookupService.getChannel(inParamBusService, inParamChannel);
                             trace.addInfo("Got the channel");
                             acc = LookupService.getSystemAccessor(channel);
                             trace.addInfo("Got The Channel and Accessor");
                             InputStream ist = new ByteArrayInputStream(lookUpRequest.getBytes());
                             XmlPayload payload = LookupService.getXmlPayload(ist);
                             if(acc != null){
                                  trace.addInfo("Got The Channel and Accessor");
                                  trace.addInfo("Executing The Webservice");
                                  lookupResult = acc.call(payload);
                                  trace.addInfo("End of Executing The Webservice");
                                  trace.addInfo(lookupResult.toString());
                             if(lookUpRequest!=null){
                                  trace.addInfo("Responce is not null");
                                  InputStream rist = lookupResult.getContent();
                                  Document lresDoc = ibuilder.parse(rist);
                                  Node n1 = lresDoc.getFirstChild();
                                  trace.addInfo(n1.getNodeName());
                                  if(n1.getNodeName().endsWith("LoadPatientFullInsuranceDetailResponse"))
                                       Node ch1 = n1.getFirstChild();
                                       trace.addInfo(ch1.getNodeName());
                                       NodeList lookupnl = ch1.getChildNodes();
                                       //trace.add
                                       for(int i=0;i<lookupnl.getLength();i++)
                                            Node n2 = TDoc.importNode(lookupnl.item(i), true);
                                            trace.addInfo("Adding Child Nodes");
                                            trace.addInfo(lookupnl.item(i).getNodeName());
                                            tRoot.appendChild(n2);
                                  else{
                                  throw new StreamTransformationException("Error in Look Up"+n1.getTextContent());
                                   //while(ch1.hasChildNodes() && !ch1.getNodeName().equals("FSCS")){
                             else{
                                  trace.addWarning("Responce is null");
                                 throw new LookupException();
                        catch(LookupException le)
                             trace.addWarning("Some Thing Wrong in LookUp"+le.getMessage());
                             throw new StreamTransformationException("Look Up Failed");                         
                           Transformer transformer = TransformerFactory.newInstance().newTransformer();
    //                       StreamResult reqResult = new StreamResult(new StringWriter());
                           DOMSource source = new DOMSource(TDoc);
                           Result result = new StreamResult(out);
                           transformer.transform(source,result);
              catch(StreamTransformationException ste)
               throw new StreamTransformationException(ste.getMessage());
              catch(Exception e){
                   trace.addInfo(e.getMessage());
              finally{

  • Endeca Utility 'backup_log_dir_for_component_AuthoringDgraph' failed

    I am getting error when i am executing baseline update script on windows. The forge is happening fine after which the script is throwing error.
    [06.25.14 21:04:38] INFO: Checking definition from AppConfig.xml against existing EAC provisioning.
    [06.25.14 21:04:39] INFO: Definition has not changed.
    [06.25.14 21:04:39] INFO: Starting baseline update script.
    [06.25.14 21:04:39] INFO: Acquired lock 'update_lock'.
    [06.25.14 21:04:39] INFO: [ITLHost] Starting shell utility 'cleanDir_processing'.
    [06.25.14 21:04:40] INFO: [ITLHost] Starting shell utility 'cleanDir_forge-output'.
    [06.25.14 21:04:41] INFO: [ITLHost] Starting shell utility 'cleanDir_dgidx-output'.
    [06.25.14 21:04:42] INFO: [ITLHost] Starting shell utility 'move_-_to_processing'.
    [06.25.14 21:04:44] INFO: [ITLHost] Starting copy utility 'fetch_config_to_input_for_forge_Forge'.
    [06.25.14 21:04:45] INFO: [ITLHost] Starting backup utility 'backup_log_dir_for_component_Forge'.
    [06.25.14 21:04:46] INFO: [ITLHost] Starting component 'Forge'.
    [06.25.14 21:04:55] INFO: [ITLHost] Starting backup utility 'backup_log_dir_for_component_Dgidx'.
    [06.25.14 21:04:56] INFO: [ITLHost] Starting component 'Dgidx'.
    [06.25.14 21:05:11] INFO: [AuthoringMDEXHost] Starting shell utility 'cleanDir_local-dgraph-input'.
    [06.25.14 21:05:12] INFO: [AuthoringMDEXHost] Starting shell utility 'rmdir_dgraph-input-old'.
    [06.25.14 21:05:13] INFO: [AuthoringMDEXHost] Starting copy utility 'copy_index_to_host_AuthoringMDEXHost_AuthoringDgraph'.
    [06.25.14 21:05:14] INFO: Applying index to dgraphs in restart group 'A'.
    [06.25.14 21:05:14] INFO: [AuthoringMDEXHost] Starting shell utility 'mkpath_dgraph-input-new'.
    [06.25.14 21:05:16] INFO: [AuthoringMDEXHost] Starting copy utility 'copy_index_to_temp_new_dgraph_input_dir_for_AuthoringDgraph'.
    [06.25.14 21:05:17] INFO: [AuthoringMDEXHost] Starting shell utility 'move_dgraph-input_to_dgraph-input-old'.
    [06.25.14 21:05:18] INFO: [AuthoringMDEXHost] Starting shell utility 'move_dgraph-input-new_to_dgraph-input'.
    [06.25.14 21:05:19] INFO: [AuthoringMDEXHost] Starting backup utility 'backup_log_dir_for_component_AuthoringDgraph'.
    [06.25.14 21:05:20] SEVERE: Utility 'backup_log_dir_for_component_AuthoringDgraph' failed.
    Occurred while executing line 5 of valid BeanShell script:
    2|
    3|    AuthoringDgraphCluster.cleanDirs();
    4|    AuthoringDgraphCluster.copyIndexToDgraphServers();
    5|    AuthoringDgraphCluster.applyIndex();
    6|
    7|    LiveDgraphCluster.cleanDirs();
    8|    LiveDgraphCluster.copyIndexToDgraphServers();
    [06.25.14 21:05:20] SEVERE: Error executing valid BeanShell script.Occurred while executing line 28 of valid BeanShell script:
    25|        Dgidx.run();
    26|
    27|        // distributed index, update Dgraphs
    28|        DistributeIndexAndApply.run();
    29|
    30|        WorkbenchManager.cleanDirs();
    31|        Forge.getPostForgeDimensions();
    [06.25.14 21:05:20] SEVERE: Caught an exception while invoking method 'run' on object 'BaselineUpdate'. Releasing locks.
    Caused by java.lang.reflect.InvocationTargetException
    sun.reflect.NativeMethodAccessorImpl invoke0 - null
    Caused by com.endeca.soleng.eac.toolkit.exception.AppControlException
    com.endeca.soleng.eac.toolkit.script.Script runBeanShellScript - Error executing
    valid BeanShell script.
    Caused by com.endeca.soleng.eac.toolkit.exception.AppControlException
    com.endeca.soleng.eac.toolkit.script.Script runBeanShellScript - Error executing
    valid BeanShell script.
    Caused by com.endeca.soleng.eac.toolkit.exception.EacComponentControlException
    com.endeca.soleng.eac.toolkit.utility.Utility runInParallel - Utility 'backup_lo
    g_dir_for_component_AuthoringDgraph' failed.
    [06.25.14 21:05:20] INFO: Released lock 'update_lock'.
    In the AuthoringDgraph log it is showing below message.
    FATAL
    06/25/14 14:39:40.954 UTC (1403707180954)
    DGRAPH
    {dgraph,baseline}
    Rapid updates are not supported for datasets that do not have a record specification model specified in the record spec file.
    Stemming should be enabled for 1 languages

    Hello,
    I believe there are two different issues that may be occurring here, so we can start with the easiest one first, the error message in AuthoringDgraph.  The error message is being triggered as it can't find a property within your records that has been flagged as a Record Specifier (aka Record Spec or RecSpec).  In the Endeca world, this is similar to what a Primary Key on a table would accomplish in a relational database.  These Record Specs are used in many ways for record retrieval, but it's not neccessary to have one put in place.
    If you don't specify a property to be a Record Spec, than Endeca will create one for you.  However - and this is where the error message comes in - you cannot perform updates to the data, known as a Partial Update.  Since Endeca hasn't been told what property will uniquely represent a specific record, it can't receive updates to update/delete them.  That's where the "Rapid Updates" comes into play.
    There are two choices for this, firstly specify a property as a Record Spec, either through Developer Studio, or through the IOC/CSV overrides if you're using the product catalog integration method (what ATG uses to send data to Endeca).  The second option is to disable Rapid Update (partial updates) functionality in your index - you'll need to comment out references to "partial" directories in the various Dgraph XML configurations in your config/scripts directory.  However, that does go against Endeca Best Practices which are to always have a Record Spec.
    And now onto the second issue, the error with the "backup_log_dir_for_component_AuthoringDgraph" script.  This isn't an uncommon type of error, and may be unrelated to the issue above.  The usual culprits for this are permission errors, inability to obtain a lock, etc.  These scripts are run by the EAC and not the application, so are stored in a different location.
    Firstly, find the installation directory for Platform Services on your system.  Inside it should be a logging directory, and you're looking for workspace/logs/utility/<appName>.backup_log_dir_for_component_AuthoringDgraph.txt inside there.  Open up that file, and it should help to clarify what the issue is - if not, then please post it's contents here and we can delve deeper.
    Best,
    - Jeff

  • IPhone 5 iOS 7.1.4 Call failed every time when the other person hung up before me

    iPhone 5 iOS 7.0.4 says Call failed every time when the other person hung up before me,
    Is there anything I can do to about that please?

    Hey there Yonidawit,
    It sounds like you are are getting a message saying that the call failed, when the person you are talking to hangs up. I have a few things to recommend here. First close all the open apps:
    iOS: Force an app to close
    http://support.apple.com/kb/ht5137
    Double-click the Home button.
    Swipe left or right until you have located the app you wish to close.
    Swipe the app up to close it.
    When you have done that restart the phone and test it out again:
    iOS: Turning off and on (restarting) and resetting
    http://support.apple.com/kb/ht1430
    If that does not resolve it, try resetting your network settings:
    Reset network settings by tapping Settings > General > Reset > Reset Network Settings. Note: This will reset all network settings including:
    previously connected Wi-Fi networks and passwords
    recently used Bluetooth accessories
    VPN and APN settings
    From: iOS: Troubleshooting Wi-Fi networks and connections
              http://support.apple.com/kb/ts1398
    If the issue persists, I would next backup your device to iTunes, and restore it:
    iOS: How to back up and restore your content
    http://support.apple.com/kb/HT1766
    Thank you for using Apple Support Communities.
    Regards,
    Sterling

  • BCP to pass an Error Message to SSIS when the BCP call fails?

    Hi,
    Within SSIS I have an Execute SQL Task which calls BCP via a source variable given the dynamic nature of the BCP call.
    When the BCP call fails it returns a number of records which give instructions on how to use BCP. SSIS then thinks that BCP has executed successfully, the component shows green and then the package continues to run.
    But of course what I want the BCP call to do is return an error message which would then trigger the standard on error event handler within SSIS. How do I do this please?
    It maybe the “Execute Process Task” could be a better SSIS component to use for this call
    Using BCP Utility in SSIS. Does anyone have experience of doing this type of thing?
    Thanks in advance,
    Kieran.
    Kieran Patrick Wood http://www.innovativebusinessintelligence.com http://uk.linkedin.com/in/kieranpatrickwood http://kieranwood.wordpress.com/

    Why don't you use the "Fast Load" option = BULK LOAD? Or the BULK LOAD stored proc?
    The only other way I can think of this can be done is by directing errors to an error file which can be interrogated from the package and then a precedence constraint used to trigger an error if there was one.
    Arthur My Blog

  • My iphone 4 says call failed and message failed?

    It was locked to orange but just had it unlocked to work on my T-mobile sim. Had this issue even before I got it unlocked. I have just bought this off of somebody and it keeps saying call failed/call ended just after I have dialled and then when I try to send a message it says message failed, however, I can get onto the internet using my signal alone and not connecting to Wi-fi. Incredibly confused as to why this isnt working. I have full signal. Done a full restore on 2 different computers still with no result. Really at the end of my tether with this, someone please help.
    Operating system 4.3.5
    Thanks

    I faced similar issue where I was getting signal, able to receive calls, use all data services but everytime i went to make a call it said "Call Failed"
    You need to try to go to Settings>>Phone>>>Show My Caller ID>> Ensure it is On.
    Even if it appears on try switching off and on.
    Mine was not allowing to switch off and on but something I did worked and then able to make calls.
    Also try Settings>>>General>>>Reset>>Rest All Settings
    I faced issue of not being able to make calls after switching my sim card from postpaid to prepaid. First it dint recognise the sim at all then on pluggin in to itunes i was able to activate and atleast receive calls, use all data only everytime i tried makinng a call it said call failed.
    When i restored to factory settings it worked. However on restoring my backup i faced the same issue.
    Apple local support was most unhelpful saying that i had to forget my data and work with a new phone without contacts, apps, etc

  • SEVERE: Utility 'cleanDir_processing' failed. Refer to utility logs in [ENDECA_CONF]/logs/shell on host ITLHost. during baseline update

    Hi Guys,
                 Suddenly my application throwing below error at "baseline" script run time.
    Please provide the remedy for it,
    [01.27.14 17:51:03] INFO: Definition has not changed.
    [01.27.14 17:51:03] INFO: Starting baseline update script.
    [01.27.14 17:51:03] INFO: Acquired lock 'update_lock'.
    [01.27.14 17:51:03] INFO: [ITLHost] Starting shell utility 'cleanDir_processing'.
    [01.27.14 17:51:04] SEVERE: Utility 'cleanDir_processing' failed. Refer to utility logs in [ENDECA_CONF]/logs/shell on host ITLHost.
    Occurred while executing line 10 of valid BeanShell script:
    7|      if (Forge.isDataReady()) {
    8|
    9|        // clean directories
    10|        Forge.cleanDirs();
    11|        PartialForge.cleanCumulativePartials();
    12|        Dgidx.cleanDirs();
    13|
    [01.27.14 17:51:04] SEVERE: Caught an exception while invoking method 'run' on object 'BaselineUpdate'. Releasing locks.
    Caused by java.lang.reflect.InvocationTargetException
    sun.reflect.NativeMethodAccessorImpl invoke0 - null
    Caused by com.endeca.soleng.eac.toolkit.exception.AppControlException
    com.endeca.soleng.eac.toolkit.script.Script runBeanShellScript - Error executing valid BeanShell script.
    Caused by com.endeca.soleng.eac.toolkit.exception.EacComponentControlException
    com.endeca.soleng.eac.toolkit.utility.Utility run - Utility 'cleanDir_processing' failed. Refer to utility logs in [ENDECA_CONF]/logs/shell on host ITLHost.
    [01.27.14 17:51:04] INFO: Released lock 'update_lock'.
    C:\Endeca\apps\Restaurants\control>
    I ran the Release_lock  then also got the same error. I have seen below error in the log file "[ENDECA_CONF]/logs/shell/ Restaurants.cleanDir_processing",
    Failed to remove directory C:\Endeca\apps\Restaurants\config\script\..\..\.\data\processing: Directory not empty at -e line 1.
    Thanks,
    swapnil

    The above issue is still there , i think this is due to some configuration in case of the pipeline. I am unable to detect it.
    Now in the below scenario My baseline index is running proper but partial update is failing with below error.
    C:\Endeca\apps\TestApp\control>partial_update.bat
    [01.30.14 17:06:34] INFO: Checking definition from AppConfig.xml against existing EAC provisioning.
    [01.30.14 17:06:35] INFO: Updating provisioning for component 'DailyReportGenerator'.
    [01.30.14 17:06:35] INFO: Updating definition for component 'DailyReportGenerator'.
    [01.30.14 17:06:35] INFO: Definition updated.
    [01.30.14 17:06:35] INFO: Starting partial update script.
    [01.30.14 17:06:35] INFO: Acquired lock 'update_lock'.
    [01.30.14 17:06:35] INFO: [ITLHost] Starting backup utility 'backup_log_dir_for_component_PartialForge'.
    [01.30.14 17:06:36] INFO: [ITLHost] Starting shell utility 'cleanDir_processing'.
    [01.30.14 17:06:38] SEVERE: Utility 'cleanDir_processing' failed. Refer to utility logs in [ENDECA_CONF]/logs/shell on host ITLHost.
    Occurred while executing line 14 of valid BeanShell script:
    11|        PartialForge.archiveLogDir();
    12|
    13|        // clean directories
    14|        PartialForge.cleanDirs();
    15|
    16|        // fetch extracted data files to forge input
    17|        PartialForge.getPartialIncomingData();
    [01.30.14 17:06:38] SEVERE: Caught an exception while invoking method 'run' on object 'PartialUpdate'. Releasing locks.
    Caused by java.lang.reflect.InvocationTargetException
    sun.reflect.NativeMethodAccessorImpl invoke0 - null
    Caused by com.endeca.soleng.eac.toolkit.exception.AppControlException
    com.endeca.soleng.eac.toolkit.script.Script runBeanShellScript - Error executing valid BeanShell script.
    Caused by com.endeca.soleng.eac.toolkit.exception.EacComponentControlException
    com.endeca.soleng.eac.toolkit.utility.Utility run - Utility 'cleanDir_processing' failed. Refer to utility logs in [ENDECA_CONF]/logs/shell on host ITLHost.
    [01.30.14 17:06:38] INFO: Released lock 'update_lock'.
    Plz recommend some remedies.
    Thanks,
    swapnil

Maybe you are looking for

  • ErrorAccount 'Acc.dep. accnt.for ordinary depreciation' could not be found.

    Hi,   When I am trying to retire an asset by scrapping using TCODE ABAVN I am getting the following error.Ho do I resolve this. Thanks. Account 'Acc.dep. accnt.for ordinary depreciation' could not be found. Message no. AU133 Diagnosis When creating t

  • Attachment List Table In sales order VA02.

    Sir, I have to develop report on attachment list in sales order on the basis of sales order number and customer no. But I'm not getting the tables. So I can develope the report. Thanks Manoj Pise

  • LR3 to LR4 upgrade from PC to mac

    i have LR3 installed in my windows based PC, i just got an imac and want to install LR4, can i just purchase an "upgrade" instead of buying the "full" version?

  • Put imac to Sleep or Shut Down

    What's up guys I'm just wondering, I use Logic pro a lot and I would like have quick access when I have a creative urge. How often should I shut down my imac as apposed to putting it to sleep?

  • A simple JSF Table CRUD - How To

    Hello, I have to build a few CRUDs for very simple tables (3 columns at most) implemented as JSF tables. I wish the "Create" function to work as: 1. By clicking the "Create" button (or link?), append an empty row at the end of the table (or perhaps i